cve_id
stringclasses
265 values
cve_published
timestamp[ns]
cve_descriptions
stringclasses
265 values
cve_metrics
dict
cve_references
listlengths
1
27
cve_configurations
listlengths
1
3
url
stringclasses
266 values
cve_tags
listlengths
1
4
domain
stringclasses
1 value
issue_owner_repo
listlengths
2
2
issue_body
stringlengths
2
8.04k
issue_title
stringlengths
3
346
issue_comments_url
stringlengths
59
81
issue_comments_count
int64
0
66
issue_created_at
timestamp[ns]
issue_updated_at
stringlengths
20
20
issue_html_url
stringlengths
40
62
issue_github_id
int64
1.01B
2.39B
issue_number
int64
21
125k
label
bool
2 classes
issue_msg
stringlengths
120
8.18k
issue_msg_n_tokens
int64
120
8.18k
issue_embedding
listlengths
3.07k
3.07k
null
null
null
null
null
null
null
null
null
[ "chakra-core", "ChakraCore" ]
ChakraCore version: commit c3ead3f Build cmd: ```shell ./build.sh --debug --static ``` Test case: ```js function foo() { for (let i = 0; i < 16; i++) { try { const s = ~"__proto__"; switch (s) { case s: try { break; } finally { } break; default: break; } } catch(e) { ; } } } let i = 0; while(i < 1024) { foo(); i++; } ``` Execute ```shell ./ch ./test.js ``` Error msg: ``` ASSERTION 453147: (ChakraCore/lib/Backend/FlowGraph.cpp, line 2078) this->func->IsLoopBodyInTry() Failure: (this->func->IsLoopBodyInTry()) Illegal instruction (core dumped) ```
ASSERTION 453147: (ChakraCore/lib/Backend/FlowGraph.cpp, line 2078) this->func->IsLoopBodyInTry()
https://api.github.com/repos/chakra-core/ChakraCore/issues/6949/comments
0
2023-09-13T09:35:01
2023-09-13T09:35:01Z
https://github.com/chakra-core/ChakraCore/issues/6949
1,894,135,788
6,949
false
This is a GitHub Issue repo:ChakraCore owner:chakra-core Title : ASSERTION 453147: (ChakraCore/lib/Backend/FlowGraph.cpp, line 2078) this->func->IsLoopBodyInTry() Issue date: --- start body --- ChakraCore version: commit c3ead3f Build cmd: ```shell ./build.sh --debug --static ``` Test case: ```js function foo() { for (let i = 0; i < 16; i++) { try { const s = ~"__proto__"; switch (s) { case s: try { break; } finally { } break; default: break; } } catch(e) { ; } } } let i = 0; while(i < 1024) { foo(); i++; } ``` Execute ```shell ./ch ./test.js ``` Error msg: ``` ASSERTION 453147: (ChakraCore/lib/Backend/FlowGraph.cpp, line 2078) this->func->IsLoopBodyInTry() Failure: (this->func->IsLoopBodyInTry()) Illegal instruction (core dumped) ``` --- end body ---
981
[ -0.024897228926420212, 0.01970154605805874, -0.011607694439589977, 0.00564992893487215, 0.005297042895108461, -0.008536836132407188, -0.02869638241827488, 0.03994368389248848, -0.03730079159140587, 0.0043322378769516945, -0.009948380291461945, 0.02857625111937523, 0.009265132248401642, -0....
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
(old topic but creating a simplified examples and ffmpeg reference) **Is it possible to generate dash `trun.version=1 (negative composite offset)` for audio segments?** Video segments use `trun.version=1` format but never able to do it on audio segments. Validators may output warnings if version=1 is not used. Mp4box can write `trun.version=1` for video segments but no audio segments. ``` mp4box -dash 3840 -frag 1920 -dash-scale 1000 -mem-frags -rap -profile dashavc264:live \ -profile-ext urn:hbbtv:dash:profile:isoff-live:2012 -min-buffer 2000 \ -bs-switching no \ -sample-groups-traf -single-traf --tfdt64 --tfdt_traf --noroll=yes --btrt=no --truns_first=yes \ --cmaf=cmf2 -subsegs-per-sidx -1 \ -segment-name "$RepresentationID$/$Number$$Init=i$" \ -out "manifest.mpd" \ "Agent327_Spring_640x360.mp4#trackID=1:id=v1:period=p0:asID=1:role=main:dur=11.520" \ "Agent327_Spring_128.mp4#trackID=1:id=a1:period=p0:asID=21:role=main:dur=11.520" ``` Ffmpeg can write `trun.version=1` for video and audio segments. (output folders `./0/, ./1/` must exist) ``` v1=[s1]scale=640x360:out_range=tv:out_color_matrix=bt709:flags=full_chroma_int+accurate_rnd,format=yuv420p,setsar=1/1[v1] ffmpeg -i "Agent327_Spring_640x360.mp4" -i "Agent327_Spring_128.mp4" -t "11.520" -preset fast \ -r 25 -keyint_min 25 -g 50 \ -filter_complex "[0:v]split=1[s1]; $v1" \ -refs 3 -bf 3 -b_strategy 1 -flags +cgop -sc_threshold 0 \ -c:v libx264 -pix_fmt yuv420p \ -map "[v1]" -crf:v:0 23 -profile:v:0 main -level:v:0 3.0 -maxrate:v:0 600k -bufsize:v:0 1200k -b:v:0 600k \ -metadata:s:v:0 "language=eng" -metadata:s:v:0 "role=main" \ -color_range tv -colorspace bt709 -color_primaries bt709 -color_trc bt709 \ -map 1:a -ar:a:0 48000 -ac:a:0 2 -c:a:0 aac -b:a:0 128k \ -metadata:s:a:0 "language=eng" -metadata:s:a:0 "role=main" \ -format_options "movflags=cmaf:write_btrt=0" \ -seg_duration 4 -use_template 1 -use_timeline 0 -dash_segment_type mp4 -frag_duration 4 \ -adaptation_sets "id=0,frag_type=duration,streams=v id=1,frag_type=duration,streams=a" \ -init_seg_name "$RepresentationID$/i.mp4" \ -media_seg_name "$RepresentationID$/$Number$.m4s" \ -movflags negative_cts_offsets+faststart \ -f dash -y "manifest.mpd" ``` Reference: https://github.com/gpac/gpac/issues/1260 https://github.com/gpac/gpac/issues/1253 Test input files: https://refapp.hbbtv.org/videos/dashtest/Agent327_Spring_128.mp4 https://refapp.hbbtv.org/videos/dashtest/Agent327_Spring_640x360.mp4
Generate dash "trun.version=1" audio segments never happens, video segments are fine
https://api.github.com/repos/gpac/gpac/issues/2586/comments
2
2023-09-04T07:40:35
2023-09-15T20:51:14Z
https://github.com/gpac/gpac/issues/2586
1,879,698,855
2,586
false
This is a GitHub Issue repo:gpac owner:gpac Title : Generate dash "trun.version=1" audio segments never happens, video segments are fine Issue date: --- start body --- (old topic but creating a simplified examples and ffmpeg reference) **Is it possible to generate dash `trun.version=1 (negative composite offset)` for audio segments?** Video segments use `trun.version=1` format but never able to do it on audio segments. Validators may output warnings if version=1 is not used. Mp4box can write `trun.version=1` for video segments but no audio segments. ``` mp4box -dash 3840 -frag 1920 -dash-scale 1000 -mem-frags -rap -profile dashavc264:live \ -profile-ext urn:hbbtv:dash:profile:isoff-live:2012 -min-buffer 2000 \ -bs-switching no \ -sample-groups-traf -single-traf --tfdt64 --tfdt_traf --noroll=yes --btrt=no --truns_first=yes \ --cmaf=cmf2 -subsegs-per-sidx -1 \ -segment-name "$RepresentationID$/$Number$$Init=i$" \ -out "manifest.mpd" \ "Agent327_Spring_640x360.mp4#trackID=1:id=v1:period=p0:asID=1:role=main:dur=11.520" \ "Agent327_Spring_128.mp4#trackID=1:id=a1:period=p0:asID=21:role=main:dur=11.520" ``` Ffmpeg can write `trun.version=1` for video and audio segments. (output folders `./0/, ./1/` must exist) ``` v1=[s1]scale=640x360:out_range=tv:out_color_matrix=bt709:flags=full_chroma_int+accurate_rnd,format=yuv420p,setsar=1/1[v1] ffmpeg -i "Agent327_Spring_640x360.mp4" -i "Agent327_Spring_128.mp4" -t "11.520" -preset fast \ -r 25 -keyint_min 25 -g 50 \ -filter_complex "[0:v]split=1[s1]; $v1" \ -refs 3 -bf 3 -b_strategy 1 -flags +cgop -sc_threshold 0 \ -c:v libx264 -pix_fmt yuv420p \ -map "[v1]" -crf:v:0 23 -profile:v:0 main -level:v:0 3.0 -maxrate:v:0 600k -bufsize:v:0 1200k -b:v:0 600k \ -metadata:s:v:0 "language=eng" -metadata:s:v:0 "role=main" \ -color_range tv -colorspace bt709 -color_primaries bt709 -color_trc bt709 \ -map 1:a -ar:a:0 48000 -ac:a:0 2 -c:a:0 aac -b:a:0 128k \ -metadata:s:a:0 "language=eng" -metadata:s:a:0 "role=main" \ -format_options "movflags=cmaf:write_btrt=0" \ -seg_duration 4 -use_template 1 -use_timeline 0 -dash_segment_type mp4 -frag_duration 4 \ -adaptation_sets "id=0,frag_type=duration,streams=v id=1,frag_type=duration,streams=a" \ -init_seg_name "$RepresentationID$/i.mp4" \ -media_seg_name "$RepresentationID$/$Number$.m4s" \ -movflags negative_cts_offsets+faststart \ -f dash -y "manifest.mpd" ``` Reference: https://github.com/gpac/gpac/issues/1260 https://github.com/gpac/gpac/issues/1253 Test input files: https://refapp.hbbtv.org/videos/dashtest/Agent327_Spring_128.mp4 https://refapp.hbbtv.org/videos/dashtest/Agent327_Spring_640x360.mp4 --- end body ---
2,736
[ -0.05941224843263626, 0.013156245462596416, -0.022129641845822334, 0.011443644762039185, -0.013132568448781967, -0.005394297186285257, -0.018199341371655464, 0.042775556445121765, -0.023376604542136192, -0.006313735153526068, -0.008168394677340984, 0.01359031442552805, 0.014079629443585873, ...
null
null
null
null
null
null
null
null
null
[ "ImageMagick", "ImageMagick" ]
### ImageMagick version 7.1.1-32 ### Operating system Linux ### Operating system, version and so on Ubuntu 24.04 LTS x86_64 ### Description ``` $ magick -version Version: ImageMagick 7.1.1-32 Q16-HDRI x86_64 22207 https://imagemagick.org Copyright: (C) 1999 ImageMagick Studio LLC License: https://imagemagick.org/script/license.php Features: Cipher DPC HDRI OpenMP(4.5) Delegates (built-in): fontconfig freetype heic png x xml zlib Compiler: gcc (11.4) ``` The screenshot below is comparison of original `hdr.avif`(left) and `out.avif`(right). As you can see, the ouput file `out.avif` is washed out, which is not desired. ![329444577-d183aa4d-d9c0-4d89-8f0b-1a59552f8bc0](https://github.com/ImageMagick/ImageMagick/assets/3142886/b9bf9f23-76ba-49a5-b41e-69016cdcbbbc) ### Steps to Reproduce 1. Save [https://raw.githubusercontent.com/AOMediaCodec/av1-avif/master/testFiles/Netflix/avif/hdr_cosmos01650_cicp9-16-9_yuv444_full_qp10.avif](https://raw.githubusercontent.com/AOMediaCodec/av1-avif/master/testFiles/Netflix/avif/hdr_cosmos01650_cicp9-16-9_yuv444_full_qp10.avif) as `hdr.avif` 2. Execute command: `magick hdr.avif -resize 50% out.avif` 3. Preview the original file and output file in Chrome `Version 124.0.6367.158 (Official Build) (64-bit)`(Any version above v85 should also produce the same result). ### Images [avif.zip](https://github.com/ImageMagick/ImageMagick/files/15282916/avif.zip) (This .zip file contains original `hdr.avif` and output `out.avif`)
Output .avif image is washed out
https://api.github.com/repos/ImageMagick/ImageMagick/issues/7308/comments
6
2024-05-11T13:21:59
2024-05-15T11:49:25Z
https://github.com/ImageMagick/ImageMagick/issues/7308
2,290,897,195
7,308
false
This is a GitHub Issue repo:ImageMagick owner:ImageMagick Title : Output .avif image is washed out Issue date: --- start body --- ### ImageMagick version 7.1.1-32 ### Operating system Linux ### Operating system, version and so on Ubuntu 24.04 LTS x86_64 ### Description ``` $ magick -version Version: ImageMagick 7.1.1-32 Q16-HDRI x86_64 22207 https://imagemagick.org Copyright: (C) 1999 ImageMagick Studio LLC License: https://imagemagick.org/script/license.php Features: Cipher DPC HDRI OpenMP(4.5) Delegates (built-in): fontconfig freetype heic png x xml zlib Compiler: gcc (11.4) ``` The screenshot below is comparison of original `hdr.avif`(left) and `out.avif`(right). As you can see, the ouput file `out.avif` is washed out, which is not desired. ![329444577-d183aa4d-d9c0-4d89-8f0b-1a59552f8bc0](https://github.com/ImageMagick/ImageMagick/assets/3142886/b9bf9f23-76ba-49a5-b41e-69016cdcbbbc) ### Steps to Reproduce 1. Save [https://raw.githubusercontent.com/AOMediaCodec/av1-avif/master/testFiles/Netflix/avif/hdr_cosmos01650_cicp9-16-9_yuv444_full_qp10.avif](https://raw.githubusercontent.com/AOMediaCodec/av1-avif/master/testFiles/Netflix/avif/hdr_cosmos01650_cicp9-16-9_yuv444_full_qp10.avif) as `hdr.avif` 2. Execute command: `magick hdr.avif -resize 50% out.avif` 3. Preview the original file and output file in Chrome `Version 124.0.6367.158 (Official Build) (64-bit)`(Any version above v85 should also produce the same result). ### Images [avif.zip](https://github.com/ImageMagick/ImageMagick/files/15282916/avif.zip) (This .zip file contains original `hdr.avif` and output `out.avif`) --- end body ---
1,675
[ -0.009106267243623734, -0.013134435750544071, 0.0010379223385825753, 0.012537415139377117, 0.0079739885404706, 0.017443956807255745, -0.004848212469369173, 0.03324096277356148, -0.011665903963148594, 0.009140579029917717, 0.00349634001031518, 0.01670282892882824, 0.0031789587810635567, 0.0...
null
null
null
null
null
null
null
null
null
[ "ImageMagick", "ImageMagick" ]
### ImageMagick version 7.1.1-10 ### Operating system Linux ### Operating system, version and so on Debian Bullseye ### Description The result image displays in the wrong orientation after converting Jpeg to WebP. libwebp version is 1.3.0 ### Steps to Reproduce 1. Open the original image in Chrome or other viewers (I tried gwenview and okular) and make sure it displays in the portrait orientation 2. Run `magick test.jpeg webp:test.webp` 3. Open webp image in Chrom or other viewer (I tried gwenview and okular) Expected behaviour: the result image displays in portrait orientation similar to the original Actual behaviour: the result imaged displays in landscape orientation ### Images ![orientation](https://github.com/ImageMagick/ImageMagick/assets/124110/2c47d7c9-7d35-4a73-ad24-af4ad0326387)
Image orientation is not correct after converting to WebP
https://api.github.com/repos/ImageMagick/ImageMagick/issues/6396/comments
4
2023-06-07T23:50:05
2023-07-02T19:31:18Z
https://github.com/ImageMagick/ImageMagick/issues/6396
1,746,836,798
6,396
false
This is a GitHub Issue repo:ImageMagick owner:ImageMagick Title : Image orientation is not correct after converting to WebP Issue date: --- start body --- ### ImageMagick version 7.1.1-10 ### Operating system Linux ### Operating system, version and so on Debian Bullseye ### Description The result image displays in the wrong orientation after converting Jpeg to WebP. libwebp version is 1.3.0 ### Steps to Reproduce 1. Open the original image in Chrome or other viewers (I tried gwenview and okular) and make sure it displays in the portrait orientation 2. Run `magick test.jpeg webp:test.webp` 3. Open webp image in Chrom or other viewer (I tried gwenview and okular) Expected behaviour: the result image displays in portrait orientation similar to the original Actual behaviour: the result imaged displays in landscape orientation ### Images ![orientation](https://github.com/ImageMagick/ImageMagick/assets/124110/2c47d7c9-7d35-4a73-ad24-af4ad0326387) --- end body ---
995
[ 0.018201064318418503, 0.01789494976401329, -0.008303359150886536, 0.020127035677433014, 0.01607101783156395, 0.02339225821197033, 0.015114408917725086, 0.0353817455470562, -0.017040379345417023, 0.01667049154639244, 0.00567587511613965, -0.018762273713946342, -0.02336674928665161, -0.00980...
null
null
null
null
null
null
null
null
null
[ "LibreDWG", "libredwg" ]
e.g. tablet_from_autocad_r10_mac.json.err this error comes from illegal layer color.index defaults, because the decode already fails before in ptr tbl VPORT.size overflow.
json-check: AcDbLayerTableRecord(1D): Color index 0
https://api.github.com/repos/LibreDWG/libredwg/issues/831/comments
1
2023-09-01T15:37:05
2023-09-04T10:23:04Z
https://github.com/LibreDWG/libredwg/issues/831
1,877,640,856
831
false
This is a GitHub Issue repo:libredwg owner:LibreDWG Title : json-check: AcDbLayerTableRecord(1D): Color index 0 Issue date: --- start body --- e.g. tablet_from_autocad_r10_mac.json.err this error comes from illegal layer color.index defaults, because the decode already fails before in ptr tbl VPORT.size overflow. --- end body ---
334
[ -0.031135980039834976, 0.003249628935009241, -0.009061174467206001, 0.015688907355070114, 0.025180846452713013, 0.03491462022066116, 0.02288343384861946, 0.04014425352215767, -0.01715501770377159, -0.00894025806337595, -0.00035566434962674975, -0.003330869833007455, 0.010988280177116394, 0...
null
null
null
null
null
null
null
null
null
[ "weidai11", "cryptopp" ]
### Crypto++ Issue Report The CryptoPP allocation code is vulnerable to infinite executions in case of memory exhaustion conditions. Submitted a PR to perform a fail safe operation (throw bad_alloc) instead of looping forever. Issue will likely affect all platforms. On top of this fix, I recommend having more precise exception handling throughout the project instead of using the generic (Exception e) as it may inform users about the precise error causes. I did not change this in the PR. See: https://github.com/weidai11/cryptopp/pull/1266 Thank you, Julien Please do not ask questions in the bug tracker. Please ask questions on the Crypto++ Users List at http://groups.google.com/forum/#!forum/cryptopp-users. Please do not ask questions about unsupported build systems, like Autotools, CMake, Conan and NuGet. They are other people's projects. We don't know anything about them. Please do not ask questions at Stack Overflow. We do not patrol Stack Overflow. We will not be able to answer your question. There is a wiki page with information on filing useful bug reports. If you have some time please visit http://www.cryptopp.com/wiki/Bug_Report on the wiki. The executive summary is: * State the operating system and version (Ubutnu 17 x86_64, Windows 7 Professional x64, etc) * State the version of the Crypto++ library (Crypto++ 7.0, Master, etc) * State how you built the library (Visual Studio, Makefile, distro provided, etc) * Show a typical command line (the output of the compiler for cryptlib.cpp) * Show the link command (the output of the linker for libcryptopp.so or cryptest.exe) * Show the exact error message you are receiving (copy and paste it); or * Clearly state the undesired behavior (and state the expected behavior)
Infinite execution behavior in main branch
https://api.github.com/repos/weidai11/cryptopp/issues/1267/comments
0
2024-03-18T19:51:25
2024-03-18T19:51:25Z
https://github.com/weidai11/cryptopp/issues/1267
2,193,146,946
1,267
false
This is a GitHub Issue repo:cryptopp owner:weidai11 Title : Infinite execution behavior in main branch Issue date: --- start body --- ### Crypto++ Issue Report The CryptoPP allocation code is vulnerable to infinite executions in case of memory exhaustion conditions. Submitted a PR to perform a fail safe operation (throw bad_alloc) instead of looping forever. Issue will likely affect all platforms. On top of this fix, I recommend having more precise exception handling throughout the project instead of using the generic (Exception e) as it may inform users about the precise error causes. I did not change this in the PR. See: https://github.com/weidai11/cryptopp/pull/1266 Thank you, Julien Please do not ask questions in the bug tracker. Please ask questions on the Crypto++ Users List at http://groups.google.com/forum/#!forum/cryptopp-users. Please do not ask questions about unsupported build systems, like Autotools, CMake, Conan and NuGet. They are other people's projects. We don't know anything about them. Please do not ask questions at Stack Overflow. We do not patrol Stack Overflow. We will not be able to answer your question. There is a wiki page with information on filing useful bug reports. If you have some time please visit http://www.cryptopp.com/wiki/Bug_Report on the wiki. The executive summary is: * State the operating system and version (Ubutnu 17 x86_64, Windows 7 Professional x64, etc) * State the version of the Crypto++ library (Crypto++ 7.0, Master, etc) * State how you built the library (Visual Studio, Makefile, distro provided, etc) * Show a typical command line (the output of the compiler for cryptlib.cpp) * Show the link command (the output of the linker for libcryptopp.so or cryptest.exe) * Show the exact error message you are receiving (copy and paste it); or * Clearly state the undesired behavior (and state the expected behavior) --- end body ---
1,939
[ -0.0527791790664196, 0.010739948600530624, -0.012072151526808739, 0.012566114775836468, 0.051521819084882736, -0.001355591113679111, -0.02814091183245182, 0.030565820634365082, -0.013756115920841694, 0.05478496477007866, -0.0012161963386461139, 0.008397367782890797, 0.01059774775058031, -0...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
hi, when i using the osmo4 to play 360 tiled video on vr headset such as [oculus quest2](https://www.oculus.com/), it can't work properly , but it can be played normally on my android phone. perhaps it is because lack of support of the underlying environment? thank you
can osmo4 support vr headset such as oculus quest2?
https://api.github.com/repos/gpac/gpac/issues/1937/comments
5
2021-10-20T01:26:30
2022-03-23T02:40:23Z
https://github.com/gpac/gpac/issues/1937
1,030,878,098
1,937
false
This is a GitHub Issue repo:gpac owner:gpac Title : can osmo4 support vr headset such as oculus quest2? Issue date: --- start body --- hi, when i using the osmo4 to play 360 tiled video on vr headset such as [oculus quest2](https://www.oculus.com/), it can't work properly , but it can be played normally on my android phone. perhaps it is because lack of support of the underlying environment? thank you --- end body ---
429
[ -0.02550029754638672, 0.026867171749472618, -0.026220478117465973, 0.0072679524309933186, 0.009222730062901974, 0.03083551861345768, -0.032334670424461365, 0.04650313779711723, -0.013073495589196682, 0.04179991036653519, -0.0203708428889513, -0.003049747319892049, 0.017504815012216568, 0.0...
null
null
null
null
null
null
null
null
null
[ "ImageMagick", "ImageMagick" ]
### ImageMagick version 7.1.1-26 ### Operating system Other (enter below) ### Operating system, version and so on FreeBSD 13.2-RELEASE-p9 ### Description We have been using OpenMP with Imagick for as long as I can remember (many years), but after our system was recently updated to PHP 8.1+ and ImageMagick 7.1.1-26, we started experiencing random instances of processes getting locked at 100% CPU usage with `truss` revealing that sched_yield(); is being called in an infinite loop on the locked processes. This appears to happen, on seemingly random occasions, when the Imagick and OpenMP modules initially load. Backtrace details for an example process that was stuck in a sched_yield() infinite loop can be seen below: ``` (gdb) thread apply all bt Thread 1 (LWP 100499 of process 33248): #0 sched_yield () at sched_yield.S:4 #1 0x00000008495aa491 in __kmp_wait_4_ptr (spinner=0x84960b51c <__kmp_initz_lock+28>, checker=1, pred=0x8495c4b50 <__kmp_bakery_check(void*, unsigned int)>, obj=0x84960b500 <__kmp_initz_lock>) at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_dispatch.cpp:2695 #2 0x00000008495c37c8 in __kmp_acquire_ticket_lock_timed_template (lck=0x0, gtid=<optimized out>) at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_lock.cpp:640 #3 __kmp_acquire_ticket_lock (lck=0x0, gtid=<optimized out>) at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_lock.cpp:645 #4 0x00000008495c9564 in __kmp_acquire_bootstrap_lock (lck=0x84960b51c <__kmp_initz_lock+28>) at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_lock.h:524 #5 __kmp_get_global_thread_id_reg () at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_runtime.cpp:253 #6 0x000000084958991e in __kmp_entry_thread () at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp.h:3401 #7 kmpc_malloc (size=35590812956) at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_alloc.cpp:1098 #8 0x00000008496041b7 in __kmp_is_address_mapped (addr=0x847bad5c8) at /usr/src/contrib/llvm-project/openmp/runtime/src/z_Linux_util.cpp:2067 #9 0x00000008495d3925 in __kmp_register_library_startup () at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_runtime.cpp:6774 #10 0x00000008495c95c4 in __kmp_do_serial_initialize () at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_runtime.cpp:6966 #11 0x00000008495c9578 in __kmp_get_global_thread_id_reg () at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_runtime.cpp:255 #12 0x00000008495b40b2 in __kmp_api_omp_init_lock (user_lock=0x84960b51c <__kmp_initz_lock+28>) at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_ftn_entry.h:1083 #13 0x000000084b24773c in MagickCoreGenesis () from /usr/local/lib/libMagickCore-7.so.10 #14 0x000000084a26bd8e in zm_startup_imagick () from /usr/local/lib/php/20230831-debug/imagick.so #15 0x000000000083b010 in zend_startup_module_ex () #16 0x000000000083bb10 in ?? () #17 0x0000000000853131 in zend_hash_apply () #18 0x000000000083b90c in zend_startup_modules () #19 0x0000000000766b0f in php_module_startup () #20 0x00000000009f51e9 in ?? () #21 0x00000000009f37a4 in ?? () #22 0x0000000000423720 in _start () (gdb) ``` We have tried updating from PHP 8.1 to 8.3, and we have tried rebuilding Image Magick with a few different options disabled, but the problem persists. Unfortunately, we cannot simply disable OpenMP support, since doing so results in a _significant_ slowdown of the convert processes. ### Steps to Reproduce Repeatedly run a `convert` operation (any operation will do) on an image in a loop, using PHP, until the process gets stuck. Then, view what's happening through the use of `truss`, which will reveal that it is stuck in a sched_yield() infinite loop. ### Images _No response_
CPU pegged at 100% and sched_yield() called in infinite loop when OpenMP enabled
https://api.github.com/repos/ImageMagick/ImageMagick/issues/7190/comments
9
2024-03-22T06:00:29
2024-04-15T15:26:44Z
https://github.com/ImageMagick/ImageMagick/issues/7190
2,201,761,538
7,190
false
This is a GitHub Issue repo:ImageMagick owner:ImageMagick Title : CPU pegged at 100% and sched_yield() called in infinite loop when OpenMP enabled Issue date: --- start body --- ### ImageMagick version 7.1.1-26 ### Operating system Other (enter below) ### Operating system, version and so on FreeBSD 13.2-RELEASE-p9 ### Description We have been using OpenMP with Imagick for as long as I can remember (many years), but after our system was recently updated to PHP 8.1+ and ImageMagick 7.1.1-26, we started experiencing random instances of processes getting locked at 100% CPU usage with `truss` revealing that sched_yield(); is being called in an infinite loop on the locked processes. This appears to happen, on seemingly random occasions, when the Imagick and OpenMP modules initially load. Backtrace details for an example process that was stuck in a sched_yield() infinite loop can be seen below: ``` (gdb) thread apply all bt Thread 1 (LWP 100499 of process 33248): #0 sched_yield () at sched_yield.S:4 #1 0x00000008495aa491 in __kmp_wait_4_ptr (spinner=0x84960b51c <__kmp_initz_lock+28>, checker=1, pred=0x8495c4b50 <__kmp_bakery_check(void*, unsigned int)>, obj=0x84960b500 <__kmp_initz_lock>) at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_dispatch.cpp:2695 #2 0x00000008495c37c8 in __kmp_acquire_ticket_lock_timed_template (lck=0x0, gtid=<optimized out>) at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_lock.cpp:640 #3 __kmp_acquire_ticket_lock (lck=0x0, gtid=<optimized out>) at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_lock.cpp:645 #4 0x00000008495c9564 in __kmp_acquire_bootstrap_lock (lck=0x84960b51c <__kmp_initz_lock+28>) at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_lock.h:524 #5 __kmp_get_global_thread_id_reg () at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_runtime.cpp:253 #6 0x000000084958991e in __kmp_entry_thread () at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp.h:3401 #7 kmpc_malloc (size=35590812956) at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_alloc.cpp:1098 #8 0x00000008496041b7 in __kmp_is_address_mapped (addr=0x847bad5c8) at /usr/src/contrib/llvm-project/openmp/runtime/src/z_Linux_util.cpp:2067 #9 0x00000008495d3925 in __kmp_register_library_startup () at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_runtime.cpp:6774 #10 0x00000008495c95c4 in __kmp_do_serial_initialize () at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_runtime.cpp:6966 #11 0x00000008495c9578 in __kmp_get_global_thread_id_reg () at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_runtime.cpp:255 #12 0x00000008495b40b2 in __kmp_api_omp_init_lock (user_lock=0x84960b51c <__kmp_initz_lock+28>) at /usr/src/contrib/llvm-project/openmp/runtime/src/kmp_ftn_entry.h:1083 #13 0x000000084b24773c in MagickCoreGenesis () from /usr/local/lib/libMagickCore-7.so.10 #14 0x000000084a26bd8e in zm_startup_imagick () from /usr/local/lib/php/20230831-debug/imagick.so #15 0x000000000083b010 in zend_startup_module_ex () #16 0x000000000083bb10 in ?? () #17 0x0000000000853131 in zend_hash_apply () #18 0x000000000083b90c in zend_startup_modules () #19 0x0000000000766b0f in php_module_startup () #20 0x00000000009f51e9 in ?? () #21 0x00000000009f37a4 in ?? () #22 0x0000000000423720 in _start () (gdb) ``` We have tried updating from PHP 8.1 to 8.3, and we have tried rebuilding Image Magick with a few different options disabled, but the problem persists. Unfortunately, we cannot simply disable OpenMP support, since doing so results in a _significant_ slowdown of the convert processes. ### Steps to Reproduce Repeatedly run a `convert` operation (any operation will do) on an image in a loop, using PHP, until the process gets stuck. Then, view what's happening through the use of `truss`, which will reveal that it is stuck in a sched_yield() infinite loop. ### Images _No response_ --- end body ---
3,939
[ -0.019047068431973457, -0.003968139179050922, -0.00662194425240159, 0.030739611014723778, 0.011505807749927044, -0.0006015052204020321, -0.027464549988508224, 0.02940373122692108, 0.004334428813308477, 0.04401222988963127, 0.011857733130455017, -0.001668952638283372, 0.01986583322286606, 0...
null
null
null
null
null
null
null
null
null
[ "chakra-core", "ChakraCore" ]
# commit c3ead3f8a6e0bb8e32e043adc091c68cba5935e9 # bug Address misalignment causes futex error # build and run os: Ubuntu 22.04.2 LTS build: ./build.sh --debug --static -j=8 run: ./DebugBuild/Debug/ch this.js Run 10 times, 3 crashes # POC ```JavaScript function classOf(object) { var string = Object.prototype.toString.call(object); return string.substring(8, string.length - 1); } function deepObjectEquals(a, b) { var aProps = Object.keys(a); aProps.sort(); var bProps = Object.keys(b); bProps.sort(); if (!deepEquals(aProps, bProps)) { return false; } for (var i = 0; i < aProps.length; i++) { if (!deepEquals(a[aProps[i]], b[aProps[i]])) { return false; } } return true; } function deepEquals(a, b) { if (a === b) { if (a === 0) return (1 / a) === (1 / b); return true; } if (typeof a != typeof b) return false; if (typeof a == 'number') return (isNaN(a) && isNaN(b)) || (a===b); if (typeof a !== 'object' && typeof a !== 'function' && typeof a !== 'symbol') return false; var objectClass = classOf(a); if (objectClass === 'Array') { if (a.length != b.length) { return false; } for (var i = 0; i < a.length; i++) { if (!deepEquals(a[i], b[i])) return false; } return true; } if (objectClass !== classOf(b)) return false; if (objectClass === 'RegExp') { return (a.toString() === b.toString()); } if (objectClass === 'Function') return true; if (objectClass == 'String' || objectClass == 'Number' || objectClass == 'Boolean' || objectClass == 'Date') { if (a.valueOf() !== b.valueOf()) return false; } return deepObjectEquals(a, b); } function opt(opt_param){ const v0 = {}; const v1 = [v0,v0,v0,v0,v0]; const v2 = v0[v0]; const v3 = class V3 { constructor(v5,v6,v7) { ({"MAX_SAFE_INTEGER":v4,"dotAll":v6,} = v2); function v8(v9,v10,v11,v12) { const v13 = {"get":v10,"set":v6}; } } get(v15,v16,v17) { const v18 = v3 >>> v16; } pop(v20,v21,v22,v23) { let {"NEGATIVE_INFINITY":v24,"byteLength":v25,"source":v26,"split":v27,...v28} = v22; } }; return v3; } function main(){ let gflag = true; let jit_a0 = opt(true); let jit_a0_0 = opt(false); for(let i=0;i<0x10;i++){opt(false);} let jit_a2 = opt(true); if (jit_a0 === undefined && jit_a2 === undefined) { opt(true); } else { if (jit_a0_0===jit_a0 && !deepEquals(jit_a0, jit_a2)) { gflag = false; } } for(let i=0;i<0x200;i++){opt(false);} let jit_a4 = opt(true); if (jit_a0 === undefined && jit_a4 === undefined) { opt(true); } else { if (gflag && jit_a0_0===jit_a0 && !deepEquals(jit_a0, jit_a4)) { fuzzilli('FUZZILLI_CRASH', 0); } } } main(); ``` # fix suggestions old in cclock.hpp line 12-16: ```c++ #if defined(_M_ARM64) #define CCLOCK_ALIGN __declspec(align(8)) #else #define CCLOCK_ALIGN #endif ``` new in cclock.hpp: ```c++ #define CCLOCK_ALIGN __declspec(align(8)) ```
Address misalignment causes futex error
https://api.github.com/repos/chakra-core/ChakraCore/issues/6971/comments
6
2024-03-20T09:23:57
2024-04-22T04:27:30Z
https://github.com/chakra-core/ChakraCore/issues/6971
2,197,040,325
6,971
false
This is a GitHub Issue repo:ChakraCore owner:chakra-core Title : Address misalignment causes futex error Issue date: --- start body --- # commit c3ead3f8a6e0bb8e32e043adc091c68cba5935e9 # bug Address misalignment causes futex error # build and run os: Ubuntu 22.04.2 LTS build: ./build.sh --debug --static -j=8 run: ./DebugBuild/Debug/ch this.js Run 10 times, 3 crashes # POC ```JavaScript function classOf(object) { var string = Object.prototype.toString.call(object); return string.substring(8, string.length - 1); } function deepObjectEquals(a, b) { var aProps = Object.keys(a); aProps.sort(); var bProps = Object.keys(b); bProps.sort(); if (!deepEquals(aProps, bProps)) { return false; } for (var i = 0; i < aProps.length; i++) { if (!deepEquals(a[aProps[i]], b[aProps[i]])) { return false; } } return true; } function deepEquals(a, b) { if (a === b) { if (a === 0) return (1 / a) === (1 / b); return true; } if (typeof a != typeof b) return false; if (typeof a == 'number') return (isNaN(a) && isNaN(b)) || (a===b); if (typeof a !== 'object' && typeof a !== 'function' && typeof a !== 'symbol') return false; var objectClass = classOf(a); if (objectClass === 'Array') { if (a.length != b.length) { return false; } for (var i = 0; i < a.length; i++) { if (!deepEquals(a[i], b[i])) return false; } return true; } if (objectClass !== classOf(b)) return false; if (objectClass === 'RegExp') { return (a.toString() === b.toString()); } if (objectClass === 'Function') return true; if (objectClass == 'String' || objectClass == 'Number' || objectClass == 'Boolean' || objectClass == 'Date') { if (a.valueOf() !== b.valueOf()) return false; } return deepObjectEquals(a, b); } function opt(opt_param){ const v0 = {}; const v1 = [v0,v0,v0,v0,v0]; const v2 = v0[v0]; const v3 = class V3 { constructor(v5,v6,v7) { ({"MAX_SAFE_INTEGER":v4,"dotAll":v6,} = v2); function v8(v9,v10,v11,v12) { const v13 = {"get":v10,"set":v6}; } } get(v15,v16,v17) { const v18 = v3 >>> v16; } pop(v20,v21,v22,v23) { let {"NEGATIVE_INFINITY":v24,"byteLength":v25,"source":v26,"split":v27,...v28} = v22; } }; return v3; } function main(){ let gflag = true; let jit_a0 = opt(true); let jit_a0_0 = opt(false); for(let i=0;i<0x10;i++){opt(false);} let jit_a2 = opt(true); if (jit_a0 === undefined && jit_a2 === undefined) { opt(true); } else { if (jit_a0_0===jit_a0 && !deepEquals(jit_a0, jit_a2)) { gflag = false; } } for(let i=0;i<0x200;i++){opt(false);} let jit_a4 = opt(true); if (jit_a0 === undefined && jit_a4 === undefined) { opt(true); } else { if (gflag && jit_a0_0===jit_a0 && !deepEquals(jit_a0, jit_a4)) { fuzzilli('FUZZILLI_CRASH', 0); } } } main(); ``` # fix suggestions old in cclock.hpp line 12-16: ```c++ #if defined(_M_ARM64) #define CCLOCK_ALIGN __declspec(align(8)) #else #define CCLOCK_ALIGN #endif ``` new in cclock.hpp: ```c++ #define CCLOCK_ALIGN __declspec(align(8)) ``` --- end body ---
3,303
[ -0.005557711236178875, -0.012341601774096489, -0.003143444424495101, 0.029892651364207268, 0.01735515147447586, 0.02649707905948162, -0.04190775752067566, 0.026206858456134796, -0.009424892254173756, 0.0014021249953657389, -0.01572991907596588, 0.03021189384162426, 0.004625380039215088, -0...
null
null
null
null
null
null
null
null
null
[ "chakra-core", "ChakraCore" ]
## Related issues - [ ] #6965 - [ ] #6964 - [ ] #6963 - [ ] #6691 - [ ] #6704 ## Progress so far - #6293 - #6533 - #6662 - #6684 - #6700 - #6706 --- @rhuanjl @ppenzin Feel free to edit this message
[Epic]: Fix jitting of generator functions
https://api.github.com/repos/chakra-core/ChakraCore/issues/6988/comments
4
2024-04-23T09:18:58
2024-05-30T22:22:39Z
https://github.com/chakra-core/ChakraCore/issues/6988
2,258,372,473
6,988
false
This is a GitHub Issue repo:ChakraCore owner:chakra-core Title : [Epic]: Fix jitting of generator functions Issue date: --- start body --- ## Related issues - [ ] #6965 - [ ] #6964 - [ ] #6963 - [ ] #6691 - [ ] #6704 ## Progress so far - #6293 - #6533 - #6662 - #6684 - #6700 - #6706 --- @rhuanjl @ppenzin Feel free to edit this message --- end body ---
381
[ -0.011048729531466961, -0.010042824782431126, -0.022876223549246788, -0.0023281830362975597, -0.014707302674651146, 0.024952931329607964, -0.02761371061205864, 0.03812703862786293, 0.0064978208392858505, 0.03325975686311722, 0.003285415004938841, -0.001705576665699482, -0.004648253787308931,...
null
null
null
null
null
null
null
null
null
[ "libming", "libming" ]
### Envrionment ```bash ubuntu  ~/lab/libming/util   master  lsb_release -a No LSB modules are available. Distributor ID: Ubuntu Description: Ubuntu 22.04.2 LTS Release: 22.04 Codename: jammy ``` ### Compiler ```bash ubuntu  ~/lab/libming/util   master  clang -v Ubuntu clang version 14.0.0-1ubuntu1.1 Target: x86_64-pc-linux-gnu Thread model: posix ``` ### Step to reproduce ```bash ./autogen.sh ./configure --disable-shared cd util make swftophp ./swftophp <PoC file> ``` ### Affected Version libming v0.4.8 (https://github.com/libming/libming/releases/tag/ming-0_4_8) ### Contents of PoC file [poc-2.zip](https://github.com/libming/libming/files/13924982/poc-2.zip) ### Expected Behavior Print error or warning messages handled within swftophp. ### Current Behavior * stdout ``` <?php $m = new SWFMovie(10); ming_setscale(1.0); $m->setRate(255.019531); $m->setDimension(2, -1); /* Note: xMin and/or yMin are not 0! */ $m->setFrames(62719); /* SWF_PLACEOBJECT */ /*Unknown block type 648*/ skipping 59 bytes /* SWF_DEFINESPRITE */ /* MovieClip 2570 */ $character2570 = new SWFMovieClip(); /* 2314 frames */ /* SWF_FREECHARACTER */ /* SWF_DEFINETEXT2 */ $character2056 = new SWFText(2); $character2056->moveTo(2815, 2058); $character2056->addString("X"); 0000: 00 00 10 00 00 0a 09 f6 0a a1 09 0a 24 0a 0a 0a ........ ....$... 0010: 16 01 02 ... header('Content-type: application/x-shockwave-flash'); $m->output(9); ?> ``` * stderr ``` header indicates a filesize of 18446744071613768297 but filesize is 144 Stream out of sync after parse of blocktype 39 (SWF_DEFINESPRITE). 105 but expecting 79. Stream out of sync after parse of blocktype 33 (SWF_DEFINETEXT2). 113 but expecting 123. outputSWF_TEXT_RECORD: can't process text record: fonts information list is NULL sudden file end: read failed @125 fileSize 144, request 42 extra garbage (i.e., we messed up in main): ================================================================= ==18453==ERROR: LeakSanitizer: detected memory leaks ... Indirect leak of 264 byte(s) in 1 object(s) allocated from: #0 0x55ba893d5478 in __interceptor_calloc (/home/ubuntu/lab/libming/util/swftophp+0xbb478) (BuildId: 8551de914f0fa059359bb0c15882dbeddbbf77b6) #1 0x55ba894aaebf in parseSWF_FREECHARACTER /home/ubuntu/lab/libming/util/parser.c:2534:3 ... SUMMARY: AddressSanitizer: 711 byte(s) leaked in 27 allocation(s). ``` You can get full log [here](https://pastebin.com/53DnWeF0). ### Analysis https://github.com/libming/libming/blob/04aee52363688426eab74f5d6180c149654a6473/util/parser.c#L39-L52 On `parseSWF_FREECHARACTER` function, just works three MACROS: `PAR_BEGIN`, `PAR_END`, `SKIP`. https://github.com/libming/libming/blob/04aee52363688426eab74f5d6180c149654a6473/util/parser.c#L39-L45 On `PAR_BEGIN` macro, we use `calloc` for allocating SWF_Parserstruct. However, this is not been freed, so a memory leak occurs. This vulnerability allows attackers to cause a Denial of Service (DoS) via a crafted SWF file.
Memory leaks on parseSWF_FREECHARACTER
https://api.github.com/repos/libming/libming/issues/308/comments
1
2024-01-12T23:32:46
2024-02-21T22:34:44Z
https://github.com/libming/libming/issues/308
2,079,882,130
308
false
This is a GitHub Issue repo:libming owner:libming Title : Memory leaks on parseSWF_FREECHARACTER Issue date: --- start body --- ### Envrionment ```bash ubuntu  ~/lab/libming/util   master  lsb_release -a No LSB modules are available. Distributor ID: Ubuntu Description: Ubuntu 22.04.2 LTS Release: 22.04 Codename: jammy ``` ### Compiler ```bash ubuntu  ~/lab/libming/util   master  clang -v Ubuntu clang version 14.0.0-1ubuntu1.1 Target: x86_64-pc-linux-gnu Thread model: posix ``` ### Step to reproduce ```bash ./autogen.sh ./configure --disable-shared cd util make swftophp ./swftophp <PoC file> ``` ### Affected Version libming v0.4.8 (https://github.com/libming/libming/releases/tag/ming-0_4_8) ### Contents of PoC file [poc-2.zip](https://github.com/libming/libming/files/13924982/poc-2.zip) ### Expected Behavior Print error or warning messages handled within swftophp. ### Current Behavior * stdout ``` <?php $m = new SWFMovie(10); ming_setscale(1.0); $m->setRate(255.019531); $m->setDimension(2, -1); /* Note: xMin and/or yMin are not 0! */ $m->setFrames(62719); /* SWF_PLACEOBJECT */ /*Unknown block type 648*/ skipping 59 bytes /* SWF_DEFINESPRITE */ /* MovieClip 2570 */ $character2570 = new SWFMovieClip(); /* 2314 frames */ /* SWF_FREECHARACTER */ /* SWF_DEFINETEXT2 */ $character2056 = new SWFText(2); $character2056->moveTo(2815, 2058); $character2056->addString("X"); 0000: 00 00 10 00 00 0a 09 f6 0a a1 09 0a 24 0a 0a 0a ........ ....$... 0010: 16 01 02 ... header('Content-type: application/x-shockwave-flash'); $m->output(9); ?> ``` * stderr ``` header indicates a filesize of 18446744071613768297 but filesize is 144 Stream out of sync after parse of blocktype 39 (SWF_DEFINESPRITE). 105 but expecting 79. Stream out of sync after parse of blocktype 33 (SWF_DEFINETEXT2). 113 but expecting 123. outputSWF_TEXT_RECORD: can't process text record: fonts information list is NULL sudden file end: read failed @125 fileSize 144, request 42 extra garbage (i.e., we messed up in main): ================================================================= ==18453==ERROR: LeakSanitizer: detected memory leaks ... Indirect leak of 264 byte(s) in 1 object(s) allocated from: #0 0x55ba893d5478 in __interceptor_calloc (/home/ubuntu/lab/libming/util/swftophp+0xbb478) (BuildId: 8551de914f0fa059359bb0c15882dbeddbbf77b6) #1 0x55ba894aaebf in parseSWF_FREECHARACTER /home/ubuntu/lab/libming/util/parser.c:2534:3 ... SUMMARY: AddressSanitizer: 711 byte(s) leaked in 27 allocation(s). ``` You can get full log [here](https://pastebin.com/53DnWeF0). ### Analysis https://github.com/libming/libming/blob/04aee52363688426eab74f5d6180c149654a6473/util/parser.c#L39-L52 On `parseSWF_FREECHARACTER` function, just works three MACROS: `PAR_BEGIN`, `PAR_END`, `SKIP`. https://github.com/libming/libming/blob/04aee52363688426eab74f5d6180c149654a6473/util/parser.c#L39-L45 On `PAR_BEGIN` macro, we use `calloc` for allocating SWF_Parserstruct. However, this is not been freed, so a memory leak occurs. This vulnerability allows attackers to cause a Denial of Service (DoS) via a crafted SWF file. --- end body ---
3,345
[ -0.034188657999038696, 0.01226785872131586, -0.00582097377628088, 0.015119231306016445, 0.06120020151138306, 0.04005831480026245, -0.024758264422416687, 0.017052602022886276, -0.04136577248573303, 0.023269986733794212, 0.016329325735569, 0.0076847979798913, -0.014242956414818764, -0.010424...
null
null
null
null
null
null
null
null
null
[ "ImageMagick", "ImageMagick" ]
### ImageMagick version 7.1.1-15 ### Operating system Linux ### Operating system, version and so on RHEL 9.2 ### Description Your Linux Binary Release RPMs are not compatible with RHEL 9.2 distributions. Using EPEL repository is not a workaround. The installation fails: ``` $ docker run -it --name dest almalinux:latest # cd /tmp # dnf install ./ImageMagick-7.1.1-15.x86_64.rpm ./ImageMagick-libs-7.1.1-15.x86_64.rpm Last metadata expiration check: 0:01:00 ago on Thu Aug 24 11:40:09 2023. Error: Problem 1: conflicting requests - nothing provides libgs.so.10()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl.so.0.7()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl.so.0.7(JXL_0)(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl_threads.so.0.7()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl_threads.so.0.7(JXL_0)(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libm.so.6(GLIBC_2.35)(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libraw_r.so.23()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 Problem 2: package ImageMagick-7.1.1-15.x86_64 requires libMagickCore-7.Q16HDRI.so.10()(64bit), but none of the providers can be installed - package ImageMagick-7.1.1-15.x86_64 requires libMagickWand-7.Q16HDRI.so.10()(64bit), but none of the providers can be installed - package ImageMagick-7.1.1-15.x86_64 requires libMagickWand-7.Q16HDRI.so.10(VERS_10.0)(64bit), but none of the providers can be installed - package ImageMagick-7.1.1-15.x86_64 requires ImageMagick-libs(x86-64) = 7.1.1-15, but none of the providers can be installed - conflicting requests - nothing provides libgs.so.10()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl.so.0.7()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl.so.0.7(JXL_0)(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl_threads.so.0.7()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl_threads.so.0.7(JXL_0)(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libm.so.6(GLIBC_2.35)(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libraw_r.so.23()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 (try to add '--skip-broken' to skip uninstallable packages or '--nobest' to use not only best candidate packages) ``` The following librairies are missing and are not available for the RHEL 9.2 distribution: - [libgs.so.10](https://pkgs.org/search/?q=libgs) is not available for RHEL 9.2, only on Fedora 38 - [libm.so.6](https://pkgs.org/search/?q=libmng)(GLIBC_2.35) is compiled on GLIBC 2.35, but RHEL 9.2 uses GLIBC 2.34, it will be usable on Fedora 38 based on GLIBC 2.37 ``` # /lib64/libc.so.6 GNU C Library (GNU libc) stable release version 2.34. ``` - [libraw_r.so.23](https://pkgs.org/search/?q=LibRaw) only available on Fedora 38, not on RHEL 9 Depending on your choices (vulnerabilities) you may target only latest stacks and up to date distributions like Fedora 38. ### Steps to Reproduce On a Docker-enabled machine, launch a RHEL 9.2 Linux distribution docker container ``` $ docker run -it --name dest almalinux:latest ``` [download your RPMs](https://imagemagick.org/script/download.php) In a shell/CMD console of the host machine copy those files to the _dest_ container ``` cd Download docker cp ImageMagick-libs-7.1.1-15.x86_64.rpm dest:/tmp docker cp ImageMagick-7.1.1-15.x86_64.rpm dest:/tmp ``` Then proceed with the _dest_ container operations: ``` # cd /tmp # dnf install ./ImageMagick-7.1.1-15.x86_64.rpm ./ImageMagick-libs-7.1.1-15.x86_64.rpm ``` You will get the given diagnostic. Finally check the availability of the missing libraries with [pkgs.org](https://pkgs.org/). ### Images _No response_
ImageMagick-7.1.1-15 RPM is not compatible with RHEL 9.2 stack
https://api.github.com/repos/ImageMagick/ImageMagick/issues/6582/comments
3
2023-08-24T14:16:16
2024-07-15T21:56:54Z
https://github.com/ImageMagick/ImageMagick/issues/6582
1,865,265,854
6,582
false
This is a GitHub Issue repo:ImageMagick owner:ImageMagick Title : ImageMagick-7.1.1-15 RPM is not compatible with RHEL 9.2 stack Issue date: --- start body --- ### ImageMagick version 7.1.1-15 ### Operating system Linux ### Operating system, version and so on RHEL 9.2 ### Description Your Linux Binary Release RPMs are not compatible with RHEL 9.2 distributions. Using EPEL repository is not a workaround. The installation fails: ``` $ docker run -it --name dest almalinux:latest # cd /tmp # dnf install ./ImageMagick-7.1.1-15.x86_64.rpm ./ImageMagick-libs-7.1.1-15.x86_64.rpm Last metadata expiration check: 0:01:00 ago on Thu Aug 24 11:40:09 2023. Error: Problem 1: conflicting requests - nothing provides libgs.so.10()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl.so.0.7()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl.so.0.7(JXL_0)(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl_threads.so.0.7()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl_threads.so.0.7(JXL_0)(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libm.so.6(GLIBC_2.35)(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libraw_r.so.23()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 Problem 2: package ImageMagick-7.1.1-15.x86_64 requires libMagickCore-7.Q16HDRI.so.10()(64bit), but none of the providers can be installed - package ImageMagick-7.1.1-15.x86_64 requires libMagickWand-7.Q16HDRI.so.10()(64bit), but none of the providers can be installed - package ImageMagick-7.1.1-15.x86_64 requires libMagickWand-7.Q16HDRI.so.10(VERS_10.0)(64bit), but none of the providers can be installed - package ImageMagick-7.1.1-15.x86_64 requires ImageMagick-libs(x86-64) = 7.1.1-15, but none of the providers can be installed - conflicting requests - nothing provides libgs.so.10()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl.so.0.7()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl.so.0.7(JXL_0)(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl_threads.so.0.7()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libjxl_threads.so.0.7(JXL_0)(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libm.so.6(GLIBC_2.35)(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 - nothing provides libraw_r.so.23()(64bit) needed by ImageMagick-libs-7.1.1-15.x86_64 (try to add '--skip-broken' to skip uninstallable packages or '--nobest' to use not only best candidate packages) ``` The following librairies are missing and are not available for the RHEL 9.2 distribution: - [libgs.so.10](https://pkgs.org/search/?q=libgs) is not available for RHEL 9.2, only on Fedora 38 - [libm.so.6](https://pkgs.org/search/?q=libmng)(GLIBC_2.35) is compiled on GLIBC 2.35, but RHEL 9.2 uses GLIBC 2.34, it will be usable on Fedora 38 based on GLIBC 2.37 ``` # /lib64/libc.so.6 GNU C Library (GNU libc) stable release version 2.34. ``` - [libraw_r.so.23](https://pkgs.org/search/?q=LibRaw) only available on Fedora 38, not on RHEL 9 Depending on your choices (vulnerabilities) you may target only latest stacks and up to date distributions like Fedora 38. ### Steps to Reproduce On a Docker-enabled machine, launch a RHEL 9.2 Linux distribution docker container ``` $ docker run -it --name dest almalinux:latest ``` [download your RPMs](https://imagemagick.org/script/download.php) In a shell/CMD console of the host machine copy those files to the _dest_ container ``` cd Download docker cp ImageMagick-libs-7.1.1-15.x86_64.rpm dest:/tmp docker cp ImageMagick-7.1.1-15.x86_64.rpm dest:/tmp ``` Then proceed with the _dest_ container operations: ``` # cd /tmp # dnf install ./ImageMagick-7.1.1-15.x86_64.rpm ./ImageMagick-libs-7.1.1-15.x86_64.rpm ``` You will get the given diagnostic. Finally check the availability of the missing libraries with [pkgs.org](https://pkgs.org/). ### Images _No response_ --- end body ---
4,174
[ -0.028800688683986664, -0.00788679625838995, -0.011151074431836605, 0.032087478786706924, 0.021881919354200363, -0.024928579106926918, 0.0026170506607741117, 0.015135744586586952, 0.008487123064696789, 0.018895292654633522, 0.01229919958859682, -0.02973119542002678, 0.002521373564377427, -...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
For some reason, download manager is not working properly in a worker. Considering both the main thread and worker thread are running on their own thread, I can't find the reason why download manager experiences degraded performance while using gpac in a worker. What happens is that when using worker and a remote source, gpac takes almost 60-70 times more to load the source. It seems that the reason lies with scheduling `fetch` promises while the same thread runs gpac. I've tried a very rough way to run `fetch` in seperate threads but I think I'm still blocked by scheduling. The correct way to achieve this seems to be through native (pthread) threading for download manager. Just how `-dm-threads` option supposed to work. I would be happy to open a PR to bring `-dm-threads` to WebAssembly version but I need some guidance on what is needed and what might be a blocking issue. @jeanlf any idea on how to proceed with this one? **My attempt at multi-threaded fetch:** ```js let workerScript = `self.onmessage = async function(event) { let {url, options} = event.data; let res = await fetch(url, options); if (res.ok) { self.postMessage({ ok: res.ok, url: res.url, status: res.status, headers: JSON.stringify([...res.headers]), redirected: res.redirected, body: res.body }, [res.body]) } else self.postMessage({ok: res.ok, status: res.status}) }`; let workerBlob = new Blob([workerScript], { type: 'application/javascript' }); fetcher._worker = new Worker(URL.createObjectURL(workerBlob)); fetcher._worker.onmessage = (event) => { if (event.data.ok) { fetcher._state = 1; fetcher._bytes = 0; fetcher._reader = event.data.body.getReader(); let final_url = null; if (event.data.redirected) final_url = event.data.url; libgpac._fetcher_set_reply(fetcher.sess, event.data.status, final_url); let headers = JSON.parse(event.data.headers); headers.forEach(([value, key]) => { libgpac._fetcher_set_header(fetcher.sess, key, value); }); libgpac._fetcher_set_header(fetcher.sess, 0, 0); } else { libgpac._fetcher_set_reply(fetcher.sess, event.data.status, null); do_log_err('fetcher for ' + url + ' failed ' + event.data.status); fetcher._state = 3; } }; fetcher._state = 0; fetcher._worker.postMessage({url, options}, [...(options.body ? [options.body] : [])]); return ret; ``` **Was placed here:** https://github.com/gpac/gpac/blob/96b474ec59b3490eed7a874b14f78ad4b3a9d2bd/src/utils/downloader.c#L7702
WebAssembly: Add threading for download manager
https://api.github.com/repos/gpac/gpac/issues/2775/comments
0
2024-03-11T06:37:13
2024-04-16T15:51:17Z
https://github.com/gpac/gpac/issues/2775
2,178,339,624
2,775
false
This is a GitHub Issue repo:gpac owner:gpac Title : WebAssembly: Add threading for download manager Issue date: --- start body --- For some reason, download manager is not working properly in a worker. Considering both the main thread and worker thread are running on their own thread, I can't find the reason why download manager experiences degraded performance while using gpac in a worker. What happens is that when using worker and a remote source, gpac takes almost 60-70 times more to load the source. It seems that the reason lies with scheduling `fetch` promises while the same thread runs gpac. I've tried a very rough way to run `fetch` in seperate threads but I think I'm still blocked by scheduling. The correct way to achieve this seems to be through native (pthread) threading for download manager. Just how `-dm-threads` option supposed to work. I would be happy to open a PR to bring `-dm-threads` to WebAssembly version but I need some guidance on what is needed and what might be a blocking issue. @jeanlf any idea on how to proceed with this one? **My attempt at multi-threaded fetch:** ```js let workerScript = `self.onmessage = async function(event) { let {url, options} = event.data; let res = await fetch(url, options); if (res.ok) { self.postMessage({ ok: res.ok, url: res.url, status: res.status, headers: JSON.stringify([...res.headers]), redirected: res.redirected, body: res.body }, [res.body]) } else self.postMessage({ok: res.ok, status: res.status}) }`; let workerBlob = new Blob([workerScript], { type: 'application/javascript' }); fetcher._worker = new Worker(URL.createObjectURL(workerBlob)); fetcher._worker.onmessage = (event) => { if (event.data.ok) { fetcher._state = 1; fetcher._bytes = 0; fetcher._reader = event.data.body.getReader(); let final_url = null; if (event.data.redirected) final_url = event.data.url; libgpac._fetcher_set_reply(fetcher.sess, event.data.status, final_url); let headers = JSON.parse(event.data.headers); headers.forEach(([value, key]) => { libgpac._fetcher_set_header(fetcher.sess, key, value); }); libgpac._fetcher_set_header(fetcher.sess, 0, 0); } else { libgpac._fetcher_set_reply(fetcher.sess, event.data.status, null); do_log_err('fetcher for ' + url + ' failed ' + event.data.status); fetcher._state = 3; } }; fetcher._state = 0; fetcher._worker.postMessage({url, options}, [...(options.body ? [options.body] : [])]); return ret; ``` **Was placed here:** https://github.com/gpac/gpac/blob/96b474ec59b3490eed7a874b14f78ad4b3a9d2bd/src/utils/downloader.c#L7702 --- end body ---
2,680
[ -0.03539044037461281, 0.01471545547246933, -0.0197122935205698, -0.01748128980398178, 0.00846558902412653, 0.00542852096259594, -0.0017687580548226833, 0.05611126869916916, -0.03429022058844566, 0.06815257668495178, 0.017420165240764618, 0.013714560307562351, 0.017252076417207718, -0.02062...
null
null
null
null
null
null
null
null
null
[ "libming", "libming" ]
Hi, developers of libming: Recently, I tested some binaries instrumented with ASAN in liming. Unfortunately, some function (all in **decompileAction** and caused by **getString or getName**) incurred crashes about address error(heap-buffer-overflow and Segment Fault): - Segment fault in 14 functions: **decompileArithmeticOp, decompileIMPLEMENTS, newVar_N(decompileNEWOBJECT), decompileNEWMETHOD, decompileNEWOBJECT, decompileGETMEMBER, decompileGETVARIABLE, decompilePUSHPARAM(decompileGETURL2), decompileIMPLEMENTS, newVar_N2(decompileINITOBJECT), decompileCAST, decompileRETURN, decompilePUSHPARAM, decompileSETTARGET**. - Heap buffer overflow in 2 functions: **decompileGETMEMBER, newVar_N**. There are all the error messages and crash inputs in my GitHub repository: https://github.com/Yhcrown/fuzzing-crashes/tree/main/validated/libming **Verification steps:** ``` git clone https://github.com/libming/libming.git cd libming export CC=clang CXX=clang++ ./autogen.sh CFLAGS+="-fsanitize=address" LDFLAGS+="-fsanitize=address" ./configure --disable-shared make -j cd util ./binary ./crash-input ``` **Environment:** Ubuntu 22.04 LTS Linux lab-pro 4.15.0-147-generic gcc (Ubuntu 11.2.0-19ubuntu1) 11.2.0 Ubuntu clang version 14.0.6 **Thanks & Best regards !**
[Bug report] Some SEGV bugs in functions of liming
https://api.github.com/repos/libming/libming/issues/248/comments
0
2022-08-19T07:33:21
2022-08-19T07:33:21Z
https://github.com/libming/libming/issues/248
1,344,061,824
248
false
This is a GitHub Issue repo:libming owner:libming Title : [Bug report] Some SEGV bugs in functions of liming Issue date: --- start body --- Hi, developers of libming: Recently, I tested some binaries instrumented with ASAN in liming. Unfortunately, some function (all in **decompileAction** and caused by **getString or getName**) incurred crashes about address error(heap-buffer-overflow and Segment Fault): - Segment fault in 14 functions: **decompileArithmeticOp, decompileIMPLEMENTS, newVar_N(decompileNEWOBJECT), decompileNEWMETHOD, decompileNEWOBJECT, decompileGETMEMBER, decompileGETVARIABLE, decompilePUSHPARAM(decompileGETURL2), decompileIMPLEMENTS, newVar_N2(decompileINITOBJECT), decompileCAST, decompileRETURN, decompilePUSHPARAM, decompileSETTARGET**. - Heap buffer overflow in 2 functions: **decompileGETMEMBER, newVar_N**. There are all the error messages and crash inputs in my GitHub repository: https://github.com/Yhcrown/fuzzing-crashes/tree/main/validated/libming **Verification steps:** ``` git clone https://github.com/libming/libming.git cd libming export CC=clang CXX=clang++ ./autogen.sh CFLAGS+="-fsanitize=address" LDFLAGS+="-fsanitize=address" ./configure --disable-shared make -j cd util ./binary ./crash-input ``` **Environment:** Ubuntu 22.04 LTS Linux lab-pro 4.15.0-147-generic gcc (Ubuntu 11.2.0-19ubuntu1) 11.2.0 Ubuntu clang version 14.0.6 **Thanks & Best regards !** --- end body ---
1,456
[ -0.012281081639230251, 0.019494032487273216, -0.008356855250895023, -0.013424985110759735, 0.04696361720561981, 0.03040878288447857, -0.039496466517448425, 0.03593765199184418, -0.017365099862217903, 0.03781238570809364, -0.027850886806845665, 0.00112603057641536, 0.0002082759456243366, 0....
null
null
null
null
null
null
null
null
null
[ "jerryscript-project", "jerryscript" ]
###### JerryScript revision 0d496966 ###### Build platform Linux-5.4.0-104-generic-x86_64-with-glibc2.29 ###### Build steps ``` ./tools/build.py --clean --debug --profile=es.next --error-messages=ON --logging=ON ``` ###### Test case ```javascript for (var {...r} = o in new Map( )); ``` ###### Output ```text ICE: Assertion '!!(flags & PARSER_PATTERN_HAS_REST_ELEMENT) == rest_found' failed at jerryscript/jerry-core/parser/js/js-parser-expr.c(parser_parse_object_initializer):3895. Error: JERRY_FATAL_FAILED_ASSERTION AddressSanitizer:DEADLYSIGNAL ================================================================= ==3270099==ERROR: AddressSanitizer: ABRT on unknown address 0x03e90031e5d3 (pc 0x7f13bff6703b bp 0x7fff3c593150 sp 0x7fff3c592ee0 T0) #0 0x7f13bff6703b in raise /build/glibc-sMfBJT/glibc-2.31/signal/../sysdeps/unix/sysv/linux/raise.c:51:1 #1 0x7f13bff46858 in abort /build/glibc-sMfBJT/glibc-2.31/stdlib/abort.c:79:7 #2 0x89ef3f in jerry_port_fatal jerryscript/jerry-port/common/jerry-port-process.c:29:5 #3 0x66b742 in jerry_fatal jerryscript/jerry-core/jrt/jrt-fatals.c:63:3 #4 0x66b4da in jerry_assert_fail jerryscript/jerry-core/jrt/jrt-fatals.c:83:3 #5 0x837a4e in parser_parse_object_initializer jerryscript/jerry-core/parser/js/js-parser-expr.c:3895:3 #6 0x8389ce in parser_parse_initializer_by_next_char jerryscript/jerry-core/parser/js/js-parser-expr.c:3944:5 #7 0x879739 in parser_parse_for_statement_start jerryscript/jerry-core/parser/js/js-parser-statm.c:1293:11 #8 0x868f9b in parser_parse_statements jerryscript/jerry-core/parser/js/js-parser-statm.c:2851:9 #9 0x69ae84 in parser_parse_source jerryscript/jerry-core/parser/js/js-parser.c:2280:5 #10 0x69629a in parser_parse_script jerryscript/jerry-core/parser/js/js-parser.c:3326:38 #11 0x4ced00 in jerry_parse_common jerryscript/jerry-core/api/jerryscript.c:412:21 #12 0x4ce473 in jerry_parse jerryscript/jerry-core/api/jerryscript.c:480:10 #13 0x89cc1f in jerryx_source_parse_script jerryscript/jerry-ext/util/sources.c:52:26 #14 0x89cd6e in jerryx_source_exec_script jerryscript/jerry-ext/util/sources.c:63:26 #15 0x4c4d84 in main jerryscript/jerry-main/main-desktop.c:156:20 #16 0x7f13bff480b2 in __libc_start_main /build/glibc-sMfBJT/glibc-2.31/csu/../csu/libc-start.c:308:16 #17 0x41c53d in _start (jerryscript/build/bin/jerry+0x41c53d) AddressSanitizer can not provide additional info. SUMMARY: AddressSanitizer: ABRT /build/glibc-sMfBJT/glibc-2.31/signal/../sysdeps/unix/sysv/linux/raise.c:51:1 in raise ==3270099==ABORTING ``` ###### Backtrace ```text bt #0 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:50 #1 0x00007ffff7c33859 in __GI_abort () at abort.c:79 #2 0x000000000089ef40 in jerry_port_fatal (code=JERRY_FATAL_FAILED_ASSERTION) at jerryscript/jerry-port/common/jerry-port-process.c:29 #3 0x000000000066b743 in jerry_fatal (code=JERRY_FATAL_FAILED_ASSERTION) at jerryscript/jerry-core/jrt/jrt-fatals.c:63 #4 0x000000000066b4db in jerry_assert_fail (assertion=0x929ea0 <str> "!!(flags & PARSER_PATTERN_HAS_REST_ELEMENT) == rest_found", file=0x929160 <str> "jerryscript/jerry-core/parser/js/js-parser-expr.c", function=0x929dc0 <__func__.parser_parse_object_initializer> "parser_parse_object_initializer", line=3895) at jerryscript/jerry-core/jrt/jrt-fatals.c:83 #5 0x0000000000837a4f in parser_parse_object_initializer (context_p=0x7fffffffcd20, flags=(PARSER_PATTERN_BINDING | PARSER_PATTERN_TARGET_ON_STACK | PARSER_PATTERN_REST_ELEMENT)) at jerryscript/jerry-core/parser/js/js-parser-expr.c:3895 #6 0x00000000008389cf in parser_parse_initializer_by_next_char (context_p=0x7fffffffcd20, flags=(PARSER_PATTERN_BINDING | PARSER_PATTERN_TARGET_ON_STACK)) at jerryscript/jerry-core/parser/js/js-parser-expr.c:3944 #7 0x000000000087973a in parser_parse_for_statement_start (context_p=0x7fffffffcd20) at jerryscript/jerry-core/parser/js/js-parser-statm.c:1293 #8 0x0000000000868f9c in parser_parse_statements (context_p=0x7fffffffcd20) at jerryscript/jerry-core/parser/js/js-parser-statm.c:2851 #9 0x000000000069ae85 in parser_parse_source (source_p=0x7fffffffd3a0, parse_opts=0, options_p=0x7fffffffd4d0) at jerryscript/jerry-core/parser/js/js-parser.c:2280 #10 0x000000000069629b in parser_parse_script (source_p=0x7fffffffd3a0, parse_opts=0, options_p=0x7fffffffd4d0) at jerryscript/jerry-core/parser/js/js-parser.c:3326 #11 0x00000000004ced01 in jerry_parse_common (source_p=0x7fffffffd3a0, options_p=0x7fffffffd4d0, parse_opts=0) at jerryscript/jerry-core/api/jerryscript.c:412 #12 0x00000000004ce474 in jerry_parse (source_p=0x607000000020 "for ( var { ... r } = o in new Map ( \"\\uFFA2\" ) ) ; ", source_size=77, options_p=0x7fffffffd4d0) at jerryscript/jerry-core/api/jerryscript.c:480 #13 0x000000000089cc20 in jerryx_source_parse_script (path_p=0x7fffffffde36 "/run/user/1001/fuzzinator/3261026/3270079-FileWriterDecorator-0e4d2c28b2ac47759841127d665b5fff/0.js") at jerryscript/jerry-ext/util/sources.c:52 #14 0x000000000089cd6f in jerryx_source_exec_script (path_p=0x7fffffffde36 "/run/user/1001/fuzzinator/3261026/3270079-FileWriterDecorator-0e4d2c28b2ac47759841127d665b5fff/0.js") at jerryscript/jerry-ext/util/sources.c:63 #15 0x00000000004c4d85 in main (argc=2, argv=0x7fffffffdb08) at jerryscript/jerry-main/main-desktop.c:156 ``` <sup>Found by [Fuzzinator](http://fuzzinator.readthedocs.io/) with [grammarinator](https://github.com/renatahodovan/grammarinator). </sup>
Assertion !!(flags & PARSER_PATTERN_HAS_REST_ELEMENT) == rest_found in parser_parse_object_initializer
https://api.github.com/repos/jerryscript-project/jerryscript/issues/4998/comments
0
2022-04-25T16:38:42
2022-04-25T16:38:42Z
https://github.com/jerryscript-project/jerryscript/issues/4998
1,214,770,024
4,998
false
This is a GitHub Issue repo:jerryscript owner:jerryscript-project Title : Assertion !!(flags & PARSER_PATTERN_HAS_REST_ELEMENT) == rest_found in parser_parse_object_initializer Issue date: --- start body --- ###### JerryScript revision 0d496966 ###### Build platform Linux-5.4.0-104-generic-x86_64-with-glibc2.29 ###### Build steps ``` ./tools/build.py --clean --debug --profile=es.next --error-messages=ON --logging=ON ``` ###### Test case ```javascript for (var {...r} = o in new Map( )); ``` ###### Output ```text ICE: Assertion '!!(flags & PARSER_PATTERN_HAS_REST_ELEMENT) == rest_found' failed at jerryscript/jerry-core/parser/js/js-parser-expr.c(parser_parse_object_initializer):3895. Error: JERRY_FATAL_FAILED_ASSERTION AddressSanitizer:DEADLYSIGNAL ================================================================= ==3270099==ERROR: AddressSanitizer: ABRT on unknown address 0x03e90031e5d3 (pc 0x7f13bff6703b bp 0x7fff3c593150 sp 0x7fff3c592ee0 T0) #0 0x7f13bff6703b in raise /build/glibc-sMfBJT/glibc-2.31/signal/../sysdeps/unix/sysv/linux/raise.c:51:1 #1 0x7f13bff46858 in abort /build/glibc-sMfBJT/glibc-2.31/stdlib/abort.c:79:7 #2 0x89ef3f in jerry_port_fatal jerryscript/jerry-port/common/jerry-port-process.c:29:5 #3 0x66b742 in jerry_fatal jerryscript/jerry-core/jrt/jrt-fatals.c:63:3 #4 0x66b4da in jerry_assert_fail jerryscript/jerry-core/jrt/jrt-fatals.c:83:3 #5 0x837a4e in parser_parse_object_initializer jerryscript/jerry-core/parser/js/js-parser-expr.c:3895:3 #6 0x8389ce in parser_parse_initializer_by_next_char jerryscript/jerry-core/parser/js/js-parser-expr.c:3944:5 #7 0x879739 in parser_parse_for_statement_start jerryscript/jerry-core/parser/js/js-parser-statm.c:1293:11 #8 0x868f9b in parser_parse_statements jerryscript/jerry-core/parser/js/js-parser-statm.c:2851:9 #9 0x69ae84 in parser_parse_source jerryscript/jerry-core/parser/js/js-parser.c:2280:5 #10 0x69629a in parser_parse_script jerryscript/jerry-core/parser/js/js-parser.c:3326:38 #11 0x4ced00 in jerry_parse_common jerryscript/jerry-core/api/jerryscript.c:412:21 #12 0x4ce473 in jerry_parse jerryscript/jerry-core/api/jerryscript.c:480:10 #13 0x89cc1f in jerryx_source_parse_script jerryscript/jerry-ext/util/sources.c:52:26 #14 0x89cd6e in jerryx_source_exec_script jerryscript/jerry-ext/util/sources.c:63:26 #15 0x4c4d84 in main jerryscript/jerry-main/main-desktop.c:156:20 #16 0x7f13bff480b2 in __libc_start_main /build/glibc-sMfBJT/glibc-2.31/csu/../csu/libc-start.c:308:16 #17 0x41c53d in _start (jerryscript/build/bin/jerry+0x41c53d) AddressSanitizer can not provide additional info. SUMMARY: AddressSanitizer: ABRT /build/glibc-sMfBJT/glibc-2.31/signal/../sysdeps/unix/sysv/linux/raise.c:51:1 in raise ==3270099==ABORTING ``` ###### Backtrace ```text bt #0 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:50 #1 0x00007ffff7c33859 in __GI_abort () at abort.c:79 #2 0x000000000089ef40 in jerry_port_fatal (code=JERRY_FATAL_FAILED_ASSERTION) at jerryscript/jerry-port/common/jerry-port-process.c:29 #3 0x000000000066b743 in jerry_fatal (code=JERRY_FATAL_FAILED_ASSERTION) at jerryscript/jerry-core/jrt/jrt-fatals.c:63 #4 0x000000000066b4db in jerry_assert_fail (assertion=0x929ea0 <str> "!!(flags & PARSER_PATTERN_HAS_REST_ELEMENT) == rest_found", file=0x929160 <str> "jerryscript/jerry-core/parser/js/js-parser-expr.c", function=0x929dc0 <__func__.parser_parse_object_initializer> "parser_parse_object_initializer", line=3895) at jerryscript/jerry-core/jrt/jrt-fatals.c:83 #5 0x0000000000837a4f in parser_parse_object_initializer (context_p=0x7fffffffcd20, flags=(PARSER_PATTERN_BINDING | PARSER_PATTERN_TARGET_ON_STACK | PARSER_PATTERN_REST_ELEMENT)) at jerryscript/jerry-core/parser/js/js-parser-expr.c:3895 #6 0x00000000008389cf in parser_parse_initializer_by_next_char (context_p=0x7fffffffcd20, flags=(PARSER_PATTERN_BINDING | PARSER_PATTERN_TARGET_ON_STACK)) at jerryscript/jerry-core/parser/js/js-parser-expr.c:3944 #7 0x000000000087973a in parser_parse_for_statement_start (context_p=0x7fffffffcd20) at jerryscript/jerry-core/parser/js/js-parser-statm.c:1293 #8 0x0000000000868f9c in parser_parse_statements (context_p=0x7fffffffcd20) at jerryscript/jerry-core/parser/js/js-parser-statm.c:2851 #9 0x000000000069ae85 in parser_parse_source (source_p=0x7fffffffd3a0, parse_opts=0, options_p=0x7fffffffd4d0) at jerryscript/jerry-core/parser/js/js-parser.c:2280 #10 0x000000000069629b in parser_parse_script (source_p=0x7fffffffd3a0, parse_opts=0, options_p=0x7fffffffd4d0) at jerryscript/jerry-core/parser/js/js-parser.c:3326 #11 0x00000000004ced01 in jerry_parse_common (source_p=0x7fffffffd3a0, options_p=0x7fffffffd4d0, parse_opts=0) at jerryscript/jerry-core/api/jerryscript.c:412 #12 0x00000000004ce474 in jerry_parse (source_p=0x607000000020 "for ( var { ... r } = o in new Map ( \"\\uFFA2\" ) ) ; ", source_size=77, options_p=0x7fffffffd4d0) at jerryscript/jerry-core/api/jerryscript.c:480 #13 0x000000000089cc20 in jerryx_source_parse_script (path_p=0x7fffffffde36 "/run/user/1001/fuzzinator/3261026/3270079-FileWriterDecorator-0e4d2c28b2ac47759841127d665b5fff/0.js") at jerryscript/jerry-ext/util/sources.c:52 #14 0x000000000089cd6f in jerryx_source_exec_script (path_p=0x7fffffffde36 "/run/user/1001/fuzzinator/3261026/3270079-FileWriterDecorator-0e4d2c28b2ac47759841127d665b5fff/0.js") at jerryscript/jerry-ext/util/sources.c:63 #15 0x00000000004c4d85 in main (argc=2, argv=0x7fffffffdb08) at jerryscript/jerry-main/main-desktop.c:156 ``` <sup>Found by [Fuzzinator](http://fuzzinator.readthedocs.io/) with [grammarinator](https://github.com/renatahodovan/grammarinator). </sup> --- end body ---
6,356
[ -0.04217992722988129, 0.010410364717245102, 0.0036365229170769453, 0.022136982530355453, 0.018487371504306793, -0.009640057571232319, -0.00677197240293026, 0.01273624412715435, -0.0379619337618351, 0.001127415569499135, -0.004988299682736397, 0.025038722902536392, 0.02506863698363304, 0.00...
null
null
null
null
null
null
null
null
null
[ "libming", "libming" ]
The "-h" option of jasper misses the option "p" for version 04aee52. + For the option "p", the following code appears during option parsing and calls the makeswf_set_dopreprocess( ) in ./util/makeswf.c:249-250 for version 3.0.6. ``` makeswf_set_dopreprocess(0); break; ``` But it does not appear in the document provided by "-h". It may prevent users from using the relevant function.
[Help document bug] For makeswf, the help document missesa an option
https://api.github.com/repos/libming/libming/issues/250/comments
0
2022-10-01T02:44:44
2022-10-01T02:44:44Z
https://github.com/libming/libming/issues/250
1,393,195,708
250
false
This is a GitHub Issue repo:libming owner:libming Title : [Help document bug] For makeswf, the help document missesa an option Issue date: --- start body --- The "-h" option of jasper misses the option "p" for version 04aee52. + For the option "p", the following code appears during option parsing and calls the makeswf_set_dopreprocess( ) in ./util/makeswf.c:249-250 for version 3.0.6. ``` makeswf_set_dopreprocess(0); break; ``` But it does not appear in the document provided by "-h". It may prevent users from using the relevant function. --- end body ---
586
[ -0.0041550081223249435, 0.026660680770874023, -0.01699301414191723, 0.047771405428647995, 0.047025442123413086, 0.03550779074430466, 0.009682586416602135, 0.045533519238233566, -0.02715301513671875, 0.04138597100973129, -0.0038491636514663696, 0.002976388204842806, 0.03944646939635277, -0....
null
null
null
null
null
null
null
null
null
[ "LibreDWG", "libredwg" ]
Hello, I need to use files like out_json.h and hash.h when installing and compiling, but the current installation method does not include these related files. Do you have a good solution for this? Thank you
The necessary files are missing when installing
https://api.github.com/repos/LibreDWG/libredwg/issues/978/comments
3
2024-05-28T10:02:02
2024-05-29T21:10:07Z
https://github.com/LibreDWG/libredwg/issues/978
2,320,645,675
978
false
This is a GitHub Issue repo:libredwg owner:LibreDWG Title : The necessary files are missing when installing Issue date: --- start body --- Hello, I need to use files like out_json.h and hash.h when installing and compiling, but the current installation method does not include these related files. Do you have a good solution for this? Thank you --- end body ---
364
[ -0.004111120011657476, 0.026856491342186928, -0.008518918417394161, -0.0030197659507393837, 0.046084098517894745, 0.03260640799999237, 0.020244793966412544, 0.03856823965907097, -0.032295603305101395, 0.0030586167704313993, -0.013753179460763931, -0.010164779610931873, -0.013901518657803535,...
null
null
null
null
null
null
null
null
null
[ "llvm", "llvm-project" ]
https://github.com/ARM-software/abi-aa/blob/main/clibabi32/clibabi32.rst#setjmph > In deference to VFP, XScale Wireless MMX, and other co-processors manipulating 8-byte aligned types, a jmp_buf must be 8-byte aligned.
[libc][arm][setjmp/longjmp] align jmp_buf to 8B
https://api.github.com/repos/llvm/llvm-project/issues/94069/comments
1
2024-05-31T22:36:21
2024-05-31T22:36:38Z
https://github.com/llvm/llvm-project/issues/94069
2,328,670,993
94,069
false
This is a GitHub Issue repo:llvm-project owner:llvm Title : [libc][arm][setjmp/longjmp] align jmp_buf to 8B Issue date: --- start body --- https://github.com/ARM-software/abi-aa/blob/main/clibabi32/clibabi32.rst#setjmph > In deference to VFP, XScale Wireless MMX, and other co-processors manipulating 8-byte aligned types, a jmp_buf must be 8-byte aligned. --- end body ---
382
[ 0.004070764873176813, 0.02684325911104679, -0.01552970428019762, 0.052629176527261734, 0.02108074724674225, 0.016732430085539818, -0.001828870503231883, 0.041342057287693024, -0.009258347563445568, 0.01378509122878313, 0.00945659913122654, 0.0160980261862278, 0.02082962915301323, 0.0074542...
null
null
null
null
null
null
null
null
null
[ "LibreDWG", "libredwg" ]
geojson parsing, return error code "0x941", what is the meaning of this code?
error code 0x941
https://api.github.com/repos/LibreDWG/libredwg/issues/919/comments
11
2024-01-09T02:52:49
2024-01-11T10:21:20Z
https://github.com/LibreDWG/libredwg/issues/919
2,071,536,689
919
false
This is a GitHub Issue repo:libredwg owner:LibreDWG Title : error code 0x941 Issue date: --- start body --- geojson parsing, return error code "0x941", what is the meaning of this code? --- end body ---
204
[ -0.03459795191884041, 0.03314824029803276, 0.001320771872997284, 0.014985010959208012, 0.019598999992012978, 0.02287479303777218, 0.030220933258533478, 0.04965266212821007, -0.01395348459482193, -0.0013913408620283008, 0.004073832184076309, -0.019682638347148895, 0.004843992181122303, -0.0...
null
null
null
null
null
null
null
null
null
[ "kubernetes", "kubernetes" ]
### What happened? When kubelet evicts a pod, the ready condition doesn’t get `NotReady` during the pod termination even if a `readinessProbe` is configured. ### What did you expect to happen? A readiness probe works during a pod termination so that the pod gets `NotReady` as early as possible. ### How can we reproduce it (as minimally and precisely as possible)? Use this `readiness.yaml`: ``` --- apiVersion: v1 kind: ConfigMap metadata: name: readiness-script-configmap data: readiness-script.sh: | #!/bin/sh handler() { rm /tmp/ready sleep 20 } touch /tmp/ready trap handler SIGTERM while true; do sleep 1 done --- apiVersion: v1 kind: Pod metadata: creationTimestamp: null labels: run: readiness-container name: readiness-test spec: containers: - command: - sh - /script/readiness-script.sh image: busybox name: readiness readinessProbe: exec: command: - cat - /tmp/ready initialDelaySeconds: 3 periodSeconds: 3 volumeMounts: - name: readiness-script mountPath: /script - command: - sh - -c - for i in `seq 100`; do dd if=/dev/random of=file${i} bs=1048576 count=1 2>/dev/null; sleep .1; done; while true; do sleep 5; done name: disk-consumer image: busybox resources: limits: ephemeral-storage: "100Mi" requests: ephemeral-storage: "100Mi" volumes: - name: readiness-script configMap: name: readiness-script-configmap ``` Create resources and wait an eviction happens: ``` $ kubectl create -f readiness.yaml; kubectl get pods readiness-test -w configmap/readiness-script-configmap created pod/readiness-test created NAME READY STATUS RESTARTS AGE readiness-test 0/2 ContainerCreating 0 0s readiness-test 1/2 Running 0 3s readiness-test 2/2 Running 0 7s readiness-test 0/2 Error 0 46s readiness-test 0/2 Error 0 47s ``` When deleting this pod, the readiness probe works during termination: ``` $ kubectl create -f readiness.yaml; (sleep 15; kubectl delete pod readiness-test) & kubectl get pods -w configmap/readiness-script-configmap created pod/readiness-test created [1] 137999 NAME READY STATUS RESTARTS AGE readiness-test 0/2 ContainerCreating 0 0s readiness-test 1/2 Running 0 2s readiness-test 2/2 Running 0 6s pod "readiness-test" deleted readiness-test 2/2 Terminating 0 15s readiness-test 1/2 Terminating 0 21s readiness-test 0/2 Terminating 0 45s readiness-test 0/2 Terminating 0 45s readiness-test 0/2 Terminating 0 45s readiness-test 0/2 Terminating 0 45s ``` ### Anything else we need to know? I guess this issue is caused as follows: At eviction, a pod phase is set to `PodFailed` internally in `podStatusFn` before stopping containers in the pod: https://github.com/kubernetes/kubernetes/blob/0d8f996aa9a1667d5994f10e03de5be21bf205e0/pkg/kubelet/kubelet.go#L2026-L2029 https://github.com/kubernetes/kubernetes/blob/0d8f996aa9a1667d5994f10e03de5be21bf205e0/pkg/kubelet/eviction/eviction_manager.go#L605-L608 Because the internal pod phase is `PodFailed`, the probe worker finishes working without probing containers at termination: https://github.com/kubernetes/kubernetes/blob/dd68c5f2409fec7176e6172d6f9d97bd6447c4da/pkg/kubelet/prober/worker.go#L203-L216 ### Kubernetes version <details> ```console $ kubectl version # paste output here ``` Client Version: v1.28.2 Kustomize Version: v5.0.4-0.20230601165947-6ce0bf390ce3 Server Version: v1.29.2 </details> ### Cloud provider <details> None. I reproduced on my laptop. </details> ### OS version <details> ```console # On Linux: $ cat /etc/os-release # paste output here $ uname -a # paste output here # On Windows: C:\> wmic os get Caption, Version, BuildNumber, OSArchitecture # paste output here ``` </details> ### Install tools <details> kind </details> ### Container runtime (CRI) and version (if applicable) <details> </details> ### Related plugins (CNI, CSI, ...) and versions (if applicable) <details> </details>
Readiness probe stops too early at eviction
https://api.github.com/repos/kubernetes/kubernetes/issues/124648/comments
6
2024-04-30T22:55:00
2024-06-17T21:31:53Z
https://github.com/kubernetes/kubernetes/issues/124648
2,272,613,226
124,648
false
This is a GitHub Issue repo:kubernetes owner:kubernetes Title : Readiness probe stops too early at eviction Issue date: --- start body --- ### What happened? When kubelet evicts a pod, the ready condition doesn’t get `NotReady` during the pod termination even if a `readinessProbe` is configured. ### What did you expect to happen? A readiness probe works during a pod termination so that the pod gets `NotReady` as early as possible. ### How can we reproduce it (as minimally and precisely as possible)? Use this `readiness.yaml`: ``` --- apiVersion: v1 kind: ConfigMap metadata: name: readiness-script-configmap data: readiness-script.sh: | #!/bin/sh handler() { rm /tmp/ready sleep 20 } touch /tmp/ready trap handler SIGTERM while true; do sleep 1 done --- apiVersion: v1 kind: Pod metadata: creationTimestamp: null labels: run: readiness-container name: readiness-test spec: containers: - command: - sh - /script/readiness-script.sh image: busybox name: readiness readinessProbe: exec: command: - cat - /tmp/ready initialDelaySeconds: 3 periodSeconds: 3 volumeMounts: - name: readiness-script mountPath: /script - command: - sh - -c - for i in `seq 100`; do dd if=/dev/random of=file${i} bs=1048576 count=1 2>/dev/null; sleep .1; done; while true; do sleep 5; done name: disk-consumer image: busybox resources: limits: ephemeral-storage: "100Mi" requests: ephemeral-storage: "100Mi" volumes: - name: readiness-script configMap: name: readiness-script-configmap ``` Create resources and wait an eviction happens: ``` $ kubectl create -f readiness.yaml; kubectl get pods readiness-test -w configmap/readiness-script-configmap created pod/readiness-test created NAME READY STATUS RESTARTS AGE readiness-test 0/2 ContainerCreating 0 0s readiness-test 1/2 Running 0 3s readiness-test 2/2 Running 0 7s readiness-test 0/2 Error 0 46s readiness-test 0/2 Error 0 47s ``` When deleting this pod, the readiness probe works during termination: ``` $ kubectl create -f readiness.yaml; (sleep 15; kubectl delete pod readiness-test) & kubectl get pods -w configmap/readiness-script-configmap created pod/readiness-test created [1] 137999 NAME READY STATUS RESTARTS AGE readiness-test 0/2 ContainerCreating 0 0s readiness-test 1/2 Running 0 2s readiness-test 2/2 Running 0 6s pod "readiness-test" deleted readiness-test 2/2 Terminating 0 15s readiness-test 1/2 Terminating 0 21s readiness-test 0/2 Terminating 0 45s readiness-test 0/2 Terminating 0 45s readiness-test 0/2 Terminating 0 45s readiness-test 0/2 Terminating 0 45s ``` ### Anything else we need to know? I guess this issue is caused as follows: At eviction, a pod phase is set to `PodFailed` internally in `podStatusFn` before stopping containers in the pod: https://github.com/kubernetes/kubernetes/blob/0d8f996aa9a1667d5994f10e03de5be21bf205e0/pkg/kubelet/kubelet.go#L2026-L2029 https://github.com/kubernetes/kubernetes/blob/0d8f996aa9a1667d5994f10e03de5be21bf205e0/pkg/kubelet/eviction/eviction_manager.go#L605-L608 Because the internal pod phase is `PodFailed`, the probe worker finishes working without probing containers at termination: https://github.com/kubernetes/kubernetes/blob/dd68c5f2409fec7176e6172d6f9d97bd6447c4da/pkg/kubelet/prober/worker.go#L203-L216 ### Kubernetes version <details> ```console $ kubectl version # paste output here ``` Client Version: v1.28.2 Kustomize Version: v5.0.4-0.20230601165947-6ce0bf390ce3 Server Version: v1.29.2 </details> ### Cloud provider <details> None. I reproduced on my laptop. </details> ### OS version <details> ```console # On Linux: $ cat /etc/os-release # paste output here $ uname -a # paste output here # On Windows: C:\> wmic os get Caption, Version, BuildNumber, OSArchitecture # paste output here ``` </details> ### Install tools <details> kind </details> ### Container runtime (CRI) and version (if applicable) <details> </details> ### Related plugins (CNI, CSI, ...) and versions (if applicable) <details> </details> --- end body ---
4,688
[ -0.019050490111112595, -0.037419673055410385, 0.0009441719157621264, -0.009394223801791668, -0.02436995320022106, -0.013770334422588348, -0.005830445792526007, 0.04142892360687256, -0.030527953058481216, -0.011523319408297539, -0.022640474140644073, 0.01832987368106842, -0.002012814162299037...
null
null
null
null
null
null
null
null
null
[ "LibreDWG", "libredwg" ]
[2.zip](https://github.com/LibreDWG/libredwg/files/11244857/2.zip) ``` (gdb) (gdb) r -v9 /home/user/out/default/crashes/2* The program being debugged has been started already. Start it from the beginning? (y or n) y Starting program: /home/user/libredwg/dwgread -v9 /home/user/out/default/crashes/2* Reading DWG file /home/user/out/default/crashes/2.dwg Program received signal SIGSEGV, Segmentation fault. 0xb6c8eb26 in acds_private (dat=0xbfffe330, dwg=0xbfffe578) at src/acds.spec:215 215 SUB_FIELD_RL (schidx.props[rcount2],index, 0); (gdb) ```
Segmentation fault in acds_private
https://api.github.com/repos/LibreDWG/libredwg/issues/695/comments
2
2023-04-16T22:53:18
2023-04-18T10:21:35Z
https://github.com/LibreDWG/libredwg/issues/695
1,670,181,746
695
false
This is a GitHub Issue repo:libredwg owner:LibreDWG Title : Segmentation fault in acds_private Issue date: --- start body --- [2.zip](https://github.com/LibreDWG/libredwg/files/11244857/2.zip) ``` (gdb) (gdb) r -v9 /home/user/out/default/crashes/2* The program being debugged has been started already. Start it from the beginning? (y or n) y Starting program: /home/user/libredwg/dwgread -v9 /home/user/out/default/crashes/2* Reading DWG file /home/user/out/default/crashes/2.dwg Program received signal SIGSEGV, Segmentation fault. 0xb6c8eb26 in acds_private (dat=0xbfffe330, dwg=0xbfffe578) at src/acds.spec:215 215 SUB_FIELD_RL (schidx.props[rcount2],index, 0); (gdb) ``` --- end body ---
731
[ -0.008513286709785461, 0.006546145770698786, -0.0017345271771773696, 0.010425479151308537, 0.050083328038454056, 0.015238928608596325, 0.031708694994449615, 0.05014193803071976, -0.01840393804013729, 0.024909788742661476, -0.025744998827576637, -0.009429086931049824, 0.028660908341407776, ...
null
null
null
null
null
null
null
null
null
[ "ChurchCRM", "CRM" ]
**Description** No QR code is displayed to complete 2FA setup after clicking "Begin Two Factor Authentication Enrollment" on page .../v2/user/current/enroll2fa. - ChurchCRM version: 5.6.0 - PHP version the server running: 8.2.17 - DB Server and Version the server is running: Percona Server MySQL 5.7.23-23 - Webhost: HostGator.com **Steps To Reproduce** Steps to reproduce the behavior: 📋 Cypress Recorder Results: ``` cy.visit('.../v2/dashboard'); cy.get('.dropdown:nth-child(6) > .nav-link').click(); cy.url().should('contains', '.../v2/user/current/enroll2fa'); ``` **Expected behavior** A QR code should display to enable the setup of two-factor authentication. **Screenshots and/or logs** 📋 Relevant screenshots: ![image](https://github.com/ChurchCRM/CRM/assets/134420868/8813d2b6-c29b-4e21-b19c-80672b1a5e80) 📋 Logs: ``` [2024-03-30T17:10:07.939728-05:00] slim-app.ERROR: Slim Application Error Type: Error Code: 0 Message: Call to undefined method Endroid\QrCode\QrCode::writeDataUri() File: .../api/routes/users/user-current.php Line: 25 Trace: #0 .../vendor/slim/slim/Slim/Handlers/Strategies/RequestResponse.php(38): refresh2fasecret(Object(Laminas\Diactoros\ServerRequest), Object(Laminas\Diactoros\Response), Array) #1 .../vendor/slim/slim/Slim/Routing/Route.php(358): Slim\Handlers\Strategies\RequestResponse->__invoke('refresh2fasecre...', Object(Laminas\Diactoros\ServerRequest), Object(Laminas\Diactoros\Response), Array) #2 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(65): Slim\Routing\Route->handle(Object(Laminas\Diactoros\ServerRequest)) #3 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(65): Slim\MiddlewareDispatcher->handle(Object(Laminas\Diactoros\ServerRequest)) #4 .../vendor/slim/slim/Slim/Routing/Route.php(315): Slim\MiddlewareDispatcher->handle(Object(Laminas\Diactoros\ServerRequest)) #5 .../vendor/slim/slim/Slim/Routing/RouteRunner.php(68): Slim\Routing\Route->run(Object(Laminas\Diactoros\ServerRequest)) #6 .../ChurchCRM/Slim/Middleware/VersionMiddleware.php(14): Slim\Routing\RouteRunner->handle(Object(Laminas\Diactoros\ServerRequest)) #7 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(168): ChurchCRM\Slim\Middleware\VersionMiddleware->__invoke(Object(Laminas\Diactoros\ServerRequest), Object(Slim\Routing\RouteRunner)) #8 .../ChurchCRM/Slim/Middleware/AuthMiddleware.php(36): Psr\Http\Server\RequestHandlerInterface@anonymous->handle(Object(Laminas\Diactoros\ServerRequest)) #9 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(168): ChurchCRM\Slim\Middleware\AuthMiddleware->__invoke(Object(Laminas\Diactoros\ServerRequest), Object(Psr\Http\Server\RequestHandlerInterface@anonymous)) #10 .../vendor/slim/slim/Slim/Middleware/BodyParsingMiddleware.php(64): Psr\Http\Server\RequestHandlerInterface@anonymous->handle(Object(Laminas\Diactoros\ServerRequest)) #11 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(121): Slim\Middleware\BodyParsingMiddleware->process(Object(Laminas\Diactoros\ServerRequest), Object(Psr\Http\Server\RequestHandlerInterface@anonymous)) #12 .../vendor/slim/slim/Slim/Middleware/ErrorMiddleware.php(76): Psr\Http\Server\RequestHandlerInterface@anonymous->handle(Object(Laminas\Diactoros\ServerRequest)) #13 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(121): Slim\Middleware\ErrorMiddleware->process(Object(Laminas\Diactoros\ServerRequest), Object(Psr\Http\Server\RequestHandlerInterface@anonymous)) #14 .../vendor/slim/http-cache/src/Cache.php(67): Psr\Http\Server\RequestHandlerInterface@anonymous->handle(Object(Laminas\Diactoros\ServerRequest)) #15 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(121): Slim\HttpCache\Cache->process(Object(Laminas\Diactoros\ServerRequest), Object(Psr\Http\Server\RequestHandlerInterface@anonymous)) #16 .../vendor/slim/http-cache/src/Cache.php(67): Psr\Http\Server\RequestHandlerInterface@anonymous->handle(Object(Laminas\Diactoros\ServerRequest)) #17 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(121): Slim\HttpCache\Cache->process(Object(Laminas\Diactoros\ServerRequest), Object(Psr\Http\Server\RequestHandlerInterface@anonymous)) #18 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(65): Psr\Http\Server\RequestHandlerInterface@anonymous->handle(Object(Laminas\Diactoros\ServerRequest)) #19 .../vendor/slim/slim/Slim/App.php(199): Slim\MiddlewareDispatcher->handle(Object(Laminas\Diactoros\ServerRequest)) #20 .../vendor/slim/slim/Slim/App.php(183): Slim\App->handle(Object(Laminas\Diactoros\ServerRequest)) #21 .../api/index.php(76): Slim\App->run() #22 {main} [] [] ``` **Debugging Steps** Developer tools log: ``` debug:1052 [Report Only] Refused to load the font 'data:application/x-font-ttf;charset=utf-8;base64,AAEAAAALAIAAAwAwT1MvMg8SBfAAAAC8AAAAYGNtYXAXVtKNAAABHAAAAFRnYXNwAAAAEAAAAXAAAAAIZ2x5ZgYydxIAAAF4AAAFNGhlYWQUJ7cIAAAGrAAAADZoaGVhB20DzAAABuQAAAAkaG10eCIABhQAAAcIAAAALGxvY2ED4AU6AAAHNAAAABhtYXhwAA8AjAAAB0wAAAAgbmFtZXsr690AAAdsAAABhnBvc3QAAwAAAAAI9AAAACAAAwPAAZAABQAAApkCzAAAAI8CmQLMAAAB6wAzAQkAAAAAAAAAAAAAAAAAAAABEAAAAAAAAAAAAAAAAAAAAABAAADpBgPA/8AAQAPAAEAAAAABAAAAAAAAAAAAAAAgAAAAAAADAAAAAwAAABwAAQADAAAAHAADAAEAAAAcAAQAOAAAAAoACAACAAIAAQAg6Qb//f//AAAAAAAg6QD//f...ABAAAAAAAEAAcAdQABAAAAAAAFAAsAFQABAAAAAAAGAAcASwABAAAAAAAKABoAigADAAEECQABAA4ABwADAAEECQACAA4AZwADAAEECQADAA4APQADAAEECQAEAA4AfAADAAEECQAFABYAIAADAAEECQAGAA4AUgADAAEECQAKADQApGZjaWNvbnMAZgBjAGkAYwBvAG4Ac1ZlcnNpb24gMS4wAFYAZQByAHMAaQBvAG4AIAAxAC4AMGZjaWNvbnMAZgBjAGkAYwBvAG4Ac2ZjaWNvbnMAZgBjAGkAYwBvAG4Ac1JlZ3VsYXIAUgBlAGcAdQBsAGEAcmZjaWNvbnMAZgBjAGkAYwBvAG4Ac0ZvbnQgZ2VuZXJhdGVkIGJ5IEljb01vb24uAEYAbwBuAHQAIABnAGUAbgBlAHIAYQB0AGUAZAAgAGIAeQAgAEkAYwBvAE0AbwBvAG4ALgAAAAMAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA=' because it violates the following Content Security Policy directive: "font-src 'self' fonts.gstatic.com". /api/system/background/csp-report:1 Failed to load resource: the server responded with a status of 406 () ``` **Desktop (please complete the following information):** - OS and Version: Windows 11 - Browser (and Version): Chrome 123.0.6312.86
User can't setup 2FA (no QR code displays).
https://api.github.com/repos/ChurchCRM/CRM/issues/6918/comments
1
2024-03-30T22:45:12
2024-05-07T14:09:44Z
https://github.com/ChurchCRM/CRM/issues/6918
2,216,689,985
6,918
false
This is a GitHub Issue repo:CRM owner:ChurchCRM Title : User can't setup 2FA (no QR code displays). Issue date: --- start body --- **Description** No QR code is displayed to complete 2FA setup after clicking "Begin Two Factor Authentication Enrollment" on page .../v2/user/current/enroll2fa. - ChurchCRM version: 5.6.0 - PHP version the server running: 8.2.17 - DB Server and Version the server is running: Percona Server MySQL 5.7.23-23 - Webhost: HostGator.com **Steps To Reproduce** Steps to reproduce the behavior: 📋 Cypress Recorder Results: ``` cy.visit('.../v2/dashboard'); cy.get('.dropdown:nth-child(6) > .nav-link').click(); cy.url().should('contains', '.../v2/user/current/enroll2fa'); ``` **Expected behavior** A QR code should display to enable the setup of two-factor authentication. **Screenshots and/or logs** 📋 Relevant screenshots: ![image](https://github.com/ChurchCRM/CRM/assets/134420868/8813d2b6-c29b-4e21-b19c-80672b1a5e80) 📋 Logs: ``` [2024-03-30T17:10:07.939728-05:00] slim-app.ERROR: Slim Application Error Type: Error Code: 0 Message: Call to undefined method Endroid\QrCode\QrCode::writeDataUri() File: .../api/routes/users/user-current.php Line: 25 Trace: #0 .../vendor/slim/slim/Slim/Handlers/Strategies/RequestResponse.php(38): refresh2fasecret(Object(Laminas\Diactoros\ServerRequest), Object(Laminas\Diactoros\Response), Array) #1 .../vendor/slim/slim/Slim/Routing/Route.php(358): Slim\Handlers\Strategies\RequestResponse->__invoke('refresh2fasecre...', Object(Laminas\Diactoros\ServerRequest), Object(Laminas\Diactoros\Response), Array) #2 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(65): Slim\Routing\Route->handle(Object(Laminas\Diactoros\ServerRequest)) #3 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(65): Slim\MiddlewareDispatcher->handle(Object(Laminas\Diactoros\ServerRequest)) #4 .../vendor/slim/slim/Slim/Routing/Route.php(315): Slim\MiddlewareDispatcher->handle(Object(Laminas\Diactoros\ServerRequest)) #5 .../vendor/slim/slim/Slim/Routing/RouteRunner.php(68): Slim\Routing\Route->run(Object(Laminas\Diactoros\ServerRequest)) #6 .../ChurchCRM/Slim/Middleware/VersionMiddleware.php(14): Slim\Routing\RouteRunner->handle(Object(Laminas\Diactoros\ServerRequest)) #7 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(168): ChurchCRM\Slim\Middleware\VersionMiddleware->__invoke(Object(Laminas\Diactoros\ServerRequest), Object(Slim\Routing\RouteRunner)) #8 .../ChurchCRM/Slim/Middleware/AuthMiddleware.php(36): Psr\Http\Server\RequestHandlerInterface@anonymous->handle(Object(Laminas\Diactoros\ServerRequest)) #9 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(168): ChurchCRM\Slim\Middleware\AuthMiddleware->__invoke(Object(Laminas\Diactoros\ServerRequest), Object(Psr\Http\Server\RequestHandlerInterface@anonymous)) #10 .../vendor/slim/slim/Slim/Middleware/BodyParsingMiddleware.php(64): Psr\Http\Server\RequestHandlerInterface@anonymous->handle(Object(Laminas\Diactoros\ServerRequest)) #11 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(121): Slim\Middleware\BodyParsingMiddleware->process(Object(Laminas\Diactoros\ServerRequest), Object(Psr\Http\Server\RequestHandlerInterface@anonymous)) #12 .../vendor/slim/slim/Slim/Middleware/ErrorMiddleware.php(76): Psr\Http\Server\RequestHandlerInterface@anonymous->handle(Object(Laminas\Diactoros\ServerRequest)) #13 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(121): Slim\Middleware\ErrorMiddleware->process(Object(Laminas\Diactoros\ServerRequest), Object(Psr\Http\Server\RequestHandlerInterface@anonymous)) #14 .../vendor/slim/http-cache/src/Cache.php(67): Psr\Http\Server\RequestHandlerInterface@anonymous->handle(Object(Laminas\Diactoros\ServerRequest)) #15 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(121): Slim\HttpCache\Cache->process(Object(Laminas\Diactoros\ServerRequest), Object(Psr\Http\Server\RequestHandlerInterface@anonymous)) #16 .../vendor/slim/http-cache/src/Cache.php(67): Psr\Http\Server\RequestHandlerInterface@anonymous->handle(Object(Laminas\Diactoros\ServerRequest)) #17 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(121): Slim\HttpCache\Cache->process(Object(Laminas\Diactoros\ServerRequest), Object(Psr\Http\Server\RequestHandlerInterface@anonymous)) #18 .../vendor/slim/slim/Slim/MiddlewareDispatcher.php(65): Psr\Http\Server\RequestHandlerInterface@anonymous->handle(Object(Laminas\Diactoros\ServerRequest)) #19 .../vendor/slim/slim/Slim/App.php(199): Slim\MiddlewareDispatcher->handle(Object(Laminas\Diactoros\ServerRequest)) #20 .../vendor/slim/slim/Slim/App.php(183): Slim\App->handle(Object(Laminas\Diactoros\ServerRequest)) #21 .../api/index.php(76): Slim\App->run() #22 {main} [] [] ``` **Debugging Steps** Developer tools log: ``` debug:1052 [Report Only] Refused to load the font 'data:application/x-font-ttf;charset=utf-8;base64,AAEAAAALAIAAAwAwT1MvMg8SBfAAAAC8AAAAYGNtYXAXVtKNAAABHAAAAFRnYXNwAAAAEAAAAXAAAAAIZ2x5ZgYydxIAAAF4AAAFNGhlYWQUJ7cIAAAGrAAAADZoaGVhB20DzAAABuQAAAAkaG10eCIABhQAAAcIAAAALGxvY2ED4AU6AAAHNAAAABhtYXhwAA8AjAAAB0wAAAAgbmFtZXsr690AAAdsAAABhnBvc3QAAwAAAAAI9AAAACAAAwPAAZAABQAAApkCzAAAAI8CmQLMAAAB6wAzAQkAAAAAAAAAAAAAAAAAAAABEAAAAAAAAAAAAAAAAAAAAABAAADpBgPA/8AAQAPAAEAAAAABAAAAAAAAAAAAAAAgAAAAAAADAAAAAwAAABwAAQADAAAAHAADAAEAAAAcAAQAOAAAAAoACAACAAIAAQAg6Qb//f//AAAAAAAg6QD//f...ABAAAAAAAEAAcAdQABAAAAAAAFAAsAFQABAAAAAAAGAAcASwABAAAAAAAKABoAigADAAEECQABAA4ABwADAAEECQACAA4AZwADAAEECQADAA4APQADAAEECQAEAA4AfAADAAEECQAFABYAIAADAAEECQAGAA4AUgADAAEECQAKADQApGZjaWNvbnMAZgBjAGkAYwBvAG4Ac1ZlcnNpb24gMS4wAFYAZQByAHMAaQBvAG4AIAAxAC4AMGZjaWNvbnMAZgBjAGkAYwBvAG4Ac2ZjaWNvbnMAZgBjAGkAYwBvAG4Ac1JlZ3VsYXIAUgBlAGcAdQBsAGEAcmZjaWNvbnMAZgBjAGkAYwBvAG4Ac0ZvbnQgZ2VuZXJhdGVkIGJ5IEljb01vb24uAEYAbwBuAHQAIABnAGUAbgBlAHIAYQB0AGUAZAAgAGIAeQAgAEkAYwBvAE0AbwBvAG4ALgAAAAMAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA=' because it violates the following Content Security Policy directive: "font-src 'self' fonts.gstatic.com". /api/system/background/csp-report:1 Failed to load resource: the server responded with a status of 406 () ``` **Desktop (please complete the following information):** - OS and Version: Windows 11 - Browser (and Version): Chrome 123.0.6312.86 --- end body ---
6,226
[ -0.007134201470762491, -0.012748796492815018, -0.0023125414736568928, 0.01637995056807995, 0.02099178172647953, -0.012722291983664036, 0.04845072701573372, 0.04947557672858238, 0.003882949473336339, 0.035834453999996185, 0.007271143142133951, -0.00840643048286438, -0.0286781657487154, -0.0...
null
null
null
null
null
null
null
null
null
[ "libming", "libming" ]
There is an allocation-size-too-big issue and the poc is in the attachment. command: ``` ./swftophp poc-1``` ``` ==22513==ERROR: AddressSanitizer: requested allocation size 0xfffffffff7efdfbf (0xfffffffff7efefc0 after adjustments for alignment, red zones etc.) exceeds maximum supported size of 0x10000000000 (thread T0) #0 0x4ad350 in malloc /home/eval/build/llvm_tools/llvm-11.0.0.src/projects/compiler-rt/lib/asan/asan_malloc_linux.cpp:145 #1 0x5212ab in readBytes /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/read.c:227:17 #2 0x51d8c1 in parseABC_STRING_INFO /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/parser.c:3068:30 #3 0x51e07d in parseABC_CONSTANT_POOL /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/parser.c:3175:5 #4 0x51fb37 in parseABC_FILE /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/parser.c:3426:3 #5 0x52033e in parseSWF_DOABC /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/parser.c:3481:3 #6 0x4fc2b0 in blockParse /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/blocktypes.c:145:14 #7 0x4fbf39 in readMovie /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/main.c:265:11 #8 0x4fb633 in main /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/main.c:350:2 #9 0x7fec449f4c86 in __libc_start_main /build/glibc-CVJwZb/glibc-2.27/csu/../csu/libc-start.c:310 ==22513==HINT: if you don't care about these errors you may set allocator_may_return_null=1 SUMMARY: AddressSanitizer: allocation-size-too-big /home/eval/build/llvm_tools/llvm-11.0.0.src/projects/compiler-rt/lib/asan/asan_malloc_linux.cpp:145 in malloc ==22513==ABORTING ``` [poc-1.zip](https://github.com/libming/libming/files/12116025/poc-1.zip)
allocation-size-too-big in readBytes
https://api.github.com/repos/libming/libming/issues/276/comments
0
2023-07-21T03:35:50
2023-07-21T03:35:50Z
https://github.com/libming/libming/issues/276
1,815,100,719
276
false
This is a GitHub Issue repo:libming owner:libming Title : allocation-size-too-big in readBytes Issue date: --- start body --- There is an allocation-size-too-big issue and the poc is in the attachment. command: ``` ./swftophp poc-1``` ``` ==22513==ERROR: AddressSanitizer: requested allocation size 0xfffffffff7efdfbf (0xfffffffff7efefc0 after adjustments for alignment, red zones etc.) exceeds maximum supported size of 0x10000000000 (thread T0) #0 0x4ad350 in malloc /home/eval/build/llvm_tools/llvm-11.0.0.src/projects/compiler-rt/lib/asan/asan_malloc_linux.cpp:145 #1 0x5212ab in readBytes /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/read.c:227:17 #2 0x51d8c1 in parseABC_STRING_INFO /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/parser.c:3068:30 #3 0x51e07d in parseABC_CONSTANT_POOL /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/parser.c:3175:5 #4 0x51fb37 in parseABC_FILE /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/parser.c:3426:3 #5 0x52033e in parseSWF_DOABC /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/parser.c:3481:3 #6 0x4fc2b0 in blockParse /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/blocktypes.c:145:14 #7 0x4fbf39 in readMovie /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/main.c:265:11 #8 0x4fb633 in main /home/eval/Desktop/libming-CVE-2018-8962/obj-aflgo/util/../../util/main.c:350:2 #9 0x7fec449f4c86 in __libc_start_main /build/glibc-CVJwZb/glibc-2.27/csu/../csu/libc-start.c:310 ==22513==HINT: if you don't care about these errors you may set allocator_may_return_null=1 SUMMARY: AddressSanitizer: allocation-size-too-big /home/eval/build/llvm_tools/llvm-11.0.0.src/projects/compiler-rt/lib/asan/asan_malloc_linux.cpp:145 in malloc ==22513==ABORTING ``` [poc-1.zip](https://github.com/libming/libming/files/12116025/poc-1.zip) --- end body ---
1,990
[ -0.025417443364858627, 0.003060217248275876, -0.001360416179522872, 0.009579476900398731, 0.048503294587135315, 0.021705301478505135, -0.022073447704315186, 0.04055746644735336, -0.02578558959066868, 0.05313580110669136, -0.002220383146777749, 0.013728794641792774, 0.01481789443641901, 0.0...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Hello, I tried to change the value of `speed` [parameter](https://github.com/gpac/gpac/wiki/inspect), but the play back speed does not change. Can you please let me know how to change playback speed when `inspect` filter is used as a sink? Here is the code I tested, I used the newest code in master. ``` import libgpac as gpac import time fs = gpac.FilterSession(0) f1 = fs.load_src(f"http://localhost/dash/coaster2/coaster2_10x5.mpd") f2 = fs.load("dashin") f3 = fs.load("tileagg") f4 = fs.load("inspect:interleave=true:deep=true:dur=15:buffer=3000:speed=0.02") f2.set_source(f1) f3.set_source(f2) f4.set_source(f3) btime = time.time() fs.run() print(f'\n\nDone, executing time: {time.time() - btime}') fs.print_graph() fs.delete() gpac.close() ``` I changed `speed` to different values, from 0.02 to 100, but the reported executing time is always about 2.4 seconds. If I set `dur` to 15 and `speed` to 1.0, should the total time report be around 15 seconds too?
inspect filter, 'speed' parameter does not work
https://api.github.com/repos/gpac/gpac/issues/2290/comments
1
2022-10-20T18:09:33
2022-11-03T21:31:49Z
https://github.com/gpac/gpac/issues/2290
1,417,050,123
2,290
false
This is a GitHub Issue repo:gpac owner:gpac Title : inspect filter, 'speed' parameter does not work Issue date: --- start body --- Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Hello, I tried to change the value of `speed` [parameter](https://github.com/gpac/gpac/wiki/inspect), but the play back speed does not change. Can you please let me know how to change playback speed when `inspect` filter is used as a sink? Here is the code I tested, I used the newest code in master. ``` import libgpac as gpac import time fs = gpac.FilterSession(0) f1 = fs.load_src(f"http://localhost/dash/coaster2/coaster2_10x5.mpd") f2 = fs.load("dashin") f3 = fs.load("tileagg") f4 = fs.load("inspect:interleave=true:deep=true:dur=15:buffer=3000:speed=0.02") f2.set_source(f1) f3.set_source(f2) f4.set_source(f3) btime = time.time() fs.run() print(f'\n\nDone, executing time: {time.time() - btime}') fs.print_graph() fs.delete() gpac.close() ``` I changed `speed` to different values, from 0.02 to 100, but the reported executing time is always about 2.4 seconds. If I set `dur` to 15 and `speed` to 1.0, should the total time report be around 15 seconds too? --- end body ---
1,771
[ 0.0033143742475658655, 0.02529481239616871, -0.015725284814834595, 0.012565146200358868, 0.0003029036452062428, -0.004819038789719343, -0.03164250776171684, 0.046668585389852524, -0.016863208264112473, 0.04686052352190018, -0.03202638402581215, 0.05434614419937134, 0.04112977534532547, -0....
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/2013/07/16/how-to-file-a-bug-properly/ Segmentation faults and timeouts **Version:** ``` MP4Box - GPAC version 1.1.0-DEV-rev1555-g339e7a736-master (c) 2000-2021 Telecom Paris distributed under LGPL v2.1+ - http://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: --prefix=/root/fuck_bin/gpac/test Features: GPAC_CONFIG_LINUX GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_JPEG GPAC_HAS_PNG GPAC_HAS_LINUX_DVB GPAC_DISABLE_3D ``` **System information** Ubuntu 20.04 focal, AMD EPYC 7742 64-Core @ 16x 2.25GHz **command:** ``` ./MP4Box -bt poc ``` [poc.zip](https://github.com/gpac/gpac/files/7708362/poc.zip) **POCs** ``` . ├── __GI_raise-__GI_abort │   ├── id_000415,sig_11,src_004550,op_havoc,rep_8 │   ├── id_000584,sig_06,src_005256,op_havoc,rep_2 │   ├── id_000633,sig_11,src_005618,op_havoc,rep_2 │   └── id_000683,sig_06,src_006013,op_havoc,rep_8 ├── __strlen_avx2-__GI___strdup │   ├── id_000433,sig_11,src_004567,op_havoc,rep_16 │   ├── id_000515,sig_11,src_004707,op_havoc,rep_4 │   ├── id_000536,sig_11,src_004861,op_havoc,rep_2 │   ├── id_000587,sig_11,src_005268,op_havoc,rep_4 │   ├── id_000602,sig_11,src_005380,op_havoc,rep_8 │   ├── id_000631,sig_11,src_005586,op_havoc,rep_4 │   ├── id_000639,sig_11,src_005671+005596,op_splice,rep_32 │   ├── id_000644,sig_11,src_005684,op_havoc,rep_8 │   ├── id_000667,sig_11,src_005818,op_havoc,rep_2 │   ├── id_000711,sig_11,src_006404+006291,op_splice,rep_2 │   └── id_000720,sig_11,src_005346+006329,op_splice,rep_4 └── timeout ├── id_000395,sig_11,src_004543+004622,op_splice,rep_8 ├── id_000448,sig_11,src_004597,op_havoc,rep_4 ├── id_000454,sig_11,src_004606,op_havoc,rep_8 ├── id_000468,sig_11,src_004611+004914,op_splice,rep_2 ├── id_000561,sig_11,src_005035,op_havoc,rep_4 ├── id_000606,sig_11,src_005410+005204,op_splice,rep_2 └── id_000698,sig_11,src_006230,op_havoc,rep_16 3 directories, 22 files ``` **Result** The result is omitted here. **gdb** The gdb result is omitted here.
Segmentation faults and timeouts
https://api.github.com/repos/gpac/gpac/issues/1989/comments
0
2021-12-14T03:12:35
2021-12-14T10:36:54Z
https://github.com/gpac/gpac/issues/1989
1,079,257,533
1,989
false
This is a GitHub Issue repo:gpac owner:gpac Title : Segmentation faults and timeouts Issue date: --- start body --- Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/2013/07/16/how-to-file-a-bug-properly/ Segmentation faults and timeouts **Version:** ``` MP4Box - GPAC version 1.1.0-DEV-rev1555-g339e7a736-master (c) 2000-2021 Telecom Paris distributed under LGPL v2.1+ - http://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: --prefix=/root/fuck_bin/gpac/test Features: GPAC_CONFIG_LINUX GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_JPEG GPAC_HAS_PNG GPAC_HAS_LINUX_DVB GPAC_DISABLE_3D ``` **System information** Ubuntu 20.04 focal, AMD EPYC 7742 64-Core @ 16x 2.25GHz **command:** ``` ./MP4Box -bt poc ``` [poc.zip](https://github.com/gpac/gpac/files/7708362/poc.zip) **POCs** ``` . ├── __GI_raise-__GI_abort │   ├── id_000415,sig_11,src_004550,op_havoc,rep_8 │   ├── id_000584,sig_06,src_005256,op_havoc,rep_2 │   ├── id_000633,sig_11,src_005618,op_havoc,rep_2 │   └── id_000683,sig_06,src_006013,op_havoc,rep_8 ├── __strlen_avx2-__GI___strdup │   ├── id_000433,sig_11,src_004567,op_havoc,rep_16 │   ├── id_000515,sig_11,src_004707,op_havoc,rep_4 │   ├── id_000536,sig_11,src_004861,op_havoc,rep_2 │   ├── id_000587,sig_11,src_005268,op_havoc,rep_4 │   ├── id_000602,sig_11,src_005380,op_havoc,rep_8 │   ├── id_000631,sig_11,src_005586,op_havoc,rep_4 │   ├── id_000639,sig_11,src_005671+005596,op_splice,rep_32 │   ├── id_000644,sig_11,src_005684,op_havoc,rep_8 │   ├── id_000667,sig_11,src_005818,op_havoc,rep_2 │   ├── id_000711,sig_11,src_006404+006291,op_splice,rep_2 │   └── id_000720,sig_11,src_005346+006329,op_splice,rep_4 └── timeout ├── id_000395,sig_11,src_004543+004622,op_splice,rep_8 ├── id_000448,sig_11,src_004597,op_havoc,rep_4 ├── id_000454,sig_11,src_004606,op_havoc,rep_8 ├── id_000468,sig_11,src_004611+004914,op_splice,rep_2 ├── id_000561,sig_11,src_005035,op_havoc,rep_4 ├── id_000606,sig_11,src_005410+005204,op_splice,rep_2 └── id_000698,sig_11,src_006230,op_havoc,rep_16 3 directories, 22 files ``` **Result** The result is omitted here. **gdb** The gdb result is omitted here. --- end body ---
3,019
[ -0.010507981292903423, 0.011004830710589886, -0.012020775116980076, 0.002745645819231868, 0.010218771174550056, 0.0035279972944408655, -0.03396369516849518, 0.028223978355526924, -0.016062306240200996, 0.032836515456438065, -0.010708204470574856, 0.024961091578006744, 0.02083798684179783, ...
null
null
null
null
null
null
null
null
null
[ "LibreDWG", "libredwg" ]
Hello, I try to convert dwg to dxf using dwg2dxf, got error as below: dwg2dxf -o ~/Downloads/64a08d92d5b6a0c965b9442fb0b48fb7389b9cec.dwg.dxf /Downloads/64a08d92d5b6a0c965b9442fb0b48fb7389b9cec.dwg Reading DWG file /Downloads/64a08d92d5b6a0c965b9442fb0b48fb7389b9cec.dwg ERROR: Failed to decode file: /Users/weiken/Downloads/64a08d92d5b6a0c965b9442fb0b48fb7389b9cec.dwg 0x100 READ ERROR 0x100 origin dwg file open with AutoCAD: ![image](https://github.com/LibreDWG/libredwg/assets/42831401/df31929c-668f-41dc-baf9-bc6fd0d59f67) ### attachment [64a08d92d5b6a0c965b9442fb0b48fb7389b9cec.dwg.zip](https://github.com/LibreDWG/libredwg/files/11949891/64a08d92d5b6a0c965b9442fb0b48fb7389b9cec.dwg.zip)
decode section_map_address: 0x0 [RLL 0] (r2004)
https://api.github.com/repos/LibreDWG/libredwg/issues/769/comments
2
2023-07-04T13:44:37
2023-07-07T06:22:41Z
https://github.com/LibreDWG/libredwg/issues/769
1,787,964,517
769
false
This is a GitHub Issue repo:libredwg owner:LibreDWG Title : decode section_map_address: 0x0 [RLL 0] (r2004) Issue date: --- start body --- Hello, I try to convert dwg to dxf using dwg2dxf, got error as below: dwg2dxf -o ~/Downloads/64a08d92d5b6a0c965b9442fb0b48fb7389b9cec.dwg.dxf /Downloads/64a08d92d5b6a0c965b9442fb0b48fb7389b9cec.dwg Reading DWG file /Downloads/64a08d92d5b6a0c965b9442fb0b48fb7389b9cec.dwg ERROR: Failed to decode file: /Users/weiken/Downloads/64a08d92d5b6a0c965b9442fb0b48fb7389b9cec.dwg 0x100 READ ERROR 0x100 origin dwg file open with AutoCAD: ![image](https://github.com/LibreDWG/libredwg/assets/42831401/df31929c-668f-41dc-baf9-bc6fd0d59f67) ### attachment [64a08d92d5b6a0c965b9442fb0b48fb7389b9cec.dwg.zip](https://github.com/LibreDWG/libredwg/files/11949891/64a08d92d5b6a0c965b9442fb0b48fb7389b9cec.dwg.zip) --- end body ---
876
[ -0.008643040433526039, 0.008507546968758106, 0.0015554976416751742, 0.008379184640944004, 0.034429535269737244, 0.011880614794790745, 0.008115329779684544, 0.05034642294049263, 0.0053983344696462154, 0.018056251108646393, 0.008635908365249634, -0.01387022528797388, 0.02012430690228939, -0....
null
null
null
null
null
null
null
null
null
[ "axiomatic-systems", "Bento4" ]
SUMMARY: AddressSanitizer: heap-buffer-overflow in /Source/C++/Core/Ap4HvccAtom.cpp:282 - Version ``` $ ./mp4dump MP4 File Dumper - Version 1.2 (Bento4 Version 1.6.0.0) (c) 2002-2011 Axiomatic Systems, LLC ``` branch [d02ef82](https://github.com/axiomatic-systems/Bento4/commit/d02ef8230a8fee4904a750eb912521c5f1c74e0b) - Platform ``` $ gcc --version gcc (Ubuntu 9.4.0-1ubuntu1~20.04.1) 9.4.0 Copyright (C) 2019 Free Software Foundation, Inc. This is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. $ uname -r 5.13.0-40-generic $ lsb_release -a No LSB modules are available. Distributor ID: Ubuntu Description: Ubuntu 20.04.4 LTS Release: 20.04 Codename: focal ``` - Steps to reproduce ``` $ mkdir build $ cd build $ cmake .. -DCMAKE_CXX_FLAGS="-fsanitize=address -g" -DCMAKE_C_FLAGS="-fsanitize=address -g" -DCMAKE_EXE_LINKER_FLAGS="-fsanitize=address" -DCMAKE_MODULE_LINKER_FLAGS="-fsanitize=address" $ make $ ./mp4dump poc ``` - Asan ``` $ ./mp4dump poc ================================================================= ==2382787==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x6030000000e6 at pc 0x562f8af4a0a7 bp 0x7ffca9b2f1e0 sp 0x7ffca9b2f1d0 READ of size 1 at 0x6030000000e6 thread T0 #0 0x562f8af4a0a6 in AP4_HvccAtom::AP4_HvccAtom(unsigned int, unsigned char const*) /home/wulearn/Bento4/Source/C++/Core/Ap4HvccAtom.cpp:282 #1 0x562f8af47394 in AP4_HvccAtom::Create(unsigned int, AP4_ByteStream&) /home/wulearn/Bento4/Source/C++/Core/Ap4HvccAtom.cpp:90 #2 0x562f8af14fa6 in AP4_AtomFactory::CreateAtomFromStream(AP4_ByteStream&, unsigned int, unsigned int, unsigned long long, AP4_Atom*&) /home/wulearn/Bento4/Source/C++/Core/Ap4AtomFactory.cpp:539 #3 0x562f8af125ee in AP4_AtomFactory::CreateAtomFromStream(AP4_ByteStream&, unsigned long long&, AP4_Atom*&) /home/wulearn/Bento4/Source/C++/Core/Ap4AtomFactory.cpp:234 #4 0x562f8af11bbd in AP4_AtomFactory::CreateAtomFromStream(AP4_ByteStream&, AP4_Atom*&) /home/wulearn/Bento4/Source/C++/Core/Ap4AtomFactory.cpp:154 #5 0x562f8af04115 in main /home/wulearn/Bento4/Source/C++/Apps/Mp4Dump/Mp4Dump.cpp:342 #6 0x7f0dcc1e30b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x240b2) #7 0x562f8af028ed in _start (/home/wulearn/Bento4/build/mp4dump+0x30f8ed) 0x6030000000e6 is located 0 bytes to the right of 22-byte region [0x6030000000d0,0x6030000000e6) allocated by thread T0 here: #0 0x7f0dcc80c787 in operator new[](unsigned long) ../../../../src/libsanitizer/asan/asan_new_delete.cc:107 #1 0x562f8af399f7 in AP4_DataBuffer::AP4_DataBuffer(unsigned int) /home/wulearn/Bento4/Source/C++/Core/Ap4DataBuffer.cpp:55 #2 0x562f8af47327 in AP4_HvccAtom::Create(unsigned int, AP4_ByteStream&) /home/wulearn/Bento4/Source/C++/Core/Ap4HvccAtom.cpp:86 #3 0x562f8af14fa6 in AP4_AtomFactory::CreateAtomFromStream(AP4_ByteStream&, unsigned int, unsigned int, unsigned long long, AP4_Atom*&) /home/wulearn/Bento4/Source/C++/Core/Ap4AtomFactory.cpp:539 #4 0x562f8af125ee in AP4_AtomFactory::CreateAtomFromStream(AP4_ByteStream&, unsigned long long&, AP4_Atom*&) /home/wulearn/Bento4/Source/C++/Core/Ap4AtomFactory.cpp:234 #5 0x562f8af11bbd in AP4_AtomFactory::CreateAtomFromStream(AP4_ByteStream&, AP4_Atom*&) /home/wulearn/Bento4/Source/C++/Core/Ap4AtomFactory.cpp:154 #6 0x562f8af04115 in main /home/wulearn/Bento4/Source/C++/Apps/Mp4Dump/Mp4Dump.cpp:342 #7 0x7f0dcc1e30b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x240b2) SUMMARY: AddressSanitizer: heap-buffer-overflow /home/wulearn/Bento4/Source/C++/Core/Ap4HvccAtom.cpp:282 in AP4_HvccAtom::AP4_HvccAtom(unsigned int, unsigned char const*) Shadow bytes around the buggy address: 0x0c067fff7fc0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c067fff7fd0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c067fff7fe0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c067fff7ff0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c067fff8000: fa fa 00 00 00 fa fa fa 00 00 00 fa fa fa 00 00 =>0x0c067fff8010: 00 fa fa fa 00 00 00 fa fa fa 00 00[06]fa fa fa 0x0c067fff8020: 00 00 06 fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c067fff8030: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c067fff8040: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c067fff8050: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c067fff8060: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa Shadow byte legend (one shadow byte represents 8 application bytes): Addressable: 00 Partially addressable: 01 02 03 04 05 06 07 Heap left redzone: fa Freed heap region: fd Stack left redzone: f1 Stack mid redzone: f2 Stack right redzone: f3 Stack after return: f5 Stack use after scope: f8 Global redzone: f9 Global init order: f6 Poisoned by user: f7 Container overflow: fc Array cookie: ac Intra object redzone: bb ASan internal: fe Left alloca redzone: ca Right alloca redzone: cb Shadow gap: cc ==2382787==ABORTING ``` poc: [poc.zip](https://github.com/axiomatic-systems/Bento4/files/8647366/poc.zip) Thanks!
heap-buffer-overflow in /Source/C++/Core/Ap4HvccAtom.cpp:282
https://api.github.com/repos/axiomatic-systems/Bento4/issues/707/comments
0
2022-05-08T16:18:46
2022-05-08T16:18:46Z
https://github.com/axiomatic-systems/Bento4/issues/707
1,228,927,927
707
false
This is a GitHub Issue repo:Bento4 owner:axiomatic-systems Title : heap-buffer-overflow in /Source/C++/Core/Ap4HvccAtom.cpp:282 Issue date: --- start body --- SUMMARY: AddressSanitizer: heap-buffer-overflow in /Source/C++/Core/Ap4HvccAtom.cpp:282 - Version ``` $ ./mp4dump MP4 File Dumper - Version 1.2 (Bento4 Version 1.6.0.0) (c) 2002-2011 Axiomatic Systems, LLC ``` branch [d02ef82](https://github.com/axiomatic-systems/Bento4/commit/d02ef8230a8fee4904a750eb912521c5f1c74e0b) - Platform ``` $ gcc --version gcc (Ubuntu 9.4.0-1ubuntu1~20.04.1) 9.4.0 Copyright (C) 2019 Free Software Foundation, Inc. This is free software; see the source for copying conditions. There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. $ uname -r 5.13.0-40-generic $ lsb_release -a No LSB modules are available. Distributor ID: Ubuntu Description: Ubuntu 20.04.4 LTS Release: 20.04 Codename: focal ``` - Steps to reproduce ``` $ mkdir build $ cd build $ cmake .. -DCMAKE_CXX_FLAGS="-fsanitize=address -g" -DCMAKE_C_FLAGS="-fsanitize=address -g" -DCMAKE_EXE_LINKER_FLAGS="-fsanitize=address" -DCMAKE_MODULE_LINKER_FLAGS="-fsanitize=address" $ make $ ./mp4dump poc ``` - Asan ``` $ ./mp4dump poc ================================================================= ==2382787==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x6030000000e6 at pc 0x562f8af4a0a7 bp 0x7ffca9b2f1e0 sp 0x7ffca9b2f1d0 READ of size 1 at 0x6030000000e6 thread T0 #0 0x562f8af4a0a6 in AP4_HvccAtom::AP4_HvccAtom(unsigned int, unsigned char const*) /home/wulearn/Bento4/Source/C++/Core/Ap4HvccAtom.cpp:282 #1 0x562f8af47394 in AP4_HvccAtom::Create(unsigned int, AP4_ByteStream&) /home/wulearn/Bento4/Source/C++/Core/Ap4HvccAtom.cpp:90 #2 0x562f8af14fa6 in AP4_AtomFactory::CreateAtomFromStream(AP4_ByteStream&, unsigned int, unsigned int, unsigned long long, AP4_Atom*&) /home/wulearn/Bento4/Source/C++/Core/Ap4AtomFactory.cpp:539 #3 0x562f8af125ee in AP4_AtomFactory::CreateAtomFromStream(AP4_ByteStream&, unsigned long long&, AP4_Atom*&) /home/wulearn/Bento4/Source/C++/Core/Ap4AtomFactory.cpp:234 #4 0x562f8af11bbd in AP4_AtomFactory::CreateAtomFromStream(AP4_ByteStream&, AP4_Atom*&) /home/wulearn/Bento4/Source/C++/Core/Ap4AtomFactory.cpp:154 #5 0x562f8af04115 in main /home/wulearn/Bento4/Source/C++/Apps/Mp4Dump/Mp4Dump.cpp:342 #6 0x7f0dcc1e30b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x240b2) #7 0x562f8af028ed in _start (/home/wulearn/Bento4/build/mp4dump+0x30f8ed) 0x6030000000e6 is located 0 bytes to the right of 22-byte region [0x6030000000d0,0x6030000000e6) allocated by thread T0 here: #0 0x7f0dcc80c787 in operator new[](unsigned long) ../../../../src/libsanitizer/asan/asan_new_delete.cc:107 #1 0x562f8af399f7 in AP4_DataBuffer::AP4_DataBuffer(unsigned int) /home/wulearn/Bento4/Source/C++/Core/Ap4DataBuffer.cpp:55 #2 0x562f8af47327 in AP4_HvccAtom::Create(unsigned int, AP4_ByteStream&) /home/wulearn/Bento4/Source/C++/Core/Ap4HvccAtom.cpp:86 #3 0x562f8af14fa6 in AP4_AtomFactory::CreateAtomFromStream(AP4_ByteStream&, unsigned int, unsigned int, unsigned long long, AP4_Atom*&) /home/wulearn/Bento4/Source/C++/Core/Ap4AtomFactory.cpp:539 #4 0x562f8af125ee in AP4_AtomFactory::CreateAtomFromStream(AP4_ByteStream&, unsigned long long&, AP4_Atom*&) /home/wulearn/Bento4/Source/C++/Core/Ap4AtomFactory.cpp:234 #5 0x562f8af11bbd in AP4_AtomFactory::CreateAtomFromStream(AP4_ByteStream&, AP4_Atom*&) /home/wulearn/Bento4/Source/C++/Core/Ap4AtomFactory.cpp:154 #6 0x562f8af04115 in main /home/wulearn/Bento4/Source/C++/Apps/Mp4Dump/Mp4Dump.cpp:342 #7 0x7f0dcc1e30b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x240b2) SUMMARY: AddressSanitizer: heap-buffer-overflow /home/wulearn/Bento4/Source/C++/Core/Ap4HvccAtom.cpp:282 in AP4_HvccAtom::AP4_HvccAtom(unsigned int, unsigned char const*) Shadow bytes around the buggy address: 0x0c067fff7fc0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c067fff7fd0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c067fff7fe0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c067fff7ff0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c067fff8000: fa fa 00 00 00 fa fa fa 00 00 00 fa fa fa 00 00 =>0x0c067fff8010: 00 fa fa fa 00 00 00 fa fa fa 00 00[06]fa fa fa 0x0c067fff8020: 00 00 06 fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c067fff8030: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c067fff8040: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c067fff8050: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c067fff8060: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa Shadow byte legend (one shadow byte represents 8 application bytes): Addressable: 00 Partially addressable: 01 02 03 04 05 06 07 Heap left redzone: fa Freed heap region: fd Stack left redzone: f1 Stack mid redzone: f2 Stack right redzone: f3 Stack after return: f5 Stack use after scope: f8 Global redzone: f9 Global init order: f6 Poisoned by user: f7 Container overflow: fc Array cookie: ac Intra object redzone: bb ASan internal: fe Left alloca redzone: ca Right alloca redzone: cb Shadow gap: cc ==2382787==ABORTING ``` poc: [poc.zip](https://github.com/axiomatic-systems/Bento4/files/8647366/poc.zip) Thanks! --- end body ---
5,551
[ -0.043146926909685135, 0.011803395114839077, -0.012091810815036297, 0.0025921324267983437, 0.030889276415109634, -0.0031905940268188715, -0.0357058122754097, 0.036224957555532455, -0.028423326089978218, 0.027745550498366356, -0.019280562177300453, 0.009164395742118359, 0.00904902908951044, ...
CVE-2023-31630
2023-05-15T15:15:12.723000
An issue in the sqlo_query_spec component of openlink virtuoso-opensource v7.2.9 allows attackers to cause a Denial of Service (DoS) via crafted SQL statements.
{ "cvssMetricV2": null, "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "LOW", "attackVector": "NETWORK", "availabilityImpact": "HIGH", "baseScore": 7.5, "baseSeverity": "HIGH", "confidentialityImpact": "NONE", "integrityImpact": "NONE", "privilegesRequired": "NONE", "scope": "UNCHANGED", "userInteraction": "NONE", "vectorString": "CVSS:3.1/AV:N/AC:L/PR:N/UI:N/S:U/C:N/I:N/A:H", "version": "3.1" }, "exploitabilityScore": 3.9, "impactScore": 3.6, "source": "nvd@nist.gov", "type": "Primary" } ] }
[ { "source": "cve@mitre.org", "tags": [ "Exploit", "Issue Tracking", "Patch" ], "url": "https://github.com/openlink/virtuoso-opensource/issues/1138" } ]
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:openlinksw:virtuoso:7.2.9:*:*:*:open_source:*:*:*", "matchCriteriaId": "75FD7F91-D201-4286-8F97-D2D1C7C9D4C3", "versionEndExcluding": null, "versionEndIncluding": null, "vers...
https://github.com/openlink/virtuoso-opensource/issues/1138
[ "Exploit", "Issue Tracking", "Patch" ]
github.com
[ "openlink", "virtuoso-opensource" ]
The PoC is generated by my DBMS fuzzer. ```SQL CREATE TABLE mvtest_pdt1 AS (SELECT 1 AS col1) WITH DATA; ``` backtrace: ```c #0 0x70d571 (sqlo_query_spec+0x41) #1 0x7da2d4 (sqlc_insert+0x13b4) #2 0x6b7387 (sql_stmt_comp+0x987) #3 0x6ba122 (sql_compile_1+0x1a62) #4 0x68276d (bif_exec+0x67d) #5 0x74b044 (ins_call_bif+0xc4) #6 0x758f09 (code_vec_run_1+0xdf9) #7 0x7ac431 (qn_input+0x3c1) #8 0x7c04fa (qr_subq_exec+0x94a) #9 0x74c001 (ins_call+0xc91) #10 0x74d9f0 (ins_call_vec+0x320) #11 0x751205 (code_vec_run_v+0xa25) #12 0x7ac3ff (qn_input+0x38f) #13 0x7bdc6e (qr_exec+0x11ee) #14 0x7c36fe (qr_rec_exec+0x34e) #15 0x4e4076 (sql_ddl_node_input_1+0xc96) #16 0x4e444e (sql_ddl_node_input+0x10e) #17 0x7b9f1b (ddl_node_input_1+0x19b) #18 0x7ba5f7 (qn_without_ac_at+0xc7) #19 0x7ac43e (qn_input+0x3ce) #20 0x7bee59 (qr_dml_array_exec+0x839) #21 0x7cb872 (sf_sql_execute+0x15d2) #22 0x7cbf4e (sf_sql_execute_w+0x17e) #23 0x7d4c0d (sf_sql_execute_wrapper+0x3d) #24 0xe1f01c (future_wrapper+0x3fc) #25 0xe2691e (_thread_boot+0x11e) #26 0x7f3405362609 (start_thread+0xd9) #27 0x7f3405132133 (clone+0x43) ``` ways to reproduce (write poc to the file '/tmp/test.sql' first): ```bash # remove the old one docker container rm virtdb_test -f # start virtuoso through docker docker run --name virtdb_test -itd --env DBA_PASSWORD=dba openlink/virtuoso-opensource-7:7.2.9 # wait the server starting sleep 10 # check whether the simple query works echo "SELECT 1;" | docker exec -i virtdb_test isql 1111 dba # run the poc docker exec -i virtdb_test isql 1111 dba < "/tmp/test.sql" ```
virtuoso 7.2.9 crashed at sqlo_query_spec
https://api.github.com/repos/openlink/virtuoso-opensource/issues/1138/comments
0
2023-04-12T19:38:44
2023-04-14T08:12:16Z
https://github.com/openlink/virtuoso-opensource/issues/1138
1,665,134,343
1,138
true
This is a GitHub Issue repo:virtuoso-opensource owner:openlink Title : virtuoso 7.2.9 crashed at sqlo_query_spec Issue date: --- start body --- The PoC is generated by my DBMS fuzzer. ```SQL CREATE TABLE mvtest_pdt1 AS (SELECT 1 AS col1) WITH DATA; ``` backtrace: ```c #0 0x70d571 (sqlo_query_spec+0x41) #1 0x7da2d4 (sqlc_insert+0x13b4) #2 0x6b7387 (sql_stmt_comp+0x987) #3 0x6ba122 (sql_compile_1+0x1a62) #4 0x68276d (bif_exec+0x67d) #5 0x74b044 (ins_call_bif+0xc4) #6 0x758f09 (code_vec_run_1+0xdf9) #7 0x7ac431 (qn_input+0x3c1) #8 0x7c04fa (qr_subq_exec+0x94a) #9 0x74c001 (ins_call+0xc91) #10 0x74d9f0 (ins_call_vec+0x320) #11 0x751205 (code_vec_run_v+0xa25) #12 0x7ac3ff (qn_input+0x38f) #13 0x7bdc6e (qr_exec+0x11ee) #14 0x7c36fe (qr_rec_exec+0x34e) #15 0x4e4076 (sql_ddl_node_input_1+0xc96) #16 0x4e444e (sql_ddl_node_input+0x10e) #17 0x7b9f1b (ddl_node_input_1+0x19b) #18 0x7ba5f7 (qn_without_ac_at+0xc7) #19 0x7ac43e (qn_input+0x3ce) #20 0x7bee59 (qr_dml_array_exec+0x839) #21 0x7cb872 (sf_sql_execute+0x15d2) #22 0x7cbf4e (sf_sql_execute_w+0x17e) #23 0x7d4c0d (sf_sql_execute_wrapper+0x3d) #24 0xe1f01c (future_wrapper+0x3fc) #25 0xe2691e (_thread_boot+0x11e) #26 0x7f3405362609 (start_thread+0xd9) #27 0x7f3405132133 (clone+0x43) ``` ways to reproduce (write poc to the file '/tmp/test.sql' first): ```bash # remove the old one docker container rm virtdb_test -f # start virtuoso through docker docker run --name virtdb_test -itd --env DBA_PASSWORD=dba openlink/virtuoso-opensource-7:7.2.9 # wait the server starting sleep 10 # check whether the simple query works echo "SELECT 1;" | docker exec -i virtdb_test isql 1111 dba # run the poc docker exec -i virtdb_test isql 1111 dba < "/tmp/test.sql" ``` --- end body ---
1,780
[ -0.00897936150431633, -0.0032727865036576986, -0.006743188481777906, 0.013944012112915516, 0.04961877316236496, 0.03774798661470413, -0.008653469383716583, 0.030453557148575783, 0.0001694464444881305, 0.03880193457007408, -0.00021473328524734825, 0.0059561943635344505, 0.014394713565707207, ...
CVE-2023-50979
2023-12-18T04:15:50.970000
Crypto++ (aka cryptopp) through 8.9.0 has a Marvin side channel during decryption with PKCS#1 v1.5 padding.
{ "cvssMetricV2": null, "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "HIGH", "attackVector": "NETWORK", "availabilityImpact": "NONE", "baseScore": 5.9, "baseSeverity": "MEDIUM", "confidentialityImpact": "HIGH", "integrityImpact": "NONE", "privilegesRequired": "NONE", "scope": "UNCHANGED", "userInteraction": "NONE", "vectorString": "CVSS:3.1/AV:N/AC:H/PR:N/UI:N/S:U/C:H/I:N/A:N", "version": "3.1" }, "exploitabilityScore": 2.2, "impactScore": 3.6, "source": "nvd@nist.gov", "type": "Primary" } ] }
[ { "source": "cve@mitre.org", "tags": [ "Exploit" ], "url": "https://github.com/weidai11/cryptopp/issues/1247" } ]
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:cryptopp:crypto\\+\\+:*:*:*:*:*:*:*:*", "matchCriteriaId": "CE3A7AA7-A0EF-4363-A9D8-D1D31498B936", "versionEndExcluding": null, "versionEndIncluding": "8.9.0", "versionStartE...
https://github.com/weidai11/cryptopp/issues/1247
[ "Exploit" ]
github.com
[ "weidai11", "cryptopp" ]
### Crypto++ Issue Report I've verified that the Crypto++ is vulnerable to the [Marvin Attack](https://people.redhat.com/~hkario/marvin/)—a timing variant of the well-known Bleichenbacher attack. I've executed the test on ArchLinux, using crypto++ 8.9.0 from the Arch repository. The reproducer is available in the [marvin-toolkit](https://github.com/tomato42/marvin-toolkit/tree/master/example/crypto%2B%2B) repository. It was compiled with a simple `g++ -o time_decrypt -lcryptopp time_decrypt.cpp`. When executed on AMD Ryzen 5 5600X, I'm able to detect a side-channel signal when performing decryption with PKCS#1 v1.5 padding. The results are statistically significant with just few hundred measurement pairs, but below I'm showing results of a run with 100k repeats to show the size of the side-channel more clearly. The test summary: ``` Sign test mean p-value: 0.2339, median p-value: 4.255e-12, min p-value: 3.18e-88 Friedman test (chisquare approximation) for all samples p-value: 0.0 Worst pair: 1(no_header_with_payload_48), 5(valid_0) Mean of differences: 3.61457e-07s, 95% CI: 2.63026e-07s, 4.577243e-07s (±9.735e-08s) Median of differences: 4.80000e-07s, 95% CI: 4.30000e-07s, 5.200001e-07s (±4.500e-08s) Trimmed mean (5%) of differences: 3.84809e-07s, 95% CI: 3.26061e-07s, 4.382825e-07s (±5.611e-08s) Trimmed mean (25%) of differences: 4.82399e-07s, 95% CI: 4.37442e-07s, 5.255875e-07s (±4.407e-08s) Trimmed mean (45%) of differences: 4.77529e-07s, 95% CI: 4.31482e-07s, 5.214141e-07s (±4.497e-08s) Trimean of differences: 4.63125e-07s, 95% CI: 4.07500e-07s, 5.025002e-07s (±4.750e-08s) Layperson explanation: Definite side-channel detected, implementation is VULNERABLE ``` ![conf_interval_plot_trim_mean_25](https://github.com/weidai11/cryptopp/assets/618246/15a4e163-c2b5-4c06-a54f-259f8161d58b) legend: ``` ID,Name 0,header_only 1,no_header_with_payload_48 2,no_padding_48 3,no_structure 4,signature_padding_8 5,valid_0 6,valid_48 7,valid_192 8,valid_246 9,valid_repeated_byte_payload_246_1 10,valid_repeated_byte_payload_246_255 11,zero_byte_in_padding_48_4 ``` Explanation for the ciphertexts is in the [`step2.py`](https://github.com/tomato42/marvin-toolkit/blob/master/step2.py) file. As can be read from the graph, there's a clear difference between valid (probes 5, 6, and 7) and invalid ciphertexts. Moreover, the probe 2 is significantly different from others, which suggests a side-channel leakage in the numerical library (the last multiplication to apply deblinder, the last Montgormery reduction, or conversion of the arbitrary precision integer to a byte string), which means that that _all_ padding modes that use the same RSADP() primitive will be similarly vulnerable, that means RSA-OAEP too. (side note: I'm not a C++ programmer, so any help to make the reproducer cleaner, and command-line configurable, like the ones for openssl, pyca-cryptography, etc. will be really appreciated)
Crypto++ vulnerable to the Marvin Attack
https://api.github.com/repos/weidai11/cryptopp/issues/1247/comments
1
2023-11-12T12:57:41
2024-01-29T00:45:42Z
https://github.com/weidai11/cryptopp/issues/1247
1,989,391,420
1,247
true
This is a GitHub Issue repo:cryptopp owner:weidai11 Title : Crypto++ vulnerable to the Marvin Attack Issue date: --- start body --- ### Crypto++ Issue Report I've verified that the Crypto++ is vulnerable to the [Marvin Attack](https://people.redhat.com/~hkario/marvin/)—a timing variant of the well-known Bleichenbacher attack. I've executed the test on ArchLinux, using crypto++ 8.9.0 from the Arch repository. The reproducer is available in the [marvin-toolkit](https://github.com/tomato42/marvin-toolkit/tree/master/example/crypto%2B%2B) repository. It was compiled with a simple `g++ -o time_decrypt -lcryptopp time_decrypt.cpp`. When executed on AMD Ryzen 5 5600X, I'm able to detect a side-channel signal when performing decryption with PKCS#1 v1.5 padding. The results are statistically significant with just few hundred measurement pairs, but below I'm showing results of a run with 100k repeats to show the size of the side-channel more clearly. The test summary: ``` Sign test mean p-value: 0.2339, median p-value: 4.255e-12, min p-value: 3.18e-88 Friedman test (chisquare approximation) for all samples p-value: 0.0 Worst pair: 1(no_header_with_payload_48), 5(valid_0) Mean of differences: 3.61457e-07s, 95% CI: 2.63026e-07s, 4.577243e-07s (±9.735e-08s) Median of differences: 4.80000e-07s, 95% CI: 4.30000e-07s, 5.200001e-07s (±4.500e-08s) Trimmed mean (5%) of differences: 3.84809e-07s, 95% CI: 3.26061e-07s, 4.382825e-07s (±5.611e-08s) Trimmed mean (25%) of differences: 4.82399e-07s, 95% CI: 4.37442e-07s, 5.255875e-07s (±4.407e-08s) Trimmed mean (45%) of differences: 4.77529e-07s, 95% CI: 4.31482e-07s, 5.214141e-07s (±4.497e-08s) Trimean of differences: 4.63125e-07s, 95% CI: 4.07500e-07s, 5.025002e-07s (±4.750e-08s) Layperson explanation: Definite side-channel detected, implementation is VULNERABLE ``` ![conf_interval_plot_trim_mean_25](https://github.com/weidai11/cryptopp/assets/618246/15a4e163-c2b5-4c06-a54f-259f8161d58b) legend: ``` ID,Name 0,header_only 1,no_header_with_payload_48 2,no_padding_48 3,no_structure 4,signature_padding_8 5,valid_0 6,valid_48 7,valid_192 8,valid_246 9,valid_repeated_byte_payload_246_1 10,valid_repeated_byte_payload_246_255 11,zero_byte_in_padding_48_4 ``` Explanation for the ciphertexts is in the [`step2.py`](https://github.com/tomato42/marvin-toolkit/blob/master/step2.py) file. As can be read from the graph, there's a clear difference between valid (probes 5, 6, and 7) and invalid ciphertexts. Moreover, the probe 2 is significantly different from others, which suggests a side-channel leakage in the numerical library (the last multiplication to apply deblinder, the last Montgormery reduction, or conversion of the arbitrary precision integer to a byte string), which means that that _all_ padding modes that use the same RSADP() primitive will be similarly vulnerable, that means RSA-OAEP too. (side note: I'm not a C++ programmer, so any help to make the reproducer cleaner, and command-line configurable, like the ones for openssl, pyca-cryptography, etc. will be really appreciated) --- end body ---
3,123
[ -0.0399344265460968, -0.018660658970475197, -0.01129016000777483, 0.02459394931793213, 0.01949070394039154, 0.03048112615942955, -0.015863096341490746, 0.025147313252091408, -0.011789724230766296, 0.1021878570318222, 0.0041463859379291534, 0.01563252881169319, 0.01970590092241764, -0.00936...
null
null
null
null
null
null
null
null
null
[ "emqx", "nanomq" ]
**Describe the bug** Hi, I found something on the NanoMQ that is contrary to the protocol specification description (protocol violation or logic bug). For tracking purposes, I will report all results under this Issue. **Environment Details** - NanoMQ version 0.21.8 - Operating system and version: Ubuntu 20.04 **Client SDK** If possible include the mqtt sdk you used to connect to nanomq Minimal C test cases are perferred. **Additional context** Add any other context about the problem here.
Several protocol violations or bugs in NanoMQ
https://api.github.com/repos/nanomq/nanomq/issues/1782/comments
47
2024-05-14T02:27:55
2024-07-30T08:43:53Z
https://github.com/nanomq/nanomq/issues/1782
2,294,214,632
1,782
false
This is a GitHub Issue repo:nanomq owner:emqx Title : Several protocol violations or bugs in NanoMQ Issue date: --- start body --- **Describe the bug** Hi, I found something on the NanoMQ that is contrary to the protocol specification description (protocol violation or logic bug). For tracking purposes, I will report all results under this Issue. **Environment Details** - NanoMQ version 0.21.8 - Operating system and version: Ubuntu 20.04 **Client SDK** If possible include the mqtt sdk you used to connect to nanomq Minimal C test cases are perferred. **Additional context** Add any other context about the problem here. --- end body ---
664
[ -0.023701336234807968, 0.007359370589256287, -0.007359370589256287, -0.003261582227423787, 0.026698056608438492, 0.02883208356797695, -0.0015494407853111625, 0.04607079550623894, 0.0024953752290457487, 0.022762969136238098, -0.004366433713585138, -0.002111325738951564, 0.010829058475792408, ...
CVE-2022-3662
2022-10-26T19:15:17.857000
A vulnerability was found in Axiomatic Bento4. It has been declared as critical. This vulnerability affects the function GetOffset of the file Ap4Sample.h of the component mp42hls. The manipulation leads to use after free. The attack can be initiated remotely. The exploit has been disclosed to the public and may be used. VDB-212002 is the identifier assigned to this vulnerability.
{ "cvssMetricV2": null, "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "LOW", "attackVector": "LOCAL", "availabilityImpact": "HIGH", "baseScore": 7.8, "baseSeverity": "HIGH", "confidentialityImpact": "HIGH", "integrityImpact": "HIGH", "privilegesRequired": "NONE", "scope": "UNCHANGED", "userInteraction": "REQUIRED", "vectorString": "CVSS:3.1/AV:L/AC:L/PR:N/UI:R/S:U/C:H/I:H/A:H", "version": "3.1" }, "exploitabilityScore": 1.8, "impactScore": 5.9, "source": "nvd@nist.gov", "type": "Primary" }, { "cvssData": { "attackComplexity": "LOW", "attackVector": "NETWORK", "availabilityImpact": "LOW", "baseScore": 7.3, "baseSeverity": "HIGH", "confidentialityImpact": "LOW", "integrityImpact": "LOW", "privilegesRequired": "NONE", "scope": "UNCHANGED", "userInteraction": "NONE", "vectorString": "CVSS:3.1/AV:N/AC:L/PR:N/UI:N/S:U/C:L/I:L/A:L", "version": "3.1" }, "exploitabilityScore": 3.9, "impactScore": 3.4, "source": "cna@vuldb.com", "type": "Secondary" } ] }
[ { "source": "cna@vuldb.com", "tags": [ "Exploit", "Third Party Advisory" ], "url": "https://github.com/axiomatic-systems/Bento4/files/9817606/mp42hls_cuaf_Ap4Sample99.zip" }, { "source": "cna@vuldb.com", "tags": [ "Exploit", "Third Party Advisory" ], "url"...
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:axiosys:bento4:1.6.0-639:*:*:*:*:*:*:*", "matchCriteriaId": "A003FBD1-339C-409D-A304-7FEE97E23250", "versionEndExcluding": null, "versionEndIncluding": null, "versionStartExc...
https://github.com/axiomatic-systems/Bento4/issues/802
[ "Exploit", "Third Party Advisory" ]
github.com
[ "axiomatic-systems", "Bento4" ]
Hi, there. There is an heap overflow in mp42hls, GetOffset, Ap4Sample.h:99, in the newest commit 5e7bb34a08272c49242196eba1cefab8af55f381. This seems to be an incomplete fix of issue #461. Here is the reproducing command: ~~~~ ./mp42hls poc ~~~~ POC: [mp42hls_cuaf_Ap4Sample99.zip](https://github.com/axiomatic-systems/Bento4/files/9817606/mp42hls_cuaf_Ap4Sample99.zip) (unzip first) Here is the reproduce trace reported by ASAN: ~~~~ ==2007234==ERROR: AddressSanitizer: heap-use-after-free on address 0x604000005dd8 at pc 0x0000005852ab bp 0x7ffc127b7960 sp 0x7ffc127b7958 READ of size 8 at 0x604000005dd8 thread T0 #0 0x5852aa in AP4_Sample::GetOffset() const /benchmark/Bento4/Source/C++/Core/Ap4Sample.h:99:48 #1 0x5852aa in AP4_LinearReader::Advance(bool) /benchmark/Bento4/Source/C++/Core/Ap4LinearReader.cpp:434:54 #2 0x585ab1 in AP4_LinearReader::ReadNextSample(unsigned int, AP4_Sample&, AP4_DataBuffer&) /benchmark/Bento4/Source/C++/Core/Ap4LinearReader.cpp:530:29 #3 0x509a31 in ReadSample(SampleReader&, AP4_Track&, AP4_Sample&, AP4_DataBuffer&, double&, double&, bool&) /benchmark/Bento4/Source/C++/Apps/Mp42Hls/Mp42Hls.cpp:1004:32 #4 0x509a31 in WriteSamples(AP4_Mpeg2TsWriter*, PackedAudioWriter*, AP4_Track*, SampleReader*, AP4_Mpeg2TsWriter::SampleStream*, AP4_Track*, SampleReader*, AP4_Mpeg2TsWriter::SampleStream*, unsigned int, unsigned char) /benchmark/Bento4/Source/C++/Apps/Mp42Hls/Mp42Hls.cpp:1289:22 #5 0x509a31 in main /benchmark/Bento4/Source/C++/Apps/Mp42Hls/Mp42Hls.cpp:2188:14 #6 0x7f33bacb6082 in __libc_start_main /build/glibc-SzIz7B/glibc-2.31/csu/../csu/libc-start.c:308:16 #7 0x41d8ed in _start ( /benchmark/Bento4/build-a/mp42hls+0x41d8ed) 0x604000005dd8 is located 8 bytes inside of 48-byte region [0x604000005dd0,0x604000005e00) freed by thread T0 here: #0 0x4f88b7 in operator delete(void*) /dependence/llvm11/llvm-11.0.0.src/projects/compiler-rt/lib/asan/asan_new_delete.cpp:160:3 #1 0x584f07 in AP4_LinearReader::SampleBuffer::~SampleBuffer() /benchmark/Bento4/Source/C++/Core/Ap4LinearReader.h:104:26 #2 0x584f07 in AP4_LinearReader::Advance(bool) /benchmark/Bento4/Source/C++/Core/Ap4LinearReader.cpp:462:17 previously allocated by thread T0 here: #0 0x4f7eb7 in operator new(unsigned long) /dependence/llvm11/llvm-11.0.0.src/projects/compiler-rt/lib/asan/asan_new_delete.cpp:99:3 #1 0x584892 in AP4_LinearReader::Advance(bool) /benchmark/Bento4/Source/C++/Core/Ap4LinearReader.cpp:422:41 SUMMARY: AddressSanitizer: heap-use-after-free /benchmark/Bento4/Source/C++/Core/Ap4Sample.h:99:48 in AP4_Sample::GetOffset() const Shadow bytes around the buggy address: 0x0c087fff8b60: fa fa fd fd fd fd fd fa fa fa fd fd fd fd fd fd 0x0c087fff8b70: fa fa fd fd fd fd fd fa fa fa fd fd fd fd fd fd 0x0c087fff8b80: fa fa fd fd fd fd fd fa fa fa fd fd fd fd fd fd 0x0c087fff8b90: fa fa fd fd fd fd fd fa fa fa fd fd fd fd fd fd 0x0c087fff8ba0: fa fa fd fd fd fd fd fa fa fa fd fd fd fd fd fd =>0x0c087fff8bb0: fa fa fd fd fd fd fd fa fa fa fd[fd]fd fd fd fd 0x0c087fff8bc0: fa fa fd fd fd fd fd fa fa fa fa fa fa fa fa fa 0x0c087fff8bd0: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c087fff8be0: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c087fff8bf0: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c087fff8c00: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa Shadow byte legend (one shadow byte represents 8 application bytes): Addressable: 00 Partially addressable: 01 02 03 04 05 06 07 Heap left redzone: fa Freed heap region: fd Stack left redzone: f1 Stack mid redzone: f2 Stack right redzone: f3 Stack after return: f5 Stack use after scope: f8 Global redzone: f9 Global init order: f6 Poisoned by user: f7 Container overflow: fc Array cookie: ac Intra object redzone: bb ASan internal: fe Left alloca redzone: ca Right alloca redzone: cb Shadow gap: cc ==2007234==ABORTING ~~~~
Concurrent heap use after free in mp42hls, GetOffset, Ap4Sample.h:99
https://api.github.com/repos/axiomatic-systems/Bento4/issues/802/comments
0
2022-10-19T06:53:27
2023-05-29T02:40:07Z
https://github.com/axiomatic-systems/Bento4/issues/802
1,414,407,802
802
true
This is a GitHub Issue repo:Bento4 owner:axiomatic-systems Title : Concurrent heap use after free in mp42hls, GetOffset, Ap4Sample.h:99 Issue date: --- start body --- Hi, there. There is an heap overflow in mp42hls, GetOffset, Ap4Sample.h:99, in the newest commit 5e7bb34a08272c49242196eba1cefab8af55f381. This seems to be an incomplete fix of issue #461. Here is the reproducing command: ~~~~ ./mp42hls poc ~~~~ POC: [mp42hls_cuaf_Ap4Sample99.zip](https://github.com/axiomatic-systems/Bento4/files/9817606/mp42hls_cuaf_Ap4Sample99.zip) (unzip first) Here is the reproduce trace reported by ASAN: ~~~~ ==2007234==ERROR: AddressSanitizer: heap-use-after-free on address 0x604000005dd8 at pc 0x0000005852ab bp 0x7ffc127b7960 sp 0x7ffc127b7958 READ of size 8 at 0x604000005dd8 thread T0 #0 0x5852aa in AP4_Sample::GetOffset() const /benchmark/Bento4/Source/C++/Core/Ap4Sample.h:99:48 #1 0x5852aa in AP4_LinearReader::Advance(bool) /benchmark/Bento4/Source/C++/Core/Ap4LinearReader.cpp:434:54 #2 0x585ab1 in AP4_LinearReader::ReadNextSample(unsigned int, AP4_Sample&, AP4_DataBuffer&) /benchmark/Bento4/Source/C++/Core/Ap4LinearReader.cpp:530:29 #3 0x509a31 in ReadSample(SampleReader&, AP4_Track&, AP4_Sample&, AP4_DataBuffer&, double&, double&, bool&) /benchmark/Bento4/Source/C++/Apps/Mp42Hls/Mp42Hls.cpp:1004:32 #4 0x509a31 in WriteSamples(AP4_Mpeg2TsWriter*, PackedAudioWriter*, AP4_Track*, SampleReader*, AP4_Mpeg2TsWriter::SampleStream*, AP4_Track*, SampleReader*, AP4_Mpeg2TsWriter::SampleStream*, unsigned int, unsigned char) /benchmark/Bento4/Source/C++/Apps/Mp42Hls/Mp42Hls.cpp:1289:22 #5 0x509a31 in main /benchmark/Bento4/Source/C++/Apps/Mp42Hls/Mp42Hls.cpp:2188:14 #6 0x7f33bacb6082 in __libc_start_main /build/glibc-SzIz7B/glibc-2.31/csu/../csu/libc-start.c:308:16 #7 0x41d8ed in _start ( /benchmark/Bento4/build-a/mp42hls+0x41d8ed) 0x604000005dd8 is located 8 bytes inside of 48-byte region [0x604000005dd0,0x604000005e00) freed by thread T0 here: #0 0x4f88b7 in operator delete(void*) /dependence/llvm11/llvm-11.0.0.src/projects/compiler-rt/lib/asan/asan_new_delete.cpp:160:3 #1 0x584f07 in AP4_LinearReader::SampleBuffer::~SampleBuffer() /benchmark/Bento4/Source/C++/Core/Ap4LinearReader.h:104:26 #2 0x584f07 in AP4_LinearReader::Advance(bool) /benchmark/Bento4/Source/C++/Core/Ap4LinearReader.cpp:462:17 previously allocated by thread T0 here: #0 0x4f7eb7 in operator new(unsigned long) /dependence/llvm11/llvm-11.0.0.src/projects/compiler-rt/lib/asan/asan_new_delete.cpp:99:3 #1 0x584892 in AP4_LinearReader::Advance(bool) /benchmark/Bento4/Source/C++/Core/Ap4LinearReader.cpp:422:41 SUMMARY: AddressSanitizer: heap-use-after-free /benchmark/Bento4/Source/C++/Core/Ap4Sample.h:99:48 in AP4_Sample::GetOffset() const Shadow bytes around the buggy address: 0x0c087fff8b60: fa fa fd fd fd fd fd fa fa fa fd fd fd fd fd fd 0x0c087fff8b70: fa fa fd fd fd fd fd fa fa fa fd fd fd fd fd fd 0x0c087fff8b80: fa fa fd fd fd fd fd fa fa fa fd fd fd fd fd fd 0x0c087fff8b90: fa fa fd fd fd fd fd fa fa fa fd fd fd fd fd fd 0x0c087fff8ba0: fa fa fd fd fd fd fd fa fa fa fd fd fd fd fd fd =>0x0c087fff8bb0: fa fa fd fd fd fd fd fa fa fa fd[fd]fd fd fd fd 0x0c087fff8bc0: fa fa fd fd fd fd fd fa fa fa fa fa fa fa fa fa 0x0c087fff8bd0: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c087fff8be0: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c087fff8bf0: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c087fff8c00: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa Shadow byte legend (one shadow byte represents 8 application bytes): Addressable: 00 Partially addressable: 01 02 03 04 05 06 07 Heap left redzone: fa Freed heap region: fd Stack left redzone: f1 Stack mid redzone: f2 Stack right redzone: f3 Stack after return: f5 Stack use after scope: f8 Global redzone: f9 Global init order: f6 Poisoned by user: f7 Container overflow: fc Array cookie: ac Intra object redzone: bb ASan internal: fe Left alloca redzone: ca Right alloca redzone: cb Shadow gap: cc ==2007234==ABORTING ~~~~ --- end body ---
4,396
[ -0.037215568125247955, 0.001705016242340207, -0.01144918892532587, -0.010011384263634682, 0.021635543555021286, -0.0050665512681007385, -0.04573589563369751, 0.04588804394006729, -0.01777096651494503, 0.03669826313853264, -0.012133858166635036, 0.006321778055280447, 0.021148668602108955, 0...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
I'm using below command to send a file using route. gpac -i ./gpac.zip:mime=application/octet-string -o route://225.1.1.0:6000 Questions: -How to set a limit the route mux rate? -This command keep sending the file forever, how to exit after send file 2x. Thanks
Question ROUTE
https://api.github.com/repos/gpac/gpac/issues/2786/comments
1
2024-03-21T03:18:41
2024-04-08T15:11:21Z
https://github.com/gpac/gpac/issues/2786
2,199,026,477
2,786
false
This is a GitHub Issue repo:gpac owner:gpac Title : Question ROUTE Issue date: --- start body --- I'm using below command to send a file using route. gpac -i ./gpac.zip:mime=application/octet-string -o route://225.1.1.0:6000 Questions: -How to set a limit the route mux rate? -This command keep sending the file forever, how to exit after send file 2x. Thanks --- end body ---
384
[ -0.06617166846990585, 0.008051594719290733, -0.01398793887346983, 0.0001554215414216742, -0.01854444481432438, 0.018726401031017303, -0.009848418645560741, 0.06932558864355087, -0.04545890539884567, 0.022699126973748207, -0.04200172424316406, 0.01537536084651947, 0.026520222425460815, -0.0...
null
null
null
null
null
null
null
null
null
[ "strukturag", "libde265" ]
libde265/arm/
cmake does not compile neon code for arm architecture
https://api.github.com/repos/strukturag/libde265/issues/445/comments
1
2024-02-06T15:36:15
2024-02-07T15:42:39Z
https://github.com/strukturag/libde265/issues/445
2,121,077,273
445
false
This is a GitHub Issue repo:libde265 owner:strukturag Title : cmake does not compile neon code for arm architecture Issue date: --- start body --- libde265/arm/ --- end body ---
179
[ -0.0044855899177491665, 0.04006289690732956, -0.021115658804774284, 0.011132989078760147, 0.06224954128265381, 0.022847751155495644, 0.00246757036074996, 0.05558561533689499, -0.01553594134747982, 0.005331802647560835, -0.01757214218378067, -0.00012519654410425574, -0.010075223632156849, 0...
null
null
null
null
null
null
null
null
null
[ "libming", "libming" ]
## Environment Ubuntu 22.04.3 LTS ## Compiler clang version 11.0.0 Target: x86_64-unknown-linux-gnu Thread model: posix ## Step to reproduce ```cd libming ./autogen.sh ./configure --disable-shared --disable-freetype cd util ./swftophp <PoC file> ``` ## Contents of PoC file ``` Unzip the file below. ``` [PoC.zip](https://github.com/libming/libming/files/13644155/PoC.zip) ## Expected behavior Print error or warning messages handled within swftophp. ## Current behavior ``` header indicates a filesize of 50529130 but filesize is 134 <?php $m = new SWFMovie(10); ming_setscale(1.0); $m->setRate(7.035156); $m->setDimension(0, 0); $m->setFrames(1799); Stream out of sync after parse of blocktype 11 (SWF_DEFINETEXT). 75 but expecting 71. /* SWF_DEFINETEXT */ $character775 = new SWFText(1); $character775->setColor(0xb3, 0x97, 0x82); $character775->moveTo(-26729, 16383); outputSWF_TEXT_RECORD: can't process text record: fonts information list is NULL $character775->addString("X"); Stream out of sync after parse of blocktype 28 (SWF_REMOVEOBJECT2). 75 but expecting 83. /* SWF_REMOVEOBJECT2 */ $m->remove($i1799); sudden file end: read failed @85 fileSize 134, request 56 extra garbage (i.e., we messed up in main): 0000: 07 07 07 07 ff 7f 39 05 55 02 07 07 07 05 05 05 .....9. U....... 0010: 0a f5 06 0a 0a 27 7a 51 5e 07 01 01 fd 05 05 05 .....'zQ ^....... 0020: 23 46 10 00 0a 07 07 07 07 07 07 05 05 05 23 46 #F...... ......#F 0030: 10 . header('Content-type: application/x-shockwave-flash'); $m->output(9); ?> ================================================================= ==1402273==ERROR: LeakSanitizer: detected memory leaks Direct leak of 48 byte(s) in 1 object(s) allocated from: #0 0x7fb3260b4c38 in __interceptor_realloc ../../../../src/libsanitizer/asan/asan_malloc_linux.cpp:164 #1 0x5639766f3f84 in parseSWF_DEFINETEXT (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x4af84) #2 0x5639766cbee4 in blockParse (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22ee4) #3 0x5639766cb3ca in readMovie (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x223ca) #4 0x5639766cbc3d in main (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22c3d) #5 0x7fb325c29d8f in __libc_start_call_main ../sysdeps/nptl/libc_start_call_main.h:58 Indirect leak of 240 byte(s) in 1 object(s) allocated from: #0 0x7fb3260b4887 in __interceptor_malloc ../../../../src/libsanitizer/asan/asan_malloc_linux.cpp:145 #1 0x5639766df499 in parseSWF_TEXTRECORD (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x36499) #2 0x5639766f40d9 in parseSWF_DEFINETEXT (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x4b0d9) #3 0x5639766cbee4 in blockParse (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22ee4) #4 0x5639766cb3ca in readMovie (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x223ca) #5 0x5639766cbc3d in main (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22c3d) #6 0x7fb325c29d8f in __libc_start_call_main ../sysdeps/nptl/libc_start_call_main.h:58 Indirect leak of 120 byte(s) in 15 object(s) allocated from: #0 0x7fb3260b4887 in __interceptor_malloc ../../../../src/libsanitizer/asan/asan_malloc_linux.cpp:145 #1 0x5639766decca in parseSWF_GLYPHENTRY (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x35cca) #2 0x5639766df528 in parseSWF_TEXTRECORD (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x36528) #3 0x5639766f40d9 in parseSWF_DEFINETEXT (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x4b0d9) #4 0x5639766cbee4 in blockParse (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22ee4) #5 0x5639766cb3ca in readMovie (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x223ca) #6 0x5639766cbc3d in main (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22c3d) #7 0x7fb325c29d8f in __libc_start_call_main ../sysdeps/nptl/libc_start_call_main.h:58 Indirect leak of 120 byte(s) in 15 object(s) allocated from: #0 0x7fb3260b4887 in __interceptor_malloc ../../../../src/libsanitizer/asan/asan_malloc_linux.cpp:145 #1 0x5639766dead1 in parseSWF_GLYPHENTRY (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x35ad1) #2 0x5639766df528 in parseSWF_TEXTRECORD (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x36528) #3 0x5639766f40d9 in parseSWF_DEFINETEXT (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x4b0d9) #4 0x5639766cbee4 in blockParse (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22ee4) #5 0x5639766cb3ca in readMovie (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x223ca) #6 0x5639766cbc3d in main (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22c3d) #7 0x7fb325c29d8f in __libc_start_call_main ../sysdeps/nptl/libc_start_call_main.h:58 SUMMARY: AddressSanitizer: 528 byte(s) leaked in 32 allocation(s). ```
Memory leaks in parseSWF_DEFINETEXT
https://api.github.com/repos/libming/libming/issues/303/comments
0
2023-12-12T05:47:16
2023-12-12T05:47:16Z
https://github.com/libming/libming/issues/303
2,037,028,499
303
false
This is a GitHub Issue repo:libming owner:libming Title : Memory leaks in parseSWF_DEFINETEXT Issue date: --- start body --- ## Environment Ubuntu 22.04.3 LTS ## Compiler clang version 11.0.0 Target: x86_64-unknown-linux-gnu Thread model: posix ## Step to reproduce ```cd libming ./autogen.sh ./configure --disable-shared --disable-freetype cd util ./swftophp <PoC file> ``` ## Contents of PoC file ``` Unzip the file below. ``` [PoC.zip](https://github.com/libming/libming/files/13644155/PoC.zip) ## Expected behavior Print error or warning messages handled within swftophp. ## Current behavior ``` header indicates a filesize of 50529130 but filesize is 134 <?php $m = new SWFMovie(10); ming_setscale(1.0); $m->setRate(7.035156); $m->setDimension(0, 0); $m->setFrames(1799); Stream out of sync after parse of blocktype 11 (SWF_DEFINETEXT). 75 but expecting 71. /* SWF_DEFINETEXT */ $character775 = new SWFText(1); $character775->setColor(0xb3, 0x97, 0x82); $character775->moveTo(-26729, 16383); outputSWF_TEXT_RECORD: can't process text record: fonts information list is NULL $character775->addString("X"); Stream out of sync after parse of blocktype 28 (SWF_REMOVEOBJECT2). 75 but expecting 83. /* SWF_REMOVEOBJECT2 */ $m->remove($i1799); sudden file end: read failed @85 fileSize 134, request 56 extra garbage (i.e., we messed up in main): 0000: 07 07 07 07 ff 7f 39 05 55 02 07 07 07 05 05 05 .....9. U....... 0010: 0a f5 06 0a 0a 27 7a 51 5e 07 01 01 fd 05 05 05 .....'zQ ^....... 0020: 23 46 10 00 0a 07 07 07 07 07 07 05 05 05 23 46 #F...... ......#F 0030: 10 . header('Content-type: application/x-shockwave-flash'); $m->output(9); ?> ================================================================= ==1402273==ERROR: LeakSanitizer: detected memory leaks Direct leak of 48 byte(s) in 1 object(s) allocated from: #0 0x7fb3260b4c38 in __interceptor_realloc ../../../../src/libsanitizer/asan/asan_malloc_linux.cpp:164 #1 0x5639766f3f84 in parseSWF_DEFINETEXT (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x4af84) #2 0x5639766cbee4 in blockParse (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22ee4) #3 0x5639766cb3ca in readMovie (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x223ca) #4 0x5639766cbc3d in main (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22c3d) #5 0x7fb325c29d8f in __libc_start_call_main ../sysdeps/nptl/libc_start_call_main.h:58 Indirect leak of 240 byte(s) in 1 object(s) allocated from: #0 0x7fb3260b4887 in __interceptor_malloc ../../../../src/libsanitizer/asan/asan_malloc_linux.cpp:145 #1 0x5639766df499 in parseSWF_TEXTRECORD (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x36499) #2 0x5639766f40d9 in parseSWF_DEFINETEXT (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x4b0d9) #3 0x5639766cbee4 in blockParse (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22ee4) #4 0x5639766cb3ca in readMovie (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x223ca) #5 0x5639766cbc3d in main (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22c3d) #6 0x7fb325c29d8f in __libc_start_call_main ../sysdeps/nptl/libc_start_call_main.h:58 Indirect leak of 120 byte(s) in 15 object(s) allocated from: #0 0x7fb3260b4887 in __interceptor_malloc ../../../../src/libsanitizer/asan/asan_malloc_linux.cpp:145 #1 0x5639766decca in parseSWF_GLYPHENTRY (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x35cca) #2 0x5639766df528 in parseSWF_TEXTRECORD (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x36528) #3 0x5639766f40d9 in parseSWF_DEFINETEXT (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x4b0d9) #4 0x5639766cbee4 in blockParse (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22ee4) #5 0x5639766cb3ca in readMovie (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x223ca) #6 0x5639766cbc3d in main (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22c3d) #7 0x7fb325c29d8f in __libc_start_call_main ../sysdeps/nptl/libc_start_call_main.h:58 Indirect leak of 120 byte(s) in 15 object(s) allocated from: #0 0x7fb3260b4887 in __interceptor_malloc ../../../../src/libsanitizer/asan/asan_malloc_linux.cpp:145 #1 0x5639766dead1 in parseSWF_GLYPHENTRY (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x35ad1) #2 0x5639766df528 in parseSWF_TEXTRECORD (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x36528) #3 0x5639766f40d9 in parseSWF_DEFINETEXT (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x4b0d9) #4 0x5639766cbee4 in blockParse (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22ee4) #5 0x5639766cb3ca in readMovie (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x223ca) #6 0x5639766cbc3d in main (/home/smb8508/Desktop/work_space/debuging/libming/util/swftophp+0x22c3d) #7 0x7fb325c29d8f in __libc_start_call_main ../sysdeps/nptl/libc_start_call_main.h:58 SUMMARY: AddressSanitizer: 528 byte(s) leaked in 32 allocation(s). ``` --- end body ---
5,422
[ -0.012809797190129757, 0.018440838903188705, -0.010858690366148949, 0.027397818863391876, 0.058088645339012146, 0.024944528937339783, -0.022968724370002747, 0.02408834733068943, -0.03879162296652794, 0.03635479882359505, 0.017749307677149773, 0.020021483302116394, 0.003924852702766657, -0....
null
null
null
null
null
null
null
null
null
[ "kubernetes", "kubernetes" ]
### What happened? I'm looking at various ways to expose a service directly from a set of hosts in a k8s cluster. [Yes, I know about the caveats here, but this is a service that doesn't play like a traditional http/rest endpoint.] I can do this with a svc that has a `port` setting and a list of `spec.externalIPs` - however, I need to maintain that list of externalIPs manually (or write a small oeprator that does so). Since these externalIPs tend to turn into individual iptbles rules (or the moral equivalent, depending on your networking stack), it's surprising that I can't give a CIDR here. That would (a) cut down on the number of IPtables rules that need evaluating, (b) require no dynamic updates. ### What did you expect to happen? I'd like to be able to say `spec.externalIPs: ["169.254.0.0/16"]` and not have to maintain the set of IPs on the svc through some external reconciliation. ### How can we reproduce it (as minimally and precisely as possible)? As above - this is a limitation currently in the validation of externalIPs rules. ### Anything else we need to know? This is related to a slew of questions people ask about exposing NodePorts on low ports. There are legitimate reasons to want to do this (again, largely around services that don't behave like a cloudy rest service, but those definitely exist), but there's always been a reasonable degree of pushback from supporting this - also for legitimate reasons. ### Kubernetes version all current ### Cloud provider n/a ### OS version n/a ### Install tools n/a ### Container runtime (CRI) and version (if applicable) n/a ### Related plugins (CNI, CSI, ...) and versions (if applicable) n/a
service.spec.externalIPs is unnecessarily limited (enhancement request)
https://api.github.com/repos/kubernetes/kubernetes/issues/124636/comments
11
2024-04-30T10:36:45
2024-05-13T18:55:27Z
https://github.com/kubernetes/kubernetes/issues/124636
2,271,108,727
124,636
false
This is a GitHub Issue repo:kubernetes owner:kubernetes Title : service.spec.externalIPs is unnecessarily limited (enhancement request) Issue date: --- start body --- ### What happened? I'm looking at various ways to expose a service directly from a set of hosts in a k8s cluster. [Yes, I know about the caveats here, but this is a service that doesn't play like a traditional http/rest endpoint.] I can do this with a svc that has a `port` setting and a list of `spec.externalIPs` - however, I need to maintain that list of externalIPs manually (or write a small oeprator that does so). Since these externalIPs tend to turn into individual iptbles rules (or the moral equivalent, depending on your networking stack), it's surprising that I can't give a CIDR here. That would (a) cut down on the number of IPtables rules that need evaluating, (b) require no dynamic updates. ### What did you expect to happen? I'd like to be able to say `spec.externalIPs: ["169.254.0.0/16"]` and not have to maintain the set of IPs on the svc through some external reconciliation. ### How can we reproduce it (as minimally and precisely as possible)? As above - this is a limitation currently in the validation of externalIPs rules. ### Anything else we need to know? This is related to a slew of questions people ask about exposing NodePorts on low ports. There are legitimate reasons to want to do this (again, largely around services that don't behave like a cloudy rest service, but those definitely exist), but there's always been a reasonable degree of pushback from supporting this - also for legitimate reasons. ### Kubernetes version all current ### Cloud provider n/a ### OS version n/a ### Install tools n/a ### Container runtime (CRI) and version (if applicable) n/a ### Related plugins (CNI, CSI, ...) and versions (if applicable) n/a --- end body ---
1,927
[ -0.013959730975329876, -0.0034372047521173954, -0.019351165741682053, -0.012206525541841984, 0.022158930078148842, -0.004116077441722155, -0.007948742248117924, 0.038649603724479675, -0.0014112971257418394, -0.010545594617724419, -0.02673308178782463, 0.035169556736946106, -0.011323331855237...
null
null
null
null
null
null
null
null
null
[ "strukturag", "libheif" ]
When there are several encoders / decoders for the same format, we might want to prefer hardware codecs over software, but they may have limitations on their maximum supported sizes, bpp, ... (see comment in: https://github.com/strukturag/libheif/pull/1072/#issuecomment-1869384941). We need to extend the plugin API to allow querying whether a format can be processed. Moreover the user must be enabled to specify a list of codecs to use, ordered by priority. Those will be queried after each other until a plugin is found that can process the data.
Provide a way to specify list of codecs in order of priority
https://api.github.com/repos/strukturag/libheif/issues/1111/comments
0
2024-01-16T11:17:42
2024-01-16T11:17:42Z
https://github.com/strukturag/libheif/issues/1111
2,083,688,526
1,111
false
This is a GitHub Issue repo:libheif owner:strukturag Title : Provide a way to specify list of codecs in order of priority Issue date: --- start body --- When there are several encoders / decoders for the same format, we might want to prefer hardware codecs over software, but they may have limitations on their maximum supported sizes, bpp, ... (see comment in: https://github.com/strukturag/libheif/pull/1072/#issuecomment-1869384941). We need to extend the plugin API to allow querying whether a format can be processed. Moreover the user must be enabled to specify a list of codecs to use, ordered by priority. Those will be queried after each other until a plugin is found that can process the data. --- end body ---
725
[ -0.0013216607039794326, 0.008409286849200726, -0.023388108238577843, -0.0015877550467848778, 0.022669125348329544, 0.024699196219444275, 0.005480486433953047, 0.048580724745988846, -0.0009057781426236033, 0.05661642178893089, -0.005550974979996681, 0.012455330230295658, -0.001987777650356292...
null
null
null
null
null
null
null
null
null
[ "espeak-ng", "espeak-ng" ]
Getting a bunch of warnings like this: ``` mbrola: Warning: m-X unkown, replaced with _-_ ``` https://odo.lv/Espeak?eslang=-vmb-hb1&speed=160&output=-w&debug=-x&text=%D7%9C%D7%A8%D7%90%D7%95%D7%AA+%D7%90%D7%AA+%D7%A2%D7%A6%D7%9E%D7%9A+%D7%9E%D7%97%D7%93%D7%A9%2C+%D7%91%D7%A2%D7%99%D7%A0%D7%99%D7%99%D7%9D+%D7%91%D7%A8%D7%99%D7%90%D7%95%D7%AA+%D7%99%D7%95%D7%AA%D7%A8&formmethod=get
Hebrew MBROLA voices have missing phonemes
https://api.github.com/repos/espeak-ng/espeak-ng/issues/1897/comments
0
2024-04-23T21:33:36
2024-04-23T21:33:36Z
https://github.com/espeak-ng/espeak-ng/issues/1897
2,259,805,211
1,897
false
This is a GitHub Issue repo:espeak-ng owner:espeak-ng Title : Hebrew MBROLA voices have missing phonemes Issue date: --- start body --- Getting a bunch of warnings like this: ``` mbrola: Warning: m-X unkown, replaced with _-_ ``` https://odo.lv/Espeak?eslang=-vmb-hb1&speed=160&output=-w&debug=-x&text=%D7%9C%D7%A8%D7%90%D7%95%D7%AA+%D7%90%D7%AA+%D7%A2%D7%A6%D7%9E%D7%9A+%D7%9E%D7%97%D7%93%D7%A9%2C+%D7%91%D7%A2%D7%99%D7%A0%D7%99%D7%99%D7%9D+%D7%91%D7%A8%D7%99%D7%90%D7%95%D7%AA+%D7%99%D7%95%D7%AA%D7%A8&formmethod=get --- end body ---
543
[ -0.02516935206949711, 0.00788025464862585, -0.018552450463175774, -0.013478096574544907, 0.030543837696313858, -0.00429609976708889, 0.00006576330633834004, 0.021204793825745583, -0.025909215211868286, 0.04137657210230827, -0.011830851435661316, 0.011935549788177013, -0.029678337275981903, ...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
when trying to create a file with video that has negative composition time offsets: mp4Box -dash 1000 -out out.mpd -dash-profile onDemand --cmaf=cmfc vvc_uhd2_hdr-ClosedGOP-7680x4320_50fps-HLG10-30s-12Mbps.mp4 the segment index box signals SAP_type=1 and the mpd signal subsegmentStartsWithSAP="1" This should for this specific file be 2, as the first decoded sample is not the first one displayed. also it may be easier to always signal 2 for dash as sap type 1 is a subset of sap type 1. My suggestion is to put the correct SapType in the mpd and segment index or always put 2
mp4Box incorrect sap type in segment index box and mpd
https://api.github.com/repos/gpac/gpac/issues/2370/comments
7
2023-01-05T13:02:22
2023-05-10T14:07:29Z
https://github.com/gpac/gpac/issues/2370
1,520,712,339
2,370
false
This is a GitHub Issue repo:gpac owner:gpac Title : mp4Box incorrect sap type in segment index box and mpd Issue date: --- start body --- when trying to create a file with video that has negative composition time offsets: mp4Box -dash 1000 -out out.mpd -dash-profile onDemand --cmaf=cmfc vvc_uhd2_hdr-ClosedGOP-7680x4320_50fps-HLG10-30s-12Mbps.mp4 the segment index box signals SAP_type=1 and the mpd signal subsegmentStartsWithSAP="1" This should for this specific file be 2, as the first decoded sample is not the first one displayed. also it may be easier to always signal 2 for dash as sap type 1 is a subset of sap type 1. My suggestion is to put the correct SapType in the mpd and segment index or always put 2 --- end body ---
756
[ -0.01659264974296093, 0.028675908222794533, -0.024690497666597366, 0.03912372142076492, 0.027135729789733887, 0.01463169977068901, -0.02773909829556942, 0.056907229125499725, 0.0013804052723571658, 0.016433868557214737, 0.02826307713985443, 0.02051454782485962, 0.03744063898921013, -0.0045...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/2013/07/16/how-to-file-a-bug-properly/ I used command: MP4Box -dash 10000 -profile onDemand video.mp4 But the result sidx subsegment_duration is random, not same value, how can i set subsegment_duration in same value, like example pic Thakns! <img width="1440" alt="スクリーンショット 2021-11-25 23 25 00" src="https://user-images.githubusercontent.com/38561128/143467732-0367c8ca-e983-49dd-b48c-e5d339a2d3ed.png"> <img width="1227" alt="スクリーンショット 2021-11-25 23 24 13" src="https://user-images.githubusercontent.com/38561128/143467599-97e659cb-68ee-44eb-8749-e2483da7fcd8.png"> ### example pic <img width="1241" alt="スクリーンショット 2021-11-26 1 06 20" src="https://user-images.githubusercontent.com/38561128/143480660-105830c8-b203-41d3-be49-1f9207ae6d42.png">
Dsah use onDemand profile segment duration is random
https://api.github.com/repos/gpac/gpac/issues/1948/comments
2
2021-11-25T15:26:40
2021-12-05T15:15:00Z
https://github.com/gpac/gpac/issues/1948
1,063,716,935
1,948
false
This is a GitHub Issue repo:gpac owner:gpac Title : Dsah use onDemand profile segment duration is random Issue date: --- start body --- Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/2013/07/16/how-to-file-a-bug-properly/ I used command: MP4Box -dash 10000 -profile onDemand video.mp4 But the result sidx subsegment_duration is random, not same value, how can i set subsegment_duration in same value, like example pic Thakns! <img width="1440" alt="スクリーンショット 2021-11-25 23 25 00" src="https://user-images.githubusercontent.com/38561128/143467732-0367c8ca-e983-49dd-b48c-e5d339a2d3ed.png"> <img width="1227" alt="スクリーンショット 2021-11-25 23 24 13" src="https://user-images.githubusercontent.com/38561128/143467599-97e659cb-68ee-44eb-8749-e2483da7fcd8.png"> ### example pic <img width="1241" alt="スクリーンショット 2021-11-26 1 06 20" src="https://user-images.githubusercontent.com/38561128/143480660-105830c8-b203-41d3-be49-1f9207ae6d42.png"> --- end body ---
1,584
[ -0.010100147686898708, -0.006409709341824055, -0.017393818125128746, 0.00429692305624485, 0.010797803290188313, 0.00838772114366293, -0.006175836082547903, 0.05536847189068794, -0.04246184602379799, 0.017615798860788345, 0.008799972012639046, 0.0278110820800066, 0.01302554365247488, -0.026...
null
null
null
null
null
null
null
null
null
[ "ImageMagick", "ImageMagick" ]
### ImageMagick version 7.1.1-26 ### Operating system Other (enter below) ### Operating system, version and so on FreeBSD 14 ### Description Wrong image orientation using MagickBooleanType MagickAutoOrientImage(MagickWand *wand) ### Steps to Reproduce Here is the EXIF information. As a result of auto orientation the image is rotated incorrectly at 90 CW but should be rotated at 90 CCW! ``` ExifTool Version Number : 12.50 File Name : pwhc7j1vpy7h Directory : . File Size : 36 MB File Modification Date/Time : 2024:05:10 21:16:26+03:00 File Access Date/Time : 2024:05:10 23:05:59+03:00 File Inode Change Date/Time : 2024:05:10 21:16:26+03:00 File Permissions : -rw-rw-rw- File Type : DNG File Type Extension : dng MIME Type : image/x-adobe-dng Exif Byte Order : Big-endian (Motorola, MM) Make : Apple Camera Model Name : iPhone 12 Pro Preview Image Start : 63556 Orientation : Rotate 90 CW Rows Per Strip : 3024 Preview Image Length : 7549126 Software : 17.5 Modify Date : 2024:04:18 13:52:39 Subfile Type : Full-resolution image Image Width : 4032 Image Height : 3024 Bits Per Sample : 12 12 12 Compression : JPEG Photometric Interpretation : Linear Raw Samples Per Pixel : 3 Planar Configuration : Chunky Tile Width : 504 Tile Length : 504 Tile Offsets : (Binary data 427 bytes, use -b option to extract) Tile Byte Counts : (Binary data 335 bytes, use -b option to extract) Linearization Table : (Binary data 21699 bytes, use -b option to extract) Black Level : 0 0 0 White Level : 65535 65535 65535 Noise Profile : 3e-05 3e-08 Default Black Render : None Profile Gain Table Map : (Binary data 49408 bytes, use -b option to extract) Exposure Time : 1/428 F Number : 2.0 Exposure Program : Program AE ISO : 25 Exif Version : 0232 Date/Time Original : 2024:04:18 13:52:39 Create Date : 2024:04:18 13:52:39 Offset Time : +03:00 Offset Time Original : +03:00 Offset Time Digitized : +03:00 Shutter Speed Value : 1/428 Aperture Value : 2.0 Brightness Value : 7.707632833 Exposure Compensation : 0 Metering Mode : Multi-segment Flash : Off, Did not fire Focal Length : 6.0 mm Subject Area : 2004 1514 2318 1393 Run Time Flags : Valid Run Time Value : 2995265885958 Run Time Scale : 1000000000 Run Time Epoch : 0 Acceleration Vector : -0.03274015709 -0.9480307104 -0.3352252245 Focus Distance Range : 3.33 - 3.37 m Live Photo Video Index : 1112547584 Sub Sec Time Original : 136 Sub Sec Time Digitized : 136 Exif Image Width : 4032 Exif Image Height : 3024 Sensing Method : One-chip color area Scene Type : Directly photographed Exposure Mode : Auto White Balance : Auto Focal Length In 35mm Format : 52 mm Lens Info : 1.539999962-6mm f/1.6-2.4 Lens Make : Apple Lens Model : iPhone 12 Pro back triple camera 6mm f/2 Composite Image : General Composite Image GPS Latitude Ref : North GPS Longitude Ref : East GPS Altitude Ref : Above Sea Level GPS Time Stamp : 10:52:38 GPS Speed Ref : km/h GPS Speed : 0 GPS Img Direction Ref : True North GPS Img Direction : 259.9790649 GPS Dest Bearing Ref : True North GPS Dest Bearing : 259.9790649 GPS Date Stamp : 2024:04:18 GPS Horizontal Positioning Error: 6.59729063 m DNG Version : 1.6.0.0 DNG Backward Version : 1.3.0.0 Unique Camera Model : iPhone13,3 back telephoto camera Color Matrix 1 : 1.210916758 -0.5171784163 -0.2446352243 -0.4480841756 1.493248224 0.007582519203 -0.04304760322 0.1864003539 0.6076977849 Color Matrix 2 : 0.8891333342 -0.2923306227 -0.1202171519 -0.421434015 1.284245372 0.1107940823 -0.1072451025 0.2506295443 0.4410457015 Analog Balance : 2.121582031 1 1.820556641 As Shot Neutral : 1 1 1 Baseline Exposure : 4.041075706 Baseline Sharpness : 1.5 Calibration Illuminant 1 : Standard Light A Calibration Illuminant 2 : D65 Noise Reduction Applied : 0.9499999881 Profile Name : Apple Embedded Color Profile Warning : [Minor] Not decoding some large array(s). Ignore minor errors to decode Profile Tone Curve : (large array of 514 float values) Run Time Since Power Up : 0:49:55 Aperture : 2.0 Image Size : 4032x3024 Megapixels : 12.2 Preview Image : (Binary data 7549126 bytes, use -b option to extract) Scale Factor To 35 mm Equivalent: 8.7 Shutter Speed : 1/428 Create Date : 2024:04:18 13:52:39.136+03:00 Date/Time Original : 2024:04:18 13:52:39.136+03:00 Modify Date : 2024:04:18 13:52:39+03:00 GPS Altitude : 622.9 m Above Sea Level GPS Date/Time : 2024:04:18 10:52:38Z GPS Latitude : 42 deg 38' 20.73" N GPS Longitude : 23 deg 23' 0.78" E Circle Of Confusion : 0.003 mm Field Of View : 38.2 deg Focal Length : 6.0 mm (35 mm equivalent: 52.0 mm) GPS Position : 42 deg 38' 20.73" N, 23 deg 23' 0.78" E Hyperfocal Distance : 5.19 m Light Value : 12.7 Lens ID : iPhone 12 Pro back triple camera 6mm f/2 ``` ### Images _No response_
Wrong image auto orientation ?
https://api.github.com/repos/ImageMagick/ImageMagick/issues/7306/comments
44
2024-05-10T22:11:41
2024-07-07T19:15:56Z
https://github.com/ImageMagick/ImageMagick/issues/7306
2,290,473,790
7,306
false
This is a GitHub Issue repo:ImageMagick owner:ImageMagick Title : Wrong image auto orientation ? Issue date: --- start body --- ### ImageMagick version 7.1.1-26 ### Operating system Other (enter below) ### Operating system, version and so on FreeBSD 14 ### Description Wrong image orientation using MagickBooleanType MagickAutoOrientImage(MagickWand *wand) ### Steps to Reproduce Here is the EXIF information. As a result of auto orientation the image is rotated incorrectly at 90 CW but should be rotated at 90 CCW! ``` ExifTool Version Number : 12.50 File Name : pwhc7j1vpy7h Directory : . File Size : 36 MB File Modification Date/Time : 2024:05:10 21:16:26+03:00 File Access Date/Time : 2024:05:10 23:05:59+03:00 File Inode Change Date/Time : 2024:05:10 21:16:26+03:00 File Permissions : -rw-rw-rw- File Type : DNG File Type Extension : dng MIME Type : image/x-adobe-dng Exif Byte Order : Big-endian (Motorola, MM) Make : Apple Camera Model Name : iPhone 12 Pro Preview Image Start : 63556 Orientation : Rotate 90 CW Rows Per Strip : 3024 Preview Image Length : 7549126 Software : 17.5 Modify Date : 2024:04:18 13:52:39 Subfile Type : Full-resolution image Image Width : 4032 Image Height : 3024 Bits Per Sample : 12 12 12 Compression : JPEG Photometric Interpretation : Linear Raw Samples Per Pixel : 3 Planar Configuration : Chunky Tile Width : 504 Tile Length : 504 Tile Offsets : (Binary data 427 bytes, use -b option to extract) Tile Byte Counts : (Binary data 335 bytes, use -b option to extract) Linearization Table : (Binary data 21699 bytes, use -b option to extract) Black Level : 0 0 0 White Level : 65535 65535 65535 Noise Profile : 3e-05 3e-08 Default Black Render : None Profile Gain Table Map : (Binary data 49408 bytes, use -b option to extract) Exposure Time : 1/428 F Number : 2.0 Exposure Program : Program AE ISO : 25 Exif Version : 0232 Date/Time Original : 2024:04:18 13:52:39 Create Date : 2024:04:18 13:52:39 Offset Time : +03:00 Offset Time Original : +03:00 Offset Time Digitized : +03:00 Shutter Speed Value : 1/428 Aperture Value : 2.0 Brightness Value : 7.707632833 Exposure Compensation : 0 Metering Mode : Multi-segment Flash : Off, Did not fire Focal Length : 6.0 mm Subject Area : 2004 1514 2318 1393 Run Time Flags : Valid Run Time Value : 2995265885958 Run Time Scale : 1000000000 Run Time Epoch : 0 Acceleration Vector : -0.03274015709 -0.9480307104 -0.3352252245 Focus Distance Range : 3.33 - 3.37 m Live Photo Video Index : 1112547584 Sub Sec Time Original : 136 Sub Sec Time Digitized : 136 Exif Image Width : 4032 Exif Image Height : 3024 Sensing Method : One-chip color area Scene Type : Directly photographed Exposure Mode : Auto White Balance : Auto Focal Length In 35mm Format : 52 mm Lens Info : 1.539999962-6mm f/1.6-2.4 Lens Make : Apple Lens Model : iPhone 12 Pro back triple camera 6mm f/2 Composite Image : General Composite Image GPS Latitude Ref : North GPS Longitude Ref : East GPS Altitude Ref : Above Sea Level GPS Time Stamp : 10:52:38 GPS Speed Ref : km/h GPS Speed : 0 GPS Img Direction Ref : True North GPS Img Direction : 259.9790649 GPS Dest Bearing Ref : True North GPS Dest Bearing : 259.9790649 GPS Date Stamp : 2024:04:18 GPS Horizontal Positioning Error: 6.59729063 m DNG Version : 1.6.0.0 DNG Backward Version : 1.3.0.0 Unique Camera Model : iPhone13,3 back telephoto camera Color Matrix 1 : 1.210916758 -0.5171784163 -0.2446352243 -0.4480841756 1.493248224 0.007582519203 -0.04304760322 0.1864003539 0.6076977849 Color Matrix 2 : 0.8891333342 -0.2923306227 -0.1202171519 -0.421434015 1.284245372 0.1107940823 -0.1072451025 0.2506295443 0.4410457015 Analog Balance : 2.121582031 1 1.820556641 As Shot Neutral : 1 1 1 Baseline Exposure : 4.041075706 Baseline Sharpness : 1.5 Calibration Illuminant 1 : Standard Light A Calibration Illuminant 2 : D65 Noise Reduction Applied : 0.9499999881 Profile Name : Apple Embedded Color Profile Warning : [Minor] Not decoding some large array(s). Ignore minor errors to decode Profile Tone Curve : (large array of 514 float values) Run Time Since Power Up : 0:49:55 Aperture : 2.0 Image Size : 4032x3024 Megapixels : 12.2 Preview Image : (Binary data 7549126 bytes, use -b option to extract) Scale Factor To 35 mm Equivalent: 8.7 Shutter Speed : 1/428 Create Date : 2024:04:18 13:52:39.136+03:00 Date/Time Original : 2024:04:18 13:52:39.136+03:00 Modify Date : 2024:04:18 13:52:39+03:00 GPS Altitude : 622.9 m Above Sea Level GPS Date/Time : 2024:04:18 10:52:38Z GPS Latitude : 42 deg 38' 20.73" N GPS Longitude : 23 deg 23' 0.78" E Circle Of Confusion : 0.003 mm Field Of View : 38.2 deg Focal Length : 6.0 mm (35 mm equivalent: 52.0 mm) GPS Position : 42 deg 38' 20.73" N, 23 deg 23' 0.78" E Hyperfocal Distance : 5.19 m Light Value : 12.7 Lens ID : iPhone 12 Pro back triple camera 6mm f/2 ``` ### Images _No response_ --- end body ---
6,987
[ 0.006815267726778984, 0.011061200872063637, 0.0004753993998747319, 0.010016048327088356, 0.02461915649473667, 0.010640236549079418, 0.004362061619758606, 0.028407834470272064, -0.017796630039811134, 0.029002992436289787, 0.011917646043002605, -0.028233643621206284, -0.007809614297002554, 0...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
I am trying to generate **LL-HLS** files from an **RTMP Stream**, but I couldn't find any examples on how to generate _multiple qualities_. It works perfectly fine, meaning the LL-HLS .m3u8 files are generated, but only a master file and two separated files. Current Command: `gpac -i rtmp://[server]:1935/live/[key] reframer:rt=on -o res/live.m3u8:segdur=2:cdur=0.5:profile=live:dmode=dynamic:llhls=sf:ts` Output Tree: res/ -- list of segments.ts -- live.m3u8 -- live_1.m3u8 -- live_2.m3u8 Is there such a command or parameters to specify different output qualities, like res/720p, res/1080p, res/480p ... ?
Generate multiple LL-HLS qualities
https://api.github.com/repos/gpac/gpac/issues/2453/comments
46
2023-05-02T10:03:31
2023-05-22T16:02:01Z
https://github.com/gpac/gpac/issues/2453
1,692,134,878
2,453
false
This is a GitHub Issue repo:gpac owner:gpac Title : Generate multiple LL-HLS qualities Issue date: --- start body --- I am trying to generate **LL-HLS** files from an **RTMP Stream**, but I couldn't find any examples on how to generate _multiple qualities_. It works perfectly fine, meaning the LL-HLS .m3u8 files are generated, but only a master file and two separated files. Current Command: `gpac -i rtmp://[server]:1935/live/[key] reframer:rt=on -o res/live.m3u8:segdur=2:cdur=0.5:profile=live:dmode=dynamic:llhls=sf:ts` Output Tree: res/ -- list of segments.ts -- live.m3u8 -- live_1.m3u8 -- live_2.m3u8 Is there such a command or parameters to specify different output qualities, like res/720p, res/1080p, res/480p ... ? --- end body ---
766
[ -0.022850101813673973, 0.016105934977531433, -0.03285856917500496, 0.009030717425048351, -0.004245900083333254, 0.0022403569892048836, 0.003941796720027924, 0.06349988281726837, -0.030779888853430748, 0.037046730518341064, 0.01636769436299801, -0.006959734950214624, 0.02040187641978264, 0....
null
null
null
null
null
null
null
null
null
[ "kubernetes", "kubernetes" ]
### What happened? The `evictionMessage` is not accounting for the restartable init containers (sidecars). /kind bug ### What did you expect to happen? When pod is evicted, requests of the sidecar containers must be added to the annotations. ### How can we reproduce it (as minimally and precisely as possible)? Evict pod with the sidecars and check out annotation. ### Anything else we need to know? /sig node /priority backlog KEP: https://github.com/kubernetes/enhancements/issues/753
[Sidecar Containers] Eviction message should account for the sidecar containers
https://api.github.com/repos/kubernetes/kubernetes/issues/124938/comments
3
2024-05-17T22:21:10
2024-05-20T12:28:38Z
https://github.com/kubernetes/kubernetes/issues/124938
2,303,643,487
124,938
false
This is a GitHub Issue repo:kubernetes owner:kubernetes Title : [Sidecar Containers] Eviction message should account for the sidecar containers Issue date: --- start body --- ### What happened? The `evictionMessage` is not accounting for the restartable init containers (sidecars). /kind bug ### What did you expect to happen? When pod is evicted, requests of the sidecar containers must be added to the annotations. ### How can we reproduce it (as minimally and precisely as possible)? Evict pod with the sidecars and check out annotation. ### Anything else we need to know? /sig node /priority backlog KEP: https://github.com/kubernetes/enhancements/issues/753 --- end body ---
708
[ -0.013673393055796623, -0.024436386302113533, -0.014593190513551235, 0.017023099586367607, 0.0001458632614230737, -0.031465280801057816, -0.026907481253147125, 0.02751152589917183, -0.041432030498981476, -0.03591325134038925, 0.010660029947757721, 0.002630686853080988, -0.003994937054812908,...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [✓ ] I looked for a similar issue and couldn't find any. - [✓ ] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [ ✓] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/2013/07/16/how-to-file-a-bug-properly/ === Using latest nightly on MacOS. The following command works … `mp4box frame_0000003.heic -add-image frame_0000003_zbuffer.hvc:ID=50:auxt=urn:mpeg:hevc:2015:auxid:2` However for my purposes the HEIC requires a Aux Subtype otherwise the depth channel is not seen. In an Apple reference HEIC this looks like the following (using "_mp4box -diso_" XML dump) … (from [https://files.michaelzfreeman.org/IMG_0019_A.HEIC](https://files.michaelzfreeman.org/IMG_0019_A.HEIC) ) ``` <AuxiliaryTypePropertyBox Size="60" Type="auxC" Version="0" Flags="0" Specification="iff" Container="ipco" aux_type="urn:mpeg:hevc:2015:auxid:2" aux_subtype="0x000000110000000D4E01B109351F7B760103C7A020" > </AuxiliaryTypePropertyBox> ``` Unless I’m missing a function can a way be added to allow the setting of “aux_subtype=“. Cheers.
Can’t create auxiliary image subtype
https://api.github.com/repos/gpac/gpac/issues/2402/comments
1
2023-02-18T14:35:44
2023-03-13T09:59:26Z
https://github.com/gpac/gpac/issues/2402
1,590,363,242
2,402
false
This is a GitHub Issue repo:gpac owner:gpac Title : Can’t create auxiliary image subtype Issue date: --- start body --- Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [✓ ] I looked for a similar issue and couldn't find any. - [✓ ] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [ ✓] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/2013/07/16/how-to-file-a-bug-properly/ === Using latest nightly on MacOS. The following command works … `mp4box frame_0000003.heic -add-image frame_0000003_zbuffer.hvc:ID=50:auxt=urn:mpeg:hevc:2015:auxid:2` However for my purposes the HEIC requires a Aux Subtype otherwise the depth channel is not seen. In an Apple reference HEIC this looks like the following (using "_mp4box -diso_" XML dump) … (from [https://files.michaelzfreeman.org/IMG_0019_A.HEIC](https://files.michaelzfreeman.org/IMG_0019_A.HEIC) ) ``` <AuxiliaryTypePropertyBox Size="60" Type="auxC" Version="0" Flags="0" Specification="iff" Container="ipco" aux_type="urn:mpeg:hevc:2015:auxid:2" aux_subtype="0x000000110000000D4E01B109351F7B760103C7A020" > </AuxiliaryTypePropertyBox> ``` Unless I’m missing a function can a way be added to allow the setting of “aux_subtype=“. Cheers. --- end body ---
1,676
[ 0.003949221223592758, 0.021935101598501205, -0.01761728711426258, 0.001748000388033688, 0.0614122673869133, 0.00445697782561183, 0.002668545115739107, 0.04549502581357956, 0.0033906882163137197, 0.02047577127814293, -0.0024090248625725508, 0.01752701960504055, 0.013510096818208694, -0.0173...
null
null
null
null
null
null
null
null
null
[ "ImageMagick", "ImageMagick" ]
### Is your feature request related to a problem? Please describe. Currently, VisualMagick does not permit to reproduce a Windows build since `CloneRepositories.sh` always retrieves the latest git checkout. This means that Windows builds that do not always use the latest ImageMagick version break from time to time - the latest example is the `oss-fuzz` directory in ImageMagick. ### Describe the solution you'd like `CloneRepositories.sh` should use the date of the latest commit in the `ImageMagick-Windows` repository to retrieve the dependencies up to this point in time. ### Describe alternatives you've considered I do not see any solution at the moment, besides to always build only the latest Windows version. ### Additional context `magickwand.js` has the option to be built from source - which few Windows users would use - that currently breaks with each new change in the build process.
Reproducible Windows builds
https://api.github.com/repos/ImageMagick/ImageMagick/issues/7041/comments
3
2024-01-18T15:20:32
2024-01-18T16:42:28Z
https://github.com/ImageMagick/ImageMagick/issues/7041
2,088,513,357
7,041
false
This is a GitHub Issue repo:ImageMagick owner:ImageMagick Title : Reproducible Windows builds Issue date: --- start body --- ### Is your feature request related to a problem? Please describe. Currently, VisualMagick does not permit to reproduce a Windows build since `CloneRepositories.sh` always retrieves the latest git checkout. This means that Windows builds that do not always use the latest ImageMagick version break from time to time - the latest example is the `oss-fuzz` directory in ImageMagick. ### Describe the solution you'd like `CloneRepositories.sh` should use the date of the latest commit in the `ImageMagick-Windows` repository to retrieve the dependencies up to this point in time. ### Describe alternatives you've considered I do not see any solution at the moment, besides to always build only the latest Windows version. ### Additional context `magickwand.js` has the option to be built from source - which few Windows users would use - that currently breaks with each new change in the build process. --- end body ---
1,053
[ -0.03254827484488487, 0.014658165164291859, -0.022223208099603653, 0.01494417805224657, 0.032605480402708054, 0.01953468658030033, -0.0029048193246126175, 0.050223879516124725, -0.014558060094714165, 0.018805352970957756, 0.03629504516720772, 0.00602057296782732, 0.0055987038649618626, -0....
null
null
null
null
null
null
null
null
null
[ "LibreDWG", "libredwg" ]
Building 0ad71990 on Ubuntu 23.04 with GCC 12.3.0 fails: ``` escape.c: In function ‘htmlescape’: escape.c:41:25: error: pointer ‘dest’ used after ‘realloc’ [-Werror=use-after-free] 41 | const int off = d - dest; | ~~^~~~~~ escape.c:45:25: note: call to ‘realloc’ here 45 | dest = (char*)realloc (dest, len); | ^~~~~~~~~~~~~~~~~~~ cc1: all warnings being treated as errors ``` Please let me know what other information I should provide to help you figure this out.
realloc UAF (gcc 12)
https://api.github.com/repos/LibreDWG/libredwg/issues/801/comments
3
2023-08-13T14:13:32
2023-08-21T08:34:50Z
https://github.com/LibreDWG/libredwg/issues/801
1,848,611,259
801
false
This is a GitHub Issue repo:libredwg owner:LibreDWG Title : realloc UAF (gcc 12) Issue date: --- start body --- Building 0ad71990 on Ubuntu 23.04 with GCC 12.3.0 fails: ``` escape.c: In function ‘htmlescape’: escape.c:41:25: error: pointer ‘dest’ used after ‘realloc’ [-Werror=use-after-free] 41 | const int off = d - dest; | ~~^~~~~~ escape.c:45:25: note: call to ‘realloc’ here 45 | dest = (char*)realloc (dest, len); | ^~~~~~~~~~~~~~~~~~~ cc1: all warnings being treated as errors ``` Please let me know what other information I should provide to help you figure this out. --- end body ---
689
[ -0.034492406994104385, 0.0304050724953413, -0.013408812694251537, 0.030316855758428574, 0.058839987963438034, 0.024465203285217285, -0.005976625718176365, 0.052664875984191895, -0.03284570947289467, 0.004094686824828386, -0.008652507327497005, 0.013967514038085938, -0.007215323392301798, 0...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
I just obtained 32 bit from https://gpac.wp.imt.fr/downloads/ To see if a bug has since been fixed. Mp4box seems to not be 32 bit. Ffmpeg runs. Installer ran. Cannot locate previous version of installer to see if my bug is there. As for the bug, frame rate specification is being ignored with one video specifically, it wants 29.08, I want 29.04 and 5fps test fails. Using: -add AVfile -fps 5 AVoutfile Also tried: …\mp4box" -add E:\copy.mp4#audio -add E:\copy.mp4#video:fps=5.0 C:\Recov\out.mp4 The first was tried on a short clip with success so shouldn't be command issue… N.
32 bit issue
https://api.github.com/repos/gpac/gpac/issues/2326/comments
11
2022-11-27T03:09:17
2022-12-12T10:14:37Z
https://github.com/gpac/gpac/issues/2326
1,465,316,125
2,326
false
This is a GitHub Issue repo:gpac owner:gpac Title : 32 bit issue Issue date: --- start body --- I just obtained 32 bit from https://gpac.wp.imt.fr/downloads/ To see if a bug has since been fixed. Mp4box seems to not be 32 bit. Ffmpeg runs. Installer ran. Cannot locate previous version of installer to see if my bug is there. As for the bug, frame rate specification is being ignored with one video specifically, it wants 29.08, I want 29.04 and 5fps test fails. Using: -add AVfile -fps 5 AVoutfile Also tried: …\mp4box" -add E:\copy.mp4#audio -add E:\copy.mp4#video:fps=5.0 C:\Recov\out.mp4 The first was tried on a short clip with success so shouldn't be command issue… N. --- end body ---
709
[ -0.02888351120054722, 0.040011655539274216, -0.019431572407484055, 0.034175973385572433, -0.018158897757530212, 0.05990884080529213, -0.011485114693641663, 0.08126494288444519, -0.01111262384802103, 0.022054523229599, 0.0052497838623821735, 0.011244547553360462, 0.02841789834201336, -0.023...
null
null
null
null
null
null
null
null
null
[ "LibreDWG", "libredwg" ]
The actual situation with parsing of POLYLINE and VERTEX entities in preR13 is this: * In preR13 is one entity for all POLYLINE and VERTEX types. This means, that the DWG file contains a flag for the (un)presence of variables in the entity and an optional flag (default without this flag is POLYLINE_2D, VERTEX_2D) which means which entity type is present (3D, MESH, PFACE) * We have naive parsing (polyline - https://github.com/LibreDWG/libredwg/blob/master/src/decode.c#L6614, vertex - https://github.com/LibreDWG/libredwg/blob/master/src/decode.c#L6630) based on entity variables which is not valid for detection. * There are issues like the entity is presented as POLYLINE_3D, but it's POLYLINE_MESH. Same with VERTEX. I created [PR](https://github.com/LibreDWG/libredwg/pull/711) for a fix of the polyline, but the issue is with the naivety of parsing. What about this? Possibilities: * Add better detection in the actual block - this means adding parsing of entity flag, which is after elevation, handle, and other common entity things, and will create detection based on this flag. * Add code for common parsing of preR13 POLYLINE and VERTEX entities - different that POLYLINE_3D (and others) in dwg.spec. And we do some convert to concrete entities. Example of failing POLYLINE MESH: [PFACE.DXF.gz](https://github.com/LibreDWG/libredwg/files/11341290/PFACE.DXF.gz) [PFACE.DWG.gz](https://github.com/LibreDWG/libredwg/files/11341291/PFACE.DWG.gz) There are other failing situations.
Wrong POLYLINE and VERTEX parsing in preR13
https://api.github.com/repos/LibreDWG/libredwg/issues/713/comments
7
2023-04-27T08:32:06
2023-05-08T09:17:14Z
https://github.com/LibreDWG/libredwg/issues/713
1,686,381,600
713
false
This is a GitHub Issue repo:libredwg owner:LibreDWG Title : Wrong POLYLINE and VERTEX parsing in preR13 Issue date: --- start body --- The actual situation with parsing of POLYLINE and VERTEX entities in preR13 is this: * In preR13 is one entity for all POLYLINE and VERTEX types. This means, that the DWG file contains a flag for the (un)presence of variables in the entity and an optional flag (default without this flag is POLYLINE_2D, VERTEX_2D) which means which entity type is present (3D, MESH, PFACE) * We have naive parsing (polyline - https://github.com/LibreDWG/libredwg/blob/master/src/decode.c#L6614, vertex - https://github.com/LibreDWG/libredwg/blob/master/src/decode.c#L6630) based on entity variables which is not valid for detection. * There are issues like the entity is presented as POLYLINE_3D, but it's POLYLINE_MESH. Same with VERTEX. I created [PR](https://github.com/LibreDWG/libredwg/pull/711) for a fix of the polyline, but the issue is with the naivety of parsing. What about this? Possibilities: * Add better detection in the actual block - this means adding parsing of entity flag, which is after elevation, handle, and other common entity things, and will create detection based on this flag. * Add code for common parsing of preR13 POLYLINE and VERTEX entities - different that POLYLINE_3D (and others) in dwg.spec. And we do some convert to concrete entities. Example of failing POLYLINE MESH: [PFACE.DXF.gz](https://github.com/LibreDWG/libredwg/files/11341290/PFACE.DXF.gz) [PFACE.DWG.gz](https://github.com/LibreDWG/libredwg/files/11341291/PFACE.DWG.gz) There are other failing situations. --- end body ---
1,665
[ -0.05015974119305611, 0.003099515801295638, -0.01926310546696186, 0.012185255065560341, 0.030360672622919083, 0.05908193439245224, -0.009095591492950916, 0.03978729993104935, 0.0006906422786414623, 0.01905817911028862, 0.02973012998700142, -0.033797137439250946, 0.04070158675312996, -0.021...
null
null
null
null
null
null
null
null
null
[ "Piwigo", "Piwigo" ]
On several test environments, I can't manage my albums. I get this in the javascript console: ``` TypeError: undefined is not an object (evaluating 'node.parent') ``` It refers to this piece of code in `albums.js` : ``` var url_split = window.location.href.split("cat_move"); var catToOpen = url_split[url_split.length - 1].split("-")[1]; console.log(catToOpen); if (catToOpen && isNumeric(catToOpen)) { nodeToGo = $('.tree').tree('getNodeById', catToOpen); goToNode(nodeToGo, nodeToGo); if (nodeToGo.children) { $(".tree").tree("openNode", nodeToGo, false); } } ``` The problem is that it tries to find an album id in the URL which is supposed to be like `http://domain.tld/admin.php?page=albums&parent_id=2#cat-2` and it assumes there is only one legit "-" in this URL. Which is totally not sure. In my case, the domain name is something like `server.ip-1-2-3-4.tld`. Good luck with that. The search for the album id should not rely on parsing the URL.
[albums] undefined node.parent
https://api.github.com/repos/Piwigo/Piwigo/issues/2064/comments
0
2023-12-12T17:42:35
2024-07-15T14:47:35Z
https://github.com/Piwigo/Piwigo/issues/2064
2,038,268,059
2,064
false
This is a GitHub Issue repo:Piwigo owner:Piwigo Title : [albums] undefined node.parent Issue date: --- start body --- On several test environments, I can't manage my albums. I get this in the javascript console: ``` TypeError: undefined is not an object (evaluating 'node.parent') ``` It refers to this piece of code in `albums.js` : ``` var url_split = window.location.href.split("cat_move"); var catToOpen = url_split[url_split.length - 1].split("-")[1]; console.log(catToOpen); if (catToOpen && isNumeric(catToOpen)) { nodeToGo = $('.tree').tree('getNodeById', catToOpen); goToNode(nodeToGo, nodeToGo); if (nodeToGo.children) { $(".tree").tree("openNode", nodeToGo, false); } } ``` The problem is that it tries to find an album id in the URL which is supposed to be like `http://domain.tld/admin.php?page=albums&parent_id=2#cat-2` and it assumes there is only one legit "-" in this URL. Which is totally not sure. In my case, the domain name is something like `server.ip-1-2-3-4.tld`. Good luck with that. The search for the album id should not rely on parsing the URL. --- end body ---
1,183
[ -0.047903865575790405, 0.033308155834674835, -0.008093133568763733, 0.00444420613348484, 0.019959943369030952, -0.04063719883561134, -0.02002231776714325, 0.04768555238842964, -0.00978505052626133, -0.001732850563712418, 0.01748054474592209, -0.009972174651920795, -0.0484340526163578, -0.0...
null
null
null
null
null
null
null
null
null
[ "ImageMagick", "ImageMagick" ]
### ImageMagick version 7.1.0-62 ### Operating system Linux ### Operating system, version and so on ubuntu 20.04 ### Description kernal: Linux 708bdfa6cac2 5.4.0-81-generic #91-Ubuntu SMP Thu Jul 15 19:09:17 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux root@708bdfa6cac2:/imgs# magick -version Version: ImageMagick 7.1.0-62 Q16-HDRI x86_64 74b3683a4:20230211 https://imagemagick.org Copyright: (C) 1999 ImageMagick Studio LLC License: https://imagemagick.org/script/license.php Features: Cipher DPC HDRI OpenMP(4.5) TCMalloc Delegates (built-in): bzlib cairo djvu fftw fontconfig freetype gslib gvc heic jbig jng jpeg lcms lqr lzma openexr pangocairo png ps raqm raw tiff webp wmf x zip zlib Compiler: gcc (9.4) **others:** root@708bdfa6cac2:/imgs# identify -list format | grep HEIC HEIC rw+ High Efficiency Image Format (1.6.1) error message: root@708bdfa6cac2:/imgs# magick a1.heic a1.jpg magick: Unsupported feature: Unsupported color conversion (4.3003) `a1.heic' @ error/heic.c/IsHEIFSuccess/139. note: the online tools is ok (https://heic.online/) ### Steps to Reproduce magick a1.heic a1.jpg ### Images [images.zip](https://github.com/ImageMagick/ImageMagick/files/10857717/images.zip)
heic convert to jpg was failed
https://api.github.com/repos/ImageMagick/ImageMagick/issues/6119/comments
2
2023-03-01T06:24:19
2023-03-02T02:05:01Z
https://github.com/ImageMagick/ImageMagick/issues/6119
1,604,330,333
6,119
false
This is a GitHub Issue repo:ImageMagick owner:ImageMagick Title : heic convert to jpg was failed Issue date: --- start body --- ### ImageMagick version 7.1.0-62 ### Operating system Linux ### Operating system, version and so on ubuntu 20.04 ### Description kernal: Linux 708bdfa6cac2 5.4.0-81-generic #91-Ubuntu SMP Thu Jul 15 19:09:17 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux root@708bdfa6cac2:/imgs# magick -version Version: ImageMagick 7.1.0-62 Q16-HDRI x86_64 74b3683a4:20230211 https://imagemagick.org Copyright: (C) 1999 ImageMagick Studio LLC License: https://imagemagick.org/script/license.php Features: Cipher DPC HDRI OpenMP(4.5) TCMalloc Delegates (built-in): bzlib cairo djvu fftw fontconfig freetype gslib gvc heic jbig jng jpeg lcms lqr lzma openexr pangocairo png ps raqm raw tiff webp wmf x zip zlib Compiler: gcc (9.4) **others:** root@708bdfa6cac2:/imgs# identify -list format | grep HEIC HEIC rw+ High Efficiency Image Format (1.6.1) error message: root@708bdfa6cac2:/imgs# magick a1.heic a1.jpg magick: Unsupported feature: Unsupported color conversion (4.3003) `a1.heic' @ error/heic.c/IsHEIFSuccess/139. note: the online tools is ok (https://heic.online/) ### Steps to Reproduce magick a1.heic a1.jpg ### Images [images.zip](https://github.com/ImageMagick/ImageMagick/files/10857717/images.zip) --- end body ---
1,415
[ -0.017623277381062508, 0.03442440181970596, 0.003848876105621457, 0.009129892103374004, 0.007684491574764252, 0.020023437216877937, 0.0065374537371098995, 0.03333703428506851, -0.013989885337650776, 0.008002744987607002, -0.0011884772684425116, -0.03938384726643562, -0.03410614654421806, 0...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
<img width="465" alt="mediainfo" src="https://user-images.githubusercontent.com/40682277/205034819-d4c0ef9d-12d6-4b67-b13d-60df67ac459b.png"> I tried to remux this mkv to mp4 using the following command `MP4Box -add input.mkv -new output.mp4` and it gave me `segmentation fault (core dumped)` My environment is Ubuntu 22.04.1 LTS (GNU/Linux 5.15.0-53-generic x86_64). And the build I tried was gpac_2.1-DEV-rev523-g4112fc35-master_amd64
bug report : segmentation fault (core dumped)
https://api.github.com/repos/gpac/gpac/issues/2334/comments
15
2022-12-01T10:57:20
2023-01-10T07:02:30Z
https://github.com/gpac/gpac/issues/2334
1,471,101,874
2,334
false
This is a GitHub Issue repo:gpac owner:gpac Title : bug report : segmentation fault (core dumped) Issue date: --- start body --- <img width="465" alt="mediainfo" src="https://user-images.githubusercontent.com/40682277/205034819-d4c0ef9d-12d6-4b67-b13d-60df67ac459b.png"> I tried to remux this mkv to mp4 using the following command `MP4Box -add input.mkv -new output.mp4` and it gave me `segmentation fault (core dumped)` My environment is Ubuntu 22.04.1 LTS (GNU/Linux 5.15.0-53-generic x86_64). And the build I tried was gpac_2.1-DEV-rev523-g4112fc35-master_amd64 --- end body ---
593
[ -0.01677410863339901, 0.02161254733800888, -0.016029734164476395, 0.010466170497238636, -0.0026630661450326443, 0.017287470400333405, -0.03501129895448685, 0.05241427570581436, -0.015131349675357342, 0.03883584588766098, -0.026925845071673393, 0.006917554885149002, 0.011467226780951023, -0...
null
null
null
null
null
null
null
null
null
[ "hwchase17", "langchain" ]
### Checked other resources - [X] I added a very descriptive title to this issue. - [X] I searched the LangChain documentation with the integrated search. - [X] I used the GitHub search to find a similar question and didn't find it. - [X] I am sure that this is a bug in LangChain rather than my code. - [X] The bug is not resolved by updating to the latest stable version of LangChain (or the specific integration package). ### Example Code Dockerfile: ``` # LLM Installs RUN pip3 install \ langchain \ langchain-community \ langchain-pinecone \ langchain-openai ``` Python Imports ``` python import langchain from langchain_community.document_loaders import PyPDFLoader, TextLoader, Docx2txtLoader, UnstructuredHTMLLoader from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain_pinecone import PineconeVectorStore from langchain_openai import OpenAIEmbeddings, OpenAI, ChatOpenAI from langchain.prompts import PromptTemplate from langchain.chains import RetrievalQAWithSourcesChain, ConversationalRetrievalChain from langchain.memory import ConversationBufferMemory from langchain.load.dump import dumps ``` ### Error Message and Stack Trace (if applicable) ``` 2024-05-30 13:28:59 from langchain_openai import OpenAIEmbeddings, OpenAI, ChatOpenAI 2024-05-30 13:28:59 File "/usr/local/lib/python3.9/site-packages/langchain_openai/__init__.py", line 1, in <module> 2024-05-30 13:28:59 from langchain_openai.chat_models import ( 2024-05-30 13:28:59 File "/usr/local/lib/python3.9/site-packages/langchain_openai/chat_models/__init__.py", line 1, in <module> 2024-05-30 13:28:59 from langchain_openai.chat_models.azure import AzureChatOpenAI 2024-05-30 13:28:59 File "/usr/local/lib/python3.9/site-packages/langchain_openai/chat_models/azure.py", line 9, in <module> 2024-05-30 13:28:59 from langchain_core.language_models.chat_models import LangSmithParams 2024-05-30 13:28:59 ImportError: cannot import name 'LangSmithParams' from 'langchain_core.language_models.chat_models' (/usr/local/lib/python3.9/site-packages/langchain_core/language_models/chat_models.py) ``` ### Description I am trying to import langchain_openai with the newest version released last night (0.1.8) and it can not find the LangSmithParams module. I move back a version with ``` langchain-openai==0.1.7 ``` and it works again. Something in this new update broke the import. ### System Info Container is running python 3.9 on Rocky Linux 8 ``` # Install dependecies RUN dnf -y install epel-release RUN dnf -y install \ httpd \ python39 \ unzip \ xz \ git-core \ ImageMagick \ wget RUN pip3 install \ psycopg2-binary \ pillow \ lxml \ pycryptodomex \ six \ pytz \ jaraco.functools \ requests \ supervisor \ flask \ flask-cors \ flask-socketio \ mako \ boto3 \ botocore==1.34.33 \ gotenberg-client \ docusign-esign \ python-dotenv \ htmldocx \ python-docx \ beautifulsoup4 \ pypandoc \ pyetherpadlite \ html2text \ PyJWT \ sendgrid \ auth0-python \ authlib \ openai==0.27.7 \ pinecone-client==3.1.0 \ pinecone-datasets==0.7.0 \ tiktoken==0.4.0 # Installing LLM requirements RUN pip3 install \ langchain \ langchain-community \ langchain-pinecone \ langchain-openai==0.1.7 \ pinecone-client \ pinecone-datasets \ unstructured \ poppler-utils \ tiktoken \ pypdf \ python-dotenv \ docx2txt ```
langchain-openai==0.1.8 is now broken
https://api.github.com/repos/langchain-ai/langchain/issues/22333/comments
10
2024-05-30T20:51:56
2024-07-03T17:07:55Z
https://github.com/langchain-ai/langchain/issues/22333
2,326,523,092
22,333
false
This is a GitHub Issue repo:langchain owner:hwchase17 Title : langchain-openai==0.1.8 is now broken Issue date: --- start body --- ### Checked other resources - [X] I added a very descriptive title to this issue. - [X] I searched the LangChain documentation with the integrated search. - [X] I used the GitHub search to find a similar question and didn't find it. - [X] I am sure that this is a bug in LangChain rather than my code. - [X] The bug is not resolved by updating to the latest stable version of LangChain (or the specific integration package). ### Example Code Dockerfile: ``` # LLM Installs RUN pip3 install \ langchain \ langchain-community \ langchain-pinecone \ langchain-openai ``` Python Imports ``` python import langchain from langchain_community.document_loaders import PyPDFLoader, TextLoader, Docx2txtLoader, UnstructuredHTMLLoader from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain_pinecone import PineconeVectorStore from langchain_openai import OpenAIEmbeddings, OpenAI, ChatOpenAI from langchain.prompts import PromptTemplate from langchain.chains import RetrievalQAWithSourcesChain, ConversationalRetrievalChain from langchain.memory import ConversationBufferMemory from langchain.load.dump import dumps ``` ### Error Message and Stack Trace (if applicable) ``` 2024-05-30 13:28:59 from langchain_openai import OpenAIEmbeddings, OpenAI, ChatOpenAI 2024-05-30 13:28:59 File "/usr/local/lib/python3.9/site-packages/langchain_openai/__init__.py", line 1, in <module> 2024-05-30 13:28:59 from langchain_openai.chat_models import ( 2024-05-30 13:28:59 File "/usr/local/lib/python3.9/site-packages/langchain_openai/chat_models/__init__.py", line 1, in <module> 2024-05-30 13:28:59 from langchain_openai.chat_models.azure import AzureChatOpenAI 2024-05-30 13:28:59 File "/usr/local/lib/python3.9/site-packages/langchain_openai/chat_models/azure.py", line 9, in <module> 2024-05-30 13:28:59 from langchain_core.language_models.chat_models import LangSmithParams 2024-05-30 13:28:59 ImportError: cannot import name 'LangSmithParams' from 'langchain_core.language_models.chat_models' (/usr/local/lib/python3.9/site-packages/langchain_core/language_models/chat_models.py) ``` ### Description I am trying to import langchain_openai with the newest version released last night (0.1.8) and it can not find the LangSmithParams module. I move back a version with ``` langchain-openai==0.1.7 ``` and it works again. Something in this new update broke the import. ### System Info Container is running python 3.9 on Rocky Linux 8 ``` # Install dependecies RUN dnf -y install epel-release RUN dnf -y install \ httpd \ python39 \ unzip \ xz \ git-core \ ImageMagick \ wget RUN pip3 install \ psycopg2-binary \ pillow \ lxml \ pycryptodomex \ six \ pytz \ jaraco.functools \ requests \ supervisor \ flask \ flask-cors \ flask-socketio \ mako \ boto3 \ botocore==1.34.33 \ gotenberg-client \ docusign-esign \ python-dotenv \ htmldocx \ python-docx \ beautifulsoup4 \ pypandoc \ pyetherpadlite \ html2text \ PyJWT \ sendgrid \ auth0-python \ authlib \ openai==0.27.7 \ pinecone-client==3.1.0 \ pinecone-datasets==0.7.0 \ tiktoken==0.4.0 # Installing LLM requirements RUN pip3 install \ langchain \ langchain-community \ langchain-pinecone \ langchain-openai==0.1.7 \ pinecone-client \ pinecone-datasets \ unstructured \ poppler-utils \ tiktoken \ pypdf \ python-dotenv \ docx2txt ``` --- end body ---
3,799
[ -0.017101500183343887, 0.007448992226272821, -0.004074425436556339, 0.02507018856704235, -0.005089567042887211, -0.06230821833014488, 0.010712689720094204, 0.032151930034160614, 0.00803105253726244, 0.03423071652650833, -0.005740921013057232, 0.016949055716395378, -0.007670729421079159, 0....
null
null
null
null
null
null
null
null
null
[ "jeecgboot", "jeecg-boot" ]
##### 版本号: ##### 分支: master?还是springboot3? ##### 前端版本: vue3版?还是 vue2版? ##### 问题描述:想要排查修复jeecg-boot jmreport/queryFieldBySql接口调用控制不严。怎么确定是否是3.4.0-3.5.3版本,是否有漏洞补丁或升级版本可解决。 ##### 截图&代码: #### 友情提示(为了提高issue处理效率): - 未按格式要求发帖、描述过于简抽象的,会被直接删掉; - 请自己初判问题描述是否清楚,是否方便我们调查处理; - 针对问题请说明是Online在线功能(需说明用的主题模板),还是生成的代码功能; - springboot3_sas分支采用 `Spring Authorization Server` 替换 `Shiro`,目前是beta版不稳定,生产项目不要使用;
想要排查修复jeecg-boot jmreport/queryFieldBySql接口调用控制不严。怎么确定是否是3.4.0-3.5.3版本,是否有漏洞补丁或升级版本可解决。
https://api.github.com/repos/jeecgboot/JeecgBoot/issues/6236/comments
1
2024-05-20T03:11:07
2024-05-20T03:40:57Z
https://github.com/jeecgboot/JeecgBoot/issues/6236
2,304,981,351
6,236
false
This is a GitHub Issue repo:jeecg-boot owner:jeecgboot Title : 想要排查修复jeecg-boot jmreport/queryFieldBySql接口调用控制不严。怎么确定是否是3.4.0-3.5.3版本,是否有漏洞补丁或升级版本可解决。 Issue date: --- start body --- ##### 版本号: ##### 分支: master?还是springboot3? ##### 前端版本: vue3版?还是 vue2版? ##### 问题描述:想要排查修复jeecg-boot jmreport/queryFieldBySql接口调用控制不严。怎么确定是否是3.4.0-3.5.3版本,是否有漏洞补丁或升级版本可解决。 ##### 截图&代码: #### 友情提示(为了提高issue处理效率): - 未按格式要求发帖、描述过于简抽象的,会被直接删掉; - 请自己初判问题描述是否清楚,是否方便我们调查处理; - 针对问题请说明是Online在线功能(需说明用的主题模板),还是生成的代码功能; - springboot3_sas分支采用 `Spring Authorization Server` 替换 `Shiro`,目前是beta版不稳定,生产项目不要使用; --- end body ---
638
[ 0.017715008929371834, -0.016892526298761368, -0.005781103391200304, 0.018031349405646324, 0.029340485110878944, 0.0006311960751190782, 0.007540741935372353, 0.039637334644794464, 0.011728285811841488, 0.05219601094722748, 0.0220172256231308, -0.01676599122583866, 0.054030779749155045, 0.01...
null
null
null
null
null
null
null
null
null
[ "kubernetes", "kubernetes" ]
### What happened? [root@m01 log]# kubectl get pod -A runtime: failed to create new OS thread (have 3 already; errno=11) runtime: may need to increase max user processes (ulimit -u) fatal error: newosproc runtime stack: runtime.throw(0x1c1d651, 0x9) /usr/local/go/src/runtime/panic.go:1116 +0x72 runtime.newosproc(0xc0000b6400) /usr/local/go/src/runtime/os_linux.go:161 +0x1c5 runtime.newm1(0xc0000b6400) /usr/local/go/src/runtime/proc.go:1843 +0xdd runtime.newm(0x1d07a08, 0xc00004a800, 0x2) /usr/local/go/src/runtime/proc.go:1822 +0x9b runtime.startm(0x0, 0xc000048001) /usr/local/go/src/runtime/proc.go:1979 +0xc9 runtime.wakep() /usr/local/go/src/runtime/proc.go:2067 +0x66 runtime.newproc.func1() /usr/local/go/src/runtime/proc.go:3561 +0x97 runtime.systemstack(0x46c414) /usr/local/go/src/runtime/asm_amd64.s:370 +0x66 runtime.mstart() /usr/local/go/src/runtime/proc.go:1116 goroutine 1 [running, locked to thread]: runtime.systemstack_switch() /usr/local/go/src/runtime/asm_amd64.s:330 fp=0xc0000b26f0 sp=0xc0000b26e8 pc=0x46c540 runtime.newproc(0x0, 0x1d078c8) /usr/local/go/src/runtime/proc.go:3554 +0x6e fp=0xc0000b2738 sp=0xc0000b26f0 pc=0x44214e runtime.init.6() /usr/local/go/src/runtime/proc.go:243 +0x35 fp=0xc0000b2758 sp=0xc0000b2738 pc=0x439e35 runtime.doInit(0x2b17e80) /usr/local/go/src/runtime/proc.go:5652 +0x8a fp=0xc0000b2788 sp=0xc0000b2758 pc=0x446fca runtime.main() /usr/local/go/src/runtime/proc.go:151 +0xd9 fp=0xc0000b27e0 sp=0xc0000b2788 pc=0x439b59 runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:1374 +0x1 fp=0xc0000b27e8 sp=0xc0000b27e0 pc=0x46e181 ### What did you expect to happen? Output Pod information. ### How can we reproduce it (as minimally and precisely as possible)? I don't know how to repeat the problem. ### Anything else we need to know? $ ulimit -u 655350 $ ulimit -n 655350 systemctl edit kubelet.service [Service] LimitNOFILE=infinity LimitNPROC=infinity ### Kubernetes version <details> ```console $ kubectl version v 1.19.16 ``` </details> ### Cloud provider <details> Physical server </details> ### OS version <details> ```console # On Linux: $ cat /etc/os-release CentOS Linux 7 (Core) $ uname -a linux m01 4.20.3-1.el7.elrepo.x86_64 x86_64 GNU/Linux ``` </details> ### Install tools <details> kubeadm </details> ### Container runtime (CRI) and version (if applicable) <details> Docker v24.0.6 </details> ### Related plugins (CNI, CSI, ...) and versions (if applicable) <details> </details>
runtime: failed to create new OS thread
https://api.github.com/repos/kubernetes/kubernetes/issues/125242/comments
6
2024-05-31T13:28:50
2024-05-31T18:54:01Z
https://github.com/kubernetes/kubernetes/issues/125242
2,327,844,151
125,242
false
This is a GitHub Issue repo:kubernetes owner:kubernetes Title : runtime: failed to create new OS thread Issue date: --- start body --- ### What happened? [root@m01 log]# kubectl get pod -A runtime: failed to create new OS thread (have 3 already; errno=11) runtime: may need to increase max user processes (ulimit -u) fatal error: newosproc runtime stack: runtime.throw(0x1c1d651, 0x9) /usr/local/go/src/runtime/panic.go:1116 +0x72 runtime.newosproc(0xc0000b6400) /usr/local/go/src/runtime/os_linux.go:161 +0x1c5 runtime.newm1(0xc0000b6400) /usr/local/go/src/runtime/proc.go:1843 +0xdd runtime.newm(0x1d07a08, 0xc00004a800, 0x2) /usr/local/go/src/runtime/proc.go:1822 +0x9b runtime.startm(0x0, 0xc000048001) /usr/local/go/src/runtime/proc.go:1979 +0xc9 runtime.wakep() /usr/local/go/src/runtime/proc.go:2067 +0x66 runtime.newproc.func1() /usr/local/go/src/runtime/proc.go:3561 +0x97 runtime.systemstack(0x46c414) /usr/local/go/src/runtime/asm_amd64.s:370 +0x66 runtime.mstart() /usr/local/go/src/runtime/proc.go:1116 goroutine 1 [running, locked to thread]: runtime.systemstack_switch() /usr/local/go/src/runtime/asm_amd64.s:330 fp=0xc0000b26f0 sp=0xc0000b26e8 pc=0x46c540 runtime.newproc(0x0, 0x1d078c8) /usr/local/go/src/runtime/proc.go:3554 +0x6e fp=0xc0000b2738 sp=0xc0000b26f0 pc=0x44214e runtime.init.6() /usr/local/go/src/runtime/proc.go:243 +0x35 fp=0xc0000b2758 sp=0xc0000b2738 pc=0x439e35 runtime.doInit(0x2b17e80) /usr/local/go/src/runtime/proc.go:5652 +0x8a fp=0xc0000b2788 sp=0xc0000b2758 pc=0x446fca runtime.main() /usr/local/go/src/runtime/proc.go:151 +0xd9 fp=0xc0000b27e0 sp=0xc0000b2788 pc=0x439b59 runtime.goexit() /usr/local/go/src/runtime/asm_amd64.s:1374 +0x1 fp=0xc0000b27e8 sp=0xc0000b27e0 pc=0x46e181 ### What did you expect to happen? Output Pod information. ### How can we reproduce it (as minimally and precisely as possible)? I don't know how to repeat the problem. ### Anything else we need to know? $ ulimit -u 655350 $ ulimit -n 655350 systemctl edit kubelet.service [Service] LimitNOFILE=infinity LimitNPROC=infinity ### Kubernetes version <details> ```console $ kubectl version v 1.19.16 ``` </details> ### Cloud provider <details> Physical server </details> ### OS version <details> ```console # On Linux: $ cat /etc/os-release CentOS Linux 7 (Core) $ uname -a linux m01 4.20.3-1.el7.elrepo.x86_64 x86_64 GNU/Linux ``` </details> ### Install tools <details> kubeadm </details> ### Container runtime (CRI) and version (if applicable) <details> Docker v24.0.6 </details> ### Related plugins (CNI, CSI, ...) and versions (if applicable) <details> </details> --- end body ---
2,880
[ -0.02526189386844635, -0.033707547932863235, 0.006737756077200174, -0.003956946078687906, 0.04028890281915665, -0.009102540090680122, -0.012299376539885998, 0.04486832767724991, 0.0017939197132363915, -0.00565233314409852, -0.0064124418422579765, -0.010660293512046337, 0.004770231433212757, ...
null
null
null
null
null
null
null
null
null
[ "dromara", "hutool" ]
### 版本情况 JDK版本: openjdk_11.0.16 hutool版本: 5.8.27(请确保最新尝试是否还有问题) ### 问题描述(包括截图) 1. 复现代码 ```java public static void main(String[] args) { String read = IoUtil.read(FileUtil.getInputStream("data.json"), StandardCharsets.UTF_8); } ``` 2. 堆栈信息 ```java Exception in thread "main" java.lang.NegativeArraySizeException: -221445712 at cn.hutool.core.io.FastByteBuffer.toArray(FastByteBuffer.java:210) at cn.hutool.core.io.FastByteArrayOutputStream.toByteArray(FastByteArrayOutputStream.java:93) at cn.hutool.core.io.IoUtil.readBytes(IoUtil.java:491) at cn.hutool.core.io.IoUtil.readBytes(IoUtil.java:457) at cn.hutool.core.io.IoUtil.read(IoUtil.java:368) ``` 3. 测试涉及到的文件(注意脱密) 比如报错的Excel文件,有问题的图片等。
读取文件超过3GB时报java.lang.NegativeArraySizeException: -221445712
https://api.github.com/repos/dromara/hutool/issues/3577/comments
1
2024-05-08T12:33:18
2024-05-09T01:28:08Z
https://github.com/dromara/hutool/issues/3577
2,285,488,997
3,577
false
This is a GitHub Issue repo:hutool owner:dromara Title : 读取文件超过3GB时报java.lang.NegativeArraySizeException: -221445712 Issue date: --- start body --- ### 版本情况 JDK版本: openjdk_11.0.16 hutool版本: 5.8.27(请确保最新尝试是否还有问题) ### 问题描述(包括截图) 1. 复现代码 ```java public static void main(String[] args) { String read = IoUtil.read(FileUtil.getInputStream("data.json"), StandardCharsets.UTF_8); } ``` 2. 堆栈信息 ```java Exception in thread "main" java.lang.NegativeArraySizeException: -221445712 at cn.hutool.core.io.FastByteBuffer.toArray(FastByteBuffer.java:210) at cn.hutool.core.io.FastByteArrayOutputStream.toByteArray(FastByteArrayOutputStream.java:93) at cn.hutool.core.io.IoUtil.readBytes(IoUtil.java:491) at cn.hutool.core.io.IoUtil.readBytes(IoUtil.java:457) at cn.hutool.core.io.IoUtil.read(IoUtil.java:368) ``` 3. 测试涉及到的文件(注意脱密) 比如报错的Excel文件,有问题的图片等。 --- end body ---
908
[ -0.029534775763750076, 0.006086916197091341, -0.0024536699056625366, 0.0035141543485224247, 0.032090526074171066, -0.01446493249386549, -0.010222994722425938, 0.053383391350507736, -0.002818506909534335, 0.010502765886485577, 0.002695634262636304, 0.01505472045391798, 0.029338181018829346, ...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
## Description There is a segment fault detected by AddressSanitizer ## System info ``` Ubuntu 20.04.2 LTS clang version 12.0.0-++20210402082642+04ba60cfe598-1~exp1~20210402063359.71 MP4Box - GPAC version 1.1.0-DEV-rev1687-g16bd902cb-master ``` ## Build command ``` ./configure --static-mp4box --prefix=`realpath ./install` --enable-sanitizer --cc=clang --cxx=clang++ ``` ## crash command ``` MP4Box -frag 1 -out /dev/null poc_file ``` ## Pocs [poc.zip](https://github.com/gpac/gpac/files/7976877/poc.zip) ## Crash output ``` ==15641==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000008 (pc 0x000000d776b8 bp 0x7ffea0e4f210 sp 0x7ffea0e4f0a0 T0) ==15641==The signal is caused by a READ memory access. ==15641==Hint: address points to the zero page. #0 0xd776b8 in mp4_mux_cenc_update /programs/mp4box/builds/build6/src/filters/mux_isom.c:3582:184 #1 0xd6ab36 in mp4_mux_config_timing /programs/mp4box/builds/build6/src/filters/mux_isom.c:5863:6 #2 0xd6ab36 in mp4_mux_process /programs/mp4box/builds/build6/src/filters/mux_isom.c:6033:3 #3 0xb7c2ae in gf_filter_process_task /programs/mp4box/builds/build6/src/filter_core/filter.c:2515:7 #4 0xb5d7eb in gf_fs_thread_proc /programs/mp4box/builds/build6/src/filter_core/filter_session.c:1756:3 #5 0xb5c694 in gf_fs_run /programs/mp4box/builds/build6/src/filter_core/filter_session.c:2000:2 #6 0x910850 in gf_media_fragment_file /programs/mp4box/builds/build6/src/media_tools/isom_tools.c:3816:6 #7 0x4e6fcc in mp4boxMain /programs/mp4box/builds/build6/applications/mp4box/main.c:6455:7 #8 0x7ff7959040b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x270b2) #9 0x41ea6d in _start (/programs/mp4box/builds/build6/bin/gcc/MP4Box+0x41ea6d) AddressSanitizer can not provide additional info. SUMMARY: AddressSanitizer: SEGV /programs/mp4box/builds/build6/src/filters/mux_isom.c:3582:184 in mp4_mux_cenc_update ==15641==ABORTING /collect/collec.sh: line 18: 15641 Aborted(core dumped) ------------------------ ```
There is a segment fault detected by AddressSanitizer
https://api.github.com/repos/gpac/gpac/issues/2089/comments
0
2022-02-01T08:21:39
2022-02-01T14:29:20Z
https://github.com/gpac/gpac/issues/2089
1,120,363,059
2,089
false
This is a GitHub Issue repo:gpac owner:gpac Title : There is a segment fault detected by AddressSanitizer Issue date: --- start body --- ## Description There is a segment fault detected by AddressSanitizer ## System info ``` Ubuntu 20.04.2 LTS clang version 12.0.0-++20210402082642+04ba60cfe598-1~exp1~20210402063359.71 MP4Box - GPAC version 1.1.0-DEV-rev1687-g16bd902cb-master ``` ## Build command ``` ./configure --static-mp4box --prefix=`realpath ./install` --enable-sanitizer --cc=clang --cxx=clang++ ``` ## crash command ``` MP4Box -frag 1 -out /dev/null poc_file ``` ## Pocs [poc.zip](https://github.com/gpac/gpac/files/7976877/poc.zip) ## Crash output ``` ==15641==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000008 (pc 0x000000d776b8 bp 0x7ffea0e4f210 sp 0x7ffea0e4f0a0 T0) ==15641==The signal is caused by a READ memory access. ==15641==Hint: address points to the zero page. #0 0xd776b8 in mp4_mux_cenc_update /programs/mp4box/builds/build6/src/filters/mux_isom.c:3582:184 #1 0xd6ab36 in mp4_mux_config_timing /programs/mp4box/builds/build6/src/filters/mux_isom.c:5863:6 #2 0xd6ab36 in mp4_mux_process /programs/mp4box/builds/build6/src/filters/mux_isom.c:6033:3 #3 0xb7c2ae in gf_filter_process_task /programs/mp4box/builds/build6/src/filter_core/filter.c:2515:7 #4 0xb5d7eb in gf_fs_thread_proc /programs/mp4box/builds/build6/src/filter_core/filter_session.c:1756:3 #5 0xb5c694 in gf_fs_run /programs/mp4box/builds/build6/src/filter_core/filter_session.c:2000:2 #6 0x910850 in gf_media_fragment_file /programs/mp4box/builds/build6/src/media_tools/isom_tools.c:3816:6 #7 0x4e6fcc in mp4boxMain /programs/mp4box/builds/build6/applications/mp4box/main.c:6455:7 #8 0x7ff7959040b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x270b2) #9 0x41ea6d in _start (/programs/mp4box/builds/build6/bin/gcc/MP4Box+0x41ea6d) AddressSanitizer can not provide additional info. SUMMARY: AddressSanitizer: SEGV /programs/mp4box/builds/build6/src/filters/mux_isom.c:3582:184 in mp4_mux_cenc_update ==15641==ABORTING /collect/collec.sh: line 18: 15641 Aborted(core dumped) ------------------------ ``` --- end body ---
2,224
[ -0.014343371614813805, 0.01596115343272686, -0.011382251977920532, 0.020106719806790352, 0.028672298416495323, 0.004918346181511879, -0.04067566245794296, 0.040242329239845276, -0.007460575085133314, 0.02259117178618908, -0.032702308148145676, 0.010320582427084446, 0.02475784346461296, 0.0...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
[POC](https://cdn.discordapp.com/attachments/1002845697801977896/1004011628288802887/null_dereference) ``` Version 06:02:21 root@ubuntu gcc ±|master ✗|→ ./MP4Box -version MP4Box - GPAC version 2.1-DEV-rev259-gfc4749f9c-master (c) 2000-2022 Telecom Paris distributed under LGPL v2.1+ - http://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: Features: GPAC_CONFIG_LINUX GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SSL GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_FAAD GPAC_HAS_JPEG GPAC_HAS_PNG GPAC_HAS_LINUX_DVB Proof of Conect ./MP4Box -info ./null_dereference ASAN Log [iso file] extra box udta found in moov, deleting [iso file] extra box udta found in moov, deleting [iso file] extra box udta found in moov, deleting [iso file] extra box udta found in moov, deleting [iso file] Unknown box type 0000 in parent minf [iso file] Unknown box type s0000s in parent stbl [ISO file] dataReferenceIndex set to 0 in sample entry, overriding to 1 [iso file] Read Box type FFFFFFEA (0xFFFFFFEA) at position 764 has size 0 but is not at root/file level. Forbidden, skipping end of parent box ! [iso file] Box "dims" (start 748) has 10 extra bytes [iso file] Read Box type 00000000 (0x00000000) at position 782 has size 0 but is not at root/file level. Forbidden, skipping end of parent box ! [iso file] Box "stsd" (start 732) has 48 extra bytes [iso file] Missing DataInformationBox [iso file] Unknown box type 00000000 in parent stbl [iso file] Unknown box type url in parent dinf [iso file] Missing dref box in dinf [iso file] Unknown top-level box type 000000 [iso file] Unknown top-level box type 00000008 [iso file] Incomplete box 00000000 - start 1574 size 808465447 [iso file] Incomplete file while reading for dump - aborting parsing # Movie Info - 2 tracks - TimeScale 3158064 Duration 00:04:16.000 (recomputed 00:00:09.000) Fragmented: no Progressive (moov before mdat) Major Brand 0000 - version 808464432 - compatible brands: 0000 0000 0000 0000 0000 0000 0000 0000 0000 0000 0000 Created: GMT Sat Sep 19 11:55:28 2065 Modified: GMT Sat Sep 19 11:54:46 2065 Copyrights: (lap) 000000000 1 UDTA types: cprt: # Track 1 Info - ID 12336 - TimeScale 16 Media Duration 00:12:51.000 (recomputed 00:00:09.000) Track flags: Disabled In Movie Size is AspectRatio Media Info: Language "``o (``o)" - Type "soun:dims" - 15 samples Alternate Group ID 12336 3GPP DIMS stream - size 0 x 0 - Profile 0 - Level 0 pathComponents: 0 - useFullRequestHost: no stream type: secondary - redundant: main+redundant AddressSanitizer:DEADLYSIGNAL ================================================================= ==8678==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000000 (pc 0x000000569547 bp 0x000000000001 sp 0x7fff81bca2d0 T0) ==8678==The signal is caused by a READ memory access. ==8678==Hint: address points to the zero page. #0 0x569546 in DumpTrackInfo /root/gpac_stable_asan/applications/mp4box/filedump.c:3361:12 #1 0x57267f in DumpMovieInfo /root/gpac_stable_asan/applications/mp4box/filedump.c:3994:3 #2 0x545302 in mp4box_main /root/gpac_stable_asan/applications/mp4box/mp4box.c:6367:9 #3 0x7f7368cafc86 in __libc_start_main /build/glibc-CVJwZb/glibc-2.27/csu/../csu/libc-start.c:310 #4 0x428b79 in _start (/root/gpac_stable_asan/bin/gcc/MP4Box+0x428b79) AddressSanitizer can not provide additional info. SUMMARY: AddressSanitizer: SEGV /root/gpac_stable_asan/applications/mp4box/filedump.c:3361:12 in DumpTrackInfo ==8678==ABORTING Root Cause gf_isom_get_dims_description@plt returns NULL data. Therefore, an error occurs when referring to the corresponding part. gf_isom_get_dims_description(file, trackNum, 1, &dims); fprintf(stderr, "\t3GPP DIMS stream - size %d x %d - Profile %d - Level %d\n", w, h, dims.profile, dims.level); fprintf(stderr, "\tpathComponents: %d - useFullRequestHost: %s\n", dims.pathComponents, dims.fullRequestHost ? "yes" : "no"); fprintf(stderr, "\tstream type: %s - redundant: %s\n", dims.streamType ? "primary" : "secondary", (dims.containsRedundant==1) ? "main" : ((dims.containsRedundant==2) ? "redundant" : "main+redundant") ); if (dims.textEncoding[0]) fprintf(stderr, "\ttext encoding %s\n", dims.textEncoding); => 0x4589f6 <DumpTrackInfo+13174>: call 0x40f500 <gf_isom_get_dims_description@plt> Guessed arguments: arg[0]: 0x8c8990 --> 0x0 arg[1]: 0x1 arg[2]: 0x1 arg[3]: 0x7ffffffd2ee0 --> 0x0 typedef struct { /*! profile*/ u8 profile; /*! level*/ u8 level; /*! number of components in path*/ u8 pathComponents; /*! full request*/ Bool fullRequestHost; /*! stream type*/ Bool streamType; /*! has redundant sample (carousel)*/ u8 containsRedundant; /*! text encoding string*/ const char *textEncoding; /*! content encoding string*/ const char *contentEncoding; /*! script string*/ const char *content_script_types; /*! mime type string*/ const char *mime_type; /*! xml schema location string*/ const char *xml_schema_loc; } GF_DIMSDescription; Impact Deinal of Service Occurrences isomedia.h L3680 filedump.c L3361 ```
Null Pointer Dereference Caused Segmentation Fault
https://api.github.com/repos/gpac/gpac/issues/2246/comments
0
2022-08-24T19:28:34
2022-09-12T09:40:46Z
https://github.com/gpac/gpac/issues/2246
1,349,923,512
2,246
false
This is a GitHub Issue repo:gpac owner:gpac Title : Null Pointer Dereference Caused Segmentation Fault Issue date: --- start body --- [POC](https://cdn.discordapp.com/attachments/1002845697801977896/1004011628288802887/null_dereference) ``` Version 06:02:21 root@ubuntu gcc ±|master ✗|→ ./MP4Box -version MP4Box - GPAC version 2.1-DEV-rev259-gfc4749f9c-master (c) 2000-2022 Telecom Paris distributed under LGPL v2.1+ - http://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: Features: GPAC_CONFIG_LINUX GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SSL GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_FAAD GPAC_HAS_JPEG GPAC_HAS_PNG GPAC_HAS_LINUX_DVB Proof of Conect ./MP4Box -info ./null_dereference ASAN Log [iso file] extra box udta found in moov, deleting [iso file] extra box udta found in moov, deleting [iso file] extra box udta found in moov, deleting [iso file] extra box udta found in moov, deleting [iso file] Unknown box type 0000 in parent minf [iso file] Unknown box type s0000s in parent stbl [ISO file] dataReferenceIndex set to 0 in sample entry, overriding to 1 [iso file] Read Box type FFFFFFEA (0xFFFFFFEA) at position 764 has size 0 but is not at root/file level. Forbidden, skipping end of parent box ! [iso file] Box "dims" (start 748) has 10 extra bytes [iso file] Read Box type 00000000 (0x00000000) at position 782 has size 0 but is not at root/file level. Forbidden, skipping end of parent box ! [iso file] Box "stsd" (start 732) has 48 extra bytes [iso file] Missing DataInformationBox [iso file] Unknown box type 00000000 in parent stbl [iso file] Unknown box type url in parent dinf [iso file] Missing dref box in dinf [iso file] Unknown top-level box type 000000 [iso file] Unknown top-level box type 00000008 [iso file] Incomplete box 00000000 - start 1574 size 808465447 [iso file] Incomplete file while reading for dump - aborting parsing # Movie Info - 2 tracks - TimeScale 3158064 Duration 00:04:16.000 (recomputed 00:00:09.000) Fragmented: no Progressive (moov before mdat) Major Brand 0000 - version 808464432 - compatible brands: 0000 0000 0000 0000 0000 0000 0000 0000 0000 0000 0000 Created: GMT Sat Sep 19 11:55:28 2065 Modified: GMT Sat Sep 19 11:54:46 2065 Copyrights: (lap) 000000000 1 UDTA types: cprt: # Track 1 Info - ID 12336 - TimeScale 16 Media Duration 00:12:51.000 (recomputed 00:00:09.000) Track flags: Disabled In Movie Size is AspectRatio Media Info: Language "``o (``o)" - Type "soun:dims" - 15 samples Alternate Group ID 12336 3GPP DIMS stream - size 0 x 0 - Profile 0 - Level 0 pathComponents: 0 - useFullRequestHost: no stream type: secondary - redundant: main+redundant AddressSanitizer:DEADLYSIGNAL ================================================================= ==8678==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000000 (pc 0x000000569547 bp 0x000000000001 sp 0x7fff81bca2d0 T0) ==8678==The signal is caused by a READ memory access. ==8678==Hint: address points to the zero page. #0 0x569546 in DumpTrackInfo /root/gpac_stable_asan/applications/mp4box/filedump.c:3361:12 #1 0x57267f in DumpMovieInfo /root/gpac_stable_asan/applications/mp4box/filedump.c:3994:3 #2 0x545302 in mp4box_main /root/gpac_stable_asan/applications/mp4box/mp4box.c:6367:9 #3 0x7f7368cafc86 in __libc_start_main /build/glibc-CVJwZb/glibc-2.27/csu/../csu/libc-start.c:310 #4 0x428b79 in _start (/root/gpac_stable_asan/bin/gcc/MP4Box+0x428b79) AddressSanitizer can not provide additional info. SUMMARY: AddressSanitizer: SEGV /root/gpac_stable_asan/applications/mp4box/filedump.c:3361:12 in DumpTrackInfo ==8678==ABORTING Root Cause gf_isom_get_dims_description@plt returns NULL data. Therefore, an error occurs when referring to the corresponding part. gf_isom_get_dims_description(file, trackNum, 1, &dims); fprintf(stderr, "\t3GPP DIMS stream - size %d x %d - Profile %d - Level %d\n", w, h, dims.profile, dims.level); fprintf(stderr, "\tpathComponents: %d - useFullRequestHost: %s\n", dims.pathComponents, dims.fullRequestHost ? "yes" : "no"); fprintf(stderr, "\tstream type: %s - redundant: %s\n", dims.streamType ? "primary" : "secondary", (dims.containsRedundant==1) ? "main" : ((dims.containsRedundant==2) ? "redundant" : "main+redundant") ); if (dims.textEncoding[0]) fprintf(stderr, "\ttext encoding %s\n", dims.textEncoding); => 0x4589f6 <DumpTrackInfo+13174>: call 0x40f500 <gf_isom_get_dims_description@plt> Guessed arguments: arg[0]: 0x8c8990 --> 0x0 arg[1]: 0x1 arg[2]: 0x1 arg[3]: 0x7ffffffd2ee0 --> 0x0 typedef struct { /*! profile*/ u8 profile; /*! level*/ u8 level; /*! number of components in path*/ u8 pathComponents; /*! full request*/ Bool fullRequestHost; /*! stream type*/ Bool streamType; /*! has redundant sample (carousel)*/ u8 containsRedundant; /*! text encoding string*/ const char *textEncoding; /*! content encoding string*/ const char *contentEncoding; /*! script string*/ const char *content_script_types; /*! mime type string*/ const char *mime_type; /*! xml schema location string*/ const char *xml_schema_loc; } GF_DIMSDescription; Impact Deinal of Service Occurrences isomedia.h L3680 filedump.c L3361 ``` --- end body ---
5,553
[ -0.02204272150993347, 0.021974949166178703, -0.0045618778094649315, 0.027684709057211876, 0.032682865858078, 0.0076243081130087376, -0.0319543220102787, 0.04059520363807678, -0.0028591156005859375, 0.020195944234728813, -0.003879925701767206, 0.019924858585000038, 0.023042352870106697, -0....
CVE-2022-22885
2022-02-16T22:15:07.773000
Hutool v5.7.18's HttpRequest was discovered to ignore all TLS/SSL certificate validation.
{ "cvssMetricV2": [ { "acInsufInfo": false, "baseSeverity": "HIGH", "cvssData": { "accessComplexity": "LOW", "accessVector": "NETWORK", "authentication": "NONE", "availabilityImpact": "PARTIAL", "baseScore": 7.5, "confidentialityImpact": "PARTIAL", "integrityImpact": "PARTIAL", "vectorString": "AV:N/AC:L/Au:N/C:P/I:P/A:P", "version": "2.0" }, "exploitabilityScore": 10, "impactScore": 6.4, "obtainAllPrivilege": false, "obtainOtherPrivilege": false, "obtainUserPrivilege": false, "source": "nvd@nist.gov", "type": "Primary", "userInteractionRequired": false } ], "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "LOW", "attackVector": "NETWORK", "availabilityImpact": "HIGH", "baseScore": 9.8, "baseSeverity": "CRITICAL", "confidentialityImpact": "HIGH", "integrityImpact": "HIGH", "privilegesRequired": "NONE", "scope": "UNCHANGED", "userInteraction": "NONE", "vectorString": "CVSS:3.1/AV:N/AC:L/PR:N/UI:N/S:U/C:H/I:H/A:H", "version": "3.1" }, "exploitabilityScore": 3.9, "impactScore": 5.9, "source": "nvd@nist.gov", "type": "Primary" } ] }
[ { "source": "cve@mitre.org", "tags": [ "Third Party Advisory" ], "url": "https://apidoc.gitee.com/dromara/hutool/cn/hutool/http/ssl/DefaultSSLInfo.html" }, { "source": "cve@mitre.org", "tags": [ "Exploit", "Issue Tracking", "Third Party Advisory" ], "url":...
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:hutool:hutool:5.7.18:*:*:*:*:*:*:*", "matchCriteriaId": "4C9E9670-0982-49B3-A11D-D4F294D3B8CD", "versionEndExcluding": null, "versionEndIncluding": null, "versionStartExcludi...
https://github.com/dromara/hutool/issues/2042
[ "Exploit", "Issue Tracking", "Third Party Advisory" ]
github.com
[ "dromara", "hutool" ]
### 版本情况 JDK版本: corretto_8_312 hutool版本: 5.7.18(请确保最新尝试是否还有问题) ### 问题描述(包括截图) 注意到 hutool HttpUtil 中 HttpConnection#setHttpsInfo 默认传入的 HostnameVerifier 是 DefaultSSLInfo.TRUST_ANY_HOSTNAME_VERIFIER,即: ```java package cn.hutool.http.ssl; import javax.net.ssl.HostnameVerifier; import javax.net.ssl.SSLSession; public class TrustAnyHostnameVerifier implements HostnameVerifier { public TrustAnyHostnameVerifier() { } public boolean verify(String hostname, SSLSession session) { return true; } } ``` 查看 javax.net.ssl.HttpsURLConnection.DefaultHostnameVerifier 代码如下: ```java /** * <code>HostnameVerifier</code> provides a callback mechanism so that * implementers of this interface can supply a policy for * handling the case where the host to connect to and * the server name from the certificate mismatch. * <p> * The default implementation will deny such connections. */ private static HostnameVerifier defaultHostnameVerifier = new DefaultHostnameVerifier(); /* * The initial default <code>HostnameVerifier</code>. Should be * updated for another other type of <code>HostnameVerifier</code> * that are created. */ private static class DefaultHostnameVerifier implements HostnameVerifier { @Override public boolean verify(String hostname, SSLSession session) { return false; } } /** * The <code>hostnameVerifier</code> for this object. */ protected HostnameVerifier hostnameVerifier = defaultHostnameVerifier; ``` 从 doc 看到 HostnameVerifier 是一种 fallback 机制,当要连接的 host 与服务端证书 server name 不匹配时才有作用。并且 javax.net.ssl.HttpsURLConnection 的默认机制是这时候拒绝此类请求。而 hutool 则默认放行所有这类请求。 如果调用 cn.hutool.http.HttpRequest 时手动传入上面的 javax.net.ssl.HttpsURLConnection.DefaultHostnameVerifier: ```java String result1 = HttpRequest.get("https://www.baidu.com") .setHostnameVerifier(HttpsURLConnection.getDefaultHostnameVerifier()) .execute().body(); System.out.println(result1); ``` 也会请求成功。按我理解这样才是合理的,如果服务端证书 server name 与 host 不匹配,应该中断请求才安全。请问我的理解是否正确?是不是其实 javax.net.ssl.HttpsURLConnection 的默认行为会有坑,hutool 才另外指定了 HostnameVerifier 呢? 谢谢。
默认的 HostnameVerifier 为什么选择 TrustAnyHostnameVerifier 呢?
https://api.github.com/repos/dromara/hutool/issues/2042/comments
2
2021-12-26T12:41:33
2024-03-12T02:30:59Z
https://github.com/dromara/hutool/issues/2042
1,088,763,905
2,042
true
This is a GitHub Issue repo:hutool owner:dromara Title : 默认的 HostnameVerifier 为什么选择 TrustAnyHostnameVerifier 呢? Issue date: --- start body --- ### 版本情况 JDK版本: corretto_8_312 hutool版本: 5.7.18(请确保最新尝试是否还有问题) ### 问题描述(包括截图) 注意到 hutool HttpUtil 中 HttpConnection#setHttpsInfo 默认传入的 HostnameVerifier 是 DefaultSSLInfo.TRUST_ANY_HOSTNAME_VERIFIER,即: ```java package cn.hutool.http.ssl; import javax.net.ssl.HostnameVerifier; import javax.net.ssl.SSLSession; public class TrustAnyHostnameVerifier implements HostnameVerifier { public TrustAnyHostnameVerifier() { } public boolean verify(String hostname, SSLSession session) { return true; } } ``` 查看 javax.net.ssl.HttpsURLConnection.DefaultHostnameVerifier 代码如下: ```java /** * <code>HostnameVerifier</code> provides a callback mechanism so that * implementers of this interface can supply a policy for * handling the case where the host to connect to and * the server name from the certificate mismatch. * <p> * The default implementation will deny such connections. */ private static HostnameVerifier defaultHostnameVerifier = new DefaultHostnameVerifier(); /* * The initial default <code>HostnameVerifier</code>. Should be * updated for another other type of <code>HostnameVerifier</code> * that are created. */ private static class DefaultHostnameVerifier implements HostnameVerifier { @Override public boolean verify(String hostname, SSLSession session) { return false; } } /** * The <code>hostnameVerifier</code> for this object. */ protected HostnameVerifier hostnameVerifier = defaultHostnameVerifier; ``` 从 doc 看到 HostnameVerifier 是一种 fallback 机制,当要连接的 host 与服务端证书 server name 不匹配时才有作用。并且 javax.net.ssl.HttpsURLConnection 的默认机制是这时候拒绝此类请求。而 hutool 则默认放行所有这类请求。 如果调用 cn.hutool.http.HttpRequest 时手动传入上面的 javax.net.ssl.HttpsURLConnection.DefaultHostnameVerifier: ```java String result1 = HttpRequest.get("https://www.baidu.com") .setHostnameVerifier(HttpsURLConnection.getDefaultHostnameVerifier()) .execute().body(); System.out.println(result1); ``` 也会请求成功。按我理解这样才是合理的,如果服务端证书 server name 与 host 不匹配,应该中断请求才安全。请问我的理解是否正确?是不是其实 javax.net.ssl.HttpsURLConnection 的默认行为会有坑,hutool 才另外指定了 HostnameVerifier 呢? 谢谢。 --- end body ---
2,515
[ 0.01825021766126156, -0.0025317552499473095, -0.00196528946980834, 0.0011724301148205996, 0.02982615865767002, 0.012770527973771095, 0.009179058484733105, 0.03283189609646797, -0.017047923058271408, -0.0012620241614058614, -0.00040654517943039536, 0.013125050812959671, -0.0008930506883189082...
null
null
null
null
null
null
null
null
null
[ "chakra-core", "ChakraCore" ]
I want to run js script in two threads . How can I do it?
How to create one more runtimes in different threads
https://api.github.com/repos/chakra-core/ChakraCore/issues/6956/comments
0
2023-10-10T05:31:59
2023-10-14T02:25:28Z
https://github.com/chakra-core/ChakraCore/issues/6956
1,934,408,999
6,956
false
This is a GitHub Issue repo:ChakraCore owner:chakra-core Title : How to create one more runtimes in different threads Issue date: --- start body --- I want to run js script in two threads . How can I do it? --- end body ---
225
[ 0.01032658014446497, -0.05351301282644272, -0.012859118171036243, -0.0013697451213374734, 0.03137259930372238, 0.001253992086276412, -0.011961154639720917, 0.0674033835530281, -0.039987437427043915, 0.01693502999842167, 0.03530118986964226, 0.014914613217115402, -0.013911419548094273, -0.0...
null
null
null
null
null
null
null
null
null
[ "ImageMagick", "ImageMagick" ]
### ImageMagick version 7.1.1-29 ### Operating system Windows ### Operating system, version and so on Windows 10 x64 ### Description Tiff image decoding seems to have totally no or totally incorrect respect of Orientation tag (274) depending of source file. I have produced 2 test files. One from LZW-compressed image and another from JPEG-compressed image. Both contain 5 frames that should produce clockwise rotation of image by changing the orientation tag in it's IFD: TopLeft -> RightTop -> BottomRight -> LeftBottom -> TopLeft When I view this images in some modern viewer, then it displays every frame according the embedded Orientation tag. (Checked in FastStone Image Viewer, XnView MP, IrfanView 64). When I convert by ImageMagick such LZW-compressed tiff image to PNG's or PDF file then it produces 5 pages **without any rotation**. When I convert such JPEG-compressed tiff image to PNG's or PDF file then it produces 5 pages **with page horizontal and vertical flipping** instead of rotation. I'm sure that decoding of source tiff _should have the correct processing of orientation tag in any compatible compression format_ of tiff page, producing the different image orientation in target file pages. Especially when converting to pdf. P.S. Also I checked these test images with ImageGlass viewer, which is based on ImageMagick's dll. It seems that viewer's programmers have implemented rotating of tiff page _after decoding_, but combining with unexpected flipping of jpeg-compressed pages **produces unpredictable results**. ### Steps to Reproduce ``` magick convert test_RotateClockwize_jpeg.tiff test_RotateClockwize_jpeg.pdf magick convert test_RotateClockwize_lzw.tiff test_RotateClockwize_lzw.pdf magick identify -verbose test_RotateClockwize_jpeg.tiff > test_RotateClockwize_jpeg.tiff.txt magick identify -verbose test_RotateClockwize_lzw.tiff > test_RotateClockwize_lzw.tiff.txt ``` ### Images [Tiff rotation.zip](https://github.com/ImageMagick/ImageMagick/files/14496282/Tiff.rotation.zip)
Orientation tag bug in tiff image decoding.
https://api.github.com/repos/ImageMagick/ImageMagick/issues/7147/comments
9
2024-03-05T14:19:46
2024-03-23T21:58:08Z
https://github.com/ImageMagick/ImageMagick/issues/7147
2,169,354,723
7,147
false
This is a GitHub Issue repo:ImageMagick owner:ImageMagick Title : Orientation tag bug in tiff image decoding. Issue date: --- start body --- ### ImageMagick version 7.1.1-29 ### Operating system Windows ### Operating system, version and so on Windows 10 x64 ### Description Tiff image decoding seems to have totally no or totally incorrect respect of Orientation tag (274) depending of source file. I have produced 2 test files. One from LZW-compressed image and another from JPEG-compressed image. Both contain 5 frames that should produce clockwise rotation of image by changing the orientation tag in it's IFD: TopLeft -> RightTop -> BottomRight -> LeftBottom -> TopLeft When I view this images in some modern viewer, then it displays every frame according the embedded Orientation tag. (Checked in FastStone Image Viewer, XnView MP, IrfanView 64). When I convert by ImageMagick such LZW-compressed tiff image to PNG's or PDF file then it produces 5 pages **without any rotation**. When I convert such JPEG-compressed tiff image to PNG's or PDF file then it produces 5 pages **with page horizontal and vertical flipping** instead of rotation. I'm sure that decoding of source tiff _should have the correct processing of orientation tag in any compatible compression format_ of tiff page, producing the different image orientation in target file pages. Especially when converting to pdf. P.S. Also I checked these test images with ImageGlass viewer, which is based on ImageMagick's dll. It seems that viewer's programmers have implemented rotating of tiff page _after decoding_, but combining with unexpected flipping of jpeg-compressed pages **produces unpredictable results**. ### Steps to Reproduce ``` magick convert test_RotateClockwize_jpeg.tiff test_RotateClockwize_jpeg.pdf magick convert test_RotateClockwize_lzw.tiff test_RotateClockwize_lzw.pdf magick identify -verbose test_RotateClockwize_jpeg.tiff > test_RotateClockwize_jpeg.tiff.txt magick identify -verbose test_RotateClockwize_lzw.tiff > test_RotateClockwize_lzw.tiff.txt ``` ### Images [Tiff rotation.zip](https://github.com/ImageMagick/ImageMagick/files/14496282/Tiff.rotation.zip) --- end body ---
2,239
[ -0.010117294266819954, 0.013442558236420155, -0.006901692599058151, 0.026602115482091904, 0.009756467305123806, 0.06446768343448639, -0.0012310553574934602, 0.04264828562736511, -0.029149126261472702, 0.03605435788631439, 0.02041853964328766, -0.03763916343450546, -0.0142986373975873, -0.0...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
## Description There is a segment fault detected by AddressSanitizer ## System info ``` Ubuntu 20.04.2 LTS clang version 12.0.0-++20210402082642+04ba60cfe598-1~exp1~20210402063359.71 MP4Box - GPAC version 1.1.0-DEV-rev1740-gbd98129fd-master ``` ## Build command ``` ./configure --static-mp4box --prefix=`realpath ./install` --enable-sanitizer --cc=clang --cxx=clang++ ``` ## crash command ``` MP4Box -flat -out /dev/null poc_file ``` ## Pocs [poc.zip](https://github.com/gpac/gpac/files/8032708/poc.zip) ## Crash output ``` ==30676==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000040 (pc 0x0000007f9b6b bp 0x7ffd0126e890 sp 0x7ffd0126c600 T0) ==30676==The signal is caused by a READ memory access. ==30676==Hint: address points to the zero page. #0 0x7f9b6b in DoWriteMeta /programs/mp4box/builds/build12/src/isomedia/isom_store.c:763:13 #1 0x7faf65 in DoWrite /programs/mp4box/builds/build12/src/isomedia/isom_store.c:928:8 #2 0x802ec2 in WriteFlat /programs/mp4box/builds/build12/src/isomedia/isom_store.c:1170:10 #3 0x800b2a in WriteToFile /programs/mp4box/builds/build12/src/isomedia/isom_store.c #4 0x7d3607 in gf_isom_write /programs/mp4box/builds/build12/src/isomedia/isom_read.c:600:8 #5 0x7d3bd3 in gf_isom_close /programs/mp4box/builds/build12/src/isomedia/isom_read.c:624:6 #6 0x4e765f in mp4boxMain /programs/mp4box/builds/build12/applications/mp4box/main.c:6547:7 #7 0x7fd809faf0b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x270b2) #8 0x41ea6d in _start (/programs/mp4box/builds/build12/bin/gcc/MP4Box+0x41ea6d) AddressSanitizer can not provide additional info. SUMMARY: AddressSanitizer: SEGV /programs/mp4box/builds/build12/src/isomedia/isom_store.c:763:13 in DoWriteMeta ==30676==ABORTING ```
There is a segment fault detected by AddressSanitizer
https://api.github.com/repos/gpac/gpac/issues/2113/comments
0
2022-02-09T12:36:07
2022-02-09T14:54:34Z
https://github.com/gpac/gpac/issues/2113
1,128,487,015
2,113
false
This is a GitHub Issue repo:gpac owner:gpac Title : There is a segment fault detected by AddressSanitizer Issue date: --- start body --- ## Description There is a segment fault detected by AddressSanitizer ## System info ``` Ubuntu 20.04.2 LTS clang version 12.0.0-++20210402082642+04ba60cfe598-1~exp1~20210402063359.71 MP4Box - GPAC version 1.1.0-DEV-rev1740-gbd98129fd-master ``` ## Build command ``` ./configure --static-mp4box --prefix=`realpath ./install` --enable-sanitizer --cc=clang --cxx=clang++ ``` ## crash command ``` MP4Box -flat -out /dev/null poc_file ``` ## Pocs [poc.zip](https://github.com/gpac/gpac/files/8032708/poc.zip) ## Crash output ``` ==30676==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000040 (pc 0x0000007f9b6b bp 0x7ffd0126e890 sp 0x7ffd0126c600 T0) ==30676==The signal is caused by a READ memory access. ==30676==Hint: address points to the zero page. #0 0x7f9b6b in DoWriteMeta /programs/mp4box/builds/build12/src/isomedia/isom_store.c:763:13 #1 0x7faf65 in DoWrite /programs/mp4box/builds/build12/src/isomedia/isom_store.c:928:8 #2 0x802ec2 in WriteFlat /programs/mp4box/builds/build12/src/isomedia/isom_store.c:1170:10 #3 0x800b2a in WriteToFile /programs/mp4box/builds/build12/src/isomedia/isom_store.c #4 0x7d3607 in gf_isom_write /programs/mp4box/builds/build12/src/isomedia/isom_read.c:600:8 #5 0x7d3bd3 in gf_isom_close /programs/mp4box/builds/build12/src/isomedia/isom_read.c:624:6 #6 0x4e765f in mp4boxMain /programs/mp4box/builds/build12/applications/mp4box/main.c:6547:7 #7 0x7fd809faf0b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x270b2) #8 0x41ea6d in _start (/programs/mp4box/builds/build12/bin/gcc/MP4Box+0x41ea6d) AddressSanitizer can not provide additional info. SUMMARY: AddressSanitizer: SEGV /programs/mp4box/builds/build12/src/isomedia/isom_store.c:763:13 in DoWriteMeta ==30676==ABORTING ``` --- end body ---
1,977
[ -0.020516473799943924, 0.006045162212103605, -0.011486899107694626, 0.024413295090198517, 0.01539826113730669, 0.004202169366180897, -0.03527496010065079, 0.04347573220729828, -0.008338906802237034, 0.013260825537145138, -0.02242126315832138, 0.016052577644586563, 0.026870619505643845, -0....
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
i am having trouble to push a dash session into an http server using dashin forward file mode. The session can be played successfully by the player using the command : ```bash gpac -i route://239.3.6.1:8092/media.mpd:ifce=docker0 aout vout ``` but when using the command: ```bash gpac -i route://239.3.6.1:8092/media.mpd:ifce=docker0 dashin:forward=file -o http://localhost:8080/media.mpd:rdirs=gmem:reqlog='*':cors=auto -logs=all@warning --max_cache_segs=40 ``` the client receives the manifest and the init segments but fails to get other segments, the server response is as follows ```bash [Socket] Asked to set socket receive buffer size to 524288 but system used 425984 [HTTPOut] Server running on port 8080 [HTTPOut] REQ#1 127.0.0.1 GET media.mpd start [HTTPOut] H2 REQ#1 GET done: reply 200 - 1859 bytes in 0 ms at 104.732 mbps [HTTPOut] REQ#2 127.0.0.1 GET v11-init.mp4 start [HTTPOut] H2 REQ#2 GET done: reply 200 - 738 bytes in 0 ms at 54.1651 mbps [HTTPOut] REQ#3 127.0.0.1 GET a22-init.mp4 start [HTTPOut] H2 REQ#3 GET done: reply 200 - 604 bytes in 0 ms at 48.8081 mbps [HTTPOut] REQ#4 127.0.0.1 GET media.mpd start [HTTPOut] H2 REQ#4 GET done: reply 200 - 1859 bytes in 0 ms at 154.917 mbps [HTTPOut] REQ#5 127.0.0.1 GET v11-889214481.m4s error 404 [HTTPOut] H2 REQ#5 GET done: reply 404 - 1859 bytes in 904 ms at 16.4487 kbps [HTTPOut] REQ#6 127.0.0.1 UNKNOWN error 501 [HTTPOut] H2 REQ#6 GET done: reply 501 - 1859 bytes in 904 ms at 16.4429 kbps [HTTPOut] REQ#7 127.0.0.1 GET media.mpd start [HTTPOut] H2 REQ#7 GET done: reply 200 - 1859 bytes in 0 ms at 109.353 mbps [HTTPOut] REQ#8 127.0.0.1 GET v11-889214481.m4s error 404 [HTTPOut] H2 REQ#8 GET done: reply 404 - 1859 bytes in 409 ms at 36.3328 kbps [HTTPOut] REQ#9 127.0.0.1 GET v11-889214481.m4s error 404 [HTTPOut] H2 REQ#9 GET done: reply 404 - 1859 bytes in 917 ms at 16.206 kbps ``` content of the manifest: ```bash cat media.mpd.3 <?xml version="1.0" encoding="UTF-8"?> <MPD availabilityStartTime="2023-11-30T08:02:24Z" maxSegmentDuration="PT1.92S" minBufferTime="PT1S" minimumUpdatePeriod="PT1S" profiles="urn:mpeg:dash:profile:isoff-live:2011" publishTime="2023-11-30T08:03:33Z" timeShiftBufferDepth="PT11S" type="dynamic" xmlns="urn:mpeg:dash:schema:mpd:2011" xmlns:cenc="urn:mpeg:cenc:2013" xmlns:dashif="https://dashif.org/" xmlns:scte35="http://www.scte.org/schemas/35/2016" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemalocation="urn:mpeg:dash:schema:mpd:2011 DASH-MPD.xsd http://dashif.org/guidelines/ContentProtection laurl.xsd"> <Period id="P0" start="PT0S"> <AdaptationSet contentType="video" id="0" maxFrameRate="50" maxHeight="576" maxWidth="720" mimeType="video/mp4" minFrameRate="50" minHeight="576" minWidth="720" par="5:4" segmentAlignment="true" startWithSAP="1"> <Role schemeIdUri="urn:mpeg:dash:role:2011" value="main" /> <SegmentTemplate duration="96" initialization="$RepresentationID$-init.mp4" media="$RepresentationID$-$Number$.m4s" startNumber="886110074" timescale="50" /> <Representation bandwidth="2000000" codecs="vvc.0.31.10" frameRate="50" height="576" id="v11" sar="64:45" scanType="progressive" width="720" /> </AdaptationSet> <AdaptationSet contentType="audio" id="1" mimeType="audio/mp4" segmentAlignment="true" startWithSAP="1"> <Role schemeIdUri="urn:mpeg:dash:role:2011" value="main" /> <SegmentTemplate duration="92160" initialization="$RepresentationID$-init.mp4" media="$RepresentationID$-$Number$.m4s" startNumber="886110074" timescale="48000" /> <AudioChannelConfiguration schemeIdUri="urn:mpeg:dash:23003:3:audio_channel_configuration:2011" value="2" /> <Representation audioSamplingRate="48000" bandwidth="128000" codecs="mp4a.40.2" id="a22" /> </AdaptationSet> </Period> </MPD> ```
Problem using dashin forward file mode
https://api.github.com/repos/gpac/gpac/issues/2751/comments
5
2024-02-07T08:08:34
2024-02-20T10:13:03Z
https://github.com/gpac/gpac/issues/2751
2,122,428,094
2,751
false
This is a GitHub Issue repo:gpac owner:gpac Title : Problem using dashin forward file mode Issue date: --- start body --- i am having trouble to push a dash session into an http server using dashin forward file mode. The session can be played successfully by the player using the command : ```bash gpac -i route://239.3.6.1:8092/media.mpd:ifce=docker0 aout vout ``` but when using the command: ```bash gpac -i route://239.3.6.1:8092/media.mpd:ifce=docker0 dashin:forward=file -o http://localhost:8080/media.mpd:rdirs=gmem:reqlog='*':cors=auto -logs=all@warning --max_cache_segs=40 ``` the client receives the manifest and the init segments but fails to get other segments, the server response is as follows ```bash [Socket] Asked to set socket receive buffer size to 524288 but system used 425984 [HTTPOut] Server running on port 8080 [HTTPOut] REQ#1 127.0.0.1 GET media.mpd start [HTTPOut] H2 REQ#1 GET done: reply 200 - 1859 bytes in 0 ms at 104.732 mbps [HTTPOut] REQ#2 127.0.0.1 GET v11-init.mp4 start [HTTPOut] H2 REQ#2 GET done: reply 200 - 738 bytes in 0 ms at 54.1651 mbps [HTTPOut] REQ#3 127.0.0.1 GET a22-init.mp4 start [HTTPOut] H2 REQ#3 GET done: reply 200 - 604 bytes in 0 ms at 48.8081 mbps [HTTPOut] REQ#4 127.0.0.1 GET media.mpd start [HTTPOut] H2 REQ#4 GET done: reply 200 - 1859 bytes in 0 ms at 154.917 mbps [HTTPOut] REQ#5 127.0.0.1 GET v11-889214481.m4s error 404 [HTTPOut] H2 REQ#5 GET done: reply 404 - 1859 bytes in 904 ms at 16.4487 kbps [HTTPOut] REQ#6 127.0.0.1 UNKNOWN error 501 [HTTPOut] H2 REQ#6 GET done: reply 501 - 1859 bytes in 904 ms at 16.4429 kbps [HTTPOut] REQ#7 127.0.0.1 GET media.mpd start [HTTPOut] H2 REQ#7 GET done: reply 200 - 1859 bytes in 0 ms at 109.353 mbps [HTTPOut] REQ#8 127.0.0.1 GET v11-889214481.m4s error 404 [HTTPOut] H2 REQ#8 GET done: reply 404 - 1859 bytes in 409 ms at 36.3328 kbps [HTTPOut] REQ#9 127.0.0.1 GET v11-889214481.m4s error 404 [HTTPOut] H2 REQ#9 GET done: reply 404 - 1859 bytes in 917 ms at 16.206 kbps ``` content of the manifest: ```bash cat media.mpd.3 <?xml version="1.0" encoding="UTF-8"?> <MPD availabilityStartTime="2023-11-30T08:02:24Z" maxSegmentDuration="PT1.92S" minBufferTime="PT1S" minimumUpdatePeriod="PT1S" profiles="urn:mpeg:dash:profile:isoff-live:2011" publishTime="2023-11-30T08:03:33Z" timeShiftBufferDepth="PT11S" type="dynamic" xmlns="urn:mpeg:dash:schema:mpd:2011" xmlns:cenc="urn:mpeg:cenc:2013" xmlns:dashif="https://dashif.org/" xmlns:scte35="http://www.scte.org/schemas/35/2016" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemalocation="urn:mpeg:dash:schema:mpd:2011 DASH-MPD.xsd http://dashif.org/guidelines/ContentProtection laurl.xsd"> <Period id="P0" start="PT0S"> <AdaptationSet contentType="video" id="0" maxFrameRate="50" maxHeight="576" maxWidth="720" mimeType="video/mp4" minFrameRate="50" minHeight="576" minWidth="720" par="5:4" segmentAlignment="true" startWithSAP="1"> <Role schemeIdUri="urn:mpeg:dash:role:2011" value="main" /> <SegmentTemplate duration="96" initialization="$RepresentationID$-init.mp4" media="$RepresentationID$-$Number$.m4s" startNumber="886110074" timescale="50" /> <Representation bandwidth="2000000" codecs="vvc.0.31.10" frameRate="50" height="576" id="v11" sar="64:45" scanType="progressive" width="720" /> </AdaptationSet> <AdaptationSet contentType="audio" id="1" mimeType="audio/mp4" segmentAlignment="true" startWithSAP="1"> <Role schemeIdUri="urn:mpeg:dash:role:2011" value="main" /> <SegmentTemplate duration="92160" initialization="$RepresentationID$-init.mp4" media="$RepresentationID$-$Number$.m4s" startNumber="886110074" timescale="48000" /> <AudioChannelConfiguration schemeIdUri="urn:mpeg:dash:23003:3:audio_channel_configuration:2011" value="2" /> <Representation audioSamplingRate="48000" bandwidth="128000" codecs="mp4a.40.2" id="a22" /> </AdaptationSet> </Period> </MPD> ``` --- end body ---
3,964
[ -0.04252973571419716, 0.019692890346050262, -0.014761030673980713, 0.009906905703246593, -0.0046986546367406845, -0.013180417008697987, -0.01702398620545864, 0.04905948415398598, -0.0389711931347847, 0.025980796664953232, 0.013543181121349335, -0.0019865636713802814, 0.011712088249623775, ...
CVE-2023-24789
2023-03-06T16:15:10.187000
jeecg-boot v3.4.4 was discovered to contain an authenticated SQL injection vulnerability via the building block report component.
{ "cvssMetricV2": null, "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "LOW", "attackVector": "NETWORK", "availabilityImpact": "HIGH", "baseScore": 8.8, "baseSeverity": "HIGH", "confidentialityImpact": "HIGH", "integrityImpact": "HIGH", "privilegesRequired": "LOW", "scope": "UNCHANGED", "userInteraction": "NONE", "vectorString": "CVSS:3.1/AV:N/AC:L/PR:L/UI:N/S:U/C:H/I:H/A:H", "version": "3.1" }, "exploitabilityScore": 2.8, "impactScore": 5.9, "source": "nvd@nist.gov", "type": "Primary" } ] }
[ { "source": "cve@mitre.org", "tags": [ "Exploit", "Third Party Advisory" ], "url": "https://github.com/jeecgboot/jeecg-boot/issues/4511" } ]
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:jeecg:jeecg:3.4.4:*:*:*:*:*:*:*", "matchCriteriaId": "83D5D7DD-38EF-48F6-8C18-F75891A91B66", "versionEndExcluding": null, "versionEndIncluding": null, "versionStartExcluding"...
https://github.com/jeecgboot/jeecg-boot/issues/4511
[ "Exploit", "Third Party Advisory" ]
github.com
[ "jeecgboot", "jeecg-boot" ]
Use the system default account password to log in to the system. In the visual design menu - report design, see Figure 1 for details. <img width="1512" alt="截屏2023-01-29 22 01 08" src="https://user-images.githubusercontent.com/32299945/215331278-68f86a40-1199-463a-b0f0-5441746c7eb1.png"> Then click New Report, see Figure 2 for details. <img width="1512" alt="截屏2023-01-29 22 03 48" src="https://user-images.githubusercontent.com/32299945/215331465-bb6f1bdf-d5c1-4c6a-a952-e34b2fd540ac.png"> Click the + sign to select a new SQL data set, see Figure 3 for details. <img width="1512" alt="截屏2023-01-29 22 07 51" src="https://user-images.githubusercontent.com/32299945/215331861-a06d00f7-f5d4-4bc8-9335-e4731f1b02a5.png"> Select and enter the corresponding data according to Figure 4, the payload is as follows: <img width="1512" alt="截屏2023-01-29 22 11 59" src="https://user-images.githubusercontent.com/32299945/215332242-94ca42ab-c4a7-4bd1-994f-1e98a42d0ec8.png"> payload:select * from sys_user WHERE id='' union SELECT 1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,(select group_concat(SCHEMA_NAME) from information_schema.SCHEMATA) According to the execution result, all the database information of the mysql server of the system can be obtained, see Figure 5 for details. <img width="1512" alt="截屏2023-01-29 22 14 53" src="https://user-images.githubusercontent.com/32299945/215332357-86bd69da-69d3-4e89-9d17-800c30a4ca53.png"> So far the vulnerability has surfaced successfully.
There is a sql injection vulnerability in the jeecg 3.4.4 building block report
https://api.github.com/repos/jeecgboot/JeecgBoot/issues/4511/comments
3
2023-01-29T14:16:41
2023-11-01T01:16:34Z
https://github.com/jeecgboot/JeecgBoot/issues/4511
1,561,308,445
4,511
true
This is a GitHub Issue repo:jeecg-boot owner:jeecgboot Title : There is a sql injection vulnerability in the jeecg 3.4.4 building block report Issue date: --- start body --- Use the system default account password to log in to the system. In the visual design menu - report design, see Figure 1 for details. <img width="1512" alt="截屏2023-01-29 22 01 08" src="https://user-images.githubusercontent.com/32299945/215331278-68f86a40-1199-463a-b0f0-5441746c7eb1.png"> Then click New Report, see Figure 2 for details. <img width="1512" alt="截屏2023-01-29 22 03 48" src="https://user-images.githubusercontent.com/32299945/215331465-bb6f1bdf-d5c1-4c6a-a952-e34b2fd540ac.png"> Click the + sign to select a new SQL data set, see Figure 3 for details. <img width="1512" alt="截屏2023-01-29 22 07 51" src="https://user-images.githubusercontent.com/32299945/215331861-a06d00f7-f5d4-4bc8-9335-e4731f1b02a5.png"> Select and enter the corresponding data according to Figure 4, the payload is as follows: <img width="1512" alt="截屏2023-01-29 22 11 59" src="https://user-images.githubusercontent.com/32299945/215332242-94ca42ab-c4a7-4bd1-994f-1e98a42d0ec8.png"> payload:select * from sys_user WHERE id='' union SELECT 1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,(select group_concat(SCHEMA_NAME) from information_schema.SCHEMATA) According to the execution result, all the database information of the mysql server of the system can be obtained, see Figure 5 for details. <img width="1512" alt="截屏2023-01-29 22 14 53" src="https://user-images.githubusercontent.com/32299945/215332357-86bd69da-69d3-4e89-9d17-800c30a4ca53.png"> So far the vulnerability has surfaced successfully. --- end body ---
1,706
[ 0.013049882836639881, -0.0034334873780608177, -0.006502389907836914, 0.011117958463728428, 0.042216669768095016, 0.026565833017230034, 0.022641846910119057, 0.035812005400657654, 0.008772588334977627, 0.04982409253716469, 0.03686441481113434, -0.01332801952958107, 0.03340649604797363, 0.00...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [X] I looked for a similar issue and couldn't find any. - [X] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [ ] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/2013/07/16/how-to-file-a-bug-properly/ I have a need to download one MP4 file and generate fragment MP4 file on the fly. I followed the example https://github.com/gpac/gpac/wiki/HAS-advanced#quick-record-examples, but got this failure. Can someone please help? ``` % gpac -i http://localhost:8000/gear4.mp4 -o grab/record.mp4:frag Unknown argument ":frag" set but not used by any filter - possible matches - range in filters httpin fin - store (inter|flat|fstart|tight|frag|sfrag) in filters mp4mx ```
Quick record example fail: Unknown argument ":frag" set but not used by any filter
https://api.github.com/repos/gpac/gpac/issues/2566/comments
5
2023-08-28T17:41:31
2023-08-31T22:33:44Z
https://github.com/gpac/gpac/issues/2566
1,870,178,965
2,566
false
This is a GitHub Issue repo:gpac owner:gpac Title : Quick record example fail: Unknown argument ":frag" set but not used by any filter Issue date: --- start body --- Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [X] I looked for a similar issue and couldn't find any. - [X] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [ ] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/2013/07/16/how-to-file-a-bug-properly/ I have a need to download one MP4 file and generate fragment MP4 file on the fly. I followed the example https://github.com/gpac/gpac/wiki/HAS-advanced#quick-record-examples, but got this failure. Can someone please help? ``` % gpac -i http://localhost:8000/gear4.mp4 -o grab/record.mp4:frag Unknown argument ":frag" set but not used by any filter - possible matches - range in filters httpin fin - store (inter|flat|fstart|tight|frag|sfrag) in filters mp4mx ``` --- end body ---
1,356
[ -0.024335786700248718, 0.03477872908115387, -0.015742115676403046, -0.012579705566167831, -0.02124330773949623, 0.0011917309602722526, -0.021771671250462532, 0.04780132696032524, -0.0636211410164833, 0.030924785882234573, -0.029432937502861023, 0.05529165267944336, 0.01902107521891594, -0....
null
null
null
null
null
null
null
null
null
[ "python", "cpython" ]
Something I discovered when working on GH-65056 and GH-113171. `IPv4Address`' and `IPv4Network`'s `is_global` and `is_private` are false at the same time for addresses in this range: ``` >>> ipaddress.IPv4Address('100.64.0.0').is_global False >>> ipaddress.IPv4Address('100.64.0.0').is_private False >>> ipaddress.IPv4Network('100.64.0.0/10').is_global False >>> ipaddress.IPv4Network('100.64.0.0/10').is_private False ``` I don't believe this is right and I'll explain why. # Historical context ## Initial additions `is_private` introduced in dc9b2555a875 ("Issue #14814: addition of the ipaddress module (stage 1 - code and tests)") on 2012-05-20. It only handled the actual private-use ranges for IPv4 (`10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`). Documented as such. ## `100.64.0.0/10` handling and `is_private` semantics changes There have been four `ipaddress` patches involving this range: 1. 22c31764262b ("#17400; ipaddress should make it easy to identify rfc6598 addresses") 3. be9c1b133b78 ("#17400: fix documentation, add cache to is_global and correctly handle 100.64.0.0/10") 4. e5019d518304 ("#17400: correct handling of 100.64.0.0/10, fixing the docs and updating NEWS") 5. 742192a4fea8 ("Issue #21386: Implement missing IPv4Address.is_global property") The first three were part of GH-61602 (https://bugs.python.org/issue17400), the fourth one just adds a missing `IPv4Address.is_global` property to match the `IPv4Network.is_global` semantics and follows the semantics established in commits 1-3. Commit 1 changed the semantics of `is_private` from "Is this a Private-use IPv4 address or a Unique-Local IPv6 address?" to roughly "Do [1] and [2] say globally reachable = false for this address". The documentation change was not quite precise A boolean, True if the address is reserved per iana-ipv4-special-registry or iana-ipv6-special-registry. but the intent of the implementation is quite clear. Commit 1 also added `is_global` that was effectively `not is_private`. In commit 3 an exception for `100.64.0.0/10` is made and both `is_global` and `is_private` are false for that range. # `is_global`/`is_private` semantics clarification As part of GH-65056 we changed the documentation to make it clear that we follow the "globally reachable" information from IANA (with some caveats, like special handling of IPv4-mapped IPv6 addresses, see 83f0f8d62f27 ("bpo-33433 Fix private address checking for IPv4 mapped IPv6. (GH-26172)")). # The problem The motivation for handling `100.64.0.0/10` like this can be found here: https://github.com/python/cpython/issues/61602#issuecomment-1093610325 > The rationale for RFC 6598 is precisely that 100.64.0.0/10 is *not* private in the common sense, so it would deserve a different treatment in the ipaddress module as well. I have to admit I don't find it convincing enough to make an exception for it. [1] says the range is not globally reachable so in my opinion `is_private` should return true for it as it does for the rest of the not globally reachable address blocks (again, with the exception of IPv4-mapped IPv6 address handling). I'd find `is_private` being false for the range surprising if I wasn't clearly aware of the semantics after reading this code multiple times. I believe it may lead to real-world issues. Additionally the behavior where `is_private` is *not* the opposite of `is_global` is likely to be surprising too. In short: IMO there should be no exception. The downside of the proposed solution: this is technically breaking backwards compatibility if some code depends on the current semantics. I'm not sure I'd classify the proposed change strictly as a bug fix. [1] https://www.iana.org/assignments/iana-ipv4-special-registry/iana-ipv4-special-registry.xhtml [2] https://www.iana.org/assignments/iana-ipv6-special-registry/iana-ipv6-special-registry.xhtml <!-- gh-linked-prs --> ### Linked PRs * gh-119813 <!-- /gh-linked-prs -->
ipaddress: Shared Address Space (100.64.0.0/10) is neither private nor global
https://api.github.com/repos/python/cpython/issues/119812/comments
6
2024-05-30T22:19:04
2024-06-04T15:13:02Z
https://github.com/python/cpython/issues/119812
2,326,617,768
119,812
false
This is a GitHub Issue repo:cpython owner:python Title : ipaddress: Shared Address Space (100.64.0.0/10) is neither private nor global Issue date: --- start body --- Something I discovered when working on GH-65056 and GH-113171. `IPv4Address`' and `IPv4Network`'s `is_global` and `is_private` are false at the same time for addresses in this range: ``` >>> ipaddress.IPv4Address('100.64.0.0').is_global False >>> ipaddress.IPv4Address('100.64.0.0').is_private False >>> ipaddress.IPv4Network('100.64.0.0/10').is_global False >>> ipaddress.IPv4Network('100.64.0.0/10').is_private False ``` I don't believe this is right and I'll explain why. # Historical context ## Initial additions `is_private` introduced in dc9b2555a875 ("Issue #14814: addition of the ipaddress module (stage 1 - code and tests)") on 2012-05-20. It only handled the actual private-use ranges for IPv4 (`10.0.0.0/8`, `172.16.0.0/12`, `192.168.0.0/16`). Documented as such. ## `100.64.0.0/10` handling and `is_private` semantics changes There have been four `ipaddress` patches involving this range: 1. 22c31764262b ("#17400; ipaddress should make it easy to identify rfc6598 addresses") 3. be9c1b133b78 ("#17400: fix documentation, add cache to is_global and correctly handle 100.64.0.0/10") 4. e5019d518304 ("#17400: correct handling of 100.64.0.0/10, fixing the docs and updating NEWS") 5. 742192a4fea8 ("Issue #21386: Implement missing IPv4Address.is_global property") The first three were part of GH-61602 (https://bugs.python.org/issue17400), the fourth one just adds a missing `IPv4Address.is_global` property to match the `IPv4Network.is_global` semantics and follows the semantics established in commits 1-3. Commit 1 changed the semantics of `is_private` from "Is this a Private-use IPv4 address or a Unique-Local IPv6 address?" to roughly "Do [1] and [2] say globally reachable = false for this address". The documentation change was not quite precise A boolean, True if the address is reserved per iana-ipv4-special-registry or iana-ipv6-special-registry. but the intent of the implementation is quite clear. Commit 1 also added `is_global` that was effectively `not is_private`. In commit 3 an exception for `100.64.0.0/10` is made and both `is_global` and `is_private` are false for that range. # `is_global`/`is_private` semantics clarification As part of GH-65056 we changed the documentation to make it clear that we follow the "globally reachable" information from IANA (with some caveats, like special handling of IPv4-mapped IPv6 addresses, see 83f0f8d62f27 ("bpo-33433 Fix private address checking for IPv4 mapped IPv6. (GH-26172)")). # The problem The motivation for handling `100.64.0.0/10` like this can be found here: https://github.com/python/cpython/issues/61602#issuecomment-1093610325 > The rationale for RFC 6598 is precisely that 100.64.0.0/10 is *not* private in the common sense, so it would deserve a different treatment in the ipaddress module as well. I have to admit I don't find it convincing enough to make an exception for it. [1] says the range is not globally reachable so in my opinion `is_private` should return true for it as it does for the rest of the not globally reachable address blocks (again, with the exception of IPv4-mapped IPv6 address handling). I'd find `is_private` being false for the range surprising if I wasn't clearly aware of the semantics after reading this code multiple times. I believe it may lead to real-world issues. Additionally the behavior where `is_private` is *not* the opposite of `is_global` is likely to be surprising too. In short: IMO there should be no exception. The downside of the proposed solution: this is technically breaking backwards compatibility if some code depends on the current semantics. I'm not sure I'd classify the proposed change strictly as a bug fix. [1] https://www.iana.org/assignments/iana-ipv4-special-registry/iana-ipv4-special-registry.xhtml [2] https://www.iana.org/assignments/iana-ipv6-special-registry/iana-ipv6-special-registry.xhtml <!-- gh-linked-prs --> ### Linked PRs * gh-119813 <!-- /gh-linked-prs --> --- end body ---
4,240
[ -0.03762134537100792, -0.018674364313483238, -0.013115976937115192, 0.02508089877665043, 0.012919086031615734, 0.012358703650534153, -0.01687205396592617, 0.012820640578866005, 0.009079709649085999, 0.03525865450501442, -0.0075727347284555435, 0.018326018005609512, 0.05079790577292442, -0....
null
null
null
null
null
null
null
null
null
[ "kubernetes", "kubernetes" ]
### What happened? Documentation states that enabling `publishNotReadyAddresses` will cause DNS records to appear for pods, even if `NotReady`. Documentation does not state that enabling this flag will cause incoming traffic to be directed to pods that are `NotReady`. ### What did you expect to happen? I expected DNS records to be created, nothing beyond that. ### How can we reproduce it (as minimally and precisely as possible)? 1. Create HTTP Server deployment with initContainer -> `sleep 30` 2. Create service for that deployment with `publishNotReadyAddresses` set to `true` 3. Continiously call HTTP Server 4. Scale HTTP Server to 10 Replicas 5. Observe incoming errors ### Anything else we need to know? It can be easily fixed by changing condition for building topology from `isReady` to `isServing && !isTerminating`. ### Kubernetes version <details> ```console $ kubectl version v1.29.1 ``` </details> ### Cloud provider <details> </details> ### OS version <details> ```console # On Linux: $ cat /etc/os-release # paste output here $ uname -a # paste output here # On Windows: C:\> wmic os get Caption, Version, BuildNumber, OSArchitecture # paste output here ``` </details> ### Install tools <details> </details> ### Container runtime (CRI) and version (if applicable) <details> </details> ### Related plugins (CNI, CSI, ...) and versions (if applicable) <details> </details>
Enabling `publishNotReadyAddresses` causes proxy to direct traffic to NotReady pods.
https://api.github.com/repos/kubernetes/kubernetes/issues/124914/comments
6
2024-05-16T18:47:44
2024-05-17T12:48:25Z
https://github.com/kubernetes/kubernetes/issues/124914
2,301,106,189
124,914
false
This is a GitHub Issue repo:kubernetes owner:kubernetes Title : Enabling `publishNotReadyAddresses` causes proxy to direct traffic to NotReady pods. Issue date: --- start body --- ### What happened? Documentation states that enabling `publishNotReadyAddresses` will cause DNS records to appear for pods, even if `NotReady`. Documentation does not state that enabling this flag will cause incoming traffic to be directed to pods that are `NotReady`. ### What did you expect to happen? I expected DNS records to be created, nothing beyond that. ### How can we reproduce it (as minimally and precisely as possible)? 1. Create HTTP Server deployment with initContainer -> `sleep 30` 2. Create service for that deployment with `publishNotReadyAddresses` set to `true` 3. Continiously call HTTP Server 4. Scale HTTP Server to 10 Replicas 5. Observe incoming errors ### Anything else we need to know? It can be easily fixed by changing condition for building topology from `isReady` to `isServing && !isTerminating`. ### Kubernetes version <details> ```console $ kubectl version v1.29.1 ``` </details> ### Cloud provider <details> </details> ### OS version <details> ```console # On Linux: $ cat /etc/os-release # paste output here $ uname -a # paste output here # On Windows: C:\> wmic os get Caption, Version, BuildNumber, OSArchitecture # paste output here ``` </details> ### Install tools <details> </details> ### Container runtime (CRI) and version (if applicable) <details> </details> ### Related plugins (CNI, CSI, ...) and versions (if applicable) <details> </details> --- end body ---
1,701
[ -0.010277806781232357, -0.029663734138011932, -0.0048835910856723785, 0.025744985789060593, -0.018703116104006767, -0.030067484825849533, -0.0004831638070754707, 0.037976231426000595, -0.030376235023140907, -0.01788374036550522, 0.00420374795794487, 0.012421243824064732, -0.02790623530745506...
CVE-2022-46489
2023-01-05T15:15:10.483000
GPAC version 2.1-DEV-rev505-gb9577e6ad-master was discovered to contain a memory leak via the gf_isom_box_parse_ex function at box_funcs.c.
{ "cvssMetricV2": null, "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "LOW", "attackVector": "LOCAL", "availabilityImpact": "HIGH", "baseScore": 5.5, "baseSeverity": "MEDIUM", "confidentialityImpact": "NONE", "integrityImpact": "NONE", "privilegesRequired": "NONE", "scope": "UNCHANGED", "userInteraction": "REQUIRED", "vectorString": "CVSS:3.1/AV:L/AC:L/PR:N/UI:R/S:U/C:N/I:N/A:H", "version": "3.1" }, "exploitabilityScore": 1.8, "impactScore": 3.6, "source": "nvd@nist.gov", "type": "Primary" } ] }
[ { "source": "cve@mitre.org", "tags": [ "Exploit", "Issue Tracking", "Third Party Advisory" ], "url": "https://github.com/gpac/gpac/issues/2328" } ]
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:gpac:gpac:*:*:*:*:*:*:*:*", "matchCriteriaId": "B4D3D58A-C3C9-4441-A84A-FB91FD19985C", "versionEndExcluding": "2.2.0", "versionEndIncluding": null, "versionStartExcluding": n...
https://github.com/gpac/gpac/issues/2328
[ "Exploit", "Issue Tracking", "Third Party Advisory" ]
github.com
[ "gpac", "gpac" ]
A memory leak has occurred when running program MP4Box, this can reproduce on the lattest commit. ### Version ``` $ ./MP4Box -version MP4Box - GPAC version 2.1-DEV-rev505-gb9577e6ad-master (c) 2000-2022 Telecom Paris distributed under LGPL v2.1+ - http://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: --static-build --extra-cflags=-fsanitize=address -g --extra-ldflags=-fsanitize=address -g Features: GPAC_CONFIG_LINUX GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SSL GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_FREETYPE GPAC_HAS_JPEG GPAC_HAS_PNG GPAC_HAS_LINUX_DVB ``` git log ``` commit b9577e6ad91ef96decbcd369227ab02b2842c77f (HEAD -> master, origin/master, origin/HEAD) Author: jeanlf <jeanlf@gpac.io> Date: Fri Nov 25 16:53:55 2022 +0100 ``` ### Verification steps ``` export CFLAGS='-fsanitize=address -g' export CC=/usr/bin/clang export CXX=/usr/bin/clang++ git clone https://github.com/gpac/gpac.git cd gpac ./configure --static-build --extra-cflags="${CFLAGS}" --extra-ldflags="${CFLAGS}" make cd bin/gcc ./MP4Box -info $poc ``` ### POC file https://github.com/HotSpurzzZ/testcases/blob/main/gpac/gpac_Direct_leak_gf_isom_box_parse_ex.mp4 ### AddressSanitizer output ``` $ ./MP4Box -info gpac_Direct_leak_gf_isom_box_parse_ex.mp4 [iso file] Failed to uncompress payload for box type !ssx (0x21737378) Error opening file gpac_Direct_leak_gf_isom_box_parse_ex.mp4: BitStream Not Compliant ================================================================= ==10575==ERROR: LeakSanitizer: detected memory leaks Direct leak of 1718840668 byte(s) in 1 object(s) allocated from: #0 0x4a186d in malloc (/root/Desktop/gpac/bin/gcc/MP4Box+0x4a186d) #1 0x7dfc41 in gf_isom_box_parse_ex /root/Desktop/gpac/src/isomedia/box_funcs.c:166:13 #2 0x7df29c in gf_isom_parse_root_box /root/Desktop/gpac/src/isomedia/box_funcs.c:38:8 Direct leak of 4096 byte(s) in 1 object(s) allocated from: #0 0x4a186d in malloc (/root/Desktop/gpac/bin/gcc/MP4Box+0x4a186d) #1 0x599d69 in gf_gz_decompress_payload /root/Desktop/gpac/src/utils/base_encoding.c:257:31 #2 0x7dfc66 in gf_isom_box_parse_ex /root/Desktop/gpac/src/isomedia/box_funcs.c:170:9 #3 0x7df29c in gf_isom_parse_root_box /root/Desktop/gpac/src/isomedia/box_funcs.c:38:8 SUMMARY: AddressSanitizer: 1718844764 byte(s) leaked in 2 allocation(s). ```
Memory leak in gf_isom_box_parse_ex function of box_funcs.c:166:13
https://api.github.com/repos/gpac/gpac/issues/2328/comments
0
2022-11-28T06:47:39
2022-11-28T10:06:52Z
https://github.com/gpac/gpac/issues/2328
1,465,902,126
2,328
true
This is a GitHub Issue repo:gpac owner:gpac Title : Memory leak in gf_isom_box_parse_ex function of box_funcs.c:166:13 Issue date: --- start body --- A memory leak has occurred when running program MP4Box, this can reproduce on the lattest commit. ### Version ``` $ ./MP4Box -version MP4Box - GPAC version 2.1-DEV-rev505-gb9577e6ad-master (c) 2000-2022 Telecom Paris distributed under LGPL v2.1+ - http://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: --static-build --extra-cflags=-fsanitize=address -g --extra-ldflags=-fsanitize=address -g Features: GPAC_CONFIG_LINUX GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SSL GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_FREETYPE GPAC_HAS_JPEG GPAC_HAS_PNG GPAC_HAS_LINUX_DVB ``` git log ``` commit b9577e6ad91ef96decbcd369227ab02b2842c77f (HEAD -> master, origin/master, origin/HEAD) Author: jeanlf <jeanlf@gpac.io> Date: Fri Nov 25 16:53:55 2022 +0100 ``` ### Verification steps ``` export CFLAGS='-fsanitize=address -g' export CC=/usr/bin/clang export CXX=/usr/bin/clang++ git clone https://github.com/gpac/gpac.git cd gpac ./configure --static-build --extra-cflags="${CFLAGS}" --extra-ldflags="${CFLAGS}" make cd bin/gcc ./MP4Box -info $poc ``` ### POC file https://github.com/HotSpurzzZ/testcases/blob/main/gpac/gpac_Direct_leak_gf_isom_box_parse_ex.mp4 ### AddressSanitizer output ``` $ ./MP4Box -info gpac_Direct_leak_gf_isom_box_parse_ex.mp4 [iso file] Failed to uncompress payload for box type !ssx (0x21737378) Error opening file gpac_Direct_leak_gf_isom_box_parse_ex.mp4: BitStream Not Compliant ================================================================= ==10575==ERROR: LeakSanitizer: detected memory leaks Direct leak of 1718840668 byte(s) in 1 object(s) allocated from: #0 0x4a186d in malloc (/root/Desktop/gpac/bin/gcc/MP4Box+0x4a186d) #1 0x7dfc41 in gf_isom_box_parse_ex /root/Desktop/gpac/src/isomedia/box_funcs.c:166:13 #2 0x7df29c in gf_isom_parse_root_box /root/Desktop/gpac/src/isomedia/box_funcs.c:38:8 Direct leak of 4096 byte(s) in 1 object(s) allocated from: #0 0x4a186d in malloc (/root/Desktop/gpac/bin/gcc/MP4Box+0x4a186d) #1 0x599d69 in gf_gz_decompress_payload /root/Desktop/gpac/src/utils/base_encoding.c:257:31 #2 0x7dfc66 in gf_isom_box_parse_ex /root/Desktop/gpac/src/isomedia/box_funcs.c:170:9 #3 0x7df29c in gf_isom_parse_root_box /root/Desktop/gpac/src/isomedia/box_funcs.c:38:8 SUMMARY: AddressSanitizer: 1718844764 byte(s) leaked in 2 allocation(s). ``` --- end body ---
2,714
[ -0.029760129749774933, 0.02946721389889717, -0.01224383246153593, -0.005338369403034449, 0.01419171504676342, 0.0008878975640982389, -0.008186964318156242, 0.04191608726978302, -0.002325010020285845, 0.006942077539861202, -0.00841397326439619, 0.0010737152770161629, 0.01574416272342205, -0...
null
null
null
null
null
null
null
null
null
[ "ImageMagick", "ImageMagick" ]
### ImageMagick version 6.9.10-23 ### Operating system Linux ### Operating system, version and so on Ubuntu 20.04 ### Description (Also tested with ImageMagick-7.1.1-21) The call `$im->Profile(profile=$blob)` does not work ### Steps to Reproduce ```perl #!/usr/bin/perl use strict; use Image::Magick; my $im=Image::Magick->new; $im->Set(size=>"100x100"); $im->ReadImage("canvas:red"); my $icc= "\0\0\2Dargl\2 \0\0mntrRGB XYZ \7\326\0\7\0\10\0\3\0\34\0/acspMSFT\0" ."\0\0\0none\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\366\326\0\1\0\0\0\0" ."\323-argl\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0" ."\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\12desc\0\0\0\374\0\0\0{cpr" ."t\0\0\1x\0\0\0005wtpt\0\0\1\260\0\0\0\24bkpt\0\0\1\304\0\0\0\24rX" ."YZ\0\0\1\330\0\0\0\24gXYZ\0\0\1\354\0\0\0\24bXYZ\0\0\2\0\0\0\0\24" ."rTRC\0\0\2\24\0\0\0\16gTRC\0\0\2\44\0\0\0\16bTRC\0\0\0024\0\0\0\16" ."desc\0\0\0\0\0\0\0!Compatible with Adobe RGB (1998)\0\0\0\0\0\0\0" ."\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0" ."\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0" ."\0\0\0\0\0\0\0text\0\0\0\0Public Domain. No Warranty, Use at own " ."risk.\0\0\0\0XYZ \0\0\0\0\0\0\363Q\0\1\0\0\0\1\26\314XYZ \0\0\0\0" ."\0\0\0\0\0\0\0\0\0\0\0\0XYZ \0\0\0\0\0\0\234\30\0\0O\245\0\0\4\374" ."XYZ \0\0\0\0\0\0004\215\0\0\240,\0\0\17\225XYZ \0\0\0\0\0\0&1\0\0" ."\20/\0\0\276\234curv\0\0\0\0\0\0\0\1\0023\0\0curv\0\0\0\0\0\0\0\1" ."\0023\0\0curv\0\0\0\0\0\0\0\1\0023\0\0"; $im->Profile(profile=>$icc); $im->Write("asdf.jpg"); ``` ### Images _No response_
PerlMagick: $im->Profile(profile=>$blob) does not work.
https://api.github.com/repos/ImageMagick/ImageMagick/issues/6864/comments
1
2023-11-14T12:24:20
2023-11-14T13:12:03Z
https://github.com/ImageMagick/ImageMagick/issues/6864
1,992,638,925
6,864
false
This is a GitHub Issue repo:ImageMagick owner:ImageMagick Title : PerlMagick: $im->Profile(profile=>$blob) does not work. Issue date: --- start body --- ### ImageMagick version 6.9.10-23 ### Operating system Linux ### Operating system, version and so on Ubuntu 20.04 ### Description (Also tested with ImageMagick-7.1.1-21) The call `$im->Profile(profile=$blob)` does not work ### Steps to Reproduce ```perl #!/usr/bin/perl use strict; use Image::Magick; my $im=Image::Magick->new; $im->Set(size=>"100x100"); $im->ReadImage("canvas:red"); my $icc= "\0\0\2Dargl\2 \0\0mntrRGB XYZ \7\326\0\7\0\10\0\3\0\34\0/acspMSFT\0" ."\0\0\0none\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\366\326\0\1\0\0\0\0" ."\323-argl\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0" ."\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\12desc\0\0\0\374\0\0\0{cpr" ."t\0\0\1x\0\0\0005wtpt\0\0\1\260\0\0\0\24bkpt\0\0\1\304\0\0\0\24rX" ."YZ\0\0\1\330\0\0\0\24gXYZ\0\0\1\354\0\0\0\24bXYZ\0\0\2\0\0\0\0\24" ."rTRC\0\0\2\24\0\0\0\16gTRC\0\0\2\44\0\0\0\16bTRC\0\0\0024\0\0\0\16" ."desc\0\0\0\0\0\0\0!Compatible with Adobe RGB (1998)\0\0\0\0\0\0\0" ."\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0" ."\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0" ."\0\0\0\0\0\0\0text\0\0\0\0Public Domain. No Warranty, Use at own " ."risk.\0\0\0\0XYZ \0\0\0\0\0\0\363Q\0\1\0\0\0\1\26\314XYZ \0\0\0\0" ."\0\0\0\0\0\0\0\0\0\0\0\0XYZ \0\0\0\0\0\0\234\30\0\0O\245\0\0\4\374" ."XYZ \0\0\0\0\0\0004\215\0\0\240,\0\0\17\225XYZ \0\0\0\0\0\0&1\0\0" ."\20/\0\0\276\234curv\0\0\0\0\0\0\0\1\0023\0\0curv\0\0\0\0\0\0\0\1" ."\0023\0\0curv\0\0\0\0\0\0\0\1\0023\0\0"; $im->Profile(profile=>$icc); $im->Write("asdf.jpg"); ``` ### Images _No response_ --- end body ---
1,801
[ -0.01922486536204815, -0.004448888823390007, -0.0020844105165451765, 0.02292046882212162, 0.008524353615939617, 0.0305949617177248, -0.03721357136964798, 0.016907064244151115, -0.028122643008828163, 0.011440918780863285, 0.025032242760062218, 0.00824106764048338, -0.012355162762105465, -0....
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at https://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Trying the following GPAC command to generate CMAF2 HLS with raw subtitles (rawsub) is creating a m3u8 playlist with an init file, which GPAC doesn't create: ``` gpac \ -i output.ts \ -i subtitle.vtt \ -o res/live.m3u8:cmaf=cmf2:rawsub=true:segdur=1:cat=auto ``` We set the `rawsub=true`, which works fine, however, the following is the subtitles playlist generated: ``` #EXTM3U #EXT-X-TARGETDURATION:1 #EXT-X-VERSION:6 #EXT-X-MEDIA-SEQUENCE:1 #EXT-X-INDEPENDENT-SEGMENTS #EXT-X-MAP:URI="subtitle_dashinit.vtt" #EXTINF:1, subtitle_dash1.vtt #EXTINF:1, subtitle_dash2.vtt #EXTINF:1, subtitle_dash3.vtt #EXTINF:1, subtitle_dash4.vtt #EXTINF:1, subtitle_dash5.vtt #EXTINF:1, subtitle_dash6.vtt #EXTINF:1, subtitle_dash7.vtt #EXTINF:1, subtitle_dash8.vtt #EXTINF:1, subtitle_dash9.vtt #EXTINF:1, subtitle_dash10.vtt #EXTINF:1, subtitle_dash11.vtt #EXTINF:1, subtitle_dash12.vtt #EXTINF:1, subtitle_dash13.vtt #EXTINF:1, subtitle_dash14.vtt #EXTINF:1, subtitle_dash15.vtt #EXTINF:1, subtitle_dash16.vtt #EXTINF:1, subtitle_dash17.vtt #EXTINF:1, subtitle_dash18.vtt #EXTINF:1, subtitle_dash19.vtt #EXTINF:1, subtitle_dash20.vtt #EXT-X-ENDLIST ``` Look to the tag `#EXT-X-MAP:URI="subtitle_dashinit.vtt"`, but the `subtitle_dashinit.vtt` file is not created, so, when we play the HLS VOD, subtitles don't show up. The following is the filter graph of that command: ``` Filters connected: fin (src=output.ts) (idx=1) -(PID output.ts) m2tsdmx (dyn_idx=4) --(PID P1V2) rfnalu (dyn_idx=6) ---(PID P1V2) dasher (dyn_idx=8) ----(PID manifest_m3u8) fout (dst=res/live.m3u8:cmaf=cmf2:rawsub=true:segdur=1:cat=auto) (idx=3) ----(PID subtitle.vtt) ufvtt (dyn_idx=12) -----(PID subtitle.vtt) writegen (dyn_idx=13) ------(PID subtitle.vtt) fout (dst=res/subtitle_dashinit.vtt:gfopt:cmaf=cmf2:rawsub=true:segdur=1:cat=auto:noinitraw:frag:xps_inband=no:psshs=moov:cmaf=cmf2) (idx=9) ----(PID P1V2) mp4mx (dyn_idx=14) -----(PID P1V2) fout (dst=res/output_dash_track256_init.mp4:gfopt:cmaf=cmf2:rawsub=true:segdur=1:cat=auto:frag:xps_inband=no:psshs=moov:mime=video/mp4:cmaf=cmf2) (idx=10) ----(PID P1A3) mp4mx (dyn_idx=15) -----(PID P1A3) fout (dst=res/output_dash_track257_init.mp4:gfopt:cmaf=cmf2:rawsub=true:segdur=1:cat=auto:frag:xps_inband=no:psshs=moov:mime=audio/mp4:cmaf=cmf2) (idx=11) --(PID P1A3) rfadts (dyn_idx=7) ---(PID P1A3) dasher (dyn_idx=8) fin (src=subtitle.vtt) (idx=2) -(PID subtitle.vtt) txtin (dyn_idx=5) ``` Note in the line `fout (dst=res/subtitle_dashinit.vtt:gfopt:cmaf=cmf2:rawsub=true:segdur=1:cat=auto:noinitraw:frag:xps_inband=no:psshs=moov:cmaf=cmf2) (idx=9)` how the flag `noinitraw` tells the `dst` to not generate an init file. Seems like GPAC shouldn't add the `#EXT-X-MAP:URI` to the subtitles playlist if `noinitraw` flag is there? The following are the attachments to run the GPAC command: - [output.txt](https://github.com/gpac/gpac/files/14563670/output.txt): Rename to output.ts - [subtitle.txt](https://github.com/gpac/gpac/files/14563673/subtitle.txt): Rename to subtitle.vtt
noinitraw flag doesn't remove the #EXT-X-MAP:URI tag in the subtitles playlist when rawsub=true for HLS
https://api.github.com/repos/gpac/gpac/issues/2777/comments
1
2024-03-11T19:27:44
2024-03-26T15:01:12Z
https://github.com/gpac/gpac/issues/2777
2,180,039,825
2,777
false
This is a GitHub Issue repo:gpac owner:gpac Title : noinitraw flag doesn't remove the #EXT-X-MAP:URI tag in the subtitles playlist when rawsub=true for HLS Issue date: --- start body --- Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at https://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Trying the following GPAC command to generate CMAF2 HLS with raw subtitles (rawsub) is creating a m3u8 playlist with an init file, which GPAC doesn't create: ``` gpac \ -i output.ts \ -i subtitle.vtt \ -o res/live.m3u8:cmaf=cmf2:rawsub=true:segdur=1:cat=auto ``` We set the `rawsub=true`, which works fine, however, the following is the subtitles playlist generated: ``` #EXTM3U #EXT-X-TARGETDURATION:1 #EXT-X-VERSION:6 #EXT-X-MEDIA-SEQUENCE:1 #EXT-X-INDEPENDENT-SEGMENTS #EXT-X-MAP:URI="subtitle_dashinit.vtt" #EXTINF:1, subtitle_dash1.vtt #EXTINF:1, subtitle_dash2.vtt #EXTINF:1, subtitle_dash3.vtt #EXTINF:1, subtitle_dash4.vtt #EXTINF:1, subtitle_dash5.vtt #EXTINF:1, subtitle_dash6.vtt #EXTINF:1, subtitle_dash7.vtt #EXTINF:1, subtitle_dash8.vtt #EXTINF:1, subtitle_dash9.vtt #EXTINF:1, subtitle_dash10.vtt #EXTINF:1, subtitle_dash11.vtt #EXTINF:1, subtitle_dash12.vtt #EXTINF:1, subtitle_dash13.vtt #EXTINF:1, subtitle_dash14.vtt #EXTINF:1, subtitle_dash15.vtt #EXTINF:1, subtitle_dash16.vtt #EXTINF:1, subtitle_dash17.vtt #EXTINF:1, subtitle_dash18.vtt #EXTINF:1, subtitle_dash19.vtt #EXTINF:1, subtitle_dash20.vtt #EXT-X-ENDLIST ``` Look to the tag `#EXT-X-MAP:URI="subtitle_dashinit.vtt"`, but the `subtitle_dashinit.vtt` file is not created, so, when we play the HLS VOD, subtitles don't show up. The following is the filter graph of that command: ``` Filters connected: fin (src=output.ts) (idx=1) -(PID output.ts) m2tsdmx (dyn_idx=4) --(PID P1V2) rfnalu (dyn_idx=6) ---(PID P1V2) dasher (dyn_idx=8) ----(PID manifest_m3u8) fout (dst=res/live.m3u8:cmaf=cmf2:rawsub=true:segdur=1:cat=auto) (idx=3) ----(PID subtitle.vtt) ufvtt (dyn_idx=12) -----(PID subtitle.vtt) writegen (dyn_idx=13) ------(PID subtitle.vtt) fout (dst=res/subtitle_dashinit.vtt:gfopt:cmaf=cmf2:rawsub=true:segdur=1:cat=auto:noinitraw:frag:xps_inband=no:psshs=moov:cmaf=cmf2) (idx=9) ----(PID P1V2) mp4mx (dyn_idx=14) -----(PID P1V2) fout (dst=res/output_dash_track256_init.mp4:gfopt:cmaf=cmf2:rawsub=true:segdur=1:cat=auto:frag:xps_inband=no:psshs=moov:mime=video/mp4:cmaf=cmf2) (idx=10) ----(PID P1A3) mp4mx (dyn_idx=15) -----(PID P1A3) fout (dst=res/output_dash_track257_init.mp4:gfopt:cmaf=cmf2:rawsub=true:segdur=1:cat=auto:frag:xps_inband=no:psshs=moov:mime=audio/mp4:cmaf=cmf2) (idx=11) --(PID P1A3) rfadts (dyn_idx=7) ---(PID P1A3) dasher (dyn_idx=8) fin (src=subtitle.vtt) (idx=2) -(PID subtitle.vtt) txtin (dyn_idx=5) ``` Note in the line `fout (dst=res/subtitle_dashinit.vtt:gfopt:cmaf=cmf2:rawsub=true:segdur=1:cat=auto:noinitraw:frag:xps_inband=no:psshs=moov:cmaf=cmf2) (idx=9)` how the flag `noinitraw` tells the `dst` to not generate an init file. Seems like GPAC shouldn't add the `#EXT-X-MAP:URI` to the subtitles playlist if `noinitraw` flag is there? The following are the attachments to run the GPAC command: - [output.txt](https://github.com/gpac/gpac/files/14563670/output.txt): Rename to output.ts - [subtitle.txt](https://github.com/gpac/gpac/files/14563673/subtitle.txt): Rename to subtitle.vtt --- end body ---
3,962
[ 0.011308244429528713, -0.010637933388352394, -0.02289506047964096, 0.03327183425426483, -0.0017617444973438978, -0.011517173610627651, -0.04885440319776535, 0.039870485663414, -0.047426726669073105, 0.024462023749947548, 0.00011636599811026827, 0.002685599261894822, 0.028274964541196823, -...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
Is there any way of using gpac to input a video file and output different qualities, such as, 480p, 720p, 1080p, and forcing some of them to be always generated, such as 480p, and the others only if the video size is at least that size? What I mean would be, if the input video is 720p, then only 480p and 720p should be generated, but not upscaling to 1080p. Then, if a video is 240p, then upscale to 480p to have that size at least, and the rest should not be generated. Among this, also do DRM accordingly
Best flow for performing selective multi-quality
https://api.github.com/repos/gpac/gpac/issues/2699/comments
2
2023-12-04T16:35:53
2023-12-05T12:41:16Z
https://github.com/gpac/gpac/issues/2699
2,024,270,799
2,699
false
This is a GitHub Issue repo:gpac owner:gpac Title : Best flow for performing selective multi-quality Issue date: --- start body --- Is there any way of using gpac to input a video file and output different qualities, such as, 480p, 720p, 1080p, and forcing some of them to be always generated, such as 480p, and the others only if the video size is at least that size? What I mean would be, if the input video is 720p, then only 480p and 720p should be generated, but not upscaling to 1080p. Then, if a video is 240p, then upscale to 480p to have that size at least, and the rest should not be generated. Among this, also do DRM accordingly --- end body ---
671
[ -0.028233429417014122, 0.028904201462864876, -0.0325324721634388, 0.0034681991674005985, -0.009276480413973331, 0.013720348477363586, -0.003843603190034628, 0.04619184136390686, -0.022257454693317413, 0.034849684685468674, 0.0167998056858778, -0.003788340836763382, 0.031892187893390656, 0....
null
null
null
null
null
null
null
null
null
[ "openlink", "virtuoso-opensource" ]
Assertion `n->n_neg == 0 || !num_is_zero (n)` failed on Virtuoso **with debug enabled**. It can be reproduced on the commit 2bbd7e0. PoC: ```sql SELECT CAST('-39999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999' AS DECIMAL) ``` Compilation Command: ```bash CC=clang-12 CXX=clang++-12 CFLAGS="-O2 -m64 -g" /root/virtuoso/configure --with-debug ``` Server log. The server was abnormally shut down due to the assertion failure. ``` 07:48:30 Initializing PRNG 07:48:30 OpenLink Virtuoso Universal Server 07:48:30 Version 07.20.3239-pthreads for Linux as of Apr 29 2024 (2bbd7e0) 07:48:30 uses OpenSSL 1.1.1f 31 Mar 2020 07:48:30 uses parts of PCRE, Html Tidy 07:48:30 Database version 3126 07:48:30 SQL Optimizer enabled (max 1000 layouts) 07:48:31 Compiler unit is timed at 0.000254 msec 07:48:31 Roll forward started 07:48:31 3 transactions, 185 bytes replayed (100 %) 07:48:31 Roll forward complete 07:48:32 PL LOG: Can't get list of vad packages in ../vad/ 07:48:32 Checkpoint started 07:48:32 Checkpoint finished, log reused 07:48:32 HTTP/WebDAV server online at 8890 07:48:32 Server online at 1111 (pid 924) virtuoso-t: numeric.c:1399: int _numeric_normalize(numeric_t): Assertion `n->n_neg == 0 || !num_is_zero (n)' failed. fish: '/root/virtuoso-original/binsrc/…' terminated by signal SIGABRT (Abort) ``` On the non-debug version, the server works normally. The output of the SQL is: `*** Error 22003: [Virtuoso Driver][Virtuoso Server]SR127: Can't convert -39999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999 to numeric`.
Assertion 'n->n_neg == 0 || !num_is_zero (n)' failed when compiled with --with-debug
https://api.github.com/repos/openlink/virtuoso-opensource/issues/1279/comments
1
2024-04-29T07:53:39
2024-05-03T10:57:39Z
https://github.com/openlink/virtuoso-opensource/issues/1279
2,268,362,852
1,279
false
This is a GitHub Issue repo:virtuoso-opensource owner:openlink Title : Assertion 'n->n_neg == 0 || !num_is_zero (n)' failed when compiled with --with-debug Issue date: --- start body --- Assertion `n->n_neg == 0 || !num_is_zero (n)` failed on Virtuoso **with debug enabled**. It can be reproduced on the commit 2bbd7e0. PoC: ```sql SELECT CAST('-39999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999' AS DECIMAL) ``` Compilation Command: ```bash CC=clang-12 CXX=clang++-12 CFLAGS="-O2 -m64 -g" /root/virtuoso/configure --with-debug ``` Server log. The server was abnormally shut down due to the assertion failure. ``` 07:48:30 Initializing PRNG 07:48:30 OpenLink Virtuoso Universal Server 07:48:30 Version 07.20.3239-pthreads for Linux as of Apr 29 2024 (2bbd7e0) 07:48:30 uses OpenSSL 1.1.1f 31 Mar 2020 07:48:30 uses parts of PCRE, Html Tidy 07:48:30 Database version 3126 07:48:30 SQL Optimizer enabled (max 1000 layouts) 07:48:31 Compiler unit is timed at 0.000254 msec 07:48:31 Roll forward started 07:48:31 3 transactions, 185 bytes replayed (100 %) 07:48:31 Roll forward complete 07:48:32 PL LOG: Can't get list of vad packages in ../vad/ 07:48:32 Checkpoint started 07:48:32 Checkpoint finished, log reused 07:48:32 HTTP/WebDAV server online at 8890 07:48:32 Server online at 1111 (pid 924) virtuoso-t: numeric.c:1399: int _numeric_normalize(numeric_t): Assertion `n->n_neg == 0 || !num_is_zero (n)' failed. fish: '/root/virtuoso-original/binsrc/…' terminated by signal SIGABRT (Abort) ``` On the non-debug version, the server works normally. The output of the SQL is: `*** Error 22003: [Virtuoso Driver][Virtuoso Server]SR127: Can't convert -39999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999999 to numeric`. --- end body ---
1,863
[ -0.005919383838772774, 0.005639068782329559, -0.01143103651702404, -0.010579169727861881, 0.05099552869796753, -0.00795803964138031, -0.014853066764771938, 0.03625895828008652, -0.0018539029406383634, 0.02526477538049221, -0.014583672396838665, -0.020095326006412506, -0.01325126551091671, ...
null
null
null
null
null
null
null
null
null
[ "ImageMagick", "ImageMagick" ]
### ImageMagick version 7.1.1-33 ### Operating system Linux ### Operating system, version and so on openSUSE Tumbleweed ### Description This version fails on one of the tests: ``` [ 216s] t/write.t ......... [ 216s] 1..32 [ 216s] AVS X image file ... [ 216s] ok 1 [ 216s] Microsoft Windows bitmap image file ... [ 216s] ReadImage() [ 216s] Image: BMP:input.bmp, signatures do not match. [ 216s] Expected: d7324c919f04f4c118da68061a5dbb3f07ebab76b471ecfb0ac822453f677983 [ 216s] Computed: a0c5566d67ad974ba31f6f5d9ef1661b2613b60ac934f2339a2efafed1ecb68b [ 216s] Depth: 16 [ 216s] not ok 2 [ 216s] Microsoft Windows 24-bit bitmap image file ... [ 216s] ReadImage() [ 216s] Image: BMP:input.bmp24, signatures do not match. [ 216s] Expected: fb6fc68beb3b1001c5ebaa671c8ac8fddea06995027127765ff508f77723cc52 [ 216s] Computed: 74136c90d3e699ea5bcbf4aa733aff0dc822b6af72fce00f0c7647bcb0d49f66 [ 216s] Depth: 16 [ 216s] not ok 3 ``` ### Steps to Reproduce Build and run tests. This worked correctly up to 7.1.1-32 but now fails in 7.1.1-33. ### Images _No response_
Test fails for Microsoft Windows (24-bit) bitmap image file
https://api.github.com/repos/ImageMagick/ImageMagick/issues/7344/comments
9
2024-05-26T12:37:38
2024-06-22T18:26:31Z
https://github.com/ImageMagick/ImageMagick/issues/7344
2,317,686,524
7,344
false
This is a GitHub Issue repo:ImageMagick owner:ImageMagick Title : Test fails for Microsoft Windows (24-bit) bitmap image file Issue date: --- start body --- ### ImageMagick version 7.1.1-33 ### Operating system Linux ### Operating system, version and so on openSUSE Tumbleweed ### Description This version fails on one of the tests: ``` [ 216s] t/write.t ......... [ 216s] 1..32 [ 216s] AVS X image file ... [ 216s] ok 1 [ 216s] Microsoft Windows bitmap image file ... [ 216s] ReadImage() [ 216s] Image: BMP:input.bmp, signatures do not match. [ 216s] Expected: d7324c919f04f4c118da68061a5dbb3f07ebab76b471ecfb0ac822453f677983 [ 216s] Computed: a0c5566d67ad974ba31f6f5d9ef1661b2613b60ac934f2339a2efafed1ecb68b [ 216s] Depth: 16 [ 216s] not ok 2 [ 216s] Microsoft Windows 24-bit bitmap image file ... [ 216s] ReadImage() [ 216s] Image: BMP:input.bmp24, signatures do not match. [ 216s] Expected: fb6fc68beb3b1001c5ebaa671c8ac8fddea06995027127765ff508f77723cc52 [ 216s] Computed: 74136c90d3e699ea5bcbf4aa733aff0dc822b6af72fce00f0c7647bcb0d49f66 [ 216s] Depth: 16 [ 216s] not ok 3 ``` ### Steps to Reproduce Build and run tests. This worked correctly up to 7.1.1-32 but now fails in 7.1.1-33. ### Images _No response_ --- end body ---
1,321
[ 0.005327433813363314, 0.01317630335688591, -0.004286441020667553, 0.026381423696875572, 0.0154744116589427, 0.034205079078674316, -0.005248188506811857, 0.04506886377930641, 0.0071752858348190784, 0.013190711848437786, -0.01069088838994503, -0.006245956756174564, 0.014004775322973728, 0.01...
null
null
null
null
null
null
null
null
null
[ "Piwigo", "Piwigo" ]
If we have multiple lines of text the icon is no longer aligned to the center ![image](https://github.com/Piwigo/Piwigo/assets/54360213/1b1813d5-d3a5-4173-8e22-e5e2bded6c22)
Admin messages, icon and text misaligned
https://api.github.com/repos/Piwigo/Piwigo/issues/2113/comments
0
2024-02-09T10:21:38
2024-02-09T10:23:36Z
https://github.com/Piwigo/Piwigo/issues/2113
2,126,847,029
2,113
false
This is a GitHub Issue repo:Piwigo owner:Piwigo Title : Admin messages, icon and text misaligned Issue date: --- start body --- If we have multiple lines of text the icon is no longer aligned to the center ![image](https://github.com/Piwigo/Piwigo/assets/54360213/1b1813d5-d3a5-4173-8e22-e5e2bded6c22) --- end body ---
323
[ -0.013049580156803131, 0.015967590734362602, -0.00862072128802538, 0.02728413976728916, 0.037563830614089966, -0.009479830972850323, -0.00656552379950881, 0.05957481637597084, -0.01519735436886549, -0.02365514077246189, 0.002521783346310258, -0.01377537939697504, -0.00391413364559412, 0.00...
CVE-2022-29770
2022-06-03T21:15:07.893000
XXL-Job v2.3.0 was discovered to contain a stored cross-site scripting (XSS) vulnerability via /xxl-job-admin/jobinfo.
{ "cvssMetricV2": [ { "acInsufInfo": false, "baseSeverity": "LOW", "cvssData": { "accessComplexity": "MEDIUM", "accessVector": "NETWORK", "authentication": "SINGLE", "availabilityImpact": "NONE", "baseScore": 3.5, "confidentialityImpact": "NONE", "integrityImpact": "PARTIAL", "vectorString": "AV:N/AC:M/Au:S/C:N/I:P/A:N", "version": "2.0" }, "exploitabilityScore": 6.8, "impactScore": 2.9, "obtainAllPrivilege": false, "obtainOtherPrivilege": false, "obtainUserPrivilege": false, "source": "nvd@nist.gov", "type": "Primary", "userInteractionRequired": true } ], "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "LOW", "attackVector": "NETWORK", "availabilityImpact": "NONE", "baseScore": 5.4, "baseSeverity": "MEDIUM", "confidentialityImpact": "LOW", "integrityImpact": "LOW", "privilegesRequired": "LOW", "scope": "CHANGED", "userInteraction": "REQUIRED", "vectorString": "CVSS:3.1/AV:N/AC:L/PR:L/UI:R/S:C/C:L/I:L/A:N", "version": "3.1" }, "exploitabilityScore": 2.3, "impactScore": 2.7, "source": "nvd@nist.gov", "type": "Primary" } ] }
[ { "source": "cve@mitre.org", "tags": [ "Exploit", "Issue Tracking", "Third Party Advisory" ], "url": "https://github.com/xuxueli/xxl-job/issues/2836" } ]
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:xuxueli:xxl-job:2.3.0:*:*:*:*:*:*:*", "matchCriteriaId": "233AAF33-91ED-425F-951A-1DAAC0B3A4C5", "versionEndExcluding": null, "versionEndIncluding": null, "versionStartExclud...
https://github.com/xuxueli/xxl-job/issues/2836
[ "Exploit", "Issue Tracking", "Third Party Advisory" ]
github.com
[ "xuxueli", "xxl-job" ]
### Which version of XXL-JOB do you using? 2.3.0 ### Expected behavior stroed XSS 1、Insert poc <script>alert(document.cookie)</script> ![image](https://user-images.githubusercontent.com/33409640/164911443-ae80fdc1-0fb8-4391-b1d1-fa83f50d624e.png) 2、press submit ![image](https://user-images.githubusercontent.com/33409640/164911508-a9c88111-2fbf-4835-89bd-69598fd53a28.png) 3、XSS attack occurs ![image](https://user-images.githubusercontent.com/33409640/164911605-e0eb3c35-7697-4c44-b29d-b3a3779d3cc3.png)
There is a stored XSS vulnerability in the task management of xxl-job
https://api.github.com/repos/xuxueli/xxl-job/issues/2836/comments
0
2022-04-23T15:02:41
2022-04-23T15:02:41Z
https://github.com/xuxueli/xxl-job/issues/2836
1,213,336,856
2,836
true
This is a GitHub Issue repo:xxl-job owner:xuxueli Title : There is a stored XSS vulnerability in the task management of xxl-job Issue date: --- start body --- ### Which version of XXL-JOB do you using? 2.3.0 ### Expected behavior stroed XSS 1、Insert poc <script>alert(document.cookie)</script> ![image](https://user-images.githubusercontent.com/33409640/164911443-ae80fdc1-0fb8-4391-b1d1-fa83f50d624e.png) 2、press submit ![image](https://user-images.githubusercontent.com/33409640/164911508-a9c88111-2fbf-4835-89bd-69598fd53a28.png) 3、XSS attack occurs ![image](https://user-images.githubusercontent.com/33409640/164911605-e0eb3c35-7697-4c44-b29d-b3a3779d3cc3.png) --- end body ---
696
[ -0.017489472404122353, 0.0008769726264290512, -0.004313131794333458, 0.009796177968382835, 0.059384237974882126, 0.0053460621275007725, 0.009425952099263668, 0.04238346219062805, 0.015208881348371506, 0.015016363002359867, 0.0033857161179184914, -0.0015938226133584976, -0.042442698031663895,...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
Potential memory leak in [ff_mx.c](https://github.com/gpac/gpac/blob/master/src/filters/ff_mx.c#L373) line 373. Calling av_dict_copy() without calling av_dict_free() to free the memory will cause a memory leak.[Doc](https://ffmpeg.org/doxygen/trunk/group__lavu__dict.html#ga1bafd682b1fbb90e48a4cc3814b820f7) says "av_dict_free() Free all the memory allocated for an [AVDictionary](https://ffmpeg.org/doxygen/trunk/structAVDictionary.html) struct and all keys and values." ![image](https://user-images.githubusercontent.com/39425646/214762607-08c18e82-4fa0-4992-81ac-93b238b67736.png)
Potential memory leak in src/filters/ff_mx.c
https://api.github.com/repos/gpac/gpac/issues/2383/comments
4
2023-01-26T05:09:03
2023-05-10T12:47:32Z
https://github.com/gpac/gpac/issues/2383
1,557,638,001
2,383
false
This is a GitHub Issue repo:gpac owner:gpac Title : Potential memory leak in src/filters/ff_mx.c Issue date: --- start body --- Potential memory leak in [ff_mx.c](https://github.com/gpac/gpac/blob/master/src/filters/ff_mx.c#L373) line 373. Calling av_dict_copy() without calling av_dict_free() to free the memory will cause a memory leak.[Doc](https://ffmpeg.org/doxygen/trunk/group__lavu__dict.html#ga1bafd682b1fbb90e48a4cc3814b820f7) says "av_dict_free() Free all the memory allocated for an [AVDictionary](https://ffmpeg.org/doxygen/trunk/structAVDictionary.html) struct and all keys and values." ![image](https://user-images.githubusercontent.com/39425646/214762607-08c18e82-4fa0-4992-81ac-93b238b67736.png) --- end body ---
733
[ -0.03502077981829643, 0.04273533821105957, -0.009768077172338963, 0.025238825008273125, 0.003957875072956085, 0.01635875552892685, 0.0009651873260736465, 0.03371651843190193, -0.02281068079173565, 0.016539130359888077, -0.01555399876087904, 0.013188292272388935, 0.003652622690424323, 0.005...
null
null
null
null
null
null
null
null
null
[ "kubernetes", "kubernetes" ]
Hello, I know that in case of node resource pressure, the controller prioritizes pods to be evicted, for example in case of node memory pressure, pods using more memory than requested are prioritised for deletion. But for HPA scale-in I struggle to find if and how pods are sorted to prioritize them for deletion. While looking at the PodDeletionCost feature PR I found this priorization comment in the code https://github.com/ahg-g/kubernetes/blob/2d88d2d993b36c231fdc7537212685544469c517/pkg/controller/controller_utils.go#L785 Does it mean that no resource usage is taken into account when the controller has to select pods to be deleted during a scale in event ? For example, if a scale-in occurs because the CPU usage is lower than the threshold set in the HPA, the controller will select pods to be deleted without prioritizing them by their current CPU usage, right ?
HPA scale-in: is pod resource usage taken into account to prioritize pods to be deleted ?
https://api.github.com/repos/kubernetes/kubernetes/issues/124821/comments
4
2024-05-11T11:26:41
2024-05-14T06:07:21Z
https://github.com/kubernetes/kubernetes/issues/124821
2,290,849,433
124,821
false
This is a GitHub Issue repo:kubernetes owner:kubernetes Title : HPA scale-in: is pod resource usage taken into account to prioritize pods to be deleted ? Issue date: --- start body --- Hello, I know that in case of node resource pressure, the controller prioritizes pods to be evicted, for example in case of node memory pressure, pods using more memory than requested are prioritised for deletion. But for HPA scale-in I struggle to find if and how pods are sorted to prioritize them for deletion. While looking at the PodDeletionCost feature PR I found this priorization comment in the code https://github.com/ahg-g/kubernetes/blob/2d88d2d993b36c231fdc7537212685544469c517/pkg/controller/controller_utils.go#L785 Does it mean that no resource usage is taken into account when the controller has to select pods to be deleted during a scale in event ? For example, if a scale-in occurs because the CPU usage is lower than the threshold set in the HPA, the controller will select pods to be deleted without prioritizing them by their current CPU usage, right ? --- end body ---
1,084
[ -0.03514720872044563, -0.005093557760119438, -0.015569363720715046, 0.03225366771221161, 0.004048299975693226, -0.0410139262676239, 0.001180477673187852, 0.03368716686964035, -0.02138298749923706, 0.03230676054954529, -0.0004326205817051232, 0.0034045539796352386, -0.0004898608895018697, 0...
null
null
null
null
null
null
null
null
null
[ "Piwigo", "Piwigo" ]
I updated to 14.0.0, which went wrong somehow, so I started with a fresh install. It seems that I am unable to edit users (or the UI is confusing). More precisely, when hovering the pen (see screenshot below), the background turns orange, but clicking does nothing. The console is clean (apart from two lines like the following, which are displayed every after every click). The Network pane (Firefox) also shows no blocked connections. Any ideas? Thank you! 20:10:39.653 false 20:10:39.655 2023/12/12 ![image](https://github.com/Piwigo/Piwigo/assets/26614/cb0ddb24-299b-482e-8249-df0899084ada)
[v14.0.0] Cannot edit user
https://api.github.com/repos/Piwigo/Piwigo/issues/2065/comments
3
2023-12-12T19:12:51
2023-12-13T21:00:08Z
https://github.com/Piwigo/Piwigo/issues/2065
2,038,400,525
2,065
false
This is a GitHub Issue repo:Piwigo owner:Piwigo Title : [v14.0.0] Cannot edit user Issue date: --- start body --- I updated to 14.0.0, which went wrong somehow, so I started with a fresh install. It seems that I am unable to edit users (or the UI is confusing). More precisely, when hovering the pen (see screenshot below), the background turns orange, but clicking does nothing. The console is clean (apart from two lines like the following, which are displayed every after every click). The Network pane (Firefox) also shows no blocked connections. Any ideas? Thank you! 20:10:39.653 false 20:10:39.655 2023/12/12 ![image](https://github.com/Piwigo/Piwigo/assets/26614/cb0ddb24-299b-482e-8249-df0899084ada) --- end body ---
747
[ -0.05543672665953636, 0.016951967030763626, -0.004774123430252075, -0.0054014334455132484, 0.010248499922454357, -0.02534041926264763, -0.0004326435155235231, 0.03793039172887802, 0.025179944932460785, -0.017316682264208794, 0.007414661813527346, -0.007462074980139732, -0.025121590122580528,...
CVE-2023-46871
2023-12-07T18:15:08.030000
GPAC version 2.3-DEV-rev602-ged8424300-master in MP4Box contains a memory leak in NewSFDouble scenegraph/vrml_tools.c:300. This vulnerability may lead to a denial of service.
{ "cvssMetricV2": null, "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "LOW", "attackVector": "NETWORK", "availabilityImpact": "LOW", "baseScore": 5.3, "baseSeverity": "MEDIUM", "confidentialityImpact": "NONE", "integrityImpact": "NONE", "privilegesRequired": "NONE", "scope": "UNCHANGED", "userInteraction": "NONE", "vectorString": "CVSS:3.1/AV:N/AC:L/PR:N/UI:N/S:U/C:N/I:N/A:L", "version": "3.1" }, "exploitabilityScore": 3.9, "impactScore": 1.4, "source": "nvd@nist.gov", "type": "Primary" } ] }
[ { "source": "cve@mitre.org", "tags": [ "Exploit", "Third Party Advisory" ], "url": "https://gist.github.com/ReturnHere/d0899bb03b8f5e8fae118f2b76888486" }, { "source": "cve@mitre.org", "tags": [ "Exploit", "Issue Tracking", "Patch", "Third Party Adviso...
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:gpac:gpac:*:*:*:*:*:*:*:*", "matchCriteriaId": "373C4F43-207C-47CB-B454-BD90A52F0ECF", "versionEndExcluding": null, "versionEndIncluding": "2.3-dev-rev602-ged8424300-master", ...
https://github.com/gpac/gpac/issues/2658
[ "Exploit", "Issue Tracking", "Patch", "Third Party Advisory" ]
github.com
[ "gpac", "gpac" ]
Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [ ] I looked for a similar issue and couldn't find any. - [ ] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [ ] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/bug-reporting/ 1、version ./MP4Box -version MP4Box - GPAC version 2.3-DEV-rev602-ged8424300-master (c) 2000-2023 Telecom Paris distributed under LGPL v2.1+ - http://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: Features: GPAC_CONFIG_LINUX GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SSL GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_JPEG GPAC_HAS_PNG GPAC_HAS_LINUX_DVB GPAC_DISABLE_3D 2、platform uname -a Linux returnzero-virtual-machine 6.2.0-35-generic #35~22.04.1-Ubuntu SMP PREEMPT_DYNAMIC Fri Oct 6 10:23:26 UTC 2 x86_64 x86_64 x86_64 GNU/Linux 3、Reproduce ./MP4Box -bt $poc 4、ASCAN ./MP4Box -bt '/home/returnzero/gpac/out/default/crashes/id:000000,sig:06,src:000008,time:167295,execs:4216,op:havoc,rep:6' [iso file] extra box maxr found in hinf, deleting [iso file] Unknown box type traI in parent moov [iso file] Box "stss" (start 9939) has 32 extra bytes [iso file] extra box maxr found in hinf, deleting [iso file] Track with no sample description box ! [iso file] Incomplete box mdat - start 11495 size 861217 [iso file] Incomplete file while reading for dump - aborting parsing [iso file] extra box maxr found in hinf, deleting [iso file] Unknown box type traI in parent moov [iso file] Box "stss" (start 9939) has 32 extra bytes [iso file] extra box maxr found in hinf, deleting [iso file] Track with no sample description box ! [iso file] Incomplete box mdat - start 11495 size 861217 [iso file] Incomplete file while reading for dump - aborting parsing MPEG-4 BIFS Scene Parsing [MP4 Loading] decoding sample 1 from track ID 8 failed Error loading scene: BitStream Not Compliant Error: BitStream Not Compliant ================================================================= ==3703==ERROR: LeakSanitizer: detected memory leaks Direct leak of 8 byte(s) in 1 object(s) allocated from: #0 0x7f0d974b4887 in __interceptor_malloc ../../../../src/libsanitizer/asan/asan_malloc_linux.cpp:145 #1 0x7f0d95d38b3b in NewSFDouble scenegraph/vrml_tools.c:300 #2 0x7f0d95d38b3b in gf_sg_vrml_field_pointer_new scenegraph/vrml_tools.c:558 SUMMARY: AddressSanitizer: 8 byte(s) leaked in 1 allocation(s). 5、Impact This vulnerability allows a remote attacker to cause a denial of service on an affected gpac MP4Box. Exploiting this vulnerability requires user interaction, as the target must access a malicious page or open a malicious file. 6、poc https://github.com/ReturnHere/CrashReport/blob/main/id%5E%25000000%2Csig%5E%2506%2Csrc%5E%25000008%2Ctime%5E%25167295%2Cexecs%5E%254216%2Cop%5E%25havoc%2Crep%5E%256
Memory leaks in NewSFDouble scenegraph/vrml_tools.c:300
https://api.github.com/repos/gpac/gpac/issues/2658/comments
2
2023-10-23T06:39:49
2023-12-09T11:54:02Z
https://github.com/gpac/gpac/issues/2658
1,956,459,097
2,658
true
This is a GitHub Issue repo:gpac owner:gpac Title : Memory leaks in NewSFDouble scenegraph/vrml_tools.c:300 Issue date: --- start body --- Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [ ] I looked for a similar issue and couldn't find any. - [ ] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [ ] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/bug-reporting/ 1、version ./MP4Box -version MP4Box - GPAC version 2.3-DEV-rev602-ged8424300-master (c) 2000-2023 Telecom Paris distributed under LGPL v2.1+ - http://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: Features: GPAC_CONFIG_LINUX GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SSL GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_JPEG GPAC_HAS_PNG GPAC_HAS_LINUX_DVB GPAC_DISABLE_3D 2、platform uname -a Linux returnzero-virtual-machine 6.2.0-35-generic #35~22.04.1-Ubuntu SMP PREEMPT_DYNAMIC Fri Oct 6 10:23:26 UTC 2 x86_64 x86_64 x86_64 GNU/Linux 3、Reproduce ./MP4Box -bt $poc 4、ASCAN ./MP4Box -bt '/home/returnzero/gpac/out/default/crashes/id:000000,sig:06,src:000008,time:167295,execs:4216,op:havoc,rep:6' [iso file] extra box maxr found in hinf, deleting [iso file] Unknown box type traI in parent moov [iso file] Box "stss" (start 9939) has 32 extra bytes [iso file] extra box maxr found in hinf, deleting [iso file] Track with no sample description box ! [iso file] Incomplete box mdat - start 11495 size 861217 [iso file] Incomplete file while reading for dump - aborting parsing [iso file] extra box maxr found in hinf, deleting [iso file] Unknown box type traI in parent moov [iso file] Box "stss" (start 9939) has 32 extra bytes [iso file] extra box maxr found in hinf, deleting [iso file] Track with no sample description box ! [iso file] Incomplete box mdat - start 11495 size 861217 [iso file] Incomplete file while reading for dump - aborting parsing MPEG-4 BIFS Scene Parsing [MP4 Loading] decoding sample 1 from track ID 8 failed Error loading scene: BitStream Not Compliant Error: BitStream Not Compliant ================================================================= ==3703==ERROR: LeakSanitizer: detected memory leaks Direct leak of 8 byte(s) in 1 object(s) allocated from: #0 0x7f0d974b4887 in __interceptor_malloc ../../../../src/libsanitizer/asan/asan_malloc_linux.cpp:145 #1 0x7f0d95d38b3b in NewSFDouble scenegraph/vrml_tools.c:300 #2 0x7f0d95d38b3b in gf_sg_vrml_field_pointer_new scenegraph/vrml_tools.c:558 SUMMARY: AddressSanitizer: 8 byte(s) leaked in 1 allocation(s). 5、Impact This vulnerability allows a remote attacker to cause a denial of service on an affected gpac MP4Box. Exploiting this vulnerability requires user interaction, as the target must access a malicious page or open a malicious file. 6、poc https://github.com/ReturnHere/CrashReport/blob/main/id%5E%25000000%2Csig%5E%2506%2Csrc%5E%25000008%2Ctime%5E%25167295%2Cexecs%5E%254216%2Cop%5E%25havoc%2Crep%5E%256 --- end body ---
3,539
[ -0.00727617833763361, -0.0027080513536930084, -0.010874289087951183, -0.013037363067269325, 0.03619151934981346, 0.0017148888437077403, -0.001045766519382596, 0.03726884722709656, -0.016227267682552338, 0.02853238396346569, -0.00802105013281107, 0.01831459254026413, 0.019341420382261276, -...
null
null
null
null
null
null
null
null
null
[ "emqx", "nanomq" ]
**Describe the bug** malformed packet may lead to unexpected behavior. This malformed packet will send SUBACK right after CONNECT, and a SUB follows. SUB clients , which should be dropped, will still be inserted into db_tree. **To Reproduce** 1. start nanomq with ` ./nanomq/nanomq start --url tls+nmq-tcp://0.0.0.0:8883 --http --cacert etc/certs/cacert.pem --cert etc/certs/cert.pem --key etc/certs/key.pem --qos_duration 1 --log_level debug --log_stdout false --log_file /tmp/nanomq_test.log` 2. get `test.py, fuzzy_test.txt, fuzzy_test.py, rest_api_test.py` and run the python script to test `python3 test.py` fuzzy file --> [fuzzy_test.txt](https://github.com/nanomq/nanomq/files/15290977/fuzzy_test.txt) test.py --> ``` from cmath import log from fileinput import close import os import subprocess import shlex iimport requests import json from collections import namedtuple base_url = "http://127.0.0.1:8081/api/v4" def test_get_api(): # not all endpoint is tested. paths = ["", "/nodes", "/brokers", "/clients", "/subscriptions", "/topic-tree", "/reload", "/configuration"] for p in paths: print("testing Get API: " + base_url + p) response = requests.get(base_url + p, auth=('admin', 'public'), headers={'Connection':'close'}) if response.status_code != 200: print("test get api failed: " + p) return False return True def test_post_api(): Param = namedtuple('Param', ['path', 'payload']) param1 = Param(path="/reload", payload={'data': { 'property_size': 64, 'max_packet_size': 5120, 'client_max_packet_size': 5, 'msq_len': 2048, 'qos_duration': 10, 'keepalive_backoff': 1250, 'allow_anonymous': False }}) param2 = Param(path="/reload", payload={'data': { 'property_size': 64, 'max_packet_size': 10240, 'client_max_packet_size': 5, 'msq_len': 2048, 'qos_duration': 90, 'keepalive_backoff': 1250, 'allow_anonymous': True }}) param_list = [param1, param2] for item in param_list: print("testing Post API: " + base_url + item.path) response = requests.post( base_url + item.path, json.dumps(item.payload), auth=('admin', 'public'), headers={'Connection':'close'}) if response.status_code != 200: print("test Post api failed: " + item.path) return False return True def rest_api_test(): return test_get_api() and test_post_api()mport time from os.path import exists from fuzzy_test import fuzzy_test from rest_api_test import rest_api_test if __name__=='__main__': print("fuzzy test start") if( False == fuzzy_test()): print("fuzzy test failed") raise AssertionError print("fuzzy test end") print("rest api test start") if False == rest_api_test(): print("rest api test failed") raise AssertionError print("rest api test end") ``` fuzzy_test.py (set right path for the files)--> ``` import time import socket import os data_path = "../../.github/scripts/fuzzy_test.txt" addr = '127.0.0.1' port = 1883 def try_connect(bytes_flow, sleep_time = 0.01): s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) while True: try: s.connect((addr, port)) s.send(bytes_flow) s.close() break except ConnectionResetError: continue except ConnectionRefusedError: return False time.sleep(sleep_time) return True def start_fuzzy_test(fuzzy_data): flag = True for hex_flow in reversed(fuzzy_data): bytes_flow = bytearray.fromhex(hex_flow) status = try_connect(bytes_flow) if status == False: print('[+] A crash was detected') flag = False if(flag == True): print('[-] No crash..') return flag def fuzzy_test(): with open(data_path, 'r') as f: fuzzy_data = f.readlines() if (start_fuzzy_test(fuzzy_data) == False): return False else: return True ``` rest_api_test.py --> ``` import requests import json from collections import namedtuple base_url = "http://127.0.0.1:8081/api/v4" def test_get_api(): # not all endpoint is tested. paths = ["", "/nodes", "/brokers", "/clients", "/subscriptions", "/topic-tree", "/reload", "/configuration"] for p in paths: print("testing Get API: " + base_url + p) response = requests.get(base_url + p, auth=('admin', 'public'), headers={'Connection':'close'}) if response.status_code != 200: print("test get api failed: " + p) return False return True def test_post_api(): Param = namedtuple('Param', ['path', 'payload']) param1 = Param(path="/reload", payload={'data': { 'property_size': 64, 'max_packet_size': 5120, 'client_max_packet_size': 5, 'msq_len': 2048, 'qos_duration': 10, 'keepalive_backoff': 1250, 'allow_anonymous': False }}) param2 = Param(path="/reload", payload={'data': { 'property_size': 64, 'max_packet_size': 10240, 'client_max_packet_size': 5, 'msq_len': 2048, 'qos_duration': 90, 'keepalive_backoff': 1250, 'allow_anonymous': True }}) param_list = [param1, param2] for item in param_list: print("testing Post API: " + base_url + item.path) response = requests.post( base_url + item.path, json.dumps(item.payload), auth=('admin', 'public'), headers={'Connection':'close'}) if response.status_code != 200: print("test Post api failed: " + item.path) return False return True def rest_api_test(): return test_get_api() and test_post_api() ```
Malformed packet may lead to unexpected behavior.
https://api.github.com/repos/nanomq/nanomq/issues/1779/comments
1
2024-05-13T07:48:06
2024-05-13T11:44:24Z
https://github.com/nanomq/nanomq/issues/1779
2,292,084,769
1,779
false
This is a GitHub Issue repo:nanomq owner:emqx Title : Malformed packet may lead to unexpected behavior. Issue date: --- start body --- **Describe the bug** malformed packet may lead to unexpected behavior. This malformed packet will send SUBACK right after CONNECT, and a SUB follows. SUB clients , which should be dropped, will still be inserted into db_tree. **To Reproduce** 1. start nanomq with ` ./nanomq/nanomq start --url tls+nmq-tcp://0.0.0.0:8883 --http --cacert etc/certs/cacert.pem --cert etc/certs/cert.pem --key etc/certs/key.pem --qos_duration 1 --log_level debug --log_stdout false --log_file /tmp/nanomq_test.log` 2. get `test.py, fuzzy_test.txt, fuzzy_test.py, rest_api_test.py` and run the python script to test `python3 test.py` fuzzy file --> [fuzzy_test.txt](https://github.com/nanomq/nanomq/files/15290977/fuzzy_test.txt) test.py --> ``` from cmath import log from fileinput import close import os import subprocess import shlex iimport requests import json from collections import namedtuple base_url = "http://127.0.0.1:8081/api/v4" def test_get_api(): # not all endpoint is tested. paths = ["", "/nodes", "/brokers", "/clients", "/subscriptions", "/topic-tree", "/reload", "/configuration"] for p in paths: print("testing Get API: " + base_url + p) response = requests.get(base_url + p, auth=('admin', 'public'), headers={'Connection':'close'}) if response.status_code != 200: print("test get api failed: " + p) return False return True def test_post_api(): Param = namedtuple('Param', ['path', 'payload']) param1 = Param(path="/reload", payload={'data': { 'property_size': 64, 'max_packet_size': 5120, 'client_max_packet_size': 5, 'msq_len': 2048, 'qos_duration': 10, 'keepalive_backoff': 1250, 'allow_anonymous': False }}) param2 = Param(path="/reload", payload={'data': { 'property_size': 64, 'max_packet_size': 10240, 'client_max_packet_size': 5, 'msq_len': 2048, 'qos_duration': 90, 'keepalive_backoff': 1250, 'allow_anonymous': True }}) param_list = [param1, param2] for item in param_list: print("testing Post API: " + base_url + item.path) response = requests.post( base_url + item.path, json.dumps(item.payload), auth=('admin', 'public'), headers={'Connection':'close'}) if response.status_code != 200: print("test Post api failed: " + item.path) return False return True def rest_api_test(): return test_get_api() and test_post_api()mport time from os.path import exists from fuzzy_test import fuzzy_test from rest_api_test import rest_api_test if __name__=='__main__': print("fuzzy test start") if( False == fuzzy_test()): print("fuzzy test failed") raise AssertionError print("fuzzy test end") print("rest api test start") if False == rest_api_test(): print("rest api test failed") raise AssertionError print("rest api test end") ``` fuzzy_test.py (set right path for the files)--> ``` import time import socket import os data_path = "../../.github/scripts/fuzzy_test.txt" addr = '127.0.0.1' port = 1883 def try_connect(bytes_flow, sleep_time = 0.01): s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) while True: try: s.connect((addr, port)) s.send(bytes_flow) s.close() break except ConnectionResetError: continue except ConnectionRefusedError: return False time.sleep(sleep_time) return True def start_fuzzy_test(fuzzy_data): flag = True for hex_flow in reversed(fuzzy_data): bytes_flow = bytearray.fromhex(hex_flow) status = try_connect(bytes_flow) if status == False: print('[+] A crash was detected') flag = False if(flag == True): print('[-] No crash..') return flag def fuzzy_test(): with open(data_path, 'r') as f: fuzzy_data = f.readlines() if (start_fuzzy_test(fuzzy_data) == False): return False else: return True ``` rest_api_test.py --> ``` import requests import json from collections import namedtuple base_url = "http://127.0.0.1:8081/api/v4" def test_get_api(): # not all endpoint is tested. paths = ["", "/nodes", "/brokers", "/clients", "/subscriptions", "/topic-tree", "/reload", "/configuration"] for p in paths: print("testing Get API: " + base_url + p) response = requests.get(base_url + p, auth=('admin', 'public'), headers={'Connection':'close'}) if response.status_code != 200: print("test get api failed: " + p) return False return True def test_post_api(): Param = namedtuple('Param', ['path', 'payload']) param1 = Param(path="/reload", payload={'data': { 'property_size': 64, 'max_packet_size': 5120, 'client_max_packet_size': 5, 'msq_len': 2048, 'qos_duration': 10, 'keepalive_backoff': 1250, 'allow_anonymous': False }}) param2 = Param(path="/reload", payload={'data': { 'property_size': 64, 'max_packet_size': 10240, 'client_max_packet_size': 5, 'msq_len': 2048, 'qos_duration': 90, 'keepalive_backoff': 1250, 'allow_anonymous': True }}) param_list = [param1, param2] for item in param_list: print("testing Post API: " + base_url + item.path) response = requests.post( base_url + item.path, json.dumps(item.payload), auth=('admin', 'public'), headers={'Connection':'close'}) if response.status_code != 200: print("test Post api failed: " + item.path) return False return True def rest_api_test(): return test_get_api() and test_post_api() ``` --- end body ---
6,096
[ -0.026153314858675003, 0.006424548104405403, 0.005424903240054846, 0.00501854345202446, 0.04183880239725113, 0.018269935622811317, -0.027372395619750023, 0.013092911802232265, -0.038815487176179886, 0.012312701903283596, -0.01622188277542591, 0.03034694865345955, 0.014604570344090462, 0.00...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/2013/07/16/how-to-file-a-bug-properly/ I have seen https://github.com/gpac/gpac/wiki/HEVC-Tile-based-adaptation-guide and successfully used kvazaar and gpac to generate tiled videos. However, I am looking for a method to stitch NxM independently encoded HEVC video stream into one mp4 file. If I can control using the same video encoding parameters, is it possible to perform the stitching? It should perform like `vstack` and `hstack` in ffmpeg. Thank you so much.
How to stitching NxM independently encoded HEVC video stream?
https://api.github.com/repos/gpac/gpac/issues/2405/comments
1
2023-02-27T15:48:21
2023-03-14T11:38:24Z
https://github.com/gpac/gpac/issues/2405
1,601,446,912
2,405
false
This is a GitHub Issue repo:gpac owner:gpac Title : How to stitching NxM independently encoded HEVC video stream? Issue date: --- start body --- Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/2013/07/16/how-to-file-a-bug-properly/ I have seen https://github.com/gpac/gpac/wiki/HEVC-Tile-based-adaptation-guide and successfully used kvazaar and gpac to generate tiled videos. However, I am looking for a method to stitch NxM independently encoded HEVC video stream into one mp4 file. If I can control using the same video encoding parameters, is it possible to perform the stitching? It should perform like `vstack` and `hstack` in ffmpeg. Thank you so much. --- end body ---
1,292
[ -0.032581355422735214, 0.028666000813245773, -0.02602313458919525, -0.01603897660970688, 0.03325255960226059, 0.02210777997970581, -0.003240656340494752, 0.0533607117831707, -0.01395544782280922, 0.03621704503893852, 0.025561682879924774, -0.03960103169083595, 0.002244338160380721, -0.0078...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
## Description There is a heap buffer overflow detected by AddressSanitizer ## System info ``` Ubuntu 20.04.2 LTS clang version 12.0.0-++20210402082642+04ba60cfe598-1~exp1~20210402063359.71 MP4Box - GPAC version 1.1.0-DEV-rev1745-g9e1378706-master ``` ## Build command ``` ./configure --static-mp4box --prefix=`realpath ./install` --enable-sanitizer --cc=clang --cxx=clang++ ``` ## crash command ``` MP4Box -out /dev/null poc_file ``` ## Pocs [poc.zip](https://github.com/gpac/gpac/files/8039045/poc.zip) ## Crash output ``` ==35757==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x602000001270 at pc 0x0000007d0889 bp 0x7fff1400a3d0 sp 0x7fff1400a3c8 READ of size 4 at 0x602000001270 thread T0 #0 0x7d0888 in convert_compact_sample_groups /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:296:18 #1 0x7c9018 in gf_isom_parse_movie_boxes_internal /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:420:7 #2 0x7c9018 in gf_isom_parse_movie_boxes /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:812:6 #3 0x7ccf16 in gf_isom_open_file /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:932:19 #4 0x4e14c6 in mp4boxMain /programs/mp4box/builds/build13/applications/mp4box/main.c:5984:12 #5 0x7fce38d730b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x270b2) #6 0x41ea6d in _start (/programs/mp4box/builds/build13/bin/gcc/MP4Box+0x41ea6d) 0x602000001271 is located 0 bytes to the right of 1-byte region [0x602000001270,0x602000001271) allocated by thread T0 here: #0 0x499ccd in malloc (/programs/mp4box/builds/build13/bin/gcc/MP4Box+0x499ccd) #1 0x1310c8e in csgp_box_read /programs/mp4box/builds/build13/src/isomedia/box_code_base.c:12613:55 #2 0x7adfad in gf_isom_box_read /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1822:9 #3 0x7adfad in gf_isom_box_parse_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:264:14 #4 0x7afda3 in gf_isom_box_array_read_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1719:7 #5 0x12e1c6b in stbl_box_read /programs/mp4box/builds/build13/src/isomedia/box_code_base.c:5069:6 #6 0x7adfad in gf_isom_box_read /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1822:9 #7 0x7adfad in gf_isom_box_parse_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:264:14 #8 0x7afda3 in gf_isom_box_array_read_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1719:7 #9 0x12d7132 in minf_box_read /programs/mp4box/builds/build13/src/isomedia/box_code_base.c:3549:6 #10 0x7adfad in gf_isom_box_read /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1822:9 #11 0x7adfad in gf_isom_box_parse_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:264:14 #12 0x7afda3 in gf_isom_box_array_read_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1719:7 #13 0x12d4665 in mdia_box_read /programs/mp4box/builds/build13/src/isomedia/box_code_base.c:3100:6 #14 0x7adfad in gf_isom_box_read /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1822:9 #15 0x7adfad in gf_isom_box_parse_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:264:14 #16 0x7afda3 in gf_isom_box_array_read_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1719:7 #17 0x12ec620 in trak_box_read /programs/mp4box/builds/build13/src/isomedia/box_code_base.c:6844:6 #18 0x7adfad in gf_isom_box_read /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1822:9 #19 0x7adfad in gf_isom_box_parse_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:264:14 #20 0x7afda3 in gf_isom_box_array_read_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1719:7 #21 0x7adfad in gf_isom_box_read /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1822:9 #22 0x7adfad in gf_isom_box_parse_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:264:14 #23 0x7ad181 in gf_isom_parse_root_box /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:38:8 #24 0x7c8b31 in gf_isom_parse_movie_boxes_internal /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:349:7 #25 0x7c8b31 in gf_isom_parse_movie_boxes /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:812:6 #26 0x7ccf16 in gf_isom_open_file /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:932:19 #27 0x4e14c6 in mp4boxMain /programs/mp4box/builds/build13/applications/mp4box/main.c:5984:12 #28 0x7fce38d730b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x270b2) SUMMARY: AddressSanitizer: heap-buffer-overflow /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:296:18 in convert_compact_sample_groups Shadow bytes around the buggy address: 0x0c047fff81f0: fa fa 00 00 fa fa 00 00 fa fa 00 00 fa fa 00 00 0x0c047fff8200: fa fa 00 05 fa fa 00 00 fa fa 00 00 fa fa 00 00 0x0c047fff8210: fa fa 00 00 fa fa 00 00 fa fa 00 00 fa fa 01 fa 0x0c047fff8220: fa fa fd fa fa fa fd fd fa fa 00 00 fa fa 00 00 0x0c047fff8230: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa 01 fa =>0x0c047fff8240: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa[01]fa 0x0c047fff8250: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa 01 fa 0x0c047fff8260: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa 01 fa 0x0c047fff8270: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa 01 fa 0x0c047fff8280: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa 00 04 0x0c047fff8290: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa 01 fa Shadow byte legend (one shadow byte represents 8 application bytes): Addressable: 00 Partially addressable: 01 02 03 04 05 06 07 Heap left redzone: fa Freed heap region: fd Stack left redzone: f1 Stack mid redzone: f2 Stack right redzone: f3 Stack after return: f5 Stack use after scope: f8 Global redzone: f9 Global init order: f6 Poisoned by user: f7 Container overflow: fc Array cookie: ac Intra object redzone: bb ASan internal: fe Left alloca redzone: ca Right alloca redzone: cb Shadow gap: cc ==35757==ABORTING ```
There is a heap buffer overflow detected by AddressSanitizer
https://api.github.com/repos/gpac/gpac/issues/2116/comments
0
2022-02-10T07:49:52
2022-02-10T20:00:56Z
https://github.com/gpac/gpac/issues/2116
1,129,589,949
2,116
false
This is a GitHub Issue repo:gpac owner:gpac Title : There is a heap buffer overflow detected by AddressSanitizer Issue date: --- start body --- ## Description There is a heap buffer overflow detected by AddressSanitizer ## System info ``` Ubuntu 20.04.2 LTS clang version 12.0.0-++20210402082642+04ba60cfe598-1~exp1~20210402063359.71 MP4Box - GPAC version 1.1.0-DEV-rev1745-g9e1378706-master ``` ## Build command ``` ./configure --static-mp4box --prefix=`realpath ./install` --enable-sanitizer --cc=clang --cxx=clang++ ``` ## crash command ``` MP4Box -out /dev/null poc_file ``` ## Pocs [poc.zip](https://github.com/gpac/gpac/files/8039045/poc.zip) ## Crash output ``` ==35757==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x602000001270 at pc 0x0000007d0889 bp 0x7fff1400a3d0 sp 0x7fff1400a3c8 READ of size 4 at 0x602000001270 thread T0 #0 0x7d0888 in convert_compact_sample_groups /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:296:18 #1 0x7c9018 in gf_isom_parse_movie_boxes_internal /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:420:7 #2 0x7c9018 in gf_isom_parse_movie_boxes /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:812:6 #3 0x7ccf16 in gf_isom_open_file /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:932:19 #4 0x4e14c6 in mp4boxMain /programs/mp4box/builds/build13/applications/mp4box/main.c:5984:12 #5 0x7fce38d730b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x270b2) #6 0x41ea6d in _start (/programs/mp4box/builds/build13/bin/gcc/MP4Box+0x41ea6d) 0x602000001271 is located 0 bytes to the right of 1-byte region [0x602000001270,0x602000001271) allocated by thread T0 here: #0 0x499ccd in malloc (/programs/mp4box/builds/build13/bin/gcc/MP4Box+0x499ccd) #1 0x1310c8e in csgp_box_read /programs/mp4box/builds/build13/src/isomedia/box_code_base.c:12613:55 #2 0x7adfad in gf_isom_box_read /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1822:9 #3 0x7adfad in gf_isom_box_parse_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:264:14 #4 0x7afda3 in gf_isom_box_array_read_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1719:7 #5 0x12e1c6b in stbl_box_read /programs/mp4box/builds/build13/src/isomedia/box_code_base.c:5069:6 #6 0x7adfad in gf_isom_box_read /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1822:9 #7 0x7adfad in gf_isom_box_parse_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:264:14 #8 0x7afda3 in gf_isom_box_array_read_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1719:7 #9 0x12d7132 in minf_box_read /programs/mp4box/builds/build13/src/isomedia/box_code_base.c:3549:6 #10 0x7adfad in gf_isom_box_read /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1822:9 #11 0x7adfad in gf_isom_box_parse_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:264:14 #12 0x7afda3 in gf_isom_box_array_read_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1719:7 #13 0x12d4665 in mdia_box_read /programs/mp4box/builds/build13/src/isomedia/box_code_base.c:3100:6 #14 0x7adfad in gf_isom_box_read /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1822:9 #15 0x7adfad in gf_isom_box_parse_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:264:14 #16 0x7afda3 in gf_isom_box_array_read_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1719:7 #17 0x12ec620 in trak_box_read /programs/mp4box/builds/build13/src/isomedia/box_code_base.c:6844:6 #18 0x7adfad in gf_isom_box_read /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1822:9 #19 0x7adfad in gf_isom_box_parse_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:264:14 #20 0x7afda3 in gf_isom_box_array_read_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1719:7 #21 0x7adfad in gf_isom_box_read /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:1822:9 #22 0x7adfad in gf_isom_box_parse_ex /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:264:14 #23 0x7ad181 in gf_isom_parse_root_box /programs/mp4box/builds/build13/src/isomedia/box_funcs.c:38:8 #24 0x7c8b31 in gf_isom_parse_movie_boxes_internal /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:349:7 #25 0x7c8b31 in gf_isom_parse_movie_boxes /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:812:6 #26 0x7ccf16 in gf_isom_open_file /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:932:19 #27 0x4e14c6 in mp4boxMain /programs/mp4box/builds/build13/applications/mp4box/main.c:5984:12 #28 0x7fce38d730b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x270b2) SUMMARY: AddressSanitizer: heap-buffer-overflow /programs/mp4box/builds/build13/src/isomedia/isom_intern.c:296:18 in convert_compact_sample_groups Shadow bytes around the buggy address: 0x0c047fff81f0: fa fa 00 00 fa fa 00 00 fa fa 00 00 fa fa 00 00 0x0c047fff8200: fa fa 00 05 fa fa 00 00 fa fa 00 00 fa fa 00 00 0x0c047fff8210: fa fa 00 00 fa fa 00 00 fa fa 00 00 fa fa 01 fa 0x0c047fff8220: fa fa fd fa fa fa fd fd fa fa 00 00 fa fa 00 00 0x0c047fff8230: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa 01 fa =>0x0c047fff8240: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa[01]fa 0x0c047fff8250: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa 01 fa 0x0c047fff8260: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa 01 fa 0x0c047fff8270: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa 01 fa 0x0c047fff8280: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa 00 04 0x0c047fff8290: fa fa 01 fa fa fa 01 fa fa fa 01 fa fa fa 01 fa Shadow byte legend (one shadow byte represents 8 application bytes): Addressable: 00 Partially addressable: 01 02 03 04 05 06 07 Heap left redzone: fa Freed heap region: fd Stack left redzone: f1 Stack mid redzone: f2 Stack right redzone: f3 Stack after return: f5 Stack use after scope: f8 Global redzone: f9 Global init order: f6 Poisoned by user: f7 Container overflow: fc Array cookie: ac Intra object redzone: bb ASan internal: fe Left alloca redzone: ca Right alloca redzone: cb Shadow gap: cc ==35757==ABORTING ``` --- end body ---
6,410
[ -0.040030479431152344, 0.017612842842936516, -0.014009245671331882, 0.012317617423832417, 0.02654009312391281, 0.013063924387097359, -0.02791898511350155, 0.0395471565425396, -0.007320914883166552, 0.023099973797798157, -0.021351484581828117, 0.007448853459209204, 0.023753881454467773, 0.0...
null
null
null
null
null
null
null
null
null
[ "LibreDWG", "libredwg" ]
for r11 handles esp.
fix read_RLL_BE
https://api.github.com/repos/LibreDWG/libredwg/issues/883/comments
0
2023-11-29T10:46:22
2023-11-29T11:43:53Z
https://github.com/LibreDWG/libredwg/issues/883
2,016,305,936
883
false
This is a GitHub Issue repo:libredwg owner:LibreDWG Title : fix read_RLL_BE Issue date: --- start body --- for r11 handles esp. --- end body ---
146
[ -0.03469650819897652, 0.014066792093217373, -0.005507319234311581, 0.02073250710964203, 0.024717703461647034, 0.04193944111466408, -0.0019382364116609097, 0.06053702160716057, 0.005080333910882473, 0.011923959478735924, -0.017443139106035233, 0.00271610077470541, 0.012999329715967178, -0.0...
null
null
null
null
null
null
null
null
null
[ "LibreDWG", "libredwg" ]
Commit d6029e974 with changes in dwg_add_Document() is bad. Decoding process is: * read numheader_vars from DWG file * dwg_add_Document (which after this commit set this variable to different values) * usage of numheader_vars As I understand, we need to move this setting to dwg_new_Document()
Issue with reading of numheader_vars
https://api.github.com/repos/LibreDWG/libredwg/issues/540/comments
5
2022-12-18T07:31:09
2022-12-22T17:46:22Z
https://github.com/LibreDWG/libredwg/issues/540
1,501,797,494
540
false
This is a GitHub Issue repo:libredwg owner:LibreDWG Title : Issue with reading of numheader_vars Issue date: --- start body --- Commit d6029e974 with changes in dwg_add_Document() is bad. Decoding process is: * read numheader_vars from DWG file * dwg_add_Document (which after this commit set this variable to different values) * usage of numheader_vars As I understand, we need to move this setting to dwg_new_Document() --- end body ---
449
[ -0.006261738948523998, 0.0023595786187797785, -0.008585133589804173, 0.01745973899960518, 0.031415343284606934, 0.05432934686541557, 0.00573230953887105, 0.05256204307079315, -0.008158543147146702, 0.020354460924863815, 0.02558019384741783, 0.0019158483482897282, 0.005766589194536209, -0.0...
CVE-2023-40781
2023-08-28T22:15:09.157000
Buffer Overflow vulnerability in Libming Libming v.0.4.8 allows a remote attacker to cause a denial of service via a crafted .swf file to the makeswf function.
{ "cvssMetricV2": null, "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "LOW", "attackVector": "NETWORK", "availabilityImpact": "HIGH", "baseScore": 6.5, "baseSeverity": "MEDIUM", "confidentialityImpact": "NONE", "integrityImpact": "NONE", "privilegesRequired": "NONE", "scope": "UNCHANGED", "userInteraction": "REQUIRED", "vectorString": "CVSS:3.1/AV:N/AC:L/PR:N/UI:R/S:U/C:N/I:N/A:H", "version": "3.1" }, "exploitabilityScore": 2.8, "impactScore": 3.6, "source": "nvd@nist.gov", "type": "Primary" } ] }
[ { "source": "cve@mitre.org", "tags": [ "Issue Tracking", "Patch" ], "url": "https://github.com/libming/libming/issues/288" } ]
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:libming:libming:0.4.8:*:*:*:*:*:*:*", "matchCriteriaId": "DD92BC79-2548-4C6F-9BDD-26C12BDF68AC", "versionEndExcluding": null, "versionEndIncluding": null, "versionStartExclud...
https://github.com/libming/libming/issues/288
[ "Issue Tracking", "Patch" ]
github.com
[ "libming", "libming" ]
A heap buffer overflow occurs when makeswf parse a invalid swf file, and the filename extension is `.swf`. ## Test Environment Ubuntu 20.04, 64 bit libming (master 04aee52) ## Steps to reproduce 1. compile libming with ASAN ``` $ CC="clang -fsanitize=address,fuzzer-no-link -g" CFLAGS+=" -fcommon" ./configure $ make ``` 2. Download the poc file from [here](https://raw.githubusercontent.com/Marsman1996/pocs/master/libming/poc-makeswf-04aee52-r_readc-HBO.swf) and run cmd `$ makeswf $POC` ## ASAN report ``` $ ./bin_asan/bin/makeswf ./poc-makeswf-04aee52-r_readc-HBO.swf Output file name: out.swf Output compression level: 9 Output SWF version: 6 ================================================================= ==5625==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x60800000013f at pc 0x0000004f15b5 bp 0x7fff376560d0 sp 0x7fff376560c8 WRITE of size 1 at 0x60800000013f thread T0 #0 0x4f15b4 in r_readc /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:264:34 #1 0x4f1a37 in getbits /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:143:18 #2 0x4f1656 in rect /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:169:9 #3 0x4efe15 in openswf /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:303:2 #4 0x4eedbe in newSWFPrebuiltClip_fromInput /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:1302:8 #5 0x4cbea3 in embed_swf /opt/disk/marsman/libming/04aee52/build_asan/util/makeswf.c:699:14 #6 0x4ca4d9 in main /opt/disk/marsman/libming/04aee52/build_asan/util/makeswf.c:401:4 #7 0x7f0aa6b3d83f in __libc_start_main /build/glibc-S7Ft5T/glibc-2.23/csu/../csu/libc-start.c:291 #8 0x41c5a8 in _start (/opt/disk/marsman/libming/04aee52/bin_asan/bin/makeswf+0x41c5a8) 0x60800000013f is located 199 bytes to the right of 88-byte region [0x608000000020,0x608000000078) allocated by thread T0 here: #0 0x4975fd in malloc /local/mnt/workspace/bcain_clang_vm-bcain-aus_3184/final/llvm-project/compiler-rt/lib/asan/asan_malloc_linux.cpp:145:3 #1 0x4ef8d8 in openswf /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:271:41 #2 0x4eedbe in newSWFPrebuiltClip_fromInput /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:1302:8 #3 0x4cbea3 in embed_swf /opt/disk/marsman/libming/04aee52/build_asan/util/makeswf.c:699:14 #4 0x4ca4d9 in main /opt/disk/marsman/libming/04aee52/build_asan/util/makeswf.c:401:4 #5 0x7f0aa6b3d83f in __libc_start_main /build/glibc-S7Ft5T/glibc-2.23/csu/../csu/libc-start.c:291 SUMMARY: AddressSanitizer: heap-buffer-overflow /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:264:34 in r_readc Shadow bytes around the buggy address: 0x0c107fff7fd0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c107fff7fe0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c107fff7ff0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c107fff8000: fa fa fa fa 00 00 00 00 00 00 00 00 00 00 00 fa 0x0c107fff8010: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa =>0x0c107fff8020: fa fa fa fa fa fa fa[fa]fa fa fa fa fa fa fa fa 0x0c107fff8030: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c107fff8040: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c107fff8050: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c107fff8060: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c107fff8070: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa Shadow byte legend (one shadow byte represents 8 application bytes): Addressable: 00 Partially addressable: 01 02 03 04 05 06 07 Heap left redzone: fa Freed heap region: fd Stack left redzone: f1 Stack mid redzone: f2 Stack right redzone: f3 Stack after return: f5 Stack use after scope: f8 Global redzone: f9 Global init order: f6 Poisoned by user: f7 Container overflow: fc Array cookie: ac Intra object redzone: bb ASan internal: fe Left alloca redzone: ca Right alloca redzone: cb Shadow gap: cc ==5625==ABORTING ```
heap-buffer-overflow in r_readc() at fromswf.c:264
https://api.github.com/repos/libming/libming/issues/288/comments
0
2023-08-09T02:43:36
2023-08-09T02:43:36Z
https://github.com/libming/libming/issues/288
1,842,363,544
288
true
This is a GitHub Issue repo:libming owner:libming Title : heap-buffer-overflow in r_readc() at fromswf.c:264 Issue date: --- start body --- A heap buffer overflow occurs when makeswf parse a invalid swf file, and the filename extension is `.swf`. ## Test Environment Ubuntu 20.04, 64 bit libming (master 04aee52) ## Steps to reproduce 1. compile libming with ASAN ``` $ CC="clang -fsanitize=address,fuzzer-no-link -g" CFLAGS+=" -fcommon" ./configure $ make ``` 2. Download the poc file from [here](https://raw.githubusercontent.com/Marsman1996/pocs/master/libming/poc-makeswf-04aee52-r_readc-HBO.swf) and run cmd `$ makeswf $POC` ## ASAN report ``` $ ./bin_asan/bin/makeswf ./poc-makeswf-04aee52-r_readc-HBO.swf Output file name: out.swf Output compression level: 9 Output SWF version: 6 ================================================================= ==5625==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x60800000013f at pc 0x0000004f15b5 bp 0x7fff376560d0 sp 0x7fff376560c8 WRITE of size 1 at 0x60800000013f thread T0 #0 0x4f15b4 in r_readc /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:264:34 #1 0x4f1a37 in getbits /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:143:18 #2 0x4f1656 in rect /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:169:9 #3 0x4efe15 in openswf /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:303:2 #4 0x4eedbe in newSWFPrebuiltClip_fromInput /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:1302:8 #5 0x4cbea3 in embed_swf /opt/disk/marsman/libming/04aee52/build_asan/util/makeswf.c:699:14 #6 0x4ca4d9 in main /opt/disk/marsman/libming/04aee52/build_asan/util/makeswf.c:401:4 #7 0x7f0aa6b3d83f in __libc_start_main /build/glibc-S7Ft5T/glibc-2.23/csu/../csu/libc-start.c:291 #8 0x41c5a8 in _start (/opt/disk/marsman/libming/04aee52/bin_asan/bin/makeswf+0x41c5a8) 0x60800000013f is located 199 bytes to the right of 88-byte region [0x608000000020,0x608000000078) allocated by thread T0 here: #0 0x4975fd in malloc /local/mnt/workspace/bcain_clang_vm-bcain-aus_3184/final/llvm-project/compiler-rt/lib/asan/asan_malloc_linux.cpp:145:3 #1 0x4ef8d8 in openswf /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:271:41 #2 0x4eedbe in newSWFPrebuiltClip_fromInput /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:1302:8 #3 0x4cbea3 in embed_swf /opt/disk/marsman/libming/04aee52/build_asan/util/makeswf.c:699:14 #4 0x4ca4d9 in main /opt/disk/marsman/libming/04aee52/build_asan/util/makeswf.c:401:4 #5 0x7f0aa6b3d83f in __libc_start_main /build/glibc-S7Ft5T/glibc-2.23/csu/../csu/libc-start.c:291 SUMMARY: AddressSanitizer: heap-buffer-overflow /opt/disk/marsman/libming/04aee52/build_asan/src/blocks/fromswf.c:264:34 in r_readc Shadow bytes around the buggy address: 0x0c107fff7fd0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c107fff7fe0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c107fff7ff0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 0x0c107fff8000: fa fa fa fa 00 00 00 00 00 00 00 00 00 00 00 fa 0x0c107fff8010: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa =>0x0c107fff8020: fa fa fa fa fa fa fa[fa]fa fa fa fa fa fa fa fa 0x0c107fff8030: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c107fff8040: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c107fff8050: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c107fff8060: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c107fff8070: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa Shadow byte legend (one shadow byte represents 8 application bytes): Addressable: 00 Partially addressable: 01 02 03 04 05 06 07 Heap left redzone: fa Freed heap region: fd Stack left redzone: f1 Stack mid redzone: f2 Stack right redzone: f3 Stack after return: f5 Stack use after scope: f8 Global redzone: f9 Global init order: f6 Poisoned by user: f7 Container overflow: fc Array cookie: ac Intra object redzone: bb ASan internal: fe Left alloca redzone: ca Right alloca redzone: cb Shadow gap: cc ==5625==ABORTING ``` --- end body ---
4,389
[ -0.030641967430710793, 0.012280993163585663, -0.010992378927767277, 0.019421767443418503, 0.05496189370751381, 0.02322353608906269, -0.025715332478284836, 0.029360473155975342, -0.02169998176395893, 0.028719725087285042, -0.022767893970012665, 0.009461704641580582, 0.0006581010529771447, 0...
null
null
null
null
null
null
null
null
null
[ "hwchase17", "langchain" ]
### Checked other resources - [X] I added a very descriptive title to this issue. - [X] I searched the LangChain documentation with the integrated search. - [X] I used the GitHub search to find a similar question and didn't find it. - [X] I am sure that this is a bug in LangChain rather than my code. - [X] The bug is not resolved by updating to the latest stable version of LangChain (or the specific integration package). ### Example Code from langchain_community.vectorstores import Chroma vectorstore = Chroma.from_documents(docs, embeddings) retriever = SelfQueryRetriever.from_llm( llm, vectorstore, document_content_description, metadata_field_info, ) ### Error Message and Stack Trace (if applicable) _No response_ ### Description I'm trying to use chroma as a vectorstore but langchain says its not supported ### System Info langchain=0.1.20 python=3.11
ValueError: Self query retriever with Vector Store type <class 'langchain_chroma.vectorstores.Chroma'> not supported.
https://api.github.com/repos/langchain-ai/langchain/issues/22272/comments
2
2024-05-29T11:59:08
2024-06-02T12:11:43Z
https://github.com/langchain-ai/langchain/issues/22272
2,323,141,447
22,272
false
This is a GitHub Issue repo:langchain owner:hwchase17 Title : ValueError: Self query retriever with Vector Store type <class 'langchain_chroma.vectorstores.Chroma'> not supported. Issue date: --- start body --- ### Checked other resources - [X] I added a very descriptive title to this issue. - [X] I searched the LangChain documentation with the integrated search. - [X] I used the GitHub search to find a similar question and didn't find it. - [X] I am sure that this is a bug in LangChain rather than my code. - [X] The bug is not resolved by updating to the latest stable version of LangChain (or the specific integration package). ### Example Code from langchain_community.vectorstores import Chroma vectorstore = Chroma.from_documents(docs, embeddings) retriever = SelfQueryRetriever.from_llm( llm, vectorstore, document_content_description, metadata_field_info, ) ### Error Message and Stack Trace (if applicable) _No response_ ### Description I'm trying to use chroma as a vectorstore but langchain says its not supported ### System Info langchain=0.1.20 python=3.11 --- end body ---
1,149
[ -0.020083216950297356, 0.023361263796687126, -0.01878306455910206, 0.02764900214970112, 0.0065318867564201355, -0.013762259855866432, 0.03125900402665138, 0.020885439589619637, -0.005183323752135038, 0.022130267694592476, 0.0011895018396899104, 0.040719691663980484, -0.04683317989110947, 0...
null
null
null
null
null
null
null
null
null
[ "openlink", "virtuoso-opensource" ]
The PoC is generated by my DBMS fuzzer. It can also be reproduced in the beta docker image. ```SQL CREATE TABLE x ( x FLOAT UNIQUE CHECK ( ( CASE WHEN 1.000000 = 1 THEN 1 ELSE ( 1.000000 , 1.000000 ) / 1 / 1 END ) AND ( SELECT CASE WHEN x < x AND x IS NULL THEN 1 ELSE ( SELECT x FROM x WHERE ( CASE WHEN x = x AND x = x AND x THEN 1.000000 ELSE x + x * 1 + 1 END ) IN ( SELECT DISTINCT 1 FROM x GROUP BY NULL , 'x' , 'x' , 'x' ) ORDER BY x / 1 DESC ) - CASE WHEN ( 1 ) AND x NOT IN ( SELECT DISTINCT x / 1 , 1 FROM x GROUP BY CASE WHEN x = 1 THEN 'x' ELSE x - x * 1.000000 + 1 END ) THEN x + 1 WHEN 1 THEN 1 ELSE 1.000000 END / 1 + 1 END ) < x + x + x ) PRIMARY KEY ) ; ``` backtrace: ```c #0 0x792e96 (sqlc_add_distinct_node+0xe6) #1 0x747a3d (sqlg_handle_select_list+0x88d) #2 0x7470e5 (sqlg_select_node+0x1f5) #3 0x749c99 (sqlg_dt_query_1+0x1779) #4 0x738ee6 (sqlg_dt_subquery+0x96) #5 0x71e57c (cmp_subq_call+0x8c) #6 0x71bd94 (pred_gen_1+0x3e4) #7 0x71b035 (sqlc_searched_case+0x3d5) #8 0x716ada (scalar_exp_generate+0x9fa) #9 0x749a8a (sqlg_dt_query_1+0x156a) #10 0x74a9e7 (sqlg_top_1+0x107) #11 0x70ffb4 (sqlo_top_select+0x164) #12 0x6b9b6f (sql_stmt_comp+0x8bf) #13 0x6be876 (sqlc_subquery_1+0x456) #14 0xd5922c (sqlc_mark_pred_deps+0x26c) #15 0xd5922c (sqlc_mark_pred_deps+0x26c) #16 0x7a7108 (sqlc_if_stmt+0xa8) #17 0x7a5a46 (sqlc_proc_stmt+0x1a6) #18 0x7a6ca2 (sqlc_compound_stmt+0x292) #19 0x7a5922 (sqlc_proc_stmt+0x82) #20 0x7aadc2 (sqlc_trigger_decl+0x2b2) #21 0x6b960e (sql_stmt_comp+0x35e) #22 0x6bc9d2 (sql_compile_1+0x1a62) #23 0x4e29aa (ddl_table_check_constraints_define_triggers+0xa1a) #24 0x4e462d (ddl_table_constraints+0xd0d) #25 0x4e5331 (sql_ddl_node_input_1+0xbc1) #26 0x4e57ee (sql_ddl_node_input+0x10e) #27 0x7bcb0b (ddl_node_input_1+0x19b) #28 0x7bd1e7 (qn_without_ac_at+0xc7) #29 0x7af05e (qn_input+0x3ce) #30 0x7c1be9 (qr_dml_array_exec+0x839) #31 0x7ce602 (sf_sql_execute+0x15d2) #32 0x7cecde (sf_sql_execute_w+0x17e) #33 0x7d799d (sf_sql_execute_wrapper+0x3d) #34 0xe214bc (future_wrapper+0x3fc) #35 0xe28dbe (_thread_boot+0x11e) #36 0x7f4cd491f609 (start_thread+0xd9) #37 0x7f4cd46ef133 (clone+0x43) ``` ways to reproduce (write poc to the file /tmp/test.sql first): ```bash # remove the old one docker container rm virtdb_test -f # start virtuoso through docker docker run --name virtdb_test -itd --env DBA_PASSWORD=dba pkleef/virtuoso-opensource-7 # wait the server starting sleep 10 # check whether the simple query works echo "SELECT 1;" | docker exec -i virtdb_test isql 1111 dba # run the poc cat /tmp/test.sql | docker exec -i virtdb_test isql 1111 dba ```
Fuzzer: Virtuoso 7.2.11 crashed at `sqlc_add_distinct_node`
https://api.github.com/repos/openlink/virtuoso-opensource/issues/1210/comments
0
2023-12-12T16:57:00
2023-12-15T14:14:11Z
https://github.com/openlink/virtuoso-opensource/issues/1210
2,038,192,181
1,210
false
This is a GitHub Issue repo:virtuoso-opensource owner:openlink Title : Fuzzer: Virtuoso 7.2.11 crashed at `sqlc_add_distinct_node` Issue date: --- start body --- The PoC is generated by my DBMS fuzzer. It can also be reproduced in the beta docker image. ```SQL CREATE TABLE x ( x FLOAT UNIQUE CHECK ( ( CASE WHEN 1.000000 = 1 THEN 1 ELSE ( 1.000000 , 1.000000 ) / 1 / 1 END ) AND ( SELECT CASE WHEN x < x AND x IS NULL THEN 1 ELSE ( SELECT x FROM x WHERE ( CASE WHEN x = x AND x = x AND x THEN 1.000000 ELSE x + x * 1 + 1 END ) IN ( SELECT DISTINCT 1 FROM x GROUP BY NULL , 'x' , 'x' , 'x' ) ORDER BY x / 1 DESC ) - CASE WHEN ( 1 ) AND x NOT IN ( SELECT DISTINCT x / 1 , 1 FROM x GROUP BY CASE WHEN x = 1 THEN 'x' ELSE x - x * 1.000000 + 1 END ) THEN x + 1 WHEN 1 THEN 1 ELSE 1.000000 END / 1 + 1 END ) < x + x + x ) PRIMARY KEY ) ; ``` backtrace: ```c #0 0x792e96 (sqlc_add_distinct_node+0xe6) #1 0x747a3d (sqlg_handle_select_list+0x88d) #2 0x7470e5 (sqlg_select_node+0x1f5) #3 0x749c99 (sqlg_dt_query_1+0x1779) #4 0x738ee6 (sqlg_dt_subquery+0x96) #5 0x71e57c (cmp_subq_call+0x8c) #6 0x71bd94 (pred_gen_1+0x3e4) #7 0x71b035 (sqlc_searched_case+0x3d5) #8 0x716ada (scalar_exp_generate+0x9fa) #9 0x749a8a (sqlg_dt_query_1+0x156a) #10 0x74a9e7 (sqlg_top_1+0x107) #11 0x70ffb4 (sqlo_top_select+0x164) #12 0x6b9b6f (sql_stmt_comp+0x8bf) #13 0x6be876 (sqlc_subquery_1+0x456) #14 0xd5922c (sqlc_mark_pred_deps+0x26c) #15 0xd5922c (sqlc_mark_pred_deps+0x26c) #16 0x7a7108 (sqlc_if_stmt+0xa8) #17 0x7a5a46 (sqlc_proc_stmt+0x1a6) #18 0x7a6ca2 (sqlc_compound_stmt+0x292) #19 0x7a5922 (sqlc_proc_stmt+0x82) #20 0x7aadc2 (sqlc_trigger_decl+0x2b2) #21 0x6b960e (sql_stmt_comp+0x35e) #22 0x6bc9d2 (sql_compile_1+0x1a62) #23 0x4e29aa (ddl_table_check_constraints_define_triggers+0xa1a) #24 0x4e462d (ddl_table_constraints+0xd0d) #25 0x4e5331 (sql_ddl_node_input_1+0xbc1) #26 0x4e57ee (sql_ddl_node_input+0x10e) #27 0x7bcb0b (ddl_node_input_1+0x19b) #28 0x7bd1e7 (qn_without_ac_at+0xc7) #29 0x7af05e (qn_input+0x3ce) #30 0x7c1be9 (qr_dml_array_exec+0x839) #31 0x7ce602 (sf_sql_execute+0x15d2) #32 0x7cecde (sf_sql_execute_w+0x17e) #33 0x7d799d (sf_sql_execute_wrapper+0x3d) #34 0xe214bc (future_wrapper+0x3fc) #35 0xe28dbe (_thread_boot+0x11e) #36 0x7f4cd491f609 (start_thread+0xd9) #37 0x7f4cd46ef133 (clone+0x43) ``` ways to reproduce (write poc to the file /tmp/test.sql first): ```bash # remove the old one docker container rm virtdb_test -f # start virtuoso through docker docker run --name virtdb_test -itd --env DBA_PASSWORD=dba pkleef/virtuoso-opensource-7 # wait the server starting sleep 10 # check whether the simple query works echo "SELECT 1;" | docker exec -i virtdb_test isql 1111 dba # run the poc cat /tmp/test.sql | docker exec -i virtdb_test isql 1111 dba ``` --- end body ---
2,843
[ -0.006176535040140152, 0.006521807052195072, -0.011648574844002724, -0.011488146148622036, 0.04924487695097923, 0.0200188085436821, -0.02222296968102455, 0.023924918845295906, -0.023059993982315063, 0.024008620530366898, -0.007065872196108103, -0.00309873023070395, 0.005018652882426977, -0...
CVE-2023-47465
2023-12-09T06:15:45.530000
An issue in GPAC v.2.2.1 and before allows a local attacker to cause a denial of service (DoS) via the ctts_box_read function of file src/isomedia/box_code_base.c.
{ "cvssMetricV2": null, "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "LOW", "attackVector": "LOCAL", "availabilityImpact": "HIGH", "baseScore": 5.5, "baseSeverity": "MEDIUM", "confidentialityImpact": "NONE", "integrityImpact": "NONE", "privilegesRequired": "LOW", "scope": "UNCHANGED", "userInteraction": "NONE", "vectorString": "CVSS:3.1/AV:L/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H", "version": "3.1" }, "exploitabilityScore": 1.8, "impactScore": 3.6, "source": "nvd@nist.gov", "type": "Primary" } ] }
[ { "source": "cve@mitre.org", "tags": [ "Issue Tracking", "Patch" ], "url": "https://github.com/gpac/gpac/issues/2652" } ]
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:gpac:gpac:*:*:*:*:*:*:*:*", "matchCriteriaId": "047BC15F-5E51-48D9-B751-9DC9311FEBCF", "versionEndExcluding": null, "versionEndIncluding": "2.2.1", "versionStartExcluding": n...
https://github.com/gpac/gpac/issues/2652
[ "Issue Tracking", "Patch" ]
github.com
[ "gpac", "gpac" ]
[Y] I looked for a similar issue and couldn't find any. [Y] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ [Y] I give enough information for contributors to reproduce my issue Description There is a integer overflow issue in isomedia/box_code_base.c:413 System info Ubuntu 22.04.2 LTS GPAC-2.2.1 Build command ./configure --enable-sanitizer && make crash command /usr/local/bin/MP4Box -saf -diod poc poc_file: [poc.zip](https://github.com/gpac/gpac/files/13040344/poc.zip) Crash output: isomedia/box_code_base.c:413:29: runtime error: negation of -2147483648 cannot be represented in type 'int'; cast to an unsigned type to negate this value to itself
Integer overflow issue in isomedia/box_code_base.c:413
https://api.github.com/repos/gpac/gpac/issues/2652/comments
1
2023-10-19T09:06:47
2023-12-08T06:37:55Z
https://github.com/gpac/gpac/issues/2652
1,951,673,630
2,652
true
This is a GitHub Issue repo:gpac owner:gpac Title : Integer overflow issue in isomedia/box_code_base.c:413 Issue date: --- start body --- [Y] I looked for a similar issue and couldn't find any. [Y] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ [Y] I give enough information for contributors to reproduce my issue Description There is a integer overflow issue in isomedia/box_code_base.c:413 System info Ubuntu 22.04.2 LTS GPAC-2.2.1 Build command ./configure --enable-sanitizer && make crash command /usr/local/bin/MP4Box -saf -diod poc poc_file: [poc.zip](https://github.com/gpac/gpac/files/13040344/poc.zip) Crash output: isomedia/box_code_base.c:413:29: runtime error: negation of -2147483648 cannot be represented in type 'int'; cast to an unsigned type to negate this value to itself --- end body ---
901
[ -0.014554640278220177, 0.02049281820654869, -0.012587214820086956, 0.00781943928450346, 0.02586374618113041, 0.009801225736737251, -0.025203151628375053, 0.05422626808285713, -0.029339052736759186, 0.024427669122815132, -0.014820314943790436, -0.0014764666557312012, 0.03188091143965721, 0....
null
null
null
null
null
null
null
null
null
[ "MonetDB", "MonetDB" ]
**Describe the bug** MonetDB Dec2023-SP2 crashes with the `PASSWORD_HASH` function. **To Reproduce** ```sql SELECT PASSWORD_HASH('1'); ``` **Backtrace of the Crash** ``` #0 0x00007fe06dbab915 in ?? () from target:/lib/x86_64-linux-gnu/libc.so.6 #1 0x00007fe06e24a2e0 in strPut () from target:/usr/local/lib/libbat.so.28 #2 0x00007fe06e2be3e9 in BUNappendmulti () from target:/usr/local/lib/libbat.so.28 #3 0x00007fe06e2bf9af in BUNappend () from target:/usr/local/lib/libbat.so.28 #4 0x00007fe06d0fa8e1 in res_col_create () from target:/usr/local/lib/libmonetdbsql.so.14 #5 0x00007fe06d0351ff in mvc_result_value () from target:/usr/local/lib/libmonetdbsql.so.14 #6 0x00007fe06cfd5a15 in mvc_scalar_value_wrap () from target:/usr/local/lib/libmonetdbsql.so.14 #7 0x00007fe06ef0a1b8 in runMALsequence () from target:/usr/local/lib/libmonetdb5.so.35 #8 0x00007fe06ef0c3b9 in runMAL () from target:/usr/local/lib/libmonetdb5.so.35 #9 0x00007fe06cfe6b67 in SQLrun () from target:/usr/local/lib/libmonetdbsql.so.14 #10 0x00007fe06cfe7d7c in SQLengineIntern () from target:/usr/local/lib/libmonetdbsql.so.14 #11 0x00007fe06cfe631a in SQLengine_ () from target:/usr/local/lib/libmonetdbsql.so.14 #12 0x00007fe06cfe691b in SQLengine () from target:/usr/local/lib/libmonetdbsql.so.14 #13 0x00007fe06ef21826 in runScenario () from target:/usr/local/lib/libmonetdb5.so.35 #14 0x00007fe06ef22140 in MSscheduleClient () from target:/usr/local/lib/libmonetdb5.so.35 #15 0x00007fe06efd1922 in doChallenge () from target:/usr/local/lib/libmonetdb5.so.35 #16 0x00007fe06e2cdeed in thread_starter () from target:/usr/local/lib/libbat.so.28 #17 0x00007fe06dc1d609 in start_thread () from target:/lib/x86_64-linux-gnu/libpthread.so.0 #18 0x00007fe06db42353 in clone () from target:/lib/x86_64-linux-gnu/libc.so.6 ``` **Software versions** - MonetDB version number: v11.49.7 (Dec2023-SP2) - OS and version: Ubuntu 20.04 - From docker image monetdb/monetdb:Dec2023-SP2 (or monetdb/dev-builds:default)
MonetDB Dec2023-SP2 crashes with the `PASSWORD_HASH` function
https://api.github.com/repos/MonetDB/MonetDB/issues/7511/comments
0
2024-05-09T20:24:55
2024-07-09T08:42:30Z
https://github.com/MonetDB/MonetDB/issues/7511
2,288,403,488
7,511
false
This is a GitHub Issue repo:MonetDB owner:MonetDB Title : MonetDB Dec2023-SP2 crashes with the `PASSWORD_HASH` function Issue date: --- start body --- **Describe the bug** MonetDB Dec2023-SP2 crashes with the `PASSWORD_HASH` function. **To Reproduce** ```sql SELECT PASSWORD_HASH('1'); ``` **Backtrace of the Crash** ``` #0 0x00007fe06dbab915 in ?? () from target:/lib/x86_64-linux-gnu/libc.so.6 #1 0x00007fe06e24a2e0 in strPut () from target:/usr/local/lib/libbat.so.28 #2 0x00007fe06e2be3e9 in BUNappendmulti () from target:/usr/local/lib/libbat.so.28 #3 0x00007fe06e2bf9af in BUNappend () from target:/usr/local/lib/libbat.so.28 #4 0x00007fe06d0fa8e1 in res_col_create () from target:/usr/local/lib/libmonetdbsql.so.14 #5 0x00007fe06d0351ff in mvc_result_value () from target:/usr/local/lib/libmonetdbsql.so.14 #6 0x00007fe06cfd5a15 in mvc_scalar_value_wrap () from target:/usr/local/lib/libmonetdbsql.so.14 #7 0x00007fe06ef0a1b8 in runMALsequence () from target:/usr/local/lib/libmonetdb5.so.35 #8 0x00007fe06ef0c3b9 in runMAL () from target:/usr/local/lib/libmonetdb5.so.35 #9 0x00007fe06cfe6b67 in SQLrun () from target:/usr/local/lib/libmonetdbsql.so.14 #10 0x00007fe06cfe7d7c in SQLengineIntern () from target:/usr/local/lib/libmonetdbsql.so.14 #11 0x00007fe06cfe631a in SQLengine_ () from target:/usr/local/lib/libmonetdbsql.so.14 #12 0x00007fe06cfe691b in SQLengine () from target:/usr/local/lib/libmonetdbsql.so.14 #13 0x00007fe06ef21826 in runScenario () from target:/usr/local/lib/libmonetdb5.so.35 #14 0x00007fe06ef22140 in MSscheduleClient () from target:/usr/local/lib/libmonetdb5.so.35 #15 0x00007fe06efd1922 in doChallenge () from target:/usr/local/lib/libmonetdb5.so.35 #16 0x00007fe06e2cdeed in thread_starter () from target:/usr/local/lib/libbat.so.28 #17 0x00007fe06dc1d609 in start_thread () from target:/lib/x86_64-linux-gnu/libpthread.so.0 #18 0x00007fe06db42353 in clone () from target:/lib/x86_64-linux-gnu/libc.so.6 ``` **Software versions** - MonetDB version number: v11.49.7 (Dec2023-SP2) - OS and version: Ubuntu 20.04 - From docker image monetdb/monetdb:Dec2023-SP2 (or monetdb/dev-builds:default) --- end body ---
2,198
[ -0.0005203947075642645, -0.014129921793937683, -0.004552592057734728, 0.006703097373247147, 0.03881937637925148, 0.02059522271156311, -0.01863771118223667, 0.03518005833029747, -0.024799736216664314, 0.01805873028934002, -0.005731234326958656, -0.004931687377393246, 0.006051742006093264, 0...
CVE-2024-22749
2024-01-25T16:15:09.010000
GPAC v2.3 was detected to contain a buffer overflow via the function gf_isom_new_generic_sample_description function in the isomedia/isom_write.c:4577
{ "cvssMetricV2": null, "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "LOW", "attackVector": "LOCAL", "availabilityImpact": "HIGH", "baseScore": 7.8, "baseSeverity": "HIGH", "confidentialityImpact": "HIGH", "integrityImpact": "HIGH", "privilegesRequired": "NONE", "scope": "UNCHANGED", "userInteraction": "REQUIRED", "vectorString": "CVSS:3.1/AV:L/AC:L/PR:N/UI:R/S:U/C:H/I:H/A:H", "version": "3.1" }, "exploitabilityScore": 1.8, "impactScore": 5.9, "source": "nvd@nist.gov", "type": "Primary" } ] }
[ { "source": "cve@mitre.org", "tags": [ "Exploit", "Issue Tracking", "Patch" ], "url": "https://github.com/gpac/gpac/issues/2713" }, { "source": "cve@mitre.org", "tags": [ "Exploit", "Third Party Advisory" ], "url": "https://github.com/hanxuer/crashes...
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:gpac:gpac:2.3:*:*:*:*:*:*:*", "matchCriteriaId": "58C43523-0CB3-474F-B88B-6ACA538645FF", "versionEndExcluding": null, "versionEndIncluding": null, "versionStartExcluding": nu...
https://github.com/gpac/gpac/issues/2713
[ "Exploit", "Issue Tracking", "Patch" ]
github.com
[ "gpac", "gpac" ]
Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at https://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: https://gpac.io/bug-reporting/ ### Description GPAC v2.3 8684dfb was detected to contain a buffer overflow via the function gf_isom_new_generic_sample_description function in the isomedia/isom_write.c:4577 ### The affect version ``` MP4Box - GPAC version 2.3-DEV-revrelease (c) 2000-2023 Telecom Paris distributed under LGPL v2.1+ - https://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: Features: GPAC_CONFIG_LINUX GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_JPEG GPAC_HAS_PNG GPAC_HAS_LINUX_DVB GPAC_DISABLE_3D ``` ### Test environment ``` $ uname -a Linux ubuntu 5.4.0-152-generic #169~18.04.1-Ubuntu SMP Wed Jun 7 22:22:24 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux ``` ### Compiler with asan ``` $ ./configure --enable-sanitizer $ make ``` ### Reproduce ./bin/gcc/MP4Box -dash 10000 ./poc poc link:https://github.com/hanxuer/crashes/raw/main/gapc/01/poc.zip ### Report GDB backtrace: *** buffer overflow detected ***: ../../gpac-asan/bin/gcc/MP4Box terminated #0 __GI_raise (sig=sig@entry=0x6) at ../sysdeps/unix/sysv/linux/raise.c:51 #1 0x00007ffff20a27f1 in __GI_abort () at abort.c:79 #2 0x00007ffff20eb837 in __libc_message (action=action@entry=(do_abort | do_backtrace), fmt=fmt@entry=0x7ffff2218869 "*** %s ***: %s terminated\n") at ../sysdeps/posix/libc_fatal.c:181 #3 0x00007ffff2196b5f in __GI___fortify_fail_abort (need_backtrace=need_backtrace@entry=0x1, msg=msg@entry=0x7ffff22187e6 "buffer overflow detected") at fortify_fail.c:33 #4 0x00007ffff2196b81 in __GI___fortify_fail (msg=msg@entry=0x7ffff22187e6 "buffer overflow detected") at fortify_fail.c:44 #5 0x00007ffff2194870 in __GI___chk_fail () at chk_fail.c:28 #6 0x00007ffff2193b02 in __strcpy_chk (dest=dest@entry=0x612000001426 "", src=src@entry=0x7ffffffe2f94 " MPEG-4 AVC|H264 Multiview Video ", destlen=destlen@entry=0x21) at strcpy_chk.c:30 #7 0x00007ffff4682517 in strcpy (__src=0x7ffffffe2f94 " MPEG-4 AVC|H264 Multiview Video ", __dest=0x612000001426 "") at /usr/include/x86_64-linux-gnu/bits/string_fortified.h:90 #8 gf_isom_new_generic_sample_description (movie=<optimized out>, trackNumber=<optimized out>, URLname=URLname@entry=0x0, URNname=URNname@entry=0x0, udesc=udesc@entry=0x7ffffffe2f60, outDescriptionIndex=outDescriptionIndex@entry=0x617000011564) at isomedia/isom_write.c:4577 #9 0x00007ffff51556ad in mp4_mux_setup_pid (filter=<optimized out>, pid=0x613000001000, is_true_pid=<optimized out>) at filters/mux_isom.c:3218 #10 0x00007ffff4da6a44 in gf_filter_pid_configure (filter=filter@entry=0x619000014a80, pid=<optimized out>, ctype=ctype@entry=GF_PID_CONF_CONNECT) at filter_core/filter_pid.c:881 #11 0x00007ffff4dadedf in gf_filter_pid_connect_task (task=0x607000000f70) at filter_core/filter_pid.c:1241 #12 0x00007ffff4de4ea1 in gf_fs_thread_proc (sess_thread=sess_thread@entry=0x616000001c10) at filter_core/filter_session.c:2100 #13 0x00007ffff4de8e94 in gf_fs_run (fsess=0x616000001b80) at filter_core/filter_session.c:2400 #14 0x00007ffff47dfe76 in gf_dasher_process (dasher=<optimized out>) at media_tools/dash_segmenter.c:1255 #15 0x00005555555d7622 in do_dash () at mp4box.c:4832 #16 0x00005555555f5bb7 in mp4box_main (argc=<optimized out>, argv=<optimized out>) at mp4box.c:6256 #17 0x00007ffff2083c87 in __libc_start_main (main=0x5555555db180 <main>, argc=0x4, argv=0x7fffffffdae8, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fffffffdad8) at ../csu/libc-start.c:310 #18 0x00005555555db23a in _start () ### Vul in source code: isom_write.c entry->Height = udesc->height; strcpy(entry->compressor_name, udesc->compressor_name); // this entry->color_table_index = -1; Thank you very much for your attention and consideration.
buffer-overflow in gf_isom_new_generic_sample_description function at isomedia/isom_write.c:4577
https://api.github.com/repos/gpac/gpac/issues/2713/comments
1
2024-01-06T06:07:11
2024-01-08T10:22:24Z
https://github.com/gpac/gpac/issues/2713
2,068,418,306
2,713
true
This is a GitHub Issue repo:gpac owner:gpac Title : buffer-overflow in gf_isom_new_generic_sample_description function at isomedia/isom_write.c:4577 Issue date: --- start body --- Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at https://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: https://gpac.io/bug-reporting/ ### Description GPAC v2.3 8684dfb was detected to contain a buffer overflow via the function gf_isom_new_generic_sample_description function in the isomedia/isom_write.c:4577 ### The affect version ``` MP4Box - GPAC version 2.3-DEV-revrelease (c) 2000-2023 Telecom Paris distributed under LGPL v2.1+ - https://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: Features: GPAC_CONFIG_LINUX GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_JPEG GPAC_HAS_PNG GPAC_HAS_LINUX_DVB GPAC_DISABLE_3D ``` ### Test environment ``` $ uname -a Linux ubuntu 5.4.0-152-generic #169~18.04.1-Ubuntu SMP Wed Jun 7 22:22:24 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux ``` ### Compiler with asan ``` $ ./configure --enable-sanitizer $ make ``` ### Reproduce ./bin/gcc/MP4Box -dash 10000 ./poc poc link:https://github.com/hanxuer/crashes/raw/main/gapc/01/poc.zip ### Report GDB backtrace: *** buffer overflow detected ***: ../../gpac-asan/bin/gcc/MP4Box terminated #0 __GI_raise (sig=sig@entry=0x6) at ../sysdeps/unix/sysv/linux/raise.c:51 #1 0x00007ffff20a27f1 in __GI_abort () at abort.c:79 #2 0x00007ffff20eb837 in __libc_message (action=action@entry=(do_abort | do_backtrace), fmt=fmt@entry=0x7ffff2218869 "*** %s ***: %s terminated\n") at ../sysdeps/posix/libc_fatal.c:181 #3 0x00007ffff2196b5f in __GI___fortify_fail_abort (need_backtrace=need_backtrace@entry=0x1, msg=msg@entry=0x7ffff22187e6 "buffer overflow detected") at fortify_fail.c:33 #4 0x00007ffff2196b81 in __GI___fortify_fail (msg=msg@entry=0x7ffff22187e6 "buffer overflow detected") at fortify_fail.c:44 #5 0x00007ffff2194870 in __GI___chk_fail () at chk_fail.c:28 #6 0x00007ffff2193b02 in __strcpy_chk (dest=dest@entry=0x612000001426 "", src=src@entry=0x7ffffffe2f94 " MPEG-4 AVC|H264 Multiview Video ", destlen=destlen@entry=0x21) at strcpy_chk.c:30 #7 0x00007ffff4682517 in strcpy (__src=0x7ffffffe2f94 " MPEG-4 AVC|H264 Multiview Video ", __dest=0x612000001426 "") at /usr/include/x86_64-linux-gnu/bits/string_fortified.h:90 #8 gf_isom_new_generic_sample_description (movie=<optimized out>, trackNumber=<optimized out>, URLname=URLname@entry=0x0, URNname=URNname@entry=0x0, udesc=udesc@entry=0x7ffffffe2f60, outDescriptionIndex=outDescriptionIndex@entry=0x617000011564) at isomedia/isom_write.c:4577 #9 0x00007ffff51556ad in mp4_mux_setup_pid (filter=<optimized out>, pid=0x613000001000, is_true_pid=<optimized out>) at filters/mux_isom.c:3218 #10 0x00007ffff4da6a44 in gf_filter_pid_configure (filter=filter@entry=0x619000014a80, pid=<optimized out>, ctype=ctype@entry=GF_PID_CONF_CONNECT) at filter_core/filter_pid.c:881 #11 0x00007ffff4dadedf in gf_filter_pid_connect_task (task=0x607000000f70) at filter_core/filter_pid.c:1241 #12 0x00007ffff4de4ea1 in gf_fs_thread_proc (sess_thread=sess_thread@entry=0x616000001c10) at filter_core/filter_session.c:2100 #13 0x00007ffff4de8e94 in gf_fs_run (fsess=0x616000001b80) at filter_core/filter_session.c:2400 #14 0x00007ffff47dfe76 in gf_dasher_process (dasher=<optimized out>) at media_tools/dash_segmenter.c:1255 #15 0x00005555555d7622 in do_dash () at mp4box.c:4832 #16 0x00005555555f5bb7 in mp4box_main (argc=<optimized out>, argv=<optimized out>) at mp4box.c:6256 #17 0x00007ffff2083c87 in __libc_start_main (main=0x5555555db180 <main>, argc=0x4, argv=0x7fffffffdae8, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fffffffdad8) at ../csu/libc-start.c:310 #18 0x00005555555db23a in _start () ### Vul in source code: isom_write.c entry->Height = udesc->height; strcpy(entry->compressor_name, udesc->compressor_name); // this entry->color_table_index = -1; Thank you very much for your attention and consideration. --- end body ---
4,772
[ -0.03547307476401329, 0.022700397297739983, -0.016210336238145828, 0.016728948801755905, 0.027842067182064056, 0.02247813530266285, -0.020714852958917618, 0.05046837776899338, -0.02477484568953514, 0.0290126483887434, -0.028908926993608475, 0.005567672662436962, 0.0366288386285305, 0.01607...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
MP4Box -dash 30000 -profile onDemand music_normal.mp4 -out music.mpd --dual play music_normal_dashinit.mp4 in windows media player when seek to the last part of the video. it is not playable and will jump back to the beggining . no problem on other player . @jeanlf in the last issue (https://github.com/gpac/gpac/issues/2488) you mentioned " Here's what I checked: dashing using 15s segments instead of 30s: playback until end OK dashing video only with 30s segments: playback until end OK dashing audio only with 30s segments: playback until end OK The segment durations in 30s Audio+Video and 30s VideoOnly are the same, and the last segments are correct. I think that's a WMP bug, there's not much we can do here. " I have set the segments to 29s 20s 15s , playback is OK. [babcock_15.m3u8.txt](https://github.com/gpac/gpac/files/12879541/babcock_15.m3u8.txt) [babcock_29.m3u8.txt](https://github.com/gpac/gpac/files/12879537/babcock_29.m3u8.txt) [babcock_30.m3u8.txt](https://github.com/gpac/gpac/files/12879538/babcock_30.m3u8.txt) babcock_15 and babcock_29 plays ok for the same mp4 I use different version of mp4box ,the segment is different . [20230922_new30.m3u8.txt](https://github.com/gpac/gpac/files/12879521/20230922_new30.m3u8.txt) [20230922_org30.m3u8.txt](https://github.com/gpac/gpac/files/12879523/20230922_org30.m3u8.txt) 20230922_new30 plays ok looks like this issue is related to segment strategy do you still have the mp4 file(timely.mp4) mentioned in https://github.com/gpac/gpac/issues/2488 ? we have a lot of cases here, however ,I could not send them to you , for it is customer data. [timely.zip](https://github.com/gpac/gpac/files/12879591/timely.zip) here is what I can provide timly_29.m3u8 plays OK while timly_30.m3u8 last few seconds is not playable.
the last few seconds of mp4 is not playable in windows media player
https://api.github.com/repos/gpac/gpac/issues/2631/comments
5
2023-10-12T09:18:55
2024-06-04T14:36:14Z
https://github.com/gpac/gpac/issues/2631
1,939,569,513
2,631
false
This is a GitHub Issue repo:gpac owner:gpac Title : the last few seconds of mp4 is not playable in windows media player Issue date: --- start body --- MP4Box -dash 30000 -profile onDemand music_normal.mp4 -out music.mpd --dual play music_normal_dashinit.mp4 in windows media player when seek to the last part of the video. it is not playable and will jump back to the beggining . no problem on other player . @jeanlf in the last issue (https://github.com/gpac/gpac/issues/2488) you mentioned " Here's what I checked: dashing using 15s segments instead of 30s: playback until end OK dashing video only with 30s segments: playback until end OK dashing audio only with 30s segments: playback until end OK The segment durations in 30s Audio+Video and 30s VideoOnly are the same, and the last segments are correct. I think that's a WMP bug, there's not much we can do here. " I have set the segments to 29s 20s 15s , playback is OK. [babcock_15.m3u8.txt](https://github.com/gpac/gpac/files/12879541/babcock_15.m3u8.txt) [babcock_29.m3u8.txt](https://github.com/gpac/gpac/files/12879537/babcock_29.m3u8.txt) [babcock_30.m3u8.txt](https://github.com/gpac/gpac/files/12879538/babcock_30.m3u8.txt) babcock_15 and babcock_29 plays ok for the same mp4 I use different version of mp4box ,the segment is different . [20230922_new30.m3u8.txt](https://github.com/gpac/gpac/files/12879521/20230922_new30.m3u8.txt) [20230922_org30.m3u8.txt](https://github.com/gpac/gpac/files/12879523/20230922_org30.m3u8.txt) 20230922_new30 plays ok looks like this issue is related to segment strategy do you still have the mp4 file(timely.mp4) mentioned in https://github.com/gpac/gpac/issues/2488 ? we have a lot of cases here, however ,I could not send them to you , for it is customer data. [timely.zip](https://github.com/gpac/gpac/files/12879591/timely.zip) here is what I can provide timly_29.m3u8 plays OK while timly_30.m3u8 last few seconds is not playable. --- end body ---
2,019
[ -0.020332666113972664, 0.0060927290469408035, -0.018117127940058708, 0.01107768900692463, -0.025266557931900024, 0.006497339345514774, -0.017897145822644234, 0.039753977209329605, 0.0029461937956511974, 0.03899975121021271, -0.007986149750649929, 0.011572649702429771, 0.010409885086119175, ...
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
[POC](https://github.com/qianshuidewajueji/poc/blob/main/gpac/gf_text_get_utf8_line_poc) ``` Version MP4Box - GPAC version 2.3-DEV-rev40-g3602a5ded-master (c) 2000-2023 Telecom Paris distributed under LGPL v2.1+ - http://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: --enable-sanitizer --verbose Features: GPAC_CONFIG_LINUX GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SSL GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_PNG GPAC_HAS_LINUX_DVB GPAC_DISABLE_3D Proof of Concept ➜ gcc git:(master) ✗ ./MP4Box -info ./gf_text_get_utf8_line_poc filters/load_text.c:362:13: runtime error: index 2048 out of bounds for type 'char [2048]' Reproduce ./configure --enable-sanitizer --enable-debug make ./MP4Box -info gf_text_get_utf8_line_poc Git Log commit 3602a5ded4e57b0044a949f985ee3792f94a9a36 (HEAD -> master, origin/master, origin/HEAD) Author: Aurelien David <aurelien.david@telecom-paristech.fr> Date: Thu Feb 9 11:24:23 2023 +0100 mp3dmx: check truncated frames (#2391) commit ea7395f39f601a7750d48d606e9d10ea0b7beefe Author: Aurelien David <aurelien.david@telecom-paristech.fr> Date: Wed Feb 8 16:52:00 2023 +0100 sgpd box entry: disallow null grouping_type (#2389) commit 8db20cb634a546c536c31caac94e1f74b778b463 Author: Aurelien David <aurelien.david@telecom-paristech.fr> Date: Tue Feb 7 18:27:19 2023 +0100 m2ts: check descs_size read from input to prevent overflow (#2388) Impact This is capable of causing crashes by using unexpected value. Occurrences load_text.c L362 if (j >= GF_ARRAY_LENGTH(szLineConv) - 1) { GF_LOG(GF_LOG_DEBUG, GF_LOG_PARSER, ("[TXTIn] Line too long to convert to utf8 (len: %d)\n", len)); break; } } szLineConv[j] = 0; // while j == GF_ARRAY_LENGTH(szLineConv) case off-by-one error strcpy(szLine, szLineConv); ```
off-by-one error in function gf_text_get_utf8_line filters/load_text.c
https://api.github.com/repos/gpac/gpac/issues/2397/comments
1
2023-02-13T09:57:01
2023-02-13T16:32:57Z
https://github.com/gpac/gpac/issues/2397
1,582,015,060
2,397
false
This is a GitHub Issue repo:gpac owner:gpac Title : off-by-one error in function gf_text_get_utf8_line filters/load_text.c Issue date: --- start body --- [POC](https://github.com/qianshuidewajueji/poc/blob/main/gpac/gf_text_get_utf8_line_poc) ``` Version MP4Box - GPAC version 2.3-DEV-rev40-g3602a5ded-master (c) 2000-2023 Telecom Paris distributed under LGPL v2.1+ - http://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: --enable-sanitizer --verbose Features: GPAC_CONFIG_LINUX GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SSL GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_PNG GPAC_HAS_LINUX_DVB GPAC_DISABLE_3D Proof of Concept ➜ gcc git:(master) ✗ ./MP4Box -info ./gf_text_get_utf8_line_poc filters/load_text.c:362:13: runtime error: index 2048 out of bounds for type 'char [2048]' Reproduce ./configure --enable-sanitizer --enable-debug make ./MP4Box -info gf_text_get_utf8_line_poc Git Log commit 3602a5ded4e57b0044a949f985ee3792f94a9a36 (HEAD -> master, origin/master, origin/HEAD) Author: Aurelien David <aurelien.david@telecom-paristech.fr> Date: Thu Feb 9 11:24:23 2023 +0100 mp3dmx: check truncated frames (#2391) commit ea7395f39f601a7750d48d606e9d10ea0b7beefe Author: Aurelien David <aurelien.david@telecom-paristech.fr> Date: Wed Feb 8 16:52:00 2023 +0100 sgpd box entry: disallow null grouping_type (#2389) commit 8db20cb634a546c536c31caac94e1f74b778b463 Author: Aurelien David <aurelien.david@telecom-paristech.fr> Date: Tue Feb 7 18:27:19 2023 +0100 m2ts: check descs_size read from input to prevent overflow (#2388) Impact This is capable of causing crashes by using unexpected value. Occurrences load_text.c L362 if (j >= GF_ARRAY_LENGTH(szLineConv) - 1) { GF_LOG(GF_LOG_DEBUG, GF_LOG_PARSER, ("[TXTIn] Line too long to convert to utf8 (len: %d)\n", len)); break; } } szLineConv[j] = 0; // while j == GF_ARRAY_LENGTH(szLineConv) case off-by-one error strcpy(szLine, szLineConv); ``` --- end body ---
2,238
[ -0.009625888429582119, 0.04087069630622864, -0.01432272233068943, 0.020660078153014183, 0.02906493842601776, 0.014397632330656052, -0.03260067477822304, 0.05072880536317825, -0.027686601504683495, -0.01808318682014942, -0.01003040000796318, 0.020465312525629997, 0.0059028794057667255, 0.01...
null
null
null
null
null
null
null
null
null
[ "ImageMagick", "ImageMagick" ]
### Is your feature request related to a problem? Please describe. _No response_ ### Describe the solution you'd like I understand ImageMagick is primarly a raster image processing library, but it support vector formats such as SVG and WPG too. Having support for ODG files - part of OpenDocument Format and created by LibreOffice Draw - would be a nice addition. ### Describe alternatives you've considered _No response_ ### Additional context _No response_
Support for ODG (OpenDocument Format Graphics)
https://api.github.com/repos/ImageMagick/ImageMagick/issues/6827/comments
3
2023-10-29T12:39:06
2023-10-30T16:57:14Z
https://github.com/ImageMagick/ImageMagick/issues/6827
1,966,974,149
6,827
false
This is a GitHub Issue repo:ImageMagick owner:ImageMagick Title : Support for ODG (OpenDocument Format Graphics) Issue date: --- start body --- ### Is your feature request related to a problem? Please describe. _No response_ ### Describe the solution you'd like I understand ImageMagick is primarly a raster image processing library, but it support vector formats such as SVG and WPG too. Having support for ODG files - part of OpenDocument Format and created by LibreOffice Draw - would be a nice addition. ### Describe alternatives you've considered _No response_ ### Additional context _No response_ --- end body ---
628
[ -0.021618666127324104, 0.028298821300268173, -0.01880238950252533, 0.03597215563058853, 0.0035373526625335217, 0.02738727256655693, 0.013543978333473206, 0.03191780298948288, -0.0288702379912138, 0.05267934128642082, 0.028516503050923347, -0.009115485474467278, 0.00554752117022872, 0.00951...
null
null
null
null
null
null
null
null
null
[ "slims", "slims9_bulian" ]
It seems to SLiMS 9 Bulian are using both PDO and MySQLi? Or an abstraction DB like ADOdb? Sometimes, I custom some codes for my library and I feel "something wrong", and I had to edit code lines , some cases it looked like PDO, and some case, it looks like MySQLi. So, I have a question: Should I use PDO or MySQLi? (for compatble in the future when I want to upgrade current SLiMS to a newer version?). SLiMS's developers will support which?
SLiMS uses PDO or MySQLi for future?
https://api.github.com/repos/slims/slims9_bulian/issues/212/comments
1
2023-09-20T04:20:54
2023-09-21T22:28:10Z
https://github.com/slims/slims9_bulian/issues/212
1,904,096,844
212
false
This is a GitHub Issue repo:slims9_bulian owner:slims Title : SLiMS uses PDO or MySQLi for future? Issue date: --- start body --- It seems to SLiMS 9 Bulian are using both PDO and MySQLi? Or an abstraction DB like ADOdb? Sometimes, I custom some codes for my library and I feel "something wrong", and I had to edit code lines , some cases it looked like PDO, and some case, it looks like MySQLi. So, I have a question: Should I use PDO or MySQLi? (for compatble in the future when I want to upgrade current SLiMS to a newer version?). SLiMS's developers will support which? --- end body ---
599
[ -0.00861549936234951, 0.025883475318551064, -0.021830862388014793, 0.0083048976957798, 0.01629919372498989, 0.036532677710056305, 0.011913794092833996, -0.0034240141976624727, -0.012719879858195782, 0.04706355556845665, 0.030379805713891983, -0.006193546112626791, 0.020499713718891144, 0.0...
CVE-2023-48088
2023-11-15T15:15:07.510000
xxl-job-admin 2.4.0 is vulnerable to Cross Site Scripting (XSS) via /xxl-job-admin/joblog/logDetailPage.
{ "cvssMetricV2": null, "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "LOW", "attackVector": "NETWORK", "availabilityImpact": "NONE", "baseScore": 5.4, "baseSeverity": "MEDIUM", "confidentialityImpact": "LOW", "integrityImpact": "LOW", "privilegesRequired": "LOW", "scope": "CHANGED", "userInteraction": "REQUIRED", "vectorString": "CVSS:3.1/AV:N/AC:L/PR:L/UI:R/S:C/C:L/I:L/A:N", "version": "3.1" }, "exploitabilityScore": 2.3, "impactScore": 2.7, "source": "nvd@nist.gov", "type": "Primary" } ] }
[ { "source": "cve@mitre.org", "tags": [ "Exploit", "Issue Tracking" ], "url": "https://github.com/xuxueli/xxl-job/issues/3329" } ]
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:xuxueli:xxl-job:2.4.0:*:*:*:*:*:*:*", "matchCriteriaId": "122ECDB6-272C-4B31-907E-E9FFB36305EA", "versionEndExcluding": null, "versionEndIncluding": null, "versionStartExclud...
https://github.com/xuxueli/xxl-job/issues/3329
[ "Exploit", "Issue Tracking" ]
github.com
[ "xuxueli", "xxl-job" ]
### Environment MySQL 5.7.44, XXL-Job-Admin 2.4.0 Virtual Machine 1: Ubuntu 22.04.3 (as XXL-Job-Admin) Virtual Machine 2: Ubuntu 22.04.3 (as XXL-Job-Executor) ### Vulnerability Information During the query of /xxl-job-admin/joblog/logDetailPage, the xxl-job-admin will query the related log directly in the machine and show it in the console in HTML format even if the log appears in <script> </script> format ### Steps to reproduce the behavior Step 1: Modify the application log in default path of XXL-Job-Executor and add malicious javascript ` cd /data/applogs/xxl-job/jobhandler/yyyy-mm-dd/` ![image](https://github.com/xuxueli/xxl-job/assets/150101148/a801ba9f-38a2-4bb0-807e-bbc3132947f1) Example malicious code `<script>alert(Test123);</script>` ![image](https://github.com/xuxueli/xxl-job/assets/150101148/dc9be914-7997-4dca-945e-6214c95204a9) Step 2: Login to the XXL-Job-Admin console by admin user and navigate to Log Query Page Check the log by querying log id ![image](https://github.com/xuxueli/xxl-job/assets/150101148/08ded6af-6a3c-4182-8651-2839ad99abc7) Step 3: Alert will show here ![image](https://github.com/xuxueli/xxl-job/assets/150101148/e751315a-6fd4-4169-b407-7c19f8457e05)
XSS attack appears in /xxl-job-admin/joblog/logDetailPage
https://api.github.com/repos/xuxueli/xxl-job/issues/3329/comments
1
2023-11-08T02:03:54
2023-11-23T09:50:55Z
https://github.com/xuxueli/xxl-job/issues/3329
1,982,577,272
3,329
true
This is a GitHub Issue repo:xxl-job owner:xuxueli Title : XSS attack appears in /xxl-job-admin/joblog/logDetailPage Issue date: --- start body --- ### Environment MySQL 5.7.44, XXL-Job-Admin 2.4.0 Virtual Machine 1: Ubuntu 22.04.3 (as XXL-Job-Admin) Virtual Machine 2: Ubuntu 22.04.3 (as XXL-Job-Executor) ### Vulnerability Information During the query of /xxl-job-admin/joblog/logDetailPage, the xxl-job-admin will query the related log directly in the machine and show it in the console in HTML format even if the log appears in <script> </script> format ### Steps to reproduce the behavior Step 1: Modify the application log in default path of XXL-Job-Executor and add malicious javascript ` cd /data/applogs/xxl-job/jobhandler/yyyy-mm-dd/` ![image](https://github.com/xuxueli/xxl-job/assets/150101148/a801ba9f-38a2-4bb0-807e-bbc3132947f1) Example malicious code `<script>alert(Test123);</script>` ![image](https://github.com/xuxueli/xxl-job/assets/150101148/dc9be914-7997-4dca-945e-6214c95204a9) Step 2: Login to the XXL-Job-Admin console by admin user and navigate to Log Query Page Check the log by querying log id ![image](https://github.com/xuxueli/xxl-job/assets/150101148/08ded6af-6a3c-4182-8651-2839ad99abc7) Step 3: Alert will show here ![image](https://github.com/xuxueli/xxl-job/assets/150101148/e751315a-6fd4-4169-b407-7c19f8457e05) --- end body ---
1,394
[ -0.016475647687911987, 0.0033297918271273375, -0.00035699387080967426, 0.008370572701096535, 0.04596071317791939, -0.016165899112820625, 0.0052657173946499825, 0.03640277311205864, 0.009152318350970745, -0.004159474279731512, 0.014882657676935196, -0.005693464539945126, -0.02830507420003414,...
null
null
null
null
null
null
null
null
null
[ "axiomatic-systems", "Bento4" ]
Hi, there. There is a reachable assertion failure in mp4info, Ap4Atom.cpp:763, in the newest commit 5b7cc250. This issue seems to be an incomplete fix of issues #197, #351, and #547. To reproduce, run: ~~~~ ./mp4info poc /dev/null ~~~~ Here is the trace reported by ASAN: ~~~~ ==3028932==ERROR: AddressSanitizer: ABRT on unknown address 0x03ea002e37c4 (pc 0x7fb712c6e00b bp 0x7fb712de3588 sp 0x7ffd90cb0930 T0) #0 0x7fb712c6e00b in raise /build/glibc-SzIz7B/glibc-2.31/signal/../sysdeps/unix/sysv/linux/raise.c:51:1 #1 0x7fb712c4d858 in abort /build/glibc-SzIz7B/glibc-2.31/stdlib/abort.c:79:7 #2 0x7fb712c4d728 in __assert_fail_base /build/glibc-SzIz7B/glibc-2.31/assert/assert.c:92:3 #3 0x7fb712c5efd5 in __assert_fail /build/glibc-SzIz7B/glibc-2.31/assert/assert.c:101:3 #4 0x5286f0 in AP4_AtomListWriter::Action(AP4_Atom*) const /benchmark/Bento4/Source/C++/Core/Ap4Atom.cpp:763:5 #5 0x5f2c92 in AP4_List<AP4_Atom>::Apply(AP4_List<AP4_Atom>::Item::Operator const&) const /benchmark/Bento4/Source/C++/Core/Ap4List.h:353:12 #6 0x5f2c92 in AP4_SampleEntry::Write(AP4_ByteStream&) /benchmark/Bento4/Source/C++/Core/Ap4SampleEntry.cpp:176:23 #7 0x523d0c in AP4_Atom::Clone() /benchmark/Bento4/Source/C++/Core/Ap4Atom.cpp:316:9 #8 0x553d57 in AP4_ContainerAtom::Clone() /home/heqing/benchmark/Bento4/Source/C++/Core/Ap4ContainerAtom.cpp:172:56 #9 0x553d57 in AP4_ContainerAtom::Clone() /benchmark/Bento4/Source/C++/Core/Ap4ContainerAtom.cpp:172:56 #10 0x553d57 in AP4_ContainerAtom::Clone() /benchmark/Bento4/Source/C++/Core/Ap4ContainerAtom.cpp:172:56 #11 0x553d57 in AP4_ContainerAtom::Clone() /benchmark/Bento4/Source/C++/Core/Ap4ContainerAtom.cpp:172:56 #12 0x5e6fa1 in AP4_SampleDescription::AP4_SampleDescription(AP4_SampleDescription::Type, unsigned int, AP4_AtomParent*) /benchmark/Bento4/Source/C++/Core/Ap4SampleDescription.cpp:138:41 #13 0x5e6fa1 in AP4_AvcSampleDescription::AP4_AvcSampleDescription(unsigned int, unsigned short, unsigned short, unsigned short, char const*, AP4_AtomParent*) /benchmark/Bento4/Source/C++/Core/Ap4SampleDescription.cpp:383:5 #14 0x5fb814 in AP4_AvcSampleEntry::ToSampleDescription() /benchmark/Bento4/Source/C++/Core/Ap4SampleEntry.cpp:1146:16 #15 0x6144a4 in AP4_StsdAtom::GetSampleDescription(unsigned int) /benchmark/Bento4/Source/C++/Core/Ap4StsdAtom.cpp:181:53 #16 0x500666 in ShowTrackInfo_Text(AP4_Movie&, AP4_Track&, AP4_ByteStream&, bool, bool, bool, bool) /benchmark/Bento4/Source/C++/Apps/Mp4Info/Mp4Info.cpp:1351:52 #17 0x500666 in ShowTrackInfo(AP4_Movie&, AP4_Track&, AP4_ByteStream&, bool, bool, bool, bool) /benchmark/Bento4/Source/C++/Apps/Mp4Info/Mp4Info.cpp:1475:13 #18 0x4ff54d in ShowTracks(AP4_Movie&, AP4_List<AP4_Track>&, AP4_ByteStream&, bool, bool, bool, bool) /benchmark/Bento4/Source/C++/Apps/Mp4Info/Mp4Info.cpp:1586:9 #19 0x4fbd0f in main /benchmark/Bento4/Source/C++/Apps/Mp4Info/Mp4Info.cpp:1868:13 #20 0x7fb712c4f082 in __libc_start_main /build/glibc-SzIz7B/glibc-2.31/csu/../csu/libc-start.c:308:16 #21 0x41d89d in _start (/benchmark/Bento4/build-a/mp4info+0x41d89d) AddressSanitizer can not provide additional info. SUMMARY: AddressSanitizer: ABRT /build/glibc-SzIz7B/glibc-2.31/signal/../sysdeps/unix/sysv/linux/raise.c:51:1 in raise ==3028932==ABORTIN ~~~~ [mp4info_assertion_Ap4Atom.cpp763.zip](https://github.com/axiomatic-systems/Bento4/files/9642730/mp4info_assertion_Ap4Atom.cpp763.zip) (unzip first)
Assertion Failure in mp4info, Ap4Atom.cpp:763
https://api.github.com/repos/axiomatic-systems/Bento4/issues/777/comments
1
2022-09-26T03:28:17
2023-05-29T02:51:15Z
https://github.com/axiomatic-systems/Bento4/issues/777
1,385,355,774
777
false
This is a GitHub Issue repo:Bento4 owner:axiomatic-systems Title : Assertion Failure in mp4info, Ap4Atom.cpp:763 Issue date: --- start body --- Hi, there. There is a reachable assertion failure in mp4info, Ap4Atom.cpp:763, in the newest commit 5b7cc250. This issue seems to be an incomplete fix of issues #197, #351, and #547. To reproduce, run: ~~~~ ./mp4info poc /dev/null ~~~~ Here is the trace reported by ASAN: ~~~~ ==3028932==ERROR: AddressSanitizer: ABRT on unknown address 0x03ea002e37c4 (pc 0x7fb712c6e00b bp 0x7fb712de3588 sp 0x7ffd90cb0930 T0) #0 0x7fb712c6e00b in raise /build/glibc-SzIz7B/glibc-2.31/signal/../sysdeps/unix/sysv/linux/raise.c:51:1 #1 0x7fb712c4d858 in abort /build/glibc-SzIz7B/glibc-2.31/stdlib/abort.c:79:7 #2 0x7fb712c4d728 in __assert_fail_base /build/glibc-SzIz7B/glibc-2.31/assert/assert.c:92:3 #3 0x7fb712c5efd5 in __assert_fail /build/glibc-SzIz7B/glibc-2.31/assert/assert.c:101:3 #4 0x5286f0 in AP4_AtomListWriter::Action(AP4_Atom*) const /benchmark/Bento4/Source/C++/Core/Ap4Atom.cpp:763:5 #5 0x5f2c92 in AP4_List<AP4_Atom>::Apply(AP4_List<AP4_Atom>::Item::Operator const&) const /benchmark/Bento4/Source/C++/Core/Ap4List.h:353:12 #6 0x5f2c92 in AP4_SampleEntry::Write(AP4_ByteStream&) /benchmark/Bento4/Source/C++/Core/Ap4SampleEntry.cpp:176:23 #7 0x523d0c in AP4_Atom::Clone() /benchmark/Bento4/Source/C++/Core/Ap4Atom.cpp:316:9 #8 0x553d57 in AP4_ContainerAtom::Clone() /home/heqing/benchmark/Bento4/Source/C++/Core/Ap4ContainerAtom.cpp:172:56 #9 0x553d57 in AP4_ContainerAtom::Clone() /benchmark/Bento4/Source/C++/Core/Ap4ContainerAtom.cpp:172:56 #10 0x553d57 in AP4_ContainerAtom::Clone() /benchmark/Bento4/Source/C++/Core/Ap4ContainerAtom.cpp:172:56 #11 0x553d57 in AP4_ContainerAtom::Clone() /benchmark/Bento4/Source/C++/Core/Ap4ContainerAtom.cpp:172:56 #12 0x5e6fa1 in AP4_SampleDescription::AP4_SampleDescription(AP4_SampleDescription::Type, unsigned int, AP4_AtomParent*) /benchmark/Bento4/Source/C++/Core/Ap4SampleDescription.cpp:138:41 #13 0x5e6fa1 in AP4_AvcSampleDescription::AP4_AvcSampleDescription(unsigned int, unsigned short, unsigned short, unsigned short, char const*, AP4_AtomParent*) /benchmark/Bento4/Source/C++/Core/Ap4SampleDescription.cpp:383:5 #14 0x5fb814 in AP4_AvcSampleEntry::ToSampleDescription() /benchmark/Bento4/Source/C++/Core/Ap4SampleEntry.cpp:1146:16 #15 0x6144a4 in AP4_StsdAtom::GetSampleDescription(unsigned int) /benchmark/Bento4/Source/C++/Core/Ap4StsdAtom.cpp:181:53 #16 0x500666 in ShowTrackInfo_Text(AP4_Movie&, AP4_Track&, AP4_ByteStream&, bool, bool, bool, bool) /benchmark/Bento4/Source/C++/Apps/Mp4Info/Mp4Info.cpp:1351:52 #17 0x500666 in ShowTrackInfo(AP4_Movie&, AP4_Track&, AP4_ByteStream&, bool, bool, bool, bool) /benchmark/Bento4/Source/C++/Apps/Mp4Info/Mp4Info.cpp:1475:13 #18 0x4ff54d in ShowTracks(AP4_Movie&, AP4_List<AP4_Track>&, AP4_ByteStream&, bool, bool, bool, bool) /benchmark/Bento4/Source/C++/Apps/Mp4Info/Mp4Info.cpp:1586:9 #19 0x4fbd0f in main /benchmark/Bento4/Source/C++/Apps/Mp4Info/Mp4Info.cpp:1868:13 #20 0x7fb712c4f082 in __libc_start_main /build/glibc-SzIz7B/glibc-2.31/csu/../csu/libc-start.c:308:16 #21 0x41d89d in _start (/benchmark/Bento4/build-a/mp4info+0x41d89d) AddressSanitizer can not provide additional info. SUMMARY: AddressSanitizer: ABRT /build/glibc-SzIz7B/glibc-2.31/signal/../sysdeps/unix/sysv/linux/raise.c:51:1 in raise ==3028932==ABORTIN ~~~~ [mp4info_assertion_Ap4Atom.cpp763.zip](https://github.com/axiomatic-systems/Bento4/files/9642730/mp4info_assertion_Ap4Atom.cpp763.zip) (unzip first) --- end body ---
3,708
[ -0.0326714850962162, 0.029791440814733505, -0.0132776303216815, -0.004188453778624535, 0.03415796160697937, -0.024635229259729385, -0.02033064514398575, 0.0415593683719635, -0.02271519973874092, 0.02307133376598358, -0.014407970942556858, 0.007459473330527544, 0.027701085433363914, 0.00914...
null
null
null
null
null
null
null
null
null
[ "jerryscript-project", "jerryscript" ]
## JerryScript commit hash 55acdf2048b390d0f56f12e64dbfb2559f0e70ad ## Build platform Ubuntu 20.04 LTS ## Build steps ``` ./tools/build.py --clean --debug --compile-flag=-fsanitize=address \ --compile-flag=-m32 --compile-flag=-fno-omit-frame-pointer \ --compile-flag=-fno-common --compile-flag=-g \ --strip=off --system-allocator=on --logging=on \ --linker-flag=-fuse-ld=gold --error-messages=on --line-info=ON \ --stack-limit=10 ``` ## poc ``` function foo(a, i) { a[i] = 1; return a[i]; } class MyArray extends class C extends Array { } { } ; o = new MyArray(); print(foo); print(1, foo(o, 0)); print(1, foo(o, 1)); o.__proto__.__proto__ = new Int32Array(2); print(undefined, foo(o, 2)); print(undefined, foo(o, 2)); ``` ## assert log ``` ICE: Assertion 'context_p->scope_stack_size == PARSER_MAXIMUM_DEPTH_OF_SCOPE_STACK' failed at /home/sakura/jerryscript/jerry-core/parser/js/js-parser-expr.c(parser_parse_class):1160. Error: ERR_FAILED_INTERNAL_ASSERTION ```
Assertion 'context_p->scope_stack_size == PARSER_MAXIMUM_DEPTH_OF_SCOPE_STACK' failed at jerry-core/parser/js/js-parser-expr.c(parser_parse_class):1160
https://api.github.com/repos/jerryscript-project/jerryscript/issues/4892/comments
0
2021-12-09T14:39:45
2021-12-22T13:50:47Z
https://github.com/jerryscript-project/jerryscript/issues/4892
1,075,648,914
4,892
false
This is a GitHub Issue repo:jerryscript owner:jerryscript-project Title : Assertion 'context_p->scope_stack_size == PARSER_MAXIMUM_DEPTH_OF_SCOPE_STACK' failed at jerry-core/parser/js/js-parser-expr.c(parser_parse_class):1160 Issue date: --- start body --- ## JerryScript commit hash 55acdf2048b390d0f56f12e64dbfb2559f0e70ad ## Build platform Ubuntu 20.04 LTS ## Build steps ``` ./tools/build.py --clean --debug --compile-flag=-fsanitize=address \ --compile-flag=-m32 --compile-flag=-fno-omit-frame-pointer \ --compile-flag=-fno-common --compile-flag=-g \ --strip=off --system-allocator=on --logging=on \ --linker-flag=-fuse-ld=gold --error-messages=on --line-info=ON \ --stack-limit=10 ``` ## poc ``` function foo(a, i) { a[i] = 1; return a[i]; } class MyArray extends class C extends Array { } { } ; o = new MyArray(); print(foo); print(1, foo(o, 0)); print(1, foo(o, 1)); o.__proto__.__proto__ = new Int32Array(2); print(undefined, foo(o, 2)); print(undefined, foo(o, 2)); ``` ## assert log ``` ICE: Assertion 'context_p->scope_stack_size == PARSER_MAXIMUM_DEPTH_OF_SCOPE_STACK' failed at /home/sakura/jerryscript/jerry-core/parser/js/js-parser-expr.c(parser_parse_class):1160. Error: ERR_FAILED_INTERNAL_ASSERTION ``` --- end body ---
1,299
[ -0.014661015942692757, 0.023344332352280617, -0.009115618653595448, -0.00895164255052805, 0.023001471534371376, -0.008824932388961315, 0.002677666489034891, 0.036820244044065475, -0.03133447468280792, 0.02675803005695343, -0.00425594300031662, -0.005687013268470764, 0.02182381972670555, 0....
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
Attempting to generate a live-profile DASH manifest using two xHE-AAC representations that have different frame sizes results in a warning that the segments are not aligned, whereas inspection of the source media files indicates that they are. Full command that invokes the error: ``` $ /Applications/GPAC.app/Contents/MacOS/MP4Box -dash 5000 -rap -out sintel/live.mpd:dual -profile live -segment-name '$RepresentationID$_$Number%03d$' sintel_12.mp4 sintel_24.mp4 [Dasher] Segments are not aligned across representations: first rep segment duration 4.92263 but new segment duration 6.91955 for the same segment 1 [Dasher] No segment alignment, switching to full profile [MPD] Generating MPD at time 2022-08-16T10:39:40.544Zsegs)(85.14 %) / MPD 29.26s 98 %.63s 94 %% [Dasher] End of Period [Dasher] End of MPD (no more active streams) ``` The media files causing the error are uploaded here: Representation 1 (12kbit/s): https://www.mediafire.com/file/pm4qg2rmwu868nr/sintel_12.mp4/file Representation 2 (24kbit/s): https://www.mediafire.com/file/qvrii06n54y0iew/sintel_24.mp4/file They were generated using a commercial xHE-AAC encoder so unfortunately I cannot give the command that produced them, but using the same build of MP4Box we can see that the random access points are located at exactly the same composition timesamps in both files: ``` $ /Applications/GPAC.app/Contents/MacOS/MP4Box -std -dts sintel_12.mp4 2>/dev/null | awk '$8 == "1" { print $0 }' Sample 1 DTS 0 CTS 0 339 1 2529 0 0 0 0 0 0 0 Sample 54 DTS 217088 CTS 217088 451 1 10114 0 0 0 0 0 0 0 Sample 108 DTS 438272 CTS 438272 568 1 18061 0 0 0 0 0 0 0 Sample 162 DTS 659456 CTS 659456 518 1 25882 0 0 0 0 0 0 0 Sample 216 DTS 880640 CTS 880640 546 1 33966 0 0 0 0 0 0 0 Sample 270 DTS 1101824 CTS 1101824 559 1 42042 0 0 0 0 0 0 0 Sample 323 DTS 1318912 CTS 1318912 557 1 49770 0 0 0 0 0 0 0 $ /Applications/GPAC.app/Contents/MacOS/MP4Box -std -dts sintel_24.mp4 2>/dev/null | awk '$8 == "1" { print $0 }' Sample 1 DTS 0 CTS 0 350 1 3821 0 0 0 0 0 0 0 Sample 107 DTS 217088 CTS 217088 516 1 18814 0 0 0 0 0 0 0 Sample 215 DTS 438272 CTS 438272 504 1 34244 0 0 0 0 0 0 0 Sample 323 DTS 659456 CTS 659456 494 1 49724 0 0 0 0 0 0 0 Sample 431 DTS 880640 CTS 880640 504 1 65142 0 0 0 0 0 0 0 Sample 539 DTS 1101824 CTS 1101824 516 1 80491 0 0 0 0 0 0 0 Sample 645 DTS 1318912 CTS 1318912 518 1 95637 0 0 0 0 0 0 0 ``` Notice that the "Sample" number is different in each representation because the xHE-AAC encoder uses different frame sizes for different bitrates. Nevertheless, the CTS for each RAP is the same in both files. Operating system: MacOS 12.5 MP4Box installed from the nightly builds: ``` $ /Applications/GPAC.app/Contents/MacOS/MP4Box -version MP4Box - GPAC version 2.1-DEV-rev259-gfc4749f9-master (c) 2000-2022 Telecom Paris distributed under LGPL v2.1+ - http://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: --extra-cflags=-Wno-deprecated -Wno-shift-negative-value Features: GPAC_CONFIG_DARWIN GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SSL GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_FAAD GPAC_HAS_MAD GPAC_HAS_LIBA52 GPAC_HAS_JPEG GPAC_HAS_PNG GPAC_HAS_FFMPEG GPAC_HAS_OPENSVC GPAC_HAS_JP2 GPAC_HAS_OPENHEVC GPAC_HAS_THEORA GPAC_HAS_VORBIS GPAC_HAS_XVID ``` Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/2013/07/16/how-to-file-a-bug-properly/
[MP4Box] DASH for xHE-AAC wrongly creates unaligned segments
https://api.github.com/repos/gpac/gpac/issues/2242/comments
3
2022-08-16T10:51:26
2022-09-12T15:58:07Z
https://github.com/gpac/gpac/issues/2242
1,340,166,212
2,242
false
This is a GitHub Issue repo:gpac owner:gpac Title : [MP4Box] DASH for xHE-AAC wrongly creates unaligned segments Issue date: --- start body --- Attempting to generate a live-profile DASH manifest using two xHE-AAC representations that have different frame sizes results in a warning that the segments are not aligned, whereas inspection of the source media files indicates that they are. Full command that invokes the error: ``` $ /Applications/GPAC.app/Contents/MacOS/MP4Box -dash 5000 -rap -out sintel/live.mpd:dual -profile live -segment-name '$RepresentationID$_$Number%03d$' sintel_12.mp4 sintel_24.mp4 [Dasher] Segments are not aligned across representations: first rep segment duration 4.92263 but new segment duration 6.91955 for the same segment 1 [Dasher] No segment alignment, switching to full profile [MPD] Generating MPD at time 2022-08-16T10:39:40.544Zsegs)(85.14 %) / MPD 29.26s 98 %.63s 94 %% [Dasher] End of Period [Dasher] End of MPD (no more active streams) ``` The media files causing the error are uploaded here: Representation 1 (12kbit/s): https://www.mediafire.com/file/pm4qg2rmwu868nr/sintel_12.mp4/file Representation 2 (24kbit/s): https://www.mediafire.com/file/qvrii06n54y0iew/sintel_24.mp4/file They were generated using a commercial xHE-AAC encoder so unfortunately I cannot give the command that produced them, but using the same build of MP4Box we can see that the random access points are located at exactly the same composition timesamps in both files: ``` $ /Applications/GPAC.app/Contents/MacOS/MP4Box -std -dts sintel_12.mp4 2>/dev/null | awk '$8 == "1" { print $0 }' Sample 1 DTS 0 CTS 0 339 1 2529 0 0 0 0 0 0 0 Sample 54 DTS 217088 CTS 217088 451 1 10114 0 0 0 0 0 0 0 Sample 108 DTS 438272 CTS 438272 568 1 18061 0 0 0 0 0 0 0 Sample 162 DTS 659456 CTS 659456 518 1 25882 0 0 0 0 0 0 0 Sample 216 DTS 880640 CTS 880640 546 1 33966 0 0 0 0 0 0 0 Sample 270 DTS 1101824 CTS 1101824 559 1 42042 0 0 0 0 0 0 0 Sample 323 DTS 1318912 CTS 1318912 557 1 49770 0 0 0 0 0 0 0 $ /Applications/GPAC.app/Contents/MacOS/MP4Box -std -dts sintel_24.mp4 2>/dev/null | awk '$8 == "1" { print $0 }' Sample 1 DTS 0 CTS 0 350 1 3821 0 0 0 0 0 0 0 Sample 107 DTS 217088 CTS 217088 516 1 18814 0 0 0 0 0 0 0 Sample 215 DTS 438272 CTS 438272 504 1 34244 0 0 0 0 0 0 0 Sample 323 DTS 659456 CTS 659456 494 1 49724 0 0 0 0 0 0 0 Sample 431 DTS 880640 CTS 880640 504 1 65142 0 0 0 0 0 0 0 Sample 539 DTS 1101824 CTS 1101824 516 1 80491 0 0 0 0 0 0 0 Sample 645 DTS 1318912 CTS 1318912 518 1 95637 0 0 0 0 0 0 0 ``` Notice that the "Sample" number is different in each representation because the xHE-AAC encoder uses different frame sizes for different bitrates. Nevertheless, the CTS for each RAP is the same in both files. Operating system: MacOS 12.5 MP4Box installed from the nightly builds: ``` $ /Applications/GPAC.app/Contents/MacOS/MP4Box -version MP4Box - GPAC version 2.1-DEV-rev259-gfc4749f9-master (c) 2000-2022 Telecom Paris distributed under LGPL v2.1+ - http://gpac.io Please cite our work in your research: GPAC Filters: https://doi.org/10.1145/3339825.3394929 GPAC: https://doi.org/10.1145/1291233.1291452 GPAC Configuration: --extra-cflags=-Wno-deprecated -Wno-shift-negative-value Features: GPAC_CONFIG_DARWIN GPAC_64_BITS GPAC_HAS_IPV6 GPAC_HAS_SSL GPAC_HAS_SOCK_UN GPAC_MINIMAL_ODF GPAC_HAS_QJS GPAC_HAS_FAAD GPAC_HAS_MAD GPAC_HAS_LIBA52 GPAC_HAS_JPEG GPAC_HAS_PNG GPAC_HAS_FFMPEG GPAC_HAS_OPENSVC GPAC_HAS_JP2 GPAC_HAS_OPENHEVC GPAC_HAS_THEORA GPAC_HAS_VORBIS GPAC_HAS_XVID ``` Thanks for reporting your issue. Please make sure these boxes are checked before submitting your issue - thank you! - [x] I looked for a similar issue and couldn't find any. - [x] I tried with the latest version of GPAC. Installers available at http://gpac.io/downloads/gpac-nightly-builds/ - [x] I give enough information for contributors to reproduce my issue (meaningful title, github labels, platform and compiler, command-line ...). I can share files anonymously with this dropbox: https://www.mediafire.com/filedrop/filedrop_hosted.php?drop=eec9e058a9486fe4e99c33021481d9e1826ca9dbc242a6cfaab0fe95da5e5d95 Detailed guidelines: http://gpac.io/2013/07/16/how-to-file-a-bug-properly/ --- end body ---
4,300
[ -0.025181574746966362, 0.010621228255331516, -0.02575715258717537, 0.00438429182395339, 0.021242456510663033, -0.0030262856744229794, -0.016466950997710228, 0.04644201695919037, -0.022717377170920372, 0.01391281932592392, 0.025343455374240875, -0.015765463933348656, 0.0203431136906147, -0....
null
null
null
null
null
null
null
null
null
[ "gpac", "gpac" ]
## Description There is a heap buffer overflow detected by AddressSanitizer ## System info ``` Ubuntu 20.04.2 LTS clang version 12.0.0-++20210402082642+04ba60cfe598-1~exp1~20210402063359.71 MP4Box - GPAC version 1.1.0-DEV-rev1687-g16bd902cb-master ``` ## Build command ``` ./configure --static-mp4box --prefix=`realpath ./install` --enable-sanitizer --cc=clang --cxx=clang++ ``` ## crash command ``` MP4Box -dxml -out /dev/null poc_file ``` ## Pocs [poc.zip](https://github.com/gpac/gpac/files/7976849/poc.zip) ## Crash output ``` ==18134==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x611000000eda at pc 0x000000cdd403 bp 0x7ffc94991ef0 sp 0x7ffc94991ee8 READ of size 1 at 0x611000000eda thread T0 #0 0xcdd402 in gf_inspect_dump_nalu_internal /programs/mp4box/builds/build6/src/filters/inspect.c:1164:24 #1 0xcd734f in gf_inspect_dump_nalu /programs/mp4box/builds/build6/src/filters/inspect.c:1206:2 #2 0x4f45e3 in dump_isom_nal_ex /programs/mp4box/builds/build6/applications/mp4box/filedump.c:1288:5 #3 0x4f8170 in dump_isom_xml /programs/mp4box/builds/build6/applications/mp4box/filedump.c:2001:9 #4 0x4e0da8 in mp4boxMain /programs/mp4box/builds/build6/applications/mp4box/main.c:6171:7 #5 0x7fec5aea30b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x270b2) #6 0x41ea6d in _start (/programs/mp4box/builds/build6/bin/gcc/MP4Box+0x41ea6d) 0x611000000eda is located 0 bytes to the right of 218-byte region [0x611000000e00,0x611000000eda) allocated by thread T0 here: #0 0x499ccd in malloc (/programs/mp4box/builds/build6/bin/gcc/MP4Box+0x499ccd) #1 0x841999 in Media_GetSample /programs/mp4box/builds/build6/src/isomedia/media.c:625:29 #2 0x7d9e90 in gf_isom_get_sample_ex /programs/mp4box/builds/build6/src/isomedia/isom_read.c:1905:6 #3 0x4f40a0 in dump_isom_nal_ex /programs/mp4box/builds/build6/applications/mp4box/filedump.c:1234:24 #4 0x4f8170 in dump_isom_xml /programs/mp4box/builds/build6/applications/mp4box/filedump.c:2001:9 #5 0x4e0da8 in mp4boxMain /programs/mp4box/builds/build6/applications/mp4box/main.c:6171:7 #6 0x7fec5aea30b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x270b2) SUMMARY: AddressSanitizer: heap-buffer-overflow /programs/mp4box/builds/build6/src/filters/inspect.c:1164:24 in gf_inspect_dump_nalu_internal Shadow bytes around the buggy address: 0x0c227fff8180: fd fd fd fd fd fd fd fd fd fd fa fa fa fa fa fa 0x0c227fff8190: fa fa fa fa fa fa fa fa fd fd fd fd fd fd fd fd 0x0c227fff81a0: fd fd fd fd fd fd fd fd fd fd fd fd fd fd fd fd 0x0c227fff81b0: fd fd fd fd fa fa fa fa fa fa fa fa fa fa fa fa 0x0c227fff81c0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 =>0x0c227fff81d0: 00 00 00 00 00 00 00 00 00 00 00[02]fa fa fa fa 0x0c227fff81e0: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c227fff81f0: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c227fff8200: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c227fff8210: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c227fff8220: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa Shadow byte legend (one shadow byte represents 8 application bytes): Addressable: 00 Partially addressable: 01 02 03 04 05 06 07 Heap left redzone: fa Freed heap region: fd Stack left redzone: f1 Stack mid redzone: f2 Stack right redzone: f3 Stack after return: f5 Stack use after scope: f8 Global redzone: f9 Global init order: f6 Poisoned by user: f7 Container overflow: fc Array cookie: ac Intra object redzone: bb ASan internal: fe Left alloca redzone: ca Right alloca redzone: cb Shadow gap: cc ==18134==ABORTING /collect/collec.sh: line 18: 18134 Aborted (core dumped) ------------------------ ```
There is a heap buffer overflow detected by AddressSanitizer
https://api.github.com/repos/gpac/gpac/issues/2087/comments
0
2022-02-01T08:18:23
2022-02-01T14:29:19Z
https://github.com/gpac/gpac/issues/2087
1,120,360,448
2,087
false
This is a GitHub Issue repo:gpac owner:gpac Title : There is a heap buffer overflow detected by AddressSanitizer Issue date: --- start body --- ## Description There is a heap buffer overflow detected by AddressSanitizer ## System info ``` Ubuntu 20.04.2 LTS clang version 12.0.0-++20210402082642+04ba60cfe598-1~exp1~20210402063359.71 MP4Box - GPAC version 1.1.0-DEV-rev1687-g16bd902cb-master ``` ## Build command ``` ./configure --static-mp4box --prefix=`realpath ./install` --enable-sanitizer --cc=clang --cxx=clang++ ``` ## crash command ``` MP4Box -dxml -out /dev/null poc_file ``` ## Pocs [poc.zip](https://github.com/gpac/gpac/files/7976849/poc.zip) ## Crash output ``` ==18134==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x611000000eda at pc 0x000000cdd403 bp 0x7ffc94991ef0 sp 0x7ffc94991ee8 READ of size 1 at 0x611000000eda thread T0 #0 0xcdd402 in gf_inspect_dump_nalu_internal /programs/mp4box/builds/build6/src/filters/inspect.c:1164:24 #1 0xcd734f in gf_inspect_dump_nalu /programs/mp4box/builds/build6/src/filters/inspect.c:1206:2 #2 0x4f45e3 in dump_isom_nal_ex /programs/mp4box/builds/build6/applications/mp4box/filedump.c:1288:5 #3 0x4f8170 in dump_isom_xml /programs/mp4box/builds/build6/applications/mp4box/filedump.c:2001:9 #4 0x4e0da8 in mp4boxMain /programs/mp4box/builds/build6/applications/mp4box/main.c:6171:7 #5 0x7fec5aea30b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x270b2) #6 0x41ea6d in _start (/programs/mp4box/builds/build6/bin/gcc/MP4Box+0x41ea6d) 0x611000000eda is located 0 bytes to the right of 218-byte region [0x611000000e00,0x611000000eda) allocated by thread T0 here: #0 0x499ccd in malloc (/programs/mp4box/builds/build6/bin/gcc/MP4Box+0x499ccd) #1 0x841999 in Media_GetSample /programs/mp4box/builds/build6/src/isomedia/media.c:625:29 #2 0x7d9e90 in gf_isom_get_sample_ex /programs/mp4box/builds/build6/src/isomedia/isom_read.c:1905:6 #3 0x4f40a0 in dump_isom_nal_ex /programs/mp4box/builds/build6/applications/mp4box/filedump.c:1234:24 #4 0x4f8170 in dump_isom_xml /programs/mp4box/builds/build6/applications/mp4box/filedump.c:2001:9 #5 0x4e0da8 in mp4boxMain /programs/mp4box/builds/build6/applications/mp4box/main.c:6171:7 #6 0x7fec5aea30b2 in __libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x270b2) SUMMARY: AddressSanitizer: heap-buffer-overflow /programs/mp4box/builds/build6/src/filters/inspect.c:1164:24 in gf_inspect_dump_nalu_internal Shadow bytes around the buggy address: 0x0c227fff8180: fd fd fd fd fd fd fd fd fd fd fa fa fa fa fa fa 0x0c227fff8190: fa fa fa fa fa fa fa fa fd fd fd fd fd fd fd fd 0x0c227fff81a0: fd fd fd fd fd fd fd fd fd fd fd fd fd fd fd fd 0x0c227fff81b0: fd fd fd fd fa fa fa fa fa fa fa fa fa fa fa fa 0x0c227fff81c0: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 =>0x0c227fff81d0: 00 00 00 00 00 00 00 00 00 00 00[02]fa fa fa fa 0x0c227fff81e0: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c227fff81f0: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c227fff8200: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c227fff8210: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x0c227fff8220: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa Shadow byte legend (one shadow byte represents 8 application bytes): Addressable: 00 Partially addressable: 01 02 03 04 05 06 07 Heap left redzone: fa Freed heap region: fd Stack left redzone: f1 Stack mid redzone: f2 Stack right redzone: f3 Stack after return: f5 Stack use after scope: f8 Global redzone: f9 Global init order: f6 Poisoned by user: f7 Container overflow: fc Array cookie: ac Intra object redzone: bb ASan internal: fe Left alloca redzone: ca Right alloca redzone: cb Shadow gap: cc ==18134==ABORTING /collect/collec.sh: line 18: 18134 Aborted (core dumped) ------------------------ ``` --- end body ---
4,108
[ -0.04576938599348068, 0.01748175546526909, -0.015331896021962166, 0.013818507082760334, 0.03940466791391373, 0.011909092776477337, -0.03134269267320633, 0.03586871549487114, -0.004264360293745995, 0.023917190730571747, -0.02241794764995575, 0.011505994014441967, 0.020890414714813232, 0.011...
null
null
null
null
null
null
null
null
null
[ "MonetDB", "MonetDB" ]
**Is your feature request related to a problem? Please describe.** When you design a dashboard in a Bi Tool (like Tableau), you don't write the exact query of the viz. You create a model with one or several tables and then you build your viz on this model (which must be at the finest granularity needed in the dashboard). However, most of the time, the granularity of your viz is different, meaning there is a lot of grouping at distinct level. **Describe the solution you'd like** Vertica has a unique way to solve this issue by building projections. To make it simple, projections are stored "synonyms" of the table that is queried, it can contains less columns, being filtered, or aggregated (so less data and faster query) and the engine choose the best fitted projection for any query. https://www.vertica.com/docs/9.2.x/HTML/Content/Authoring/AdministratorsGuide/Projections/WorkingWithProjections.htm https://www.vertica.com/docs/9.3.x/HTML/Content/Authoring/SQLReferenceManual/Statements/CREATEPROJECTIONLiveAggregate.htm Of course, it takes more space disk but it's way better for dashboarding. **Describe alternatives you've considered** N/A **Additional context** N/A
Projection (like Vertica) support
https://api.github.com/repos/MonetDB/MonetDB/issues/7491/comments
12
2024-03-27T05:47:41
2024-03-28T20:16:38Z
https://github.com/MonetDB/MonetDB/issues/7491
2,209,921,389
7,491
false
This is a GitHub Issue repo:MonetDB owner:MonetDB Title : Projection (like Vertica) support Issue date: --- start body --- **Is your feature request related to a problem? Please describe.** When you design a dashboard in a Bi Tool (like Tableau), you don't write the exact query of the viz. You create a model with one or several tables and then you build your viz on this model (which must be at the finest granularity needed in the dashboard). However, most of the time, the granularity of your viz is different, meaning there is a lot of grouping at distinct level. **Describe the solution you'd like** Vertica has a unique way to solve this issue by building projections. To make it simple, projections are stored "synonyms" of the table that is queried, it can contains less columns, being filtered, or aggregated (so less data and faster query) and the engine choose the best fitted projection for any query. https://www.vertica.com/docs/9.2.x/HTML/Content/Authoring/AdministratorsGuide/Projections/WorkingWithProjections.htm https://www.vertica.com/docs/9.3.x/HTML/Content/Authoring/SQLReferenceManual/Statements/CREATEPROJECTIONLiveAggregate.htm Of course, it takes more space disk but it's way better for dashboarding. **Describe alternatives you've considered** N/A **Additional context** N/A --- end body ---
1,340
[ -0.021450776606798172, -0.015227495692670345, -0.030966082587838173, 0.01943647861480713, -0.011687441729009151, 0.01576865091919899, 0.0050958748906850815, 0.036858655512332916, -0.036648206412792206, 0.03487442433834076, 0.025539502501487732, -0.027147933840751648, 0.03062034398317337, -...
CVE-2023-42278
2023-09-08T22:15:12.037000
hutool v5.8.21 was discovered to contain a buffer overflow via the component JSONUtil.parse().
{ "cvssMetricV2": null, "cvssMetricV30": null, "cvssMetricV31": [ { "cvssData": { "attackComplexity": "LOW", "attackVector": "NETWORK", "availabilityImpact": "HIGH", "baseScore": 7.5, "baseSeverity": "HIGH", "confidentialityImpact": "NONE", "integrityImpact": "NONE", "privilegesRequired": "NONE", "scope": "UNCHANGED", "userInteraction": "NONE", "vectorString": "CVSS:3.1/AV:N/AC:L/PR:N/UI:N/S:U/C:N/I:N/A:H", "version": "3.1" }, "exploitabilityScore": 3.9, "impactScore": 3.6, "source": "nvd@nist.gov", "type": "Primary" } ] }
[ { "source": "cve@mitre.org", "tags": [ "Exploit", "Issue Tracking", "Vendor Advisory" ], "url": "https://github.com/dromara/hutool/issues/3289" } ]
[ { "nodes": [ { "cpeMatch": [ { "criteria": "cpe:2.3:a:hutool:hutool:5.8.21:*:*:*:*:*:*:*", "matchCriteriaId": "EBFA93D9-F18F-4939-BF26-89603CD7148D", "versionEndExcluding": null, "versionEndIncluding": null, "versionStartExcludi...
https://github.com/dromara/hutool/issues/3289
[ "Exploit", "Issue Tracking", "Vendor Advisory" ]
github.com
[ "dromara", "hutool" ]
### 版本情况 JDK版本: 1.8.0_362 hutool版本: 5.8.21 ### 问题描述(包括截图) 1. 复现代码 ```java import cn.hutool.json.JSONUtil; public class JSONOUtilTest { public static void main(String[] args) { String s = "{\"G\":00,[,,[0E5,6E9,6E5,6E9,6E5,6E9,6E5,6E9,6E9,6E5,true,6E5,6E9,6E5,6E9,6956,EE,5E9,6E5,RE,6E9,6E9,6E5,6E9,6E5,6E9,6E5,6E9,6E5,6E962756779,4141697],]}"; new JSONOUtilTest().testHutoolJSON(s); new JSONOUtilTest().testGson(s); } // hutool-json测试 public void testHutoolJSON (String str) { JSON json = JSONUtil.parse(str); } // gson测试 public void testGson (String str) { Gson gson = new GsonBuilder().setPrettyPrinting().create(); JSONObject entries = gson.fromJson(str, JSONObject.class); System.out.println(entries.toStringPretty()); } } ``` 2. 堆栈信息 我们的服务使用了hutool-json来将字符串解析为json,当我们接收到的字符串为`"{\"G\":00,[,,[0E5,6E9,6E5,6E9,6E5,6E9,6E5,6E9,6E9,6E5,true,6E5,6E9,6E5,6E9,6956,EE,5E9,6E5,RE,6E9,6E9,6E5,6E9,6E5,6E9,6E5,6E9,6E5,6E962756779,4141697],]}"`时,`JSONUtil.parse()`挂起一段时间,然后服务崩溃,报错: ``` Exception in thread "main" java.lang.OutOfMemoryError: Java heap space at java.util.Arrays.copyOf(Arrays.java:3332) at java.lang.AbstractStringBuilder.ensureCapacityInternal(AbstractStringBuilder.java:124) at java.lang.AbstractStringBuilder.append(AbstractStringBuilder.java:649) at java.lang.StringBuffer.append(StringBuffer.java:387) at java.io.StringWriter.write(StringWriter.java:77) at cn.hutool.json.serialize.JSONWriter.writeRaw(JSONWriter.java:392) at cn.hutool.json.serialize.JSONWriter.writeValueDirect(JSONWriter.java:231) at cn.hutool.json.serialize.JSONWriter.writeField(JSONWriter.java:196) at cn.hutool.json.JSONArray.lambda$write$2cc9e97d$1(JSONArray.java:561) at cn.hutool.json.JSONArray$$Lambda$6/379110473.accept(Unknown Source) at cn.hutool.core.collection.CollUtil.forEach(CollUtil.java:2690) at cn.hutool.core.collection.CollUtil.forEach(CollUtil.java:2674) at cn.hutool.json.JSONArray.write(JSONArray.java:561) at cn.hutool.json.serialize.JSONWriter.writeObjValue(JSONWriter.java:257) at cn.hutool.json.serialize.JSONWriter.writeValueDirect(JSONWriter.java:239) at cn.hutool.json.serialize.JSONWriter.writeField(JSONWriter.java:196) at cn.hutool.json.JSONArray.lambda$write$2cc9e97d$1(JSONArray.java:561) at cn.hutool.json.JSONArray$$Lambda$6/379110473.accept(Unknown Source) at cn.hutool.core.collection.CollUtil.forEach(CollUtil.java:2690) at cn.hutool.core.collection.CollUtil.forEach(CollUtil.java:2674) at cn.hutool.json.JSONArray.write(JSONArray.java:561) at cn.hutool.json.JSONArray.write(JSONArray.java:543) at cn.hutool.json.JSON.toJSONString(JSON.java:120) at cn.hutool.json.JSONArray.toString(JSONArray.java:522) at cn.hutool.json.JSONParser.parseTo(JSONParser.java:69) at cn.hutool.json.ObjectMapper.mapFromTokener(ObjectMapper.java:243) at cn.hutool.json.ObjectMapper.mapFromStr(ObjectMapper.java:219) at cn.hutool.json.ObjectMapper.map(ObjectMapper.java:98) at cn.hutool.json.JSONObject.<init>(JSONObject.java:210) at cn.hutool.json.JSONObject.<init>(JSONObject.java:187) at cn.hutool.json.JSONUtil.parseObj(JSONUtil.java:112) at cn.hutool.json.JSONUtil.parse(JSONUtil.java:227) ``` 经测试,我们发现,如果使用gson解析上述字符串,gson能够捕获异常: ``` Exception in thread "main" com.google.gson.JsonSyntaxException: duplicate key: G at com.google.gson.internal.bind.MapTypeAdapterFactory$Adapter.read(MapTypeAdapterFactory.java:190) at com.google.gson.internal.bind.MapTypeAdapterFactory$Adapter.read(MapTypeAdapterFactory.java:145) at com.google.gson.Gson.fromJson(Gson.java:932) at com.google.gson.Gson.fromJson(Gson.java:897) at com.google.gson.Gson.fromJson(Gson.java:846) at com.google.gson.Gson.fromJson(Gson.java:817) at JSONOUtilTest.testGson(JSONOUtilTest.java:43) at JSONOUtilTest.main(JSONOUtilTest.java:54) ``` 3. 测试涉及到的文件(注意脱密) 见复现代码。 4. 分析 通过与gson运行结果的对比,`JSONUtil.parse()`方法解析特定输入时,会导致服务挂起和崩溃,存在安全隐患。
`JSONUtil.parse()`方法解析特定输入时,会导致服务挂起和崩溃,存在安全隐患
https://api.github.com/repos/dromara/hutool/issues/3289/comments
5
2023-09-04T02:58:25
2023-09-14T07:12:52Z
https://github.com/dromara/hutool/issues/3289
1,879,420,551
3,289
true
This is a GitHub Issue repo:hutool owner:dromara Title : `JSONUtil.parse()`方法解析特定输入时,会导致服务挂起和崩溃,存在安全隐患 Issue date: --- start body --- ### 版本情况 JDK版本: 1.8.0_362 hutool版本: 5.8.21 ### 问题描述(包括截图) 1. 复现代码 ```java import cn.hutool.json.JSONUtil; public class JSONOUtilTest { public static void main(String[] args) { String s = "{\"G\":00,[,,[0E5,6E9,6E5,6E9,6E5,6E9,6E5,6E9,6E9,6E5,true,6E5,6E9,6E5,6E9,6956,EE,5E9,6E5,RE,6E9,6E9,6E5,6E9,6E5,6E9,6E5,6E9,6E5,6E962756779,4141697],]}"; new JSONOUtilTest().testHutoolJSON(s); new JSONOUtilTest().testGson(s); } // hutool-json测试 public void testHutoolJSON (String str) { JSON json = JSONUtil.parse(str); } // gson测试 public void testGson (String str) { Gson gson = new GsonBuilder().setPrettyPrinting().create(); JSONObject entries = gson.fromJson(str, JSONObject.class); System.out.println(entries.toStringPretty()); } } ``` 2. 堆栈信息 我们的服务使用了hutool-json来将字符串解析为json,当我们接收到的字符串为`"{\"G\":00,[,,[0E5,6E9,6E5,6E9,6E5,6E9,6E5,6E9,6E9,6E5,true,6E5,6E9,6E5,6E9,6956,EE,5E9,6E5,RE,6E9,6E9,6E5,6E9,6E5,6E9,6E5,6E9,6E5,6E962756779,4141697],]}"`时,`JSONUtil.parse()`挂起一段时间,然后服务崩溃,报错: ``` Exception in thread "main" java.lang.OutOfMemoryError: Java heap space at java.util.Arrays.copyOf(Arrays.java:3332) at java.lang.AbstractStringBuilder.ensureCapacityInternal(AbstractStringBuilder.java:124) at java.lang.AbstractStringBuilder.append(AbstractStringBuilder.java:649) at java.lang.StringBuffer.append(StringBuffer.java:387) at java.io.StringWriter.write(StringWriter.java:77) at cn.hutool.json.serialize.JSONWriter.writeRaw(JSONWriter.java:392) at cn.hutool.json.serialize.JSONWriter.writeValueDirect(JSONWriter.java:231) at cn.hutool.json.serialize.JSONWriter.writeField(JSONWriter.java:196) at cn.hutool.json.JSONArray.lambda$write$2cc9e97d$1(JSONArray.java:561) at cn.hutool.json.JSONArray$$Lambda$6/379110473.accept(Unknown Source) at cn.hutool.core.collection.CollUtil.forEach(CollUtil.java:2690) at cn.hutool.core.collection.CollUtil.forEach(CollUtil.java:2674) at cn.hutool.json.JSONArray.write(JSONArray.java:561) at cn.hutool.json.serialize.JSONWriter.writeObjValue(JSONWriter.java:257) at cn.hutool.json.serialize.JSONWriter.writeValueDirect(JSONWriter.java:239) at cn.hutool.json.serialize.JSONWriter.writeField(JSONWriter.java:196) at cn.hutool.json.JSONArray.lambda$write$2cc9e97d$1(JSONArray.java:561) at cn.hutool.json.JSONArray$$Lambda$6/379110473.accept(Unknown Source) at cn.hutool.core.collection.CollUtil.forEach(CollUtil.java:2690) at cn.hutool.core.collection.CollUtil.forEach(CollUtil.java:2674) at cn.hutool.json.JSONArray.write(JSONArray.java:561) at cn.hutool.json.JSONArray.write(JSONArray.java:543) at cn.hutool.json.JSON.toJSONString(JSON.java:120) at cn.hutool.json.JSONArray.toString(JSONArray.java:522) at cn.hutool.json.JSONParser.parseTo(JSONParser.java:69) at cn.hutool.json.ObjectMapper.mapFromTokener(ObjectMapper.java:243) at cn.hutool.json.ObjectMapper.mapFromStr(ObjectMapper.java:219) at cn.hutool.json.ObjectMapper.map(ObjectMapper.java:98) at cn.hutool.json.JSONObject.<init>(JSONObject.java:210) at cn.hutool.json.JSONObject.<init>(JSONObject.java:187) at cn.hutool.json.JSONUtil.parseObj(JSONUtil.java:112) at cn.hutool.json.JSONUtil.parse(JSONUtil.java:227) ``` 经测试,我们发现,如果使用gson解析上述字符串,gson能够捕获异常: ``` Exception in thread "main" com.google.gson.JsonSyntaxException: duplicate key: G at com.google.gson.internal.bind.MapTypeAdapterFactory$Adapter.read(MapTypeAdapterFactory.java:190) at com.google.gson.internal.bind.MapTypeAdapterFactory$Adapter.read(MapTypeAdapterFactory.java:145) at com.google.gson.Gson.fromJson(Gson.java:932) at com.google.gson.Gson.fromJson(Gson.java:897) at com.google.gson.Gson.fromJson(Gson.java:846) at com.google.gson.Gson.fromJson(Gson.java:817) at JSONOUtilTest.testGson(JSONOUtilTest.java:43) at JSONOUtilTest.main(JSONOUtilTest.java:54) ``` 3. 测试涉及到的文件(注意脱密) 见复现代码。 4. 分析 通过与gson运行结果的对比,`JSONUtil.parse()`方法解析特定输入时,会导致服务挂起和崩溃,存在安全隐患。 --- end body ---
4,236
[ 0.00036030402407050133, 0.010487066581845284, -0.0013552720192819834, 0.016395486891269684, 0.043459370732307434, 0.0030052096117287874, -0.01480882614850998, 0.03357674553990364, -0.011469285003840923, -0.0008339409250766039, 0.01338083203881979, 0.026701221242547035, 0.02558300271630287, ...
null
null
null
null
null
null
null
null
null
[ "ImageMagick", "ImageMagick" ]
### ImageMagick version 7.1.1-27 ### Operating system Windows ### Operating system, version and so on Windows 11 ### Description When I use ImageMagick to convert tiff to tiff attached file, the output file is rotated. I didn't expect it. With ImageMagick version 7.0.10-35, this doesn't happen. Please help me investigate and explanation for it. Thank you. [sample.zip](https://github.com/ImageMagick/ImageMagick/files/14767669/sample.zip) ### Steps to Reproduce Use `convert.exe <input file> <output file> ` command line. ### Images _No response_
TIFF image is rotated 180 degree
https://api.github.com/repos/ImageMagick/ImageMagick/issues/7199/comments
3
2024-03-27T04:41:32
2024-04-06T12:25:13Z
https://github.com/ImageMagick/ImageMagick/issues/7199
2,209,839,333
7,199
false
This is a GitHub Issue repo:ImageMagick owner:ImageMagick Title : TIFF image is rotated 180 degree Issue date: --- start body --- ### ImageMagick version 7.1.1-27 ### Operating system Windows ### Operating system, version and so on Windows 11 ### Description When I use ImageMagick to convert tiff to tiff attached file, the output file is rotated. I didn't expect it. With ImageMagick version 7.0.10-35, this doesn't happen. Please help me investigate and explanation for it. Thank you. [sample.zip](https://github.com/ImageMagick/ImageMagick/files/14767669/sample.zip) ### Steps to Reproduce Use `convert.exe <input file> <output file> ` command line. ### Images _No response_ --- end body ---
713
[ 0.002086521126329899, -0.006090138573199511, -0.0005334595334716141, 0.021820707246661186, -0.0035899763461202383, 0.039538584649562836, -0.006886893417686224, 0.050113141536712646, -0.014286641031503677, 0.029354773461818695, 0.0034495522268116474, -0.04075966402888298, -0.00672204792499542...