*.7z filter=lfs diff=lfs merge=lfs -text *.arrow filter=lfs diff=lfs merge=lfs -text *.bin filter=lfs diff=lfs merge=lfs -text *.bz2 filter=lfs diff=lfs merge=lfs -text *.ckpt filter=lfs diff=lfs merge=lfs -text *.ftz filter=lfs diff=lfs merge=lfs -text *.gz filter=lfs diff=lfs merge=lfs -text *.h5 filter=lfs diff=lfs merge=lfs -text *.joblib filter=lfs diff=lfs merge=lfs -text *.lfs.* filter=lfs diff=lfs merge=lfs -text *.lz4 filter=lfs diff=lfs merge=lfs -text *.mds filter=lfs diff=lfs merge=lfs -text *.mlmodel filter=lfs diff=lfs merge=lfs -text *.model filter=lfs diff=lfs merge=lfs -text *.msgpack filter=lfs diff=lfs merge=lfs -text *.npy filter=lfs diff=lfs merge=lfs -text *.npz filter=lfs diff=lfs merge=lfs -text *.onnx filter=lfs diff=lfs merge=lfs -text *.ot filter=lfs diff=lfs merge=lfs -text *.parquet filter=lfs diff=lfs merge=lfs -text *.pb filter=lfs diff=lfs merge=lfs -text *.pickle filter=lfs diff=lfs merge=lfs -text *.pkl filter=lfs diff=lfs merge=lfs -text *.pt filter=lfs diff=lfs merge=lfs -text *.pth filter=lfs diff=lfs merge=lfs -text *.rar filter=lfs diff=lfs merge=lfs -text *.safetensors filter=lfs diff=lfs merge=lfs -text saved_model/**/* filter=lfs diff=lfs merge=lfs -text *.tar.* filter=lfs diff=lfs merge=lfs -text *.tar filter=lfs diff=lfs merge=lfs -text *.tflite filter=lfs diff=lfs merge=lfs -text *.tgz filter=lfs diff=lfs merge=lfs -text *.wasm filter=lfs diff=lfs merge=lfs -text *.xz filter=lfs diff=lfs merge=lfs -text *.zip filter=lfs diff=lfs merge=lfs -text *.zst filter=lfs diff=lfs merge=lfs -text *tfevents* filter=lfs diff=lfs merge=lfs -text # Audio files - uncompressed *.pcm filter=lfs diff=lfs merge=lfs -text *.sam filter=lfs diff=lfs merge=lfs -text *.raw filter=lfs diff=lfs merge=lfs -text # Audio files - compressed *.aac filter=lfs diff=lfs merge=lfs -text *.flac filter=lfs diff=lfs merge=lfs -text *.mp3 filter=lfs diff=lfs merge=lfs -text *.ogg filter=lfs diff=lfs merge=lfs -text *.wav filter=lfs diff=lfs merge=lfs -text # Image files - uncompressed *.bmp filter=lfs diff=lfs merge=lfs -text *.gif filter=lfs diff=lfs merge=lfs -text *.png filter=lfs diff=lfs merge=lfs -text *.tiff filter=lfs diff=lfs merge=lfs -text # Image files - compressed *.jpg filter=lfs diff=lfs merge=lfs -text *.jpeg filter=lfs diff=lfs merge=lfs -text *.webp filter=lfs diff=lfs merge=lfs -text # Video files - compressed *.mp4 filter=lfs diff=lfs merge=lfs -text *.webm filter=lfs diff=lfs merge=lfs -text 2025/2DMamba_[[:space:]]Efficient[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]for[[:space:]]Image[[:space:]]Representation[[:space:]]with[[:space:]]Applications[[:space:]]on[[:space:]]Giga-Pixel[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Classification/3dcdbbd2-63a3-41ca-bed4-b0071bd9d91f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D[[:space:]]Convex[[:space:]]Splatting_[[:space:]]Radiance[[:space:]]Field[[:space:]]Rendering[[:space:]]with[[:space:]]3D[[:space:]]Smooth[[:space:]]Convexes/f5ed5ddb-3e23-43ba-b2ea-85540f2f015f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D[[:space:]]Dental[[:space:]]Model[[:space:]]Segmentation[[:space:]]with[[:space:]]Geometrical[[:space:]]Boundary[[:space:]]Preserving/532cf63b-00fc-4ff9-ad5b-7735e7795d09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D[[:space:]]Gaussian[[:space:]]Head[[:space:]]Avatars[[:space:]]with[[:space:]]Expressive[[:space:]]Dynamic[[:space:]]Appearances[[:space:]]by[[:space:]]Compact[[:space:]]Tensorial[[:space:]]Representations/627db675-5c1d-42be-a429-641d6eb44826_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D[[:space:]]Gaussian[[:space:]]Inpainting[[:space:]]with[[:space:]]Depth-Guided[[:space:]]Cross-View[[:space:]]Consistency/e8b63c5d-11f1-40a3-993f-7cfad108cb38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D[[:space:]]Occupancy[[:space:]]Prediction[[:space:]]with[[:space:]]Low-Resolution[[:space:]]Queries[[:space:]]via[[:space:]]Prototype-aware[[:space:]]View[[:space:]]Transformation/47811d03-929d-4f21-805c-f1072e579c38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D[[:space:]]Prior[[:space:]]Is[[:space:]]All[[:space:]]You[[:space:]]Need_[[:space:]]Cross-Task[[:space:]]Few-shot[[:space:]]2D[[:space:]]Gaze[[:space:]]Estimation/768d9909-7302-4b31-81de-9e15acdeb35e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D[[:space:]]Student[[:space:]]Splatting[[:space:]]and[[:space:]]Scooping/5f61e07c-5e93-420e-a919-2f54a9dc803a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D-AVS_[[:space:]]LiDAR-based[[:space:]]3D[[:space:]]Auto-Vocabulary[[:space:]]Segmentation/17e84a16-1ce0-4997-bd4d-dc9b8b0e89ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D-GRAND_[[:space:]]A[[:space:]]Million-Scale[[:space:]]Dataset[[:space:]]for[[:space:]]3D-LLMs[[:space:]]with[[:space:]]Better[[:space:]]Grounding[[:space:]]and[[:space:]]Less[[:space:]]Hallucination/85529eff-9044-4d08-8d4c-7fd264bba320_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D-GSW_[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Robust[[:space:]]Watermarking/c8fc89b6-7a4d-4610-af82-1a446e24a6b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D-HGS_[[:space:]]3D[[:space:]]Half-Gaussian[[:space:]]Splatting/756891c7-db71-4b8d-abcc-7c843cd22944_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D-LLaVA_[[:space:]]Towards[[:space:]]Generalist[[:space:]]3D[[:space:]]LMMs[[:space:]]with[[:space:]]Omni[[:space:]]Superpoint[[:space:]]Transformer/39b23d12-60ca-4604-960c-965e629e5821_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D-MVP_[[:space:]]3D[[:space:]]Multiview[[:space:]]Pretraining[[:space:]]for[[:space:]]Manipulation/1e7b2619-295e-41d1-97fc-d9acf2fa341e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D-Mem_[[:space:]]3D[[:space:]]Scene[[:space:]]Memory[[:space:]]for[[:space:]]Embodied[[:space:]]Exploration[[:space:]]and[[:space:]]Reasoning/d0002279-be38-4de0-a167-db2dd5083a0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D-SLNR_[[:space:]]A[[:space:]]Super[[:space:]]Lightweight[[:space:]]Neural[[:space:]]Representation[[:space:]]for[[:space:]]Large-scale[[:space:]]3D[[:space:]]Mapping/4284846f-ef79-4d62-9c0f-2fa490a19f60_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3DEnhancer_[[:space:]]Consistent[[:space:]]Multi-View[[:space:]]Diffusion[[:space:]]for[[:space:]]3D[[:space:]]Enhancement/3f4f27a6-0bf6-4821-b4d9-5b5287d3f5c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3DGUT_[[:space:]]Enabling[[:space:]]Distorted[[:space:]]Cameras[[:space:]]and[[:space:]]Secondary[[:space:]]Rays[[:space:]]in[[:space:]]Gaussian[[:space:]]Splatting/4d137404-b3cb-4e57-a972-c50723461082_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3DTopia-XL_[[:space:]]Scaling[[:space:]]High-quality[[:space:]]3D[[:space:]]Asset[[:space:]]Generation[[:space:]]via[[:space:]]Primitive[[:space:]]Diffusion/75fbc8e4-4797-48af-9122-0c9fec5fabaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/4D[[:space:]]LangSplat_[[:space:]]4D[[:space:]]Language[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]via[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/402df206-80ec-4c4d-b554-9a26536c90ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/4D-Fly_[[:space:]]Fast[[:space:]]4D[[:space:]]Reconstruction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Monocular[[:space:]]Video/bec555e7-c847-4ed7-801a-a559d1ce7b48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/4DGC_[[:space:]]Rate-Aware[[:space:]]4D[[:space:]]Gaussian[[:space:]]Compression[[:space:]]for[[:space:]]Efficient[[:space:]]Streamable[[:space:]]Free-Viewpoint[[:space:]]Video/d2b94c9c-2214-43ee-865b-39c67991e81c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/4DTAM_[[:space:]]Non-Rigid[[:space:]]Tracking[[:space:]]and[[:space:]]Mapping[[:space:]]via[[:space:]]Dynamic[[:space:]]Surface[[:space:]]Gaussians/4cb7c164-1232-44e5-9451-158b5d7048fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/4Deform_[[:space:]]Neural[[:space:]]Surface[[:space:]]Deformation[[:space:]]for[[:space:]]Robust[[:space:]]Shape[[:space:]]Interpolation/a98600ea-d152-4146-b6fe-1c8c3f931210_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/4Real-Video_[[:space:]]Learning[[:space:]]Generalizable[[:space:]]Photo-Realistic[[:space:]]4D[[:space:]]Video[[:space:]]Diffusion/0cf19282-fa7f-4680-bc87-175d439267a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/5%_100%_[[:space:]]Breaking[[:space:]]Performance[[:space:]]Shackles[[:space:]]of[[:space:]]Full[[:space:]]Fine-Tuning[[:space:]]on[[:space:]]Visual[[:space:]]Recognition[[:space:]]Tasks/1a75cdda-a594-42aa-b287-0cf484866607_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Bias-Free[[:space:]]Training[[:space:]]Paradigm[[:space:]]for[[:space:]]More[[:space:]]General[[:space:]]AI-generated[[:space:]]Image[[:space:]]Detection/323dbdcc-b693-4070-92b9-d9dcaf560a4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Closer[[:space:]]Look[[:space:]]at[[:space:]]Time[[:space:]]Steps[[:space:]]is[[:space:]]Worthy[[:space:]]of[[:space:]]Triple[[:space:]]Speed-Up[[:space:]]for[[:space:]]Diffusion[[:space:]]Model[[:space:]]Training/cce7ea3d-6315-432c-a78d-5b778092ad35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Comprehensive[[:space:]]Study[[:space:]]of[[:space:]]Decoder-Only[[:space:]]LLMs[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/f029460e-1f93-4e71-9a1a-d734965310bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Data-Centric[[:space:]]Revisit[[:space:]]of[[:space:]]Pre-Trained[[:space:]]Vision[[:space:]]Models[[:space:]]for[[:space:]]Robot[[:space:]]Learning/00996750-7955-4b08-8199-1c2e8100b73e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Dataset[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]in[[:space:]]the[[:space:]]Presence[[:space:]]of[[:space:]]Unknowns/23aebd3c-4a4f-4148-b060-1061f0f8c69a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Distractor-Aware[[:space:]]Memory[[:space:]]for[[:space:]]Visual[[:space:]]Object[[:space:]]Tracking[[:space:]]with[[:space:]]SAM2/4e976258-ab89-4b67-a5da-a2728d6871ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Flag[[:space:]]Decomposition[[:space:]]for[[:space:]]Hierarchical[[:space:]]Datasets/df09f1c7-a9b0-4c23-b900-ec225450f19c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Focused[[:space:]]Human[[:space:]]Body[[:space:]]Model[[:space:]]for[[:space:]]Accurate[[:space:]]Anthropometric[[:space:]]Measurements[[:space:]]Extraction/468856f8-8572-45cb-8162-0cfa7d6d7227_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]General[[:space:]]Adaptive[[:space:]]Dual-level[[:space:]]Weighting[[:space:]]Mechanism[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Pansharpening/b84d8fe3-e105-4d16-bcf4-2ecaed6a1524_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Hubness[[:space:]]Perspective[[:space:]]on[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Graph-Based[[:space:]]Multi-View[[:space:]]Clustering/895e46af-6cc5-4cc4-b2d9-ddcd4e7c57f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Lightweight[[:space:]]UDF[[:space:]]Learning[[:space:]]Framework[[:space:]]for[[:space:]]3D[[:space:]]Reconstruction[[:space:]]Based[[:space:]]on[[:space:]]Local[[:space:]]Shape[[:space:]]Functions/3ec10370-a835-4f8b-9dd6-310bcae1dc70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]New[[:space:]]Statistical[[:space:]]Model[[:space:]]of[[:space:]]Star[[:space:]]Speckles[[:space:]]for[[:space:]]Learning[[:space:]]to[[:space:]]Detect[[:space:]]and[[:space:]]Characterize[[:space:]]Exoplanets[[:space:]]in[[:space:]]Direct[[:space:]]Imaging[[:space:]]Observations/329b88c6-b793-419e-a0b2-e0ed7257493a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Physics-Informed[[:space:]]Blur[[:space:]]Learning[[:space:]]Framework[[:space:]]for[[:space:]]Imaging[[:space:]]Systems/a0d4bfdc-a5ad-4533-a724-3bee0f7d0259_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Polarization-Aided[[:space:]]Transformer[[:space:]]for[[:space:]]Image[[:space:]]Deblurring[[:space:]]via[[:space:]]Motion[[:space:]]Vector[[:space:]]Decomposition/d1140f05-15bd-4a94-91b7-a27e1828a326_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Regularization-Guided[[:space:]]Equivariant[[:space:]]Approach[[:space:]]for[[:space:]]Image[[:space:]]Restoration/9db9042c-ddce-4703-8663-1a5f43849d91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Selective[[:space:]]Re-learning[[:space:]]Mechanism[[:space:]]for[[:space:]]Hyperspectral[[:space:]]Fusion[[:space:]]Imaging/6a213956-431d-4057-8ec5-43422f931878_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Semantic[[:space:]]Knowledge[[:space:]]Complementarity[[:space:]]based[[:space:]]Decoupling[[:space:]]Framework[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Class-imbalanced[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/a07e730c-be1d-4068-a8a4-93e06f295be0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Simple[[:space:]]Data[[:space:]]Augmentation[[:space:]]for[[:space:]]Feature[[:space:]]Distribution[[:space:]]Skewed[[:space:]]Federated[[:space:]]Learning/e0ad170f-ffdd-4ee6-8dbc-8fd4c6ea1cbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Simple[[:space:]]yet[[:space:]]Effective[[:space:]]Layout[[:space:]]Token[[:space:]]in[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Document[[:space:]]Understanding/5f860180-3af4-4fff-abf7-4bbdf8e78ae2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Stitch[[:space:]]in[[:space:]]Time[[:space:]]Saves[[:space:]]Nine_[[:space:]]Small[[:space:]]VLM[[:space:]]is[[:space:]]a[[:space:]]Precise[[:space:]]Guidance[[:space:]]for[[:space:]]Accelerating[[:space:]]Large[[:space:]]VLMs/81ce4c52-dac7-4add-aaa7-e959289e0870_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Tale[[:space:]]of[[:space:]]Two[[:space:]]Classes_[[:space:]]Adapting[[:space:]]Supervised[[:space:]]Contrastive[[:space:]]Learning[[:space:]]to[[:space:]]Binary[[:space:]]Imbalanced[[:space:]]Datasets/7be69b4e-6240-4897-87a0-9e54582fed6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Theory[[:space:]]of[[:space:]]Learning[[:space:]]Unified[[:space:]]Model[[:space:]]via[[:space:]]Knowledge[[:space:]]Integration[[:space:]]from[[:space:]]Label[[:space:]]Space[[:space:]]Varying[[:space:]]Domains/71226fdf-7aa9-4bda-88d3-c62f01e56520_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Unified[[:space:]]Approach[[:space:]]to[[:space:]]Interpreting[[:space:]]Self-supervised[[:space:]]Pre-training[[:space:]]Methods[[:space:]]for[[:space:]]3D[[:space:]]Point[[:space:]]Clouds[[:space:]]via[[:space:]]Interactions/99ce7f16-2914-4f50-bc65-0054f0b31f07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Heterogeneous[[:space:]]Semi-supervised[[:space:]]Learning/341a6b6d-e6ba-48c2-98b8-5f373e8e2473_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Unified[[:space:]]Image-Dense[[:space:]]Annotation[[:space:]]Generation[[:space:]]Model[[:space:]]for[[:space:]]Underwater[[:space:]]Scenes/c77f859c-1439-4915-ba1a-a9314ac3d9a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Unified[[:space:]]Latent[[:space:]]Schrodinger[[:space:]]Bridge[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Unsupervised[[:space:]]Anomaly[[:space:]]Detection[[:space:]]and[[:space:]]Localization/0371bea6-a128-4eff-9f4f-dffd7eab7a85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Unified[[:space:]]Model[[:space:]]for[[:space:]]Compressed[[:space:]]Sensing[[:space:]]MRI[[:space:]]Across[[:space:]]Undersampling[[:space:]]Patterns/e66975d9-0abf-4453-9e0f-887ea6234025_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Unified,[[:space:]]Resilient,[[:space:]]and[[:space:]]Explainable[[:space:]]Adversarial[[:space:]]Patch[[:space:]]Detector/b9c12ba3-81c5-4427-ad44-e661e361941c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Universal[[:space:]]Scale-Adaptive[[:space:]]Deformable[[:space:]]Transformer[[:space:]]for[[:space:]]Image[[:space:]]Restoration[[:space:]]across[[:space:]]Diverse[[:space:]]Artifacts/36bcd46e-32c4-4da4-aff4-67b68f83d335_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A3_[[:space:]]Few-shot[[:space:]]Prompt[[:space:]]Learning[[:space:]]of[[:space:]]Unlearnable[[:space:]]Examples[[:space:]]with[[:space:]]Cross-Modal[[:space:]]Adversarial[[:space:]]Feature[[:space:]]Alignment/641deceb-eb2b-45b8-97bb-8e2f7fe97a5e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A4A_[[:space:]]Adapter[[:space:]]for[[:space:]]Adapter[[:space:]]Transfer[[:space:]]via[[:space:]]All-for-All[[:space:]]Mapping[[:space:]]for[[:space:]]Cross-Architecture[[:space:]]Models/66a56aa1-a030-49e2-8d93-323c4a070c6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AA-CLIP_[[:space:]]Enhancing[[:space:]]Zero-Shot[[:space:]]Anomaly[[:space:]]Detection[[:space:]]via[[:space:]]Anomaly-Aware[[:space:]]CLIP/1a15f5c2-189a-4bf7-936d-11fcae85b257_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ABBSPO_[[:space:]]Adaptive[[:space:]]Bounding[[:space:]]Box[[:space:]]Scaling[[:space:]]and[[:space:]]Symmetric[[:space:]]Prior[[:space:]]based[[:space:]]Orientation[[:space:]]Prediction[[:space:]]for[[:space:]]Detecting[[:space:]]Aerial[[:space:]]Image[[:space:]]Objects/4a21cd8b-8cbb-4db3-b4a2-0a741c570b46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ABC-Former_[[:space:]]Auxiliary[[:space:]]Bimodal[[:space:]]Cross-domain[[:space:]]Transformer[[:space:]]with[[:space:]]Interactive[[:space:]]Channel[[:space:]]Attention[[:space:]]for[[:space:]]White[[:space:]]Balance/531c4970-08d2-402e-8972-1a2e62e92c0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AC3D_[[:space:]]Analyzing[[:space:]]and[[:space:]]Improving[[:space:]]3D[[:space:]]Camera[[:space:]]Control[[:space:]]in[[:space:]]Video[[:space:]]Diffusion[[:space:]]Transformers/61cdf9ca-5840-4279-a8c2-d8f7abcf95b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ACAttack_[[:space:]]Adaptive[[:space:]]Cross[[:space:]]Attacking[[:space:]]RGB-T[[:space:]]Tracker[[:space:]]via[[:space:]]Multi-Modal[[:space:]]Response[[:space:]]Decoupling/6f190354-412a-4587-a766-f43d48cafb75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ACE_[[:space:]]Anti-Editing[[:space:]]Concept[[:space:]]Erasure[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Models/b5a87e68-e26d-46f8-a43c-98bb87af5dc7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ACL_[[:space:]]Activating[[:space:]]Capability[[:space:]]of[[:space:]]Linear[[:space:]]Attention[[:space:]]for[[:space:]]Image[[:space:]]Restoration/b3cdf41c-47b4-40a8-888d-383f4742ab99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ADD_[[:space:]]Attribution-Driven[[:space:]]Data[[:space:]]Augmentation[[:space:]]Framework[[:space:]]for[[:space:]]Boosting[[:space:]]Image[[:space:]]Super-Resolution/e29ca394-45e6-4bf5-b8ea-58a74dbe24fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ADU_[[:space:]]Adaptive[[:space:]]Detection[[:space:]]of[[:space:]]Unknown[[:space:]]Categories[[:space:]]in[[:space:]]Black-Box[[:space:]]Domain[[:space:]]Adaptation/a3a5f623-9c8e-40c7-93d2-175e52310d81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AFL_[[:space:]]A[[:space:]]Single-Round[[:space:]]Analytic[[:space:]]Approach[[:space:]]for[[:space:]]Federated[[:space:]]Learning[[:space:]]with[[:space:]]Pre-trained[[:space:]]Models/1961e1d7-9e21-46e8-8ee1-ddb51cc88578_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AG-VPReID_[[:space:]]A[[:space:]]Challenging[[:space:]]Large-Scale[[:space:]]Benchmark[[:space:]]for[[:space:]]Aerial-Ground[[:space:]]Video-based[[:space:]]Person[[:space:]]Re-Identification/4cdeaea8-7584-41e0-a7e3-68147957d6c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AI-Face_[[:space:]]A[[:space:]]Million-Scale[[:space:]]Demographically[[:space:]]Annotated[[:space:]]AI-Generated[[:space:]]Face[[:space:]]Dataset[[:space:]]and[[:space:]]Fairness[[:space:]]Benchmark/69a93785-ab72-45f7-87fa-9182ea82821d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AIGV-Assessor_[[:space:]]Benchmarking[[:space:]]and[[:space:]]Evaluating[[:space:]]the[[:space:]]Perceptual[[:space:]]Quality[[:space:]]of[[:space:]]Text-to-Video[[:space:]]Generation[[:space:]]with[[:space:]]LMM/6da59f8b-88b1-4dd7-9656-385f5fb4c136_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AIM-Fair_[[:space:]]Advancing[[:space:]]Algorithmic[[:space:]]Fairness[[:space:]]via[[:space:]]Selectively[[:space:]]Fine-Tuning[[:space:]]Biased[[:space:]]Models[[:space:]]with[[:space:]]Contextual[[:space:]]Synthetic[[:space:]]Data/e3f123cb-93dd-4a70-bf37-7825bb88c80c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AIpparel_[[:space:]]A[[:space:]]Multimodal[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Digital[[:space:]]Garments/cb5e1214-45b4-45fb-973d-1b1e69c31e9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AKiRa_[[:space:]]Augmentation[[:space:]]Kit[[:space:]]on[[:space:]]Rays[[:space:]]for[[:space:]]Optical[[:space:]]Video[[:space:]]Generation/a6c7a556-f741-4830-97d0-e5bff3fbd90e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ALIEN_[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations[[:space:]]for[[:space:]]Human[[:space:]]Motion[[:space:]]Prediction[[:space:]]under[[:space:]]Arbitrary[[:space:]]Latency/36dd2f51-bc2c-4f32-bce3-0f31cc93b437_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AMO[[:space:]]Sampler_[[:space:]]Enhancing[[:space:]]Text[[:space:]]Rendering[[:space:]]with[[:space:]]Overshooting/36ba09a6-cb41-4431-b966-1bb4ac405663_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AMR-Transformer_[[:space:]]Enabling[[:space:]]Efficient[[:space:]]Long-range[[:space:]]Interaction[[:space:]]for[[:space:]]Complex[[:space:]]Neural[[:space:]]Fluid[[:space:]]Simulation/fde66b48-7438-40ff-858c-23379a28df70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ANNEXE_[[:space:]]Unified[[:space:]]Analyzing,[[:space:]]Answering,[[:space:]]and[[:space:]]Pixel[[:space:]]Grounding[[:space:]]for[[:space:]]Egocentric[[:space:]]Interaction/3ad66852-c283-4b9f-ae72-5f19b6b801d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/APHQ-ViT_[[:space:]]Post-Training[[:space:]]Quantization[[:space:]]with[[:space:]]Average[[:space:]]Perturbation[[:space:]]Hessian[[:space:]]Based[[:space:]]Reconstruction[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/959d7612-e866-4f28-9fa9-c45c76adc117_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/APT_[[:space:]]Adaptive[[:space:]]Personalized[[:space:]]Training[[:space:]]for[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Limited[[:space:]]Data/e1426338-b974-42f5-bbf7-281f898f4fda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AR-Diffusion_[[:space:]]Asynchronous[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Auto-Regressive[[:space:]]Diffusion/c0879fca-0022-40d5-b619-70f13ce4b0b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ARKit[[:space:]]LabelMaker_[[:space:]]A[[:space:]]New[[:space:]]Scale[[:space:]]for[[:space:]]Indoor[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/f04d603b-51d3-4844-a6bb-a0dc57e1988d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ARM_[[:space:]]Appearance[[:space:]]Reconstruction[[:space:]]Model[[:space:]]for[[:space:]]Relightable[[:space:]]3D[[:space:]]Generation/b42d0494-c967-413d-a0db-d7029a7a1cb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ART_[[:space:]]Anonymous[[:space:]]Region[[:space:]]Transformer[[:space:]]for[[:space:]]Variable[[:space:]]Multi-Layer[[:space:]]Transparent[[:space:]]Image[[:space:]]Generation/3c46533d-a542-49b4-875c-2472e2ae6f45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ASAP_[[:space:]]Advancing[[:space:]]Semantic[[:space:]]Alignment[[:space:]]Promotes[[:space:]]Multi-Modal[[:space:]]Manipulation[[:space:]]Detecting[[:space:]]and[[:space:]]Grounding/f7fca8ee-e28d-4985-93bf-eb9e003b5e6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ASHiTA_[[:space:]]Automatic[[:space:]]Scene-grounded[[:space:]]HIerarchical[[:space:]]Task[[:space:]]Analysis/d2f3fd11-c995-44f8-bb50-b38a1023e5bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ASIGN_[[:space:]]An[[:space:]]Anatomy-aware[[:space:]]Spatial[[:space:]]Imputation[[:space:]]Graphic[[:space:]]Network[[:space:]]for[[:space:]]3D[[:space:]]Spatial[[:space:]]Transcriptomics/f188ab42-f900-4f45-99a8-fdccbbb31e53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ATA_[[:space:]]Adaptive[[:space:]]Transformation[[:space:]]Agent[[:space:]]for[[:space:]]Text-Guided[[:space:]]Subject-Position[[:space:]]Variable[[:space:]]Background[[:space:]]Inpainting/14199472-5326-4a35-8c14-097431f77de1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ATP-LLaVA_[[:space:]]Adaptive[[:space:]]Token[[:space:]]Pruning[[:space:]]for[[:space:]]Large[[:space:]]Vision[[:space:]]Language[[:space:]]Models/8d1f0df9-f203-48c9-9342-281daba45228_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ATP_[[:space:]]Adaptive[[:space:]]Threshold[[:space:]]Pruning[[:space:]]for[[:space:]]Efficient[[:space:]]Data[[:space:]]Encoding[[:space:]]in[[:space:]]Quantum[[:space:]]Neural[[:space:]]Networks/79b52e33-68a9-4a4f-ba48-65f0d682cbf4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AToM_[[:space:]]Aligning[[:space:]]Text-to-Motion[[:space:]]Model[[:space:]]at[[:space:]]Event-Level[[:space:]]with[[:space:]]GPT-4Vision[[:space:]]Reward/53c55dc0-83e6-442b-888e-b224dec66e72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AVF-MAE++_[[:space:]]Scaling[[:space:]]Affective[[:space:]]Video[[:space:]]Facial[[:space:]]Masked[[:space:]]Autoencoders[[:space:]]via[[:space:]]Efficient[[:space:]]Audio-Visual[[:space:]]Self-Supervised[[:space:]]Learning/83d6d292-6d99-48f9-bbdc-1103770d2580_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AVQACL_[[:space:]]A[[:space:]]Novel[[:space:]]Benchmark[[:space:]]for[[:space:]]Audio-Visual[[:space:]]Question[[:space:]]Answering[[:space:]]Continual[[:space:]]Learning/b674e7c6-26b4-4ecd-89fa-8c07ae0c7e73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Acc3D_[[:space:]]Accelerating[[:space:]]Single[[:space:]]Image[[:space:]]to[[:space:]]3D[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Edge[[:space:]]Consistency[[:space:]]Guided[[:space:]]Score[[:space:]]Distillation/17bd19a6-d3a1-4433-b775-c08e6ecd64ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Accelerating[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]via[[:space:]]Increment-Calibrated[[:space:]]Caching[[:space:]]with[[:space:]]Channel-Aware[[:space:]]Singular[[:space:]]Value[[:space:]]Decomposition/7888968c-48cf-4d4f-a959-67393862ef73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Accelerating[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]by[[:space:]]Searching[[:space:]]Optimal[[:space:]]Vision[[:space:]]Token[[:space:]]Reduction/a2ebe21a-dd1c-4482-ab11-128a62fd12ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Accurate[[:space:]]Differential[[:space:]]Operators[[:space:]]for[[:space:]]Hybrid[[:space:]]Neural[[:space:]]Fields/c44d2865-cbb6-462e-9dd7-d5a3d2a2b9e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Accurate[[:space:]]Scene[[:space:]]Text[[:space:]]Recognition[[:space:]]with[[:space:]]Efficient[[:space:]]Model[[:space:]]Scaling[[:space:]]and[[:space:]]Cloze[[:space:]]Self-Distillation/8b6df24f-3fc1-4d1c-b1ba-505adaf9019f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Acquire[[:space:]]and[[:space:]]then[[:space:]]Adapt_[[:space:]]Squeezing[[:space:]]out[[:space:]]Text-to-Image[[:space:]]Model[[:space:]]for[[:space:]]Image[[:space:]]Restoration/07b6f944-c037-4310-b515-257765e135f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Action[[:space:]]Detail[[:space:]]Matters_[[:space:]]Refining[[:space:]]Video[[:space:]]Recognition[[:space:]]with[[:space:]]Local[[:space:]]Action[[:space:]]Queries/b601f295-f3c7-4954-9e4e-6e0e21d54f88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Activating[[:space:]]Sparse[[:space:]]Part[[:space:]]Concepts[[:space:]]for[[:space:]]3D[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning/fabc8d8a-8823-4cd5-b9dd-08fc72902ce3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Active[[:space:]]Data[[:space:]]Curation[[:space:]]Effectively[[:space:]]Distills[[:space:]]Large-Scale[[:space:]]Multimodal[[:space:]]Models/975ddfce-0026-4a44-8121-44c16bac99b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Active[[:space:]]Event-based[[:space:]]Stereo[[:space:]]Vision/facc70e9-e2cd-4cca-ac3d-9def9aed72a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Active[[:space:]]Hyperspectral[[:space:]]Imaging[[:space:]]Using[[:space:]]an[[:space:]]Event[[:space:]]Camera/a03728ba-8a8f-49b7-934d-19cc07606a93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ActiveGAMER_[[:space:]]Active[[:space:]]GAussian[[:space:]]Mapping[[:space:]]through[[:space:]]Efficient[[:space:]]Rendering/bf80f70a-9b60-4e61-a832-fff7dd165d0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AdMiT_[[:space:]]Adaptive[[:space:]]Multi-Source[[:space:]]Tuning[[:space:]]in[[:space:]]Dynamic[[:space:]]Environments/d96341ee-6b6d-4f3e-84a0-f056b4651277_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AdaCM^2_[[:space:]]On[[:space:]]Understanding[[:space:]]Extremely[[:space:]]Long-Term[[:space:]]Video[[:space:]]with[[:space:]]Adaptive[[:space:]]Cross-Modality[[:space:]]Memory[[:space:]]Reduction/d9da8c6c-84df-4285-8556-905770af79d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AdaDARE-gamma_[[:space:]]Balancing[[:space:]]Stability[[:space:]]and[[:space:]]Plasticity[[:space:]]in[[:space:]]Multi-modal[[:space:]]LLMs[[:space:]]through[[:space:]]Efficient[[:space:]]Adaptation/40308d8d-f99a-4e31-9931-f5a321e85728_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AdaMMS_[[:space:]]Model[[:space:]]Merging[[:space:]]for[[:space:]]Heterogeneous[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]with[[:space:]]Unsupervised[[:space:]]Coefficient[[:space:]]Optimization/4a08ffb8-8904-4ef5-923a-c10f3aeb33f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AdaptCMVC_[[:space:]]Robust[[:space:]]Adaption[[:space:]]to[[:space:]]Incremental[[:space:]]Views[[:space:]]in[[:space:]]Continual[[:space:]]Multi-view[[:space:]]Clustering/73638180-8aa5-4d61-9e73-03b5c7fd1d88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adapter[[:space:]]Merging[[:space:]]with[[:space:]]Centroid[[:space:]]Prototype[[:space:]]Mapping[[:space:]]for[[:space:]]Scalable[[:space:]]Class-Incremental[[:space:]]Learning/22e4c0ac-bef5-4540-8a32-a7412685f6a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adapting[[:space:]]Dense[[:space:]]Matching[[:space:]]for[[:space:]]Homography[[:space:]]Estimation[[:space:]]with[[:space:]]Grid-based[[:space:]]Acceleration/2f8fff98-4047-461c-a364-929dae730a0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adapting[[:space:]]Pre-trained[[:space:]]3D[[:space:]]Models[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Video[[:space:]]Understanding[[:space:]]via[[:space:]]Cross-frame[[:space:]]Spatio-temporal[[:space:]]Perception/2318884c-d71a-4947-ab84-f1c482a132fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adapting[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]with[[:space:]]Feature[[:space:]]Difference[[:space:]]Instruction[[:space:]]for[[:space:]]Generic[[:space:]]Image[[:space:]]Restoration/0abe3b3e-a1ff-4d75-954a-4898bc338177_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adapting[[:space:]]to[[:space:]]Observation[[:space:]]Length[[:space:]]of[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]via[[:space:]]Contrastive[[:space:]]Learning/2aff36b8-d0fd-49cd-ab28-36f875e41f15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adapting[[:space:]]to[[:space:]]the[[:space:]]Unknown_[[:space:]]Training-Free[[:space:]]Audio-Visual[[:space:]]Event[[:space:]]Perception[[:space:]]with[[:space:]]Dynamic[[:space:]]Thresholds/e3af0000-7a87-46f4-897e-0e5ba81f78d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Dropout_[[:space:]]Unleashing[[:space:]]Dropout[[:space:]]across[[:space:]]Layers[[:space:]]for[[:space:]]Generalizable[[:space:]]Image[[:space:]]Super-Resolution/753ef088-ca7a-48f3-810f-83fa0edbc376_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Keyframe[[:space:]]Sampling[[:space:]]for[[:space:]]Long[[:space:]]Video[[:space:]]Understanding/f9180f95-5574-49b3-8480-091edde2c028_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Markup[[:space:]]Language[[:space:]]Generation[[:space:]]for[[:space:]]Contextually-Grounded[[:space:]]Visual[[:space:]]Document[[:space:]]Understanding/dbe47fad-c46f-4091-84d0-ae0a68fd8d7a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Non-Uniform[[:space:]]Timestep[[:space:]]Sampling[[:space:]]for[[:space:]]Accelerating[[:space:]]Diffusion[[:space:]]Model[[:space:]]Training/22adf80a-55fe-4236-9782-243fa367cc68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Parameter[[:space:]]Selection[[:space:]]for[[:space:]]Tuning[[:space:]]Vision-Language[[:space:]]Models/1c740f1f-79f0-4120-87dd-dfad61af0ea1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Part[[:space:]]Learning[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery_[[:space:]]A[[:space:]]Plug-and-Play[[:space:]]Enhancement/ec1ae1d1-cc72-4130-b4f0-88367d3f1765_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Rectangular[[:space:]]Convolution[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Pansharpening/5353ef4c-f997-4ff3-8a75-44de03885284_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Unimodal[[:space:]]Regulation[[:space:]]for[[:space:]]Balanced[[:space:]]Multimodal[[:space:]]Information[[:space:]]Acquisition/b7e4b615-c7a4-4e19-8ac2-4c23de2bddf6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adv-CPG_[[:space:]]A[[:space:]]Customized[[:space:]]Portrait[[:space:]]Generation[[:space:]]Framework[[:space:]]with[[:space:]]Facial[[:space:]]Adversarial[[:space:]]Attacks/935ee611-fc85-4dc5-b64f-76a0954a862e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Advancing[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]in[[:space:]]GNeRFs_[[:space:]]The[[:space:]]IL2-NeRF[[:space:]]Attack/da079a0c-be2d-4202-8367-e1269fa454f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Advancing[[:space:]]Generalizable[[:space:]]Tumor[[:space:]]Segmentation[[:space:]]with[[:space:]]Anomaly-Aware[[:space:]]Open-Vocabulary[[:space:]]Attention[[:space:]]Maps[[:space:]]and[[:space:]]Frozen[[:space:]]Foundation[[:space:]]Diffusion[[:space:]]Models/93e36109-47c6-4476-8f14-d6e66b1e3bcb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Advancing[[:space:]]Manga[[:space:]]Analysis_[[:space:]]Comprehensive[[:space:]]Segmentation[[:space:]]Annotations[[:space:]]for[[:space:]]the[[:space:]]Manga109[[:space:]]Dataset/70b24633-f352-44f7-8401-0b43e88806e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Advancing[[:space:]]Multiple[[:space:]]Instance[[:space:]]Learning[[:space:]]with[[:space:]]Continual[[:space:]]Learning[[:space:]]for[[:space:]]Whole[[:space:]]Slide[[:space:]]Imaging/fb82e95a-5f8c-4a5e-8aef-bec3818e8ab7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Advancing[[:space:]]Myopia[[:space:]]To[[:space:]]Holism_[[:space:]]Fully[[:space:]]Contrastive[[:space:]]Language-Image[[:space:]]Pre-training/81e20a7b-d971-4004-bd5c-d28406f525e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Advancing[[:space:]]Semantic[[:space:]]Future[[:space:]]Prediction[[:space:]]through[[:space:]]Multimodal[[:space:]]Visual[[:space:]]Sequence[[:space:]]Transformers/b877dd04-fd12-45ca-87e6-f2c286a76d4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adventurer_[[:space:]]Optimizing[[:space:]]Vision[[:space:]]Mamba[[:space:]]Architecture[[:space:]]Designs[[:space:]]for[[:space:]]Efficiency/636527d7-e914-46d3-98df-b024266f2083_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adversarial[[:space:]]Diffusion[[:space:]]Compression[[:space:]]for[[:space:]]Real-World[[:space:]]Image[[:space:]]Super-Resolution/b17a4f99-8cde-4333-92df-317d8c405cc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adversarial[[:space:]]Domain[[:space:]]Prompt[[:space:]]Tuning[[:space:]]and[[:space:]]Generation[[:space:]]for[[:space:]]Single[[:space:]]Domain[[:space:]]Generalization/d9b65f1b-19c1-4339-b638-f48e19307f40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AeSPa[[:space:]]_[[:space:]]Attention-guided[[:space:]]Self-supervised[[:space:]]Parallel[[:space:]]Imaging[[:space:]]for[[:space:]]MRI[[:space:]]Reconstruction/c88c1381-21d7-49b0-8456-dec7f96f9352_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AerialMegaDepth_[[:space:]]Learning[[:space:]]Aerial-Ground[[:space:]]Reconstruction[[:space:]]and[[:space:]]View[[:space:]]Synthesis/8c898bc5-7457-44c4-b163-0e1cdbc6691e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AeroGen_[[:space:]]Enhancing[[:space:]]Remote[[:space:]]Sensing[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Diffusion-Driven[[:space:]]Data[[:space:]]Generation/7a102fa8-5370-49c8-b769-090b3862ed57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AesthetiQ_[[:space:]]Enhancing[[:space:]]Graphic[[:space:]]Layout[[:space:]]Design[[:space:]]via[[:space:]]Aesthetic-Aware[[:space:]]Preference[[:space:]]Alignment[[:space:]]of[[:space:]]Multi-modal[[:space:]]Large[[:space:]]Language[[:space:]]Models/3334f6bb-e372-41c5-8b84-459985af2808_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Aesthetic[[:space:]]Post-Training[[:space:]]Diffusion[[:space:]]Models[[:space:]]from[[:space:]]Generic[[:space:]]Preferences[[:space:]]with[[:space:]]Step-by-step[[:space:]]Preference[[:space:]]Optimization/707d74d2-e68b-4378-811f-7b9f92af548c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AffordDP_[[:space:]]Generalizable[[:space:]]Diffusion[[:space:]]Policy[[:space:]]with[[:space:]]Transferable[[:space:]]Affordance/2ded0a52-8c20-46f4-8402-2cfb1a6b2523_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AirRoom_[[:space:]]Objects[[:space:]]Matter[[:space:]]in[[:space:]]Room[[:space:]]Reidentification/038d25fb-3df5-4833-8312-a08baf2819c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Alias-Free[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models_[[:space:]]Improving[[:space:]]Fractional[[:space:]]Shift[[:space:]]Equivariance[[:space:]]of[[:space:]]Diffusion[[:space:]]Latent[[:space:]]Space/80428174-fe1b-4452-b39a-871579acfdc8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Align-A-Video_[[:space:]]Deterministic[[:space:]]Reward[[:space:]]Tuning[[:space:]]of[[:space:]]Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Consistent[[:space:]]Video[[:space:]]Editing/272da2cc-697d-4dc9-87a9-5f36c73e4a51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Align-KD_[[:space:]]Distilling[[:space:]]Cross-Modal[[:space:]]Alignment[[:space:]]Knowledge[[:space:]]for[[:space:]]Mobile[[:space:]]Vision-Language[[:space:]]Large[[:space:]]Model[[:space:]]Enhancement/d17ce722-02e7-4757-92ce-b29d7e6b2292_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Align3R_[[:space:]]Aligned[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation[[:space:]]for[[:space:]]Dynamic[[:space:]]Videos/491353b3-bac9-434c-9a57-2358cfac27a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AlignMamba_[[:space:]]Enhancing[[:space:]]Multimodal[[:space:]]Mamba[[:space:]]with[[:space:]]Local[[:space:]]and[[:space:]]Global[[:space:]]Cross-modal[[:space:]]Alignment/a5861339-b784-4b27-81bb-8f157e39a2fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Alignment,[[:space:]]Mining[[:space:]]and[[:space:]]Fusion_[[:space:]]Representation[[:space:]]Alignment[[:space:]]with[[:space:]]Hard[[:space:]]Negative[[:space:]]Mining[[:space:]]and[[:space:]]Selective[[:space:]]Knowledge[[:space:]]Fusion[[:space:]]for[[:space:]]Medical[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/65891794-3f1e-430f-9b6c-d393e6392deb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/All[[:space:]]Languages[[:space:]]Matter_[[:space:]]Evaluating[[:space:]]LMMs[[:space:]]on[[:space:]]Culturally[[:space:]]Diverse[[:space:]]100[[:space:]]Languages/d9e05aff-1334-4f29-8894-c24ce537b99b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/All-Day[[:space:]]Multi-Camera[[:space:]]Multi-Target[[:space:]]Tracking/4f6f86d0-47bd-45fb-93f1-63cacc4cc890_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/All-Optical[[:space:]]Nonlinear[[:space:]]Diffractive[[:space:]]Deep[[:space:]]Network[[:space:]]for[[:space:]]Ultrafast[[:space:]]Image[[:space:]]Denoising/13c893bd-684b-4393-9092-3b65681d800a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/All-directional[[:space:]]Disparity[[:space:]]Estimation[[:space:]]for[[:space:]]Real-world[[:space:]]QPD[[:space:]]Images/0f1a6651-c6fe-4511-b31c-d629c8fe9c56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AlphaPre_[[:space:]]Amplitude-Phase[[:space:]]Disentanglement[[:space:]]Model[[:space:]]for[[:space:]]Precipitation[[:space:]]Nowcasting/1d1b598a-ac2a-42b4-8969-c87e38539381_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/An[[:space:]]End-to-End[[:space:]]Robust[[:space:]]Point[[:space:]]Cloud[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]Network[[:space:]]with[[:space:]]Single-Step[[:space:]]Conditional[[:space:]]Diffusion[[:space:]]Models/aeac79b7-3a1a-4deb-94e1-ec254a4ea9e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/An[[:space:]]Image-like[[:space:]]Diffusion[[:space:]]Method[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection/4d7e1ab6-dcd7-4795-9aa9-d018084ca7eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Analyzing[[:space:]]the[[:space:]]Synthetic-to-Real[[:space:]]Domain[[:space:]]Gap[[:space:]]in[[:space:]]3D[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation/79a38a4d-7df5-495e-9145-7e40493cd17f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Anatomical[[:space:]]Consistency[[:space:]]and[[:space:]]Adaptive[[:space:]]Prior-informed[[:space:]]Transformation[[:space:]]for[[:space:]]Multi-contrast[[:space:]]MR[[:space:]]Image[[:space:]]Synthesis[[:space:]]via[[:space:]]Diffusion[[:space:]]Model/73930d31-9303-4698-85d2-1b40a2279b0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Anchor-Aware[[:space:]]Similarity[[:space:]]Cohesion[[:space:]]in[[:space:]]Target[[:space:]]Frames[[:space:]]Enables[[:space:]]Predicting[[:space:]]Temporal[[:space:]]Moment[[:space:]]Boundaries[[:space:]]in[[:space:]]2D/b7981468-d86f-4efa-8fd3-ce681a02518d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AniDoc_[[:space:]]Animation[[:space:]]Creation[[:space:]]Made[[:space:]]Easier/cfb9ba0e-7968-4c23-9933-c4dc758bc21c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AniGS_[[:space:]]Animatable[[:space:]]Gaussian[[:space:]]Avatar[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]with[[:space:]]Inconsistent[[:space:]]Gaussian[[:space:]]Reconstruction/ef806cac-9a21-49d6-9c03-b51761e865ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AniGrad_[[:space:]]Anisotropic[[:space:]]Gradient-Adaptive[[:space:]]Sampling[[:space:]]for[[:space:]]3D[[:space:]]Reconstruction[[:space:]]From[[:space:]]Monocular[[:space:]]Video/6e7c8cbc-7bea-4a82-bede-18625f53a954_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AniMer_[[:space:]]Animal[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Estimation[[:space:]]Using[[:space:]]Family[[:space:]]Aware[[:space:]]Transformer/e8320de9-2373-4f94-af6d-e0477421e70a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AniMo_[[:space:]]Species-Aware[[:space:]]Model[[:space:]]for[[:space:]]Text-Driven[[:space:]]Animal[[:space:]]Motion[[:space:]]Generation/d1bdf116-08c1-4ae7-be83-53b6b984d11b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Animate[[:space:]]and[[:space:]]Sound[[:space:]]an[[:space:]]Image/05ad8fc9-2f1c-48f9-a872-f58540e7568e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnimateAnything_[[:space:]]Consistent[[:space:]]and[[:space:]]Controllable[[:space:]]Animation[[:space:]]for[[:space:]]Video[[:space:]]Generation/bcea80c1-c59f-49ef-ac7c-d37012d488e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Annotation[[:space:]]Ambiguity[[:space:]]Aware[[:space:]]Semi-Supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/4a018a65-3437-4e7d-b298-342b324a7eb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnomalyNCD_[[:space:]]Towards[[:space:]]Novel[[:space:]]Anomaly[[:space:]]Class[[:space:]]Discovery[[:space:]]in[[:space:]]Industrial[[:space:]]Scenarios/11e82997-538e-497e-b7ec-1eafddcfb0ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Anomize_[[:space:]]Better[[:space:]]Open[[:space:]]Vocabulary[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/4fc9dbe6-936e-4bc5-a4dd-aba17209b462_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Antidote_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Mitigating[[:space:]]LVLM[[:space:]]Hallucinations[[:space:]]in[[:space:]]Counterfactual[[:space:]]Presupposition[[:space:]]and[[:space:]]Object[[:space:]]Perception/6db1ce8e-fd18-4a0e-ba53-a6c28e2d6b2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Any-Resolution[[:space:]]AI-Generated[[:space:]]Image[[:space:]]Detection[[:space:]]by[[:space:]]Spectral[[:space:]]Learning/01237cda-8864-438e-b109-9bbcccc117a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Any3DIS_[[:space:]]Class-Agnostic[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation[[:space:]]by[[:space:]]2D[[:space:]]Mask[[:space:]]Tracking/ce1981cd-2be3-4f58-a52d-00c95bdfd1ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Any6D_[[:space:]]Model-free[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation[[:space:]]of[[:space:]]Novel[[:space:]]Objects/88be08d0-0c92-46e8-8d1e-d73a0d61086d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnyCam_[[:space:]]Learning[[:space:]]to[[:space:]]Recover[[:space:]]Camera[[:space:]]Poses[[:space:]]and[[:space:]]Intrinsics[[:space:]]from[[:space:]]Casual[[:space:]]Videos/fbf99415-72e0-4933-bc99-8d3b71f9ff34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnyDressing_[[:space:]]Customizable[[:space:]]Multi-Garment[[:space:]]Virtual[[:space:]]Dressing[[:space:]]via[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models/2f6b4d56-cac4-4272-ac83-8d6926be3b7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnyEdit_[[:space:]]Mastering[[:space:]]Unified[[:space:]]High-Quality[[:space:]]Image[[:space:]]Editing[[:space:]]for[[:space:]]Any[[:space:]]Idea/9fcd57e4-069d-4375-aa36-ed6b52100a2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnyMap_[[:space:]]Learning[[:space:]]a[[:space:]]General[[:space:]]Camera[[:space:]]Model[[:space:]]for[[:space:]]Structure-from-Motion[[:space:]]with[[:space:]]Unknown[[:space:]]Distortion[[:space:]]in[[:space:]]Dynamic[[:space:]]Scenes/e465c9e0-7671-45e8-bbd9-2949c9f79bfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnyMoLe_[[:space:]]Any[[:space:]]Character[[:space:]]Motion[[:space:]]In-betweening[[:space:]]Leveraging[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/7e8e676d-22d3-4beb-8ba2-13aaedc45940_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnySat_[[:space:]]One[[:space:]]Earth[[:space:]]Observation[[:space:]]Model[[:space:]]for[[:space:]]Many[[:space:]]Resolutions,[[:space:]]Scales,[[:space:]]and[[:space:]]Modalities/341f0685-ded4-4173-9751-e3ffd6c66a1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Anyattack_[[:space:]]Towards[[:space:]]Large-scale[[:space:]]Self-supervised[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]on[[:space:]]Vision-language[[:space:]]Models/438383d8-f00d-4b52-b5ba-083f19728f0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Apollo_[[:space:]]An[[:space:]]Exploration[[:space:]]of[[:space:]]Video[[:space:]]Understanding[[:space:]]in[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models/26f1c714-9db0-4a2b-b2b0-ca71b5fae411_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Apply[[:space:]]Hierarchical-Chain-of-Generation[[:space:]]to[[:space:]]Complex[[:space:]]Attributes[[:space:]]Text-to-3D[[:space:]]Generation/0c828a06-e51c-4ad9-97a1-8a8708e42c1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Arbitrary-steps[[:space:]]Image[[:space:]]Super-resolution[[:space:]]via[[:space:]]Diffusion[[:space:]]Inversion/ab19a24c-693b-415b-ac78-ea7cbddafc79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Arc2Avatar_[[:space:]]Generating[[:space:]]Expressive[[:space:]]3D[[:space:]]Avatars[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]via[[:space:]]ID[[:space:]]Guidance/f7535bb3-6c42-42e7-aaa4-5b0adb7838b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ArcPro_[[:space:]]Architectural[[:space:]]Programs[[:space:]]for[[:space:]]Structured[[:space:]]3D[[:space:]]Abstraction[[:space:]]of[[:space:]]Sparse[[:space:]]Points/88d9983c-45d8-484a-bfc2-8a2ae621ecf2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Are[[:space:]]Images[[:space:]]Indistinguishable[[:space:]]to[[:space:]]Humans[[:space:]]Also[[:space:]]Indistinguishable[[:space:]]to[[:space:]]Classifiers_/a10f1c5f-a2ad-4e40-8dbe-b762a4ff40ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Are[[:space:]]Spatial-Temporal[[:space:]]Graph[[:space:]]Convolution[[:space:]]Networks[[:space:]]for[[:space:]]Human[[:space:]]Action[[:space:]]Recognition[[:space:]]Over-Parameterized_/88bf08f4-5749-497b-a1d1-ebcd94f5a09f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Argus_[[:space:]]A[[:space:]]Compact[[:space:]]and[[:space:]]Versatile[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Vision/5930bf5a-acd0-4237-83b9-75790121bbdb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Argus_[[:space:]]Vision-Centric[[:space:]]Reasoning[[:space:]]with[[:space:]]Grounded[[:space:]]Chain-of-Thought/128acd83-dd58-4ab7-9b5a-030184652612_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Around[[:space:]]the[[:space:]]World[[:space:]]in[[:space:]]80[[:space:]]Timesteps_[[:space:]]A[[:space:]]Generative[[:space:]]Approach[[:space:]]to[[:space:]]Global[[:space:]]Visual[[:space:]]Geolocation/13410d71-cb2a-4fa8-b7e0-f3c80090cc77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ArtFormer_[[:space:]]Controllable[[:space:]]Generation[[:space:]]of[[:space:]]Diverse[[:space:]]3D[[:space:]]Articulated[[:space:]]Objects/331a9c6f-f94e-4edc-a70d-ca90256772fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ArtiFade_[[:space:]]Learning[[:space:]]to[[:space:]]Generate[[:space:]]High-quality[[:space:]]Subject[[:space:]]from[[:space:]]Blemished[[:space:]]Images/220dc224-ed64-4d94-8e9a-7402ee1617ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ArtiScene_[[:space:]]Language-Driven[[:space:]]Artistic[[:space:]]3D[[:space:]]Scene[[:space:]]Generation[[:space:]]Through[[:space:]]Image[[:space:]]Intermediary/19d3c1a8-9a67-4aed-9896-d751d9865c8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Articulated[[:space:]]Kinematics[[:space:]]Distillation[[:space:]]from[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/2d03bda4-d2a1-45e9-a44c-67dfdd0d6eb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ArticulatedGS_[[:space:]]Self-supervised[[:space:]]Digital[[:space:]]Twin[[:space:]]Modeling[[:space:]]of[[:space:]]Articulated[[:space:]]Objects[[:space:]]using[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/8a170f33-77e0-4001-9a38-b2307c10d04d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Assessing[[:space:]]and[[:space:]]Learning[[:space:]]Alignment[[:space:]]of[[:space:]]Unimodal[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Models/387769ad-5e6e-4ef5-a3f0-a0929424982d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Associative[[:space:]]Transformer/ee277769-dba9-4193-85d0-2f3197d7eac7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Asynchronous[[:space:]]Collaborative[[:space:]]Graph[[:space:]]Representation[[:space:]]for[[:space:]]Frames[[:space:]]and[[:space:]]Events/32dbae9f-d8f1-4316-83ba-916b86e1868b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Attend[[:space:]]to[[:space:]]Not[[:space:]]Attended_[[:space:]]Structure-then-Detail[[:space:]]Token[[:space:]]Merging[[:space:]]for[[:space:]]Post-training[[:space:]]DiT[[:space:]]Acceleration/d4c604c8-4776-40df-aa9d-1d3cc316b935_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Attention[[:space:]]Distillation_[[:space:]]A[[:space:]]Unified[[:space:]]Approach[[:space:]]to[[:space:]]Visual[[:space:]]Characteristics[[:space:]]Transfer/b2695df0-b620-4afd-b530-6626ac988e55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Attention[[:space:]]IoU_[[:space:]]Examining[[:space:]]Biases[[:space:]]in[[:space:]]CelebA[[:space:]]using[[:space:]]Attention[[:space:]]Maps/6356502c-2327-4a96-9868-40f3cf353575_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Attraction[[:space:]]Diminishing[[:space:]]and[[:space:]]Distributing[[:space:]]for[[:space:]]Few-Shot[[:space:]]Class-Incremental[[:space:]]Learning/9c29868b-9e31-4eac-acb4-9863a05f0fa3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Attribute-Missing[[:space:]]Multi-view[[:space:]]Graph[[:space:]]Clustering/ca6a563a-ee33-4e4f-ab78-41e9062e12c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Attribute-formed[[:space:]]Class-specific[[:space:]]Concept[[:space:]]Space_[[:space:]]Endowing[[:space:]]Language[[:space:]]Bottleneck[[:space:]]Model[[:space:]]with[[:space:]]Better[[:space:]]Interpretability[[:space:]]and[[:space:]]Scalability/ceb0790c-111c-4a8b-a926-f9d7c3f4908d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AudCast_[[:space:]]Audio-Driven[[:space:]]Human[[:space:]]Video[[:space:]]Generation[[:space:]]by[[:space:]]Cascaded[[:space:]]Diffusion[[:space:]]Transformers/1fb6851a-374a-4557-acc1-20e9f55890bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Audio-Visual[[:space:]]Instance[[:space:]]Segmentation/437703d1-8a6a-48df-9b9e-657ac1d7dd43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Audio-Visual[[:space:]]Semantic[[:space:]]Graph[[:space:]]Network[[:space:]]for[[:space:]]Audio-Visual[[:space:]]Event[[:space:]]Localization/d63d6e32-de1f-498e-827e-fe6fe076d40d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Augmented[[:space:]]Deep[[:space:]]Contexts[[:space:]]for[[:space:]]Spatially[[:space:]]Embedded[[:space:]]Video[[:space:]]Coding/9ae43478-4052-472f-a4d0-fae77bc158fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Augmenting[[:space:]]Multimodal[[:space:]]LLMs[[:space:]]with[[:space:]]Self-Reflective[[:space:]]Tokens[[:space:]]for[[:space:]]Knowledge-based[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/cd88c679-837a-470d-8700-c20d63033e8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Augmenting[[:space:]]Perceptual[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Image[[:space:]]Quality[[:space:]]Predictors/220bdcd1-f2bc-4e08-969a-5c4c011e1f7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AuraFusion360_[[:space:]]Augmented[[:space:]]Unseen[[:space:]]Region[[:space:]]Alignment[[:space:]]for[[:space:]]Reference-based[[:space:]]360deg[[:space:]]Unbounded[[:space:]]Scene[[:space:]]Inpainting/c5606fa4-1d66-43e2-b01a-b00a76838288_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Auto[[:space:]]Cherry-Picker_[[:space:]]Learning[[:space:]]from[[:space:]]High-quality[[:space:]]Generative[[:space:]]Data[[:space:]]Driven[[:space:]]by[[:space:]]Language/85189887-c99d-45dc-9c6c-e88a1f076ea8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Auto-Encoded[[:space:]]Supervision[[:space:]]for[[:space:]]Perceptual[[:space:]]Image[[:space:]]Super-Resolution/5dd53cad-1fb4-4c5c-b377-cef994cdc8d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AutoLUT_[[:space:]]LUT-Based[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]with[[:space:]]Automatic[[:space:]]Sampling[[:space:]]and[[:space:]]Adaptive[[:space:]]Residual[[:space:]]Learning/47fef72d-e172-4f09-844d-453d37400759_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AutoPresent_[[:space:]]Designing[[:space:]]Structured[[:space:]]Visuals[[:space:]]from[[:space:]]Scratch/289716aa-b120-4644-a304-4816a6c1c13e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AutoSSVH_[[:space:]]Exploring[[:space:]]Automated[[:space:]]Frame[[:space:]]Sampling[[:space:]]for[[:space:]]Efficient[[:space:]]Self-Supervised[[:space:]]Video[[:space:]]Hashing/71a2b65b-a4ed-45e7-9e7d-b1b6b51c1606_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AutoURDF_[[:space:]]Unsupervised[[:space:]]Robot[[:space:]]Modeling[[:space:]]from[[:space:]]Point[[:space:]]Cloud[[:space:]]Frames[[:space:]]Using[[:space:]]Cluster[[:space:]]Registration/b9553c65-8626-4958-8f03-20375fd68e44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Automated[[:space:]]Generation[[:space:]]of[[:space:]]Challenging[[:space:]]Multiple-Choice[[:space:]]Questions[[:space:]]for[[:space:]]Vision[[:space:]]Language[[:space:]]Model[[:space:]]Evaluation/47911ae5-7f43-434b-8395-e460ec774b60_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Automated[[:space:]]Proof[[:space:]]of[[:space:]]Polynomial[[:space:]]Inequalities[[:space:]]via[[:space:]]Reinforcement[[:space:]]Learning/cc3d6f4e-1f70-4b2d-97f7-4876ecc42e8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Automatic[[:space:]]Joint[[:space:]]Structured[[:space:]]Pruning[[:space:]]and[[:space:]]Quantization[[:space:]]for[[:space:]]Efficient[[:space:]]Neural[[:space:]]Network[[:space:]]Training[[:space:]]and[[:space:]]Compression/d41d59cf-b5aa-48ce-9a89-ac293f74abd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Automatic[[:space:]]Spectral[[:space:]]Calibration[[:space:]]of[[:space:]]Hyperspectral[[:space:]]Images_[[:space:]]Method,[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark/3991cd1d-7fdd-4107-9330-15f23e9b6e35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Autoregressive[[:space:]]Distillation[[:space:]]of[[:space:]]Diffusion[[:space:]]Transformers/30fe01dc-6601-4f2a-abcf-187532285ad5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Autoregressive[[:space:]]Sequential[[:space:]]Pretraining[[:space:]]for[[:space:]]Visual[[:space:]]Tracking/c4c08443-0cc8-4e36-bd95-a17330b1c6ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AvatarArtist_[[:space:]]Open-Domain[[:space:]]4D[[:space:]]Avatarization/d7c26c4b-676d-42f9-8419-39786d1dbfe6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BACON_[[:space:]]Improving[[:space:]]Clarity[[:space:]]of[[:space:]]Image[[:space:]]Captions[[:space:]]via[[:space:]]Bag-of-Concept[[:space:]]Graphs/8c3abc3e-93f5-4793-8009-1ee8f2b037f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BADGR_[[:space:]]Bundle[[:space:]]Adjustment[[:space:]]Diffusion[[:space:]]Conditioned[[:space:]]by[[:space:]]Gradients[[:space:]]for[[:space:]]Wide-Baseline[[:space:]]Floor[[:space:]]Plan[[:space:]]Reconstruction/9f03394e-0304-4f26-8e39-3f7ffb1da3c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BARD-GS_[[:space:]]Blur-Aware[[:space:]]Reconstruction[[:space:]]of[[:space:]]Dynamic[[:space:]]Scenes[[:space:]]via[[:space:]]Gaussian[[:space:]]Splatting/ac47de9d-73ff-480b-a66a-11b911e65974_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BASKET_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Video[[:space:]]Dataset[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Skill[[:space:]]Estimation/b29844b0-5214-47ff-9a9d-e26577935c65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BEVDiffuser_[[:space:]]Plug-and-Play[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]BEV[[:space:]]Denoising[[:space:]]with[[:space:]]Ground-Truth[[:space:]]Guidance/57d1ada5-4ed9-4887-8e4e-f2c6887f11c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BF-STVSR_[[:space:]]B-Splines[[:space:]]and[[:space:]]Fourier---Best[[:space:]]Friends[[:space:]]for[[:space:]]High[[:space:]]Fidelity[[:space:]]Spatial-Temporal[[:space:]]Video[[:space:]]Super-Resolution/3223e1f4-7d94-4347-82cc-6ec964cb0c74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BFANet_[[:space:]]Revisiting[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Boundary[[:space:]]Feature[[:space:]]Analysis/82fbd4e7-82b3-430e-80cf-009e2c800cba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BG-Triangle_[[:space:]]Bezier[[:space:]]Gaussian[[:space:]]Triangle[[:space:]]for[[:space:]]3D[[:space:]]Vectorization[[:space:]]and[[:space:]]Rendering/129f58af-a09f-453d-811c-9ef1f9006e65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BHViT_[[:space:]]Binarized[[:space:]]Hybrid[[:space:]]Vision[[:space:]]Transformer/6e75461b-1c95-481f-92de-996ed64d2223_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BIGS_[[:space:]]Bimanual[[:space:]]Category-agnostic[[:space:]]Interaction[[:space:]]Reconstruction[[:space:]]from[[:space:]]Monocular[[:space:]]Videos[[:space:]]via[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/a8f17bd9-f7a3-49ed-999f-a2c946e3f525_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BIMBA_[[:space:]]Selective-Scan[[:space:]]Compression[[:space:]]for[[:space:]]Long-Range[[:space:]]Video[[:space:]]Question[[:space:]]Answering/276bcb5d-3e7c-4cf3-9998-9a4d8d472f15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BIOMEDICA_[[:space:]]An[[:space:]]Open[[:space:]]Biomedical[[:space:]]Image-Caption[[:space:]]Archive,[[:space:]]Dataset,[[:space:]]and[[:space:]]Vision-Language[[:space:]]Models[[:space:]]Derived[[:space:]]from[[:space:]]Scientific[[:space:]]Literature/7da43da1-ca9e-48cf-86cd-195f9dd3efb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BIP3D_[[:space:]]Bridging[[:space:]]2D[[:space:]]Images[[:space:]]and[[:space:]]3D[[:space:]]Perception[[:space:]]for[[:space:]]Embodied[[:space:]]Intelligence/39e7694a-6cd6-4e45-bbda-300b6e75c734_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BLADE_[[:space:]]Single-view[[:space:]]Body[[:space:]]Mesh[[:space:]]Estimation[[:space:]]through[[:space:]]Accurate[[:space:]]Depth[[:space:]]Estimation/4ddb892b-dc77-4058-bfef-f3478f9456e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BOE-ViT_[[:space:]]Boosting[[:space:]]Orientation[[:space:]]Estimation[[:space:]]with[[:space:]]Equivariance[[:space:]]in[[:space:]]Self-Supervised[[:space:]]3D[[:space:]]Subtomogram[[:space:]]Alignment/ccb85c11-ba4e-4b8c-8218-c21e437a278e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BOLT_[[:space:]]Boost[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Model[[:space:]]Without[[:space:]]Training[[:space:]]for[[:space:]]Long-form[[:space:]]Video[[:space:]]Understanding/49ff0268-a890-45e3-9d66-3b5444c9283c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BOOTPLACE_[[:space:]]Bootstrapped[[:space:]]Object[[:space:]]Placement[[:space:]]with[[:space:]]Detection[[:space:]]Transformers/52f6625f-611c-4f94-8616-341138c295f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BWFormer_[[:space:]]Building[[:space:]]Wireframe[[:space:]]Reconstruction[[:space:]]from[[:space:]]Airborne[[:space:]]LiDAR[[:space:]]Point[[:space:]]Cloud[[:space:]]with[[:space:]]Transformer/4becc537-14ca-4148-bc38-b79915963628_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BadToken_[[:space:]]Token-level[[:space:]]Backdoor[[:space:]]Attacks[[:space:]]to[[:space:]]Multi-modal[[:space:]]Large[[:space:]]Language[[:space:]]Models/572f4b6d-7ccf-4325-96a8-040036f12a15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Balanced[[:space:]]Direction[[:space:]]from[[:space:]]Multifarious[[:space:]]Choices_[[:space:]]Arithmetic[[:space:]]Meta-Learning[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/57b88193-e564-457a-ac0c-f1145fa4a051_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Balanced[[:space:]]Rate-Distortion[[:space:]]Optimization[[:space:]]in[[:space:]]Learned[[:space:]]Image[[:space:]]Compression/6d2ad8ce-e46b-44bd-af79-ccbdb4c0fa56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Balancing[[:space:]]Two[[:space:]]Classifiers[[:space:]]via[[:space:]]A[[:space:]]Simplex[[:space:]]ETF[[:space:]]Structure[[:space:]]for[[:space:]]Model[[:space:]]Calibration/a6c860ea-d21f-40a6-b40b-37b30585cb75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bayesian[[:space:]]Prompt[[:space:]]Flow[[:space:]]Learning[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Anomaly[[:space:]]Detection/b958079c-6eb4-4a8f-b125-b9c958f9bb8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bayesian[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/9baebd5d-b536-4152-82cb-58b5dd6689d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Be[[:space:]]More[[:space:]]Specific_[[:space:]]Evaluating[[:space:]]Object-centric[[:space:]]Realism[[:space:]]in[[:space:]]Synthetic[[:space:]]Images/ef2a8afc-f330-4de9-abc2-1a6d9e97dc88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Believing[[:space:]]is[[:space:]]Seeing_[[:space:]]Unobserved[[:space:]]Object[[:space:]]Detection[[:space:]]using[[:space:]]Generative[[:space:]]Models/d2fc6493-b964-4d8a-8029-42d28c745ab3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Benchmarking[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]via[[:space:]]Directed[[:space:]]Scene[[:space:]]Graph[[:space:]]for[[:space:]]Comprehensive[[:space:]]Image[[:space:]]Captioning/b7584ffc-b264-4c11-8aa1-d670d278e02c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Benchmarking[[:space:]]Object[[:space:]]Detectors[[:space:]]under[[:space:]]Real-World[[:space:]]Distribution[[:space:]]Shifts[[:space:]]in[[:space:]]Satellite[[:space:]]Imagery/3d20364c-dd31-4da5-9906-f86101d5954e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Background[[:space:]]Shift_[[:space:]]Rethinking[[:space:]]Instance[[:space:]]Replay[[:space:]]in[[:space:]]Continual[[:space:]]Semantic[[:space:]]Segmentation/b683cfd3-120e-48b8-834a-7fb0b210ed01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Clean[[:space:]]Training[[:space:]]Data_[[:space:]]A[[:space:]]Versatile[[:space:]]and[[:space:]]Model-Agnostic[[:space:]]Framework[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection[[:space:]]with[[:space:]]Contaminated[[:space:]]Training[[:space:]]Data/61665345-4aac-4897-b5e3-036662e647f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Generation_[[:space:]]A[[:space:]]Diffusion-based[[:space:]]Low-level[[:space:]]Feature[[:space:]]Extractor[[:space:]]for[[:space:]]Detecting[[:space:]]AI-generated[[:space:]]Images/157abeba-883e-4b7c-9b07-e34c1403c974_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Human[[:space:]]Perception_[[:space:]]Understanding[[:space:]]Multi-Object[[:space:]]World[[:space:]]from[[:space:]]Monocular[[:space:]]View/840a32fb-8ccc-4f62-aad8-190ba7bea159_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Image[[:space:]]Classification_[[:space:]]A[[:space:]]Video[[:space:]]Benchmark[[:space:]]and[[:space:]]Dual-Branch[[:space:]]Hybrid[[:space:]]Discrimination[[:space:]]Framework[[:space:]]for[[:space:]]Compositional[[:space:]]Zero-Shot[[:space:]]Learning/6e350201-6bec-4ef4-b2b1-6c55a53bd843_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Local[[:space:]]Sharpness_[[:space:]]Communication-Efficient[[:space:]]Global[[:space:]]Sharpness-aware[[:space:]]Minimization[[:space:]]for[[:space:]]Federated[[:space:]]Learning/0dc01fcd-8f8c-4ed5-9c4b-4a338a08f6c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Sight_[[:space:]]Towards[[:space:]]Cognitive[[:space:]]Alignment[[:space:]]in[[:space:]]LVLM[[:space:]]via[[:space:]]Enriched[[:space:]]Visual[[:space:]]Knowledge/fbac1277-76d2-4974-82cd-9098b1fccc02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Single-Modal[[:space:]]Boundary_[[:space:]]Cross-Modal[[:space:]]Anomaly[[:space:]]Detection[[:space:]]through[[:space:]]Visual[[:space:]]Prototype[[:space:]]and[[:space:]]Harmonization/b1ddd5e3-ee03-4577-8891-4b4e1f190baa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Words_[[:space:]]Augmenting[[:space:]]Discriminative[[:space:]]Richness[[:space:]]via[[:space:]]Diffusions[[:space:]]in[[:space:]]Unsupervised[[:space:]]Prompt[[:space:]]Learning/4f7706f2-55fa-4f80-b140-80ab091081f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BiLoRA_[[:space:]]Almost-Orthogonal[[:space:]]Parameter[[:space:]]Spaces[[:space:]]for[[:space:]]Continual[[:space:]]Learning/f04994c2-4557-4d91-97e9-e74150c4d903_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BiM-VFI_[[:space:]]Bidirectional[[:space:]]Motion[[:space:]]Field-Guided[[:space:]]Frame[[:space:]]Interpolation[[:space:]]for[[:space:]]Video[[:space:]]with[[:space:]]Non-uniform[[:space:]]Motions/391ba5e7-3012-4525-a5a2-a95b08b8971f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bias[[:space:]]for[[:space:]]Action_[[:space:]]Video[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations[[:space:]]with[[:space:]]Bias[[:space:]]Modulation/f7a1591d-031d-4e2e-a0f2-3310032d1acc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BimArt_[[:space:]]A[[:space:]]Unified[[:space:]]Approach[[:space:]]for[[:space:]]the[[:space:]]Synthesis[[:space:]]of[[:space:]]3D[[:space:]]Bimanual[[:space:]]Interaction[[:space:]]with[[:space:]]Articulated[[:space:]]Objects/33f75bf8-3999-400f-a1dc-70b30ebe09ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Binarized[[:space:]]Mamba-Transformer[[:space:]]for[[:space:]]Lightweight[[:space:]]Quad[[:space:]]Bayer[[:space:]]HybridEVS[[:space:]]Demosaicing/9d10d303-4bfc-4340-b3e6-38bf1fce0add_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Binarized[[:space:]]Neural[[:space:]]Network[[:space:]]for[[:space:]]Multi-spectral[[:space:]]Image[[:space:]]Fusion/cf45dc70-89ce-4e2f-9acd-f581c76dcc49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BioX-CPath_[[:space:]]Biologically-driven[[:space:]]Explainable[[:space:]]Diagnostics[[:space:]]for[[:space:]]Multistain[[:space:]]IHC[[:space:]]Computational[[:space:]]Pathology/db4ad025-7e28-494f-be0d-dc35c7f0f44c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BiomedCoOp_[[:space:]]Learning[[:space:]]to[[:space:]]Prompt[[:space:]]for[[:space:]]Biomedical[[:space:]]Vision-Language[[:space:]]Models/155f5e71-e8cd-4b01-93c5-4ba3403dc70e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Birth[[:space:]]and[[:space:]]Death[[:space:]]of[[:space:]]a[[:space:]]Rose/a4640967-4c69-4562-a120-510929dfa81a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BizGen_[[:space:]]Advancing[[:space:]]Article-level[[:space:]]Visual[[:space:]]Text[[:space:]]Rendering[[:space:]]for[[:space:]]Infographics[[:space:]]Generation/ed39bd7d-522d-4724-b48f-3fc6eb5ee25e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Black[[:space:]]Hole-Driven[[:space:]]Identity[[:space:]]Absorbing[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/90114cf9-0603-437a-a8f1-2f354335df95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Black[[:space:]]Swan_[[:space:]]Abductive[[:space:]]and[[:space:]]Defeasible[[:space:]]Video[[:space:]]Reasoning[[:space:]]in[[:space:]]Unpredictable[[:space:]]Events/908835a5-9aab-4c65-ac7c-c0490d1eb1df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Black-Box[[:space:]]Forgery[[:space:]]Attacks[[:space:]]on[[:space:]]Semantic[[:space:]]Watermarks[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/ea69783c-e5d1-4330-9825-c381b9bf4cff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BlenderGym_[[:space:]]Benchmarking[[:space:]]Foundational[[:space:]]Model[[:space:]]Systems[[:space:]]for[[:space:]]Graphics[[:space:]]Editing/eb2a8fb6-8865-4ad1-bfdc-12a8dacc0741_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Blind[[:space:]]Bitstream-corrupted[[:space:]]Video[[:space:]]Recovery[[:space:]]via[[:space:]]Metadata-guided[[:space:]]Diffusion[[:space:]]Model/495a93a3-ae96-4942-93bf-5958f750616e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BlobGEN-Vid_[[:space:]]Compositional[[:space:]]Text-to-Video[[:space:]]Generation[[:space:]]with[[:space:]]Blob[[:space:]]Video[[:space:]]Representations/6976921e-69b0-4753-8652-9e7010eba2fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BlockDance_[[:space:]]Reuse[[:space:]]Structurally[[:space:]]Similar[[:space:]]Spatio-Temporal[[:space:]]Features[[:space:]]to[[:space:]]Accelerate[[:space:]]Diffusion[[:space:]]Transformers/4f957fb1-12b7-4cbc-9aae-847efa4ee1ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Blood[[:space:]]Flow[[:space:]]Speed[[:space:]]Estimation[[:space:]]with[[:space:]]Optical[[:space:]]Coherence[[:space:]]Tomography[[:space:]]Angiography[[:space:]]Images/d29ec0b4-630e-43b8-af71-0d27472915dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BlueLM-V-3B_[[:space:]]Algorithm[[:space:]]and[[:space:]]System[[:space:]]Co-Design[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]on[[:space:]]Mobile[[:space:]]Devices/bc30def9-6b04-4755-b336-e585d917cd9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Blurred[[:space:]]LiDAR[[:space:]]for[[:space:]]Sharper[[:space:]]3D_[[:space:]]Robust[[:space:]]Handheld[[:space:]]3D[[:space:]]Scanning[[:space:]]with[[:space:]]Diffuse[[:space:]]LiDAR[[:space:]]and[[:space:]]RGB/cb7e99ae-f74b-49a4-a495-94f62b3c2b2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Blurry-Edges_[[:space:]]Photon-Limited[[:space:]]Depth[[:space:]]Estimation[[:space:]]from[[:space:]]Defocused[[:space:]]Boundaries/07d6169f-210a-403e-bb2f-775bee4b6020_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boltzmann[[:space:]]Attention[[:space:]]Sampling[[:space:]]for[[:space:]]Image[[:space:]]Analysis[[:space:]]with[[:space:]]Small[[:space:]]Objects/20fc079e-62c4-47d2-8684-3fb72c584a2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BooW-VTON_[[:space:]]Boosting[[:space:]]In-the-Wild[[:space:]]Virtual[[:space:]]Try-On[[:space:]]via[[:space:]]Mask-Free[[:space:]]Pseudo[[:space:]]Data[[:space:]]Training/de11fdf9-7a0a-46d0-a2ae-9ee2c6295e52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boost[[:space:]]Your[[:space:]]Human[[:space:]]Image[[:space:]]Generation[[:space:]]Model[[:space:]]via[[:space:]]Direct[[:space:]]Preference[[:space:]]Optimization/5c5df61e-21e6-4ef6-93b4-ad0df36841e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boost[[:space:]]the[[:space:]]Inference[[:space:]]with[[:space:]]Co-training_[[:space:]]A[[:space:]]Depth-guided[[:space:]]Mutual[[:space:]]Learning[[:space:]]Framework[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Medical[[:space:]]Polyp[[:space:]]Segmentation/217938b7-5f6c-415d-b6b3-3f6d6c079f67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]Adversarial[[:space:]]Transferability[[:space:]]through[[:space:]]Augmentation[[:space:]]in[[:space:]]Hypothesis[[:space:]]Space/75f50a4b-5e6f-4ecb-a5c4-e02d1dfbf27e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]Domain[[:space:]]Incremental[[:space:]]Learning_[[:space:]]Selecting[[:space:]]the[[:space:]]Optimal[[:space:]]Parameters[[:space:]]is[[:space:]]All[[:space:]]You[[:space:]]Need/dbf55c45-02d8-4273-874f-49d14b66bc8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]Point-Supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization[[:space:]]through[[:space:]]Integrating[[:space:]]Query[[:space:]]Reformation[[:space:]]and[[:space:]]Optimal[[:space:]]Transport/c2a0db74-43f3-496e-9072-8751c3fa27a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]the[[:space:]]Dual-Stream[[:space:]]Architecture[[:space:]]in[[:space:]]Ultra-High[[:space:]]Resolution[[:space:]]Segmentation[[:space:]]with[[:space:]]Resolution-Biased[[:space:]]Uncertainty[[:space:]]Estimation/e1207d24-0d5e-460d-af2a-b7620343ffda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bootstrap[[:space:]]Your[[:space:]]Own[[:space:]]Views_[[:space:]]Masked[[:space:]]Ego-Exo[[:space:]]Modeling[[:space:]]for[[:space:]]Fine-grained[[:space:]]View-invariant[[:space:]]Video[[:space:]]Representations/49d5397c-dbff-41cc-90fa-0e5057e8a3b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Brain-Inspired[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks[[:space:]]for[[:space:]]Energy-Efficient[[:space:]]Object[[:space:]]Detection/76d2dc19-d363-4394-b985-2a71aba018ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Breaking[[:space:]]the[[:space:]]Low-Rank[[:space:]]Dilemma[[:space:]]of[[:space:]]Linear[[:space:]]Attention/0e720e7a-aad2-4126-a669-d3c76e88e3d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Breaking[[:space:]]the[[:space:]]Memory[[:space:]]Barrier[[:space:]]of[[:space:]]Contrastive[[:space:]]Loss[[:space:]]via[[:space:]]Tile-Based[[:space:]]Strategy/9fc7b8bc-e02a-4191-bee7-7c88b5538195_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BrepGiff_[[:space:]]Lightweight[[:space:]]Generation[[:space:]]of[[:space:]]Complex[[:space:]]B-rep[[:space:]]with[[:space:]]3D[[:space:]]GAT[[:space:]]Diffusion/3e31ad72-3f7f-4f3a-a815-ec3c607364b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridge[[:space:]]Frame[[:space:]]and[[:space:]]Event_[[:space:]]Common[[:space:]]Spatiotemporal[[:space:]]Fusion[[:space:]]for[[:space:]]High-Dynamic[[:space:]]Scene[[:space:]]Optical[[:space:]]Flow/bf14a18b-401b-404c-a3b2-75268bda526a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridge[[:space:]]the[[:space:]]Gap_[[:space:]]From[[:space:]]Weak[[:space:]]to[[:space:]]Full[[:space:]]Supervision[[:space:]]for[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization[[:space:]]with[[:space:]]PseudoFormer/b4bc4b17-5e9d-41f3-a2ce-415013684094_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]Gait[[:space:]]Recognition[[:space:]]and[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]Sequence[[:space:]]Modeling/136d07ea-62db-4312-9b2a-f03c2cd38f30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]Modalities_[[:space:]]Improving[[:space:]]Universal[[:space:]]Multimodal[[:space:]]Retrieval[[:space:]]by[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/13b1783d-8ad2-43ae-8182-3c6059b9c476_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]Past[[:space:]]and[[:space:]]Future_[[:space:]]End-to-End[[:space:]]Autonomous[[:space:]]Driving[[:space:]]with[[:space:]]Historical[[:space:]]Prediction[[:space:]]and[[:space:]]Planning/581fdaae-0f43-4223-861e-e8fae7adedb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]Viewpoint[[:space:]]Gaps_[[:space:]]Geometric[[:space:]]Reasoning[[:space:]]Boosts[[:space:]]Semantic[[:space:]]Correspondence/e51209a8-45f8-4823-90b4-3a1a3cc851a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]the[[:space:]]Gap[[:space:]]between[[:space:]]Gaussian[[:space:]]Diffusion[[:space:]]Models[[:space:]]and[[:space:]]Universal[[:space:]]Quantization[[:space:]]for[[:space:]]Image[[:space:]]Compression/3236a89e-cee7-41fa-a024-b3a58198759e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]the[[:space:]]Vision-Brain[[:space:]]Gap[[:space:]]with[[:space:]]an[[:space:]]Uncertainty-Aware[[:space:]]Blur[[:space:]]Prior/0cc4bbc6-832b-4b60-97ae-22ce76a00b2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bringing[[:space:]]CLIP[[:space:]]to[[:space:]]the[[:space:]]Clinic_[[:space:]]Dynamic[[:space:]]Soft[[:space:]]Labels[[:space:]]and[[:space:]]Negation-Aware[[:space:]]Learning[[:space:]]for[[:space:]]Medical[[:space:]]Analysis/2fa5a86b-cfe9-4357-affa-217eba2436b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Buffer[[:space:]]Anytime_[[:space:]]Zero-Shot[[:space:]]Video[[:space:]]Depth[[:space:]]and[[:space:]]Normal[[:space:]]from[[:space:]]Image[[:space:]]Priors/33132efc-4829-4646-9bac-a3c3441ad085_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Building[[:space:]]Vision[[:space:]]Models[[:space:]]upon[[:space:]]Heat[[:space:]]Conduction/d4ff1b0b-04d5-4bc1-9e6d-94d640bae102_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Building[[:space:]]a[[:space:]]Mind[[:space:]]Palace_[[:space:]]Structuring[[:space:]]Environment-Grounded[[:space:]]Semantic[[:space:]]Graphs[[:space:]]for[[:space:]]Effective[[:space:]]Long[[:space:]]Video[[:space:]]Analysis[[:space:]]with[[:space:]]LLMs/8fe84805-0a2c-4530-a88c-2fd10242c919_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ByTheWay_[[:space:]]Boost[[:space:]]Your[[:space:]]Text-to-Video[[:space:]]Generation[[:space:]]Model[[:space:]]to[[:space:]]Higher[[:space:]]Quality[[:space:]]in[[:space:]]a[[:space:]]Training-free[[:space:]]Way/09422adb-d2e2-407d-b382-a88ee3e15bba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CAD-Llama_[[:space:]]Leveraging[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Computer-Aided[[:space:]]Design[[:space:]]Parametric[[:space:]]3D[[:space:]]Model[[:space:]]Generation/7ad4689e-e865-4429-9b11-6f6e6eb7445c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CADCrafter_[[:space:]]Generating[[:space:]]Computer-Aided[[:space:]]Design[[:space:]]Models[[:space:]]from[[:space:]]Unconstrained[[:space:]]Images/3e257b8e-f293-43ec-af1f-613c65f899d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CADDreamer_[[:space:]]CAD[[:space:]]Object[[:space:]]Generation[[:space:]]from[[:space:]]Single-view[[:space:]]Images/3933cb0c-8844-4553-95cf-a17192766870_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CADRef_[[:space:]]Robust[[:space:]]Out-of-Distribution[[:space:]]Detection[[:space:]]via[[:space:]]Class-Aware[[:space:]]Decoupled[[:space:]]Relative[[:space:]]Feature[[:space:]]Leveraging/2e07023a-8bb1-40f7-9bd1-324af5a637cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CALICO_[[:space:]]Part-Focused[[:space:]]Semantic[[:space:]]Co-Segmentation[[:space:]]with[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/962ce1bb-3916-479b-86fc-f94ae7a24616_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CAP-Net_[[:space:]]A[[:space:]]Unified[[:space:]]Network[[:space:]]for[[:space:]]6D[[:space:]]Pose[[:space:]]and[[:space:]]Size[[:space:]]Estimation[[:space:]]of[[:space:]]Categorical[[:space:]]Articulated[[:space:]]Parts[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]RGB-D[[:space:]]Image/0bca52ad-de15-47da-9953-9fb29f9855a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CAP4D_[[:space:]]Creating[[:space:]]Animatable[[:space:]]4D[[:space:]]Portrait[[:space:]]Avatars[[:space:]]with[[:space:]]Morphable[[:space:]]Multi-View[[:space:]]Diffusion[[:space:]]Models/cb8f417b-f654-4729-9474-3e8bd890e6a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CARE[[:space:]]Transformer_[[:space:]]Mobile-Friendly[[:space:]]Linear[[:space:]]Visual[[:space:]]Transformer[[:space:]]via[[:space:]]Decoupled[[:space:]]Dual[[:space:]]Interaction/1763756c-cf7b-4dba-b43a-6e6eed28da83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CARL_[[:space:]]A[[:space:]]Framework[[:space:]]for[[:space:]]Equivariant[[:space:]]Image[[:space:]]Registration/6281d56e-e5f8-4dd0-ac12-68395ad5223c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CASAGPT_[[:space:]]Cuboid[[:space:]]Arrangement[[:space:]]and[[:space:]]Scene[[:space:]]Assembly[[:space:]]for[[:space:]]Interior[[:space:]]Design/bace6851-472a-4fa4-a602-6cf71cd7a977_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CASP_[[:space:]]Compression[[:space:]]of[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models[[:space:]]Based[[:space:]]on[[:space:]]Attention[[:space:]]Sparsity/cd4182c3-7bac-40c1-9a8a-fa47fb9daa6b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CASP_[[:space:]]Consistency-aware[[:space:]]Audio-induced[[:space:]]Saliency[[:space:]]Prediction[[:space:]]Model[[:space:]]for[[:space:]]Omnidirectional[[:space:]]Video/878d16ba-8c28-4f23-86d7-1532e7b18f33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CAT4D_[[:space:]]Create[[:space:]]Anything[[:space:]]in[[:space:]]4D[[:space:]]with[[:space:]]Multi-View[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/0f794068-6b77-4bd2-9fa2-c2b86ce1cab1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CATANet_[[:space:]]Efficient[[:space:]]Content-Aware[[:space:]]Token[[:space:]]Aggregation[[:space:]]for[[:space:]]Lightweight[[:space:]]Image[[:space:]]Super-Resolution/da9594c6-0714-466a-a000-a7999e3303d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CAV-MAE[[:space:]]Sync_[[:space:]]Improving[[:space:]]Contrastive[[:space:]]Audio-Visual[[:space:]]Mask[[:space:]]Autoencoders[[:space:]]via[[:space:]]Fine-Grained[[:space:]]Alignment/ac5daa03-60a0-4a2e-893c-2461bcb92532_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CCIN_[[:space:]]Compositional[[:space:]]Conflict[[:space:]]Identification[[:space:]]and[[:space:]]Neutralization[[:space:]]for[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/35c4fe97-84b9-45f4-a804-9a6dbcc48240_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CDI_[[:space:]]Copyrighted[[:space:]]Data[[:space:]]Identification[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/1d82cbd3-0183-42f3-bd30-42cd3c7c7e2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CGMatch_[[:space:]]A[[:space:]]Different[[:space:]]Perspective[[:space:]]of[[:space:]]Semi-supervised[[:space:]]Learning/fd435828-c9b8-46b5-85c0-ffee072717ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CH3Depth_[[:space:]]Efficient[[:space:]]and[[:space:]]Flexible[[:space:]]Depth[[:space:]]Foundation[[:space:]]Model[[:space:]]with[[:space:]]Flow[[:space:]]Matching/8a4ee827-a8c5-4c91-8ca6-fd4d182ddcac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CL-LoRA_[[:space:]]Continual[[:space:]]Low-Rank[[:space:]]Adaptation[[:space:]]for[[:space:]]Rehearsal-Free[[:space:]]Class-Incremental[[:space:]]Learning/a6d48c97-f125-4f66-ac49-756ef97d1392_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CL-MoE_[[:space:]]Enhancing[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]with[[:space:]]Dual[[:space:]]Momentum[[:space:]]Mixture-of-Experts[[:space:]]for[[:space:]]Continual[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/e58f9270-0014-48be-b85b-73cb64f7ff43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CLIP[[:space:]]Under[[:space:]]the[[:space:]]Microscope_[[:space:]]A[[:space:]]Fine-Grained[[:space:]]Analysis[[:space:]]of[[:space:]]Multi-Object[[:space:]]Representation/3fd6077e-e9cd-4ee1-9c36-3d02e4308b34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CLIP[[:space:]]is[[:space:]]Almost[[:space:]]All[[:space:]]You[[:space:]]Need_[[:space:]]Towards[[:space:]]Parameter-Efficient[[:space:]]Scene[[:space:]]Text[[:space:]]Retrieval[[:space:]]without[[:space:]]OCR/be221c82-1123-4958-b065-9e56a1c165fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CLIP[[:space:]]is[[:space:]]Strong[[:space:]]Enough[[:space:]]to[[:space:]]Fight[[:space:]]Back_[[:space:]]Test-time[[:space:]]Counterattacks[[:space:]]towards[[:space:]]Zero-shot[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]of[[:space:]]CLIP/6e055581-3cab-438c-aef5-95d068c083f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CLIP-driven[[:space:]]Coarse-to-fine[[:space:]]Semantic[[:space:]]Guidance[[:space:]]for[[:space:]]Fine-grained[[:space:]]Open-set[[:space:]]Semi-supervised[[:space:]]Learning/9ce9b318-5f64-4388-87fa-a22c3aef407c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CLOC_[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Ordinal[[:space:]]Classification[[:space:]]with[[:space:]]Multi-Margin[[:space:]]N-pair[[:space:]]Loss/e846628e-5411-48c1-9cdc-8c44871e0687_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CMMLoc_[[:space:]]Advancing[[:space:]]Text-to-PointCloud[[:space:]]Localization[[:space:]]with[[:space:]]Cauchy-Mixture-Model[[:space:]]Based[[:space:]]Framework/1dfeea1e-91b2-402e-8dc7-843047c17875_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CO-SPY_[[:space:]]Combining[[:space:]]Semantic[[:space:]]and[[:space:]]Pixel[[:space:]]Features[[:space:]]to[[:space:]]Detect[[:space:]]Synthetic[[:space:]]Images[[:space:]]by[[:space:]]AI/895e1d5b-a728-4c68-9fc4-06c986132a06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/COAP_[[:space:]]Memory-Efficient[[:space:]]Training[[:space:]]with[[:space:]]Correlation-Aware[[:space:]]Gradient[[:space:]]Projection/53c2222b-25b5-492e-9176-3ffa15fd8c4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/COB-GS_[[:space:]]Clear[[:space:]]Object[[:space:]]Boundaries[[:space:]]in[[:space:]]3DGS[[:space:]]Segmentation[[:space:]]Based[[:space:]]on[[:space:]]Boundary-Adaptive[[:space:]]Gaussian[[:space:]]Splitting/1dc33ea0-9c6e-473e-8e73-541b3680da0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/COBRA_[[:space:]]COmBinatorial[[:space:]]Retrieval[[:space:]]Augmentation[[:space:]]for[[:space:]]Few-Shot[[:space:]]Adaptation/161fb516-c2c6-45bc-b0e0-90d841895985_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CORE4D_[[:space:]]A[[:space:]]4D[[:space:]]Human-Object-Human[[:space:]]Interaction[[:space:]]Dataset[[:space:]]for[[:space:]]Collaborative[[:space:]]Object[[:space:]]REarrangement/c6db2d6c-6ee0-4911-b8c6-79b57897b35d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/COSMIC_[[:space:]]Clique-Oriented[[:space:]]Semantic[[:space:]]Multi-space[[:space:]]Integration[[:space:]]for[[:space:]]Robust[[:space:]]CLIP[[:space:]]Test-Time[[:space:]]Adaptation/680d2abd-2333-4aae-8645-882dd224440d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/COSMOS_[[:space:]]Cross-Modality[[:space:]]Self-Distillation[[:space:]]for[[:space:]]Vision[[:space:]]Language[[:space:]]Pre-training/2bbf48d8-e332-4dc3-883d-a0c0a2c937bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/COUNTS_[[:space:]]Benchmarking[[:space:]]Object[[:space:]]Detectors[[:space:]]and[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]under[[:space:]]Distribution[[:space:]]Shifts/c334fb63-5558-4496-967c-c86ceb9ab3f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CPath-Omni_[[:space:]]A[[:space:]]Unified[[:space:]]Multimodal[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Patch[[:space:]]and[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Analysis[[:space:]]in[[:space:]]Computational[[:space:]]Pathology/ef06b74c-d1a0-409f-8bc1-29348738e571_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CRISP_[[:space:]]Object[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Estimation[[:space:]]with[[:space:]]Test-Time[[:space:]]Adaptation/2da9705a-b1da-443a-bd91-631670c37326_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CSC-PA_[[:space:]]Cross-image[[:space:]]Semantic[[:space:]]Correlation[[:space:]]via[[:space:]]Prototype[[:space:]]Attentions[[:space:]]for[[:space:]]Single-network[[:space:]]Semi-supervised[[:space:]]Breast[[:space:]]Tumor[[:space:]]Segmentation/be954a07-555f-4d65-b666-6097b7d0edb3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CTRL-D_[[:space:]]Controllable[[:space:]]Dynamic[[:space:]]3D[[:space:]]Scene[[:space:]]Editing[[:space:]]with[[:space:]]Personalized[[:space:]]2D[[:space:]]Diffusion/72cc7e5b-03ae-4580-89f7-87467e032211_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CTRL-O_[[:space:]]Language-Controllable[[:space:]]Object-Centric[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/ee0a3002-4312-460d-9bfa-5f6fd080f431_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CXPMRG-Bench_[[:space:]]Pre-training[[:space:]]and[[:space:]]Benchmarking[[:space:]]for[[:space:]]X-ray[[:space:]]Medical[[:space:]]Report[[:space:]]Generation[[:space:]]on[[:space:]]CheXpert[[:space:]]Plus[[:space:]]Dataset/c9b0bf65-7355-47aa-9d02-f7ecad349422_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CaMuViD_[[:space:]]Calibration-Free[[:space:]]Multi-View[[:space:]]Detection/c8014b9a-f82f-4722-8921-f863b5d2bda9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CacheQuant_[[:space:]]Comprehensively[[:space:]]Accelerated[[:space:]]Diffusion[[:space:]]Models/cf4665fc-d069-4342-9c5e-95c15ceb3458_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Calibrated[[:space:]]Multi-Preference[[:space:]]Optimization[[:space:]]for[[:space:]]Aligning[[:space:]]Diffusion[[:space:]]Models/ec76af2e-064a-48a0-abe6-7911a91221c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CamFreeDiff_[[:space:]]Camera-free[[:space:]]Image[[:space:]]to[[:space:]]Panorama[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion[[:space:]]Model/a6b50bbd-9f45-4d2a-908f-d210df71485e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CamPoint_[[:space:]]Boosting[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation[[:space:]]with[[:space:]]Virtual[[:space:]]Camera/16ed545c-f7f6-4b99-bdc7-95f6994d97d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Camera[[:space:]]Resection[[:space:]]from[[:space:]]Known[[:space:]]Line[[:space:]]Pencils[[:space:]]and[[:space:]]a[[:space:]]Radially[[:space:]]Distorted[[:space:]]Scanline/72c028c9-d5ff-443f-892b-78a1cd9a2298_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Camouflage[[:space:]]Anything_[[:space:]]Learning[[:space:]]to[[:space:]]Hide[[:space:]]using[[:space:]]Controlled[[:space:]]Out-painting[[:space:]]and[[:space:]]Representation[[:space:]]Engineering/c1644fc2-14d9-4887-88ea-c36155ce48f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Can[[:space:]]Generative[[:space:]]Video[[:space:]]Models[[:space:]]Help[[:space:]]Pose[[:space:]]Estimation_/49fefa3a-b1ac-457b-9db4-a63dd6fd7b4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Can[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]Correct[[:space:]]Semantic[[:space:]]Grounding[[:space:]]Errors[[:space:]]By[[:space:]]Themselves_/f80463a1-e907-4552-b464-0249757e80d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Can[[:space:]]Machines[[:space:]]Understand[[:space:]]Composition_[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Photographic[[:space:]]Image[[:space:]]Composition[[:space:]]Embedding[[:space:]]and[[:space:]]Understanding/34f1fde4-af95-4567-ac77-bf46b296783c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Can[[:space:]]Text-to-Video[[:space:]]Generation[[:space:]]help[[:space:]]Video-Language[[:space:]]Alignment_/a0988973-5da0-422c-8966-cc914358f7bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Can't[[:space:]]Slow[[:space:]]Me[[:space:]]Down_[[:space:]]Learning[[:space:]]Robust[[:space:]]and[[:space:]]Hardware-Adaptive[[:space:]]Object[[:space:]]Detectors[[:space:]]against[[:space:]]Latency[[:space:]]Attacks[[:space:]]for[[:space:]]Edge[[:space:]]Devices/358c15a0-1a7c-414a-a5e9-d4dc22a374cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CarPlanner_[[:space:]]Consistent[[:space:]]Auto-regressive[[:space:]]Trajectory[[:space:]]Planning[[:space:]]for[[:space:]]Large-Scale[[:space:]]Reinforcement[[:space:]]Learning[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/d414c93b-a7f5-4be3-a5c4-e2221f3c1cee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CaricatureBooth_[[:space:]]Data-Free[[:space:]]Interactive[[:space:]]Caricature[[:space:]]Generation[[:space:]]in[[:space:]]a[[:space:]]Photo[[:space:]]Booth/11a6e206-d825-486f-bd24-1a86424cc9eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Category-Agnostic[[:space:]]Neural[[:space:]]Object[[:space:]]Rigging/fa215e46-46df-4902-8cc2-d682acf2a35f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Causal[[:space:]]Composition[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Closed-loop[[:space:]]Traffic[[:space:]]Generation/269a9e6f-c4bb-41d1-bf85-d32fd39ad305_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Certified[[:space:]]Human[[:space:]]Trajectory[[:space:]]Prediction/8c8a5965-baee-4909-bfbf-e339d49281b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Chain[[:space:]]of[[:space:]]Attack_[[:space:]]On[[:space:]]the[[:space:]]Robustness[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models[[:space:]]Against[[:space:]]Transfer-Based[[:space:]]Adversarial[[:space:]]Attacks/4e78f406-66b0-4728-91fc-98189e13420e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Chain[[:space:]]of[[:space:]]Semantics[[:space:]]Programming[[:space:]]in[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Representation[[:space:]]for[[:space:]]3D[[:space:]]Vision[[:space:]]Grounding/0f72fb5d-a469-4fd4-b3fb-2e6e19e60074_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ChainHOI_[[:space:]]Joint-based[[:space:]]Kinematic[[:space:]]Chain[[:space:]]Modeling[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Generation/4550042c-6e98-4dbc-b87f-acb949bc7a09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Change3D_[[:space:]]Revisiting[[:space:]]Change[[:space:]]Detection[[:space:]]and[[:space:]]Captioning[[:space:]]from[[:space:]]A[[:space:]]Video[[:space:]]Modeling[[:space:]]Perspective/bf102372-b0ac-450f-98e7-1e9b37a55194_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Channel[[:space:]]Consistency[[:space:]]Prior[[:space:]]and[[:space:]]Self-Reconstruction[[:space:]]Strategy[[:space:]]Based[[:space:]]Unsupervised[[:space:]]Image[[:space:]]Deraining/1cd88811-9c34-4363-ae8a-2a52dd5a816f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Channel-wise[[:space:]]Noise[[:space:]]Scheduled[[:space:]]Diffusion[[:space:]]for[[:space:]]Inverse[[:space:]]Rendering[[:space:]]in[[:space:]]Indoor[[:space:]]Scenes/e287505c-edc5-4f3b-9021-f507eb27467f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Chapter-Llama_[[:space:]]Efficient[[:space:]]Chaptering[[:space:]]in[[:space:]]Hour-Long[[:space:]]Videos[[:space:]]with[[:space:]]LLMs/c5a48e4c-7b19-4801-9ffe-45905e053adb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Charm_[[:space:]]The[[:space:]]Missing[[:space:]]Piece[[:space:]]in[[:space:]]ViT[[:space:]]Fine-Tuning[[:space:]]for[[:space:]]Image[[:space:]]Aesthetic[[:space:]]Assessment/0ec61778-fb6e-453d-bfdb-3d054143034b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Chat-based[[:space:]]Person[[:space:]]Retrieval[[:space:]]via[[:space:]]Dialogue-Refined[[:space:]]Cross-Modal[[:space:]]Alignment/6896dcb7-a954-4e52-90ab-4e44608cdd32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Chat2SVG_[[:space:]]Vector[[:space:]]Graphics[[:space:]]Generation[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]and[[:space:]]Image[[:space:]]Diffusion[[:space:]]Models/03e11a55-7ca3-4174-a887-3307219befcc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ChatGarment_[[:space:]]Garment[[:space:]]Estimation,[[:space:]]Generation[[:space:]]and[[:space:]]Editing[[:space:]]via[[:space:]]Large[[:space:]]Language[[:space:]]Models/425b09e1-924e-4a43-b277-eb8a41387d21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ChatGen_[[:space:]]Automatic[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]From[[:space:]]FreeStyle[[:space:]]Chatting/0c8e9c4a-231c-4c6d-af84-cd3eb4170579_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ChatHuman_[[:space:]]Chatting[[:space:]]about[[:space:]]3D[[:space:]]Humans[[:space:]]with[[:space:]]Tools/a1962ae6-92af-4828-9f1c-48e34ea37fec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CheXWorld_[[:space:]]Exploring[[:space:]]Image[[:space:]]World[[:space:]]Modeling[[:space:]]for[[:space:]]Radiograph[[:space:]]Representation[[:space:]]Learning/f79a692c-48ff-4b64-8cb4-11cdf9303afb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CheXwhatsApp_[[:space:]]A[[:space:]]Dataset[[:space:]]for[[:space:]]Exploring[[:space:]]Challenges[[:space:]]in[[:space:]]the[[:space:]]Diagnosis[[:space:]]of[[:space:]]Chest[[:space:]]X-rays[[:space:]]through[[:space:]]Mobile[[:space:]]Devices/8cf11147-8ad4-4f89-a3ee-6b5492c8beb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cheb-GR_[[:space:]]Rethinking[[:space:]]K-nearest[[:space:]]Neighbor[[:space:]]Search[[:space:]]in[[:space:]]Re-ranking[[:space:]]for[[:space:]]Person[[:space:]]Re-identification/9e933435-e0fc-412e-b745-089375b60f23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Chebyshev[[:space:]]Attention[[:space:]]Depth[[:space:]]Permutation[[:space:]]Texture[[:space:]]Network[[:space:]]with[[:space:]]Latent[[:space:]]Texture[[:space:]]Attribute[[:space:]]Loss/44aa142c-41b8-482e-a039-e4a4fc955e21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CheckManual_[[:space:]]A[[:space:]]New[[:space:]]Challenge[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Manual-based[[:space:]]Appliance[[:space:]]Manipulation/c1aa26f1-f8e2-45ab-92dc-640acaf534e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CholecTrack20_[[:space:]]A[[:space:]]Multi-Perspective[[:space:]]Tracking[[:space:]]Dataset[[:space:]]for[[:space:]]Surgical[[:space:]]Tools/3671c43b-afb0-4272-8573-cad641013bba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Circumventing[[:space:]]Shortcuts[[:space:]]in[[:space:]]Audio-visual[[:space:]]Deepfake[[:space:]]Detection[[:space:]]Datasets[[:space:]]with[[:space:]]Unsupervised[[:space:]]Learning/371069ba-8973-4b9c-bb98-ecda46029053_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CityWalker_[[:space:]]Learning[[:space:]]Embodied[[:space:]]Urban[[:space:]]Navigation[[:space:]]from[[:space:]]Web-Scale[[:space:]]Videos/a9edaa72-0d38-4c4e-a25a-9480eb60d753_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Classic[[:space:]]Video[[:space:]]Denoising[[:space:]]in[[:space:]]a[[:space:]]Machine[[:space:]]Learning[[:space:]]World_[[:space:]]Robust,[[:space:]]Fast,[[:space:]]and[[:space:]]Controllable/079f9f38-09f8-4442-836f-ae05472182ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Classifier-Free[[:space:]]Guidance[[:space:]]Inside[[:space:]]the[[:space:]]Attraction[[:space:]]Basin[[:space:]]May[[:space:]]Cause[[:space:]]Memorization/a1a30dde-51b7-469a-a404-0c4a54031538_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Classifier-guided[[:space:]]CLIP[[:space:]]Distillation[[:space:]]for[[:space:]]Unsupervised[[:space:]]Multi-label[[:space:]]Classification/325c3ac3-6589-4c70-b2e1-2f7ac7877553_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Classifier-to-Bias_[[:space:]]Toward[[:space:]]Unsupervised[[:space:]]Automatic[[:space:]]Bias[[:space:]]Detection[[:space:]]for[[:space:]]Visual[[:space:]]Classifiers/303d2f58-e46e-4b7a-806e-0d4e64d53df2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CleanDIFT_[[:space:]]Diffusion[[:space:]]Features[[:space:]]without[[:space:]]Noise/17f652ff-055a-4bf1-a266-9144370345a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ClearSight_[[:space:]]Visual[[:space:]]Signal[[:space:]]Enhancement[[:space:]]for[[:space:]]Object[[:space:]]Hallucination[[:space:]]Mitigation[[:space:]]in[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/cc596546-7afd-468d-96ec-564a706a95d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ClimbingCap_[[:space:]]Multi-Modal[[:space:]]Dataset[[:space:]]and[[:space:]]Method[[:space:]]for[[:space:]]Rock[[:space:]]Climbing[[:space:]]in[[:space:]]World[[:space:]]Coordinate/56189f24-e199-449b-838e-e830c93475ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Closed-Loop[[:space:]]Supervised[[:space:]]Fine-Tuning[[:space:]]of[[:space:]]Tokenized[[:space:]]Traffic[[:space:]]Models/2dc3d2cc-243e-4ca0-9955-c4c727b1c79c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Closest[[:space:]]Neighbors[[:space:]]are[[:space:]]Harmful[[:space:]]for[[:space:]]Lightweight[[:space:]]Masked[[:space:]]Auto-encoders/ce070f35-5bfa-41af-9465-b1f84c742005_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Co-op_[[:space:]]Correspondence-based[[:space:]]Novel[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/5a137c7c-5457-4cd6-bb8c-541c7e90dee7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoA_[[:space:]]Towards[[:space:]]Real[[:space:]]Image[[:space:]]Dehazing[[:space:]]via[[:space:]]Compression-and-Adaptation/06a0b6a5-a641-41a6-99e5-3f3c8a8c1a70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoCoGaussian_[[:space:]]Leveraging[[:space:]]Circle[[:space:]]of[[:space:]]Confusion[[:space:]]for[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Defocused[[:space:]]Images/c9f7dcb9-6d18-430a-9426-253bcaf437ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoE_[[:space:]]Chain-of-Explanation[[:space:]]via[[:space:]]Automatic[[:space:]]Visual[[:space:]]Concept[[:space:]]Circuit[[:space:]]Description[[:space:]]and[[:space:]]Polysemanticity[[:space:]]Quantification/ed717be1-08cc-4195-89dc-b7ee977099ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoLLM_[[:space:]]A[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/02729911-7a78-4d55-af78-f50b5e45a135_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoMBO_[[:space:]]Conflict[[:space:]]Mitigation[[:space:]]via[[:space:]]Branched[[:space:]]Optimization[[:space:]]for[[:space:]]Class[[:space:]]Incremental[[:space:]]Segmentation/a4944725-d45e-4b4f-a404-3fd5f2ee3b6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoMM_[[:space:]]A[[:space:]]Coherent[[:space:]]Interleaved[[:space:]]Image-Text[[:space:]]Dataset[[:space:]]for[[:space:]]Multimodal[[:space:]]Understanding[[:space:]]and[[:space:]]Generation/4d3dc3b4-7f37-464f-b503-0ded8f6008de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoMapGS_[[:space:]]Covisibility[[:space:]]Map-based[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Sparse[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/025e1d28-cc10-46ad-8c05-b01e6a4dff3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoMatcher_[[:space:]]Multi-View[[:space:]]Collaborative[[:space:]]Feature[[:space:]]Matching/b3062684-7794-47f2-b9d1-53d25767d94f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoSDH_[[:space:]]Communication-Efficient[[:space:]]Collaborative[[:space:]]Perception[[:space:]]via[[:space:]]Supply-Demand[[:space:]]Awareness[[:space:]]and[[:space:]]Intermediate-Late[[:space:]]Hybridization/82cb35c7-62f2-4e6d-b4eb-be441dfd4b29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoSER_[[:space:]]Towards[[:space:]]Consistent[[:space:]]Dense[[:space:]]Multiview[[:space:]]Text-to-Image[[:space:]]Generator[[:space:]]for[[:space:]]3D[[:space:]]Creation/911def72-ec15-4b90-8a72-6946e2b20044_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoSpace_[[:space:]]Benchmarking[[:space:]]Continuous[[:space:]]Space[[:space:]]Perception[[:space:]]Ability[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/45f50ce2-59a4-414e-86e0-abbb59b51daa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoT-VLA_[[:space:]]Visual[[:space:]]Chain-of-Thought[[:space:]]Reasoning[[:space:]]for[[:space:]]Vision-Language-Action[[:space:]]Models/4f2ea357-dbc3-435d-b0c3-c4ec69dfb496_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Coarse[[:space:]]Correspondences[[:space:]]Boost[[:space:]]Spatial-Temporal[[:space:]]Reasoning[[:space:]]in[[:space:]]Multimodal[[:space:]]Language[[:space:]]Model/53ff4ccd-60e6-479a-bd87-54a994f71bf8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CocoER_[[:space:]]Aligning[[:space:]]Multi-Level[[:space:]]Feature[[:space:]]by[[:space:]]Competition[[:space:]]and[[:space:]]Coordination[[:space:]]for[[:space:]]Emotion[[:space:]]Recognition/91abd7f6-c62e-4947-ae7d-c03321b50ab1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Code-as-Monitor_[[:space:]]Constraint-aware[[:space:]]Visual[[:space:]]Programming[[:space:]]for[[:space:]]Reactive[[:space:]]and[[:space:]]Proactive[[:space:]]Robotic[[:space:]]Failure[[:space:]]Detection/3086dac5-625c-4ed2-b32d-12ee6b15f252_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Coeff-Tuning_[[:space:]]A[[:space:]]Graph[[:space:]]Filter[[:space:]]Subspace[[:space:]]View[[:space:]]for[[:space:]]Tuning[[:space:]]Attention-Based[[:space:]]Large[[:space:]]Models/e7a23930-dc20-4352-a4dc-e2a2616c610f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Coherent[[:space:]]3D[[:space:]]Portrait[[:space:]]Video[[:space:]]Reconstruction[[:space:]]via[[:space:]]Triplane[[:space:]]Fusion/2088b6b2-40e1-4341-a0c5-c5303c3acd65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ColabSfM_[[:space:]]Collaborative[[:space:]]Structure-from-Motion[[:space:]]by[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/52ce366e-9afd-441c-9c7e-e1d8e8bedeb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Collaborative[[:space:]]Decoding[[:space:]]Makes[[:space:]]Visual[[:space:]]Auto-Regressive[[:space:]]Modeling[[:space:]]Efficient/9a935c24-e29c-43fe-a54b-aa3325d4f954_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Collaborative[[:space:]]Tree[[:space:]]Search[[:space:]]for[[:space:]]Enhancing[[:space:]]Embodied[[:space:]]Multi-Agent[[:space:]]Collaboration/509a3bd5-5db7-45e0-bb5b-ba942b10cc92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Color[[:space:]]Alignment[[:space:]]in[[:space:]]Diffusion/4b8bcc3c-a5c5-4a38-89da-b85960cd1b00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ComRoPE_[[:space:]]Scalable[[:space:]]and[[:space:]]Robust[[:space:]]Rotary[[:space:]]Position[[:space:]]Embedding[[:space:]]Parameterized[[:space:]]by[[:space:]]Trainable[[:space:]]Commuting[[:space:]]Angle[[:space:]]Matrices/287888de-80a8-43c2-8d0f-f61f04d80e4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ComfyBench_[[:space:]]Benchmarking[[:space:]]LLM-based[[:space:]]Agents[[:space:]]in[[:space:]]ComfyUI[[:space:]]for[[:space:]]Autonomously[[:space:]]Designing[[:space:]]Collaborative[[:space:]]AI[[:space:]]Systems/9eeb1c75-c4c6-45dd-9896-e14c13fb270d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Common3D_[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]of[[:space:]]3D[[:space:]]Morphable[[:space:]]Models[[:space:]]for[[:space:]]Common[[:space:]]Objects[[:space:]]in[[:space:]]Neural[[:space:]]Feature[[:space:]]Space/ecdc4e97-8d79-41e2-abfb-5c0cebf6aa2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Commonsense[[:space:]]Video[[:space:]]Question[[:space:]]Answering[[:space:]]through[[:space:]]Video-Grounded[[:space:]]Entailment[[:space:]]Tree[[:space:]]Reasoning/e1080574-1050-4c6e-8509-597d8efaff00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Community[[:space:]]Forensics_[[:space:]]Using[[:space:]]Thousands[[:space:]]of[[:space:]]Generators[[:space:]]to[[:space:]]Train[[:space:]]Fake[[:space:]]Image[[:space:]]Detectors/f9d7ae16-50eb-4d82-828a-d6dd35d593a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CompGS_[[:space:]]Unleashing[[:space:]]2D[[:space:]]Compositionality[[:space:]]for[[:space:]]Compositional[[:space:]]Text-to-3D[[:space:]]via[[:space:]]Dynamically[[:space:]]Optimizing[[:space:]]3D[[:space:]]Gaussians/d9350216-461c-42c5-bc6e-3da77f71b98b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Compass[[:space:]]Control_[[:space:]]Multi[[:space:]]Object[[:space:]]Orientation[[:space:]]Control[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/9022e55b-ce42-4998-ae5e-db01a18e1b74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Complementary[[:space:]]Advantages_[[:space:]]Exploiting[[:space:]]Cross-Field[[:space:]]Frequency[[:space:]]Correlation[[:space:]]for[[:space:]]NIR-Assisted[[:space:]]Image[[:space:]]Denoising/c9ee5d0d-57af-4b5c-aacf-8ca5596906ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Completion[[:space:]]as[[:space:]]Enhancement_[[:space:]]A[[:space:]]Degradation-Aware[[:space:]]Selective[[:space:]]Image[[:space:]]Guided[[:space:]]Network[[:space:]]for[[:space:]]Depth[[:space:]]Completion/5b8a9777-b777-4fb5-9ba2-85ebeb5a648a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Complexity[[:space:]]Experts[[:space:]]are[[:space:]]Task-Discriminative[[:space:]]Learners[[:space:]]for[[:space:]]Any[[:space:]]Image[[:space:]]Restoration/811da5ba-4d1d-41b9-ac9b-e649712870d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Composing[[:space:]]Parts[[:space:]]for[[:space:]]Expressive[[:space:]]Object[[:space:]]Generation/edecee1f-cf3a-417c-a3d7-2ea937e0cfb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Compositional[[:space:]]Caching[[:space:]]for[[:space:]]Training-free[[:space:]]Open-vocabulary[[:space:]]Attribute[[:space:]]Detection/5997b1e0-b306-408f-91ce-2ab833ab2009_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Compositional[[:space:]]Targeted[[:space:]]Multi-Label[[:space:]]Universal[[:space:]]Perturbations/616cefba-8863-49bf-987c-2e525bd606b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Comprehensive[[:space:]]Information[[:space:]]Bottleneck[[:space:]]for[[:space:]]Unveiling[[:space:]]Universal[[:space:]]Attribution[[:space:]]to[[:space:]]Interpret[[:space:]]Vision[[:space:]]Transformers/abcc8427-3423-466d-afdc-b93969a43358_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Comprehensive[[:space:]]Relighting_[[:space:]]Generalizable[[:space:]]and[[:space:]]Consistent[[:space:]]Monocular[[:space:]]Human[[:space:]]Relighting[[:space:]]and[[:space:]]Harmonization/743d9653-d63f-4670-9d09-7ac226ec0ec0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ConMo_[[:space:]]Controllable[[:space:]]Motion[[:space:]]Disentanglement[[:space:]]and[[:space:]]Recomposition[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Motion[[:space:]]Transfer/c2885eb1-351e-456d-a404-fbc497fbcfb8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ConText-CIR_[[:space:]]Learning[[:space:]]from[[:space:]]Concepts[[:space:]]in[[:space:]]Text[[:space:]]for[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/73b8c8cd-f044-41da-aa48-a90d48bc8088_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Concept[[:space:]]Lancet_[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Compositional[[:space:]]Representation[[:space:]]Transplant/f841493e-d79e-460e-aa9e-12d127f2b90d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Concept[[:space:]]Replacer_[[:space:]]Replacing[[:space:]]Sensitive[[:space:]]Concepts[[:space:]]in[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Precision[[:space:]]Localization/5067da12-4c9b-4788-9049-b1876c6814e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ConceptGuard_[[:space:]]Continual[[:space:]]Personalized[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]with[[:space:]]Forgetting[[:space:]]and[[:space:]]Confusion[[:space:]]Mitigation/0e05d313-4c25-4c2b-bd5f-194172347d23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Condensing[[:space:]]Action[[:space:]]Segmentation[[:space:]]Datasets[[:space:]]via[[:space:]]Generative[[:space:]]Network[[:space:]]Inversion/0246e620-f269-40ea-be73-71e3fd5a8f44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Conditional[[:space:]]Balance_[[:space:]]Improving[[:space:]]Multi-Conditioning[[:space:]]Trade-Offs[[:space:]]in[[:space:]]Image[[:space:]]Generation/7d5085f0-d75a-4d87-a2e3-31d6f6559539_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Conformal[[:space:]]Prediction[[:space:]]and[[:space:]]MLLM[[:space:]]aided[[:space:]]Uncertainty[[:space:]]Quantification[[:space:]]in[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/e097b95c-4b11-4486-8696-895cf9360636_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Conformal[[:space:]]Prediction[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Models/c8e7fc0c-d793-4229-945d-edca1f580c2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Conical[[:space:]]Visual[[:space:]]Concentration[[:space:]]for[[:space:]]Efficient[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/6c773821-4351-4330-be62-0db1059c3c4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Consistency[[:space:]]Posterior[[:space:]]Sampling[[:space:]]for[[:space:]]Diverse[[:space:]]Image[[:space:]]Synthesis/a39e3a43-a316-4111-b28f-ca77cfbed254_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Consistency-aware[[:space:]]Self-Training[[:space:]]for[[:space:]]Iterative-based[[:space:]]Stereo[[:space:]]Matching/eded84c0-5c5a-4124-844b-df362791e377_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Consistent[[:space:]]Normal[[:space:]]Orientation[[:space:]]for[[:space:]]3D[[:space:]]Point[[:space:]]Clouds[[:space:]]via[[:space:]]Least[[:space:]]Squares[[:space:]]on[[:space:]]Delaunay[[:space:]]Graph/9294c835-6671-4352-a70d-7e0ea9dc36c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Consistent[[:space:]]and[[:space:]]Controllable[[:space:]]Image[[:space:]]Animation[[:space:]]with[[:space:]]Motion[[:space:]]Diffusion[[:space:]]Models/fb421ad6-a632-4c5c-91d8-21c22c9ca47e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Context-Aware[[:space:]]Multimodal[[:space:]]Pretraining/50cb43b9-8260-41c5-bdf9-6cfb0ed1db6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Context-Enhanced[[:space:]]Memory-Refined[[:space:]]Transformer[[:space:]]for[[:space:]]Online[[:space:]]Action[[:space:]]Detection/44a538fe-24ad-4d1c-bfb0-075863cf64c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Contextual[[:space:]]AD[[:space:]]Narration[[:space:]]with[[:space:]]Interleaved[[:space:]]Multimodal[[:space:]]Sequence/90cc95d4-3e98-4a2a-a509-0adbb1c3322b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Continual[[:space:]]SFT[[:space:]]Matches[[:space:]]Multimodal[[:space:]]RLHF[[:space:]]with[[:space:]]Negative[[:space:]]Supervision/04b076de-bb15-4d43-941a-1e294fd72c43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Continuous[[:space:]]3D[[:space:]]Perception[[:space:]]Model[[:space:]]with[[:space:]]Persistent[[:space:]]State/9e7a19c4-a18f-4652-8168-716957cf6171_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Continuous[[:space:]]Adverse[[:space:]]Weather[[:space:]]Removal[[:space:]]via[[:space:]]Degradation-Aware[[:space:]]Distillation/7b965121-849f-4e7a-87ed-d0fedc4b87d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Continuous[[:space:]]Locomotive[[:space:]]Crowd[[:space:]]Behavior[[:space:]]Generation/c16f65cb-587d-4c8f-a98e-d4aaf2b37e09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Continuous[[:space:]]Space-Time[[:space:]]Video[[:space:]]Resampling[[:space:]]with[[:space:]]Invertible[[:space:]]Motion[[:space:]]Steganography/116a1f84-5d94-4d3e-80d1-a123d4ff5a81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Continuous,[[:space:]]Subject-Specific[[:space:]]Attribute[[:space:]]Control[[:space:]]in[[:space:]]T2I[[:space:]]Models[[:space:]]by[[:space:]]Identifying[[:space:]]Semantic[[:space:]]Directions/0dbecd97-a25d-4567-8b2e-2742f15e53d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ControlFace_[[:space:]]Harnessing[[:space:]]Facial[[:space:]]Parametric[[:space:]]Control[[:space:]]for[[:space:]]Face[[:space:]]Rigging/eecb4db3-cba8-4683-a0fd-d03e73698ae0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Controllable[[:space:]]Human[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Personalized[[:space:]]Multi-Garments/73629037-2959-4d37-8d87-aa914dba3e74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Convex[[:space:]]Combination[[:space:]]Star[[:space:]]Shape[[:space:]]Prior[[:space:]]for[[:space:]]Data-driven[[:space:]]Image[[:space:]]Semantic[[:space:]]Segmentation/f454ef56-3d66-4ba5-bcc4-acbc6606d477_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Convex[[:space:]]Relaxation[[:space:]]for[[:space:]]Robust[[:space:]]Vanishing[[:space:]]Point[[:space:]]Estimation[[:space:]]in[[:space:]]Manhattan[[:space:]]World/44cadb13-ebbf-40d9-ba6e-92f2034fed61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CorrBEV_[[:space:]]Multi-View[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]by[[:space:]]Correlation[[:space:]]Learning[[:space:]]with[[:space:]]Multi-modal[[:space:]]Prototypes/a8270820-98d7-488b-8610-33a54ac930a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Correcting[[:space:]]Deviations[[:space:]]from[[:space:]]Normality_[[:space:]]A[[:space:]]Reformulated[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Multi-Class[[:space:]]Unsupervised[[:space:]]Anomaly[[:space:]]Detection/9641b761-64b3-4520-b52d-bcde65584d80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Correlative[[:space:]]and[[:space:]]Discriminative[[:space:]]Label[[:space:]]Grouping[[:space:]]for[[:space:]]Multi-Label[[:space:]]Visual[[:space:]]Prompt[[:space:]]Tuning/39f0115b-146e-4d5c-826a-c5738a000d76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CountLLM_[[:space:]]Towards[[:space:]]Generalizable[[:space:]]Repetitive[[:space:]]Action[[:space:]]Counting[[:space:]]via[[:space:]]Large[[:space:]]Language[[:space:]]Model/4c3763e8-43fb-436e-a528-929202946132_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Crab_[[:space:]]A[[:space:]]Unified[[:space:]]Audio-Visual[[:space:]]Scene[[:space:]]Understanding[[:space:]]Model[[:space:]]with[[:space:]]Explicit[[:space:]]Cooperation/5dbe6623-7825-4fd2-88e5-7a044ba10885_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CraftsMan3D_[[:space:]]High-fidelity[[:space:]]Mesh[[:space:]]Generation[[:space:]]with[[:space:]]3D[[:space:]]Native[[:space:]]Diffusion[[:space:]]and[[:space:]]Interactive[[:space:]]Geometry[[:space:]]Refiner/5144ef73-499f-4ff7-be02-431979e029aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Creating[[:space:]]Your[[:space:]]Editable[[:space:]]3D[[:space:]]Photorealistic[[:space:]]Avatar[[:space:]]with[[:space:]]Tetrahedron-constrained[[:space:]]Gaussian[[:space:]]Splatting/7a8da946-5567-4a09-8a39-a81e768a31e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Critic-V_[[:space:]]VLM[[:space:]]Critics[[:space:]]Help[[:space:]]Catch[[:space:]]VLM[[:space:]]Errors[[:space:]]in[[:space:]]Multimodal[[:space:]]Reasoning/cd5ff46c-7b81-4de5-b8a8-425d98ac1b4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CroCoDL_[[:space:]]Cross-device[[:space:]]Collaborative[[:space:]]Dataset[[:space:]]for[[:space:]]Localization/862167b1-fded-4c5f-bda5-1d5d5be3f7b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cropper_[[:space:]]Vision-Language[[:space:]]Model[[:space:]]for[[:space:]]Image[[:space:]]Cropping[[:space:]]through[[:space:]]In-Context[[:space:]]Learning/9ebbbdbc-807f-4eb5-8589-07406247ef13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-Modal[[:space:]]3D[[:space:]]Representation[[:space:]]with[[:space:]]Multi-View[[:space:]]Images[[:space:]]and[[:space:]]Point[[:space:]]Clouds/7c2e669e-d8aa-43ed-8f93-6689e539a89b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-Modal[[:space:]]Distillation[[:space:]]for[[:space:]]2D_3D[[:space:]]Multi-Object[[:space:]]Discovery[[:space:]]from[[:space:]]2D[[:space:]]Motion/1050431c-cf52-4081-ac40-d13b37c0dbb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-Modal[[:space:]]Interactive[[:space:]]Perception[[:space:]]Network[[:space:]]with[[:space:]]Mamba[[:space:]]for[[:space:]]Lung[[:space:]]Tumor[[:space:]]Segmentation[[:space:]]in[[:space:]]PET-CT[[:space:]]Images/8d2ce17c-ea33-41ec-b54c-1c11725e9e4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-Modal[[:space:]]and[[:space:]]Uncertainty-Aware[[:space:]]Agglomeration[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/6f4b0ff7-e69b-480f-92cd-485e26500989_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-Rejective[[:space:]]Open-Set[[:space:]]SAR[[:space:]]Image[[:space:]]Registration/c56ea334-2697-4c4f-a580-b1a5767bf55c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-View[[:space:]]Completion[[:space:]]Models[[:space:]]are[[:space:]]Zero-shot[[:space:]]Correspondence[[:space:]]Estimators/820d410f-c25a-4caf-af81-d2d85693f5e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-modal[[:space:]]Causal[[:space:]]Relation[[:space:]]Alignment[[:space:]]for[[:space:]]Video[[:space:]]Question[[:space:]]Grounding/69af920b-ebdc-473e-b85b-a424b0303738_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-modal[[:space:]]Information[[:space:]]Flow[[:space:]]in[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/ede4923d-afb7-4330-aab9-7701dc99b29e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CrossOver_[[:space:]]3D[[:space:]]Scene[[:space:]]Cross-Modal[[:space:]]Alignment/30a98fdc-b413-45a6-9d56-4fee165c9382_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CrossSDF_[[:space:]]3D[[:space:]]Reconstruction[[:space:]]of[[:space:]]Thin[[:space:]]Structures[[:space:]]From[[:space:]]Cross-Sections/c4551bf4-a6e0-4190-b85f-0dcc28810d5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CryptoFace_[[:space:]]End-to-End[[:space:]]Encrypted[[:space:]]Face[[:space:]]Recognition/a98f1c53-38ec-4f98-a7f6-ab50dc580eb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cubify[[:space:]]Anything_[[:space:]]Scaling[[:space:]]Indoor[[:space:]]3D[[:space:]]Object[[:space:]]Detection/f86f6986-935c-47a1-a6ec-168925cdf55e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Curriculum[[:space:]]Coarse-to-Fine[[:space:]]Selection[[:space:]]for[[:space:]]High-IPC[[:space:]]Dataset[[:space:]]Distillation/ae462d90-2b47-4942-8fa7-31d5c47be7b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Curriculum[[:space:]]Direct[[:space:]]Preference[[:space:]]Optimization[[:space:]]for[[:space:]]Diffusion[[:space:]]and[[:space:]]Consistency[[:space:]]Models/f7f05e77-b097-4ce8-bc16-947cdf589509_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CustAny_[[:space:]]Customizing[[:space:]]Anything[[:space:]]from[[:space:]]A[[:space:]]Single[[:space:]]Example/e6b6cb88-e6a1-4882-bd0e-763fef1dfe5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CustomKD_[[:space:]]Customizing[[:space:]]Large[[:space:]]Vision[[:space:]]Foundation[[:space:]]for[[:space:]]Edge[[:space:]]Model[[:space:]]Improvement[[:space:]]via[[:space:]]Knowledge[[:space:]]Distillation/ac0e9354-793a-441e-bbc5-8bc5c657ea48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Customized[[:space:]]Condition[[:space:]]Controllable[[:space:]]Generation[[:space:]]for[[:space:]]Video[[:space:]]Soundtrack/4b96ee2d-0826-4f4e-bf2c-89e8c69eaab0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/D2SP_[[:space:]]Dynamic[[:space:]]Dual-Stage[[:space:]]Purification[[:space:]]Framework[[:space:]]for[[:space:]]Dual[[:space:]]Noise[[:space:]]Mitigation[[:space:]]in[[:space:]]Vision-based[[:space:]]Affective[[:space:]]Recognition./d51c9ffc-e4f0-4278-81d6-188ca27dcd4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DA-VPT_[[:space:]]Semantic-Guided[[:space:]]Visual[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/7f620e8d-ccd8-48df-bf31-61c533d75484_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DAGSM_[[:space:]]Disentangled[[:space:]]Avatar[[:space:]]Generation[[:space:]]with[[:space:]]GS-enhanced[[:space:]]Mesh/d381eee2-084b-4498-a38b-f3cfc3cbf2ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DAMM-Diffusion_[[:space:]]Learning[[:space:]]Divergence-Aware[[:space:]]Multi-Modal[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Nanoparticles[[:space:]]Distribution[[:space:]]Prediction/ba2a92d8-b83a-475d-9908-9f49265091bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DART_[[:space:]]Disease-aware[[:space:]]Image-Text[[:space:]]Alignment[[:space:]]and[[:space:]]Self-correcting[[:space:]]Re-alignment[[:space:]]for[[:space:]]Trustworthy[[:space:]]Radiology[[:space:]]Report[[:space:]]Generation/59213d27-0669-4b4b-af29-22066c8ce267_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DCEvo_[[:space:]]Discriminative[[:space:]]Cross-Dimensional[[:space:]]Evolutionary[[:space:]]Learning[[:space:]]for[[:space:]]Infrared[[:space:]]and[[:space:]]Visible[[:space:]]Image[[:space:]]Fusion/763440ec-b962-4ebf-a768-f7b9b84869ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DEAL_[[:space:]]Data-Efficient[[:space:]]Adversarial[[:space:]]Learning[[:space:]]for[[:space:]]High-Quality[[:space:]]Infrared[[:space:]]Imaging/0ca5f5ac-624c-46ac-8cb8-0af4ab790b5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DEFOM-Stereo_[[:space:]]Depth[[:space:]]Foundation[[:space:]]Model[[:space:]]Based[[:space:]]Stereo[[:space:]]Matching/a273891a-d56b-467e-b26c-a4f446917898_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DEIM_[[:space:]]DETR[[:space:]]with[[:space:]]Improved[[:space:]]Matching[[:space:]]for[[:space:]]Fast[[:space:]]Convergence/700c05ab-02e8-4cb8-9e53-6b338c23c887_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DELT_[[:space:]]A[[:space:]]Simple[[:space:]]Diversity-driven[[:space:]]EarlyLate[[:space:]]Training[[:space:]]for[[:space:]]Dataset[[:space:]]Distillation/e8a5d8b9-5b79-4bc1-a6b9-6a9791dc1a29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DFM_[[:space:]]Differentiable[[:space:]]Feature[[:space:]]Matching[[:space:]]for[[:space:]]Anomaly[[:space:]]Detection/9175d952-b477-4cda-a5ee-a6da1e7d0fcf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DFormerv2_[[:space:]]Geometry[[:space:]]Self-Attention[[:space:]]for[[:space:]]RGBD[[:space:]]Semantic[[:space:]]Segmentation/0fdee2be-7d7f-422e-9322-3b0880ae6ccf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DH-Set_[[:space:]]Improving[[:space:]]Vision-Language[[:space:]]Alignment[[:space:]]with[[:space:]]Diverse[[:space:]]and[[:space:]]Hybrid[[:space:]]Set-Embeddings[[:space:]]Learning/ef351c8a-139e-43f0-a060-e64d0e5c0ae4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DI-PCG_[[:space:]]Diffusion-based[[:space:]]Efficient[[:space:]]Inverse[[:space:]]Procedural[[:space:]]Content[[:space:]]Generation[[:space:]]for[[:space:]]High-quality[[:space:]]3D[[:space:]]Asset[[:space:]]Creation/cfb9f507-5765-4b35-9b5a-8085f99378d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DIFFER_[[:space:]]Disentangling[[:space:]]Identity[[:space:]]Features[[:space:]]via[[:space:]]Semantic[[:space:]]Cues[[:space:]]for[[:space:]]Clothes-Changing[[:space:]]Person[[:space:]]Re-ID/3b7ba7cf-6163-4d2e-86bf-cbbc5edc2518_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DIFIX3D+_[[:space:]]Improving[[:space:]]3D[[:space:]]Reconstructions[[:space:]]with[[:space:]]Single-Step[[:space:]]Diffusion[[:space:]]Models/0b6e4b23-2e2d-4870-ad99-ad18b5a2098d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DINOv2[[:space:]]Meets[[:space:]]Text_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Image-[[:space:]]and[[:space:]]Pixel-Level[[:space:]]Vision-Language[[:space:]]Alignment/0ae8bb33-9135-4f87-8554-172250bea954_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DIO_[[:space:]]Decomposable[[:space:]]Implicit[[:space:]]4D[[:space:]]Occupancy-Flow[[:space:]]World[[:space:]]Model/d0138108-6a70-45cc-90a1-4a873e3afa90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DIV-FF_[[:space:]]Dynamic[[:space:]]Image-Video[[:space:]]Feature[[:space:]]Fields[[:space:]]For[[:space:]]Environment[[:space:]]Understanding[[:space:]]in[[:space:]]Egocentric[[:space:]]Videos/db6de755-dfbf-42cc-be3a-c53a009a70e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DKC_[[:space:]]Differentiated[[:space:]]Knowledge[[:space:]]Consolidation[[:space:]]for[[:space:]]Cloth-Hybrid[[:space:]]Lifelong[[:space:]]Person[[:space:]]Re-identification/438cdc80-aa4a-4f61-8343-f73afb57065b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DKDM_[[:space:]]Data-Free[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Any[[:space:]]Architecture/7db1f727-99aa-4e9f-b990-2b91062cf1ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DL2G_[[:space:]]Degradation-guided[[:space:]]Local-to-Global[[:space:]]Restoration[[:space:]]for[[:space:]]Eyeglass[[:space:]]Reflection[[:space:]]Removal/634c3408-ddc9-4cc9-8eca-7d0befde896a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DNF_[[:space:]]Unconditional[[:space:]]4D[[:space:]]Generation[[:space:]]with[[:space:]]Dictionary-based[[:space:]]Neural[[:space:]]Fields/53c5c554-ebb1-4f7b-887a-570d92a80cfe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DOF-GS_[[:space:]]Adjustable[[:space:]]Depth-of-Field[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Post-Capture[[:space:]]Refocusing,[[:space:]]Defocus[[:space:]]Rendering[[:space:]]and[[:space:]]Blur[[:space:]]Removal/ea16eb05-5a7b-42ab-a447-73536a9b5a01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DORNet_[[:space:]]A[[:space:]]Degradation[[:space:]]Oriented[[:space:]]and[[:space:]]Regularized[[:space:]]Network[[:space:]]for[[:space:]]Blind[[:space:]]Depth[[:space:]]Super-Resolution/01f4bf5e-c5b0-4353-8d95-14b50fe846d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DPC_[[:space:]]Dual-Prompt[[:space:]]Collaboration[[:space:]]for[[:space:]]Tuning[[:space:]]Vision-Language[[:space:]]Models/1e67dc6e-7609-4c80-b113-6cee1b4ca6f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DPFlow_[[:space:]]Adaptive[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation[[:space:]]with[[:space:]]a[[:space:]]Dual-Pyramid[[:space:]]Framework/f757fd8f-04a7-462a-82d4-ccd928a35069_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DPSeg_[[:space:]]Dual-Prompt[[:space:]]Cost[[:space:]]Volume[[:space:]]Learning[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/86a67dbb-2114-4a44-a090-481aa3138956_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DPU_[[:space:]]Dynamic[[:space:]]Prototype[[:space:]]Updating[[:space:]]for[[:space:]]Multimodal[[:space:]]Out-of-Distribution[[:space:]]Detection/e3a8a150-3d53-4522-ba7e-737fcc4ea249_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DRAWER_[[:space:]]Digital[[:space:]]Reconstruction[[:space:]]and[[:space:]]Articulation[[:space:]]With[[:space:]]Environment[[:space:]]Realism/0f504618-279a-4b47-9314-2eedbabe1dec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DRiVE_[[:space:]]Diffusion-based[[:space:]]Rigging[[:space:]]Empowers[[:space:]]Generation[[:space:]]of[[:space:]]Versatile[[:space:]]and[[:space:]]Expressive[[:space:]]Characters/619e1290-28a8-4d29-b2f5-0c4afb4e7953_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DSPNet_[[:space:]]Dual-vision[[:space:]]Scene[[:space:]]Perception[[:space:]]for[[:space:]]Robust[[:space:]]3D[[:space:]]Question[[:space:]]Answering/a73d5dc4-9870-4bd2-9d61-7da9455a779f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DSV-LFS_[[:space:]]Unifying[[:space:]]LLM-Driven[[:space:]]Semantic[[:space:]]Cues[[:space:]]with[[:space:]]Visual[[:space:]]Features[[:space:]]for[[:space:]]Robust[[:space:]]Few-Shot[[:space:]]Segmentation/cf5e37e4-d9ee-4b3a-a594-a07c84894c09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DTGBrepGen_[[:space:]]A[[:space:]]Novel[[:space:]]B-rep[[:space:]]Generative[[:space:]]Model[[:space:]]through[[:space:]]Decoupling[[:space:]]Topology[[:space:]]and[[:space:]]Geometry/9740d615-86d0-470b-b129-7e407f643398_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DTOS_[[:space:]]Dynamic[[:space:]]Time[[:space:]]Object[[:space:]]Sensing[[:space:]]with[[:space:]]Large[[:space:]]Multimodal[[:space:]]Model/8c34a6af-a63a-4f4d-835b-0b89efd1f7e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DUNE_[[:space:]]Distilling[[:space:]]a[[:space:]]Universal[[:space:]]Encoder[[:space:]]from[[:space:]]Heterogeneous[[:space:]]2D[[:space:]]and[[:space:]]3D[[:space:]]Teachers/19ff7f32-275f-4372-87e6-aa8494012a0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DV-Matcher_[[:space:]]Deformation-based[[:space:]]Non-rigid[[:space:]]Point[[:space:]]Cloud[[:space:]]Matching[[:space:]]Guided[[:space:]]by[[:space:]]Pre-trained[[:space:]]Visual[[:space:]]Features/d71acb8b-f5a9-47d2-9e6b-c7fc9b0be98b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DVHGNN_[[:space:]]Multi-Scale[[:space:]]Dilated[[:space:]]Vision[[:space:]]HGNN[[:space:]]for[[:space:]]Efficient[[:space:]]Vision[[:space:]]Recognition/a09a1b1d-0b95-48b8-8b43-e73928aa7de2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DViN_[[:space:]]Dynamic[[:space:]]Visual[[:space:]]Routing[[:space:]]Network[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Referring[[:space:]]Expression[[:space:]]Comprehension/e25b1240-453b-4dfe-a141-9d05d69078b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/D^2iT_[[:space:]]Dynamic[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]for[[:space:]]Accurate[[:space:]]Image[[:space:]]Generation/6a06742a-9d7b-41e5-a45e-5f1feb16c423_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/D^3-Human_[[:space:]]Dynamic[[:space:]]Disentangled[[:space:]]Digital[[:space:]]Human[[:space:]]from[[:space:]]Monocular[[:space:]]Video/d06113aa-884c-4df6-8a2b-2ae67d7a11c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/D^3CTTA_[[:space:]]Domain-Dependent[[:space:]]Decorrelation[[:space:]]for[[:space:]]Continual[[:space:]]Test-Time[[:space:]]Adaption[[:space:]]of[[:space:]]3D[[:space:]]LiDAR[[:space:]]Segmentation/d0e0e7e5-a473-4121-9384-68a7bbfc7c91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/D^3_[[:space:]]Scaling[[:space:]]Up[[:space:]]Deepfake[[:space:]]Detection[[:space:]]by[[:space:]]Learning[[:space:]]from[[:space:]]Discrepancy/cee11b62-05f7-4681-9618-b60281674d06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DaCapo_[[:space:]]Score[[:space:]]Distillation[[:space:]]as[[:space:]]Stacked[[:space:]]Bridge[[:space:]]for[[:space:]]Fast[[:space:]]and[[:space:]]High-quality[[:space:]]3D[[:space:]]Editing/fe3021df-c5c4-4637-b0f5-bf5990db3a7a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DarkIR_[[:space:]]Robust[[:space:]]Low-Light[[:space:]]Image[[:space:]]Restoration/11a2434f-c11a-47a9-a3f6-5bdae6348a55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DashGaussian_[[:space:]]Optimizing[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]in[[:space:]]200[[:space:]]Seconds/7084f6a7-a041-4c15-9c9c-ed72f7e963f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Data[[:space:]]Distributional[[:space:]]Properties[[:space:]]As[[:space:]]Inductive[[:space:]]Bias[[:space:]]for[[:space:]]Systematic[[:space:]]Generalization/8b3d00b3-f2b7-42c4-bbf6-19b930d77c33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Data[[:space:]]Synthesis[[:space:]]with[[:space:]]Diverse[[:space:]]Styles[[:space:]]for[[:space:]]Face[[:space:]]Recognition[[:space:]]via[[:space:]]3DMM-Guided[[:space:]]Diffusion/342ddb85-a1de-438b-ae96-f0582ca9cc59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Data-Free[[:space:]]Group-Wise[[:space:]]Fully[[:space:]]Quantized[[:space:]]Winograd[[:space:]]Convolution[[:space:]]via[[:space:]]Learnable[[:space:]]Scales/732a9d8a-0242-4cc4-b580-a55f0b048d1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Data-free[[:space:]]Universal[[:space:]]Adversarial[[:space:]]Perturbation[[:space:]]with[[:space:]]Pseudo-semantic[[:space:]]Prior/9a465eb1-a358-4ea4-bf28-ee7afca7ed25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dataset[[:space:]]Distillation[[:space:]]with[[:space:]]Neural[[:space:]]Characteristic[[:space:]]Function_[[:space:]]A[[:space:]]Minmax[[:space:]]Perspective/eacc91a1-3e9e-4487-85f8-88bc28202e0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeCLIP_[[:space:]]Decoupled[[:space:]]Learning[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Dense[[:space:]]Perception/535947d7-06b5-4d7c-a448-67fd50dc47db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeCafNet_[[:space:]]Delegate[[:space:]]and[[:space:]]Conquer[[:space:]]for[[:space:]]Efficient[[:space:]]Temporal[[:space:]]Grounding[[:space:]]in[[:space:]]Long[[:space:]]Videos/92a9452b-34b6-4812-87d3-0fd330dc7b0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeClotH_[[:space:]]Decomposable[[:space:]]3D[[:space:]]Cloth[[:space:]]and[[:space:]]Human[[:space:]]Body[[:space:]]Reconstruction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/b2e5ae65-3c32-4f3c-b78a-7880a4bd5fc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeDe_[[:space:]]Detecting[[:space:]]Backdoor[[:space:]]Samples[[:space:]]for[[:space:]]SSL[[:space:]]Encoders[[:space:]]via[[:space:]]Decoders/4aecf957-e78a-49f4-b0d7-05a5cdfb5d95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeNVeR_[[:space:]]Deformable[[:space:]]Neural[[:space:]]Vessel[[:space:]]Representations[[:space:]]for[[:space:]]Unsupervised[[:space:]]Video[[:space:]]Vessel[[:space:]]Segmentation/c3a27120-e20d-4a79-a039-08cd9557b7c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeRS_[[:space:]]Towards[[:space:]]Extremely[[:space:]]Efficient[[:space:]]Upcycled[[:space:]]Mixture-of-Experts[[:space:]]Models/f2f8bbd9-c99a-4a2e-bce3-b46b58ef0f46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeSiRe-GS_[[:space:]]4D[[:space:]]Street[[:space:]]Gaussians[[:space:]]for[[:space:]]Static-Dynamic[[:space:]]Decomposition[[:space:]]and[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]for[[:space:]]Urban[[:space:]]Driving[[:space:]]Scenes/d29cab3c-4daf-4d84-8d8a-55223190029a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeSplat_[[:space:]]Decomposed[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Distractor-Free[[:space:]]Rendering/afcce868-2040-49c5-a663-b4b87e65b5e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/De^2Gaze_[[:space:]]Deformable[[:space:]]and[[:space:]]Decoupled[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]3D[[:space:]]Gaze[[:space:]]Estimation/4285b006-efb3-48d7-ab2e-5610d0c653d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Debiasing[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]via[[:space:]]Noise-Aware[[:space:]]Preference[[:space:]]Optimization/a9c6d1cd-fae5-48dd-a608-c33c78561e53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decentralized[[:space:]]Diffusion[[:space:]]Models/931a8a2c-b730-4445-8cd6-6be4a5c30098_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decision[[:space:]]SpikeFormer_[[:space:]]Spike-Driven[[:space:]]Transformer[[:space:]]for[[:space:]]Decision[[:space:]]Making/2f9d286b-5ec9-4366-a640-378a43e60185_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decoder[[:space:]]Gradient[[:space:]]Shield_[[:space:]]Provable[[:space:]]and[[:space:]]High-Fidelity[[:space:]]Prevention[[:space:]]of[[:space:]]Gradient-Based[[:space:]]Box-Free[[:space:]]Watermark[[:space:]]Removal/4ea3cb35-b0d8-4336-9e87-4a0cf7228744_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decompositional[[:space:]]Neural[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]with[[:space:]]Generative[[:space:]]Diffusion[[:space:]]Prior/f3625071-30fa-4298-a33b-ceafd260d38c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decouple[[:space:]]Distortion[[:space:]]from[[:space:]]Perception_[[:space:]]Region[[:space:]]Adaptive[[:space:]]Diffusion[[:space:]]for[[:space:]]Extreme-low[[:space:]]Bitrate[[:space:]]Perception[[:space:]]Image[[:space:]]Compression/3fe4186f-a87d-4174-8f8c-f63734fc78ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decouple-Then-Merge_[[:space:]]Finetune[[:space:]]Diffusion[[:space:]]Models[[:space:]]as[[:space:]]Multi-Task[[:space:]]Learning/d6660b78-0b48-4809-b422-f208263c4c94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decoupled[[:space:]]Distillation[[:space:]]to[[:space:]]Erase_[[:space:]]A[[:space:]]General[[:space:]]Unlearning[[:space:]]Method[[:space:]]for[[:space:]]Any[[:space:]]Class-centric[[:space:]]Tasks/67ddadf2-525f-43e5-abd3-c05ee2940502_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decoupled[[:space:]]Motion[[:space:]]Expression[[:space:]]Video[[:space:]]Segmentation/c65b08f7-2c41-4983-a9c0-eece9fe97c32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DecoupledGaussian_[[:space:]]Object-Scene[[:space:]]Decoupling[[:space:]]for[[:space:]]Physics-Based[[:space:]]Interaction/4d6395a0-4fb9-492a-9fba-545d8c0df116_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decoupling[[:space:]]Fine[[:space:]]Detail[[:space:]]and[[:space:]]Global[[:space:]]Geometry[[:space:]]for[[:space:]]Compressed[[:space:]]Depth[[:space:]]Map[[:space:]]Super-Resolution/9a0a1d3a-47d8-4a93-9451-a52e0f68b026_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decoupling[[:space:]]Training-Free[[:space:]]Guided[[:space:]]Diffusion[[:space:]]by[[:space:]]ADMM/5de317b5-15b4-44a6-a032-7b51d7994305_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Deep[[:space:]]Change[[:space:]]Monitoring_[[:space:]]A[[:space:]]Hyperbolic[[:space:]]Representative[[:space:]]Learning[[:space:]]Framework[[:space:]]and[[:space:]]a[[:space:]]Dataset[[:space:]]for[[:space:]]Long-term[[:space:]]Fine-grained[[:space:]]Tree[[:space:]]Change[[:space:]]Detection/a2da7e4d-5251-421c-b0bb-c914d1af1e87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Deep[[:space:]]Fair[[:space:]]Multi-View[[:space:]]Clustering[[:space:]]with[[:space:]]Attention[[:space:]]KAN/623e4bc3-e225-4f72-bbd0-1fe1ef817d42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeepCompress-ViT_[[:space:]]Rethinking[[:space:]]Model[[:space:]]Compression[[:space:]]to[[:space:]]Enhance[[:space:]]Efficiency[[:space:]]of[[:space:]]Vision[[:space:]]Transformers[[:space:]]at[[:space:]]the[[:space:]]Edge/3bd82ff4-b9fe-43da-a28f-c5a1697c4ab3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeepLA-Net_[[:space:]]Very[[:space:]]Deep[[:space:]]Local[[:space:]]Aggregation[[:space:]]Networks[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Analysis/a85e9f83-617a-4a59-8124-97eefae14b9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DefMamba_[[:space:]]Deformable[[:space:]]Visual[[:space:]]State[[:space:]]Space[[:space:]]Model/deb50b63-d2d0-4e53-81b5-8019ea8427b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DefectFill_[[:space:]]Realistic[[:space:]]Defect[[:space:]]Generation[[:space:]]with[[:space:]]Inpainting[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Visual[[:space:]]Inspection/41a38a7f-989e-4bd9-87f8-8ae83802b924_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeformCL_[[:space:]]Learning[[:space:]]Deformable[[:space:]]Centerline[[:space:]]Representation[[:space:]]for[[:space:]]Vessel[[:space:]]Extraction[[:space:]]in[[:space:]]3D[[:space:]]Medical[[:space:]]Image/3df80607-397e-48ec-af7c-1eb858969941_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Deformable[[:space:]]Radial[[:space:]]Kernel[[:space:]]Splatting/3fdb69e1-a3c3-48aa-8f3c-430c977f4a43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Degradation-Aware[[:space:]]Feature[[:space:]]Perturbation[[:space:]]for[[:space:]]All-in-One[[:space:]]Image[[:space:]]Restoration/dd15326b-1f78-4e60-9709-e2819267a8c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DejaVid_[[:space:]]Encoder-Agnostic[[:space:]]Learned[[:space:]]Temporal[[:space:]]Matching[[:space:]]for[[:space:]]Video[[:space:]]Classification/b75bb180-204a-4d50-a1dd-1f6046fc0819_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Denoising[[:space:]]Functional[[:space:]]Maps_[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Shape[[:space:]]Correspondence/2c4c8aae-c014-423b-a2aa-0d67eb2103ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dense[[:space:]]Dispersed[[:space:]]Structured[[:space:]]Light[[:space:]]for[[:space:]]Hyperspectral[[:space:]]3D[[:space:]]Imaging[[:space:]]of[[:space:]]Dynamic[[:space:]]Scenes/e6f27c19-909f-43c9-8505-0dcb89ae13c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dense[[:space:]]Match[[:space:]]Summarization[[:space:]]for[[:space:]]Faster[[:space:]]Two-view[[:space:]]Estimation/cc201b1e-b30f-4c1e-aa8e-8a269c75e076_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dense-SfM_[[:space:]]Structure[[:space:]]from[[:space:]]Motion[[:space:]]with[[:space:]]Dense[[:space:]]Consistent[[:space:]]Matching/39fcbf6f-a095-433d-809d-0e3787aefd05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Depth[[:space:]]Any[[:space:]]Camera_[[:space:]]Zero-Shot[[:space:]]Metric[[:space:]]Depth[[:space:]]Estimation[[:space:]]from[[:space:]]Any[[:space:]]Camera/d2be3c71-7e44-4cd1-ae50-da4952c81a8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Depth-Guided[[:space:]]Bundle[[:space:]]Sampling[[:space:]]for[[:space:]]Efficient[[:space:]]Generalizable[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field[[:space:]]Reconstruction/8a97af03-ecf0-4b44-846e-a010c9f066c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DepthCrafter_[[:space:]]Generating[[:space:]]Consistent[[:space:]]Long[[:space:]]Depth[[:space:]]Sequences[[:space:]]for[[:space:]]Open-world[[:space:]]Videos/7820448d-f4cd-4ec7-b539-aef2413229a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DepthCues_[[:space:]]Evaluating[[:space:]]Monocular[[:space:]]Depth[[:space:]]Perception[[:space:]]in[[:space:]]Large[[:space:]]Vision[[:space:]]Models/ba2594dd-8284-4f56-b7c2-2adbb7e80d5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DepthSplat_[[:space:]]Connecting[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]and[[:space:]]Depth/3c38ee77-2762-4a1c-ad37-7a37f09eab09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Derivative-Free[[:space:]]Diffusion[[:space:]]Manifold-Constrained[[:space:]]Gradient[[:space:]]for[[:space:]]Unified[[:space:]]XAI/065052e0-b498-4bc6-9515-464f2840472f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Descriptor-In-Pixel[[:space:]]_[[:space:]]Point-Feature[[:space:]]Tracking[[:space:]]For[[:space:]]Pixel[[:space:]]Processor[[:space:]]Arrays/4c7b942d-4d91-4aec-b4fd-a208ea5f65b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Design2GarmentCode_[[:space:]]Turning[[:space:]]Design[[:space:]]Concepts[[:space:]]to[[:space:]]Tangible[[:space:]]Garments[[:space:]]Through[[:space:]]Program[[:space:]]Synthesis/d74dfa3b-ff23-42b1-9fc2-1b69e990b9a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DesignDiffusion_[[:space:]]High-Quality[[:space:]]Text-to-Design[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/eb2e95ca-796c-48bb-9edd-bb8243db35c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Detail-Preserving[[:space:]]Latent[[:space:]]Diffusion[[:space:]]for[[:space:]]Stable[[:space:]]Shadow[[:space:]]Removal/a2941634-46f8-4a96-a7ab-f9421ddf1c2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Detect[[:space:]]Any[[:space:]]Mirrors_[[:space:]]Boosting[[:space:]]Learning[[:space:]]Reliability[[:space:]]on[[:space:]]Large-Scale[[:space:]]Unlabeled[[:space:]]Data[[:space:]]with[[:space:]]an[[:space:]]Iterative[[:space:]]Data[[:space:]]Engine/64ac5be5-9942-4491-856e-a27352561435_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Detect-and-Guide_[[:space:]]Self-regulation[[:space:]]of[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Safe[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]via[[:space:]]Guideline[[:space:]]Token[[:space:]]Optimization/2a9b10bf-47cd-4616-ab01-1e841be5bccf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Detecting[[:space:]]Adversarial[[:space:]]Data[[:space:]]Using[[:space:]]Perturbation[[:space:]]Forgery/a87e2166-e7a0-46b8-94b2-0c0c61135018_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Detecting[[:space:]]Backdoor[[:space:]]Attacks[[:space:]]in[[:space:]]Federated[[:space:]]Learning[[:space:]]via[[:space:]]Direction[[:space:]]Alignment[[:space:]]Inspection/13985764-96b8-48db-82c6-84b1d3a23329_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Detecting[[:space:]]Open[[:space:]]World[[:space:]]Objects[[:space:]]via[[:space:]]Partial[[:space:]]Attribute[[:space:]]Assignment/9a893be0-4b48-49d9-8973-6218a7e4b1e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Detecting[[:space:]]Out-of-Distribution[[:space:]]Through[[:space:]]the[[:space:]]Lens[[:space:]]of[[:space:]]Neural[[:space:]]Collapse/65f987d8-3fd0-42f5-b253-2ec1ecfb0a2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Detection-Friendly[[:space:]]Nonuniformity[[:space:]]Correction_[[:space:]]A[[:space:]]Union[[:space:]]Framework[[:space:]]for[[:space:]]Infrared[[:space:]]UAV[[:space:]]Target[[:space:]]Detection/c87793a3-7dc6-4ae0-b5e0-3251be204e39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Deterministic[[:space:]]Certification[[:space:]]of[[:space:]]Graph[[:space:]]Neural[[:space:]]Networks[[:space:]]against[[:space:]]Graph[[:space:]]Poisoning[[:space:]]Attacks[[:space:]]with[[:space:]]Arbitrary[[:space:]]Perturbations/0497332e-bd66-4105-bea8-f5b6eb06db3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Deterministic[[:space:]]Image-to-Image[[:space:]]Translation[[:space:]]via[[:space:]]Denoising[[:space:]]Brownian[[:space:]]Bridge[[:space:]]Models[[:space:]]with[[:space:]]Dual[[:space:]]Approximators/f031c8e4-7831-43c8-a97d-74d4fb079432_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Deterministic-to-Stochastic[[:space:]]Diverse[[:space:]]Latent[[:space:]]Feature[[:space:]]Mapping[[:space:]]for[[:space:]]Human[[:space:]]Motion[[:space:]]Synthesis/85360d39-ddc7-43b2-b5d2-e1635ded4be9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Devil[[:space:]]is[[:space:]]in[[:space:]]the[[:space:]]Detail_[[:space:]]Towards[[:space:]]Injecting[[:space:]]Fine[[:space:]]Details[[:space:]]of[[:space:]]Image[[:space:]]Prompt[[:space:]]in[[:space:]]Image[[:space:]]Generation[[:space:]]via[[:space:]]Conflict-free[[:space:]]Guidance[[:space:]]and[[:space:]]Stratified[[:space:]]Attention/33468e61-bdab-44f8-8e48-6c0ba3c7248c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Devils[[:space:]]in[[:space:]]Middle[[:space:]]Layers[[:space:]]of[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models_[[:space:]]Interpreting,[[:space:]]Detecting[[:space:]]and[[:space:]]Mitigating[[:space:]]Object[[:space:]]Hallucinations[[:space:]]via[[:space:]]Attention[[:space:]]Lens/c159f2ec-9cd0-43d8-97ae-a644ae3c367b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DexGrasp[[:space:]]Anything_[[:space:]]Towards[[:space:]]Universal[[:space:]]Robotic[[:space:]]Dexterous[[:space:]]Grasping[[:space:]]with[[:space:]]Physics[[:space:]]Awareness/fb2fbb43-15d0-4a10-bd84-0e5530360297_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DexHandDiff_[[:space:]]Interaction-aware[[:space:]]Diffusion[[:space:]]Planning[[:space:]]for[[:space:]]Adaptive[[:space:]]Dexterous[[:space:]]Manipulation/776fbd8b-c951-4932-8e6a-abc5a13d15dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiC_[[:space:]]Rethinking[[:space:]]Conv3x3[[:space:]]Designs[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/ec13b3df-7c93-4573-84f4-921e88721107_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiET-GS_[[:space:]]Diffusion[[:space:]]Prior[[:space:]]and[[:space:]]Event[[:space:]]Stream-Assisted[[:space:]]Motion[[:space:]]Deblurring[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/78e0ff72-d618-47c9-b699-d7747a415149_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiGIT_[[:space:]]Multi-Dilated[[:space:]]Gated[[:space:]]Encoder[[:space:]]and[[:space:]]Central-Adjacent[[:space:]]Region[[:space:]]Integrated[[:space:]]Decoder[[:space:]]for[[:space:]]Temporal[[:space:]]Action[[:space:]]Detection[[:space:]]Transformer/140a65e2-34a7-4d35-8450-c8ca2647aff6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiG_[[:space:]]Scalable[[:space:]]and[[:space:]]Efficient[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Gated[[:space:]]Linear[[:space:]]Attention/283f0e42-855f-4918-822c-d5c1219d1790_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiN_[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Robust[[:space:]]Medical[[:space:]]VQA[[:space:]]with[[:space:]]Semantic[[:space:]]Noisy[[:space:]]Labels/3d88adc0-2039-46a2-83ba-83308b59f70d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiSRT-In-Bed_[[:space:]]Diffusion-Based[[:space:]]Sim-to-Real[[:space:]]Transfer[[:space:]]Framework[[:space:]]for[[:space:]]In-Bed[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/e57c9036-ad52-4521-ad16-43207f593def_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiSciPLE_[[:space:]]Learning[[:space:]]Interpretable[[:space:]]Programs[[:space:]]for[[:space:]]Scientific[[:space:]]Visual[[:space:]]Discovery/01086b46-2a5a-4e5d-8038-ecfacbe0a732_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiTASK_[[:space:]]Multi-Task[[:space:]]Fine-Tuning[[:space:]]with[[:space:]]Diffeomorphic[[:space:]]Transformations/3a6773a0-380b-4fb7-a04b-955503269a41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiTCtrl_[[:space:]]Exploring[[:space:]]Attention[[:space:]]Control[[:space:]]in[[:space:]]Multi-Modal[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]for[[:space:]]Tuning-Free[[:space:]]Multi-Prompt[[:space:]]Longer[[:space:]]Video[[:space:]]Generation/b01bdc90-e200-44b3-add7-aa954711fd9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DifIISR_[[:space:]]A[[:space:]]Diffusion[[:space:]]Model[[:space:]]with[[:space:]]Gradient[[:space:]]Guidance[[:space:]]for[[:space:]]Infrared[[:space:]]Image[[:space:]]Super-Resolution/7a53821b-b747-463f-8310-ef76d61abca9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diff-Palm_[[:space:]]Realistic[[:space:]]Palmprint[[:space:]]Generation[[:space:]]with[[:space:]]Polynomial[[:space:]]Creases[[:space:]]and[[:space:]]Intra-Class[[:space:]]Variation[[:space:]]Controllable[[:space:]]Diffusion[[:space:]]Models/c6971867-defc-4d26-9a52-6e7c2c442721_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diff2Flow_[[:space:]]Training[[:space:]]Flow[[:space:]]Matching[[:space:]]Models[[:space:]]via[[:space:]]Diffusion[[:space:]]Model[[:space:]]Alignment/fbb4858e-0a86-4812-87cf-57f08dfebd6c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffCAM_[[:space:]]Data-Driven[[:space:]]Saliency[[:space:]]Maps[[:space:]]by[[:space:]]Capturing[[:space:]]Feature[[:space:]]Differences/bf2fdff3-427c-4aae-800f-e13a073ec24c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffFNO_[[:space:]]Diffusion[[:space:]]Fourier[[:space:]]Neural[[:space:]]Operator/3a04b4ef-d06b-4b04-9301-be75af2eacee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffLO_[[:space:]]Semantic-Aware[[:space:]]LiDAR[[:space:]]Odometry[[:space:]]with[[:space:]]Diffusion-Based[[:space:]]Refinement/0ca80052-baf1-4186-ab94-c84b3011f6df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffLocks_[[:space:]]Generating[[:space:]]3D[[:space:]]Hair[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]using[[:space:]]Diffusion[[:space:]]Models/8616985f-9ce1-4bd9-83a9-440f3fb3d484_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffPortrait360_[[:space:]]Consistent[[:space:]]Portrait[[:space:]]Diffusion[[:space:]]for[[:space:]]360[[:space:]]View[[:space:]]Synthesis/300f1ac6-d9a8-4527-94a8-d8a1857b23e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffSensei_[[:space:]]Bridging[[:space:]]Multi-Modal[[:space:]]LLMs[[:space:]]and[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Customized[[:space:]]Manga[[:space:]]Generation/fada47c4-9d9c-4ce9-8477-3cca25b37313_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffVsgg_[[:space:]]Diffusion-Driven[[:space:]]Online[[:space:]]Video[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/5f1209e6-1544-45b7-80b5-0030e5048d70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Difference[[:space:]]Inversion_[[:space:]]Interpolate[[:space:]]and[[:space:]]Isolate[[:space:]]the[[:space:]]Difference[[:space:]]with[[:space:]]Token[[:space:]]Consistency[[:space:]]for[[:space:]]Image[[:space:]]Analogy[[:space:]]Generation/7ebc23f8-a8e8-4361-9a63-f7a091962245_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Differentiable[[:space:]]Inverse[[:space:]]Rendering[[:space:]]with[[:space:]]Interpretable[[:space:]]Basis[[:space:]]BRDFs/a699140a-0702-4e27-b997-85573651f535_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion[[:space:]]Bridge_[[:space:]]Leveraging[[:space:]]Diffusion[[:space:]]Model[[:space:]]to[[:space:]]Reduce[[:space:]]the[[:space:]]Modality[[:space:]]Gap[[:space:]]Between[[:space:]]Text[[:space:]]and[[:space:]]Vision[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Image[[:space:]]Captioning/a385942f-3738-4670-9bc4-4dd50307bb7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion[[:space:]]Model[[:space:]]is[[:space:]]Effectively[[:space:]]Its[[:space:]]Own[[:space:]]Teacher/0fc82481-a954-458e-b9b5-30892d3cb276_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion[[:space:]]Renderer_[[:space:]]Neural[[:space:]]Inverse[[:space:]]and[[:space:]]Forward[[:space:]]Rendering[[:space:]]with[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/65119a53-c0fc-4b0a-9ea3-a2e6c2c9c1d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion[[:space:]]Self-Distillation[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Customized[[:space:]]Image[[:space:]]Generation/a8e48178-d4e9-44cb-a83c-4fab97048f04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion-4K_[[:space:]]Ultra-High-Resolution[[:space:]]Image[[:space:]]Synthesis[[:space:]]with[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models/4935daf0-51e8-442b-8d44-f641c7d877f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion-based[[:space:]]Event[[:space:]]Generation[[:space:]]for[[:space:]]High-Quality[[:space:]]Image[[:space:]]Deblurring/d2f268e0-066c-441a-8577-0e87bbf4c0b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion-based[[:space:]]Realistic[[:space:]]Listening[[:space:]]Head[[:space:]]Generation[[:space:]]via[[:space:]]Hybrid[[:space:]]Motion[[:space:]]Modeling/54547168-012b-4a96-999b-4f0bfc23f076_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffusionDrive_[[:space:]]Truncated[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]End-to-End[[:space:]]Autonomous[[:space:]]Driving/186eb20a-812b-4b73-a9f1-a45f01468323_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffusionSfM_[[:space:]]Predicting[[:space:]]Structure[[:space:]]and[[:space:]]Motion[[:space:]]via[[:space:]]Ray[[:space:]]Origin[[:space:]]and[[:space:]]Endpoint[[:space:]]Diffusion/a0e804a0-3c59-43c5-bb05-b2160d8636bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Digital[[:space:]]Twin[[:space:]]Catalog_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Photorealistic[[:space:]]3D[[:space:]]Object[[:space:]]Digital[[:space:]]Twin[[:space:]]Dataset/2d29b0d5-5c08-413f-9ad0-63d98b0c6539_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dinomaly_[[:space:]]The[[:space:]]Less[[:space:]]Is[[:space:]]More[[:space:]]Philosophy[[:space:]]in[[:space:]]Multi-Class[[:space:]]Unsupervised[[:space:]]Anomaly[[:space:]]Detection/269a00de-90dc-4a5c-8e52-737661810fb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DirectTriGS_[[:space:]]Triplane-based[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Field[[:space:]]Representation[[:space:]]for[[:space:]]3D[[:space:]]Generation/cb304427-02fe-443c-baf8-93b01f1651af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Directional[[:space:]]Label[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Learning[[:space:]]from[[:space:]]Noisy[[:space:]]Labels/133ce87a-4a21-4fe5-b113-05841acf1ebd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Disco4D_[[:space:]]Disentangled[[:space:]]4D[[:space:]]Human[[:space:]]Generation[[:space:]]and[[:space:]]Animation[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/5f039bdf-5dd5-496e-b90f-b0a10473c8cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiscoVLA_[[:space:]]Discrepancy[[:space:]]Reduction[[:space:]]in[[:space:]]Vision,[[:space:]]Language,[[:space:]]and[[:space:]]Alignment[[:space:]]for[[:space:]]Parameter-Efficient[[:space:]]Video-Text[[:space:]]Retrieval/9d7a1765-54c3-41c7-9719-3ce7d4a7b6fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Discovering[[:space:]]Fine-Grained[[:space:]]Visual-Concept[[:space:]]Relations[[:space:]]by[[:space:]]Disentangled[[:space:]]Optimal[[:space:]]Transport[[:space:]]Concept[[:space:]]Bottleneck[[:space:]]Models/353387ac-1b53-4c3c-ac08-91f95db61c80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Discovering[[:space:]]Hidden[[:space:]]Visual[[:space:]]Concepts[[:space:]]Beyond[[:space:]]Linguistic[[:space:]]Input[[:space:]]in[[:space:]]Infant[[:space:]]Learning/50337d5c-a1b4-4b7f-9096-7ccfa12467ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Discrete[[:space:]]to[[:space:]]Continuous_[[:space:]]Generating[[:space:]]Smooth[[:space:]]Transition[[:space:]]Poses[[:space:]]from[[:space:]]Sign[[:space:]]Language[[:space:]]Observations/4e14bfdb-c7fc-42ce-bc5c-af5b00971815_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Disentangled[[:space:]]Pose[[:space:]]and[[:space:]]Appearance[[:space:]]Guidance[[:space:]]for[[:space:]]Multi-Pose[[:space:]]Generation/eaadccfc-d365-4d2d-83e8-71935f857eca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Disentangling[[:space:]]Safe[[:space:]]and[[:space:]]Unsafe[[:space:]]Image[[:space:]]Corruptions[[:space:]]via[[:space:]]Anisotropy[[:space:]]and[[:space:]]Locality/3a52d538-e6a0-40a8-904e-0b81bce10c6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiskVPS_[[:space:]]Vanishing[[:space:]]Point[[:space:]]Detector[[:space:]]via[[:space:]]Hough[[:space:]]Transform[[:space:]]in[[:space:]]a[[:space:]]Disk[[:space:]]Region/80e52809-79f5-4a63-9312-c7df737fcd10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dispider_[[:space:]]Enabling[[:space:]]Video[[:space:]]LLMs[[:space:]]with[[:space:]]Active[[:space:]]Real-Time[[:space:]]Interaction[[:space:]]via[[:space:]]Disentangled[[:space:]]Perception,[[:space:]]Decision,[[:space:]]and[[:space:]]Reaction/509357f7-50b0-4a3e-be02-5d53b2c39c5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dissecting[[:space:]]and[[:space:]]Mitigating[[:space:]]Diffusion[[:space:]]Bias[[:space:]]via[[:space:]]Mechanistic[[:space:]]Interpretability/3f305fc0-29c2-45c0-bbbf-2f6d947ae1f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Distilled[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Incomplete[[:space:]]Multimodal[[:space:]]Survival[[:space:]]Prediction/5aae416b-ca2c-40f8-8583-f2902bfc3349_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Distilling[[:space:]]Long-tailed[[:space:]]Datasets/f09fbc7b-8c7e-4bee-803c-2d67e3d05b6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Distilling[[:space:]]Monocular[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Fine-grained[[:space:]]Depth[[:space:]]Completion/1663a87e-e3f7-4145-9d6a-27f93678241e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Distilling[[:space:]]Multi-modal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/2bea0a6d-f2ee-4f71-9581-707f2552ee73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Distilling[[:space:]]Spatially-Heterogeneous[[:space:]]Distortion[[:space:]]Perception[[:space:]]for[[:space:]]Blind[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment/c7a78868-95dc-4d35-97c7-04bbcfc1063d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Distilling[[:space:]]Spectral[[:space:]]Graph[[:space:]]for[[:space:]]Object-Context[[:space:]]Aware[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/7ae5aeda-7866-414e-b296-550ab2d843cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DistinctAD_[[:space:]]Distinctive[[:space:]]Audio[[:space:]]Description[[:space:]]Generation[[:space:]]in[[:space:]]Contexts/d4975861-cea8-4f94-9bb1-e02706824334_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Distinguish[[:space:]]Then[[:space:]]Exploit_[[:space:]]Source-free[[:space:]]Open[[:space:]]Set[[:space:]]Domain[[:space:]]Adaptation[[:space:]]via[[:space:]]Weight[[:space:]]Barcode[[:space:]]Estimation[[:space:]]and[[:space:]]Sparse[[:space:]]Label[[:space:]]Assignment/099209ec-4ec9-4836-8000-10ddecb15989_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Distraction[[:space:]]is[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]Jailbreaking/1375bb02-e050-41f2-9b1d-ed16de66c742_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Distribution[[:space:]]Prototype[[:space:]]Diffusion[[:space:]]Learning[[:space:]]for[[:space:]]Open-set[[:space:]]Supervised[[:space:]]Anomaly[[:space:]]Detection/5f6492f6-61fa-40e8-b92f-42fc55a187bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DivPrune_[[:space:]]Diversity-based[[:space:]]Visual[[:space:]]Token[[:space:]]Pruning[[:space:]]for[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models/f2b05006-c955-46f7-a5af-8b7909376348_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiverseFlow_[[:space:]]Sample-Efficient[[:space:]]Diverse[[:space:]]Mode[[:space:]]Coverage[[:space:]]in[[:space:]]Flows/276f7b10-7b9f-439d-a85a-8ad9f72195cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Divide[[:space:]]and[[:space:]]Conquer_[[:space:]]Heterogeneous[[:space:]]Noise[[:space:]]Integration[[:space:]]for[[:space:]]Diffusion-based[[:space:]]Adversarial[[:space:]]Purification/94f1ab27-85c1-491d-abe8-0116716b932b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Divot_[[:space:]]Diffusion[[:space:]]Powers[[:space:]]Video[[:space:]]Tokenizer[[:space:]]for[[:space:]]Comprehension[[:space:]]and[[:space:]]Generation/dd9cf0f7-0275-41db-a729-40fdeab212b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DnLUT_[[:space:]]Ultra-Efficient[[:space:]]Color[[:space:]]Image[[:space:]]Denoising[[:space:]]via[[:space:]]Channel-Aware[[:space:]]Lookup[[:space:]]Tables/efb58f94-4255-404d-9c8c-4e560774a0e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Do[[:space:]]Computer[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models[[:space:]]Learn[[:space:]]the[[:space:]]Low-level[[:space:]]Characteristics[[:space:]]of[[:space:]]the[[:space:]]Human[[:space:]]Visual[[:space:]]System_/4607e0a5-8aa8-470e-b0e4-8cfed36f43c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Do[[:space:]]ImageNet-trained[[:space:]]Models[[:space:]]Learn[[:space:]]Shortcuts_[[:space:]]The[[:space:]]Impact[[:space:]]of[[:space:]]Frequency[[:space:]]Shortcuts[[:space:]]on[[:space:]]Generalization/f684a2ee-5aa0-48af-b0b3-fb7e06614e21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Do[[:space:]]Visual[[:space:]]Imaginations[[:space:]]Improve[[:space:]]Vision-and-Language[[:space:]]Navigation[[:space:]]Agents_/6684e0af-b484-43ef-8443-50d4bfdefdd1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Do[[:space:]]We[[:space:]]Always[[:space:]]Need[[:space:]]the[[:space:]]Simplicity[[:space:]]Bias_[[:space:]]Looking[[:space:]]for[[:space:]]Optimal[[:space:]]Inductive[[:space:]]Biases[[:space:]]in[[:space:]]the[[:space:]]Wild/f8b7dd25-2878-4e2f-ac24-ae160cb54c72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Do[[:space:]]We[[:space:]]Really[[:space:]]Need[[:space:]]Curated[[:space:]]Malicious[[:space:]]Data[[:space:]]for[[:space:]]Safety[[:space:]]Alignment[[:space:]]in[[:space:]]Multi-modal[[:space:]]Large[[:space:]]Language[[:space:]]Models_/7cc50b70-a1c6-4460-8b07-4c06c5509d66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Do[[:space:]]Your[[:space:]]Best[[:space:]]and[[:space:]]Get[[:space:]]Enough[[:space:]]Rest[[:space:]]for[[:space:]]Continual[[:space:]]Learning/364acbb3-eca1-422d-be23-a76c46789b30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DoF-Gaussian_[[:space:]]Controllable[[:space:]]Depth-of-Field[[:space:]]for[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/1ef9852f-e1e4-4571-8156-b74884315f2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DocLayLLM_[[:space:]]An[[:space:]]Efficient[[:space:]]Multi-modal[[:space:]]Extension[[:space:]]of[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Text-rich[[:space:]]Document[[:space:]]Understanding/587c40bc-a1fe-44db-8881-11814790688b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DocSAM_[[:space:]]Unified[[:space:]]Document[[:space:]]Image[[:space:]]Segmentation[[:space:]]via[[:space:]]Query[[:space:]]Decomposition[[:space:]]and[[:space:]]Heterogeneous[[:space:]]Mixed[[:space:]]Learning/cd34a0c4-c9ec-4e5a-a2a8-007f0052f160_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DocVLM_[[:space:]]Make[[:space:]]Your[[:space:]]VLM[[:space:]]an[[:space:]]Efficient[[:space:]]Reader/ec872868-b641-4059-95e8-62856f47ad39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Docopilot_[[:space:]]Improving[[:space:]]Multimodal[[:space:]]Models[[:space:]]for[[:space:]]Document-Level[[:space:]]Understanding/2cc56609-3eb1-412a-98df-28e82ae9cd58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Document[[:space:]]Haystacks_[[:space:]]Vision-Language[[:space:]]Reasoning[[:space:]]Over[[:space:]]Piles[[:space:]]of[[:space:]]1000+[[:space:]]Documents/1b0623c9-cc0a-4950-bcf6-716e67ed00b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Domain[[:space:]]Adaptive[[:space:]]Diabetic[[:space:]]Retinopathy[[:space:]]Grading[[:space:]]with[[:space:]]Model[[:space:]]Absence[[:space:]]and[[:space:]]Flowing[[:space:]]Data/8fc9ee64-90f9-4f33-8fb1-64a5f7b319a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Domain[[:space:]]Generalization[[:space:]]in[[:space:]]CLIP[[:space:]]via[[:space:]]Learning[[:space:]]with[[:space:]]Diverse[[:space:]]Text[[:space:]]Prompts/51839507-6fb2-45c2-a66c-0644bd7936ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Don't[[:space:]]Shake[[:space:]]the[[:space:]]Wheel_[[:space:]]Momentum-Aware[[:space:]]Planning[[:space:]]in[[:space:]]End-to-End[[:space:]]Autonomous[[:space:]]Driving/85994b24-3b2b-443a-a1bf-eabbda00b06b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Doppelgangers[[:space:]]and[[:space:]]Adversarial[[:space:]]Vulnerability/85f12467-2b9c-44c2-8a35-40f17a213b17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Doppelgangers++_[[:space:]]Improved[[:space:]]Visual[[:space:]]Disambiguation[[:space:]]with[[:space:]]Geometric[[:space:]]3D[[:space:]]Features/a4009708-9357-4078-b5aa-d3dd7d42ccfe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DoraCycle_[[:space:]]Domain-Oriented[[:space:]]Adaptation[[:space:]]of[[:space:]]Unified[[:space:]]Generative[[:space:]]Model[[:space:]]in[[:space:]]Multimodal[[:space:]]Cycles/fb2a505a-21d1-4afa-86a3-162bb4136d23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dora_[[:space:]]Sampling[[:space:]]and[[:space:]]Benchmarking[[:space:]]for[[:space:]]3D[[:space:]]Shape[[:space:]]Variational[[:space:]]Auto-Encoders/6b478709-2cef-4587-8775-92fc50ace470_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dr.[[:space:]]Splat_[[:space:]]Directly[[:space:]]Referring[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]via[[:space:]]Direct[[:space:]]Language[[:space:]]Embedding[[:space:]]Registration/04973834-dfc5-4d84-8238-2be21678c244_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DrVideo_[[:space:]]Document[[:space:]]Retrieval[[:space:]]Based[[:space:]]Long[[:space:]]Video[[:space:]]Understanding/81160fea-84bd-416d-98f4-7b4c205b85c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dragin3D_[[:space:]]Image[[:space:]]Editing[[:space:]]by[[:space:]]Dragging[[:space:]]in[[:space:]]3D[[:space:]]Space/63363afe-acf7-462c-bf4b-9a484c6ae8ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DreamCache_[[:space:]]Finetuning-Free[[:space:]]Lightweight[[:space:]]Personalized[[:space:]]Image[[:space:]]Generation[[:space:]]via[[:space:]]Feature[[:space:]]Caching/484a2287-cf26-4bf8-9e41-034501dc6452_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DreamOmni_[[:space:]]Unified[[:space:]]Image[[:space:]]Generation[[:space:]]and[[:space:]]Editing/ff84d3e1-2391-4334-90b8-2fe54fd66a09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DreamRelation_[[:space:]]Bridging[[:space:]]Customization[[:space:]]and[[:space:]]Relation[[:space:]]Generation/dc0bd82f-ae71-4d19-a1ab-522a355a226a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DreamText_[[:space:]]High[[:space:]]Fidelity[[:space:]]Scene[[:space:]]Text[[:space:]]Synthesis/6bc8ea68-7656-478a-8d11-90857d6ca90e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DreamTrack_[[:space:]]Dreaming[[:space:]]the[[:space:]]Future[[:space:]]for[[:space:]]Multimodal[[:space:]]Visual[[:space:]]Object[[:space:]]Tracking/2e4cde12-368f-4feb-b1b3-a980bb1eb002_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DriveDreamer4D_[[:space:]]World[[:space:]]Models[[:space:]]Are[[:space:]]Effective[[:space:]]Data[[:space:]]Machines[[:space:]]for[[:space:]]4D[[:space:]]Driving[[:space:]]Scene[[:space:]]Representation/aa9a31fd-776c-4a2a-a4f7-87c23be589e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DriveGEN_[[:space:]]Generalized[[:space:]]and[[:space:]]Robust[[:space:]]3D[[:space:]]Detection[[:space:]]in[[:space:]]Driving[[:space:]]via[[:space:]]Controllable[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Generation/d5fb6699-8753-41df-8e0e-5b82ff7bf5a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DriveGPT4-V2_[[:space:]]Harnessing[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]Capabilities[[:space:]]for[[:space:]]Enhanced[[:space:]]Closed-Loop[[:space:]]Autonomous[[:space:]]Driving/31ff1a5c-7171-4c25-a1d0-ee2bacb8057b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DriveScape_[[:space:]]High-Resolution[[:space:]]Driving[[:space:]]Video[[:space:]]Generation[[:space:]]by[[:space:]]Multi-View[[:space:]]Feature[[:space:]]Fusion/6698f995-6b34-4c34-b673-bcdca2c64dfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Driving[[:space:]]by[[:space:]]the[[:space:]]Rules_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Integrating[[:space:]]Traffic[[:space:]]Sign[[:space:]]Regulations[[:space:]]into[[:space:]]Vectorized[[:space:]]HD[[:space:]]Map/cbe7a2b6-7e9c-4185-bab6-ca66fed01b85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DrivingSphere_[[:space:]]Building[[:space:]]a[[:space:]]High-fidelity[[:space:]]4D[[:space:]]World[[:space:]]for[[:space:]]Closed-loop[[:space:]]Simulation/e8e4b9c2-6cbc-49b7-a622-765862a5895e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EchoONE_[[:space:]]Segmenting[[:space:]]Multiple[[:space:]]Echocardiography[[:space:]]Planes[[:space:]]in[[:space:]]One[[:space:]]Model/6a4744ba-41ab-4cef-9867-0b6e14ce4cbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EchoTraffic_[[:space:]]Enhancing[[:space:]]Traffic[[:space:]]Anomaly[[:space:]]Understanding[[:space:]]with[[:space:]]Audio-Visual[[:space:]]Insights/15c90442-69f5-4365-9226-debd8012bfbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EchoWorld_[[:space:]]Learning[[:space:]]Motion-Aware[[:space:]]World[[:space:]]Models[[:space:]]for[[:space:]]Echocardiography[[:space:]]Probe[[:space:]]Guidance/4c2c578f-b6e6-4bb8-bca5-82651b3619a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Edge-SD-SR_[[:space:]]Low[[:space:]]Latency[[:space:]]and[[:space:]]Parameter[[:space:]]Efficient[[:space:]]On-device[[:space:]]Super-Resolution[[:space:]]with[[:space:]]Stable[[:space:]]Diffusion[[:space:]]via[[:space:]]Bidirectional[[:space:]]Conditioning/356ad591-f3e9-404d-b0ed-f572cb702b4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EdgeDiff_[[:space:]]Edge-aware[[:space:]]Diffusion[[:space:]]Network[[:space:]]for[[:space:]]Building[[:space:]]Reconstruction[[:space:]]from[[:space:]]Point[[:space:]]Clouds/850fd50e-a122-4c67-a4c5-2d9652be792a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EdgeMovingNet_[[:space:]]Edge-preserving[[:space:]]Point[[:space:]]Cloud[[:space:]]Reconstruction[[:space:]]via[[:space:]]Joint[[:space:]]Geometry[[:space:]]Features/23220cc3-545c-45e3-9457-2e2b54ee6ca0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EdgeTAM_[[:space:]]On-Device[[:space:]]Track[[:space:]]Anything[[:space:]]Model/d2952bf6-f427-43c5-bdc7-9830644a3f4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Edit[[:space:]]Away[[:space:]]and[[:space:]]My[[:space:]]Face[[:space:]]Will[[:space:]]not[[:space:]]Stay_[[:space:]]Personal[[:space:]]Biometric[[:space:]]Defense[[:space:]]against[[:space:]]Malicious[[:space:]]Generative[[:space:]]Editing/4ba5acfe-4d07-46ec-8ebf-879276845fd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EditAR_[[:space:]]Unified[[:space:]]Conditional[[:space:]]Generation[[:space:]]with[[:space:]]Autoregressive[[:space:]]Models/c2effc0c-948f-4818-b255-f19a3566163e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EditSplat_[[:space:]]Multi-View[[:space:]]Fusion[[:space:]]and[[:space:]]Attention-Guided[[:space:]]Optimization[[:space:]]for[[:space:]]View-Consistent[[:space:]]3D[[:space:]]Scene[[:space:]]Editing[[:space:]]with[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/1ab31d7f-9c4f-4c08-b4c6-33f61bccd908_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Effective[[:space:]]Cloud[[:space:]]Removal[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Images[[:space:]]by[[:space:]]an[[:space:]]Improved[[:space:]]Mean-Reverting[[:space:]]Denoising[[:space:]]Model[[:space:]]with[[:space:]]Elucidated[[:space:]]Design[[:space:]]Space/32f0a22f-5e81-428b-90d1-44eed6fa57e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Effective[[:space:]]SAM[[:space:]]Combination[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/a00a7bba-9aca-43bb-ace6-d0318d92c333_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EffiDec3D_[[:space:]]An[[:space:]]Optimized[[:space:]]Decoder[[:space:]]for[[:space:]]High-Performance[[:space:]]and[[:space:]]Efficient[[:space:]]3D[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/544d4697-1ded-4cd9-a237-cab459d082a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]ANN-Guided[[:space:]]Distillation_[[:space:]]Aligning[[:space:]]Rate-based[[:space:]]Features[[:space:]]of[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks[[:space:]]through[[:space:]]Hybrid[[:space:]]Block-wise[[:space:]]Replacement/a92373f1-c17a-4d84-bf48-bb7df7e5d854_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Data[[:space:]]Driven[[:space:]]Mixture-of-Expert[[:space:]]Extraction[[:space:]]from[[:space:]]Trained[[:space:]]Networks/b95664b7-62bd-4911-ad95-8c49b95940ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Decoupled[[:space:]]Feature[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]via[[:space:]]Hierarchical[[:space:]]Compression/afa8d552-54de-4467-a8bd-3d634daa1b37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Depth[[:space:]]Estimation[[:space:]]for[[:space:]]Unstable[[:space:]]Stereo[[:space:]]Camera[[:space:]]Systems[[:space:]]on[[:space:]]AR[[:space:]]Glasses/19ca0a1e-d485-4999-bb0d-5d02efca768c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Diffusion[[:space:]]as[[:space:]]Low[[:space:]]Light[[:space:]]Enhancer/8c4782df-63da-413d-aa18-7a635c0f8202_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Editing[[:space:]]via[[:space:]]4D[[:space:]]Gaussian-based[[:space:]]Static-Dynamic[[:space:]]Separation/fe503b2a-8baf-4bb5-aa0c-7e2878ceee99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Event-Based[[:space:]]Object[[:space:]]Detection_[[:space:]]A[[:space:]]Hybrid[[:space:]]Neural[[:space:]]Network[[:space:]]with[[:space:]]Spatial[[:space:]]and[[:space:]]Temporal[[:space:]]Attention/77c6d014-83fc-41fc-8f14-734704ecb3fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Fine-Tuning[[:space:]]and[[:space:]]Concept[[:space:]]Suppression[[:space:]]for[[:space:]]Pruned[[:space:]]Diffusion[[:space:]]Models/353625d3-f3f3-443b-8ae6-9a0efea45b8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Long[[:space:]]Video[[:space:]]Tokenization[[:space:]]via[[:space:]]Coordinate-based[[:space:]]Patch[[:space:]]Reconstruction/9c0883ab-6511-4236-81b8-36eba3f0168a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Motion-Aware[[:space:]]Video[[:space:]]MLLM/7ffd2a27-ad7c-466b-a41d-bedc0031c7ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Personalization[[:space:]]of[[:space:]]Quantized[[:space:]]Diffusion[[:space:]]Model[[:space:]]without[[:space:]]Backpropagation/1ff8e052-bc28-4f01-9a5c-02b178b96d78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Test-time[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Sensitivity-Guided[[:space:]]Pruning/0c09441c-998a-4565-9e99-39bd1a65e05f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Transfer[[:space:]]Learning[[:space:]]for[[:space:]]Video-language[[:space:]]Foundation[[:space:]]Models/c6082731-485c-4df9-a314-5b595ac25ef4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Video[[:space:]]Face[[:space:]]Enhancement[[:space:]]with[[:space:]]Enhanced[[:space:]]Spatial-Temporal[[:space:]]Consistency/7e225650-66e7-4da7-801c-2c51cddccf77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Video[[:space:]]Super-Resolution[[:space:]]for[[:space:]]Real-time[[:space:]]Rendering[[:space:]]with[[:space:]]Decoupled[[:space:]]G-buffer[[:space:]]Guidance/50ec4cdf-fb40-4b75-b5bc-93dd44213244_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Visual[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]for[[:space:]]Image[[:space:]]Deblurring/67dc408e-a03a-4b59-86fb-e68899316f98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EfficientLLaVA_[[:space:]]Generalizable[[:space:]]Auto-Pruning[[:space:]]for[[:space:]]Large[[:space:]]Vision-language[[:space:]]Models/3bccfadc-3b86-4efc-b5e1-577725d3509f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EfficientViM_[[:space:]]Efficient[[:space:]]Vision[[:space:]]Mamba[[:space:]]with[[:space:]]Hidden[[:space:]]State[[:space:]]Mixer[[:space:]]based[[:space:]]State[[:space:]]Space[[:space:]]Duality/f8b0247c-b417-4c33-a9fc-1cfd54f2104e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Effortless[[:space:]]Active[[:space:]]Labeling[[:space:]]for[[:space:]]Long-Term[[:space:]]Test-Time[[:space:]]Adaptation/b910b4bd-e9d6-42f4-85c4-055c5541d27d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ego4o_[[:space:]]Egocentric[[:space:]]Human[[:space:]]Motion[[:space:]]Capture[[:space:]]and[[:space:]]Understanding[[:space:]]from[[:space:]]Multi-Modal[[:space:]]Input/bdf4a325-0005-40e3-b338-db7cea09365b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EgoLM_[[:space:]]Multi-Modal[[:space:]]Language[[:space:]]Model[[:space:]]of[[:space:]]Egocentric[[:space:]]Motions/9c6a11c3-f8bb-48a3-bf2a-e918227422d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EgoLife_[[:space:]]Towards[[:space:]]Egocentric[[:space:]]Life[[:space:]]Assistant/43ca0f16-dcb8-4428-bc83-1fac13a90c1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EgoPressure_[[:space:]]A[[:space:]]Dataset[[:space:]]for[[:space:]]Hand[[:space:]]Pressure[[:space:]]and[[:space:]]Pose[[:space:]]Estimation[[:space:]]in[[:space:]]Egocentric[[:space:]]Vision/ed82ff29-db1e-411e-bac2-f38592f010bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EgoTextVQA_[[:space:]]Towards[[:space:]]Egocentric[[:space:]]Scene-Text[[:space:]]Aware[[:space:]]Video[[:space:]]Question[[:space:]]Answering/e8489a7b-2b1a-4394-abdf-e73b8a521bd7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EigenGS[[:space:]]Representation_[[:space:]]From[[:space:]]Eigenspace[[:space:]]to[[:space:]]Gaussian[[:space:]]Image[[:space:]]Space/e97db05b-154d-4100-a164-a988ff86955d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Electromyography-Informed[[:space:]]Facial[[:space:]]Expression[[:space:]]Reconstruction[[:space:]]for[[:space:]]Physiological-Based[[:space:]]Synthesis[[:space:]]and[[:space:]]Analysis/a8a37130-c841-4d02-b897-49bb7c51e249_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Embodied[[:space:]]Scene[[:space:]]Understanding[[:space:]]for[[:space:]]Vision[[:space:]]Language[[:space:]]Models[[:space:]]via[[:space:]]MetaVQA/d182c28e-0b3b-4417-add9-9981bfd95c77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Embracing[[:space:]]Collaboration[[:space:]]Over[[:space:]]Competition_[[:space:]]Condensing[[:space:]]Multiple[[:space:]]Prompts[[:space:]]for[[:space:]]Visual[[:space:]]In-Context[[:space:]]Learning/f03c6303-9a89-4209-9afe-1def85a59c19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EmoDubber_[[:space:]]Towards[[:space:]]High[[:space:]]Quality[[:space:]]and[[:space:]]Emotion[[:space:]]Controllable[[:space:]]Movie[[:space:]]Dubbing/63403c47-dde8-48fd-98f5-f7d1a3ecb149_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EmoEdit_[[:space:]]Evoking[[:space:]]Emotions[[:space:]]through[[:space:]]Image[[:space:]]Manipulation/714e8be9-b2ff-416e-8db6-a40fdadf9fa1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EmotiveTalk_[[:space:]]Expressive[[:space:]]Talking[[:space:]]Head[[:space:]]Generation[[:space:]]through[[:space:]]Audio[[:space:]]Information[[:space:]]Decoupling[[:space:]]and[[:space:]]Emotional[[:space:]]Video[[:space:]]Diffusion/d6e74091-6ebd-4568-a571-015f2e2bd2e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Emphasizing[[:space:]]Discriminative[[:space:]]Features[[:space:]]for[[:space:]]Dataset[[:space:]]Distillation[[:space:]]in[[:space:]]Complex[[:space:]]Scenarios/dc6751d1-ffdb-4ae6-b664-3365209a4428_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Empowering[[:space:]]LLMs[[:space:]]to[[:space:]]Understand[[:space:]]and[[:space:]]Generate[[:space:]]Complex[[:space:]]Vector[[:space:]]Graphics/f76f8102-3e6b-4709-b74e-d569ca0c7184_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Empowering[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]with[[:space:]]3D[[:space:]]Situation[[:space:]]Awareness/d5e9b999-444d-4142-a658-922c4ae5d3cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Empowering[[:space:]]Vector[[:space:]]Graphics[[:space:]]with[[:space:]]Consistently[[:space:]]Arbitrary[[:space:]]Viewing[[:space:]]and[[:space:]]View-dependent[[:space:]]Visibility/65cac4ba-7559-4b69-a79f-8c652f0ade6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Encapsulated[[:space:]]Composition[[:space:]]of[[:space:]]Text-to-Image[[:space:]]and[[:space:]]Text-to-Video[[:space:]]Models[[:space:]]for[[:space:]]High-Quality[[:space:]]Video[[:space:]]Synthesis/a1cd2b87-8a80-4784-b442-2201801e4938_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/End-to-End[[:space:]]HOI[[:space:]]Reconstruction[[:space:]]Transformer[[:space:]]with[[:space:]]Graph-based[[:space:]]Encoding/0ae72291-4823-4b36-b1f4-ae8d6ae0b2ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/End-to-End[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations[[:space:]]for[[:space:]]Classification/2a36d7b8-6b8d-43a2-806c-ec5699774233_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enduring,[[:space:]]Efficient[[:space:]]and[[:space:]]Robust[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]Attack[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving[[:space:]]via[[:space:]]Optimization-Driven[[:space:]]Multi-Frame[[:space:]]Perturbation[[:space:]]Framework/6af8cba3-6d1d-43cf-be6c-9b224f253939_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EnergyMoGen_[[:space:]]Compositional[[:space:]]Human[[:space:]]Motion[[:space:]]Generation[[:space:]]with[[:space:]]Energy-Based[[:space:]]Diffusion[[:space:]]Model[[:space:]]in[[:space:]]Latent[[:space:]]Space/9b5a14d1-0c8a-4a03-9c02-4105c534c4d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhanced[[:space:]]Contrastive[[:space:]]Learning[[:space:]]with[[:space:]]Multi-view[[:space:]]Longitudinal[[:space:]]Data[[:space:]]for[[:space:]]Chest[[:space:]]X-ray[[:space:]]Report[[:space:]]Generation/01e446a3-a51b-4451-b41c-ebf00395dc24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhanced[[:space:]]OoD[[:space:]]Detection[[:space:]]through[[:space:]]Cross-Modal[[:space:]]Alignment[[:space:]]of[[:space:]]Multi-Modal[[:space:]]Representations/993054cf-839c-4d32-8159-11fba1ebdae6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhanced[[:space:]]Visual-Semantic[[:space:]]Interaction[[:space:]]with[[:space:]]Tailored[[:space:]]Prompts[[:space:]]for[[:space:]]Pedestrian[[:space:]]Attribute[[:space:]]Recognition/a360def4-3cb3-4af3-8eb9-7293a0985214_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhanced[[:space:]]then[[:space:]]Progressive[[:space:]]Fusion[[:space:]]with[[:space:]]View[[:space:]]Graph[[:space:]]for[[:space:]]Multi-View[[:space:]]Clustering/c551b676-570a-4ac6-b8c4-21db341159bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]3D[[:space:]]Gaze[[:space:]]Estimation[[:space:]]in[[:space:]]the[[:space:]]Wild[[:space:]]using[[:space:]]Weak[[:space:]]Supervision[[:space:]]with[[:space:]]Gaze[[:space:]]Following[[:space:]]Labels/a4ff8470-73eb-452c-b454-46c7c0fd993b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Adversarial[[:space:]]Transferability[[:space:]]with[[:space:]]Checkpoints[[:space:]]of[[:space:]]a[[:space:]]Single[[:space:]]Model's[[:space:]]Training/13456f12-1c01-490f-a682-4804456e112b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Creative[[:space:]]Generation[[:space:]]on[[:space:]]Stable[[:space:]]Diffusion-based[[:space:]]Models/3a0f79af-ac94-4315-b506-26ec5a89c72e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Dance-to-Music[[:space:]]Generation[[:space:]]via[[:space:]]Negative[[:space:]]Conditioning[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Model/5781ac55-8c35-45b6-9198-a0b3776094cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Dataset[[:space:]]Distillation[[:space:]]via[[:space:]]Non-Critical[[:space:]]Region[[:space:]]Refinement/3c74811d-e323-4125-95c8-ddc36051f849_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Diversity[[:space:]]for[[:space:]]Data-free[[:space:]]Quantization/3c9b1dea-51bc-4b58-b282-2fa2dc23bc03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Facial[[:space:]]Privacy[[:space:]]Protection[[:space:]]via[[:space:]]Weakening[[:space:]]Diffusion[[:space:]]Purification/549230a9-ec45-481d-a791-6b5163c0d5c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FAM[[:space:]]Diffusion_[[:space:]]Frequency[[:space:]]and[[:space:]]Attention[[:space:]]Modulation[[:space:]]for[[:space:]]High-Resolution[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Stable[[:space:]]Diffusion/3e37a3c1-9f99-405d-bc64-2d39aebb425d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FASTer_[[:space:]]Focal[[:space:]]token[[:space:]]Acquiring-and-Scaling[[:space:]]Transformer[[:space:]]for[[:space:]]Long-term[[:space:]]3D[[:space:]]Objection[[:space:]]Detection/477b168d-6ce0-48df-ad21-535adab91f4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FATE_[[:space:]]Full-head[[:space:]]Gaussian[[:space:]]Avatar[[:space:]]with[[:space:]]Textural[[:space:]]Editing[[:space:]]from[[:space:]]Monocular[[:space:]]Video/093aeb8d-1f70-4306-9802-b580064da890_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FDS_[[:space:]]Frequency-Aware[[:space:]]Denoising[[:space:]]Score[[:space:]]for[[:space:]]Text-Guided[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Image[[:space:]]Editing/be8eafed-3e00-4305-ab01-f74556639e96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FFR_[[:space:]]Frequency[[:space:]]Feature[[:space:]]Rectification[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/3620d1f3-27b4-44f7-ae22-5b90b5f77a1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FFaceNeRF_[[:space:]]Few-shot[[:space:]]Face[[:space:]]Editing[[:space:]]in[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/9dbc92ea-1e7f-420e-b319-42509a0129d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FG^2_[[:space:]]Fine-Grained[[:space:]]Cross-View[[:space:]]Localization[[:space:]]by[[:space:]]Fine-Grained[[:space:]]Feature[[:space:]]Matching/2d3f03bd-3c76-4e0f-914e-2720e5cc3fad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FIFA_[[:space:]]Fine-grained[[:space:]]Inter-frame[[:space:]]Attention[[:space:]]for[[:space:]]Driver's[[:space:]]Video[[:space:]]Gaze[[:space:]]Estimation/20ee7bef-7a87-4aa4-8319-2c09b2893956_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FIMA-Q_[[:space:]]Post-Training[[:space:]]Quantization[[:space:]]for[[:space:]]Vision[[:space:]]Transformers[[:space:]]by[[:space:]]Fisher[[:space:]]Information[[:space:]]Matrix[[:space:]]Approximation/e9717c35-1f77-46ad-9350-49c7c5216475_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FINECAPTION_[[:space:]]Compositional[[:space:]]Image[[:space:]]Captioning[[:space:]]Focusing[[:space:]]on[[:space:]]Wherever[[:space:]]You[[:space:]]Want[[:space:]]at[[:space:]]Any[[:space:]]Granularity/0db6f7cd-2703-4d38-a868-9404e9fb7332_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FIRE_[[:space:]]Robust[[:space:]]Detection[[:space:]]of[[:space:]]Diffusion-Generated[[:space:]]Images[[:space:]]via[[:space:]]Frequency-Guided[[:space:]]Reconstruction[[:space:]]Error/b72f6d0f-44bf-4fb2-8d26-217f2ba3422f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FIction_[[:space:]]4D[[:space:]]Future[[:space:]]Interaction[[:space:]]Prediction[[:space:]]from[[:space:]]Video/f9d00e26-c1ef-4eaa-ac97-7ad68f5431ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FLAIR_[[:space:]]VLM[[:space:]]with[[:space:]]Fine-grained[[:space:]]Language-informed[[:space:]]Image[[:space:]]Representations/7b174c25-5a18-4075-8bda-8c790f2d02b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FLAME_[[:space:]]Frozen[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]Enable[[:space:]]Data-Efficient[[:space:]]Language-Image[[:space:]]Pre-training/f455399c-0cf4-4e4b-9e94-386dfa46498e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FLARE_[[:space:]]Feed-forward[[:space:]]Geometry,[[:space:]]Appearance[[:space:]]and[[:space:]]Camera[[:space:]]Estimation[[:space:]]from[[:space:]]Uncalibrated[[:space:]]Sparse[[:space:]]Views/97ed67d5-0bee-44a8-b85c-f1500309792c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FLAVC_[[:space:]]Learned[[:space:]]Video[[:space:]]Compression[[:space:]]with[[:space:]]Feature[[:space:]]Level[[:space:]]Attention/2f2dc54d-25fe-4c8b-b224-44ed7622a38d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FOCUS_[[:space:]]Knowledge-enhanced[[:space:]]Adaptive[[:space:]]Visual[[:space:]]Compression[[:space:]]for[[:space:]]Few-shot[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Classification/9032ceb9-c148-487d-859e-96c2c44f2055_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FRAMES-VQA_[[:space:]]Benchmarking[[:space:]]Fine-Tuning[[:space:]]Robustness[[:space:]]across[[:space:]]Multi-Modal[[:space:]]Shifts[[:space:]]in[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/6d4d0ec7-45de-4d8c-a80b-268960516fd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FRAME_[[:space:]]Floor-aligned[[:space:]]Representation[[:space:]]for[[:space:]]Avatar[[:space:]]Motion[[:space:]]from[[:space:]]Egocentric[[:space:]]Video/95315ae8-8940-443b-b27a-d65001fc70e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FRESA_[[:space:]]Feedforward[[:space:]]Reconstruction[[:space:]]of[[:space:]]Personalized[[:space:]]Skinned[[:space:]]Avatars[[:space:]]from[[:space:]]Few[[:space:]]Images/1c334927-b801-4747-bd12-ed0953c344da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FSBench_[[:space:]]A[[:space:]]Figure[[:space:]]Skating[[:space:]]Benchmark[[:space:]]for[[:space:]]Advancing[[:space:]]Artistic[[:space:]]Sports[[:space:]]Understanding/27730523-8de5-4899-bb19-7c9df5e62723_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FSFM_[[:space:]]A[[:space:]]Generalizable[[:space:]]Face[[:space:]]Security[[:space:]]Foundation[[:space:]]Model[[:space:]]via[[:space:]]Self-Supervised[[:space:]]Facial[[:space:]]Representation[[:space:]]Learning/f07b1021-4b39-4ced-9714-eac497748e59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FSHNet_[[:space:]]Fully[[:space:]]Sparse[[:space:]]Hybrid[[:space:]]Network[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/0d33325e-218a-4ce8-b625-b5aa9cf481f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FSboard_[[:space:]]Over[[:space:]]3[[:space:]]Million[[:space:]]Characters[[:space:]]of[[:space:]]ASL[[:space:]]Fingerspelling[[:space:]]Collected[[:space:]]via[[:space:]]Smartphones/4326e803-f6c5-446c-a3ea-3cf12fe898b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/F^3OCUS[[:space:]]-[[:space:]]Federated[[:space:]]Finetuning[[:space:]]of[[:space:]]Vision-Language[[:space:]]Foundation[[:space:]]Models[[:space:]]with[[:space:]]Optimal[[:space:]]Client[[:space:]]Layer[[:space:]]Updating[[:space:]]Strategy[[:space:]]via[[:space:]]Multi-objective[[:space:]]Meta-Heuristics/9ae8604b-cbaa-4f7e-a06a-596949f1978c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Face[[:space:]]Forgery[[:space:]]Video[[:space:]]Detection[[:space:]]via[[:space:]]Temporal[[:space:]]Forgery[[:space:]]Cue[[:space:]]Unraveling/62b2f1b5-ae73-442a-94a3-e6079ef2f31b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FaceBench_[[:space:]]A[[:space:]]Multi-View[[:space:]]Multi-Level[[:space:]]Facial[[:space:]]Attribute[[:space:]]VQA[[:space:]]Dataset[[:space:]]for[[:space:]]Benchmarking[[:space:]]Face[[:space:]]Perception[[:space:]]MLLMs/a5fabdae-90a5-4133-afe8-0aef3e92b13f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FactCheXcker_[[:space:]]Mitigating[[:space:]]Measurement[[:space:]]Hallucinations[[:space:]]in[[:space:]]Chest[[:space:]]X-ray[[:space:]]Report[[:space:]]Generation[[:space:]]Models/cbda5d01-f1d8-49a2-a659-8fddae1b1038_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Factored-NeuS_[[:space:]]Reconstructing[[:space:]]Surfaces,[[:space:]]Illumination,[[:space:]]and[[:space:]]Materials[[:space:]]of[[:space:]]Possibly[[:space:]]Glossy[[:space:]]Objects/00cf5fc8-67cd-4926-8e07-03c6b52a818e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FaithDiff_[[:space:]]Unleashing[[:space:]]Diffusion[[:space:]]Priors[[:space:]]for[[:space:]]Faithful[[:space:]]Image[[:space:]]Super-resolution/8bb104c0-a9b0-4dc4-aac3-66b0a0d6999e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fancy123_[[:space:]]One[[:space:]]Image[[:space:]]to[[:space:]]High-Quality[[:space:]]3D[[:space:]]Mesh[[:space:]]Generation[[:space:]]via[[:space:]]Plug-and-Play[[:space:]]Deformation/7d8719c8-236a-45a1-86e7-8c8821381958_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fast[[:space:]]and[[:space:]]Accurate[[:space:]]Gigapixel[[:space:]]Pathological[[:space:]]Image[[:space:]]Classification[[:space:]]with[[:space:]]Hierarchical[[:space:]]Distillation[[:space:]]Multi-Instance[[:space:]]Learning/a1ccc860-6f38-4808-89b7-527a477bbe29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fast3R_[[:space:]]Towards[[:space:]]3D[[:space:]]Reconstruction[[:space:]]of[[:space:]]1000+[[:space:]]Images[[:space:]]in[[:space:]]One[[:space:]]Forward[[:space:]]Pass/2fa150f4-3fed-43f5-887d-26d025351bc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FastVLM_[[:space:]]Efficient[[:space:]]Vision[[:space:]]Encoding[[:space:]]for[[:space:]]Vision[[:space:]]Language[[:space:]]Models/8d0b4bb8-c92c-432b-b7c6-63fb3a1ca776_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Faster[[:space:]]Parameter-Efficient[[:space:]]Tuning[[:space:]]with[[:space:]]Token[[:space:]]Redundancy[[:space:]]Reduction/b0bace7b-f431-4750-91a6-a42b4d9dc8e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Feat2GS_[[:space:]]Probing[[:space:]]Visual[[:space:]]Foundation[[:space:]]Models[[:space:]]with[[:space:]]Gaussian[[:space:]]Splatting/6ac035b3-e3e3-4cff-bfe8-b53368183dbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Feature[[:space:]]Information[[:space:]]Driven[[:space:]]Position[[:space:]]Gaussian[[:space:]]Distribution[[:space:]]Estimation[[:space:]]for[[:space:]]Tiny[[:space:]]Object[[:space:]]Detection/7ce98538-99ab-4828-8e03-f7f9dee132d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Feature[[:space:]]Selection[[:space:]]for[[:space:]]Latent[[:space:]]Factor[[:space:]]Models/7e062dcb-59b2-4607-96da-99b55a2e6916_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Feature[[:space:]]Spectrum[[:space:]]Learning[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Change[[:space:]]Detection/fae65852-ac43-4e5a-8d15-59b7b4b5c70e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Feature-Preserving[[:space:]]Mesh[[:space:]]Decimation[[:space:]]for[[:space:]]Normal[[:space:]]Integration/bba150d2-1552-4b97-aa14-eeb6faabd18a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Feature4X_[[:space:]]Bridging[[:space:]]Any[[:space:]]Monocular[[:space:]]Video[[:space:]]to[[:space:]]4D[[:space:]]Agentic[[:space:]]AI[[:space:]]with[[:space:]]Versatile[[:space:]]Gaussian[[:space:]]Feature[[:space:]]Fields/fd4e1353-7335-4d8e-a4c7-940bdf92aab0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedAWA_[[:space:]]Adaptive[[:space:]]Optimization[[:space:]]of[[:space:]]Aggregation[[:space:]]Weights[[:space:]]in[[:space:]]Federated[[:space:]]Learning[[:space:]]Using[[:space:]]Client[[:space:]]Vectors/a469d4f1-1fda-4a0a-8342-17bfa293f28a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedBiP_[[:space:]]Heterogeneous[[:space:]]One-Shot[[:space:]]Federated[[:space:]]Learning[[:space:]]with[[:space:]]Personalized[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models/0ed597f2-b8d7-49a9-b3be-02d2ea0ed5b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedCALM_[[:space:]]Conflict-aware[[:space:]]Layer-wise[[:space:]]Mitigation[[:space:]]for[[:space:]]Selective[[:space:]]Aggregation[[:space:]]in[[:space:]]Deeper[[:space:]]Personalized[[:space:]]Federated[[:space:]]Learning/1c036a7b-34be-473f-8a90-56d50794a12f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedCS_[[:space:]]Coreset[[:space:]]Selection[[:space:]]for[[:space:]]Federated[[:space:]]Learning/13885f33-fe13-4839-bcd7-9058a64976e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedMIA_[[:space:]]An[[:space:]]Effective[[:space:]]Membership[[:space:]]Inference[[:space:]]Attack[[:space:]]Exploiting[[:space:]]'All[[:space:]]for[[:space:]]One'[[:space:]]Principle[[:space:]]in[[:space:]]Federated[[:space:]]Learning/e38dea68-001b-432a-810c-4953d472354a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedSPA_[[:space:]]Generalizable[[:space:]]Federated[[:space:]]Graph[[:space:]]Learning[[:space:]]under[[:space:]]Homophily[[:space:]]Heterogeneity/e80c4081-86f2-4f9d-8d7c-eb349d6d40a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Federated[[:space:]]Learning[[:space:]]with[[:space:]]Domain[[:space:]]Shift[[:space:]]Eraser/7de3aa95-e3aa-4d9f-8611-6c70bcb529c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FeedEdit_[[:space:]]Text-Based[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Dynamic[[:space:]]Feedback[[:space:]]Regulation/35fef65c-6843-480a-8a11-f4a52a5a1637_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ferret_[[:space:]]An[[:space:]]Efficient[[:space:]]Online[[:space:]]Continual[[:space:]]Learning[[:space:]]Framework[[:space:]]under[[:space:]]Varying[[:space:]]Memory[[:space:]]Constraints/56b462c7-5e10-4270-a4b9-ab748d41ad76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Few-Shot[[:space:]]Recognition[[:space:]]via[[:space:]]Stage-Wise[[:space:]]Retrieval-Augmented[[:space:]]Finetuning/dc585257-676b-4adf-876f-823827d1244b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Few-shot[[:space:]]Implicit[[:space:]]Function[[:space:]]Generation[[:space:]]via[[:space:]]Equivariance/40a82437-6241-47fe-a55a-6d55522b9825_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Few-shot[[:space:]]Personalized[[:space:]]Scanpath[[:space:]]Prediction/23c1c5ba-bf51-4bb4-947a-523ea02f84e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FiRe_[[:space:]]Fixed-points[[:space:]]of[[:space:]]Restoration[[:space:]]Priors[[:space:]]for[[:space:]]Solving[[:space:]]Inverse[[:space:]]Problems/7cbc1099-ba4e-4310-b131-2927aea31492_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FilmComposer_[[:space:]]LLM-Driven[[:space:]]Music[[:space:]]Production[[:space:]]for[[:space:]]Silent[[:space:]]Film[[:space:]]Clips/a1924ad4-889a-45b7-b151-5805ab628df3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Filter[[:space:]]Images[[:space:]]First,[[:space:]]Generate[[:space:]]Instructions[[:space:]]Later_[[:space:]]Pre-Instruction[[:space:]]Data[[:space:]]Selection[[:space:]]for[[:space:]]Visual[[:space:]]Instruction[[:space:]]Tuning/71aaae33-8513-4df9-9728-41d6ee1347c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Finding[[:space:]]Local[[:space:]]Diffusion[[:space:]]Schrodinger[[:space:]]Bridge[[:space:]]using[[:space:]]Kolmogorov-Arnold[[:space:]]Network/745e7c66-ba5d-4acb-8a31-fb6ce188f276_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fine-Grained[[:space:]]Erasure[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion-based[[:space:]]Foundation[[:space:]]Models/66a8f222-b7af-40e1-b35a-fd525fd29ff6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fine-Grained[[:space:]]Image-Text[[:space:]]Correspondence[[:space:]]with[[:space:]]Cost[[:space:]]Aggregation[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Part[[:space:]]Segmentation/7212f9fe-3cf2-4695-ba56-29d90b16a026_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FineLIP_[[:space:]]Extending[[:space:]]CLIP's[[:space:]]Reach[[:space:]]via[[:space:]]Fine-Grained[[:space:]]Alignment[[:space:]]with[[:space:]]Longer[[:space:]]Text[[:space:]]Inputs/2c937ee0-3b0e-468f-85e0-c38703c58f7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FinePhys_[[:space:]]Fine-grained[[:space:]]Human[[:space:]]Action[[:space:]]Generation[[:space:]]by[[:space:]]Explicitly[[:space:]]Incorporating[[:space:]]Physical[[:space:]]Laws[[:space:]]for[[:space:]]Effective[[:space:]]Skeletal[[:space:]]Guidance/6acb2047-2b14-4153-aa31-644a4162ebe8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FineVQ_[[:space:]]Fine-Grained[[:space:]]User[[:space:]]Generated[[:space:]]Content[[:space:]]Video[[:space:]]Quality[[:space:]]Assessment/de34e382-77ec-4d1c-86d9-9e2539b53cce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Finer-CAM_[[:space:]]Spotting[[:space:]]the[[:space:]]Difference[[:space:]]Reveals[[:space:]]Finer[[:space:]]Details[[:space:]]for[[:space:]]Visual[[:space:]]Explanation/71b1cfde-d55f-4589-9551-177375791ac9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fingerprinting[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Probabilistic[[:space:]]Models/6e22ca4f-2d2c-43e1-acdb-ac49c8f93530_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Finsler[[:space:]]Multi-Dimensional[[:space:]]Scaling_[[:space:]]Manifold[[:space:]]Learning[[:space:]]for[[:space:]]Asymmetric[[:space:]]Dimensionality[[:space:]]Reduction[[:space:]]and[[:space:]]Embedding/bd3f829a-3cb6-4c47-a17e-cfe83243c346_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FireEdit_[[:space:]]Fine-grained[[:space:]]Instruction-based[[:space:]]Image[[:space:]]Editing[[:space:]]via[[:space:]]Region-aware[[:space:]]Vision[[:space:]]Language[[:space:]]Model/9ae051ea-855d-4406-887e-17845c492db5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FirePlace_[[:space:]]Geometric[[:space:]]Refinements[[:space:]]of[[:space:]]LLM[[:space:]]Common[[:space:]]Sense[[:space:]]Reasoning[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Placement/2d9eb4e8-dc35-430c-ba31-3544cf31c83a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fish-Vista_[[:space:]]A[[:space:]]Multi-Purpose[[:space:]]Dataset[[:space:]]for[[:space:]]Understanding[[:space:]]&[[:space:]]Identification[[:space:]]of[[:space:]]Traits[[:space:]]from[[:space:]]Images/6a78cb30-05e6-4057-bd67-96a012a922cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FisherTune_[[:space:]]Fisher-Guided[[:space:]]Robust[[:space:]]Tuning[[:space:]]of[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models[[:space:]]for[[:space:]]Domain[[:space:]]Generalized[[:space:]]Segmentation/358dc5c4-3215-4e7b-a412-af201b89e5d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fitted[[:space:]]Neural[[:space:]]Lossless[[:space:]]Image[[:space:]]Compression/b8e3351a-d8f9-4455-a341-cdaf822c6322_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Flash-Split_[[:space:]]2D[[:space:]]Reflection[[:space:]]Removal[[:space:]]with[[:space:]]Flash[[:space:]]Cues[[:space:]]and[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Separation/67294752-bef5-41b2-b364-0f75370031d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Flash3D_[[:space:]]Super-scaling[[:space:]]Point[[:space:]]Transformers[[:space:]]through[[:space:]]Joint[[:space:]]Hardware-Geometry[[:space:]]Locality/ce686829-e0f5-45b4-9183-dd0704f3066a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlashGS_[[:space:]]Efficient[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Large-scale[[:space:]]and[[:space:]]High-resolution[[:space:]]Rendering/e01d4bca-80e5-429d-87f3-cd3db7ee5401_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlashSloth[[:space:]]_[[:space:]]Lightning[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]via[[:space:]]Embedded[[:space:]]Visual[[:space:]]Compression/d4b19b9c-9d1d-498e-9b19-b5c396475d3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlexDrive_[[:space:]]Toward[[:space:]]Trajectory[[:space:]]Flexibility[[:space:]]in[[:space:]]Driving[[:space:]]Scene[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Reconstruction[[:space:]]and[[:space:]]Rendering/2be603d4-a79b-4590-8429-8b7af1b01129_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlexGS_[[:space:]]Train[[:space:]]Once,[[:space:]]Deploy[[:space:]]Everywhere[[:space:]]with[[:space:]]Many-in-One[[:space:]]Flexible[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/c797b236-3d9e-40bb-b0b1-d0dc0f5a3502_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlexUOD_[[:space:]]The[[:space:]]Answer[[:space:]]to[[:space:]]Real-world[[:space:]]Unsupervised[[:space:]]Image[[:space:]]Outlier[[:space:]]Detection/d07dc3fe-2b9b-4302-b969-68dfe2c11eab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlexiDiT_[[:space:]]Your[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]Can[[:space:]]Easily[[:space:]]Generate[[:space:]]High-Quality[[:space:]]Samples[[:space:]]with[[:space:]]Less[[:space:]]Compute/ba9432e2-eef1-4061-9e8a-050a8d67d402_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Flexible[[:space:]]Frame[[:space:]]Selection[[:space:]]for[[:space:]]Efficient[[:space:]]Video[[:space:]]Reasoning/4dcc1958-37a2-4aae-b081-4baf3e1441f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Flexible[[:space:]]Group[[:space:]]Count[[:space:]]Enables[[:space:]]Hassle-Free[[:space:]]Structured[[:space:]]Pruning/f8a1d8b3-81d9-415d-9304-66df03966ce5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlipSketch_[[:space:]]Flipping[[:space:]]Static[[:space:]]Drawings[[:space:]]to[[:space:]]Text-Guided[[:space:]]Sketch[[:space:]]Animations/1d72e70c-7046-494b-9893-b6181e9cdb1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FloVD_[[:space:]]Optical[[:space:]]Flow[[:space:]]Meets[[:space:]]Video[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Enhanced[[:space:]]Camera-Controlled[[:space:]]Video[[:space:]]Synthesis/966774cf-bb68-4461-b5a3-b31cc1ba4561_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Floating[[:space:]]No[[:space:]]More_[[:space:]]Object-Ground[[:space:]]Reconstruction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/234d2cb0-7a7a-45a1-b505-1e44b0d20706_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Florence-VL_[[:space:]]Enhancing[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]Generative[[:space:]]Vision[[:space:]]Encoder[[:space:]]and[[:space:]]Depth-Breadth[[:space:]]Fusion/6a04e9fb-e863-4e01-95ba-2bebbd8a8b84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Flow-NeRF_[[:space:]]Joint[[:space:]]Learning[[:space:]]of[[:space:]]Geometry,[[:space:]]Poses,[[:space:]]and[[:space:]]Dense[[:space:]]Flow[[:space:]]within[[:space:]]Unified[[:space:]]Neural[[:space:]]Representations/87139e27-1b9a-406f-ae0b-8a239d8ca4b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlowRAM_[[:space:]]Grounding[[:space:]]Flow[[:space:]]Matching[[:space:]]Policy[[:space:]]with[[:space:]]Region-Aware[[:space:]]Mamba[[:space:]]Framework[[:space:]]for[[:space:]]Robotic[[:space:]]Manipulation/219f1c70-9f4a-41f6-a31f-cb2715c328ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Flowing[[:space:]]from[[:space:]]Words[[:space:]]to[[:space:]]Pixels_[[:space:]]A[[:space:]]Noise-Free[[:space:]]Framework[[:space:]]for[[:space:]]Cross-Modality[[:space:]]Evolution/0581b6c1-94c8-4009-b458-0f373f9cc05d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Floxels_[[:space:]]Fast[[:space:]]Unsupervised[[:space:]]Voxel[[:space:]]Based[[:space:]]Scene[[:space:]]Flow[[:space:]]Estimation/88cffa8c-8150-4222-9b07-76a43f6aba38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FluidNexus_[[:space:]]3D[[:space:]]Fluid[[:space:]]Reconstruction[[:space:]]and[[:space:]]Prediction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Video/e4503ea8-3d5b-4337-98a2-56ea5dbbd12d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FluxSpace_[[:space:]]Disentangled[[:space:]]Semantic[[:space:]]Editing[[:space:]]in[[:space:]]Rectified[[:space:]]Flow[[:space:]]Models/14236874-d228-4c7e-9b78-b1e5ea0f1948_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Focal[[:space:]]Split_[[:space:]]Untethered[[:space:]]Snapshot[[:space:]]Depth[[:space:]]from[[:space:]]Differential[[:space:]]Defocus/8add8a0d-471c-44c0-87fe-2571b98821d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Focus-N-Fix_[[:space:]]Region-Aware[[:space:]]Fine-Tuning[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/8727fce4-4e88-40ef-abb2-a1db80ac26e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Focusing[[:space:]]on[[:space:]]Tracks[[:space:]]for[[:space:]]Online[[:space:]]Multi-Object[[:space:]]Tracking/67d31135-4d8b-4b67-bf27-5c8f033b0ea5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Foley-Flow_[[:space:]]Coordinated[[:space:]]Video-to-Audio[[:space:]]Generation[[:space:]]with[[:space:]]Masked[[:space:]]Audio-Visual[[:space:]]Alignment[[:space:]]and[[:space:]]Dynamic[[:space:]]Conditional[[:space:]]Flows/8ceb2ede-3593-42a3-b921-c7aa1ce66414_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Font-Agent_[[:space:]]Enhancing[[:space:]]Font[[:space:]]Understanding[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Models/707c109f-359c-49ee-8480-2050ec105040_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Forensic[[:space:]]Self-Descriptions[[:space:]]Are[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Detection,[[:space:]]Open-Set[[:space:]]Source[[:space:]]Attribution,[[:space:]]and[[:space:]]Clustering[[:space:]]of[[:space:]]AI-generated[[:space:]]Images/87976d40-931c-47a2-bcaf-b745052fbd36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Forensics[[:space:]]Adapter_[[:space:]]Adapting[[:space:]]CLIP[[:space:]]for[[:space:]]Generalizable[[:space:]]Face[[:space:]]Forgery[[:space:]]Detection/ec785d8c-afb5-4f14-9f38-489e7dc539d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Forensics-Bench_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Forgery[[:space:]]Detection[[:space:]]Benchmark[[:space:]]Suite[[:space:]]for[[:space:]]Large[[:space:]]Vision[[:space:]]Language[[:space:]]Models/83c663b5-660a-4692-b13d-48208f86daf4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ForestLPR_[[:space:]]LiDAR[[:space:]]Place[[:space:]]Recognition[[:space:]]in[[:space:]]Forests[[:space:]]Attentioning[[:space:]]Multiple[[:space:]]BEV[[:space:]]Density[[:space:]]Images/3c3183c5-6a09-4af1-a678-a67924209220_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Forming[[:space:]]Auxiliary[[:space:]]High-confident[[:space:]]Instance-level[[:space:]]Loss[[:space:]]to[[:space:]]Promote[[:space:]]Learning[[:space:]]from[[:space:]]Label[[:space:]]Proportions/e6c1b7f1-cd1a-45f7-93ae-0283147fc948_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fortifying[[:space:]]Federated[[:space:]]Learning[[:space:]]Towards[[:space:]]Trustworthiness[[:space:]]via[[:space:]]Auditable[[:space:]]Data[[:space:]]Valuation[[:space:]]and[[:space:]]Verifiable[[:space:]]Client[[:space:]]Contribution/cb677217-8984-47eb-bd4b-bb49c3b6070a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FoundHand_[[:space:]]Large-Scale[[:space:]]Domain-Specific[[:space:]]Learning[[:space:]]for[[:space:]]Controllable[[:space:]]Hand[[:space:]]Image[[:space:]]Generation/3fc7fb54-1364-4b29-a608-c8709d1e7905_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FoundationStereo_[[:space:]]Zero-Shot[[:space:]]Stereo[[:space:]]Matching/734bacd4-e447-4f2c-a566-c37065b51bda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Foundations[[:space:]]of[[:space:]]the[[:space:]]Theory[[:space:]]of[[:space:]]Performance-Based[[:space:]]Ranking/c7153ec5-6f67-49d7-959e-763d5058e7dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Foveated[[:space:]]Instance[[:space:]]Segmentation/1c308ffa-0a64-4b5b-acf7-4ce7a04a15e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fractal[[:space:]]Calibration[[:space:]]for[[:space:]]Long-tailed[[:space:]]Object[[:space:]]Detection/667776a5-1a09-42ab-b116-7192711cfdcd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Free[[:space:]]Lunch[[:space:]]Enhancements[[:space:]]for[[:space:]]Multi-modal[[:space:]]Crowd[[:space:]]Counting/592f2751-8f2f-452b-bae9-54f0aba0aa5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Free[[:space:]]on[[:space:]]the[[:space:]]Fly_[[:space:]]Enhancing[[:space:]]Flexibility[[:space:]]in[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]with[[:space:]]Online[[:space:]]EM/0815d6a9-01e9-40b1-9aaa-f036cad150fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Free-viewpoint[[:space:]]Human[[:space:]]Animation[[:space:]]with[[:space:]]Pose-correlated[[:space:]]Reference[[:space:]]Selection/f0181534-1c14-4a13-9f88-4bf3f60875a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Free360_[[:space:]]Layered[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Unbounded[[:space:]]360-Degree[[:space:]]View[[:space:]]Synthesis[[:space:]]from[[:space:]]Extremely[[:space:]]Sparse[[:space:]]and[[:space:]]Unposed[[:space:]]Views/3b609597-ff5b-48f0-aa88-d17f12adcf41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeCloth_[[:space:]]Free-form[[:space:]]Generation[[:space:]]Enhances[[:space:]]Challenging[[:space:]]Clothed[[:space:]]Human[[:space:]]Modeling/4141aa76-1005-436d-8cb0-f6d77273eee7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeGave_[[:space:]]3D[[:space:]]Physics[[:space:]]Learning[[:space:]]from[[:space:]]Dynamic[[:space:]]Videos[[:space:]]by[[:space:]]Gaussian[[:space:]]Velocity/f143731d-7b79-4838-b6b7-1020d6921417_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreePCA_[[:space:]]Integrating[[:space:]]Consistency[[:space:]]Information[[:space:]]across[[:space:]]Long-short[[:space:]]Frames[[:space:]]in[[:space:]]Training-free[[:space:]]Long[[:space:]]Video[[:space:]]Generation[[:space:]]via[[:space:]]Principal[[:space:]]Component[[:space:]]Analysis/381516e2-ec54-415c-a358-bfdcfcbf1449_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeScene_[[:space:]]Mixed[[:space:]]Graph[[:space:]]Diffusion[[:space:]]for[[:space:]]3D[[:space:]]Scene[[:space:]]Synthesis[[:space:]]from[[:space:]]Free[[:space:]]Prompts/0b354f5d-dd04-4e78-b087-e955350ac914_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeSim_[[:space:]]Toward[[:space:]]Free-viewpoint[[:space:]]Camera[[:space:]]Simulation[[:space:]]in[[:space:]]Driving[[:space:]]Scenes/d18396fa-b946-43a6-855d-113b04859445_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeTimeGS_[[:space:]]Free[[:space:]]Gaussian[[:space:]]Primitives[[:space:]]at[[:space:]]Anytime[[:space:]]Anywhere[[:space:]]for[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Reconstruction/0863a1ab-42f5-4bd1-87fc-eaaa221406f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeUV_[[:space:]]Ground-Truth-Free[[:space:]]Realistic[[:space:]]Facial[[:space:]]UV[[:space:]]Texture[[:space:]]Recovery[[:space:]]via[[:space:]]Cross-Assembly[[:space:]]Inference[[:space:]]Strategy/adc5da15-9676-45bb-8ac5-74b6ae1c1717_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreqDebias_[[:space:]]Towards[[:space:]]Generalizable[[:space:]]Deepfake[[:space:]]Detection[[:space:]]via[[:space:]]Consistency-Driven[[:space:]]Frequency[[:space:]]Debiasing/0e52a600-27ae-4d07-b6aa-0fadf6110e7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Frequency[[:space:]]Dynamic[[:space:]]Convolution[[:space:]]for[[:space:]]Dense[[:space:]]Image[[:space:]]Prediction/c58c8d7f-298f-4286-8814-86e3184dcbfe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Frequency-Biased[[:space:]]Synergistic[[:space:]]Design[[:space:]]for[[:space:]]Image[[:space:]]Compression[[:space:]]and[[:space:]]Compensation/6c3be626-b963-48da-929d-fb16dda6136d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Alexnet[[:space:]]to[[:space:]]Transformers_[[:space:]]Measuring[[:space:]]the[[:space:]]Non-linearity[[:space:]]of[[:space:]]Deep[[:space:]]Neural[[:space:]]Networks[[:space:]]with[[:space:]]Affine[[:space:]]Optimal[[:space:]]Transport/09181469-48dc-41c1-aa5c-d4125e19dcc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Elements[[:space:]]to[[:space:]]Design_[[:space:]]A[[:space:]]Layered[[:space:]]Approach[[:space:]]for[[:space:]]Automatic[[:space:]]Graphic[[:space:]]Design[[:space:]]Composition/e26f69db-15f4-419d-88c4-2a02ec184415_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Faces[[:space:]]to[[:space:]]Voices_[[:space:]]Learning[[:space:]]Hierarchical[[:space:]]Representations[[:space:]]for[[:space:]]High-quality[[:space:]]Video-to-Speech/b37694a7-c54b-4fe3-950c-612eb2dcb8bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Head[[:space:]]to[[:space:]]Tail_[[:space:]]Efficient[[:space:]]Black-box[[:space:]]Model[[:space:]]Inversion[[:space:]]Attack[[:space:]]via[[:space:]]Long-tailed[[:space:]]Learning/28fcc9c8-3c8d-4c79-997a-0ae994a6cdd3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Head[[:space:]]to[[:space:]]Tail_[[:space:]]Towards[[:space:]]Balanced[[:space:]]Representation[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]through[[:space:]]Adaptive[[:space:]]Data[[:space:]]Calibration/e432652c-1f9a-4c26-b84e-bad8c918b3d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Laboratory[[:space:]]to[[:space:]]Real[[:space:]]World_[[:space:]]A[[:space:]]New[[:space:]]Benchmark[[:space:]]Towards[[:space:]]Privacy-Preserved[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification/380f6e9e-2288-454b-8de0-71124c8876c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Multimodal[[:space:]]LLMs[[:space:]]to[[:space:]]Generalist[[:space:]]Embodied[[:space:]]Agents_[[:space:]]Methods[[:space:]]and[[:space:]]Lessons/1b7e8e76-ff54-4987-930c-d6f90da4a70a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Poses[[:space:]]to[[:space:]]Identity_[[:space:]]Training-Free[[:space:]]Person[[:space:]]Re-Identification[[:space:]]via[[:space:]]Feature[[:space:]]Centralization/3e6e9709-b16e-4f18-80e0-a44471331402_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Prototypes[[:space:]]to[[:space:]]General[[:space:]]Distributions_[[:space:]]An[[:space:]]Efficient[[:space:]]Curriculum[[:space:]]for[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling/c6159d46-4939-46db-b94d-59083f02c29b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Slow[[:space:]]Bidirectional[[:space:]]to[[:space:]]Fast[[:space:]]Autoregressive[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/b2ffd5fd-b5ac-4897-b2a4-1d563ffb24b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Sparse[[:space:]]Signal[[:space:]]to[[:space:]]Smooth[[:space:]]Motion_[[:space:]]Real-Time[[:space:]]Motion[[:space:]]Generation[[:space:]]with[[:space:]]Rolling[[:space:]]Prediction[[:space:]]Models/c452ffb7-04a2-44c1-a8b2-ae53277622ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Sparse[[:space:]]to[[:space:]]Dense_[[:space:]]Camera[[:space:]]Relocalization[[:space:]]with[[:space:]]Scene-Specific[[:space:]]Detector[[:space:]]from[[:space:]]Feature[[:space:]]Gaussian[[:space:]]Splatting/50245dab-5a19-477e-b581-c7447aec1672_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Words[[:space:]]to[[:space:]]Structured[[:space:]]Visuals_[[:space:]]A[[:space:]]Benchmark[[:space:]]and[[:space:]]Framework[[:space:]]for[[:space:]]Text-to-Diagram[[:space:]]Generation[[:space:]]and[[:space:]]Editing/b43a58e1-768b-4ee2-8a44-1df51a1a6e84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Zero[[:space:]]to[[:space:]]Detail_[[:space:]]Deconstructing[[:space:]]Ultra-High-Definition[[:space:]]Image[[:space:]]Restoration[[:space:]]from[[:space:]]Progressive[[:space:]]Spectral[[:space:]]Perspective/cb1f1623-d084-495b-add2-3e5db807da57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FrugalNeRF_[[:space:]]Fast[[:space:]]Convergence[[:space:]]for[[:space:]]Extreme[[:space:]]Few-shot[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]without[[:space:]]Learned[[:space:]]Priors/f71ab808-e3e1-40b9-8ace-08ca6f2900ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FruitNinja_[[:space:]]3D[[:space:]]Object[[:space:]]Interior[[:space:]]Texture[[:space:]]Generation[[:space:]]with[[:space:]]Gaussian[[:space:]]Splatting/050f0e16-3e87-44be-a032-3e2e34eea118_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Full-DoF[[:space:]]Egomotion[[:space:]]Estimation[[:space:]]for[[:space:]]Event[[:space:]]Cameras[[:space:]]Using[[:space:]]Geometric[[:space:]]Solvers/7ad63892-5cd4-4647-a2fc-f6e2b06118b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Functionality[[:space:]]Understanding[[:space:]]and[[:space:]]Segmentation[[:space:]]in[[:space:]]3D[[:space:]]Scenes/18987ad0-ed7e-436c-91af-77496a1471e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fuzzy[[:space:]]Multimodal[[:space:]]Learning[[:space:]]for[[:space:]]Trusted[[:space:]]Cross-modal[[:space:]]Retrieval/8e74d6b4-e89a-4f43-8305-1bc23fa5aab3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/G3Flow_[[:space:]]Generative[[:space:]]3D[[:space:]]Semantic[[:space:]]Flow[[:space:]]for[[:space:]]Pose-aware[[:space:]]and[[:space:]]Generalizable[[:space:]]Object[[:space:]]Manipulation/15a9c302-d3eb-4ccb-99e6-ea38fcbe428a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GA3CE_[[:space:]]Unconstrained[[:space:]]3D[[:space:]]Gaze[[:space:]]Estimation[[:space:]]with[[:space:]]Gaze-Aware[[:space:]]3D[[:space:]]Context[[:space:]]Encoding/841c0218-e13c-41c1-a424-0cd450aedace_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GAF_[[:space:]]Gaussian[[:space:]]Avatar[[:space:]]Reconstruction[[:space:]]from[[:space:]]Monocular[[:space:]]Videos[[:space:]]via[[:space:]]Multi-view[[:space:]]Diffusion/5739633e-e898-4f2f-81e3-fa1b00f5b35d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GASP_[[:space:]]Gaussian[[:space:]]Avatars[[:space:]]with[[:space:]]Synthetic[[:space:]]Priors/d247ca07-2946-4be9-899f-324e8ca02a97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GBC-Splat_[[:space:]]Generalizable[[:space:]]Gaussian-Based[[:space:]]Clothed[[:space:]]Human[[:space:]]Digitalization[[:space:]]under[[:space:]]Sparse[[:space:]]RGB[[:space:]]Cameras/20951abf-de1b-48a7-bf0f-e1bd9dc0e545_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GBlobs_[[:space:]]Explicit[[:space:]]Local[[:space:]]Structure[[:space:]]via[[:space:]]Gaussian[[:space:]]Blobs[[:space:]]for[[:space:]]Improved[[:space:]]Cross-Domain[[:space:]]LiDAR-based[[:space:]]3D[[:space:]]Object[[:space:]]Detection/5f1dad4e-b7bf-42ef-9853-4a1d03484c87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GCC_[[:space:]]Generative[[:space:]]Color[[:space:]]Constancy[[:space:]]via[[:space:]]Diffusing[[:space:]]a[[:space:]]Color[[:space:]]Checker/19398c97-5525-41b6-9cec-21c633b65f2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GCE-Pose_[[:space:]]Global[[:space:]]Context[[:space:]]Enhancement[[:space:]]for[[:space:]]Category-level[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/db8de138-0542-4261-9ff0-0d7ee9664e38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GEAL_[[:space:]]Generalizable[[:space:]]3D[[:space:]]Affordance[[:space:]]Learning[[:space:]]with[[:space:]]Cross-Modal[[:space:]]Consistency/27619916-f232-40d9-8ed7-1570950e227c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GEM_[[:space:]]A[[:space:]]Generalizable[[:space:]]Ego-Vision[[:space:]]Multimodal[[:space:]]World[[:space:]]Model[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Ego-Motion,[[:space:]]Object[[:space:]]Dynamics,[[:space:]]and[[:space:]]Scene[[:space:]]Composition[[:space:]]Control/69cbc9f4-e195-4400-aa12-a3f08979efa4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GEN3C_[[:space:]]3D-Informed[[:space:]]World-Consistent[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Precise[[:space:]]Camera[[:space:]]Control/be6efe74-4722-4e39-b990-ff6d7aaa58e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GENIUS_[[:space:]]A[[:space:]]Generative[[:space:]]Framework[[:space:]]for[[:space:]]Universal[[:space:]]Multimodal[[:space:]]Search/2dc13f62-0b07-4c28-98f9-8e8b8fc34c34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GENMANIP_[[:space:]]LLM-driven[[:space:]]Simulation[[:space:]]for[[:space:]]Generalizable[[:space:]]Instruction-Following[[:space:]]Manipulation/26fffdd7-add8-4f3f-9503-fb9b8b381d57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GET_[[:space:]]Unlocking[[:space:]]the[[:space:]]Multi-modal[[:space:]]Potential[[:space:]]of[[:space:]]CLIP[[:space:]]for[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/bf0f86bd-1187-4fc3-beb0-0601502bd2a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GFlowVLM_[[:space:]]Enhancing[[:space:]]Multi-step[[:space:]]Reasoning[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]Generative[[:space:]]Flow[[:space:]]Networks/38b5b823-6ad9-48ae-a690-8f89656872f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GG-SSMs_[[:space:]]Graph-Generating[[:space:]]State[[:space:]]Space[[:space:]]Models/525d6b57-74d2-4b85-93e8-6da80d5894ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GIFStream_[[:space:]]4D[[:space:]]Gaussian-based[[:space:]]Immersive[[:space:]]Video[[:space:]]with[[:space:]]Feature[[:space:]]Stream/33c4cb1b-90b4-4850-be8f-6584ce453927_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GIF_[[:space:]]Generative[[:space:]]Inspiration[[:space:]]for[[:space:]]Face[[:space:]]Recognition[[:space:]]at[[:space:]]Scale/4188b325-1be7-4257-a7d1-1b7186da592a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GIVEPose_[[:space:]]Gradual[[:space:]]Intra-class[[:space:]]Variation[[:space:]]Elimination[[:space:]]for[[:space:]]RGB-based[[:space:]]Category-Level[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/74284d56-ff7f-49e5-87fb-07e23d96fd61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GLASS_[[:space:]]Guided[[:space:]]Latent[[:space:]]Slot[[:space:]]Diffusion[[:space:]]for[[:space:]]Object-Centric[[:space:]]Learning/d8c60ee5-3478-4708-b2ca-c21eda0f7b68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GLUS_[[:space:]]Global-Local[[:space:]]Reasoning[[:space:]]Unified[[:space:]]into[[:space:]]A[[:space:]]Single[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Video[[:space:]]Segmentation/ef07b75d-66d5-4945-96d1-5f5a99ca7b27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GLane3D_[[:space:]]Detecting[[:space:]]Lanes[[:space:]]with[[:space:]]Graph[[:space:]]of[[:space:]]3D[[:space:]]Keypoints/4714bdb4-dc28-4cae-a7ae-7ffafc50426c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GO-N3RDet_[[:space:]]Geometry[[:space:]]Optimized[[:space:]]NeRF-enhanced[[:space:]]3D[[:space:]]Object[[:space:]]Detector/7c1da70a-81e6-441c-89b7-f7eb7da1940f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GOAL_[[:space:]]Global-local[[:space:]]Object[[:space:]]Alignment[[:space:]]Learning/455f3e8d-bd3d-4cf1-9666-034870b03ec3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GPAvatar_[[:space:]]High-fidelity[[:space:]]Head[[:space:]]Avatars[[:space:]]by[[:space:]]Learning[[:space:]]Efficient[[:space:]]Gaussian[[:space:]]Projections/002ddf96-d838-45a0-9433-9314dd546b65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GPS[[:space:]]as[[:space:]]a[[:space:]]Control[[:space:]]Signal[[:space:]]for[[:space:]]Image[[:space:]]Generation/fb020720-5cd1-4ec1-b851-4b2ba4a4bb47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GPVK-VL_[[:space:]]Geometry-Preserving[[:space:]]Virtual[[:space:]]Keyframes[[:space:]]for[[:space:]]Visual[[:space:]]Localization[[:space:]]under[[:space:]]Large[[:space:]]Viewpoint[[:space:]]Changes/82762777-8953-425c-9e90-277d69d21c6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GRAE-3DMOT_[[:space:]]Geometry[[:space:]]Relation-Aware[[:space:]]Encoder[[:space:]]for[[:space:]]Online[[:space:]]3D[[:space:]]Multi-Object[[:space:]]Tracking/fe0e12b2-f8a5-4541-af49-2841a78d6212_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GRAPHGPT-O_[[:space:]]Synergistic[[:space:]]Multimodal[[:space:]]Comprehension[[:space:]]and[[:space:]]Generation[[:space:]]on[[:space:]]Graphs/ae3321d5-e1b5-4b83-8869-bfb6b2f6f9c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GREAT_[[:space:]]Geometry-Intention[[:space:]]Collaborative[[:space:]]Inference[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Object[[:space:]]Affordance[[:space:]]Grounding/b5c8369b-de9c-4d99-93b1-f6601c114ce4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GROVE_[[:space:]]A[[:space:]]Generalized[[:space:]]Reward[[:space:]]for[[:space:]]Learning[[:space:]]Open-Vocabulary[[:space:]]Physical[[:space:]]Skill/b9a8d3a2-7981-46c6-ae28-01d1e2e7cc72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GS-2DGS_[[:space:]]Geometrically[[:space:]]Supervised[[:space:]]2DGS[[:space:]]for[[:space:]]Reflective[[:space:]]Object[[:space:]]Reconstruction/79957440-026b-451e-9f83-4c70bd515bfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GS-DiT_[[:space:]]Advancing[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Dynamic[[:space:]]3D[[:space:]]Gaussian[[:space:]]Fields[[:space:]]through[[:space:]]Efficient[[:space:]]Dense[[:space:]]3D[[:space:]]Point[[:space:]]Tracking/ceff57aa-9b85-401c-a708-5eb29ea3d4fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GUI-Xplore_[[:space:]]Empowering[[:space:]]Generalizable[[:space:]]GUI[[:space:]]Agents[[:space:]]with[[:space:]]One[[:space:]]Exploration/96223897-44a0-4065-9d17-2c58ffaf497e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaPT-DAR_[[:space:]]Category-level[[:space:]]Garments[[:space:]]Pose[[:space:]]Tracking[[:space:]]via[[:space:]]Integrated[[:space:]]2D[[:space:]]Deformation[[:space:]]and[[:space:]]3D[[:space:]]Reconstruction/d027132e-3bf9-4d7c-9bd4-f692f19995eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gain[[:space:]]from[[:space:]]Neighbors_[[:space:]]Boosting[[:space:]]Model[[:space:]]Robustness[[:space:]]in[[:space:]]the[[:space:]]Wild[[:space:]]via[[:space:]]Adversarial[[:space:]]Perturbations[[:space:]]Toward[[:space:]]Neighboring[[:space:]]Classes/7c63026d-2d9d-4b66-8d1b-a2c2921ef0e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Galaxy[[:space:]]Walker_[[:space:]]Geometry-aware[[:space:]]VLMs[[:space:]]For[[:space:]]Galaxy-scale[[:space:]]Understanding/04de0f5c-8b5f-4f0a-bb58-7578988a37cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GarmentPile_[[:space:]]Point-Level[[:space:]]Visual[[:space:]]Affordance[[:space:]]Guided[[:space:]]Retrieval[[:space:]]and[[:space:]]Adaptation[[:space:]]for[[:space:]]Cluttered[[:space:]]Garments[[:space:]]Manipulation/73202280-8f0d-40ee-8d2f-4c610e438f1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GauCho_[[:space:]]Gaussian[[:space:]]Distributions[[:space:]]with[[:space:]]Cholesky[[:space:]]Decomposition[[:space:]]for[[:space:]]Oriented[[:space:]]Object[[:space:]]Detection/baba7653-b2d6-4a64-a240-46fb4bbdd4fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GauSTAR_[[:space:]]Gaussian[[:space:]]Surface[[:space:]]Tracking[[:space:]]and[[:space:]]Reconstruction/11bb7a23-25ce-4ef1-b886-826f7b1153ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussHDR_[[:space:]]High[[:space:]]Dynamic[[:space:]]Range[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]via[[:space:]]Learning[[:space:]]Unified[[:space:]]3D[[:space:]]and[[:space:]]2D[[:space:]]Local[[:space:]]Tone[[:space:]]Mapping/e14a29dd-740e-4f78-9681-b8b924cc1d69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussTR_[[:space:]]Foundation[[:space:]]Model-Aligned[[:space:]]Gaussian[[:space:]]Transformer[[:space:]]for[[:space:]]Self-Supervised[[:space:]]3D[[:space:]]Spatial[[:space:]]Understanding/819561f2-210f-4b94-b03b-de849e04dfec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gaussian[[:space:]]Eigen[[:space:]]Models[[:space:]]for[[:space:]]Human[[:space:]]Heads/2c1696ee-36e9-4806-aaa1-feae8c398ad6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gaussian[[:space:]]Splashing_[[:space:]]Unified[[:space:]]Particles[[:space:]]for[[:space:]]Versatile[[:space:]]Motion[[:space:]]Synthesis[[:space:]]and[[:space:]]Rendering/6796e064-8f63-454c-a668-20a4fe162433_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gaussian[[:space:]]Splatting[[:space:]]Feature[[:space:]]Fields[[:space:]]for[[:space:]](Privacy-Preserving)[[:space:]]Visual[[:space:]]Localization/8785c1b3-3a2c-4b74-9182-c74acd59a715_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Efficient[[:space:]]Satellite[[:space:]]Image[[:space:]]Photogrammetry/825fbfda-a75e-4fca-b8a1-998d52a87f24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussianFormer-2_[[:space:]]Probabilistic[[:space:]]Gaussian[[:space:]]Superposition[[:space:]]for[[:space:]]Efficient[[:space:]]3D[[:space:]]Occupancy[[:space:]]Prediction/a73b2ed6-43d9-4834-8360-b8cadd6f9853_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussianIP_[[:space:]]Identity-Preserving[[:space:]]Realistic[[:space:]]3D[[:space:]]Human[[:space:]]Generation[[:space:]]via[[:space:]]Human-Centric[[:space:]]Diffusion[[:space:]]Prior/e4417971-cb30-44b1-83c3-d7223e397594_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussianSpa_[[:space:]]An[[:space:]]'Optimizing-Sparsifying'[[:space:]]Simplification[[:space:]]Framework[[:space:]]for[[:space:]]Compact[[:space:]]and[[:space:]]High-Quality[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/92653a9e-0441-4df9-bec6-d6792327d348_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussianUDF_[[:space:]]Inferring[[:space:]]Unsigned[[:space:]]Distance[[:space:]]Functions[[:space:]]through[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/cc720cf9-8b93-45e3-ba1c-285231dc2941_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussianWorld_[[:space:]]Gaussian[[:space:]]World[[:space:]]Model[[:space:]]for[[:space:]]Streaming[[:space:]]3D[[:space:]]Occupancy[[:space:]]Prediction/8c4a5e8a-78b9-406e-8a9f-9a67644df1b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gaze-LLE_[[:space:]]Gaze[[:space:]]Target[[:space:]]Estimation[[:space:]]via[[:space:]]Large-Scale[[:space:]]Learned[[:space:]]Encoders/0c5cb704-88c0-44b0-a9fd-1cc849c31acf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GazeGene_[[:space:]]Large-scale[[:space:]]Synthetic[[:space:]]Gaze[[:space:]]Dataset[[:space:]]with[[:space:]]3D[[:space:]]Eyeball[[:space:]]Annotations/0340f6ba-f654-499d-a449-278c45c4ae21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gazing[[:space:]]Into[[:space:]]Missteps_[[:space:]]Leveraging[[:space:]]Eye-Gaze[[:space:]]for[[:space:]]Unsupervised[[:space:]]Mistake[[:space:]]Detection[[:space:]]in[[:space:]]Egocentric[[:space:]]Videos[[:space:]]of[[:space:]]Skilled[[:space:]]Human[[:space:]]Activities/b5c8a7bd-3790-4667-bb6b-1a0b6e103c3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gazing[[:space:]]at[[:space:]]Rewards_[[:space:]]Eye[[:space:]]Movements[[:space:]]as[[:space:]]a[[:space:]]Lens[[:space:]]into[[:space:]]Human[[:space:]]and[[:space:]]AI[[:space:]]Decision-Making[[:space:]]in[[:space:]]Hybrid[[:space:]]Visual[[:space:]]Foraging/84b60c8f-4259-403d-bbe3-ef55dd238828_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gen3DEval_[[:space:]]Using[[:space:]]vLLMs[[:space:]]for[[:space:]]Automatic[[:space:]]Evaluation[[:space:]]of[[:space:]]Generated[[:space:]]3D[[:space:]]Objects/7305263f-e2cc-4b0d-96ef-36ee756fb132_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GenAssets_[[:space:]]Generating[[:space:]]in-the-wild[[:space:]]3D[[:space:]]Assets[[:space:]]in[[:space:]]Latent[[:space:]]Space/205bb4f6-3a46-4d03-9a48-3c1a7f69cf02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GenDeg_[[:space:]]Diffusion-based[[:space:]]Degradation[[:space:]]Synthesis[[:space:]]for[[:space:]]Generalizable[[:space:]]All-In-One[[:space:]]Image[[:space:]]Restoration/b4fb985e-0447-4144-8fd3-3575bae6c402_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GenFusion_[[:space:]]Closing[[:space:]]the[[:space:]]Loop[[:space:]]between[[:space:]]Reconstruction[[:space:]]and[[:space:]]Generation[[:space:]]via[[:space:]]Videos/73ed4fdc-aae9-40bf-822a-c5433b756d42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GenPC_[[:space:]]Zero-shot[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion[[:space:]]via[[:space:]]3D[[:space:]]Generative[[:space:]]Priors/33065446-96bd-487f-842a-0ad7ebde7571_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GenVDM_[[:space:]]Generating[[:space:]]Vector[[:space:]]Displacement[[:space:]]Maps[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Image/8d964c4a-db05-4b2a-9f2e-ebfc143e1e1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalizable[[:space:]]Object[[:space:]]Keypoint[[:space:]]Localization[[:space:]]from[[:space:]]Generative[[:space:]]Priors/36248270-69b3-4db8-b5d8-67f377d993c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalized[[:space:]]Diffusion[[:space:]]Detector_[[:space:]]Mining[[:space:]]Robust[[:space:]]Features[[:space:]]from[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Domain-Generalized[[:space:]]Detection/9415c498-7b1c-4b14-a48b-6f96313f1e11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalized[[:space:]]Few-shot[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation[[:space:]]with[[:space:]]Vision-Language[[:space:]]Model/4d12541d-b94d-4ecc-89f9-33d0d766ec07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalized[[:space:]]Gaussian[[:space:]]Entropy[[:space:]]Model[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Attribute[[:space:]]Compression[[:space:]]with[[:space:]]Dynamic[[:space:]]Likelihood[[:space:]]Intervals/84766216-4935-47fb-beab-09b4479a5fe6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalized[[:space:]]Recorrupted-to-Recorrupted_[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]Beyond[[:space:]]Gaussian[[:space:]]Noise/4a399214-37f3-46f8-855d-0f3fd3139d1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalized[[:space:]]Zero-Shot[[:space:]]Classification[[:space:]]via[[:space:]]Semantics-Free[[:space:]]Inter-Class[[:space:]]Feature[[:space:]]Generation/4c6af842-822a-4f56-88b9-2c47f8a724af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalizing[[:space:]]Deepfake[[:space:]]Video[[:space:]]Detection[[:space:]]with[[:space:]]Plug-and-Play_[[:space:]]Video-Level[[:space:]]Blending[[:space:]]and[[:space:]]Spatiotemporal[[:space:]]Adapter[[:space:]]Tuning/48b07e5e-c255-49a3-8ff8-83d6d753b112_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generating[[:space:]]3D-Consistent[[:space:]]Videos[[:space:]]from[[:space:]]Unposed[[:space:]]Internet[[:space:]]Photos/790bb30a-ea83-4877-950f-b62c93b4c597_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generating[[:space:]]6DoF[[:space:]]Object[[:space:]]Manipulation[[:space:]]Trajectories[[:space:]]from[[:space:]]Action[[:space:]]Description[[:space:]]in[[:space:]]Egocentric[[:space:]]Vision/a7bb5ce1-7d9c-4738-bf7a-97b31482722c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generating[[:space:]]Multimodal[[:space:]]Driving[[:space:]]Scenes[[:space:]]via[[:space:]]Next-Scene[[:space:]]Prediction/1aabe5c9-9237-4556-aa02-4c15cda1e42a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Densification_[[:space:]]Learning[[:space:]]to[[:space:]]Densify[[:space:]]Gaussians[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Generalizable[[:space:]]3D[[:space:]]Reconstruction/215c0109-6e38-4d85-ad1d-7392385bad3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Unbounded[[:space:]]3D[[:space:]]City[[:space:]]Generation/bc06109b-e030-4ac7-8ca6-a35b49065ee7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Hard[[:space:]]Example[[:space:]]Augmentation[[:space:]]for[[:space:]]Semantic[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation/8ee300b0-345b-4162-b735-be71099a4813_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Image[[:space:]]Layer[[:space:]]Decomposition[[:space:]]with[[:space:]]Visual[[:space:]]Effects/11d057aa-65a1-4fa7-a909-a2fd048d22a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Inbetweening[[:space:]]through[[:space:]]Frame-wise[[:space:]]Conditions-Driven[[:space:]]Video[[:space:]]Generation/3ff2b572-eacd-49c1-8a28-878d8c332139_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Map[[:space:]]Priors[[:space:]]for[[:space:]]Collaborative[[:space:]]BEV[[:space:]]Semantic[[:space:]]Segmentation/36370215-ddbe-432c-8353-c63d16f3f636_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Modeling[[:space:]]of[[:space:]]Class[[:space:]]Probability[[:space:]]for[[:space:]]Multi-Modal[[:space:]]Representation[[:space:]]Learning/c5d2944a-daf5-4608-8271-564aa7374577_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Multimodal[[:space:]]Pretraining[[:space:]]with[[:space:]]Discrete[[:space:]]Diffusion[[:space:]]Timestep[[:space:]]Tokens/86a5b2d7-cf9b-474e-9670-fe50842b8517_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Multiview[[:space:]]Relighting[[:space:]]for[[:space:]]3D[[:space:]]Reconstruction[[:space:]]under[[:space:]]Extreme[[:space:]]Illumination[[:space:]]Variation/3ac11737-3d24-418e-a146-137fbf30be6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Omnimatte_[[:space:]]Learning[[:space:]]to[[:space:]]Decompose[[:space:]]Video[[:space:]]into[[:space:]]Layers/5f3a827c-9a41-4a7b-9ee3-c0a0bbc081a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Photography_[[:space:]]Scene-Consistent[[:space:]]Camera[[:space:]]Control[[:space:]]for[[:space:]]Realistic[[:space:]]Text-to-Image[[:space:]]Synthesis/978efb00-1bc8-42f1-8084-40af3f1940f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Photomontage/0b6084aa-fa74-448a-b438-c03b03f778ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Sparse-View[[:space:]]Gaussian[[:space:]]Splatting/27218994-b0dd-41df-b98c-cee3df8a76bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Video[[:space:]]Propagation/4e0e48f6-0d84-44e1-8cb2-586713761f5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Zero-Shot[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/34978457-438a-44a4-8e52-8d1c37bd09c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GeoAvatar_[[:space:]]Geometrically-Consistent[[:space:]]Multi-Person[[:space:]]Avatar[[:space:]]Reconstruction[[:space:]]from[[:space:]]Sparse[[:space:]]Multi-View[[:space:]]Videos/f397471e-cfcf-48f7-9378-eb4b6f702a0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GeoDepth_[[:space:]]From[[:space:]]Point-to-Depth[[:space:]]to[[:space:]]Plane-to-Depth[[:space:]]Modeling[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/537c4a85-d199-4566-a929-9bc9ad24de25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GeoMM_[[:space:]]On[[:space:]]Geodesic[[:space:]]Perspective[[:space:]]for[[:space:]]Multi-modal[[:space:]]Learning/d4315402-247c-421b-b312-673d8b6cdab6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Geometric[[:space:]]Knowledge-Guided[[:space:]]Localized[[:space:]]Global[[:space:]]Distribution[[:space:]]Alignment[[:space:]]for[[:space:]]Federated[[:space:]]Learning/2e8bcf78-83dd-44ff-95cd-e28155d181e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Geometry[[:space:]]Field[[:space:]]Splatting[[:space:]]with[[:space:]]Gaussian[[:space:]]Surfels/3a0d3f71-20ab-4575-84f9-85c9f121d8c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Geometry[[:space:]]in[[:space:]]Style_[[:space:]]3D[[:space:]]Stylization[[:space:]]via[[:space:]]Surface[[:space:]]Normal[[:space:]]Deformation/e5ea4625-10ae-4517-8bd1-667690326f38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Geometry-guided[[:space:]]Online[[:space:]]3D[[:space:]]Video[[:space:]]Synthesis[[:space:]]with[[:space:]]Multi-View[[:space:]]Temporal[[:space:]]Consistency/fd6ec0d4-1321-4b1f-92b0-c5b96b417f72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ges3ViG[[:space:]]_[[:space:]]Incorporating[[:space:]]Pointing[[:space:]]Gestures[[:space:]]into[[:space:]]Language-Based[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding[[:space:]]for[[:space:]]Embodied[[:space:]]Reference[[:space:]]Understanding/ec35ce3f-41d5-4f23-9f54-b9db850c483a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GigaHands_[[:space:]]A[[:space:]]Massive[[:space:]]Annotated[[:space:]]Dataset[[:space:]]of[[:space:]]Bimanual[[:space:]]Hand[[:space:]]Activities/7e78827a-b786-4f2c-b29a-be8f35f69a64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GliaNet_[[:space:]]Adaptive[[:space:]]Neural[[:space:]]Network[[:space:]]Structure[[:space:]]Learning[[:space:]]with[[:space:]]Glia-Driven/14a6eb3f-fa56-4a0c-8c03-16b148e4e1f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Global-Local[[:space:]]Tree[[:space:]]Search[[:space:]]in[[:space:]]VLMs[[:space:]]for[[:space:]]3D[[:space:]]Indoor[[:space:]]Scene[[:space:]]Generation/ff74ed7c-49e9-43a6-b17b-55256feeb33c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Glossy[[:space:]]Object[[:space:]]Reconstruction[[:space:]]with[[:space:]]Cost-effective[[:space:]]Polarized[[:space:]]Acquisition/6d010cc8-fb53-4d56-9087-d9ceb52e7326_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GlyphMastero_[[:space:]]A[[:space:]]Glyph[[:space:]]Encoder[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Scene[[:space:]]Text[[:space:]]Editing/cf25e8de-63ec-4bb5-a4d0-e916a98c19ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Go-with-the-Flow_[[:space:]]Motion-Controllable[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models[[:space:]]Using[[:space:]]Real-Time[[:space:]]Warped[[:space:]]Noise/8005106b-2bb8-494d-8579-b7bbda4809e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GoLF-NRT_[[:space:]]Integrating[[:space:]]Global[[:space:]]Context[[:space:]]and[[:space:]]Local[[:space:]]Geometry[[:space:]]for[[:space:]]Few-Shot[[:space:]]View[[:space:]]Synthesis/2ba13e93-be62-4898-abd0-91ade550876e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GoalFlow_[[:space:]]Goal-Driven[[:space:]]Flow[[:space:]]Matching[[:space:]]for[[:space:]]Multimodal[[:space:]]Trajectories[[:space:]]Generation[[:space:]]in[[:space:]]End-to-End[[:space:]]Autonomous[[:space:]]Driving/17a20542-ff14-4f3e-9dd4-82ee3d6b0f53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Goku_[[:space:]]Flow[[:space:]]Based[[:space:]]Video[[:space:]]Generative[[:space:]]Foundation[[:space:]]Models/c8ea5822-d85e-4da4-8c6b-983d1a9f6203_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Golden[[:space:]]Cudgel[[:space:]]Network[[:space:]]for[[:space:]]Real-Time[[:space:]]Semantic[[:space:]]Segmentation/8768da67-8697-4f82-be5d-8dc99774d481_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Good,[[:space:]]Cheap,[[:space:]]and[[:space:]]Fast_[[:space:]]Overfitted[[:space:]]Image[[:space:]]Compression[[:space:]]with[[:space:]]Wasserstein[[:space:]]Distortion/3adcbb11-ea71-4691-b359-0fd4fc14ffb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gradient[[:space:]]Inversion[[:space:]]Attacks[[:space:]]on[[:space:]]Parameter-Efficient[[:space:]]Fine-Tuning/66a6c1ec-dd02-4f6c-9f73-01c221a373e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gradient-Guided[[:space:]]Annealing[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/54a78395-d6cb-4eba-b998-5fcc900a7c04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Graph[[:space:]]Neural[[:space:]]Network[[:space:]]Combining[[:space:]]Event[[:space:]]Stream[[:space:]]and[[:space:]]Periodic[[:space:]]Aggregation[[:space:]]for[[:space:]]Low-Latency[[:space:]]Event-based[[:space:]]Vision/69080078-d8c4-41bd-8b17-567af86a81c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Graph-Embedded[[:space:]]Structure-Aware[[:space:]]Perceptual[[:space:]]Hashing[[:space:]]for[[:space:]]Neural[[:space:]]Network[[:space:]]Protection[[:space:]]and[[:space:]]Piracy[[:space:]]Detection/81ff9977-894e-42bc-883f-f88b07af7671_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GraphI2P_[[:space:]]Image-to-Point[[:space:]]Cloud[[:space:]]Registration[[:space:]]with[[:space:]]Exploring[[:space:]]Pattern[[:space:]]of[[:space:]]Correspondence[[:space:]]via[[:space:]]Graph[[:space:]]Learning/475e6395-6dd9-4e6b-899b-49834cc2ed20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GraphMimic_[[:space:]]Graph-to-Graphs[[:space:]]Generative[[:space:]]Modeling[[:space:]]from[[:space:]]Videos[[:space:]]for[[:space:]]Policy[[:space:]]Learning/f50e1b72-b680-4690-826b-78b41eebe54e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gromov-Wasserstein[[:space:]]Problem[[:space:]]with[[:space:]]Cyclic[[:space:]]Symmetry/0a96047c-2069-4b88-a709-f9f3103c679a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GroomLight_[[:space:]]Hybrid[[:space:]]Inverse[[:space:]]Rendering[[:space:]]for[[:space:]]Relightable[[:space:]]Human[[:space:]]Hair[[:space:]]Appearance[[:space:]]Modeling/2dc488ec-f2d0-4f6c-92c4-cd397dac71f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ground-V_[[:space:]]Teaching[[:space:]]VLMs[[:space:]]to[[:space:]]Ground[[:space:]]Complex[[:space:]]Instructions[[:space:]]in[[:space:]]Pixels/4281951d-dd6b-48e4-b77f-533a1a5b91a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Grounding[[:space:]]3D[[:space:]]Object[[:space:]]Affordance[[:space:]]with[[:space:]]Language[[:space:]]Instructions,[[:space:]]Visual[[:space:]]Observations[[:space:]]and[[:space:]]Interactions/786d156f-e5d3-4bbd-8256-689fac95f5f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GroundingFace_[[:space:]]Fine-grained[[:space:]]Face[[:space:]]Understanding[[:space:]]via[[:space:]]Pixel[[:space:]]Grounding[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model/f5bd3f61-02af-42fb-bb37-51ddb15cb87b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GroupMamba_[[:space:]]Efficient[[:space:]]Group-Based[[:space:]]Visual[[:space:]]State[[:space:]]Space[[:space:]]Model/03367207-7a12-4748-bafa-c86ed5cef1d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GuardSplat_[[:space:]]Efficient[[:space:]]and[[:space:]]Robust[[:space:]]Watermarking[[:space:]]for[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/ba537cb2-c839-4de4-a8a5-413ad7f278cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Guiding[[:space:]]Human-Object[[:space:]]Interactions[[:space:]]with[[:space:]]Rich[[:space:]]Geometry[[:space:]]and[[:space:]]Relations/bfc07a77-2141-47f3-9ed6-ef9387efab79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gyro-based[[:space:]]Neural[[:space:]]Single[[:space:]]Image[[:space:]]Deblurring/714fe0b2-02ad-4937-888a-61b4baa95f16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/H-MoRe_[[:space:]]Learning[[:space:]]Human-centric[[:space:]]Motion[[:space:]]Representation[[:space:]]for[[:space:]]Action[[:space:]]Analysis/df417d26-7ab1-44c1-bff6-aa0e5ec4bf9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/H2ST_[[:space:]]Hierarchical[[:space:]]Two-Sample[[:space:]]Tests[[:space:]]for[[:space:]]Continual[[:space:]]Out-of-Distribution[[:space:]]Detection/a2486f9b-1014-4f9c-94d3-0e7afe0c17a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HD-EPIC_[[:space:]]A[[:space:]]Highly-Detailed[[:space:]]Egocentric[[:space:]]Video[[:space:]]Dataset/4d6a8539-9197-42b7-bc98-8498a8ad8008_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HEIE_[[:space:]]MLLM-Based[[:space:]]Hierarchical[[:space:]]Explainable[[:space:]]AIGC[[:space:]]Image[[:space:]]Implausibility[[:space:]]Evaluator/b6aa41d1-d2e0-4c29-bb32-b123ea7eb344_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HELVIPAD_[[:space:]]A[[:space:]]Real-World[[:space:]]Dataset[[:space:]]for[[:space:]]Omnidirectional[[:space:]]Stereo[[:space:]]Depth[[:space:]]Estimation/b116d203-048e-4e20-8807-791ae4570b9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HERA_[[:space:]]Hybrid[[:space:]]Explicit[[:space:]]Representation[[:space:]]for[[:space:]]Ultra-Realistic[[:space:]]Head[[:space:]]Avatars/03f3617f-8082-4b24-a4aa-32201f4cf567_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HIIF_[[:space:]]Hierarchical[[:space:]]Encoding[[:space:]]based[[:space:]]Implicit[[:space:]]Image[[:space:]]Function[[:space:]]for[[:space:]]Continuous[[:space:]]Super-resolution/c3be542c-58c6-4a4c-bca6-51957fe9c5e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HMAR_[[:space:]]Efficient[[:space:]]Hierarchical[[:space:]]Masked[[:space:]]Auto-Regressive[[:space:]]Image[[:space:]]Generation/f00d2ab8-c591-4ffc-892f-0010111e7a82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HOIGPT_[[:space:]]Learning[[:space:]]Long-Sequence[[:space:]]Hand-Object[[:space:]]Interaction[[:space:]]with[[:space:]]Language[[:space:]]Models/8211d8d9-8a79-4393-bbcf-5557f8692253_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HOIGen-1M_[[:space:]]A[[:space:]]Large-scale[[:space:]]Dataset[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Video[[:space:]]Generation/fbf17c45-20d8-459a-930e-b0ffab71f04e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HOP_[[:space:]]Heterogeneous[[:space:]]Topology-based[[:space:]]Multimodal[[:space:]]Entanglement[[:space:]]for[[:space:]]Co-Speech[[:space:]]Gesture[[:space:]]Generation/a2876681-dd28-4da7-8d1e-95e824b0946a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HORP_[[:space:]]Human-Object[[:space:]]Relation[[:space:]]Priors[[:space:]]Guided[[:space:]]HOI[[:space:]]Detection/268f7358-9a36-4db5-97fa-4c857a46a203_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HOT3D_[[:space:]]Hand[[:space:]]and[[:space:]]Object[[:space:]]Tracking[[:space:]]in[[:space:]]3D[[:space:]]from[[:space:]]Egocentric[[:space:]]Multi-View[[:space:]]Videos/50e52961-2d9c-46c8-9083-0b8eb58b11b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HOTFormerLoc_[[:space:]]Hierarchical[[:space:]]Octree[[:space:]]Transformer[[:space:]]for[[:space:]]Versatile[[:space:]]Lidar[[:space:]]Place[[:space:]]Recognition[[:space:]]Across[[:space:]]Ground[[:space:]]and[[:space:]]Aerial[[:space:]]Views/95eb10cf-f3ce-423c-80ec-7138bdb3a7f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HOT_[[:space:]]Hadamard-based[[:space:]]Optimized[[:space:]]Training/2e6717df-5519-4e65-a86c-ab99b9a0fcf3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HRAvatar_[[:space:]]High-Quality[[:space:]]and[[:space:]]Relightable[[:space:]]Gaussian[[:space:]]Head[[:space:]]Avatar/28bf049a-7ec8-4f9e-86c7-b846e29433e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HSI-GPT_[[:space:]]A[[:space:]]General-Purpose[[:space:]]Large[[:space:]]Scene-Motion-Language[[:space:]]Model[[:space:]]for[[:space:]]Human[[:space:]]Scene[[:space:]]Interaction/5709f2d7-442a-4e2d-83f3-03d6e4c1ac10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HSI_[[:space:]]A[[:space:]]Holistic[[:space:]]Style[[:space:]]Injector[[:space:]]for[[:space:]]Arbitrary[[:space:]]Style[[:space:]]Transfer/41428334-a734-45f0-bb7e-5c97b022fb0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HUNet_[[:space:]]Homotopy[[:space:]]Unfolding[[:space:]]Network[[:space:]]for[[:space:]]Image[[:space:]]Compressive[[:space:]]Sensing/d39648a6-6de5-430c-b58a-4a9fb2429894_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HUSH_[[:space:]]Holistic[[:space:]]Panoramic[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding[[:space:]]using[[:space:]]Spherical[[:space:]]Harmonics/97f23795-8409-47c7-84d5-0a54f2e0cf08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HVI_[[:space:]]A[[:space:]]New[[:space:]]Color[[:space:]]Space[[:space:]]for[[:space:]]Low-light[[:space:]]Image[[:space:]]Enhancement/fb26c59a-ceff-4c84-b45a-37305360f119_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HaWoR_[[:space:]]World-Space[[:space:]]Hand[[:space:]]Motion[[:space:]]Reconstruction[[:space:]]from[[:space:]]Egocentric[[:space:]]Videos/8248b1ef-a4cb-4a6b-b755-69d7d731b50c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HalLoc_[[:space:]]Token-level[[:space:]]Localization[[:space:]]of[[:space:]]Hallucinations[[:space:]]for[[:space:]]Vision[[:space:]]Language[[:space:]]Models/4e0d9ffb-10b1-4974-a0fc-49454fa3c060_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hallo3_[[:space:]]Highly[[:space:]]Dynamic[[:space:]]and[[:space:]]Realistic[[:space:]]Portrait[[:space:]]Image[[:space:]]Animation[[:space:]]with[[:space:]]Video[[:space:]]Diffusion[[:space:]]Transformer/e3060eba-cff1-4652-b4a0-a4c17ee0c6c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hand-held[[:space:]]Object[[:space:]]Reconstruction[[:space:]]from[[:space:]]RGB[[:space:]]Video[[:space:]]with[[:space:]]Dynamic[[:space:]]Interaction/dded505e-7a87-442b-8311-6518e4566d4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HandOS_[[:space:]]3D[[:space:]]Hand[[:space:]]Reconstruction[[:space:]]in[[:space:]]One[[:space:]]Stage/67f344fa-dbb4-47ec-a09b-c519711e461d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Handling[[:space:]]Spatial-Temporal[[:space:]]Data[[:space:]]Heterogeneity[[:space:]]for[[:space:]]Federated[[:space:]]Continual[[:space:]]Learning[[:space:]]via[[:space:]]Tail[[:space:]]Anchor/42dc91fb-c4d3-4e4f-b573-0a5cd271887c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hardware-Rasterized[[:space:]]Ray-Based[[:space:]]Gaussian[[:space:]]Splatting/10726f3c-bba0-405b-8a4c-3ea1bbfc7ad5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HarmonySet_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Dataset[[:space:]]for[[:space:]]Understanding[[:space:]]Video-Music[[:space:]]Semantic[[:space:]]Alignment[[:space:]]and[[:space:]]Temporal[[:space:]]Synchronization/3b4247a1-f62f-4fb7-bf00-0904b9ef6192_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Harnessing[[:space:]]Frequency[[:space:]]Spectrum[[:space:]]Insights[[:space:]]for[[:space:]]Image[[:space:]]Copyright[[:space:]]Protection[[:space:]]Against[[:space:]]Diffusion[[:space:]]Models/841da8ea-fe78-44db-9802-7c8dbc486e99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Harnessing[[:space:]]Frozen[[:space:]]Unimodal[[:space:]]Encoders[[:space:]]for[[:space:]]Flexible[[:space:]]Multimodal[[:space:]]Alignment/95034206-bd2e-4eb8-84ae-f288e0286d7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Harnessing[[:space:]]Global-Local[[:space:]]Collaborative[[:space:]]Adversarial[[:space:]]Perturbation[[:space:]]for[[:space:]]Anti-Customization/dfb6b0e8-139d-4544-a2e8-49ebf861e54a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hash3D_[[:space:]]Training-free[[:space:]]Acceleration[[:space:]]for[[:space:]]3D[[:space:]]Generation/a82a32f0-73d5-44ba-84a0-645a8b56b806_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hazy[[:space:]]Low-Quality[[:space:]]Satellite[[:space:]]Video[[:space:]]Restoration[[:space:]]Via[[:space:]]Learning[[:space:]]Optimal[[:space:]]Joint[[:space:]]Degradation[[:space:]]Patterns[[:space:]]and[[:space:]]Continuous-Scale[[:space:]]Super-Resolution[[:space:]]Reconstruction/d69f029e-977a-4bc6-bda7-4b5ff76e43ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HeMoRa_[[:space:]]Unsupervised[[:space:]]Heuristic[[:space:]]Consensus[[:space:]]Sampling[[:space:]]for[[:space:]]Robust[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/7809c293-9621-474d-8667-5a6763c16148_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hearing[[:space:]]Anywhere[[:space:]]in[[:space:]]Any[[:space:]]Environment/0cd2be96-a7ee-45d3-b540-7b22dab9096e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hearing[[:space:]]Hands_[[:space:]]Generating[[:space:]]Sounds[[:space:]]from[[:space:]]Physical[[:space:]]Interactions[[:space:]]in[[:space:]]3D[[:space:]]Scenes/42293e7f-093a-4192-9b70-2e12d7884996_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HeatFormer_[[:space:]]A[[:space:]]Neural[[:space:]]Optimizer[[:space:]]for[[:space:]]Multiview[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/913c55c1-160f-4c7c-913a-4a267f7a2df4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Heterogeneous[[:space:]]Skeleton-Based[[:space:]]Action[[:space:]]Representation[[:space:]]Learning/2ee9cdcd-1d71-446c-b6fe-9224ef26b1a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HiFi-Portrait_[[:space:]]Zero-shot[[:space:]]Identity-preserved[[:space:]]Portrait[[:space:]]Generation[[:space:]]with[[:space:]]High-fidelity[[:space:]]Multi-face[[:space:]]Fusion/37a31756-f1a0-4e30-8e50-d2d185f8d953_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HiLoTs_[[:space:]]High-Low[[:space:]]Temporal[[:space:]]Sensitive[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]LiDAR[[:space:]]Segmentation[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/680b028c-74d0-4896-b6a2-ce4af4995456_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HiMoR_[[:space:]]Monocular[[:space:]]Deformable[[:space:]]Gaussian[[:space:]]Reconstruction[[:space:]]with[[:space:]]Hierarchical[[:space:]]Motion[[:space:]]Representation/39cfe680-b5ae-4b81-bb33-6bd997dd7676_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HiPART_[[:space:]]Hierarchical[[:space:]]Pose[[:space:]]AutoRegressive[[:space:]]Transformer[[:space:]]for[[:space:]]Occluded[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/87b052ed-3952-4e71-870e-02400bbdf0d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HiRes-LLaVA_[[:space:]]Restoring[[:space:]]Fragmentation[[:space:]]Input[[:space:]]in[[:space:]]High-Resolution[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/36293265-0dfc-49c1-a4b8-6074f9251af2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hiding[[:space:]]Images[[:space:]]in[[:space:]]Diffusion[[:space:]]Models[[:space:]]by[[:space:]]Editing[[:space:]]Learned[[:space:]]Score[[:space:]]Functions/f806bc87-fb8e-46d2-beac-1b6d7ee1e54d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HierarQ_[[:space:]]Task-Aware[[:space:]]Hierarchical[[:space:]]Q-Former[[:space:]]for[[:space:]]Enhanced[[:space:]]Video[[:space:]]Understanding/e5b68db9-f1b3-464b-b4cc-109a4832e576_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]Adaptive[[:space:]]Filtering[[:space:]]Network[[:space:]]for[[:space:]]Text[[:space:]]Image[[:space:]]Specular[[:space:]]Highlight[[:space:]]Removal/75b8694b-da80-4cc0-8cfe-a49eece06817_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]Compact[[:space:]]Clustering[[:space:]]Attention[[:space:]](COCA)[[:space:]]for[[:space:]]Unsupervised[[:space:]]Object-Centric[[:space:]]Learning/6b31bbd3-f0ae-4c94-a4bd-b3b5ebfd28fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]Features[[:space:]]Matter_[[:space:]]A[[:space:]]Deep[[:space:]]Exploration[[:space:]]of[[:space:]]Progressive[[:space:]]Parameterization[[:space:]]Method[[:space:]]for[[:space:]]Dataset[[:space:]]Distillation/839ebd88-9bb8-4357-acb3-ca52fc8ad79b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]Flow[[:space:]]Diffusion[[:space:]]for[[:space:]]Efficient[[:space:]]Frame[[:space:]]Interpolation/c3e4cec7-6f5e-4aa1-819f-3d8bb1ff4021_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]Gaussian[[:space:]]Mixture[[:space:]]Model[[:space:]]Splatting[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Part[[:space:]]Controllable[[:space:]]3D[[:space:]]Generation/caeffd00-5e8a-47c7-b322-443159cf3d78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]Knowledge[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Multi-task[[:space:]]Test-Time[[:space:]]Adaptation/e1370638-674a-40b7-928c-4df9b57c7384_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/High[[:space:]]Dynamic[[:space:]]Range[[:space:]]Video[[:space:]]Compression_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]and[[:space:]]A[[:space:]]Learned[[:space:]]Bit-depth[[:space:]]Scalable[[:space:]]Compression[[:space:]]Algorithm/a2a2f3ed-511d-4c5a-8bbb-0782861dfb18_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/High[[:space:]]Temporal[[:space:]]Consistency[[:space:]]through[[:space:]]Semantic[[:space:]]Similarity[[:space:]]Propagation[[:space:]]in[[:space:]]Semi-Supervised[[:space:]]Video[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]for[[:space:]]Autonomous[[:space:]]Flight/5e169a18-1d74-4617-ae8b-01275601db89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/High-Fidelity[[:space:]]Lightweight[[:space:]]Mesh[[:space:]]Reconstruction[[:space:]]from[[:space:]]Point[[:space:]]Clouds/b6bccb41-1727-4cba-b8ce-510276cd267b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/High-Fidelity[[:space:]]Relightable[[:space:]]Monocular[[:space:]]Portrait[[:space:]]Animation[[:space:]]with[[:space:]]Lighting-Controllable[[:space:]]Video[[:space:]]Diffusion[[:space:]]Model/35f957ca-9aac-4688-a640-18d009677e4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/High-fidelity[[:space:]]3D[[:space:]]Object[[:space:]]Generation[[:space:]]from[[:space:]]Single[[:space:]]Image[[:space:]]with[[:space:]]RGBN-Volume[[:space:]]Gaussian[[:space:]]Reconstruction[[:space:]]Model/624a621f-35a4-4b95-a1e9-0315a5cd0051_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/High-quality[[:space:]]Point[[:space:]]Cloud[[:space:]]Oriented[[:space:]]Normal[[:space:]]Estimation[[:space:]]via[[:space:]]Hybrid[[:space:]]Angular[[:space:]]and[[:space:]]Euclidean[[:space:]]Distance[[:space:]]Encoding/815076bc-5862-412f-bdbf-4e3e3df4426f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Higher-Order[[:space:]]Ratio[[:space:]]Cycles[[:space:]]for[[:space:]]Fast[[:space:]]and[[:space:]]Globally[[:space:]]Optimal[[:space:]]Shape[[:space:]]Matching/7b6743f1-f248-4de2-8fe9-b4d5688fc8ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HistoFS_[[:space:]]Non-IID[[:space:]]Histopathologic[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Classification[[:space:]]via[[:space:]]Federated[[:space:]]Style[[:space:]]Transfer[[:space:]]with[[:space:]]RoI-Preserving/752a608c-dc4e-4d64-86f3-2a1f5341e163_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HoGS_[[:space:]]Unified[[:space:]]Near[[:space:]]and[[:space:]]Far[[:space:]]Object[[:space:]]Reconstruction[[:space:]]via[[:space:]]Homogeneous[[:space:]]Gaussian[[:space:]]Splatting/be0a5430-87f7-4bd3-9631-e2ed5908577c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Image[[:space:]]Quality[[:space:]]Assessment_[[:space:]]From[[:space:]]Human[[:space:]]to[[:space:]]Machine[[:space:]]Preference/9c43c7f6-7667-4c47-8440-32e15bc27376_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Image[[:space:]]Quality[[:space:]]Assessment_[[:space:]]Investigating[[:space:]]Causal[[:space:]]Perceptual[[:space:]]Effects[[:space:]]with[[:space:]]Abductive[[:space:]]Counterfactual[[:space:]]Inference/5e0e415c-9773-49a3-97ea-205f1760b866_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Image[[:space:]]Reconstruction[[:space:]]from[[:space:]]Readout-Multiplexed[[:space:]]Single-Photon[[:space:]]Detector[[:space:]]Arrays/960040c5-ed98-43eb-9b66-fa83e44db35d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Image[[:space:]]Referenced[[:space:]]Sketch[[:space:]]Colorization[[:space:]]Based[[:space:]]on[[:space:]]Animation[[:space:]]Creation[[:space:]]Workflow/9a0c23ba-5904-4499-9efc-651298f2730b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Image[[:space:]]is[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]to[[:space:]]Empower[[:space:]]Large-scale[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]In-Domain[[:space:]]Generation/e387f9f2-242c-46fd-9d3c-463d01e38cba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Imagine[[:space:]]and[[:space:]]Seek_[[:space:]]Improving[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval[[:space:]]with[[:space:]]an[[:space:]]Imagined[[:space:]]Proxy/f227006c-210b-4aac-822f-93d2a98acc36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ImagineFSL_[[:space:]]Self-Supervised[[:space:]]Pretraining[[:space:]]Matters[[:space:]]on[[:space:]]Imagined[[:space:]]Base[[:space:]]Set[[:space:]]for[[:space:]]VLM-based[[:space:]]Few-shot[[:space:]]Learning/b527a5f4-addd-424b-8f6f-3155526e27dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Img-Diff_[[:space:]]Contrastive[[:space:]]Data[[:space:]]Synthesis[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/e6b1bb30-e9d5-4708-960e-31ff7e5e9a5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Immune_[[:space:]]Improving[[:space:]]Safety[[:space:]]Against[[:space:]]Jailbreaks[[:space:]]in[[:space:]]Multi-modal[[:space:]]LLMs[[:space:]]via[[:space:]]Inference-Time[[:space:]]Alignment/66fe35a0-e089-4c9c-afba-719727b0b2df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Implicit[[:space:]]Bias[[:space:]]Injection[[:space:]]Attacks[[:space:]]against[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/5a713a88-9e78-41c3-b9a4-4778994a6f99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Implicit[[:space:]]Correspondence[[:space:]]Learning[[:space:]]for[[:space:]]Image-to-Point[[:space:]]Cloud[[:space:]]Registration/3b6935ad-40bc-4903-80f6-f62148a4f3d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improve[[:space:]]Representation[[:space:]]for[[:space:]]Imbalanced[[:space:]]Regression[[:space:]]through[[:space:]]Geometric[[:space:]]Constraints/e4bdbcad-b68b-4255-a63d-f2d51b9212bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improved[[:space:]]Monocular[[:space:]]Depth[[:space:]]Prediction[[:space:]]Using[[:space:]]Distance[[:space:]]Transform[[:space:]]Over[[:space:]]Pre-semantic[[:space:]]Contours[[:space:]]with[[:space:]]Self-supervised[[:space:]]Neural[[:space:]]Networks/18d0caf1-d799-41ef-95e4-0a0e2105060c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improved[[:space:]]Video[[:space:]]VAE[[:space:]]for[[:space:]]Latent[[:space:]]Video[[:space:]]Diffusion[[:space:]]Model/d6e6ed64-039e-4e0b-a0ce-69a2cf999362_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Accuracy[[:space:]]and[[:space:]]Calibration[[:space:]]via[[:space:]]Differentiated[[:space:]]Deep[[:space:]]Mutual[[:space:]]Learning/6a2c48a3-aafc-4ad2-9262-8700829f154e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Adversarial[[:space:]]Transferability[[:space:]]on[[:space:]]Vision[[:space:]]Transformers[[:space:]]via[[:space:]]Forward[[:space:]]Propagation[[:space:]]Refinement/c0d9e11b-700c-40d7-8dca-46321613cb22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Autoregressive[[:space:]]Visual[[:space:]]Generation[[:space:]]with[[:space:]]Cluster-Oriented[[:space:]]Token[[:space:]]Prediction/2639ed20-911e-44d7-9142-a8fcfe0f6597_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Diffusion[[:space:]]Inverse[[:space:]]Problem[[:space:]]Solving[[:space:]]with[[:space:]]Decoupled[[:space:]]Noise[[:space:]]Annealing/d107e339-bf08-4731-b903-e7f3360d585d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Editability[[:space:]]in[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Layer-wise[[:space:]]Memory/5ef2c05f-1b0c-4899-806c-c1cd90e73e44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]Localized[[:space:]]Points[[:space:]]Management/6436626a-e39e-4d87-9a2d-0ebf83ce94c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Personalized[[:space:]]Search[[:space:]]with[[:space:]]Regularized[[:space:]]Low-Rank[[:space:]]Parameter[[:space:]]Updates/673cbe39-bd0b-47dd-99b2-3c870c8defd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Sliced-Wasserstein[[:space:]]Feature[[:space:]]Alignment[[:space:]]and[[:space:]]Uniformity/ca5cab4a-bdaf-48a0-9763-5efeeaa13237_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Sound[[:space:]]Source[[:space:]]Localization[[:space:]]with[[:space:]]Joint[[:space:]]Slot[[:space:]]Attention[[:space:]]on[[:space:]]Image[[:space:]]and[[:space:]]Audio/feaa7f56-7770-404f-a305-b468e86df1e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Transferable[[:space:]]Targeted[[:space:]]Attacks[[:space:]]with[[:space:]]Feature[[:space:]]Tuning[[:space:]]Mixup/5356bb72-29a6-442b-984c-adce3c009504_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Visual[[:space:]]and[[:space:]]Downstream[[:space:]]Performance[[:space:]]of[[:space:]]Low-Light[[:space:]]Enhancer[[:space:]]with[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models[[:space:]]Collaboration/0e31d729-0f45-4249-a70f-af6d4389a18f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]the[[:space:]]Training[[:space:]]of[[:space:]]Data-Efficient[[:space:]]GANs[[:space:]]via[[:space:]]Quality[[:space:]]Aware[[:space:]]Dynamic[[:space:]]Discriminator[[:space:]]Rejection[[:space:]]Sampling/9b5acaea-53bf-4055-87a9-1a5bf68ccd00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]the[[:space:]]Transferability[[:space:]]of[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]on[[:space:]]Face[[:space:]]Recognition[[:space:]]with[[:space:]]Diverse[[:space:]]Parameters[[:space:]]Augmentation/aa0474ee-612f-4595-88a1-c6ba9ec15b27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Imputation-free[[:space:]]and[[:space:]]Alignment-free_[[:space:]]Incomplete[[:space:]]Multi-view[[:space:]]Clustering[[:space:]]Driven[[:space:]]by[[:space:]]Consensus[[:space:]]Semantic[[:space:]]Learning/380dd383-b00d-4370-9caa-3f9395391322_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InPO_[[:space:]]Inversion[[:space:]]Preference[[:space:]]Optimization[[:space:]]with[[:space:]]Reparametrized[[:space:]]DDIM[[:space:]]for[[:space:]]Efficient[[:space:]]Diffusion[[:space:]]Model[[:space:]]Alignment/14854e80-d5af-4094-8f56-d57259d0057b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IncEventGS_[[:space:]]Pose-Free[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Event[[:space:]]Camera/c5e31ebc-2f5f-4333-a4a5-98739dcdbce4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Incomplete[[:space:]]Multi-View[[:space:]]Multi-label[[:space:]]Learning[[:space:]]via[[:space:]]Disentangled[[:space:]]Representation[[:space:]]and[[:space:]]Label[[:space:]]Semantic[[:space:]]Embedding/419f8c28-377b-4f29-8c8b-a2cba093a85b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Incomplete[[:space:]]Multi-modal[[:space:]]Brain[[:space:]]Tumor[[:space:]]Segmentation[[:space:]]via[[:space:]]Learnable[[:space:]]Sorting[[:space:]]State[[:space:]]Space[[:space:]]Model/58ab88d4-b760-40c9-8213-6c50aa676cf3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Incorporating[[:space:]]Dense[[:space:]]Knowledge[[:space:]]Alignment[[:space:]]into[[:space:]]Unified[[:space:]]Multimodal[[:space:]]Representation[[:space:]]Models/f0305c72-f982-4ece-ab22-ea79e8e3955f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Incremental[[:space:]]Object[[:space:]]Keypoint[[:space:]]Learning/d3b12539-b476-45cb-a752-bae4990c3622_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IndoorGS_[[:space:]]Geometric[[:space:]]Cues[[:space:]]Guided[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Indoor[[:space:]]Scene[[:space:]]Reconstruction/848b3be8-e7e7-4fb9-82b5-d4028969c01d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Inference-Scale[[:space:]]Complexity[[:space:]]in[[:space:]]ANN-SNN[[:space:]]Conversion[[:space:]]for[[:space:]]High-Performance[[:space:]]and[[:space:]]Low-Power[[:space:]]Applications/6b151e98-3001-4e0f-b77a-ea1941036ad2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Infighting[[:space:]]in[[:space:]]the[[:space:]]Dark_[[:space:]]Multi-Label[[:space:]]Backdoor[[:space:]]Attack[[:space:]]in[[:space:]]Federated[[:space:]]Learning/bab57aa6-5f36-4253-ae3a-1e5e1f4632d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Infinity_[[:space:]]Scaling[[:space:]]Bitwise[[:space:]]AutoRegressive[[:space:]]Modeling[[:space:]]for[[:space:]]High-Resolution[[:space:]]Image[[:space:]]Synthesis/d1db9220-1c65-401b-9202-6e181bfed6b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InsTaG_[[:space:]]Learning[[:space:]]Personalized[[:space:]]3D[[:space:]]Talking[[:space:]]Head[[:space:]]from[[:space:]]Few-Second[[:space:]]Video/1014c4e5-3a69-4fa3-98b3-a879e2f45651_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Insight-V_[[:space:]]Exploring[[:space:]]Long-Chain[[:space:]]Visual[[:space:]]Reasoning[[:space:]]with[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/cb8122c4-a3cd-447b-9677-ef086abe8658_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InsightEdit_[[:space:]]Towards[[:space:]]Better[[:space:]]Instruction[[:space:]]Following[[:space:]]for[[:space:]]Image[[:space:]]Editing/fb96c75a-1b6e-4864-bf83-3b9ecc089e1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Insightful[[:space:]]Instance[[:space:]]Features[[:space:]]for[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation/275c7e3b-7085-4434-b01d-e129139897fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Inst3D-LMM_[[:space:]]Instance-Aware[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding[[:space:]]with[[:space:]]Multi-modal[[:space:]]Instruction[[:space:]]Tuning/ada2a42d-606c-400d-85d5-a68b46a5f62b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Instance-wise[[:space:]]Supervision-level[[:space:]]Optimization[[:space:]]in[[:space:]]Active[[:space:]]Learning/6ab274ce-c563-45f8-9f14-82dfa36b95f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InstanceCap_[[:space:]]Improving[[:space:]]Text-to-Video[[:space:]]Generation[[:space:]]via[[:space:]]Instance-aware[[:space:]]Structured[[:space:]]Caption/63b6efae-0d11-487a-b293-b1b579ef5966_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InstanceGaussian_[[:space:]]Appearance-Semantic[[:space:]]Joint[[:space:]]Gaussian[[:space:]]Representation[[:space:]]for[[:space:]]3D[[:space:]]Instance-Level[[:space:]]Perception/e6f2e48d-8efc-4d47-81e4-de4b789d0ecb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Instant[[:space:]]Adversarial[[:space:]]Purification[[:space:]]with[[:space:]]Adversarial[[:space:]]Consistency[[:space:]]Distillation/661ff35f-282c-423b-9185-bb6d8ff5da4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Instant[[:space:]]Gaussian[[:space:]]Stream_[[:space:]]Fast[[:space:]]and[[:space:]]Generalizable[[:space:]]Streaming[[:space:]]of[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]via[[:space:]]Gaussian[[:space:]]Splatting/c9f2cc0e-b9b5-4129-a1a3-c1fc202a4666_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Instant3dit_[[:space:]]Multiview[[:space:]]Inpainting[[:space:]]for[[:space:]]Fast[[:space:]]Editing[[:space:]]of[[:space:]]3D[[:space:]]Objects/671f9dcf-f80a-4cc0-9e9d-72da3c00b249_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Instruct-CLIP_[[:space:]]Improving[[:space:]]Instruction-Guided[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Automated[[:space:]]Data[[:space:]]Refinement[[:space:]]Using[[:space:]]Contrastive[[:space:]]Learning/a04fb8db-d9e4-4e5e-a8c0-00327c1bd422_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Instruction-based[[:space:]]Image[[:space:]]Manipulation[[:space:]]by[[:space:]]Watching[[:space:]]How[[:space:]]Things[[:space:]]Move/6db53ffc-36dc-4f4f-8a58-b9c5bb9ce24b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Integral[[:space:]]Fast[[:space:]]Fourier[[:space:]]Color[[:space:]]Constancy/9fd420e6-b4fa-48bd-940e-f8dcc99e9a7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InterAct_[[:space:]]Advancing[[:space:]]Large-Scale[[:space:]]Versatile[[:space:]]3D[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Generation/4511c287-17f9-4b70-a6d5-d77c48a5192a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InterDyn_[[:space:]]Controllable[[:space:]]Interactive[[:space:]]Dynamics[[:space:]]with[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/1657842e-f939-4237-87d3-49f995d53e68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InterMimic_[[:space:]]Towards[[:space:]]Universal[[:space:]]Whole-Body[[:space:]]Control[[:space:]]for[[:space:]]Physics-Based[[:space:]]Human-Object[[:space:]]Interactions/c17df58e-4f4a-4403-b950-698942e97f1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InteractAnything_[[:space:]]Zero-shot[[:space:]]Human[[:space:]]Object[[:space:]]Interaction[[:space:]]Synthesis[[:space:]]via[[:space:]]LLM[[:space:]]Feedback[[:space:]]and[[:space:]]Object[[:space:]]Affordance[[:space:]]Parsing/bba3c5bb-a0b8-4050-9fca-51eeaf9696e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InteractVLM_[[:space:]]3D[[:space:]]Interaction[[:space:]]Reasoning[[:space:]]from[[:space:]]2D[[:space:]]Foundational[[:space:]]Models/9f03277a-37b8-49e7-9fad-13e2e5f171b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InteractionMap_[[:space:]]Improving[[:space:]]Online[[:space:]]Vectorized[[:space:]]HDMap[[:space:]]Construction[[:space:]]with[[:space:]]Interaction/447efac7-303b-4880-a09c-b0e941c7cbb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Interactive[[:space:]]Medical[[:space:]]Image[[:space:]]Analysis[[:space:]]with[[:space:]]Concept-based[[:space:]]Similarity[[:space:]]Reasoning/a8ba1cd8-3d09-4bb8-8912-43eb50d5dbf1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Interactive[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation_[[:space:]]A[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]and[[:space:]]Baseline/72ec2f30-d729-4ccc-bc2d-90a1d1bd3d41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Interleaved-Modal[[:space:]]Chain-of-Thought/3edf2a1c-2168-41b7-a536-b8e17bc35145_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Interpretable[[:space:]]Generative[[:space:]]Models[[:space:]]through[[:space:]]Post-hoc[[:space:]]Concept[[:space:]]Bottlenecks/86d682d5-19b8-4583-bdec-4fcac9608ef2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Interpretable[[:space:]]Image[[:space:]]Classification[[:space:]]via[[:space:]]Non-parametric[[:space:]]Part[[:space:]]Prototype[[:space:]]Learning/d624a8da-e662-4173-965d-8a4cb70d4b42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Interpreting[[:space:]]Object-level[[:space:]]Foundation[[:space:]]Models[[:space:]]via[[:space:]]Visual[[:space:]]Precision[[:space:]]Search/396da971-babe-4c65-8e62-f7e2e9a705a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Inversion[[:space:]]Circle[[:space:]]Interpolation_[[:space:]]Diffusion-based[[:space:]]Image[[:space:]]Augmentation[[:space:]]for[[:space:]]Data-scarce[[:space:]]Classification/8d0fa3cc-8894-4a2c-8c38-0871b4963e22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Investigating[[:space:]]the[[:space:]]Role[[:space:]]of[[:space:]]Weight[[:space:]]Decay[[:space:]]in[[:space:]]Enhancing[[:space:]]Nonconvex[[:space:]]SGD/6f3ba934-d928-4457-b997-636845fb91ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Invisible[[:space:]]Backdoor[[:space:]]Attack[[:space:]]against[[:space:]]Self-supervised[[:space:]]Learning/9e544bcf-a6c1-4e2f-8db4-edb23e04ad07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Is[[:space:]]Your[[:space:]]World[[:space:]]Simulator[[:space:]]a[[:space:]]Good[[:space:]]Story[[:space:]]Presenter_[[:space:]]A[[:space:]]Consecutive[[:space:]]Events-Based[[:space:]]Benchmark[[:space:]]for[[:space:]]Future[[:space:]]Long[[:space:]]Video[[:space:]]Generation/82ae2dd6-8bd2-4433-85da-e99fcdc0f32c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Is[[:space:]]`Right'[[:space:]]Right_[[:space:]]Enhancing[[:space:]]Object[[:space:]]Orientation[[:space:]]Understanding[[:space:]]in[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]through[[:space:]]Egocentric[[:space:]]Instruction[[:space:]]Tuning/931219ab-c5f9-40e3-be39-009575d02724_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Is[[:space:]]this[[:space:]]Generated[[:space:]]Person[[:space:]]Existed[[:space:]]in[[:space:]]Real-world_[[:space:]]Fine-grained[[:space:]]Detecting[[:space:]]and[[:space:]]Calibrating[[:space:]]Abnormal[[:space:]]Human-body/3e7c9e40-6b2d-4e33-8d94-be53ee6c955a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/It's[[:space:]]a[[:space:]](Blind)[[:space:]]Match![[:space:]]Towards[[:space:]]Vision-Language[[:space:]]Correspondence[[:space:]]without[[:space:]]Parallel[[:space:]]Data/792e72be-058c-4275-af91-0375a10913d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IterIS_[[:space:]]Iterative[[:space:]]Inference-Solving[[:space:]]Alignment[[:space:]]for[[:space:]]LoRA[[:space:]]Merging/aa4d371a-0aca-4267-ba3e-284cb27d1446_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Iterative[[:space:]]Predictor-Critic[[:space:]]Code[[:space:]]Decoding[[:space:]]for[[:space:]]Real-World[[:space:]]Image[[:space:]]Dehazing/aa39bec7-c712-43a6-b67b-9677b4a1ca91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/JTD-UAV_[[:space:]]MLLM-Enhanced[[:space:]]Joint[[:space:]]Tracking[[:space:]]and[[:space:]]Description[[:space:]]Framework[[:space:]]for[[:space:]]Anti-UAV[[:space:]]Systems/ae7b4ca8-6617-48dc-a873-4c96791b5893_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Jailbreaking[[:space:]]the[[:space:]]Non-Transferable[[:space:]]Barrier[[:space:]]via[[:space:]]Test-Time[[:space:]]Data[[:space:]]Disguising/8f2429ba-5896-4fed-8787-5387f7a7c2b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/JamMa_[[:space:]]Ultra-lightweight[[:space:]]Local[[:space:]]Feature[[:space:]]Matching[[:space:]]with[[:space:]]Joint[[:space:]]Mamba/2224a678-5059-4a23-984a-11d935314c94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/JanusFlow_[[:space:]]Harmonizing[[:space:]]Autoregression[[:space:]]and[[:space:]]Rectified[[:space:]]Flow[[:space:]]for[[:space:]]Unified[[:space:]]Multimodal[[:space:]]Understanding[[:space:]]and[[:space:]]Generation/eb20bb5f-dac6-4842-a2a9-efce319159b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Janus_[[:space:]]Decoupling[[:space:]]Visual[[:space:]]Encoding[[:space:]]for[[:space:]]Unified[[:space:]]Multimodal[[:space:]]Understanding[[:space:]]and[[:space:]]Generation/7cd1746f-524c-4616-81f4-45bf9fae27c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/JarvisIR_[[:space:]]Elevating[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Perception[[:space:]]with[[:space:]]Intelligent[[:space:]]Image[[:space:]]Restoration/564f75fb-9069-40bc-9b81-1abe0d6bf292_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/JiSAM_[[:space:]]Alleviate[[:space:]]Labeling[[:space:]]Burden[[:space:]]and[[:space:]]Corner[[:space:]]Case[[:space:]]Problems[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving[[:space:]]via[[:space:]]Minimal[[:space:]]Real-World[[:space:]]Data/823b10e8-3181-4ff7-b9a6-5804631e0889_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Joint[[:space:]]Optimization[[:space:]]of[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]and[[:space:]]Continuous[[:space:]]Camera[[:space:]]Motion[[:space:]]from[[:space:]]a[[:space:]]Monocular[[:space:]]Video/a8f83c1c-f509-45cf-b2ea-aadc3620aeb8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Joint[[:space:]]Out-of-Distribution[[:space:]]Filtering[[:space:]]and[[:space:]]Data[[:space:]]Discovery[[:space:]]Active[[:space:]]Learning/aac94046-b8e6-4f93-9064-909ab1881fba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Joint[[:space:]]Scheduling[[:space:]]of[[:space:]]Causal[[:space:]]Prompts[[:space:]]and[[:space:]]Tasks[[:space:]]for[[:space:]]Multi-Task[[:space:]]Learning/8a2d4082-c63a-40b3-8fbe-ff724b130f71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Joint[[:space:]]Vision-Language[[:space:]]Social[[:space:]]Bias[[:space:]]Removal[[:space:]]for[[:space:]]CLIP/c95ec5bf-da93-4a01-a9f6-607caf7c59f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Just[[:space:]]Dance[[:space:]]with[[:space:]]pi![[:space:]]A[[:space:]]Poly-modal[[:space:]]Inductor[[:space:]]for[[:space:]]Weakly-supervised[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/db166284-5d6c-4fa4-8503-d5c1e8a51cc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/K-LoRA_[[:space:]]Unlocking[[:space:]]Training-Free[[:space:]]Fusion[[:space:]]of[[:space:]]Any[[:space:]]Subject[[:space:]]and[[:space:]]Style[[:space:]]LoRAs/59aa36ff-613a-44bd-b40f-d27a0d59242b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/K-Sort[[:space:]]Arena_[[:space:]]Efficient[[:space:]]and[[:space:]]Reliable[[:space:]]Benchmarking[[:space:]]for[[:space:]]Generative[[:space:]]Models[[:space:]]via[[:space:]]K-wise[[:space:]]Human[[:space:]]Preferences/505e45e9-24c1-4647-bd08-aac830907d33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/KAC_[[:space:]]Kolmogorov-Arnold[[:space:]]Classifier[[:space:]]for[[:space:]]Continual[[:space:]]Learning/7096d1ce-7f5b-4fc1-b83c-db5cfffee53f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/KMD_[[:space:]]Koopman[[:space:]]Multi-modality[[:space:]]Decomposition[[:space:]]for[[:space:]]Generalized[[:space:]]Brain[[:space:]]Tumor[[:space:]]Segmentation[[:space:]]under[[:space:]]Incomplete[[:space:]]Modalities/f72d8ab5-06c2-478a-8b72-f3c4adbba67d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/KVQ_[[:space:]]Boosting[[:space:]]Video[[:space:]]Quality[[:space:]]Assessment[[:space:]]via[[:space:]]Saliency-guided[[:space:]]Local[[:space:]]Perception/0194f279-bd4f-4761-b072-83ef5df76103_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Keep[[:space:]]the[[:space:]]Balance_[[:space:]]A[[:space:]]Parameter-Efficient[[:space:]]Symmetrical[[:space:]]Framework[[:space:]]for[[:space:]]RGB+X[[:space:]]Semantic[[:space:]]Segmentation/41c47a3f-3456-4660-aff0-54055430d717_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/KeyFace_[[:space:]]Expressive[[:space:]]Audio-Driven[[:space:]]Facial[[:space:]]Animation[[:space:]]for[[:space:]]Long[[:space:]]Sequences[[:space:]]via[[:space:]]KeyFrame[[:space:]]Interpolation/f8d52aaa-94e3-4294-99a5-e958dc674cb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Keyframe-Guided[[:space:]]Creative[[:space:]]Video[[:space:]]Inpainting/9a903872-5fd2-4add-9adc-038439170239_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Kiss3DGen_[[:space:]]Repurposing[[:space:]]Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]3D[[:space:]]Asset[[:space:]]Generation/9c832955-6d9c-4608-8893-0637d353c75f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Knowledge[[:space:]]Bridger_[[:space:]]Towards[[:space:]]Training-Free[[:space:]]Missing[[:space:]]Modality[[:space:]]Completion/bacfb3f4-1431-47f5-b918-bcc32369e76c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Knowledge[[:space:]]Memorization[[:space:]]and[[:space:]]Rumination[[:space:]]for[[:space:]]Pre-trained[[:space:]]Model-based[[:space:]]Class-Incremental[[:space:]]Learning/65dd4671-8f56-4447-95e9-3ac9ce0fe75f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Knowledge-Aligned[[:space:]]Counterfactual-Enhancement[[:space:]]Diffusion[[:space:]]Perception[[:space:]]for[[:space:]]Unsupervised[[:space:]]Cross-Domain[[:space:]]Visual[[:space:]]Emotion[[:space:]]Recognition/8b0ea7da-c4ac-4a10-be97-40e9deca50c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Koala-36M_[[:space:]]A[[:space:]]Large-scale[[:space:]]Video[[:space:]]Dataset[[:space:]]Improving[[:space:]]Consistency[[:space:]]between[[:space:]]Fine-grained[[:space:]]Conditions[[:space:]]and[[:space:]]Video[[:space:]]Content/1b749a0f-fa1b-4379-8b7e-8ee83e2b459b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/L-SWAG_[[:space:]]Layer-Sample[[:space:]]Wise[[:space:]]Activation[[:space:]]with[[:space:]]Gradients[[:space:]]Information[[:space:]]for[[:space:]]Zero-Shot[[:space:]]NAS[[:space:]]on[[:space:]]Vision[[:space:]]Transformers/b8aa8c44-7a5c-4de0-84ba-9d21a2c3b306_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LAL_[[:space:]]Enhancing[[:space:]]3D[[:space:]]Human[[:space:]]Motion[[:space:]]Prediction[[:space:]]with[[:space:]]Latency-aware[[:space:]]Auxiliary[[:space:]]Learning/3ae387f7-0485-4c58-bca6-f5c011874a7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LATTE-MV_[[:space:]]Learning[[:space:]]to[[:space:]]Anticipate[[:space:]]Table[[:space:]]Tennis[[:space:]]Hits[[:space:]]from[[:space:]]Monocular[[:space:]]Videos/c0a17a51-392d-458a-9c93-6081f559a5bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LC-Mamba_[[:space:]]Local[[:space:]]and[[:space:]]Continuous[[:space:]]Mamba[[:space:]]with[[:space:]]Shifted[[:space:]]Windows[[:space:]]for[[:space:]]Frame[[:space:]]Interpolation/7b5eda24-8f58-46d4-8aec-4dbd24c95a31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LEDiff_[[:space:]]Latent[[:space:]]Exposure[[:space:]]Diffusion[[:space:]]for[[:space:]]HDR[[:space:]]Generation/c5833ecd-e35d-46e2-be94-8fae8009ac53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LIM_[[:space:]]Large[[:space:]]Interpolator[[:space:]]Model[[:space:]]for[[:space:]]Dynamic[[:space:]]Reconstruction/785d307c-d149-4146-b0a9-19ef0d2abfa2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LION-FS_[[:space:]]Fast[[:space:]]&[[:space:]]Slow[[:space:]]Video-Language[[:space:]]Thinker[[:space:]]as[[:space:]]Online[[:space:]]Video[[:space:]]Assistant/4a478460-614d-459c-8250-8e050024e906_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LIRM_[[:space:]]Large[[:space:]]Inverse[[:space:]]Rendering[[:space:]]Model[[:space:]]for[[:space:]]Progressive[[:space:]]Reconstruction[[:space:]]of[[:space:]]Shape,[[:space:]]Materials[[:space:]]and[[:space:]]View-dependent[[:space:]]Radiance[[:space:]]Fields/3af0fb20-148f-4134-aa5a-1962e388882b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LITA-GS_[[:space:]]Illumination-Agnostic[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]via[[:space:]]Reference-Free[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]and[[:space:]]Physical[[:space:]]Priors/f1bcca35-8b9e-4aa1-881c-77f0235a614d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLAVIDAL_[[:space:]]A[[:space:]]Large[[:space:]]LAnguage[[:space:]]VIsion[[:space:]]Model[[:space:]]for[[:space:]]Daily[[:space:]]Activities[[:space:]]of[[:space:]]Living/7ce4a10c-274a-47eb-8a33-316845726b5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLM-driven[[:space:]]Multimodal[[:space:]]and[[:space:]]Multi-Identity[[:space:]]Listening[[:space:]]Head[[:space:]]Generation/1aa1ddf7-0138-4c52-9baa-83bff9d63ed4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLMDet_[[:space:]]Learning[[:space:]]Strong[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detectors[[:space:]]under[[:space:]]the[[:space:]]Supervision[[:space:]]of[[:space:]]Large[[:space:]]Language[[:space:]]Models/4d23c274-7e3d-43ad-a39c-81607c9de2b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLaVA-Critic_[[:space:]]Learning[[:space:]]to[[:space:]]Evaluate[[:space:]]Multimodal[[:space:]]Models/4b85d8d1-b36c-4f27-885b-51d0286bec1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLaVA-ST_[[:space:]]A[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Spatial-Temporal[[:space:]]Understanding/ab7735b7-654d-48a4-9305-824782d014af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LMO_[[:space:]]Linear[[:space:]]Mamba[[:space:]]Operator[[:space:]]for[[:space:]]MRI[[:space:]]Reconstruction/9fbc671c-cff1-4b05-a1ed-9c4dc3eda656_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LOCORE_[[:space:]]Image[[:space:]]Re-ranking[[:space:]]with[[:space:]]Long-Context[[:space:]]Sequence[[:space:]]Modeling/c5c5ef45-c279-4f38-b1e1-ab6163e4f07e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LOD-GS_[[:space:]]Achieving[[:space:]]Levels[[:space:]]of[[:space:]]Detail[[:space:]]using[[:space:]]Scalable[[:space:]]Gaussian[[:space:]]Soup/f511c703-910b-460d-88d5-62176f899e85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LOGICZSL_[[:space:]]Exploring[[:space:]]Logic-induced[[:space:]]Representation[[:space:]]for[[:space:]]Compositional[[:space:]]Zero-shot[[:space:]]Learning/16ff8f23-6bf2-40a9-b818-df1fa1d76e90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LP-Diff_[[:space:]]Towards[[:space:]]Improved[[:space:]]Restoration[[:space:]]of[[:space:]]Real-World[[:space:]]Degraded[[:space:]]License[[:space:]]Plate/d6986a79-814c-4611-82d7-fff08e5b4e07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LPOSS_[[:space:]]Label[[:space:]]Propagation[[:space:]]Over[[:space:]]Patches[[:space:]]and[[:space:]]Pixels[[:space:]]for[[:space:]]Open-vocabulary[[:space:]]Semantic[[:space:]]Segmentation/db4382ef-18be-4775-86d3-7af80bae74f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LSNet_[[:space:]]See[[:space:]]Large,[[:space:]]Focus[[:space:]]Small/55a67021-327a-4efa-95c0-898f1654d3ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LSceneLLM_[[:space:]]Enhancing[[:space:]]Large[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding[[:space:]]Using[[:space:]]Adaptive[[:space:]]Visual[[:space:]]Preferences/dfc2b03e-147a-4edc-ae47-1add32fcb998_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LT3SD_[[:space:]]Latent[[:space:]]Trees[[:space:]]for[[:space:]]3D[[:space:]]Scene[[:space:]]Diffusion/bbb3b2cc-5666-49fe-8b34-12dddae15e35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LUCAS_[[:space:]]Layered[[:space:]]Universal[[:space:]]Codec[[:space:]]Avatars/210caf04-d9d0-4423-b213-74f79204e3f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LaTexBlend_[[:space:]]Scaling[[:space:]]Multi-concept[[:space:]]Customized[[:space:]]Generation[[:space:]]with[[:space:]]Latent[[:space:]]Textual[[:space:]]Blending/f788d6dd-86b8-4b3e-b7ad-d0f8877301bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LaVin-DiT_[[:space:]]Large[[:space:]]Vision[[:space:]]Diffusion[[:space:]]Transformer/2b4c707a-0607-4a9c-b99f-9f4529ff86bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Label[[:space:]]Shift[[:space:]]Meets[[:space:]]Online[[:space:]]Learning_[[:space:]]Ensuring[[:space:]]Consistent[[:space:]]Adaptation[[:space:]]with[[:space:]]Universal[[:space:]]Dynamic[[:space:]]Regret/b8dd89de-37de-404a-94b6-146245fbbb33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LamRA_[[:space:]]Large[[:space:]]Multimodal[[:space:]]Model[[:space:]]as[[:space:]]Your[[:space:]]Advanced[[:space:]]Retrieval[[:space:]]Assistant/3bd12f60-cfee-4e07-b6b1-a9ab31fbecbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Language[[:space:]]Guided[[:space:]]Concept[[:space:]]Bottleneck[[:space:]]Models[[:space:]]for[[:space:]]Interpretable[[:space:]]Continual[[:space:]]Learning/31ceef46-4eee-4c37-a8c8-e65787d2b1db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Language-Assisted[[:space:]]Debiasing[[:space:]]and[[:space:]]Smoothing[[:space:]]for[[:space:]]Foundation[[:space:]]Model-Based[[:space:]]Semi-Supervised[[:space:]]Learning/dbac4839-2486-47a6-94a8-a111ef6de12e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Language-Guided[[:space:]]Audio-Visual[[:space:]]Learning[[:space:]]for[[:space:]]Long-Term[[:space:]]Sports[[:space:]]Assessment/973d7f68-7215-4a2b-bcee-e31972121a37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Language-Guided[[:space:]]Image[[:space:]]Tokenization[[:space:]]for[[:space:]]Generation/4b55ef5e-8da9-4ee1-9dee-eda77068a153_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Language-Guided[[:space:]]Salient[[:space:]]Object[[:space:]]Ranking/d66035f3-264c-41f0-b933-6ee1c9292951_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Large[[:space:]]Self-Supervised[[:space:]]Models[[:space:]]Bridge[[:space:]]the[[:space:]]Gap[[:space:]]in[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detection/9f7f4ed3-aae7-4649-84f7-aae41f69ca4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Large-Scale[[:space:]]Text-to-Image[[:space:]]Model[[:space:]]with[[:space:]]Inpainting[[:space:]]is[[:space:]]a[[:space:]]Zero-Shot[[:space:]]Subject-Driven[[:space:]]Image[[:space:]]Generator/514ee3a2-bf1d-4d64-a892-2e9ace397642_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Large-scale[[:space:]]Multi-view[[:space:]]Tensor[[:space:]]Clustering[[:space:]]with[[:space:]]Implicit[[:space:]]Linear[[:space:]]Kernels/f4b7746b-0722-4273-aff7-4e43959f3558_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Latent[[:space:]]Drifting[[:space:]]in[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Counterfactual[[:space:]]Medical[[:space:]]Image[[:space:]]Synthesis/41a90e55-2b29-4bee-9d11-c11fa8e8901a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Latent[[:space:]]Space[[:space:]]Imaging/fa005c22-0e46-49ac-8b27-40b1938a68e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Latent[[:space:]]Space[[:space:]]Super-Resolution[[:space:]]for[[:space:]]Higher-Resolution[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/db5c15b3-784d-4b78-a0bc-10f31574b9c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LatentHOI_[[:space:]]On[[:space:]]the[[:space:]]Generalizable[[:space:]]Hand[[:space:]]Object[[:space:]]Motion[[:space:]]Generation[[:space:]]with[[:space:]]Latent[[:space:]]Hand[[:space:]]Diffusion./9544e81f-acd8-44cb-a46f-7a0efb99083d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Layer-[[:space:]]and[[:space:]]Timestep-Adaptive[[:space:]]Differentiable[[:space:]]Token[[:space:]]Compression[[:space:]]Ratios[[:space:]]for[[:space:]]Efficient[[:space:]]Diffusion[[:space:]]Transformers/db600d17-6e1d-451a-943a-f87901eef700_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Layered[[:space:]]Image[[:space:]]Vectorization[[:space:]]via[[:space:]]Semantic[[:space:]]Simplification/d78fb716-8b96-467d-a3cb-c2cc0a2aa0c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Layered[[:space:]]Motion[[:space:]]Fusion_[[:space:]]Lifting[[:space:]]Motion[[:space:]]Segmentation[[:space:]]to[[:space:]]3D[[:space:]]in[[:space:]]Egocentric[[:space:]]Videos/74447121-1d02-42a4-a4cf-f605bae52864_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LayoutVLM_[[:space:]]Differentiable[[:space:]]Optimization[[:space:]]of[[:space:]]3D[[:space:]]Layout[[:space:]]via[[:space:]]Vision-Language[[:space:]]Models/76acd092-bccd-4ebd-96b7-d34f440b9db7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LeanGaussian_[[:space:]]Breaking[[:space:]]Pixel[[:space:]]or[[:space:]]Point[[:space:]]Cloud[[:space:]]Correspondence[[:space:]]in[[:space:]]Modeling[[:space:]]3D[[:space:]]Gaussians/0aae6331-af21-474f-9bb6-b38c118338a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learnable[[:space:]]Infinite[[:space:]]Taylor[[:space:]]Gaussian[[:space:]]for[[:space:]]Dynamic[[:space:]]View[[:space:]]Rendering/bf361c76-f15e-47cc-a464-fa4b72fc0fcb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learned[[:space:]]Binocular-Encoding[[:space:]]Optics[[:space:]]for[[:space:]]RGBD[[:space:]]Imaging[[:space:]]Using[[:space:]]Joint[[:space:]]Stereo[[:space:]]and[[:space:]]Focus[[:space:]]Cues/24daed97-f4bc-4fd5-a1fa-264d5651d9a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learned[[:space:]]Image[[:space:]]Compression[[:space:]]with[[:space:]]Dictionary-based[[:space:]]Entropy[[:space:]]Model/e1e0b0b5-a635-4a6d-840d-9969f366c2ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]4D[[:space:]]Panoptic[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]from[[:space:]]Rich[[:space:]]2D[[:space:]]Visual[[:space:]]Scene/81b0b32d-ccb3-4e5a-af84-4cd817d148b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Affine[[:space:]]Correspondences[[:space:]]by[[:space:]]Integrating[[:space:]]Geometric[[:space:]]Constraints/bc7e7c38-4cc7-4e02-af6b-6fb37d6b54bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Audio-guided[[:space:]]Video[[:space:]]Representation[[:space:]]with[[:space:]]Gated[[:space:]]Attention[[:space:]]for[[:space:]]Video-Text[[:space:]]Retrieval/5d6ec6cd-b367-490c-9e25-aaae57c48217_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Bijective[[:space:]]Surface[[:space:]]Parameterization[[:space:]]for[[:space:]]Inferring[[:space:]]Signed[[:space:]]Distance[[:space:]]Functions[[:space:]]from[[:space:]]Sparse[[:space:]]Point[[:space:]]Clouds[[:space:]]with[[:space:]]Grid[[:space:]]Deformation/67d252df-3f53-4b82-8955-d840a4c4ef82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Class[[:space:]]Prototypes[[:space:]]for[[:space:]]Unified[[:space:]]Sparse-Supervised[[:space:]]3D[[:space:]]Object[[:space:]]Detection/7c52d9e8-6d15-4f39-af9b-ac310db1aa46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Compatible[[:space:]]Multi-Prize[[:space:]]Subnetworks[[:space:]]for[[:space:]]Asymmetric[[:space:]]Retrieval/43366f20-4f29-4f3f-ae43-f47add4525d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Conditional[[:space:]]Space-Time[[:space:]]Prompt[[:space:]]Distributions[[:space:]]for[[:space:]]Video[[:space:]]Class-Incremental[[:space:]]Learning/c4eceaf1-21ea-4fef-9221-dd320cd69762_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Dynamic[[:space:]]Collaborative[[:space:]]Network[[:space:]]for[[:space:]]Semi-supervised[[:space:]]3D[[:space:]]Vessel[[:space:]]Segmentation/78a208b8-f2c3-4cdb-915d-7fa8430f44cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Endogenous[[:space:]]Attention[[:space:]]for[[:space:]]Incremental[[:space:]]Object[[:space:]]Detection/6662ac3e-2139-4b36-8a03-5a0177961377_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Extremely[[:space:]]High[[:space:]]Density[[:space:]]Crowds[[:space:]]as[[:space:]]Active[[:space:]]Matters/8b7ad141-4864-425c-879f-c2e01c30831d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Flow[[:space:]]Fields[[:space:]]in[[:space:]]Attention[[:space:]]for[[:space:]]Controllable[[:space:]]Person[[:space:]]Image[[:space:]]Generation/a299c738-798e-411b-af48-68e3301b9b51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Hazing[[:space:]]to[[:space:]]Dehazing_[[:space:]]Towards[[:space:]]Realistic[[:space:]]Haze[[:space:]]Generation[[:space:]]for[[:space:]]Real-World[[:space:]]Image[[:space:]]Dehazing/72975592-1226-4db3-8741-f1638997e13f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Heterogeneous[[:space:]]Tissues[[:space:]]with[[:space:]]Mixture[[:space:]]of[[:space:]]Experts[[:space:]]for[[:space:]]Gigapixel[[:space:]]Whole[[:space:]]Slide[[:space:]]Images/8eafdbdd-d4c1-4345-9723-f2a65ebc4305_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Occlusion-Robust[[:space:]]Vision[[:space:]]Transformers[[:space:]]for[[:space:]]Real-Time[[:space:]]UAV[[:space:]]Tracking/f1567ef1-3598-45e2-a22a-5520868621ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Partonomic[[:space:]]3D[[:space:]]Reconstruction[[:space:]]from[[:space:]]Image[[:space:]]Collections/390b2788-2b46-4081-bf6f-c74d43c381dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Person-Specific[[:space:]]Animatable[[:space:]]Face[[:space:]]Models[[:space:]]from[[:space:]]In-the-Wild[[:space:]]Images[[:space:]]via[[:space:]]a[[:space:]]Shared[[:space:]]Base[[:space:]]Model/27ef5491-e71e-4c3e-8fb7-b054f9c3176a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Phase[[:space:]]Distortion[[:space:]]with[[:space:]]Selective[[:space:]]State[[:space:]]Space[[:space:]]Models[[:space:]]for[[:space:]]Video[[:space:]]Turbulence[[:space:]]Mitigation/458bd0fc-387e-45db-8b53-0b3979214d75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Physics[[:space:]]From[[:space:]]Video_[[:space:]]Unsupervised[[:space:]]Physical[[:space:]]Parameter[[:space:]]Estimation[[:space:]]for[[:space:]]Continuous[[:space:]]Dynamical[[:space:]]Systems/553e0856-21d7-441c-8089-bdc6a1a088a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Physics-Based[[:space:]]Full-Body[[:space:]]Human[[:space:]]Reaching[[:space:]]and[[:space:]]Grasping[[:space:]]from[[:space:]]Brief[[:space:]]Walking[[:space:]]References/94a2cbf3-ff58-4e90-a2bc-90c7e7958765_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Video[[:space:]]Depth[[:space:]]from[[:space:]]Video[[:space:]]Diffusion[[:space:]]Priors/83db1c3e-bcb6-42f8-9c3f-d0b120bfced8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Textual[[:space:]]Prompts[[:space:]]for[[:space:]]Open-World[[:space:]]Semi-Supervised[[:space:]]Learning/fa60e83d-2352-4bdd-bc8a-eb2e98b3371c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Visual[[:space:]]Composition[[:space:]]through[[:space:]]Improved[[:space:]]Semantic[[:space:]]Guidance/947a6bf8-5b60-49e1-a6fc-8f2eee8a04d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Visual[[:space:]]Generative[[:space:]]Priors[[:space:]]without[[:space:]]Text/dceacf57-2e12-4f00-9377-b14706a6c56a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]from[[:space:]]Neighbors_[[:space:]]Category[[:space:]]Extrapolation[[:space:]]for[[:space:]]Long-Tail[[:space:]]Learning/34e4b580-3998-46fb-a168-caa0c5389636_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]from[[:space:]]Streaming[[:space:]]Video[[:space:]]with[[:space:]]Orthogonal[[:space:]]Gradients/625f98aa-3c9b-475f-af89-74b5b06e42aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]from[[:space:]]Synchronization_[[:space:]]Self-Supervised[[:space:]]Uncalibrated[[:space:]]Multi-View[[:space:]]Person[[:space:]]Association[[:space:]]in[[:space:]]Challenging[[:space:]]Scenes/a794a984-6867-4cf6-b2f7-e961747b1dfb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]on[[:space:]]Model[[:space:]]Weights[[:space:]]using[[:space:]]Tree[[:space:]]Experts/bc8731fa-aff4-4f3f-a44a-643a61f5bb92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]to[[:space:]]Detect[[:space:]]Objects[[:space:]]from[[:space:]]Multi-Agent[[:space:]]LiDAR[[:space:]]Scans[[:space:]]without[[:space:]]Manual[[:space:]]Labels/33fca6d6-64e1-410d-81ee-53537a01b1d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]to[[:space:]]Filter[[:space:]]Outlier[[:space:]]Edges[[:space:]]in[[:space:]]Global[[:space:]]SfM/69cd11c0-a9a2-46d6-891e-8428eb4117ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]to[[:space:]]Highlight[[:space:]]Audio[[:space:]]by[[:space:]]Watching[[:space:]]Movies/0456d20e-76c1-4a48-ba97-95d87044933e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]to[[:space:]]Normalize[[:space:]]on[[:space:]]the[[:space:]]SPD[[:space:]]Manifold[[:space:]]under[[:space:]]Bures-Wasserstein[[:space:]]Geometry/8b03520d-03a4-429e-a71d-67264d2dde28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]to[[:space:]]Sample[[:space:]]Effective[[:space:]]and[[:space:]]Diverse[[:space:]]Prompts[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/bfe68aba-3688-4e4d-9440-c9e6b827607e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]with[[:space:]]Noisy[[:space:]]Triplet[[:space:]]Correspondence[[:space:]]for[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/02e41919-626f-4dcc-8941-77e58eb80bb3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning-enabled[[:space:]]Polynomial[[:space:]]Lyapunov[[:space:]]Function[[:space:]]Synthesis[[:space:]]via[[:space:]]High-Accuracy[[:space:]]Counterexample-Guided[[:space:]]Framework/4a95e59a-012e-443f-8352-0acc712ee9ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LesionLocator_[[:space:]]Zero-Shot[[:space:]]Universal[[:space:]]Tumor[[:space:]]Segmentation[[:space:]]and[[:space:]]Tracking[[:space:]]in[[:space:]]3D[[:space:]]Whole-Body[[:space:]]Imaging/8a825843-1118-46c6-a7a5-252fcad04337_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Less[[:space:]]Attention[[:space:]]is[[:space:]]More_[[:space:]]Prompt[[:space:]]Transformer[[:space:]]for[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/7aa4ddfe-4b99-479e-a93b-0b355fdc5dc3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Less[[:space:]]is[[:space:]]More_[[:space:]]Efficient[[:space:]]Image[[:space:]]Vectorization[[:space:]]with[[:space:]]Adaptive[[:space:]]Parameterization/4e3426e3-951b-4f3b-a78b-5baebc0d5429_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Less[[:space:]]is[[:space:]]More_[[:space:]]Efficient[[:space:]]Model[[:space:]]Merging[[:space:]]with[[:space:]]Binary[[:space:]]Task[[:space:]]Switch/db461d3e-2cfe-4c59-bb54-ce152ae4475a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lessons[[:space:]]and[[:space:]]Insights[[:space:]]from[[:space:]]a[[:space:]]Unifying[[:space:]]Study[[:space:]]of[[:space:]]Parameter-Efficient[[:space:]]Fine-Tuning[[:space:]](PEFT)[[:space:]]in[[:space:]]Visual[[:space:]]Recognition/218e7e73-f1f1-4a91-b316-fd9e9cbd3d56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Let[[:space:]]Humanoids[[:space:]]Hike![[:space:]]Integrative[[:space:]]Skill[[:space:]]Development[[:space:]]on[[:space:]]Complex[[:space:]]Trails/7abcf823-5bf8-45d2-90be-77e04436f359_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Let[[:space:]]Samples[[:space:]]Speak_[[:space:]]Mitigating[[:space:]]Spurious[[:space:]]Correlation[[:space:]]by[[:space:]]Exploiting[[:space:]]the[[:space:]]Clusterness[[:space:]]of[[:space:]]Samples/2127432c-0935-44ce-a1df-ac717391303e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Let's[[:space:]]Chorus_[[:space:]]Partner-aware[[:space:]]Hybrid[[:space:]]Song-Driven[[:space:]]3D[[:space:]]Head[[:space:]]Animation/e1fcf447-e255-418d-a377-7cd7706a5fd0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Let's[[:space:]]Verify[[:space:]]and[[:space:]]Reinforce[[:space:]]Image[[:space:]]Generation[[:space:]]Step[[:space:]]by[[:space:]]Step/e2d0e8b0-904b-4de9-a066-c05ee06035b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]3D[[:space:]]Geometric[[:space:]]Priors[[:space:]]in[[:space:]]2D[[:space:]]Rotation[[:space:]]Symmetry[[:space:]]Detection/ee1890d6-90b6-4764-974a-c9711a2efd28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]Global[[:space:]]Stereo[[:space:]]Consistency[[:space:]]for[[:space:]]Category-Level[[:space:]]Shape[[:space:]]and[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation[[:space:]]from[[:space:]]Stereo[[:space:]]Images/51a46f2f-5d68-4976-88fd-ac97bcb2a02f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]Perturbation[[:space:]]Robustness[[:space:]]to[[:space:]]Enhance[[:space:]]Out-of-Distribution[[:space:]]Detection/51f8478e-ab69-4cc4-a2ad-f681d2b75f92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]SD[[:space:]]Map[[:space:]]to[[:space:]]Augment[[:space:]]HD[[:space:]]Map-based[[:space:]]Trajectory[[:space:]]Prediction/6060753a-f3ad-4032-857e-a89003d4ce21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]Temporal[[:space:]]Cues[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Multi-View[[:space:]]3D[[:space:]]Object[[:space:]]Detection/4461cd39-b5e0-4901-bebf-c6baffda2bb8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LeviTor_[[:space:]]3D[[:space:]]Trajectory[[:space:]]Oriented[[:space:]]Image-to-Video[[:space:]]Synthesis/54a7a744-8873-4038-a283-e934ff53ba1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LiDAR-RT_[[:space:]]Gaussian-based[[:space:]]Ray[[:space:]]Tracing[[:space:]]for[[:space:]]Dynamic[[:space:]]LiDAR[[:space:]]Re-simulation/d9f62dc2-e400-47d8-bd79-a26cd8986e3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LiMoE_[[:space:]]Mixture[[:space:]]of[[:space:]]LiDAR[[:space:]]Representation[[:space:]]Learners[[:space:]]from[[:space:]]Automotive[[:space:]]Scenes/e8ba7f6b-c77e-4eef-b07c-569aa41d32c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LiSu_[[:space:]]A[[:space:]]Dataset[[:space:]]and[[:space:]]Method[[:space:]]for[[:space:]]LiDAR[[:space:]]Surface[[:space:]]Normal[[:space:]]Estimation/9fb4568c-01c3-42a1-b5d7-a6dd0c6d6c20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LiVOS_[[:space:]]Light[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation[[:space:]]with[[:space:]]Gated[[:space:]]Linear[[:space:]]Matching/cd329740-2ad8-411b-899b-c64035f6b437_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Libra-Merging_[[:space:]]Importance-redundancy[[:space:]]and[[:space:]]Pruning-merging[[:space:]]Trade-off[[:space:]]for[[:space:]]Acceleration[[:space:]]Plug-in[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Model/e6a57655-3a52-4cdb-8d0e-269c38563f6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LibraGrad_[[:space:]]Balancing[[:space:]]Gradient[[:space:]]Flow[[:space:]]for[[:space:]]Universally[[:space:]]Better[[:space:]]Vision[[:space:]]Transformer[[:space:]]Attributions/058fd863-1420-48e9-a759-b236dd2d0e5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LidarGait++_[[:space:]]Learning[[:space:]]Local[[:space:]]Features[[:space:]]and[[:space:]]Size[[:space:]]Awareness[[:space:]]from[[:space:]]LiDAR[[:space:]]Point[[:space:]]Clouds[[:space:]]for[[:space:]]3D[[:space:]]Gait[[:space:]]Recognition/9e1d7c0f-746b-4a74-be17-8ffa9545c48b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lifelong[[:space:]]Knowledge[[:space:]]Editing[[:space:]]for[[:space:]]Vision[[:space:]]Language[[:space:]]Models[[:space:]]with[[:space:]]Low-Rank[[:space:]]Mixture-of-Experts/e9e824cf-cade-45d1-8dfe-70051bb2ad9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lift3D[[:space:]]Policy_[[:space:]]Lifting[[:space:]]2D[[:space:]]Foundation[[:space:]]Models[[:space:]]for[[:space:]]Robust[[:space:]]3D[[:space:]]Robotic[[:space:]]Manipulation/5e030147-ac58-4508-8c8a-ae68b8d68e12_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lifting[[:space:]]Motion[[:space:]]to[[:space:]]the[[:space:]]3D[[:space:]]World[[:space:]]via[[:space:]]2D[[:space:]]Diffusion/944f10cb-8e82-4917-8a36-ba9e1a612c52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lifting[[:space:]]the[[:space:]]Veil[[:space:]]on[[:space:]]Visual[[:space:]]Information[[:space:]]Flow[[:space:]]in[[:space:]]MLLMs_[[:space:]]Unlocking[[:space:]]Pathways[[:space:]]to[[:space:]]Faster[[:space:]]Inference/07d628d7-4812-46bf-add9-6b3ca12e8221_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Light[[:space:]]Transport-aware[[:space:]]Diffusion[[:space:]]Posterior[[:space:]]Sampling[[:space:]]for[[:space:]]Single-View[[:space:]]Reconstruction[[:space:]]of[[:space:]]3D[[:space:]]Volumes/c310586e-46db-48c8-a068-00677be78c8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Light3R-SfM_[[:space:]]Towards[[:space:]]Feed-forward[[:space:]]Structure-from-Motion/4fc21371-c6fe-431f-8abc-44d8e09afe45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LightLoc_[[:space:]]Learning[[:space:]]Outdoor[[:space:]]LiDAR[[:space:]]Localization[[:space:]]at[[:space:]]Light[[:space:]]Speed/657d1a5b-994b-409b-8c47-41cf2fa75bf4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LinGen_[[:space:]]Towards[[:space:]]High-Resolution[[:space:]]Minute-Length[[:space:]]Text-to-Video[[:space:]]Generation[[:space:]]with[[:space:]]Linear[[:space:]]Computational[[:space:]]Complexity/d48d8f54-93ec-4a50-a63b-b5afb0ff9e2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LineArt_[[:space:]]A[[:space:]]Knowledge-guided[[:space:]]Training-free[[:space:]]High-quality[[:space:]]Appearance[[:space:]]Transfer[[:space:]]for[[:space:]]Design[[:space:]]Drawing[[:space:]]with[[:space:]]Diffusion[[:space:]]Model/516191a9-cd6d-4d4c-843e-129a3297f520_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Linear[[:space:]]Attention[[:space:]]Modeling[[:space:]]for[[:space:]]Learned[[:space:]]Image[[:space:]]Compression/eee572fd-9c73-414d-85cf-1d7029fde4fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Linguistics-aware[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling[[:space:]]for[[:space:]]Self-supervised[[:space:]]Scene[[:space:]]Text[[:space:]]Recognition/636303a3-2964-4519-9d50-a20476d3bec5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Link[[:space:]]to[[:space:]]the[[:space:]]Past_[[:space:]]Temporal[[:space:]]Propagation[[:space:]]for[[:space:]]Fast[[:space:]]3D[[:space:]]Human[[:space:]]Reconstruction[[:space:]]from[[:space:]]Monocular[[:space:]]Video/92d2c26e-df65-4816-8e80-00f5e6e8e0eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Link-based[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]One-Shot[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/533b7f0e-f520-43bd-8d64-4f44a24c68b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LiveCC_[[:space:]]Learning[[:space:]]Video[[:space:]]LLM[[:space:]]with[[:space:]]Streaming[[:space:]]Speech[[:space:]]Transcription[[:space:]]at[[:space:]]Scale/0c746694-2f94-4e98-9d80-dbbfabb0faa2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LoKi_[[:space:]]Low-dimensional[[:space:]]KAN[[:space:]]for[[:space:]]Efficient[[:space:]]Fine-tuning[[:space:]]Image[[:space:]]Models/7a886681-df09-461c-9c9d-a2b8b34f52e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LoRA[[:space:]]Recycle_[[:space:]]Unlocking[[:space:]]Tuning-Free[[:space:]]Few-Shot[[:space:]]Adaptability[[:space:]]in[[:space:]]Visual[[:space:]]Foundation[[:space:]]Models[[:space:]]by[[:space:]]Recycling[[:space:]]Pre-Tuned[[:space:]]LoRAs/bbb0aa5f-0878-44e3-8ad0-d0b5365c159b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LoRA[[:space:]]Subtraction[[:space:]]for[[:space:]]Drift-Resistant[[:space:]]Space[[:space:]]in[[:space:]]Exemplar-Free[[:space:]]Continual[[:space:]]Learning/4e4fc5d6-3890-4dde-8eb3-45f8c6f95639_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LoRACLR_[[:space:]]Contrastive[[:space:]]Adaptation[[:space:]]for[[:space:]]Customization[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/673322f2-8957-4356-b9fb-1d4c360750eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LoRASculpt_[[:space:]]Sculpting[[:space:]]LoRA[[:space:]]for[[:space:]]Harmonizing[[:space:]]General[[:space:]]and[[:space:]]Specialized[[:space:]]Knowledge[[:space:]]in[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/f107eb7c-9341-43bb-ba1b-b507a31496e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LoTUS_[[:space:]]Large-Scale[[:space:]]Machine[[:space:]]Unlearning[[:space:]]with[[:space:]]a[[:space:]]Taste[[:space:]]of[[:space:]]Uncertainty/07ba1d75-72b6-485e-a4c6-9b5239687680_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Locality-Aware[[:space:]]Zero-Shot[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection/9f4b1aab-de28-48aa-94c8-d2fab07b2646_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Localized[[:space:]]Concept[[:space:]]Erasure[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]Using[[:space:]]Training-Free[[:space:]]Gated[[:space:]]Low-Rank[[:space:]]Adaptation/52133e51-ba45-420c-b454-e034fcd93e2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Localizing[[:space:]]Events[[:space:]]in[[:space:]]Videos[[:space:]]with[[:space:]]Multimodal[[:space:]]Queries/52638e88-b204-4200-b749-f4b3b9922eb3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Locally[[:space:]]Orderless[[:space:]]Images[[:space:]]for[[:space:]]Optimization[[:space:]]in[[:space:]]Differentiable[[:space:]]Rendering/25796109-d8b7-4dbe-b9ac-7cd6ccd2e7e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Logits[[:space:]]DeConfusion[[:space:]]with[[:space:]]CLIP[[:space:]]for[[:space:]]Few-Shot[[:space:]]Learning/aae6d653-510f-43f4-af6e-56eb5f0c6d65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LogoSP_[[:space:]]Local-global[[:space:]]Grouping[[:space:]]of[[:space:]]Superpoints[[:space:]]for[[:space:]]Unsupervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]of[[:space:]]3D[[:space:]]Point[[:space:]]Clouds/dac64916-bde8-44e0-8f01-88a49a636548_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Long[[:space:]]Video[[:space:]]Diffusion[[:space:]]Generation[[:space:]]with[[:space:]]Segmented[[:space:]]Cross-Attention[[:space:]]and[[:space:]]Content-Rich[[:space:]]Video[[:space:]]Data[[:space:]]Curation/344850cb-63cd-4531-95fd-5cae79fc795c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LongDiff_[[:space:]]Training-Free[[:space:]]Long[[:space:]]Video[[:space:]]Generation[[:space:]]in[[:space:]]One[[:space:]]Go/9a8c2597-cc2a-42fd-b18d-e6a0fccd26f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LongVALE_[[:space:]]Vision-Audio-Language-Event[[:space:]]Benchmark[[:space:]]Towards[[:space:]]Time-Aware[[:space:]]Omni-Modal[[:space:]]Perception[[:space:]]of[[:space:]]Long[[:space:]]Videos/64249c2e-d687-4336-919a-b15d2943579c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LookCloser_[[:space:]]Frequency-aware[[:space:]]Radiance[[:space:]]Field[[:space:]]for[[:space:]]Tiny-Detail[[:space:]]Scene/efe99b9f-4188-4aee-ac53-d63ae477c686_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LookingGlass_[[:space:]]Generative[[:space:]]Anamorphoses[[:space:]]via[[:space:]]Laplacian[[:space:]]Pyramid[[:space:]]Warping/e96ce183-1050-412d-83f5-ba0552df7c59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lost[[:space:]]in[[:space:]]Translation,[[:space:]]Found[[:space:]]in[[:space:]]Context_[[:space:]]Sign[[:space:]]Language[[:space:]]Translation[[:space:]]with[[:space:]]Contextual[[:space:]]Cues/c707390d-a78f-4b90-9ab3-6b828b92cff5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LotusFilter_[[:space:]]Fast[[:space:]]Diverse[[:space:]]Nearest[[:space:]]Neighbor[[:space:]]Search[[:space:]]via[[:space:]]a[[:space:]]Learned[[:space:]]Cutoff[[:space:]]Table/f73686ed-93d5-41f8-aeb6-ce24ea418d64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Low-Biased[[:space:]]General[[:space:]]Annotated[[:space:]]Dataset[[:space:]]Generation/a2b5e883-fb71-4c38-b0d7-56e9ccb20762_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Low-Rank[[:space:]]Adaptation[[:space:]]in[[:space:]]Multilinear[[:space:]]Operator[[:space:]]Networks[[:space:]]for[[:space:]]Security-Preserving[[:space:]]Incremental[[:space:]]Learning/01a0abf2-2a43-40c7-8aae-8c849b746d8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LumiNet_[[:space:]]Latent[[:space:]]Intrinsics[[:space:]]Meets[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Indoor[[:space:]]Scene[[:space:]]Relighting/f195d1d0-bb88-4991-8489-d290707469fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Luminance-GS_[[:space:]]Adapting[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]to[[:space:]]Challenging[[:space:]]Lighting[[:space:]]Conditions[[:space:]]with[[:space:]]View-Adaptive[[:space:]]Curve[[:space:]]Adjustment/7290f587-8d40-48b4-b120-44ccf4c5c994_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lux[[:space:]]Post[[:space:]]Facto_[[:space:]]Learning[[:space:]]Portrait[[:space:]]Performance[[:space:]]Relighting[[:space:]]with[[:space:]]Conditional[[:space:]]Video[[:space:]]Diffusion[[:space:]]and[[:space:]]a[[:space:]]Hybrid[[:space:]]Dataset/dad3b90e-47f1-491b-9555-0d26e2b808ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/M-LLM[[:space:]]Based[[:space:]]Video[[:space:]]Frame[[:space:]]Selection[[:space:]]for[[:space:]]Efficient[[:space:]]Video[[:space:]]Understanding/32972605-143b-4cef-a945-c4c3e9da0557_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/M3GYM_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Multimodal[[:space:]]Multi-view[[:space:]]Multi-person[[:space:]]Pose[[:space:]]Dataset[[:space:]]for[[:space:]]Fitness[[:space:]]Activity[[:space:]]Understanding[[:space:]]in[[:space:]]Real-world[[:space:]]Settings/22098186-7b16-47cc-ae0a-cb64461a395b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/M3amba_[[:space:]]Memory[[:space:]]Mamba[[:space:]]is[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]for[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Classification/c8b27f5a-66e9-4f77-9dd7-2d06e0d896ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MAC-Ego3D_[[:space:]]Multi-Agent[[:space:]]Gaussian[[:space:]]Consensus[[:space:]]for[[:space:]]Real-Time[[:space:]]Collaborative[[:space:]]Ego-Motion[[:space:]]and[[:space:]]Photorealistic[[:space:]]3D[[:space:]]Reconstruction/2c9e5bee-8729-4d76-abda-1922c7d4b4df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MAD_[[:space:]]Memory-Augmented[[:space:]]Detection[[:space:]]of[[:space:]]3D[[:space:]]Objects/69713549-401a-4757-aebf-72f705d8c26d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MAGE[[:space:]]_[[:space:]]Single[[:space:]]Image[[:space:]]to[[:space:]]Material-Aware[[:space:]]3D[[:space:]]via[[:space:]]the[[:space:]]Multi-View[[:space:]]G-Buffer[[:space:]]Estimation[[:space:]]Model/8741b222-9f6f-43af-b77e-cd4e9c59fd70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MAGiC-SLAM_[[:space:]]Multi-Agent[[:space:]]Gaussian[[:space:]]Globally[[:space:]]Consistent[[:space:]]SLAM/5353dce4-2947-46da-bfec-6595dd647f07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MANTA_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Multi-View[[:space:]]and[[:space:]]Visual-Text[[:space:]]Anomaly[[:space:]]Detection[[:space:]]Dataset[[:space:]]for[[:space:]]Tiny[[:space:]]Objects/7e24860b-dbed-4a13-8431-778c5b74310c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MANTA_[[:space:]]Diffusion[[:space:]]Mamba[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Effective[[:space:]]Stochastic[[:space:]]Long-Term[[:space:]]Dense[[:space:]]Action[[:space:]]Anticipation/bdce2589-2657-44d0-888c-bf877dfae01e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MAP_[[:space:]]Unleashing[[:space:]]Hybrid[[:space:]]Mamba-Transformer[[:space:]]Vision[[:space:]]Backbone's[[:space:]]Potential[[:space:]]with[[:space:]]Masked[[:space:]]Autoregressive[[:space:]]Pretraining/7f53e4ea-012a-4a9b-a7f9-6e384b56698c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MAR-3D_[[:space:]]Progressive[[:space:]]Masked[[:space:]]Auto-regressor[[:space:]]for[[:space:]]High-Resolution[[:space:]]3D[[:space:]]Generation/b73aebb0-6b68-429c-8ccd-f34d73b47472_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MARBLE_[[:space:]]Material[[:space:]]Recomposition[[:space:]]and[[:space:]]Blending[[:space:]]in[[:space:]]CLIP-Space/4765a6fc-379b-4453-8337-95960dbb18c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MARVEL-40M+_[[:space:]]Multi-Level[[:space:]]Visual[[:space:]]Elaboration[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Text-to-3D[[:space:]]Content[[:space:]]Creation/9944745e-7ca9-41f7-9420-5a047d1244c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MASH-VLM_[[:space:]]Mitigating[[:space:]]Action-Scene[[:space:]]Hallucination[[:space:]]in[[:space:]]Video-LLMs[[:space:]]through[[:space:]]Disentangled[[:space:]]Spatial-Temporal[[:space:]]Representations/f5b58337-2ea4-4764-8fff-973b5a9daa66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MASt3R-SLAM_[[:space:]]Real-Time[[:space:]]Dense[[:space:]]SLAM[[:space:]]with[[:space:]]3D[[:space:]]Reconstruction[[:space:]]Priors/fa0f3e39-63c9-4696-88ca-0a3c331d4cb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MATCHA_[[:space:]]Towards[[:space:]]Matching[[:space:]]Anything/dfa42bee-f25e-41a0-90b7-46d2d98b585e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MAtCha[[:space:]]Gaussians_[[:space:]]Atlas[[:space:]]of[[:space:]]Charts[[:space:]]for[[:space:]]High-Quality[[:space:]]Geometry[[:space:]]and[[:space:]]Photorealism[[:space:]]From[[:space:]]Sparse[[:space:]]Views/27cf4031-ed23-4607-a117-00325e28e4cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MBQ_[[:space:]]Modality-Balanced[[:space:]]Quantization[[:space:]]for[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/05a766ac-7649-4e06-896d-a3f1afd772e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MCCD_[[:space:]]Multi-Agent[[:space:]]Collaboration-based[[:space:]]Compositional[[:space:]]Diffusion[[:space:]]for[[:space:]]Complex[[:space:]]Text-to-Image[[:space:]]Generation/4ac57744-529d-4387-9877-5ee6dd81e10b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MC^2_[[:space:]]Multi-concept[[:space:]]Guidance[[:space:]]for[[:space:]]Customized[[:space:]]Multi-concept[[:space:]]Generation/be68938c-3316-462f-9c52-2c0f9f5b3e3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MDP_[[:space:]]Multidimensional[[:space:]]Vision[[:space:]]Model[[:space:]]Pruning[[:space:]]with[[:space:]]Latency[[:space:]]Constraint/e596af1a-fc99-42c9-9eea-a82d3a7526f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MEAT_[[:space:]]Multiview[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Human[[:space:]]Generation[[:space:]]on[[:space:]]Megapixels[[:space:]]with[[:space:]]Mesh[[:space:]]Attention/3912b515-bb12-4f40-b43d-1904a9bf9af5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MEET_[[:space:]]Towards[[:space:]]Memory-Efficient[[:space:]]Temporal[[:space:]]Sparse[[:space:]]Deep[[:space:]]Neural[[:space:]]Networks/30a49dd0-7038-4752-ba5d-337ef2d1abbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MEGA_[[:space:]]Masked[[:space:]]Generative[[:space:]]Autoencoder[[:space:]]for[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/ed7c32bb-0ac3-45e7-9fd0-b6d7ea8f00f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MERGE_[[:space:]]Multi-faceted[[:space:]]Hierarchical[[:space:]]Graph-based[[:space:]]GNN[[:space:]]for[[:space:]]Gene[[:space:]]Expression[[:space:]]Prediction[[:space:]]from[[:space:]]Whole[[:space:]]Slide[[:space:]]Histopathology[[:space:]]Images/0ae85dca-0be9-4689-92e7-1ca344d548d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MESC-3D_Mining[[:space:]]Effective[[:space:]]Semantic[[:space:]]Cues[[:space:]]for[[:space:]]3D[[:space:]]Reconstruction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/cb65a750-352c-4d9f-83ed-2acbaca88403_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MET3R_[[:space:]]Measuring[[:space:]]Multi-View[[:space:]]Consistency[[:space:]]in[[:space:]]Generated[[:space:]]Images/11c18021-c5de-4af6-9522-5775a90a67ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/METASCENES_[[:space:]]Towards[[:space:]]Automated[[:space:]]Replica[[:space:]]Creation[[:space:]]for[[:space:]]Real-world[[:space:]]3D[[:space:]]Scans/c8194c3b-8f6c-4a36-933e-a3aa04369dba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MExD_[[:space:]]An[[:space:]]Expert-Infused[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Whole-Slide[[:space:]]Image[[:space:]]Classification/e872b28c-0e8f-45bf-831a-85c5a58ad3a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MFogHub_[[:space:]]Bridging[[:space:]]Multi-Regional[[:space:]]and[[:space:]]Multi-Satellite[[:space:]]Data[[:space:]]for[[:space:]]Global[[:space:]]Marine[[:space:]]Fog[[:space:]]Detection[[:space:]]and[[:space:]]Forecasting/d0f4fa1e-a049-4615-ad66-4d2c07772828_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MG-MotionLLM_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Motion[[:space:]]Comprehension[[:space:]]and[[:space:]]Generation[[:space:]]across[[:space:]]Multiple[[:space:]]Granularities/18085a81-554a-47d8-9b65-cd5d6dff2f14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MI-DETR_[[:space:]]An[[:space:]]Object[[:space:]]Detection[[:space:]]Model[[:space:]]with[[:space:]]Multi-time[[:space:]]Inquiries[[:space:]]Mechanism/ffffefe5-3be2-429b-8aa2-e4ad6155648a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MICAS_[[:space:]]Multi-grained[[:space:]]In-Context[[:space:]]Adaptive[[:space:]]Sampling[[:space:]]for[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Processing/dabc3652-423c-42af-aad6-8708983d890f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MIDI_[[:space:]]Multi-Instance[[:space:]]Diffusion[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]to[[:space:]]3D[[:space:]]Scene[[:space:]]Generation/e98b5e2c-7ae9-494e-acbb-1bfcaf93a0c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MIMO_[[:space:]]A[[:space:]]Medical[[:space:]]Vision[[:space:]]Language[[:space:]]Model[[:space:]]with[[:space:]]Visual[[:space:]]Referring[[:space:]]Multimodal[[:space:]]Input[[:space:]]and[[:space:]]Pixel[[:space:]]Grounding[[:space:]]Multimodal[[:space:]]Output/a502bd0b-7b65-42ba-acb1-65d49aa49741_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MIMO_[[:space:]]Controllable[[:space:]]Character[[:space:]]Video[[:space:]]Synthesis[[:space:]]with[[:space:]]Spatial[[:space:]]Decomposed[[:space:]]Modeling/ab381371-79de-4207-b08e-f64b406139a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MINIMA_[[:space:]]Modality[[:space:]]Invariant[[:space:]]Image[[:space:]]Matching/20812d95-d03c-418f-9657-e443e7b0d5ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MIRE_[[:space:]]Matched[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations/fa9317b1-ab37-4d80-ac87-75e3a249f5ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MITracker_[[:space:]]Multi-View[[:space:]]Integration[[:space:]]for[[:space:]]Visual[[:space:]]Object[[:space:]]Tracking/e01542c1-da01-4082-bead-1cb60de4dcc0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MLLM-as-a-Judge[[:space:]]for[[:space:]]Image[[:space:]]Safety[[:space:]]without[[:space:]]Human[[:space:]]Labeling/5da87c3e-e4f6-4a61-8bc6-11015ad42f4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MLVU_[[:space:]]Benchmarking[[:space:]]Multi-task[[:space:]]Long[[:space:]]Video[[:space:]]Understanding/7360b821-f979-45d0-83ba-d0126cf9fbf1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MM-OR_[[:space:]]A[[:space:]]Large[[:space:]]Multimodal[[:space:]]Operating[[:space:]]Room[[:space:]]Dataset[[:space:]]for[[:space:]]Semantic[[:space:]]Understanding[[:space:]]of[[:space:]]High-Intensity[[:space:]]Surgical[[:space:]]Environments/03b170ff-dfa7-4417-b986-baf49874f562_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MMAR_[[:space:]]Towards[[:space:]]Lossless[[:space:]]Multi-Modal[[:space:]]Auto-Regressive[[:space:]]Probabilistic[[:space:]]Modeling/170c47d9-a69b-4f0f-bf3b-b55212733e7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MMAudio_[[:space:]]Taming[[:space:]]Multimodal[[:space:]]Joint[[:space:]]Training[[:space:]]for[[:space:]]High-Quality[[:space:]]Video-to-Audio[[:space:]]Synthesis/00f29c12-52ca-4627-b326-3baa2c992685_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MMRL_[[:space:]]Multi-Modal[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/18d065d1-f3e0-4f80-a3d1-e973c6171fb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MMTL-UniAD_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Multimodal[[:space:]]and[[:space:]]Multi-Task[[:space:]]Learning[[:space:]]in[[:space:]]Assistive[[:space:]]Driving[[:space:]]Perception/c3e4edc3-82ce-48fd-9872-59c13458ae64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MMVU_[[:space:]]Measuring[[:space:]]Expert-Level[[:space:]]Multi-Discipline[[:space:]]Video[[:space:]]Understanding/b92f1a2d-32e9-47c5-8092-d4405920fb0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MNE-SLAM_[[:space:]]Multi-Agent[[:space:]]Neural[[:space:]]SLAM[[:space:]]for[[:space:]]Mobile[[:space:]]Robots/218cca4d-6191-46c7-a55c-ccc1f5f18adc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MODA_[[:space:]]Motion-Drift[[:space:]]Augmentation[[:space:]]for[[:space:]]Inertial[[:space:]]Human[[:space:]]Motion[[:space:]]Analysis/e513efad-f0e2-408c-88ac-ddd3ed9e19ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MODfinity_[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]with[[:space:]]Multimodal[[:space:]]Information[[:space:]]Flow[[:space:]]Intertwining/090c1a16-c957-4c49-b1be-2aed110c7b72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MOS-Attack_[[:space:]]A[[:space:]]Scalable[[:space:]]Multi-objective[[:space:]]Adversarial[[:space:]]Attack[[:space:]]Framework/d497afdc-6c2b-4be9-b6d8-8171e91dac04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MOS_[[:space:]]Modeling[[:space:]]Object-Scene[[:space:]]Associations[[:space:]]in[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/4830638f-e099-4da7-8a6d-e48b2bbe55f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MOVIS_[[:space:]]Enhancing[[:space:]]Multi-Object[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]for[[:space:]]Indoor[[:space:]]Scenes/566ac67a-affe-4911-bb73-a514511cc7f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MP-GUI_[[:space:]]Modality[[:space:]]Perception[[:space:]]with[[:space:]]MLLMs[[:space:]]for[[:space:]]GUI[[:space:]]Understanding/5361db1c-479f-4068-b7ee-742b791a1150_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MP-SfM_[[:space:]]Monocular[[:space:]]Surface[[:space:]]Priors[[:space:]]for[[:space:]]Robust[[:space:]]Structure-from-Motion/4b8c0907-4c4b-4ce3-9712-6c9c7f903b70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MPDrive_[[:space:]]Improving[[:space:]]Spatial[[:space:]]Understanding[[:space:]]with[[:space:]]Marker-Based[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/5098bba8-fb85-42e7-b544-cc495e7739ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MTADiffusion_[[:space:]]Mask[[:space:]]Text[[:space:]]Alignment[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Object[[:space:]]Inpainting/42b0f5a2-ecc1-48b8-966a-82cdca9ad816_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MUST_[[:space:]]The[[:space:]]First[[:space:]]Dataset[[:space:]]and[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Multispectral[[:space:]]UAV[[:space:]]Single[[:space:]]Object[[:space:]]Tracking/09dfe5ab-7cf6-429a-837a-6dbbe81c72e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MUSt3R_[[:space:]]Multi-view[[:space:]]Network[[:space:]]for[[:space:]]Stereo[[:space:]]3D[[:space:]]Reconstruction/7314b898-a9d2-4c73-bc23-bc9c30d48894_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MV-DUSt3R+_[[:space:]]Single-Stage[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]from[[:space:]]Sparse[[:space:]]Views[[:space:]]In[[:space:]]2[[:space:]]Seconds/820c0049-4532-4dd6-bd58-036af6ba3051_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MV-MATH_[[:space:]]Evaluating[[:space:]]Multimodal[[:space:]]Math[[:space:]]Reasoning[[:space:]]in[[:space:]]Multi-Visual[[:space:]]Contexts/d3b05f53-0492-437b-81f0-5ec268366fc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MV-SSM_[[:space:]]Multi-View[[:space:]]State[[:space:]]Space[[:space:]]Modeling[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/54604212-fae5-41ca-a535-67e6096b267f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MVBoost_[[:space:]]Boost[[:space:]]3D[[:space:]]Reconstruction[[:space:]]with[[:space:]]Multi-View[[:space:]]Refinement/541a0189-7907-4182-9eca-640c2e0e8dc5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MVDoppler-Pose_[[:space:]]Multi-Modal[[:space:]]Multi-View[[:space:]]mmWave[[:space:]]Sensing[[:space:]]for[[:space:]]Long-Distance[[:space:]]Self-Occluded[[:space:]]Human[[:space:]]Walking[[:space:]]Pose[[:space:]]Estimation/a2cfc937-ed41-49e3-a686-a61a709b2189_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MVGenMaster_[[:space:]]Scaling[[:space:]]Multi-View[[:space:]]Generation[[:space:]]from[[:space:]]Any[[:space:]]Image[[:space:]]via[[:space:]]3D[[:space:]]Priors[[:space:]]Enhanced[[:space:]]Diffusion[[:space:]]Model/58310381-3356-4da2-80b4-504e92148788_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MVPaint_[[:space:]]Synchronized[[:space:]]Multi-View[[:space:]]Diffusion[[:space:]]for[[:space:]]Painting[[:space:]]Anything[[:space:]]3D/dd4975b0-e470-468c-b265-8d686ef5e4b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MVPortrait_[[:space:]]Text-Guided[[:space:]]Motion[[:space:]]and[[:space:]]Emotion[[:space:]]Control[[:space:]]for[[:space:]]Multi-view[[:space:]]Vivid[[:space:]]Portrait[[:space:]]Animation/3220b978-c65a-4cc8-800d-fb811ce4b06a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MVSAnywhere_[[:space:]]Zero-Shot[[:space:]]Multi-View[[:space:]]Stereo/e78c7625-2c9c-482e-9feb-fbd466c0c8fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/M^3-VOS_[[:space:]]Multi-Phase,[[:space:]]Multi-Transition,[[:space:]]and[[:space:]]Multi-Scenery[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/01627f7d-341a-4c93-88f0-ce1d95d337a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaDCoW_[[:space:]]Marginal[[:space:]]Distortion[[:space:]]Correction[[:space:]]for[[:space:]]Wide-Angle[[:space:]]Photography[[:space:]]with[[:space:]]Arbitrary[[:space:]]Objects/2667a534-8127-4f5b-8e19-69235ca5373f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaIR_[[:space:]]A[[:space:]]Locality-[[:space:]]and[[:space:]]Continuity-Preserving[[:space:]]Mamba[[:space:]]for[[:space:]]Image[[:space:]]Restoration/196e0583-53c6-4998-a232-b611af107504_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaRI_[[:space:]]Material[[:space:]]Retrieval[[:space:]]Integration[[:space:]]across[[:space:]]Domains/54470e0e-7427-4bdd-82d4-b3bc7f7dade4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaSS13K_[[:space:]]A[[:space:]]Matting-level[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]Benchmark/a70f3a74-c0e5-4377-ab83-76ff548ca35b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MagicArticulate_[[:space:]]Make[[:space:]]Your[[:space:]]3D[[:space:]]Models[[:space:]]Articulation-Ready/8f2aa2ea-1110-4410-a357-1640024d93d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MagicQuill_[[:space:]]An[[:space:]]Intelligent[[:space:]]Interactive[[:space:]]Image[[:space:]]Editing[[:space:]]System/b889e858-507e-421d-918e-5669f7d14b62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Magma_[[:space:]]A[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Multimodal[[:space:]]AI[[:space:]]Agents/d9faaf8d-1341-4cb4-a504-8d422fb0e683_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Maintaining[[:space:]]Consistent[[:space:]]Inter-Class[[:space:]]Topology[[:space:]]in[[:space:]]Continual[[:space:]]Test-Time[[:space:]]Adaptation/f462acf4-ea2d-4fe3-82a1-3ff446dd3058_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Make[[:space:]]It[[:space:]]Count_[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]with[[:space:]]an[[:space:]]Accurate[[:space:]]Number[[:space:]]of[[:space:]]Objects/f4af7cd0-6071-477a-9fae-2e207cb142ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Make-It-Animatable_[[:space:]]An[[:space:]]Efficient[[:space:]]Framework[[:space:]]for[[:space:]]Authoring[[:space:]]Animation-Ready[[:space:]]3D[[:space:]]Characters/2bca9ca2-829e-4be2-9f45-b2f23686846d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Making[[:space:]]Old[[:space:]]Film[[:space:]]Great[[:space:]]Again_[[:space:]]Degradation-aware[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]for[[:space:]]Old[[:space:]]Film[[:space:]]Restoration/e3994b4d-58d2-4995-8826-fc9b6c39a2de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mamba[[:space:]]as[[:space:]]a[[:space:]]Bridge_[[:space:]]Where[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models[[:space:]]Meet[[:space:]]Vision[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Domain-Generalized[[:space:]]Semantic[[:space:]]Segmentation/054a620e-98cd-43cc-a458-596df88d5013_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mamba-Adaptor_[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]Adaptor[[:space:]]for[[:space:]]Visual[[:space:]]Recognition/0789c523-cb52-4fd5-bf32-ee6941b42756_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mamba-Reg_[[:space:]]Vision[[:space:]]Mamba[[:space:]]Also[[:space:]]Needs[[:space:]]Registers/25d5e916-ae94-4735-bf5f-31eac437fc78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mamba4D_[[:space:]]Efficient[[:space:]]4D[[:space:]]Point[[:space:]]Cloud[[:space:]]Video[[:space:]]Understanding[[:space:]]with[[:space:]]Disentangled[[:space:]]Spatial-Temporal[[:space:]]State[[:space:]]Space[[:space:]]Models/8ffd020b-02c2-4cf2-a126-01574ecbbc47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MambaIC_[[:space:]]State[[:space:]]Space[[:space:]]Models[[:space:]]for[[:space:]]High-Performance[[:space:]]Learned[[:space:]]Image[[:space:]]Compression/6fee6e7b-620e-4a86-8795-52db348c6ee4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MambaIRv2_[[:space:]]Attentive[[:space:]]State[[:space:]]Space[[:space:]]Restoration/6db7444f-1e6a-4960-9ecf-eef2b3246194_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MambaOut_[[:space:]]Do[[:space:]]We[[:space:]]Really[[:space:]]Need[[:space:]]Mamba[[:space:]]for[[:space:]]Vision_/2520899f-fef7-4a37-ae58-2e9cf07e23dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MambaVLT_[[:space:]]Time-Evolving[[:space:]]Multimodal[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]for[[:space:]]Vision-Language[[:space:]]Tracking/f0e30704-d74b-4f92-8212-dfb21c1db099_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MambaVO_[[:space:]]Deep[[:space:]]Visual[[:space:]]Odometry[[:space:]]Based[[:space:]]on[[:space:]]Sequential[[:space:]]Matching[[:space:]]Refinement[[:space:]]and[[:space:]]Training[[:space:]]Smoothing/0c535f77-8eb6-47be-89e2-4c61ac14c555_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MambaVision_[[:space:]]A[[:space:]]Hybrid[[:space:]]Mamba-Transformer[[:space:]]Vision[[:space:]]Backbone/adfb5a35-45b2-4117-acf0-4d54bb3696ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MammAlps_[[:space:]]A[[:space:]]Multi-view[[:space:]]Video[[:space:]]Behavior[[:space:]]Monitoring[[:space:]]Dataset[[:space:]]of[[:space:]]Wild[[:space:]]Mammals[[:space:]]in[[:space:]]the[[:space:]]Swiss[[:space:]]Alps/1b6d83ee-3a3b-40d8-9966-769372bd739b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MangaNinja_[[:space:]]Line[[:space:]]Art[[:space:]]Colorization[[:space:]]with[[:space:]]Precise[[:space:]]Reference[[:space:]]Following/4cffc8c0-9fb5-413d-9891-8a6fb39e1ee8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mani-GS_[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Manipulation[[:space:]]with[[:space:]]Triangular[[:space:]]Mesh/84bdfa54-cbb6-4585-bd48-54e24b64b429_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ManiVideo_[[:space:]]Generating[[:space:]]Hand-Object[[:space:]]Manipulation[[:space:]]Video[[:space:]]with[[:space:]]Dexterous[[:space:]]and[[:space:]]Generalizable[[:space:]]Grasping/44143190-6020-4b64-be80-a4066ebfe905_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ManipTrans_[[:space:]]Efficient[[:space:]]Dexterous[[:space:]]Bimanual[[:space:]]Manipulation[[:space:]]Transfer[[:space:]]via[[:space:]]Residual[[:space:]]Learning/0dae64e7-ffc6-4a26-bbc7-963a7bf8caae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MarkushGrapher_[[:space:]]Joint[[:space:]]Visual[[:space:]]and[[:space:]]Textual[[:space:]]Recognition[[:space:]]of[[:space:]]Markush[[:space:]]Structures/9216a15f-6e86-4fc8-98bc-f6947280e462_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Marten_[[:space:]]Visual[[:space:]]Question[[:space:]]Answering[[:space:]]with[[:space:]]Mask[[:space:]]Generation[[:space:]]for[[:space:]]Multi-modal[[:space:]]Document[[:space:]]Understanding/feabbd7e-711c-4b39-85ea-57eafd52a519_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mask-Adapter_[[:space:]]The[[:space:]]Devil[[:space:]]is[[:space:]]in[[:space:]]the[[:space:]]Masks[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Segmentation/6dc5a052-f20b-494f-8f7c-4affac317e8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaskGWM_[[:space:]]A[[:space:]]Generalizable[[:space:]]Driving[[:space:]]World[[:space:]]Model[[:space:]]with[[:space:]]Video[[:space:]]Mask[[:space:]]Reconstruction/74991f13-75ef-4210-a3bf-186d35376e02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaskGaussian_[[:space:]]Adaptive[[:space:]]3D[[:space:]]Gaussian[[:space:]]Representation[[:space:]]from[[:space:]]Probabilistic[[:space:]]Masks/24e9d976-ef8e-405f-a59c-189da91265ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mask^2DiT_[[:space:]]Dual[[:space:]]Mask-based[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]for[[:space:]]Multi-Scene[[:space:]]Long[[:space:]]Video[[:space:]]Generation/56a829d6-84bc-41e2-b1c7-8cf27554a61b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Masked[[:space:]]Point-Entity[[:space:]]Contrast[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/862e6312-7eb6-490b-8b6d-03c530d7cc86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Masked[[:space:]]Scene[[:space:]]Modeling_[[:space:]]Narrowing[[:space:]]the[[:space:]]Gap[[:space:]]Between[[:space:]]Supervised[[:space:]]and[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]in[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/721b9157-13fe-41cc-b9c0-642df6baa15e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Masking[[:space:]]meets[[:space:]]Supervision_[[:space:]]A[[:space:]]Strong[[:space:]]Learning[[:space:]]Alliance/8fe2b7c5-43df-4837-b74c-43dcff90ef33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MatAnyone_[[:space:]]Stable[[:space:]]Video[[:space:]]Matting[[:space:]]with[[:space:]]Consistent[[:space:]]Memory[[:space:]]Propagation/bb1ab0d1-f66f-4983-9145-4ce24228a380_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Material[[:space:]]Anything_[[:space:]]Generating[[:space:]]Materials[[:space:]]for[[:space:]]Any[[:space:]]3D[[:space:]]Object[[:space:]]via[[:space:]]Diffusion/3227c5ce-b15a-45a8-9acf-37fb5537d3a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Matrix-Free[[:space:]]Shared[[:space:]]Intrinsics[[:space:]]Bundle[[:space:]]Adjustment/abfaba39-0b35-4a85-86bd-7f1763c34dfe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Matrix3D_[[:space:]]Large[[:space:]]Photogrammetry[[:space:]]Model[[:space:]]All-in-One/2c73988c-d2c6-4328-8f54-886217065830_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MeGA_[[:space:]]Hybrid[[:space:]]Mesh-Gaussian[[:space:]]Head[[:space:]]Avatar[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Rendering[[:space:]]and[[:space:]]Head[[:space:]]Editing/c7f36af5-a80e-4d22-83eb-ee70486dbebb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MedUnifier_[[:space:]]Unifying[[:space:]]Vision-and-Language[[:space:]]Pre-training[[:space:]]on[[:space:]]Medical[[:space:]]Data[[:space:]]with[[:space:]]Vision[[:space:]]Generation[[:space:]]Task[[:space:]]using[[:space:]]Discrete[[:space:]]Visual[[:space:]]Representations/ce6811e5-b6ae-42a7-9a43-3ea7701234e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Medusa_[[:space:]]A[[:space:]]Multi-Scale[[:space:]]High-order[[:space:]]Contrastive[[:space:]]Dual-Diffusion[[:space:]]Approach[[:space:]]for[[:space:]]Multi-View[[:space:]]Clustering/ded8d5ad-68e8-4ae8-83fc-09a7ef8eb34e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MegaSaM_[[:space:]]Accurate,[[:space:]]Fast[[:space:]]and[[:space:]]Robust[[:space:]]Structure[[:space:]]and[[:space:]]Motion[[:space:]]from[[:space:]]Casual[[:space:]]Dynamic[[:space:]]Videos/59637ec6-3587-41d6-b419-dbecf7856256_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MegaSynth_[[:space:]]Scaling[[:space:]]Up[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]with[[:space:]]Synthesized[[:space:]]Data/63fbcd7b-98da-4a5e-b94f-6067496751c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Memories[[:space:]]of[[:space:]]Forgotten[[:space:]]Concepts/c45512ae-4c62-4557-aaf8-c6fdf90a5b3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MergeVQ_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Visual[[:space:]]Generation[[:space:]]and[[:space:]]Representation[[:space:]]with[[:space:]]Disentangled[[:space:]]Token[[:space:]]Merging[[:space:]]and[[:space:]]Quantization/24f1bc6d-65ae-4319-9598-4036ab3192c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mesh[[:space:]]Mamba_[[:space:]]A[[:space:]]Unified[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]for[[:space:]]Saliency[[:space:]]Prediction[[:space:]]in[[:space:]]Non-Textured[[:space:]]and[[:space:]]Textured[[:space:]]Meshes/64303da7-8b9b-4da3-a7a9-8e2c33f6a5af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MeshArt_[[:space:]]Generating[[:space:]]Articulated[[:space:]]Meshes[[:space:]]with[[:space:]]Structure-Guided[[:space:]]Transformers/e25a637e-0ee4-4d4a-a672-02acc4b0482c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MeshGen_[[:space:]]Generating[[:space:]]PBR[[:space:]]Textured[[:space:]]Mesh[[:space:]]with[[:space:]]Render-Enhanced[[:space:]]Auto-Encoder[[:space:]]and[[:space:]]Generative[[:space:]]Data[[:space:]]Augmentation/7de70104-4b45-4af4-a3c6-07116cbeba97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Meta-Learning[[:space:]]Hyperparameters[[:space:]]for[[:space:]]Parameter[[:space:]]Efficient[[:space:]]Fine-Tuning/b6a33a48-9ee7-413f-ad90-02951fb80e15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MetaShadow_[[:space:]]Object-Centered[[:space:]]Shadow[[:space:]]Detection,[[:space:]]Removal,[[:space:]]and[[:space:]]Synthesis/7e1e126f-121e-4788-aad5-39862db294e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MetaWriter_[[:space:]]Personalized[[:space:]]Handwritten[[:space:]]Text[[:space:]]Recognition[[:space:]]Using[[:space:]]Meta-Learned[[:space:]]Prompt[[:space:]]Tuning/6e942c09-bb74-4797-8880-85c7f3819d5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MetricGrids_[[:space:]]Arbitrary[[:space:]]Nonlinear[[:space:]]Approximation[[:space:]]with[[:space:]]Elementary[[:space:]]Metric[[:space:]]Grids[[:space:]]based[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representation/70d68c29-f28b-494b-b022-8243ea208a27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MicroVQA_[[:space:]]A[[:space:]]Multimodal[[:space:]]Reasoning[[:space:]]Benchmark[[:space:]]for[[:space:]]Microscopy-Based[[:space:]]Scientific[[:space:]]Research/655b1b6c-888e-4518-b82f-f90a28314a97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mimic[[:space:]]In-Context[[:space:]]Learning[[:space:]]for[[:space:]]Multimodal[[:space:]]Tasks/342f7b46-95a1-4c6b-9725-7cd96012a666_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mimir_[[:space:]]Improving[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Precise[[:space:]]Text[[:space:]]Understanding/764fca72-c139-441b-bf46-165a068b73da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mind[[:space:]]the[[:space:]]Gap_[[:space:]]Confidence[[:space:]]Discrepancy[[:space:]]Can[[:space:]]Guide[[:space:]]Federated[[:space:]]Semi-Supervised[[:space:]]Learning[[:space:]]Across[[:space:]]Pseudo-Mismatch/9473851e-e573-4d38-a2a2-fc057d561d10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mind[[:space:]]the[[:space:]]Gap_[[:space:]]Detecting[[:space:]]Black-box[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]in[[:space:]]the[[:space:]]Making[[:space:]]through[[:space:]]Query[[:space:]]Update[[:space:]]Analysis/18052269-1311-4c13-bd15-a0860057171b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mind[[:space:]]the[[:space:]]Time_[[:space:]]Temporally-Controlled[[:space:]]Multi-Event[[:space:]]Video[[:space:]]Generation/3dcdf987-470f-4b6e-a90c-c91deefa1ea8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mind[[:space:]]the[[:space:]]Trojan[[:space:]]Horse_[[:space:]]Image[[:space:]]Prompt[[:space:]]Adapter[[:space:]]Enabling[[:space:]]Scalable[[:space:]]and[[:space:]]Deceptive[[:space:]]Jailbreaking/ed2577b7-e83e-4319-83ce-1c1ef3ee4e34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Minding[[:space:]]Fuzzy[[:space:]]Regions_[[:space:]]A[[:space:]]Data-driven[[:space:]]Alternating[[:space:]]Learning[[:space:]]Paradigm[[:space:]]for[[:space:]]Stable[[:space:]]Lesion[[:space:]]Segmentation/ea50b849-e9dc-47c1-9024-838d547bcf34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Minimal[[:space:]]Interaction[[:space:]]Seperated[[:space:]]Tuning_[[:space:]]A[[:space:]]New[[:space:]]Paradigm[[:space:]]for[[:space:]]Visual[[:space:]]Adaptation/638c7bfa-ab10-402b-9d80-7df3ec15a82d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Minimizing[[:space:]]Labeled,[[:space:]]Maximizing[[:space:]]Unlabeled_[[:space:]]An[[:space:]]Image-Driven[[:space:]]Approach[[:space:]]for[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation/0317cd73-4f67-431e-ad47-565aef690904_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Minority-Focused[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]via[[:space:]]Prompt[[:space:]]Optimization/5e174202-139a-45c3-8235-8d5060a9825f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MirrorVerse_[[:space:]]Pushing[[:space:]]Diffusion[[:space:]]Models[[:space:]]to[[:space:]]Realistically[[:space:]]Reflect[[:space:]]the[[:space:]]World/66dab947-b500-4c37-8602-6c512c3ac396_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Missing[[:space:]]Target-Relevant[[:space:]]Information[[:space:]]Prediction[[:space:]]with[[:space:]]World[[:space:]]Model[[:space:]]for[[:space:]]Accurate[[:space:]]Zero-Shot[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/015315b9-8695-4cba-bf9d-c3dd3048d132_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mitigating[[:space:]]Ambiguities[[:space:]]in[[:space:]]3D[[:space:]]Classification[[:space:]]with[[:space:]]Gaussian[[:space:]]Splatting/55fe5378-4941-4756-ba41-deb1dd96cfda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mitigating[[:space:]]Hallucinations[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]via[[:space:]]DPO_[[:space:]]On-Policy[[:space:]]Data[[:space:]]Hold[[:space:]]the[[:space:]]Key/1ea369ad-e84c-42b0-93ae-d223abc3d42a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mitigating[[:space:]]Object[[:space:]]Hallucinations[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]Assembly[[:space:]]of[[:space:]]Global[[:space:]]and[[:space:]]Local[[:space:]]Attention/385c091f-69a4-4671-87ce-3159dfbbe39a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mitigating[[:space:]]the[[:space:]]Human-Robot[[:space:]]Domain[[:space:]]Discrepancy[[:space:]]in[[:space:]]Visual[[:space:]]Pre-training[[:space:]]for[[:space:]]Robotic[[:space:]]Manipulation/4db7c8fc-c442-4bbc-856c-93a749cee3d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MixerMDM_[[:space:]]Learnable[[:space:]]Composition[[:space:]]of[[:space:]]Human[[:space:]]Motion[[:space:]]Diffusion[[:space:]]Models/43876985-3a23-430f-9d61-6f01f1109fe1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mixture[[:space:]]of[[:space:]]Submodules[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Person[[:space:]]Search/5c6a6072-e07b-47dd-bb4f-6054fd7ca144_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoDec-GS_[[:space:]]Global-to-Local[[:space:]]Motion[[:space:]]Decomposition[[:space:]]and[[:space:]]Temporal[[:space:]]Interval[[:space:]]Adjustment[[:space:]]for[[:space:]]Compact[[:space:]]Dynamic[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/103786fa-2967-4bd3-9d34-20e9ec6b5fe8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoEE_[[:space:]]Mixture[[:space:]]of[[:space:]]Emotion[[:space:]]Experts[[:space:]]for[[:space:]]Audio-Driven[[:space:]]Portrait[[:space:]]Animation/1e61482c-8133-4996-92ff-368bad4c06d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoEdit_[[:space:]]On[[:space:]]Learning[[:space:]]Quantity[[:space:]]Perception[[:space:]]for[[:space:]]Multi-object[[:space:]]Image[[:space:]]Editing/9b6df6b0-504e-4917-a214-f40a1c2704a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoFlow_[[:space:]]One-Step[[:space:]]Flow[[:space:]]Matching[[:space:]]for[[:space:]]Human[[:space:]]Trajectory[[:space:]]Forecasting[[:space:]]via[[:space:]]Implicit[[:space:]]Maximum[[:space:]]Likelihood[[:space:]]Estimation[[:space:]]based[[:space:]]Distillation/a498e01a-84f8-4c3c-9bb6-8bb6686b5ed8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoGe_[[:space:]]Unlocking[[:space:]]Accurate[[:space:]]Monocular[[:space:]]Geometry[[:space:]]Estimation[[:space:]]for[[:space:]]Open-Domain[[:space:]]Images[[:space:]]with[[:space:]]Optimal[[:space:]]Training[[:space:]]Supervision/a3a9f8d1-532a-4e16-ab95-f3b166306e23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoManipVLA_[[:space:]]Transferring[[:space:]]Vision-language-action[[:space:]]Models[[:space:]]for[[:space:]]General[[:space:]]Mobile[[:space:]]Manipulation/4ee3797c-85a9-423f-8553-06831d41c281_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoST_[[:space:]]Efficient[[:space:]]Monarch[[:space:]]Sparse[[:space:]]Tuning[[:space:]]for[[:space:]]3D[[:space:]]Representation[[:space:]]Learning/45a2aaf0-f48f-4e98-b6a8-f3cb9833357c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoSca_[[:space:]]Dynamic[[:space:]]Gaussian[[:space:]]Fusion[[:space:]]from[[:space:]]Casual[[:space:]]Videos[[:space:]]via[[:space:]]4D[[:space:]]Motion[[:space:]]Scaffolds/d8b08541-b054-4da8-8d87-3d779b31ae6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoVE-KD_[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]VLMs[[:space:]]with[[:space:]]Mixture[[:space:]]of[[:space:]]Visual[[:space:]]Encoders/ab6c14bc-6079-4f8f-a5e7-4099ae1c02b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MobileH2R_[[:space:]]Learning[[:space:]]Generalizable[[:space:]]Human[[:space:]]to[[:space:]]Mobile[[:space:]]Robot[[:space:]]Handover[[:space:]]Exclusively[[:space:]]from[[:space:]]Scalable[[:space:]]and[[:space:]]Diverse[[:space:]]Synthetic[[:space:]]Data/a524cc5a-81b5-4d27-b9b7-4c9c0e7be696_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MobileMamba_[[:space:]]Lightweight[[:space:]]Multi-Receptive[[:space:]]Visual[[:space:]]Mamba[[:space:]]Network/bd7a42c9-fc43-4c95-9aad-4230d7af037a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MobilePortrait_[[:space:]]Real-Time[[:space:]]One-Shot[[:space:]]Neural[[:space:]]Head[[:space:]]Avatars[[:space:]]on[[:space:]]Mobile[[:space:]]Devices/82153080-e100-436f-814d-8c7d9e2fc092_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ModeSeq_[[:space:]]Taming[[:space:]]Sparse[[:space:]]Multimodal[[:space:]]Motion[[:space:]]Prediction[[:space:]]with[[:space:]]Sequential[[:space:]]Mode[[:space:]]Modeling/570d0237-a7d5-44c1-8921-c6090d078afd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Model[[:space:]]Diagnosis[[:space:]]and[[:space:]]Correction[[:space:]]via[[:space:]]Linguistic[[:space:]]and[[:space:]]Implicit[[:space:]]Attribute[[:space:]]Editing/5e292c16-7716-4b95-9058-dc676d0e9fd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Model[[:space:]]Poisoning[[:space:]]Attacks[[:space:]]to[[:space:]]Federated[[:space:]]Learning[[:space:]]via[[:space:]]Multi-Round[[:space:]]Consistency/f7032ae1-d982-4bcf-b557-155cfc88dbec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Modeling[[:space:]]Multiple[[:space:]]Normal[[:space:]]Action[[:space:]]Representations[[:space:]]for[[:space:]]Error[[:space:]]Detection[[:space:]]in[[:space:]]Procedural[[:space:]]Tasks/afa210cf-7905-4e7f-9041-77ee05151b18_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Modeling[[:space:]]Thousands[[:space:]]of[[:space:]]Human[[:space:]]Annotators[[:space:]]for[[:space:]]Generalizable[[:space:]]Text-to-Image[[:space:]]Person[[:space:]]Re-identification/8c31cc27-90d4-4bff-9cfe-9c1bc6e6486b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Molmo[[:space:]]and[[:space:]]PixMo_[[:space:]]Open[[:space:]]Weights[[:space:]]and[[:space:]]Open[[:space:]]Data[[:space:]]for[[:space:]]State-of-the-Art[[:space:]]Vision-Language[[:space:]]Models/a2e8d8f6-bc6a-4ef7-a351-4818e08f787b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MonSter_[[:space:]]Marry[[:space:]]Monodepth[[:space:]]to[[:space:]]Stereo[[:space:]]Unleashes[[:space:]]Power/e578532e-21be-4275-87ef-3418ad0e542f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mono-InternVL_[[:space:]]Pushing[[:space:]]the[[:space:]]Boundaries[[:space:]]of[[:space:]]Monolithic[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]with[[:space:]]Endogenous[[:space:]]Visual[[:space:]]Pre-training/6a7b1a05-1f2f-4300-88df-174c12a656ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mono2Stereo_[[:space:]]A[[:space:]]Benchmark[[:space:]]and[[:space:]]Empirical[[:space:]]Study[[:space:]]for[[:space:]]Stereo[[:space:]]Conversion/c3cbdb17-27b5-4fbd-8053-30a7782d6cad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mono3DVLT_[[:space:]]Monocular-Video-Based[[:space:]]3D[[:space:]]Visual[[:space:]]Language[[:space:]]Tracking/1a85fd1f-bfc1-4251-a4d1-37097db31dbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MonoDGP_[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Decoupled-Query[[:space:]]and[[:space:]]Geometry-Error[[:space:]]Priors/60af5e9e-87d7-4c6e-b44c-f0754208e228_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MonoInstance_[[:space:]]Enhancing[[:space:]]Monocular[[:space:]]Priors[[:space:]]via[[:space:]]Multi-view[[:space:]]Instance[[:space:]]Alignment[[:space:]]for[[:space:]]Neural[[:space:]]Rendering[[:space:]]and[[:space:]]Reconstruction/40b3ea90-399a-4a93-9f06-e8fe2179eba6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MonoPlace3D_[[:space:]]Learning[[:space:]]3D-Aware[[:space:]]Object[[:space:]]Placement[[:space:]]for[[:space:]]3D[[:space:]]Monocular[[:space:]]Detection/90ccf5c3-784d-4d1f-8f6e-a70e82f67c82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MonoSplat_[[:space:]]Generalizable[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Monocular[[:space:]]Depth[[:space:]]Foundation[[:space:]]Models/4e608f0d-d688-43cc-8a07-cf059c925983_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MonoTAKD_[[:space:]]Teaching[[:space:]]Assistant[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection/bde17624-d7bc-4f61-aadc-d5ec00c07475_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Monocular[[:space:]]and[[:space:]]Generalizable[[:space:]]Gaussian[[:space:]]Talking[[:space:]]Head[[:space:]]Animation/957deefb-000b-4681-bb27-d1432a54d0ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Morpheus_[[:space:]]Text-Driven[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splat[[:space:]]Shape[[:space:]]and[[:space:]]Color[[:space:]]Stylization/cc07fc60-8c4a-4475-853e-f1d270a6e790_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mosaic[[:space:]]of[[:space:]]Modalities_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Benchmark[[:space:]]for[[:space:]]Multimodal[[:space:]]Graph[[:space:]]Learning/4d0e5184-aac4-472c-aed6-148fb174b963_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mosaic3D_[[:space:]]Foundation[[:space:]]Dataset[[:space:]]and[[:space:]]Model[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Segmentation/02f69ed3-d686-4af9-a9d8-cef40e18af30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotiF_[[:space:]]Making[[:space:]]Text[[:space:]]Count[[:space:]]in[[:space:]]Image[[:space:]]Animation[[:space:]]with[[:space:]]Motion[[:space:]]Focal[[:space:]]Loss/e9eb2951-ced9-4f02-b3e0-850f6e246ac4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Motion[[:space:]]Modes_[[:space:]]What[[:space:]]Could[[:space:]]Happen[[:space:]]Next_/d7676079-5376-4e10-b5b2-6fd4f8d86e11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Motion[[:space:]]Prompting_[[:space:]]Controlling[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Motion[[:space:]]Trajectories/c2d6eafe-71b1-4826-bc55-231cb5a5be73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Motion-Grounded[[:space:]]Video[[:space:]]Reasoning_[[:space:]]Understanding[[:space:]]and[[:space:]]Perceiving[[:space:]]Motion[[:space:]]at[[:space:]]Pixel[[:space:]]Level/0d9bb875-d59f-4000-b2df-cea87f169b45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotionBench_[[:space:]]Benchmarking[[:space:]]and[[:space:]]Improving[[:space:]]Fine-grained[[:space:]]Video[[:space:]]Motion[[:space:]]Understanding[[:space:]]for[[:space:]]Vision[[:space:]]Language[[:space:]]Models/2531e09f-c0cc-4575-bdd2-afd6f59adcb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotionMap_[[:space:]]Representing[[:space:]]Multimodality[[:space:]]in[[:space:]]Human[[:space:]]Pose[[:space:]]Forecasting/03443153-ba6c-453c-b407-5055757e25d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotionPRO_[[:space:]]Exploring[[:space:]]the[[:space:]]Role[[:space:]]of[[:space:]]Pressure[[:space:]]in[[:space:]]Human[[:space:]]MoCap[[:space:]]and[[:space:]]Beyond/4af579d6-0d4f-48f3-8bd4-98104375cb04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotionPro_[[:space:]]A[[:space:]]Precise[[:space:]]Motion[[:space:]]Controller[[:space:]]for[[:space:]]Image-to-Video[[:space:]]Generation/4cf139b7-8c09-4a8e-8708-f72958265fbf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotionStone_[[:space:]]Decoupled[[:space:]]Motion[[:space:]]Intensity[[:space:]]Modulation[[:space:]]with[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]for[[:space:]]Image-to-Video[[:space:]]Generation/dbd35dd9-1a36-436b-9349-9b6b2f8d8860_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Motions[[:space:]]as[[:space:]]Queries_[[:space:]]One-Stage[[:space:]]Multi-Person[[:space:]]Holistic[[:space:]]Human[[:space:]]Motion[[:space:]]Capture/5c248acd-8bdd-413f-96d8-ee7cc25fe524_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Move-in-2D_[[:space:]]2D-Conditioned[[:space:]]Human[[:space:]]Motion[[:space:]]Generation/b761f725-21fb-42dd-8447-08e2ace25ddf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Movie[[:space:]]Weaver_[[:space:]]Tuning-Free[[:space:]]Multi-Concept[[:space:]]Video[[:space:]]Personalization[[:space:]]with[[:space:]]Anchored[[:space:]]Prompts/316b1746-3cf2-4d89-b96f-251c9c4c1da9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MovieBench_[[:space:]]A[[:space:]]Hierarchical[[:space:]]Movie[[:space:]]Level[[:space:]]Dataset[[:space:]]for[[:space:]]Long[[:space:]]Video[[:space:]]Generation/b94cdb85-d55c-4785-9574-84a773e8d39d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mr.[[:space:]]DETR_[[:space:]]Instructive[[:space:]]Multi-Route[[:space:]]Training[[:space:]]for[[:space:]]Detection[[:space:]]Transformers/c8af3d9b-f62a-4e9b-a50f-9d6a49aa5823_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MuTri_[[:space:]]Multi-view[[:space:]]Tri-alignment[[:space:]]for[[:space:]]OCT[[:space:]]to[[:space:]]OCTA[[:space:]]3D[[:space:]]Image[[:space:]]Translation/3360afc4-6a1c-40c7-8579-562d36d32db4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Granularity[[:space:]]Class[[:space:]]Prototype[[:space:]]Topology[[:space:]]Distillation[[:space:]]for[[:space:]]Class-Incremental[[:space:]]Source-Free[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/f66b1fde-5282-42ec-a7c6-6d77dfe68a44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Group[[:space:]]Proportional[[:space:]]Representations[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Models/d798f629-1742-470a-8b34-daa903c01895_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Label[[:space:]]Prototype[[:space:]]Visual[[:space:]]Spatial[[:space:]]Search[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/65eff5b9-8eab-4fef-9461-c37095b19342_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Layer[[:space:]]Visual[[:space:]]Feature[[:space:]]Fusion[[:space:]]in[[:space:]]Multimodal[[:space:]]LLMs_[[:space:]]Methods,[[:space:]]Analysis,[[:space:]]and[[:space:]]Best[[:space:]]Practices/58e6103d-ff55-4865-90aa-f53cbfa8c551_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Modal[[:space:]]Aerial-Ground[[:space:]]Cross-View[[:space:]]Place[[:space:]]Recognition[[:space:]]with[[:space:]]Neural[[:space:]]ODEs/a3ef34a2-da20-42cd-9ce3-173f9564ad2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Modal[[:space:]]Contrastive[[:space:]]Masked[[:space:]]Autoencoders_[[:space:]]A[[:space:]]Two-Stage[[:space:]]Progressive[[:space:]]Pre-training[[:space:]]Approach[[:space:]]for[[:space:]]RGBD[[:space:]]Datasets/2860d8ed-ff38-4bde-8a34-5148c8747d03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Modal[[:space:]]Synergistic[[:space:]]Implicit[[:space:]]Image[[:space:]]Enhancement[[:space:]]for[[:space:]]Efficient[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation/ba4d661d-4b76-4508-be16-37bf6a411d28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Resolution[[:space:]]Pathology-Language[[:space:]]Pre-training[[:space:]]Model[[:space:]]with[[:space:]]Text-Guided[[:space:]]Visual[[:space:]]Representation/a1372db4-18b9-4a63-8835-259cd9b88763_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Scale[[:space:]]Neighborhood[[:space:]]Occupancy[[:space:]]Masked[[:space:]]Autoencoder[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]in[[:space:]]LiDAR[[:space:]]Point[[:space:]]Clouds/c1355a3a-5b51-4a58-9750-aa17cf5261d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Sensor[[:space:]]Object[[:space:]]Anomaly[[:space:]]Detection_[[:space:]]Unifying[[:space:]]Appearance,[[:space:]]Geometry,[[:space:]]and[[:space:]]Internal[[:space:]]Properties/e21933d2-56b8-4a95-ba9e-dee5731808ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-View[[:space:]]Pose-Agnostic[[:space:]]Change[[:space:]]Localization[[:space:]]with[[:space:]]Zero[[:space:]]Labels/24a301db-8aa2-4356-bf93-c7e01d4c7955_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-focal[[:space:]]Conditioned[[:space:]]Latent[[:space:]]Diffusion[[:space:]]for[[:space:]]Person[[:space:]]Image[[:space:]]Synthesis/a83cb6af-6baa-40cb-92b8-a97d92e4f738_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-modal[[:space:]]Contrastive[[:space:]]Learning[[:space:]]with[[:space:]]Negative[[:space:]]Sampling[[:space:]]Calibration[[:space:]]for[[:space:]]Phenotypic[[:space:]]Drug[[:space:]]Discovery/5625be24-a6d8-4187-a306-6e2e5e7e0620_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-modal[[:space:]]Knowledge[[:space:]]Distillation-based[[:space:]]Human[[:space:]]Trajectory[[:space:]]Forecasting/9c907734-c291-4eab-86c6-02181c988f72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-modal[[:space:]]Medical[[:space:]]Diagnosis[[:space:]]via[[:space:]]Large-small[[:space:]]Model[[:space:]]Collaboration/4e6dce4c-0787-480c-80e0-a7af364390a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-modal[[:space:]]Topology-embedded[[:space:]]Graph[[:space:]]Learning[[:space:]]for[[:space:]]Spatially[[:space:]]Resolved[[:space:]]Genes[[:space:]]Prediction[[:space:]]from[[:space:]]Pathology[[:space:]]Images[[:space:]]with[[:space:]]Prior[[:space:]]Gene[[:space:]]Similarity[[:space:]]Information/00bc1419-63f2-43ea-86b0-40bd6a90022d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-modal[[:space:]]Vision[[:space:]]Pre-training[[:space:]]for[[:space:]]Medical[[:space:]]Image[[:space:]]Analysis/2bdcd4b1-2112-45e6-9279-1f826ba774c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-party[[:space:]]Collaborative[[:space:]]Attention[[:space:]]Control[[:space:]]for[[:space:]]Image[[:space:]]Customization/177a98a9-39ae-4048-9623-9202d392360a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-subject[[:space:]]Open-set[[:space:]]Personalization[[:space:]]in[[:space:]]Video[[:space:]]Generation/be63cff0-29d5-4070-9c64-ac4ff58772c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-view[[:space:]]Reconstruction[[:space:]]via[[:space:]]SfM-guided[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/c65d6087-85ad-428e-9026-f49493e3dcf0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MultiGO_[[:space:]]Towards[[:space:]]Multi-level[[:space:]]Geometry[[:space:]]Learning[[:space:]]for[[:space:]]Monocular[[:space:]]3D[[:space:]]Textured[[:space:]]Human[[:space:]]Reconstruction/3eca82a2-58fa-4f1d-8330-eea738ed9063_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OSDFace_[[:space:]]One-Step[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Face[[:space:]]Restoration/1f53a280-0134-4654-b52d-d587c0851c41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OSLoPrompt_[[:space:]]Bridging[[:space:]]Low-Supervision[[:space:]]Challenges[[:space:]]and[[:space:]]Open-Set[[:space:]]Domain[[:space:]]Generalization[[:space:]]in[[:space:]]CLIP/c25cdcea-e372-4470-8e56-509c5a3db52d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OSMamba_[[:space:]]Omnidirectional[[:space:]]Spectral[[:space:]]Mamba[[:space:]]with[[:space:]]Dual-Domain[[:space:]]Prior[[:space:]]Generator[[:space:]]for[[:space:]]Exposure[[:space:]]Correction/d8b4b8b9-a830-461a-8c32-66e274fbdd8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OSV_[[:space:]]One[[:space:]]Step[[:space:]]is[[:space:]]Enough[[:space:]]for[[:space:]]High-Quality[[:space:]]Image[[:space:]]to[[:space:]]Video[[:space:]]Generation/164be1bc-9556-4fb3-9e86-4b279e901686_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OVO-Bench_[[:space:]]How[[:space:]]Far[[:space:]]is[[:space:]]Your[[:space:]]Video-LLMs[[:space:]]from[[:space:]]Real-World[[:space:]]Online[[:space:]]Video[[:space:]]Understanding_/fd77a437-e99e-4040-94de-ef1d9bc36179_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OW-OVD_[[:space:]]Unified[[:space:]]Open[[:space:]]World[[:space:]]and[[:space:]]Open[[:space:]]Vocabulary[[:space:]]Object[[:space:]]Detection/0eec5a88-9587-43d3-a7d4-26ac0e09d927_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Object[[:space:]]Detection[[:space:]]using[[:space:]]Event[[:space:]]Camera_[[:space:]]A[[:space:]]MoE[[:space:]]Heat[[:space:]]Conduction[[:space:]]based[[:space:]]Detector[[:space:]]and[[:space:]]A[[:space:]]New[[:space:]]Benchmark[[:space:]]Dataset/1ec7f5cb-f521-4c84-b7b4-7504b783f3e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Object-Centric[[:space:]]Prompt-Driven[[:space:]]Vision-Language-Action[[:space:]]Model[[:space:]]for[[:space:]]Robotic[[:space:]]Manipulation/2b3d2731-29bf-4fe5-a556-07cbb31fd742_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Object-Shot[[:space:]]Enhanced[[:space:]]Grounding[[:space:]]Network[[:space:]]for[[:space:]]Egocentric[[:space:]]Video/cef57492-dc63-4ddb-9c34-f275ae7f5a12_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Object-aware[[:space:]]Sound[[:space:]]Source[[:space:]]Localization[[:space:]]via[[:space:]]Audio-Visual[[:space:]]Scene[[:space:]]Understanding/c6118b29-9fd0-4c0e-82ff-626a5de37c61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ObjectMover_[[:space:]]Generative[[:space:]]Object[[:space:]]Movement[[:space:]]with[[:space:]]Video[[:space:]]Prior/30279836-e12f-449e-82ef-79ffe18be3fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OccMamba_[[:space:]]Semantic[[:space:]]Occupancy[[:space:]]Prediction[[:space:]]with[[:space:]]State[[:space:]]Space[[:space:]]Models/6cdd5490-e841-47b9-81f6-b9aa357887a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Occlusion-aware[[:space:]]Text-Image-Point[[:space:]]Cloud[[:space:]]Pretraining[[:space:]]for[[:space:]]Open-World[[:space:]]3D[[:space:]]Object[[:space:]]Recognition/73b16b04-6826-4e05-8352-a4b07b235461_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Octopus_[[:space:]]Alleviating[[:space:]]Hallucination[[:space:]]via[[:space:]]Dynamic[[:space:]]Contrastive[[:space:]]Decoding/ee076920-0b97-4ee1-ba36-19a9830c4fd0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Odd-One-Out_[[:space:]]Anomaly[[:space:]]Detection[[:space:]]by[[:space:]]Comparing[[:space:]]with[[:space:]]Neighbors/aea02b8e-3b61-4cb6-b631-3b24b77a578c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OffsetOPT_[[:space:]]Explicit[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]without[[:space:]]Normals/ee30e813-30ff-4e72-b940-992a57326174_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Olympus_[[:space:]]A[[:space:]]Universal[[:space:]]Task[[:space:]]Router[[:space:]]for[[:space:]]Computer[[:space:]]Vision[[:space:]]Tasks/7bb8f602-56a6-4698-a1e5-8252ba98873c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Omni-ID_[[:space:]]Holistic[[:space:]]Identity[[:space:]]Representation[[:space:]]Designed[[:space:]]for[[:space:]]Generative[[:space:]]Tasks/01c29e21-21c8-4c25-840b-9401db0971e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Omni-RGPT_[[:space:]]Unifying[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Region-level[[:space:]]Understanding[[:space:]]via[[:space:]]Token[[:space:]]Marks/13b7deff-faff-4529-bb7a-2fa6601f895a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Omni-Scene_[[:space:]]Omni-Gaussian[[:space:]]Representation[[:space:]]for[[:space:]]Ego-Centric[[:space:]]Sparse-View[[:space:]]Scene[[:space:]]Reconstruction/3915ca3f-b934-42f0-9d15-eb1feff5eb97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniDocBench_[[:space:]]Benchmarking[[:space:]]Diverse[[:space:]]PDF[[:space:]]Document[[:space:]]Parsing[[:space:]]with[[:space:]]Comprehensive[[:space:]]Annotations/c2f97f8c-4f05-4b67-9a5c-e556bab8b8fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniDrive_[[:space:]]A[[:space:]]Holistic[[:space:]]Vision-Language[[:space:]]Dataset[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving[[:space:]]with[[:space:]]Counterfactual[[:space:]]Reasoning/381c574d-2d0d-46d6-82a0-4a845117ca49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniFlow_[[:space:]]Any-to-Any[[:space:]]Generation[[:space:]]with[[:space:]]Multi-Modal[[:space:]]Rectified[[:space:]]Flows/2fa47e03-0cfc-4b8d-a066-5d6ba35480ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniGen_[[:space:]]Unified[[:space:]]Image[[:space:]]Generation/63a64e76-af83-4941-8133-701bdf22d04a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniGuard_[[:space:]]Hybrid[[:space:]]Manipulation[[:space:]]Localization[[:space:]]via[[:space:]]Augmented[[:space:]]Versatile[[:space:]]Deep[[:space:]]Image[[:space:]]Watermarking/900e313c-8954-49c3-bb99-5ea17c3e52ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniMMI_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Multi-modal[[:space:]]Interaction[[:space:]]Benchmark[[:space:]]in[[:space:]]Streaming[[:space:]]Video[[:space:]]Contexts/54cc6a60-9222-4434-8f2a-7fddd5dda439_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniManip_[[:space:]]Towards[[:space:]]General[[:space:]]Robotic[[:space:]]Manipulation[[:space:]]via[[:space:]]Object-Centric[[:space:]]Interaction[[:space:]]Primitives[[:space:]]as[[:space:]]Spatial[[:space:]]Constraints/e993d05a-0941-457c-a4d4-4a73cd52c119_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniSplat_[[:space:]]Taming[[:space:]]Feed-Forward[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Omnidirectional[[:space:]]Images[[:space:]]with[[:space:]]Editable[[:space:]]Capabilities/97b11af0-089e-4dd6-b1b6-3ace07d07ae1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniStereo_[[:space:]]Real-time[[:space:]]Omnidireactional[[:space:]]Depth[[:space:]]Estimation[[:space:]]with[[:space:]]Multiview[[:space:]]Fisheye[[:space:]]Cameras/b71aa929-847f-477a-abd0-9fb8ae8a4c4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniStyle_[[:space:]]Filtering[[:space:]]High[[:space:]]Quality[[:space:]]Style[[:space:]]Transfer[[:space:]]Data[[:space:]]at[[:space:]]Scale/0faa52f6-da0d-4f4a-a712-c671d6958dec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Omnia[[:space:]]de[[:space:]]EgoTempo_[[:space:]]Benchmarking[[:space:]]Temporal[[:space:]]Understanding[[:space:]]of[[:space:]]Multi-Modal[[:space:]]LLMs[[:space:]]in[[:space:]]Egocentric[[:space:]]Videos/d1d822d5-04cb-453e-9c98-330331afcf5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Omnidirectional[[:space:]]Multi-Object[[:space:]]Tracking/750188f0-23b6-4a03-a966-6991f89d6979_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On[[:space:]]Denoising[[:space:]]Walking[[:space:]]Videos[[:space:]]for[[:space:]]Gait[[:space:]]Recognition/06fdd4d6-9460-444a-ae49-70086d6eabff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On[[:space:]]the[[:space:]]Consistency[[:space:]]of[[:space:]]Video[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]in[[:space:]]Temporal[[:space:]]Comprehension/048ce3d3-f5af-498d-9251-263c126a0f24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On[[:space:]]the[[:space:]]Generalization[[:space:]]of[[:space:]]Handwritten[[:space:]]Text[[:space:]]Recognition[[:space:]]Models/e6119616-bfed-4041-a055-5bcc5c86887f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On[[:space:]]the[[:space:]]Out-Of-Distribution[[:space:]]Generalization[[:space:]]of[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models/8bb84a02-6234-4de0-a53e-63ff9ccac5f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On[[:space:]]the[[:space:]]Zero-shot[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models_[[:space:]]A[[:space:]]Truly[[:space:]]Zero-shot[[:space:]]and[[:space:]]Training-free[[:space:]]Approach/fb120035-926f-4721-be81-96f5982afe5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On-Device[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]of[[:space:]]Low-Latency[[:space:]]Monocular[[:space:]]Depth[[:space:]]from[[:space:]]Only[[:space:]]Events/149fa0e5-0021-4314-8d49-14443c5f26a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Once-Tuning-Multiple-Variants_[[:space:]]Tuning[[:space:]]Once[[:space:]]and[[:space:]]Expanded[[:space:]]as[[:space:]]Multiple[[:space:]]Vision-Language[[:space:]]Model[[:space:]]Variants/c96bb747-b59f-480c-b4e2-ad5984ad0286_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One[[:space:]]Diffusion[[:space:]]to[[:space:]]Generate[[:space:]]Them[[:space:]]All/39f2b67a-8dca-4c74-afb0-f03cafb48718_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One[[:space:]]Model[[:space:]]for[[:space:]]ALL_[[:space:]]Low-Level[[:space:]]Task[[:space:]]Interaction[[:space:]]Is[[:space:]]a[[:space:]]Key[[:space:]]to[[:space:]]Task-Agnostic[[:space:]]Image[[:space:]]Fusion/7330d623-3a08-4f74-a6bc-b193043f6a4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One[[:space:]]is[[:space:]]Plenty_[[:space:]]A[[:space:]]Polymorphic[[:space:]]Feature[[:space:]]Interpreter[[:space:]]for[[:space:]]Immutable[[:space:]]Heterogeneous[[:space:]]Collaborative[[:space:]]Perception/68f7077a-bdce-464d-bdb3-8f3cdc56629f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One-Minute[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Test-Time[[:space:]]Training/85b949af-0141-4071-bd55-d06f27c00c00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One-Step[[:space:]]Event-Driven[[:space:]]High-Speed[[:space:]]Autofocus/95c4e4e2-58e5-478a-93c0-9df909a54839_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One-Way[[:space:]]Ticket_[[:space:]]Time-Independent[[:space:]]Unified[[:space:]]Encoder[[:space:]]for[[:space:]]Distilling[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/ac1c83ad-bcc1-45af-9c0a-a97fd945d78e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One-for-More_[[:space:]]Continual[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Anomaly[[:space:]]Detection/892dae6a-bf04-4ad9-ae18-435e0c0e29b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One-shot[[:space:]]3D[[:space:]]Object[[:space:]]Canonicalization[[:space:]]based[[:space:]]on[[:space:]]Geometric[[:space:]]and[[:space:]]Semantic[[:space:]]Consistency/3782866b-49c0-422d-8cd8-fdf103853cff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One2Any_[[:space:]]One-Reference[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation[[:space:]]for[[:space:]]Any[[:space:]]Object/c397ac7e-9148-43c1-9790-b52d2ebc44a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Online[[:space:]]Task-Free[[:space:]]Continual[[:space:]]Learning[[:space:]]via[[:space:]]Dynamic[[:space:]]Expansionable[[:space:]]Memory[[:space:]]Distribution/2cfed467-fdcb-4c0f-b2d0-ae9911448aed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Online[[:space:]]Video[[:space:]]Understanding_[[:space:]]OVBench[[:space:]]and[[:space:]]VideoChat-Online/fe148595-0571-4e69-9ade-91cc32b5b1f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OnlineAnySeg_[[:space:]]Online[[:space:]]Zero-Shot[[:space:]]3D[[:space:]]Segmentation[[:space:]]by[[:space:]]Visual[[:space:]]Foundation[[:space:]]Model[[:space:]]Guided[[:space:]]2D[[:space:]]Mask[[:space:]]Merging/2ebdb5b0-7893-47c9-a0a1-ee31fc2ea499_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Open[[:space:]]Ad-hoc[[:space:]]Categorization[[:space:]]with[[:space:]]Contextualized[[:space:]]Feature[[:space:]]Learning/33489ea7-a360-4626-8851-19431b042cd1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Open[[:space:]]Set[[:space:]]Label[[:space:]]Shift[[:space:]]with[[:space:]]Test[[:space:]]Time[[:space:]]Out-of-Distribution[[:space:]]Reference/62dafbab-6030-4613-b956-0f9c1ad74993_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Open-Canopy_[[:space:]]Towards[[:space:]]Very[[:space:]]High[[:space:]]Resolution[[:space:]]Forest[[:space:]]Monitoring/575c5c03-fe20-4401-942a-b3eb118f493e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Open-Vocabulary[[:space:]]Functional[[:space:]]3D[[:space:]]Scene[[:space:]]Graphs[[:space:]]for[[:space:]]Real-World[[:space:]]Indoor[[:space:]]Spaces/61972c7a-d07f-4089-b601-bf6f9f7f2ea7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Open-World[[:space:]]Amodal[[:space:]]Appearance[[:space:]]Completion/bf3404cf-4f7e-47a5-b4a6-f6416083a542_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Open-World[[:space:]]Objectness[[:space:]]Modeling[[:space:]]Unifies[[:space:]]Novel[[:space:]]Object[[:space:]]Detection/f1eb016c-6e97-4c60-82d7-b9dfe3137560_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OpenHumanVid_[[:space:]]A[[:space:]]Large-Scale[[:space:]]High-Quality[[:space:]]Dataset[[:space:]]for[[:space:]]Enhancing[[:space:]]Human-Centric[[:space:]]Video[[:space:]]Generation/998be6f6-773c-45bd-b3d3-ad8066822b83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OpenING_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Benchmark[[:space:]]for[[:space:]]Judging[[:space:]]Open-ended[[:space:]]Interleaved[[:space:]]Image-Text[[:space:]]Generation/00b52ae3-9e6d-4fc5-bc84-3f5001aeccbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OpenMIBOOD_[[:space:]]Open[[:space:]]Medical[[:space:]]Imaging[[:space:]]Benchmarks[[:space:]]for[[:space:]]Out-Of-Distribution[[:space:]]Detection/6bac23c2-9091-4794-a3b6-66db01a71a7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OpenSDI_[[:space:]]Spotting[[:space:]]Diffusion-Generated[[:space:]]Images[[:space:]]in[[:space:]]the[[:space:]]Open[[:space:]]World/2cd568d7-c9b1-49e7-97bb-6031d30775f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Opportunistic[[:space:]]Single-Photon[[:space:]]Time[[:space:]]of[[:space:]]Flight/2b23d974-9eb1-4829-a21c-2ff9fe8b9c14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Optical-Flow[[:space:]]Guided[[:space:]]Prompt[[:space:]]Optimization[[:space:]]for[[:space:]]Coherent[[:space:]]Video[[:space:]]Generation/0533ad04-4e42-407b-a8ab-95e3d6387a53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OpticalNet_[[:space:]]An[[:space:]]Optical[[:space:]]Imaging[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]Beyond[[:space:]]the[[:space:]]Diffraction[[:space:]]Limit/5b60e093-1635-437f-8a0a-0f0e6e5c2e2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Optimal[[:space:]]Transport-Guided[[:space:]]Source-Free[[:space:]]Adaptation[[:space:]]for[[:space:]]Face[[:space:]]Anti-Spoofing/e314b1b5-d65e-4f4e-9531-3f48989c3084_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Optimizing[[:space:]]for[[:space:]]the[[:space:]]Shortest[[:space:]]Path[[:space:]]in[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Model/243a450f-2864-496d-8498-20dac8940e08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Optimus-2_[[:space:]]Multimodal[[:space:]]Minecraft[[:space:]]Agent[[:space:]]with[[:space:]]Goal-Observation-Action[[:space:]]Conditioned[[:space:]]Policy/f74f85db-6e8a-4b9d-a0c2-5138ed2a23f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OralXrays-9_[[:space:]]Towards[[:space:]]Hospital-Scale[[:space:]]Panoramic[[:space:]]X-ray[[:space:]]Anomaly[[:space:]]Detection[[:space:]]via[[:space:]]Personalized[[:space:]]Multi-Object[[:space:]]Query-Aware[[:space:]]Mining/39a94426-c516-4610-b342-efd63087b1b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Order-One[[:space:]]Rolling[[:space:]]Shutter[[:space:]]Cameras/8ed549fa-f401-431d-8ebb-355540ad8478_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Order-Robust[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning_[[:space:]]Graph-Driven[[:space:]]Dynamic[[:space:]]Similarity[[:space:]]Grouping/fa4ce536-151f-4a7d-bc0e-470ce40b0b3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ouroboros3D_[[:space:]]Image-to-3D[[:space:]]Generation[[:space:]]via[[:space:]]3D-aware[[:space:]]Recursive[[:space:]]Diffusion/52aaef83-3169-40e3-8075-74d97387cf89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OverLoCK_[[:space:]]An[[:space:]]Overview-first-Look-Closely-next[[:space:]]ConvNet[[:space:]]with[[:space:]]Context-Mixing[[:space:]]Dynamic[[:space:]]Kernels/6dae3b32-d997-4ec5-b63f-f3db7139f971_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Overcoming[[:space:]]Shortcut[[:space:]]Problem[[:space:]]in[[:space:]]VLM[[:space:]]for[[:space:]]Robust[[:space:]]Out-of-Distribution[[:space:]]Detection/3da23e83-5657-4685-84e7-0747989795c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PACT_[[:space:]]Pruning[[:space:]]and[[:space:]]Clustering-Based[[:space:]]Token[[:space:]]Reduction[[:space:]]for[[:space:]]Faster[[:space:]]Visual[[:space:]]Language[[:space:]]Models/7f7da60e-2bb5-4b48-88d8-dfd80e2b742c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PARC_[[:space:]]A[[:space:]]Quantitative[[:space:]]Framework[[:space:]]Uncovering[[:space:]]the[[:space:]]Symmetries[[:space:]]within[[:space:]]Vision[[:space:]]Language[[:space:]]Models/891f26d3-09d6-4eb3-bac3-760662ea1547_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PAVE_[[:space:]]Patching[[:space:]]and[[:space:]]Adapting[[:space:]]Video[[:space:]]Large[[:space:]]Language[[:space:]]Models/fdbd73e8-c052-465d-90b0-bc236be2f444_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PBR-NeRF_[[:space:]]Inverse[[:space:]]Rendering[[:space:]]with[[:space:]]Physics-Based[[:space:]]Neural[[:space:]]Fields/2256078d-467d-4452-aebc-8e0734315b11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PCDreamer_[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion[[:space:]]Through[[:space:]]Multi-view[[:space:]]Diffusion[[:space:]]Priors/b3611bb5-dc66-4d7d-86c4-43f57daf8b05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PCM[[:space:]]_[[:space:]]Picard[[:space:]]Consistency[[:space:]]Model[[:space:]]for[[:space:]]Fast[[:space:]]Parallel[[:space:]]Sampling[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/f990dc5e-e9ef-4ca8-a142-436eacd432d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PDFactor_[[:space:]]Learning[[:space:]]Tri-Perspective[[:space:]]View[[:space:]]Policy[[:space:]]Diffusion[[:space:]]Field[[:space:]]for[[:space:]]Multi-Task[[:space:]]Robotic[[:space:]]Manipulation/278e9816-f4cd-45fe-a497-dbeeff1d2ec4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PEACE_[[:space:]]Empowering[[:space:]]Geologic[[:space:]]Map[[:space:]]Holistic[[:space:]]Understanding[[:space:]]with[[:space:]]MLLMs/0090d196-b3ab-4d5e-bd0a-18da0215ff48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PEER[[:space:]]Pressure_[[:space:]]Model-to-Model[[:space:]]Regularization[[:space:]]for[[:space:]]Single[[:space:]]Source[[:space:]]Domain[[:space:]]Generalization/47d1f8af-f2fc-4822-905f-796daedf8fb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PERSE_[[:space:]]Personalized[[:space:]]3D[[:space:]]Generative[[:space:]]Avatars[[:space:]]from[[:space:]]A[[:space:]]Single[[:space:]]Portrait/1f628a7e-aafa-4276-baab-32f9773ad341_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PGC_[[:space:]]Physics-Based[[:space:]]Gaussian[[:space:]]Cloth[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Pose/58ebb58c-0fe8-4996-abaa-65774f1d5e93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PHGC_[[:space:]]Procedural[[:space:]]Heterogeneous[[:space:]]Graph[[:space:]]Completion[[:space:]]for[[:space:]]Natural[[:space:]]Language[[:space:]]Task[[:space:]]Verification[[:space:]]in[[:space:]]Egocentric[[:space:]]Videos/244e7873-2ef1-46f6-9629-bb9167c2dca4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PI-HMR_[[:space:]]Towards[[:space:]]Robust[[:space:]]In-bed[[:space:]]Temporal[[:space:]]Human[[:space:]]Shape[[:space:]]Reconstruction[[:space:]]with[[:space:]]Contact[[:space:]]Pressure[[:space:]]Sensing/0f4e8682-34be-4f51-8dac-a6651a5fb384_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PIAD_[[:space:]]Pose[[:space:]]and[[:space:]]Illumination[[:space:]]agnostic[[:space:]]Anomaly[[:space:]]Detection/cc542083-e02d-4338-9d98-c88c39ba3de2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PICD_[[:space:]]Versatile[[:space:]]Perceptual[[:space:]]Image[[:space:]]Compression[[:space:]]with[[:space:]]Diffusion[[:space:]]Rendering/11144333-f344-4d28-b36c-6d6c5218dc9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PICO_[[:space:]]Reconstructing[[:space:]]3D[[:space:]]People[[:space:]]In[[:space:]]Contact[[:space:]]with[[:space:]]Objects/8dbab643-c765-440f-ae6a-e93ec0c621fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PIDLoc_[[:space:]]Cross-View[[:space:]]Pose[[:space:]]Optimization[[:space:]]Network[[:space:]]Inspired[[:space:]]by[[:space:]]PID[[:space:]]Controllers/dbe1be5d-172e-44d9-bcfc-5079019591f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PIDSR_[[:space:]]Complementary[[:space:]]Polarized[[:space:]]Image[[:space:]]Demosaicing[[:space:]]and[[:space:]]Super-Resolution/3a6f368a-5534-4faf-8844-7e759660d5b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PLeaS[[:space:]]-[[:space:]]Merging[[:space:]]Models[[:space:]]with[[:space:]]Permutations[[:space:]]and[[:space:]]Least[[:space:]]Squares/92dd9250-15de-4856-8fea-f042c7e8eecb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PMA_[[:space:]]Towards[[:space:]]Parameter-Efficient[[:space:]]Point[[:space:]]Cloud[[:space:]]Understanding[[:space:]]via[[:space:]]Point[[:space:]]Mamba[[:space:]]Adapter/c2e98814-5096-44f9-a81b-1bc9ce0c6c28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PMNI_[[:space:]]Pose-free[[:space:]]Multi-view[[:space:]]Normal[[:space:]]Integration[[:space:]]for[[:space:]]Reflective[[:space:]]and[[:space:]]Textureless[[:space:]]Surface[[:space:]]Reconstruction/a8da9ead-4360-49e9-8fac-6640016f5fff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PO3AD_[[:space:]]Predicting[[:space:]]Point[[:space:]]Offsets[[:space:]]toward[[:space:]]Better[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Anomaly[[:space:]]Detection/8436db93-fdb0-40fb-9387-e16535fe0768_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/POMP_[[:space:]]Physics-consistent[[:space:]]Motion[[:space:]]Generative[[:space:]]Model[[:space:]]through[[:space:]]Phase[[:space:]]Manifolds/0aa9451b-65fe-4d03-86d6-7bbb541c8a38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/POPEN_[[:space:]]Preference-Based[[:space:]]Optimization[[:space:]]and[[:space:]]Ensemble[[:space:]]for[[:space:]]LVLM-Based[[:space:]]Reasoning[[:space:]]Segmentation/44a0bbcf-cff3-4a72-90d6-df07026ffcfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/POSTA_[[:space:]]A[[:space:]]Go-to[[:space:]]Framework[[:space:]]for[[:space:]]Customized[[:space:]]Artistic[[:space:]]Poster[[:space:]]Generation/d3076757-d2c5-4500-8ba7-8c5093daf54a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/POT_[[:space:]]Prototypical[[:space:]]Optimal[[:space:]]Transport[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/62f1107d-30c8-4ab1-8fb4-acd8c2119a74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/POp-GS_[[:space:]]Next[[:space:]]Best[[:space:]]View[[:space:]]in[[:space:]]3D-Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]P-Optimality/6c18ef63-531c-4d64-ac45-b3fddad1ddfd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PQPP_[[:space:]]A[[:space:]]Joint[[:space:]]Benchmark[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Prompt[[:space:]]and[[:space:]]Query[[:space:]]Performance[[:space:]]Prediction/92393a70-b9c5-431c-948d-c986b59932cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PRaDA_[[:space:]]Projective[[:space:]]Radial[[:space:]]Distortion[[:space:]]Averaging/2dbcae95-43c3-4b23-8948-426169a2bdbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PS-Diffusion_[[:space:]]Photorealistic[[:space:]]Subject-Driven[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Disentangled[[:space:]]Control[[:space:]]and[[:space:]]Attention/e7808b12-21a1-4de1-95e9-90f108091990_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PS-EIP_[[:space:]]Robust[[:space:]]Photometric[[:space:]]Stereo[[:space:]]Based[[:space:]]on[[:space:]]Event[[:space:]]Interval[[:space:]]Profile/0867d3e2-87ee-43b0-900a-d729d056956c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PSA-SSL_[[:space:]]Pose[[:space:]]and[[:space:]]Size-aware[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]on[[:space:]]LiDAR[[:space:]]Point[[:space:]]Clouds/dc943ad9-b7d8-4337-9471-37907832bd9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PSBD_[[:space:]]Prediction[[:space:]]Shift[[:space:]]Uncertainty[[:space:]]Unlocks[[:space:]]Backdoor[[:space:]]Detection/4feea4fa-2e1b-4f34-97ce-fa9b48a0e741_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PSHuman_[[:space:]]Photorealistic[[:space:]]Single-image[[:space:]]3D[[:space:]]Human[[:space:]]Reconstruction[[:space:]]using[[:space:]]Cross-Scale[[:space:]]Multiview[[:space:]]Diffusion[[:space:]]and[[:space:]]Explicit[[:space:]]Remeshing/674cc41e-2aac-42de-a649-e2b21f22d4e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PTDiffusion_[[:space:]]Free[[:space:]]Lunch[[:space:]]for[[:space:]]Generating[[:space:]]Optical[[:space:]]Illusion[[:space:]]Hidden[[:space:]]Pictures[[:space:]]with[[:space:]]Phase-Transferred[[:space:]]Diffusion[[:space:]]Model/48fd3635-4df8-4a90-81c4-8dcba983626a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PUP[[:space:]]3D-GS_[[:space:]]Principled[[:space:]]Uncertainty[[:space:]]Pruning[[:space:]]for[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/10cba95b-d3ab-459d-91b7-2ba7d33715fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PURA_[[:space:]]Parameter[[:space:]]Update-Recovery[[:space:]]Test-Time[[:space:]]Adaption[[:space:]]for[[:space:]]RGB-T[[:space:]]Tracking/947aeca0-b3c7-4b73-8302-ff5a7fd6c7e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PVC_[[:space:]]Progressive[[:space:]]Visual[[:space:]]Token[[:space:]]Compression[[:space:]]for[[:space:]]Unified[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Processing[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/b2a57b69-4b09-41e0-ae3c-258c99b641b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Paint[[:space:]]by[[:space:]]Inpaint_[[:space:]]Learning[[:space:]]to[[:space:]]Add[[:space:]]Image[[:space:]]Objects[[:space:]]by[[:space:]]Removing[[:space:]]Them[[:space:]]First/c9f8639f-3f64-49d1-8a81-50b810780c3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PanDA_[[:space:]]Towards[[:space:]]Panoramic[[:space:]]Depth[[:space:]]Anything[[:space:]]with[[:space:]]Unlabeled[[:space:]]Panoramas[[:space:]]and[[:space:]]Mobius[[:space:]]Spatial[[:space:]]Augmentation/37fa5da7-896f-4193-a74d-3939aac05fa9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PanSplat_[[:space:]]4K[[:space:]]Panorama[[:space:]]Synthesis[[:space:]]with[[:space:]]Feed-Forward[[:space:]]Gaussian[[:space:]]Splatting/836bf3c9-5855-4e47-9751-f306ec81b690_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PanoGS_[[:space:]]Gaussian-based[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]for[[:space:]]3D[[:space:]]Open[[:space:]]Vocabulary[[:space:]]Scene[[:space:]]Understanding/123721ee-951a-4a6a-a80f-6a011ac91def_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Panorama[[:space:]]Generation[[:space:]]From[[:space:]]NFoV[[:space:]]Image[[:space:]]Done[[:space:]]Right/c0f42e0e-32c5-4aee-92d9-63f8fc3a1786_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ParaHome_[[:space:]]Parameterizing[[:space:]]Everyday[[:space:]]Home[[:space:]]Activities[[:space:]]Towards[[:space:]]3D[[:space:]]Generative[[:space:]]Modeling[[:space:]]of[[:space:]]Human-Object[[:space:]]Interactions/22287db8-4cf3-42bd-9396-440c03e26639_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Parallel[[:space:]]Sequence[[:space:]]Modeling[[:space:]]via[[:space:]]Generalized[[:space:]]Spatial[[:space:]]Propagation[[:space:]]Network/b2ad4b19-425e-4153-a7a9-eab51ff2a540_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Parallelized[[:space:]]Autoregressive[[:space:]]Visual[[:space:]]Generation/34a3f5a8-b744-446b-b4e7-8bd469471b05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Parameter[[:space:]]Efficient[[:space:]]Mamba[[:space:]]Tuning[[:space:]]via[[:space:]]Projector-targeted[[:space:]]Diagonal-centric[[:space:]]Linear[[:space:]]Transformation/d7d69aa2-1a41-4ef4-a2aa-758d624162e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Parameter-efficient[[:space:]]Fine-tuning[[:space:]]in[[:space:]]Hyperspherical[[:space:]]Space[[:space:]]for[[:space:]]Open-vocabulary[[:space:]]Semantic[[:space:]]Segmentation/34b3c106-f028-46e1-aa69-9a8e58755b5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Parameterized[[:space:]]Blur[[:space:]]Kernel[[:space:]]Prior[[:space:]]Learning[[:space:]]for[[:space:]]Local[[:space:]]Motion[[:space:]]Deblurring/422dce86-8786-4d15-8349-ecd870bf9c68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Parametric[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion[[:space:]]for[[:space:]]Polygonal[[:space:]]Surface[[:space:]]Reconstruction/e3e0c32e-a97f-4db1-bca8-4faf5cb6cec1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PartGen_[[:space:]]Part-level[[:space:]]3D[[:space:]]Generation[[:space:]]and[[:space:]]Reconstruction[[:space:]]with[[:space:]]Multi-view[[:space:]]Diffusion[[:space:]]Models/47fb04da-4ed5-4ade-aa34-af576836193a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PartRM_[[:space:]]Modeling[[:space:]]Part-Level[[:space:]]Dynamics[[:space:]]with[[:space:]]Large[[:space:]]Cross-State[[:space:]]Reconstruction[[:space:]]Model/0ad92b0d-639a-4a60-93c4-45bd0ade58c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PassionSR_[[:space:]]Post-Training[[:space:]]Quantization[[:space:]]with[[:space:]]Adaptive[[:space:]]Scale[[:space:]]in[[:space:]]One-Step[[:space:]]Diffusion[[:space:]]based[[:space:]]Image[[:space:]]Super-Resolution/12edb9f7-ce33-4011-ae44-3a97b6f825f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Patch[[:space:]]Matters_[[:space:]]Training-free[[:space:]]Fine-grained[[:space:]]Image[[:space:]]Caption[[:space:]]Enhancement[[:space:]]via[[:space:]]Local[[:space:]]Perception/ad7b7d61-e53d-436e-98be-4d8c90eb49b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PatchDEMUX_[[:space:]]A[[:space:]]Certifiably[[:space:]]Robust[[:space:]]Framework[[:space:]]for[[:space:]]Multi-label[[:space:]]Classifiers[[:space:]]Against[[:space:]]Adversarial[[:space:]]Patches/b0e9e99e-06e6-468a-a8a3-673f58598de7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PatchDPO_[[:space:]]Patch-level[[:space:]]DPO[[:space:]]for[[:space:]]Finetuning-free[[:space:]]Personalized[[:space:]]Image[[:space:]]Generation/201fb178-5c6a-42cf-95fe-3e03de95821f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PatchGuard_[[:space:]]Adversarially[[:space:]]Robust[[:space:]]Anomaly[[:space:]]Detection[[:space:]]and[[:space:]]Localization[[:space:]]through[[:space:]]Vision[[:space:]]Transformers[[:space:]]and[[:space:]]Pseudo[[:space:]]Anomalies/6310723f-d72e-495d-87c7-6b59d06d7ca9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PatchVSR_[[:space:]]Breaking[[:space:]]Video[[:space:]]Diffusion[[:space:]]Resolution[[:space:]]Limits[[:space:]]with[[:space:]]Patch-wise[[:space:]]Video[[:space:]]Super-Resolution/7794bc7e-86d2-49a8-8adc-3430e39dd287_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pathways[[:space:]]on[[:space:]]the[[:space:]]Image[[:space:]]Manifold_[[:space:]]Image[[:space:]]Editing[[:space:]]via[[:space:]]Video[[:space:]]Generation/f625c957-1581-449f-8cd5-463f4552272c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Patient-Level[[:space:]]Anatomy[[:space:]]Meets[[:space:]]Scanning-Level[[:space:]]Physics_[[:space:]]Personalized[[:space:]]Federated[[:space:]]Low-Dose[[:space:]]CT[[:space:]]Denoising[[:space:]]Empowered[[:space:]]by[[:space:]]Large[[:space:]]Language[[:space:]]Model/b8f485b4-2d7f-474b-acf2-8dc4a4e3047c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pattern[[:space:]]Analogies_[[:space:]]Learning[[:space:]]to[[:space:]]Perform[[:space:]]Programmatic[[:space:]]Image[[:space:]]Edits[[:space:]]by[[:space:]]Analogy/86d3746e-1656-4dc5-ab96-b8f696f1d699_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pay[[:space:]]Attention[[:space:]]to[[:space:]]the[[:space:]]Foreground[[:space:]]in[[:space:]]Object-Centric[[:space:]]Learning/b4bf29e9-ae1b-4c69-a186-15cc519fb707_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PerLA_[[:space:]]Perceptive[[:space:]]3D[[:space:]]Language[[:space:]]Assistant/54882f4a-04ec-4ab3-b66a-a66c8e83acf5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Percept,[[:space:]]Memory,[[:space:]]and[[:space:]]Imagine_[[:space:]]World[[:space:]]Feature[[:space:]]Simulating[[:space:]]for[[:space:]]Open-Domain[[:space:]]Unknown[[:space:]]Object[[:space:]]Detection/b34e32e0-849d-4c9e-8711-6ce6de1dc89c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Perception[[:space:]]Tokens[[:space:]]Enhance[[:space:]]Visual[[:space:]]Reasoning[[:space:]]in[[:space:]]Multimodal[[:space:]]Language[[:space:]]Models/a1c25909-0376-4a01-a3cb-0e20c3229024_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Perceptual[[:space:]]Inductive[[:space:]]Bias[[:space:]]Is[[:space:]]What[[:space:]]You[[:space:]]Need[[:space:]]Before[[:space:]]Contrastive[[:space:]]Learning/cdaa1dd9-bb97-42b0-9c82-a5fe76697ec3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Perceptual[[:space:]]Video[[:space:]]Compression[[:space:]]with[[:space:]]Neural[[:space:]]Wrapping/4f847378-cfb2-4455-91dd-a70a0227cf83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Perceptually[[:space:]]Accurate[[:space:]]3D[[:space:]]Talking[[:space:]]Head[[:space:]]Generation_[[:space:]]New[[:space:]]Definitions,[[:space:]]Speech-Mesh[[:space:]]Representation,[[:space:]]and[[:space:]]Evaluation[[:space:]]Metrics/f0ca3209-a68f-41a0-a209-e41298a3de25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Period-LLM_[[:space:]]Extending[[:space:]]the[[:space:]]Periodic[[:space:]]Capability[[:space:]]of[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model/96cc4bcd-798e-44b6-b645-30ba105eaf18_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Person[[:space:]]De-reidentification_[[:space:]]A[[:space:]]Variation-guided[[:space:]]Identity[[:space:]]Shift[[:space:]]Modeling/1180818e-c94d-44fa-b5df-b1fc8919ac06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PersonaBooth_[[:space:]]Personalized[[:space:]]Text-to-Motion[[:space:]]Generation/62946df9-a553-4225-bd84-84b1acdb1a5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PersonaHOI_[[:space:]]Effortlessly[[:space:]]Improving[[:space:]]Face[[:space:]]Personalization[[:space:]]in[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Generation/00e10d7f-1169-4801-84c3-c1efaaab378c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Personalized[[:space:]]Preference[[:space:]]Fine-tuning[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/005a06ec-5d3e-419e-a758-dd73f721c5f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Perturb-and-Revise_[[:space:]]Flexible[[:space:]]3D[[:space:]]Editing[[:space:]]with[[:space:]]Generative[[:space:]]Trajectories/11693e39-cb59-4d3c-8f05-aea184580288_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PhD_[[:space:]]A[[:space:]]ChatGPT-Prompted[[:space:]]Visual[[:space:]]Hallucination[[:space:]]Evaluation[[:space:]]Dataset/0cae22af-0103-455d-a35f-1a510beabf2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Phoenix_[[:space:]]A[[:space:]]Motion-based[[:space:]]Self-Reflection[[:space:]]Framework[[:space:]]for[[:space:]]Fine-grained[[:space:]]Robotic[[:space:]]Action[[:space:]]Correction/497d038e-fa23-45f9-aee3-d3bfa187eaf5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PhyS-EdiT_[[:space:]]Physics-aware[[:space:]]Semantic[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Text[[:space:]]Description/7ee131d2-1c01-43c3-8f06-0c7615cc10d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PhyT2V_[[:space:]]LLM-Guided[[:space:]]Iterative[[:space:]]Self-Refinement[[:space:]]for[[:space:]]Physics-Grounded[[:space:]]Text-to-Video[[:space:]]Generation/d3a4c2ae-5aa4-4658-a693-f3535a28ce44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PhysAnimator_[[:space:]]Physics-Guided[[:space:]]Generative[[:space:]]Cartoon[[:space:]]Animation/d8df7b02-5bad-4fed-8afa-940086807a3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PhysGen3D_[[:space:]]Crafting[[:space:]]a[[:space:]]Miniature[[:space:]]Interactive[[:space:]]World[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/1bf0def3-a278-47bf-9ffd-13a93c05e545_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PhysVLM_[[:space:]]Enabling[[:space:]]Visual[[:space:]]Language[[:space:]]Models[[:space:]]to[[:space:]]Understand[[:space:]]Robotic[[:space:]]Physical[[:space:]]Reachability/14176105-f8ea-4fe1-b4ab-02889a03b21d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Physical[[:space:]]Plausibility-aware[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]via[[:space:]]Locomotion[[:space:]]Embodiment/88abcf25-51f5-40e1-9600-4a2e38b6cc9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PhysicsGen_[[:space:]]Can[[:space:]]Generative[[:space:]]Models[[:space:]]Learn[[:space:]]from[[:space:]]Images[[:space:]]to[[:space:]]Predict[[:space:]]Complex[[:space:]]Physical[[:space:]]Relations_/d8bec7dd-2170-41a3-a07a-e8a3813c711e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PillarHist_[[:space:]]A[[:space:]]Quantization-aware[[:space:]]Pillar[[:space:]]Feature[[:space:]]Encoder[[:space:]]based[[:space:]]on[[:space:]]Height-aware[[:space:]]Histogram/18874676-27a7-4b87-b788-f15decae2835_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pioneering[[:space:]]4-Bit[[:space:]]FP[[:space:]]Quantization[[:space:]]for[[:space:]]Diffusion[[:space:]]Models_[[:space:]]Mixup-Sign[[:space:]]Quantization[[:space:]]and[[:space:]]Timestep-Aware[[:space:]]Fine-Tuning/6042cf5b-516c-4fe8-bfad-ecfc854fc855_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pippo_[[:space:]]High-Resolution[[:space:]]Multi-View[[:space:]]Humans[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/2679a0a3-d3fd-49c2-aeba-9db847b48300_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pixel-aligned[[:space:]]RGB-NIR[[:space:]]Stereo[[:space:]]Imaging[[:space:]]and[[:space:]]Dataset[[:space:]]for[[:space:]]Robot[[:space:]]Vision/640cd516-44b5-47ce-aa6a-b65d88b8da0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pixel-level[[:space:]]and[[:space:]]Semantic-level[[:space:]]Adjustable[[:space:]]Super-resolution_[[:space:]]A[[:space:]]Dual-LoRA[[:space:]]Approach/6dc2236b-e40d-40ee-98e0-34b3414c88ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PlanarSplatting_[[:space:]]Accurate[[:space:]]Planar[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]in[[:space:]]3[[:space:]]Minutes/3af07a07-d377-440f-9ac8-47f29938e643_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Playing[[:space:]]the[[:space:]]Fool_[[:space:]]Jailbreaking[[:space:]]LLMs[[:space:]]and[[:space:]]Multimodal[[:space:]]LLMs[[:space:]]with[[:space:]]Out-of-Distribution[[:space:]]Strategy/013bcc47-e134-4c31-adf0-ced109b7aaa0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Plug-and-Play[[:space:]]Interpretable[[:space:]]Responsible[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]via[[:space:]]Dual-Space[[:space:]]Multi-facet[[:space:]]Concept[[:space:]]Control/2c24c8f1-1822-42ba-8659-5feab5b2d726_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Plug-and-Play[[:space:]]PPO_[[:space:]]An[[:space:]]Adaptive[[:space:]]Point[[:space:]]Prompt[[:space:]]Optimizer[[:space:]]Making[[:space:]]SAM[[:space:]]Greater/fb8f3fbf-9e73-432d-9130-bbe1d55d00ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Plug-and-Play[[:space:]]Versatile[[:space:]]Compressed[[:space:]]Video[[:space:]]Enhancement/e1321f09-cc21-461b-8861-4f42cf1e3c7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Point[[:space:]]Cloud[[:space:]]Upsampling[[:space:]]Using[[:space:]]Conditional[[:space:]]Diffusion[[:space:]]Module[[:space:]]with[[:space:]]Adaptive[[:space:]]Noise[[:space:]]Suppression/96fbf65a-6eae-49b1-9bdc-b206a300f6d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Point[[:space:]]Clouds[[:space:]]Meets[[:space:]]Physics_[[:space:]]Dynamic[[:space:]]Acoustic[[:space:]]Field[[:space:]]Fitting[[:space:]]Network[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Understanding/7aa2860d-04df-426a-bde9-40b0ed00d4aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Point-Cache_[[:space:]]Test-time[[:space:]]Dynamic[[:space:]]and[[:space:]]Hierarchical[[:space:]]Cache[[:space:]]for[[:space:]]Robust[[:space:]]and[[:space:]]Generalizable[[:space:]]Point[[:space:]]Cloud[[:space:]]Analysis/b2000dba-b471-4537-8cba-488e6bd7eaf3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Point-to-Region[[:space:]]Loss[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Point-Based[[:space:]]Crowd[[:space:]]Counting/7bc00c8f-f0e1-486b-af9f-cab0b77e9a84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Point2RBox-v2_[[:space:]]Rethinking[[:space:]]Point-supervised[[:space:]]Oriented[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Spatial[[:space:]]Layout[[:space:]]Among[[:space:]]Instances/d3a3a010-6834-441e-beb3-f3f784483984_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PointLoRA_[[:space:]]Low-Rank[[:space:]]Adaptation[[:space:]]with[[:space:]]Token[[:space:]]Selection[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Learning/d2184bbd-78d6-4278-aed5-2e0864e5914c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PointSR_[[:space:]]Self-Regularized[[:space:]]Point[[:space:]]Supervision[[:space:]]for[[:space:]]Drone-View[[:space:]]Object[[:space:]]Detection/f94115e7-2e5c-44d9-81fe-c067d1bde80f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PolarFree_[[:space:]]Polarization-based[[:space:]]Reflection-Free[[:space:]]Imaging/e1567ff5-b726-4ec2-baeb-81cce9b4c3bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PolarNeXt_[[:space:]]Rethink[[:space:]]Instance[[:space:]]Segmentation[[:space:]]with[[:space:]]Polar[[:space:]]Representation/a5f92f60-4941-4386-a4db-1f556681a113_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Polarized[[:space:]]Color[[:space:]]Screen[[:space:]]Matting/4bc8f905-0a33-46ab-838e-6d058731820a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Poly-Autoregressive[[:space:]]Prediction[[:space:]]for[[:space:]]Modeling[[:space:]]Interactions/43d051a6-6821-4e3b-bd9a-a4c244857daa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Population[[:space:]]Normalization[[:space:]]for[[:space:]]Federated[[:space:]]Learning/b6b8182e-2812-44b6-99cb-82e1b12982cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pos3R_[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation[[:space:]]for[[:space:]]Unseen[[:space:]]Objects[[:space:]]Made[[:space:]]Easy/162f30eb-880f-42fb-95ca-56f10c93b7f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pose[[:space:]]Priors[[:space:]]from[[:space:]]Language[[:space:]]Models/9f701210-ac31-4714-b2b0-3e6ed2c2dfd9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pose-Guided[[:space:]]Temporal[[:space:]]Enhancement[[:space:]]for[[:space:]]Robust[[:space:]]Low-Resolution[[:space:]]Hand[[:space:]]Reconstruction/61fa7d4b-03fc-4f0b-b84c-af353e8bd6fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PoseBH_[[:space:]]Prototypical[[:space:]]Multi-Dataset[[:space:]]Training[[:space:]]Beyond[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/6abac681-ec51-468a-ba82-3e1de6acd774_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PoseTraj_[[:space:]]Pose-Aware[[:space:]]Trajectory[[:space:]]Control[[:space:]]in[[:space:]]Video[[:space:]]Diffusion/8cdccae4-d7bc-4e3e-9398-db2b0c310587_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Positive2Negative_[[:space:]]Breaking[[:space:]]the[[:space:]]Information-Lossy[[:space:]]Barrier[[:space:]]in[[:space:]]Self-Supervised[[:space:]]Single[[:space:]]Image[[:space:]]Denoising/867e9811-fe1d-4217-8c41-8e621b6da43b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Post-pre-training[[:space:]]for[[:space:]]Modality[[:space:]]Alignment[[:space:]]in[[:space:]]Vision-Language[[:space:]]Foundation[[:space:]]Models/1196e091-a513-40f9-85ef-51604d8681b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PosterMaker_[[:space:]]Towards[[:space:]]High-Quality[[:space:]]Product[[:space:]]Poster[[:space:]]Generation[[:space:]]with[[:space:]]Accurate[[:space:]]Text[[:space:]]Rendering/b07ff10a-9a3d-44d4-80c1-c5cc5cfdc9e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PosterO_[[:space:]]Structuring[[:space:]]Layout[[:space:]]Trees[[:space:]]to[[:space:]]Enable[[:space:]]Language[[:space:]]Models[[:space:]]in[[:space:]]Generalized[[:space:]]Content-Aware[[:space:]]Layout[[:space:]]Generation/28953876-24ed-4c02-8fbd-02d889f7bc8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Potential[[:space:]]Field[[:space:]]Based[[:space:]]Deep[[:space:]]Metric[[:space:]]Learning/c5f58da8-3e8b-4fa3-accf-6f5c33627040_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pow3R_[[:space:]]Empowering[[:space:]]Unconstrained[[:space:]]3D[[:space:]]Reconstruction[[:space:]]with[[:space:]]Camera[[:space:]]and[[:space:]]Scene[[:space:]]Priors/fb9435c5-f90b-40a7-b204-e0d114acb0c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PrEditor3D_[[:space:]]Fast[[:space:]]and[[:space:]]Precise[[:space:]]3D[[:space:]]Shape[[:space:]]Editing/0f1a120c-8d95-4d35-a553-6cd4183aad0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Practical[[:space:]]Solutions[[:space:]]to[[:space:]]the[[:space:]]Relative[[:space:]]Pose[[:space:]]of[[:space:]]Three[[:space:]]Calibrated[[:space:]]Cameras/3fec933e-2dde-4be1-8ad1-28710156467f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Precise[[:space:]]Event[[:space:]]Spotting[[:space:]]in[[:space:]]Sports[[:space:]]Videos_[[:space:]]Solving[[:space:]]Long-Range[[:space:]]Dependency[[:space:]]and[[:space:]]Class[[:space:]]Imbalance/f70c89aa-2e9b-4112-89fc-f4837e98e688_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Precise,[[:space:]]Fast,[[:space:]]and[[:space:]]Low-cost[[:space:]]Concept[[:space:]]Erasure[[:space:]]in[[:space:]]Value[[:space:]]Space_[[:space:]]Orthogonal[[:space:]]Complement[[:space:]]Matters/cf17c55d-6d37-44b9-b018-ae4df3534d26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PreciseCam_[[:space:]]Precise[[:space:]]Camera[[:space:]]Control[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/f1c3bc77-cc18-4660-8f04-34981ca422eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Preconditioners[[:space:]]for[[:space:]]the[[:space:]]Stochastic[[:space:]]Training[[:space:]]of[[:space:]]Neural[[:space:]]Fields/b8f2c34c-ad8b-45bc-bfac-d02839f8ad19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Preserve[[:space:]]or[[:space:]]Modify_[[:space:]]Context-Aware[[:space:]]Evaluation[[:space:]]for[[:space:]]Balancing[[:space:]]Preservation[[:space:]]and[[:space:]]Modification[[:space:]]in[[:space:]]Text-Guided[[:space:]]Image[[:space:]]Editing/29ea49ed-c354-4bc9-af72-1f4bc52aa1a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Preserving[[:space:]]Clusters[[:space:]]in[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/c4c315b6-c9de-42ff-a0ca-93a857f517af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prior[[:space:]]Does[[:space:]]Matter_[[:space:]]Visual[[:space:]]Navigation[[:space:]]via[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Bridge[[:space:]]Models/ee705ec4-0cf2-476b-b84e-1441c5f20d50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prior-free[[:space:]]3D[[:space:]]Object[[:space:]]Tracking/56b7fc45-5f9f-43f7-bec3-8d552616dec5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProAPO_[[:space:]]Progressively[[:space:]]Automatic[[:space:]]Prompt[[:space:]]Optimization[[:space:]]for[[:space:]]Visual[[:space:]]Classification/a3c1ad48-aaba-47a6-aeeb-2ef7d80a2260_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProHOC_[[:space:]]Probabilistic[[:space:]]Hierarchical[[:space:]]Out-of-Distribution[[:space:]]Classification[[:space:]]via[[:space:]]Multi-Depth[[:space:]]Networks/ea6afc41-0a54-4fa5-8ade-1924337e5c45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProKeR_[[:space:]]A[[:space:]]Kernel[[:space:]]Perspective[[:space:]]on[[:space:]]Few-Shot[[:space:]]Adaptation[[:space:]]of[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/db02e676-2ec5-438d-b07d-f0e3ebad98e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProReflow_[[:space:]]Progressive[[:space:]]Reflow[[:space:]]with[[:space:]]Decomposed[[:space:]]Velocity/ab413bbc-dea1-4ace-8805-5da24de9f86f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProbPose_[[:space:]]A[[:space:]]Probabilistic[[:space:]]Approach[[:space:]]to[[:space:]]2D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/8ccd2594-9394-4eab-a971-5f2753ccb58b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Probabilistic[[:space:]]Prompt[[:space:]]Distribution[[:space:]]Learning[[:space:]]for[[:space:]]Animal[[:space:]]Pose[[:space:]]Estimation/1be3cde5-60cb-4a57-a201-7a6dd25687f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Probability[[:space:]]Density[[:space:]]Geodesics[[:space:]]in[[:space:]]Image[[:space:]]Diffusion[[:space:]]Latent[[:space:]]Space/ae7d5292-3106-44ab-bc09-b7c04df42c2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProbeSDF_[[:space:]]Light[[:space:]]Field[[:space:]]Probes[[:space:]]For[[:space:]]Neural[[:space:]]Surface[[:space:]]Reconstruction/c8ccc978-962a-4a2e-a885-9ef25eda0c9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Probing[[:space:]]the[[:space:]]Mid-level[[:space:]]Vision[[:space:]]Capabilities[[:space:]]of[[:space:]]Self-Supervised[[:space:]]Learning/eaa35116-af32-4a96-9d09-fe46bcf9854f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prof.[[:space:]]Robot_[[:space:]]Differentiable[[:space:]]Robot[[:space:]]Rendering[[:space:]]Without[[:space:]]Static[[:space:]]and[[:space:]]Self-Collisions/f5b2ba36-3dfc-406e-a34c-8757ef9ca5cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Progress-Aware[[:space:]]Video[[:space:]]Frame[[:space:]]Captioning/1ef336dd-a39c-4288-9c6c-56f22b842b12_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Progressive[[:space:]]Correspondence[[:space:]]Regenerator[[:space:]]for[[:space:]]Robust[[:space:]]3D[[:space:]]Registration/efc41f9b-f851-485a-97ce-67b61ab26c11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Progressive[[:space:]]Focused[[:space:]]Transformer[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Super-Resolution/a3bd7da0-3f79-431d-b8a1-9a72e56058b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Progressive[[:space:]]Rendering[[:space:]]Distillation_[[:space:]]Adapting[[:space:]]Stable[[:space:]]Diffusion[[:space:]]for[[:space:]]Instant[[:space:]]Text-to-Mesh[[:space:]]Generation[[:space:]]without[[:space:]]3D[[:space:]]Data/64be8fea-91f9-4a05-a3e4-a20e757ee9ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProjAttacker_[[:space:]]A[[:space:]]Configurable[[:space:]]Physical[[:space:]]Adversarial[[:space:]]Attack[[:space:]]for[[:space:]]Face[[:space:]]Recognition[[:space:]]via[[:space:]]Projector/4340b920-658a-4e45-bed9-f9bb84d4699c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Project-Probe-Aggregate_[[:space:]]Efficient[[:space:]]Fine-Tuning[[:space:]]for[[:space:]]Group[[:space:]]Robustness/5e368750-4958-49b1-b258-1956bba6e2a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prometheus_[[:space:]]3D-Aware[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Feed-Forward[[:space:]]Text-to-3D[[:space:]]Scene[[:space:]]Generation/f7239594-bff6-466c-a34f-3ffb021b8e29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prompt-CAM_[[:space:]]Making[[:space:]]Vision[[:space:]]Transformers[[:space:]]Interpretable[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Analysis/88b2374b-97f1-4093-aedd-5058cdff0e06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prompt2Perturb[[:space:]](P2P)_[[:space:]]Text-Guided[[:space:]]Diffusion-Based[[:space:]]Adversarial[[:space:]]Attack[[:space:]]on[[:space:]]Breast[[:space:]]Ultrasound[[:space:]]Images/98c943d3-1292-44ba-a2e7-68b21f76d5b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PromptHMR_[[:space:]]Promptable[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/5b944c82-9867-4899-8aca-e1b6957dfaa1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PromptHash_Affinity-Prompted[[:space:]]Collaborative[[:space:]]Cross-Modal[[:space:]]Learning[[:space:]]for[[:space:]]Adaptive[[:space:]]Hashing[[:space:]]Retrieval/64d85235-28eb-4b3d-9010-9b6834181937_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prompting[[:space:]]Depth[[:space:]]Anything[[:space:]]for[[:space:]]4K[[:space:]]Resolution[[:space:]]Accurate[[:space:]]Metric[[:space:]]Depth[[:space:]]Estimation/5ba6a389-8293-4209-b8d7-39d2a373b186_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prosody-Enhanced[[:space:]]Acoustic[[:space:]]Pre-training[[:space:]]and[[:space:]]Acoustic-Disentangled[[:space:]]Prosody[[:space:]]Adapting[[:space:]]for[[:space:]]Movie[[:space:]]Dubbing/424fa348-ae41-4000-988b-67f0bf11eea6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Protecting[[:space:]]Your[[:space:]]Video[[:space:]]Content_[[:space:]]Disrupting[[:space:]]Automated[[:space:]]Video-based[[:space:]]LLM[[:space:]]Annotations/3e55fb27-6e40-4d5b-80e6-37b2f1422af5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProtoDepth_[[:space:]]Unsupervised[[:space:]]Continual[[:space:]]Depth[[:space:]]Completion[[:space:]]with[[:space:]]Prototypes/739ff3ca-b993-42db-9cbe-a91a1fb89d52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prototype-Based[[:space:]]Image[[:space:]]Prompting[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Histopathological[[:space:]]Image[[:space:]]Segmentation/3d418639-aae6-43b4-aad7-bbaf79e8d3c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Provoking[[:space:]]Multi-modal[[:space:]]Few-Shot[[:space:]]LVLM[[:space:]]via[[:space:]]Exploration-Exploitation[[:space:]]In-Context[[:space:]]Learning/b91a6ee3-ca22-45bc-888a-12a0ccc4e9d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Proximal[[:space:]]Algorithm[[:space:]]Unrolling_[[:space:]]Flexible[[:space:]]and[[:space:]]Efficient[[:space:]]Reconstruction[[:space:]]Networks[[:space:]]for[[:space:]]Single-Pixel[[:space:]]Imaging/b1aa96ba-e721-4e19-b13a-1d18da180bc8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProxyTransformation_[[:space:]]Preshaping[[:space:]]Point[[:space:]]Cloud[[:space:]]Manifold[[:space:]]With[[:space:]]Proxy[[:space:]]Attention[[:space:]]For[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding/4b7e6ed5-1e73-47b0-831f-cd411124477b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pseudo[[:space:]]Visible[[:space:]]Feature[[:space:]]Fine-Grained[[:space:]]Fusion[[:space:]]for[[:space:]]Thermal[[:space:]]Object[[:space:]]Detection/6a6b4e19-6c9a-4a0e-90ee-27b158772a20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pursuing[[:space:]]Temporal-Consistent[[:space:]]Video[[:space:]]Virtual[[:space:]]Try-On[[:space:]]via[[:space:]]Dynamic[[:space:]]Pose[[:space:]]Interaction/0dbff5db-f8cb-4f92-8abf-9eb7e5e15468_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PyTorchGeoNodes_[[:space:]]Enabling[[:space:]]Differentiable[[:space:]]Shape[[:space:]]Programs[[:space:]]for[[:space:]]3D[[:space:]]Shape[[:space:]]Reconstruction/795f18d0-96c4-47cf-9ef2-701d590c4ab0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Q-Bench-Video_[[:space:]]Benchmark[[:space:]]the[[:space:]]Video[[:space:]]Quality[[:space:]]Understanding[[:space:]]of[[:space:]]LMMs/f97aa133-81bf-4c87-9369-c5d1f7aad162_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Q-DiT_[[:space:]]Accurate[[:space:]]Post-Training[[:space:]]Quantization[[:space:]]for[[:space:]]Diffusion[[:space:]]Transformers/5289ce42-2d7e-4bc4-96f9-cf1d2e0175bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Q-Eval-100K_[[:space:]]Evaluating[[:space:]]Visual[[:space:]]Quality[[:space:]]and[[:space:]]Alignment[[:space:]]Level[[:space:]]for[[:space:]]Text-to-Vision[[:space:]]Content/4ac7661d-7792-4463-bffb-e8cb82d66bee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Q-PART_[[:space:]]Quasi-Periodic[[:space:]]Adaptive[[:space:]]Regression[[:space:]]with[[:space:]]Test-time[[:space:]]Training[[:space:]]for[[:space:]]Pediatric[[:space:]]Left[[:space:]]Ventricular[[:space:]]Ejection[[:space:]]Fraction[[:space:]]Regression/9b3ea12c-1e95-4e84-9ddd-5170e5c7771b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/QMambaBSR_[[:space:]]Burst[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]with[[:space:]]Query[[:space:]]State[[:space:]]Space[[:space:]]Model/ccb55e80-eeaa-40e8-8e8b-d128ea3974f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/QuCOOP_[[:space:]]A[[:space:]]Versatile[[:space:]]Framework[[:space:]]for[[:space:]]Solving[[:space:]]Composite[[:space:]]and[[:space:]]Binary-Parametrised[[:space:]]Problems[[:space:]]on[[:space:]]Quantum[[:space:]]Annealers/c311c3bc-3c21-4b1a-a4f8-f00774ef973f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Quad-Pixel[[:space:]]Image[[:space:]]Defocus[[:space:]]Deblurring_[[:space:]]A[[:space:]]New[[:space:]]Benchmark[[:space:]]and[[:space:]]Model/c721a9dd-029d-4470-b5ba-1f618e42b739_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Quaffure_[[:space:]]Real-Time[[:space:]]Quasi-Static[[:space:]]Neural[[:space:]]Hair[[:space:]]Simulation/40107a29-fe45-40f8-a607-7d120945adc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Quantization[[:space:]]without[[:space:]]Tears/7acb9c84-5b2e-438a-8a19-e3c09310038a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/QuartDepth_[[:space:]]Post-Training[[:space:]]Quantization[[:space:]]for[[:space:]]Real-Time[[:space:]]Depth[[:space:]]Estimation[[:space:]]on[[:space:]]the[[:space:]]Edge/7bffb5f3-e473-422d-9ee3-fd3a6408a1a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Query[[:space:]]Efficient[[:space:]]Black-Box[[:space:]]Visual[[:space:]]Prompting[[:space:]]with[[:space:]]Subspace[[:space:]]Learning/a238a6bd-550a-424d-9ec1-def2baa9176b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Question-Aware[[:space:]]Gaussian[[:space:]]Experts[[:space:]]for[[:space:]]Audio-Visual[[:space:]]Question[[:space:]]Answering/8c821211-5fd4-495c-9b23-97b0652c108a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/R-SCoRe_[[:space:]]Revisiting[[:space:]]Scene[[:space:]]Coordinate[[:space:]]Regression[[:space:]]for[[:space:]]Robust[[:space:]]Large-Scale[[:space:]]Visual[[:space:]]Localization/de480991-e387-4648-af00-268474a4dd11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/R-TPT_[[:space:]]Improving[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models[[:space:]]through[[:space:]]Test-Time[[:space:]]Prompt[[:space:]]Tuning/db61575c-1280-4b24-9ca2-b88169374737_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/R2C_[[:space:]]Mapping[[:space:]]Room[[:space:]]to[[:space:]]Chessboard[[:space:]]to[[:space:]]Unlock[[:space:]]LLM[[:space:]]As[[:space:]]Low-Level[[:space:]]Action[[:space:]]Planner/47843699-e3e2-4b14-a3de-2c78f86cbdfa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RADIOv2.5_[[:space:]]Improved[[:space:]]Baselines[[:space:]]for[[:space:]]Agglomerative[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models/f67d7907-0e00-4bb3-98df-d2ff222bce3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RAD_[[:space:]]Region-Aware[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Image[[:space:]]Inpainting/c3fc017b-fff9-4d44-a432-47bdefc13495_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RAEncoder_[[:space:]]A[[:space:]]Label-Free[[:space:]]Reversible[[:space:]]Adversarial[[:space:]]Examples[[:space:]]Encoder[[:space:]]for[[:space:]]Dataset[[:space:]]Intellectual[[:space:]]Property[[:space:]]Protection/d38fb843-2f59-4406-8bd2-f601b0c7b2ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RANGE_[[:space:]]Retrieval[[:space:]]Augmented[[:space:]]Neural[[:space:]]Fields[[:space:]]for[[:space:]]Multi-Resolution[[:space:]]Geo-Embeddings/8a549df3-3427-4bdd-bb2f-4a67986ec4e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RAP_[[:space:]]Retrieval-Augmented[[:space:]]Personalization[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/7309d551-2983-4b1a-9d3d-bd90263e9ffc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RASP_[[:space:]]Revisiting[[:space:]]3D[[:space:]]Anamorphic[[:space:]]Art[[:space:]]for[[:space:]]Shadow-Guided[[:space:]]Packing[[:space:]]of[[:space:]]Irregular[[:space:]]Objects/aef7411d-a9ca-4b1d-a6a0-4ab29fc800b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RC-AutoCalib_[[:space:]]An[[:space:]]End-to-End[[:space:]]Radar-Camera[[:space:]]Automatic[[:space:]]Calibration[[:space:]]Network/a8eca73f-f319-47b0-b493-af40482783f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RCP-Bench_[[:space:]]Benchmarking[[:space:]]Robustness[[:space:]]for[[:space:]]Collaborative[[:space:]]Perception[[:space:]]Under[[:space:]]Diverse[[:space:]]Corruptions/9170f020-ef41-4783-9545-58730495f413_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RDD_[[:space:]]Robust[[:space:]]Feature[[:space:]]Detector[[:space:]]and[[:space:]]Descriptor[[:space:]]using[[:space:]]Deformable[[:space:]]Transformer/dd7235df-3118-4324-8db3-9be17c94a765_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RELOCATE_[[:space:]]A[[:space:]]Simple[[:space:]]Training-Free[[:space:]]Baseline[[:space:]]for[[:space:]]Visual[[:space:]]Query[[:space:]]Localization[[:space:]]Using[[:space:]]Region-Based[[:space:]]Representations/0b95343c-eb8a-4ffd-99e0-a7265d714a94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RENO_[[:space:]]Real-Time[[:space:]]Neural[[:space:]]Compression[[:space:]]for[[:space:]]3D[[:space:]]LiDAR[[:space:]]Point[[:space:]]Clouds/39b4a5e0-b6f4-432e-8b8b-9cf6f16fe276_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/REWIND_[[:space:]]Real-Time[[:space:]]Egocentric[[:space:]]Whole-Body[[:space:]]Motion[[:space:]]Diffusion[[:space:]]with[[:space:]]Exemplar-Based[[:space:]]Identity[[:space:]]Conditioning/ffd639f7-1ce7-4a48-8034-72429670d484_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RGBAvatar_[[:space:]]Reduced[[:space:]]Gaussian[[:space:]]Blendshapes[[:space:]]for[[:space:]]Online[[:space:]]Modeling[[:space:]]of[[:space:]]Head[[:space:]]Avatars/7be001b2-a469-482b-aaf4-ac68ca4fcfc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RICCARDO_[[:space:]]Radar[[:space:]]Hit[[:space:]]Prediction[[:space:]]and[[:space:]]Convolution[[:space:]]for[[:space:]]Camera-Radar[[:space:]]3D[[:space:]]Object[[:space:]]Detection/76e4b9f5-1122-461a-b5e5-9b30b5946e13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RL-RC-DoT_[[:space:]]A[[:space:]]Block-level[[:space:]]RL[[:space:]]agent[[:space:]]for[[:space:]]Task-Aware[[:space:]]Video[[:space:]]Compression/567c1b1c-404a-4519-9663-668d9f3e827b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RLAIF-V_[[:space:]]Open-Source[[:space:]]AI[[:space:]]Feedback[[:space:]]Leads[[:space:]]to[[:space:]]Super[[:space:]]GPT-4V[[:space:]]Trustworthiness/efc9b31b-68d9-4964-963a-be0c77b65098_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RNG_[[:space:]]Relightable[[:space:]]Neural[[:space:]]Gaussians/e6ee837e-39a9-4722-8b41-7807f788f5b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ROCKET-1_[[:space:]]Mastering[[:space:]]Open-World[[:space:]]Interaction[[:space:]]with[[:space:]]Visual-Temporal[[:space:]]Context[[:space:]]Prompting/b498a6d4-6e42-4bc9-a519-aefbfa4fb148_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ROD-MLLM_[[:space:]]Towards[[:space:]]More[[:space:]]Reliable[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/a3e2cba9-0d24-4767-95eb-0b9f77a83aa5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ROICtrl_[[:space:]]Boosting[[:space:]]Instance[[:space:]]Control[[:space:]]for[[:space:]]Visual[[:space:]]Generation/6078b586-a6e7-4108-8ef5-0c83dcaffdb3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ROLL_[[:space:]]Robust[[:space:]]Noisy[[:space:]]Pseudo-label[[:space:]]Learning[[:space:]]for[[:space:]]Multi-View[[:space:]]Clustering[[:space:]]with[[:space:]]Noisy[[:space:]]Correspondence/0ddd580a-0372-4643-8dc8-4c76cd7cab9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RORem_[[:space:]]Training[[:space:]]a[[:space:]]Robust[[:space:]]Object[[:space:]]Remover[[:space:]]with[[:space:]]Human-in-the-Loop/db70faae-6e6b-4a93-b974-cf66bc30fd1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ROS-SAM_[[:space:]]High-Quality[[:space:]]Interactive[[:space:]]Segmentation[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Moving[[:space:]]Object/665dcf8c-03d1-4e77-b2d1-eb1870230beb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RSAR_[[:space:]]Restricted[[:space:]]State[[:space:]]Angle[[:space:]]Resolver[[:space:]]and[[:space:]]Rotated[[:space:]]SAR[[:space:]]Benchmark/cd4f1f10-5d8f-4540-88c9-a775ac305b65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RUBIK_[[:space:]]A[[:space:]]Structured[[:space:]]Benchmark[[:space:]]for[[:space:]]Image[[:space:]]Matching[[:space:]]across[[:space:]]Geometric[[:space:]]Challenges/4055a777-cb70-4e8d-a56c-68c85be745af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RaCFormer_[[:space:]]Towards[[:space:]]High-Quality[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Query-based[[:space:]]Radar-Camera[[:space:]]Fusion/aef1ab03-d0c9-4bbc-a602-26c872f2441c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RaSS_[[:space:]]Improving[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Samplers[[:space:]]with[[:space:]]Reinforced[[:space:]]Active[[:space:]]Sampling[[:space:]]Scheduler/18f05ee5-f239-414b-8fd8-16228f6440cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Radio[[:space:]]Frequency[[:space:]]Ray[[:space:]]Tracing[[:space:]]with[[:space:]]Neural[[:space:]]Object[[:space:]]Representation[[:space:]]for[[:space:]]Enhanced[[:space:]]RF[[:space:]]Modeling/2ab3b9d7-2ec3-4fe2-8f34-0e4410dfd2e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RainyGS_[[:space:]]Efficient[[:space:]]Rain[[:space:]]Synthesis[[:space:]]with[[:space:]]Physically-Based[[:space:]]Gaussian[[:space:]]Splatting/7e287e43-86a7-461d-9e10-a6d28a8e1343_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RandAR_[[:space:]]Decoder-only[[:space:]]Autoregressive[[:space:]]Visual[[:space:]]Generation[[:space:]]in[[:space:]]Random[[:space:]]Orders/a0e6e964-828c-4e39-ae4e-c31e79523926_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Random[[:space:]]Conditioning[[:space:]]for[[:space:]]Diffusion[[:space:]]Model[[:space:]]Compression[[:space:]]with[[:space:]]Distillation/8a08399e-2554-4d80-9b75-a8d09da1f5a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rashomon[[:space:]]Sets[[:space:]]for[[:space:]]Prototypical-Part[[:space:]]Networks_[[:space:]]Editing[[:space:]]Interpretable[[:space:]]Models[[:space:]]in[[:space:]]Real-Time/c8eb1c46-b8d1-4ec2-91e7-c2141aad5bf0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rate-In_[[:space:]]Information-Driven[[:space:]]Adaptive[[:space:]]Dropout[[:space:]]Rates[[:space:]]for[[:space:]]Improved[[:space:]]Inference-Time[[:space:]]Uncertainty[[:space:]]Estimation/7956c40c-7c1f-4a16-b6c1-dc84d8132ff2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RayFlow_[[:space:]]Instance-Aware[[:space:]]Diffusion[[:space:]]Acceleration[[:space:]]via[[:space:]]Adaptive[[:space:]]Flow[[:space:]]Trajectories/4bf269b9-9ce6-46d8-8062-320205ed488e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Re-HOLD_[[:space:]]Video[[:space:]]Hand[[:space:]]Object[[:space:]]Interaction[[:space:]]Reenactment[[:space:]]via[[:space:]]adaptive[[:space:]]Layout-instructed[[:space:]]Diffusion[[:space:]]Model/ef07d2c8-a886-4f21-822c-a325057b9979_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Re-thinking[[:space:]]Temporal[[:space:]]Search[[:space:]]for[[:space:]]Long-Form[[:space:]]Video[[:space:]]Understanding/55643346-dcc1-4145-86b7-17d5f78d76f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReCap_[[:space:]]Better[[:space:]]Gaussian[[:space:]]Relighting[[:space:]]with[[:space:]]Cross-Environment[[:space:]]Captures/99bfbb8f-ec2d-409e-88b3-99b66cea3553_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReCapture_[[:space:]]Generative[[:space:]]Video[[:space:]]Camera[[:space:]]Controls[[:space:]]for[[:space:]]User-Provided[[:space:]]Videos[[:space:]]using[[:space:]]Masked[[:space:]]Video[[:space:]]Fine-Tuning/3544a413-5d2e-4b09-b3f1-1ba3e94814d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReCon_[[:space:]]Enhancing[[:space:]]True[[:space:]]Correspondence[[:space:]]Discrimination[[:space:]]through[[:space:]]Relation[[:space:]]Consistency[[:space:]]for[[:space:]]Robust[[:space:]]Noisy[[:space:]]Correspondence[[:space:]]Learning/283c5fe1-41e8-4a3d-a138-86287d321e7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReDiffDet_[[:space:]]Rotation-equivariant[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Oriented[[:space:]]Object[[:space:]]Detection/3c120bfe-a23a-4b2e-9ea9-d61f7b316287_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReNeg_[[:space:]]Learning[[:space:]]Negative[[:space:]]Embedding[[:space:]]with[[:space:]]Reward[[:space:]]Guidance/2dd8ea42-76d9-478f-9833-0cfcc98fc526_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RePerformer_[[:space:]]Immersive[[:space:]]Human-centric[[:space:]]Volumetric[[:space:]]Videos[[:space:]]from[[:space:]]Playback[[:space:]]to[[:space:]]Photoreal[[:space:]]Reperformance/9cffa49b-aca9-4d4a-bd20-c62f1c84ee48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReRAW_[[:space:]]RGB-to-RAW[[:space:]]Image[[:space:]]Reconstruction[[:space:]]via[[:space:]]Stratified[[:space:]]Sampling[[:space:]]for[[:space:]]Efficient[[:space:]]Object[[:space:]]Detection[[:space:]]on[[:space:]]the[[:space:]]Edge/192f840d-b880-40aa-932e-804498bbc19f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReSpec_[[:space:]]Relevance[[:space:]]and[[:space:]]Specificity[[:space:]]Grounded[[:space:]]Online[[:space:]]Filtering[[:space:]]for[[:space:]]Learning[[:space:]]on[[:space:]]Video-Text[[:space:]]Data[[:space:]]Streams/d405ceb2-3c9a-4a00-a126-f6a33cdf8ddd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReVisionLLM_[[:space:]]Recursive[[:space:]]Vision-Language[[:space:]]Model[[:space:]]for[[:space:]]Temporal[[:space:]]Grounding[[:space:]]in[[:space:]]Hour-Long[[:space:]]Videos/7d14cbaf-b7ec-45d8-a41e-130ae614bbd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReWind_[[:space:]]Understanding[[:space:]]Long[[:space:]]Videos[[:space:]]with[[:space:]]Instructed[[:space:]]Learnable[[:space:]]Memory/6a8a4575-34ea-4c66-9261-563393bbc7df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Real-IAD[[:space:]]D3_[[:space:]]A[[:space:]]Real-World[[:space:]]2D_Pseudo-3D_3D[[:space:]]Dataset[[:space:]]for[[:space:]]Industrial[[:space:]]Anomaly[[:space:]]Detection/6617953f-620e-4286-93ff-17cea3c0c6a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Real-time[[:space:]]Free-view[[:space:]]Human[[:space:]]Rendering[[:space:]]from[[:space:]]Sparse-view[[:space:]]RGB[[:space:]]Videos[[:space:]]using[[:space:]]Double[[:space:]]Unprojected[[:space:]]Textures/b84b54f9-0440-4d6f-8edc-b9ef1fc24a58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Real-time[[:space:]]High-fidelity[[:space:]]Gaussian[[:space:]]Human[[:space:]]Avatars[[:space:]]with[[:space:]]Position-based[[:space:]]Interpolation[[:space:]]of[[:space:]]Spatially[[:space:]]Distributed[[:space:]]MLPs/58c0c3bd-700d-43fd-8915-2c029980720d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RealEdit_[[:space:]]Reddit[[:space:]]Edits[[:space:]]As[[:space:]]a[[:space:]]Large-scale[[:space:]]Empirical[[:space:]]Dataset[[:space:]]for[[:space:]]Image[[:space:]]Transformations/d56fb203-14eb-4e52-858d-fc16e1736d74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Realistic[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models/e27e89eb-afdc-4d20-92c2-aa9ca234c954_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reanimating[[:space:]]Images[[:space:]]using[[:space:]]Neural[[:space:]]Representations[[:space:]]of[[:space:]]Dynamic[[:space:]]Stimuli/874e2680-5896-4c9b-91c9-09f3d766cc1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reason-before-Retrieve_[[:space:]]One-Stage[[:space:]]Reflective[[:space:]]Chain-of-Thoughts[[:space:]]for[[:space:]]Training-Free[[:space:]]Zero-Shot[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/e9bf520c-9e09-45e1-bbbc-16dbe7011696_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReasonGrounder_[[:space:]]LVLM-Guided[[:space:]]Hierarchical[[:space:]]Feature[[:space:]]Splatting[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding[[:space:]]and[[:space:]]Reasoning/50e72652-8c72-4132-a8d0-6766a40b493c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reasoning[[:space:]]Mamba_[[:space:]]Hypergraph-Guided[[:space:]]Region[[:space:]]Relation[[:space:]]Calculating[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Affordance[[:space:]]Grounding/2d000e5c-3508-4351-a8f3-3e76a3d1254d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reasoning[[:space:]]in[[:space:]]Visual[[:space:]]Navigation[[:space:]]of[[:space:]]End-to-end[[:space:]]Trained[[:space:]]Agents_[[:space:]]A[[:space:]]Dynamical[[:space:]]Systems[[:space:]]Approach/b51b3d15-3a18-4e71-89be-8f3addad2ca2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reasoning[[:space:]]to[[:space:]]Attend_[[:space:]]Try[[:space:]]to[[:space:]]Understand[[:space:]]How[[:space:]]_SEG_[[:space:]]Token[[:space:]]Works/ab784d25-ef1c-45d0-9ca5-96598b096585_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Recognition-Synergistic[[:space:]]Scene[[:space:]]Text[[:space:]]Editing/9cce98ca-d165-4377-805b-790ca32b1bbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReconDreamer_[[:space:]]Crafting[[:space:]]World[[:space:]]Models[[:space:]]for[[:space:]]Driving[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]via[[:space:]]Online[[:space:]]Restoration/faf64427-1086-4008-9126-875ef2b87c83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reconciling[[:space:]]Stochastic[[:space:]]and[[:space:]]Deterministic[[:space:]]Strategies[[:space:]]for[[:space:]]Zero-shot[[:space:]]Image[[:space:]]Restoration[[:space:]]using[[:space:]]Diffusion[[:space:]]Model[[:space:]]in[[:space:]]Dual/7694bc21-a628-4801-90b0-a7b28ba946a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reconstructing[[:space:]]Animals[[:space:]]and[[:space:]]the[[:space:]]Wild/5ae9e06d-d523-45d1-936f-355ee6e5907f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reconstructing[[:space:]]Close[[:space:]]Human[[:space:]]Interaction[[:space:]]with[[:space:]]Appearance[[:space:]]and[[:space:]]Proxemics[[:space:]]Reasoning/c72601e6-8b68-41ff-a9b2-75acd1e703a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reconstructing[[:space:]]Humans[[:space:]]with[[:space:]]a[[:space:]]Biomechanically[[:space:]]Accurate[[:space:]]Skeleton/bb366ea0-7534-437c-a136-20f531d259d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reconstructing[[:space:]]In-the-Wild[[:space:]]Open-Vocabulary[[:space:]]Human-Object[[:space:]]Interactions/7b6776a3-2e2a-4f44-bedd-e340ad61e164_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reconstructing[[:space:]]People,[[:space:]]Places,[[:space:]]and[[:space:]]Cameras/0e73ff7c-c575-4d20-a8f8-ee3180fae024_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reconstruction[[:space:]]vs.[[:space:]]Generation_[[:space:]]Taming[[:space:]]Optimization[[:space:]]Dilemma[[:space:]]in[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models/28c08e8e-377b-4778-9024-89cecb50aaa5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Recover[[:space:]]and[[:space:]]Match_[[:space:]]Open-Vocabulary[[:space:]]Multi-Label[[:space:]]Recognition[[:space:]]through[[:space:]]Knowledge-Constrained[[:space:]]Optimal[[:space:]]Transport/43c2cc68-57b8-487e-9318-fc614f9b2bc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Recovering[[:space:]]Dynamic[[:space:]]3D[[:space:]]Sketches[[:space:]]from[[:space:]]Videos/3f090001-6ce4-4e3b-9401-74e0009bd850_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rectification-specific[[:space:]]Supervision[[:space:]]and[[:space:]]Constrained[[:space:]]Estimator[[:space:]]for[[:space:]]Online[[:space:]]Stereo[[:space:]]Rectification/0458f371-9013-4b13-bcad-4e72b3fd6b6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rectified[[:space:]]Diffusion[[:space:]]Guidance[[:space:]]for[[:space:]]Conditional[[:space:]]Generation/6bb3a343-3f41-4113-bdf8-40055d4fdafd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Recurrence-Enhanced[[:space:]]Vision-and-Language[[:space:]]Transformers[[:space:]]for[[:space:]]Robust[[:space:]]Multimodal[[:space:]]Document[[:space:]]Retrieval/8cc873ea-c75d-45ba-84f7-b1cc5abe7f86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Recurrent[[:space:]]Feature[[:space:]]Mining[[:space:]]and[[:space:]]Keypoint[[:space:]]Mixup[[:space:]]Padding[[:space:]]for[[:space:]]Category-Agnostic[[:space:]]Pose[[:space:]]Estimation/a43072a1-3e09-499d-b9fe-6e1d72d76075_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Redefining[[:space:]]_Creative_[[:space:]]in[[:space:]]Dictionary_[[:space:]]Towards[[:space:]]an[[:space:]]Enhanced[[:space:]]Semantic[[:space:]]Understanding[[:space:]]of[[:space:]]Creative[[:space:]]Generation/a6ae9285-7fbb-4728-aef6-d6a09c5bf8a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reducing[[:space:]]Class-wise[[:space:]]Confusion[[:space:]]for[[:space:]]Incremental[[:space:]]Learning[[:space:]]with[[:space:]]Disentangled[[:space:]]Manifolds/07248f3f-e491-4078-9a53-2b88f44a101d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ref-GS_[[:space:]]Directional[[:space:]]Factorization[[:space:]]for[[:space:]]2D[[:space:]]Gaussian[[:space:]]Splatting/6d7640fe-8cc4-4b89-aefa-7b1f3a6aa1ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RefPose_[[:space:]]Leveraging[[:space:]]Reference[[:space:]]Geometric[[:space:]]Correspondences[[:space:]]for[[:space:]]Accurate[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation[[:space:]]of[[:space:]]Unseen[[:space:]]Objects/d3e9731a-2dd5-414e-b189-d505a88e86ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reference-Based[[:space:]]3D-Aware[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Triplanes/a0a0edad-d72e-46a4-93f3-3fc9ff27cbf0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Relation-Rich[[:space:]]Visual[[:space:]]Document[[:space:]]Generator[[:space:]]for[[:space:]]Visual[[:space:]]Information[[:space:]]Extraction/e752f7aa-0541-4e00-a977-a255e1fa7667_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Relation3D[[:space:]]_[[:space:]]Enhancing[[:space:]]Relation[[:space:]]Modeling[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Instance[[:space:]]Segmentation/28183270-d14f-4654-8966-4d31c871cc7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RelationField_[[:space:]]Relate[[:space:]]Anything[[:space:]]in[[:space:]]Radiance[[:space:]]Fields/daa7a061-0189-45bd-bd6e-811777a5750e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Relative[[:space:]]Pose[[:space:]]Estimation[[:space:]]through[[:space:]]Affine[[:space:]]Corrections[[:space:]]of[[:space:]]Monocular[[:space:]]Depth[[:space:]]Priors/30c5e61f-4532-43a0-911e-a0f24e957bb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reloc3r_[[:space:]]Large-Scale[[:space:]]Training[[:space:]]of[[:space:]]Relative[[:space:]]Camera[[:space:]]Pose[[:space:]]Regression[[:space:]]for[[:space:]]Generalizable,[[:space:]]Fast,[[:space:]]and[[:space:]]Accurate[[:space:]]Visual[[:space:]]Localization/ee055fc6-de7a-4dc4-88b1-9c07aaedda1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Remote[[:space:]]Photoplethysmography[[:space:]]in[[:space:]]Real-World[[:space:]]and[[:space:]]Extreme[[:space:]]Lighting[[:space:]]Scenarios/9c41485d-1c85-4e79-b90f-4b7f3a6e81b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Removing[[:space:]]Reflections[[:space:]]from[[:space:]]RAW[[:space:]]Photos/7b6f4efa-0c4a-46d6-b345-c43942dd07a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reproducible[[:space:]]Vision-Language[[:space:]]Models[[:space:]]Meet[[:space:]]Concepts[[:space:]]Out[[:space:]]of[[:space:]]Pre-Training/01891411-e11f-4190-a260-e532e8e57380_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Repurposing[[:space:]]Pre-trained[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Event-based[[:space:]]Video[[:space:]]Interpolation/e9a01722-a15a-4f0e-ab86-be55c66a49f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Repurposing[[:space:]]Stable[[:space:]]Diffusion[[:space:]]Attention[[:space:]]for[[:space:]]Training-Free[[:space:]]Unsupervised[[:space:]]Interactive[[:space:]]Segmentation/4b62901e-7739-4fda-9556-570a7b198b3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ResCLIP_[[:space:]]Residual[[:space:]]Attention[[:space:]]for[[:space:]]Training-free[[:space:]]Dense[[:space:]]Vision-language[[:space:]]Inference/d91c539a-0263-4f21-b714-a8dfa8e9be0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Resilient[[:space:]]Sensor[[:space:]]Fusion[[:space:]]Under[[:space:]]Adverse[[:space:]]Sensor[[:space:]]Failures[[:space:]]via[[:space:]]Multi-Modal[[:space:]]Expert[[:space:]]Fusion/9a19d9d1-5d81-457f-9199-d8f534a2804e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RestorGS_[[:space:]]Depth-aware[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Efficient[[:space:]]3D[[:space:]]Scene[[:space:]]Restoration/63c6018a-b4b6-488c-a72c-f3e08fe599af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Retaining[[:space:]]Knowledge[[:space:]]and[[:space:]]Enhancing[[:space:]]Long-Text[[:space:]]Representations[[:space:]]in[[:space:]]CLIP[[:space:]]through[[:space:]]Dual-Teacher[[:space:]]Distillation/f74769fb-445b-4b03-bc83-3f4ed0e3ece4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Correspondence-based[[:space:]]Category-Level[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/ad43205b-b19f-4ffd-bbc9-db6043026644_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Decoder[[:space:]]Design_[[:space:]]Improving[[:space:]]Biomarker[[:space:]]Segmentation[[:space:]]Using[[:space:]]Depth-to-Space[[:space:]]Restoration[[:space:]]and[[:space:]]Residual[[:space:]]Linear[[:space:]]Attention/5bce38d4-e047-415d-a6c6-007ab056b9c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Diffusion[[:space:]]for[[:space:]]Text-Driven[[:space:]]Human[[:space:]]Motion[[:space:]]Generation_[[:space:]]Redundant[[:space:]]Representations,[[:space:]]Evaluation,[[:space:]]and[[:space:]]Masked[[:space:]]Autoregression/09b81429-0fac-4a9a-94fe-727cad400488_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]End-to-End[[:space:]]2D[[:space:]]to[[:space:]]3D[[:space:]]Scene[[:space:]]Segmentation[[:space:]]in[[:space:]]Gaussian[[:space:]]Splatting/4e25be1e-4424-404b-bf37-a5e285fb3e55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Epistemic[[:space:]]and[[:space:]]Aleatoric[[:space:]]Uncertainty[[:space:]]for[[:space:]]Active[[:space:]]Open-Set[[:space:]]Annotation_[[:space:]]An[[:space:]]Energy-Based[[:space:]]Approach/10abeb08-de67-449d-8076-ebf6c3a92a89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Few-Shot[[:space:]]Adaptation[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models[[:space:]]in[[:space:]]Two[[:space:]]Stages/c0cd0b00-fe43-4e96-acf8-330c1fe9d2a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Lanes[[:space:]]and[[:space:]]Points[[:space:]]in[[:space:]]Complex[[:space:]]Scenarios[[:space:]]for[[:space:]]Monocular[[:space:]]3D[[:space:]]Lane[[:space:]]Detection/d7cc5255-9b64-465e-a450-baef9e2ee11b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Noisy[[:space:]]Video-Text[[:space:]]Retrieval[[:space:]]via[[:space:]]Relation-aware[[:space:]]Alignment/f1a62f24-f0a6-4eb9-9d05-5a9a7b29250f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Personalized[[:space:]]Aesthetics[[:space:]]Assessment_[[:space:]]Employing[[:space:]]Physique[[:space:]]Aesthetics[[:space:]]Assessment[[:space:]]as[[:space:]]An[[:space:]]Exemplification/27681670-d5a0-4394-b7cd-05840add3200_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Query-based[[:space:]]Transformer[[:space:]]for[[:space:]]Continual[[:space:]]Image[[:space:]]Segmentation/e644b11e-e07a-4b4e-bd5d-adb2f2e84c8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Reconstruction[[:space:]]and[[:space:]]Denoising[[:space:]]in[[:space:]]the[[:space:]]Dark_[[:space:]]New[[:space:]]Perspective,[[:space:]]General[[:space:]]Architecture[[:space:]]and[[:space:]]Beyond/4d1b3a86-7e02-4ae4-b691-f521baae205d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Spiking[[:space:]]Self-Attention[[:space:]]Mechanism_[[:space:]]Implementing[[:space:]]a-XNOR[[:space:]]Similarity[[:space:]]Calculation[[:space:]]in[[:space:]]Spiking[[:space:]]Transformers/c6c26e66-597b-40a6-9ec6-3b7ad2d57585_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Temporal[[:space:]]Fusion[[:space:]]with[[:space:]]a[[:space:]]Unified[[:space:]]Gradient[[:space:]]Descent[[:space:]]View[[:space:]]for[[:space:]]3D[[:space:]]Semantic[[:space:]]Occupancy[[:space:]]Prediction/feebf5ad-b6f5-4b14-83a6-239dc30e9923_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Token[[:space:]]Reduction[[:space:]]with[[:space:]]Parameter-Efficient[[:space:]]Fine-Tuning[[:space:]]in[[:space:]]ViT[[:space:]]for[[:space:]]Pixel-Level[[:space:]]Tasks/eb127117-5bec-4f6c-bb50-b26dea016ae1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Training[[:space:]]for[[:space:]]De-biasing[[:space:]]Text-to-Image[[:space:]]Generation_[[:space:]]Unlocking[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Stable[[:space:]]Diffusion/a5b61659-5ec4-43a0-aa69-ae89743b52c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Vision-Language[[:space:]]Model[[:space:]]in[[:space:]]Face[[:space:]]Forensics_[[:space:]]Multi-Modal[[:space:]]Interpretable[[:space:]]Forged[[:space:]]Face[[:space:]]Detector/eda2db57-f778-410b-afb1-425beb814441_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]the[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]of[[:space:]]Multi-Exit[[:space:]]Neural[[:space:]]Networks[[:space:]]in[[:space:]]an[[:space:]]Attack-Defense[[:space:]]Game/a27c6176-fc23-45c8-9f0a-f03bb13fbc0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Retrieving[[:space:]]Semantics[[:space:]]from[[:space:]]the[[:space:]]Deep_[[:space:]]an[[:space:]]RAG[[:space:]]Solution[[:space:]]for[[:space:]]Gesture[[:space:]]Synthesis/88c9feef-7c53-43c5-b296-29917beb1f13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revealing[[:space:]]Key[[:space:]]Details[[:space:]]to[[:space:]]See[[:space:]]Differences_[[:space:]]A[[:space:]]Novel[[:space:]]Prototypical[[:space:]]Perspective[[:space:]]for[[:space:]]Skeleton-based[[:space:]]Action[[:space:]]Recognition/70ef18cb-9617-4c87-bb50-9e4a33f73835_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reversible[[:space:]]Decoupling[[:space:]]Network[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Reflection[[:space:]]Removal/cb5f7811-9f1a-4946-be0b-c9ea0e84d324_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reversing[[:space:]]Flow[[:space:]]for[[:space:]]Image[[:space:]]Restoration/50b19daf-f826-4e2a-a90c-45ca7f4541ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revisiting[[:space:]]Audio-Visual[[:space:]]Segmentation[[:space:]]with[[:space:]]Vision-Centric[[:space:]]Transformer/305970d3-f01d-4446-9c30-4c797eece8d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revisiting[[:space:]]Backdoor[[:space:]]Attacks[[:space:]]against[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]from[[:space:]]Domain[[:space:]]Shift/6695c6a6-7bff-4bf3-b4ec-29d3a4cf8e2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revisiting[[:space:]]Fairness[[:space:]]in[[:space:]]Multitask[[:space:]]Learning_[[:space:]]A[[:space:]]Performance-Driven[[:space:]]Approach[[:space:]]for[[:space:]]Variance[[:space:]]Reduction/2d811a26-5244-423b-9cba-a3cd310a8421_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revisiting[[:space:]]Generative[[:space:]]Replay[[:space:]]for[[:space:]]Class[[:space:]]Incremental[[:space:]]Object[[:space:]]Detection/a864c1cc-6dc5-45d9-9ce6-e60f6f67bb80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revisiting[[:space:]]MAE[[:space:]]Pre-training[[:space:]]for[[:space:]]3D[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/15979a23-fd47-46cd-af1e-fb88f4aedb80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revisiting[[:space:]]Source-Free[[:space:]]Domain[[:space:]]Adaptation_[[:space:]]Insights[[:space:]]into[[:space:]]Representativeness,[[:space:]]Generalization,[[:space:]]and[[:space:]]Variety/fce4f98f-b621-4d4b-a2c2-66dece952d30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reward[[:space:]]Fine-Tuning[[:space:]]Two-Step[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Learning[[:space:]]Differentiable[[:space:]]Latent-Space[[:space:]]Surrogate[[:space:]]Reward/f2bbd944-cf24-4e8a-ac35-22aec0a3d276_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RigGS_[[:space:]]Rigging[[:space:]]of[[:space:]]3D[[:space:]]Gaussians[[:space:]]for[[:space:]]Modeling[[:space:]]Articulated[[:space:]]Objects[[:space:]]in[[:space:]]Videos/54f9c1c4-65af-45fc-8387-c8ac4ee992af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RipVIS_[[:space:]]Rip[[:space:]]Currents[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation[[:space:]]Benchmark[[:space:]]for[[:space:]]Beach[[:space:]]Monitoring[[:space:]]and[[:space:]]Safety/807a66bb-8e84-4c82-b66c-db38c6d68b64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RivuletMLP_[[:space:]]An[[:space:]]MLP-based[[:space:]]Architecture[[:space:]]for[[:space:]]Efficient[[:space:]]Compressed[[:space:]]Video[[:space:]]Quality[[:space:]]Enhancement/ae3db28c-b348-4bce-a6b5-d5ac66d964b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoGSplat_[[:space:]]Learning[[:space:]]Robust[[:space:]]Generalizable[[:space:]]Human[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Sparse[[:space:]]Multi-View[[:space:]]Images/daff6056-61e3-48e6-9096-04a409e5ac1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoadSocial_[[:space:]]A[[:space:]]Diverse[[:space:]]VideoQA[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Road[[:space:]]Event[[:space:]]Understanding[[:space:]]from[[:space:]]Social[[:space:]]Video[[:space:]]Narratives/687c97d8-c696-4322-8051-d52846fa352d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RobSense_[[:space:]]A[[:space:]]Robust[[:space:]]Multi-modal[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]with[[:space:]]Static,[[:space:]]Temporal,[[:space:]]and[[:space:]]Incomplete[[:space:]]Data[[:space:]]Adaptability/65de715c-0b6a-4e44-9e84-319fd880b85e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboBrain_[[:space:]]A[[:space:]]Unified[[:space:]]Brain[[:space:]]Model[[:space:]]for[[:space:]]Robotic[[:space:]]Manipulation[[:space:]]from[[:space:]]Abstract[[:space:]]to[[:space:]]Concrete/6a11b5b0-980d-4763-bdfc-a1fb3870ef4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboGround_[[:space:]]Robotic[[:space:]]Manipulation[[:space:]]with[[:space:]]Grounded[[:space:]]Vision-Language[[:space:]]Priors/5b8d841a-e59c-4d3e-9e93-e0944653f688_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboPEPP_[[:space:]]Vision-Based[[:space:]]Robot[[:space:]]Pose[[:space:]]and[[:space:]]Joint[[:space:]]Angle[[:space:]]Estimation[[:space:]]through[[:space:]]Embedding[[:space:]]Predictive[[:space:]]Pre-Training/ea394626-30e1-4fc3-b8ce-9098cf315490_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboSense_[[:space:]]Large-scale[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Egocentric[[:space:]]Robot[[:space:]]Perception[[:space:]]and[[:space:]]Navigation[[:space:]]in[[:space:]]Crowded[[:space:]]and[[:space:]]Unstructured[[:space:]]Environments/91c96d3f-b83a-45b4-86fe-41a3239662a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboSpatial_[[:space:]]Teaching[[:space:]]Spatial[[:space:]]Understanding[[:space:]]to[[:space:]]2D[[:space:]]and[[:space:]]3D[[:space:]]Vision-Language[[:space:]]Models[[:space:]]for[[:space:]]Robotics/5d4e51df-4129-4910-98dc-9b9a6f947e23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboTwin_[[:space:]]Dual-Arm[[:space:]]Robot[[:space:]]Benchmark[[:space:]]with[[:space:]]Generative[[:space:]]Digital[[:space:]]Twins/09e5b98b-6169-4d08-91e8-766e76ef341a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robotic[[:space:]]Visual[[:space:]]Instruction/44bdfa3f-35a5-4ed8-984f-3dc432136497_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]3D[[:space:]]Shape[[:space:]]Reconstruction[[:space:]]in[[:space:]]Zero-Shot[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]in[[:space:]]the[[:space:]]Wild/f6d9eb1a-8d1b-44fe-a100-7679dcaf2867_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]Audio-Visual[[:space:]]Segmentation[[:space:]]via[[:space:]]Audio-Guided[[:space:]]Visual[[:space:]]Convergent[[:space:]]Alignment/0080b1bb-bfcf-445a-92e5-ee398a4ea469_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]Message[[:space:]]Embedding[[:space:]]via[[:space:]]Attention[[:space:]]Flow-Based[[:space:]]Steganography/7719faf7-1cb1-4964-811e-ac0e77b1d88a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]Multi-Object[[:space:]]4D[[:space:]]Generation[[:space:]]for[[:space:]]In-the-wild[[:space:]]Videos/5608bb67-dff7-4952-a0b9-ba92a1ee0b94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]Multimodal[[:space:]]Survival[[:space:]]Prediction[[:space:]]with[[:space:]]Conditional[[:space:]]Latent[[:space:]]Differentiation[[:space:]]Variational[[:space:]]AutoEncoder/67d06a41-5db8-4d3b-a154-67ea148145c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust-MVTON_[[:space:]]Learning[[:space:]]Cross-Pose[[:space:]]Feature[[:space:]]Alignment[[:space:]]and[[:space:]]Fusion[[:space:]]for[[:space:]]Robust[[:space:]]Multi-View[[:space:]]Virtual[[:space:]]Try-On/e4a7a6ed-e1ee-46d8-ba87-bfe4f3f1f820_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoomPainter_[[:space:]]View-Integrated[[:space:]]Diffusion[[:space:]]for[[:space:]]Consistent[[:space:]]Indoor[[:space:]]Scene[[:space:]]Texturing/cdff11d5-3d3b-4439-92c3-e436eb05d496_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoomTour3D_[[:space:]]Geometry-Aware[[:space:]]Video-Instruction[[:space:]]Tuning[[:space:]]for[[:space:]]Embodied[[:space:]]Navigation/0dcfb182-7bce-42ad-a99b-7e9e0d669261_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rotation-Equivariant[[:space:]]Self-Supervised[[:space:]]Method[[:space:]]in[[:space:]]Image[[:space:]]Denoising/144f4971-bf0f-4465-8bbc-8b32834cb27e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/S2D-LFE_[[:space:]]Sparse-to-Dense[[:space:]]Light[[:space:]]Field[[:space:]]Event[[:space:]]Generation/f17a4408-3f4e-4639-92d7-cce01abd5b00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/S2Gaussian_[[:space:]]Sparse-View[[:space:]]Super-Resolution[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/f341023e-9ff1-40bd-a8ab-93db55d08132_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/S4-Driver_[[:space:]]Scalable[[:space:]]Self-Supervised[[:space:]]Driving[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]with[[:space:]]Spatio-Temporal[[:space:]]Visual[[:space:]]Representation/20e2679c-f6b9-4a84-9623-55af78c15f58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SACB-Net_[[:space:]]Spatial-awareness[[:space:]]Convolutions[[:space:]]for[[:space:]]Medical[[:space:]]Image[[:space:]]Registration/91fc8aba-6a9b-4595-ad8c-fb4d819f6f44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAIST_[[:space:]]Segment[[:space:]]Any[[:space:]]Infrared[[:space:]]Small[[:space:]]Target[[:space:]]Model[[:space:]]Guided[[:space:]]by[[:space:]]Contrastive[[:space:]]Language-Image[[:space:]]Pretraining/230aeda6-22e8-4ab6-b4a7-19b2bc75fac3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SALAD_[[:space:]]Skeleton-aware[[:space:]]Latent[[:space:]]Diffusion[[:space:]]for[[:space:]]Text-driven[[:space:]]Motion[[:space:]]Generation[[:space:]]and[[:space:]]Editing/56cf0cf2-143c-481b-820b-cc5bf0c85862_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SALOVA_[[:space:]]Segment-Augmented[[:space:]]Long[[:space:]]Video[[:space:]]Assistant[[:space:]]for[[:space:]]Targeted[[:space:]]Retrieval[[:space:]]and[[:space:]]Routing[[:space:]]in[[:space:]]Long-Form[[:space:]]Video[[:space:]]Analysis/df919c36-d04e-4ca1-a80f-80ab910d4f79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAM-I2V_[[:space:]]Upgrading[[:space:]]SAM[[:space:]]to[[:space:]]Support[[:space:]]Promptable[[:space:]]Video[[:space:]]Segmentation[[:space:]]with[[:space:]]Less[[:space:]]than[[:space:]]0.2%[[:space:]]Training[[:space:]]Cost/ca4e7e79-4820-4add-9b9e-7558abdc9127_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAM-REF_[[:space:]]Introducing[[:space:]]Image-Prompt[[:space:]]Synergy[[:space:]]during[[:space:]]Interaction[[:space:]]for[[:space:]]Detail[[:space:]]Enhancement[[:space:]]in[[:space:]]the[[:space:]]Segment[[:space:]]Anything[[:space:]]Model/a959cab0-5bb3-4e06-9751-4a034dc7813b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAM2-LOVE_[[:space:]]Segment[[:space:]]Anything[[:space:]]Model[[:space:]]2[[:space:]]in[[:space:]]Language-aided[[:space:]]Audio-Visual[[:space:]]Scenes/dc4d562e-b5c0-4f65-86f2-b983c6d324de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAM2Object_[[:space:]]Consolidating[[:space:]]View[[:space:]]Consistency[[:space:]]via[[:space:]]SAM2[[:space:]]for[[:space:]]Zero-Shot[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation/663b1e37-f7c9-4cbb-87f6-8d16e84f1d5e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAMBLE_[[:space:]]Shape-Specific[[:space:]]Point[[:space:]]Cloud[[:space:]]Sampling[[:space:]]for[[:space:]]an[[:space:]]Optimal[[:space:]]Trade-Off[[:space:]]Between[[:space:]]Local[[:space:]]Detail[[:space:]]and[[:space:]]Global[[:space:]]Uniformity/10d62ba6-9e83-4677-8d7a-788b8bf292ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAMWISE_[[:space:]]Infusing[[:space:]]Wisdom[[:space:]]in[[:space:]]SAM2[[:space:]]for[[:space:]]Text-Driven[[:space:]]Video[[:space:]]Segmentation/ded78dc8-8abb-4efd-9adb-f4bf8c96b90f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAR3D_[[:space:]]Autoregressive[[:space:]]3D[[:space:]]Object[[:space:]]Generation[[:space:]]and[[:space:]]Understanding[[:space:]]via[[:space:]]Multi-scale[[:space:]]3D[[:space:]]VQVAE/742f6d19-8530-4fec-8c8e-94f82f4b2720_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SASep_[[:space:]]Saliency-Aware[[:space:]]Structured[[:space:]]Separation[[:space:]]of[[:space:]]Geometry[[:space:]]and[[:space:]]Feature[[:space:]]for[[:space:]]Open[[:space:]]Set[[:space:]]Learning[[:space:]]on[[:space:]]Point[[:space:]]Clouds/eddf9d0f-a790-4fc2-8177-c2291f25615a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAT-HMR_[[:space:]]Real-Time[[:space:]]Multi-Person[[:space:]]3D[[:space:]]Mesh[[:space:]]Estimation[[:space:]]via[[:space:]]Scale-Adaptive[[:space:]]Tokens/eab4bc33-3ff9-4b95-b35e-cb1cb0f44fba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SATA_[[:space:]]Spatial[[:space:]]Autocorrelation[[:space:]]Token[[:space:]]Analysis[[:space:]]for[[:space:]]Enhancing[[:space:]]the[[:space:]]Robustness[[:space:]]of[[:space:]]Vision[[:space:]]Transformers/5e97f7eb-5402-4736-8030-7dfbcb7e2397_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SCAP_[[:space:]]Transductive[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]via[[:space:]]Supportive[[:space:]]Clique-based[[:space:]]Attribute[[:space:]]Prompting/f1fd7643-840d-469e-b4c4-32935040de8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SCFlow2_[[:space:]]Plug-and-Play[[:space:]]Object[[:space:]]Pose[[:space:]]Refiner[[:space:]]with[[:space:]]Shape-Constraint[[:space:]]Scene[[:space:]]Flow/bd2181f3-adaf-406c-ac89-ebbf9c811d2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SCSA_[[:space:]]A[[:space:]]Plug-and-Play[[:space:]]Semantic[[:space:]]Continuous-Sparse[[:space:]]Attention[[:space:]]for[[:space:]]Arbitrary[[:space:]]Semantic[[:space:]]Style[[:space:]]Transfer/fae97f07-7418-476a-be91-97b643e54a78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SCSegamba_[[:space:]]Lightweight[[:space:]]Structure-Aware[[:space:]]Vision[[:space:]]Mamba[[:space:]]for[[:space:]]Crack[[:space:]]Segmentation[[:space:]]in[[:space:]]Structures/2870bb13-5ce1-4a70-8677-d49f0f5cde37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SDBF_[[:space:]]Steep-Decision-Boundary[[:space:]]Fingerprinting[[:space:]]for[[:space:]]Hard-Label[[:space:]]Tampering[[:space:]]Detection[[:space:]]of[[:space:]]DNN[[:space:]]Models/6e3f109f-6d3c-482c-aac1-fe12d944b4cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SDGOCC_[[:space:]]Semantic[[:space:]]and[[:space:]]Depth-Guided[[:space:]]Bird's-Eye[[:space:]]View[[:space:]]Transformation[[:space:]]for[[:space:]]3D[[:space:]]Multimodal[[:space:]]Occupancy[[:space:]]Prediction/14986d14-2105-4389-9a81-6b64509e2379_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SEAL_[[:space:]]Semantic[[:space:]]Attention[[:space:]]Learning[[:space:]]for[[:space:]]Long[[:space:]]Video[[:space:]]Representation/382a607d-24da-4509-8433-748d58e2f990_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SEC-Prompt_SEmantic[[:space:]]Complementary[[:space:]]Prompting[[:space:]]for[[:space:]]Few-Shot[[:space:]]Class-Incremental[[:space:]]Learning/fea93fa7-25dd-4a35-b985-6366f8b59366_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SEEN-DA_[[:space:]]SEmantic[[:space:]]ENtropy[[:space:]]guided[[:space:]]Domain-aware[[:space:]]Attention[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detection/48c4d107-49aa-405e-ad25-28dd7ace0585_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SET_[[:space:]]Spectral[[:space:]]Enhancement[[:space:]]for[[:space:]]Tiny[[:space:]]Object[[:space:]]Detection/70fd0953-a894-43a1-b9c6-1b75dbe9ffb2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SF2T_[[:space:]]Self-supervised[[:space:]]Fragment[[:space:]]Finetuning[[:space:]]of[[:space:]]Video-LLMs[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Understanding/d4e22b54-75f3-40d9-971f-0e4ee1868232_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SF3D_[[:space:]]Stable[[:space:]]Fast[[:space:]]3D[[:space:]]Mesh[[:space:]]Reconstruction[[:space:]]with[[:space:]]UV-unwrapping[[:space:]]and[[:space:]]Illumination[[:space:]]Disentanglement/3b32e463-bad2-49dc-a4f1-d9f41d2a1f1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SFDM_[[:space:]]Robust[[:space:]]Decomposition[[:space:]]of[[:space:]]Geometry[[:space:]]and[[:space:]]Reflectance[[:space:]]for[[:space:]]Realistic[[:space:]]Face[[:space:]]Rendering[[:space:]]from[[:space:]]Sparse-view[[:space:]]Images/54e9e069-f5ec-481a-a9c4-0b86f22aae79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SGC-Net_[[:space:]]Stratified[[:space:]]Granular[[:space:]]Comparison[[:space:]]Network[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]HOI[[:space:]]Detection/de53cd58-a64d-437d-bffd-56e874a1fc12_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SGCR_[[:space:]]Spherical[[:space:]]Gaussians[[:space:]]for[[:space:]]Efficient[[:space:]]3D[[:space:]]Curve[[:space:]]Reconstruction/b3444c69-4274-4c13-82cc-b440a9e78c57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SGFormer_[[:space:]]Satellite-Ground[[:space:]]Fusion[[:space:]]for[[:space:]]3D[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion/fa8f1162-46b6-4b8d-ab34-ea04e929ede5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SGSST_[[:space:]]Scaling[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Style[[:space:]]Transfer/5b7056b2-fc5b-410d-b257-8d1a879ecc29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SIDA_[[:space:]]Social[[:space:]]Media[[:space:]]Image[[:space:]]Deepfake[[:space:]]Detection,[[:space:]]Localization[[:space:]]and[[:space:]]Explanation[[:space:]]with[[:space:]]Large[[:space:]]Multimodal[[:space:]]Model/d78f919f-a503-4ae4-a9af-9158aa41b47d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SILMM_[[:space:]]Self-Improving[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models[[:space:]]for[[:space:]]Compositional[[:space:]]Text-to-Image[[:space:]]Generation/d6a00039-382c-4d20-89db-a164b6033fad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SINR_[[:space:]]Sparsity[[:space:]]Driven[[:space:]]Compressed[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations/a675d8dc-7039-4fed-b302-de9ebbbd1aed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SIR-DIFF_[[:space:]]Sparse[[:space:]]Image[[:space:]]Sets[[:space:]]Restoration[[:space:]]with[[:space:]]Multi-View[[:space:]]Diffusion[[:space:]]Model/77b2d674-0341-4ee8-8978-dbba54ab0a91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SKDream_[[:space:]]Controllable[[:space:]]Multi-view[[:space:]]and[[:space:]]3D[[:space:]]Generation[[:space:]]with[[:space:]]Arbitrary[[:space:]]Skeletons/7b2caa84-1e0b-47fd-a90e-b20604143574_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SKE-Layout_[[:space:]]Spatial[[:space:]]Knowledge[[:space:]]Enhanced[[:space:]]Layout[[:space:]]Generation[[:space:]]with[[:space:]]LLMs/acd7c271-45e8-43e6-a57e-fd2570599f21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SLADE_[[:space:]]Shielding[[:space:]]against[[:space:]]Dual[[:space:]]Exploits[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/18172bcf-3306-49a6-8b62-dc7f40cb507d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SLAM3R_[[:space:]]Real-Time[[:space:]]Dense[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]from[[:space:]]Monocular[[:space:]]RGB[[:space:]]Videos/2aa67e72-2cf5-4fb7-b0f5-0552ddd9811a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SLVR_[[:space:]]Super-Light[[:space:]]Visual[[:space:]]Reconstruction[[:space:]]via[[:space:]]Blueprint[[:space:]]Controllable[[:space:]]Convolutions[[:space:]]and[[:space:]]Exploring[[:space:]]Feature[[:space:]]Diversity[[:space:]]Representation/c7aef326-130c-4b70-8496-8a5c0eb43ee8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SMILE_[[:space:]]Infusing[[:space:]]Spatial[[:space:]]and[[:space:]]Motion[[:space:]]Semantics[[:space:]]in[[:space:]]Masked[[:space:]]Video[[:space:]]Learning/6f3fc16f-e100-416b-9a89-673ecf151cb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SMTPD_[[:space:]]A[[:space:]]New[[:space:]]Benchmark[[:space:]]for[[:space:]]Temporal[[:space:]]Prediction[[:space:]]of[[:space:]]Social[[:space:]]Media[[:space:]]Popularity/3f53df77-cd61-4f83-8a2e-0eb089935375_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SOAP_[[:space:]]Vision-Centric[[:space:]]3D[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion[[:space:]]with[[:space:]]Scene-Adaptive[[:space:]]Decoder[[:space:]]and[[:space:]]Occluded[[:space:]]Region-Aware[[:space:]]View[[:space:]]Projection/399b8632-9ed1-4ab5-bef3-002cbfb4c36d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SOGS_[[:space:]]Second-Order[[:space:]]Anchor[[:space:]]for[[:space:]]Advanced[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/c50dd783-fbc4-44df-8116-14786787e34e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SOLAMI_[[:space:]]Social[[:space:]]Vision-Language-Action[[:space:]]Modeling[[:space:]]for[[:space:]]Immersive[[:space:]]Interaction[[:space:]]with[[:space:]]3D[[:space:]]Autonomous[[:space:]]Characters/268c1c21-7560-456f-8ae1-2ae010e66b03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SOLVE_[[:space:]]Synergy[[:space:]]of[[:space:]]Language-Vision[[:space:]]and[[:space:]]End-to-End[[:space:]]Networks[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/0ded0517-3ff0-4f0f-aa7e-4de5fe3d35e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SP3D_[[:space:]]Boosting[[:space:]]Sparsely-Supervised[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Accurate[[:space:]]Cross-Modal[[:space:]]Semantic[[:space:]]Prompts/dd67abde-1a08-4102-9911-744b35ee1a19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SPA-VL_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Safety[[:space:]]Preference[[:space:]]Alignment[[:space:]]Dataset[[:space:]]for[[:space:]]Vision[[:space:]]Language[[:space:]]Models/22cd485b-929e-4951-969d-47a8232e9e43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SPAR3D_[[:space:]]Stable[[:space:]]Point-Aware[[:space:]]Reconstruction[[:space:]]of[[:space:]]3D[[:space:]]Objects[[:space:]]from[[:space:]]Single[[:space:]]Images/3000205b-bdbd-4656-8ffc-c648c9a58288_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SPARC_[[:space:]]Score[[:space:]]Prompting[[:space:]]and[[:space:]]Adaptive[[:space:]]Fusion[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Multi-Label[[:space:]]Recognition[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models/4c63b13f-f035-4408-a7c4-a7f777a69d2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SPARS3R_[[:space:]]Semantic[[:space:]]Prior[[:space:]]Alignment[[:space:]]and[[:space:]]Regularization[[:space:]]for[[:space:]]Sparse[[:space:]]3D[[:space:]]Reconstruction/ad2ca2d2-676a-49f1-84fe-0212a262ed5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SPC-GS_[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]Semantic-Prompt[[:space:]]Consistency[[:space:]]for[[:space:]]Indoor[[:space:]]Open-World[[:space:]]Free-view[[:space:]]Synthesis[[:space:]]from[[:space:]]Sparse[[:space:]]Inputs/d312a2c0-c1b5-4f3c-8ba0-826d0bc6f893_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SPMTrack_[[:space:]]Spatio-Temporal[[:space:]]Parameter-Efficient[[:space:]]Fine-Tuning[[:space:]]with[[:space:]]Mixture[[:space:]]of[[:space:]]Experts[[:space:]]for[[:space:]]Scalable[[:space:]]Visual[[:space:]]Tracking/4755d93e-1740-4309-977d-6ab1138fe42c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SSHNet_[[:space:]]Unsupervised[[:space:]]Cross-modal[[:space:]]Homography[[:space:]]Estimation[[:space:]]via[[:space:]]Problem[[:space:]]Reformulation[[:space:]]and[[:space:]]Split[[:space:]]Optimization/ae624eae-7853-4995-bcaa-12d94352fbaa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STAA-SNN_[[:space:]]Spatial-Temporal[[:space:]]Attention[[:space:]]Aggregator[[:space:]]for[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks/5435aeb8-d79e-4d0b-a257-264caa74e5cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STAR-Edge_[[:space:]]Structure-aware[[:space:]]Local[[:space:]]Spherical[[:space:]]Curve[[:space:]]Representation[[:space:]]for[[:space:]]Thin-walled[[:space:]]Edge[[:space:]]Extraction[[:space:]]from[[:space:]]Unstructured[[:space:]]Point[[:space:]]Clouds/f9f356eb-ac90-4c6e-8326-94d9a1d7d184_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STCOcc_[[:space:]]Sparse[[:space:]]Spatial-Temporal[[:space:]]Cascade[[:space:]]Renovation[[:space:]]for[[:space:]]3D[[:space:]]Occupancy[[:space:]]and[[:space:]]Scene[[:space:]]Flow[[:space:]]Prediction/8219c456-c593-40fc-a736-867ed4131468_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STDD_[[:space:]]Spatio-Temporal[[:space:]]Dual[[:space:]]Diffusion[[:space:]]for[[:space:]]Video[[:space:]]Generation/6f423d46-a407-4dbd-85d4-7aa0894a46ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STEPS_[[:space:]]Sequential[[:space:]]Probability[[:space:]]Tensor[[:space:]]Estimation[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Hard[[:space:]]Prompt[[:space:]]Search/d4154c3e-e954-409d-acb1-94c4437c4fa3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STEP_[[:space:]]Enhancing[[:space:]]Video-LLMs'[[:space:]]Compositional[[:space:]]Reasoning[[:space:]]by[[:space:]]Spatio-Temporal[[:space:]]Graph-guided[[:space:]]Self-Training/d113dfa2-5fa3-47d5-b263-22bb5886698e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STEREO_[[:space:]]A[[:space:]]Two-Stage[[:space:]]Framework[[:space:]]for[[:space:]]Adversarially[[:space:]]Robust[[:space:]]Concept[[:space:]]Erasing[[:space:]]from[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/e40006c2-d2fd-4576-a2c9-5d9a49b32da8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STING-BEE_[[:space:]]Towards[[:space:]]Vision-Language[[:space:]]Model[[:space:]]for[[:space:]]Real-World[[:space:]]X-ray[[:space:]]Baggage[[:space:]]Security[[:space:]]Inspection/cfcd6d70-4afd-4c18-92a2-5c3587a10dec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STINR_[[:space:]]Deciphering[[:space:]]Spatial[[:space:]]Transcriptomics[[:space:]]via[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representation/a40a9ec1-abc4-4c0f-999d-14cdbdd09eae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STOP_[[:space:]]Integrated[[:space:]]Spatial-Temporal[[:space:]]Dynamic[[:space:]]Prompting[[:space:]]for[[:space:]]Video[[:space:]]Understanding/9d941640-725f-4c0e-8246-c2e201f96bee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STPro_[[:space:]]Spatial[[:space:]]and[[:space:]]Temporal[[:space:]]Progressive[[:space:]]Learning[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Spatio-Temporal[[:space:]]Grounding/f34924bd-55df-4f96-be22-f66d6fe3d0c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STiL_[[:space:]]Semi-supervised[[:space:]]Tabular-Image[[:space:]]Learning[[:space:]]for[[:space:]]Comprehensive[[:space:]]Task-Relevant[[:space:]]Information[[:space:]]Exploration[[:space:]]in[[:space:]]Multimodal[[:space:]]Classification/53e07557-08f3-44c3-bb9e-3b90270a0ed0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SUM[[:space:]]Parts_[[:space:]]Benchmarking[[:space:]]Part-Level[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]of[[:space:]]Urban[[:space:]]Meshes/53ef3190-1f71-4686-9f44-dccf7ae86a0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SURGEON_[[:space:]]Memory-Adaptive[[:space:]]Fully[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]via[[:space:]]Dynamic[[:space:]]Activation[[:space:]]Sparsity/41142d7e-4f3d-4224-8c97-ab43102564f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SVDC_[[:space:]]Consistent[[:space:]]Direct[[:space:]]Time-of-Flight[[:space:]]Video[[:space:]]Depth[[:space:]]Completion[[:space:]]with[[:space:]]Frequency[[:space:]]Selective[[:space:]]Fusion/a2dd34ec-f187-4483-b676-d347e61cf844_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SVFR_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Generalized[[:space:]]Video[[:space:]]Face[[:space:]]Restoration/fc349add-9e72-49d1-93de-20cefb31be3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SVG-IR_[[:space:]]Spatially-Varying[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Inverse[[:space:]]Rendering/80580159-7dd0-49b5-ba90-b2f7cc91455d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SVLTA_[[:space:]]Benchmarking[[:space:]]Vision-Language[[:space:]]Temporal[[:space:]]Alignment[[:space:]]via[[:space:]]Synthetic[[:space:]]Video[[:space:]]Situation/b5e7d9b5-4d64-4516-b2a2-37fa4cda932c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/S^3-Face_[[:space:]]SSS-Compliant[[:space:]]Facial[[:space:]]Reflectance[[:space:]]Estimation[[:space:]]via[[:space:]]Diffusion[[:space:]]Priors/11635983-6d0d-4bb1-8c33-7a1045909cc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SaMam_[[:space:]]Style-aware[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]for[[:space:]]Arbitrary[[:space:]]Image[[:space:]]Style[[:space:]]Transfer/1c8cb17b-3dfc-49dd-ac06-cca252793f15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Saliuitl_[[:space:]]Ensemble[[:space:]]Salience[[:space:]]Guided[[:space:]]Recovery[[:space:]]of[[:space:]]Adversarial[[:space:]]Patches[[:space:]]against[[:space:]]CNNs/22f430fa-194a-455e-8375-338e0c52189c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Samba_[[:space:]]A[[:space:]]Unified[[:space:]]Mamba-based[[:space:]]Framework[[:space:]]for[[:space:]]General[[:space:]]Salient[[:space:]]Object[[:space:]]Detection/24e1fb80-7004-4fb3-b82a-a834fd98c335_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sample-[[:space:]]and[[:space:]]Parameter-Efficient[[:space:]]Auto-Regressive[[:space:]]Image[[:space:]]Models/58b934cb-8cd1-4944-bba9-1b386e3e0220_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sampling[[:space:]]Innovation-Based[[:space:]]Adaptive[[:space:]]Compressive[[:space:]]Sensing/270eb60d-82ca-4c8f-804e-9eb5c53f177b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SapiensID_[[:space:]]Foundation[[:space:]]for[[:space:]]Human[[:space:]]Recognition/436e1ae5-df46-441d-8f75-46d66bab57db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Satellite[[:space:]]Observations[[:space:]]Guided[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Accurate[[:space:]]Meteorological[[:space:]]States[[:space:]]at[[:space:]]Arbitrary[[:space:]]Resolution/48f3dfa4-8440-4128-91c9-4a4337a23aee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Satellite[[:space:]]to[[:space:]]GroundScape[[:space:]]-[[:space:]]Large-scale[[:space:]]Consistent[[:space:]]Ground[[:space:]]View[[:space:]]Generation[[:space:]]from[[:space:]]Satellite[[:space:]]Views/56bdbc5b-2d91-4a58-870d-522f2ea0edad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ScaMo_[[:space:]]Exploring[[:space:]]the[[:space:]]Scaling[[:space:]]Law[[:space:]]in[[:space:]]Autoregressive[[:space:]]Motion[[:space:]]Generation[[:space:]]Model/c7d4dd82-2498-4295-8b9c-b9f9ea436680_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scalable[[:space:]]Autoregressive[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/5ce52cb7-3938-40e0-abc8-b8e27e318313_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scalable[[:space:]]Video-to-Dataset[[:space:]]Generation[[:space:]]for[[:space:]]Cross-Platform[[:space:]]Mobile[[:space:]]Agents/d2653592-8485-47d6-9a30-de055f273633_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scale[[:space:]]Efficient[[:space:]]Training[[:space:]]for[[:space:]]Large[[:space:]]Datasets/e61cafeb-ce6d-4052-a69b-7b3b71fcf252_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ScaleLSD_[[:space:]]Scalable[[:space:]]Deep[[:space:]]Line[[:space:]]Segment[[:space:]]Detection[[:space:]]Streamlined/500f7ba0-2387-4785-89c6-87f074b3a7a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]Down[[:space:]]Text[[:space:]]Encoders[[:space:]]of[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/b010ff1b-39b9-46a6-b393-3da00dba696e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]Inference[[:space:]]Time[[:space:]]Compute[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/39e57438-984c-4358-b2b0-512a959f0fb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]Mesh[[:space:]]Generation[[:space:]]via[[:space:]]Compressive[[:space:]]Tokenization/5de76a12-300e-473f-bf22-647884a76403_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]Properties[[:space:]]of[[:space:]]Diffusion[[:space:]]Models[[:space:]]For[[:space:]]Perceptual[[:space:]]Tasks/c429c96f-e821-4e97-8458-1d1be4e1968f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]Vision[[:space:]]Pre-Training[[:space:]]to[[:space:]]4K[[:space:]]Resolution/56420e12-c6f7-42fc-8e3a-15dcc0c1aae4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]up[[:space:]]Image[[:space:]]Segmentation[[:space:]]across[[:space:]]Data[[:space:]]and[[:space:]]Tasks/138f59b0-ddc2-4d76-ab4a-a5634830350f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scenario[[:space:]]Dreamer_[[:space:]]Vectorized[[:space:]]Latent[[:space:]]Diffusion[[:space:]]for[[:space:]]Generating[[:space:]]Driving[[:space:]]Simulation[[:space:]]Environments/348d3c95-8b5c-44ba-8cc7-59df00f37d57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scene[[:space:]]Map-based[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Navigation[[:space:]]Instruction[[:space:]]Generation/6a38956b-17d1-4ba7-a0de-4c53d3237d6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scene[[:space:]]Splatter_[[:space:]]Momentum[[:space:]]3D[[:space:]]Scene[[:space:]]Generation[[:space:]]from[[:space:]]Single[[:space:]]Image[[:space:]]with[[:space:]]Video[[:space:]]Diffusion[[:space:]]Model/535a91ed-e764-4196-a77f-a7c50addc72d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scene-Centric[[:space:]]Unsupervised[[:space:]]Panoptic[[:space:]]Segmentation/15f78b82-f214-4b03-a700-fffc804b7bcf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scene-agnostic[[:space:]]Pose[[:space:]]Regression[[:space:]]for[[:space:]]Visual[[:space:]]Localization/162f73a8-ed2e-4293-ab83-1d17b773a662_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scene4U_[[:space:]]Hierarchical[[:space:]]Layered[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]from[[:space:]]Single[[:space:]]Panoramic[[:space:]]Image[[:space:]]for[[:space:]]Your[[:space:]]Immerse[[:space:]]Exploration/0539e349-10d0-40c5-9eb2-a497f951cc9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SceneCrafter_[[:space:]]Controllable[[:space:]]Multi-View[[:space:]]Driving[[:space:]]Scene[[:space:]]Editing/da28ea54-8cfc-411b-986b-78e5dcf83d75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SceneDiffuser++_[[:space:]]City-Scale[[:space:]]Traffic[[:space:]]Simulation[[:space:]]via[[:space:]]a[[:space:]]Generative[[:space:]]World[[:space:]]Model/6e81cf2b-e181-4c35-ae88-e59a97b9f6c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SceneFactor_[[:space:]]Factored[[:space:]]Latent[[:space:]]3D[[:space:]]Diffusion[[:space:]]for[[:space:]]Controllable[[:space:]]3D[[:space:]]Scene[[:space:]]Generation/673ba6ea-6a0a-4209-9ee5-ca27cde672e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SceneTAP_[[:space:]]Scene-Coherent[[:space:]]Typographic[[:space:]]Adversarial[[:space:]]Planner[[:space:]]against[[:space:]]Vision-Language[[:space:]]Models[[:space:]]in[[:space:]]Real-World[[:space:]]Environments/2d1aa1d0-76db-47c6-906d-55662b27fec9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Schedule[[:space:]]On[[:space:]]the[[:space:]]Fly_[[:space:]]Diffusion[[:space:]]Time[[:space:]]Prediction[[:space:]]for[[:space:]]Faster[[:space:]]and[[:space:]]Better[[:space:]]Image[[:space:]]Generation/dc08fc21-acc6-4605-80dd-7f4fab947ac2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Science-T2I_[[:space:]]Addressing[[:space:]]Scientific[[:space:]]Illusions[[:space:]]in[[:space:]]Image[[:space:]]Synthesis/b7f91f72-91c2-4882-bf2d-7da90a61c4d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ScribbleLight_[[:space:]]Single[[:space:]]Image[[:space:]]Indoor[[:space:]]Relighting[[:space:]]with[[:space:]]Scribbles/d02b376e-d354-4535-8aae-1c981dc30f71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SeCap_[[:space:]]Self-Calibrating[[:space:]]and[[:space:]]Adaptive[[:space:]]Prompts[[:space:]]for[[:space:]]Cross-view[[:space:]]Person[[:space:]]Re-Identification[[:space:]]in[[:space:]]Aerial-Ground[[:space:]]Networks/e3d5bf6c-adf7-407c-8671-d1c3b804379c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sea-ing[[:space:]]in[[:space:]]Low-light/ca1ae0e7-5e5c-472a-8584-f4fb3a220690_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SeaLion_[[:space:]]Semantic[[:space:]]Part-Aware[[:space:]]Latent[[:space:]]Point[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]3D[[:space:]]Generation/dc573eb4-ded7-4541-b7fb-a31729ff5996_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Search[[:space:]]and[[:space:]]Detect_[[:space:]]Training-Free[[:space:]]Long[[:space:]]Tail[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Web-Image[[:space:]]Retrieval/32861376-f7e7-4ecf-ae19-0d8ac140d38c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Secret[[:space:]]Lies[[:space:]]in[[:space:]]Color_[[:space:]]Enhancing[[:space:]]AI-Generated[[:space:]]Images[[:space:]]Detection[[:space:]]with[[:space:]]Color[[:space:]]Distribution[[:space:]]Analysis/61620578-b258-4a4b-bdb2-5534b2589aeb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/See[[:space:]]Further[[:space:]]When[[:space:]]Clear_[[:space:]]Curriculum[[:space:]]Consistency[[:space:]]Model/649869c9-1492-4fa5-b9e6-fde74a0ff1c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SeeGround_[[:space:]]See[[:space:]]and[[:space:]]Ground[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding/055671e7-4547-45ec-a1e9-5f22c89aefd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SeedVR_[[:space:]]Seeding[[:space:]]Infinity[[:space:]]in[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]Towards[[:space:]]Generic[[:space:]]Video[[:space:]]Restoration/d036330b-fab1-4aa4-9e53-3cb3f151ecb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeing[[:space:]]A[[:space:]]3D[[:space:]]World[[:space:]]in[[:space:]]A[[:space:]]Grain[[:space:]]of[[:space:]]Sand/54bd7f05-ddcc-4e4f-a2c8-7435aae76e4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeing[[:space:]]Far[[:space:]]and[[:space:]]Clearly_[[:space:]]Mitigating[[:space:]]Hallucinations[[:space:]]in[[:space:]]MLLMs[[:space:]]with[[:space:]]Attention[[:space:]]Causal[[:space:]]Decoding/eea6560f-c338-43f7-aa8e-8536f49f9237_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeing[[:space:]]More[[:space:]]with[[:space:]]Less_[[:space:]]Human-like[[:space:]]Representations[[:space:]]in[[:space:]]Vision[[:space:]]Models/1f09357f-b362-4b33-8c75-2dae5b80d48d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeing[[:space:]]Speech[[:space:]]and[[:space:]]Sound_[[:space:]]Distinguishing[[:space:]]and[[:space:]]Locating[[:space:]]Audio[[:space:]]Sources[[:space:]]in[[:space:]]Visual[[:space:]]Scenes/917b0c17-3017-4d18-852a-502eb64e4184_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeing[[:space:]]What[[:space:]]Matters_[[:space:]]Empowering[[:space:]]CLIP[[:space:]]with[[:space:]]Patch[[:space:]]Generation-to-Selection/f1bdc995-bdc3-421e-be93-ea9df8640b51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeing[[:space:]]is[[:space:]]Not[[:space:]]Believing_[[:space:]]Adversarial[[:space:]]Natural[[:space:]]Object[[:space:]]Optimization[[:space:]]for[[:space:]]Hard-Label[[:space:]]3D[[:space:]]Scene[[:space:]]Attacks/3cdf7947-e315-4616-9445-c5c62bdc4bf8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeing[[:space:]]the[[:space:]]Abstract_[[:space:]]Translating[[:space:]]the[[:space:]]Abstract[[:space:]]Language[[:space:]]for[[:space:]]Vision[[:space:]]Language[[:space:]]Models/d6183534-ef29-4f52-b58a-e7e56ad32a4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seek[[:space:]]Common[[:space:]]Ground[[:space:]]While[[:space:]]Reserving[[:space:]]Differences_[[:space:]]Semi-Supervised[[:space:]]Image-Text[[:space:]]Sentiment[[:space:]]Recognition/129ec86b-7981-45f7-94dd-dd117a01e0d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeking[[:space:]]Consistent[[:space:]]Flat[[:space:]]Minima[[:space:]]for[[:space:]]Better[[:space:]]Domain[[:space:]]Generalization[[:space:]]via[[:space:]]Refining[[:space:]]Loss[[:space:]]Landscapes/eabf1cb1-416d-4833-b397-be038a9ae636_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Simpler[[:space:]]Diffusion_[[:space:]]1.5[[:space:]]FID[[:space:]]on[[:space:]]ImageNet512[[:space:]]with[[:space:]]Pixel-space[[:space:]]Diffusion/3cf16242-5862-4aa6-864f-5aba8865f5f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Simplification[[:space:]]Is[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]against[[:space:]]Out-of-Distribution[[:space:]]Overconfidence/c4dc8f46-aeb5-4477-a94f-497bd92530cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Simulator[[:space:]]HC_[[:space:]]Regression-based[[:space:]]Online[[:space:]]Simulation[[:space:]]of[[:space:]]Starting[[:space:]]Problem-Solution[[:space:]]Pairs[[:space:]]for[[:space:]]Homotopy[[:space:]]Continuation[[:space:]]in[[:space:]]Geometric[[:space:]]Vision/d90c2114-786f-423a-b72c-4f3391aa43d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SinGS_[[:space:]]Animatable[[:space:]]Single-Image[[:space:]]Human[[:space:]]Gaussian[[:space:]]Splats[[:space:]]with[[:space:]]Kinematic[[:space:]]Priors/a1907643-3cec-4712-b625-aae92326d106_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Single[[:space:]]Domain[[:space:]]Generalization[[:space:]]for[[:space:]]Few-Shot[[:space:]]Counting[[:space:]]via[[:space:]]Universal[[:space:]]Representation[[:space:]]Matching/ef6b53d9-c810-43e9-a737-a386bc2ad915_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Six-CD_[[:space:]]Benchmarking[[:space:]]Concept[[:space:]]Removals[[:space:]]for[[:space:]]Text-to-image[[:space:]]Diffusion[[:space:]]Models/17b4b8e3-2cbd-49ce-875b-c3c7f736c51d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sketch[[:space:]]Down[[:space:]]the[[:space:]]FLOPs_[[:space:]]Towards[[:space:]]Efficient[[:space:]]Networks[[:space:]]for[[:space:]]Human[[:space:]]Sketch/d51d8c32-98f5-44d4-9674-ee9bc3d4af26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SketchAgent_[[:space:]]Language-Driven[[:space:]]Sequential[[:space:]]Sketch[[:space:]]Generation/76b8d485-0aca-43aa-b3b2-9c94133cc069_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SketchFusion_[[:space:]]Learning[[:space:]]Universal[[:space:]]Sketch[[:space:]]Features[[:space:]]through[[:space:]]Fusing[[:space:]]Foundation[[:space:]]Models/1fc738f6-a4cf-4a97-bcd3-ea7682cab3c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SketchVideo_[[:space:]]Sketch-based[[:space:]]Video[[:space:]]Generation[[:space:]]and[[:space:]]Editing/86ad126c-969d-47b0-8162-a7cc53739116_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sketchtopia_[[:space:]]A[[:space:]]Dataset[[:space:]]and[[:space:]]Foundational[[:space:]]Agents[[:space:]]for[[:space:]]Benchmarking[[:space:]]Asynchronous[[:space:]]Multimodal[[:space:]]Communication[[:space:]]with[[:space:]]Iconic[[:space:]]Feedback/fff37445-0c8d-4c24-a3bf-32cf9fcbe8ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sketchy[[:space:]]Bounding-box[[:space:]]Supervision[[:space:]]for[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation/fed9c100-3d53-4c29-809d-ba6f2288846b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SkillMimic_[[:space:]]Learning[[:space:]]Basketball[[:space:]]Interaction[[:space:]]Skills[[:space:]]from[[:space:]]Demonstrations/70ea78a5-d7df-4937-ac48-382497691b79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Skip[[:space:]]Tuning_[[:space:]]Pre-trained[[:space:]]Vision-Language[[:space:]]Models[[:space:]]are[[:space:]]Effective[[:space:]]and[[:space:]]Efficient[[:space:]]Adapters[[:space:]]Themselves/e5453c9b-356b-43d9-9454-788a3f8b346d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SkySense-O_[[:space:]]Towards[[:space:]]Open-World[[:space:]]Remote[[:space:]]Sensing[[:space:]]Interpretation[[:space:]]with[[:space:]]Vision-Centric[[:space:]]Visual-Language[[:space:]]Modeling/2415dd78-f506-4cd3-a12c-42283424590a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SleeperMark_[[:space:]]Towards[[:space:]]Robust[[:space:]]Watermark[[:space:]]against[[:space:]]Fine-Tuning[[:space:]]Text-to-image[[:space:]]Diffusion[[:space:]]Models/2530599b-741a-485a-addd-c98f8e367f9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SlideChat_[[:space:]]A[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Assistant[[:space:]]for[[:space:]]Whole-Slide[[:space:]]Pathology[[:space:]]Image[[:space:]]Understanding/5542bef9-2edf-4ac5-b8df-cf3f63aeaa6c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SmartCLIP_[[:space:]]Modular[[:space:]]Vision-language[[:space:]]Alignment[[:space:]]with[[:space:]]Identification[[:space:]]Guarantees/28f50741-407e-4ef2-b9fa-f2bf47df9fd3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SmartEraser_[[:space:]]Remove[[:space:]]Anything[[:space:]]from[[:space:]]Images[[:space:]]using[[:space:]]Masked-Region[[:space:]]Guidance/ab3bbba3-1da3-4883-858e-f047afa69964_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SnapGen-V_[[:space:]]Generating[[:space:]]a[[:space:]]Five-Second[[:space:]]Video[[:space:]]within[[:space:]]Five[[:space:]]Seconds[[:space:]]on[[:space:]]a[[:space:]]Mobile[[:space:]]Device/2089e6e3-1b02-4449-b418-985fa5c0c175_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SnapGen_[[:space:]]Taming[[:space:]]High-Resolution[[:space:]]Text-to-Image[[:space:]]Models[[:space:]]for[[:space:]]Mobile[[:space:]]Devices[[:space:]]with[[:space:]]Efficient[[:space:]]Architectures[[:space:]]and[[:space:]]Training/d6336171-7440-4517-9fbd-fdd52a399434_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SnowMaster_[[:space:]]Comprehensive[[:space:]]Real-world[[:space:]]Image[[:space:]]Desnowing[[:space:]]via[[:space:]]MLLM[[:space:]]with[[:space:]]Multi-Model[[:space:]]Feedback[[:space:]]Optimization/01611a60-cb24-4ca8-93fd-c47c38025543_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SoMA_[[:space:]]Singular[[:space:]]Value[[:space:]]Decomposed[[:space:]]Minor[[:space:]]Components[[:space:]]Adaptation[[:space:]]for[[:space:]]Domain[[:space:]]Generalizable[[:space:]]Representation[[:space:]]Learning/282504c5-c667-4202-94ca-c92386507635_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SocialGesture_[[:space:]]Delving[[:space:]]into[[:space:]]Multi-person[[:space:]]Gesture[[:space:]]Understanding/17e9a440-617f-4e5f-9712-f768185c9d5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SocialMOIF_[[:space:]]Multi-Order[[:space:]]Intention[[:space:]]Fusion[[:space:]]for[[:space:]]Pedestrian[[:space:]]Trajectory[[:space:]]Prediction/74c8547f-5c13-42b2-b27c-77ef8f52058e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Soft[[:space:]]Self-labeling[[:space:]]and[[:space:]]Potts[[:space:]]Relaxations[[:space:]]for[[:space:]]Weakly-supervised[[:space:]]Segmentation/4b6c17bc-bca2-47a9-b3a7-cd9f02074980_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SoftShadow_[[:space:]]Leveraging[[:space:]]Soft[[:space:]]Masks[[:space:]]for[[:space:]]Penumbra-Aware[[:space:]]Shadow[[:space:]]Removal/1a57ac9e-bdab-4cc5-a82c-bea51caf54d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SoftVQ-VAE_[[:space:]]Efficient[[:space:]]1-Dimensional[[:space:]]Continuous[[:space:]]Tokenizer/315fadef-93cf-49d1-961b-c20538f1f4de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Solving[[:space:]]Instance[[:space:]]Detection[[:space:]]from[[:space:]]an[[:space:]]Open-World[[:space:]]Perspective/ab6d1961-27f9-4a61-af19-064b80505b12_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sonata_[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]of[[:space:]]Reliable[[:space:]]Point[[:space:]]Representations/3e06b74b-3b4f-4789-9e14-04119c218bb3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sonic_[[:space:]]Shifting[[:space:]]Focus[[:space:]]to[[:space:]]Global[[:space:]]Audio[[:space:]]Perception[[:space:]]in[[:space:]]Portrait[[:space:]]Animation/108dd369-6bdf-4c82-9e76-1c8b1e3be7c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sound[[:space:]]Bridge_[[:space:]]Associating[[:space:]]Egocentric[[:space:]]and[[:space:]]Exocentric[[:space:]]Videos[[:space:]]via[[:space:]]Audio[[:space:]]Cues/37c295df-9911-4189-b46a-8795fdcf6a6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SoundVista_[[:space:]]Novel-View[[:space:]]Ambient[[:space:]]Sound[[:space:]]Synthesis[[:space:]]via[[:space:]]Visual-Acoustic[[:space:]]Binding/9c309271-ab78-4453-a939-59e8aa02fe41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sparse[[:space:]]Point[[:space:]]Cloud[[:space:]]Patches[[:space:]]Rendering[[:space:]]via[[:space:]]Splitting[[:space:]]2D[[:space:]]Gaussians/dd278d9d-da45-4b1e-bcc2-33b0924a1fa1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sparse[[:space:]]Voxels[[:space:]]Rasterization_[[:space:]]Real-time[[:space:]]High-fidelity[[:space:]]Radiance[[:space:]]Field[[:space:]]Rendering/a6bd5a3d-4547-4194-a227-b9c1f40e864c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sparse2DGS_[[:space:]]Geometry-Prioritized[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]from[[:space:]]Sparse[[:space:]]Views/3c5dfe86-080c-4ec1-9ca9-65bf2b3453a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SparseAlign_[[:space:]]a[[:space:]]Fully[[:space:]]Sparse[[:space:]]Framework[[:space:]]for[[:space:]]Cooperative[[:space:]]Object[[:space:]]Detection/e9f3f305-e62d-4aad-9255-49e76e43c9cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spatial[[:space:]]Transport[[:space:]]Optimization[[:space:]]by[[:space:]]Repositioning[[:space:]]Attention[[:space:]]Map[[:space:]]for[[:space:]]Training-Free[[:space:]]Text-to-Image[[:space:]]Synthesis/68174923-d215-4833-8530-d036bb11b7fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spatial-Temporal[[:space:]]Graph[[:space:]]Diffusion[[:space:]]Policy[[:space:]]with[[:space:]]Kinematic[[:space:]]Modeling[[:space:]]for[[:space:]]Bimanual[[:space:]]Robotic[[:space:]]Manipulation/8300c221-08df-4f6a-8aec-775be3434ff5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spatial457_[[:space:]]A[[:space:]]Diagnostic[[:space:]]Benchmark[[:space:]]for[[:space:]]6D[[:space:]]Spatial[[:space:]]Reasoning[[:space:]]of[[:space:]]Large[[:space:]]Mutimodal[[:space:]]Models/96f4562f-945b-497d-9468-cd476fb5f769_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpatialCLIP_[[:space:]]Learning[[:space:]]3D-aware[[:space:]]Image[[:space:]]Representations[[:space:]]from[[:space:]]Spatially[[:space:]]Discriminative[[:space:]]Language/dad5650f-d216-4b30-a0c7-dd7ca6658593_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpatialDreamer_[[:space:]]Self-supervised[[:space:]]Stereo[[:space:]]Video[[:space:]]Synthesis[[:space:]]from[[:space:]]Monocular[[:space:]]Input/24175650-223e-41c7-b804-6cc94c2c5191_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpatialLLM_[[:space:]]A[[:space:]]Compound[[:space:]]3D-Informed[[:space:]]Design[[:space:]]towards[[:space:]]Spatially-Intelligent[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models/cf47bc9d-8b67-4bc0-a3fe-ba2e3465761d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spatiotemporal[[:space:]]Decoupling[[:space:]]for[[:space:]]Efficient[[:space:]]Vision-Based[[:space:]]Occupancy[[:space:]]Forecasting/9ae564d7-647a-4ca8-a4dc-af9f575b1e21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spatiotemporal[[:space:]]Skip[[:space:]]Guidance[[:space:]]for[[:space:]]Enhanced[[:space:]]Video[[:space:]]Diffusion[[:space:]]Sampling/4ec7da94-3031-45c8-a8d6-19bf226e3cba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpecTRe-GS_[[:space:]]Modeling[[:space:]]Highly[[:space:]]Specular[[:space:]]Surfaces[[:space:]]with[[:space:]]Reflected[[:space:]]Nearby[[:space:]]Objects[[:space:]]by[[:space:]]Tracing[[:space:]]Rays[[:space:]]in[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/05871852-d282-4279-bc18-a4b518c95c0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spectral[[:space:]]Informed[[:space:]]Mamba[[:space:]]for[[:space:]]Robust[[:space:]]Point[[:space:]]Cloud[[:space:]]Processing/9acc5c2f-4c09-4911-b4a0-485ec3edcc53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spectral[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]for[[:space:]]Rotation-Invariant[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/da806762-c7b4-4ed7-80ed-cc0683ab510c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpectroMotion_[[:space:]]Dynamic[[:space:]]3D[[:space:]]Reconstruction[[:space:]]of[[:space:]]Specular[[:space:]]Scenes/5055bce6-6fec-402b-adcc-413e1a1becdd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Speedy-Splat_[[:space:]]Fast[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]Sparse[[:space:]]Pixels[[:space:]]and[[:space:]]Sparse[[:space:]]Primitives/08166209-021b-4f8f-8611-9f7790e8bc2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SphereUFormer_[[:space:]]A[[:space:]]U-Shaped[[:space:]]Transformer[[:space:]]for[[:space:]]Spherical[[:space:]]360[[:space:]]Perception/d3dc320a-7961-421e-ae82-56124645a3a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spherical[[:space:]]Manifold[[:space:]]Guided[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Panoramic[[:space:]]Image[[:space:]]Generation/25742454-f26d-401b-a2ab-0a30d46da30a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spiking[[:space:]]Transformer[[:space:]]with[[:space:]]Spatial-Temporal[[:space:]]Attention/de8548d3-fd04-4e18-9d13-559b7231ddb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spiking[[:space:]]Transformer_[[:space:]]Introducing[[:space:]]Accurate[[:space:]]Addition-Only[[:space:]]Spiking[[:space:]]Self-Attention[[:space:]]for[[:space:]]Transformer/24b0ab2b-a4ff-49fc-865c-b8d7799e49bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpiritSight[[:space:]]Agent_[[:space:]]Advanced[[:space:]]GUI[[:space:]]Agent[[:space:]]with[[:space:]]One[[:space:]]Look/c38ba9da-d4d0-4e2a-a186-fdd0cc0687b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spk2SRImgNet_[[:space:]]Super-Resolve[[:space:]]Dynamic[[:space:]]Scene[[:space:]]from[[:space:]]Spike[[:space:]]Stream[[:space:]]via[[:space:]]Motion[[:space:]]Aligned[[:space:]]Collaborative[[:space:]]Filtering/e0bcc965-cf40-409b-a203-db12626ad68b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SplatAD_[[:space:]]Real-Time[[:space:]]Lidar[[:space:]]and[[:space:]]Camera[[:space:]]Rendering[[:space:]]with[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/58aed1a8-44d5-4314-9f37-1456921ac406_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SplatFlow_[[:space:]]Multi-View[[:space:]]Rectified[[:space:]]Flow[[:space:]]Model[[:space:]]for[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Synthesis/a1da8541-9ef2-4901-8219-e22e7fdfc1fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SplatFlow_[[:space:]]Self-Supervised[[:space:]]Dynamic[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]in[[:space:]]Neural[[:space:]]Motion[[:space:]]Flow[[:space:]]Field[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/8f50ebde-225c-4574-a537-44bd7c2cfecd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Splatter-360_[[:space:]]Generalizable[[:space:]]360[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Wide-baseline[[:space:]]Panoramic[[:space:]]Images/695c9e66-9d36-4a85-a2d0-e99f9431d57c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SplineGS_[[:space:]]Robust[[:space:]]Motion-Adaptive[[:space:]]Spline[[:space:]]for[[:space:]]Real-Time[[:space:]]Dynamic[[:space:]]3D[[:space:]]Gaussians[[:space:]]from[[:space:]]Monocular[[:space:]]Video/c20079e6-f30b-47d1-a98c-7f27a87d2799_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Split[[:space:]]Adaptation[[:space:]]for[[:space:]]Pre-trained[[:space:]]Vision[[:space:]]Transformers/b5d6478d-8866-4dba-8ddb-9d84bd75792a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spotting[[:space:]]the[[:space:]]Unexpected[[:space:]](STU)_[[:space:]]A[[:space:]]3D[[:space:]]LiDAR[[:space:]]Dataset[[:space:]]for[[:space:]]Anomaly[[:space:]]Segmentation[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/7019444b-0523-423c-8391-a6c740f37e7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stabilizing[[:space:]]and[[:space:]]Accelerating[[:space:]]Autofocus[[:space:]]with[[:space:]]Expert[[:space:]]Trajectory[[:space:]]Regularized[[:space:]]Deep[[:space:]]Reinforcement[[:space:]]Learning/a699e1d4-e683-4371-bc56-144538826827_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stable[[:space:]]Flow_[[:space:]]Vital[[:space:]]Layers[[:space:]]for[[:space:]]Training-Free[[:space:]]Image[[:space:]]Editing/a190527f-6eb2-4017-9d47-b894a9bba078_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stable-SCore_[[:space:]]A[[:space:]]Stable[[:space:]]Registration-based[[:space:]]Framework[[:space:]]for[[:space:]]3D[[:space:]]Shape[[:space:]]Correspondence/2db343d7-4cba-4977-a0a7-2b8d62575881_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StableAnimator_[[:space:]]High-Quality[[:space:]]Identity-Preserving[[:space:]]Human[[:space:]]Image[[:space:]]Animation/2c7a20b7-fa25-44af-8aa8-92c6dda76d0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stacking[[:space:]]Brick[[:space:]]by[[:space:]]Brick_[[:space:]]Aligned[[:space:]]Feature[[:space:]]Isolation[[:space:]]for[[:space:]]Incremental[[:space:]]Face[[:space:]]Forgery[[:space:]]Detection/c804cdcc-bfdd-498b-8c75-677eb6a386e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StageDesigner_[[:space:]]Artistic[[:space:]]Stage[[:space:]]Generation[[:space:]]for[[:space:]]Scenography[[:space:]]via[[:space:]]Theater[[:space:]]Scripts/51bd5233-c684-4944-ae56-67deda6596de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Star[[:space:]]with[[:space:]]Bilinear[[:space:]]Mapping/357c7c79-0a0e-4fd1-af25-7fd18a906d89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StarGen_[[:space:]]A[[:space:]]Spatiotemporal[[:space:]]Autoregression[[:space:]]Framework[[:space:]]with[[:space:]]Video[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Scalable[[:space:]]and[[:space:]]Controllable[[:space:]]Scene[[:space:]]Generation/77316d2c-f071-4798-8983-b33ccc9587f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StarVector_[[:space:]]Generating[[:space:]]Scalable[[:space:]]Vector[[:space:]]Graphics[[:space:]]Code[[:space:]]from[[:space:]]Images[[:space:]]and[[:space:]]Text/5cd98bdc-523f-4b08-8843-98de99fe8c3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StdGEN_[[:space:]]Semantic-Decomposed[[:space:]]3D[[:space:]]Character[[:space:]]Generation[[:space:]]from[[:space:]]Single[[:space:]]Images/8cee1d2e-c124-4be4-b93a-135e9506ea92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Steady[[:space:]]Progress[[:space:]]Beats[[:space:]]Stagnation_[[:space:]]Mutual[[:space:]]Aid[[:space:]]of[[:space:]]Foundation[[:space:]]and[[:space:]]Conventional[[:space:]]Models[[:space:]]in[[:space:]]Mixed[[:space:]]Domain[[:space:]]Semi-Supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/a5661569-28fb-42f2-bd7a-af7d6d06652f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stealthy[[:space:]]Backdoor[[:space:]]Attack[[:space:]]in[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]Vision[[:space:]]Encoders[[:space:]]for[[:space:]]Large[[:space:]]Vision[[:space:]]Language[[:space:]]Models/e4906c56-5e5f-4a1e-8aaf-e38f290137b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Steepest[[:space:]]Descent[[:space:]]Density[[:space:]]Control[[:space:]]for[[:space:]]Compact[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/22061661-df9f-4b16-97aa-329dd78a31c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Steering[[:space:]]Away[[:space:]]from[[:space:]]Harm_[[:space:]]An[[:space:]]Adaptive[[:space:]]Approach[[:space:]]to[[:space:]]Defending[[:space:]]Vision[[:space:]]Language[[:space:]]Model[[:space:]]Against[[:space:]]Jailbreaks/1316ef2a-f22b-4e12-b277-a480d5583fd0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stereo[[:space:]]Anywhere_[[:space:]]Robust[[:space:]]Zero-Shot[[:space:]]Deep[[:space:]]Stereo[[:space:]]Matching[[:space:]]Even[[:space:]]Where[[:space:]]Either[[:space:]]Stereo[[:space:]]or[[:space:]]Mono[[:space:]]Fail/7d5379a5-ef0a-4596-9ee5-f53d205d1d3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stereo4D_[[:space:]]Learning[[:space:]]How[[:space:]]Things[[:space:]]Move[[:space:]]in[[:space:]]3D[[:space:]]from[[:space:]]Internet[[:space:]]Stereo[[:space:]]Videos/cf886b95-9dd2-43a7-9d71-f112ad40fb4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StickMotion_[[:space:]]Generating[[:space:]]3D[[:space:]]Human[[:space:]]Motions[[:space:]]by[[:space:]]Drawing[[:space:]]a[[:space:]]Stickman/7f0c2a31-ac0b-47ea-92e5-ccf8b0dc73f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stochastic[[:space:]]Human[[:space:]]Motion[[:space:]]Prediction[[:space:]]with[[:space:]]Memory[[:space:]]of[[:space:]]Action[[:space:]]Transition[[:space:]]and[[:space:]]Action[[:space:]]Characteristic/3a6806f8-3c22-44c0-b354-c2885bf7fddf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stop[[:space:]]Learning[[:space:]]it[[:space:]]all[[:space:]]to[[:space:]]Mitigate[[:space:]]Visual[[:space:]]Hallucination,[[:space:]]Focus[[:space:]]on[[:space:]]the[[:space:]]Hallucination[[:space:]]Target./70e90cd9-a6bd-49ad-98f9-8c56e5becef2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stop[[:space:]]Walking[[:space:]]in[[:space:]]Circles![[:space:]]Bailing[[:space:]]Out[[:space:]]Early[[:space:]]in[[:space:]]Projected[[:space:]]Gradient[[:space:]]Descent/c80dcb74-fe73-4c6b-a75d-b4361bb46dd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StoryGPT-V_[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]as[[:space:]]Consistent[[:space:]]Story[[:space:]]Visualizers/c95cf5c2-c8f6-446d-9f60-c0110fdd0dfb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StreamingT2V_[[:space:]]Consistent,[[:space:]]Dynamic,[[:space:]]and[[:space:]]Extendable[[:space:]]Long[[:space:]]Video[[:space:]]Generation[[:space:]]from[[:space:]]Text/abd21718-bb53-4530-9d4a-41857d6cd2a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StreetCrafter_[[:space:]]Street[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]Controllable[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/e564ce0d-1063-4b88-94b9-3c94311b40cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stretching[[:space:]]Each[[:space:]]Dollar_[[:space:]]Diffusion[[:space:]]Training[[:space:]]from[[:space:]]Scratch[[:space:]]on[[:space:]]a[[:space:]]Micro-Budget/9ea3710b-4706-43ba-bb97-0e2eeba19ff6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Structure[[:space:]]from[[:space:]]Collision/9f6e24b4-d561-4508-b379-3406c68d3bfa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Structure-Aware[[:space:]]Correspondence[[:space:]]Learning[[:space:]]for[[:space:]]Relative[[:space:]]Pose[[:space:]]Estimation/afbf80ec-15b1-4cf1-97ab-c5371bb60343_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Structure-from-Motion[[:space:]]with[[:space:]]a[[:space:]]Non-Parametric[[:space:]]Camera[[:space:]]Model/631ebec5-c70c-4739-bd1e-3bf4348b63b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Structured[[:space:]]3D[[:space:]]Latents[[:space:]]for[[:space:]]Scalable[[:space:]]and[[:space:]]Versatile[[:space:]]3D[[:space:]]Generation/a5e5a861-e756-4882-aa40-f0a5d3d98d7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Style[[:space:]]Evolving[[:space:]]along[[:space:]]Chain-of-Thought[[:space:]]for[[:space:]]Unknown-Domain[[:space:]]Object[[:space:]]Detection/a0b630c9-33f6-4dce-88fa-5eeef37c0790_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Style[[:space:]]Quantization[[:space:]]for[[:space:]]Data-Efficient[[:space:]]GAN[[:space:]]Training/c13161f0-8754-43d2-9e10-0719a4ae9501_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Style-Editor_[[:space:]]Text-driven[[:space:]]Object-centric[[:space:]]Style[[:space:]]Editing/f027dd38-5233-4a32-8582-7d87594c85d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StyleMaster_[[:space:]]Stylize[[:space:]]Your[[:space:]]Video[[:space:]]with[[:space:]]Artistic[[:space:]]Generation[[:space:]]and[[:space:]]Translation/450785d4-75dc-4838-b85b-d61f824214ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StyleSSP_[[:space:]]Sampling[[:space:]]StartPoint[[:space:]]Enhancement[[:space:]]for[[:space:]]Training-free[[:space:]]Diffusion-based[[:space:]]Method[[:space:]]for[[:space:]]Style[[:space:]]Transfer/33385e28-6e48-42ce-86f2-023740b8c602_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StyleStudio_[[:space:]]Text-Driven[[:space:]]Style[[:space:]]Transfer[[:space:]]with[[:space:]]Selective[[:space:]]Control[[:space:]]of[[:space:]]Style[[:space:]]Elements/49d61e99-dda0-4c60-9259-683273883f4c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Subnet-Aware[[:space:]]Dynamic[[:space:]]Supernet[[:space:]]Training[[:space:]]for[[:space:]]Neural[[:space:]]Architecture[[:space:]]Search/bd906641-40af-4bee-99fc-c15d851c38a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Subspace[[:space:]]Constraint[[:space:]]and[[:space:]]Contribution[[:space:]]Estimation[[:space:]]for[[:space:]]Heterogeneous[[:space:]]Federated[[:space:]]Learning/29deb285-7b46-4b14-900f-081368d6cf84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sufficient[[:space:]]Invariant[[:space:]]Learning[[:space:]]for[[:space:]]Distribution[[:space:]]Shift/7012c5eb-1b29-4508-940e-8bd07fe1b3fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SuperLightNet_[[:space:]]Lightweight[[:space:]]Parameter[[:space:]]Aggregation[[:space:]]Network[[:space:]]for[[:space:]]Multimodal[[:space:]]Brain[[:space:]]Tumor[[:space:]]Segmentation/f1c3c18d-30ba-4372-bd89-c4b20db51826_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SuperPC_[[:space:]]A[[:space:]]Single[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion,[[:space:]]Upsampling,[[:space:]]Denoising,[[:space:]]and[[:space:]]Colorization/dc2b6ee9-e477-41c1-84b2-0afcfb16b5d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Supervising[[:space:]]Sound[[:space:]]Localization[[:space:]]by[[:space:]]In-the-wild[[:space:]]Egomotion/5008b1bd-4fa0-4575-9279-9b3625f8266c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SwiftEdit_[[:space:]]Lightning[[:space:]]Fast[[:space:]]Text-Guided[[:space:]]Image[[:space:]]Editing[[:space:]]via[[:space:]]One-Step[[:space:]]Diffusion/feff8562-2a05-49d4-93a5-ed9c0df9c42f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SymDPO_[[:space:]]Boosting[[:space:]]In-Context[[:space:]]Learning[[:space:]]of[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models[[:space:]]with[[:space:]]Symbol[[:space:]]Demonstration[[:space:]]Direct[[:space:]]Preference[[:space:]]Optimization/a3d27b6a-1a00-49bd-b5ed-86deb04e57e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Symbolic[[:space:]]Representation[[:space:]]for[[:space:]]Any-to-Any[[:space:]]Generative[[:space:]]Tasks/4031952f-21ea-418b-856e-1b156bcc6f86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Symmetry[[:space:]]Strikes[[:space:]]Back_[[:space:]]From[[:space:]]Single-Image[[:space:]]Symmetry[[:space:]]Detection[[:space:]]to[[:space:]]3D[[:space:]]Generation/eeeba0ae-3616-4b50-94f5-fc8829b805cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SynTab-LLaVA_[[:space:]]Enhancing[[:space:]]Multimodal[[:space:]]Table[[:space:]]Understanding[[:space:]]with[[:space:]]Decoupled[[:space:]]Synthesis/ff8b01a5-2eb6-4d70-a1a6-937e33586ac7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SyncSDE_[[:space:]]A[[:space:]]Probabilistic[[:space:]]Framework[[:space:]]for[[:space:]]Diffusion[[:space:]]Synchronization/86fdbc8b-2647-4506-91ff-2afc13365972_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SyncVP_[[:space:]]Joint[[:space:]]Diffusion[[:space:]]for[[:space:]]Synchronous[[:space:]]Multi-Modal[[:space:]]Video[[:space:]]Prediction/b538c49d-bf5e-4316-a072-c15dd13269d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Synchronized[[:space:]]Video-to-Audio[[:space:]]Generation[[:space:]]via[[:space:]]Mel[[:space:]]Quantization-Continuum[[:space:]]Decomposition/c00cd712-1c5f-4d86-ac9d-fbc8fd76640d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SynerGen-VL_[[:space:]]Towards[[:space:]]Synergistic[[:space:]]Image[[:space:]]Understanding[[:space:]]and[[:space:]]Generation[[:space:]]with[[:space:]]Vision[[:space:]]Experts[[:space:]]and[[:space:]]Token[[:space:]]Folding/0bf15063-2ebf-4964-8c01-f776609b4c6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Synergizing[[:space:]]Motion[[:space:]]and[[:space:]]Appearance_[[:space:]]Multi-Scale[[:space:]]Compensatory[[:space:]]Codebooks[[:space:]]for[[:space:]]Talking[[:space:]]Head[[:space:]]Video[[:space:]]Generation/9b9071bf-2058-49c3-8193-7adc543c1a95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SynthLight_[[:space:]]Portrait[[:space:]]Relighting[[:space:]]with[[:space:]]Diffusion[[:space:]]Model[[:space:]]by[[:space:]]Learning[[:space:]]to[[:space:]]Re-render[[:space:]]Synthetic[[:space:]]Faces/693fd0bd-2d03-4c20-9807-7bced42e37cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Synthetic[[:space:]]Data[[:space:]]is[[:space:]]an[[:space:]]Elegant[[:space:]]GIFT[[:space:]]for[[:space:]]Continual[[:space:]]Vision-Language[[:space:]]Models/d8da82c7-10fd-4d08-bb82-9bc22666923f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Synthetic[[:space:]]Prior[[:space:]]for[[:space:]]Few-Shot[[:space:]]Drivable[[:space:]]Head[[:space:]]Avatar[[:space:]]Inversion/799d747e-4522-42a9-8f87-5d99000c2dbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Synthetic[[:space:]]Visual[[:space:]]Genome/3e487c9c-46f1-4215-b7f6-f8b4cff30da5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Synthetic-to-Real[[:space:]]Self-supervised[[:space:]]Robust[[:space:]]Depth[[:space:]]Estimation[[:space:]]via[[:space:]]Learning[[:space:]]with[[:space:]]Motion[[:space:]]and[[:space:]]Structure[[:space:]]Priors/ab20bb29-13ef-4e64-8bb2-fb8fba84c1fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/T-CIL_[[:space:]]Temperature[[:space:]]Scaling[[:space:]]using[[:space:]]Adversarial[[:space:]]Perturbation[[:space:]]for[[:space:]]Calibration[[:space:]]in[[:space:]]Class-Incremental[[:space:]]Learning/5971b22b-1d27-4f85-b8aa-f9df16a06bfa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/T-FAKE_[[:space:]]Synthesizing[[:space:]]Thermal[[:space:]]Images[[:space:]]for[[:space:]]Facial[[:space:]]Landmarking/d2467b88-e387-42d6-8ed4-acc34b1b1dc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/T2ICount_[[:space:]]Enhancing[[:space:]]Cross-modal[[:space:]]Understanding[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Counting/17e180a1-eb0b-4e4f-88a4-461cd408357b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/T2ISafety_[[:space:]]Benchmark[[:space:]]for[[:space:]]Assessing[[:space:]]Fairness,[[:space:]]Toxicity,[[:space:]]and[[:space:]]Privacy[[:space:]]in[[:space:]]Image[[:space:]]Generation/920f875b-51c0-49dd-ba56-421ad5aa343d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/T2SG_[[:space:]]Traffic[[:space:]]Topology[[:space:]]Scene[[:space:]]Graph[[:space:]]for[[:space:]]Topology[[:space:]]Reasoning[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/2876f4c7-b2bc-46ab-9e22-a8ee681d7cd0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/T2V-CompBench_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Benchmark[[:space:]]for[[:space:]]Compositional[[:space:]]Text-to-video[[:space:]]Generation/54e6a884-52bb-4955-96e0-6ddf8f72f1b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TADFormer_[[:space:]]Task-Adaptive[[:space:]]Dynamic[[:space:]]TransFormer[[:space:]]for[[:space:]]Efficient[[:space:]]Multi-Task[[:space:]]Learning/30dfce79-8709-4d67-ab76-174ace28e888_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TAET_[[:space:]]Two-Stage[[:space:]]Adversarial[[:space:]]Equalization[[:space:]]Training[[:space:]]on[[:space:]]Long-Tailed[[:space:]]Distributions/4476cf03-3c08-4623-986a-a92479d57f0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TAGA_[[:space:]]Self-supervised[[:space:]]Learning[[:space:]]for[[:space:]]Template-free[[:space:]]Animatable[[:space:]]Gaussian[[:space:]]Articulated[[:space:]]Model/e7cef16f-7c47-4c54-903b-2d640c6e7390_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TAMT_[[:space:]]Temporal-Aware[[:space:]]Model[[:space:]]Tuning[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Few-Shot[[:space:]]Action[[:space:]]Recognition/7caf9c86-26da-41ce-9aca-6e2c4dcc4ad4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TANGO_[[:space:]]Training-free[[:space:]]Embodied[[:space:]]AI[[:space:]]Agents[[:space:]]for[[:space:]]Open-world[[:space:]]Tasks/0a8d6e42-42fa-4911-84af-bd3526112049_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TAPT_[[:space:]]Test-Time[[:space:]]Adversarial[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Robust[[:space:]]Inference[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models/4adafa67-4896-4739-bdf7-1f5ba54cc07b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TAROT_[[:space:]]Towards[[:space:]]Essentially[[:space:]]Domain-Invariant[[:space:]]Robustness[[:space:]]with[[:space:]]Theoretical[[:space:]]Justification/ec4c8b5a-9e2a-4f15-83e9-9d7b778961e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TASTE-Rob_[[:space:]]Advancing[[:space:]]Video[[:space:]]Generation[[:space:]]of[[:space:]]Task-Oriented[[:space:]]Hand-Object[[:space:]]Interaction[[:space:]]for[[:space:]]Generalizable[[:space:]]Robotic[[:space:]]Manipulation/924d8cf7-7b9c-4d7b-995b-3d251b7a8e03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TCFG_[[:space:]]Tangential[[:space:]]Damping[[:space:]]Classifier-free[[:space:]]Guidance/8113438a-9689-405a-b6bc-61aa320e1894_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TFCustom_[[:space:]]Customized[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Time-Aware[[:space:]]Frequency[[:space:]]Feature[[:space:]]Guidance/1a56ac6e-2cb9-45e4-af1e-434f2df3991f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TIDE_[[:space:]]Training[[:space:]]Locally[[:space:]]Interpretable[[:space:]]Domain[[:space:]]Generalization[[:space:]]Models[[:space:]]Enables[[:space:]]Test-time[[:space:]]Correction/e47fec81-867b-4ede-a8aa-c8b31b2fa2b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TIMotion_[[:space:]]Temporal[[:space:]]and[[:space:]]Interactive[[:space:]]Framework[[:space:]]for[[:space:]]Efficient[[:space:]]Human-Human[[:space:]]Motion[[:space:]]Generation/50cb4aae-7f81-4aeb-985d-8b68faf24e85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TKG-DM_[[:space:]]Training-free[[:space:]]Chroma[[:space:]]Key[[:space:]]Content[[:space:]]Generation[[:space:]]Diffusion[[:space:]]Model/95b0d4d9-07cb-43c6-ad69-b00fd2844ccb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TSAM_[[:space:]]Temporal[[:space:]]SAM[[:space:]]Augmented[[:space:]]with[[:space:]]Multimodal[[:space:]]Prompts[[:space:]]for[[:space:]]Referring[[:space:]]Audio-Visual[[:space:]]Segmentation/8535e16b-e9a4-4280-8556-88908a875920_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TSD-SR_[[:space:]]One-Step[[:space:]]Diffusion[[:space:]]with[[:space:]]Target[[:space:]]Score[[:space:]]Distillation[[:space:]]for[[:space:]]Real-World[[:space:]]Image[[:space:]]Super-Resolution/f7583ef3-e58b-4f63-b84a-7f9f17f0f37a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TSP-Mamba_[[:space:]]The[[:space:]]Travelling[[:space:]]Salesman[[:space:]]Problem[[:space:]]Meets[[:space:]]Mamba[[:space:]]for[[:space:]]Image[[:space:]]Super-resolution[[:space:]]and[[:space:]]Beyond/989548d5-80da-4477-a3a5-42129f121eff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TacoDepth_[[:space:]]Towards[[:space:]]Efficient[[:space:]]Radar-Camera[[:space:]]Depth[[:space:]]Estimation[[:space:]]with[[:space:]]One-stage[[:space:]]Fusion/60893af7-fdb2-4c8a-88de-cc24df777172_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TailedCore_[[:space:]]Few-Shot[[:space:]]Sampling[[:space:]]for[[:space:]]Unsupervised[[:space:]]Long-Tail[[:space:]]Noisy[[:space:]]Anomaly[[:space:]]Detection/5a75e7a0-a353-42d1-89a1-a561d7342d9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Take[[:space:]]the[[:space:]]Bull[[:space:]]by[[:space:]]the[[:space:]]Horns_[[:space:]]Learning[[:space:]]to[[:space:]]Segment[[:space:]]Hard[[:space:]]Samples/c1e057f3-058c-42b0-8d35-8e21af099503_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Taming[[:space:]]Teacher[[:space:]]Forcing[[:space:]]for[[:space:]]Masked[[:space:]]Autoregressive[[:space:]]Video[[:space:]]Generation/e2161c88-f390-4144-acf4-52130bcde3bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Taming[[:space:]]Video[[:space:]]Diffusion[[:space:]]Prior[[:space:]]with[[:space:]]Scene-Grounding[[:space:]]Guidance[[:space:]]for[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Sparse[[:space:]]Inputs/b3776566-4465-43cf-8d8c-c85e0e038e48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TaoAvatar_[[:space:]]Real-Time[[:space:]]Lifelike[[:space:]]Full-Body[[:space:]]Talking[[:space:]]Avatars[[:space:]]for[[:space:]]Augmented[[:space:]]Reality[[:space:]]via[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/45c8e882-5f08-40bd-989e-2c96f33116f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Targeted[[:space:]]Forgetting[[:space:]]of[[:space:]]Image[[:space:]]Subgroups[[:space:]]in[[:space:]]CLIP[[:space:]]Models/d5d3c532-3ca2-4b7b-a059-b24be7107188_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tartan[[:space:]]IMU_[[:space:]]A[[:space:]]Light[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Inertial[[:space:]]Positioning[[:space:]]in[[:space:]]Robotics/183cae4a-40b3-4a32-ae82-e543fd104c41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Task[[:space:]]Preference[[:space:]]Optimization_[[:space:]]Improving[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]with[[:space:]]Vision[[:space:]]Task[[:space:]]Alignment/775c99dd-79ec-4cb9-9585-2a300953790b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Task[[:space:]]Singular[[:space:]]Vectors_[[:space:]]Reducing[[:space:]]Task[[:space:]]Interference[[:space:]]in[[:space:]]Model[[:space:]]Merging/a88c8ab6-3362-4dfb-a40f-5aa669be1c2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Task-Agnostic[[:space:]]Guided[[:space:]]Feature[[:space:]]Expansion[[:space:]]for[[:space:]]Class-Incremental[[:space:]]Learning/3797d0e9-aa1c-4a42-ad74-2a4a0b89a0f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Task-Aware[[:space:]]Clustering[[:space:]]for[[:space:]]Prompting[[:space:]]Vision-Language[[:space:]]Models/aabdfb34-0949-4921-9337-b3ad9044174e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Task-Specific[[:space:]]Gradient[[:space:]]Adaptation[[:space:]]for[[:space:]]Few-Shot[[:space:]]One-Class[[:space:]]Classification/9921c614-541a-4ded-93b9-8ac26e16d0a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Task-aware[[:space:]]Cross-modal[[:space:]]Feature[[:space:]]Refinement[[:space:]]Transformer[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Visual[[:space:]]Grounding/3f001702-7969-4b52-b9c0-e5b93c68bbfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Task-driven[[:space:]]Image[[:space:]]Fusion[[:space:]]with[[:space:]]Learnable[[:space:]]Fusion[[:space:]]Loss/37d9af9d-2102-464a-b77a-9ecf64006f8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Taste[[:space:]]More,[[:space:]]Taste[[:space:]]Better_[[:space:]]Diverse[[:space:]]Data[[:space:]]and[[:space:]]Strong[[:space:]]Model[[:space:]]Boost[[:space:]]Semi-Supervised[[:space:]]Crowd[[:space:]]Counting/97690d50-f3d9-442e-9ba4-c2196e1e5a9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Taxonomy-Aware[[:space:]]Evaluation[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models/777684c8-8659-4ecd-b9f3-bb5e792a1348_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Teaching[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]to[[:space:]]Regress[[:space:]]Accurate[[:space:]]Image[[:space:]]Quality[[:space:]]Scores[[:space:]]Using[[:space:]]Score[[:space:]]Distribution/6e8f8571-95f4-493d-9b70-2616ff293e01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Teller_[[:space:]]Real-Time[[:space:]]Streaming[[:space:]]Audio-Driven[[:space:]]Portrait[[:space:]]Animation[[:space:]]with[[:space:]]Autoregressive[[:space:]]Motion[[:space:]]Generation/acf23ef4-3e0f-4da0-a811-56761959df29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Temporal[[:space:]]Action[[:space:]]Detection[[:space:]]Model[[:space:]]Compression[[:space:]]by[[:space:]]Progressive[[:space:]]Block[[:space:]]Drop/c193762d-37bf-4775-b666-dc5ed17dc22a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Temporal[[:space:]]Alignment-Free[[:space:]]Video[[:space:]]Matching[[:space:]]for[[:space:]]Few-shot[[:space:]]Action[[:space:]]Recognition/ea700305-465d-4c66-9110-1ea42533dcb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Temporal[[:space:]]Score[[:space:]]Analysis[[:space:]]for[[:space:]]Understanding[[:space:]]and[[:space:]]Correcting[[:space:]]Diffusion[[:space:]]Artifacts/b10327a5-6a8f-4c6a-a274-2ff13e185549_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Temporal[[:space:]]Separation[[:space:]]with[[:space:]]Entropy[[:space:]]Regularization[[:space:]]for[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]in[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks/511efbd9-e72e-4675-94ed-5687d3419119_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Temporally[[:space:]]Consistent[[:space:]]Object-Centric[[:space:]]Learning[[:space:]]by[[:space:]]Contrasting[[:space:]]Slots/f1085bc3-3cb8-4ee6-8cb0-6147e12071f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TensoFlow_[[:space:]]Tensorial[[:space:]]Flow-based[[:space:]]Sampler[[:space:]]for[[:space:]]Inverse[[:space:]]Rendering/ade03037-3358-4b0c-84b6-3b9ceb264f42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Test-Time[[:space:]]Backdoor[[:space:]]Detection[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]Models/2e7a4e4a-a106-4d5b-99fb-be8cb267958c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Test-Time[[:space:]]Domain[[:space:]]Generalization[[:space:]]via[[:space:]]Universe[[:space:]]Learning_[[:space:]]A[[:space:]]Multi-Graph[[:space:]]Matching[[:space:]]Approach[[:space:]]for[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/eebac6bc-8348-490f-a98e-64c5f1658ac6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Test-Time[[:space:]]Fine-Tuning[[:space:]]of[[:space:]]Image[[:space:]]Compression[[:space:]]Models[[:space:]]for[[:space:]]Multi-Task[[:space:]]Adaptability/aab7c6c2-25ab-43c4-8b18-560bbb6a0c36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Test-Time[[:space:]]Visual[[:space:]]In-Context[[:space:]]Tuning/1106c942-047b-40f5-b114-2de12258d183_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Test-time[[:space:]]Augmentation[[:space:]]Improves[[:space:]]Efficiency[[:space:]]in[[:space:]]Conformal[[:space:]]Prediction/69bebec2-b46a-4233-826c-ee6981af40a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TexGarment_[[:space:]]Consistent[[:space:]]Garment[[:space:]]UV[[:space:]]Texture[[:space:]]Generation[[:space:]]via[[:space:]]Efficient[[:space:]]3D[[:space:]]Structure-Guided[[:space:]]Diffusion[[:space:]]Transformer/a4a25bb5-b3b2-4fab-baad-be3f4246d157_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TexGaussian_[[:space:]]Generating[[:space:]]High-quality[[:space:]]PBR[[:space:]]Material[[:space:]]via[[:space:]]Octree-based[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/80f51f64-f3e1-48f6-bd31-effee985225f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Text[[:space:]]Augmented[[:space:]]Correlation[[:space:]]Transformer[[:space:]]For[[:space:]]Few-shot[[:space:]]Classification[[:space:]]&[[:space:]]Segmentation/dedbc069-ae6a-4770-9618-33a3a7706ba4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Text[[:space:]]Embedding[[:space:]]is[[:space:]]Not[[:space:]]All[[:space:]]You[[:space:]]Need_[[:space:]]Attention[[:space:]]Control[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Semantic[[:space:]]Alignment[[:space:]]with[[:space:]]Text[[:space:]]Self-Attention[[:space:]]Maps/986776a8-b76f-4d8b-9abf-09cb8a9ba5b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Text-Driven[[:space:]]Fashion[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Compositional[[:space:]]Concept[[:space:]]Learning[[:space:]]and[[:space:]]Counterfactual[[:space:]]Abduction/0570630b-8b25-4f74-887f-1f4fc2b92f31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Text-guided[[:space:]]Sparse[[:space:]]Voxel[[:space:]]Pruning[[:space:]]for[[:space:]]Efficient[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding/4a2bb82b-b995-4ea1-8fd6-89e0dbc418a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Textured[[:space:]]Gaussians[[:space:]]for[[:space:]]Enhanced[[:space:]]3D[[:space:]]Scene[[:space:]]Appearance[[:space:]]Modeling/d26b8141-c9cc-4aa4-b4d4-43213dcfab1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Art[[:space:]]of[[:space:]]Deception_[[:space:]]Color[[:space:]]Visual[[:space:]]Illusions[[:space:]]and[[:space:]]Diffusion[[:space:]]Models/32f81902-9541-4c54-8edb-2ca63c430edb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Change[[:space:]]You[[:space:]]Want[[:space:]]To[[:space:]]Detect_[[:space:]]Semantic[[:space:]]Change[[:space:]]Detection[[:space:]]In[[:space:]]Earth[[:space:]]Observation[[:space:]]With[[:space:]]Hybrid[[:space:]]Data[[:space:]]Generationf/301f1377-a4d4-4bd4-b05d-943883c2d8ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Devil[[:space:]]is[[:space:]]in[[:space:]]Low-Level[[:space:]]Features[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Few-Shot[[:space:]]Segmentation/9337f673-c412-4595-8bb5-7eba2f00a28b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Devil[[:space:]]is[[:space:]]in[[:space:]]Temporal[[:space:]]Token_[[:space:]]High[[:space:]]Quality[[:space:]]Video[[:space:]]Reasoning[[:space:]]Segmentation/a6997446-be6b-4107-b5e4-8b2f54d90026_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Devil[[:space:]]is[[:space:]]in[[:space:]]the[[:space:]]Prompts_[[:space:]]Retrieval-Augmented[[:space:]]Prompt[[:space:]]Optimization[[:space:]]for[[:space:]]Text-to-Video[[:space:]]Generation/bf61a946-383a-4ffe-b29d-37540410b684_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Illusion[[:space:]]of[[:space:]]Unlearning_[[:space:]]The[[:space:]]Unstable[[:space:]]Nature[[:space:]]of[[:space:]]Machine[[:space:]]Unlearning[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/ca9a5cea-4cef-406e-b3c3-a1c279377219_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Impact[[:space:]]Label[[:space:]]Noise[[:space:]]and[[:space:]]Choice[[:space:]]of[[:space:]]Threshold[[:space:]]has[[:space:]]on[[:space:]]Cross-Entropy[[:space:]]and[[:space:]]Soft-Dice[[:space:]]in[[:space:]]Image[[:space:]]Segmentation/a28916d9-88e3-4c4a-afb0-99e422d85d63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Language[[:space:]]of[[:space:]]Motion_[[:space:]]Unifying[[:space:]]Verbal[[:space:]]and[[:space:]]Non-verbal[[:space:]]Language[[:space:]]of[[:space:]]3D[[:space:]]Human[[:space:]]Motion/5856b279-35bc-44f9-b231-826fe1b3d735_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]PanAf-FGBG[[:space:]]Dataset_[[:space:]]Understanding[[:space:]]the[[:space:]]Impact[[:space:]]of[[:space:]]Backgrounds[[:space:]]in[[:space:]]Wildlife[[:space:]]Behaviour[[:space:]]Recognition/2e0201f3-4718-4a61-af3c-ec2f92f1137d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Photographer's[[:space:]]Eye_[[:space:]]Teaching[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]to[[:space:]]See,[[:space:]]and[[:space:]]Critique[[:space:]]Like[[:space:]]Photographers/e9b6e548-ef9a-4478-aeda-c759b34b1d36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Power[[:space:]]of[[:space:]]Context_[[:space:]]How[[:space:]]Multimodality[[:space:]]Improves[[:space:]]Image[[:space:]]Super-Resolution/10472ffe-820d-4b89-beac-a0259c3bd355_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Scene[[:space:]]Language_[[:space:]]Representing[[:space:]]Scenes[[:space:]]with[[:space:]]Programs,[[:space:]]Words,[[:space:]]and[[:space:]]Embeddings/dc337ba4-1560-4ebb-a36f-4fd2fd0e7e1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Theoretical[[:space:]]Insights[[:space:]]in[[:space:]]Model[[:space:]]Inversion[[:space:]]Robustness[[:space:]]and[[:space:]]Conditional[[:space:]]Entropy[[:space:]]Maximization[[:space:]]for[[:space:]]Collaborative[[:space:]]Inference[[:space:]]Systems/64dad82b-15cb-48dd-921a-3abe9eba9f91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Theory-Inspired[[:space:]]Deep[[:space:]]Multi-View[[:space:]]Multi-Label[[:space:]]Learning[[:space:]]with[[:space:]]Incomplete[[:space:]]Views[[:space:]]and[[:space:]]Noisy[[:space:]]Labels/fa88376e-5594-45d4-abf8-c70bbaaceb86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Thin-Shell-SfT_[[:space:]]Fine-Grained[[:space:]]Monocular[[:space:]]Non-rigid[[:space:]]3D[[:space:]]Surface[[:space:]]Tracking[[:space:]]with[[:space:]]Neural[[:space:]]Deformation[[:space:]]Fields/513fc02c-a19b-4f3d-a5da-fa87c09303bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Think[[:space:]]Small,[[:space:]]Act[[:space:]]Big_[[:space:]]Primitive[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Lifelong[[:space:]]Robot[[:space:]]Manipulation/10e65ac5-a283-4c50-9bfb-b60d470045bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Thinking[[:space:]]in[[:space:]]Space_[[:space:]]How[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]See,[[:space:]]Remember,[[:space:]]and[[:space:]]Recall[[:space:]]Spaces/d0be0117-af91-4243-862d-0b38b354af27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Three[[:space:]]Cars[[:space:]]Approaching[[:space:]]within[[:space:]]100m![[:space:]]Enhancing[[:space:]]Distant[[:space:]]Geometry[[:space:]]by[[:space:]]Tri-Axis[[:space:]]Voxel[[:space:]]Scanning[[:space:]]for[[:space:]]Camera-based[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion/15e849e2-fb9f-4292-a517-246d0d780878_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Three-view[[:space:]]Focal[[:space:]]Length[[:space:]]Recovery[[:space:]]From[[:space:]]Homographies/0672ff2d-7e74-4085-acaa-100fd8ab0f1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Through-The-Mask_[[:space:]]Mask-based[[:space:]]Motion[[:space:]]Trajectories[[:space:]]for[[:space:]]Image-to-Video[[:space:]]Generation/6bd7168e-72b4-45a3-9dd1-742afa951d3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tightening[[:space:]]Robustness[[:space:]]Verification[[:space:]]of[[:space:]]MaxPool-based[[:space:]]Neural[[:space:]]Networks[[:space:]]via[[:space:]]Minimizing[[:space:]]the[[:space:]]Over-Approximation[[:space:]]Zone/5db53854-4963-4ab3-a30d-5494f691a070_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tiled[[:space:]]Diffusion/357e19f3-7983-410d-8404-04d3f5441bf3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Time[[:space:]]of[[:space:]]the[[:space:]]Flight[[:space:]]of[[:space:]]the[[:space:]]Gaussians_[[:space:]]Optimizing[[:space:]]Depth[[:space:]]Indirectly[[:space:]]in[[:space:]]Dynamic[[:space:]]Radiance[[:space:]]Fields/2aed119b-6fd8-4ca9-bb16-6aa628c20f4c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TimeTracker_[[:space:]]Event-based[[:space:]]Continuous[[:space:]]Point[[:space:]]Tracking[[:space:]]for[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation[[:space:]]with[[:space:]]Non-linear[[:space:]]Motion/88861a6e-0902-4642-90b0-264ddf09a3e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Timestep[[:space:]]Embedding[[:space:]]Tells_[[:space:]]It's[[:space:]]Time[[:space:]]to[[:space:]]Cache[[:space:]]for[[:space:]]Video[[:space:]]Diffusion[[:space:]]Model/ff68a3e5-08ba-4671-bcd2-c07f7f71ad3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TinyFusion_[[:space:]]Diffusion[[:space:]]Transformers[[:space:]]Learned[[:space:]]Shallow/dc3182bc-cb48-4197-a057-d50681c20a7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Token[[:space:]]Cropr_[[:space:]]Faster[[:space:]]ViTs[[:space:]]for[[:space:]]Quite[[:space:]]a[[:space:]]Few[[:space:]]Tasks/ecf31b86-cbcf-4340-830f-3120ba5f8090_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TokenFlow_[[:space:]]Unified[[:space:]]Image[[:space:]]Tokenizer[[:space:]]for[[:space:]]Multimodal[[:space:]]Understanding[[:space:]]and[[:space:]]Generation/d48967ea-7366-42fc-bd50-9715bc3c9183_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TokenHSI_[[:space:]]Unified[[:space:]]Synthesis[[:space:]]of[[:space:]]Physical[[:space:]]Human-Scene[[:space:]]Interactions[[:space:]]through[[:space:]]Task[[:space:]]Tokenization/84edbcb4-fd2b-474d-b025-1c8499da3c6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TokenMotion_[[:space:]]Decoupled[[:space:]]Motion[[:space:]]Control[[:space:]]via[[:space:]]Token[[:space:]]Disentanglement[[:space:]]for[[:space:]]Human-centric[[:space:]]Video[[:space:]]Generation/bbdb01e8-3fc9-4da1-82d2-e6c27c9cdbce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tokenize[[:space:]]Image[[:space:]]Patches_[[:space:]]Global[[:space:]]Context[[:space:]]Fusion[[:space:]]for[[:space:]]Effective[[:space:]]Haze[[:space:]]Removal[[:space:]]in[[:space:]]Large[[:space:]]Images/43967ce3-b2b0-49b2-ab13-3ef98a81c1b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TopNet_[[:space:]]Transformer-Efficient[[:space:]]Occupancy[[:space:]]Prediction[[:space:]]Network[[:space:]]for[[:space:]]Octree-Structured[[:space:]]Point[[:space:]]Cloud[[:space:]]Geometry[[:space:]]Compression/96bf9958-0407-4978-98c0-c80598e41530_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TopV_[[:space:]]Compatible[[:space:]]Token[[:space:]]Pruning[[:space:]]with[[:space:]]Inference[[:space:]]Time[[:space:]]Optimization[[:space:]]for[[:space:]]Fast[[:space:]]and[[:space:]]Low-Memory[[:space:]]Multimodal[[:space:]]Vision[[:space:]]Language[[:space:]]Model/d5e968a0-e5ec-4910-8d52-140b478f8777_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TopoCellGen_[[:space:]]Generating[[:space:]]Histopathology[[:space:]]Cell[[:space:]]Topology[[:space:]]with[[:space:]]a[[:space:]]Diffusion[[:space:]]Model/b31a3388-d9b3-4099-938a-b999359bfc35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tora_[[:space:]]Trajectory-oriented[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]for[[:space:]]Video[[:space:]]Generation/d8158b59-c91a-42d2-ab7d-bb0e24d0c8c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Touch2Shape_[[:space:]]Touch-Conditioned[[:space:]]3D[[:space:]]Diffusion[[:space:]]for[[:space:]]Shape[[:space:]]Exploration[[:space:]]and[[:space:]]Reconstruction/52d4d95b-4d0c-4c6a-b83c-f606459577e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Toward[[:space:]]Generalized[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment_[[:space:]]Relaxing[[:space:]]the[[:space:]]Perfect[[:space:]]Reference[[:space:]]Quality[[:space:]]Assumption/d63da5b3-7eeb-4d6b-ae8e-6b6cf865ce4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Toward[[:space:]]Real-world[[:space:]]BEV[[:space:]]Perception_[[:space:]]Depth[[:space:]]Uncertainty[[:space:]]Estimation[[:space:]]via[[:space:]]Gaussian[[:space:]]Splatting/a3e9f757-72e1-422e-b438-98253206edde_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Toward[[:space:]]Robust[[:space:]]Neural[[:space:]]Reconstruction[[:space:]]from[[:space:]]Sparse[[:space:]]Point[[:space:]]Sets/a0c593be-dad8-497a-a686-1f7045054962_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]All-in-One[[:space:]]Medical[[:space:]]Image[[:space:]]Re-Identification/393bb6d7-09d2-4dea-ad4c-c8581a9fcf4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Autonomous[[:space:]]Micromobility[[:space:]]through[[:space:]]Scalable[[:space:]]Urban[[:space:]]Simulation/07143474-d1eb-47e1-be6e-49c80b409d0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Better[[:space:]]Alignment_[[:space:]]Training[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Reinforcement[[:space:]]Learning[[:space:]]Against[[:space:]]Sparse[[:space:]]Rewards/5e067f03-90b5-4181-ae26-9d6bf0c85f11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Consistent[[:space:]]Multi-Task[[:space:]]Learning_[[:space:]]Unlocking[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Task-Specific[[:space:]]Parameters/eeb3bebc-ff8d-4a37-ac12-d1d885ac06c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Continual[[:space:]]Universal[[:space:]]Segmentation/a9bde470-002d-4667-b7bf-55b061743579_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Cost-Effective[[:space:]]Learning_[[:space:]]A[[:space:]]Synergy[[:space:]]of[[:space:]]Semi-Supervised[[:space:]]and[[:space:]]Active[[:space:]]Learning/9cd929f2-344c-47ad-b770-ef1c451a6879_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Effective[[:space:]]and[[:space:]]Sparse[[:space:]]Adversarial[[:space:]]Attack[[:space:]]on[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks[[:space:]]via[[:space:]]Breaking[[:space:]]Invisible[[:space:]]Surrogate[[:space:]]Gradients/4ca5560d-4397-4519-b5a3-f7cc5fee6129_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Efficient[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Zero-shot[[:space:]]Amodal[[:space:]]Segmentation/de9e8468-cee8-43f8-a8aa-1c93c5bece80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Enhanced[[:space:]]Image[[:space:]]Inpainting_[[:space:]]Mitigating[[:space:]]Unwanted[[:space:]]Object[[:space:]]Insertion[[:space:]]and[[:space:]]Preserving[[:space:]]Color[[:space:]]Consistency/340d24fb-c31c-4c11-9606-f6a3258142a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Explainable[[:space:]]and[[:space:]]Unprecedented[[:space:]]Accuracy[[:space:]]in[[:space:]]Matching[[:space:]]Challenging[[:space:]]Finger[[:space:]]Crease[[:space:]]Patterns/cb786974-9798-4a9f-be6d-a024540300f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Explicit[[:space:]]Geometry-Reflectance[[:space:]]Collaboration[[:space:]]for[[:space:]]Generalized[[:space:]]LiDAR[[:space:]]Segmentation[[:space:]]in[[:space:]]Adverse[[:space:]]Weather/da2daad5-e88a-4151-9554-f77f3d85350e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Fine-Grained[[:space:]]Interpretability_[[:space:]]Counterfactual[[:space:]]Explanations[[:space:]]for[[:space:]]Misclassification[[:space:]]with[[:space:]]Saliency[[:space:]]Partition/57f2df9f-d8fd-46ef-a6f6-9e2727906da9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]General[[:space:]]Visual-Linguistic[[:space:]]Face[[:space:]]Forgery[[:space:]]Detection/2ce1c645-14f0-4400-aa12-ee93bd1fdd46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Generalizable[[:space:]]Scene[[:space:]]Change[[:space:]]Detection/51aeaff8-0c49-44e3-9465-f7be51500eb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Generalizable[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]using[[:space:]]Dual-Level[[:space:]]Representation[[:space:]]Learning[[:space:]]and[[:space:]]Adaptive[[:space:]]Prompting/39b94b32-1e78-46e3-9882-fada112468fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]High-fidelity[[:space:]]3D[[:space:]]Talking[[:space:]]Avatar[[:space:]]with[[:space:]]Personalized[[:space:]]Dynamic[[:space:]]Texture/9072c048-2ab7-4a23-a29c-17cfd9c3776e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Human-Understandable[[:space:]]Multi-Dimensional[[:space:]]Concept[[:space:]]Discovery/8aaff5e8-96e7-4a51-b7aa-c16fd54d99db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Improved[[:space:]]Text-Aligned[[:space:]]Codebook[[:space:]]Learning_[[:space:]]Multi-Hierarchical[[:space:]]Codebook-Text[[:space:]]Alignment[[:space:]]with[[:space:]]Long[[:space:]]Text/54e5aafc-26c0-471e-82ec-6f66c164f208_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]In-the-wild[[:space:]]3D[[:space:]]Plane[[:space:]]Reconstruction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/3e70b58c-8048-419b-9c5e-c010c365112e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Long-Horizon[[:space:]]Vision-Language[[:space:]]Navigation_[[:space:]]Platform,[[:space:]]Benchmark[[:space:]]and[[:space:]]Method/f8e00c20-c399-4943-ba5f-2f21e7cd34f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Lossless[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representation[[:space:]]via[[:space:]]Bit[[:space:]]Plane[[:space:]]Decomposition/ae98aa38-0799-4cf0-8278-e51e284c686f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Million-Scale[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]Evaluation[[:space:]]With[[:space:]]Stronger[[:space:]]Individual[[:space:]]Attacks/aadccabb-b9a4-4cc0-b5ad-5f6d2ae4d8e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]More[[:space:]]General[[:space:]]Video-based[[:space:]]Deepfake[[:space:]]Detection[[:space:]]through[[:space:]]Facial[[:space:]]Component[[:space:]]Guided[[:space:]]Adaptation[[:space:]]for[[:space:]]Foundation[[:space:]]Model/497debf1-0704-487c-ab11-dd5eb612f677_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Natural[[:space:]]Language-Based[[:space:]]Document[[:space:]]Image[[:space:]]Retrieval_[[:space:]]New[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark/841120a0-d6bb-459d-a156-d6377d8a75df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Open-Vocabulary[[:space:]]Audio-Visual[[:space:]]Event[[:space:]]Localization/82a11163-c9e8-4f77-a1c2-be6e45704e9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Optimizing[[:space:]]Large-Scale[[:space:]]Multi-Graph[[:space:]]Matching[[:space:]]in[[:space:]]Bioimaging/30ccd85c-0078-47a9-b9a7-0df7820b8456_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Practical[[:space:]]Real-Time[[:space:]]Neural[[:space:]]Video[[:space:]]Compression/315892ee-79ad-4f5c-b02c-e7385f8c1136_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Precise[[:space:]]Embodied[[:space:]]Dialogue[[:space:]]Localization[[:space:]]via[[:space:]]Causality[[:space:]]Guided[[:space:]]Diffusion/413c8a38-fdf9-4ed2-9099-2c93ad02747c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Precise[[:space:]]Scaling[[:space:]]Laws[[:space:]]for[[:space:]]Video[[:space:]]Diffusion[[:space:]]Transformers/98709fd6-4514-4666-88d7-e12222090c9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]RAW[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]Diverse[[:space:]]Conditions/8568ff82-db8c-40d1-8299-457fbb808815_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Realistic[[:space:]]Example-based[[:space:]]Modeling[[:space:]]via[[:space:]]3D[[:space:]]Gaussian[[:space:]]Stitching/ad7f16a4-c967-4249-8d4e-4d8abe9722dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Satellite[[:space:]]Image[[:space:]]Road[[:space:]]Graph[[:space:]]Extraction_[[:space:]]A[[:space:]]Global-Scale[[:space:]]Dataset[[:space:]]and[[:space:]]A[[:space:]]Novel[[:space:]]Method/d4e690fd-dbed-41f2-a9f3-cc8a43f638cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Scalable[[:space:]]Human-aligned[[:space:]]Benchmark[[:space:]]for[[:space:]]Text-guided[[:space:]]Image[[:space:]]Editing/c5d5d015-164a-4395-a1f6-266905bbebb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Smart[[:space:]]Point-and-Shoot[[:space:]]Photography/f69f109c-aa3e-4d84-8a0a-11b339c021bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Source-Free[[:space:]]Machine[[:space:]]Unlearning/57cf3af9-e9ea-4b42-9700-b97d288b9980_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Stable[[:space:]]and[[:space:]]Storage-efficient[[:space:]]Dataset[[:space:]]Distillation_[[:space:]]Matching[[:space:]]Convexified[[:space:]]Trajectory/3e39a654-8188-4750-9dec-e0302e98fac5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Training-free[[:space:]]Anomaly[[:space:]]Detection[[:space:]]with[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Foundation[[:space:]]Models/8ebb0929-887b-4f6c-aa5b-025362a33658_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Transformer-Based[[:space:]]Aligned[[:space:]]Generation[[:space:]]with[[:space:]]Self-Coherence[[:space:]]Guidance/0430b3b2-c194-4129-b2be-acaa49b5720f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Unbiased[[:space:]]and[[:space:]]Robust[[:space:]]Spatio-Temporal[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]and[[:space:]]Anticipation/61a4ff36-4fd8-43f1-a4cb-46547e183676_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Understanding[[:space:]]How[[:space:]]Knowledge[[:space:]]Evolves[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/feb88a82-d0b8-411b-8ce3-184786042b8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Understanding[[:space:]]and[[:space:]]Quantifying[[:space:]]Uncertainty[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/c3ad3744-8ee2-4117-a4a1-7d23c58e3fe4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Universal[[:space:]]AI-Generated[[:space:]]Image[[:space:]]Detection[[:space:]]by[[:space:]]Variational[[:space:]]Information[[:space:]]Bottleneck[[:space:]]Network/c23bce64-f04c-4e24-b700-3d6473e34043_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Universal[[:space:]]Dataset[[:space:]]Distillation[[:space:]]via[[:space:]]Task-Driven[[:space:]]Diffusion/cd8b4c9e-1814-423f-bdad-2b631dc6a75a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Universal[[:space:]]Soccer[[:space:]]Video[[:space:]]Understanding/bd5f526c-bc30-45cc-bbcd-e151dfc21f1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Visual[[:space:]]Discrimination[[:space:]]and[[:space:]]Reasoning[[:space:]]of[[:space:]]Real-World[[:space:]]Physical[[:space:]]Dynamics_[[:space:]]Physics-Grounded[[:space:]]Anomaly[[:space:]]Detection/e496d86c-0d1a-42d1-b098-7f552a70c4ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Zero-Shot[[:space:]]Anomaly[[:space:]]Detection[[:space:]]and[[:space:]]Reasoning[[:space:]]with[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/9698d3f1-26f6-4c2e-aef0-60a71f074ba9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]a[[:space:]]Universal[[:space:]]Synthetic[[:space:]]Video[[:space:]]Detector_[[:space:]]From[[:space:]]Face[[:space:]]or[[:space:]]Background[[:space:]]Manipulations[[:space:]]to[[:space:]]Fully[[:space:]]AI-Generated[[:space:]]Content/49360ca2-47e3-4c7b-81b1-6a87621996c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tra-MoE_[[:space:]]Learning[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]Model[[:space:]]from[[:space:]]Multiple[[:space:]]Domains[[:space:]]for[[:space:]]Adaptive[[:space:]]Policy[[:space:]]Conditioning/d41e3854-0a29-4aee-85fc-0fcf3b667c89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TraF-Align_[[:space:]]Trajectory-aware[[:space:]]Feature[[:space:]]Alignment[[:space:]]for[[:space:]]Asynchronous[[:space:]]Multi-agent[[:space:]]Perception/0f12efc0-153e-4cda-95dd-b16491feb48e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Track[[:space:]]Any[[:space:]]Anomalous[[:space:]]Object_A[[:space:]]Granular[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection[[:space:]]Pipeline/1ddd1714-9e44-4427-b96d-74620c052e99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Track4Gen_[[:space:]]Teaching[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models[[:space:]]to[[:space:]]Track[[:space:]]Points[[:space:]]Improves[[:space:]]Video[[:space:]]Generation/52b14d4e-8da8-4b5e-a32c-278d64dd36e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tracktention_[[:space:]]Leveraging[[:space:]]Point[[:space:]]Tracking[[:space:]]to[[:space:]]Attend[[:space:]]Videos[[:space:]]Faster[[:space:]]and[[:space:]]Better/55c83ffe-0183-42a7-a82c-e5f5408552e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Training[[:space:]]Data[[:space:]]Provenance[[:space:]]Verification_[[:space:]]Did[[:space:]]Your[[:space:]]Model[[:space:]]Use[[:space:]]Synthetic[[:space:]]Data[[:space:]]from[[:space:]]My[[:space:]]Generative[[:space:]]Model[[:space:]]for[[:space:]]Training_/95a43e45-ca4b-40d2-9bb6-79d7e388dedb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Training-free[[:space:]]Dense-Aligned[[:space:]]Diffusion[[:space:]]Guidance[[:space:]]for[[:space:]]Modular[[:space:]]Conditional[[:space:]]Image[[:space:]]Synthesis/237b334f-3be2-4e40-8ac1-466a51400496_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Training-free[[:space:]]Neural[[:space:]]Architecture[[:space:]]Search[[:space:]]through[[:space:]]Variance[[:space:]]of[[:space:]]Knowledge[[:space:]]of[[:space:]]Deep[[:space:]]Network[[:space:]]Weights/740054ef-86e4-4bac-b7b3-53d57b1f3a66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Trajectory[[:space:]]Mamba_[[:space:]]Efficient[[:space:]]Attention-Mamba[[:space:]]Forecasting[[:space:]]Model[[:space:]]Based[[:space:]]on[[:space:]]Selective[[:space:]]SSM/14500b01-a80f-4244-a6f9-1c844f2302da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TransPixeler_[[:space:]]Advancing[[:space:]]Text-to-Video[[:space:]]Generation[[:space:]]with[[:space:]]Transparency/a57aba62-4ac6-4f75-a668-a4f223b5c05f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Transfer[[:space:]]Your[[:space:]]Perspective_[[:space:]]Controllable[[:space:]]3D[[:space:]]Generation[[:space:]]from[[:space:]]Any[[:space:]]Viewpoint[[:space:]]in[[:space:]]a[[:space:]]Driving[[:space:]]Scene/0e711492-3ab2-4748-8147-c897ce7d76b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Transformers[[:space:]]without[[:space:]]Normalization/05ce47c0-7126-423f-aeff-039d314c6fa4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Traversing[[:space:]]Distortion-Perception[[:space:]]Tradeoff[[:space:]]using[[:space:]]a[[:space:]]Single[[:space:]]Score-Based[[:space:]]Generative[[:space:]]Model/dcdd686b-eeb1-4343-a7e6-20b5cd329198_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TreeMeshGPT_[[:space:]]Artistic[[:space:]]Mesh[[:space:]]Generation[[:space:]]with[[:space:]]Autoregressive[[:space:]]Tree[[:space:]]Sequencing/cc43c3b2-b6bf-472a-9dd6-c74d4f03cee9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TriTex_[[:space:]]Learning[[:space:]]Texture[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Mesh[[:space:]]via[[:space:]]Triplane[[:space:]]Semantic[[:space:]]Features/f0b8cf58-1685-4de9-9df0-451922c97e91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tripartite[[:space:]]Weight-Space[[:space:]]Ensemble[[:space:]]for[[:space:]]Few-Shot[[:space:]]Class-Incremental[[:space:]]Learning/6894bed0-33d2-4729-8056-07cf7489b89f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tuning[[:space:]]the[[:space:]]Frequencies_[[:space:]]Robust[[:space:]]Training[[:space:]]for[[:space:]]Sinusoidal[[:space:]]Neural[[:space:]]Networks/79e8b6d6-50a7-4215-bd2a-fb5524e17539_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Turbo3D_[[:space:]]Ultra-fast[[:space:]]Text-to-3D[[:space:]]Generation/bd4523a6-3f35-4dc6-9b48-90da3f36ebea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TurboFill_[[:space:]]Adapting[[:space:]]Few-step[[:space:]]Text-to-image[[:space:]]Model[[:space:]]for[[:space:]]Fast[[:space:]]Image[[:space:]]Inpainting/6302d669-b0e3-4a35-adf2-a42f09f692e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Twinner_[[:space:]]Shining[[:space:]]Light[[:space:]]on[[:space:]]Digital[[:space:]]Twins[[:space:]]in[[:space:]]a[[:space:]]Few[[:space:]]Snaps/0293a039-5500-41cc-90fa-fe4f8710c6e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Two[[:space:]]by[[:space:]]Two_[[:space:]]Learning[[:space:]]Multi-Task[[:space:]]Pairwise[[:space:]]Objects[[:space:]]Assembly[[:space:]]for[[:space:]]Generalizable[[:space:]]Robot[[:space:]]Manipulation/57fa886e-a784-460f-94a7-de5f04b8ab96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Two[[:space:]]is[[:space:]]Better[[:space:]]than[[:space:]]One_[[:space:]]Efficient[[:space:]]Ensemble[[:space:]]Defense[[:space:]]for[[:space:]]Robust[[:space:]]and[[:space:]]Compact[[:space:]]Models/f8d31d9f-8dff-45eb-9d02-43167ab5a77a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Type-R_[[:space:]]Automatically[[:space:]]Retouching[[:space:]]Typos[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/6f176b19-a787-4b85-9ebe-081a3f4759be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/U-Know-DiffPAN_[[:space:]]An[[:space:]]Uncertainty-aware[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]Diffusion[[:space:]]Framework[[:space:]]with[[:space:]]Details[[:space:]]Enhancement[[:space:]]for[[:space:]]PAN-Sharpening/ca20f1f2-4ee9-4b83-80ca-6e2c71410bad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UA-Pose_[[:space:]]Uncertainty-Aware[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]and[[:space:]]Online[[:space:]]Object[[:space:]]Completion[[:space:]]with[[:space:]]Partial[[:space:]]References/0d3743ff-3af5-445c-ae4c-ba5a33b9d7b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UCM-VeID[[:space:]]V2_[[:space:]]A[[:space:]]Richer[[:space:]]Dataset[[:space:]]and[[:space:]]A[[:space:]]Pre-training[[:space:]]Method[[:space:]]for[[:space:]]UAV[[:space:]]Cross-Modality[[:space:]]Vehicle[[:space:]]Re-Identification/9b7d21db-8ee1-4718-8ddf-98653daa1c9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UCOD-DPL_[[:space:]]Unsupervised[[:space:]]Camouflaged[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Dynamic[[:space:]]Pseudo-label[[:space:]]Learning/dec1b26d-4dd9-4c06-a461-c03d0e31f356_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UHD-processer_[[:space:]]Unified[[:space:]]UHD[[:space:]]Image[[:space:]]Restoration[[:space:]]with[[:space:]]Progressive[[:space:]]Frequency[[:space:]]Learning[[:space:]]and[[:space:]]Degradation-aware[[:space:]]Prompts/2e2768b6-2887-4bd5-935c-d0b880ebe319_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UIBDiffusion_[[:space:]]Universal[[:space:]]Imperceptible[[:space:]]Backdoor[[:space:]]Attack[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/7f0574b2-e282-4c3e-a072-3d64d026983a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UMFN_[[:space:]]Unified[[:space:]]Multi-Domain[[:space:]]Face[[:space:]]Normalization[[:space:]]for[[:space:]]Joint[[:space:]]Cross-domain[[:space:]]Prototype[[:space:]]Learning[[:space:]]and[[:space:]]Heterogeneous[[:space:]]Face[[:space:]]Recognition/3e3f9780-5e1d-4d68-aee1-68dcb88c5368_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UMotion_[[:space:]]Uncertainty-driven[[:space:]]Human[[:space:]]Motion[[:space:]]Estimation[[:space:]]from[[:space:]]Inertial[[:space:]]and[[:space:]]Ultra-wideband[[:space:]]Units/20c112a8-f19a-4ffc-8e5b-a5d02b15caa1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UNEM_[[:space:]]UNrolled[[:space:]]Generalized[[:space:]]EM[[:space:]]for[[:space:]]Transductive[[:space:]]Few-Shot[[:space:]]Learning/592279b8-3f9e-4b39-91e5-638d8677cb8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UNIALIGN_[[:space:]]Scaling[[:space:]]Multimodal[[:space:]]Alignment[[:space:]]within[[:space:]]One[[:space:]]Unified[[:space:]]Model/7e109366-fc0f-438c-9b91-2ba7f07e59e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UNIC-Adapter_[[:space:]]Unified[[:space:]]Image-instruction[[:space:]]Adapter[[:space:]]with[[:space:]]Multi-modal[[:space:]]Transformer[[:space:]]for[[:space:]]Image[[:space:]]Generation/409eb6c5-d571-4968-a329-2c1ebf51e321_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UNICL-SAM_[[:space:]]Uncertainty-Driven[[:space:]]In-Context[[:space:]]Segmentation[[:space:]]with[[:space:]]Part[[:space:]]Prototype[[:space:]]Discovery/fc453b95-4cba-4bb9-95fb-6c8d34e4ccaa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UNOPose_[[:space:]]Unseen[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]an[[:space:]]Unposed[[:space:]]RGB-D[[:space:]]Reference[[:space:]]Image/94d3840c-f3fe-4b2b-9388-48ccee10c9d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UPME_[[:space:]]An[[:space:]]Unsupervised[[:space:]]Peer[[:space:]]Review[[:space:]]Framework[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]Evaluation/42fa34eb-b512-4300-b1c0-40e45c00118f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/URWKV_[[:space:]]Unified[[:space:]]RWKV[[:space:]]Model[[:space:]]with[[:space:]]Multi-state[[:space:]]Perspective[[:space:]]for[[:space:]]Low-light[[:space:]]Image[[:space:]]Restoration/e077664a-8838-4206-9246-0d60f57e032f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/USP-Gaussian_[[:space:]]Unifying[[:space:]]Spike-based[[:space:]]Image[[:space:]]Reconstruction,[[:space:]]Pose[[:space:]]Correction[[:space:]]and[[:space:]]Gaussian[[:space:]]Splatting/62c2f0f5-8ea0-4959-8e97-f59984bffdfe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UVGS_[[:space:]]Reimagining[[:space:]]Unstructured[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]using[[:space:]]UV[[:space:]]Mapping/77087a0b-d1cb-4230-a846-7760f7510d17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UWAV_[[:space:]]Uncertainty-weighted[[:space:]]Weakly-supervised[[:space:]]Audio-Visual[[:space:]]Video[[:space:]]Parsing/82f7769f-b767-4619-96ea-539823053afb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UltraFusion_[[:space:]]Ultra[[:space:]]High[[:space:]]Dynamic[[:space:]]Imaging[[:space:]]using[[:space:]]Exposure[[:space:]]Fusion/c2c26ec1-c8af-49f8-87ba-b3c749531067_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UnCommon[[:space:]]Objects[[:space:]]in[[:space:]]3D/4d2bd976-41fd-45e5-bfc6-8a4578f73725_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unbiased[[:space:]]Video[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]via[[:space:]]Visual[[:space:]]and[[:space:]]Semantic[[:space:]]Dual[[:space:]]Debiasing/168f707b-a7fa-4e7c-93c1-aac7eef770eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unbiasing[[:space:]]through[[:space:]]Textual[[:space:]]Descriptions_[[:space:]]Mitigating[[:space:]]Representation[[:space:]]Bias[[:space:]]in[[:space:]]Video[[:space:]]Benchmarks/5edfcfce-ad23-41c6-9a43-d3cb9f8bcb90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unboxed_[[:space:]]Geometrically[[:space:]]and[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Video[[:space:]]Outpainting/3a973275-71b1-47b0-9c32-b162722059fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Uncertain[[:space:]]Multimodal[[:space:]]Intention[[:space:]]and[[:space:]]Emotion[[:space:]]Understanding[[:space:]]in[[:space:]]the[[:space:]]Wild/50195041-da23-4b5c-8721-3858d2b20dc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Uncertainty[[:space:]]Meets[[:space:]]Diversity_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Active[[:space:]]Learning[[:space:]]Framework[[:space:]]for[[:space:]]Indoor[[:space:]]3D[[:space:]]Object[[:space:]]Detection/d2e21437-1e7e-4dc1-839a-f40a2bf9489e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Uncertainty[[:space:]]Weighted[[:space:]]Gradients[[:space:]]for[[:space:]]Model[[:space:]]Calibration/792a362b-85b5-4092-b750-5040ac391b64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Uncertainty-Instructed[[:space:]]Structure[[:space:]]Injection[[:space:]]for[[:space:]]Generalizable[[:space:]]HD[[:space:]]Map[[:space:]]Construction/308073a9-0b23-46de-a5fd-816ea2a49dc7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Uncertainty-guided[[:space:]]Perturbation[[:space:]]for[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]Diffusion[[:space:]]Model/10430da4-a80e-4106-a1b5-f1210ade542e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Understanding[[:space:]]Fine-tuning[[:space:]]CLIP[[:space:]]for[[:space:]]Open-vocabulary[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]in[[:space:]]Hyperbolic[[:space:]]Space/4933d0fc-2175-4f82-8204-a2ef1c424606_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Understanding[[:space:]]Multi-Task[[:space:]]Activities[[:space:]]from[[:space:]]Single-Task[[:space:]]Videos/b633800a-1331-4862-80d6-5ccbeba54c12_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Understanding[[:space:]]Multi-layered[[:space:]]Transmission[[:space:]]Matrices/0c5698ec-c1a0-4efc-b1d6-15a1169745e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Uni-Renderer_[[:space:]]Unifying[[:space:]]Rendering[[:space:]]and[[:space:]]Inverse[[:space:]]Rendering[[:space:]]Via[[:space:]]Dual[[:space:]]Stream[[:space:]]Diffusion/ef5cd413-b783-4606-bcb0-152603271134_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Uni4D_[[:space:]]Unifying[[:space:]]Visual[[:space:]]Foundation[[:space:]]Models[[:space:]]for[[:space:]]4D[[:space:]]Modeling[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Video/b0bfb21a-5579-4e48-8be6-e0a25d4d029f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniAP_[[:space:]]Unifying[[:space:]]Inter-[[:space:]]and[[:space:]]Intra-Layer[[:space:]]Automatic[[:space:]]Parallelism[[:space:]]by[[:space:]]Mixed[[:space:]]Integer[[:space:]]Quadratic[[:space:]]Programming/b1ddd4d6-5b56-44ba-82ed-6323538fa782_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniGoal_[[:space:]]Towards[[:space:]]Universal[[:space:]]Zero-shot[[:space:]]Goal-oriented[[:space:]]Navigation/88e424e9-ab5f-4a75-83c1-543aa336d4d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniGraspTransformer_[[:space:]]Simplified[[:space:]]Policy[[:space:]]Distillation[[:space:]]for[[:space:]]Scalable[[:space:]]Dexterous[[:space:]]Robotic[[:space:]]Grasping/65b5c418-123a-4ff6-a237-35673bf5c1e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniHOPE_[[:space:]]A[[:space:]]Unified[[:space:]]Approach[[:space:]]for[[:space:]]Hand-Only[[:space:]]and[[:space:]]Hand-Object[[:space:]]Pose[[:space:]]Estimation/7a47a703-2a00-46c3-bcad-6df0cd9883eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniK3D_[[:space:]]Universal[[:space:]]Camera[[:space:]]Monocular[[:space:]]3D[[:space:]]Estimation/e52b9d8c-87d8-48de-bee5-89602db2d48a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniMamba_[[:space:]]Unified[[:space:]]Spatial-Channel[[:space:]]Representation[[:space:]]Learning[[:space:]]with[[:space:]]Group-Efficient[[:space:]]Mamba[[:space:]]for[[:space:]]LiDAR-based[[:space:]]3D[[:space:]]Object[[:space:]]Detection/fd620e18-8be3-43d4-9cf9-619e7a7651ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniNet_[[:space:]]A[[:space:]]Contrastive[[:space:]]Learning-guided[[:space:]]Unified[[:space:]]Framework[[:space:]]with[[:space:]]Feature[[:space:]]Selection[[:space:]]for[[:space:]]Anomaly[[:space:]]Detection/a5e75eb4-b19a-4b89-9c94-9cd8b5d84d84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniPhy_[[:space:]]Learning[[:space:]]a[[:space:]]Unified[[:space:]]Constitutive[[:space:]]Model[[:space:]]for[[:space:]]Inverse[[:space:]]Physics[[:space:]]Simulation/d1af2cfc-e229-417a-b20f-f5d73ab936cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniPose_[[:space:]]A[[:space:]]Unified[[:space:]]Multimodal[[:space:]]Framework[[:space:]]for[[:space:]]Human[[:space:]]Pose[[:space:]]Comprehension,[[:space:]]Generation[[:space:]]and[[:space:]]Editing/71440e3f-578c-413e-b26f-77becb7c1883_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniPre3D_[[:space:]]Unified[[:space:]]Pre-training[[:space:]]of[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Models[[:space:]]with[[:space:]]Cross-Modal[[:space:]]Gaussian[[:space:]]Splatting/065c1667-8be9-4f1b-8f0e-622104aa14d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniReal_[[:space:]]Universal[[:space:]]Image[[:space:]]Generation[[:space:]]and[[:space:]]Editing[[:space:]]via[[:space:]]Learning[[:space:]]Real-world[[:space:]]Dynamics/42a0c187-3d08-4766-816f-e1f791733ec9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniRestore_[[:space:]]Unified[[:space:]]Perceptual[[:space:]]and[[:space:]]Task-Oriented[[:space:]]Image[[:space:]]Restoration[[:space:]]Model[[:space:]]Using[[:space:]]Diffusion[[:space:]]Prior/8fb572e0-013e-4981-af9b-583d1ad35994_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniSTD_[[:space:]]Towards[[:space:]]Unified[[:space:]]Spatio-Temporal[[:space:]]Learning[[:space:]]across[[:space:]]Diverse[[:space:]]Disciplines/80bdfb16-a78d-42a8-bf67-28cebe65325b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniScene_[[:space:]]Unified[[:space:]]Occupancy-centric[[:space:]]Driving[[:space:]]Scene[[:space:]]Generation/0017146a-8707-411c-85ad-7152430bc1f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniVAD_[[:space:]]A[[:space:]]Training-free[[:space:]]Unified[[:space:]]Model[[:space:]]for[[:space:]]Few-shot[[:space:]]Visual[[:space:]]Anomaly[[:space:]]Detection/627df775-91ce-408e-8449-22d57138eaf8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unified[[:space:]]Dense[[:space:]]Prediction[[:space:]]of[[:space:]]Video[[:space:]]Diffusion/dd7cba54-fc52-4483-9411-769418f8500f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unified[[:space:]]Medical[[:space:]]Lesion[[:space:]]Segmentation[[:space:]]via[[:space:]]Self-referring[[:space:]]Indicator/13078c5c-c857-4b31-8dd0-f10c61b36192_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unified[[:space:]]Reconstruction[[:space:]]of[[:space:]]Static[[:space:]]and[[:space:]]Dynamic[[:space:]]Scenes[[:space:]]from[[:space:]]Events/302f511b-8ca3-412b-b164-b3d3cfb6ea42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unified[[:space:]]Uncertainty-Aware[[:space:]]Diffusion[[:space:]]for[[:space:]]Multi-Agent[[:space:]]Trajectory[[:space:]]Modeling/375f7360-3378-4a01-a453-b000e7afdba3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unity[[:space:]]in[[:space:]]Diversity_[[:space:]]Video[[:space:]]Editing[[:space:]]via[[:space:]]Gradient-Latent[[:space:]]Purification/ccbbd6b7-d70e-4c6d-a124-58b67c4e6fda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Universal[[:space:]]Actions[[:space:]]for[[:space:]]Enhanced[[:space:]]Embodied[[:space:]]Foundation[[:space:]]Models/a40c66d5-718c-4d10-be4b-56cf25bac6f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Universal[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/88e9f666-f77c-4764-bbc8-f5811af7dfd9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Universal[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/60251398-7562-410c-9dc1-fca08d462183_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unlearning[[:space:]]through[[:space:]]Knowledge[[:space:]]Overwriting_[[:space:]]Reversible[[:space:]]Federated[[:space:]]Unlearning[[:space:]]via[[:space:]]Selective[[:space:]]Sparse[[:space:]]Adapter/a68e8495-ee94-42e7-9b38-0dfd4680d80c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unleashing[[:space:]]In-context[[:space:]]Learning[[:space:]]of[[:space:]]Autoregressive[[:space:]]Models[[:space:]]for[[:space:]]Few-shot[[:space:]]Image[[:space:]]Manipulation/21c264e1-770d-4a45-9664-5d827344ad71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unleashing[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Consistency[[:space:]]Learning[[:space:]]for[[:space:]]Detecting[[:space:]]and[[:space:]]Grounding[[:space:]]Multi-Modal[[:space:]]Media[[:space:]]Manipulation/ad94865f-89e2-49ae-a251-c93119de6450_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unleashing[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Multi-modal[[:space:]]Foundation[[:space:]]Models[[:space:]]and[[:space:]]Video[[:space:]]Diffusion[[:space:]]for[[:space:]]4D[[:space:]]Dynamic[[:space:]]Physical[[:space:]]Scene[[:space:]]Simulation/030b3319-f432-4841-861f-4634eff93eba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unlocking[[:space:]]Generalization[[:space:]]Power[[:space:]]in[[:space:]]LiDAR[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/f477815c-f14c-4741-96fe-9cd60e395a9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unlocking[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Unlabeled[[:space:]]Data[[:space:]]in[[:space:]]Semi-Supervised[[:space:]]Domain[[:space:]]Generalization/19a8b2cc-c3c3-4c19-b807-05947690f856_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unraveling[[:space:]]Normal[[:space:]]Anatomy[[:space:]]via[[:space:]]Fluid-Driven[[:space:]]Anomaly[[:space:]]Randomization/3a596ec3-4195-48c3-9c0b-9faebb22bc84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unseen[[:space:]]Visual[[:space:]]Anomaly[[:space:]]Generation/77395d70-580c-431b-a2a6-8fbec8b1d95a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unsupervised[[:space:]]Continual[[:space:]]Domain[[:space:]]Shift[[:space:]]Learning[[:space:]]with[[:space:]]Multi-Prototype[[:space:]]Modeling/3ed5fc8d-6d71-4b17-a19b-59231da21217_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unsupervised[[:space:]]Discovery[[:space:]]of[[:space:]]Facial[[:space:]]Landmarks[[:space:]]and[[:space:]]Head[[:space:]]Pose/9f710d7d-c184-4cb3-bedd-6388e0afd314_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unsupervised[[:space:]]Foundation[[:space:]]Model-Agnostic[[:space:]]Slide-Level[[:space:]]Representation[[:space:]]Learning/59d43776-cd86-4dae-9816-56040f71fac5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unveil[[:space:]]Inversion[[:space:]]and[[:space:]]Invariance[[:space:]]in[[:space:]]Flow[[:space:]]Transformer[[:space:]]for[[:space:]]Versatile[[:space:]]Image[[:space:]]Editing/74170e85-8800-4580-ac3b-a78b1c818c38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unveiling[[:space:]]Differences[[:space:]]in[[:space:]]Generative[[:space:]]Models_[[:space:]]A[[:space:]]Scalable[[:space:]]Differential[[:space:]]Clustering[[:space:]]Approach/83aafb74-01db-43d3-a4ee-4fe261e052bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unveiling[[:space:]]Visual[[:space:]]Perception[[:space:]]in[[:space:]]Language[[:space:]]Models_[[:space:]]An[[:space:]]Attention[[:space:]]Head[[:space:]]Analysis[[:space:]]Approach/de09c5d1-6c3a-48ee-af79-a9fd3e25880d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unveiling[[:space:]]the[[:space:]]Ignorance[[:space:]]of[[:space:]]MLLMs_[[:space:]]Seeing[[:space:]]Clearly,[[:space:]]Answering[[:space:]]Incorrectly/5f14f37b-97d8-4d0e-b655-7b7fb20cec9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unveiling[[:space:]]the[[:space:]]Mist[[:space:]]over[[:space:]]3D[[:space:]]Vision-Language[[:space:]]Understanding_[[:space:]]Object-centric[[:space:]]Evaluation[[:space:]]with[[:space:]]Chain-of-Analysis/867a7d0f-edf5-4255-9e3a-7c58b2702120_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UrbanCAD_[[:space:]]Towards[[:space:]]Highly[[:space:]]Controllable[[:space:]]and[[:space:]]Photorealistic[[:space:]]3D[[:space:]]Vehicles[[:space:]]for[[:space:]]Urban[[:space:]]Scene[[:space:]]Simulation/fdb5d6fa-b5e9-4f1d-b440-3b82e60761af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Using[[:space:]]Diffusion[[:space:]]Priors[[:space:]]for[[:space:]]Video[[:space:]]Amodal[[:space:]]Segmentation/9e6cdc81-53a6-4da2-bd44-46151ed7f5db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Using[[:space:]]Powerful[[:space:]]Prior[[:space:]]Knowledge[[:space:]]of[[:space:]]Diffusion[[:space:]]Model[[:space:]]in[[:space:]]Deep[[:space:]]Unfolding[[:space:]]Networks[[:space:]]for[[:space:]]Image[[:space:]]Compressive[[:space:]]Sensing/77fb2416-b473-4795-b20b-d0f5fa5ee2ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/V-Stylist_[[:space:]]Video[[:space:]]Stylization[[:space:]]via[[:space:]]Collaboration[[:space:]]and[[:space:]]Reflection[[:space:]]of[[:space:]]MLLM[[:space:]]Agents/5decddf5-9654-42db-b0eb-e890d0e13076_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/V2V3D_[[:space:]]View-to-View[[:space:]]Denoised[[:space:]]3D[[:space:]]Reconstruction[[:space:]]for[[:space:]]Light[[:space:]]Field[[:space:]]Microscopy/5a825645-08b4-4e0a-9cc8-0369ebf747ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/V2X-R_[[:space:]]Cooperative[[:space:]]LiDAR-4D[[:space:]]Radar[[:space:]]Fusion[[:space:]]with[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/1039e525-4293-4870-b998-e5dcb0286026_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VASparse_[[:space:]]Towards[[:space:]]Efficient[[:space:]]Visual[[:space:]]Hallucination[[:space:]]Mitigation[[:space:]]via[[:space:]]Visual-Aware[[:space:]]Token[[:space:]]Sparsification/22122987-b014-4ca2-a5b5-dcf675c1cb86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VDocRAG_[[:space:]]Retrieval-Augmented[[:space:]]Generation[[:space:]]over[[:space:]]Visually-Rich[[:space:]]Documents/23b281bf-cd19-499d-b429-500f23b3b8c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VELOCITI_[[:space:]]Benchmarking[[:space:]]Video-Language[[:space:]]Compositional[[:space:]]Reasoning[[:space:]]with[[:space:]]Strict[[:space:]]Entailment/dc15c09d-8c20-43f6-aea9-c9a3af05db6c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VERA_[[:space:]]Explainable[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection[[:space:]]via[[:space:]]Verbalized[[:space:]]Learning[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models/afa9b429-c167-49c2-bbf0-c8175fba54ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VEU-Bench_[[:space:]]Towards[[:space:]]Comprehensive[[:space:]]Understanding[[:space:]]of[[:space:]]Video[[:space:]]Editing/f60333f8-92d8-4a87-87a7-0be0a0f6e37d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VGGT_[[:space:]]Visual[[:space:]]Geometry[[:space:]]Grounded[[:space:]]Transformer/732e7f72-2d06-46d2-9bd9-3fee5a917d0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VILA-M3_[[:space:]]Enhancing[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]Medical[[:space:]]Expert[[:space:]]Knowledge/6fe122c6-0da2-4ede-b2fa-094a89c9eec2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VIRES_[[:space:]]Video[[:space:]]Instance[[:space:]]Repainting[[:space:]]via[[:space:]]Sketch[[:space:]]and[[:space:]]Text[[:space:]]Guided[[:space:]]Generation/f8dad9d3-96fc-4e9f-bd03-d6d4630343c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VISCO_[[:space:]]Benchmarking[[:space:]]Fine-Grained[[:space:]]Critique[[:space:]]and[[:space:]]Correction[[:space:]]Towards[[:space:]]Self-Improvement[[:space:]]in[[:space:]]Visual[[:space:]]Reasoning/fa92df7c-8eb2-4a1f-bc26-8097a895209e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VISTA3D_[[:space:]]A[[:space:]]Unified[[:space:]]Segmentation[[:space:]]Foundation[[:space:]]Model[[:space:]]For[[:space:]]3D[[:space:]]Medical[[:space:]]Imaging/58f5da73-2585-4c16-b897-4c0cbe53ebb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VISTA_[[:space:]]Enhancing[[:space:]]Long-Duration[[:space:]]and[[:space:]]High-Resolution[[:space:]]Video[[:space:]]Understanding[[:space:]]by[[:space:]]Video[[:space:]]Spatiotemporal[[:space:]]Augmentation/d8578d04-441c-47c1-bdf9-92fa460dab3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VISTREAM_[[:space:]]Improving[[:space:]]Computation[[:space:]]Efficiency[[:space:]]of[[:space:]]Visual[[:space:]]Streaming[[:space:]]Perception[[:space:]]via[[:space:]]Law-of-Charge-Conservation[[:space:]]Inspired[[:space:]]Spiking[[:space:]]Neural[[:space:]]Network/7a2db10a-7210-473e-9050-b4c58de4e556_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VITED_[[:space:]]Video[[:space:]]Temporal[[:space:]]Evidence[[:space:]]Distillation/7319cae9-b231-498c-ba19-3ef6ad34d61f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VI^3NR_[[:space:]]Variance[[:space:]]Informed[[:space:]]Initialization[[:space:]]for[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations/c6456133-3182-4f8d-a4e2-73f4f2e72fce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VL-RewardBench_[[:space:]]A[[:space:]]Challenging[[:space:]]Benchmark[[:space:]]for[[:space:]]Vision-Language[[:space:]]Generative[[:space:]]Reward[[:space:]]Models/3e5a73b1-9f9b-4b8d-83b0-4a96fd6d591d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VL2Lite_[[:space:]]Task-Specific[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]from[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]to[[:space:]]Lightweight[[:space:]]Networks/3c547a61-ab40-47a5-aabe-6e1e5bb0a777_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VLMs-Guided[[:space:]]Representation[[:space:]]Distillation[[:space:]]for[[:space:]]Efficient[[:space:]]Vision-Based[[:space:]]Reinforcement[[:space:]]Learning/5e041dc8-bfca-4db4-a5df-ddf0affe123f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VLOGGER_[[:space:]]Multimodal[[:space:]]Diffusion[[:space:]]for[[:space:]]Embodied[[:space:]]Avatar[[:space:]]Synthesis/e32a2703-b32a-4a6f-b45c-33f8b72e19be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VLog_[[:space:]]Video-Language[[:space:]]Models[[:space:]]by[[:space:]]Generative[[:space:]]Retrieval[[:space:]]of[[:space:]]Narration[[:space:]]Vocabulary/5cce3931-2141-492f-94ca-887e72e0b54b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VLsI_[[:space:]]Verbalized[[:space:]]Layers-to-Interactions[[:space:]]from[[:space:]]Large[[:space:]]to[[:space:]]Small[[:space:]]Vision[[:space:]]Language[[:space:]]Models/d257a4d1-091b-48d7-b804-775fddd94e7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VODiff_[[:space:]]Controlling[[:space:]]Object[[:space:]]Visibility[[:space:]]Order[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Generation/3dac837e-c8db-4c3a-a872-8c96ea7194f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VSNet_[[:space:]]Focusing[[:space:]]on[[:space:]]the[[:space:]]Linguistic[[:space:]]Characteristics[[:space:]]of[[:space:]]Sign[[:space:]]Language/1135c1a4-3af0-45ac-a497-2c9557104d8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VTON[[:space:]]360_[[:space:]]High-Fidelity[[:space:]]Virtual[[:space:]]Try-On[[:space:]]from[[:space:]]Any[[:space:]]Viewing[[:space:]]Direction/4af7cfba-2431-4854-85f1-f3f911dcc811_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VTON-HandFit_[[:space:]]Virtual[[:space:]]Try-on[[:space:]]for[[:space:]]Arbitrary[[:space:]]Hand[[:space:]]Pose[[:space:]]Guided[[:space:]]by[[:space:]]Hand[[:space:]]Priors[[:space:]]Embedding/00462254-bb4b-4c24-9d0a-9a8f2054757a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/V^2Dial_[[:space:]]Unification[[:space:]]of[[:space:]]Video[[:space:]]and[[:space:]]Visual[[:space:]]Dialog[[:space:]]via[[:space:]]Multimodal[[:space:]]Experts/a2fbca87-c18a-41dc-864b-f66927e9f633_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Variance-Based[[:space:]]Membership[[:space:]]Inference[[:space:]]Attacks[[:space:]]Against[[:space:]]Large-Scale[[:space:]]Image[[:space:]]Captioning[[:space:]]Models/26bae7f8-8e37-4862-840e-6578a63d1d11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VasTSD_[[:space:]]Learning[[:space:]]3D[[:space:]]Vascular[[:space:]]Tree-state[[:space:]]Space[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Angiography[[:space:]]Synthesis/221d762b-beb5-4562-9aa6-3ebe0030c356_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VerbDiff_[[:space:]]Text-Only[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Enhanced[[:space:]]Interaction[[:space:]]Awareness/ad57dfe8-0121-4da2-a044-841b75f6e661_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViCaS_[[:space:]]A[[:space:]]Dataset[[:space:]]for[[:space:]]Combining[[:space:]]Holistic[[:space:]]and[[:space:]]Pixel-level[[:space:]]Video[[:space:]]Understanding[[:space:]]using[[:space:]]Captions[[:space:]]with[[:space:]]Grounded[[:space:]]Segmentation/ae891128-2b76-44ec-8a3e-060beba26fb9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViKIENet_[[:space:]]Towards[[:space:]]Efficient[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Virtual[[:space:]]Key[[:space:]]Instance[[:space:]]Enhanced[[:space:]]Network/c314dea2-42fe-4d4d-8bb4-7010c813e409_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViUniT_[[:space:]]Visual[[:space:]]Unit[[:space:]]Tests[[:space:]]for[[:space:]]More[[:space:]]Robust[[:space:]]Visual[[:space:]]Programming/8d692184-2a4f-4cca-806a-220832f3b327_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vid2Avatar-Pro_[[:space:]]Authentic[[:space:]]Avatar[[:space:]]from[[:space:]]Videos[[:space:]]in[[:space:]]the[[:space:]]Wild[[:space:]]via[[:space:]]Universal[[:space:]]Prior/61a0578e-e851-4106-a096-325f4e7911db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vid2Sim_[[:space:]]Generalizable,[[:space:]]Video-based[[:space:]]Reconstruction[[:space:]]of[[:space:]]Appearance,[[:space:]]Geometry[[:space:]]and[[:space:]]Physics[[:space:]]for[[:space:]]Mesh-free[[:space:]]Simulation/48cd44e1-bff7-400d-aceb-8a86fa95ccb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vid2Sim_[[:space:]]Realistic[[:space:]]and[[:space:]]Interactive[[:space:]]Simulation[[:space:]]from[[:space:]]Video[[:space:]]for[[:space:]]Urban[[:space:]]Navigation/a9b54426-b8df-4fc1-983c-717562672544_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VidBot_[[:space:]]Learning[[:space:]]Generalizable[[:space:]]3D[[:space:]]Actions[[:space:]]from[[:space:]]In-the-Wild[[:space:]]2D[[:space:]]Human[[:space:]]Videos[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Robotic[[:space:]]Manipulation/c9488d81-95eb-4b83-bf3b-3ddc8f3b7707_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VidComposition_[[:space:]]Can[[:space:]]MLLMs[[:space:]]Analyze[[:space:]]Compositions[[:space:]]in[[:space:]]Compiled[[:space:]]Videos_/6c9efc65-d2ba-48e3-a8d2-fe7c5c1d7646_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VidHalluc_[[:space:]]Evaluating[[:space:]]Temporal[[:space:]]Hallucinations[[:space:]]in[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Video[[:space:]]Understanding/c6c9e644-f51f-4e6b-aa79-2d374850db15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VidMuse_[[:space:]]A[[:space:]]Simple[[:space:]]Video-to-Music[[:space:]]Generation[[:space:]]Framework[[:space:]]with[[:space:]]Long-Short-Term[[:space:]]Modeling/e0ce05e4-9f59-4588-812e-59896766b6d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VidSeg_[[:space:]]Training-free[[:space:]]Video[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]based[[:space:]]on[[:space:]]Diffusion[[:space:]]Models/a8a75be1-b6e5-46b8-a340-187193161579_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VidTwin_[[:space:]]Video[[:space:]]VAE[[:space:]]with[[:space:]]Decoupled[[:space:]]Structure[[:space:]]and[[:space:]]Dynamics/396b7efb-b08c-4897-8d56-4437ed850d42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video[[:space:]]Depth[[:space:]]Anything_[[:space:]]Consistent[[:space:]]Depth[[:space:]]Estimation[[:space:]]for[[:space:]]Super-Long[[:space:]]Videos/3d97efd2-8c6d-451d-8d75-d43ab643dfda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video[[:space:]]Depth[[:space:]]without[[:space:]]Video[[:space:]]Models/adc6ac4e-2f14-4dd6-b43a-b385177519a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video[[:space:]]Language[[:space:]]Model[[:space:]]Pretraining[[:space:]]with[[:space:]]Spatio-temporal[[:space:]]Masking/592a1785-c0f4-44c0-ad31-218ccb7da263_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video[[:space:]]Motion[[:space:]]Transfer[[:space:]]with[[:space:]]Diffusion[[:space:]]Transformers/f0f4362e-9924-4b87-aeff-45869e95fb16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video[[:space:]]Summarization[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Models/49bb8211-5bbe-4fd6-a3e5-e9eae1a8e8d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video-3D[[:space:]]LLM_[[:space:]]Learning[[:space:]]Position-Aware[[:space:]]Video[[:space:]]Representation[[:space:]]for[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/c7916639-0a81-44c5-a96b-970d900fe738_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video-Bench_[[:space:]]Human-Aligned[[:space:]]Video[[:space:]]Generation[[:space:]]Benchmark/b48e59c4-b36d-49ef-a7ca-02fe49c0d380_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video-ColBERT_[[:space:]]Contextualized[[:space:]]Late[[:space:]]Interaction[[:space:]]for[[:space:]]Text-to-Video[[:space:]]Retrieval/6c2b31d1-930f-426e-b6df-3bce096a9320_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video-Guided[[:space:]]Foley[[:space:]]Sound[[:space:]]Generation[[:space:]]with[[:space:]]Multimodal[[:space:]]Controls/27ba25cd-566f-43d3-badb-af29c2dceb60_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video-MME_[[:space:]]The[[:space:]]First-Ever[[:space:]]Comprehensive[[:space:]]Evaluation[[:space:]]Benchmark[[:space:]]of[[:space:]]Multi-modal[[:space:]]LLMs[[:space:]]in[[:space:]]Video[[:space:]]Analysis/8df19168-ac1f-43c7-9bef-05050364c876_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video-Panda_[[:space:]]Parameter-efficient[[:space:]]Alignment[[:space:]]for[[:space:]]Encoder-free[[:space:]]Video-Language[[:space:]]Models/8e708f5e-f4c2-49cb-b40d-85c403ad2f23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video-XL_[[:space:]]Extra-Long[[:space:]]Vision[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Hour-Scale[[:space:]]Video[[:space:]]Understanding/410e6e98-5f80-4e5e-9672-9d5b8fa4130a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoAutoArena_[[:space:]]An[[:space:]]Automated[[:space:]]Arena[[:space:]]for[[:space:]]Evaluating[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models[[:space:]]in[[:space:]]Video[[:space:]]Analysis[[:space:]]through[[:space:]]User[[:space:]]Simulation/5a43a242-bd5d-4961-aed4-ae68b47ff795_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoComp_[[:space:]]Advancing[[:space:]]Fine-Grained[[:space:]]Compositional[[:space:]]and[[:space:]]Temporal[[:space:]]Alignment[[:space:]]in[[:space:]]Video-Text[[:space:]]Models/85b1b586-c77a-4209-812f-b601b6870f0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoDPO_[[:space:]]Omni-Preference[[:space:]]Alignment[[:space:]]for[[:space:]]Video[[:space:]]Diffusion[[:space:]]Generation/2f70dda9-0787-49af-abce-676cbf00394e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoDirector_[[:space:]]Precise[[:space:]]Video[[:space:]]Editing[[:space:]]via[[:space:]]Text-to-Video[[:space:]]Models/8f2614cf-ca8d-409c-94f9-479e72f1f29f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoEspresso_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Chain-of-Thought[[:space:]]Dataset[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Video[[:space:]]Reasoning[[:space:]]via[[:space:]]Core[[:space:]]Frame[[:space:]]Selection/eff006bb-cde1-4b27-a0ef-89b5fd58945a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoGEM_[[:space:]]Training-free[[:space:]]Action[[:space:]]Grounding[[:space:]]in[[:space:]]Videos/0a3fbcb1-73fe-47da-ac43-25ceb4ef9ba0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoGLaMM[[:space:]]_[[:space:]]A[[:space:]]Large[[:space:]]Multimodal[[:space:]]Model[[:space:]]for[[:space:]]Pixel-Level[[:space:]]Visual[[:space:]]Grounding[[:space:]]in[[:space:]]Videos/a7f6c402-bb8e-45ec-b9bf-5b07a3082f52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoGigaGAN_[[:space:]]Towards[[:space:]]Detail-rich[[:space:]]Video[[:space:]]Super-Resolution/52ec8084-0006-41dd-a7b0-3a3880698e67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoGuide_[[:space:]]Improving[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models[[:space:]]without[[:space:]]Training[[:space:]]Through[[:space:]]a[[:space:]]Teacher's[[:space:]]Guide/8fb0f104-e76b-45ee-932c-212b94049d7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoHandles_[[:space:]]Editing[[:space:]]3D[[:space:]]Object[[:space:]]Compositions[[:space:]]in[[:space:]]Videos[[:space:]]Using[[:space:]]Video[[:space:]]Generative[[:space:]]Priors/2cf37ed5-ea0a-4c05-acd5-0c34fb8789ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoICL_[[:space:]]Confidence-based[[:space:]]Iterative[[:space:]]In-context[[:space:]]Learning[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Video[[:space:]]Understanding/3dc9da78-cea1-471c-b4db-ee52c6853efe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoMage_[[:space:]]Multi-Subject[[:space:]]and[[:space:]]Motion[[:space:]]Customization[[:space:]]of[[:space:]]Text-to-Video[[:space:]]Diffusion[[:space:]]Models/80815955-c6e2-4c64-a2b6-3555fa9f4dc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoRefer[[:space:]]Suite_[[:space:]]Advancing[[:space:]]Spatial-Temporal[[:space:]]Object[[:space:]]Understanding[[:space:]]with[[:space:]]Video[[:space:]]LLM/1eb2e581-034c-4440-aa2e-5d6157a422fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoSPatS_[[:space:]]Video[[:space:]]SPatiotemporal[[:space:]]Splines[[:space:]]for[[:space:]]Disentangled[[:space:]]Occlusion,[[:space:]]Appearance[[:space:]]and[[:space:]]Motion[[:space:]]Modeling[[:space:]]and[[:space:]]Editing/47539518-9a92-424c-830b-8495e559a706_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoScene_[[:space:]]Distilling[[:space:]]Video[[:space:]]Diffusion[[:space:]]Model[[:space:]]to[[:space:]]Generate[[:space:]]3D[[:space:]]Scenes[[:space:]]in[[:space:]]One[[:space:]]Step/3d9c7184-8bc7-4684-b4b3-dee716e5015a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoTree_[[:space:]]Adaptive[[:space:]]Tree-based[[:space:]]Video[[:space:]]Representation[[:space:]]for[[:space:]]LLM[[:space:]]Reasoning[[:space:]]on[[:space:]]Long[[:space:]]Videos/40b3e4b5-66bb-4e61-88c1-0c383821b986_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoWorld_[[:space:]]Exploring[[:space:]]Knowledge[[:space:]]Learning[[:space:]]from[[:space:]]Unlabeled[[:space:]]Videos/d7364fb4-c126-4443-875f-c1f7daf2bf13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Viewpoint[[:space:]]Rosetta[[:space:]]Stone_[[:space:]]Unlocking[[:space:]]Unpaired[[:space:]]Ego-Exo[[:space:]]Videos[[:space:]]for[[:space:]]View-invariant[[:space:]]Representation[[:space:]]Learning/ae8678aa-d4cb-4d54-aa26-b58e33e1e461_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViiNeuS_[[:space:]]Volumetric[[:space:]]Initialization[[:space:]]for[[:space:]]Implicit[[:space:]]Neural[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]of[[:space:]]Urban[[:space:]]Scenes[[:space:]]with[[:space:]]Limited[[:space:]]Image[[:space:]]Overlap/23124bd5-ecfd-490d-a48b-e03c3ec06af3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VinTAGe_[[:space:]]Joint[[:space:]]Video[[:space:]]and[[:space:]]Text[[:space:]]Conditioning[[:space:]]for[[:space:]]Holistic[[:space:]]Audio[[:space:]]Generation/a74d1a60-3b5d-4c9c-a324-7df50d716e1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VinaBench_[[:space:]]Benchmark[[:space:]]for[[:space:]]Faithful[[:space:]]and[[:space:]]Consistent[[:space:]]Visual[[:space:]]Narratives/e5948aac-c2d3-4df8-a7f3-82cb99be6728_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vision-Guided[[:space:]]Action_[[:space:]]Enhancing[[:space:]]3D[[:space:]]Human[[:space:]]Motion[[:space:]]Prediction[[:space:]]with[[:space:]]Gaze-informed[[:space:]]Affordance[[:space:]]in[[:space:]]3D[[:space:]]Scenes/9b42c369-4ef9-4c45-9192-f6b0a17d8bda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vision-Language[[:space:]]Embodiment[[:space:]]for[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/959b3b25-9f17-45a6-aab1-48ac53aa864f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vision-Language[[:space:]]Gradient[[:space:]]Descent-driven[[:space:]]All-in-One[[:space:]]Deep[[:space:]]Unfolding[[:space:]]Networks/bd772a11-bc94-4fb3-8aee-976fd29748e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vision-Language[[:space:]]Model[[:space:]]IP[[:space:]]Protection[[:space:]]via[[:space:]]Prompt-based[[:space:]]Learning/e2d53e64-2ba0-4eec-9a41-e03d1858d841_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vision-Language[[:space:]]Models[[:space:]]Do[[:space:]]Not[[:space:]]Understand[[:space:]]Negation/f44bf4ea-b507-4a1f-8e9b-87600bcf741b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VisionArena_[[:space:]]230k[[:space:]]Real[[:space:]]World[[:space:]]User-VLM[[:space:]]Conversations[[:space:]]with[[:space:]]Preference[[:space:]]Labels/f703962d-2329-4c48-bd31-19be8745b28f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VisionPAD_[[:space:]]A[[:space:]]Vision-Centric[[:space:]]Pre-training[[:space:]]Paradigm[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/4e6d5045-e8c8-4646-a828-f050db70849c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VisionZip_[[:space:]]Longer[[:space:]]is[[:space:]]Better[[:space:]]but[[:space:]]Not[[:space:]]Necessary[[:space:]]in[[:space:]]Vision[[:space:]]Language[[:space:]]Models/d6a76b28-eeb8-41ac-bee1-8e5731b2110b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Agentic[[:space:]]AI[[:space:]]for[[:space:]]Spatial[[:space:]]Reasoning[[:space:]]with[[:space:]]a[[:space:]]Dynamic[[:space:]]API/18209ccb-89a4-4fe8-8dec-31f34f20c734_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Consensus[[:space:]]Prompting[[:space:]]for[[:space:]]Co-Salient[[:space:]]Object[[:space:]]Detection/3f1a6dd0-b317-40d4-8fff-a7533b787af0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Lexicon_[[:space:]]Rich[[:space:]]Image[[:space:]]Features[[:space:]]in[[:space:]]Language[[:space:]]Space/18fabbbb-6482-437e-bf79-727cc6d61f01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Co-Speech[[:space:]]Gesture[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Implicit[[:space:]]Motion-Audio[[:space:]]Entanglement/22309812-5e22-4431-b27e-c647daff1c38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Persona_[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Full-Body[[:space:]]Human[[:space:]]Customization/ec6e8648-d94c-4b41-88eb-4e6b2f8bf9bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Prompting[[:space:]]for[[:space:]]One-shot[[:space:]]Controllable[[:space:]]Video[[:space:]]Editing[[:space:]]without[[:space:]]Inversion/0ace3724-f28b-4285-9197-7b13c2d364ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Representation[[:space:]]Learning[[:space:]]through[[:space:]]Causal[[:space:]]Intervention[[:space:]]for[[:space:]]Controllable[[:space:]]Image[[:space:]]Editing/e50e95b3-c676-4213-8a7e-0a922b2af357_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]and[[:space:]]Semantic[[:space:]]Prompt[[:space:]]Collaboration[[:space:]]for[[:space:]]Generalized[[:space:]]Zero-Shot[[:space:]]Learning/84096cac-4565-47ea-80af-e134222f8001_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual-Instructed[[:space:]]Degradation[[:space:]]Diffusion[[:space:]]for[[:space:]]All-in-One[[:space:]]Image[[:space:]]Restoration/fd008495-f795-478a-845e-9e57b9805e9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VladVA_[[:space:]]Discriminative[[:space:]]Fine-tuning[[:space:]]of[[:space:]]LVLMs/d52c1f74-3033-44a3-a3ee-dbdde51746cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VoCo-LLaMA_[[:space:]]Towards[[:space:]]Vision[[:space:]]Compression[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Models/56026415-39a0-4683-89a1-38cbea5aa03b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VolFormer_[[:space:]]Explore[[:space:]]More[[:space:]]Comprehensive[[:space:]]Cube[[:space:]]Interaction[[:space:]]for[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Restoration[[:space:]]and[[:space:]]Beyond/9a7bca8f-c213-464c-8f52-51f264b27773_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Volume[[:space:]]Tells_[[:space:]]Dual[[:space:]]Cycle-Consistent[[:space:]]Diffusion[[:space:]]for[[:space:]]3D[[:space:]]Fluorescence[[:space:]]Microscopy[[:space:]]De-noising[[:space:]]and[[:space:]]Super-Resolution/6a906115-1038-43f5-a1c0-9f109ae3f1e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Volumetric[[:space:]]Surfaces_[[:space:]]Representing[[:space:]]Fuzzy[[:space:]]Geometries[[:space:]]with[[:space:]]Layered[[:space:]]Meshes/fa52d6eb-803c-41b8-a857-f48c6d428120_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Volumetrically[[:space:]]Consistent[[:space:]]3D[[:space:]]Gaussian[[:space:]]Rasterization/69609d3e-37b4-4870-a5ed-a89ea97b5f6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VoteFlow_[[:space:]]Enforcing[[:space:]]Local[[:space:]]Rigidity[[:space:]]in[[:space:]]Self-Supervised[[:space:]]Scene[[:space:]]Flow/40638b97-89cb-4926-aea4-86a14a1df0c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VoxelSplat_[[:space:]]Dynamic[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]as[[:space:]]an[[:space:]]Effective[[:space:]]Loss[[:space:]]for[[:space:]]Occupancy[[:space:]]and[[:space:]]Flow[[:space:]]Prediction/1bf99923-2cf4-4b2b-81a1-520dfd077810_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WAVE_[[:space:]]Weight[[:space:]]Templates[[:space:]]for[[:space:]]Adaptive[[:space:]]Initialization[[:space:]]of[[:space:]]Variable-sized[[:space:]]Models/e318d02a-7917-46e5-937d-512cf627caef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WF-VAE_[[:space:]]Enhancing[[:space:]]Video[[:space:]]VAE[[:space:]]by[[:space:]]Wavelet-Driven[[:space:]]Energy[[:space:]]Flow[[:space:]]for[[:space:]]Latent[[:space:]]Video[[:space:]]Diffusion[[:space:]]Model/c4444e54-5b1e-4b6a-8bbf-270c7b1f71f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WISE_[[:space:]]A[[:space:]]Framework[[:space:]]for[[:space:]]Gigapixel[[:space:]]Whole-Slide-Image[[:space:]]Lossless[[:space:]]Compression/dd1ff0be-5929-40de-958f-cc648f085993_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WISH_[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Instance[[:space:]]Segmentation[[:space:]]using[[:space:]]Heterogeneous[[:space:]]Labels/b7d1d0cf-4f88-4e86-ac05-f766fb67ed07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WISNet_[[:space:]]Pseudo[[:space:]]Label[[:space:]]Generation[[:space:]]on[[:space:]]Unbalanced[[:space:]]and[[:space:]]Patch[[:space:]]Annotated[[:space:]]Waste[[:space:]]Images/aefc5043-bf51-4b0a-9f78-eaac1d704325_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Watermarking[[:space:]]One[[:space:]]for[[:space:]]All_[[:space:]]A[[:space:]]Robust[[:space:]]Watermarking[[:space:]]Scheme[[:space:]]Against[[:space:]]Partial[[:space:]]Image[[:space:]]Theft/3e960148-a90b-4396-91d7-681fc2bb6c2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Wav2Sem_[[:space:]]Plug-and-Play[[:space:]]Audio[[:space:]]Semantic[[:space:]]Decoupling[[:space:]]for[[:space:]]3D[[:space:]]Speech-Driven[[:space:]]Facial[[:space:]]Animation/1f33851d-335f-4269-b141-04605eb61c66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Wavelet[[:space:]]and[[:space:]]Prototype[[:space:]]Augmented[[:space:]]Query-based[[:space:]]Transformer[[:space:]]for[[:space:]]Pixel-level[[:space:]]Surface[[:space:]]Defect[[:space:]]Detection/140ae139-3395-48bf-b577-835e4298a027_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WeGen_[[:space:]]A[[:space:]]Unified[[:space:]]Model[[:space:]]for[[:space:]]Interactive[[:space:]]Multimodal[[:space:]]Generation[[:space:]]as[[:space:]]We[[:space:]]Chat/71a20a6d-ca16-41a1-ae4a-04622bd31849_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WeakMCN_[[:space:]]Multi-task[[:space:]]Collaborative[[:space:]]Network[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Referring[[:space:]]Expression[[:space:]]Comprehension[[:space:]]and[[:space:]]Segmentation/22f7bfc0-1476-4304-88f4-9b3898f4fdab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]via[[:space:]]Progressive[[:space:]]Confidence[[:space:]]Region[[:space:]]Expansion/3bbafa66-04c2-4e9f-b4ee-20b90c846e75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Weakly[[:space:]]Supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization[[:space:]]via[[:space:]]Dual-Prior[[:space:]]Collaborative[[:space:]]Learning[[:space:]]Guided[[:space:]]by[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/462776bb-18ba-409c-a604-30cec6df7bb8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WeatherGen_[[:space:]]A[[:space:]]Unified[[:space:]]Diverse[[:space:]]Weather[[:space:]]Generator[[:space:]]for[[:space:]]LiDAR[[:space:]]Point[[:space:]]Clouds[[:space:]]via[[:space:]]Spider[[:space:]]Mamba[[:space:]]Diffusion/313200e5-9037-4a00-b1a9-6dfc4e25370e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/What[[:space:]]Makes[[:space:]]a[[:space:]]Good[[:space:]]Dataset[[:space:]]for[[:space:]]Knowledge[[:space:]]Distillation_/1b108190-3b13-4b10-9cf7-b78a907351a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/What's[[:space:]]in[[:space:]]the[[:space:]]Image_[[:space:]]A[[:space:]]Deep-Dive[[:space:]]into[[:space:]]the[[:space:]]Vision[[:space:]]of[[:space:]]Vision[[:space:]]Language[[:space:]]Models/b54660f1-1823-48ec-804a-e9659e000387_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/When[[:space:]]Domain[[:space:]]Generalization[[:space:]]meets[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery_[[:space:]]An[[:space:]]Adaptive[[:space:]]Task-Arithmetic[[:space:]]Driven[[:space:]]Approach/9bb4f2b3-9e26-4366-8f22-1e02a28a58d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/When[[:space:]]the[[:space:]]Future[[:space:]]Becomes[[:space:]]the[[:space:]]Past_[[:space:]]Taming[[:space:]]Temporal[[:space:]]Correspondence[[:space:]]for[[:space:]]Self-supervised[[:space:]]Video[[:space:]]Representation[[:space:]]Learning/170f8e64-f3c8-4443-85ce-3fdea3a5090c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Where[[:space:]]the[[:space:]]Devil[[:space:]]Hides_[[:space:]]Deepfake[[:space:]]Detectors[[:space:]]Can[[:space:]]No[[:space:]]Longer[[:space:]]Be[[:space:]]Trusted/0a9ed4bb-ee16-4460-9151-4efa35dd9254_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Where's[[:space:]]the[[:space:]]Liability[[:space:]]in[[:space:]]the[[:space:]]Generative[[:space:]]Era_[[:space:]]Recovery-based[[:space:]]Black-Box[[:space:]]Detection[[:space:]]of[[:space:]]AI-Generated[[:space:]]Content/0f33d032-4a6c-4e30-9d20-20935beea5f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Which[[:space:]]Viewpoint[[:space:]]Shows[[:space:]]it[[:space:]]Best_[[:space:]]Language[[:space:]]for[[:space:]]Weakly[[:space:]]Supervising[[:space:]]View[[:space:]]Selection[[:space:]]in[[:space:]]Multi-view[[:space:]]Instructional[[:space:]]Videos/55424da2-46d1-49d3-8022-feb7ecca531c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WiLoR_[[:space:]]End-to-end[[:space:]]3D[[:space:]]Hand[[:space:]]Localization[[:space:]]and[[:space:]]Reconstruction[[:space:]]in-the-wild/f948aa54-add5-4383-b304-884998c61be1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WildAvatar_[[:space:]]Learning[[:space:]]In-the-wild[[:space:]]3D[[:space:]]Avatars[[:space:]]from[[:space:]]the[[:space:]]Web/c2d030d4-e668-42c7-bb13-b16f8d5a511b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WildGS-SLAM_[[:space:]]Monocular[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]SLAM[[:space:]]in[[:space:]]Dynamic[[:space:]]Environments/76862b6f-6552-4501-9c2f-df98e6b2a3a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WonderWorld_[[:space:]]Interactive[[:space:]]3D[[:space:]]Scene[[:space:]]Generation[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/8b19ecda-2614-4cf0-8480-0def34c5da26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Wonderland_[[:space:]]Navigating[[:space:]]3D[[:space:]]Scenes[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/36525cf0-e9d2-48b8-92e1-904c7261d1e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Words[[:space:]]or[[:space:]]Vision_[[:space:]]Do[[:space:]]Vision-Language[[:space:]]Models[[:space:]]Have[[:space:]]Blind[[:space:]]Faith[[:space:]]in[[:space:]]Text_/b5906f56-1e18-4565-973a-910fa73da3f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/World-consistent[[:space:]]Video[[:space:]]Diffusion[[:space:]]with[[:space:]]Explicit[[:space:]]3D[[:space:]]Modeling/f965e6f9-4f7e-45b6-b5ba-1a5254faa1f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/X-Dyna_[[:space:]]Expressive[[:space:]]Dynamic[[:space:]]Human[[:space:]]Image[[:space:]]Animation/54e18dc9-a7a4-45f6-a159-cce110d54e3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/XLRS-Bench_[[:space:]]Could[[:space:]]Your[[:space:]]Multimodal[[:space:]]LLMs[[:space:]]Understand[[:space:]]Extremely[[:space:]]Large[[:space:]]Ultra-High-Resolution[[:space:]]Remote[[:space:]]Sensing[[:space:]]Imagery_/48f9d345-be75-472b-b622-b9ab56f15e22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Yo'Chameleon_[[:space:]]Personalized[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Generation/f509dede-230e-4c36-b7bc-c33d9434f7ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/You[[:space:]]See[[:space:]]it,[[:space:]]You[[:space:]]Got[[:space:]]it_[[:space:]]Learning[[:space:]]3D[[:space:]]Creation[[:space:]]on[[:space:]]Pose-Free[[:space:]]Videos[[:space:]]at[[:space:]]Scale/d328176d-78f8-40e2-af97-c59496f92dfd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Your[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Model[[:space:]]Only[[:space:]]Needs[[:space:]]A[[:space:]]Few[[:space:]]Attention[[:space:]]Heads[[:space:]]For[[:space:]]Visual[[:space:]]Grounding/c4998c37-caf9-4895-94a1-69270509d598_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Your[[:space:]]Scale[[:space:]]Factors[[:space:]]are[[:space:]]My[[:space:]]Weapon_[[:space:]]Targeted[[:space:]]Bit-Flip[[:space:]]Attacks[[:space:]]on[[:space:]]Vision[[:space:]]Transformers[[:space:]]via[[:space:]]Scale[[:space:]]Factor[[:space:]]Manipulation/a65b7d6e-4de4-4e3c-8176-d7304d63cc93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Your[[:space:]]ViT[[:space:]]is[[:space:]]Secretly[[:space:]]an[[:space:]]Image[[:space:]]Segmentation[[:space:]]Model/f1eb4195-9d6c-4fad-afc5-61fd8da4cbf4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Z-Magic_[[:space:]]Zero-shot[[:space:]]Multiple[[:space:]]Attributes[[:space:]]Guided[[:space:]]Image[[:space:]]Creator/bc3be614-b79a-4138-8d79-61b6837b148f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-1-to-A_[[:space:]]Zero-Shot[[:space:]]One[[:space:]]Image[[:space:]]to[[:space:]]Animatable[[:space:]]Head[[:space:]]Avatars[[:space:]]Using[[:space:]]Video[[:space:]]Diffusion/0fc00f29-85c4-4d3f-9811-f7a9693e7075_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-Shot[[:space:]]4D[[:space:]]Lidar[[:space:]]Panoptic[[:space:]]Segmentation/a80b3827-5425-448f-86f7-67b4d3d14684_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-Shot[[:space:]]Blind-spot[[:space:]]Image[[:space:]]Denoising[[:space:]]via[[:space:]]Implicit[[:space:]]Neural[[:space:]]Sampling/c741fbc2-fdcf-4407-a5b9-a0f010311969_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-Shot[[:space:]]Head[[:space:]]Swapping[[:space:]]in[[:space:]]Real-World[[:space:]]Scenarios/b4b95ca2-0483-4a83-9c74-3188680aff1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-Shot[[:space:]]Image[[:space:]]Restoration[[:space:]]Using[[:space:]]Few-Step[[:space:]]Guidance[[:space:]]of[[:space:]]Consistency[[:space:]]Models[[:space:]](and[[:space:]]Beyond)/52b76595-5101-4a72-a9c3-969757386cb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-Shot[[:space:]]Monocular[[:space:]]Scene[[:space:]]Flow[[:space:]]Estimation[[:space:]]in[[:space:]]the[[:space:]]Wild/c1d2cef9-1607-407d-a7e6-02fd82029d4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-Shot[[:space:]]Novel[[:space:]]View[[:space:]]and[[:space:]]Depth[[:space:]]Synthesis[[:space:]]with[[:space:]]Multi-View[[:space:]]Geometric[[:space:]]Diffusion/a0b00b4d-84fa-43b8-82fa-0c235320a434_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-Shot[[:space:]]Styled[[:space:]]Text[[:space:]]Image[[:space:]]Generation,[[:space:]]but[[:space:]]Make[[:space:]]It[[:space:]]Autoregressive/5fdf7c0c-13de-4025-ae8e-f164575dde46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-shot[[:space:]]3D[[:space:]]Question[[:space:]]Answering[[:space:]]via[[:space:]]Voxel-based[[:space:]]Dynamic[[:space:]]Token[[:space:]]Compression/8d8ac0d9-d50c-4ccf-adca-5ae5694b3711_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-shot[[:space:]]RGB-D[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration[[:space:]]with[[:space:]]Pre-trained[[:space:]]Large[[:space:]]Vision[[:space:]]Model/cc1335fb-3748-4356-b884-7218ee5de324_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ZeroGrasp_[[:space:]]Zero-Shot[[:space:]]Shape[[:space:]]Reconstruction[[:space:]]Enabled[[:space:]]Robotic[[:space:]]Grasping/eebebb2c-dd4f-4c4d-bbde-d36484d550d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ZeroVO_[[:space:]]Visual[[:space:]]Odometry[[:space:]]with[[:space:]]Minimal[[:space:]]Assumptions/07d97fcd-beb6-4db8-9d15-727f79d06230_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ZoomLDM_[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Multi-scale[[:space:]]Image[[:space:]]Generation/64d1b96f-e41e-4ee5-88fe-92864a01628b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/beta-FFT_[[:space:]]Nonlinear[[:space:]]Interpolation[[:space:]]and[[:space:]]Differentiated[[:space:]]Training[[:space:]]Strategies[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/d97dc35c-1298-4ebe-af81-4eec695a335e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/dFLMoE_[[:space:]]Decentralized[[:space:]]Federated[[:space:]]Learning[[:space:]]via[[:space:]]Mixture[[:space:]]of[[:space:]]Experts[[:space:]]for[[:space:]]Medical[[:space:]]Data[[:space:]]Analysis/19b2c42f-ce51-44ed-b899-64702962cff1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/g3D-LF_[[:space:]]Generalizable[[:space:]]3D-Language[[:space:]]Feature[[:space:]]Fields[[:space:]]for[[:space:]]Embodied[[:space:]]Tasks/319bbfa4-f317-493b-a350-5e2125370b5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/h-Edit_[[:space:]]Effective[[:space:]]and[[:space:]]Flexible[[:space:]]Diffusion-Based[[:space:]]Editing[[:space:]]via[[:space:]]Doob's[[:space:]]h-Transform/a31f06fe-0eb5-4f9e-ab73-37feb4a38682_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/iG-6DoF_[[:space:]]Model-free[[:space:]]6DoF[[:space:]]Pose[[:space:]]Estimation[[:space:]]for[[:space:]]Unseen[[:space:]]Object[[:space:]]via[[:space:]]Iterative[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/56588c3b-cd81-42c5-a2a1-c7cdd5a9d9cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/iSegMan_[[:space:]]Interactive[[:space:]]Segment-and-Manipulate[[:space:]]3D[[:space:]]Gaussians/41e89e4d-a900-4b4c-acdd-4a77ec356f1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/nnWNet_[[:space:]]Rethinking[[:space:]]the[[:space:]]Use[[:space:]]of[[:space:]]Transformers[[:space:]]in[[:space:]]Biomedical[[:space:]]Image[[:space:]]Segmentation[[:space:]]and[[:space:]]Calling[[:space:]]for[[:space:]]a[[:space:]]Unified[[:space:]]Evaluation[[:space:]]Benchmark/1b40876e-3c13-4c66-a6cf-340a49d08b05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/pFedMxF_[[:space:]]Personalized[[:space:]]Federated[[:space:]]Class-Incremental[[:space:]]Learning[[:space:]]with[[:space:]]Mixture[[:space:]]of[[:space:]]Frequency[[:space:]]Aggregation/af5803f4-69bd-43df-bcea-1df37a7dcf4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/v-CLR_[[:space:]]View-Consistent[[:space:]]Learning[[:space:]]for[[:space:]]Open-World[[:space:]]Instance[[:space:]]Segmentation/ef2f822f-1ae9-4363-af6b-d621ee5b7ef6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/vesselFM_[[:space:]]A[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Universal[[:space:]]3D[[:space:]]Blood[[:space:]]Vessel[[:space:]]Segmentation/ae7b8187-15d3-4443-9213-c4542fccaa23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/1-Lipschitz[[:space:]]Layers[[:space:]]Compared_[[:space:]]Memory[[:space:]]Speed[[:space:]]and[[:space:]]Certifiable[[:space:]]Robustness/dc5df91e-4a9d-43cc-9e47-6667f322a373_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/2S-UDF_[[:space:]]A[[:space:]]Novel[[:space:]]Two-stage[[:space:]]UDF[[:space:]]Learning[[:space:]]Method[[:space:]]for[[:space:]]Robust[[:space:]]Non-watertight[[:space:]]Model[[:space:]]Reconstruction[[:space:]]from[[:space:]]Multi-view[[:space:]]Images/42a5d7da-3ecd-41f5-9527-9d0c2dfbf201_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/360+x_[[:space:]]A[[:space:]]Panoptic[[:space:]]Multi-modal[[:space:]]Scene[[:space:]]Understanding[[:space:]]Dataset/1b673a67-1eed-49d6-b7e1-f0b4a9d871e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/360DVD_[[:space:]]Controllable[[:space:]]Panorama[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]360-Degree[[:space:]]Video[[:space:]]Diffusion[[:space:]]Model/b8d3f2a2-b423-4011-8c8e-f2db936aeb7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/360Loc_[[:space:]]A[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Omnidirectional[[:space:]]Visual[[:space:]]Localization[[:space:]]with[[:space:]]Cross-device[[:space:]]Queries/15eb225d-3032-419c-84b0-35d6ec576cbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D[[:space:]]Building[[:space:]]Reconstruction[[:space:]]from[[:space:]]Monocular[[:space:]]Remote[[:space:]]Sensing[[:space:]]Images[[:space:]]with[[:space:]]Multi-level[[:space:]]Supervisions/3c7a0eb0-f8af-4281-92e0-1ce5ec55d350_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D[[:space:]]Face[[:space:]]Reconstruction[[:space:]]with[[:space:]]the[[:space:]]Geometric[[:space:]]Guidance[[:space:]]of[[:space:]]Facial[[:space:]]Part[[:space:]]Segmentation/29911afb-57cf-4105-bc3b-b432a117add8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D[[:space:]]Face[[:space:]]Tracking[[:space:]]from[[:space:]]2D[[:space:]]Video[[:space:]]through[[:space:]]Iterative[[:space:]]Dense[[:space:]]UV[[:space:]]to[[:space:]]Image[[:space:]]Flow/658f2c5b-c12d-479a-94f3-61e9ffc8e1df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D[[:space:]]Facial[[:space:]]Expressions[[:space:]]through[[:space:]]Analysis-by-Neural-Synthesis/444ecb6f-5ab4-45bb-9d08-b2b359c08da3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D[[:space:]]Feature[[:space:]]Tracking[[:space:]]via[[:space:]]Event[[:space:]]Camera/a6809d22-03e7-4639-b845-8393b79ecc8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D[[:space:]]Geometry-Aware[[:space:]]Deformable[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Dynamic[[:space:]]View[[:space:]]Synthesis/f455d128-070e-4b9d-a550-262379c7f3f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D[[:space:]]Human[[:space:]]Pose[[:space:]]Perception[[:space:]]from[[:space:]]Egocentric[[:space:]]Stereo[[:space:]]Videos/95770e99-65fc-4fd7-9de3-96977a97b4b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D[[:space:]]LiDAR[[:space:]]Mapping[[:space:]]in[[:space:]]Dynamic[[:space:]]Environments[[:space:]]using[[:space:]]a[[:space:]]4D[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representation/024f2dc1-2c03-4b1e-a716-0e0aea35b1de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D[[:space:]]Multi-frame[[:space:]]Fusion[[:space:]]for[[:space:]]Video[[:space:]]Stabilization/475676e5-4dd7-4a8c-bd05-cc44ef21267a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D[[:space:]]Neural[[:space:]]Edge[[:space:]]Reconstruction/12034c9b-4470-4339-9189-38596581605f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D[[:space:]]Paintbrush_[[:space:]]Local[[:space:]]Stylization[[:space:]]of[[:space:]]3D[[:space:]]Shapes[[:space:]]with[[:space:]]Cascaded[[:space:]]Score[[:space:]]Distillation/669e6bfe-eb9e-4f5b-a53c-23335eda80fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D-Aware[[:space:]]Face[[:space:]]Editing[[:space:]]via[[:space:]]Warping-Guided[[:space:]]Latent[[:space:]]Direction[[:space:]]Learning/d47f630a-17d8-4298-a368-699d1959d603_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D-LFM_[[:space:]]Lifting[[:space:]]Foundation[[:space:]]Model/5d227142-e6b0-440e-bad4-facab1940a16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3D-SceneDreamer_[[:space:]]Text-Driven[[:space:]]3D-Consistent[[:space:]]Scene[[:space:]]Generation/e514998b-c539-47e4-bf66-6c5fccc605eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3DFIRES_[[:space:]]Few[[:space:]]Image[[:space:]]3D[[:space:]]REconstruction[[:space:]]for[[:space:]]Scenes[[:space:]]with[[:space:]]Hidden[[:space:]]Surfaces/9acb5370-2e99-4481-9b63-bbd93724edf4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3DGS-Avatar_[[:space:]]Animatable[[:space:]]Avatars[[:space:]]via[[:space:]]Deformable[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/0f8abe9e-31c6-4dc4-9520-66dabe1eb0cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3DGStream_[[:space:]]On-the-Fly[[:space:]]Training[[:space:]]of[[:space:]]3D[[:space:]]Gaussians[[:space:]]for[[:space:]]Efficient[[:space:]]Streaming[[:space:]]of[[:space:]]Photo-Realistic[[:space:]]Free-Viewpoint[[:space:]]Videos/94b406dc-a25e-4b49-8259-ce68b53e5886_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3DInAction_[[:space:]]Understanding[[:space:]]Human[[:space:]]Actions[[:space:]]in[[:space:]]3D[[:space:]]Point[[:space:]]Clouds/5d416e0e-fbb0-491c-8e72-ffffff1eb68b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3DSFLabelling_[[:space:]]Boosting[[:space:]]3D[[:space:]]Scene[[:space:]]Flow[[:space:]]Estimation[[:space:]]by[[:space:]]Pseudo[[:space:]]Auto-labelling/128911ab-0f2f-4697-8895-080e5b45c36b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3DToonify_[[:space:]]Creating[[:space:]]Your[[:space:]]High-Fidelity[[:space:]]3D[[:space:]]Stylized[[:space:]]Avatar[[:space:]]Easily[[:space:]]from[[:space:]]2D[[:space:]]Portrait[[:space:]]Images/2c81e75a-1abf-4d0b-aa9d-80d61a8cb264_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/3DiffTection_[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Geometry-Aware[[:space:]]Diffusion[[:space:]]Features/0c8075a5-3d90-4e50-a4ef-fbf63dd9f1bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/4D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Real-Time[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Rendering/c4bfa810-f46e-49e4-9e19-ae1a9e3dcad6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/4D-DRESS_[[:space:]]A[[:space:]]4D[[:space:]]Dataset[[:space:]]of[[:space:]]Real-World[[:space:]]Human[[:space:]]Clothing[[:space:]]With[[:space:]]Semantic[[:space:]]Annotations/cd2548ca-a539-45a3-adaf-36a364d6da68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/4D-fy_[[:space:]]Text-to-4D[[:space:]]Generation[[:space:]]Using[[:space:]]Hybrid[[:space:]]Score[[:space:]]Distillation[[:space:]]Sampling/7295c7f1-f21d-431e-aa65-0a0fb95fe12c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/4K4D_[[:space:]]Real-Time[[:space:]]4D[[:space:]]View[[:space:]]Synthesis[[:space:]]at[[:space:]]4K[[:space:]]Resolution/6d74ca33-515d-4b03-96e1-8cccfa68be60_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/6D-Diff_[[:space:]]A[[:space:]]Keypoint[[:space:]]Diffusion[[:space:]]Framework[[:space:]]for[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/1d3927f2-2533-4713-91b0-b3f9e13c8aed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Backpack[[:space:]]Full[[:space:]]of[[:space:]]Skills_[[:space:]]Egocentric[[:space:]]Video[[:space:]]Understanding[[:space:]]with[[:space:]]Diverse[[:space:]]Task[[:space:]]Perspectives/b19cc746-13a8-4ac2-b79c-a1691351681c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Bayesian[[:space:]]Approach[[:space:]]to[[:space:]]OOD[[:space:]]Robustness[[:space:]]in[[:space:]]Image[[:space:]]Classification/d0e61d97-e025-4ae4-a494-3d44cf79404b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Call[[:space:]]to[[:space:]]Reflect[[:space:]]on[[:space:]]Evaluation[[:space:]]Practices[[:space:]]for[[:space:]]Age[[:space:]]Estimation_[[:space:]]Comparative[[:space:]]Analysis[[:space:]]of[[:space:]]the[[:space:]]State-of-the-Art[[:space:]]and[[:space:]]a[[:space:]]Unified[[:space:]]Benchmark/393f0825-ecb8-44e6-bdfa-5dde4b82ecdb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Category[[:space:]]Agnostic[[:space:]]Model[[:space:]]for[[:space:]]Visual[[:space:]]Rearrangment/4b328694-69ab-47a2-83d3-ce2efe00b0f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Closer[[:space:]]Look[[:space:]]at[[:space:]]the[[:space:]]Few-Shot[[:space:]]Adaptation[[:space:]]of[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/69aa9b91-03a6-4c14-a53b-96602951c67b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Conditional[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Probabilistic[[:space:]]Model[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Upsampling/80eefa2b-3d90-4d98-ab03-f2521d12efac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Dual-Augmentor[[:space:]]Framework[[:space:]]for[[:space:]]Domain[[:space:]]Generalization[[:space:]]in[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/2e309a58-2e8d-4563-8890-368854bbd34f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Dynamic[[:space:]]Kernel[[:space:]]Prior[[:space:]]Model[[:space:]]for[[:space:]]Unsupervised[[:space:]]Blind[[:space:]]Image[[:space:]]Super-Resolution/eb377d27-ee40-49e4-9796-048cc8e1c35d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]General[[:space:]]and[[:space:]]Efficient[[:space:]]Training[[:space:]]for[[:space:]]Transformer[[:space:]]via[[:space:]]Token[[:space:]]Expansion/6b319bca-e10d-4650-be77-29bcc4ffd8dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Generative[[:space:]]Approach[[:space:]]for[[:space:]]Wikipedia-Scale[[:space:]]Visual[[:space:]]Entity[[:space:]]Recognition/25668f69-d21b-4819-83f7-a45db4e2f055_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Noisy[[:space:]]Elephant[[:space:]]in[[:space:]]the[[:space:]]Room_[[:space:]]Is[[:space:]]Your[[:space:]]Out-of-Distribution[[:space:]]Detector[[:space:]]Robust[[:space:]]to[[:space:]]Label[[:space:]]Noise_/8eab7491-89d9-4dd1-8ee1-40b9bd851b01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Pedestrian[[:space:]]is[[:space:]]Worth[[:space:]]One[[:space:]]Prompt_[[:space:]]Towards[[:space:]]Language[[:space:]]Guidance[[:space:]]Person[[:space:]]Re-Identification/85746221-1e2b-4579-be8b-1626ff544e58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Physics-informed[[:space:]]Low-rank[[:space:]]Deep[[:space:]]Neural[[:space:]]Network[[:space:]]for[[:space:]]Blind[[:space:]]and[[:space:]]Universal[[:space:]]Lens[[:space:]]Aberration[[:space:]]Correction/aeee8578-b512-4a62-8ec3-b06e011ce338_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Picture[[:space:]]is[[:space:]]Worth[[:space:]]More[[:space:]]Than[[:space:]]77[[:space:]]Text[[:space:]]Tokens_[[:space:]]Evaluating[[:space:]]CLIP-Style[[:space:]]Models[[:space:]]on[[:space:]]Dense[[:space:]]Captions/f5f3c319-3887-4d17-8f43-fce0198c0c77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Recipe[[:space:]]for[[:space:]]Scaling[[:space:]]up[[:space:]]Text-to-Video[[:space:]]Generation[[:space:]]with[[:space:]]Text-free[[:space:]]Videos/873ecf7b-8814-4ac3-a70f-20982249ac1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Semi-supervised[[:space:]]Nighttime[[:space:]]Dehazing[[:space:]]Baseline[[:space:]]with[[:space:]]Spatial-Frequency[[:space:]]Aware[[:space:]]and[[:space:]]Realistic[[:space:]]Brightness[[:space:]]Constraint/4a280801-3209-4899-b345-f6dbc9c9ec52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Simple[[:space:]]Baseline[[:space:]]for[[:space:]]Efficient[[:space:]]Hand[[:space:]]Mesh[[:space:]]Reconstruction/edb8cbac-0e71-45cd-8cdb-45284f946ab7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Simple[[:space:]]Recipe[[:space:]]for[[:space:]]Contrastively[[:space:]]Pre-training[[:space:]]Video-First[[:space:]]Encoders[[:space:]]Beyond[[:space:]]16[[:space:]]Frames/e81a3abe-11ba-459a-b183-aa765dce41a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Simple[[:space:]]Recipe[[:space:]]for[[:space:]]Language-guided[[:space:]]Domain[[:space:]]Generalized[[:space:]]Segmentation/b9db7707-a86d-4d4c-b962-58bd8f08eecd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Simple[[:space:]]and[[:space:]]Effective[[:space:]]Point-based[[:space:]]Network[[:space:]]for[[:space:]]Event[[:space:]]Camera[[:space:]]6-DOFs[[:space:]]Pose[[:space:]]Relocalization/1bfdaa4b-618c-45a6-9de6-a1e062fefbcd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Stealthy[[:space:]]Wrongdoer_[[:space:]]Feature-Oriented[[:space:]]Reconstruction[[:space:]]Attack[[:space:]]against[[:space:]]Split[[:space:]]Learning/de2cb66c-4154-4b26-aa15-91e83e19d783_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Study[[:space:]]of[[:space:]]Dropout-Induced[[:space:]]Modality[[:space:]]Bias[[:space:]]on[[:space:]]Robustness[[:space:]]to[[:space:]]Missing[[:space:]]Video[[:space:]]Frames[[:space:]]for[[:space:]]Audio-Visual[[:space:]]Speech[[:space:]]Recognition/6897fec6-4bb7-4167-a28e-16a34134af6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Subspace-Constrained[[:space:]]Tyler's[[:space:]]Estimator[[:space:]]and[[:space:]]its[[:space:]]Applications[[:space:]]to[[:space:]]Structure[[:space:]]from[[:space:]]Motion/02a10508-95ff-4550-b14e-3121c8c91065_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Theory[[:space:]]of[[:space:]]Joint[[:space:]]Light[[:space:]]and[[:space:]]Heat[[:space:]]Transport[[:space:]]for[[:space:]]Lambertian[[:space:]]Scenes/c36c78d9-fdc0-45ae-af17-5820282f52ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Unified[[:space:]]Approach[[:space:]]for[[:space:]]Text-[[:space:]]and[[:space:]]Image-guided[[:space:]]4D[[:space:]]Scene[[:space:]]Generation/cf3caca8-62ec-4e3a-95a4-3e99a1fed94a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Unified[[:space:]]Diffusion[[:space:]]Framework[[:space:]]for[[:space:]]Scene-aware[[:space:]]Human[[:space:]]Motion[[:space:]]Estimation[[:space:]]from[[:space:]]Sparse[[:space:]]Signals/fd76c794-5741-41ca-abc9-25667a78bd7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Human-centric[[:space:]]Point[[:space:]]Cloud[[:space:]]Video[[:space:]]Understanding/d82e42af-26eb-4efa-b300-6b0b675aeba6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Microscopy[[:space:]]Defocus[[:space:]]Deblur[[:space:]]with[[:space:]]Multi-Pyramid[[:space:]]Transformer[[:space:]]and[[:space:]]Contrastive[[:space:]]Learning/fc1f4bb9-40e6-4cf0-af7b-880fdb6df934_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Unified[[:space:]]and[[:space:]]Interpretable[[:space:]]Emotion[[:space:]]Representation[[:space:]]and[[:space:]]Expression[[:space:]]Generation/71d78822-b622-4f9b-9743-2192e460a912_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Versatile[[:space:]]Framework[[:space:]]for[[:space:]]Continual[[:space:]]Test-Time[[:space:]]Domain[[:space:]]Adaptation_[[:space:]]Balancing[[:space:]]Discriminability[[:space:]]and[[:space:]]Generalizability/9de8f2cf-38ad-409c-9455-eb072224a83f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Video[[:space:]]is[[:space:]]Worth[[:space:]]256[[:space:]]Bases_[[:space:]]Spatial-Temporal[[:space:]]Expectation-Maximization[[:space:]]Inversion[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Video[[:space:]]Editing/acfeaf84-a215-4ec6-9f05-b5770d69d868_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A[[:space:]]Vision[[:space:]]Check-up[[:space:]]for[[:space:]]Language[[:space:]]Models/5c418f34-27eb-4e7b-83f9-aa566ee2d4cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A&B[[:space:]]BNN_[[:space:]]Add&Bit-Operation-Only[[:space:]]Hardware-Friendly[[:space:]]Binary[[:space:]]Neural[[:space:]]Network/2a3b310f-0433-461d-b9a2-9b529e344a29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A-Teacher_[[:space:]]Asymmetric[[:space:]]Network[[:space:]]for[[:space:]]3D[[:space:]]Semi-Supervised[[:space:]]Object[[:space:]]Detection/dfef3ace-a664-44c6-b8d3-66486174276f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/A2XP_[[:space:]]Towards[[:space:]]Private[[:space:]]Domain[[:space:]]Generalization/b2795817-60ba-4c2b-9576-800b707bc53a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AAMDM_[[:space:]]Accelerated[[:space:]]Auto-regressive[[:space:]]Motion[[:space:]]Diffusion[[:space:]]Model/6ce91fc3-a31e-4571-b529-526b5c85cd75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ACT-Diffusion_[[:space:]]Efficient[[:space:]]Adversarial[[:space:]]Consistency[[:space:]]Training[[:space:]]for[[:space:]]One-step[[:space:]]Diffusion[[:space:]]Models/ead19f55-a0b6-4ee4-92db-db601fba1e21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ADA-Track_[[:space:]]End-to-End[[:space:]]Multi-Camera[[:space:]]3D[[:space:]]Multi-Object[[:space:]]Tracking[[:space:]]with[[:space:]]Alternating[[:space:]]Detection[[:space:]]and[[:space:]]Association/708be312-7784-4948-b222-435f1ecb2969_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ADFactory_[[:space:]]An[[:space:]]Effective[[:space:]]Framework[[:space:]]for[[:space:]]Generalizing[[:space:]]Optical[[:space:]]Flow[[:space:]]with[[:space:]]NeRF/ac2cde0b-17a5-49cd-b192-6b3449cf43c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AEROBLADE_[[:space:]]Training-Free[[:space:]]Detection[[:space:]]of[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Images[[:space:]]Using[[:space:]]Autoencoder[[:space:]]Reconstruction[[:space:]]Error/fa52d00e-2bdb-4c68-9d21-d51d792b3e0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AETTA_[[:space:]]Label-Free[[:space:]]Accuracy[[:space:]]Estimation[[:space:]]for[[:space:]]Test-Time[[:space:]]Adaptation/1f4b2010-09b3-4d99-a08a-c480904f4094_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AHIVE_[[:space:]]Anatomy-aware[[:space:]]Hierarchical[[:space:]]Vision[[:space:]]Encoding[[:space:]]for[[:space:]]Interactive[[:space:]]Radiology[[:space:]]Report[[:space:]]Retrieval/18adc2ee-bf82-4eef-90c2-a84519926cee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AIDE_[[:space:]]An[[:space:]]Automatic[[:space:]]Data[[:space:]]Engine[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/7177958d-83c0-4be5-8a16-bcf9a8d8ba67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ALGM_[[:space:]]Adaptive[[:space:]]Local-then-Global[[:space:]]Token[[:space:]]Merging[[:space:]]for[[:space:]]Efficient[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Plain[[:space:]]Vision[[:space:]]Transformers/4e0ebc3b-233f-42e9-9c2d-2fe1f617fcf2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AM-RADIO_[[:space:]]Agglomerative[[:space:]]Vision[[:space:]]Foundation[[:space:]]Model[[:space:]]Reduce[[:space:]]All[[:space:]]Domains[[:space:]]Into[[:space:]]One/34c69b0a-7815-45b6-a286-9e16ea17fec3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AMU-Tuning_[[:space:]]Effective[[:space:]]Logit[[:space:]]Bias[[:space:]]for[[:space:]]CLIP-based[[:space:]]Few-shot[[:space:]]Learning/81add02f-293a-4533-9ed2-aaa46d9b21df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ANIM_[[:space:]]Accurate[[:space:]]Neural[[:space:]]Implicit[[:space:]]Model[[:space:]]for[[:space:]]Human[[:space:]]Reconstruction[[:space:]]from[[:space:]]a[[:space:]]single[[:space:]]RGB-D[[:space:]]Image/bfa7080b-2c69-4512-98cd-99bed9cddc32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/APISR_[[:space:]]Anime[[:space:]]Production[[:space:]]Inspired[[:space:]]Real-World[[:space:]]Anime[[:space:]]Super-Resolution/338cfb38-7fae-4b7c-9e30-04043a7f3f04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/APSeg_[[:space:]]Auto-Prompt[[:space:]]Network[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Few-Shot[[:space:]]Semantic[[:space:]]Segmentation/3d57ef06-65e6-4696-b1bf-ec7e58abd4df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ARTrackV2_[[:space:]]Prompting[[:space:]]Autoregressive[[:space:]]Tracker[[:space:]]Where[[:space:]]to[[:space:]]Look[[:space:]]and[[:space:]]How[[:space:]]to[[:space:]]Describe/e3bf9a13-15a5-4137-bcfa-fd777ce1cc8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ASAM_[[:space:]]Boosting[[:space:]]Segment[[:space:]]Anything[[:space:]]Model[[:space:]]with[[:space:]]Adversarial[[:space:]]Tuning/36197f30-615e-421d-90e2-4cfca63aa45d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ASH_[[:space:]]Animatable[[:space:]]Gaussian[[:space:]]Splats[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Photoreal[[:space:]]Human[[:space:]]Rendering/c850f432-239a-4971-9e65-66978117a509_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AUEditNet_[[:space:]]Dual-Branch[[:space:]]Facial[[:space:]]Action[[:space:]]Unit[[:space:]]Intensity[[:space:]]Manipulation[[:space:]]with[[:space:]]Implicit[[:space:]]Disentanglement/d7d9e32a-6643-4cdd-a9aa-a5c3c2e6c216_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AV-RIR_[[:space:]]Audio-Visual[[:space:]]Room[[:space:]]Impulse[[:space:]]Response[[:space:]]Estimation/aebb747f-b0e2-42cb-8c4c-5f3e35e70aa5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AV2AV_[[:space:]]Direct[[:space:]]Audio-Visual[[:space:]]Speech[[:space:]]to[[:space:]]Audio-Visual[[:space:]]Speech[[:space:]]Translation[[:space:]]with[[:space:]]Unified[[:space:]]Audio-Visual[[:space:]]Speech[[:space:]]Representation/93042e6e-c40a-42a8-92de-824606de4fc5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AVFF_[[:space:]]Audio-Visual[[:space:]]Feature[[:space:]]Fusion[[:space:]]for[[:space:]]Video[[:space:]]Deepfake[[:space:]]Detection/bd9c3a94-112c-4ed8-af56-c89e41616b62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AVID_[[:space:]]Any-Length[[:space:]]Video[[:space:]]Inpainting[[:space:]]with[[:space:]]Diffusion[[:space:]]Model/01215d18-082c-4bb1-b4fb-5bd3ac631fb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AZ-NAS_[[:space:]]Assembling[[:space:]]Zero-Cost[[:space:]]Proxies[[:space:]]for[[:space:]]Network[[:space:]]Architecture[[:space:]]Search/5fb164d8-6043-4215-b007-312a2ca972dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Abductive[[:space:]]Ego-View[[:space:]]Accident[[:space:]]Video[[:space:]]Understanding[[:space:]]for[[:space:]]Safe[[:space:]]Driving[[:space:]]Perception/3f0d502b-93b9-4eae-ae39-19540f4dd5f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Absolute[[:space:]]Pose[[:space:]]from[[:space:]]One[[:space:]]or[[:space:]]Two[[:space:]]Scaled[[:space:]]and[[:space:]]Oriented[[:space:]]Features/36a2b532-51a6-4f6b-b2da-9d323017fefe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Accelerating[[:space:]]Diffusion[[:space:]]Sampling[[:space:]]with[[:space:]]Optimized[[:space:]]Time[[:space:]]Steps/179101b6-f67f-430d-8f1e-d09c0b3c0618_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Accelerating[[:space:]]Neural[[:space:]]Field[[:space:]]Training[[:space:]]via[[:space:]]Soft[[:space:]]Mining/e4435969-1550-465a-b628-4796a397d012_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Accept[[:space:]]the[[:space:]]Modality[[:space:]]Gap_[[:space:]]An[[:space:]]Exploration[[:space:]]in[[:space:]]the[[:space:]]Hyperbolic[[:space:]]Space/ee933822-9cc9-424f-ab35-040ed4e81569_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Accurate[[:space:]]Spatial[[:space:]]Gene[[:space:]]Expression[[:space:]]Prediction[[:space:]]by[[:space:]]Integrating[[:space:]]Multi-Resolution[[:space:]]Features/5bd07931-16f9-4ab8-9e74-8b514577d81c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Accurate[[:space:]]Training[[:space:]]Data[[:space:]]for[[:space:]]Occupancy[[:space:]]Map[[:space:]]Prediction[[:space:]]in[[:space:]]Automated[[:space:]]Driving[[:space:]]Using[[:space:]]Evidence[[:space:]]Theory/4cf8e312-3cbd-417d-b6a6-b6c16142d795_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Action[[:space:]]Detection[[:space:]]via[[:space:]]an[[:space:]]Image[[:space:]]Diffusion[[:space:]]Process/f5378932-0225-4869-95c0-87be841834c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Action[[:space:]]Scene[[:space:]]Graphs[[:space:]]for[[:space:]]Long-Form[[:space:]]Understanding[[:space:]]of[[:space:]]Egocentric[[:space:]]Videos/79af1fe8-797d-4357-82bf-e8b31cc80335_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Action-slot_[[:space:]]Visual[[:space:]]Action-centric[[:space:]]Representations[[:space:]]for[[:space:]]Multi-label[[:space:]]Atomic[[:space:]]Activity[[:space:]]Recognition[[:space:]]in[[:space:]]Traffic[[:space:]]Scenes/4236109b-ec04-4f7c-8bb8-2d430fef699d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Active[[:space:]]Domain[[:space:]]Adaptation[[:space:]]with[[:space:]]False[[:space:]]Negative[[:space:]]Prediction[[:space:]]for[[:space:]]Object[[:space:]]Detection/58ea045e-fd64-449c-ae09-c731b85ca038_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Active[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/91754107-467c-4976-b0ed-7103b0d69794_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Active[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Knowledge[[:space:]]Aggregation[[:space:]]and[[:space:]]Distillation[[:space:]]from[[:space:]]Large[[:space:]]Models/94548157-11d0-4a27-a091-5a619bf905e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Active[[:space:]]Open-Vocabulary[[:space:]]Recognition_[[:space:]]Let[[:space:]]Intelligent[[:space:]]Moving[[:space:]]Mitigate[[:space:]]CLIP[[:space:]]Limitations/071dcffd-08a7-47c1-840a-4cd878cce7b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Active[[:space:]]Prompt[[:space:]]Learning[[:space:]]in[[:space:]]Vision[[:space:]]Language[[:space:]]Models/263a2d2c-f3d6-42c8-a71b-86436f8722c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ActiveDC_[[:space:]]Distribution[[:space:]]Calibration[[:space:]]for[[:space:]]Active[[:space:]]Finetuning/44ff82e9-3c5d-413b-a7b7-cfe81422c55c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Activity-Biometrics_[[:space:]]Person[[:space:]]Identification[[:space:]]from[[:space:]]Daily[[:space:]]Activities/03cdf98b-1b76-4751-b2b0-17da087ce104_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AdaBM_[[:space:]]On-the-Fly[[:space:]]Adaptive[[:space:]]Bit[[:space:]]Mapping[[:space:]]for[[:space:]]Image[[:space:]]Super-Resolution/9521aabb-8a2d-4a09-8cda-cf0dc9be1225_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AdaRevD_[[:space:]]Adaptive[[:space:]]Patch[[:space:]]Exiting[[:space:]]Reversible[[:space:]]Decoder[[:space:]]Pushes[[:space:]]the[[:space:]]Limit[[:space:]]of[[:space:]]Image[[:space:]]Deblurring/c1019b46-f612-4347-b4a2-3b83e874ca97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AdaShift_[[:space:]]Learning[[:space:]]Discriminative[[:space:]]Self-Gated[[:space:]]Neural[[:space:]]Feature[[:space:]]Activation[[:space:]]With[[:space:]]an[[:space:]]Adaptive[[:space:]]Shift[[:space:]]Factor/c2adbf30-c2b3-413a-bc73-5d255dc82737_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adapt[[:space:]]Before[[:space:]]Comparison_[[:space:]]A[[:space:]]New[[:space:]]Perspective[[:space:]]on[[:space:]]Cross-Domain[[:space:]]Few-Shot[[:space:]]Segmentation/66e7543b-64e8-420d-ba3c-448b4ad53fb9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adapt[[:space:]]or[[:space:]]Perish_[[:space:]]Adaptive[[:space:]]Sparse[[:space:]]Transformer[[:space:]]with[[:space:]]Attentive[[:space:]]Feature[[:space:]]Refinement[[:space:]]for[[:space:]]Image[[:space:]]Restoration/d8d8aa60-6e12-41e2-a447-0b3aab1a3bdb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adapters[[:space:]]Strike[[:space:]]Back/7a648b9f-e6a6-4f34-b960-5790f28a62ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adapting[[:space:]]Short-Term[[:space:]]Transformers[[:space:]]for[[:space:]]Action[[:space:]]Detection[[:space:]]in[[:space:]]Untrimmed[[:space:]]Videos/1853ca32-2923-41ce-93cd-53d8ed6b7c2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adapting[[:space:]]Visual-Language[[:space:]]Models[[:space:]]for[[:space:]]Generalizable[[:space:]]Anomaly[[:space:]]Detection[[:space:]]in[[:space:]]Medical[[:space:]]Images/b03cf887-e43a-44dc-81c2-2e2163fe5964_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adapting[[:space:]]to[[:space:]]Length[[:space:]]Shift_[[:space:]]FlexiLength[[:space:]]Network[[:space:]]for[[:space:]]Trajectory[[:space:]]Prediction/e65b1c61-b4b2-403b-9f5b-4e1277d70ce6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adaptive[[:space:]]Bidirectional[[:space:]]Displacement[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/7d709263-8879-466c-b35a-21cff2b32e69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adaptive[[:space:]]Fusion[[:space:]]of[[:space:]]Single-View[[:space:]]and[[:space:]]Multi-View[[:space:]]Depth[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/3f1cab6f-6d22-4752-b4b0-acfc7d74494f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adaptive[[:space:]]Hyper-graph[[:space:]]Aggregation[[:space:]]for[[:space:]]Modality-Agnostic[[:space:]]Federated[[:space:]]Learning/dff48f54-c257-45c8-ad10-d1b037ab2d20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adaptive[[:space:]]Multi-Modal[[:space:]]Cross-Entropy[[:space:]]Loss[[:space:]]for[[:space:]]Stereo[[:space:]]Matching/277222c8-7877-4ce3-a9e7-aaf91839c40d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adaptive[[:space:]]Random[[:space:]]Feature[[:space:]]Regularization[[:space:]]on[[:space:]]Fine-tuning[[:space:]]Deep[[:space:]]Neural[[:space:]]Networks/b91a9c1d-3cbe-4860-bf2e-b06709be4a8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adaptive[[:space:]]Slot[[:space:]]Attention_[[:space:]]Object[[:space:]]Discovery[[:space:]]with[[:space:]]Dynamic[[:space:]]Slot[[:space:]]Number/0b3cb444-8f36-45ee-af39-022345b2755f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adaptive[[:space:]]Softassign[[:space:]]via[[:space:]]Hadamard-Equipped[[:space:]]Sinkhorn/e64dfa59-d678-46ad-9fa9-d4b2e90f34db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adaptive[[:space:]]VIO_[[:space:]]Deep[[:space:]]Visual-Inertial[[:space:]]Odometry[[:space:]]with[[:space:]]Online[[:space:]]Continual[[:space:]]Learning/9fc1fda4-5e98-4f33-9aaf-60fff7f89340_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Addressing[[:space:]]Background[[:space:]]Context[[:space:]]Bias[[:space:]]in[[:space:]]Few-Shot[[:space:]]Segmentation[[:space:]]through[[:space:]]Iterative[[:space:]]Modulation/5ff9bc5d-47fd-4926-bc03-51300425c6f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Advancing[[:space:]]Saliency[[:space:]]Ranking[[:space:]]with[[:space:]]Human[[:space:]]Fixations_[[:space:]]Dataset[[:space:]]Models[[:space:]]and[[:space:]]Benchmarks/def71ca4-38fb-42b7-9683-34aeb7347ccb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adversarial[[:space:]]Backdoor[[:space:]]Attack[[:space:]]by[[:space:]]Naturalistic[[:space:]]Data[[:space:]]Poisoning[[:space:]]on[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/1dce66f6-c37f-40e2-a71f-323b7f5bc857_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adversarial[[:space:]]Distillation[[:space:]]Based[[:space:]]on[[:space:]]Slack[[:space:]]Matching[[:space:]]and[[:space:]]Attribution[[:space:]]Region[[:space:]]Alignment/d8258235-cebe-4d20-8e40-f16c3b7cd3aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adversarial[[:space:]]Score[[:space:]]Distillation_[[:space:]]When[[:space:]]score[[:space:]]distillation[[:space:]]meets[[:space:]]GAN/44f1cef7-b6cf-4f93-abf6-a8a951fd353d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adversarial[[:space:]]Text[[:space:]]to[[:space:]]Continuous[[:space:]]Image[[:space:]]Generation/3b845f83-13a1-4a9f-9dc3-f0b795e9e960_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Adversarially[[:space:]]Robust[[:space:]]Few-shot[[:space:]]Learning[[:space:]]via[[:space:]]Parameter[[:space:]]Co-distillation[[:space:]]of[[:space:]]Similarity[[:space:]]and[[:space:]]Class[[:space:]]Concept[[:space:]]Learners/7be5cc09-0482-4610-8751-58cbe467cad7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Aerial[[:space:]]Lifting_[[:space:]]Neural[[:space:]]Urban[[:space:]]Semantic[[:space:]]and[[:space:]]Building[[:space:]]Instance[[:space:]]Lifting[[:space:]]from[[:space:]]Aerial[[:space:]]Imagery/6e6a4db9-c6c0-4c40-acc5-1c0a12aa84e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Affine[[:space:]]Equivariant[[:space:]]Networks[[:space:]]Based[[:space:]]on[[:space:]]Differential[[:space:]]Invariants/9fed17fb-1938-4636-9e01-3afae7065759_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AiOS_[[:space:]]All-in-One-Stage[[:space:]]Expressive[[:space:]]Human[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Estimation/515a19f4-402c-4fd4-8da1-00256976c97f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AirPlanes_[[:space:]]Accurate[[:space:]]Plane[[:space:]]Estimation[[:space:]]via[[:space:]]3D-Consistent[[:space:]]Embeddings/ac37ebb0-f18f-4022-af4e-1925aaec6f50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Alchemist_[[:space:]]Parametric[[:space:]]Control[[:space:]]of[[:space:]]Material[[:space:]]Properties[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/41cd119b-5d81-4c42-b290-2bcd7160d859_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Align[[:space:]]Before[[:space:]]Adapt_[[:space:]]Leveraging[[:space:]]Entity-to-Region[[:space:]]Alignments[[:space:]]for[[:space:]]Generalizable[[:space:]]Video[[:space:]]Action[[:space:]]Recognition/3c259d58-bb16-4c6a-bcb7-0420b28381ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Align[[:space:]]Your[[:space:]]Gaussians_[[:space:]]Text-to-4D[[:space:]]with[[:space:]]Dynamic[[:space:]]3D[[:space:]]Gaussians[[:space:]]and[[:space:]]Composed[[:space:]]Diffusion[[:space:]]Models/cd0fb40f-f126-463f-8026-3d79b1819435_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Align[[:space:]]and[[:space:]]Aggregate_[[:space:]]Compositional[[:space:]]Reasoning[[:space:]]with[[:space:]]Video[[:space:]]Alignment[[:space:]]and[[:space:]]Answer[[:space:]]Aggregation[[:space:]]for[[:space:]]Video[[:space:]]Question-Answering/b11e9f42-4e77-4647-9a93-e15b35d35256_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AlignMiF_[[:space:]]Geometry-Aligned[[:space:]]Multimodal[[:space:]]Implicit[[:space:]]Field[[:space:]]for[[:space:]]LiDAR-Camera[[:space:]]Joint[[:space:]]Synthesis/c16ae51e-7d16-4105-abad-d73327069de7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AlignSAM_[[:space:]]Aligning[[:space:]]Segment[[:space:]]Anything[[:space:]]Model[[:space:]]to[[:space:]]Open[[:space:]]Context[[:space:]]via[[:space:]]Reinforcement[[:space:]]Learning/20ca7782-fc65-46b1-8bc1-e41fba5c191e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Aligning[[:space:]]Logits[[:space:]]Generatively[[:space:]]for[[:space:]]Principled[[:space:]]Black-Box[[:space:]]Knowledge[[:space:]]Distillation/f3346329-f111-4bdb-a5a8-7fd18084c7a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Aligning[[:space:]]and[[:space:]]Prompting[[:space:]]Everything[[:space:]]All[[:space:]]at[[:space:]]Once[[:space:]]for[[:space:]]Universal[[:space:]]Visual[[:space:]]Perception/a8e8f864-6df7-48db-a604-d2c449b57cd9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/All[[:space:]]Rivers[[:space:]]Run[[:space:]]to[[:space:]]the[[:space:]]Sea_[[:space:]]Private[[:space:]]Learning[[:space:]]with[[:space:]]Asymmetric[[:space:]]Flows/1653546f-7d0a-4233-a716-a1945d78d0de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/All[[:space:]]in[[:space:]]One[[:space:]]Framework[[:space:]]for[[:space:]]Multimodal[[:space:]]Re-identification[[:space:]]in[[:space:]]the[[:space:]]Wild/b566a6aa-5df2-4227-96f3-33406fe83f14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AllSpark_[[:space:]]Reborn[[:space:]]Labeled[[:space:]]Features[[:space:]]from[[:space:]]Unlabeled[[:space:]]in[[:space:]]Transformer[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation/04b433f7-a189-48ad-af1b-1e042d322f8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Alpha[[:space:]]Invariance_[[:space:]]On[[:space:]]Inverse[[:space:]]Scaling[[:space:]]Between[[:space:]]Distance[[:space:]]and[[:space:]]Volume[[:space:]]Density[[:space:]]in[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/f2632435-497e-407d-85fd-8a3463e49035_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Alpha-CLIP_[[:space:]]A[[:space:]]CLIP[[:space:]]Model[[:space:]]Focusing[[:space:]]on[[:space:]]Wherever[[:space:]]You[[:space:]]Want/7c50f7c2-98b5-4c0a-8769-9f339d346eb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Amodal[[:space:]]Completion[[:space:]]via[[:space:]]Progressive[[:space:]]Mixed[[:space:]]Context[[:space:]]Diffusion/a4d1f1b9-01c6-4390-9ee6-30ec50f6f9c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Amodal[[:space:]]Ground[[:space:]]Truth[[:space:]]and[[:space:]]Completion[[:space:]]in[[:space:]]the[[:space:]]Wild/a5eee9b3-e428-4eb9-8d90-3f93409c29da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/An[[:space:]]Aggregation-Free[[:space:]]Federated[[:space:]]Learning[[:space:]]for[[:space:]]Tackling[[:space:]]Data[[:space:]]Heterogeneity/268798e2-f54d-4170-a3e4-a4d750365147_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/An[[:space:]]Asymmetric[[:space:]]Augmented[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]Method[[:space:]]for[[:space:]]Unsupervised[[:space:]]Fine-Grained[[:space:]]Image[[:space:]]Hashing/61c48b1c-a7a2-41f2-ac2a-071bd45619a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/An[[:space:]]Edit[[:space:]]Friendly[[:space:]]DDPM[[:space:]]Noise[[:space:]]Space_[[:space:]]Inversion[[:space:]]and[[:space:]]Manipulations/10445918-ea5a-42fd-b870-2cd4e21d83e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/An[[:space:]]Empirical[[:space:]]Study[[:space:]]of[[:space:]]Scaling[[:space:]]Law[[:space:]]for[[:space:]]Scene[[:space:]]Text[[:space:]]Recognition/aed0e1fe-9168-4968-9bf1-52042e93cd52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/An[[:space:]]Empirical[[:space:]]Study[[:space:]]of[[:space:]]the[[:space:]]Generalization[[:space:]]Ability[[:space:]]of[[:space:]]Lidar[[:space:]]3D[[:space:]]Object[[:space:]]Detectors[[:space:]]to[[:space:]]Unseen[[:space:]]Domains/cc042981-247d-4de1-bcc2-9f6657bcaf6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/An[[:space:]]Interactive[[:space:]]Navigation[[:space:]]Method[[:space:]]with[[:space:]]Effect-oriented[[:space:]]Affordance/a01574d3-028b-4a77-b448-a4286fc778bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/An[[:space:]]N-Point[[:space:]]Linear[[:space:]]Solver[[:space:]]for[[:space:]]Line[[:space:]]and[[:space:]]Motion[[:space:]]Estimation[[:space:]]with[[:space:]]Event[[:space:]]Cameras/c3709262-53cf-4026-80f5-d2f2ba343e58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/An[[:space:]]Upload-Efficient[[:space:]]Scheme[[:space:]]for[[:space:]]Transferring[[:space:]]Knowledge[[:space:]]From[[:space:]]a[[:space:]]Server-Side[[:space:]]Pre-trained[[:space:]]Generator[[:space:]]to[[:space:]]Clients[[:space:]]in[[:space:]]Heterogeneous[[:space:]]Federated[[:space:]]Learning/f3dd7eac-2ba8-4ee5-afa3-9baa9d9ff112_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Analyzing[[:space:]]and[[:space:]]Improving[[:space:]]the[[:space:]]Training[[:space:]]Dynamics[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/f079a68e-db72-484c-ac7a-d8d2c4ced0f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Anatomically[[:space:]]Constrained[[:space:]]Implicit[[:space:]]Face[[:space:]]Models/efb3eaea-3e2e-4f6a-914c-24c28578ed0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Anchor-based[[:space:]]Robust[[:space:]]Finetuning[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models/1390530a-39c4-4324-adf5-415f1a67b69f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Animatable[[:space:]]Gaussians_[[:space:]]Learning[[:space:]]Pose-dependent[[:space:]]Gaussian[[:space:]]Maps[[:space:]]for[[:space:]]High-fidelity[[:space:]]Human[[:space:]]Avatar[[:space:]]Modeling/a536ec2f-48cc-4908-b864-0e49bf59959d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Animate[[:space:]]Anyone_[[:space:]]Consistent[[:space:]]and[[:space:]]Controllable[[:space:]]Image-to-Video[[:space:]]Synthesis[[:space:]]for[[:space:]]Character[[:space:]]Animation/7af3f4f9-eea5-48fd-af87-57f7424d228f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Animating[[:space:]]General[[:space:]]Image[[:space:]]with[[:space:]]Large[[:space:]]Visual[[:space:]]Motion[[:space:]]Model/dec1332b-680c-4517-9c8e-f8d40f749836_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Anomaly[[:space:]]Heterogeneity[[:space:]]Learning[[:space:]]for[[:space:]]Open-set[[:space:]]Supervised[[:space:]]Anomaly[[:space:]]Detection/605486ed-fd9d-424f-b782-d9b752432630_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Anomaly[[:space:]]Score_[[:space:]]Evaluating[[:space:]]Generative[[:space:]]Models[[:space:]]and[[:space:]]Individual[[:space:]]Generated[[:space:]]Images[[:space:]]based[[:space:]]on[[:space:]]Complexity[[:space:]]and[[:space:]]Vulnerability/9d674bb1-be04-44ce-8659-6f017f96a9d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Any-Shift[[:space:]]Prompting[[:space:]]for[[:space:]]Generalization[[:space:]]over[[:space:]]Distributions/cb211c45-426e-4f9b-8779-fce28d0f8544_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AnyDoor_[[:space:]]Zero-shot[[:space:]]Object-level[[:space:]]Image[[:space:]]Customization/39434d5d-7288-447f-ae12-51d506f96375_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AnyScene_[[:space:]]Customized[[:space:]]Image[[:space:]]Synthesis[[:space:]]with[[:space:]]Composited[[:space:]]Foreground/d0edfba9-513e-41e1-b983-ddfbe29cbeae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AnySkill_[[:space:]]Learning[[:space:]]Open-Vocabulary[[:space:]]Physical[[:space:]]Skill[[:space:]]for[[:space:]]Interactive[[:space:]]Agents/a7069391-34d0-4857-b2b1-9a96e1054941_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ArGue_[[:space:]]Attribute-Guided[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/d1e69041-3267-49cb-aecd-3ded66296260_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Arbitrary[[:space:]]Motion[[:space:]]Style[[:space:]]Transfer[[:space:]]with[[:space:]]Multi-condition[[:space:]]Motion[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Model/2204b1d6-527c-4122-ae14-8c947b8cb878_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Arbitrary-Scale[[:space:]]Image[[:space:]]Generation[[:space:]]and[[:space:]]Upsampling[[:space:]]using[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Model[[:space:]]and[[:space:]]Implicit[[:space:]]Neural[[:space:]]Decoder/fe269b4c-61f7-401d-9c37-9201888b2314_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Are[[:space:]]Conventional[[:space:]]SNNs[[:space:]]Really[[:space:]]Efficient_[[:space:]]A[[:space:]]Perspective[[:space:]]from[[:space:]]Network[[:space:]]Quantization/85577dd5-78ab-49aa-8713-ebd59e6d37ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ArtAdapter_[[:space:]]Text-to-Image[[:space:]]Style[[:space:]]Transfer[[:space:]]using[[:space:]]Multi-Level[[:space:]]Style[[:space:]]Encoder[[:space:]]and[[:space:]]Explicit[[:space:]]Adaptation/d1263937-1748-4a3b-bcbd-08ebe32e6322_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Artist-Friendly[[:space:]]Relightable[[:space:]]and[[:space:]]Animatable[[:space:]]Neural[[:space:]]Heads/426aec2b-740b-4d65-afe7-f7f84cf59071_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/As-Plausible-As-Possible_[[:space:]]Plausibility-Aware[[:space:]]Mesh[[:space:]]Deformation[[:space:]]Using[[:space:]]2D[[:space:]]Diffusion[[:space:]]Priors/4e094905-9f6c-467e-bf33-19d0a0756b2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AssistGUI_[[:space:]]Task-Oriented[[:space:]]PC[[:space:]]Graphical[[:space:]]User[[:space:]]Interface[[:space:]]Automation/1bc34cbb-6088-46a3-bbf3-2b80efa7be58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Asymmetric[[:space:]]Masked[[:space:]]Distillation[[:space:]]for[[:space:]]Pre-Training[[:space:]]Small[[:space:]]Foundation[[:space:]]Models/174f35d1-536c-4e88-a4c8-f050bd4e73ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Atlantis_[[:space:]]Enabling[[:space:]]Underwater[[:space:]]Depth[[:space:]]Estimation[[:space:]]with[[:space:]]Stable[[:space:]]Diffusion/6ee5a57a-1151-4c47-a668-9cf2ff8822e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Atom-Level[[:space:]]Optical[[:space:]]Chemical[[:space:]]Structure[[:space:]]Recognition[[:space:]]with[[:space:]]Limited[[:space:]]Supervision/2300cae5-558b-4533-86bf-7857793b35ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Attack[[:space:]]To[[:space:]]Defend_[[:space:]]Exploiting[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]for[[:space:]]Detecting[[:space:]]Poisoned[[:space:]]Models/0044eddf-31b8-4b07-a08a-593c68789ae9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Attention[[:space:]]Calibration[[:space:]]for[[:space:]]Disentangled[[:space:]]Text-to-Image[[:space:]]Personalization/3adf8495-abf8-46c4-ba52-cebc73d9910e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Attention-Driven[[:space:]]Training-Free[[:space:]]Efficiency[[:space:]]Enhancement[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/09b68574-1296-44ba-b1e1-347e732dc703_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Attention-Propagation[[:space:]]Network[[:space:]]for[[:space:]]Egocentric[[:space:]]Heatmap[[:space:]]to[[:space:]]3D[[:space:]]Pose[[:space:]]Lifting/3fc9607e-cf5f-4163-9079-5285814ddfd3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Attentive[[:space:]]Illumination[[:space:]]Decomposition[[:space:]]Model[[:space:]]for[[:space:]]Multi-Illuminant[[:space:]]White[[:space:]]Balancing/71839592-8a76-4b86-a4be-62ddd718b067_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AttriHuman-3D_[[:space:]]Editable[[:space:]]3D[[:space:]]Human[[:space:]]Avatar[[:space:]]Generation[[:space:]]with[[:space:]]Attribute[[:space:]]Decomposition[[:space:]]and[[:space:]]Indexing/246ab2b7-ff0c-4150-836a-1c85b1a49a28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Attribute-Guided[[:space:]]Pedestrian[[:space:]]Retrieval_[[:space:]]Bridging[[:space:]]Person[[:space:]]Re-ID[[:space:]]with[[:space:]]Internal[[:space:]]Attribute[[:space:]]Variability/4b1b2abb-9625-4b5b-8c89-a8396c234ccb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Audio-Visual[[:space:]]Segmentation[[:space:]]via[[:space:]]Unlabeled[[:space:]]Frame[[:space:]]Exploitation/1e5486df-b5aa-4290-863b-3453e21cc0f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Authentic[[:space:]]Hand[[:space:]]Avatar[[:space:]]from[[:space:]]a[[:space:]]Phone[[:space:]]Scan[[:space:]]via[[:space:]]Universal[[:space:]]Hand[[:space:]]Model/b67e1412-9624-4b79-9f24-0cf518e4262f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Auto[[:space:]]MC-Reward_[[:space:]]Automated[[:space:]]Dense[[:space:]]Reward[[:space:]]Design[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Minecraft/12cb1254-8f36-4bbc-8924-54e4a6cd5b59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Auto-Train-Once_[[:space:]]Controller[[:space:]]Network[[:space:]]Guided[[:space:]]Automatic[[:space:]]Network[[:space:]]Pruning[[:space:]]from[[:space:]]Scratch/9d567cdf-19ed-44c6-8da1-04cbfb615dd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AutoAD[[:space:]]III_[[:space:]]The[[:space:]]Prequel[[:space:]]-[[:space:]]Back[[:space:]]to[[:space:]]the[[:space:]]Pixels/eba48e55-d15b-49ad-8013-f5dd76886f81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Automatic[[:space:]]Controllable[[:space:]]Colorization[[:space:]]via[[:space:]]Imagination/192cc558-e70e-466d-930c-444fb2d4db32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Autoregressive[[:space:]]Queries[[:space:]]for[[:space:]]Adaptive[[:space:]]Tracking[[:space:]]with[[:space:]]Spatio-Temporal[[:space:]]Transformers/05288a05-25ea-4c2c-a667-1563dbb7a0fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/AvatarGPT_[[:space:]]All-in-One[[:space:]]Framework[[:space:]]for[[:space:]]Motion[[:space:]]Understanding[[:space:]]Planning[[:space:]]Generation[[:space:]]and[[:space:]]Beyond/11c49f76-0621-42ef-ab06-7002d815c153_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BA-SAM_[[:space:]]Scalable[[:space:]]Bias-Mode[[:space:]]Attention[[:space:]]Mask[[:space:]]for[[:space:]]Segment[[:space:]]Anything[[:space:]]Model/8efa50ea-7a84-4fa0-89a7-2381828ef213_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BANF_[[:space:]]Band-Limited[[:space:]]Neural[[:space:]]Fields[[:space:]]for[[:space:]]Levels[[:space:]]of[[:space:]]Detail[[:space:]]Reconstruction/62fd3eb5-a5a9-4412-8ea4-e242e3bd48a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BEHAVIOR[[:space:]]Vision[[:space:]]Suite_[[:space:]]Customizable[[:space:]]Dataset[[:space:]]Generation[[:space:]]via[[:space:]]Simulation/0ceda5c5-4cd8-4631-9188-51e4bce35a17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BEM_[[:space:]]Balanced[[:space:]]and[[:space:]]Entropy-based[[:space:]]Mix[[:space:]]for[[:space:]]Long-Tailed[[:space:]]Semi-Supervised[[:space:]]Learning/94dd21fa-7db5-4e3e-8753-38486e83bf17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BEVNeXt_[[:space:]]Reviving[[:space:]]Dense[[:space:]]BEV[[:space:]]Frameworks[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/61d4e713-4252-453d-9433-b90e2f33b138_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BEVSpread_[[:space:]]Spread[[:space:]]Voxel[[:space:]]Pooling[[:space:]]for[[:space:]]Bird's-Eye-View[[:space:]]Representation[[:space:]]in[[:space:]]Vision-based[[:space:]]Roadside[[:space:]]3D[[:space:]]Object[[:space:]]Detection/b7207def-fab4-4d0f-8d23-4a7429f8ae4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BIVDiff_[[:space:]]A[[:space:]]Training-Free[[:space:]]Framework[[:space:]]for[[:space:]]General-Purpose[[:space:]]Video[[:space:]]Synthesis[[:space:]]via[[:space:]]Bridging[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/8315e157-6287-4b93-b614-a0fff49316d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BOTH2Hands_[[:space:]]Inferring[[:space:]]3D[[:space:]]Hands[[:space:]]from[[:space:]]Both[[:space:]]Text[[:space:]]Prompts[[:space:]]and[[:space:]]Body[[:space:]]Dynamics/5bab2d73-8a85-4134-9b80-dd30fc5e514d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BSNet_[[:space:]]Box-Supervised[[:space:]]Simulation-assisted[[:space:]]Mean[[:space:]]Teacher[[:space:]]for[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation/21d1e057-b626-4eed-8f72-c8449e680335_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BT-Adapter_[[:space:]]Video[[:space:]]Conversation[[:space:]]is[[:space:]]Feasible[[:space:]]Without[[:space:]]Video[[:space:]]Instruction[[:space:]]Tuning/e7d23b18-f6c2-4306-96c3-038bc12b390d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Back[[:space:]]to[[:space:]]3D_[[:space:]]Few-Shot[[:space:]]3D[[:space:]]Keypoint[[:space:]]Detection[[:space:]]with[[:space:]]Back-Projected[[:space:]]2D[[:space:]]Features/3aeb6d6c-b15f-4066-ae11-b76ef3816e6c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Backdoor[[:space:]]Defense[[:space:]]via[[:space:]]Test-Time[[:space:]]Detecting[[:space:]]and[[:space:]]Repairing/f037146e-7097-445e-aaf1-4bc481b6dcb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Backpropagation-free[[:space:]]Network[[:space:]]for[[:space:]]3D[[:space:]]Test-time[[:space:]]Adaptation/22d87301-a399-4ae2-97ec-74778c8ab6f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BadCLIP_[[:space:]]Dual-Embedding[[:space:]]Guided[[:space:]]Backdoor[[:space:]]Attack[[:space:]]on[[:space:]]Multimodal[[:space:]]Contrastive[[:space:]]Learning/7f5d57ff-d2fb-484f-bc5d-914da7132779_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BadCLIP_[[:space:]]Trigger-Aware[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Backdoor[[:space:]]Attacks[[:space:]]on[[:space:]]CLIP/6ae79c99-4739-4f18-ac9c-46987ebf9d69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Balancing[[:space:]]Act_[[:space:]]Distribution-Guided[[:space:]]Debiasing[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/d5569d2e-eacc-47fc-bf8c-9cd4d233ea27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Batch[[:space:]]Normalization[[:space:]]Alleviates[[:space:]]the[[:space:]]Spectral[[:space:]]Bias[[:space:]]in[[:space:]]Coordinate[[:space:]]Networks/62fb5ef8-4da4-447b-9c58-4d2b6b8a5310_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bayes'[[:space:]]Rays_[[:space:]]Uncertainty[[:space:]]Quantification[[:space:]]for[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/4bd792c2-be62-4e8d-a89d-bc8f8ae8f426_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bayesian[[:space:]]Differentiable[[:space:]]Physics[[:space:]]for[[:space:]]Cloth[[:space:]]Digitalization/813af3e2-7d50-44f6-b865-37d2db7290ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bayesian[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]3D[[:space:]]Shape[[:space:]]Reconstruction/76158074-f74a-4868-b6b7-6c962eb8f00b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bayesian[[:space:]]Exploration[[:space:]]of[[:space:]]Pre-trained[[:space:]]Models[[:space:]]for[[:space:]]Low-shot[[:space:]]Image[[:space:]]Classification/bd225d4e-a76a-4f9e-941a-80966f47431d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Behind[[:space:]]the[[:space:]]Veil_[[:space:]]Enhanced[[:space:]]Indoor[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]with[[:space:]]Occluded[[:space:]]Surfaces[[:space:]]Completion/886d4931-cdc7-4727-9193-711bc1746bb2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Benchmarking[[:space:]]Audio[[:space:]]Visual[[:space:]]Segmentation[[:space:]]for[[:space:]]Long-Untrimmed[[:space:]]Videos/9bec05a3-5fcc-42b1-af2d-5714f93c115f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Benchmarking[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representation[[:space:]]and[[:space:]]Geometric[[:space:]]Rendering[[:space:]]in[[:space:]]Real-Time[[:space:]]RGB-D[[:space:]]SLAM/787daa9f-f928-446f-8637-38949ecba35f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Benchmarking[[:space:]]Segmentation[[:space:]]Models[[:space:]]with[[:space:]]Mask-Preserved[[:space:]]Attribute[[:space:]]Editing/b4010002-853e-445b-92d6-1f57993fe177_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Benchmarking[[:space:]]the[[:space:]]Robustness[[:space:]]of[[:space:]]Temporal[[:space:]]Action[[:space:]]Detection[[:space:]]Models[[:space:]]Against[[:space:]]Temporal[[:space:]]Corruptions/da2e7e3b-01fb-4135-a15d-0daebff17c74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BerfScene_[[:space:]]Bev-conditioned[[:space:]]Equivariant[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Infinite[[:space:]]3D[[:space:]]Scene[[:space:]]Generation/7eee3cff-6b1d-49c9-91f3-0e692b114ffe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Beyond[[:space:]]Average_[[:space:]]Individualized[[:space:]]Visual[[:space:]]Scanpath[[:space:]]Prediction/b8f9fd8c-68de-4817-84b6-44251e9f3557_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Beyond[[:space:]]First-Order[[:space:]]Tweedie_[[:space:]]Solving[[:space:]]Inverse[[:space:]]Problems[[:space:]]using[[:space:]]Latent[[:space:]]Diffusion/355906fe-ff73-41bb-9243-9fe72af42c31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Beyond[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]for[[:space:]]Image[[:space:]]Recognition[[:space:]]with[[:space:]]Task-Driven[[:space:]]Perceptual[[:space:]]Loss/a8e7255c-e598-4898-897a-0fc918b04e39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Beyond[[:space:]]Seen[[:space:]]Primitive[[:space:]]Concepts[[:space:]]and[[:space:]]Attribute-Object[[:space:]]Compositional[[:space:]]Learning/fd98ab2c-3027-4cce-94aa-350b6b407557_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Beyond[[:space:]]Text_[[:space:]]Frozen[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]in[[:space:]]Visual[[:space:]]Signal[[:space:]]Comprehension/95422830-0ab1-4426-a931-ecf8faf8d1f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Beyond[[:space:]]Textual[[:space:]]Constraints_[[:space:]]Learning[[:space:]]Novel[[:space:]]Diffusion[[:space:]]Conditions[[:space:]]with[[:space:]]Fewer[[:space:]]Examples/73daca31-b344-4472-8cda-2a6e9612e21a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bezier[[:space:]]Everywhere[[:space:]]All[[:space:]]at[[:space:]]Once_[[:space:]]Learning[[:space:]]Drivable[[:space:]]Lanes[[:space:]]as[[:space:]]Bezier[[:space:]]Graphs/b3ec81d2-040f-4d76-966d-263fb4798be0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bi-Causal_[[:space:]]Group[[:space:]]Activity[[:space:]]Recognition[[:space:]]via[[:space:]]Bidirectional[[:space:]]Causality/dde19080-01d1-4193-86cb-199d46df2998_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bi-SSC_[[:space:]]Geometric-Semantic[[:space:]]Bidirectional[[:space:]]Fusion[[:space:]]for[[:space:]]Camera-based[[:space:]]3D[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion/1c178ba3-9415-42e7-8686-3a4535d7eb15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bi-level[[:space:]]Learning[[:space:]]of[[:space:]]Task-Specific[[:space:]]Decoders[[:space:]]for[[:space:]]Joint[[:space:]]Registration[[:space:]]and[[:space:]]One-Shot[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/a60cda2e-0f4f-44b1-b7fb-ea68a6684b8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BiPer_[[:space:]]Binary[[:space:]]Neural[[:space:]]Networks[[:space:]]using[[:space:]]a[[:space:]]Periodic[[:space:]]Function/2973eb99-2ffe-4e14-9eee-0a4468e17b8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BiTT_[[:space:]]Bi-directional[[:space:]]Texture[[:space:]]Reconstruction[[:space:]]of[[:space:]]Interacting[[:space:]]Two[[:space:]]Hands[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/1c1f0705-0d98-48d8-9ce3-6b52bc92aacd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bidirectional[[:space:]]Autoregessive[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Dance[[:space:]]Generation/01bc04fd-2a23-4ff3-94b3-9435f67b3762_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bidirectional[[:space:]]Multi-Scale[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations[[:space:]]for[[:space:]]Image[[:space:]]Deraining/cc5ad6ac-2601-4b85-8c8b-523a6b1a3de8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BigGait_[[:space:]]Learning[[:space:]]Gait[[:space:]]Representation[[:space:]]You[[:space:]]Want[[:space:]]by[[:space:]]Large[[:space:]]Vision[[:space:]]Models/a0d78a49-5e51-4a61-a0c6-0f6e35012a7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bilateral[[:space:]]Adaptation[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection[[:space:]]with[[:space:]]Occlusion-Robustness/cea6c2cd-e86f-4b1c-8fa6-9018648eea6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bilateral[[:space:]]Event[[:space:]]Mining[[:space:]]and[[:space:]]Complementary[[:space:]]for[[:space:]]Event[[:space:]]Stream[[:space:]]Super-Resolution/7b3f59de-9611-4417-87a3-226bd3a5733c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bilateral[[:space:]]Propagation[[:space:]]Network[[:space:]]for[[:space:]]Depth[[:space:]]Completion/76ebbd49-2d53-422a-8124-0a20b492bbf7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BilevelPruning_[[:space:]]Unified[[:space:]]Dynamic[[:space:]]and[[:space:]]Static[[:space:]]Channel[[:space:]]Pruning[[:space:]]for[[:space:]]Convolutional[[:space:]]Neural[[:space:]]Networks/23f36021-449c-45ca-b04b-a9d131baad71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Binarized[[:space:]]Low-light[[:space:]]Raw[[:space:]]Video[[:space:]]Enhancement/e641e3ba-adbe-4c16-bc51-28a02790df47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Binding[[:space:]]Touch[[:space:]]to[[:space:]]Everything_[[:space:]]Learning[[:space:]]Unified[[:space:]]Multimodal[[:space:]]Tactile[[:space:]]Representations/92d6ae22-1528-4a49-8499-1cdea780c3d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BioCLIP_[[:space:]]A[[:space:]]Vision[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]the[[:space:]]Tree[[:space:]]of[[:space:]]Life/22913e53-bfb3-4a82-b45a-da30ca8ed74a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Blind[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment[[:space:]]Based[[:space:]]on[[:space:]]Geometric[[:space:]]Order[[:space:]]Learning/66ae1283-9142-4df4-baf7-48ee3d97507d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BlockGCN_[[:space:]]Redefine[[:space:]]Topology[[:space:]]Awareness[[:space:]]for[[:space:]]Skeleton-Based[[:space:]]Action[[:space:]]Recognition/5b0452b0-c20d-4e99-a4d6-83a9275bb61b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Blur-aware[[:space:]]Spatio-temporal[[:space:]]Sparse[[:space:]]Transformer[[:space:]]for[[:space:]]Video[[:space:]]Deblurring/8dad5a72-9be2-4ad7-8c86-fdde749df3a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Blur2Blur_[[:space:]]Blur[[:space:]]Conversion[[:space:]]for[[:space:]]Unsupervised[[:space:]]Image[[:space:]]Deblurring[[:space:]]on[[:space:]]Unknown[[:space:]]Domains/ad00db43-88f4-46bf-9e4e-82e7d5f5c614_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BoQ_[[:space:]]A[[:space:]]Place[[:space:]]is[[:space:]]Worth[[:space:]]a[[:space:]]Bag[[:space:]]of[[:space:]]Learnable[[:space:]]Queries/e41b26dc-a9ce-4a5a-b1a1-43508891d319_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BodyMAP[[:space:]]-[[:space:]]Jointly[[:space:]]Predicting[[:space:]]Body[[:space:]]Mesh[[:space:]]and[[:space:]]3D[[:space:]]Applied[[:space:]]Pressure[[:space:]]Map[[:space:]]for[[:space:]]People[[:space:]]in[[:space:]]Bed/d5a6d03f-3023-4b42-96f6-ae14bcda95cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Boosting[[:space:]]Adversarial[[:space:]]Training[[:space:]]via[[:space:]]Fisher-Rao[[:space:]]Norm-based[[:space:]]Regularization/4683d66d-b567-4109-9717-f6812e19131c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Boosting[[:space:]]Adversarial[[:space:]]Transferability[[:space:]]by[[:space:]]Block[[:space:]]Shuffle[[:space:]]and[[:space:]]Rotation/4192bf0c-80ac-4f8b-b688-7a42bf1bd730_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Boosting[[:space:]]Continual[[:space:]]Learning[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models[[:space:]]via[[:space:]]Mixture-of-Experts[[:space:]]Adapters/1f777fcb-226e-4be1-9a29-ee73f950297c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Boosting[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Moving[[:space:]]Average[[:space:]]Sampling[[:space:]]in[[:space:]]Frequency[[:space:]]Domain/479b7c60-4f32-4d47-84bf-2c2bc5f596a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Boosting[[:space:]]Flow-based[[:space:]]Generative[[:space:]]Super-Resolution[[:space:]]Models[[:space:]]via[[:space:]]Learned[[:space:]]Prior/ed8c064a-8024-48ce-8332-36c862bc0dd4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Boosting[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment[[:space:]]through[[:space:]]Efficient[[:space:]]Transformer[[:space:]]Adaptation[[:space:]]with[[:space:]]Local[[:space:]]Feature[[:space:]]Enhancement/499200e1-9e2b-4df2-9a52-aad590b31a86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Boosting[[:space:]]Image[[:space:]]Restoration[[:space:]]via[[:space:]]Priors[[:space:]]from[[:space:]]Pre-trained[[:space:]]Models/7088bae4-1b3d-41b0-bb19-52616e29d0c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Boosting[[:space:]]Neural[[:space:]]Representations[[:space:]]for[[:space:]]Videos[[:space:]]with[[:space:]]a[[:space:]]Conditional[[:space:]]Decoder/98ba4945-43ba-4750-9f2f-7e07f88dbb0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Boosting[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Zero-Shot[[:space:]]Day-Night[[:space:]]Domain[[:space:]]Adaptation/8f812f2f-96dd-427e-bd76-6e9f43c92de6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Boosting[[:space:]]Order-Preserving[[:space:]]and[[:space:]]Transferability[[:space:]]for[[:space:]]Neural[[:space:]]Architecture[[:space:]]Search_[[:space:]]a[[:space:]]Joint[[:space:]]Architecture[[:space:]]Refined[[:space:]]Search[[:space:]]and[[:space:]]Fine-tuning[[:space:]]Approach/34b628a0-9851-4bc5-b0a2-9a54cf7ad83a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Boosting[[:space:]]Self-Supervision[[:space:]]for[[:space:]]Single-View[[:space:]]Scene[[:space:]]Completion[[:space:]]via[[:space:]]Knowledge[[:space:]]Distillation/e2df5207-cf8b-4ad3-8495-70be141466e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Boosting[[:space:]]Spike[[:space:]]Camera[[:space:]]Image[[:space:]]Reconstruction[[:space:]]from[[:space:]]a[[:space:]]Perspective[[:space:]]of[[:space:]]Dealing[[:space:]]with[[:space:]]Spike[[:space:]]Fluctuations/9333cec8-4f68-4ccf-944e-f12f169909b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bootstrapping[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Radars[[:space:]]with[[:space:]]Self-Supervised[[:space:]]Learning/835eba7b-7297-4066-8d01-cce434324482_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bootstrapping[[:space:]]Chest[[:space:]]CT[[:space:]]Image[[:space:]]Understanding[[:space:]]by[[:space:]]Distilling[[:space:]]Knowledge[[:space:]]from[[:space:]]X-ray[[:space:]]Expert[[:space:]]Models/d544bf6c-97c6-4c47-aca0-a186cf072afb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bootstrapping[[:space:]]SparseFormers[[:space:]]from[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models/37a2bac7-a81b-4d15-b995-a00c5fbac3bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Brain[[:space:]]Decodes[[:space:]]Deep[[:space:]]Nets/c607ff23-b249-47f5-ac89-9667e168b4a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/BrainWash_[[:space:]]A[[:space:]]Poisoning[[:space:]]Attack[[:space:]]to[[:space:]]Forget[[:space:]]in[[:space:]]Continual[[:space:]]Learning/c1579b87-1101-45ec-bba3-a699ecec78db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Breathing[[:space:]]Life[[:space:]]Into[[:space:]]Sketches[[:space:]]Using[[:space:]]Text-to-Video[[:space:]]Priors/d481c8c7-d0d3-44c9-becf-4899884c473e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bridging[[:space:]]Remote[[:space:]]Sensors[[:space:]]with[[:space:]]Multisensor[[:space:]]Geospatial[[:space:]]Foundation[[:space:]]Models/be6c893a-15d2-4fa2-a450-a9e186313338_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bridging[[:space:]]the[[:space:]]Gap[[:space:]]Between[[:space:]]End-to-End[[:space:]]and[[:space:]]Two-Step[[:space:]]Text[[:space:]]Spotting/4d537b0e-fe6e-4521-9d58-585be7996785_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bridging[[:space:]]the[[:space:]]Gap_[[:space:]]A[[:space:]]Unified[[:space:]]Video[[:space:]]Comprehension[[:space:]]Framework[[:space:]]for[[:space:]]Moment[[:space:]]Retrieval[[:space:]]and[[:space:]]Highlight[[:space:]]Detection/d05266ca-afd4-4eb4-99fa-7d45290453e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bridging[[:space:]]the[[:space:]]Synthetic-to-Authentic[[:space:]]Gap_[[:space:]]Distortion-Guided[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Blind[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment/f59282c5-9fd2-4b2d-9b3b-f99573216f5e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Bring[[:space:]]Event[[:space:]]into[[:space:]]RGB[[:space:]]and[[:space:]]LiDAR_[[:space:]]Hierarchical[[:space:]]Visual-Motion[[:space:]]Fusion[[:space:]]for[[:space:]]Scene[[:space:]]Flow/2dc0a68a-1633-4ec2-8fca-caaf26c6f1b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Brush2Prompt_[[:space:]]Contextual[[:space:]]Prompt[[:space:]]Generator[[:space:]]for[[:space:]]Object[[:space:]]Inpainting/c813c161-9b68-46fb-aa94-a00dca52f48e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Building[[:space:]]Bridges[[:space:]]across[[:space:]]Spatial[[:space:]]and[[:space:]]Temporal[[:space:]]Resolutions_[[:space:]]Reference-Based[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Change[[:space:]]Priors[[:space:]]and[[:space:]]Conditional[[:space:]]Diffusion[[:space:]]Model/b16790f2-ece9-4b5c-846b-57d8d808aa94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Building[[:space:]]Optimal[[:space:]]Neural[[:space:]]Architectures[[:space:]]using[[:space:]]Interpretable[[:space:]]Knowledge/3ece47c5-a161-4e91-85de-28b874de35ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Building[[:space:]]Vision-Language[[:space:]]Models[[:space:]]on[[:space:]]Solid[[:space:]]Foundations[[:space:]]with[[:space:]]Masked[[:space:]]Distillation/cd973856-6b4c-4361-995d-7b60f17eb048_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Building[[:space:]]a[[:space:]]Strong[[:space:]]Pre-Training[[:space:]]Baseline[[:space:]]for[[:space:]]Universal[[:space:]]3D[[:space:]]Large-Scale[[:space:]]Perception/340bea8b-9694-4ca9-8be9-f35d985ee791_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Byzantine-robust[[:space:]]Decentralized[[:space:]]Federated[[:space:]]Learning[[:space:]]via[[:space:]]Dual-domain[[:space:]]Clustering[[:space:]]and[[:space:]]Trust[[:space:]]Bootstrapping/babc6988-4ad0-4fa7-aa11-1365e832e1d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/C2KD_[[:space:]]Bridging[[:space:]]the[[:space:]]Modality[[:space:]]Gap[[:space:]]for[[:space:]]Cross-Modal[[:space:]]Knowledge[[:space:]]Distillation/57667940-04a0-4dcb-93ad-49be03026370_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/C3Net_[[:space:]]Compound[[:space:]]Conditioned[[:space:]]ControlNet[[:space:]]for[[:space:]]Multimodal[[:space:]]Content[[:space:]]Generation/f4a0d0eb-a9dc-47a3-bd64-55a043f52442_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/C3_[[:space:]]High-Performance[[:space:]]and[[:space:]]Low-Complexity[[:space:]]Neural[[:space:]]Compression[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]or[[:space:]]Video/84624c39-82a1-4443-8975-16cbabed153f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CA-Jaccard_[[:space:]]Camera-aware[[:space:]]Jaccard[[:space:]]Distance[[:space:]]for[[:space:]]Person[[:space:]]Re-identification/558dd556-dd05-474f-9671-81ecc6e7cfb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CAD-SIGNet_[[:space:]]CAD[[:space:]]Language[[:space:]]Inference[[:space:]]from[[:space:]]Point[[:space:]]Clouds[[:space:]]using[[:space:]]Layer-wise[[:space:]]Sketch[[:space:]]Instance[[:space:]]Guided[[:space:]]Attention/3823178a-ee31-4ec3-af97-a5a24d05ab4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CADTalk_[[:space:]]An[[:space:]]Algorithm[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Semantic[[:space:]]Commenting[[:space:]]of[[:space:]]CAD[[:space:]]Programs/391aefbf-0c73-4152-9c58-c7fecb9f3273_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CAD_[[:space:]]Photorealistic[[:space:]]3D[[:space:]]Generation[[:space:]]via[[:space:]]Adversarial[[:space:]]Distillation/9a36d695-ed87-4c5c-bc45-e4da6d115ea5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CAGE_[[:space:]]Controllable[[:space:]]Articulation[[:space:]]GEneration/3cd6b273-a325-4248-9fe3-d566e6fa2243_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CAM[[:space:]]Back[[:space:]]Again_[[:space:]]Large[[:space:]]Kernel[[:space:]]CNNs[[:space:]]from[[:space:]]a[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Object[[:space:]]Localization[[:space:]]Perspective/00b6a6d7-c9f7-46c9-939f-a9adb3993725_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CAMEL_[[:space:]]CAusal[[:space:]]Motion[[:space:]]Enhancement[[:space:]]Tailored[[:space:]]for[[:space:]]Lifting[[:space:]]Text-driven[[:space:]]Video[[:space:]]Editing/41cd71a9-cd5b-49c7-9903-f499c6c389fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CAMixerSR_[[:space:]]Only[[:space:]]Details[[:space:]]Need[[:space:]]More[[:space:]]'Attention/a38c66a5-bbef-4d57-83da-77b56f57c853_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CAPE_[[:space:]]CAM[[:space:]]as[[:space:]]a[[:space:]]Probabilistic[[:space:]]Ensemble[[:space:]]for[[:space:]]Enhanced[[:space:]]DNN[[:space:]]Interpretation/52076ad3-3d15-4156-aeb2-9d06141f7d91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CARZero_[[:space:]]Cross-Attention[[:space:]]Alignment[[:space:]]for[[:space:]]Radiology[[:space:]]Zero-Shot[[:space:]]Classification/1b8ec5c2-4db4-4c5c-b25d-c899b7d55f99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CAT-DM_[[:space:]]Controllable[[:space:]]Accelerated[[:space:]]Virtual[[:space:]]Try-on[[:space:]]with[[:space:]]Diffusion[[:space:]]Model/4cb589b0-0220-4077-89fe-31ab3136f309_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CAT-Seg_[[:space:]]Cost[[:space:]]Aggregation[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/a4ec65eb-0581-4814-bfd0-b764bedce7cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CAT_[[:space:]]Exploiting[[:space:]]Inter-Class[[:space:]]Dynamics[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detection/c5917b02-5182-44a8-8aa8-2a264ae3c5d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CCEdit_[[:space:]]Creative[[:space:]]and[[:space:]]Controllable[[:space:]]Video[[:space:]]Editing[[:space:]]via[[:space:]]Diffusion[[:space:]]Models/2102154b-48d3-44f7-a727-a83d71b0e21e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CDFormer_[[:space:]]When[[:space:]]Degradation[[:space:]]Prediction[[:space:]]Embraces[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Blind[[:space:]]Image[[:space:]]Super-Resolution/f3a7dc88-4072-497a-8481-c26fec46d7cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CDMAD_[[:space:]]Class-Distribution-Mismatch-Aware[[:space:]]Debiasing[[:space:]]for[[:space:]]Class-Imbalanced[[:space:]]Semi-Supervised[[:space:]]Learning/f99edae5-80d6-4ab7-8468-d7bb1ff88b6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CFAT_[[:space:]]Unleashing[[:space:]]Triangular[[:space:]]Windows[[:space:]]for[[:space:]]Image[[:space:]]Super-resolution/dcbeb9f3-4524-422c-a605-2aaf9526c456_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CFPL-FAS_[[:space:]]Class[[:space:]]Free[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Generalizable[[:space:]]Face[[:space:]]Anti-spoofing/c38d1a1a-23a9-42a1-a1a4-d67a6e383aa0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CG-HOI_[[:space:]]Contact-Guided[[:space:]]3D[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Generation/d353881d-a77d-48d8-a75b-8231ca212967_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CGI-DM_[[:space:]]Digital[[:space:]]Copyright[[:space:]]Authentication[[:space:]]for[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Contrasting[[:space:]]Gradient[[:space:]]Inversion/9cfb7f2c-0b8e-4f56-91dc-16d1d72951d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CHAIN_[[:space:]]Enhancing[[:space:]]Generalization[[:space:]]in[[:space:]]Data-Efficient[[:space:]]GANs[[:space:]]via[[:space:]]lipsCHitz[[:space:]]continuity[[:space:]]constrAIned[[:space:]]Normalization/359c8b0b-8bcf-451d-97db-2d20ac8ce5f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CLIB-FIQA_[[:space:]]Face[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment[[:space:]]with[[:space:]]Confidence[[:space:]]Calibration/f3b783a1-c416-4b39-ad2b-92972d1cce68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CLIP[[:space:]]as[[:space:]]RNN_[[:space:]]Segment[[:space:]]Countless[[:space:]]Visual[[:space:]]Concepts[[:space:]]without[[:space:]]Training[[:space:]]Endeavor/7c0c2bfd-19d8-4266-bf91-75c74451272e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CLIP-BEVFormer_[[:space:]]Enhancing[[:space:]]Multi-View[[:space:]]Image-Based[[:space:]]BEV[[:space:]]Detector[[:space:]]with[[:space:]]Ground[[:space:]]Truth[[:space:]]Flow/68bace04-d789-4850-8373-ad0e34607268_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CLIP-Driven[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]via[[:space:]]Cross-Modality[[:space:]]Contrastive[[:space:]]Learning/1aa7225a-7f88-4b68-9c1a-c52a0d96e100_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CLIP-KD_[[:space:]]An[[:space:]]Empirical[[:space:]]Study[[:space:]]of[[:space:]]CLIP[[:space:]]Model[[:space:]]Distillation/c8c5b3e7-8be4-47eb-8b48-38292fa7be73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CLIPtone_[[:space:]]Unsupervised[[:space:]]Learning[[:space:]]for[[:space:]]Text-based[[:space:]]Image[[:space:]]Tone[[:space:]]Adjustment/b4464280-fff8-481b-8f9f-f447b0653c6c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CLOAF_[[:space:]]CoLlisiOn-Aware[[:space:]]Human[[:space:]]Flow/f794988a-200f-4dbd-834a-0352735f6561_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CLOVA_[[:space:]]A[[:space:]]Closed-LOop[[:space:]]Visual[[:space:]]Assistant[[:space:]]with[[:space:]]Tool[[:space:]]Usage[[:space:]]and[[:space:]]Update/e7ad4d42-5be2-4b8d-a5b8-95a60c92cb04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CLiC_[[:space:]]Concept[[:space:]]Learning[[:space:]]in[[:space:]]Context/24d7a6cb-6135-457d-a7dd-abc151005c8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CMA_[[:space:]]A[[:space:]]Chromaticity[[:space:]]Map[[:space:]]Adapter[[:space:]]for[[:space:]]Robust[[:space:]]Detection[[:space:]]of[[:space:]]Screen-Recapture[[:space:]]Document[[:space:]]Images/841a813a-c12e-4db3-bfc4-b56f8feef465_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CN-RMA_[[:space:]]Combined[[:space:]]Network[[:space:]]with[[:space:]]Ray[[:space:]]Marching[[:space:]]Aggregation[[:space:]]for[[:space:]]3D[[:space:]]Indoor[[:space:]]Object[[:space:]]Detection[[:space:]]from[[:space:]]Multi-view[[:space:]]Images/c59b5353-80b1-493e-90f7-e3bbb94bdd5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CNC-Net_[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]for[[:space:]]CNC[[:space:]]Machining[[:space:]]Operations/069c134e-7571-4c23-9082-ebb469f6a085_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/COCONut_[[:space:]]Modernizing[[:space:]]COCO[[:space:]]Segmentation/81ade5ff-cdbc-4e23-a38a-d872475f5bc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/COLMAP-Free[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/ff19612d-9627-475d-8a7f-27d57e340fbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CONFORM_[[:space:]]Contrast[[:space:]]is[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/1fb6d076-a53f-4779-a3d2-bd91cf0e2fdf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CORE-MPI_[[:space:]]Consistency[[:space:]]Object[[:space:]]Removal[[:space:]]with[[:space:]]Embedding[[:space:]]MultiPlane[[:space:]]Image/60b8a591-52aa-4a24-bf1e-58754a6d0086_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CORES_[[:space:]]Convolutional[[:space:]]Response-based[[:space:]]Score[[:space:]]for[[:space:]]Out-of-distribution[[:space:]]Detection/1a94b2d1-5e35-4cef-8fd5-f19da6b5711d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/COTR_[[:space:]]Compact[[:space:]]Occupancy[[:space:]]TRansformer[[:space:]]for[[:space:]]Vision-based[[:space:]]3D[[:space:]]Occupancy[[:space:]]Prediction/eb34c6d4-23cc-40a0-9aca-13ada8ca4389_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CPGA_[[:space:]]Coding[[:space:]]Priors-Guided[[:space:]]Aggregation[[:space:]]Network[[:space:]]for[[:space:]]Compressed[[:space:]]Video[[:space:]]Quality[[:space:]]Enhancement/5e7647ee-fb90-4f50-bc9a-d57e3b32948d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CPLIP_[[:space:]]Zero-Shot[[:space:]]Learning[[:space:]]for[[:space:]]Histopathology[[:space:]]with[[:space:]]Comprehensive[[:space:]]Vision-Language[[:space:]]Alignment/1fd92ab6-a7af-4638-b886-cfcbea2727b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CPP-Net_[[:space:]]Embracing[[:space:]]Multi-Scale[[:space:]]Feature[[:space:]]Fusion[[:space:]]into[[:space:]]Deep[[:space:]]Unfolding[[:space:]]CP-PPA[[:space:]]Network[[:space:]]for[[:space:]]Compressive[[:space:]]Sensing/93c4b0e5-9f07-4bac-849d-f5aac95cfb31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CPR-Coach_[[:space:]]Recognizing[[:space:]]Composite[[:space:]]Error[[:space:]]Actions[[:space:]]based[[:space:]]on[[:space:]]Single-class[[:space:]]Training/ba3ff7f4-baa9-4e99-aa79-603dec2cf257_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CPR_[[:space:]]Retrieval[[:space:]]Augmented[[:space:]]Generation[[:space:]]for[[:space:]]Copyright[[:space:]]Protection/98920e94-9369-4b96-b4e9-3ac23ab44c65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CRKD_[[:space:]]Enhanced[[:space:]]Camera-Radar[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Cross-modality[[:space:]]Knowledge[[:space:]]Distillation/188128bb-bf2e-4f28-b53a-eeebd3794ec0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CSTA_[[:space:]]CNN-based[[:space:]]Spatiotemporal[[:space:]]Attention[[:space:]]for[[:space:]]Video[[:space:]]Summarization/2583048d-7544-4123-a76c-0210a7ac76b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CURSOR_[[:space:]]Scalable[[:space:]]Mixed-Order[[:space:]]Hypergraph[[:space:]]Matching[[:space:]]with[[:space:]]CUR[[:space:]]Decomposition/7c646744-a471-43c6-aa9f-56ea7893a8ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CVT-xRF_[[:space:]]Contrastive[[:space:]]In-Voxel[[:space:]]Transformer[[:space:]]for[[:space:]]3D[[:space:]]Consistent[[:space:]]Radiance[[:space:]]Fields[[:space:]]from[[:space:]]Sparse[[:space:]]Inputs/19f5d386-a7aa-4981-be7e-d3520f44e426_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/C^2RV_[[:space:]]Cross-Regional[[:space:]]and[[:space:]]Cross-View[[:space:]]Learning[[:space:]]for[[:space:]]Sparse-View[[:space:]]CBCT[[:space:]]Reconstruction/8cd77b48-5973-4a75-8ca2-9ad11dc6cdf0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CaDeT_[[:space:]]a[[:space:]]Causal[[:space:]]Disentanglement[[:space:]]Approach[[:space:]]for[[:space:]]Robust[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/635614aa-2851-44dc-aac1-851316510d32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CaKDP_[[:space:]]Category-aware[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]and[[:space:]]Pruning[[:space:]]Framework[[:space:]]for[[:space:]]Lightweight[[:space:]]3D[[:space:]]Object[[:space:]]Detection/97b76402-227a-4d17-8da7-f0a9c88a8b8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Cache[[:space:]]Me[[:space:]]if[[:space:]]You[[:space:]]Can_[[:space:]]Accelerating[[:space:]]Diffusion[[:space:]]Models[[:space:]]through[[:space:]]Block[[:space:]]Caching/639b301d-8ff8-4116-8dbe-9d87b8fc33d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Calibrating[[:space:]]Multi-modal[[:space:]]Representations_[[:space:]]A[[:space:]]Pursuit[[:space:]]of[[:space:]]Group[[:space:]]Robustness[[:space:]]without[[:space:]]Annotations/bc46c562-f63d-45f7-ba29-d62bbff876a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Cam4DOcc_[[:space:]]Benchmark[[:space:]]for[[:space:]]Camera-Only[[:space:]]4D[[:space:]]Occupancy[[:space:]]Forecasting[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Applications/a8789e84-0e91-4383-98e7-dfdf61320805_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Can[[:space:]]Biases[[:space:]]in[[:space:]]ImageNet[[:space:]]Models[[:space:]]Explain[[:space:]]Generalization_/432d6ac6-5b85-4715-ada0-c3c2b886ae64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Can[[:space:]]I[[:space:]]Trust[[:space:]]Your[[:space:]]Answer_[[:space:]]Visually[[:space:]]Grounded[[:space:]]Video[[:space:]]Question[[:space:]]Answering/8e25ee23-cef5-4232-ad05-6d03475cea56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Can[[:space:]]Language[[:space:]]Beat[[:space:]]Numerical[[:space:]]Regression_[[:space:]]Language-Based[[:space:]]Multimodal[[:space:]]Trajectory[[:space:]]Prediction/cef7e796-3795-4ec6-99e3-8c3c58edb8cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Can[[:space:]]Protective[[:space:]]Perturbation[[:space:]]Safeguard[[:space:]]Personal[[:space:]]Data[[:space:]]from[[:space:]]Being[[:space:]]Exploited[[:space:]]by[[:space:]]Stable[[:space:]]Diffusion_/72015268-e947-4088-ae68-68432eacdfac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Can't[[:space:]]Make[[:space:]]an[[:space:]]Omelette[[:space:]]Without[[:space:]]Breaking[[:space:]]Some[[:space:]]Eggs_[[:space:]]Plausible[[:space:]]Action[[:space:]]Anticipation[[:space:]]Using[[:space:]]Large[[:space:]]Video-Language[[:space:]]Models/8ba84968-c082-4799-8554-b4cb4a64c1e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CapHuman_[[:space:]]Capture[[:space:]]Your[[:space:]]Moments[[:space:]]in[[:space:]]Parallel[[:space:]]Universes/12660aec-3746-418e-8572-8215643e13cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CapsFusion_[[:space:]]Rethinking[[:space:]]Image-Text[[:space:]]Data[[:space:]]at[[:space:]]Scale/3b81e6ca-3019-4096-8c7a-a2dfca18ccd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Capturing[[:space:]]Closely[[:space:]]Interacted[[:space:]]Two-Person[[:space:]]Motions[[:space:]]with[[:space:]]Reaction[[:space:]]Priors/ac1c2825-5afa-41dd-a7cf-8e6e7b171564_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Carve3D_[[:space:]]Improving[[:space:]]Multi-view[[:space:]]Reconstruction[[:space:]]Consistency[[:space:]]for[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]RL[[:space:]]Finetuning/c6e99d0d-11ec-439f-933e-28c6cf4c7d13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Category-Level[[:space:]]Multi-Part[[:space:]]Multi-Joint[[:space:]]3D[[:space:]]Shape[[:space:]]Assembly/108b424d-a705-49b3-a064-a4bbbda53a58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Causal[[:space:]]Mode[[:space:]]Multiplexer_[[:space:]]A[[:space:]]Novel[[:space:]]Framework[[:space:]]for[[:space:]]Unbiased[[:space:]]Multispectral[[:space:]]Pedestrian[[:space:]]Detection/2a4a18fa-9f7d-485e-9b18-8f26ecfedc33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Causal-CoG_[[:space:]]A[[:space:]]Causal-Effect[[:space:]]Look[[:space:]]at[[:space:]]Context[[:space:]]Generation[[:space:]]for[[:space:]]Boosting[[:space:]]Multi-modal[[:space:]]Language[[:space:]]Models/4184b885-ee49-45d9-bf73-603c65cf19ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CausalPC_[[:space:]]Improving[[:space:]]the[[:space:]]Robustness[[:space:]]of[[:space:]]Point[[:space:]]Cloud[[:space:]]Classification[[:space:]]by[[:space:]]Causal[[:space:]]Effect[[:space:]]Identification/5f4cc680-a33a-4422-b8f5-34806556998b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ChAda-ViT[[:space:]]_[[:space:]]Channel[[:space:]]Adaptive[[:space:]]Attention[[:space:]]for[[:space:]]Joint[[:space:]]Representation[[:space:]]Learning[[:space:]]of[[:space:]]Heterogeneous[[:space:]]Microscopy[[:space:]]Images/46f2b02d-18e3-49fb-b893-230e07676ea4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Characteristics[[:space:]]Matching[[:space:]]Based[[:space:]]Hash[[:space:]]Codes[[:space:]]Generation[[:space:]]for[[:space:]]Efficient[[:space:]]Fine-grained[[:space:]]Image[[:space:]]Retrieval/09c2c96d-09e2-4f18-b45a-5ddb79e89669_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Chat-UniVi_[[:space:]]Unified[[:space:]]Visual[[:space:]]Representation[[:space:]]Empowers[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]with[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Understanding/5adf87ef-478a-4110-a7a6-494a86dc2d61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ChatPose_[[:space:]]Chatting[[:space:]]about[[:space:]]3D[[:space:]]Human[[:space:]]Pose/0857ab71-c255-4085-9d13-081190a050ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ChatScene_[[:space:]]Knowledge-Enabled[[:space:]]Safety-Critical[[:space:]]Scenario[[:space:]]Generation[[:space:]]for[[:space:]]Autonomous[[:space:]]Vehicles/83d361b4-df11-4bac-a57c-3535ecf660d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Check[[:space:]]Locate[[:space:]]Rectify_[[:space:]]A[[:space:]]Training-Free[[:space:]]Layout[[:space:]]Calibration[[:space:]]System[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/0836cc28-3272-4455-a8ed-fc7577658e7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Choose[[:space:]]What[[:space:]]You[[:space:]]Need_[[:space:]]Disentangled[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Scene[[:space:]]Text[[:space:]]Recognition[[:space:]]Removal[[:space:]]and[[:space:]]Editing/cdb5f8dd-44fa-4a14-8942-3d30e4df61b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Cinematic[[:space:]]Behavior[[:space:]]Transfer[[:space:]]via[[:space:]]NeRF-based[[:space:]]Differentiable[[:space:]]Filming/a96830f6-8421-455a-9b16-851eabd4acb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Circuit[[:space:]]Design[[:space:]]and[[:space:]]Efficient[[:space:]]Simulation[[:space:]]of[[:space:]]Quantum[[:space:]]Inner[[:space:]]Product[[:space:]]and[[:space:]]Empirical[[:space:]]Studies[[:space:]]of[[:space:]]Its[[:space:]]Effect[[:space:]]on[[:space:]]Near-Term[[:space:]]Hybrid[[:space:]]Quantum-Classic[[:space:]]Machine[[:space:]]Learning/3dc83798-a296-4f4e-94e9-5bc1e086dded_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CityDreamer_[[:space:]]Compositional[[:space:]]Generative[[:space:]]Model[[:space:]]of[[:space:]]Unbounded[[:space:]]3D[[:space:]]Cities/54d4f587-8ced-4d40-9064-353c0383190b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Class[[:space:]]Incremental[[:space:]]Learning[[:space:]]with[[:space:]]Multi-Teacher[[:space:]]Distillation/30bfcded-9978-4e62-8f7c-5b9a7d9d161f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Class[[:space:]]Tokens[[:space:]]Infusion[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/9ea25d24-57f4-49c3-9a3f-5e0a8915d528_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Classes[[:space:]]Are[[:space:]]Not[[:space:]]Equal_[[:space:]]An[[:space:]]Empirical[[:space:]]Study[[:space:]]on[[:space:]]Image[[:space:]]Recognition[[:space:]]Fairness/0ad83484-1e22-42f1-9b82-169c14e9fddb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Clockwork[[:space:]]Diffusion_[[:space:]]Efficient[[:space:]]Generation[[:space:]]With[[:space:]]Model-Step[[:space:]]Distillation/f7c712d9-50b0-44be-8250-c73bc62fe859_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Close[[:space:]]Imitation[[:space:]]of[[:space:]]Expert[[:space:]]Retouching[[:space:]]for[[:space:]]Black-and-White[[:space:]]Photography/10dfdc64-4213-4d3e-948a-49d26f06901a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Closely[[:space:]]Interactive[[:space:]]Human[[:space:]]Reconstruction[[:space:]]with[[:space:]]Proxemics[[:space:]]and[[:space:]]Physics-Guided[[:space:]]Adaption/8c62e97d-18b0-4326-b880-c0060d52279e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Cloud-Device[[:space:]]Collaborative[[:space:]]Learning[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/6b6c75d9-8ea6-4124-8ccd-262606c0aef4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Clustering[[:space:]]Propagation[[:space:]]for[[:space:]]Universal[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/1dfb4348-085f-450d-9102-eaffa19075e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Clustering[[:space:]]for[[:space:]]Protein[[:space:]]Representation[[:space:]]Learning/d43cffe0-7da8-4318-9eae-e8627b81568e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Co-Speech[[:space:]]Gesture[[:space:]]Video[[:space:]]Generation[[:space:]]via[[:space:]]Motion-Decoupled[[:space:]]Diffusion[[:space:]]Model/640d8a1b-df4b-4600-a505-17ff2e70c853_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CoDeF_[[:space:]]Content[[:space:]]Deformation[[:space:]]Fields[[:space:]]for[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Video[[:space:]]Processing/bcd837e5-9f2b-4233-beb9-0ac80f96b7b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CoDe_[[:space:]]An[[:space:]]Explicit[[:space:]]Content[[:space:]]Decoupling[[:space:]]Framework[[:space:]]for[[:space:]]Image[[:space:]]Restoration/a75c834a-fc2c-4bfc-96da-fd5a490bcbf4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CoDi-2_[[:space:]]In-Context[[:space:]]Interleaved[[:space:]]and[[:space:]]Interactive[[:space:]]Any-to-Any[[:space:]]Generation/21ccf41e-b7fb-44dc-902e-05126c8e046d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CoDi_[[:space:]]Conditional[[:space:]]Diffusion[[:space:]]Distillation[[:space:]]for[[:space:]]Higher-Fidelity[[:space:]]and[[:space:]]Faster[[:space:]]Image[[:space:]]Generation/66c2ea7c-4861-45fc-aba8-6ede5543d2e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CoG-DQA_[[:space:]]Chain-of-Guiding[[:space:]]Learning[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Diagram[[:space:]]Question[[:space:]]Answering/d1f5efc3-df10-474b-bacc-d873cfe00891_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CoGS_[[:space:]]Controllable[[:space:]]Gaussian[[:space:]]Splatting/f9a3a03e-751a-422b-8fd8-0d9081ac7c55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CoSeR_[[:space:]]Bridging[[:space:]]Image[[:space:]]and[[:space:]]Language[[:space:]]for[[:space:]]Cognitive[[:space:]]Super-Resolution/27088af5-5565-4082-acff-20ef9d99edac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Coarse-to-Fine[[:space:]]Latent[[:space:]]Diffusion[[:space:]]for[[:space:]]Pose-Guided[[:space:]]Person[[:space:]]Image[[:space:]]Synthesis/bc0c6a6a-f762-4b6b-a355-40cb7723e5a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Codebook[[:space:]]Transfer[[:space:]]with[[:space:]]Part-of-Speech[[:space:]]for[[:space:]]Vector-Quantized[[:space:]]Image[[:space:]]Modeling/f0aa42c6-fc8c-41d0-bda1-d05bd6fd27aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CodedEvents_[[:space:]]Optimal[[:space:]]Point-Spread-Function[[:space:]]Engineering[[:space:]]for[[:space:]]3D-Tracking[[:space:]]with[[:space:]]Event[[:space:]]Cameras/7a4b1bf2-993c-4d93-a6f8-b9a3e1deb9e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CogAgent_[[:space:]]A[[:space:]]Visual[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]GUI[[:space:]]Agents/c3998dba-218e-4a25-84f8-11b68de947e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Coherence[[:space:]]As[[:space:]]Texture[[:space:]]-[[:space:]]Passive[[:space:]]Textureless[[:space:]]3D[[:space:]]Reconstruction[[:space:]]by[[:space:]]Self-interference/0b351f70-d996-4f9d-be78-8fbaad9acffa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Coherent[[:space:]]Temporal[[:space:]]Synthesis[[:space:]]for[[:space:]]Incremental[[:space:]]Action[[:space:]]Segmentation/7b90f260-43b8-408a-80ec-57ca826cefb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Collaborating[[:space:]]Foundation[[:space:]]Models[[:space:]]for[[:space:]]Domain[[:space:]]Generalized[[:space:]]Semantic[[:space:]]Segmentation/50d9bcdb-7f47-4735-ba1b-a2a937a246c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Collaborative[[:space:]]Learning[[:space:]]of[[:space:]]Anomalies[[:space:]]with[[:space:]]Privacy[[:space:]](CLAP)[[:space:]]for[[:space:]]Unsupervised[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection_[[:space:]]A[[:space:]]New[[:space:]]Baseline/8173347a-7cf4-4313-82d0-8cbb5c34fce8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Collaborative[[:space:]]Semantic[[:space:]]Occupancy[[:space:]]Prediction[[:space:]]with[[:space:]]Hybrid[[:space:]]Feature[[:space:]]Fusion[[:space:]]in[[:space:]]Connected[[:space:]]Automated[[:space:]]Vehicles/7bc1e7c8-31b3-4d88-819e-197fb66e20d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Color[[:space:]]Shift[[:space:]]Estimation-and-Correction[[:space:]]for[[:space:]]Image[[:space:]]Enhancement/84164354-e2df-4472-b332-2c8561bb89ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ColorPCR_[[:space:]]Color[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration[[:space:]]with[[:space:]]Multi-Stage[[:space:]]Geometric-Color[[:space:]]Fusion/9523f2e5-9c5a-4fab-9ef9-9e1a7cc97873_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Combining[[:space:]]Frame[[:space:]]and[[:space:]]GOP[[:space:]]Embeddings[[:space:]]for[[:space:]]Neural[[:space:]]Video[[:space:]]Representation/df3e696c-bb68-4d61-96a0-4dd2b761336b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CommonCanvas_[[:space:]]Open[[:space:]]Diffusion[[:space:]]Models[[:space:]]Trained[[:space:]]on[[:space:]]Creative-Commons[[:space:]]Images/1a525e34-9ad6-4be8-9d81-1c53de6a4ce1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Commonsense[[:space:]]Prototype[[:space:]]for[[:space:]]Outdoor[[:space:]]Unsupervised[[:space:]]3D[[:space:]]Object[[:space:]]Detection/3c63c694-e1f0-409d-b2f2-1078b2b37122_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Communication-Efficient[[:space:]]Collaborative[[:space:]]Perception[[:space:]]via[[:space:]]Information[[:space:]]Filling[[:space:]]with[[:space:]]Codebook/3f3feaa7-03ea-4554-9aae-b1bdab7b79ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Communication-Efficient[[:space:]]Federated[[:space:]]Learning[[:space:]]with[[:space:]]Accelerated[[:space:]]Client[[:space:]]Gradient/5d548808-060b-4e7c-91ee-f15a63bea155_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Compact[[:space:]]3D[[:space:]]Gaussian[[:space:]]Representation[[:space:]]for[[:space:]]Radiance[[:space:]]Field/93e4bf44-f951-41bb-8fef-299805e5c998_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Comparing[[:space:]]the[[:space:]]Decision-Making[[:space:]]Mechanisms[[:space:]]by[[:space:]]Transformers[[:space:]]and[[:space:]]CNNs[[:space:]]via[[:space:]]Explanation[[:space:]]Methods/4314ba42-4b91-4e21-b423-d816f4403659_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Complementing[[:space:]]Event[[:space:]]Streams[[:space:]]and[[:space:]]RGB[[:space:]]Frames[[:space:]]for[[:space:]]Hand[[:space:]]Mesh[[:space:]]Reconstruction/7b82d029-6717-41d5-8ee3-576b0ef382cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Composed[[:space:]]Video[[:space:]]Retrieval[[:space:]]via[[:space:]]Enriched[[:space:]]Context[[:space:]]and[[:space:]]Discriminative[[:space:]]Embeddings/b971f37e-0551-425d-9ab8-778b9b58aceb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Composing[[:space:]]Object[[:space:]]Relations[[:space:]]and[[:space:]]Attributes[[:space:]]for[[:space:]]Image-Text[[:space:]]Matching/7c507623-5fd9-444b-a698-2c977880eb80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Compositional[[:space:]]Chain-of-Thought[[:space:]]Prompting[[:space:]]for[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models/5db7b8f5-5d63-467d-a5bb-cd8f5f5dd4ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Compositional[[:space:]]Video[[:space:]]Understanding[[:space:]]with[[:space:]]Spatiotemporal[[:space:]]Structure-based[[:space:]]Transformers/21a973f2-46b5-476b-a179-7e05f00886c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Compressed[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Accelerated[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/d4f3822e-53da-4c09-bccb-c82d7d09696f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ConCon-Chi_[[:space:]]Concept-Context[[:space:]]Chimera[[:space:]]Benchmark[[:space:]]for[[:space:]]Personalized[[:space:]]Vision-Language[[:space:]]Tasks/beeaf857-30d8-4892-8522-374e7828bbf0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ConTex-Human_[[:space:]]Free-View[[:space:]]Rendering[[:space:]]of[[:space:]]Human[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]with[[:space:]]Texture-Consistent[[:space:]]Synthesis/c37f4de3-7d4e-4403-b639-7ebefe43d5b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Concept[[:space:]]Weaver_[[:space:]]Enabling[[:space:]]Multi-Concept[[:space:]]Fusion[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Models/1502b6f3-857d-4b41-9069-7c9296e4da0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Condition-Aware[[:space:]]Neural[[:space:]]Network[[:space:]]for[[:space:]]Controlled[[:space:]]Image[[:space:]]Generation/a90d6c3d-2fc7-452c-9f37-13562cf744c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Confronting[[:space:]]Ambiguity[[:space:]]in[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Score-Based[[:space:]]Diffusion[[:space:]]on[[:space:]]SE(3)/e0874087-2f50-4673-8a10-961b1c2b9193_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ConsistDreamer_[[:space:]]3D-Consistent[[:space:]]2D[[:space:]]Diffusion[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Scene[[:space:]]Editing/9b88ec28-510a-4c3b-82da-210f2f3b754c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ConsistNet_[[:space:]]Enforcing[[:space:]]3D[[:space:]]Consistency[[:space:]]for[[:space:]]Multi-view[[:space:]]Images[[:space:]]Diffusion/d7219c9b-e787-4bfc-b64a-5f8c537b4b18_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Consistency[[:space:]]and[[:space:]]Uncertainty_[[:space:]]Identifying[[:space:]]Unreliable[[:space:]]Responses[[:space:]]From[[:space:]]Black-Box[[:space:]]Vision-Language[[:space:]]Models[[:space:]]for[[:space:]]Selective[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/908abb11-a9f4-4585-a061-bcaeccc8c911_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Consistent[[:space:]]Prompting[[:space:]]for[[:space:]]Rehearsal-Free[[:space:]]Continual[[:space:]]Learning/b1222bdc-b40a-4da7-87de-62f02b50254c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Consistent3D_[[:space:]]Towards[[:space:]]Consistent[[:space:]]High-Fidelity[[:space:]]Text-to-3D[[:space:]]Generation[[:space:]]with[[:space:]]Deterministic[[:space:]]Sampling[[:space:]]Prior/57eac6ef-18b7-4642-9a5f-6a3fc3a35f35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Constrained[[:space:]]Layout[[:space:]]Generation[[:space:]]with[[:space:]]Factor[[:space:]]Graphs/2b964aa3-943a-4eaf-8c75-4aebb1828102_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Construct[[:space:]]to[[:space:]]Associate_[[:space:]]Cooperative[[:space:]]Context[[:space:]]Learning[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation/a06b85b7-3da4-40f5-ab15-0ec7d75ae39c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Constructing[[:space:]]and[[:space:]]Exploring[[:space:]]Intermediate[[:space:]]Domains[[:space:]]in[[:space:]]Mixed[[:space:]]Domain[[:space:]]Semi-supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/263c0481-75da-4d3d-829f-8aa7247da3d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Content-Adaptive[[:space:]]Non-Local[[:space:]]Convolution[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Pansharpening/8c0ec6e7-2688-4ac8-b208-edad6037f6c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Content-Style[[:space:]]Decoupling[[:space:]]for[[:space:]]Unsupervised[[:space:]]Makeup[[:space:]]Transfer[[:space:]]without[[:space:]]Generating[[:space:]]Pseudo[[:space:]]Ground[[:space:]]Truth/6696fdfd-a6a7-4e81-a030-4ec5b3a72538_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Context-Aware[[:space:]]Integration[[:space:]]of[[:space:]]Language[[:space:]]and[[:space:]]Visual[[:space:]]References[[:space:]]for[[:space:]]Natural[[:space:]]Language[[:space:]]Tracking/19052759-7cee-40bf-80dc-3fcaea3aa7bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Context-Guided[[:space:]]Spatio-Temporal[[:space:]]Video[[:space:]]Grounding/2a5fac1c-b682-4780-9135-cd5a0f6a5378_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Context-based[[:space:]]and[[:space:]]Diversity-driven[[:space:]]Specificity[[:space:]]in[[:space:]]Compositional[[:space:]]Zero-Shot[[:space:]]Learning/64f20b99-2c44-4f0f-b6c1-64c000b8fee8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ContextSeg_[[:space:]]Sketch[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]by[[:space:]]Querying[[:space:]]the[[:space:]]Context[[:space:]]with[[:space:]]Attention/3997e135-75a3-41fc-9a0f-211eb5b261d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Contextrast_[[:space:]]Contextual[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/e11df540-003e-4b10-bc97-3890e8c1d4fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Contextual[[:space:]]Augmented[[:space:]]Global[[:space:]]Contrast[[:space:]]for[[:space:]]Multimodal[[:space:]]Intent[[:space:]]Recognition/299b20f5-9f89-4d25-9243-4fb88caf4a9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Continual[[:space:]]Forgetting[[:space:]]for[[:space:]]Pre-trained[[:space:]]Vision[[:space:]]Models/1ce5c7b8-b8d1-4f83-9f6f-beb1956228f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Continual[[:space:]]Learning[[:space:]]for[[:space:]]Motion[[:space:]]Prediction[[:space:]]Model[[:space:]]via[[:space:]]Meta-Representation[[:space:]]Learning[[:space:]]and[[:space:]]Optimal[[:space:]]Memory[[:space:]]Buffer[[:space:]]Retention[[:space:]]Strategy/37f6d397-3f22-47ac-9074-aec13ad14ebd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Continual[[:space:]]Segmentation[[:space:]]with[[:space:]]Disentangled[[:space:]]Objectness[[:space:]]Learning[[:space:]]and[[:space:]]Class[[:space:]]Recognition/c85c47d1-edf9-40a7-896c-5edfe301a43d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Continual[[:space:]]Self-supervised[[:space:]]Learning_[[:space:]]Towards[[:space:]]Universal[[:space:]]Multi-modal[[:space:]]Medical[[:space:]]Data[[:space:]]Representation[[:space:]]Learning/e01fd6f2-120d-43e5-be82-9d5e7abb6b78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Continual-MAE_[[:space:]]Adaptive[[:space:]]Distribution[[:space:]]Masked[[:space:]]Autoencoders[[:space:]]for[[:space:]]Continual[[:space:]]Test-Time[[:space:]]Adaptation/641b187a-769b-4771-9889-b591e5298bf2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Continuous[[:space:]]Optical[[:space:]]Zooming_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Arbitrary-Scale[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]in[[:space:]]Real[[:space:]]World/fa8a22bb-a963-4314-b273-4aa7351163fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Continuous[[:space:]]Pose[[:space:]]for[[:space:]]Monocular[[:space:]]Cameras[[:space:]]in[[:space:]]Neural[[:space:]]Implicit[[:space:]]Representation/725e2e94-1f54-4b98-906d-8bf49c56d72c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Contrasting[[:space:]]Intra-Modal[[:space:]]and[[:space:]]Ranking[[:space:]]Cross-Modal[[:space:]]Hard[[:space:]]Negatives[[:space:]]to[[:space:]]Enhance[[:space:]]Visio-Linguistic[[:space:]]Compositional[[:space:]]Understanding/31e7d46d-aeb4-4eb5-88ca-e6496cca9953_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Contrastive[[:space:]]Denoising[[:space:]]Score[[:space:]]for[[:space:]]Text-guided[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Image[[:space:]]Editing/4a42ce98-38d8-4ba5-ad73-aa77d582ceff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]DeepFake[[:space:]]Classification[[:space:]]and[[:space:]]Localization[[:space:]]via[[:space:]]Multi-Label[[:space:]]Ranking/2f9203df-33f0-4172-af12-a42a9471222a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Contrastive[[:space:]]Mean-Shift[[:space:]]Learning[[:space:]]for[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/625f2c2c-6b5f-405b-b65f-580759bbdf5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Contrastive[[:space:]]Pre-Training[[:space:]]with[[:space:]]Multi-View[[:space:]]Fusion[[:space:]]for[[:space:]]No-Reference[[:space:]]Point[[:space:]]Cloud[[:space:]]Quality[[:space:]]Assessment/d8a3c45c-559a-4b51-b7a9-eafdf6103e8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Control4D_[[:space:]]Efficient[[:space:]]4D[[:space:]]Portrait[[:space:]]Editing[[:space:]]with[[:space:]]Text/737d554f-4ad5-41ca-9dad-d0ae83df4db5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ControlRoom3D_[[:space:]]Room[[:space:]]Generation[[:space:]]using[[:space:]]Semantic[[:space:]]Proxy[[:space:]]Rooms/feb3d3a9-2169-4026-8aea-d9b9838548b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ConvoFusion_[[:space:]]Multi-Modal[[:space:]]Conversational[[:space:]]Diffusion[[:space:]]for[[:space:]]Co-Speech[[:space:]]Gesture[[:space:]]Synthesis/a39eeaed-f519-4221-9042-f902cd3ced28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Convolutional[[:space:]]Prompting[[:space:]]meets[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Continual[[:space:]]Learning/da817dce-1b4e-4dd4-885f-44f0afabd062_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Cooperation[[:space:]]Does[[:space:]]Matter_[[:space:]]Exploring[[:space:]]Multi-Order[[:space:]]Bilateral[[:space:]]Relations[[:space:]]for[[:space:]]Audio-Visual[[:space:]]Segmentation/b33ef8bd-6465-4afa-9c96-981ce5893efc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CoralSCOP_[[:space:]]Segment[[:space:]]any[[:space:]]COral[[:space:]]Image[[:space:]]on[[:space:]]this[[:space:]]Planet/e3723d97-53da-4060-ac0d-8bb58495473f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CorrMatch_[[:space:]]Label[[:space:]]Propagation[[:space:]]via[[:space:]]Correlation[[:space:]]Matching[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation/00f77946-c539-482a-af7e-2ab27c8f2c3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Correcting[[:space:]]Diffusion[[:space:]]Generation[[:space:]]through[[:space:]]Resampling/436258dc-faeb-4588-8c13-f5badc49d6bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Correlation-Decoupled[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Multimodal[[:space:]]Sentiment[[:space:]]Analysis[[:space:]]with[[:space:]]Incomplete[[:space:]]Modalities/0ded1b06-f6d1-4b5a-a087-9978c148fc67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Correlation-aware[[:space:]]Coarse-to-fine[[:space:]]MLPs[[:space:]]for[[:space:]]Deformable[[:space:]]Medical[[:space:]]Image[[:space:]]Registration/6666014c-41f1-4b39-9faa-3f4a8dbfd8fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Correspondence-Free[[:space:]]Non-Rigid[[:space:]]Point[[:space:]]Set[[:space:]]Registration[[:space:]]Using[[:space:]]Unsupervised[[:space:]]Clustering[[:space:]]Analysis/d25533ac-25e3-4eb3-a021-abad315d6b6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CosalPure_[[:space:]]Learning[[:space:]]Concept[[:space:]]from[[:space:]]Group[[:space:]]Images[[:space:]]for[[:space:]]Robust[[:space:]]Co-Saliency[[:space:]]Detection/0455de5f-2efd-4061-a5d7-e31e88d3db36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CosmicMan_[[:space:]]A[[:space:]]Text-to-Image[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Humans/913dcb76-28cb-42db-a173-6f5bdbc64555_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Countering[[:space:]]Personalized[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]with[[:space:]]Influence[[:space:]]Watermarks/af03448f-370c-4565-9c20-d18d73c09d01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Coupled[[:space:]]Laplacian[[:space:]]Eigenmaps[[:space:]]for[[:space:]]Locally-Aware[[:space:]]3D[[:space:]]Rigid[[:space:]]Point[[:space:]]Cloud[[:space:]]Matching/6009d91a-dda3-423a-9a8e-58ad132601c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CricaVPR_[[:space:]]Cross-image[[:space:]]Correlation-aware[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Visual[[:space:]]Place[[:space:]]Recognition/514442b0-a00e-41eb-ab83-6a4a1d9746cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CroSel_[[:space:]]Cross[[:space:]]Selection[[:space:]]of[[:space:]]Confident[[:space:]]Pseudo[[:space:]]Labels[[:space:]]for[[:space:]]Partial-Label[[:space:]]Learning/ca7b302c-444d-4de9-95f4-b1b465d96faa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Cross[[:space:]]Initialization[[:space:]]for[[:space:]]Face[[:space:]]Personalization[[:space:]]of[[:space:]]Text-to-Image[[:space:]]Models/465159c7-b12d-4156-9e3d-cef2e406404d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Cross-Dimension[[:space:]]Affinity[[:space:]]Distillation[[:space:]]for[[:space:]]3D[[:space:]]EM[[:space:]]Neuron[[:space:]]Segmentation/c37084dd-6670-4b27-bc19-c9066d040b1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Cross-Domain[[:space:]]Few-Shot[[:space:]]Segmentation[[:space:]]via[[:space:]]Iterative[[:space:]]Support-Query[[:space:]]Correspondence[[:space:]]Mining/b86f1fc0-3e8a-451c-b29d-fbe7a969c1e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Cross-spectral[[:space:]]Gated-RGB[[:space:]]Stereo[[:space:]]Depth[[:space:]]Estimation/6da8d8f1-936b-47c5-91bf-db93f5089560_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Cross-view[[:space:]]and[[:space:]]Cross-pose[[:space:]]Completion[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Understanding/a14d26e8-6ddb-4d25-ab9c-0e5aad04d266_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CrossKD_[[:space:]]Cross-Head[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Object[[:space:]]Detection/57b74a0c-96cd-4c3d-a66c-5fd02c791467_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CrossMAE_[[:space:]]Cross-Modality[[:space:]]Masked[[:space:]]Autoencoders[[:space:]]for[[:space:]]Region-Aware[[:space:]]Audio-Visual[[:space:]]Pre-Training/03a006bd-74fa-4920-acce-3594eb7ba17a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CrowdDiff_[[:space:]]Multi-hypothesis[[:space:]]Crowd[[:space:]]Density[[:space:]]Estimation[[:space:]]using[[:space:]]Diffusion[[:space:]]Models/1ff131fb-ddeb-4e5b-9d35-ce31f9899307_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CuVLER_[[:space:]]Enhanced[[:space:]]Unsupervised[[:space:]]Object[[:space:]]Discoveries[[:space:]]through[[:space:]]Exhaustive[[:space:]]Self-Supervised[[:space:]]Transformers/4d215f61-769c-4800-b06d-b3961ba79e94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Curriculum[[:space:]]Point[[:space:]]Prompting[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation/075c5e4b-b01b-4abc-b136-08025043ab2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CurveCloudNet_[[:space:]]Processing[[:space:]]Point[[:space:]]Clouds[[:space:]]with[[:space:]]1D[[:space:]]Structure/5b20e8d9-b45d-4ba8-99f8-4404e77eb348_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CustomListener_[[:space:]]Text-guided[[:space:]]Responsive[[:space:]]Interaction[[:space:]]for[[:space:]]User-friendly[[:space:]]Listening[[:space:]]Head[[:space:]]Generation/b15be453-449f-4fbc-9b66-c4c248d2f5c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Customization[[:space:]]Assistant[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/f2800bdf-7fb4-4cb1-95e7-afc5a0cd6d77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Customize[[:space:]]your[[:space:]]NeRF_[[:space:]]Adaptive[[:space:]]Source[[:space:]]Driven[[:space:]]3D[[:space:]]Scene[[:space:]]Editing[[:space:]]via[[:space:]]Local-Global[[:space:]]Iterative[[:space:]]Training/be392c3e-c3f0-409f-a89b-9b6682002aa6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CyberDemo_[[:space:]]Augmenting[[:space:]]Simulated[[:space:]]Human[[:space:]]Demonstration[[:space:]]for[[:space:]]Real-World[[:space:]]Dexterous[[:space:]]Manipulation/7cab242e-31f8-41e8-b3c7-4f48154a173e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/CycleINR_[[:space:]]Cycle[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representation[[:space:]]for[[:space:]]Arbitrary-Scale[[:space:]]Volumetric[[:space:]]Super-Resolution[[:space:]]of[[:space:]]Medical[[:space:]]Data/bd64dc49-3074-4208-8fab-d09a8b6280ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Cyclic[[:space:]]Learning[[:space:]]for[[:space:]]Binaural[[:space:]]Audio[[:space:]]Generation[[:space:]]and[[:space:]]Localization/8da89f36-ebd5-4f0b-9c3b-f992e3b643b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/D3T_[[:space:]]Distinctive[[:space:]]Dual-Domain[[:space:]]Teacher[[:space:]]Zigzagging[[:space:]]Across[[:space:]]RGB-Thermal[[:space:]]Gap[[:space:]]for[[:space:]]Domain-Adaptive[[:space:]]Object[[:space:]]Detection/cfa93ca1-34a1-4ddf-ae42-06a03a1b23c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/D3still_[[:space:]]Decoupled[[:space:]]Differential[[:space:]]Distillation[[:space:]]for[[:space:]]Asymmetric[[:space:]]Image[[:space:]]Retrieval/8eab4750-31d8-4a6c-bcd6-bff8e283a741_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DAP_[[:space:]]A[[:space:]]Dynamic[[:space:]]Adversarial[[:space:]]Patch[[:space:]]for[[:space:]]Evading[[:space:]]Person[[:space:]]Detectors/35b4bd50-92da-4cd7-8072-6ea4aa285109_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DART_[[:space:]]Implicit[[:space:]]Doppler[[:space:]]Tomography[[:space:]]for[[:space:]]Radar[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/2cbea64d-dbe3-4274-a1a8-ec69ec128093_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DAVE[[:space:]]-[[:space:]]A[[:space:]]Detect-and-Verify[[:space:]]Paradigm[[:space:]]for[[:space:]]Low-Shot[[:space:]]Counting/3cf3aa5e-a391-469f-ae54-b942788e1339_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DEADiff_[[:space:]]An[[:space:]]Efficient[[:space:]]Stylization[[:space:]]Diffusion[[:space:]]Model[[:space:]]with[[:space:]]Disentangled[[:space:]]Representations/71fd2364-e697-4c71-b447-f89b171cc4ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DETRs[[:space:]]Beat[[:space:]]YOLOs[[:space:]]on[[:space:]]Real-time[[:space:]]Object[[:space:]]Detection/fa1c5c05-839d-4bca-bad8-a84eca62dcc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DGC-GNN_[[:space:]]Leveraging[[:space:]]Geometry[[:space:]]and[[:space:]]Color[[:space:]]Cues[[:space:]]for[[:space:]]Visual[[:space:]]Descriptor-Free[[:space:]]2D-3D[[:space:]]Matching/7181bcf2-53e3-4c8b-98f9-ba10d011a7ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DIBS_[[:space:]]Enhancing[[:space:]]Dense[[:space:]]Video[[:space:]]Captioning[[:space:]]with[[:space:]]Unlabeled[[:space:]]Videos[[:space:]]via[[:space:]]Pseudo[[:space:]]Boundary[[:space:]]Enrichment[[:space:]]and[[:space:]]Online[[:space:]]Refinement/e7374e1c-a8a5-4e55-9121-46b052cbd829_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DIEM_[[:space:]]Decomposition-Integration[[:space:]]Enhancing[[:space:]]Multimodal[[:space:]]Insights/e6a52ccd-c91b-4a3d-aedd-529c16822a37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DIMAT_[[:space:]]Decentralized[[:space:]]Iterative[[:space:]]Merging-And-Training[[:space:]]for[[:space:]]Deep[[:space:]]Learning[[:space:]]Models/612c25a2-297b-471c-ba0f-a362f1da8e81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DIOD_[[:space:]]Self-Distillation[[:space:]]Meets[[:space:]]Object[[:space:]]Discovery/90e29a4c-6fee-4935-ad3d-86a32ab07c46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DIRECT-3D_[[:space:]]Learning[[:space:]]Direct[[:space:]]Text-to-3D[[:space:]]Generation[[:space:]]on[[:space:]]Massive[[:space:]]Noisy[[:space:]]3D[[:space:]]Data/75419a3a-ec62-4a3b-81cd-ec1cf1f76f8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DITTO_[[:space:]]Dual[[:space:]]and[[:space:]]Integrated[[:space:]]Latent[[:space:]]Topologies[[:space:]]for[[:space:]]Implicit[[:space:]]3D[[:space:]]Reconstruction/d86a9170-dc87-430c-b042-670dd2b7e498_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DL3DV-10K_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Scene[[:space:]]Dataset[[:space:]]for[[:space:]]Deep[[:space:]]Learning-based[[:space:]]3D[[:space:]]Vision/ec7c3bbf-0982-4295-9b0d-098b7d155cf7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DMR_[[:space:]]Decomposed[[:space:]]Multi-Modality[[:space:]]Representations[[:space:]]for[[:space:]]Frames[[:space:]]and[[:space:]]Events[[:space:]]Fusion[[:space:]]in[[:space:]]Visual[[:space:]]Reinforcement[[:space:]]Learning/03ab7d7f-9b4c-4427-95fc-f86bc29bccc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DNGaussian_[[:space:]]Optimizing[[:space:]]Sparse-View[[:space:]]3D[[:space:]]Gaussian[[:space:]]Radiance[[:space:]]Fields[[:space:]]with[[:space:]]Global-Local[[:space:]]Depth[[:space:]]Normalization/1a9db803-e2f2-4baa-b317-c3b6e8782fc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DPHMs_[[:space:]]Diffusion[[:space:]]Parametric[[:space:]]Head[[:space:]]Models[[:space:]]for[[:space:]]Depth-based[[:space:]]Tracking/6df4c78e-cb6d-4ccb-9bb5-4a1c3998fe9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DPMesh_[[:space:]]Exploiting[[:space:]]Diffusion[[:space:]]Prior[[:space:]]for[[:space:]]Occluded[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/2ef08625-b811-4446-b47f-dfd0c9d6c350_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DREAM_[[:space:]]Diffusion[[:space:]]Rectification[[:space:]]and[[:space:]]Estimation-Adaptive[[:space:]]Models/ca3ac3b4-8d9d-438b-89fe-f03b26019420_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DRESS_[[:space:]]Instructing[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]to[[:space:]]Align[[:space:]]and[[:space:]]Interact[[:space:]]with[[:space:]]Humans[[:space:]]via[[:space:]]Natural[[:space:]]Language[[:space:]]Feedback/d0e46456-301b-4156-b0bf-3cb6a5a75453_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DS-NeRV_[[:space:]]Implicit[[:space:]]Neural[[:space:]]Video[[:space:]]Representation[[:space:]]with[[:space:]]Decomposed[[:space:]]Static[[:space:]]and[[:space:]]Dynamic[[:space:]]Codes/4cc8eade-fc01-46d0-9ccf-bb2afce11e1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DSGG_[[:space:]]Dense[[:space:]]Relation[[:space:]]Transformer[[:space:]]for[[:space:]]an[[:space:]]End-to-end[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/bf8a7e76-c990-48bb-b2df-ec1eb993eb08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DSL-FIQA_[[:space:]]Assessing[[:space:]]Facial[[:space:]]Image[[:space:]]Quality[[:space:]]via[[:space:]]Dual-Set[[:space:]]Degradation[[:space:]]Learning[[:space:]]and[[:space:]]Landmark-Guided[[:space:]]Transformer/604506f4-72ad-4cfc-b223-1e1ccec58325_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DUDF_[[:space:]]Differentiable[[:space:]]Unsigned[[:space:]]Distance[[:space:]]Fields[[:space:]]with[[:space:]]Hyperbolic[[:space:]]Scaling/aa1ca586-6720-4bb4-8688-4fdd533c8e6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DUSt3R_[[:space:]]Geometric[[:space:]]3D[[:space:]]Vision[[:space:]]Made[[:space:]]Easy/cb0ea578-b53d-4e78-8060-55294dfa8875_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DVMNet_[[:space:]]Computing[[:space:]]Relative[[:space:]]Pose[[:space:]]for[[:space:]]Unseen[[:space:]]Objects[[:space:]]Beyond[[:space:]]Hypotheses/06c1535e-2bca-43ff-90ee-2e46db4221b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DYSON_[[:space:]]Dynamic[[:space:]]Feature[[:space:]]Space[[:space:]]Self-Organization[[:space:]]for[[:space:]]Online[[:space:]]Task-Free[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning/b7930789-5c6e-4cee-a58f-5d7de62aa068_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/D^4_[[:space:]]Dataset[[:space:]]Distillation[[:space:]]via[[:space:]]Disentangled[[:space:]]Diffusion[[:space:]]Model/18b8088d-f767-443b-a861-75cde9aac1f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DaReNeRF_[[:space:]]Direction-aware[[:space:]]Representation[[:space:]]for[[:space:]]Dynamic[[:space:]]Scenes/95612d3e-41f7-4a07-9ca3-10a60caf6281_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DanceCamera3D_[[:space:]]3D[[:space:]]Camera[[:space:]]Movement[[:space:]]Synthesis[[:space:]]with[[:space:]]Music[[:space:]]and[[:space:]]Dance/af8a6037-75df-451c-9bcf-699b3dbf310c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dancing[[:space:]]with[[:space:]]Still[[:space:]]Images_[[:space:]]Video[[:space:]]Distillation[[:space:]]via[[:space:]]Static-Dynamic[[:space:]]Disentanglement/81243964-6292-4614-8b0a-70a60e8557ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Data[[:space:]]Poisoning[[:space:]]based[[:space:]]Backdoor[[:space:]]Attacks[[:space:]]to[[:space:]]Contrastive[[:space:]]Learning/925f06d2-d047-4677-a8ab-ac3094bc700a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Data[[:space:]]Valuation[[:space:]]and[[:space:]]Detections[[:space:]]in[[:space:]]Federated[[:space:]]Learning/f7965d6c-09c0-45d4-9575-f693c1b3b12a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Data-Efficient[[:space:]]Multimodal[[:space:]]Fusion[[:space:]]on[[:space:]]a[[:space:]]Single[[:space:]]GPU/887fba96-0363-4724-b495-9d8dc3ffcaa9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Data-Efficient[[:space:]]Unsupervised[[:space:]]Interpolation[[:space:]]Without[[:space:]]Any[[:space:]]Intermediate[[:space:]]Frame[[:space:]]for[[:space:]]4D[[:space:]]Medical[[:space:]]Images/ce6b20bd-5ec1-4355-a413-05986761e070_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Data-Free[[:space:]]Quantization[[:space:]]via[[:space:]]Pseudo-label[[:space:]]Filtering/21213099-e757-40fc-9662-495f362b8e06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Day-Night[[:space:]]Cross-domain[[:space:]]Vehicle[[:space:]]Re-identification/bb5f7894-ef08-47ec-ad8e-b30049e8ca48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/De-Diffusion[[:space:]]Makes[[:space:]]Text[[:space:]]a[[:space:]]Strong[[:space:]]Cross-Modal[[:space:]]Interface/eec5e079-ea28-4f5d-b742-7def29dca08b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/De-confounded[[:space:]]Data-free[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Handling[[:space:]]Distribution[[:space:]]Shifts/c02fb1b6-b6c6-409f-8312-b6d7f941575a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DeCoTR_[[:space:]]Enhancing[[:space:]]Depth[[:space:]]Completion[[:space:]]with[[:space:]]2D[[:space:]]and[[:space:]]3D[[:space:]]Attentions/f982ae62-1545-464a-a3ac-b81d945472ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DeIL_[[:space:]]Direct-and-Inverse[[:space:]]CLIP[[:space:]]for[[:space:]]Open-World[[:space:]]Few-Shot[[:space:]]Learning/97b5609a-bcc8-4d50-b491-9ea0a9a7f683_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DeMatch_[[:space:]]Deep[[:space:]]Decomposition[[:space:]]of[[:space:]]Motion[[:space:]]Field[[:space:]]for[[:space:]]Two-View[[:space:]]Correspondence[[:space:]]Learning/f480a9fc-54b1-4273-9e18-0a256f105276_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Decentralized[[:space:]]Directed[[:space:]]Collaboration[[:space:]]for[[:space:]]Personalized[[:space:]]Federated[[:space:]]Learning/59b74956-2b6a-4df3-af11-cb7e98721be6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Deciphering[[:space:]]'What'[[:space:]]and[[:space:]]'Where'[[:space:]]Visual[[:space:]]Pathways[[:space:]]from[[:space:]]Spectral[[:space:]]Clustering[[:space:]]of[[:space:]]Layer-Distributed[[:space:]]Neural[[:space:]]Representations/4e404140-94d8-4403-b30a-e177f90ed113_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Decompose-and-Compose_[[:space:]]A[[:space:]]Compositional[[:space:]]Approach[[:space:]]to[[:space:]]Mitigating[[:space:]]Spurious[[:space:]]Correlation/77f83950-8bde-4376-9f63-a5877c8aed16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Decomposing[[:space:]]Disease[[:space:]]Descriptions[[:space:]]for[[:space:]]Enhanced[[:space:]]Pathology[[:space:]]Detection_[[:space:]]A[[:space:]]Multi-Aspect[[:space:]]Vision-Language[[:space:]]Pre-training[[:space:]]Framework/d6fe0a96-979e-4681-a6aa-75c50e5c899a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DeconfuseTrack_[[:space:]]Dealing[[:space:]]with[[:space:]]Confusion[[:space:]]for[[:space:]]Multi-Object[[:space:]]Tracking/41382006-cbe0-4b49-b99c-e7e97bc731bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Decoupled[[:space:]]Pseudo-labeling[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection/ed0578b6-60e0-4f07-91df-f0f6d0aad924_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Decoupling[[:space:]]Static[[:space:]]and[[:space:]]Hierarchical[[:space:]]Motion[[:space:]]Perception[[:space:]]for[[:space:]]Referring[[:space:]]Video[[:space:]]Segmentation/65ec04d0-1a72-4af2-b17e-ef410873f595_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Deep[[:space:]]Equilibrium[[:space:]]Diffusion[[:space:]]Restoration[[:space:]]with[[:space:]]Parallel[[:space:]]Sampling/ecb35b5c-fbb7-49bb-8cf1-049b3f64c8a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Deep[[:space:]]Generative[[:space:]]Model[[:space:]]based[[:space:]]Rate-Distortion[[:space:]]for[[:space:]]Image[[:space:]]Downscaling[[:space:]]Assessment/c009d89f-84c0-4494-89f6-dcd1e100930b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Deep[[:space:]]Imbalanced[[:space:]]Regression[[:space:]]via[[:space:]]Hierarchical[[:space:]]Classification[[:space:]]Adjustment/6a6e4749-3b81-43f6-962d-f3c7b4677ff0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Deep[[:space:]]Single[[:space:]]Image[[:space:]]Camera[[:space:]]Calibration[[:space:]]by[[:space:]]Heatmap[[:space:]]Regression[[:space:]]to[[:space:]]Recover[[:space:]]Fisheye[[:space:]]Images[[:space:]]Under[[:space:]]Manhattan[[:space:]]World[[:space:]]Assumption/bfb16168-91b9-4657-b8bb-7684c699ec0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Deep[[:space:]]Video[[:space:]]Inverse[[:space:]]Tone[[:space:]]Mapping[[:space:]]Based[[:space:]]on[[:space:]]Temporal[[:space:]]Clues/b9524d3c-11ec-46e3-8d3c-17e16de5090d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Deep-TROJ_[[:space:]]An[[:space:]]Inference[[:space:]]Stage[[:space:]]Trojan[[:space:]]Insertion[[:space:]]Algorithm[[:space:]]through[[:space:]]Efficient[[:space:]]Weight[[:space:]]Replacement[[:space:]]Attack/0bedb42e-1fc3-4e2f-93fd-a69abdda1f0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DeepCache_[[:space:]]Accelerating[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Free/1b67d234-85be-47e6-9b5e-fb48901f17ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Defense[[:space:]]Against[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]on[[:space:]]No-Reference[[:space:]]Image[[:space:]]Quality[[:space:]]Models[[:space:]]with[[:space:]]Gradient[[:space:]]Norm[[:space:]]Regularization/ddbea5fa-9706-4c3a-8918-f1a75a53e09f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Defense[[:space:]]without[[:space:]]Forgetting_[[:space:]]Continual[[:space:]]Adversarial[[:space:]]Defense[[:space:]]with[[:space:]]Anisotropic[[:space:]]&[[:space:]]Isotropic[[:space:]]Pseudo[[:space:]]Replay/476e33a1-540d-4939-912e-ab9d21de4a4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Deformable[[:space:]]3D[[:space:]]Gaussians[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Monocular[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Reconstruction/d73366f4-fbde-4393-9ed7-fe2a0db8b014_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Deformable[[:space:]]One-shot[[:space:]]Face[[:space:]]Stylization[[:space:]]via[[:space:]]DINO[[:space:]]Semantic[[:space:]]Guidance/502de8d3-25a7-459d-a69a-05b9721cdcc7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Degrees[[:space:]]of[[:space:]]Freedom[[:space:]]Matter_[[:space:]]Inferring[[:space:]]Dynamics[[:space:]]from[[:space:]]Point[[:space:]]Trajectories/330de71f-3ac9-4294-a062-920f7f57e295_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DeiT-LT_[[:space:]]Distillation[[:space:]]Strikes[[:space:]]Back[[:space:]]for[[:space:]]Vision[[:space:]]Transformer[[:space:]]Training[[:space:]]on[[:space:]]Long-Tailed[[:space:]]Datasets/5f804b38-ad3f-41ee-986c-c5b0ed56673e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Delving[[:space:]]into[[:space:]]the[[:space:]]Trajectory[[:space:]]Long-tail[[:space:]]Distribution[[:space:]]for[[:space:]]Muti-object[[:space:]]Tracking/c6ec9c57-9a01-469b-a0ea-75d2dbada0b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DemoCaricature_[[:space:]]Democratising[[:space:]]Caricature[[:space:]]Generation[[:space:]]with[[:space:]]a[[:space:]]Rough[[:space:]]Sketch/c459bfbf-e498-4953-bb86-29a4b248cf14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DemoFusion_[[:space:]]Democratising[[:space:]]High-Resolution[[:space:]]Image[[:space:]]Generation[[:space:]]With[[:space:]]No[[:space:]]$$$/0ef8d512-0135-4f16-9af5-a5116b9fadaa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Denoising[[:space:]]Point[[:space:]]Clouds[[:space:]]in[[:space:]]Latent[[:space:]]Space[[:space:]]via[[:space:]]Graph[[:space:]]Convolution[[:space:]]and[[:space:]]Invertible[[:space:]]Neural[[:space:]]Network/95c9a539-f8df-4964-bfa6-455010a438d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dense[[:space:]]Optical[[:space:]]Tracking_[[:space:]]Connecting[[:space:]]the[[:space:]]Dots/f75c217f-09dc-4f38-b41b-821d95f9fe02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dense[[:space:]]Vision[[:space:]]Transformer[[:space:]]Compression[[:space:]]with[[:space:]]Few[[:space:]]Samples/de29c657-196a-471d-8810-f201c847fe2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Density-Adaptive[[:space:]]Model[[:space:]]Based[[:space:]]on[[:space:]]Motif[[:space:]]Matrix[[:space:]]for[[:space:]]Multi-Agent[[:space:]]Trajectory[[:space:]]Prediction/d79d2a84-99c1-4961-99d4-f02d7a166286_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Density-Guided[[:space:]]Semi-Supervised[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Dual-Space[[:space:]]Hardness[[:space:]]Sampling/560d7ff6-5245-4ed4-a448-5c0b60bfae43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Density-guided[[:space:]]Translator[[:space:]]Boosts[[:space:]]Synthetic-to-Real[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Segmentation[[:space:]]of[[:space:]]3D[[:space:]]Point[[:space:]]Clouds/2592e7d3-ce6f-42d9-aba4-5a3c7f12de5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Depth[[:space:]]Anything_[[:space:]]Unleashing[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]Large-Scale[[:space:]]Unlabeled[[:space:]]Data/24061cf8-83f1-4672-86d6-403a59e082da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Depth[[:space:]]Information[[:space:]]Assisted[[:space:]]Collaborative[[:space:]]Mutual[[:space:]]Promotion[[:space:]]Network[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Dehazing/aa348a70-1865-4fce-aa76-b439c0bc325c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Depth[[:space:]]Prompting[[:space:]]for[[:space:]]Sensor-Agnostic[[:space:]]Depth[[:space:]]Estimation/e30ba754-55c6-4d0f-a5c1-22852e4c0ba5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Depth-Aware[[:space:]]Concealed[[:space:]]Crop[[:space:]]Detection[[:space:]]in[[:space:]]Dense[[:space:]]Agricultural[[:space:]]Scenes/d1c47714-cee3-43a3-b4d3-79abaede45c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Depth-aware[[:space:]]Test-Time[[:space:]]Training[[:space:]]for[[:space:]]Zero-shot[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/65dceece-1db7-4fcd-814c-60a339346e11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Describing[[:space:]]Differences[[:space:]]in[[:space:]]Image[[:space:]]Sets[[:space:]]with[[:space:]]Natural[[:space:]]Language/d580900a-936e-44e9-a1d4-5372c72a6722_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Descriptor[[:space:]]and[[:space:]]Word[[:space:]]Soups_[[:space:]]Overcoming[[:space:]]the[[:space:]]Parameter[[:space:]]Efficiency[[:space:]]Accuracy[[:space:]]Tradeoff[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Few-shot[[:space:]]Learning/c15f55bf-6538-41ab-bc65-4ef6a900c2c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Desigen_[[:space:]]A[[:space:]]Pipeline[[:space:]]for[[:space:]]Controllable[[:space:]]Design[[:space:]]Template[[:space:]]Generation/d6d33070-cf90-4cfd-ac3f-cbc826fcb03a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Design2Cloth_[[:space:]]3D[[:space:]]Cloth[[:space:]]Generation[[:space:]]from[[:space:]]2D[[:space:]]Masks/88df92c0-6b74-4bed-954e-cb428f074ae8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DetCLIPv3_[[:space:]]Towards[[:space:]]Versatile[[:space:]]Generative[[:space:]]Open-vocabulary[[:space:]]Object[[:space:]]Detection/0573ff8a-eb0c-4194-b3df-43a072fba9e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DetDiffusion_[[:space:]]Synergizing[[:space:]]Generative[[:space:]]and[[:space:]]Perceptive[[:space:]]Models[[:space:]]for[[:space:]]Enhanced[[:space:]]Data[[:space:]]Generation[[:space:]]and[[:space:]]Perception/074bf15a-e86f-4ee0-8d92-1d1b0f69280b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Detector-Free[[:space:]]Structure[[:space:]]from[[:space:]]Motion/417ee905-2f67-4e2f-8cdc-a779b8765b0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Detours[[:space:]]for[[:space:]]Navigating[[:space:]]Instructional[[:space:]]Videos/0398ae01-660a-4ad0-b068-fdf6fc8e32fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Device-Wise[[:space:]]Federated[[:space:]]Network[[:space:]]Pruning/76319bed-c6f8-4ca1-ac34-7d50c8228eeb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dexterous[[:space:]]Grasp[[:space:]]Transformer/d0573b18-8acd-4d63-8574-13e95bcfd478_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiG-IN_[[:space:]]Diffusion[[:space:]]Guidance[[:space:]]for[[:space:]]Investigating[[:space:]]Networks[[:space:]]-[[:space:]]Uncovering[[:space:]]Classifier[[:space:]]Differences[[:space:]]Neuron[[:space:]]Visualisations[[:space:]]and[[:space:]]Visual[[:space:]]Counterfactual[[:space:]]Explanations/f7baa00b-2199-4507-88c6-64d21466dd0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiLiGenRT_[[:space:]]A[[:space:]]Photometric[[:space:]]Stereo[[:space:]]Dataset[[:space:]]with[[:space:]]Quantified[[:space:]]Roughness[[:space:]]and[[:space:]]Translucency/e9a75f89-ecec-4b08-abd4-96f5ab891a42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiPrompT_[[:space:]]Disentangled[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Multiple[[:space:]]Latent[[:space:]]Domain[[:space:]]Generalization[[:space:]]in[[:space:]]Federated[[:space:]]Learning/c01dc149-7a95-4397-910d-b877ac3448bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiSR-NeRF_[[:space:]]Diffusion-Guided[[:space:]]View-Consistent[[:space:]]Super-Resolution[[:space:]]NeRF/7373538e-8ab5-41ca-8a06-60c8bb9b6df5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiVAS_[[:space:]]Video[[:space:]]and[[:space:]]Audio[[:space:]]Synchronization[[:space:]]with[[:space:]]Dynamic[[:space:]]Frame[[:space:]]Rates/9b096c0f-8df9-444c-afcd-de7d03458f04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiVa-360_[[:space:]]The[[:space:]]Dynamic[[:space:]]Visual[[:space:]]Dataset[[:space:]]for[[:space:]]Immersive[[:space:]]Neural[[:space:]]Fields/e645629c-91b8-496b-a41c-3e43ca5ea2cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiaLoc_[[:space:]]An[[:space:]]Iterative[[:space:]]Approach[[:space:]]to[[:space:]]Embodied[[:space:]]Dialog[[:space:]]Localization/179b504d-09d2-473f-b9b3-783f80e0ca69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DifFlow3D_[[:space:]]Toward[[:space:]]Robust[[:space:]]Uncertainty-Aware[[:space:]]Scene[[:space:]]Flow[[:space:]]Estimation[[:space:]]with[[:space:]]Iterative[[:space:]]Diffusion-Based[[:space:]]Refinement/5086da18-cc18-4526-89e5-d9d74b1da6a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diff-BGM_[[:space:]]A[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Video[[:space:]]Background[[:space:]]Music[[:space:]]Generation/b379253a-d735-4be9-a83d-954f33a995d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diff-Plugin_[[:space:]]Revitalizing[[:space:]]Details[[:space:]]for[[:space:]]Diffusion-based[[:space:]]Low-level[[:space:]]Tasks/3fa63abc-51cd-4f76-baaf-d21aa2b74566_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffAM_[[:space:]]Diffusion-based[[:space:]]Adversarial[[:space:]]Makeup[[:space:]]Transfer[[:space:]]for[[:space:]]Facial[[:space:]]Privacy[[:space:]]Protection/d51186f7-1646-48a5-b2f8-5b8310dccd0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffAgent_[[:space:]]Fast[[:space:]]and[[:space:]]Accurate[[:space:]]Text-to-Image[[:space:]]API[[:space:]]Selection[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Model/6b2bf5c9-39b8-444e-a6de-547bb86d25f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffAssemble_[[:space:]]A[[:space:]]Unified[[:space:]]Graph-Diffusion[[:space:]]Model[[:space:]]for[[:space:]]2D[[:space:]]and[[:space:]]3D[[:space:]]Reassembly/dfa277e8-a9be-45b6-b0a2-ad02523e3b5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffAvatar_[[:space:]]Simulation-Ready[[:space:]]Garment[[:space:]]Optimization[[:space:]]with[[:space:]]Differentiable[[:space:]]Simulation/111b816f-37d8-4562-9644-ed0f87327709_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffCast_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]via[[:space:]]Residual[[:space:]]Diffusion[[:space:]]for[[:space:]]Precipitation[[:space:]]Nowcasting/b76402ef-cd5e-4515-b237-77dd609dddfa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffEditor_[[:space:]]Boosting[[:space:]]Accuracy[[:space:]]and[[:space:]]Flexibility[[:space:]]on[[:space:]]Diffusion-based[[:space:]]Image[[:space:]]Editing/c4be10b0-d3c7-44df-a8d8-ca8f1866ae73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffForensics_[[:space:]]Leveraging[[:space:]]Diffusion[[:space:]]Prior[[:space:]]to[[:space:]]Image[[:space:]]Forgery[[:space:]]Detection[[:space:]]and[[:space:]]Localization/dd69e8e1-a787-44cc-a0bb-eebe8fad25d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffHuman_[[:space:]]Probabilistic[[:space:]]Photorealistic[[:space:]]3D[[:space:]]Reconstruction[[:space:]]of[[:space:]]Humans/717f8a4d-1490-464d-990b-0c833c744bdf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffInDScene_[[:space:]]Diffusion-based[[:space:]]High-Quality[[:space:]]3D[[:space:]]Indoor[[:space:]]Scene[[:space:]]Generation/ee3731cc-e428-424f-ac97-19f5a651a497_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffLoc_[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Outdoor[[:space:]]LiDAR[[:space:]]Localization/ccf0426b-d606-44bd-9bd2-7f56c4e8b27b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffMOT_[[:space:]]A[[:space:]]Real-time[[:space:]]Diffusion-based[[:space:]]Multiple[[:space:]]Object[[:space:]]Tracker[[:space:]]with[[:space:]]Non-linear[[:space:]]Prediction/a252e67a-8a63-47af-8c72-09d7c805b33d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffMorpher_[[:space:]]Unleashing[[:space:]]the[[:space:]]Capability[[:space:]]of[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Image[[:space:]]Morphing/ddfe44ac-b07b-48b2-856d-4a202ac19d16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffPerformer_[[:space:]]Iterative[[:space:]]Learning[[:space:]]of[[:space:]]Consistent[[:space:]]Latent[[:space:]]Guidance[[:space:]]for[[:space:]]Diffusion-based[[:space:]]Human[[:space:]]Video[[:space:]]Generation/28c47348-bfe8-4d57-b5f1-38478506c63c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffPortrait3D_[[:space:]]Controllable[[:space:]]Diffusion[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Portrait[[:space:]]View[[:space:]]Synthesis/a08fba18-efe1-43be-82e9-0ad1039550e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffSCI_[[:space:]]Zero-Shot[[:space:]]Snapshot[[:space:]]Compressive[[:space:]]Imaging[[:space:]]via[[:space:]]Iterative[[:space:]]Spectral[[:space:]]Diffusion[[:space:]]Model/7b9ac18b-9ffe-44ea-a491-1f9f2789a1d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffSHEG_[[:space:]]A[[:space:]]Diffusion-Based[[:space:]]Approach[[:space:]]for[[:space:]]Real-Time[[:space:]]Speech-driven[[:space:]]Holistic[[:space:]]3D[[:space:]]Expression[[:space:]]and[[:space:]]Gesture[[:space:]]Generation/0bdc2c2c-f12d-452d-9894-6cd0a08f564b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffSal_[[:space:]]Joint[[:space:]]Audio[[:space:]]and[[:space:]]Video[[:space:]]Learning[[:space:]]for[[:space:]]Diffusion[[:space:]]Saliency[[:space:]]Prediction/1be379d2-cb46-4f75-90ba-becc71d34311_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffeomorphic[[:space:]]Template[[:space:]]Registration[[:space:]]for[[:space:]]Atmospheric[[:space:]]Turbulence[[:space:]]Mitigation/ceb03540-f310-429d-887d-5ecec376f7b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Differentiable[[:space:]]Display[[:space:]]Photometric[[:space:]]Stereo/d3e4269a-7eaa-4d07-802e-3e1df46f0f18_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Differentiable[[:space:]]Information[[:space:]]Bottleneck[[:space:]]for[[:space:]]Deterministic[[:space:]]Multi-view[[:space:]]Clustering/87f255cb-c0bc-4b09-a754-1212cabbbb99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Differentiable[[:space:]]Micro-Mesh[[:space:]]Construction/21e3ba39-4c40-4f3e-b9f9-2584925d37c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Differentiable[[:space:]]Neural[[:space:]]Surface[[:space:]]Refinement[[:space:]]for[[:space:]]Modeling[[:space:]]Transparent[[:space:]]Objects/c541e1d4-ef5f-4350-8fd9-b9b2f8e2f3ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Differentiable[[:space:]]Point-based[[:space:]]Inverse[[:space:]]Rendering/f6f931d9-4a05-4b6d-afa4-7310fe825b4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffuScene_[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Generative[[:space:]]Indoor[[:space:]]Scene[[:space:]]Synthesis/ffa0cf9b-1aa1-4d13-9bd8-8aaa571f83c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffuse[[:space:]]Attend[[:space:]]and[[:space:]]Segment_[[:space:]]Unsupervised[[:space:]]Zero-Shot[[:space:]]Segmentation[[:space:]]using[[:space:]]Stable[[:space:]]Diffusion/4954dd4c-3f96-4bbe-8277-4bc7e62a2175_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffuseMix_[[:space:]]Label-Preserving[[:space:]]Data[[:space:]]Augmentation[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/8071d802-25ad-47b6-bbcb-e2e1e6464454_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffusion[[:space:]]3D[[:space:]]Features[[:space:]](Diff3F)_[[:space:]]Decorating[[:space:]]Untextured[[:space:]]Shapes[[:space:]]with[[:space:]]Distilled[[:space:]]Semantic[[:space:]]Features/a6577b61-bfa0-4a1f-b03d-d467385c0aa8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffusion[[:space:]]Handles[[:space:]]Enabling[[:space:]]3D[[:space:]]Edits[[:space:]]for[[:space:]]Diffusion[[:space:]]Models[[:space:]]by[[:space:]]Lifting[[:space:]]Activations[[:space:]]to[[:space:]]3D/ecc6b20c-f2b2-4922-8863-a91075d90f84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffusion[[:space:]]Model[[:space:]]Alignment[[:space:]]Using[[:space:]]Direct[[:space:]]Preference[[:space:]]Optimization/987f9854-6fe5-49b4-b372-b0cb6aad3146_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffusion[[:space:]]Models[[:space:]]Without[[:space:]]Attention/f955b1eb-1a93-4a2c-9ac7-834c4d9c8f01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffusion[[:space:]]Reflectance[[:space:]]Map_[[:space:]]Single-Image[[:space:]]Stochastic[[:space:]]Inverse[[:space:]]Rendering[[:space:]]of[[:space:]]Illumination[[:space:]]and[[:space:]]Reflectance/c5983a42-9ef0-4f8c-be8f-0e2e14ffefb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffusion[[:space:]]Time-step[[:space:]]Curriculum[[:space:]]for[[:space:]]One[[:space:]]Image[[:space:]]to[[:space:]]3D[[:space:]]Generation/7c0ce23e-a3f9-4a4a-8c20-b613c8b96b23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffusion-EDFs_[[:space:]]Bi-equivariant[[:space:]]Denoising[[:space:]]Generative[[:space:]]Modeling[[:space:]]on[[:space:]]SE(3)[[:space:]]for[[:space:]]Visual[[:space:]]Robotic[[:space:]]Manipulation/52f82fa2-ac2e-4a06-9c30-c53d2332e090_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffusion-ES_[[:space:]]Gradient-free[[:space:]]Planning[[:space:]]with[[:space:]]Diffusion[[:space:]]for[[:space:]]Autonomous[[:space:]]and[[:space:]]Instruction-guided[[:space:]]Driving/43ebe2b2-cb73-4bae-94cb-0ee53fcf183d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffusion-FOF_[[:space:]]Single-View[[:space:]]Clothed[[:space:]]Human[[:space:]]Reconstruction[[:space:]]via[[:space:]]Diffusion-Based[[:space:]]Fourier[[:space:]]Occupancy[[:space:]]Field/5df2c032-a4c3-44ae-9ec6-ef870e381184_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffusion-based[[:space:]]Blind[[:space:]]Text[[:space:]]Image[[:space:]]Super-Resolution/259b8054-46e6-4b09-8798-d6a9d17ff4ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diffusion-driven[[:space:]]GAN[[:space:]]Inversion[[:space:]]for[[:space:]]Multi-Modal[[:space:]]Face[[:space:]]Image[[:space:]]Generation/16e371eb-d1c0-448f-9b66-c461b326c329_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffusionAvatars_[[:space:]]Deferred[[:space:]]Diffusion[[:space:]]for[[:space:]]High-fidelity[[:space:]]3D[[:space:]]Head[[:space:]]Avatars/f0395d92-f4eb-49de-9647-6df99fd5c9fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffusionGAN3D_[[:space:]]Boosting[[:space:]]Text-guided[[:space:]]3D[[:space:]]Generation[[:space:]]and[[:space:]]Domain[[:space:]]Adaptation[[:space:]]by[[:space:]]Combining[[:space:]]3D[[:space:]]GANs[[:space:]]and[[:space:]]Diffusion[[:space:]]Priors/f2dcd952-4b8a-40f0-8067-cdc51a42ca84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffusionLight_[[:space:]]Light[[:space:]]Probes[[:space:]]for[[:space:]]Free[[:space:]]by[[:space:]]Painting[[:space:]]a[[:space:]]Chrome[[:space:]]Ball/080f05c4-3850-4167-b0d2-2bd0163fb515_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffusionMTL_[[:space:]]Learning[[:space:]]Multi-Task[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Model[[:space:]]from[[:space:]]Partially[[:space:]]Annotated[[:space:]]Data/75343206-586a-42ff-9024-46fc8dc72ba6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffusionPoser_[[:space:]]Real-time[[:space:]]Human[[:space:]]Motion[[:space:]]Reconstruction[[:space:]]From[[:space:]]Arbitrary[[:space:]]Sparse[[:space:]]Sensors[[:space:]]Using[[:space:]]Autoregressive[[:space:]]Diffusion/952bac99-8ecf-4b71-ad88-832cead66ee5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffusionRegPose_[[:space:]]Enhancing[[:space:]]Multi-Person[[:space:]]Pose[[:space:]]Estimation[[:space:]]using[[:space:]]a[[:space:]]Diffusion-Based[[:space:]]End-to-End[[:space:]]Regression[[:space:]]Approach/45a9933f-1d4f-4d3e-ae0c-6827b8a02541_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiffusionTrack_[[:space:]]Point[[:space:]]Set[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Visual[[:space:]]Object[[:space:]]Tracking/7b0d2455-4e75-40b8-8756-9715b55680d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Digital[[:space:]]Life[[:space:]]Project_[[:space:]]Autonomous[[:space:]]3D[[:space:]]Characters[[:space:]]with[[:space:]]Social[[:space:]]Intelligence/00a429e9-f44f-4904-9892-c15808ba921c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Direct2.5_[[:space:]]Diverse[[:space:]]Text-to-3D[[:space:]]Generation[[:space:]]via[[:space:]]Multi-view[[:space:]]2.5D[[:space:]]Diffusion/30b48a1c-28ae-41da-9a35-dc60f8eeeb63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DisCo_[[:space:]]Disentangled[[:space:]]Control[[:space:]]for[[:space:]]Realistic[[:space:]]Human[[:space:]]Dance[[:space:]]Generation/a211c450-7970-4654-805f-c9c397894570_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Discontinuity-preserving[[:space:]]Normal[[:space:]]Integration[[:space:]]with[[:space:]]Auxiliary[[:space:]]Edges/2e62790b-129d-48fd-9c00-bac72ea3c67d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Discover[[:space:]]and[[:space:]]Mitigate[[:space:]]Multiple[[:space:]]Biased[[:space:]]Subgroups[[:space:]]in[[:space:]]Image[[:space:]]Classifiers/738228b9-2d49-4aac-bde9-d40e5b5bdfd1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Discovering[[:space:]]Syntactic[[:space:]]Interaction[[:space:]]Clues[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection/02ea9264-bafc-48b7-9ede-7e7a9802d51d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Discovering[[:space:]]and[[:space:]]Mitigating[[:space:]]Visual[[:space:]]Biases[[:space:]]through[[:space:]]Keyword[[:space:]]Explanation/4f7bc3e0-e7ad-40ea-a0be-4ba05eb80c75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Discriminability-Driven[[:space:]]Channel[[:space:]]Selection[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection/c49dec9d-a770-4428-8640-8dd9cdf8efd1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Discriminative[[:space:]]Pattern[[:space:]]Calibration[[:space:]]Mechanism[[:space:]]for[[:space:]]Source-Free[[:space:]]Domain[[:space:]]Adaptation/070d5af3-ef14-4cb2-b56e-b393e303d0a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Discriminative[[:space:]]Probing[[:space:]]and[[:space:]]Tuning[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/fa3d2c78-96f5-4372-8980-4a2d08b869f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Discriminative[[:space:]]Sample-Guided[[:space:]]and[[:space:]]Parameter-Efficient[[:space:]]Feature[[:space:]]Space[[:space:]]Adaptation[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Few-Shot[[:space:]]Learning/b8d256a3-e90f-4912-8144-91f916218cfd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Disentangled[[:space:]]Pre-training[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection/278c8d62-1997-481b-bc39-dd5beb5cddd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Disentangled[[:space:]]Prompt[[:space:]]Representation[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/1d1fa3b8-bc8c-4a56-8d76-8d89b4724b8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dispel[[:space:]]Darkness[[:space:]]for[[:space:]]Better[[:space:]]Fusion_[[:space:]]A[[:space:]]Controllable[[:space:]]Visual[[:space:]]Enhancer[[:space:]]based[[:space:]]on[[:space:]]Cross-modal[[:space:]]Conditional[[:space:]]Adversarial[[:space:]]Learning/7fdf9593-44c0-4b0f-83d9-374db648d7e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dispersed[[:space:]]Structured[[:space:]]Light[[:space:]]for[[:space:]]Hyperspectral[[:space:]]3D[[:space:]]Imaging/9879dad6-cd7c-4e38-a8c9-7f67c82458fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Distilled[[:space:]]Datamodel[[:space:]]with[[:space:]]Reverse[[:space:]]Gradient[[:space:]]Matching/d7ddb1ad-ea43-4b90-af8a-200010b66805_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Distilling[[:space:]]CLIP[[:space:]]with[[:space:]]Dual[[:space:]]Guidance[[:space:]]for[[:space:]]Learning[[:space:]]Discriminative[[:space:]]Human[[:space:]]Body[[:space:]]Shape[[:space:]]Representation/405307b0-9f01-428e-9ebb-a82cb17127ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Distilling[[:space:]]ODE[[:space:]]Solvers[[:space:]]of[[:space:]]Diffusion[[:space:]]Models[[:space:]]into[[:space:]]Smaller[[:space:]]Steps/34bc8c1f-6bae-401f-b053-29bbd4b10e09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Distilling[[:space:]]Semantic[[:space:]]Priors[[:space:]]from[[:space:]]SAM[[:space:]]to[[:space:]]Efficient[[:space:]]Image[[:space:]]Restoration[[:space:]]Models/12789ef0-4815-4206-8000-05e64804a250_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Distilling[[:space:]]Vision-Language[[:space:]]Models[[:space:]]on[[:space:]]Millions[[:space:]]of[[:space:]]Videos/8828befa-4d70-45fc-845d-fab6cc126c02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Distraction[[:space:]]is[[:space:]]All[[:space:]]You[[:space:]]Need_[[:space:]]Memory-Efficient[[:space:]]Image[[:space:]]Immunization[[:space:]]against[[:space:]]Diffusion-Based[[:space:]]Image[[:space:]]Editing/fbee6dc0-d9de-41e0-8606-40da1c87d444_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DistriFusion_[[:space:]]Distributed[[:space:]]Parallel[[:space:]]Inference[[:space:]]for[[:space:]]High-Resolution[[:space:]]Diffusion[[:space:]]Models/550b64fe-21a7-42c7-94d5-2c5bf12182fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Distribution-aware[[:space:]]Knowledge[[:space:]]Prototyping[[:space:]]for[[:space:]]Non-exemplar[[:space:]]Lifelong[[:space:]]Person[[:space:]]Re-identification/dc186928-c593-4047-b92c-da912985b7fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Distributionally[[:space:]]Generative[[:space:]]Augmentation[[:space:]]for[[:space:]]Fair[[:space:]]Facial[[:space:]]Attribute[[:space:]]Classification/c0595e7c-a382-440e-bead-c24efb7626f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DiverGen_[[:space:]]Improving[[:space:]]Instance[[:space:]]Segmentation[[:space:]]by[[:space:]]Learning[[:space:]]Wider[[:space:]]Data[[:space:]]Distribution[[:space:]]with[[:space:]]More[[:space:]]Diverse[[:space:]]Generative[[:space:]]Data/eeaaaaf5-aec0-4c96-93e2-08c29cdb4989_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diversified[[:space:]]and[[:space:]]Personalized[[:space:]]Multi-rater[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/33362df4-5f93-4123-ae10-d7274a7869ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Diversity-aware[[:space:]]Channel[[:space:]]Pruning[[:space:]]for[[:space:]]StyleGAN[[:space:]]Compression/6280f2c0-b9b7-4582-bd87-cf4cd31b00db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Do[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Encoders[[:space:]]Represent[[:space:]]the[[:space:]]World[[:space:]]Similarly_/ae3d16c2-a97b-4212-9cd5-7de96bf7cd1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Do[[:space:]]You[[:space:]]Remember_[[:space:]]Dense[[:space:]]Video[[:space:]]Captioning[[:space:]]with[[:space:]]Cross-Modal[[:space:]]Memory[[:space:]]Retrieval/4957fb9d-28db-4cb6-9204-7a204c6377bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DocRes_[[:space:]]A[[:space:]]Generalist[[:space:]]Model[[:space:]]Toward[[:space:]]Unifying[[:space:]]Document[[:space:]]Image[[:space:]]Restoration[[:space:]]Tasks/40d81229-9353-4bff-95b0-1ebce990f90a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Domain[[:space:]]Gap[[:space:]]Embeddings[[:space:]]for[[:space:]]Generative[[:space:]]Dataset[[:space:]]Augmentation/5494a5ac-2304-4ac2-9420-1dfdf881d3ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Domain[[:space:]]Prompt[[:space:]]Learning[[:space:]]with[[:space:]]Quaternion[[:space:]]Networks/303fb3fe-6ec6-468a-b337-88d6468b4619_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Domain[[:space:]]Separation[[:space:]]Graph[[:space:]]Neural[[:space:]]Networks[[:space:]]for[[:space:]]Saliency[[:space:]]Object[[:space:]]Ranking/c949ebe7-5184-416b-ac8d-3fdfd4559972_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Domain-Agnostic[[:space:]]Mutual[[:space:]]Prompting[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/bfac0f1b-64f8-4bbb-a613-d2a4de55522d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Domain-Rectifying[[:space:]]Adapter[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Few-Shot[[:space:]]Segmentation/8ac4cde6-1d67-48b7-84c5-43bdd0eb339e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Domain-Specific[[:space:]]Block[[:space:]]Selection[[:space:]]and[[:space:]]Paired-View[[:space:]]Pseudo-Labeling[[:space:]]for[[:space:]]Online[[:space:]]Test-Time[[:space:]]Adaptation/079ac40d-fd93-4b1d-81cd-f677205c98f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Don't[[:space:]]Drop[[:space:]]Your[[:space:]]Samples![[:space:]]Coherence-Aware[[:space:]]Training[[:space:]]Benefits[[:space:]]Conditional[[:space:]]Diffusion/637e6810-2de6-4b0a-a2c4-29351a197c24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Don't[[:space:]]Look[[:space:]]into[[:space:]]the[[:space:]]Dark_[[:space:]]Latent[[:space:]]Codes[[:space:]]for[[:space:]]Pluralistic[[:space:]]Image[[:space:]]Inpainting/494a7f7e-9c2e-4263-b47b-ceb5f8af3405_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Doodle[[:space:]]Your[[:space:]]3D_[[:space:]]From[[:space:]]Abstract[[:space:]]Freehand[[:space:]]Sketches[[:space:]]to[[:space:]]Precise[[:space:]]3D[[:space:]]Shapes/5a8a1158-d0a9-4102-a948-1a3cfa06964c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Doubly[[:space:]]Abductive[[:space:]]Counterfactual[[:space:]]Inference[[:space:]]for[[:space:]]Text-based[[:space:]]Image[[:space:]]Editing/66d374d5-de32-4685-9e1b-369e41f0cd30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dr.[[:space:]]Bokeh_[[:space:]]DiffeRentiable[[:space:]]Occlusion-aware[[:space:]]Bokeh[[:space:]]Rendering/5fcfe7c4-67be-418b-8051-c0d8252d33a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dr.Hair_[[:space:]]Reconstructing[[:space:]]Scalp-Connected[[:space:]]Hair[[:space:]]Strands[[:space:]]without[[:space:]]Pre-Training[[:space:]]via[[:space:]]Differentiable[[:space:]]Rendering[[:space:]]of[[:space:]]Line[[:space:]]Segments/f26898fb-be59-43a7-b03a-72aa594cf404_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dr2Net_[[:space:]]Dynamic[[:space:]]Reversible[[:space:]]Dual-Residual[[:space:]]Networks[[:space:]]for[[:space:]]Memory-Efficient[[:space:]]Finetuning/bb31d1fd-a351-4f02-9c95-013676b062f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Drag[[:space:]]Your[[:space:]]Noise_[[:space:]]Interactive[[:space:]]Point-based[[:space:]]Editing[[:space:]]via[[:space:]]Diffusion[[:space:]]Semantic[[:space:]]Propagation/33b40dba-662b-4e33-ba32-c5c3bb0270ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DragDiffusion_[[:space:]]Harnessing[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Interactive[[:space:]]Point-based[[:space:]]Image[[:space:]]Editing/7e564e37-cfd2-4951-a16b-3676cb8458fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Draw[[:space:]]Step[[:space:]]by[[:space:]]Step_[[:space:]]Reconstructing[[:space:]]CAD[[:space:]]Construction[[:space:]]Sequences[[:space:]]from[[:space:]]Point[[:space:]]Clouds[[:space:]]via[[:space:]]Multimodal[[:space:]]Diffusion./21e0e672-572d-41b3-afcf-11ff132ca2ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DreamAvatar_[[:space:]]Text-and-Shape[[:space:]]Guided[[:space:]]3D[[:space:]]Human[[:space:]]Avatar[[:space:]]Generation[[:space:]]via[[:space:]]Diffusion[[:space:]]Models/0a621ba0-89c8-43ad-b979-25d8e4971daf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DreamComposer_[[:space:]]Controllable[[:space:]]3D[[:space:]]Object[[:space:]]Generation[[:space:]]via[[:space:]]Multi-View[[:space:]]Conditions/623a866d-dd7b-4618-9c06-30d2dbe4d9fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DreamControl_[[:space:]]Control-Based[[:space:]]Text-to-3D[[:space:]]Generation[[:space:]]with[[:space:]]3D[[:space:]]Self-Prior/5e088b57-2ab4-4bba-b801-fb4c346da8ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DreamMatcher_[[:space:]]Appearance[[:space:]]Matching[[:space:]]Self-Attention[[:space:]]for[[:space:]]Semantically-Consistent[[:space:]]Text-to-Image[[:space:]]Personalization/77ec074b-ee13-4a73-92e0-0314eb77a866_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DreamPropeller_[[:space:]]Supercharge[[:space:]]Text-to-3D[[:space:]]Generation[[:space:]]with[[:space:]]Parallel[[:space:]]Sampling/2c7bcaaf-b186-4337-9555-9c5b84e9339d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DreamSalon_[[:space:]]A[[:space:]]Staged[[:space:]]Diffusion[[:space:]]Framework[[:space:]]for[[:space:]]Preserving[[:space:]]Identity-Context[[:space:]]in[[:space:]]Editable[[:space:]]Face[[:space:]]Generation/9c5ddb99-4556-4772-b051-134150e5e01b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DreamVideo_[[:space:]]Composing[[:space:]]Your[[:space:]]Dream[[:space:]]Videos[[:space:]]with[[:space:]]Customized[[:space:]]Subject[[:space:]]and[[:space:]]Motion/eed5e8f0-114a-45b5-a61b-330f597e3df6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DriveTrack_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Long-Range[[:space:]]Point[[:space:]]Tracking[[:space:]]in[[:space:]]Real-World[[:space:]]Videos/5015a4e5-1f68-44d9-98d4-1e9ac131076b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DriveWorld_[[:space:]]4D[[:space:]]Pre-trained[[:space:]]Scene[[:space:]]Understanding[[:space:]]via[[:space:]]World[[:space:]]Models[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/f6d47dee-999b-4ee9-803f-ad19884b432a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Driving[[:space:]]Everywhere[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]Policy[[:space:]]Adaptation/ced507f2-8d79-4aab-b5d5-f34cf5a10308_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Driving[[:space:]]into[[:space:]]the[[:space:]]Future_[[:space:]]Multiview[[:space:]]Visual[[:space:]]Forecasting[[:space:]]and[[:space:]]Planning[[:space:]]with[[:space:]]World[[:space:]]Model[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/98637d83-1edf-45c1-9f4b-bb0fcf9036f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Driving-Video[[:space:]]Dehazing[[:space:]]with[[:space:]]Non-Aligned[[:space:]]Regularization[[:space:]]for[[:space:]]Safety[[:space:]]Assistance/659f5fcf-ba6b-4574-9bde-69e03d711590_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DrivingGaussian_[[:space:]]Composite[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Surrounding[[:space:]]Dynamic[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Scenes/2fa13699-c688-48ef-8047-c2e9149f5fee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DuPL_[[:space:]]Dual[[:space:]]Student[[:space:]]with[[:space:]]Trustworthy[[:space:]]Progressive[[:space:]]Learning[[:space:]]for[[:space:]]Robust[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/b0124abb-e45a-494e-b1c3-05b932ca8aa9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dual[[:space:]]DETRs[[:space:]]for[[:space:]]Multi-Label[[:space:]]Temporal[[:space:]]Action[[:space:]]Detection/e8d6bf87-033b-4784-ad0d-1d05a9e91fd7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dual[[:space:]]Memory[[:space:]]Networks_[[:space:]]A[[:space:]]Versatile[[:space:]]Adaptation[[:space:]]Approach[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/68cfc1da-46b9-4da9-afd3-e9d63cc8a523_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dual[[:space:]]Pose-invariant[[:space:]]Embeddings_[[:space:]]Learning[[:space:]]Category[[:space:]]and[[:space:]]Object-specific[[:space:]]Discriminative[[:space:]]Representations[[:space:]]for[[:space:]]Recognition[[:space:]]and[[:space:]]Retrieval/fe7eee9b-e2f8-4f61-ab5b-7360a789de79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dual[[:space:]]Prior[[:space:]]Unfolding[[:space:]]for[[:space:]]Snapshot[[:space:]]Compressive[[:space:]]Imaging/580b3e38-9f77-4de9-91bf-38ec72151d82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dual[[:space:]]Prototype[[:space:]]Attention[[:space:]]for[[:space:]]Unsupervised[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/8eb69291-e000-4aef-b767-3cb0557c2a9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dual-Consistency[[:space:]]Model[[:space:]]Inversion[[:space:]]for[[:space:]]Non-Exemplar[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning/520c826f-8cbd-4a53-a958-cc72e9444d9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dual-Enhanced[[:space:]]Coreset[[:space:]]Selection[[:space:]]with[[:space:]]Class-wise[[:space:]]Collaboration[[:space:]]for[[:space:]]Online[[:space:]]Blurry[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning/48bd2bc8-3786-4f52-a6a5-f8f54adedd75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dual-Scale[[:space:]]Transformer[[:space:]]for[[:space:]]Large-Scale[[:space:]]Single-Pixel[[:space:]]Imaging/ed35fef7-f429-4f05-a98f-0561d4629006_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dual-View[[:space:]]Visual[[:space:]]Contextualization[[:space:]]for[[:space:]]Web[[:space:]]Navigation/de6a3bb6-eed0-49d4-a7a6-45305702ad87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DualAD_[[:space:]]Disentangling[[:space:]]the[[:space:]]Dynamic[[:space:]]and[[:space:]]Static[[:space:]]World[[:space:]]for[[:space:]]End-to-End[[:space:]]Driving/e39ad68c-b907-4cdc-8833-1262a3710de8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DyBluRF_[[:space:]]Dynamic[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]from[[:space:]]Blurry[[:space:]]Monocular[[:space:]]Video/afb0babd-edc6-46be-958f-06a1a659d81f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/DynVideo-E_[[:space:]]Harnessing[[:space:]]Dynamic[[:space:]]NeRF[[:space:]]for[[:space:]]Large-Scale[[:space:]]Motion-[[:space:]]and[[:space:]]View-Change[[:space:]]Human-Centric[[:space:]]Video[[:space:]]Editing/438d63b7-240c-46d3-b411-e0269993152e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dynamic[[:space:]]Adapter[[:space:]]Meets[[:space:]]Prompt[[:space:]]Tuning_[[:space:]]Parameter-Efficient[[:space:]]Transfer[[:space:]]Learning[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Analysis/4ddb62ba-f226-497a-9e27-67c4e57e02eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dynamic[[:space:]]Cues-Assisted[[:space:]]Transformer[[:space:]]for[[:space:]]Robust[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/18419ae7-38a2-405c-bef7-af804859abb3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dynamic[[:space:]]Graph[[:space:]]Representation[[:space:]]with[[:space:]]Knowledge-aware[[:space:]]Attention[[:space:]]for[[:space:]]Histopathology[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Analysis/118e445e-c9f5-4f50-822f-70a81f6990cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dynamic[[:space:]]Inertial[[:space:]]Poser[[:space:]](DynaIP)_[[:space:]]Part-Based[[:space:]]Motion[[:space:]]Dynamics[[:space:]]Learning[[:space:]]for[[:space:]]Enhanced[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]Sparse[[:space:]]Inertial[[:space:]]Sensors/e78c5d5c-a228-4890-b0b1-bf22bb915983_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dynamic[[:space:]]LiDAR[[:space:]]Re-simulation[[:space:]]using[[:space:]]Compositional[[:space:]]Neural[[:space:]]Fields/f502bf31-c27b-4e80-a82d-c752f70c1d4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dynamic[[:space:]]Policy-Driven[[:space:]]Adaptive[[:space:]]Multi-Instance[[:space:]]Learning[[:space:]]for[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Classification/f4320555-dfb9-4c7f-a1cd-50844dd9dcc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dynamic[[:space:]]Prompt[[:space:]]Optimizing[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/af258c73-c151-4d10-a555-01670475daeb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dynamic[[:space:]]Support[[:space:]]Information[[:space:]]Mining[[:space:]]for[[:space:]]Category-Agnostic[[:space:]]Pose[[:space:]]Estimation/f434778e-bd29-466f-86aa-9af5c2b18a03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Dysen-VDM_[[:space:]]Empowering[[:space:]]Dynamics-aware[[:space:]]Text-to-Video[[:space:]]Diffusion[[:space:]]with[[:space:]]LLMs/6e2a1ed0-806d-4437-b8f6-29e3e3bc395a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/E-GPS_[[:space:]]Explainable[[:space:]]Geometry[[:space:]]Problem[[:space:]]Solving[[:space:]]via[[:space:]]Top-Down[[:space:]]Solver[[:space:]]and[[:space:]]Bottom-Up[[:space:]]Generator/d88813ea-6996-4a82-937f-f36899d585e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EAGLE_[[:space:]]Eigen[[:space:]]Aggregation[[:space:]]Learning[[:space:]]for[[:space:]]Object-Centric[[:space:]]Unsupervised[[:space:]]Semantic[[:space:]]Segmentation/dc4d6ab5-3d0d-4512-8238-9679352b211b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EASE-DETR_[[:space:]]Easing[[:space:]]the[[:space:]]Competition[[:space:]]among[[:space:]]Object[[:space:]]Queries/9ff0d4e0-f729-41a1-9284-a3462c62983a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ECLIPSE_[[:space:]]A[[:space:]]Resource-Efficient[[:space:]]Text-to-Image[[:space:]]Prior[[:space:]]for[[:space:]]Image[[:space:]]Generations/83a8980c-9503-4c5f-99c0-a6eb3b04e979_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ECLIPSE_[[:space:]]Efficient[[:space:]]Continual[[:space:]]Learning[[:space:]]in[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]with[[:space:]]Visual[[:space:]]Prompt[[:space:]]Tuning/9f39aafa-e0ef-4ad0-a765-ad142df3abb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ECoDepth_[[:space:]]Effective[[:space:]]Conditioning[[:space:]]of[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/7e3c5935-f184-4aa8-a48b-682e7477329f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EFHQ_[[:space:]]Multi-purpose[[:space:]]ExtremePose-Face-HQ[[:space:]]dataset/65a10ed9-cef6-42e8-a141-8586b125aea4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EFormer_[[:space:]]Enhanced[[:space:]]Transformer[[:space:]]towards[[:space:]]Semantic-Contour[[:space:]]Features[[:space:]]of[[:space:]]Foreground[[:space:]]for[[:space:]]Portraits[[:space:]]Matting/652e59cf-51c6-401e-a44e-7a8d54a66599_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EGTR_[[:space:]]Extracting[[:space:]]Graph[[:space:]]from[[:space:]]Transformer[[:space:]]for[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/174c2199-8aa8-4e45-8d2e-205eb2e2e099_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EMAGE_[[:space:]]Towards[[:space:]]Unified[[:space:]]Holistic[[:space:]]Co-Speech[[:space:]]Gesture[[:space:]]Generation[[:space:]]via[[:space:]]Expressive[[:space:]]Masked[[:space:]]Audio[[:space:]]Gesture[[:space:]]Modeling/f8a85da7-9221-4a37-baf0-e5a3b46705e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EMCAD_[[:space:]]Efficient[[:space:]]Multi-scale[[:space:]]Convolutional[[:space:]]Attention[[:space:]]Decoding[[:space:]]for[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/32047b53-ccf5-41e9-bc53-4e712089f958_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EMOPortraits_[[:space:]]Emotion-enhanced[[:space:]]Multimodal[[:space:]]One-shot[[:space:]]Head[[:space:]]Avatars/ab68d1cd-73a3-4a4c-9896-716aeaaff436_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ERMVP_[[:space:]]Communication-Efficient[[:space:]]and[[:space:]]Collaboration-Robust[[:space:]]Multi-Vehicle[[:space:]]Perception[[:space:]]in[[:space:]]Challenging[[:space:]]Environments/197e9226-8724-4a9f-ac9a-1556459fa452_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ES3_[[:space:]]Evolving[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]of[[:space:]]Robust[[:space:]]Audio-Visual[[:space:]]Speech[[:space:]]Representations/09b7366b-807f-4dfb-a9c8-45edf9472928_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ESCAPE_[[:space:]]Encoding[[:space:]]Super-keypoints[[:space:]]for[[:space:]]Category-Agnostic[[:space:]]Pose[[:space:]]Estimation/64b3bbb9-9580-410a-adfe-c8a2bd5b8bca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ESR-NeRF_[[:space:]]Emissive[[:space:]]Source[[:space:]]Reconstruction[[:space:]]Using[[:space:]]LDR[[:space:]]Multi-view[[:space:]]Images/2c6d481f-7a23-410c-a62a-93255bbc656e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EVCap_[[:space:]]Retrieval-Augmented[[:space:]]Image[[:space:]]Captioning[[:space:]]with[[:space:]]External[[:space:]]Visual-Name[[:space:]]Memory[[:space:]]for[[:space:]]Open-World[[:space:]]Comprehension/003cc44f-e449-40f1-aec0-baf1c6235010_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EVS-assisted[[:space:]]Joint[[:space:]]Deblurring[[:space:]]Rolling-Shutter[[:space:]]Correction[[:space:]]and[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation[[:space:]]through[[:space:]]Sensor[[:space:]]Inverse[[:space:]]Modeling/2babbe27-3090-44dc-8f9d-2d16b50cc5ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Each[[:space:]]Test[[:space:]]Image[[:space:]]Deserves[[:space:]]A[[:space:]]Specific[[:space:]]Prompt_[[:space:]]Continual[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]2D[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/b547c627-f0bf-4b29-9f5d-10851e261fb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EarthLoc_[[:space:]]Astronaut[[:space:]]Photography[[:space:]]Localization[[:space:]]by[[:space:]]Indexing[[:space:]]Earth[[:space:]]from[[:space:]]Space/09848c3c-ff6b-4e53-9236-ab0f624d8495_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EasyDrag_[[:space:]]Efficient[[:space:]]Point-based[[:space:]]Manipulation[[:space:]]on[[:space:]]Diffusion[[:space:]]Models/3c97514b-8a67-4bdc-8795-3e7cfc02ceca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Eclipse_[[:space:]]Disambiguating[[:space:]]Illumination[[:space:]]and[[:space:]]Materials[[:space:]]using[[:space:]]Unintended[[:space:]]Shadows/a9a99e7b-5bee-4200-8036-13e74c55e36e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Edge-Aware[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation[[:space:]]Network[[:space:]]with[[:space:]]Intelligent[[:space:]]Semantic[[:space:]]Prior/ef85b90b-3f1b-4322-ac5a-c5d6ae60802c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Edit[[:space:]]One[[:space:]]for[[:space:]]All_[[:space:]]Interactive[[:space:]]Batch[[:space:]]Image[[:space:]]Editing/d0962876-52e1-4fef-ac92-9ddbf4fa5230_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EditGuard_[[:space:]]Versatile[[:space:]]Image[[:space:]]Watermarking[[:space:]]for[[:space:]]Tamper[[:space:]]Localization[[:space:]]and[[:space:]]Copyright[[:space:]]Protection/43cb9c2b-dd1a-4670-9150-d4345483687e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Editable[[:space:]]Scene[[:space:]]Simulation[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving[[:space:]]via[[:space:]]Collaborative[[:space:]]LLM-Agents/49eb43d9-92f1-4849-a7c5-3da61380ffc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Effective[[:space:]]Video[[:space:]]Mirror[[:space:]]Detection[[:space:]]with[[:space:]]Inconsistent[[:space:]]Motion[[:space:]]Cues/423d0156-aefb-4c1d-9839-9825d419ff81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]3D[[:space:]]Implicit[[:space:]]Head[[:space:]]Avatar[[:space:]]with[[:space:]]Mesh-anchored[[:space:]]Hash[[:space:]]Table[[:space:]]Blendshapes/afaaa8c5-698e-4fcb-9bde-21aa2469de9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Dataset[[:space:]]Distillation[[:space:]]via[[:space:]]Minimax[[:space:]]Diffusion/75146b44-b76a-42b0-8456-4c37446317b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Deformable[[:space:]]ConvNets_[[:space:]]Rethinking[[:space:]]Dynamic[[:space:]]and[[:space:]]Sparse[[:space:]]Operator[[:space:]]for[[:space:]]Vision[[:space:]]Applications/d839f1b7-f6aa-4faa-bb5a-7c661a9aa4e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Detection[[:space:]]of[[:space:]]Long[[:space:]]Consistent[[:space:]]Cycles[[:space:]]and[[:space:]]its[[:space:]]Application[[:space:]]to[[:space:]]Distributed[[:space:]]Synchronization/c0dd6b15-09a8-4eba-bcfb-24b8083dd375_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Hyperparameter[[:space:]]Optimization[[:space:]]with[[:space:]]Adaptive[[:space:]]Fidelity[[:space:]]Identification/fa155004-89c9-4599-9ab7-9a5d6edc5fbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]LoFTR_[[:space:]]Semi-Dense[[:space:]]Local[[:space:]]Feature[[:space:]]Matching[[:space:]]with[[:space:]]Sparse-Like[[:space:]]Speed/14a55345-a1a3-4ac4-a5a2-7c153c54b832_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Meshflow[[:space:]]and[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation[[:space:]]from[[:space:]]Event[[:space:]]Cameras/141a2ee1-d1ef-416d-8c61-bc462b1074b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Model[[:space:]]Stealing[[:space:]]Defense[[:space:]]with[[:space:]]Noise[[:space:]]Transition[[:space:]]Matrix/5181a0c8-f9af-47c9-b829-4e394817b1e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Multi-scale[[:space:]]Network[[:space:]]with[[:space:]]Learnable[[:space:]]Discrete[[:space:]]Wavelet[[:space:]]Transform[[:space:]]for[[:space:]]Blind[[:space:]]Motion[[:space:]]Deblurring/4af0ea1e-f85d-4e0b-ab29-04d268e0386c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Multitask[[:space:]]Dense[[:space:]]Predictor[[:space:]]via[[:space:]]Binarization/b91bc222-326c-4762-aabe-433ec8965e13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Privacy-Preserving[[:space:]]Visual[[:space:]]Localization[[:space:]]Using[[:space:]]3D[[:space:]]Ray[[:space:]]Clouds/61345f4c-613d-4e27-91d1-09fcd6a52a24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Scene[[:space:]]Recovery[[:space:]]Using[[:space:]]Luminous[[:space:]]Flux[[:space:]]Prior/0327d2f4-6c3a-4e74-a5a8-041807da336a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Solution[[:space:]]of[[:space:]]Point-Line[[:space:]]Absolute[[:space:]]Pose/8561d28f-e735-46cc-9618-c6c206bdafc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Stitchable[[:space:]]Task[[:space:]]Adaptation/03126c3f-71ba-4532-921c-1c95d792e515_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models/f86a8c24-7089-4003-88ca-fa1a5c70fe2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]Vision-Language[[:space:]]Pre-training[[:space:]]by[[:space:]]Cluster[[:space:]]Masking/24f9f648-410c-4f99-8615-dc6d0380efad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficient[[:space:]]and[[:space:]]Effective[[:space:]]Weakly-Supervised[[:space:]]Action[[:space:]]Segmentation[[:space:]]via[[:space:]]Action-Transition-Aware[[:space:]]Boundary[[:space:]]Alignment/1ae3fdc0-77c3-4dc9-87e2-e3b0568d2efa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EfficientDreamer_[[:space:]]High-Fidelity[[:space:]]and[[:space:]]Robust[[:space:]]3D[[:space:]]Creation[[:space:]]via[[:space:]]Orthogonal-view[[:space:]]Diffusion[[:space:]]Priors/6a261e68-4855-488d-9a23-298bfd3be3c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EfficientSAM_[[:space:]]Leveraged[[:space:]]Masked[[:space:]]Image[[:space:]]Pretraining[[:space:]]for[[:space:]]Efficient[[:space:]]Segment[[:space:]]Anything/f1e649e9-e5d4-48b3-8018-af1f58065ff2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Efficiently[[:space:]]Assemble[[:space:]]Normalization[[:space:]]Layers[[:space:]]and[[:space:]]Regularization[[:space:]]for[[:space:]]Federated[[:space:]]Domain[[:space:]]Generalization/16ce0297-575e-48b3-acb7-4a9aa0af4fe6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Ego-Exo4D_[[:space:]]Understanding[[:space:]]Skilled[[:space:]]Human[[:space:]]Activity[[:space:]]from[[:space:]]First-[[:space:]]and[[:space:]]Third-Person[[:space:]]Perspectives/742959dc-4498-434e-af9a-b1cc7f6e5bd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EgoExoLearn_[[:space:]]A[[:space:]]Dataset[[:space:]]for[[:space:]]Bridging[[:space:]]Asynchronous[[:space:]]Ego-[[:space:]]and[[:space:]]Exo-centric[[:space:]]View[[:space:]]of[[:space:]]Procedural[[:space:]]Activities[[:space:]]in[[:space:]]Real[[:space:]]World/134e2089-370f-4ad6-9686-84a1d7af0ed4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EgoGen_[[:space:]]An[[:space:]]Egocentric[[:space:]]Synthetic[[:space:]]Data[[:space:]]Generator/717ad579-fa9d-4ce5-996d-77c046c2cdbf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EgoThink_[[:space:]]Evaluating[[:space:]]First-Person[[:space:]]Perspective[[:space:]]Thinking[[:space:]]Capability[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models/2ff0d147-b907-475e-8aa3-5584b71f3014_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Egocentric[[:space:]]Whole-Body[[:space:]]Motion[[:space:]]Capture[[:space:]]with[[:space:]]FisheyeViT[[:space:]]and[[:space:]]Diffusion-Based[[:space:]]Motion[[:space:]]Refinement/90ece1d3-5ea4-4b56-9fe8-5a7d9a524eef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ElasticDiffusion_[[:space:]]Training-free[[:space:]]Arbitrary[[:space:]]Size[[:space:]]Image[[:space:]]Generation[[:space:]]through[[:space:]]Global-Local[[:space:]]Content[[:space:]]Separation/de40da54-b1b7-40fa-9aec-0f4f1bc4f406_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Elite360D_[[:space:]]Towards[[:space:]]Efficient[[:space:]]360[[:space:]]Depth[[:space:]]Estimation[[:space:]]via[[:space:]]Semantic-[[:space:]]and[[:space:]]Distance-Aware[[:space:]]Bi-Projection[[:space:]]Fusion/b933bca2-9ff4-42fc-9bd7-be64c39e1936_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Embodied[[:space:]]Multi-Modal[[:space:]]Agent[[:space:]]trained[[:space:]]by[[:space:]]an[[:space:]]LLM[[:space:]]from[[:space:]]a[[:space:]]Parallel[[:space:]]TextWorld/911fda92-445d-4ff5-8ef4-3ff0a7daf4bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EmbodiedScan_[[:space:]]A[[:space:]]Holistic[[:space:]]Multi-Modal[[:space:]]3D[[:space:]]Perception[[:space:]]Suite[[:space:]]Towards[[:space:]]Embodied[[:space:]]AI/ab5e0a2a-2a5f-41d4-a08d-01fd4564b83b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Embracing[[:space:]]Unimodal[[:space:]]Aleatoric[[:space:]]Uncertainty[[:space:]]for[[:space:]]Robust[[:space:]]Multimodal[[:space:]]Fusion/b864c4f0-5e7f-441f-a2a0-028202223b38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Emergent[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]from[[:space:]]Off-the-shelf[[:space:]]Vision-Language[[:space:]]Models/bf254b56-9ba7-4039-9f52-56365463bb00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EmoGen_[[:space:]]Emotional[[:space:]]Image[[:space:]]Content[[:space:]]Generation[[:space:]]with[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/a0b8035c-47d5-4f15-8dd7-134ed07d94b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EmoVIT_[[:space:]]Revolutionizing[[:space:]]Emotion[[:space:]]Insights[[:space:]]with[[:space:]]Visual[[:space:]]Instruction[[:space:]]Tuning/337f29a8-74df-4960-89a4-6eff2797f1cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Emotional[[:space:]]Speech-driven[[:space:]]3D[[:space:]]Body[[:space:]]Animation[[:space:]]via[[:space:]]Disentangled[[:space:]]Latent[[:space:]]Diffusion/8fe0dd5b-5118-4e0a-ac09-f70da9816cdc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Empowering[[:space:]]Resampling[[:space:]]Operation[[:space:]]for[[:space:]]Ultra-High-Definition[[:space:]]Image[[:space:]]Enhancement[[:space:]]with[[:space:]]Model-Aware[[:space:]]Guidance/5e31bf7f-5e48-4030-90cc-b526f724e932_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Emu[[:space:]]Edit_[[:space:]]Precise[[:space:]]Image[[:space:]]Editing[[:space:]]via[[:space:]]Recognition[[:space:]]and[[:space:]]Generation[[:space:]]Tasks/b6371d78-25ca-4b11-84b1-a566ed9d62f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/En3D_[[:space:]]An[[:space:]]Enhanced[[:space:]]Generative[[:space:]]Model[[:space:]]for[[:space:]]Sculpting[[:space:]]3D[[:space:]]Humans[[:space:]]from[[:space:]]2D[[:space:]]Synthetic[[:space:]]Data/cb7c8896-3bc3-4bc4-a3cb-b2aedebff579_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/End-to-End[[:space:]]Spatio-Temporal[[:space:]]Action[[:space:]]Localisation[[:space:]]with[[:space:]]Video[[:space:]]Transformers/62e0dda5-0752-47db-aad5-cb1b83945b43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/End-to-End[[:space:]]Temporal[[:space:]]Action[[:space:]]Detection[[:space:]]with[[:space:]]1B[[:space:]]Parameters[[:space:]]Across[[:space:]]1000[[:space:]]Frames/aed9af76-c260-41b9-be21-a2e11ea6ab0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Endow[[:space:]]SAM[[:space:]]with[[:space:]]Keen[[:space:]]Eyes_[[:space:]]Temporal-spatial[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Video[[:space:]]Camouflaged[[:space:]]Object[[:space:]]Detection/e8cb1a37-360c-44d3-92de-d83afed15e31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhance[[:space:]]Image[[:space:]]Classification[[:space:]]via[[:space:]]Inter-Class[[:space:]]Image[[:space:]]Mixup[[:space:]]with[[:space:]]Diffusion[[:space:]]Model/d24eac86-8eb2-4146-81d8-2b9455b8e75f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhanced[[:space:]]Motion-Text[[:space:]]Alignment[[:space:]]for[[:space:]]Image-to-Video[[:space:]]Transfer[[:space:]]Learning/1b431374-7f10-4d10-9e33-68850035e408_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhancing[[:space:]]3D[[:space:]]Fidelity[[:space:]]of[[:space:]]Text-to-3D[[:space:]]using[[:space:]]Cross-View[[:space:]]Correspondences/88cbf137-ede4-46fe-8d77-a1c81c7f200e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhancing[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]2D[[:space:]]Detection-Guided[[:space:]]Query[[:space:]]Anchors/268a1773-8b82-479c-85e1-46d912320b66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhancing[[:space:]]Intrinsic[[:space:]]Features[[:space:]]for[[:space:]]Debiasing[[:space:]]via[[:space:]]Investigating[[:space:]]Class-Discerning[[:space:]]Common[[:space:]]Attributes[[:space:]]in[[:space:]]Bias-Contrastive[[:space:]]Pair/1dd44ec2-211e-443c-8085-75dc4cb05104_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhancing[[:space:]]Multimodal[[:space:]]Cooperation[[:space:]]via[[:space:]]Sample-level[[:space:]]Modality[[:space:]]Valuation/8b972527-54af-428a-86fe-7057ea0431d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhancing[[:space:]]Post-training[[:space:]]Quantization[[:space:]]Calibration[[:space:]]through[[:space:]]Contrastive[[:space:]]Learning/15cf0763-9136-4a22-a42d-58fa5f702e3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhancing[[:space:]]Quality[[:space:]]of[[:space:]]Compressed[[:space:]]Images[[:space:]]by[[:space:]]Mitigating[[:space:]]Enhancement[[:space:]]Bias[[:space:]]Towards[[:space:]]Compression[[:space:]]Domain/ad9fabfe-decf-42b2-a14d-11e9dadbee5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhancing[[:space:]]Video[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Implicit[[:space:]]Resampling-based[[:space:]]Alignment/239000b0-f1bc-4dcf-98a2-d77dc17d1f38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhancing[[:space:]]Vision-Language[[:space:]]Pre-training[[:space:]]with[[:space:]]Rich[[:space:]]Supervisions/4041375d-c564-4256-87ca-292bf5ad7174_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhancing[[:space:]]Visual[[:space:]]Continual[[:space:]]Learning[[:space:]]with[[:space:]]Language-Guided[[:space:]]Supervision/9689a5ac-d80c-496b-bde8-9585711aa8d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhancing[[:space:]]Visual[[:space:]]Document[[:space:]]Understanding[[:space:]]with[[:space:]]Contrastive[[:space:]]Learning[[:space:]]in[[:space:]]Large[[:space:]]Visual-Language[[:space:]]Models/2b6fe286-8569-46aa-ae08-d7968e9557e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Enhancing[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]OOD[[:space:]]Detection[[:space:]]via[[:space:]]Sample-Aware[[:space:]]Model[[:space:]]Selection/fea43db8-f35a-4040-ab41-ff34748546df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Ensemble[[:space:]]Diversity[[:space:]]Facilitates[[:space:]]Adversarial[[:space:]]Transferability/549675ea-3d4d-49aa-bd5f-53847d4faf8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Entangled[[:space:]]View-Epipolar[[:space:]]Information[[:space:]]Aggregation[[:space:]]for[[:space:]]Generalizable[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/5673c47f-7284-4f19-adc1-03f8804ca62d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Entity-NeRF_[[:space:]]Detecting[[:space:]]and[[:space:]]Removing[[:space:]]Moving[[:space:]]Entities[[:space:]]in[[:space:]]Urban[[:space:]]Scenes/e0674e7f-b495-4e35-b2ec-16ed1fed89b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EpiDiff_[[:space:]]Enhancing[[:space:]]Multi-View[[:space:]]Synthesis[[:space:]]via[[:space:]]Localized[[:space:]]Epipolar-Constrained[[:space:]]Diffusion/281936a8-16a2-43d8-b95d-b8611363c42b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Epistemic[[:space:]]Uncertainty[[:space:]]Quantification[[:space:]]For[[:space:]]Pre-Trained[[:space:]]Neural[[:space:]]Networks/e620e7a3-0bfa-4d8e-9482-4cf6f739ff6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Equivariant[[:space:]]Multi-Modality[[:space:]]Image[[:space:]]Fusion/f6eac607-6ac3-4b13-8b48-cea53380492b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Equivariant[[:space:]]Plug-and-Play[[:space:]]Image[[:space:]]Reconstruction/d5e7673e-6b3a-455f-92cc-9afd948c9915_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Error[[:space:]]Detection[[:space:]]in[[:space:]]Egocentric[[:space:]]Procedural[[:space:]]Task[[:space:]]Videos/273ba7e7-7a77-4f69-abb2-2ed1f5263964_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EscherNet_[[:space:]]A[[:space:]]Generative[[:space:]]Model[[:space:]]for[[:space:]]Scalable[[:space:]]View[[:space:]]Synthesis/64a46bfd-45ec-40dc-b985-14f2c2ca4528_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Estimating[[:space:]]Extreme[[:space:]]3D[[:space:]]Image[[:space:]]Rotations[[:space:]]using[[:space:]]Cascaded[[:space:]]Attention/75aa5f98-2245-4e74-9d5b-bf3192ead917_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Estimating[[:space:]]Noisy[[:space:]]Class[[:space:]]Posterior[[:space:]]with[[:space:]]Part-level[[:space:]]Labels[[:space:]]for[[:space:]]Noisy[[:space:]]Label[[:space:]]Learning/e9b72fe4-eef5-4e13-99dc-8d499a99b190_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EvDiG_[[:space:]]Event-guided[[:space:]]Direct[[:space:]]and[[:space:]]Global[[:space:]]Components[[:space:]]Separation/1708f432-207d-46c5-b08d-b83a0c2a3d65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EvalCrafter_[[:space:]]Benchmarking[[:space:]]and[[:space:]]Evaluating[[:space:]]Large[[:space:]]Video[[:space:]]Generation[[:space:]]Models/1fc4bf1e-ef59-4bcd-b687-b0a88a0d1514_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Evaluating[[:space:]]Transferability[[:space:]]in[[:space:]]Retrieval[[:space:]]Tasks_[[:space:]]An[[:space:]]Approach[[:space:]]Using[[:space:]]MMD[[:space:]]and[[:space:]]Kernel[[:space:]]Methods/02e0decb-ace4-47e1-8366-263b96f8472e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Event[[:space:]]Stream-based[[:space:]]Visual[[:space:]]Object[[:space:]]Tracking_[[:space:]]A[[:space:]]High-Resolution[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]and[[:space:]]A[[:space:]]Novel[[:space:]]Baseline/1852b65e-0321-481c-9a5e-4b2efc7d28f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Event-assisted[[:space:]]Low-Light[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/7bef663d-c883-4594-a9a0-2a4535fa8542_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Event-based[[:space:]]Structure-from-Orbit/8e27958a-d27b-4343-84c7-97aa4c0523f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Event-based[[:space:]]Visible[[:space:]]and[[:space:]]Infrared[[:space:]]Fusion[[:space:]]via[[:space:]]Multi-task[[:space:]]Collaboration/48044879-0495-4a08-9d94-acc4d38cb667_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EventDance_[[:space:]]Unsupervised[[:space:]]Source-free[[:space:]]Cross-modal[[:space:]]Adaptation[[:space:]]for[[:space:]]Event-based[[:space:]]Object[[:space:]]Recognition/683bab1b-c15d-49ab-9f3f-42a572f008dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EventEgo3D_[[:space:]]3D[[:space:]]Human[[:space:]]Motion[[:space:]]Capture[[:space:]]from[[:space:]]Egocentric[[:space:]]Event[[:space:]]Streams/4e57e97f-744b-4a6e-a517-a2e4349a0c48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/EventPS_[[:space:]]Real-Time[[:space:]]Photometric[[:space:]]Stereo[[:space:]]Using[[:space:]]an[[:space:]]Event[[:space:]]Camera/779326a6-3300-420f-a747-5c4e2a953ced_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Evidential[[:space:]]Active[[:space:]]Recognition_[[:space:]]Intelligent[[:space:]]and[[:space:]]Prudent[[:space:]]Open-World[[:space:]]Embodied[[:space:]]Perception/724b8513-72a9-4583-9e30-d919d3bd2d37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ExACT_[[:space:]]Language-guided[[:space:]]Conceptual[[:space:]]Reasoning[[:space:]]and[[:space:]]Uncertainty[[:space:]]Estimation[[:space:]]for[[:space:]]Event-based[[:space:]]Action[[:space:]]Recognition[[:space:]]and[[:space:]]More/8741c2c6-24a7-42e2-9280-f24194b801ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ExMap_[[:space:]]Leveraging[[:space:]]Explainability[[:space:]]Heatmaps[[:space:]]for[[:space:]]Unsupervised[[:space:]]Group[[:space:]]Robustness[[:space:]]to[[:space:]]Spurious[[:space:]]Correlations/16dc7660-6733-46c5-a74f-f97238ace615_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Exact[[:space:]]Fusion[[:space:]]via[[:space:]]Feature[[:space:]]Distribution[[:space:]]Matching[[:space:]]for[[:space:]]Few-shot[[:space:]]Image[[:space:]]Generation/383032d6-9ddc-491c-a6f7-9fee8092aecd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Expandable[[:space:]]Subspace[[:space:]]Ensemble[[:space:]]for[[:space:]]Pre-Trained[[:space:]]Model-Based[[:space:]]Class-Incremental[[:space:]]Learning/15a73ef8-3b89-412a-9282-8bb56cde17cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Explaining[[:space:]]CLIP's[[:space:]]Performance[[:space:]]Disparities[[:space:]]on[[:space:]]Data[[:space:]]from[[:space:]]Blind_Low[[:space:]]Vision[[:space:]]Users/7ede4b0d-2f64-4c14-b446-168d9441109b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Explaining[[:space:]]the[[:space:]]Implicit[[:space:]]Neural[[:space:]]Canvas_[[:space:]]Connecting[[:space:]]Pixels[[:space:]]to[[:space:]]Neurons[[:space:]]by[[:space:]]Tracing[[:space:]]their[[:space:]]Contributions/d351a8b1-c795-407a-8cb5-b4e8a0babb71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Exploiting[[:space:]]Diffusion[[:space:]]Prior[[:space:]]for[[:space:]]Generalizable[[:space:]]Dense[[:space:]]Prediction/3c6a67ef-491d-4ea1-a0c2-625f214ed8ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Exploiting[[:space:]]Inter-sample[[:space:]]and[[:space:]]Inter-feature[[:space:]]Relations[[:space:]]in[[:space:]]Dataset[[:space:]]Distillation/3bbffd2b-9cbd-48c1-8428-d848b7dce88d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Exploiting[[:space:]]Style[[:space:]]Latent[[:space:]]Flows[[:space:]]for[[:space:]]Generalizing[[:space:]]Deepfake[[:space:]]Video[[:space:]]Detection/212e9436-5799-47a5-bb8b-48d19d0e74c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Exploring[[:space:]]Efficient[[:space:]]Asymmetric[[:space:]]Blind-Spots[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Denoising[[:space:]]in[[:space:]]Real-World[[:space:]]Scenarios/6f5a6702-1aca-4573-ba13-8e95c098b031_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Exploring[[:space:]]Orthogonality[[:space:]]in[[:space:]]Open[[:space:]]World[[:space:]]Object[[:space:]]Detection/45ae888d-ee36-4818-b36c-bef9017fdbb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Exploring[[:space:]]Pose-Aware[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]via[[:space:]]Hybrid[[:space:]]Learning/b23b5922-86cc-44d9-865e-b91e916a21a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Exploring[[:space:]]Region-Word[[:space:]]Alignment[[:space:]]in[[:space:]]Built-in[[:space:]]Detector[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detection/55f5d9cd-1bf5-4850-bf9e-c1400753753c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Exploring[[:space:]]Regional[[:space:]]Clues[[:space:]]in[[:space:]]CLIP[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Semantic[[:space:]]Segmentation/53402281-3036-4667-91cf-0ec1693c77fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Exploring[[:space:]]Vision[[:space:]]Transformers[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Motion-Language[[:space:]]Models[[:space:]]with[[:space:]]Motion[[:space:]]Patches/1d5ebf27-65c9-4621-8012-13ef51d13bbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Exploring[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Large[[:space:]]Foundation[[:space:]]Models[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]HOI[[:space:]]Detection/d5547767-c641-4aae-bbe2-2c6592b09b8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Exploring[[:space:]]the[[:space:]]Transferability[[:space:]]of[[:space:]]Visual[[:space:]]Prompting[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/43ecb574-1499-4004-92fa-f8b90c2d73a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ExtDM_[[:space:]]Distribution[[:space:]]Extrapolation[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Video[[:space:]]Prediction/f0582e79-2b5d-449e-8d73-8be7642dcfa5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Extend[[:space:]]Your[[:space:]]Own[[:space:]]Correspondences_[[:space:]]Unsupervised[[:space:]]Distant[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration[[:space:]]by[[:space:]]Progressive[[:space:]]Distance[[:space:]]Extension/d3551fc9-12cc-46fe-9e42-c63b5a92858d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ExtraNeRF_[[:space:]]Visibility-Aware[[:space:]]View[[:space:]]Extrapolation[[:space:]]of[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/87e4f6c8-6575-4898-b7b4-3736b35dd355_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Extreme[[:space:]]Point[[:space:]]Supervised[[:space:]]Instance[[:space:]]Segmentation/5f4193a5-ed43-4768-ae66-02b838445b8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Eyes[[:space:]]Wide[[:space:]]Shut_[[:space:]]Exploring[[:space:]]the[[:space:]]Visual[[:space:]]Shortcomings[[:space:]]of[[:space:]]Multimodal[[:space:]]LLMs/f0ff9eb4-73e2-40fd-9d59-b530115cfb51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/F3Loc_[[:space:]]Fusion[[:space:]]and[[:space:]]Filtering[[:space:]]for[[:space:]]Floorplan[[:space:]]Localization/c7538241-d569-4598-88f7-92c84ee9a991_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FACT_[[:space:]]Frame-Action[[:space:]]Cross-Attention[[:space:]]Temporal[[:space:]]Modeling[[:space:]]for[[:space:]]Efficient[[:space:]]Action[[:space:]]Segmentation/c523fbcc-3b00-45db-8d2c-4deb1c98b1ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FADES_[[:space:]]Fair[[:space:]]Disentanglement[[:space:]]with[[:space:]]Sensitive[[:space:]]Relevance/4fc7b521-a35a-4fd1-86b0-61ba8750ba8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FAR_[[:space:]]Flexible[[:space:]]Accurate[[:space:]]and[[:space:]]Robust[[:space:]]6DoF[[:space:]]Relative[[:space:]]Camera[[:space:]]Pose[[:space:]]Estimation/9d58ef90-4dd8-493c-a893-df207e69e8b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FC-GNN_[[:space:]]Recovering[[:space:]]Reliable[[:space:]]and[[:space:]]Accurate[[:space:]]Correspondences[[:space:]]from[[:space:]]Interferences/ee5c3e0e-661f-4ca8-87cd-af6419db01e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FCS_[[:space:]]Feature[[:space:]]Calibration[[:space:]]and[[:space:]]Separation[[:space:]]for[[:space:]]Non-Exemplar[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning/1b89a2b8-b8c6-4063-a490-d53a4c530f86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FFF_[[:space:]]Fixing[[:space:]]Flawed[[:space:]]Foundations[[:space:]]in[[:space:]]Contrastive[[:space:]]Pre-Training[[:space:]]Results[[:space:]]in[[:space:]]Very[[:space:]]Strong[[:space:]]Vision-Language[[:space:]]Models/7c7a2192-e85b-4937-b25a-4800654db47b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FINER_[[:space:]]Flexible[[:space:]]Spectral-bias[[:space:]]Tuning[[:space:]]in[[:space:]]Implicit[[:space:]]NEural[[:space:]]Representation[[:space:]]by[[:space:]]Variable-periodic[[:space:]]Activation[[:space:]]Functions/9fa89bd5-b080-412b-9d0c-b0241716eed1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FISBe_[[:space:]]A[[:space:]]Real-World[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]for[[:space:]]Instance[[:space:]]Segmentation[[:space:]]of[[:space:]]Long-Range[[:space:]]Thin[[:space:]]Filamentous[[:space:]]Structures/04598530-3c8a-4c03-845f-c98cf6c19c1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FLHetBench_[[:space:]]Benchmarking[[:space:]]Device[[:space:]]and[[:space:]]State[[:space:]]Heterogeneity[[:space:]]in[[:space:]]Federated[[:space:]]Learning/934e0f78-7c23-4c0c-8a86-bfd62f29a921_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FMA-Net_[[:space:]]Flow-Guided[[:space:]]Dynamic[[:space:]]Filtering[[:space:]]and[[:space:]]Iterative[[:space:]]Feature[[:space:]]Refinement[[:space:]]with[[:space:]]Multi-Attention[[:space:]]for[[:space:]]Joint[[:space:]]Video[[:space:]]Super-Resolution[[:space:]]and[[:space:]]Deblurring/3d4e529e-cd32-4756-bd87-c1527d71f50b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FREE_[[:space:]]Faster[[:space:]]and[[:space:]]Better[[:space:]]Data-Free[[:space:]]Meta-Learning/1c74194c-f473-44c9-9c3d-0da3115ced97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FRESCO_[[:space:]]Spatial-Temporal[[:space:]]Correspondence[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Video[[:space:]]Translation/3efa9392-f500-4533-b815-75f8557d8c39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FSC_[[:space:]]Few-point[[:space:]]Shape[[:space:]]Completion/7b0ec71c-2515-4313-99e2-a97d62caaf47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FSRT_[[:space:]]Facial[[:space:]]Scene[[:space:]]Representation[[:space:]]Transformer[[:space:]]for[[:space:]]Face[[:space:]]Reenactment[[:space:]]from[[:space:]]Factorized[[:space:]]Appearance[[:space:]]Head-pose[[:space:]]and[[:space:]]Facial[[:space:]]Expression[[:space:]]Features/1b901cfd-c0e2-4cd9-8da2-1a3ba74a4458_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Face2Diffusion[[:space:]]for[[:space:]]Fast[[:space:]]and[[:space:]]Editable[[:space:]]Face[[:space:]]Personalization/f632820f-d711-4217-a336-87b76631633b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FaceChain-ImagineID_[[:space:]]Freely[[:space:]]Crafting[[:space:]]High-Fidelity[[:space:]]Diverse[[:space:]]Talking[[:space:]]Faces[[:space:]]from[[:space:]]Disentangled[[:space:]]Audio/7816997c-f2e0-4c89-bb81-a5eb1da3b37c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FaceChain-SuDe_[[:space:]]Building[[:space:]]Derived[[:space:]]Class[[:space:]]to[[:space:]]Inherit[[:space:]]Category[[:space:]]Attributes[[:space:]]for[[:space:]]One-shot[[:space:]]Subject-Driven[[:space:]]Generation/c3b455ce-f254-4de1-9c75-9d79469b0fd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FaceCom_[[:space:]]Towards[[:space:]]High-fidelity[[:space:]]3D[[:space:]]Facial[[:space:]]Shape[[:space:]]Completion[[:space:]]via[[:space:]]Optimization[[:space:]]and[[:space:]]Inpainting[[:space:]]Guidance/72c222b9-0fba-4d5a-82c7-6366b6edf854_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FaceLift_[[:space:]]Semi-supervised[[:space:]]3D[[:space:]]Facial[[:space:]]Landmark[[:space:]]Localization/829d33a1-bbc5-411a-a5e5-e7b2cfae4dcb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FaceTalk_[[:space:]]Audio-Driven[[:space:]]Motion[[:space:]]Diffusion[[:space:]]for[[:space:]]Neural[[:space:]]Parametric[[:space:]]Head[[:space:]]Models/d00fd6c9-d9fa-4eb5-b3b6-658635d33044_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Faces[[:space:]]that[[:space:]]Speak_[[:space:]]Jointly[[:space:]]Synthesising[[:space:]]Talking[[:space:]]Face[[:space:]]and[[:space:]]Speech[[:space:]]from[[:space:]]Text/17c5f08d-4f2b-43a8-ac8a-a2993bf46384_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Facial[[:space:]]Identity[[:space:]]Anonymization[[:space:]]via[[:space:]]Intrinsic[[:space:]]and[[:space:]]Extrinsic[[:space:]]Attention[[:space:]]Distraction/406cc9c1-aaf8-4532-9d92-4a88c61a3031_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fair[[:space:]]Federated[[:space:]]Learning[[:space:]]under[[:space:]]Domain[[:space:]]Skew[[:space:]]with[[:space:]]Local[[:space:]]Consistency[[:space:]]and[[:space:]]Domain[[:space:]]Diversity/17bd0235-517f-480a-9268-75e0428f026d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fair-VPT_[[:space:]]Fair[[:space:]]Visual[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Image[[:space:]]Classification/8d86d210-2ac0-48a9-a8c4-a6efde50a6cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FairCLIP_[[:space:]]Harnessing[[:space:]]Fairness[[:space:]]in[[:space:]]Vision-Language[[:space:]]Learning/324be6f5-f5ef-4628-962e-3be337cbf9ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FairDeDup_[[:space:]]Detecting[[:space:]]and[[:space:]]Mitigating[[:space:]]Vision-Language[[:space:]]Fairness[[:space:]]Disparities[[:space:]]in[[:space:]]Semantic[[:space:]]Dataset[[:space:]]Deduplication/e1b314ff-cd9d-40b3-b595-4cd707f552b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FairRAG_[[:space:]]Fair[[:space:]]Human[[:space:]]Generation[[:space:]]via[[:space:]]Fair[[:space:]]Retrieval[[:space:]]Augmentation/6ce903f1-f7e0-498d-bf3e-524402498ce9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fairy_[[:space:]]Fast[[:space:]]Parallelized[[:space:]]Instruction-Guided[[:space:]]Video-to-Video[[:space:]]Synthesis/ddbd53b7-486d-46ac-8c44-a69701df2a35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FakeInversion_[[:space:]]Learning[[:space:]]to[[:space:]]Detect[[:space:]]Images[[:space:]]from[[:space:]]Unseen[[:space:]]Text-to-Image[[:space:]]Models[[:space:]]by[[:space:]]Inverting[[:space:]]Stable[[:space:]]Diffusion/0be1a18b-e360-404f-8cdf-e3976467595b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fantastic[[:space:]]Animals[[:space:]]and[[:space:]]Where[[:space:]]to[[:space:]]Find[[:space:]]Them_[[:space:]]Segment[[:space:]]Any[[:space:]]Marine[[:space:]]Animal[[:space:]]with[[:space:]]Dual[[:space:]]SAM/2296cdc8-3cd4-48b2-bf98-71b35a1cd60a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fast[[:space:]]Adaptation[[:space:]]for[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Meta-Optimization/2a31ff28-04b4-4651-82c1-3ef1ea8b5d75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fast[[:space:]]ODE-based[[:space:]]Sampling[[:space:]]for[[:space:]]Diffusion[[:space:]]Models[[:space:]]in[[:space:]]Around[[:space:]]5[[:space:]]Steps/e1f0787b-c415-47ab-b97b-c9766a721372_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FastMAC_[[:space:]]Stochastic[[:space:]]Spectral[[:space:]]Sampling[[:space:]]of[[:space:]]Correspondence[[:space:]]Graph/ccae9a7e-53c3-4d53-be49-5449bcfa1596_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Feature[[:space:]]3DGS_[[:space:]]Supercharging[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]to[[:space:]]Enable[[:space:]]Distilled[[:space:]]Feature[[:space:]]Fields/f7c3270c-4ce6-427e-ad79-284cbc77f870_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Feature[[:space:]]Re-Embedding_[[:space:]]Towards[[:space:]]Foundation[[:space:]]Model-Level[[:space:]]Performance[[:space:]]in[[:space:]]Computational[[:space:]]Pathology/d21f2e35-be84-4426-8136-022aa40ea111_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FedAS_[[:space:]]Bridging[[:space:]]Inconsistency[[:space:]]in[[:space:]]Personalized[[:space:]]Federated[[:space:]]Learning/9b1bb845-c367-4772-a2d8-60d175068d90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FedHCA2_[[:space:]]Towards[[:space:]]Hetero-Client[[:space:]]Federated[[:space:]]Multi-Task[[:space:]]Learning/c4928416-1c1d-47ef-b05b-f1109e5f927b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FedMef_[[:space:]]Towards[[:space:]]Memory-efficient[[:space:]]Federated[[:space:]]Dynamic[[:space:]]Pruning/c8aa173f-035b-4aad-bb11-535583ea2581_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FedSOL_[[:space:]]Stabilized[[:space:]]Orthogonal[[:space:]]Learning[[:space:]]with[[:space:]]Proximal[[:space:]]Restrictions[[:space:]]in[[:space:]]Federated[[:space:]]Learning/7015b4ca-ccfc-485b-b148-b92c527ab3c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FedSelect_[[:space:]]Personalized[[:space:]]Federated[[:space:]]Learning[[:space:]]with[[:space:]]Customized[[:space:]]Selection[[:space:]]of[[:space:]]Parameters[[:space:]]for[[:space:]]Fine-Tuning/f7b37c2a-fe8c-4d96-b416-e604d12f39cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FedUV_[[:space:]]Uniformity[[:space:]]and[[:space:]]Variance[[:space:]]for[[:space:]]Heterogeneous[[:space:]]Federated[[:space:]]Learning/d6a4d273-cdd3-41e0-86aa-ed0112dd1001_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Federated[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/f3fc064f-c328-458a-88a6-ad2dee21b2d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Federated[[:space:]]Online[[:space:]]Adaptation[[:space:]]for[[:space:]]Deep[[:space:]]Stereo/62df9f40-83ce-44e3-a33b-2d66717f19dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Feedback-Guided[[:space:]]Autonomous[[:space:]]Driving/0d986129-c627-4c85-9257-33f667ee2c8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Few-Shot[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Foundation[[:space:]]Models/faf33279-d6e0-47a3-9ec9-4715e89aa51b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Few-shot[[:space:]]Learner[[:space:]]Parameterization[[:space:]]by[[:space:]]Diffusion[[:space:]]Time-steps/a295ba58-fb2d-4ec1-ae5e-a895f5c6fd02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Finding[[:space:]]Lottery[[:space:]]Tickets[[:space:]]in[[:space:]]Vision[[:space:]]Models[[:space:]]via[[:space:]]Data-driven[[:space:]]Spectral[[:space:]]Foresight[[:space:]]Pruning/6df29a5c-967f-42c8-8704-c5a145a0e1ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fine-Grained[[:space:]]Bipartite[[:space:]]Concept[[:space:]]Factorization[[:space:]]for[[:space:]]Clustering/745a776e-b351-44af-bc79-1ad7f0217662_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fine-grained[[:space:]]Prototypical[[:space:]]Voting[[:space:]]with[[:space:]]Heterogeneous[[:space:]]Mixup[[:space:]]for[[:space:]]Semi-supervised[[:space:]]2D-3D[[:space:]]Cross-modal[[:space:]]Retrieval/900e6233-1d99-4df5-8f50-e8b39ec40929_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FinePOSE_[[:space:]]Fine-Grained[[:space:]]Prompt-Driven[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Diffusion[[:space:]]Models/0f3ef071-1760-47ad-8e0c-3cd5b6874507_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FineParser_[[:space:]]A[[:space:]]Fine-grained[[:space:]]Spatio-temporal[[:space:]]Action[[:space:]]Parser[[:space:]]for[[:space:]]Human-centric[[:space:]]Action[[:space:]]Quality[[:space:]]Assessment/aad00275-93c4-402a-9697-cc65df976a34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FineSports_[[:space:]]A[[:space:]]Multi-person[[:space:]]Hierarchical[[:space:]]Sports[[:space:]]Video[[:space:]]Dataset[[:space:]]for[[:space:]]Fine-grained[[:space:]]Action[[:space:]]Understanding/4ac4ce85-0ae4-489b-9937-4f3284230cef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Finsler-Laplace-Beltrami[[:space:]]Operators[[:space:]]with[[:space:]]Application[[:space:]]to[[:space:]]Shape[[:space:]]Analysis/fba21631-133f-4aca-ab9a-82fe7ef910bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fitting[[:space:]]Flats[[:space:]]to[[:space:]]Flats/7201f2e1-2e53-406e-a7b2-94b9ab19f582_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fixed[[:space:]]Point[[:space:]]Diffusion[[:space:]]Models/913a4950-3d02-446e-b8fc-2b778db371d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FlashAvatar_[[:space:]]High-fidelity[[:space:]]Head[[:space:]]Avatar[[:space:]]with[[:space:]]Efficient[[:space:]]Gaussian[[:space:]]Embedding/585d834d-e373-412d-b174-e53ff2f326a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FlashEval_[[:space:]]Towards[[:space:]]Fast[[:space:]]and[[:space:]]Accurate[[:space:]]Evaluation[[:space:]]of[[:space:]]Text-to-image[[:space:]]Diffusion[[:space:]]Generative[[:space:]]Models/52f11be6-e125-49fa-ad99-8e8f4c7dbba0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Flatten[[:space:]]Long-Range[[:space:]]Loss[[:space:]]Landscapes[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Few-Shot[[:space:]]Learning/8096018e-7ad6-4d2a-98c1-6123296a020c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Flattening[[:space:]]the[[:space:]]Parent[[:space:]]Bias_[[:space:]]Hierarchical[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]in[[:space:]]the[[:space:]]Poincare[[:space:]]Ball/aed233a0-3a0e-4c07-83e1-cce89f9a5575_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Flexible[[:space:]]Biometrics[[:space:]]Recognition_[[:space:]]Bridging[[:space:]]the[[:space:]]Multimodality[[:space:]]Gap[[:space:]]through[[:space:]]Attention[[:space:]]Alignment[[:space:]]and[[:space:]]Prompt[[:space:]]Tuning/b2f31bce-e1b9-45c3-9f08-4d8aaddc2451_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Flexible[[:space:]]Depth[[:space:]]Completion[[:space:]]for[[:space:]]Sparse[[:space:]]and[[:space:]]Varying[[:space:]]Point[[:space:]]Densities/47f63bcc-47bf-4199-aa6b-1b57a1bb20c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Florence-2_[[:space:]]Advancing[[:space:]]a[[:space:]]Unified[[:space:]]Representation[[:space:]]for[[:space:]]a[[:space:]]Variety[[:space:]]of[[:space:]]Vision[[:space:]]Tasks/293a6920-940d-42c0-9936-8fe855143e2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Flow-Guided[[:space:]]Online[[:space:]]Stereo[[:space:]]Rectification[[:space:]]for[[:space:]]Wide[[:space:]]Baseline[[:space:]]Stereo/bdc60839-e12b-4b1e-a407-8eb884cb30db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FlowDiffuser_[[:space:]]Advancing[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/2e45e764-5be4-4148-8bbf-124d92f4a3d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FlowIE_[[:space:]]Efficient[[:space:]]Image[[:space:]]Enhancement[[:space:]]via[[:space:]]Rectified[[:space:]]Flow/7abe66a8-d9bc-4089-915b-2f4310a63a8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FlowTrack_[[:space:]]Revisiting[[:space:]]Optical[[:space:]]Flow[[:space:]]for[[:space:]]Long-Range[[:space:]]Dense[[:space:]]Tracking/bb3bf568-9590-4030-a8ac-5d5c2ee20629_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FlowVQTalker_[[:space:]]High-Quality[[:space:]]Emotional[[:space:]]Talking[[:space:]]Face[[:space:]]Generation[[:space:]]through[[:space:]]Normalizing[[:space:]]Flow[[:space:]]and[[:space:]]Quantization/464775fd-b3d3-4d41-b272-16472210d70c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FlowVid_[[:space:]]Taming[[:space:]]Imperfect[[:space:]]Optical[[:space:]]Flows[[:space:]]for[[:space:]]Consistent[[:space:]]Video-to-Video[[:space:]]Synthesis/9117dec3-c84c-4173-95dd-9d0f0f6db320_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FlowerFormer_[[:space:]]Empowering[[:space:]]Neural[[:space:]]Architecture[[:space:]]Encoding[[:space:]]using[[:space:]]a[[:space:]]Flow-aware[[:space:]]Graph[[:space:]]Transformer/393542d8-ad13-4fd5-89e4-127b4eb2a5eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FocSAM_[[:space:]]Delving[[:space:]]Deeply[[:space:]]into[[:space:]]Focused[[:space:]]Objects[[:space:]]in[[:space:]]Segmenting[[:space:]]Anything/eedb2f1c-9606-4bbd-8547-60df221c2ddc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Focus[[:space:]]on[[:space:]]Hiders_[[:space:]]Exploring[[:space:]]Hidden[[:space:]]Threats[[:space:]]for[[:space:]]Enhancing[[:space:]]Adversarial[[:space:]]Training/853d1890-f8d4-4012-a429-d974c72f2364_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Focus[[:space:]]on[[:space:]]Your[[:space:]]Instruction_[[:space:]]Fine-grained[[:space:]]and[[:space:]]Multi-instruction[[:space:]]Image[[:space:]]Editing[[:space:]]by[[:space:]]Attention[[:space:]]Modulation/a4b8826f-be52-4da9-9206-84db185a3716_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FocusMAE_[[:space:]]Gallbladder[[:space:]]Cancer[[:space:]]Detection[[:space:]]from[[:space:]]Ultrasound[[:space:]]Videos[[:space:]]with[[:space:]]Focused[[:space:]]Masked[[:space:]]Autoencoders/d3e4f665-9d72-4916-b27a-d8d7ec970086_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fooling[[:space:]]Polarization-Based[[:space:]]Vision[[:space:]]using[[:space:]]Locally[[:space:]]Controllable[[:space:]]Polarizing[[:space:]]Projection/6f5214c5-f58b-4da0-941c-e006cb80bdb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Forecasting[[:space:]]of[[:space:]]3D[[:space:]]Whole-body[[:space:]]Human[[:space:]]Poses[[:space:]]with[[:space:]]Grasping[[:space:]]Objects/68d92d49-7815-40b3-a288-a75ba8152481_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Forgery-aware[[:space:]]Adaptive[[:space:]]Transformer[[:space:]]for[[:space:]]Generalizable[[:space:]]Synthetic[[:space:]]Image[[:space:]]Detection/6a233e50-7f88-4f87-a294-e0a34f1917af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FoundationPose_[[:space:]]Unified[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation[[:space:]]and[[:space:]]Tracking[[:space:]]of[[:space:]]Novel[[:space:]]Objects/570e504a-7237-437d-9d39-2854874f110f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fourier[[:space:]]Priors-Guided[[:space:]]Diffusion[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Joint[[:space:]]Low-Light[[:space:]]Enhancement[[:space:]]and[[:space:]]Deblurring/762cd2b6-5730-451f-bfe6-58f66f9e9cd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fourier-basis[[:space:]]Functions[[:space:]]to[[:space:]]Bridge[[:space:]]Augmentation[[:space:]]Gap_[[:space:]]Rethinking[[:space:]]Frequency[[:space:]]Augmentation[[:space:]]in[[:space:]]Image[[:space:]]Classification/0e2a91ae-b644-4df0-9ffc-953a784b25b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FreGS_[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]Progressive[[:space:]]Frequency[[:space:]]Regularization/24144024-e1aa-40ee-8c8f-78782f719ab2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Free3D_[[:space:]]Consistent[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]without[[:space:]]3D[[:space:]]Representation/0e560a0a-b2e4-4ade-a9da-8e8d7ed77f71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FreeControl_[[:space:]]Training-Free[[:space:]]Spatial[[:space:]]Control[[:space:]]of[[:space:]]Any[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Model[[:space:]]with[[:space:]]Any[[:space:]]Condition/faaa3c00-7fe7-447c-a8ba-e9572ca694ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FreeCustom_[[:space:]]Tuning-Free[[:space:]]Customized[[:space:]]Image[[:space:]]Generation[[:space:]]for[[:space:]]Multi-Concept[[:space:]]Composition/3c58fab1-6d23-4445-8740-355018745526_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FreeDrag_[[:space:]]Feature[[:space:]]Dragging[[:space:]]for[[:space:]]Reliable[[:space:]]Point-based[[:space:]]Image[[:space:]]Editing/6acbd8b9-7507-4c7d-9a8e-a653798d93b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FreeKD_[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]via[[:space:]]Semantic[[:space:]]Frequency[[:space:]]Prompt/f6215e21-7a18-42b1-902f-66826f469135_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FreeMan_[[:space:]]Towards[[:space:]]Benchmarking[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]under[[:space:]]Real-World[[:space:]]Conditions/9d4857fb-d215-4984-828c-375cc3836eb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FreePoint_[[:space:]]Unsupervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Instance[[:space:]]Segmentation/f87fa0a6-b1d7-45a3-adf8-7a0458cf79d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FreeU_[[:space:]]Free[[:space:]]Lunch[[:space:]]in[[:space:]]Diffusion[[:space:]]U-Net/39c7b46e-4e7e-48ae-89be-fb2b57bd95fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Frequency[[:space:]]Decoupling[[:space:]]for[[:space:]]Motion[[:space:]]Magnification[[:space:]]via[[:space:]]Multi-Level[[:space:]]Isomorphic[[:space:]]Architecture/81e5ab63-5bf5-4d82-a4a9-5970a0475a9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Frequency-Adaptive[[:space:]]Dilated[[:space:]]Convolution[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/8796d900-f53c-45ce-8d63-1eaa269be9b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Frequency-aware[[:space:]]Event-based[[:space:]]Video[[:space:]]Deblurring[[:space:]]for[[:space:]]Real-World[[:space:]]Motion[[:space:]]Blur/38a2be9d-fa72-476c-beb7-fc02cdcaaef7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Friendly[[:space:]]Sharpness-Aware[[:space:]]Minimization/41c5e64f-2759-4aa6-bfb7-e08490116ff6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/From[[:space:]]Activation[[:space:]]to[[:space:]]Initialization_[[:space:]]Scaling[[:space:]]Insights[[:space:]]for[[:space:]]Optimizing[[:space:]]Neural[[:space:]]Fields/e3bfe168-69aa-45e5-bc3f-2af60adbb5fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/From[[:space:]]Audio[[:space:]]to[[:space:]]Photoreal[[:space:]]Embodiment_[[:space:]]Synthesizing[[:space:]]Humans[[:space:]]in[[:space:]]Conversations/18467a11-fdca-43eb-8167-8f89d94559b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/From[[:space:]]Coarse[[:space:]]to[[:space:]]Fine-Grained[[:space:]]Open-Set[[:space:]]Recognition/f0f3bb89-a43b-4874-8692-c7b0a5e03025_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/From[[:space:]]Correspondences[[:space:]]to[[:space:]]Pose_[[:space:]]Non-minimal[[:space:]]Certifiably[[:space:]]Optimal[[:space:]]Relative[[:space:]]Pose[[:space:]]without[[:space:]]Disambiguation/062ecfd0-0ad9-42ce-bd67-d6b8dacaf7d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/From[[:space:]]Feature[[:space:]]to[[:space:]]Gaze_[[:space:]]A[[:space:]]Generalizable[[:space:]]Replacement[[:space:]]of[[:space:]]Linear[[:space:]]Layer[[:space:]]for[[:space:]]Gaze[[:space:]]Estimation/cda1547f-9b3b-4666-ade8-5f9490564243_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/From[[:space:]]Isolated[[:space:]]Islands[[:space:]]to[[:space:]]Pangea_[[:space:]]Unifying[[:space:]]Semantic[[:space:]]Space[[:space:]]for[[:space:]]Human[[:space:]]Action[[:space:]]Understanding/ea7e3d38-613b-484c-baa7-df86560fbe03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/From[[:space:]]Pixels[[:space:]]to[[:space:]]Graphs_[[:space:]]Open-Vocabulary[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]with[[:space:]]Vision-Language[[:space:]]Models/c3a62b99-15b1-4c20-9aff-452efe4f42c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/From[[:space:]]SAM[[:space:]]to[[:space:]]CAMs_[[:space:]]Exploring[[:space:]]Segment[[:space:]]Anything[[:space:]]Model[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/2fdf147c-23f7-4ebd-8aa3-54bf04fa03b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/From[[:space:]]Variance[[:space:]]to[[:space:]]Veracity_[[:space:]]Unbundling[[:space:]]and[[:space:]]Mitigating[[:space:]]Gradient[[:space:]]Variance[[:space:]]in[[:space:]]Differentiable[[:space:]]Bundle[[:space:]]Adjustment[[:space:]]Layers/e37e4f06-e2c9-4819-8374-0d2b054facdc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/From[[:space:]]a[[:space:]]Bird's[[:space:]]Eye[[:space:]]View[[:space:]]to[[:space:]]See_[[:space:]]Joint[[:space:]]Camera[[:space:]]and[[:space:]]Subject[[:space:]]Registration[[:space:]]without[[:space:]]the[[:space:]]Camera[[:space:]]Calibration/267ec1c3-deec-453d-b617-bad60d70f7c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/From-Ground-To-Objects_[[:space:]]Coarse-to-Fine[[:space:]]Self-supervised[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation[[:space:]]of[[:space:]]Dynamic[[:space:]]Objects[[:space:]]with[[:space:]]Ground[[:space:]]Contact[[:space:]]Prior/d7cf041f-f794-403f-8502-965bfd70d59b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Frozen[[:space:]]CLIP_[[:space:]]A[[:space:]]Strong[[:space:]]Backbone[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/b96cb7e0-37bb-41fc-8a8b-857712a4b71d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Frozen[[:space:]]Feature[[:space:]]Augmentation[[:space:]]for[[:space:]]Few-Shot[[:space:]]Image[[:space:]]Classification/eeca3f4b-8f36-41c3-9116-0f8de03fd671_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fully[[:space:]]Convolutional[[:space:]]Slice-to-Volume[[:space:]]Reconstruction[[:space:]]for[[:space:]]Single-Stack[[:space:]]MRI/ef93a893-2bdf-4dbc-bd8a-45bae35db358_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fully[[:space:]]Exploiting[[:space:]]Every[[:space:]]Real[[:space:]]Sample_[[:space:]]SuperPixel[[:space:]]Sample[[:space:]]Gradient[[:space:]]Model[[:space:]]Stealing/21c794c3-a535-4bb8-85c5-c2d6581a621a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fully[[:space:]]Geometric[[:space:]]Panoramic[[:space:]]Localization/215fae5e-584c-4f87-ac89-3dfd8572591a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fun[[:space:]]with[[:space:]]Flags_[[:space:]]Robust[[:space:]]Principal[[:space:]]Directions[[:space:]]via[[:space:]]Flag[[:space:]]Manifolds/40977cf4-70c4-4d74-80af-1b87f6152117_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Functional[[:space:]]Diffusion/41f7f373-b855-42db-810e-84d65a2220e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Fusing[[:space:]]Personal[[:space:]]and[[:space:]]Environmental[[:space:]]Cues[[:space:]]for[[:space:]]Identification[[:space:]]and[[:space:]]Segmentation[[:space:]]of[[:space:]]First-Person[[:space:]]Camera[[:space:]]Wearers[[:space:]]in[[:space:]]Third-Person[[:space:]]Views/98057eaf-0d33-4d55-9ea7-2e5890eb7c2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/FutureHuman3D_[[:space:]]Forecasting[[:space:]]Complex[[:space:]]Long-Term[[:space:]]3D[[:space:]]Human[[:space:]]Behavior[[:space:]]from[[:space:]]Video[[:space:]]Observations/a4a5fe46-7340-498c-8481-e31218e7bc9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/G-FARS_[[:space:]]Gradient-Field-based[[:space:]]Auto-Regressive[[:space:]]Sampling[[:space:]]for[[:space:]]3D[[:space:]]Part[[:space:]]Grouping/c493864b-a480-4e35-86f2-0ab5a04ed8ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/G-HOP_[[:space:]]Generative[[:space:]]Hand-Object[[:space:]]Prior[[:space:]]for[[:space:]]Interaction[[:space:]]Reconstruction[[:space:]]and[[:space:]]Grasp[[:space:]]Synthesis/a622bb95-f046-4f88-88f6-8bd0a31c2112_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/G-NeRF_[[:space:]]Geometry-enhanced[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]from[[:space:]]Single-View[[:space:]]Images/775a482f-6b71-49c2-a3e0-abce53e3faa2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/G3DR_[[:space:]]Generative[[:space:]]3D[[:space:]]Reconstruction[[:space:]]in[[:space:]]ImageNet/90a24c97-098a-47c9-be0d-e86fa1798e57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GAFusion_[[:space:]]Adaptive[[:space:]]Fusing[[:space:]]LiDAR[[:space:]]and[[:space:]]Camera[[:space:]]with[[:space:]]Multiple[[:space:]]Guidance[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/77703d70-6214-4bbf-9b36-1818e6da9507_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GALA_[[:space:]]Generating[[:space:]]Animatable[[:space:]]Layered[[:space:]]Assets[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Scan/e43b62e9-f190-452b-9e46-572a4104becc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GARField_[[:space:]]Group[[:space:]]Anything[[:space:]]with[[:space:]]Radiance[[:space:]]Fields/148fe4f1-18f5-4c85-a114-d914bbeba083_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GART_[[:space:]]Gaussian[[:space:]]Articulated[[:space:]]Template[[:space:]]Models/91cad794-89db-45ee-a0f5-291658b9b32e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GAvatar_[[:space:]]Animatable[[:space:]]3D[[:space:]]Gaussian[[:space:]]Avatars[[:space:]]with[[:space:]]Implicit[[:space:]]Mesh[[:space:]]Learning/021808b9-82d3-441e-9ca7-47b3b59672a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GDA_[[:space:]]Generalized[[:space:]]Diffusion[[:space:]]for[[:space:]]Robust[[:space:]]Test-time[[:space:]]Adaptation/6fb18cfd-4bdc-4de0-84b3-0847fad4e772_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GEARS_[[:space:]]Local[[:space:]]Geometry-aware[[:space:]]Hand-object[[:space:]]Interaction[[:space:]]Synthesis/48e81588-0e10-4df6-8a9d-c53ef50a084b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GES[[:space:]]_[[:space:]]Generalized[[:space:]]Exponential[[:space:]]Splatting[[:space:]]for[[:space:]]Efficient[[:space:]]Radiance[[:space:]]Field[[:space:]]Rendering/2be06ffd-5c00-41a5-a036-bf504d16c3d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GLACE_[[:space:]]Global[[:space:]]Local[[:space:]]Accelerated[[:space:]]Coordinate[[:space:]]Encoding/dd3d6039-8dd3-4a9f-b112-489ca17ad1a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GLID_[[:space:]]Pre-training[[:space:]]a[[:space:]]Generalist[[:space:]]Encoder-Decoder[[:space:]]Vision[[:space:]]Model/301c4048-2666-4085-b33b-98f7bc62b72d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GLOW_[[:space:]]Global[[:space:]]Layout[[:space:]]Aware[[:space:]]Attacks[[:space:]]on[[:space:]]Object[[:space:]]Detection/1722d0d2-50f9-4007-8536-c2c92f066840_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GLaMM_[[:space:]]Pixel[[:space:]]Grounding[[:space:]]Large[[:space:]]Multimodal[[:space:]]Model/579f9567-b4a9-45df-86c9-3b5030deea5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GLiDR_[[:space:]]Topologically[[:space:]]Regularized[[:space:]]Graph[[:space:]]Generative[[:space:]]Network[[:space:]]for[[:space:]]Sparse[[:space:]]LiDAR[[:space:]]Point[[:space:]]Clouds/dea84620-ce7d-45ad-9896-d1bce2c20ee1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GOAT-Bench_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Multi-Modal[[:space:]]Lifelong[[:space:]]Navigation/93de72e3-c056-4d1c-ae09-174407bd55a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GOV-NeSF_[[:space:]]Generalizable[[:space:]]Open-Vocabulary[[:space:]]Neural[[:space:]]Semantic[[:space:]]Fields/90f2be93-b16b-4819-8711-580d838cb77e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GP-NeRF_[[:space:]]Generalized[[:space:]]Perception[[:space:]]NeRF[[:space:]]for[[:space:]]Context-Aware[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/786d35c5-32eb-4384-bc72-11c30027b8fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GPLD3D_[[:space:]]Latent[[:space:]]Diffusion[[:space:]]of[[:space:]]3D[[:space:]]Shape[[:space:]]Generative[[:space:]]Models[[:space:]]by[[:space:]]Enforcing[[:space:]]Geometric[[:space:]]and[[:space:]]Physical[[:space:]]Priors/2c0d0857-f959-4b5b-851d-61ea0518f7a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GPS-Gaussian_[[:space:]]Generalizable[[:space:]]Pixel-wise[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Real-time[[:space:]]Human[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/d80f2432-62bb-4d59-a543-d7a12c972968_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GPT-4V(ision)[[:space:]]is[[:space:]]a[[:space:]]Human-Aligned[[:space:]]Evaluator[[:space:]]for[[:space:]]Text-to-3D[[:space:]]Generation/a3660d3f-9c59-4d9f-b7f2-795fb06ae9a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GPT4Point_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Point-Language[[:space:]]Understanding[[:space:]]and[[:space:]]Generation/f8542736-8911-4c2d-be22-bd501d54c8bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GRAM_[[:space:]]Global[[:space:]]Reasoning[[:space:]]for[[:space:]]Multi-Page[[:space:]]VQA/bd3538ed-fd76-4274-b7c4-f2ff80c7e1bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GROUNDHOG_[[:space:]]Grounding[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]to[[:space:]]Holistic[[:space:]]Segmentation/bdfcf2de-41e8-4d27-af7d-65821e6b79f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GS-IR_[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Inverse[[:space:]]Rendering/d0b63333-3b55-4d7f-93c1-cfbd4adae3a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GS-SLAM_[[:space:]]Dense[[:space:]]Visual[[:space:]]SLAM[[:space:]]with[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/93048e35-cf42-49e6-8897-d6803305631d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GSNeRF_[[:space:]]Generalizable[[:space:]]Semantic[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]with[[:space:]]Enhanced[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/e670ccdd-f8c3-4912-9d6e-b86fa014dde2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GSVA_[[:space:]]Generalized[[:space:]]Segmentation[[:space:]]via[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/12e5e8f5-7030-48a1-9213-1240824636f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/G^3-LQ_[[:space:]]Marrying[[:space:]]Hyperbolic[[:space:]]Alignment[[:space:]]with[[:space:]]Explicit[[:space:]]Semantic-Geometric[[:space:]]Modeling[[:space:]]for[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding/7e2af7bd-e579-4e10-8895-d1ef28fe0cc7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Garment[[:space:]]Recovery[[:space:]]with[[:space:]]Shape[[:space:]]and[[:space:]]Deformation[[:space:]]Priors/83e37ee5-1c15-40f2-ac5d-8e3141ed428d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Gated[[:space:]]Fields_[[:space:]]Learning[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]from[[:space:]]Gated[[:space:]]Videos/babd13ae-5bce-44f5-bd0f-203ff88804c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GauHuman_[[:space:]]Articulated[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Monocular[[:space:]]Human[[:space:]]Videos/8c2dc361-8082-42c7-a19c-2ac9cac55b4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Gaussian[[:space:]]Head[[:space:]]Avatar_[[:space:]]Ultra[[:space:]]High-fidelity[[:space:]]Head[[:space:]]Avatar[[:space:]]via[[:space:]]Dynamic[[:space:]]Gaussians/24a06516-83b5-41f4-8bb5-a4da984c01e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Gaussian[[:space:]]Shading_[[:space:]]Provable[[:space:]]Performance-Lossless[[:space:]]Image[[:space:]]Watermarking[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/cf093428-bde9-4e71-b54b-2506850cecda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Gaussian[[:space:]]Shadow[[:space:]]Casting[[:space:]]for[[:space:]]Neural[[:space:]]Characters/e6fedb1c-caf9-484c-be44-7dbe074386f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Gaussian[[:space:]]Shell[[:space:]]Maps[[:space:]]for[[:space:]]Efficient[[:space:]]3D[[:space:]]Human[[:space:]]Generation/c6318955-2d79-4389-9712-ec90ab24e73b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Gaussian[[:space:]]Splatting[[:space:]]SLAM/32db8396-043e-45f8-8764-d01308949b32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Gaussian-Flow_[[:space:]]4D[[:space:]]Reconstruction[[:space:]]with[[:space:]]Dynamic[[:space:]]3D[[:space:]]Gaussian[[:space:]]Particle/3e2469b7-4042-4a31-bb56-8ba156724067_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GaussianAvatar_[[:space:]]Towards[[:space:]]Realistic[[:space:]]Human[[:space:]]Avatar[[:space:]]Modeling[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Video[[:space:]]via[[:space:]]Animatable[[:space:]]3D[[:space:]]Gaussians/30b3a92a-861e-47e6-8246-2e90d1765e38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GaussianAvatars_[[:space:]]Photorealistic[[:space:]]Head[[:space:]]Avatars[[:space:]]with[[:space:]]Rigged[[:space:]]3D[[:space:]]Gaussians/41411f0e-3b8e-40c3-bbbd-d16b793f1be0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GaussianDreamer_[[:space:]]Fast[[:space:]]Generation[[:space:]]from[[:space:]]Text[[:space:]]to[[:space:]]3D[[:space:]]Gaussians[[:space:]]by[[:space:]]Bridging[[:space:]]2D[[:space:]]and[[:space:]]3D[[:space:]]Diffusion[[:space:]]Models/68c13cd6-fa8d-4333-a8a0-12d60a7c6cac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GaussianEditor_[[:space:]]Editing[[:space:]]3D[[:space:]]Gaussians[[:space:]]Delicately[[:space:]]with[[:space:]]Text[[:space:]]Instructions/22c39c85-c3be-4622-b7a4-f891bcbdf14e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GaussianEditor_[[:space:]]Swift[[:space:]]and[[:space:]]Controllable[[:space:]]3D[[:space:]]Editing[[:space:]]with[[:space:]]Gaussian[[:space:]]Splatting/f458e425-2ba9-468a-94e0-95be47f99560_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GaussianShader_[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]Shading[[:space:]]Functions[[:space:]]for[[:space:]]Reflective[[:space:]]Surfaces/e3ac7488-6672-4bab-abdc-38e188c8adc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Gear-NeRF_[[:space:]]Free-Viewpoint[[:space:]]Rendering[[:space:]]and[[:space:]]Tracking[[:space:]]with[[:space:]]Motion-aware[[:space:]]Spatio-Temporal[[:space:]]Sampling/8ba31972-a909-4d60-aef6-d492fd479dbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GenFlow_[[:space:]]Generalizable[[:space:]]Recurrent[[:space:]]Flow[[:space:]]for[[:space:]]6D[[:space:]]Pose[[:space:]]Refinement[[:space:]]of[[:space:]]Novel[[:space:]]Objects/ff0e87f9-2cf2-45f0-8b79-ad267112b5a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GenH2R_[[:space:]]Learning[[:space:]]Generalizable[[:space:]]Human-to-Robot[[:space:]]Handover[[:space:]]via[[:space:]]Scalable[[:space:]]Simulation[[:space:]]Demonstration[[:space:]]and[[:space:]]Imitation/40314ff6-3af8-45b8-b8fc-6f9ebdc7d993_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GenN2N_[[:space:]]Generative[[:space:]]NeRF2NeRF[[:space:]]Translation/720fe634-ca22-4125-bab8-c49c422e71c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GenNBV_[[:space:]]Generalizable[[:space:]]Next-Best-View[[:space:]]Policy[[:space:]]for[[:space:]]Active[[:space:]]3D[[:space:]]Reconstruction/171b2e0e-83b0-4811-a381-9c85825e51ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GenTron_[[:space:]]Diffusion[[:space:]]Transformers[[:space:]]for[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Generation/2bb6a697-7387-4091-9b21-cb7c548eb341_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GenZI_[[:space:]]Zero-Shot[[:space:]]3D[[:space:]]Human-Scene[[:space:]]Interaction[[:space:]]Generation/3823a91d-f3af-49d1-a8df-4cb9a1c22737_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GeneAvatar_[[:space:]]Generic[[:space:]]Expression-Aware[[:space:]]Volumetric[[:space:]]Head[[:space:]]Avatar[[:space:]]Editing[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/2d50005a-8c8d-4327-bf7a-e48503a1d824_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/General[[:space:]]Object[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Images[[:space:]]and[[:space:]]Videos[[:space:]]at[[:space:]]Scale/0bc99999-be81-4577-bff4-eaa67b1ee51f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/General[[:space:]]Point[[:space:]]Model[[:space:]]Pretraining[[:space:]]with[[:space:]]Autoencoding[[:space:]]and[[:space:]]Autoregressive/a64a8093-c9f6-4338-aa51-464da1052ad9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generalizable[[:space:]]Face[[:space:]]Landmarking[[:space:]]Guided[[:space:]]by[[:space:]]Conditional[[:space:]]Face[[:space:]]Warping/33b95235-e8cb-4b6b-a79e-5de2b0524965_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generalizable[[:space:]]Novel-View[[:space:]]Synthesis[[:space:]]using[[:space:]]a[[:space:]]Stereo[[:space:]]Camera/516ec60a-018b-4106-92f4-813c127e6028_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generalizable[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Classification[[:space:]]with[[:space:]]Fine-Grained[[:space:]]Visual-Semantic[[:space:]]Interaction/c4c5691b-0982-48b4-b41b-bc14bcf8455f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generalized[[:space:]]Event[[:space:]]Cameras/e213eaf6-87b7-49bd-b232-571683e80e75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generalized[[:space:]]Large-Scale[[:space:]]Data[[:space:]]Condensation[[:space:]]via[[:space:]]Various[[:space:]]Backbone[[:space:]]and[[:space:]]Statistical[[:space:]]Matching/1c0e887f-135a-4bea-8bad-3d161e8c65c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generalized[[:space:]]Predictive[[:space:]]Model[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/554856ec-712f-446a-b09c-0cce647f2af6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generalizing[[:space:]]6-DoF[[:space:]]Grasp[[:space:]]Detection[[:space:]]via[[:space:]]Domain[[:space:]]Prior[[:space:]]Knowledge/65c67c8c-a659-4259-9daf-ad24fd3765f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generate[[:space:]]Like[[:space:]]Experts_[[:space:]]Multi-Stage[[:space:]]Font[[:space:]]Generation[[:space:]]by[[:space:]]Incorporating[[:space:]]Font[[:space:]]Transfer[[:space:]]Process[[:space:]]into[[:space:]]Diffusion[[:space:]]Models/bfb6c4d4-ce31-4c6d-ba23-750f3d411ecc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generate[[:space:]]Subgoal[[:space:]]Images[[:space:]]before[[:space:]]Act_[[:space:]]Unlocking[[:space:]]the[[:space:]]Chain-of-Thought[[:space:]]Reasoning[[:space:]]in[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Robot[[:space:]]Manipulation[[:space:]]with[[:space:]]Multimodal[[:space:]]Prompts/d0cc6546-2994-4121-92a3-5e7ff79756f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generating[[:space:]]Content[[:space:]]for[[:space:]]HDR[[:space:]]Deghosting[[:space:]]from[[:space:]]Frequency[[:space:]]View/2b6f8409-576d-47cf-ab70-c1b4b9e76809_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generating[[:space:]]Enhanced[[:space:]]Negatives[[:space:]]for[[:space:]]Training[[:space:]]Language-Based[[:space:]]Object[[:space:]]Detectors/2970baed-7bcf-45c6-83b6-e79faa1f9b77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generating[[:space:]]Handwritten[[:space:]]Mathematical[[:space:]]Expressions[[:space:]]From[[:space:]]Symbol[[:space:]]Graphs_[[:space:]]An[[:space:]]End-to-End[[:space:]]Pipeline/f14645bb-d996-46da-9c4b-81a14daf5714_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generating[[:space:]]Human[[:space:]]Motion[[:space:]]in[[:space:]]3D[[:space:]]Scenes[[:space:]]from[[:space:]]Text[[:space:]]Descriptions/507f8570-ad9c-455e-838f-8fa1bda75b31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generating[[:space:]]Illustrated[[:space:]]Instructions/dd71c86a-390f-4c46-b6ae-0a13f2787fa0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generating[[:space:]]Non-Stationary[[:space:]]Textures[[:space:]]using[[:space:]]Self-Rectification/3ddc2997-ec2e-480d-b380-c7012481f817_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generative[[:space:]]3D[[:space:]]Part[[:space:]]Assembly[[:space:]]via[[:space:]]Part-Whole-Hierarchy[[:space:]]Message[[:space:]]Passing/40f117f5-c539-42c1-b970-aab5fe53da07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generative[[:space:]]Image[[:space:]]Dynamics/34d8fada-6dfb-4124-8738-6ec5f099992d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generative[[:space:]]Latent[[:space:]]Coding[[:space:]]for[[:space:]]Ultra-Low[[:space:]]Bitrate[[:space:]]Image[[:space:]]Compression/a71ce275-c18c-4862-b210-a735f3bb6944_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generative[[:space:]]Multi-modal[[:space:]]Models[[:space:]]are[[:space:]]Good[[:space:]]Class[[:space:]]Incremental[[:space:]]Learners/a19d9ccd-1dff-4a86-83d4-a672741b4e0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generative[[:space:]]Multimodal[[:space:]]Models[[:space:]]are[[:space:]]In-Context[[:space:]]Learners/1a0461d9-1d47-4e83-8d07-3f1f380c508b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generative[[:space:]]Powers[[:space:]]of[[:space:]]Ten/2ccba12d-1f5a-40cf-8686-79a45cb775c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generative[[:space:]]Proxemics_[[:space:]]A[[:space:]]Prior[[:space:]]for[[:space:]]3D[[:space:]]Social[[:space:]]Interaction[[:space:]]from[[:space:]]Images/b086fde4-b439-4f25-92c2-a2dc5214a0a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generative[[:space:]]Quanta[[:space:]]Color[[:space:]]Imaging/9125a07a-d19f-455b-bcb9-5f2fd8cc25e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generative[[:space:]]Region-Language[[:space:]]Pretraining[[:space:]]for[[:space:]]Open-Ended[[:space:]]Object[[:space:]]Detection/bb1e00ce-acc5-43f7-b817-49b140b640c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generative[[:space:]]Rendering_[[:space:]]Controllable[[:space:]]4D-Guided[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]2D[[:space:]]Diffusion[[:space:]]Models/038dcdfb-1eda-4a35-919e-6cd09ccad282_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Generative[[:space:]]Unlearning[[:space:]]for[[:space:]]Any[[:space:]]Identity/c3c2d3ee-c7c2-4590-9220-bed4fbf9b764_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GenesisTex_[[:space:]]Adapting[[:space:]]Image[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]to[[:space:]]Texture[[:space:]]Space/ba009eff-00cf-485a-8674-0d1aa83b6a01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Genuine[[:space:]]Knowledge[[:space:]]from[[:space:]]Practice_[[:space:]]Diffusion[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Video[[:space:]]Adverse[[:space:]]Weather[[:space:]]Removal/254c789f-0e30-45ba-9906-00c14a08b4b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GeoAuxNet_[[:space:]]Towards[[:space:]]Universal[[:space:]]3D[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Multi-sensor[[:space:]]Point[[:space:]]Clouds/f223747a-d407-4fbb-9660-d95973ceefec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GeoChat_[[:space:]]Grounded[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Model[[:space:]]for[[:space:]]Remote[[:space:]]Sensing/d9275ce8-58db-4c85-a4fc-e4300f4bc853_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GeoReF_[[:space:]]Geometric[[:space:]]Alignment[[:space:]]Across[[:space:]]Shape[[:space:]]Variation[[:space:]]for[[:space:]]Category-level[[:space:]]Object[[:space:]]Pose[[:space:]]Refinement/966fabe9-a4db-4529-94bf-c710cc01e232_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Geometrically-driven[[:space:]]Aggregation[[:space:]]for[[:space:]]Zero-shot[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Understanding/570ecd4d-8254-48b6-9b45-52b56da82573_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Geometry[[:space:]]Transfer[[:space:]]for[[:space:]]Stylizing[[:space:]]Radiance[[:space:]]Fields/a606e334-3ec3-4971-9ea0-d0e3501674fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Geometry-aware[[:space:]]Reconstruction[[:space:]]and[[:space:]]Fusion-refined[[:space:]]Rendering[[:space:]]for[[:space:]]Generalizable[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/cb549adb-f27f-4afe-8cd5-93b9bac6890a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GigaPose_[[:space:]]Fast[[:space:]]and[[:space:]]Robust[[:space:]]Novel[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]One[[:space:]]Correspondence/5375364a-62df-4ab4-a760-301a7885758e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GigaTraj_[[:space:]]Predicting[[:space:]]Long-term[[:space:]]Trajectories[[:space:]]of[[:space:]]Hundreds[[:space:]]of[[:space:]]Pedestrians[[:space:]]in[[:space:]]Gigapixel[[:space:]]Complex[[:space:]]Scenes/d99e0a37-c868-45d2-a75e-fc3c3c270a81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GlitchBench_[[:space:]]Can[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models[[:space:]]Detect[[:space:]]Video[[:space:]]Game[[:space:]]Glitches_/9ccff5ba-748e-48d2-a577-f26b4d3855f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Global[[:space:]]Latent[[:space:]]Neural[[:space:]]Rendering/0ad91f30-3e0e-4c5a-93f0-eb2a13c4bff3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Global[[:space:]]and[[:space:]]Hierarchical[[:space:]]Geometry[[:space:]]Consistency[[:space:]]Priors[[:space:]]for[[:space:]]Few-shot[[:space:]]NeRFs[[:space:]]in[[:space:]]Indoor[[:space:]]Scenes/3cd408ee-dcc0-415c-b5c5-e2643c1eae01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Global[[:space:]]and[[:space:]]Local[[:space:]]Prompts[[:space:]]Cooperation[[:space:]]via[[:space:]]Optimal[[:space:]]Transport[[:space:]]for[[:space:]]Federated[[:space:]]Learning/aa17552d-8104-4d22-83cb-7c75bd4fc6e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GoMAvatar_[[:space:]]Efficient[[:space:]]Animatable[[:space:]]Human[[:space:]]Modeling[[:space:]]from[[:space:]]Monocular[[:space:]]Video[[:space:]]Using[[:space:]]Gaussians-on-Mesh/ac3709f4-dae5-48ba-9391-e162a6b5ae8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GoMVS_[[:space:]]Geometrically[[:space:]]Consistent[[:space:]]Cost[[:space:]]Aggregation[[:space:]]for[[:space:]]Multi-View[[:space:]]Stereo/f4079120-bb95-42b3-a218-9ecb0e334b67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Going[[:space:]]Beyond[[:space:]]Multi-Task[[:space:]]Dense[[:space:]]Prediction[[:space:]]with[[:space:]]Synergy[[:space:]]Embedding[[:space:]]Models/0e180367-dcb5-4e94-ad3f-2ed1a33e8a3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GoodSAM_[[:space:]]Bridging[[:space:]]Domain[[:space:]]and[[:space:]]Capacity[[:space:]]Gaps[[:space:]]via[[:space:]]Segment[[:space:]]Anything[[:space:]]Model[[:space:]]for[[:space:]]Distortion-aware[[:space:]]Panoramic[[:space:]]Semantic[[:space:]]Segmentation/ae7f3ccf-e36a-45c7-aada-2336e28238a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GraCo_[[:space:]]Granularity-Controllable[[:space:]]Interactive[[:space:]]Segmentation/fb0c13f0-276c-4d23-95c7-34863ab22f94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Gradient[[:space:]]Alignment[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Face[[:space:]]Anti-Spoofing/5568b606-7ede-43d8-a317-b0e6312edf5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Gradient[[:space:]]Reweighting_[[:space:]]Towards[[:space:]]Imbalanced[[:space:]]Class-Incremental[[:space:]]Learning/745473f2-6f24-4209-b782-184fd09e70ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Gradient-based[[:space:]]Parameter[[:space:]]Selection[[:space:]]for[[:space:]]Efficient[[:space:]]Fine-Tuning/9422dcaa-bbe7-40c1-a0e6-32328cc18644_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GraphDreamer_[[:space:]]Compositional[[:space:]]3D[[:space:]]Scene[[:space:]]Synthesis[[:space:]]from[[:space:]]Scene[[:space:]]Graphs/f36b22f5-bf0d-4471-91c5-e8018fd104c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GreedyViG_[[:space:]]Dynamic[[:space:]]Axial[[:space:]]Graph[[:space:]]Construction[[:space:]]for[[:space:]]Efficient[[:space:]]Vision[[:space:]]GNNs/f1afdbb5-c725-49ec-8709-7167b416e967_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Grid[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Text-to-Video[[:space:]]Generation/651fc616-c102-404a-a7ff-a3446bb603a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Grounded[[:space:]]Question-Answering[[:space:]]in[[:space:]]Long[[:space:]]Egocentric[[:space:]]Videos/088fdb5a-f4a3-42d1-8cb1-3dbf5a0c99df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Grounded[[:space:]]Text-to-Image[[:space:]]Synthesis[[:space:]]with[[:space:]]Attention[[:space:]]Refocusing/223453e2-d227-4243-bc3b-2c605fd8f6d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Grounding[[:space:]]Everything_[[:space:]]Emerging[[:space:]]Localization[[:space:]]Properties[[:space:]]in[[:space:]]Vision-Language[[:space:]]Transformers/a5f53d17-fae6-4c22-9136-8cfecc733e3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Grounding[[:space:]]and[[:space:]]Enhancing[[:space:]]Grid-based[[:space:]]Models[[:space:]]for[[:space:]]Neural[[:space:]]Fields/fc09a9bd-91c9-478d-838e-cce9af12c6a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/GroupContrast_[[:space:]]Semantic-aware[[:space:]]Self-supervised[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]3D[[:space:]]Understanding/032174fe-80c4-4761-b8ac-7cd874b8a683_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Groupwise[[:space:]]Query[[:space:]]Specialization[[:space:]]and[[:space:]]Quality-Aware[[:space:]]Multi-Assignment[[:space:]]for[[:space:]]Transformer-based[[:space:]]Visual[[:space:]]Relationship[[:space:]]Detection/c0fdbffe-853d-422d-8224-e744af39f332_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Guess[[:space:]]The[[:space:]]Unseen_[[:space:]]Dynamic[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]from[[:space:]]Partial[[:space:]]2D[[:space:]]Glimpses/1c024b6a-587a-49d5-a92a-c7570e8adcac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Guided[[:space:]]Slot[[:space:]]Attention[[:space:]]for[[:space:]]Unsupervised[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/8b9bee56-f2c2-48a3-9e95-fa426d512916_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/H-ViT_[[:space:]]A[[:space:]]Hierarchical[[:space:]]Vision[[:space:]]Transformer[[:space:]]for[[:space:]]Deformable[[:space:]]Image[[:space:]]Registration/511259f2-917c-45c5-ae4b-fa2765a91d8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HAVE-FUN_[[:space:]]Human[[:space:]]Avatar[[:space:]]Reconstruction[[:space:]]from[[:space:]]Few-Shot[[:space:]]Unconstrained[[:space:]]Images/87eaf448-1431-43d3-9df9-0677aa190d22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HDQMF_[[:space:]]Holographic[[:space:]]Feature[[:space:]]Decomposition[[:space:]]Using[[:space:]]Quantum[[:space:]]Algorithms/c8688281-a45b-4fb9-9f12-34e6d8a59b8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HDRFlow_[[:space:]]Real-Time[[:space:]]HDR[[:space:]]Video[[:space:]]Reconstruction[[:space:]]with[[:space:]]Large[[:space:]]Motions/ecf6e681-21e4-4d03-bd6c-8ed70ae68e73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HEAL-SWIN_[[:space:]]A[[:space:]]Vision[[:space:]]Transformer[[:space:]]On[[:space:]]The[[:space:]]Sphere/68d8bc40-5dcf-48b0-a2b4-3877b2aa65d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HHMR_[[:space:]]Holistic[[:space:]]Hand[[:space:]]Mesh[[:space:]]Recovery[[:space:]]by[[:space:]]Enhancing[[:space:]]the[[:space:]]Multimodal[[:space:]]Controllability[[:space:]]of[[:space:]]Graph[[:space:]]Diffusion[[:space:]]Models/cddb5d11-2f01-4767-9266-5c94dca3487d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HIG_[[:space:]]Hierarchical[[:space:]]Interlacement[[:space:]]Graph[[:space:]]Approach[[:space:]]to[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]in[[:space:]]Video[[:space:]]Understanding/0e916748-82e8-4f94-8449-d9230b318d93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HIMap_[[:space:]]HybrId[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]End-to-end[[:space:]]Vectorized[[:space:]]HD[[:space:]]Map[[:space:]]Construction/debda2b8-a011-4d6e-8305-40d7e7710f01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HINTED_[[:space:]]Hard[[:space:]]Instance[[:space:]]Enhanced[[:space:]]Detector[[:space:]]with[[:space:]]Mixed-Density[[:space:]]Feature[[:space:]]Fusion[[:space:]]for[[:space:]]Sparsely-Supervised[[:space:]]3D[[:space:]]Object[[:space:]]Detection/93fa2d80-cc26-4c13-91e1-dc447644e188_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HIPTrack_[[:space:]]Visual[[:space:]]Tracking[[:space:]]with[[:space:]]Historical[[:space:]]Prompts/cda3de52-4002-462c-8355-bf489f7ee0f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HIR-Diff_[[:space:]]Unsupervised[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Restoration[[:space:]]Via[[:space:]]Improved[[:space:]]Diffusion[[:space:]]Models/e380b0e8-f54d-4c04-83f2-95f80b8fe08b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HIT_[[:space:]]Estimating[[:space:]]Internal[[:space:]]Human[[:space:]]Implicit[[:space:]]Tissues[[:space:]]from[[:space:]]the[[:space:]]Body[[:space:]]Surface/6fa7624d-59a1-45de-ae21-76c4223036da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HIVE_[[:space:]]Harnessing[[:space:]]Human[[:space:]]Feedback[[:space:]]for[[:space:]]Instructional[[:space:]]Visual[[:space:]]Editing/1d573b3b-ec1a-435a-aa5d-bad1c5d507a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HMD-Poser_[[:space:]]On-Device[[:space:]]Real-time[[:space:]]Human[[:space:]]Motion[[:space:]]Tracking[[:space:]]from[[:space:]]Scalable[[:space:]]Sparse[[:space:]]Observations/e945f1af-be0e-4551-893d-6f337c666f76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HOI-M^3_[[:space:]]Capture[[:space:]]Multiple[[:space:]]Humans[[:space:]]and[[:space:]]Objects[[:space:]]Interaction[[:space:]]within[[:space:]]Contextual[[:space:]]Environment/3013a567-29e3-401a-9aa5-62ec8d526640_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HOIAnimator_[[:space:]]Generating[[:space:]]Text-prompt[[:space:]]Human-object[[:space:]]Animations[[:space:]]using[[:space:]]Novel[[:space:]]Perceptive[[:space:]]Diffusion[[:space:]]Models/2f4133dc-677f-4527-adc3-7a67da3a82a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HOIDiffusion_[[:space:]]Generating[[:space:]]Realistic[[:space:]]3D[[:space:]]Hand-Object[[:space:]]Interaction[[:space:]]Data/502319ba-2c03-408f-b167-e13f97d7e939_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HOISDF_[[:space:]]Constraining[[:space:]]3D[[:space:]]Hand-Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]Global[[:space:]]Signed[[:space:]]Distance[[:space:]]Fields/4fa67722-1e88-49e3-b878-6ab8b9db5a29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HOIST-Former_[[:space:]]Hand-held[[:space:]]Objects[[:space:]]Identification[[:space:]]Segmentation[[:space:]]and[[:space:]]Tracking[[:space:]]in[[:space:]]the[[:space:]]Wild/c8373694-9562-477c-a08b-5b7c50523e63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HOLD_[[:space:]]Category-agnostic[[:space:]]3D[[:space:]]Reconstruction[[:space:]]of[[:space:]]Interacting[[:space:]]Hands[[:space:]]and[[:space:]]Objects[[:space:]]from[[:space:]]Video/eb697109-44c1-4a45-8b14-531f102f2ed1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HPL-ESS_[[:space:]]Hybrid[[:space:]]Pseudo-Labeling[[:space:]]for[[:space:]]Unsupervised[[:space:]]Event-based[[:space:]]Semantic[[:space:]]Segmentation/384b55d1-c6ad-4e4d-acf7-29620c73d054_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HPNet_[[:space:]]Dynamic[[:space:]]Trajectory[[:space:]]Forecasting[[:space:]]with[[:space:]]Historical[[:space:]]Prediction[[:space:]]Attention/c472fdb7-44ae-4af7-a387-45c8cfa33b07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HRVDA_[[:space:]]High-Resolution[[:space:]]Visual[[:space:]]Document[[:space:]]Assistant/bfcf0ad6-6ca0-4058-b471-e387b0e58342_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HUGS_[[:space:]]Holistic[[:space:]]Urban[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding[[:space:]]via[[:space:]]Gaussian[[:space:]]Splatting/851501ba-f367-4a36-a756-dd4441dbd60a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HUGS_[[:space:]]Human[[:space:]]Gaussian[[:space:]]Splats/1f980a6c-3984-4059-8020-79c463d4d933_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HUNTER_[[:space:]]Unsupervised[[:space:]]Human-centric[[:space:]]3D[[:space:]]Detection[[:space:]]via[[:space:]]Transferring[[:space:]]Knowledge[[:space:]]from[[:space:]]Synthetic[[:space:]]Instances[[:space:]]to[[:space:]]Real[[:space:]]Scenes/0297e577-980d-4001-b2df-ec3f8b18faf4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Habitat[[:space:]]Synthetic[[:space:]]Scenes[[:space:]]Dataset[[:space:]](HSSD-200)_[[:space:]]An[[:space:]]Analysis[[:space:]]of[[:space:]]3D[[:space:]]Scene[[:space:]]Scale[[:space:]]and[[:space:]]Realism[[:space:]]Tradeoffs[[:space:]]for[[:space:]]ObjectGoal[[:space:]]Navigation/182135e9-0c67-4e67-8698-0bd8f82ab5cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HalluciDoctor_[[:space:]]Mitigating[[:space:]]Hallucinatory[[:space:]]Toxicity[[:space:]]in[[:space:]]Visual[[:space:]]Instruction[[:space:]]Data/026e27cd-6105-4876-8e7b-c047d2e416e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hallucination[[:space:]]Augmented[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model/63c1bb26-1845-40bd-bcf4-12e6c8eaa948_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HallusionBench_[[:space:]]An[[:space:]]Advanced[[:space:]]Diagnostic[[:space:]]Suite[[:space:]]for[[:space:]]Entangled[[:space:]]Language[[:space:]]Hallucination[[:space:]]and[[:space:]]Visual[[:space:]]Illusion[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/ca9a69b2-d464-47f1-b8d8-a3691c340ef5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HanDiffuser_[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]With[[:space:]]Realistic[[:space:]]Hand[[:space:]]Appearances/6d90203e-3dd1-4a78-8695-f4be65f86ce8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HandBooster_[[:space:]]Boosting[[:space:]]3D[[:space:]]Hand-Mesh[[:space:]]Reconstruction[[:space:]]by[[:space:]]Conditional[[:space:]]Synthesis[[:space:]]and[[:space:]]Sampling[[:space:]]of[[:space:]]Hand-Object[[:space:]]Interactions/74c63997-b175-4bd2-919e-5c9321457a62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HandDiff_[[:space:]]3D[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]Diffusion[[:space:]]on[[:space:]]Image-Point[[:space:]]Cloud/5b515a93-0b26-4fae-9411-13fa8bdc6c72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HardMo_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Hardcase[[:space:]]Dataset[[:space:]]for[[:space:]]Motion[[:space:]]Capture/46ce041e-7fd5-42e4-b34d-1b6e5a64ddc8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HarmonyView_[[:space:]]Harmonizing[[:space:]]Consistency[[:space:]]and[[:space:]]Diversity[[:space:]]in[[:space:]]One-Image-to-3D/708d70bc-3b8d-4c9c-8e2d-144b342fd138_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Harnessing[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Training-free[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/65b55c3d-4578-48c5-be98-7ff310626a28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Harnessing[[:space:]]Meta-Learning[[:space:]]for[[:space:]]Improving[[:space:]]Full-Frame[[:space:]]Video[[:space:]]Stabilization/8080e91f-8bae-44ee-81dc-6d9e3e8aefd1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Harnessing[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]MLLMs[[:space:]]for[[:space:]]Transferable[[:space:]]Text-to-Image[[:space:]]Person[[:space:]]ReID/1d4dba02-b26f-4c21-b1e0-ddf74503ef5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HashPoint_[[:space:]]Accelerated[[:space:]]Point[[:space:]]Searching[[:space:]]and[[:space:]]Sampling[[:space:]]for[[:space:]]Neural[[:space:]]Rendering/f82d3f13-9ed0-425e-9fc8-35f9a8ee7ce1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hearing[[:space:]]Anything[[:space:]]Anywhere/c48185f6-be50-44d7-b5ca-b7adf7b2eb39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HiFi4G_[[:space:]]High-Fidelity[[:space:]]Human[[:space:]]Performance[[:space:]]Rendering[[:space:]]via[[:space:]]Compact[[:space:]]Gaussian[[:space:]]Splatting/b4dc085d-582d-4a44-b3ba-0ca7f2dc53d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HiKER-SGG_[[:space:]]Hierarchical[[:space:]]Knowledge[[:space:]]Enhanced[[:space:]]Robust[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/c3833f6a-706f-4970-a336-0b6bd0afa2a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HiLo_[[:space:]]Detailed[[:space:]]and[[:space:]]Robust[[:space:]]3D[[:space:]]Clothed[[:space:]]Human[[:space:]]Reconstruction[[:space:]]with[[:space:]]High-and[[:space:]]Low-Frequency[[:space:]]Information[[:space:]]of[[:space:]]Parametric[[:space:]]Models/c4c07d9f-e561-4b84-9dce-8b47bf7979c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HiPose_[[:space:]]Hierarchical[[:space:]]Binary[[:space:]]Surface[[:space:]]Encoding[[:space:]]and[[:space:]]Correspondence[[:space:]]Pruning[[:space:]]for[[:space:]]RGB-D[[:space:]]6DoF[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/30097c35-18c1-4540-9ddc-275a4d99ebf5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hide[[:space:]]in[[:space:]]Thicket_[[:space:]]Generating[[:space:]]Imperceptible[[:space:]]and[[:space:]]Rational[[:space:]]Adversarial[[:space:]]Perturbations[[:space:]]on[[:space:]]3D[[:space:]]Point[[:space:]]Clouds/dd773576-b805-466a-a968-de9ae3eeed21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hierarchical[[:space:]]Correlation[[:space:]]Clustering[[:space:]]and[[:space:]]Tree[[:space:]]Preserving[[:space:]]Embedding/83af6c64-48b5-4260-a582-3e227f120a6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hierarchical[[:space:]]Diffusion[[:space:]]Policy[[:space:]]for[[:space:]]Kinematics-Aware[[:space:]]Multi-Task[[:space:]]Robotic[[:space:]]Manipulation/bb817413-5153-4514-9909-1e53f1667aa4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hierarchical[[:space:]]Histogram[[:space:]]Threshold[[:space:]]Segmentation[[:space:]]-[[:space:]]Auto-terminating[[:space:]]High-detail[[:space:]]Oversegmentation/f42e7890-3166-4c68-b634-a7b88109f2a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hierarchical[[:space:]]Intra-modal[[:space:]]Correlation[[:space:]]Learning[[:space:]]for[[:space:]]Label-free[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation/72f631d1-b3bf-403a-b303-f0b11cefb77e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hierarchical[[:space:]]Patch[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]High-Resolution[[:space:]]Video[[:space:]]Generation/00cd7eaa-0f1b-4c43-b631-dbacac43f0df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hierarchical[[:space:]]Spatio-temporal[[:space:]]Decoupling[[:space:]]for[[:space:]]Text-to-Video[[:space:]]Generation/df9fcab5-771d-47fd-b92b-b272fb6684b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/High-Quality[[:space:]]Facial[[:space:]]Geometry[[:space:]]and[[:space:]]Appearance[[:space:]]Capture[[:space:]]at[[:space:]]Home/93a851dc-a119-4208-8e4f-f4a67c59b777_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/High-fidelity[[:space:]]Person-centric[[:space:]]Subject-to-Image[[:space:]]Synthesis/3bf133ff-e8e6-4715-b977-e0ae3b7e358d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Higher-order[[:space:]]Relational[[:space:]]Reasoning[[:space:]]for[[:space:]]Pedestrian[[:space:]]Trajectory[[:space:]]Prediction/cd4ba151-d7d7-4d85-9742-4a27cb55d343_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Holistic[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Understanding[[:space:]]by[[:space:]]Bird's-Eye-View[[:space:]]Injected[[:space:]]Multi-Modal[[:space:]]Large[[:space:]]Models/920e308d-d05c-41d9-bdeb-4199270e61b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Holistic[[:space:]]Features[[:space:]]are[[:space:]]almost[[:space:]]Sufficient[[:space:]]for[[:space:]]Text-to-Video[[:space:]]Retrieval/fc4fce8c-f904-4cad-9e1d-ea219621799d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Holo-Relighting_[[:space:]]Controllable[[:space:]]Volumetric[[:space:]]Portrait[[:space:]]Relighting[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/e0a5375b-03d6-4718-a1b4-7a740993a53a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HoloVIC_[[:space:]]Large-scale[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Multi-Sensor[[:space:]]Holographic[[:space:]]Intersection[[:space:]]and[[:space:]]Vehicle-Infrastructure[[:space:]]Cooperative/841a9503-341f-48ae-8b1b-7664aec2f2fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Holodeck_[[:space:]]Language[[:space:]]Guided[[:space:]]Generation[[:space:]]of[[:space:]]3D[[:space:]]Embodied[[:space:]]AI[[:space:]]Environments/8f13cbf1-a7bf-4aa1-b2b4-10f1c5a5d913_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Holoported[[:space:]]Characters_[[:space:]]Real-time[[:space:]]Free-viewpoint[[:space:]]Rendering[[:space:]]of[[:space:]]Humans[[:space:]]from[[:space:]]Sparse[[:space:]]RGB[[:space:]]Cameras/09cbd2e2-9671-48f5-82a7-d1e11c46b15f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HomoFormer_[[:space:]]Homogenized[[:space:]]Transformer[[:space:]]for[[:space:]]Image[[:space:]]Shadow[[:space:]]Removal/be6aec7e-e056-4ec7-9c28-88745b08ad6c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Honeybee_[[:space:]]Locality-enhanced[[:space:]]Projector[[:space:]]for[[:space:]]Multimodal[[:space:]]LLM/a2db7d97-a947-4331-9703-ca4cd234e2d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hourglass[[:space:]]Tokenizer[[:space:]]for[[:space:]]Efficient[[:space:]]Transformer-Based[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/62159953-1cff-47ab-8ead-cd4be7bcfdf7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HouseCat6D[[:space:]]-[[:space:]]A[[:space:]]Large-Scale[[:space:]]Multi-Modal[[:space:]]Category[[:space:]]Level[[:space:]]6D[[:space:]]Object[[:space:]]Perception[[:space:]]Dataset[[:space:]]with[[:space:]]Household[[:space:]]Objects[[:space:]]in[[:space:]]Realistic[[:space:]]Scenarios/665d1a5f-e7eb-480b-bb60-0d028af7db95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/How[[:space:]]Far[[:space:]]Can[[:space:]]We[[:space:]]Compress[[:space:]]Instant-NGP-Based[[:space:]]NeRF_/c9bd0bdb-a841-416a-abc9-7befa2206d22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/How[[:space:]]to[[:space:]]Configure[[:space:]]Good[[:space:]]In-Context[[:space:]]Sequence[[:space:]]for[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/4c20ee97-68c1-4673-b27d-4513a8694a13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/How[[:space:]]to[[:space:]]Handle[[:space:]]Sketch-Abstraction[[:space:]]in[[:space:]]Sketch-Based[[:space:]]Image[[:space:]]Retrieval_/17529ee1-6658-494d-9abf-b5823b5d4118_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/How[[:space:]]to[[:space:]]Make[[:space:]]Cross[[:space:]]Encoder[[:space:]]a[[:space:]]Good[[:space:]]Teacher[[:space:]]for[[:space:]]Efficient[[:space:]]Image-Text[[:space:]]Retrieval_/9c4e4b15-6263-4beb-a02d-77c459c82fdb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/How[[:space:]]to[[:space:]]Train[[:space:]]Neural[[:space:]]Field[[:space:]]Representations_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Study[[:space:]]and[[:space:]]Benchmark/ac267515-5aa6-4f02-bc63-1c67db24823b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HumMUSS_[[:space:]]Human[[:space:]]Motion[[:space:]]Understanding[[:space:]]using[[:space:]]State[[:space:]]Space[[:space:]]Models/39f4a6be-1aec-4965-81fc-bd6018bb1abc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Human[[:space:]]Gaussian[[:space:]]Splatting_[[:space:]]Real-time[[:space:]]Rendering[[:space:]]of[[:space:]]Animatable[[:space:]]Avatars/d8777f80-9d29-44a0-ac1f-d00b40c490af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Human[[:space:]]Motion[[:space:]]Prediction[[:space:]]Under[[:space:]]Unexpected[[:space:]]Perturbation/c6b0b159-e27b-483e-9c8c-9fedd8c81c64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HumanGaussian_[[:space:]]Text-Driven[[:space:]]3D[[:space:]]Human[[:space:]]Generation[[:space:]]with[[:space:]]Gaussian[[:space:]]Splatting/626bb88a-87b0-4baf-8672-57480e464884_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HumanNeRF-SE_[[:space:]]A[[:space:]]Simple[[:space:]]yet[[:space:]]Effective[[:space:]]Approach[[:space:]]to[[:space:]]Animate[[:space:]]HumanNeRF[[:space:]]with[[:space:]]Diverse[[:space:]]Poses/8e0e7e52-ddd1-4e45-a93b-e71d2d8e1b4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HumanNorm_[[:space:]]Learning[[:space:]]Normal[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]High-quality[[:space:]]and[[:space:]]Realistic[[:space:]]3D[[:space:]]Human[[:space:]]Generation/74d89ea0-3627-4f6d-9d8f-1f21601954ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HumanRef_[[:space:]]Single[[:space:]]Image[[:space:]]to[[:space:]]3D[[:space:]]Human[[:space:]]Generation[[:space:]]via[[:space:]]Reference-Guided[[:space:]]Diffusion/f2246bf9-efa5-4f6c-b1af-a360949eaa2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hunting[[:space:]]Attributes_[[:space:]]Context[[:space:]]Prototype-Aware[[:space:]]Learning[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/f59f57ae-78d8-4724-bf6a-cd507a048b9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hybrid[[:space:]]Functional[[:space:]]Maps[[:space:]]for[[:space:]]Crease-Aware[[:space:]]Non-Isometric[[:space:]]Shape[[:space:]]Matching/e7b23608-bf5d-4f38-a389-698bf5a5754c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hybrid[[:space:]]Proposal[[:space:]]Refiner_[[:space:]]Revisiting[[:space:]]DETR[[:space:]]Series[[:space:]]from[[:space:]]the[[:space:]]Faster[[:space:]]R-CNN[[:space:]]Perspective/6b5a5f0a-8571-47fa-a375-73292507c760_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HybridNeRF_[[:space:]]Efficient[[:space:]]Neural[[:space:]]Rendering[[:space:]]via[[:space:]]Adaptive[[:space:]]Volumetric[[:space:]]Surfaces/2831e038-4a4d-474b-ba3a-a6c6384a8807_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hyper-MD_[[:space:]]Mesh[[:space:]]Denoising[[:space:]]with[[:space:]]Customized[[:space:]]Parameters[[:space:]]Aware[[:space:]]of[[:space:]]Noise[[:space:]]Intensity[[:space:]]and[[:space:]]Geometric[[:space:]]Characteristics/65d58485-f6c5-4358-a508-b51c490542ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HyperDreamBooth_[[:space:]]HyperNetworks[[:space:]]for[[:space:]]Fast[[:space:]]Personalization[[:space:]]of[[:space:]]Text-to-Image[[:space:]]Models/24c0434a-be24-460f-b064-6a60613f5ec4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/HyperSDFusion_[[:space:]]Bridging[[:space:]]Hierarchical[[:space:]]Structures[[:space:]]in[[:space:]]Language[[:space:]]and[[:space:]]Geometry[[:space:]]for[[:space:]]Enhanced[[:space:]]3D[[:space:]]Text2Shape[[:space:]]Generation/d954c1ed-d30e-4495-bf06-38c7e0703ac8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hyperbolic[[:space:]]Anomaly[[:space:]]Detection/1fe414da-57be-40ce-9a0e-c4407d1e96d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hyperbolic[[:space:]]Learning[[:space:]]with[[:space:]]Synthetic[[:space:]]Captions[[:space:]]for[[:space:]]Open-World[[:space:]]Detection/be394cdf-7cae-49f4-a6ef-7e6cf530f46e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Hyperspherical[[:space:]]Classification[[:space:]]with[[:space:]]Dynamic[[:space:]]Label-to-Prototype[[:space:]]Assignment/946de759-bfb7-4497-ba23-e498f21ac53b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/I'M[[:space:]]HOI_[[:space:]]Inertia-aware[[:space:]]Monocular[[:space:]]Capture[[:space:]]of[[:space:]]3D[[:space:]]Human-Object[[:space:]]Interactions/6171cab9-bd0a-4cc5-894f-266582754e09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/IBD-SLAM_[[:space:]]Learning[[:space:]]Image-Based[[:space:]]Depth[[:space:]]Fusion[[:space:]]for[[:space:]]Generalizable[[:space:]]SLAM/7da646fb-9888-4057-bd40-105953ae148e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ICON_[[:space:]]Incremental[[:space:]]CONfidence[[:space:]]for[[:space:]]Joint[[:space:]]Pose[[:space:]]and[[:space:]]Radiance[[:space:]]Field[[:space:]]Optimization/c5351303-6ef1-4de9-b49d-c02384d47e0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ICP-Flow_[[:space:]]LiDAR[[:space:]]Scene[[:space:]]Flow[[:space:]]Estimation[[:space:]]with[[:space:]]ICP/a16b5fcf-453a-4af3-a713-e1fb4dd4c4b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ID-Blau_[[:space:]]Image[[:space:]]Deblurring[[:space:]]by[[:space:]]Implicit[[:space:]]Diffusion-based[[:space:]]reBLurring[[:space:]]AUgmentation/1bb6c9f5-cbef-4dab-a848-77095f3884e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ID-like[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Few-Shot[[:space:]]Out-of-Distribution[[:space:]]Detection/1978164c-5b76-4029-88cd-f63922e7aae2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/IDGuard_[[:space:]]Robust[[:space:]]General[[:space:]]Identity-centric[[:space:]]POI[[:space:]]Proactive[[:space:]]Defense[[:space:]]Against[[:space:]]Face[[:space:]]Editing[[:space:]]Abuse/21adf346-ab3f-47ee-8115-2ddd6d490bf4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/IIRP-Net_[[:space:]]Iterative[[:space:]]Inference[[:space:]]Residual[[:space:]]Pyramid[[:space:]]Network[[:space:]]for[[:space:]]Enhanced[[:space:]]Image[[:space:]]Registration/1032fd2e-6071-4293-8dfb-f7e0812ea637_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/IMPRINT_[[:space:]]Generative[[:space:]]Object[[:space:]]Compositing[[:space:]]by[[:space:]]Learning[[:space:]]Identity-Preserving[[:space:]]Representation/85578736-e613-4e09-9d4c-44fef315136a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/IPoD_[[:space:]]Implicit[[:space:]]Field[[:space:]]Learning[[:space:]]with[[:space:]]Point[[:space:]]Diffusion[[:space:]]for[[:space:]]Generalizable[[:space:]]3D[[:space:]]Object[[:space:]]Reconstruction[[:space:]]from[[:space:]]Single[[:space:]]RGB-D[[:space:]]Images/b6aa672f-1205-41cc-b079-5a5e9ff5bca0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/IQ-VFI_[[:space:]]Implicit[[:space:]]Quadratic[[:space:]]Motion[[:space:]]Estimation[[:space:]]for[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation/9125b19f-55ed-4de1-b008-82c7c09e9f5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/IReNe_[[:space:]]Instant[[:space:]]Recoloring[[:space:]]of[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/276f8885-84ca-482c-bf1d-5197c933cb67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/IS-Fusion_[[:space:]]Instance-Scene[[:space:]]Collaborative[[:space:]]Fusion[[:space:]]for[[:space:]]Multimodal[[:space:]]3D[[:space:]]Object[[:space:]]Detection/57930fea-c38e-42a5-87c8-9c073e7f9def_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Identifying[[:space:]]Important[[:space:]]Group[[:space:]]of[[:space:]]Pixels[[:space:]]using[[:space:]]Interactions/44f3006b-1277-47e3-94b1-a13d5d002f5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Image[[:space:]]Neural[[:space:]]Field[[:space:]]Diffusion[[:space:]]Models/e632bc2a-df53-472b-b65b-a754db96e990_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Image[[:space:]]Processing[[:space:]]GNN_[[:space:]]Breaking[[:space:]]Rigidity[[:space:]]in[[:space:]]Super-Resolution/dacb6475-02d0-43fb-ade3-b56b50c0b0c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Image[[:space:]]Restoration[[:space:]]by[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Iteratively[[:space:]]Preconditioned[[:space:]]Guidance/aad8aa4f-e5bd-4107-bdf3-47998cec58ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Image[[:space:]]Sculpting_[[:space:]]Precise[[:space:]]Object[[:space:]]Editing[[:space:]]with[[:space:]]3D[[:space:]]Geometry[[:space:]]Control/98207aad-0c67-4b4c-bfe7-6d5597bcdd58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Image-Text[[:space:]]Co-Decomposition[[:space:]]for[[:space:]]Text-Supervised[[:space:]]Semantic[[:space:]]Segmentation/903a1b45-2229-4628-ac3b-aaa39b0c7384_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Image-to-Image[[:space:]]Matching[[:space:]]via[[:space:]]Foundation[[:space:]]Models_[[:space:]]A[[:space:]]New[[:space:]]Perspective[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/bc2d94c7-b97d-48d5-b633-2ae9bb2d7d08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ImageNet-D_[[:space:]]Benchmarking[[:space:]]Neural[[:space:]]Network[[:space:]]Robustness[[:space:]]on[[:space:]]Diffusion[[:space:]]Synthetic[[:space:]]Object/40bed3ab-f671-40b9-b3ce-7de57541124c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Imagine[[:space:]]Before[[:space:]]Go_[[:space:]]Self-Supervised[[:space:]]Generative[[:space:]]Map[[:space:]]for[[:space:]]Object[[:space:]]Goal[[:space:]]Navigation/48dfab61-319b-4d14-baa1-3fef35834727_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Implicit[[:space:]]Discriminative[[:space:]]Knowledge[[:space:]]Learning[[:space:]]for[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification/b65ca941-cca6-4d10-a9af-cf9b0fc08ec1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Implicit[[:space:]]Event-RGBD[[:space:]]Neural[[:space:]]SLAM/f9bb5455-cdd9-4999-bf90-4f084d0f2dbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Implicit[[:space:]]Motion[[:space:]]Function/52fb1b5d-2b35-4b67-9ebd-49238a4058c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improved[[:space:]]Baselines[[:space:]]with[[:space:]]Visual[[:space:]]Instruction[[:space:]]Tuning/82854e3f-7839-4652-bd17-7a90bb94eee7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improved[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representation[[:space:]]with[[:space:]]Fourier[[:space:]]Reparameterized[[:space:]]Training/cb3f94d3-0c17-46df-9543-5d0cd59a45b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improved[[:space:]]Self-Training[[:space:]]for[[:space:]]Test-Time[[:space:]]Adaptation/dc47adbd-48af-4fc3-a882-67b2a96aeb19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improved[[:space:]]Visual[[:space:]]Grounding[[:space:]]through[[:space:]]Self-Consistent[[:space:]]Explanations/10bbab27-77ad-411f-b190-f41874ca342d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improved[[:space:]]Zero-Shot[[:space:]]Classification[[:space:]]by[[:space:]]Adapting[[:space:]]VLMs[[:space:]]with[[:space:]]Text[[:space:]]Descriptions/54ac6e92-bc80-4537-bc0c-01614b64b2c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Bird's[[:space:]]Eye[[:space:]]View[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]by[[:space:]]Task[[:space:]]Decomposition/64afc4ea-9f99-49c3-ab03-ca65285f08d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Depth[[:space:]]Completion[[:space:]]via[[:space:]]Depth[[:space:]]Feature[[:space:]]Upsampling/0ed38fa6-aa40-45ef-a200-06326c698546_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Distant[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]Using[[:space:]]2D[[:space:]]Box[[:space:]]Supervision/34f5dadc-28a4-4c87-abd0-683d3827eb2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Generalization[[:space:]]via[[:space:]]Meta-Learning[[:space:]]on[[:space:]]Hard[[:space:]]Samples/ceb0ccb0-32b0-4bdb-8488-9b62f8a01ad9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Generalized[[:space:]]Zero-Shot[[:space:]]Learning[[:space:]]by[[:space:]]Exploring[[:space:]]the[[:space:]]Diverse[[:space:]]Semantics[[:space:]]from[[:space:]]External[[:space:]]Class[[:space:]]Names/eeeede11-1438-46b6-b66c-0eb160635025_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Graph[[:space:]]Contrastive[[:space:]]Learning[[:space:]]via[[:space:]]Adaptive[[:space:]]Positive[[:space:]]Sampling/39e494d0-f5b1-4e1a-9fa4-55c8eca880a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Image[[:space:]]Restoration[[:space:]]through[[:space:]]Removing[[:space:]]Degradations[[:space:]]in[[:space:]]Textual[[:space:]]Representations/c7ee9b17-7c3b-4549-9cfe-7a14b5397358_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Out-of-Distribution[[:space:]]Generalization[[:space:]]in[[:space:]]Graphs[[:space:]]via[[:space:]]Hierarchical[[:space:]]Semantic[[:space:]]Environments/bf45557d-430d-42a3-9a68-3f81e133f486_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Physics-Augmented[[:space:]]Continuum[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field-Based[[:space:]]Geometry-Agnostic[[:space:]]System[[:space:]]Identification[[:space:]]with[[:space:]]Lagrangian[[:space:]]Particle[[:space:]]Optimization/4b87738e-8519-4923-9186-4d7b85e26126_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Plasticity[[:space:]]in[[:space:]]Online[[:space:]]Continual[[:space:]]Learning[[:space:]]via[[:space:]]Collaborative[[:space:]]Learning/ea24163b-c4cd-4a88-9ea1-8a8c1cf095e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Semantic[[:space:]]Correspondence[[:space:]]with[[:space:]]Viewpoint-Guided[[:space:]]Spherical[[:space:]]Maps/22b6ed4b-b424-4fe5-8f1f-d4f3f6fa485d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Single[[:space:]]Domain-Generalized[[:space:]]Object[[:space:]]Detection_[[:space:]]A[[:space:]]Focus[[:space:]]on[[:space:]]Diversification[[:space:]]and[[:space:]]Alignment/8c4702fb-5780-47bb-bdfa-6d2896061b4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Spectral[[:space:]]Snapshot[[:space:]]Reconstruction[[:space:]]with[[:space:]]Spectral-Spatial[[:space:]]Rectification/1d035099-0d94-4bf4-acc9-0b3e57ed6aca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Subject-Driven[[:space:]]Image[[:space:]]Synthesis[[:space:]]with[[:space:]]Subject-Agnostic[[:space:]]Guidance/e95992ac-5461-4e8c-ba60-c4005512a188_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Training[[:space:]]Efficiency[[:space:]]of[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Multi-Stage[[:space:]]Framework[[:space:]]and[[:space:]]Tailored[[:space:]]Multi-Decoder[[:space:]]Architecture/ab8cb3ef-8706-4683-ab23-b47ea618549e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Transferable[[:space:]]Targeted[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]with[[:space:]]Model[[:space:]]Self-Enhancement/bed83424-cf5b-430d-ae99-b81617d05b70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Unsupervised[[:space:]]Hierarchical[[:space:]]Representation[[:space:]]with[[:space:]]Reinforcement[[:space:]]Learning/303e0d84-56db-4046-87a5-f6c50e1e3b69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]Visual[[:space:]]Recognition[[:space:]]with[[:space:]]Hyperbolical[[:space:]]Visual[[:space:]]Hierarchy[[:space:]]Mapping/9a46de44-a8c1-46de-8924-4659484e6064_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Improving[[:space:]]the[[:space:]]Generalization[[:space:]]of[[:space:]]Segmentation[[:space:]]Foundation[[:space:]]Model[[:space:]]under[[:space:]]Distribution[[:space:]]Shift[[:space:]]via[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Adaptation/0c9a1639-459a-4ab7-b55c-2fff676c9ecb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/In[[:space:]]Search[[:space:]]of[[:space:]]a[[:space:]]Data[[:space:]]Transformation[[:space:]]That[[:space:]]Accelerates[[:space:]]Neural[[:space:]]Field[[:space:]]Training/e8b54a6a-20cf-4428-8c04-7ff202724221_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/In-Context[[:space:]]Matting/d72f8b5e-8660-4264-a6c3-5f329153b637_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/In-N-Out_[[:space:]]Faithful[[:space:]]3D[[:space:]]GAN[[:space:]]Inversion[[:space:]]with[[:space:]]Volumetric[[:space:]]Decomposition[[:space:]]for[[:space:]]Face[[:space:]]Editing/937da690-a378-4303-a4f6-0fbaa725c37d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/In-distribution[[:space:]]Public[[:space:]]Data[[:space:]]Synthesis[[:space:]]with[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Differentially[[:space:]]Private[[:space:]]Image[[:space:]]Classification/35a6ec9e-03c5-459f-83ad-bafb99a44b46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/In2SET_[[:space:]]Intra-Inter[[:space:]]Similarity[[:space:]]Exploiting[[:space:]]Transformer[[:space:]]for[[:space:]]Dual-Camera[[:space:]]Compressive[[:space:]]Hyperspectral[[:space:]]Imaging/9683a098-8483-47af-abcf-584b3e2dad43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/InNeRF360_[[:space:]]Text-Guided[[:space:]]3D-Consistent[[:space:]]Object[[:space:]]Inpainting[[:space:]]on[[:space:]]360-degree[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/09378d81-d6b3-4b15-9d8b-0c6cf0d263a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/InceptionNeXt_[[:space:]]When[[:space:]]Inception[[:space:]]Meets[[:space:]]ConvNeXt/66d1a177-6c15-49db-8496-bf969a4b7f0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Incorporating[[:space:]]Geo-Diverse[[:space:]]Knowledge[[:space:]]into[[:space:]]Prompting[[:space:]]for[[:space:]]Increased[[:space:]]Geographical[[:space:]]Robustness[[:space:]]in[[:space:]]Object[[:space:]]Recognition/a3586110-0f62-4a60-8f39-dd762f0a5105_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Incremental[[:space:]]Nuclei[[:space:]]Segmentation[[:space:]]from[[:space:]]Histopathological[[:space:]]Images[[:space:]]via[[:space:]]Future-class[[:space:]]Awareness[[:space:]]and[[:space:]]Compatibility-inspired[[:space:]]Distillation/9fa3bbf9-7154-47a0-a450-b1f92451140e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Incremental[[:space:]]Residual[[:space:]]Concept[[:space:]]Bottleneck[[:space:]]Models/249f4872-e5d6-464e-98af-dc5a4e7c1958_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/InfLoRA_[[:space:]]Interference-Free[[:space:]]Low-Rank[[:space:]]Adaptation[[:space:]]for[[:space:]]Continual[[:space:]]Learning/318c0ec9-6df7-445c-afc8-eff054c0abb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Infer[[:space:]]from[[:space:]]What[[:space:]]You[[:space:]]Have[[:space:]]Seen[[:space:]]Before_[[:space:]]Temporally-dependent[[:space:]]Classifier[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Video[[:space:]]Segmentation/b88320a5-3d99-4155-a16b-84fc816465a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Infinigen[[:space:]]Indoors_[[:space:]]Photorealistic[[:space:]]Indoor[[:space:]]Scenes[[:space:]]using[[:space:]]Procedural[[:space:]]Generation/6d8274a0-6d1c-4618-99cc-d4d3f62a5569_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Infrared[[:space:]]Adversarial[[:space:]]Car[[:space:]]Stickers/1b21f991-5369-4064-a439-e126cdaa2289_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Infrared[[:space:]]Small[[:space:]]Target[[:space:]]Detection[[:space:]]with[[:space:]]Scale[[:space:]]and[[:space:]]Location[[:space:]]Sensitivity/cc1c597c-b63f-4045-ac03-6fa4f61276e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/InitNO_[[:space:]]Boosting[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Initial[[:space:]]Noise[[:space:]]Optimization/ee9f1a65-ba75-4383-9c8b-786fd3d7ea61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Initialization[[:space:]]Matters[[:space:]]for[[:space:]]Adversarial[[:space:]]Transfer[[:space:]]Learning/22e1735b-5de4-4162-9aa3-d030e26b0bff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Ink[[:space:]]Dot-Oriented[[:space:]]Differentiable[[:space:]]Optimization[[:space:]]for[[:space:]]Neural[[:space:]]Image[[:space:]]Halftoning/1f12ac3c-71f4-433c-8965-472dc5121501_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Inlier[[:space:]]Confidence[[:space:]]Calibration[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/55d4a26e-41c5-4e7c-ab72-057a3c3654dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Insect-Foundation_[[:space:]]A[[:space:]]Foundation[[:space:]]Model[[:space:]]and[[:space:]]Large-scale[[:space:]]1M[[:space:]]Dataset[[:space:]]for[[:space:]]Visual[[:space:]]Insect[[:space:]]Understanding/15f153cc-fc03-4374-8076-9262f9d4ae0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Insights[[:space:]]from[[:space:]]the[[:space:]]Use[[:space:]]of[[:space:]]Previously[[:space:]]Unseen[[:space:]]Neural[[:space:]]Architecture[[:space:]]Search[[:space:]]Datasets/0098ce82-8d53-4ba3-8f97-4927160fad84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/InstaGen_[[:space:]]Enhancing[[:space:]]Object[[:space:]]Detection[[:space:]]by[[:space:]]Training[[:space:]]on[[:space:]]Synthetic[[:space:]]Dataset/44612d68-2249-4388-abaf-3d6e6fc8d9fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Instance[[:space:]]Tracking[[:space:]]in[[:space:]]3D[[:space:]]Scenes[[:space:]]from[[:space:]]Egocentric[[:space:]]Videos/7d6af3cb-cad2-4fee-942f-c5f4b1f4a49b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Instance-Adaptive[[:space:]]and[[:space:]]Geometric-Aware[[:space:]]Keypoint[[:space:]]Learning[[:space:]]for[[:space:]]Category-Level[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/2e346385-1067-495a-b77c-624a4cea3be4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Instance-Aware[[:space:]]Group[[:space:]]Quantization[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/ecc41151-6303-4521-9423-fe8c05533db0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Instance-aware[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Occluded[[:space:]]Human[[:space:]]Mesh[[:space:]]Reconstruction/fc45c75f-97ff-434b-b185-fb7682fd134e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Instance-aware[[:space:]]Exploration-Verification-Exploitation[[:space:]]for[[:space:]]Instance[[:space:]]ImageGoal[[:space:]]Navigation/22b62f8c-49cb-4dc8-b958-1da0c2cefd6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Instance-based[[:space:]]Max-margin[[:space:]]for[[:space:]]Practical[[:space:]]Few-shot[[:space:]]Recognition/661d9576-3318-460a-b4ae-7562837e1a63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Instance-level[[:space:]]Expert[[:space:]]Knowledge[[:space:]]and[[:space:]]Aggregate[[:space:]]Discriminative[[:space:]]Attention[[:space:]]for[[:space:]]Radiology[[:space:]]Report[[:space:]]Generation/67919adf-151b-4ec6-a2cc-6cf5e3e354ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/InstanceDiffusion_[[:space:]]Instance-level[[:space:]]Control[[:space:]]for[[:space:]]Image[[:space:]]Generation/979a8cba-a244-4c62-a32e-4d0bcb8d10d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/InstantBooth_[[:space:]]Personalized[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]without[[:space:]]Test-Time[[:space:]]Finetuning/bea3b9ed-e9c3-4e34-b981-6ada87e364bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Instantaneous[[:space:]]Perception[[:space:]]of[[:space:]]Moving[[:space:]]Objects[[:space:]]in[[:space:]]3D/528ec1d1-cd49-4f16-9f99-7eaa3281251a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Instruct[[:space:]]4D-to-4D_[[:space:]]Editing[[:space:]]4D[[:space:]]Scenes[[:space:]]as[[:space:]]Pseudo-3D[[:space:]]Scenes[[:space:]]Using[[:space:]]2D[[:space:]]Diffusion/c71052ce-4671-4ad1-9d21-b2e7d64253f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Instruct-Imagen_[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Multi-modal[[:space:]]Instruction/7233e90b-8d83-44c8-8a2a-00eb83bc144a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Instruct-ReID_[[:space:]]A[[:space:]]Multi-purpose[[:space:]]Person[[:space:]]Re-identification[[:space:]]Task[[:space:]]with[[:space:]]Instructions/3affda84-4bde-4c59-8075-137058b12704_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/InstructDiffusion_[[:space:]]A[[:space:]]Generalist[[:space:]]Modeling[[:space:]]Interface[[:space:]]for[[:space:]]Vision[[:space:]]Tasks/c99bbd74-c29a-4dea-b23f-d1f53e796d99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/InstructVideo_[[:space:]]Instructing[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Human[[:space:]]Feedback/84d0fc0b-5daa-451b-b2a3-4234b5d0f28f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Integrating[[:space:]]Efficient[[:space:]]Optimal[[:space:]]Transport[[:space:]]and[[:space:]]Functional[[:space:]]Maps[[:space:]]For[[:space:]]Unsupervised[[:space:]]Shape[[:space:]]Correspondence[[:space:]]Learning/ae83c0fa-5d87-4e18-8f15-f30a184d72c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Intelligent[[:space:]]Grimm[[:space:]]-[[:space:]]Open-ended[[:space:]]Visual[[:space:]]Storytelling[[:space:]]via[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models/126206c2-5bdf-4ccc-8a0a-06a25baf9956_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Intensity-Robust[[:space:]]Autofocus[[:space:]]for[[:space:]]Spike[[:space:]]Camera/9cad1a17-df16-45f4-99a2-491426f180c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Inter-X_[[:space:]]Towards[[:space:]]Versatile[[:space:]]Human-Human[[:space:]]Interaction[[:space:]]Analysis/47ac4442-3f8e-4524-97d5-ac578cc53f1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/InterHandGen_[[:space:]]Two-Hand[[:space:]]Interaction[[:space:]]Generation[[:space:]]via[[:space:]]Cascaded[[:space:]]Reverse[[:space:]]Diffusion/1b47c03b-e23c-41ad-a0ee-248a8c6765b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/InteractDiffusion_[[:space:]]Interaction[[:space:]]Control[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/f34942c0-2fa8-4ed8-88ee-0d1cd97e5ddc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Interactive[[:space:]]Continual[[:space:]]Learning_[[:space:]]Fast[[:space:]]and[[:space:]]Slow[[:space:]]Thinking/e7e30e56-eb02-4f66-9210-f693b2413289_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Interactive3D_[[:space:]]Create[[:space:]]What[[:space:]]You[[:space:]]Want[[:space:]]by[[:space:]]Interactive[[:space:]]3D[[:space:]]Generation/fd874c4a-afb5-4fc7-ab3a-3ff82d88db44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/InternVL_[[:space:]]Scaling[[:space:]]up[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models[[:space:]]and[[:space:]]Aligning[[:space:]]for[[:space:]]Generic[[:space:]]Visual-Linguistic[[:space:]]Tasks/95a44f72-908f-4ae2-89bc-02e0755ecc7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Interpretable[[:space:]]Measures[[:space:]]of[[:space:]]Conceptual[[:space:]]Similarity[[:space:]]by[[:space:]]Complexity-Constrained[[:space:]]Descriptive[[:space:]]Auto-Encoding/d4756190-a2f3-4167-b5ce-109ea65f4bf3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Intraoperative[[:space:]]2D_3D[[:space:]]Image[[:space:]]Registration[[:space:]]via[[:space:]]Differentiable[[:space:]]X-ray[[:space:]]Rendering/2347fdeb-f017-4284-87d2-311b5576393f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Intriguing[[:space:]]Properties[[:space:]]of[[:space:]]Diffusion[[:space:]]Models_[[:space:]]An[[:space:]]Empirical[[:space:]]Study[[:space:]]of[[:space:]]the[[:space:]]Natural[[:space:]]Attack[[:space:]]Capability[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Generative[[:space:]]Models/18b5c875-3add-43ca-8d09-62424539991b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Intrinsic[[:space:]]Image[[:space:]]Diffusion[[:space:]]for[[:space:]]Indoor[[:space:]]Single-view[[:space:]]Material[[:space:]]Estimation/bb5a353b-93e5-4149-b027-f27149203dee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/IntrinsicAvatar_[[:space:]]Physically[[:space:]]Based[[:space:]]Inverse[[:space:]]Rendering[[:space:]]of[[:space:]]Dynamic[[:space:]]Humans[[:space:]]from[[:space:]]Monocular[[:space:]]Videos[[:space:]]via[[:space:]]Explicit[[:space:]]Ray[[:space:]]Tracing/39bf7692-af06-4f9d-aae4-42f873ed1d70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Inverse[[:space:]]Rendering[[:space:]]of[[:space:]]Glossy[[:space:]]Objects[[:space:]]via[[:space:]]the[[:space:]]Neural[[:space:]]Plenoptic[[:space:]]Function[[:space:]]and[[:space:]]Radiance[[:space:]]Fields/23f34a75-31ba-44e6-a9ab-c706e127d2e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Inversion-Free[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Language-Guided[[:space:]]Diffusion[[:space:]]Models/2e86c04e-1fe1-447c-91e8-2d914f4bf6ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Investigating[[:space:]]Compositional[[:space:]]Challenges[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models[[:space:]]for[[:space:]]Visual[[:space:]]Grounding/73ebfedb-e337-4a97-875d-44130de34e7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Investigating[[:space:]]and[[:space:]]Mitigating[[:space:]]the[[:space:]]Side[[:space:]]Effects[[:space:]]of[[:space:]]Noisy[[:space:]]Views[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Clustering[[:space:]]Algorithms[[:space:]]in[[:space:]]Practical[[:space:]]Multi-View[[:space:]]Scenarios/296c5e3a-0e1f-4e58-b870-c1a7ec20a408_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Is[[:space:]]Ego[[:space:]]Status[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]for[[:space:]]Open-Loop[[:space:]]End-to-End[[:space:]]Autonomous[[:space:]]Driving_/86d09793-b278-4d07-ac0f-5340b547cf36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Is[[:space:]]Vanilla[[:space:]]MLP[[:space:]]in[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field[[:space:]]Enough[[:space:]]for[[:space:]]Few-shot[[:space:]]View[[:space:]]Synthesis_/e09e1938-b750-43a2-bb40-f46d36ac614f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/It's[[:space:]]All[[:space:]]About[[:space:]]Your[[:space:]]Sketch_[[:space:]]Democratising[[:space:]]Sketch[[:space:]]Control[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/92ff8967-87c2-48b1-bfeb-03e6a355c73b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Iterated[[:space:]]Learning[[:space:]]Improves[[:space:]]Compositionality[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/c32a4552-51b5-4663-b99c-db7713f93b6b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/JDEC_[[:space:]]JPEG[[:space:]]Decoding[[:space:]]via[[:space:]]Enhanced[[:space:]]Continuous[[:space:]]Cosine[[:space:]]Coefficients/27f544e1-0ebf-49f5-9075-a3186a9d99b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/JRDB-PanoTrack_[[:space:]]An[[:space:]]Open-world[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]and[[:space:]]Tracking[[:space:]]Robotic[[:space:]]Dataset[[:space:]]in[[:space:]]Crowded[[:space:]]Human[[:space:]]Environments/406910bd-7952-4292-b4e2-140744032f5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/JRDB-Social_[[:space:]]A[[:space:]]Multifaceted[[:space:]]Robotic[[:space:]]Dataset[[:space:]]for[[:space:]]Understanding[[:space:]]of[[:space:]]Context[[:space:]]and[[:space:]]Dynamics[[:space:]]of[[:space:]]Human[[:space:]]Interactions[[:space:]]Within[[:space:]]Social[[:space:]]Groups/11d99d1c-9420-4de0-aea8-4a503d551d2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Jack[[:space:]]of[[:space:]]All[[:space:]]Tasks[[:space:]]Master[[:space:]]of[[:space:]]Many_[[:space:]]Designing[[:space:]]General-Purpose[[:space:]]Coarse-to-Fine[[:space:]]Vision-Language[[:space:]]Model/167ea0c1-9747-41cd-a16f-62e408e4f09d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/JeDi_[[:space:]]Joint-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Finetuning-Free[[:space:]]Personalized[[:space:]]Text-to-Image[[:space:]]Generation/7dab8307-ee57-4132-96d3-222282bda0db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/JoAPR_[[:space:]]Cleaning[[:space:]]the[[:space:]]Lens[[:space:]]of[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/faaeccfc-977a-48e1-9008-dfb0148d6d55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Joint[[:space:]]Reconstruction[[:space:]]of[[:space:]]3D[[:space:]]Human[[:space:]]and[[:space:]]Object[[:space:]]via[[:space:]]Contact-Based[[:space:]]Refinement[[:space:]]Transformer/dd38f93f-59cb-4dc8-b8ec-e7c9f8e4551d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Joint-Task[[:space:]]Regularization[[:space:]]for[[:space:]]Partially[[:space:]]Labeled[[:space:]]Multi-Task[[:space:]]Learning/99c5d09d-e2e9-4cea-82f7-968af58bf0d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Joint2Human_[[:space:]]High-Quality[[:space:]]3D[[:space:]]Human[[:space:]]Generation[[:space:]]via[[:space:]]Compact[[:space:]]Spherical[[:space:]]Embedding[[:space:]]of[[:space:]]3D[[:space:]]Joints/8082490c-6f81-4aa6-82e4-b76bdf8f9443_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/JointSQ_[[:space:]]Joint[[:space:]]Sparsification-Quantization[[:space:]]for[[:space:]]Distributed[[:space:]]Learning/96355f00-9794-442f-aaf6-b0018c6ec8bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Jointly[[:space:]]Training[[:space:]]and[[:space:]]Pruning[[:space:]]CNNs[[:space:]]via[[:space:]]Learnable[[:space:]]Agent[[:space:]]Guidance[[:space:]]and[[:space:]]Alignment/5a156a1c-059a-408c-a35e-3572fe293f39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Just[[:space:]]Add[[:space:]]_![[:space:]]Pose[[:space:]]Induced[[:space:]]Video[[:space:]]Transformers[[:space:]]for[[:space:]]Understanding[[:space:]]Activities[[:space:]]of[[:space:]]Daily[[:space:]]Living/4d57a961-3422-4860-8ed5-39fac46697f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/KD-DETR_[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Detection[[:space:]]Transformer[[:space:]]with[[:space:]]Consistent[[:space:]]Distillation[[:space:]]Points[[:space:]]Sampling/6f20064b-c3c9-449a-b4f5-131f53075b89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/KITRO_[[:space:]]Refining[[:space:]]Human[[:space:]]Mesh[[:space:]]by[[:space:]]2D[[:space:]]Clues[[:space:]]and[[:space:]]Kinematic-tree[[:space:]]Rotation/46b45f5b-b183-46f6-a225-f5b7078984cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/KP-RED_[[:space:]]Exploiting[[:space:]]Semantic[[:space:]]Keypoints[[:space:]]for[[:space:]]Joint[[:space:]]3D[[:space:]]Shape[[:space:]]Retrieval[[:space:]]and[[:space:]]Deformation/69a28ad2-695b-47fd-85de-1daf3d9953a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/KPConvX_[[:space:]]Modernizing[[:space:]]Kernel[[:space:]]Point[[:space:]]Convolution[[:space:]]with[[:space:]]Kernel[[:space:]]Attention/52c793bc-43cc-4724-a336-4e36ddb335b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/KTPFormer_[[:space:]]Kinematics[[:space:]]and[[:space:]]Trajectory[[:space:]]Prior[[:space:]]Knowledge-Enhanced[[:space:]]Transformer[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/1c1ad04d-093b-41ab-833b-1018c31bf8f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/KVQ_[[:space:]]Kwai[[:space:]]Video[[:space:]]Quality[[:space:]]Assessment[[:space:]]for[[:space:]]Short-form[[:space:]]Videos/ac8c14df-e58b-440b-821c-26d5dd010f64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Kandinsky[[:space:]]Conformal[[:space:]]Prediction_[[:space:]]Efficient[[:space:]]Calibration[[:space:]]of[[:space:]]Image[[:space:]]Segmentation[[:space:]]Algorithms/818151a0-bd42-4648-b996-432e21e6d8ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Kernel[[:space:]]Adaptive[[:space:]]Convolution[[:space:]]for[[:space:]]Scene[[:space:]]Text[[:space:]]Detection[[:space:]]via[[:space:]]Distance[[:space:]]Map[[:space:]]Prediction/4d1dddc2-37a8-42c2-988c-7a6df1ddf55a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/KeyPoint[[:space:]]Relative[[:space:]]Position[[:space:]]Encoding[[:space:]]for[[:space:]]Face[[:space:]]Recognition/6584e83c-ac98-42da-ac32-f42b79f313cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Know[[:space:]]Your[[:space:]]Neighbors_[[:space:]]Improving[[:space:]]Single-View[[:space:]]Reconstruction[[:space:]]via[[:space:]]Spatial[[:space:]]Vision-Language[[:space:]]Reasoning/390b3fb4-20df-49e0-ade3-0a7ab9e80dc3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Knowledge-Enhanced[[:space:]]Dual-stream[[:space:]]Zero-shot[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/24b35b7f-5f59-4e24-84f9-be98666c6d47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Koala_[[:space:]]Key[[:space:]]Frame-Conditioned[[:space:]]Long[[:space:]]Video-LLM/80db729a-94ba-4cf1-8539-8da85474977c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/L-MAGIC_[[:space:]]Language[[:space:]]Model[[:space:]]Assisted[[:space:]]Generation[[:space:]]of[[:space:]]Images[[:space:]]with[[:space:]]Coherence/f3e59faf-bf52-4d27-91e4-3d2051082f96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/L0-Sampler_[[:space:]]An[[:space:]]L0[[:space:]]Model[[:space:]]Guided[[:space:]]Volume[[:space:]]Sampling[[:space:]]for[[:space:]]NeRF/754ee175-248f-48d1-a54a-fd097bf352fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/L2B_[[:space:]]Learning[[:space:]]to[[:space:]]Bootstrap[[:space:]]Robust[[:space:]]Models[[:space:]]for[[:space:]]Combating[[:space:]]Label[[:space:]]Noise/bf15d88a-6052-4b9f-a559-7ad50946da26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/L4D-Track_[[:space:]]Language-to-4D[[:space:]]Modeling[[:space:]]Towards[[:space:]]6-DoF[[:space:]]Tracking[[:space:]]and[[:space:]]Shape[[:space:]]Reconstruction[[:space:]]in[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Stream/de084167-0bc6-4fc8-bcc3-d870d435c9a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LAA-Net_[[:space:]]Localized[[:space:]]Artifact[[:space:]]Attention[[:space:]]Network[[:space:]]for[[:space:]]Quality-Agnostic[[:space:]]and[[:space:]]Generalizable[[:space:]]Deepfake[[:space:]]Detection/2f5f5240-ee33-4e5f-a2e8-af8274bc0a1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LAENeRF_[[:space:]]Local[[:space:]]Appearance[[:space:]]Editing[[:space:]]for[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/5c6c265c-7e58-4cec-af54-828e52acfd31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LAFS_[[:space:]]Landmark-based[[:space:]]Facial[[:space:]]Self-supervised[[:space:]]Learning[[:space:]]for[[:space:]]Face[[:space:]]Recognition/e9e9e89c-eba1-40b9-b0fa-b9045d32f556_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LAKE-RED_[[:space:]]Camouflaged[[:space:]]Images[[:space:]]Generation[[:space:]]by[[:space:]]Latent[[:space:]]Background[[:space:]]Knowledge[[:space:]]Retrieval-Augmented[[:space:]]Diffusion/12722ca1-ee01-4244-ac76-dd0f56c3ada2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LAMP_[[:space:]]Learn[[:space:]]A[[:space:]]Motion[[:space:]]Pattern[[:space:]]for[[:space:]]Few-Shot[[:space:]]Video[[:space:]]Generation/a527aaf0-4fd2-4061-87f8-d8fd7f439d09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LAN_[[:space:]]Learning[[:space:]]to[[:space:]]Adapt[[:space:]]Noise[[:space:]]for[[:space:]]Image[[:space:]]Denoising/74bb68f0-698b-4fd3-96c1-35383b7bb71f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LASA_[[:space:]]Instance[[:space:]]Reconstruction[[:space:]]from[[:space:]]Real[[:space:]]Scans[[:space:]]using[[:space:]]A[[:space:]]Large-scale[[:space:]]Aligned[[:space:]]Shape[[:space:]]Annotation[[:space:]]Dataset/e66a9d83-37a7-45da-8099-3af5a6b6fd8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LASIL_[[:space:]]Learner-Aware[[:space:]]Supervised[[:space:]]Imitation[[:space:]]Learning[[:space:]]For[[:space:]]Long-term[[:space:]]Microscopic[[:space:]]Traffic[[:space:]]Simulation/44a62bf3-d838-46a8-bbd5-a317e9666352_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LASO_[[:space:]]Language-guided[[:space:]]Affordance[[:space:]]Segmentation[[:space:]]on[[:space:]]3D[[:space:]]Object/4a958b96-aed8-4e26-96c9-3b681d296992_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LDP_[[:space:]]Language-driven[[:space:]]Dual-Pixel[[:space:]]Image[[:space:]]Defocus[[:space:]]Deblurring[[:space:]]Network/2d70b048-ecc7-45e4-b194-27af8843d7f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LEAD_[[:space:]]Exploring[[:space:]]Logit[[:space:]]Space[[:space:]]Evolution[[:space:]]for[[:space:]]Model[[:space:]]Selection/cac4214c-e8ea-4c89-8778-e0818b32709b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LEAD_[[:space:]]Learning[[:space:]]Decomposition[[:space:]]for[[:space:]]Source-free[[:space:]]Universal[[:space:]]Domain[[:space:]]Adaptation/68e710b5-658b-4bb7-a5db-feae6ea998dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LEAP-VO_[[:space:]]Long-term[[:space:]]Effective[[:space:]]Any[[:space:]]Point[[:space:]]Tracking[[:space:]]for[[:space:]]Visual[[:space:]]Odometry/03c2e012-3152-4602-8d61-0831334e6fb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LEDITS++_[[:space:]]Limitless[[:space:]]Image[[:space:]]Editing[[:space:]]using[[:space:]]Text-to-Image[[:space:]]Models/f8bff481-3f32-424c-92f1-b66a29a1df02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LED_[[:space:]]A[[:space:]]Large-scale[[:space:]]Real-world[[:space:]]Paired[[:space:]]Dataset[[:space:]]for[[:space:]]Event[[:space:]]Camera[[:space:]]Denoising/c132a07a-cde4-41ed-ba0e-fbc21a44ee59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LEMON_[[:space:]]Learning[[:space:]]3D[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Relation[[:space:]]from[[:space:]]2D[[:space:]]Images/2d6d94d5-881a-4f31-92c0-fc2ad5248536_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LEOD_[[:space:]]Label-Efficient[[:space:]]Object[[:space:]]Detection[[:space:]]for[[:space:]]Event[[:space:]]Cameras/dfaca580-4023-4cf1-b4f1-9b27b658deef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LION_[[:space:]]Empowering[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]with[[:space:]]Dual-Level[[:space:]]Visual[[:space:]]Knowledge/ee456ca8-65df-48b8-a118-352557caef54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LISA_[[:space:]]Reasoning[[:space:]]Segmentation[[:space:]]via[[:space:]]Large[[:space:]]Language[[:space:]]Model/92907404-c48e-45be-bfa1-4c27e2c673e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LL3DA_[[:space:]]Visual[[:space:]]Interactive[[:space:]]Instruction[[:space:]]Tuning[[:space:]]for[[:space:]]Omni-3D[[:space:]]Understanding[[:space:]]Reasoning[[:space:]]and[[:space:]]Planning/e5af2afb-e504-4d6c-99d0-a3f72bf3d45e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LLM4SGG_[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/a2404cd0-4571-42a9-a243-243f9def1b0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LLMs[[:space:]]are[[:space:]]Good[[:space:]]Action[[:space:]]Recognizers/8e366b59-bab5-4965-a328-ae40f7bddcd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LLMs[[:space:]]are[[:space:]]Good[[:space:]]Sign[[:space:]]Language[[:space:]]Translators/09537dfb-17f4-403d-a9c1-d3890b661d09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LLaFS_[[:space:]]When[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]Meet[[:space:]]Few-Shot[[:space:]]Segmentation/aef3f26f-2462-4dec-ad92-ab4a881db4c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LLaMA-Excitor_[[:space:]]General[[:space:]]Instruction[[:space:]]Tuning[[:space:]]via[[:space:]]Indirect[[:space:]]Feature[[:space:]]Interaction/54e482bc-9979-46be-8235-fd3729928e88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LMDrive_[[:space:]]Closed-Loop[[:space:]]End-to-End[[:space:]]Driving[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Models/03e61345-cacb-4ace-90cd-06b4f58ee7fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LORS_[[:space:]]Low-rank[[:space:]]Residual[[:space:]]Structure[[:space:]]for[[:space:]]Parameter-Efficient[[:space:]]Network[[:space:]]Stacking/2ca03dd9-d123-452f-a1c7-2252d386e5e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LOTUS_[[:space:]]Evasive[[:space:]]and[[:space:]]Resilient[[:space:]]Backdoor[[:space:]]Attacks[[:space:]]through[[:space:]]Sub-Partitioning/6b71bad9-734e-4c81-9b75-13613bad7df8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LP++_[[:space:]]A[[:space:]]Surprisingly[[:space:]]Strong[[:space:]]Linear[[:space:]]Probe[[:space:]]for[[:space:]]Few-Shot[[:space:]]CLIP/d1daeba3-e67f-49fe-b6c7-0613f97689a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LPSNet_[[:space:]]End-to-End[[:space:]]Human[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Estimation[[:space:]]with[[:space:]]Lensless[[:space:]]Imaging/3f527010-7aaa-4b2e-ab0a-89e9d8eb4cef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LQMFormer_[[:space:]]Language-aware[[:space:]]Query[[:space:]]Mask[[:space:]]Transformer[[:space:]]for[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation/c24a3bea-55f3-42b0-b8e6-0dc981e7639b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LSK3DNet_[[:space:]]Towards[[:space:]]Effective[[:space:]]and[[:space:]]Efficient[[:space:]]3D[[:space:]]Perception[[:space:]]with[[:space:]]Large[[:space:]]Sparse[[:space:]]Kernels/4a7b9d7e-a440-4471-acbb-bbb97d0681c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LTA-PCS_[[:space:]]Learnable[[:space:]]Task-Agnostic[[:space:]]Point[[:space:]]Cloud[[:space:]]Sampling/c6e80379-91d5-42d2-871e-0e2f7374da5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LTGC_[[:space:]]Long-tail[[:space:]]Recognition[[:space:]]via[[:space:]]Leveraging[[:space:]]LLMs-driven[[:space:]]Generated[[:space:]]Content/2e9dbf01-8020-47fa-831f-ee0b1bfd9643_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LTM_[[:space:]]Lightweight[[:space:]]Textured[[:space:]]Mesh[[:space:]]Extraction[[:space:]]and[[:space:]]Refinement[[:space:]]of[[:space:]]Large[[:space:]]Unbounded[[:space:]]Scenes[[:space:]]for[[:space:]]Efficient[[:space:]]Storage[[:space:]]and[[:space:]]Real-time[[:space:]]Rendering/b965bd98-b1d7-4faf-8f20-171bc6260c15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LUWA[[:space:]]Dataset_[[:space:]]Learning[[:space:]]Lithic[[:space:]]Use-Wear[[:space:]]Analysis[[:space:]]on[[:space:]]Microscopic[[:space:]]Images/962a883b-464b-4830-ab92-8bd978dfedca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LaMPilot_[[:space:]]An[[:space:]]Open[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving[[:space:]]with[[:space:]]Language[[:space:]]Model[[:space:]]Programs/c2cae23c-06f6-48c3-8fe3-d3a2dc677cb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LaRE^2_[[:space:]]Latent[[:space:]]Reconstruction[[:space:]]Error[[:space:]]Based[[:space:]]Method[[:space:]]for[[:space:]]Diffusion-Generated[[:space:]]Image[[:space:]]Detection/fbde4cbd-2dff-4663-bad3-50e64c26068e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Label[[:space:]]Propagation[[:space:]]for[[:space:]]Zero-shot[[:space:]]Classification[[:space:]]with[[:space:]]Vision-Language[[:space:]]Models/f789bcea-03da-4382-8fd5-69048545fa3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Label-Efficient[[:space:]]Group[[:space:]]Robustness[[:space:]]via[[:space:]]Out-of-Distribution[[:space:]]Concept[[:space:]]Curation/71c0abd3-bdae-44a0-ba02-b4e8dc4dac4c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Lane2Seq_[[:space:]]Towards[[:space:]]Unified[[:space:]]Lane[[:space:]]Detection[[:space:]]via[[:space:]]Sequence[[:space:]]Generation/0b70847d-c1fa-4821-9d6c-0fc3de9d8952_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LaneCPP_[[:space:]]Continuous[[:space:]]3D[[:space:]]Lane[[:space:]]Detection[[:space:]]using[[:space:]]Physical[[:space:]]Priors/7a911166-f711-4534-9bf2-2cb8d0604f4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LangSplat_[[:space:]]3D[[:space:]]Language[[:space:]]Gaussian[[:space:]]Splatting/591ba264-2503-43cc-8f70-dc10b9821047_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Language[[:space:]]Embedded[[:space:]]3D[[:space:]]Gaussians[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Scene[[:space:]]Understanding/887164f8-5ac1-4491-be92-996acbb7aa0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Language[[:space:]]Model[[:space:]]Guided[[:space:]]Interpretable[[:space:]]Video[[:space:]]Action[[:space:]]Reasoning/f807e59b-d0ea-4264-a8c5-fca1bb6bb589_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Language[[:space:]]Models[[:space:]]as[[:space:]]Black-Box[[:space:]]Optimizers[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/40b8d27b-a1d9-46c0-919c-6079601bc412_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Language-Driven[[:space:]]Anchors[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Adversarial[[:space:]]Robustness/5bd9567c-d374-4782-9305-9ad2d0f7be37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Language-aware[[:space:]]Visual[[:space:]]Semantic[[:space:]]Distillation[[:space:]]for[[:space:]]Video[[:space:]]Question[[:space:]]Answering/f0849498-5020-4fb0-b9e6-adb27b80af64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Language-conditioned[[:space:]]Detection[[:space:]]Transformer/4052d6d0-e8f9-4912-885e-f4d1033b20c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Language-driven[[:space:]]All-in-one[[:space:]]Adverse[[:space:]]Weather[[:space:]]Removal/f89b79b0-6a22-4bde-912e-e3cada02d463_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Language-driven[[:space:]]Grasp[[:space:]]Detection/4da4eafd-acd2-41ed-99a7-76c7e2b26e40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Language-driven[[:space:]]Object[[:space:]]Fusion[[:space:]]into[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]with[[:space:]]Pose-Conditioned[[:space:]]Dataset[[:space:]]Updates/af532a3c-a441-40c6-a819-aabe1e6f5939_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Language-guided[[:space:]]Image[[:space:]]Reflection[[:space:]]Separation/2783b2be-5a5b-4ab1-8b98-8dc1d3acc65d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Language-only[[:space:]]Training[[:space:]]of[[:space:]]Zero-shot[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/35e5e309-9782-4a25-8b32-4eac66825f83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Laplacian-guided[[:space:]]Entropy[[:space:]]Model[[:space:]]in[[:space:]]Neural[[:space:]]Codec[[:space:]]with[[:space:]]Blur-dissipated[[:space:]]Synthesis/3c1b4fc3-944d-4df5-a8e9-c0d68cc7d2d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Large[[:space:]]Language[[:space:]]Models[[:space:]]are[[:space:]]Good[[:space:]]Prompt[[:space:]]Learners[[:space:]]for[[:space:]]Low-Shot[[:space:]]Image[[:space:]]Classification/9c680495-254d-41be-9282-c0c5d7c9a2e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Latency[[:space:]]Correction[[:space:]]for[[:space:]]Event-guided[[:space:]]Deblurring[[:space:]]and[[:space:]]Frame[[:space:]]Interpolation/29c2331e-d68b-477e-86d2-3c185688337a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Latent[[:space:]]Modulated[[:space:]]Function[[:space:]]for[[:space:]]Computational[[:space:]]Optimal[[:space:]]Continuous[[:space:]]Image[[:space:]]Representation/637c4edb-98b1-40b4-8db4-8d4aa7226fc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Layout-Agnostic[[:space:]]Scene[[:space:]]Text[[:space:]]Image[[:space:]]Synthesis[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/cffc0cc1-9751-4215-8102-3e240ce1394d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LayoutFormer_[[:space:]]Hierarchical[[:space:]]Text[[:space:]]Detection[[:space:]]Towards[[:space:]]Scene[[:space:]]Text[[:space:]]Understanding/09e4cb52-2d02-4e8e-9918-66498721ecee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LayoutLLM_[[:space:]]Layout[[:space:]]Instruction[[:space:]]Tuning[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Document[[:space:]]Understanding/a3bb284c-623e-4013-87c6-088a22a81e99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LeGO_[[:space:]]Leveraging[[:space:]]a[[:space:]]Surface[[:space:]]Deformation[[:space:]]Network[[:space:]]for[[:space:]]Animatable[[:space:]]Stylized[[:space:]]Face[[:space:]]Generation[[:space:]]with[[:space:]]One[[:space:]]Example/6836350f-601e-42fa-b6e0-49050e114ece_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Leak[[:space:]]and[[:space:]]Learn_[[:space:]]An[[:space:]]Attacker's[[:space:]]Cookbook[[:space:]]to[[:space:]]Train[[:space:]]Using[[:space:]]Leaked[[:space:]]Data[[:space:]]from[[:space:]]Federated[[:space:]]Learning/c7bd675a-7448-4fdd-958a-3d9af33c562c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learn[[:space:]]from[[:space:]]View[[:space:]]Correlation_[[:space:]]An[[:space:]]Anchor[[:space:]]Enhancement[[:space:]]Strategy[[:space:]]for[[:space:]]Multi-view[[:space:]]Clustering/4eed6c41-c8e8-44a9-84e6-f4256d6b5e7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learn[[:space:]]to[[:space:]]Rectify[[:space:]]the[[:space:]]Bias[[:space:]]of[[:space:]]CLIP[[:space:]]for[[:space:]]Unsupervised[[:space:]]Semantic[[:space:]]Segmentation/4af240bb-fb73-4b13-b278-60a5b44fd376_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learnable[[:space:]]Earth[[:space:]]Parser_[[:space:]]Discovering[[:space:]]3D[[:space:]]Prototypes[[:space:]]in[[:space:]]Aerial[[:space:]]Scans/5317043d-970e-43ef-b2f4-dc56567d2d02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learned[[:space:]]Lossless[[:space:]]Image[[:space:]]Compression[[:space:]]based[[:space:]]on[[:space:]]Bit[[:space:]]Plane[[:space:]]Slicing/493cad4c-eae1-44ac-813f-9e0c42d64707_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learned[[:space:]]Representation-Guided[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Large-Image[[:space:]]Generation/5695d81c-8814-42a9-899f-2c8baf3ccbb3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learned[[:space:]]Scanpaths[[:space:]]Aid[[:space:]]Blind[[:space:]]Panoramic[[:space:]]Video[[:space:]]Quality[[:space:]]Assessment/4097298f-874f-411d-9a63-b8f9db525955_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learned[[:space:]]Trajectory[[:space:]]Embedding[[:space:]]for[[:space:]]Subspace[[:space:]]Clustering/37fd4122-aae1-41d4-895f-bd90782ed866_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Adaptive[[:space:]]Spatial[[:space:]]Coherent[[:space:]]Correlations[[:space:]]for[[:space:]]Speech-Preserving[[:space:]]Facial[[:space:]]Expression[[:space:]]Manipulation/199bb55b-1c98-451b-b065-6d0fccbafb25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Background[[:space:]]Prompts[[:space:]]to[[:space:]]Discover[[:space:]]Implicit[[:space:]]Knowledge[[:space:]]for[[:space:]]Open[[:space:]]Vocabulary[[:space:]]Object[[:space:]]Detection/57bc7ecd-3bfc-4e56-9aff-1de8b9713c30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]CNN[[:space:]]on[[:space:]]ViT_[[:space:]]A[[:space:]]Hybrid[[:space:]]Model[[:space:]]to[[:space:]]Explicitly[[:space:]]Class-specific[[:space:]]Boundaries[[:space:]]for[[:space:]]Domain[[:space:]]Adaptation/f88da0d8-496d-49e1-8c98-717f50be5563_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Continual[[:space:]]Compatible[[:space:]]Representation[[:space:]]for[[:space:]]Re-indexing[[:space:]]Free[[:space:]]Lifelong[[:space:]]Person[[:space:]]Re-identification/68315fca-b218-4982-8c8a-e271e8d926b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Continuous[[:space:]]3D[[:space:]]Words[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/276644cc-5935-4e67-b150-37736c673436_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Correlation[[:space:]]Structures[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/3e02995b-1bff-4a49-8649-a6ad79f7017d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Coupled[[:space:]]Dictionaries[[:space:]]from[[:space:]]Unpaired[[:space:]]Data[[:space:]]for[[:space:]]Image[[:space:]]Super-Resolution/14c29cf4-a5f6-481c-be92-86ff4e2cfb38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Degradation-Independent[[:space:]]Representations[[:space:]]for[[:space:]]Camera[[:space:]]ISP[[:space:]]Pipelines/9184b776-09c8-4040-9313-78bc29f34068_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Degradation-unaware[[:space:]]Representation[[:space:]]with[[:space:]]Prior-based[[:space:]]Latent[[:space:]]Transformations[[:space:]]for[[:space:]]Blind[[:space:]]Face[[:space:]]Restoration/00121fc6-ccac-4774-942c-2646f683b09a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Diffusion[[:space:]]Texture[[:space:]]Priors[[:space:]]for[[:space:]]Image[[:space:]]Restoration/c4fdbfda-8abc-47c5-aabf-b69c62343d36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Discriminative[[:space:]]Dynamics[[:space:]]with[[:space:]]Label[[:space:]]Corruption[[:space:]]for[[:space:]]Noisy[[:space:]]Label[[:space:]]Detection/9326fcaf-0cf3-4eae-8525-ad22e7243842_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Disentangled[[:space:]]Identifiers[[:space:]]for[[:space:]]Action-Customized[[:space:]]Text-to-Image[[:space:]]Generation/9257612f-7f23-4c8a-85fd-19ebf525ae6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Dynamic[[:space:]]Tetrahedra[[:space:]]for[[:space:]]High-Quality[[:space:]]Talking[[:space:]]Head[[:space:]]Synthesis/53f925b0-849e-40c5-8f29-bb11fe50f467_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Equi-angular[[:space:]]Representations[[:space:]]for[[:space:]]Online[[:space:]]Continual[[:space:]]Learning/1f2bf4e3-a137-4287-b431-b0fd185698d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Group[[:space:]]Activity[[:space:]]Features[[:space:]]Through[[:space:]]Person[[:space:]]Attribute[[:space:]]Prediction/58724bbe-6f1a-4e5e-a5d2-557cd815fe3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Inclusion[[:space:]]Matching[[:space:]]for[[:space:]]Animation[[:space:]]Paint[[:space:]]Bucket[[:space:]]Colorization/3a58aaab-2a4a-445b-9bfe-6c6c88da7d90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Instance-Aware[[:space:]]Correspondences[[:space:]]for[[:space:]]Robust[[:space:]]Multi-Instance[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration[[:space:]]in[[:space:]]Cluttered[[:space:]]Scenes/2ddc8710-2e01-4d04-9c1a-627c67a89c1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Intra-view[[:space:]]and[[:space:]]Cross-view[[:space:]]Geometric[[:space:]]Knowledge[[:space:]]for[[:space:]]Stereo[[:space:]]Matching/2798eade-4610-47df-96d1-47a92d0431fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Large-Factor[[:space:]]EM[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]with[[:space:]]Generative[[:space:]]Priors/246e675a-bddb-4921-a098-1fe2784b0d78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Multi-Dimensional[[:space:]]Human[[:space:]]Preference[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/0686b0c0-a9d4-4259-972e-6fc4bad8bf88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Object[[:space:]]State[[:space:]]Changes[[:space:]]in[[:space:]]Videos_[[:space:]]An[[:space:]]Open-World[[:space:]]Perspective/fe5e8aba-0eb4-418b-9856-ff82ba536a86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Occupancy[[:space:]]for[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection/9f9633e2-2ceb-4ed8-94da-20b12a33cfae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]SO(3)-Invariant[[:space:]]Semantic[[:space:]]Correspondence[[:space:]]via[[:space:]]Local[[:space:]]Shape[[:space:]]Transform/6e54c33e-941d-421a-b1b4-ad1a347fd128_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Spatial[[:space:]]Adaptation[[:space:]]and[[:space:]]Temporal[[:space:]]Coherence[[:space:]]in[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Video[[:space:]]Super-Resolution/34c7419c-4229-4f4a-ad73-efdadc875a8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Spatial[[:space:]]Features[[:space:]]from[[:space:]]Audio-Visual[[:space:]]Correspondence[[:space:]]in[[:space:]]Egocentric[[:space:]]Videos/99328dcb-7d90-4f7e-aee5-421c7b55c1a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Structure-from-Motion[[:space:]]with[[:space:]]Graph[[:space:]]Attention[[:space:]]Networks/15e89f9b-21e7-4dd4-a256-0f698de897a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Transferable[[:space:]]Negative[[:space:]]Prompts[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection/5385f0b8-2f8e-4840-b283-fc050f4ae0b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Triangular[[:space:]]Distribution[[:space:]]in[[:space:]]Visual[[:space:]]World/7d1e6df8-db40-4bc4-aab8-58f2635d78ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Vision[[:space:]]from[[:space:]]Models[[:space:]]Rivals[[:space:]]Learning[[:space:]]Vision[[:space:]]from[[:space:]]Data/b7bae39a-3af5-4fdf-bc33-43a030ca35b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]Visual[[:space:]]Prompt[[:space:]]for[[:space:]]Gait[[:space:]]Recognition/250e53d0-2beb-4a17-986d-3fa0bac7cb32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]by[[:space:]]Correction_[[:space:]]Efficient[[:space:]]Tuning[[:space:]]Task[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Generative[[:space:]]Vision-Language[[:space:]]Reasoning/b024fa39-4df3-49c4-9b05-b3492e317bba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]for[[:space:]]Transductive[[:space:]]Threshold[[:space:]]Calibration[[:space:]]in[[:space:]]Open-World[[:space:]]Recognition/487c5882-cae9-4d51-9135-8fd112247491_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]from[[:space:]]Observer[[:space:]]Gaze_[[:space:]]Zero-Shot[[:space:]]Attention[[:space:]]Prediction[[:space:]]Oriented[[:space:]]by[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Recognition/111dd706-825f-4535-846b-ae607fa98e28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]from[[:space:]]One[[:space:]]Continuous[[:space:]]Video[[:space:]]Stream/a8db15d8-1180-483d-9687-c41b3735ca04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]from[[:space:]]Synthetic[[:space:]]Human[[:space:]]Group[[:space:]]Activities/804774d4-d5b8-4f49-9724-9eca71443605_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]the[[:space:]]3D[[:space:]]Fauna[[:space:]]of[[:space:]]the[[:space:]]Web/dd8fb285-42f8-4fbe-a149-b011c10a95c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Control[[:space:]]Camera[[:space:]]Exposure[[:space:]]via[[:space:]]Reinforcement[[:space:]]Learning/e1678f2e-e73f-4113-9734-14288276a0a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Count[[:space:]]without[[:space:]]Annotations/d11673d4-2f49-4551-9017-c5ac90ac3421_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Localize[[:space:]]Objects[[:space:]]Improves[[:space:]]Spatial[[:space:]]Reasoning[[:space:]]in[[:space:]]Visual-LLMs/7cd84fbe-5225-4590-b9bf-3163d999924f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Navigate[[:space:]]Efficiently[[:space:]]and[[:space:]]Precisely[[:space:]]in[[:space:]]Real[[:space:]]Environments/620f678d-eca2-450f-9ffd-9c056c561f0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Predict[[:space:]]Activity[[:space:]]Progress[[:space:]]by[[:space:]]Self-Supervised[[:space:]]Video[[:space:]]Alignment/6059053b-0714-4369-b348-437b5de261d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Produce[[:space:]]Semi-dense[[:space:]]Correspondences[[:space:]]for[[:space:]]Visual[[:space:]]Localization/3fb41c79-8021-41c2-9eec-bd21aa2928ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Rank[[:space:]]Patches[[:space:]]for[[:space:]]Unbiased[[:space:]]Image[[:space:]]Redundancy[[:space:]]Reduction/77fab6af-6f9d-408d-9dae-8e7bad3bdc39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Rematch[[:space:]]Mismatched[[:space:]]Pairs[[:space:]]for[[:space:]]Robust[[:space:]]Cross-Modal[[:space:]]Retrieval/fd35e1ac-d332-4502-a9ba-6902d0ab95b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Remove[[:space:]]Wrinkled[[:space:]]Transparent[[:space:]]Film[[:space:]]with[[:space:]]Polarized[[:space:]]Prior/5e25a466-41ef-4842-a8b6-ee1cd66325a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Segment[[:space:]]Referred[[:space:]]Objects[[:space:]]from[[:space:]]Narrated[[:space:]]Egocentric[[:space:]]Videos/13b15e2b-026f-47a4-a5c8-eb8241c9c8e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Select[[:space:]]Views[[:space:]]for[[:space:]]Efficient[[:space:]]Multi-View[[:space:]]Understanding/6f1ed185-f83a-4de1-ba41-bd094a6eccd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Transform[[:space:]]Dynamically[[:space:]]for[[:space:]]Better[[:space:]]Adversarial[[:space:]]Transferability/20178bd8-b1c9-4630-a7e2-243d8c197a2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]to[[:space:]]Visually[[:space:]]Localize[[:space:]]Sound[[:space:]]Sources[[:space:]]from[[:space:]]Mixtures[[:space:]]without[[:space:]]Prior[[:space:]]Source[[:space:]]Knowledge/563474b5-06bb-4f08-b441-cc9dcf4bf078_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]with[[:space:]]Structural[[:space:]]Labels[[:space:]]for[[:space:]]Learning[[:space:]]with[[:space:]]Noisy[[:space:]]Labels/7cf38337-93d1-4b9d-ac46-50f09fbdcf75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]with[[:space:]]Unreliability_[[:space:]]Fast[[:space:]]Few-shot[[:space:]]Voxel[[:space:]]Radiance[[:space:]]Fields[[:space:]]with[[:space:]]Relative[[:space:]]Geometric[[:space:]]Consistency/c9a1770c-6e1c-4c1e-859d-d8ada4bc155c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Learning[[:space:]]without[[:space:]]Exact[[:space:]]Guidance_[[:space:]]Updating[[:space:]]Large-scale[[:space:]]High-resolution[[:space:]]Land[[:space:]]Cover[[:space:]]Maps[[:space:]]from[[:space:]]Low-resolution[[:space:]]Historical[[:space:]]Labels/29f9f1d2-f3f8-470f-b021-c1331e97eff4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LeftRefill_[[:space:]]Filling[[:space:]]Right[[:space:]]Canvas[[:space:]]based[[:space:]]on[[:space:]]Left[[:space:]]Reference[[:space:]]through[[:space:]]Generalized[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Model/908a3914-e513-461a-933d-8314753d5fc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Let's[[:space:]]Think[[:space:]]Outside[[:space:]]the[[:space:]]Box_[[:space:]]Exploring[[:space:]]Leap-of-Thought[[:space:]]in[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]with[[:space:]]Creative[[:space:]]Humor[[:space:]]Generation/1c2a1ffb-0d92-41f0-bf24-e35e48d25ef2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Leveraging[[:space:]]Camera[[:space:]]Triplets[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Accurate[[:space:]]Structure-from-Motion/a217b793-2629-4522-9467-ef3937164206_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Leveraging[[:space:]]Cross-Modal[[:space:]]Neighbor[[:space:]]Representation[[:space:]]for[[:space:]]Improved[[:space:]]CLIP[[:space:]]Classification/a3f70eee-6e9c-43f6-ae6b-71319cbe101f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Leveraging[[:space:]]Frame[[:space:]]Affinity[[:space:]]for[[:space:]]sRGB-to-RAW[[:space:]]Video[[:space:]]De-rendering/e20bf1ce-fabe-4b72-8cff-e86075845d22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Leveraging[[:space:]]Predicate[[:space:]]and[[:space:]]Triplet[[:space:]]Learning[[:space:]]for[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/24e4e41a-71a3-4905-abbc-d74d18828d82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Leveraging[[:space:]]Vision-Language[[:space:]]Models[[:space:]]for[[:space:]]Improving[[:space:]]Domain[[:space:]]Generalization[[:space:]]in[[:space:]]Image[[:space:]]Classification/229a1582-b514-420e-a818-a61870d35925_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LiDAR-Net_[[:space:]]A[[:space:]]Real-scanned[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Dataset[[:space:]]for[[:space:]]Indoor[[:space:]]Scenes/966078f8-f41c-47ed-993a-88a5e306b677_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LiDAR-based[[:space:]]Person[[:space:]]Re-identification/9f64a85c-6e54-4870-b464-e9b92b89bbaa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LiDAR4D_[[:space:]]Dynamic[[:space:]]Neural[[:space:]]Fields[[:space:]]for[[:space:]]Novel[[:space:]]Space-time[[:space:]]View[[:space:]]LiDAR[[:space:]]Synthesis/6a80c1df-3dbd-46be-ba3b-340e6d1a9ff3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LiSA_[[:space:]]LiDAR[[:space:]]Localization[[:space:]]with[[:space:]]Semantic[[:space:]]Awareness/a6e01367-c5ea-4fd3-9c99-a8b4204b195f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LidaRF_[[:space:]]Delving[[:space:]]into[[:space:]]Lidar[[:space:]]for[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field[[:space:]]on[[:space:]]Street[[:space:]]Scenes/19449483-1443-4082-8fc3-3172cf73192a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Lift3D_[[:space:]]Zero-Shot[[:space:]]Lifting[[:space:]]of[[:space:]]Any[[:space:]]2D[[:space:]]Vision[[:space:]]Model[[:space:]]to[[:space:]]3D/64a1e91a-3998-4c4f-88c0-fa295e87487e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Light[[:space:]]the[[:space:]]Night_[[:space:]]A[[:space:]]Multi-Condition[[:space:]]Diffusion[[:space:]]Framework[[:space:]]for[[:space:]]Unpaired[[:space:]]Low-Light[[:space:]]Enhancement[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/3c5d62fd-566f-4b4e-b672-b6c485895512_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LightIt_[[:space:]]Illumination[[:space:]]Modeling[[:space:]]and[[:space:]]Control[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/868fb9af-d342-49ca-8347-bf69246ed9a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LightOctree_[[:space:]]Lightweight[[:space:]]3D[[:space:]]Spatially-Coherent[[:space:]]Indoor[[:space:]]Lighting[[:space:]]Estimation/85e43bfd-6016-4a5f-8090-cfaaca5d82c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Linguistic-Aware[[:space:]]Patch[[:space:]]Slimming[[:space:]]Framework[[:space:]]for[[:space:]]Fine-grained[[:space:]]Cross-Modal[[:space:]]Alignment/a0c284e3-7206-4bd7-9c2c-c573232b4f5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Link-Context[[:space:]]Learning[[:space:]]for[[:space:]]Multimodal[[:space:]]LLMs/7b53bc01-b5db-4ae3-8df2-fd7ba2b64ea0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LiveHPS_[[:space:]]LiDAR-based[[:space:]]Scene-level[[:space:]]Human[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Estimation[[:space:]]in[[:space:]]Free[[:space:]]Environment/f51d0186-d594-42f9-bafb-0477cd3ba680_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Living[[:space:]]Scenes_[[:space:]]Multi-object[[:space:]]Relocalization[[:space:]]and[[:space:]]Reconstruction[[:space:]]in[[:space:]]Changing[[:space:]]3D[[:space:]]Environments/eea804c4-7895-4458-857a-a833325570c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LoCoNet_[[:space:]]Long-Short[[:space:]]Context[[:space:]]Network[[:space:]]for[[:space:]]Active[[:space:]]Speaker[[:space:]]Detection/9becb59b-f5a5-44f2-b34a-6673d16ce641_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LoS_[[:space:]]Local[[:space:]]Structure-Guided[[:space:]]Stereo[[:space:]]Matching/0463a5aa-1195-4447-b561-341d57363a83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LoSh_[[:space:]]Long-Short[[:space:]]Text[[:space:]]Joint[[:space:]]Prediction[[:space:]]Network[[:space:]]for[[:space:]]Referring[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/e8271093-2d1c-4f14-acb4-fc5637edd2e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LocLLM_[[:space:]]Exploiting[[:space:]]Generalizable[[:space:]]Human[[:space:]]Keypoint[[:space:]]Localization[[:space:]]via[[:space:]]Large[[:space:]]Language[[:space:]]Model/78387429-3567-4197-b729-579a6cf9f388_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Local-consistent[[:space:]]Transformation[[:space:]]Learning[[:space:]]for[[:space:]]Rotation-invariant[[:space:]]Point[[:space:]]Cloud[[:space:]]Analysis/58ba37e4-8697-465f-a6b2-c85def15d573_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Localization[[:space:]]Is[[:space:]]All[[:space:]]You[[:space:]]Evaluate_[[:space:]]Data[[:space:]]Leakage[[:space:]]in[[:space:]]Online[[:space:]]Mapping[[:space:]]Datasets[[:space:]]and[[:space:]]How[[:space:]]to[[:space:]]Fix[[:space:]]It/52c0e7fe-82d8-4e86-9b00-a96bb781d580_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Locally[[:space:]]Adaptive[[:space:]]Neural[[:space:]]3D[[:space:]]Morphable[[:space:]]Models/66bed243-d612-4938-9894-d6bc5f16cc49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Lodge_[[:space:]]A[[:space:]]Coarse[[:space:]]to[[:space:]]Fine[[:space:]]Diffusion[[:space:]]Network[[:space:]]for[[:space:]]Long[[:space:]]Dance[[:space:]]Generation[[:space:]]Guided[[:space:]]by[[:space:]]the[[:space:]]Characteristic[[:space:]]Dance[[:space:]]Primitives/73139e23-f68b-4e30-b2d5-dea6b579e628_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Logarithmic[[:space:]]Lenses_[[:space:]]Exploring[[:space:]]Log[[:space:]]RGB[[:space:]]Data[[:space:]]for[[:space:]]Image[[:space:]]Classification/87f8d42b-2cfc-4218-8d72-a0e4ab06801d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Logit[[:space:]]Standardization[[:space:]]in[[:space:]]Knowledge[[:space:]]Distillation/c1884c69-29a6-47cb-8f7d-8c7cd62adf57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Long-Tail[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning[[:space:]]via[[:space:]]Independent[[:space:]]Sub-prototype[[:space:]]Construction/cf2028db-7327-4e5c-9e3b-eca734712662_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Long-Tailed[[:space:]]Anomaly[[:space:]]Detection[[:space:]]with[[:space:]]Learnable[[:space:]]Class[[:space:]]Names/58e9402f-24d2-471f-a9aa-16cc1bbda4d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Look-Up[[:space:]]Table[[:space:]]Compression[[:space:]]for[[:space:]]Efficient[[:space:]]Image[[:space:]]Restoration/951d278e-fd02-4d93-8676-d0acdf8a0060_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Lookahead[[:space:]]Exploration[[:space:]]with[[:space:]]Neural[[:space:]]Radiance[[:space:]]Representation[[:space:]]for[[:space:]]Continuous[[:space:]]Vision-Language[[:space:]]Navigation/eb609c4f-d979-48db-bba0-a78cdd0883c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Looking[[:space:]]3D_[[:space:]]Anomaly[[:space:]]Detection[[:space:]]with[[:space:]]2D-3D[[:space:]]Alignment/e91f3907-81ec-443d-84dd-b18fbcd2513e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Looking[[:space:]]Similar[[:space:]]Sounding[[:space:]]Different_[[:space:]]Leveraging[[:space:]]Counterfactual[[:space:]]Cross-Modal[[:space:]]Pairs[[:space:]]for[[:space:]]Audiovisual[[:space:]]Representation[[:space:]]Learning/851cd8f2-a0c1-4b1d-812d-f5f74d20112a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Loopy-SLAM_[[:space:]]Dense[[:space:]]Neural[[:space:]]SLAM[[:space:]]with[[:space:]]Loop[[:space:]]Closures/e396a372-627b-4b6d-a39d-a26a3fc16d70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Loose[[:space:]]Inertial[[:space:]]Poser_[[:space:]]Motion[[:space:]]Capture[[:space:]]with[[:space:]]IMU-attached[[:space:]]Loose-Wear[[:space:]]Jacket/b47b2eaf-32ef-403f-9b3f-28d55dc638eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Low-Latency[[:space:]]Neural[[:space:]]Stereo[[:space:]]Streaming/61321c8e-c52b-4a1d-a58a-2e827b236812_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Low-Rank[[:space:]]Approximation[[:space:]]for[[:space:]]Sparse[[:space:]]Attention[[:space:]]in[[:space:]]Multi-Modal[[:space:]]LLMs/36c0dcbd-16c4-4abb-b836-7f5feb207485_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Low-Rank[[:space:]]Knowledge[[:space:]]Decomposition[[:space:]]for[[:space:]]Medical[[:space:]]Foundation[[:space:]]Models/cd90ca30-af3c-41e5-9e21-a01a10a015ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Low-Rank[[:space:]]Rescaled[[:space:]]Vision[[:space:]]Transformer[[:space:]]Fine-Tuning_[[:space:]]A[[:space:]]Residual[[:space:]]Design[[:space:]]Approach/a3181dbe-fb78-4dc0-9ccd-288dcfaf8ab3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Low-Res[[:space:]]Leads[[:space:]]the[[:space:]]Way_[[:space:]]Improving[[:space:]]Generalization[[:space:]]for[[:space:]]Super-Resolution[[:space:]]by[[:space:]]Self-Supervised[[:space:]]Learning/5f6102dc-8c14-4ec5-a3a7-79adfabcbd3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Low-Resource[[:space:]]Vision[[:space:]]Challenges[[:space:]]for[[:space:]]Foundation[[:space:]]Models/ec2f902c-1ca9-4f26-a452-643619949e3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Low-power[[:space:]]Continuous[[:space:]]Remote[[:space:]]Behavioral[[:space:]]Localization[[:space:]]with[[:space:]]Event[[:space:]]Cameras/58bcc8b6-7726-4726-a8ec-4ed106d59c9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LowRankOcc_[[:space:]]Tensor[[:space:]]Decomposition[[:space:]]and[[:space:]]Low-Rank[[:space:]]Recovery[[:space:]]for[[:space:]]Vision-based[[:space:]]3D[[:space:]]Semantic[[:space:]]Occupancy[[:space:]]Prediction/cf425786-1232-45e8-b9d7-4d86d5488166_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/LucidDreamer_[[:space:]]Towards[[:space:]]High-Fidelity[[:space:]]Text-to-3D[[:space:]]Generation[[:space:]]via[[:space:]]Interval[[:space:]]Score[[:space:]]Matching/b73f6bba-051c-4948-a330-433213ba70d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/M&M[[:space:]]VTO_[[:space:]]Multi-Garment[[:space:]]Virtual[[:space:]]Try-On[[:space:]]and[[:space:]]Editing/3ed09f1d-21f4-4998-aba0-bc67193a7784_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/M3-UDA_[[:space:]]A[[:space:]]New[[:space:]]Benchmark[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Fetal[[:space:]]Cardiac[[:space:]]Structure[[:space:]]Detection/a525fc09-81a4-4ce7-a2cb-a6f8b92908e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MA-LMM_[[:space:]]Memory-Augmented[[:space:]]Large[[:space:]]Multimodal[[:space:]]Model[[:space:]]for[[:space:]]Long-Term[[:space:]]Video[[:space:]]Understanding/cc944389-cd77-40d8-a237-626c240f5d82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MACE_[[:space:]]Mass[[:space:]]Concept[[:space:]]Erasure[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/409c2ab5-3089-4830-9a40-22c3d6dc3f99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MADTP_[[:space:]]Multimodal[[:space:]]Alignment-Guided[[:space:]]Dynamic[[:space:]]Token[[:space:]]Pruning[[:space:]]for[[:space:]]Accelerating[[:space:]]Vision-Language[[:space:]]Transformer/f559c5da-9a39-47d1-9108-c7c1b7c3ae3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MAFA_[[:space:]]Managing[[:space:]]False[[:space:]]Negatives[[:space:]]for[[:space:]]Vision-Language[[:space:]]Pre-training/1de43d9d-087a-42ba-b9e7-483ac6938163_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MAGICK_[[:space:]]A[[:space:]]Large-scale[[:space:]]Captioned[[:space:]]Dataset[[:space:]]from[[:space:]]Matting[[:space:]]Generated[[:space:]]Images[[:space:]]using[[:space:]]Chroma[[:space:]]Keying/c76293fd-e2e5-46f1-9a33-8bdcca957d71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MANUS_[[:space:]]Markerless[[:space:]]Grasp[[:space:]]Capture[[:space:]]using[[:space:]]Articulated[[:space:]]3D[[:space:]]Gaussians/3e62ba3a-a4d3-4394-894d-0f8cf1c9ff51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MAPLM_[[:space:]]A[[:space:]]Real-World[[:space:]]Large-Scale[[:space:]]Vision-Language[[:space:]]Benchmark[[:space:]]for[[:space:]]Map[[:space:]]and[[:space:]]Traffic[[:space:]]Scene[[:space:]]Understanding/f82167dd-82b6-4af4-99d1-4cfe22f42c49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MAPSeg_[[:space:]]Unified[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Heterogeneous[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation[[:space:]]Based[[:space:]]on[[:space:]]3D[[:space:]]Masked[[:space:]]Autoencoding[[:space:]]and[[:space:]]Pseudo-Labeling/e9e12de9-ef56-42a2-81f3-9874ebdf07d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MAP_[[:space:]]MAsk-Pruning[[:space:]]for[[:space:]]Source-Free[[:space:]]Model[[:space:]]Intellectual[[:space:]]Property[[:space:]]Protection/b3499203-2ef2-4c1d-945a-48fd676c7ad6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MART_[[:space:]]Masked[[:space:]]Affective[[:space:]]RepresenTation[[:space:]]Learning[[:space:]]via[[:space:]]Masked[[:space:]]Temporal[[:space:]]Distribution[[:space:]]Distillation/4c51332b-db2b-4277-b05a-487c0d015613_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MAS_[[:space:]]Multi-view[[:space:]]Ancestral[[:space:]]Sampling[[:space:]]for[[:space:]]3D[[:space:]]Motion[[:space:]]Generation[[:space:]]Using[[:space:]]2D[[:space:]]Diffusion/a718f813-b12f-4547-8595-cacf280b176f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MCD_[[:space:]]Diverse[[:space:]]Large-Scale[[:space:]]Multi-Campus[[:space:]]Dataset[[:space:]]for[[:space:]]Robot[[:space:]]Perception/768ef51a-64e3-48ba-b682-c8562ab95460_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MCNet_[[:space:]]Rethinking[[:space:]]the[[:space:]]Core[[:space:]]Ingredients[[:space:]]for[[:space:]]Accurate[[:space:]]and[[:space:]]Efficient[[:space:]]Homography[[:space:]]Estimation/1d9682f9-5ba5-4901-bb01-2629e74e9589_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MCPNet_[[:space:]]An[[:space:]]Interpretable[[:space:]]Classifier[[:space:]]via[[:space:]]Multi-Level[[:space:]]Concept[[:space:]]Prototypes/b759412a-49de-47e1-a864-32fb2efaf307_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MESA_[[:space:]]Matching[[:space:]]Everything[[:space:]]by[[:space:]]Segmenting[[:space:]]Anything/ef6b4e60-652b-4b83-9072-5c69c14da440_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MFP_[[:space:]]Making[[:space:]]Full[[:space:]]Use[[:space:]]of[[:space:]]Probability[[:space:]]Maps[[:space:]]for[[:space:]]Interactive[[:space:]]Image[[:space:]]Segmentation/03c635cc-b8c0-4048-874c-e675789d3583_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MGMap_[[:space:]]Mask-Guided[[:space:]]Learning[[:space:]]for[[:space:]]Online[[:space:]]Vectorized[[:space:]]HD[[:space:]]Map[[:space:]]Construction/df982ac5-c82a-41e9-bed5-55da6a5a7f57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MICap_[[:space:]]A[[:space:]]Unified[[:space:]]Model[[:space:]]for[[:space:]]Identity-Aware[[:space:]]Movie[[:space:]]Descriptions/c0b6be6c-f6c6-4a55-b35c-ab09b28d831f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MIGC_[[:space:]]Multi-Instance[[:space:]]Generation[[:space:]]Controller[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Synthesis/3a6ad60e-a46d-481d-bd5e-9eac04181b17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MLIP_[[:space:]]Enhancing[[:space:]]Medical[[:space:]]Visual[[:space:]]Representation[[:space:]]with[[:space:]]Divergence[[:space:]]Encoder[[:space:]]and[[:space:]]Knowledge-guided[[:space:]]Contrastive[[:space:]]Learning/06a42161-ea27-4cac-b3b8-f6524b04af19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MLP[[:space:]]Can[[:space:]]Be[[:space:]]A[[:space:]]Good[[:space:]]Transformer[[:space:]]Learner/60b7e6ff-a7f4-4de8-9703-8987efa2f15d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MM-Narrator_[[:space:]]Narrating[[:space:]]Long-form[[:space:]]Videos[[:space:]]with[[:space:]]Multimodal[[:space:]]In-Context[[:space:]]Learning/00f72bf2-5236-4bea-90f2-53181f2061ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MMA-Diffusion_[[:space:]]MultiModal[[:space:]]Attack[[:space:]]on[[:space:]]Diffusion[[:space:]]Models/e67a3224-afb9-4f24-8651-4aeaa2ab1c4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MMA_[[:space:]]Multi-Modal[[:space:]]Adapter[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/e961d9f0-1f13-47c4-993a-e582adbc19d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MMCert_[[:space:]]Provable[[:space:]]Defense[[:space:]]against[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]to[[:space:]]Multi-modal[[:space:]]Models/d1bb501f-5c8a-4c12-9b29-9281cc3dc7df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MMMU_[[:space:]]A[[:space:]]Massive[[:space:]]Multi-discipline[[:space:]]Multimodal[[:space:]]Understanding[[:space:]]and[[:space:]]Reasoning[[:space:]]Benchmark[[:space:]]for[[:space:]]Expert[[:space:]]AGI/e2fc2ca7-c8ec-4ec2-a2d7-51a500ea6331_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MMM_[[:space:]]Generative[[:space:]]Masked[[:space:]]Motion[[:space:]]Model/85e4fa9e-8264-42be-a0d1-b15cd771fde1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MMSum_[[:space:]]A[[:space:]]Dataset[[:space:]]for[[:space:]]Multimodal[[:space:]]Summarization[[:space:]]and[[:space:]]Thumbnail[[:space:]]Generation[[:space:]]of[[:space:]]Videos/aa0895f6-2b74-4ccf-a1a4-94faa0741a54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MMVP_[[:space:]]A[[:space:]]Multimodal[[:space:]]MoCap[[:space:]]Dataset[[:space:]]with[[:space:]]Vision[[:space:]]and[[:space:]]Pressure[[:space:]]Sensors/f240f2ca-5c83-4ba3-9dff-31fce497d292_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MOHO_[[:space:]]Learning[[:space:]]Single-view[[:space:]]Hand-held[[:space:]]Object[[:space:]]Reconstruction[[:space:]]with[[:space:]]Multi-view[[:space:]]Occlusion-Aware[[:space:]]Supervision/a88a5ca5-44e9-4410-b201-3cbc4ff8d0a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MP5_[[:space:]]A[[:space:]]Multi-modal[[:space:]]Open-ended[[:space:]]Embodied[[:space:]]System[[:space:]]in[[:space:]]Minecraft[[:space:]]via[[:space:]]Active[[:space:]]Perception/9495ea09-ee67-4ff3-88d5-b32844218e98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MPOD123_[[:space:]]One[[:space:]]Image[[:space:]]to[[:space:]]3D[[:space:]]Content[[:space:]]Generation[[:space:]]Using[[:space:]]Mask-enhanced[[:space:]]Progressive[[:space:]]Outline-to-Detail[[:space:]]Optimization/c8f0faf8-2cb3-4ba4-873d-d738e4f28ea4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MR-VNet_[[:space:]]Media[[:space:]]Restoration[[:space:]]using[[:space:]]Volterra[[:space:]]Networks/fbe8e5a0-3768-4285-91ff-346d6c0daa83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MRC-Net_[[:space:]]6-DoF[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]MultiScale[[:space:]]Residual[[:space:]]Correlation/00fcc36d-0770-44e6-832c-2956ff110bf2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MRFP_[[:space:]]Learning[[:space:]]Generalizable[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]from[[:space:]]Sim-2-Real[[:space:]]with[[:space:]]Multi-Resolution[[:space:]]Feature[[:space:]]Perturbation/23f504f1-f4d7-46c4-a7f0-39f2317e41e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MRFS_[[:space:]]Mutually[[:space:]]Reinforcing[[:space:]]Image[[:space:]]Fusion[[:space:]]and[[:space:]]Segmentation/e8baa93f-03c8-457a-a651-36e52113ac26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MS-DETR_[[:space:]]Efficient[[:space:]]DETR[[:space:]]Training[[:space:]]with[[:space:]]Mixed[[:space:]]Supervision/986e8bcb-0ee2-4e56-a873-2642c4c985c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MS-MANO_[[:space:]]Enabling[[:space:]]Hand[[:space:]]Pose[[:space:]]Tracking[[:space:]]with[[:space:]]Biomechanical[[:space:]]Constraints/69ac812c-b80c-4d31-9dd8-880a34282402_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MSU-4S[[:space:]]-[[:space:]]The[[:space:]]Michigan[[:space:]]State[[:space:]]University[[:space:]]Four[[:space:]]Seasons[[:space:]]Dataset/a66b5f21-4507-46f1-a0d1-1ef0c30cdc01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MTLoRA_[[:space:]]Low-Rank[[:space:]]Adaptation[[:space:]]Approach[[:space:]]for[[:space:]]Efficient[[:space:]]Multi-Task[[:space:]]Learning/99c241d6-df32-40bc-a2f2-ba80a908d648_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MTMMC_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Real-World[[:space:]]Multi-Modal[[:space:]]Camera[[:space:]]Tracking[[:space:]]Benchmark/48bff917-9ab4-465b-a919-50cf75a33a18_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MULAN_[[:space:]]A[[:space:]]Multi[[:space:]]Layer[[:space:]]Annotated[[:space:]]Dataset[[:space:]]for[[:space:]]Controllable[[:space:]]Text-to-Image[[:space:]]Generation/d8c0b754-df2d-465b-8cdf-f00231973266_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MULDE_[[:space:]]Multiscale[[:space:]]Log-Density[[:space:]]Estimation[[:space:]]via[[:space:]]Denoising[[:space:]]Score[[:space:]]Matching[[:space:]]for[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/08123875-abb0-49d0-8bc6-c948bca7cc38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MULTIFLOW_[[:space:]]Shifting[[:space:]]Towards[[:space:]]Task-Agnostic[[:space:]]Vision-Language[[:space:]]Pruning/35a0aa5b-f3db-4c52-a33d-533c454d28c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MV-Adapter_[[:space:]]Multimodal[[:space:]]Video[[:space:]]Transfer[[:space:]]Learning[[:space:]]for[[:space:]]Video[[:space:]]Text[[:space:]]Retrieval/a454014a-ab12-4a17-a125-00f614577ecf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MVBench_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Multi-modal[[:space:]]Video[[:space:]]Understanding[[:space:]]Benchmark/675f0d81-7d92-4461-947a-a69f25eade23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MVCPS-NeuS_[[:space:]]Multi-view[[:space:]]Constrained[[:space:]]Photometric[[:space:]]Stereo[[:space:]]for[[:space:]]Neural[[:space:]]Surface[[:space:]]Reconstruction/1dbfd8d7-9d22-4683-92ab-e724af57e8e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MVD-Fusion_[[:space:]]Single-view[[:space:]]3D[[:space:]]via[[:space:]]Depth-consistent[[:space:]]Multi-view[[:space:]]Generation/b572cdd4-7159-497a-ad63-00e16a163d80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MVHumanNet_[[:space:]]A[[:space:]]Large-scale[[:space:]]Dataset[[:space:]]of[[:space:]]Multi-view[[:space:]]Daily[[:space:]]Dressing[[:space:]]Human[[:space:]]Captures/026fb4f9-3031-4b2b-b3be-c156c7e1f856_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MVIP-NeRF_[[:space:]]Multi-view[[:space:]]3D[[:space:]]Inpainting[[:space:]]on[[:space:]]NeRF[[:space:]]Scenes[[:space:]]via[[:space:]]Diffusion[[:space:]]Prior/39c5bdce-c626-470c-a8ba-3ec72310c11d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MaGGIe_[[:space:]]Masked[[:space:]]Guided[[:space:]]Gradual[[:space:]]Human[[:space:]]Instance[[:space:]]Matting/63745a55-5728-4dbf-b8f2-684d2b004a83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Magic[[:space:]]Tokens_[[:space:]]Select[[:space:]]Diverse[[:space:]]Tokens[[:space:]]for[[:space:]]Multi-modal[[:space:]]Object[[:space:]]Re-Identification/0f812672-2b6b-4d88-9f10-d927071810a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MagicAnimate_[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Human[[:space:]]Image[[:space:]]Animation[[:space:]]using[[:space:]]Diffusion[[:space:]]Model/6b3684fc-dac5-4398-ad8c-b218c1b116ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Make[[:space:]]Me[[:space:]]a[[:space:]]BNN_[[:space:]]A[[:space:]]Simple[[:space:]]Strategy[[:space:]]for[[:space:]]Estimating[[:space:]]Bayesian[[:space:]]Uncertainty[[:space:]]from[[:space:]]Pre-trained[[:space:]]Models/047c6914-07f0-44d9-9485-92c687d0cef2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Make[[:space:]]Pixels[[:space:]]Dance_[[:space:]]High-Dynamic[[:space:]]Video[[:space:]]Generation/529ce227-c61e-4b80-93e6-5410c587adba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Make-It-Vivid_[[:space:]]Dressing[[:space:]]Your[[:space:]]Animatable[[:space:]]Biped[[:space:]]Cartoon[[:space:]]Characters[[:space:]]from[[:space:]]Text/58b63fa9-7178-44d0-99a9-17b04dc63e3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Make-Your-Anchor_[[:space:]]A[[:space:]]Diffusion-based[[:space:]]2D[[:space:]]Avatar[[:space:]]Generation[[:space:]]Framework/e05de9a5-59fd-4b1a-92a8-d6568ebf1863_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Makeup[[:space:]]Prior[[:space:]]Models[[:space:]]for[[:space:]]3D[[:space:]]Facial[[:space:]]Makeup[[:space:]]Estimation[[:space:]]and[[:space:]]Applications/746e97d3-87d4-4062-b730-87501be66c5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Making[[:space:]]Vision[[:space:]]Transformers[[:space:]]Truly[[:space:]]Shift-Equivariant/18d34818-931f-40d6-a0de-63b54f47d2d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Making[[:space:]]Visual[[:space:]]Sense[[:space:]]of[[:space:]]Oracle[[:space:]]Bones[[:space:]]for[[:space:]]You[[:space:]]and[[:space:]]Me/00ac0fca-2d81-4c05-8adf-98330dfbb884_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ManiFPT_[[:space:]]Defining[[:space:]]and[[:space:]]Analyzing[[:space:]]Fingerprints[[:space:]]of[[:space:]]Generative[[:space:]]Models/b8270718-402c-4c1c-a2ab-9760365a764b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ManipLLM_[[:space:]]Embodied[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Object-Centric[[:space:]]Robotic[[:space:]]Manipulation/ec60057b-8873-45a8-a497-bfd2469ce96b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Map-Relative[[:space:]]Pose[[:space:]]Regression[[:space:]]for[[:space:]]Visual[[:space:]]Re-Localization/a775e99d-bcc0-4720-a3af-96305e8d3e34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MarkovGen_[[:space:]]Structured[[:space:]]Prediction[[:space:]]for[[:space:]]Efficient[[:space:]]Text-to-Image[[:space:]]Generation/3615aa89-4599-4756-bc9d-a3197df24bbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mask[[:space:]]Grounding[[:space:]]for[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation/a251d9ca-ef79-40a3-9a34-1b4d20442172_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mask4Align_[[:space:]]Aligned[[:space:]]Entity[[:space:]]Prompting[[:space:]]with[[:space:]]Color[[:space:]]Masks[[:space:]]for[[:space:]]Multi-Entity[[:space:]]Localization[[:space:]]Problems/88c29475-72a6-4c04-b48e-d123b1122837_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MaskCLR_[[:space:]]Attention-Guided[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Robust[[:space:]]Action[[:space:]]Representation[[:space:]]Learning/f1a4a268-dbe8-4c85-897e-b8c6d3a4af5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MaskClustering_[[:space:]]View[[:space:]]Consensus[[:space:]]based[[:space:]]Mask[[:space:]]Graph[[:space:]]Clustering[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation/e2da6c88-7b90-4c6f-9c91-97cf4484fcb9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MaskINT_[[:space:]]Video[[:space:]]Editing[[:space:]]via[[:space:]]Interpolative[[:space:]]Non-autoregressive[[:space:]]Masked[[:space:]]Transformers/136f93fe-8d14-4636-b12a-d9987d54cfc8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MaskPLAN_[[:space:]]Masked[[:space:]]Generative[[:space:]]Layout[[:space:]]Planning[[:space:]]from[[:space:]]Partial[[:space:]]Input/73d685d1-19a3-4f65-bcd3-c18a88da845a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Masked[[:space:]]AutoDecoder[[:space:]]is[[:space:]]Effective[[:space:]]Multi-Task[[:space:]]Vision[[:space:]]Generalist/a847d220-3bd5-4109-8339-59af9c363004_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Masked[[:space:]]Autoencoders[[:space:]]for[[:space:]]Microscopy[[:space:]]are[[:space:]]Scalable[[:space:]]Learners[[:space:]]of[[:space:]]Cellular[[:space:]]Biology/14cc5bd6-ca74-4214-8ea1-96e9794f8948_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Masked[[:space:]]Spatial[[:space:]]Propagation[[:space:]]Network[[:space:]]for[[:space:]]Sparsity-Adaptive[[:space:]]Depth[[:space:]]Refinement/8f9f8232-7c1d-4718-a2e5-c49e0df9c9eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Masked[[:space:]]and[[:space:]]Shuffled[[:space:]]Blind[[:space:]]Spot[[:space:]]Denoising[[:space:]]for[[:space:]]Real-World[[:space:]]Images/96f99f3d-075c-40ce-9788-4e8434e975d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MatFuse_[[:space:]]Controllable[[:space:]]Material[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/a758fcc7-5c1b-4b85-8963-dd145cbfd94a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MatSynth_[[:space:]]A[[:space:]]Modern[[:space:]]PBR[[:space:]]Materials[[:space:]]Dataset/de7466b3-dd79-43ec-b835-db0afa069874_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MatchU_[[:space:]]Matching[[:space:]]Unseen[[:space:]]Objects[[:space:]]for[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation[[:space:]]from[[:space:]]RGB-D[[:space:]]Images/52279bc7-f87a-430f-b619-ffcc2213929a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Matching[[:space:]]2D[[:space:]]Images[[:space:]]in[[:space:]]3D_[[:space:]]Metric[[:space:]]Relative[[:space:]]Pose[[:space:]]from[[:space:]]Metric[[:space:]]Correspondences/57be3ca7-ab34-49c2-b2df-a7c982c8278d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Matching[[:space:]]Anything[[:space:]]by[[:space:]]Segmenting[[:space:]]Anything/e0e8ad99-f19e-429d-9adc-1dc2cc5efa38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Material[[:space:]]Palette_[[:space:]]Extraction[[:space:]]of[[:space:]]Materials[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/301aa835-4fce-4dcd-a01a-3109a66df09c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MaxQ_[[:space:]]Multi-Axis[[:space:]]Query[[:space:]]for[[:space:]]N_M[[:space:]]Sparsity[[:space:]]Network/30829115-c102-4d53-a0c7-a9f302892f64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MeLFusion_[[:space:]]Synthesizing[[:space:]]Music[[:space:]]from[[:space:]]Image[[:space:]]and[[:space:]]Language[[:space:]]Cues[[:space:]]using[[:space:]]Diffusion[[:space:]]Models/555ebec9-10d0-485e-aeb2-531ac957a6c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MeaCap_[[:space:]]Memory-Augmented[[:space:]]Zero-shot[[:space:]]Image[[:space:]]Captioning/5d600fb7-5e74-47eb-8150-a5a90dc121bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mean-Shift[[:space:]]Feature[[:space:]]Transformer/5faa2e5e-a1a1-46cc-a332-c9fdee18a759_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MedBN_[[:space:]]Robust[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]against[[:space:]]Malicious[[:space:]]Test[[:space:]]Samples/850f4dcd-fdf0-426e-86e9-7ec70d1f8f0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MedM2G_[[:space:]]Unifying[[:space:]]Medical[[:space:]]Multi-Modal[[:space:]]Generation[[:space:]]via[[:space:]]Cross-Guided[[:space:]]Diffusion[[:space:]]with[[:space:]]Visual[[:space:]]Invariant/936806c4-794e-47bc-97d9-bd36b1d43d40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MemFlow_[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation[[:space:]]and[[:space:]]Prediction[[:space:]]with[[:space:]]Memory/18215ed0-4f78-4136-9be0-b1ddf794fad9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MemSAM_[[:space:]]Taming[[:space:]]Segment[[:space:]]Anything[[:space:]]Model[[:space:]]for[[:space:]]Echocardiography[[:space:]]Video[[:space:]]Segmentation/d683f69b-9574-4b64-acb3-2e502b2644ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MemoNav_[[:space:]]Working[[:space:]]Memory[[:space:]]Model[[:space:]]for[[:space:]]Visual[[:space:]]Navigation/8d2f0268-d30d-431c-b00c-a92129507d55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Memory-Scalable[[:space:]]and[[:space:]]Simplified[[:space:]]Functional[[:space:]]Map[[:space:]]Learning/a56a9f00-a5ef-4c37-9d8b-7a54ddfb40f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Memory-based[[:space:]]Adapters[[:space:]]for[[:space:]]Online[[:space:]]3D[[:space:]]Scene[[:space:]]Perception/7fe858e1-364c-4d23-a01e-5264ed012837_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MeshGPT_[[:space:]]Generating[[:space:]]Triangle[[:space:]]Meshes[[:space:]]with[[:space:]]Decoder-Only[[:space:]]Transformers/984387c8-2c42-4a23-b178-91396f98a080_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MeshPose_[[:space:]]Unifying[[:space:]]DensePose[[:space:]]and[[:space:]]3D[[:space:]]Body[[:space:]]Mesh[[:space:]]Reconstruction/dab398d2-0b34-4b54-b61a-d73012e1ae94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Meta-Point[[:space:]]Learning[[:space:]]and[[:space:]]Refining[[:space:]]for[[:space:]]Category-Agnostic[[:space:]]Pose[[:space:]]Estimation/dacaf1c6-29cb-4704-bb5f-d7194c594cbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MetaCloak_[[:space:]]Preventing[[:space:]]Unauthorized[[:space:]]Subject-driven[[:space:]]Text-to-image[[:space:]]Diffusion-based[[:space:]]Synthesis[[:space:]]via[[:space:]]Meta-learning/3671c28b-c29e-4377-9186-15c1029ca85d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MiKASA_[[:space:]]Multi-Key-Anchor[[:space:]]&[[:space:]]Scene-Aware[[:space:]]Transformer[[:space:]]for[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding/cfc094ba-23ef-46aa-ae5b-3b0259a07719_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MicroCinema_[[:space:]]A[[:space:]]Divide-and-Conquer[[:space:]]Approach[[:space:]]for[[:space:]]Text-to-Video[[:space:]]Generation/3b9b77a8-aa29-4238-967e-8c7cc870eb1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MicroDiffusion_[[:space:]]Implicit[[:space:]]Representation-Guided[[:space:]]Diffusion[[:space:]]for[[:space:]]3D[[:space:]]Reconstruction[[:space:]]from[[:space:]]Limited[[:space:]]2D[[:space:]]Microscopy[[:space:]]Projections/4e8cfb1b-5397-4102-a70c-2893efae0718_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MimicDiffusion_[[:space:]]Purifying[[:space:]]Adversarial[[:space:]]Perturbation[[:space:]]via[[:space:]]Mimicking[[:space:]]Clean[[:space:]]Diffusion[[:space:]]Model/4e2422e2-5148-48b2-a5dd-807dc3d59476_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mind[[:space:]]Artist_[[:space:]]Creating[[:space:]]Artistic[[:space:]]Snapshots[[:space:]]with[[:space:]]Human[[:space:]]Thought/8e434a69-cb32-444a-a0c3-6870e0733789_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mind[[:space:]]Marginal[[:space:]]Non-Crack[[:space:]]Regions_[[:space:]]Clustering-Inspired[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Crack[[:space:]]Segmentation/5b518349-f076-4186-8665-0eeba06cad03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mind[[:space:]]The[[:space:]]Edge_[[:space:]]Refining[[:space:]]Depth[[:space:]]Edges[[:space:]]in[[:space:]]Sparsely-Supervised[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/cf8747b2-cbfb-4883-ac96-1be1842f3bdb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MindBridge_[[:space:]]A[[:space:]]Cross-Subject[[:space:]]Brain[[:space:]]Decoding[[:space:]]Framework/8194ff07-2ccc-4a54-8e05-92db3268d3e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Minimal[[:space:]]Perspective[[:space:]]Autocalibration/bb9a8e2d-ad67-4724-9dc7-89e635527f92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mining[[:space:]]Supervision[[:space:]]for[[:space:]]Dynamic[[:space:]]Regions[[:space:]]in[[:space:]]Self-Supervised[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/956c8173-a48b-4d9f-8e80-b72b4ab748f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mip-Splatting_[[:space:]]Alias-free[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/d23328c0-f0c2-4f05-9a2d-85cac5e1975a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MirageRoom_[[:space:]]3D[[:space:]]Scene[[:space:]]Segmentation[[:space:]]with[[:space:]]2D[[:space:]]Pre-trained[[:space:]]Models[[:space:]]by[[:space:]]Mirage[[:space:]]Projection/4edfa839-bbb0-402e-baaa-0ed0f2c7665a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mirasol3B_[[:space:]]A[[:space:]]Multimodal[[:space:]]Autoregressive[[:space:]]Model[[:space:]]for[[:space:]]Time-Aligned[[:space:]]and[[:space:]]Contextual[[:space:]]Modalities/93379e12-4fc3-4e4f-a049-cc73820d02c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Misalignment-Robust[[:space:]]Frequency[[:space:]]Distribution[[:space:]]Loss[[:space:]]for[[:space:]]Image[[:space:]]Transformation/788e118c-82cc-47ed-8324-13761274e510_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mitigating[[:space:]]Motion[[:space:]]Blur[[:space:]]in[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]with[[:space:]]Events[[:space:]]and[[:space:]]Frames/af5719a9-38dd-4605-9a60-2597af61bbbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mitigating[[:space:]]Noisy[[:space:]]Correspondence[[:space:]]by[[:space:]]Geometrical[[:space:]]Structure[[:space:]]Consistency[[:space:]]Learning/7478478e-683a-4005-8421-d0c9b47f1d87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mitigating[[:space:]]Object[[:space:]]Dependencies_[[:space:]]Improving[[:space:]]Point[[:space:]]Cloud[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]through[[:space:]]Object[[:space:]]Exchange/636a66f8-694e-49c8-afc3-72baf0cba69f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mitigating[[:space:]]Object[[:space:]]Hallucinations[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]through[[:space:]]Visual[[:space:]]Contrastive[[:space:]]Decoding/ebd62a33-defa-4cbc-9000-e0fc62b9f9f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mixed-Precision[[:space:]]Quantization[[:space:]]for[[:space:]]Federated[[:space:]]Learning[[:space:]]on[[:space:]]Resource-Constrained[[:space:]]Heterogeneous[[:space:]]Devices/73ae0fa7-4629-4a48-914b-622a9b3dd8db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MoCha-Stereo_[[:space:]]Motif[[:space:]]Channel[[:space:]]Attention[[:space:]]Network[[:space:]]for[[:space:]]Stereo[[:space:]]Matching/4530b10d-4512-4bda-a843-106ce3f64cb3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MoDE_[[:space:]]CLIP[[:space:]]Data[[:space:]]Experts[[:space:]]via[[:space:]]Clustering/eb9da03b-f692-420c-b683-649c2509b3f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MoML_[[:space:]]Online[[:space:]]Meta[[:space:]]Adaptation[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Motion[[:space:]]Prediction/7c173a27-26d9-4d23-9886-e897ffaec33a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MoMask_[[:space:]]Generative[[:space:]]Masked[[:space:]]Modeling[[:space:]]of[[:space:]]3D[[:space:]]Human[[:space:]]Motions/98c35ea3-8ff1-4af9-9ab5-93e757967171_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MoPE-CLIP_[[:space:]]Structured[[:space:]]Pruning[[:space:]]for[[:space:]]Efficient[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]Module-wise[[:space:]]Pruning[[:space:]]Error[[:space:]]Metric/0c4e3e94-ee65-4505-a8ed-1b4431529e58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MoReVQA_[[:space:]]Exploring[[:space:]]Modular[[:space:]]Reasoning[[:space:]]Models[[:space:]]for[[:space:]]Video[[:space:]]Question[[:space:]]Answering/d3d21623-2e6a-4fec-8c82-2d8b15ae239d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MoSAR_[[:space:]]Monocular[[:space:]]Semi-Supervised[[:space:]]Model[[:space:]]for[[:space:]]Avatar[[:space:]]Reconstruction[[:space:]]using[[:space:]]Differentiable[[:space:]]Shading/413ad762-48cb-417e-acfe-ed12b836e3b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MoST_[[:space:]]Motion[[:space:]]Style[[:space:]]Transformer[[:space:]]Between[[:space:]]Diverse[[:space:]]Action[[:space:]]Contents/62b4a486-e5f2-4615-bb52-cfe4e7ff8fb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MoST_[[:space:]]Multi-Modality[[:space:]]Scene[[:space:]]Tokenization[[:space:]]for[[:space:]]Motion[[:space:]]Prediction/159455e4-b26e-4130-8067-b2db1a750fac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MobileCLIP_[[:space:]]Fast[[:space:]]Image-Text[[:space:]]Models[[:space:]]through[[:space:]]Multi-Modal[[:space:]]Reinforced[[:space:]]Training/e9462882-e474-4e6d-9222-d3c6af5d8400_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mocap[[:space:]]Everyone[[:space:]]Everywhere_[[:space:]]Lightweight[[:space:]]Motion[[:space:]]Capture[[:space:]]With[[:space:]]Smartwatches[[:space:]]and[[:space:]]a[[:space:]]Head-Mounted[[:space:]]Camera/b417fe93-12ff-4b32-95f8-fefd5b6c565d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ModaVerse_[[:space:]]Efficiently[[:space:]]Transforming[[:space:]]Modalities[[:space:]]with[[:space:]]LLMs/59e8b62c-f803-401f-9e32-a8e5f7fa5b08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Modality-Agnostic[[:space:]]Structural[[:space:]]Image[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Deformable[[:space:]]Multi-Modality[[:space:]]Medical[[:space:]]Image[[:space:]]Registration/d9351dac-86cf-4a76-909a-077577555cb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Modality-Collaborative[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Action[[:space:]]Recognition/275b3e32-0a69-4b84-9c42-92e0b8c100ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Modality-agnostic[[:space:]]Domain[[:space:]]Generalizable[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation[[:space:]]by[[:space:]]Multi-Frequency[[:space:]]in[[:space:]]Multi-Scale[[:space:]]Attention/46b76a9c-a9a5-4cdc-ab93-2f7db49e75d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Model[[:space:]]Adaptation[[:space:]]for[[:space:]]Time[[:space:]]Constrained[[:space:]]Embodied[[:space:]]Control/61d09335-e26d-4b52-a64a-ff6577ee6ac4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Model[[:space:]]Inversion[[:space:]]Robustness_[[:space:]]Can[[:space:]]Transfer[[:space:]]Learning[[:space:]]Help_/3c91d711-cbb6-4c79-9847-ebca75834ddc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Modeling[[:space:]]Collaborator_[[:space:]]Enabling[[:space:]]Subjective[[:space:]]Vision[[:space:]]Classification[[:space:]]With[[:space:]]Minimal[[:space:]]Human[[:space:]]Effort[[:space:]]via[[:space:]]LLM[[:space:]]Tool-Use/026c61f9-72f9-4060-8c91-78636eacdab3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Modeling[[:space:]]Dense[[:space:]]Multimodal[[:space:]]Interactions[[:space:]]Between[[:space:]]Biological[[:space:]]Pathways[[:space:]]and[[:space:]]Histology[[:space:]]for[[:space:]]Survival[[:space:]]Prediction/54be8f8d-e032-4eb4-86d6-4a30352114f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Modeling[[:space:]]Multimodal[[:space:]]Social[[:space:]]Interactions_[[:space:]]New[[:space:]]Challenges[[:space:]]and[[:space:]]Baselines[[:space:]]with[[:space:]]Densely[[:space:]]Aligned[[:space:]]Representations/f1604caa-e055-4a42-b1a8-8411dbb9383c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Modular[[:space:]]Blind[[:space:]]Video[[:space:]]Quality[[:space:]]Assessment/bf4c2f3f-c824-43a2-a396-e847bb354846_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Molecular[[:space:]]Data[[:space:]]Programming_[[:space:]]Towards[[:space:]]Molecule[[:space:]]Pseudo-labeling[[:space:]]with[[:space:]]Systematic[[:space:]]Weak[[:space:]]Supervision/1caa568c-ac44-4c4b-b3cd-2ce3f7e4edfd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Monkey_[[:space:]]Image[[:space:]]Resolution[[:space:]]and[[:space:]]Text[[:space:]]Label[[:space:]]Are[[:space:]]Important[[:space:]]Things[[:space:]]for[[:space:]]Large[[:space:]]Multi-modal[[:space:]]Models/0d4d85ca-c69d-4862-814d-c0c33d7e7c01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MonoCD_[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Complementary[[:space:]]Depths/2b41fc58-3837-4503-a1d7-9c207b304eb3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MonoDiff_[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]and[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/2c10152f-6e75-4823-925a-53baf2c92758_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MonoHair_[[:space:]]High-Fidelity[[:space:]]Hair[[:space:]]Modeling[[:space:]]from[[:space:]]a[[:space:]]Monocular[[:space:]]Video/367b3ddd-c262-4573-9440-4177404ec44c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MonoNPHM_[[:space:]]Dynamic[[:space:]]Head[[:space:]]Reconstruction[[:space:]]from[[:space:]]Monocular[[:space:]]Videos/ea6778f9-6d18-44a2-b776-ad24811f9bb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Monocular[[:space:]]Identity-Conditioned[[:space:]]Facial[[:space:]]Reflectance[[:space:]]Reconstruction/6c543276-6057-4276-91cb-56e0bbac748d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Morphable[[:space:]]Diffusion_[[:space:]]3D-Consistent[[:space:]]Diffusion[[:space:]]for[[:space:]]Single-image[[:space:]]Avatar[[:space:]]Creation/6f15a8a0-3dd0-43b5-9fbc-9e09a2e70b80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MorpheuS_[[:space:]]Neural[[:space:]]Dynamic[[:space:]]360deg[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]from[[:space:]]Monocular[[:space:]]RGB-D[[:space:]]Video/21212102-bec4-4226-bdd8-0cdb8bc02a95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Morphological[[:space:]]Prototyping[[:space:]]for[[:space:]]Unsupervised[[:space:]]Slide[[:space:]]Representation[[:space:]]Learning[[:space:]]in[[:space:]]Computational[[:space:]]Pathology/88e1d8ae-85f9-4205-be5f-33860465699e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mosaic-SDF[[:space:]]for[[:space:]]3D[[:space:]]Generative[[:space:]]Models/592b0c38-10dd-416a-ad02-004982e68675_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Motion[[:space:]]Blur[[:space:]]Decomposition[[:space:]]with[[:space:]]Cross-shutter[[:space:]]Guidance/0ec1753f-20a3-48e0-913c-0373e233fcd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Motion[[:space:]]Diversification[[:space:]]Networks/d63824af-ec30-4353-98e0-14d56905cdee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Motion-adaptive[[:space:]]Separable[[:space:]]Collaborative[[:space:]]Filters[[:space:]]for[[:space:]]Blind[[:space:]]Motion[[:space:]]Deblurring/1dfab0bf-8b5a-4887-ae5d-e986a894cbcc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Motion2VecSets_[[:space:]]4D[[:space:]]Latent[[:space:]]Vector[[:space:]]Set[[:space:]]Diffusion[[:space:]]for[[:space:]]Non-rigid[[:space:]]Shape[[:space:]]Reconstruction[[:space:]]and[[:space:]]Tracking/13e05fa7-58bb-425e-841c-2f720eb6a8c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MotionEditor_[[:space:]]Editing[[:space:]]Video[[:space:]]Motion[[:space:]]via[[:space:]]Content-Aware[[:space:]]Diffusion/e2a94018-a9d8-4efb-af9c-d35bdd00e973_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Move[[:space:]]Anything[[:space:]]with[[:space:]]Layered[[:space:]]Scene[[:space:]]Diffusion/aff98016-ee3e-4f1c-b9a8-e559e922233a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Move[[:space:]]as[[:space:]]You[[:space:]]Say[[:space:]]Interact[[:space:]]as[[:space:]]You[[:space:]]Can_[[:space:]]Language-guided[[:space:]]Human[[:space:]]Motion[[:space:]]Generation[[:space:]]with[[:space:]]Scene[[:space:]]Affordance/63d40f73-31a4-4c03-9015-7428005fa39b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MovieChat_[[:space:]]From[[:space:]]Dense[[:space:]]Token[[:space:]]to[[:space:]]Sparse[[:space:]]Memory[[:space:]]for[[:space:]]Long[[:space:]]Video[[:space:]]Understanding/87b1b3a8-af53-4dbf-8cc6-34b0b0d0a75a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MuGE_[[:space:]]Multiple[[:space:]]Granularity[[:space:]]Edge[[:space:]]Detection/a7e2ae2d-2070-4adb-a467-fe719afb4075_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MuRF_[[:space:]]Multi-Baseline[[:space:]]Radiance[[:space:]]Fields/265b8c06-80fc-4e85-8d50-8ff0b6769d10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Mudslide_[[:space:]]A[[:space:]]Universal[[:space:]]Nuclear[[:space:]]Instance[[:space:]]Segmentation[[:space:]]Method/5a64aa1e-710a-4a50-b851-b34aaaa3cbbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-Attribute[[:space:]]Interactions[[:space:]]Matter[[:space:]]for[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding/477f7931-e1b2-442e-a45b-11cf7d648a49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-Level[[:space:]]Neural[[:space:]]Scene[[:space:]]Graphs[[:space:]]for[[:space:]]Dynamic[[:space:]]Urban[[:space:]]Environments/10ec9eda-ceb6-4b88-aa59-8a3b78a66785_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-Modal[[:space:]]Hallucination[[:space:]]Control[[:space:]]by[[:space:]]Visual[[:space:]]Information[[:space:]]Grounding/1e5cae18-a58b-4415-90f8-bc4858438f19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-Modal[[:space:]]Proxy[[:space:]]Learning[[:space:]]Towards[[:space:]]Personalized[[:space:]]Visual[[:space:]]Multiple[[:space:]]Clustering/f9afe157-eaaf-440c-afcf-7604121ce960_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-Object[[:space:]]Tracking[[:space:]]in[[:space:]]the[[:space:]]Dark/4c20e9a2-1106-45cf-a36a-0161982e6015_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-Scale[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Anti-Aliased[[:space:]]Rendering/20977810-85e5-40f1-a4c6-cf52c533cece_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-Scale[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection[[:space:]]by[[:space:]]Multi-Grained[[:space:]]Spatio-Temporal[[:space:]]Representation[[:space:]]Learning/82eacc56-0492-4ac2-b935-f0e92d5461ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-Session[[:space:]]SLAM[[:space:]]with[[:space:]]Differentiable[[:space:]]Wide-Baseline[[:space:]]Pose[[:space:]]Optimization/04406ae4-0193-431d-9be9-c73afd625413_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-Space[[:space:]]Alignments[[:space:]]Towards[[:space:]]Universal[[:space:]]LiDAR[[:space:]]Segmentation/b0f5aa24-c6e7-4f07-a0f8-7432b1bda5de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-Task[[:space:]]Dense[[:space:]]Prediction[[:space:]]via[[:space:]]Mixture[[:space:]]of[[:space:]]Low-Rank[[:space:]]Experts/4b27d7d8-e816-4974-8903-c5d6fd2b2750_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-View[[:space:]]Attentive[[:space:]]Contextualization[[:space:]]for[[:space:]]Multi-View[[:space:]]3D[[:space:]]Object[[:space:]]Detection/cb273d05-390f-4fb9-aa74-74b5a46402e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-agent[[:space:]]Collaborative[[:space:]]Perception[[:space:]]via[[:space:]]Motion-aware[[:space:]]Robust[[:space:]]Communication[[:space:]]Network/8f321c57-13fe-4536-adc3-b4b1a5e6a84e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-agent[[:space:]]Long-term[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Forecasting[[:space:]]via[[:space:]]Interaction-aware[[:space:]]Trajectory[[:space:]]Conditioning/12609e19-efed-4425-b5ad-32eb26a81347_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-criteria[[:space:]]Token[[:space:]]Fusion[[:space:]]with[[:space:]]One-step-ahead[[:space:]]Attention[[:space:]]for[[:space:]]Efficient[[:space:]]Vision[[:space:]]Transformers/c38e4100-b4f3-4d92-aff9-1fdadd7cfa0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-modal[[:space:]]In-Context[[:space:]]Learning[[:space:]]Makes[[:space:]]an[[:space:]]Ego-evolving[[:space:]]Scene[[:space:]]Text[[:space:]]Recognizer/bf855ad8-e560-4eb0-9886-16b2ff2266f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-modal[[:space:]]Instruction[[:space:]]Tuned[[:space:]]LLMs[[:space:]]with[[:space:]]Fine-grained[[:space:]]Visual[[:space:]]Perception/1060b587-378f-4876-81ef-a69e77b5ca70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-modal[[:space:]]Learning[[:space:]]for[[:space:]]Geospatial[[:space:]]Vegetation[[:space:]]Forecasting/42a10fad-a12f-4501-8672-d7229d356fee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-scale[[:space:]]Dynamic[[:space:]]and[[:space:]]Hierarchical[[:space:]]Relationship[[:space:]]Modeling[[:space:]]for[[:space:]]Facial[[:space:]]Action[[:space:]]Units[[:space:]]Recognition/75181bea-24ed-4d67-89b7-aa0610066eda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multi-view[[:space:]]Aggregation[[:space:]]Network[[:space:]]for[[:space:]]Dichotomous[[:space:]]Image[[:space:]]Segmentation/cce93abc-7f7e-4582-a2d5-107b621a8f82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MultiDiff_[[:space:]]Consistent[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/2aa2e82a-dff4-4750-bb74-ff506ca72a62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MultiPLY_[[:space:]]A[[:space:]]Multisensory[[:space:]]Object-Centric[[:space:]]Embodied[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]in[[:space:]]3D[[:space:]]World/6cdd4677-266b-499a-a637-8d38b4cfc6f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MultiPhys_[[:space:]]Multi-Person[[:space:]]Physics-aware[[:space:]]3D[[:space:]]Motion[[:space:]]Estimation/8bf95065-c5b4-4afa-86d1-804b7f567a4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MultiPly_[[:space:]]Reconstruction[[:space:]]of[[:space:]]Multiple[[:space:]]People[[:space:]]from[[:space:]]Monocular[[:space:]]Video[[:space:]]in[[:space:]]the[[:space:]]Wild/9118ef01-a3a3-4d0e-a512-cab61af53519_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multiagent[[:space:]]Multitraversal[[:space:]]Multimodal[[:space:]]Self-Driving_[[:space:]]Open[[:space:]]MARS[[:space:]]Dataset/daf2a5e1-efd2-4f29-b37c-4c73f81b7002_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multimodal[[:space:]]Industrial[[:space:]]Anomaly[[:space:]]Detection[[:space:]]by[[:space:]]Crossmodal[[:space:]]Feature[[:space:]]Mapping/37da17b3-6824-4605-81f8-c01f9eebd49d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multimodal[[:space:]]Pathway_[[:space:]]Improve[[:space:]]Transformers[[:space:]]with[[:space:]]Irrelevant[[:space:]]Data[[:space:]]from[[:space:]]Other[[:space:]]Modalities/31541608-9780-4c92-8dfb-6ccd462a5a0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multimodal[[:space:]]Prompt[[:space:]]Perceiver_[[:space:]]Empower[[:space:]]Adaptiveness[[:space:]]Generalizability[[:space:]]and[[:space:]]Fidelity[[:space:]]for[[:space:]]All-in-One[[:space:]]Image[[:space:]]Restoration/dd270751-648d-4a9b-b4c9-9b85d339c5c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multimodal[[:space:]]Representation[[:space:]]Learning[[:space:]]by[[:space:]]Alternating[[:space:]]Unimodal[[:space:]]Adaptation/9f47df9d-5a19-4287-a41d-2d4ffe80f5db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multimodal[[:space:]]Sense-Informed[[:space:]]Forecasting[[:space:]]of[[:space:]]3D[[:space:]]Human[[:space:]]Motions/b5ef95c4-a9a0-4681-a4a8-822e1b33a34a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multiplane[[:space:]]Prior[[:space:]]Guided[[:space:]]Few-Shot[[:space:]]Aerial[[:space:]]Scene[[:space:]]Rendering/15e4c819-6e34-4bd4-b379-6775e97abc28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multiple[[:space:]]View[[:space:]]Geometry[[:space:]]Transformers[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/07b64eb8-929d-4d4b-b3f7-dd15c8bf344f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multiscale[[:space:]]Vision[[:space:]]Transformers[[:space:]]Meet[[:space:]]Bipartite[[:space:]]Matching[[:space:]]for[[:space:]]Efficient[[:space:]]Single-stage[[:space:]]Action[[:space:]]Localization/eeb58ba1-b49b-44b5-b8bc-8e43b4f9d2bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multiview[[:space:]]Aerial[[:space:]]Visual[[:space:]]RECognition[[:space:]](MAVREC)_[[:space:]]Can[[:space:]]Multi-view[[:space:]]Improve[[:space:]]Aerial[[:space:]]Visual[[:space:]]Perception_/5ed3fe81-108b-429b-8fc0-08881954a458_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Multiway[[:space:]]Point[[:space:]]Cloud[[:space:]]Mosaicking[[:space:]]with[[:space:]]Diffusion[[:space:]]and[[:space:]]Global[[:space:]]Optimization/d99639eb-7fa3-4be0-a146-25a4bc301b9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/MuseChat_[[:space:]]A[[:space:]]Conversational[[:space:]]Music[[:space:]]Recommendation[[:space:]]System[[:space:]]for[[:space:]]Videos/f0f6aab1-0b41-4348-8488-4140eb3d81e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NAPGuard_[[:space:]]Towards[[:space:]]Detecting[[:space:]]Naturalistic[[:space:]]Adversarial[[:space:]]Patches/139c013e-88a7-4162-8220-0b4f45d02eeb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NARUTO_[[:space:]]Neural[[:space:]]Active[[:space:]]Reconstruction[[:space:]]from[[:space:]]Uncertain[[:space:]]Target[[:space:]]Observations/b788f3e6-255f-4cf1-93ae-ce6f871b55a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NAYER_[[:space:]]Noisy[[:space:]]Layer[[:space:]]Data[[:space:]]Generation[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Effective[[:space:]]Data-free[[:space:]]Knowledge[[:space:]]Distillation/400ba431-dcaa-4545-88f2-8b0361ba2923_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NB-GTR_[[:space:]]Narrow-Band[[:space:]]Guided[[:space:]]Turbulence[[:space:]]Removal/c76eed54-1b6a-4a11-9752-38ffa0dd036b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NC-SDF_[[:space:]]Enhancing[[:space:]]Indoor[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]Using[[:space:]]Neural[[:space:]]SDFs[[:space:]]with[[:space:]]View-Dependent[[:space:]]Normal[[:space:]]Compensation/d308c00a-4405-4e9c-a230-0021d3336660_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NC-TTT_[[:space:]]A[[:space:]]Noise[[:space:]]Constrastive[[:space:]]Approach[[:space:]]for[[:space:]]Test-Time[[:space:]]Training/bd059d62-00e8-44dd-afc5-8486d7ebc6ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NEAT_[[:space:]]Distilling[[:space:]]3D[[:space:]]Wireframes[[:space:]]from[[:space:]]Neural[[:space:]]Attraction[[:space:]]Fields/2ecd2fa3-e051-4f87-90aa-cbfb130cb94d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NECA_[[:space:]]Neural[[:space:]]Customizable[[:space:]]Human[[:space:]]Avatar/1fd371b3-1c77-4438-ace5-2f95f345e90e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NICE_[[:space:]]Neurogenesis[[:space:]]Inspired[[:space:]]Contextual[[:space:]]Encoding[[:space:]]for[[:space:]]Replay-free[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning/e2ee9981-d8f0-4b73-ae4d-2cddbed7fe01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NIFTY_[[:space:]]Neural[[:space:]]Object[[:space:]]Interaction[[:space:]]Fields[[:space:]]for[[:space:]]Guided[[:space:]]Human[[:space:]]Motion[[:space:]]Synthesis/a26a4987-a6c2-4a19-82a7-dc8df7105be2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NIVeL_[[:space:]]Neural[[:space:]]Implicit[[:space:]]Vector[[:space:]]Layers[[:space:]]for[[:space:]]Text-to-Vector[[:space:]]Generation/2b4a188b-d422-4fff-bac9-a5ee93dcab15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NOPE_[[:space:]]Novel[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/eaddf8a4-618b-423e-9b47-d54ca4506fe6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NRDF_[[:space:]]Neural[[:space:]]Riemannian[[:space:]]Distance[[:space:]]Fields[[:space:]]for[[:space:]]Learning[[:space:]]Articulated[[:space:]]Pose[[:space:]]Priors/3c4744e1-d7a1-4bb7-9c52-5e6d82e2c55c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NTO3D_[[:space:]]Neural[[:space:]]Target[[:space:]]Object[[:space:]]3D[[:space:]]Reconstruction[[:space:]]with[[:space:]]Segment[[:space:]]Anything/d75ec439-fb66-44d1-8326-a3b002c61709_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NViST_[[:space:]]In[[:space:]]the[[:space:]]Wild[[:space:]]New[[:space:]]View[[:space:]]Synthesis[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]with[[:space:]]Transformers/ef8ae36f-fce2-4270-87ba-de527ac2b27e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Named[[:space:]]Entity[[:space:]]Driven[[:space:]]Zero-Shot[[:space:]]Image[[:space:]]Manipulation/29b9477a-ce33-450d-b6a0-e8383eaf57c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Narrative[[:space:]]Action[[:space:]]Evaluation[[:space:]]with[[:space:]]Prompt-Guided[[:space:]]Multimodal[[:space:]]Interaction/9f79c1fa-6fc9-4b63-b0a8-abfc48152309_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Naturally[[:space:]]Supervised[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding[[:space:]]with[[:space:]]Language-Regularized[[:space:]]Concept[[:space:]]Learners/098143a6-2834-44b7-90c6-1c263dd6e214_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Navigate[[:space:]]Beyond[[:space:]]Shortcuts_[[:space:]]Debiased[[:space:]]Learning[[:space:]]Through[[:space:]]the[[:space:]]Lens[[:space:]]of[[:space:]]Neural[[:space:]]Collapse/c60a89d3-50ed-43e1-9489-517c526d3ab0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Navigating[[:space:]]Beyond[[:space:]]Dropout_[[:space:]]An[[:space:]]Intriguing[[:space:]]Solution[[:space:]]towards[[:space:]]Generalizable[[:space:]]Image[[:space:]]Super[[:space:]]Resolution/306f09fc-f3eb-4f7a-8e1c-e7932830585d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NeISF_[[:space:]]Neural[[:space:]]Incident[[:space:]]Stokes[[:space:]]Field[[:space:]]for[[:space:]]Geometry[[:space:]]and[[:space:]]Material[[:space:]]Estimation/0ccf1001-f8bd-45cf-af5d-a6678647b774_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NeLF-Pro_[[:space:]]Neural[[:space:]]Light[[:space:]]Field[[:space:]]Probes[[:space:]]for[[:space:]]Multi-Scale[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/582c225d-f7ae-4dbf-b523-c3d7d61b55fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NeRF[[:space:]]Analogies_[[:space:]]Example-Based[[:space:]]Visual[[:space:]]Attribute[[:space:]]Transfer[[:space:]]for[[:space:]]NeRFs/18430570-b9c7-4f0b-bc14-bce6d48fd341_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NeRF[[:space:]]Director_[[:space:]]Revisiting[[:space:]]View[[:space:]]Selection[[:space:]]in[[:space:]]Neural[[:space:]]Volume[[:space:]]Rendering/af270895-9751-4bb4-a280-a3905369976f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NeRF[[:space:]]On-the-go_[[:space:]]Exploiting[[:space:]]Uncertainty[[:space:]]for[[:space:]]Distractor-free[[:space:]]NeRFs[[:space:]]in[[:space:]]the[[:space:]]Wild/84919b4e-7dcd-414d-8d02-3749ba7af706_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NeRF-HuGS_[[:space:]]Improved[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]in[[:space:]]Non-static[[:space:]]Scenes[[:space:]]Using[[:space:]]Heuristics-Guided[[:space:]]Segmentation/ddc8147a-c1a1-4504-bf06-6c58b9172cc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NeRFCodec_[[:space:]]Neural[[:space:]]Feature[[:space:]]Compression[[:space:]]Meets[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Memory-Efficient[[:space:]]Scene[[:space:]]Representation/db013d96-7fde-4b15-9ee1-0e7d5955b6d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NeRFDeformer_[[:space:]]NeRF[[:space:]]Transformation[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]View[[:space:]]via[[:space:]]3D[[:space:]]Scene[[:space:]]Flows/1998c11c-e67a-4171-a47c-34a4c98322a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NeRFiller_[[:space:]]Completing[[:space:]]Scenes[[:space:]]via[[:space:]]Generative[[:space:]]3D[[:space:]]Inpainting/da0b62be-9013-4de6-91c8-9a9b6f214742_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NeRSP_[[:space:]]Neural[[:space:]]3D[[:space:]]Reconstruction[[:space:]]for[[:space:]]Reflective[[:space:]]Objects[[:space:]]with[[:space:]]Sparse[[:space:]]Polarized[[:space:]]Images/f1b474cc-2bda-4c5d-865c-2de0b6709deb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Nearest[[:space:]]is[[:space:]]Not[[:space:]]Dearest_[[:space:]]Towards[[:space:]]Practical[[:space:]]Defense[[:space:]]against[[:space:]]Quantization-conditioned[[:space:]]Backdoor[[:space:]]Attacks/c54b5229-6942-4a6c-8508-d94dc7a237d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neighbor[[:space:]]Relations[[:space:]]Matter[[:space:]]in[[:space:]]Video[[:space:]]Scene[[:space:]]Detection/a316b7f0-bcaf-4927-aefa-20ac75981971_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NetTrack_[[:space:]]Tracking[[:space:]]Highly[[:space:]]Dynamic[[:space:]]Objects[[:space:]]with[[:space:]]a[[:space:]]Net/f1a6d7e4-8a95-456e-b5da-d1f46f82a3cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NeuRAD_[[:space:]]Neural[[:space:]]Rendering[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/12a13340-3f2d-4a26-8d60-37631995ae35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]3D[[:space:]]Strokes_[[:space:]]Creating[[:space:]]Stylized[[:space:]]3D[[:space:]]Scenes[[:space:]]with[[:space:]]Vectorized[[:space:]]3D[[:space:]]Strokes/3e7598ea-ea1f-4afc-8f9a-7d996cea28d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Clustering[[:space:]]based[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/2abb19c7-e11d-4368-8914-fd2dbd1d600f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Directional[[:space:]]Encoding[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Accurate[[:space:]]View-Dependent[[:space:]]Appearance[[:space:]]Modeling/552e4ff1-c267-4ccb-88c6-60d36dd2aabc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Exposure[[:space:]]Fusion[[:space:]]for[[:space:]]High-Dynamic[[:space:]]Range[[:space:]]Object[[:space:]]Detection/42458e65-8ed8-4db6-b81f-a6eda14112d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Fields[[:space:]]as[[:space:]]Distributions_[[:space:]]Signal[[:space:]]Processing[[:space:]]Beyond[[:space:]]Euclidean[[:space:]]Space/c97644a4-1ac7-4fbe-aba4-282c59672a20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Implicit[[:space:]]Morphing[[:space:]]of[[:space:]]Face[[:space:]]Images/f26aec0f-b093-422a-8a0e-5af02762c80b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Implicit[[:space:]]Representation[[:space:]]for[[:space:]]Building[[:space:]]Digital[[:space:]]Twins[[:space:]]of[[:space:]]Unknown[[:space:]]Articulated[[:space:]]Objects/848c6059-aa0d-44c9-bee4-b2531169ebca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Lineage/87080a35-305e-4c2f-a498-76819ef092c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Markov[[:space:]]Random[[:space:]]Field[[:space:]]for[[:space:]]Stereo[[:space:]]Matching/8b2d487d-ab55-439d-921c-bdfc83b6068f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Modes_[[:space:]]Self-supervised[[:space:]]Learning[[:space:]]of[[:space:]]Nonlinear[[:space:]]Modal[[:space:]]Subspaces/4a66b7ef-3737-4f24-954f-c1bd2cc05233_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Parametric[[:space:]]Gaussians[[:space:]]for[[:space:]]Monocular[[:space:]]Non-Rigid[[:space:]]Object[[:space:]]Reconstruction/ff5e7807-0eb5-4e81-8af8-7a562296e9aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Point[[:space:]]Cloud[[:space:]]Diffusion[[:space:]]for[[:space:]]Disentangled[[:space:]]3D[[:space:]]Shape[[:space:]]and[[:space:]]Appearance[[:space:]]Generation/2dd6324f-6731-40c1-b624-a25e0c7a8a63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Redshift_[[:space:]]Random[[:space:]]Networks[[:space:]]are[[:space:]]not[[:space:]]Random[[:space:]]Functions/4b6baacb-5087-4bc9-9d7c-97d6c86e62ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Refinement[[:space:]]for[[:space:]]Absolute[[:space:]]Pose[[:space:]]Regression[[:space:]]with[[:space:]]Feature[[:space:]]Synthesis/4a995ad7-316e-4c98-aee2-e34bc35bbac3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Sign[[:space:]]Actors_[[:space:]]A[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]3D[[:space:]]Sign[[:space:]]Language[[:space:]]Production[[:space:]]from[[:space:]]Text/7b4b673b-5e27-491b-b032-42bc7e800f04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Spline[[:space:]]Fields[[:space:]]for[[:space:]]Burst[[:space:]]Image[[:space:]]Fusion[[:space:]]and[[:space:]]Layer[[:space:]]Separation/2d6040ef-69be-45ce-bd97-3b8a70b0f2be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Super-Resolution[[:space:]]for[[:space:]]Real-time[[:space:]]Rendering[[:space:]]with[[:space:]]Radiance[[:space:]]Demodulation/50925100-2364-4853-ae01-3e2ff14a1826_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Underwater[[:space:]]Scene[[:space:]]Representation/a0f59c25-ca6a-4e0b-b937-c936d88cfe40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Video[[:space:]]Compression[[:space:]]with[[:space:]]Feature[[:space:]]Modulation/a216f4e1-2542-4139-85a6-fc1a68182c93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Neural[[:space:]]Visibility[[:space:]]Field[[:space:]]for[[:space:]]Uncertainty-Driven[[:space:]]Active[[:space:]]Mapping/466fcb11-2b97-45ba-8c7f-f34d71fd32b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NightCC_[[:space:]]Nighttime[[:space:]]Color[[:space:]]Constancy[[:space:]]via[[:space:]]Adaptive[[:space:]]Channel[[:space:]]Masking/bb217ef5-d1e2-433c-90ef-5ef0145f65f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/No[[:space:]]More[[:space:]]Ambiguity[[:space:]]in[[:space:]]360deg[[:space:]]Room[[:space:]]Layout[[:space:]]via[[:space:]]Bi-Layout[[:space:]]Estimation/d0053989-45c0-4abd-a024-d4089b4b81c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/No[[:space:]]Time[[:space:]]to[[:space:]]Train_[[:space:]]Empowering[[:space:]]Non-Parametric[[:space:]]Networks[[:space:]]for[[:space:]]Few-shot[[:space:]]3D[[:space:]]Scene[[:space:]]Segmentation/53f342a3-4322-4441-862c-16fc1e76012e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NoiseCLR_[[:space:]]A[[:space:]]Contrastive[[:space:]]Learning[[:space:]]Approach[[:space:]]for[[:space:]]Unsupervised[[:space:]]Discovery[[:space:]]of[[:space:]]Interpretable[[:space:]]Directions[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/121eba10-dcc2-40af-a11e-0abc3ed2d088_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/NoiseCollage_[[:space:]]A[[:space:]]Layout-Aware[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Model[[:space:]]Based[[:space:]]on[[:space:]]Noise[[:space:]]Cropping[[:space:]]and[[:space:]]Merging/33cdc9fb-7ebe-458f-bec3-1cf01b0abb59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Noisy[[:space:]]One-point[[:space:]]Homographies[[:space:]]are[[:space:]]Surprisingly[[:space:]]Good/f2180c9f-e49e-4c0b-a915-0eda8f690d31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Noisy-Correspondence[[:space:]]Learning[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Person[[:space:]]Re-identification/c98f44f3-22b1-4601-a013-59adaa62d76d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Non-Rigid[[:space:]]Structure-from-Motion_[[:space:]]Temporally-Smooth[[:space:]]Procrustean[[:space:]]Alignment[[:space:]]and[[:space:]]Spatially-Variant[[:space:]]Deformation[[:space:]]Modeling/07f188fb-22a7-4e25-8703-693ecf683871_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Non-autoregressive[[:space:]]Sequence-to-Sequence[[:space:]]Vision-Language[[:space:]]Models/43821a3b-e4f9-4f3f-97dc-b7b065fa619e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Normalizing[[:space:]]Flows[[:space:]]on[[:space:]]the[[:space:]]Product[[:space:]]Space[[:space:]]of[[:space:]]SO(3)[[:space:]]Manifolds[[:space:]]for[[:space:]]Probabilistic[[:space:]]Human[[:space:]]Pose[[:space:]]Modeling/35f57cb3-d51e-4d54-8c57-ccab25d6f74e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Not[[:space:]]All[[:space:]]Classes[[:space:]]Stand[[:space:]]on[[:space:]]Same[[:space:]]Embeddings_[[:space:]]Calibrating[[:space:]]a[[:space:]]Semantic[[:space:]]Distance[[:space:]]with[[:space:]]Metric[[:space:]]Tensor/6cdec78b-9ee0-4dd4-8861-4c133b15c8de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Not[[:space:]]All[[:space:]]Prompts[[:space:]]Are[[:space:]]Secure_[[:space:]]A[[:space:]]Switchable[[:space:]]Backdoor[[:space:]]Attack[[:space:]]Against[[:space:]]Pre-trained[[:space:]]Vision[[:space:]]Transfomers/ce5e96be-b4eb-4c2f-b6d4-3380b1e5d1fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Not[[:space:]]All[[:space:]]Voxels[[:space:]]Are[[:space:]]Equal_[[:space:]]Hardness-Aware[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion[[:space:]]with[[:space:]]Self-Distillation/dca30e01-8645-4b40-8efc-d4e4762ecd9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Novel[[:space:]]Class[[:space:]]Discovery[[:space:]]for[[:space:]]Ultra-Fine-Grained[[:space:]]Visual[[:space:]]Categorization/56edef84-e1c3-426e-97fb-a5ed288f6523_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]View-Dependent[[:space:]]Effects[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/f8c24963-2b00-4b36-9692-8952e9bb794d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OA-CNNs_[[:space:]]Omni-Adaptive[[:space:]]Sparse[[:space:]]CNNs[[:space:]]for[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation/9268543d-1734-4bb8-8640-27c3af8215dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OAKINK2_[[:space:]]A[[:space:]]Dataset[[:space:]]of[[:space:]]Bimanual[[:space:]]Hands-Object[[:space:]]Manipulation[[:space:]]in[[:space:]]Complex[[:space:]]Task[[:space:]]Completion/631ab617-5489-415e-b29b-3c34aa4efa18_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OCAI_[[:space:]]Improving[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation[[:space:]]by[[:space:]]Occlusion[[:space:]]and[[:space:]]Consistency[[:space:]]Aware[[:space:]]Interpolation/21da4538-e316-4132-9939-3b0a88f01192_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ODCR_[[:space:]]Orthogonal[[:space:]]Decoupling[[:space:]]Contrastive[[:space:]]Regularization[[:space:]]for[[:space:]]Unpaired[[:space:]]Image[[:space:]]Dehazing/55aeff40-8986-4a53-add6-e6606c1f12f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ODIN_[[:space:]]A[[:space:]]Single[[:space:]]Model[[:space:]]for[[:space:]]2D[[:space:]]and[[:space:]]3D[[:space:]]Segmentation/52879bdd-ecfa-4ccd-8e8c-6451e648eab7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ODM_[[:space:]]A[[:space:]]Text-Image[[:space:]]Further[[:space:]]Alignment[[:space:]]Pre-training[[:space:]]Approach[[:space:]]for[[:space:]]Scene[[:space:]]Text[[:space:]]Detection[[:space:]]and[[:space:]]Spotting/e1a61ef9-3f69-4ae4-87b3-85fa75225abb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OED_[[:space:]]Towards[[:space:]]One-stage[[:space:]]End-to-End[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/1c493e64-66e5-49c5-b16e-c7a0058a78ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OHTA_[[:space:]]One-shot[[:space:]]Hand[[:space:]]Avatar[[:space:]]via[[:space:]]Data-driven[[:space:]]Implicit[[:space:]]Priors/4961aa57-2680-4a2d-bfac-c9cc02068b0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OMG-Seg_[[:space:]]Is[[:space:]]One[[:space:]]Model[[:space:]]Good[[:space:]]Enough[[:space:]]For[[:space:]]All[[:space:]]Segmentation_/ffdcdd06-9c8a-4644-a85a-ff6d4707c4a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OMG_[[:space:]]Towards[[:space:]]Open-vocabulary[[:space:]]Motion[[:space:]]Generation[[:space:]]via[[:space:]]Mixture[[:space:]]of[[:space:]]Controllers/7eb3f111-a53c-4e0c-85b7-a61d858dc1ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OOSTraj_[[:space:]]Out-of-Sight[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]With[[:space:]]Vision-Positioning[[:space:]]Denoising/10db9e40-3089-46c5-ab6a-f22c3e826aaa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OPERA_[[:space:]]Alleviating[[:space:]]Hallucination[[:space:]]in[[:space:]]Multi-Modal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]via[[:space:]]Over-Trust[[:space:]]Penalty[[:space:]]and[[:space:]]Retrospection-Allocation/fdf56278-05c8-4cbb-b470-257247db9a7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OST_[[:space:]]Refining[[:space:]]Text[[:space:]]Knowledge[[:space:]]with[[:space:]]Optimal[[:space:]]Spatio-Temporal[[:space:]]Descriptor[[:space:]]for[[:space:]]General[[:space:]]Video[[:space:]]Recognition/64c9d1f7-c14f-4b1d-b79c-4f03f94c0125_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OTE_[[:space:]]Exploring[[:space:]]Accurate[[:space:]]Scene[[:space:]]Text[[:space:]]Recognition[[:space:]]Using[[:space:]]One[[:space:]]Token/68c8027a-8443-4e66-a7d0-4a79e2c067b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OVER-NAV_[[:space:]]Elevating[[:space:]]Iterative[[:space:]]Vision-and-Language[[:space:]]Navigation[[:space:]]with[[:space:]]Open-Vocabulary[[:space:]]Detection[[:space:]]and[[:space:]]StructurEd[[:space:]]Representation/0199164c-1c39-4ae0-a577-d6420be428a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OVFoodSeg_[[:space:]]Elevating[[:space:]]Open-Vocabulary[[:space:]]Food[[:space:]]Image[[:space:]]Segmentation[[:space:]]via[[:space:]]Image-Informed[[:space:]]Textual[[:space:]]Representation/bb4c87ea-a3e2-4059-aee9-2e9b6ee52838_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OVMR_[[:space:]]Open-Vocabulary[[:space:]]Recognition[[:space:]]with[[:space:]]Multi-Modal[[:space:]]References/050c068b-e537-43e0-8921-54b14a0eef81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Object[[:space:]]Dynamics[[:space:]]Modeling[[:space:]]with[[:space:]]Hierarchical[[:space:]]Point[[:space:]]Cloud-based[[:space:]]Representations/34914316-b8ec-45e1-98ed-d3896f4abbb2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]the[[:space:]]Aggregation[[:space:]]of[[:space:]]Diffusion[[:space:]]Features/531f2d48-abeb-407f-accb-57912c5242f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Object[[:space:]]Recognition[[:space:]]as[[:space:]]Next[[:space:]]Token[[:space:]]Prediction/cb03de9f-afd0-4bc5-8e22-fde6f5e68ec8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Objects[[:space:]]as[[:space:]]Volumes_[[:space:]]A[[:space:]]Stochastic[[:space:]]Geometry[[:space:]]View[[:space:]]of[[:space:]]Opaque[[:space:]]Solids/18cff6e5-4cd9-463b-8e25-fec189228c16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Observation-Guided[[:space:]]Diffusion[[:space:]]Probabilistic[[:space:]]Models/38920473-3049-4e1b-8d67-38d2c932284b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Omni-Q_[[:space:]]Omni-Directional[[:space:]]Scene[[:space:]]Understanding[[:space:]]for[[:space:]]Unsupervised[[:space:]]Visual[[:space:]]Grounding/193048ef-0f36-42fa-8325-ace9d40f5c7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Omni-SMoLA_[[:space:]]Boosting[[:space:]]Generalist[[:space:]]Multimodal[[:space:]]Models[[:space:]]with[[:space:]]Soft[[:space:]]Mixture[[:space:]]of[[:space:]]Low-rank[[:space:]]Experts/4738c0c2-0779-4d25-9622-85d6ce099342_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OmniGlue_[[:space:]]Generalizable[[:space:]]Feature[[:space:]]Matching[[:space:]]with[[:space:]]Foundation[[:space:]]Model[[:space:]]Guidance/74b2d84c-62bd-4ea8-a532-0a61889217fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OmniLocalRF_[[:space:]]Omnidirectional[[:space:]]Local[[:space:]]Radiance[[:space:]]Fields[[:space:]]from[[:space:]]Dynamic[[:space:]]Videos/52139af7-1717-41df-afe6-f56c3c54f26e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OmniMedVQA_[[:space:]]A[[:space:]]New[[:space:]]Large-Scale[[:space:]]Comprehensive[[:space:]]Evaluation[[:space:]]Benchmark[[:space:]]for[[:space:]]Medical[[:space:]]LVLM/15cdd46a-ffd2-463b-8b5b-87ce2495c410_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OmniMotionGPT_[[:space:]]Animal[[:space:]]Motion[[:space:]]Generation[[:space:]]with[[:space:]]Limited[[:space:]]Data/71e013e9-b966-4a64-a5ef-11514d9454a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OmniParser_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Text[[:space:]]Spotting[[:space:]]Key[[:space:]]Information[[:space:]]Extraction[[:space:]]and[[:space:]]Table[[:space:]]Recognition/ca62ec37-a9f3-4138-a128-c4ca363f3ca0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OmniSDF_[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]using[[:space:]]Omnidirectional[[:space:]]Signed[[:space:]]Distance[[:space:]]Functions[[:space:]]and[[:space:]]Adaptive[[:space:]]Binoctrees/4fd9da50-9f80-4271-abeb-cebb55d7a55f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OmniSeg3D_[[:space:]]Omniversal[[:space:]]3D[[:space:]]Segmentation[[:space:]]via[[:space:]]Hierarchical[[:space:]]Contrastive[[:space:]]Learning/c4562839-a2a2-438b-b364-a5f771331479_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OmniVec2[[:space:]]-[[:space:]]A[[:space:]]Novel[[:space:]]Transformer[[:space:]]based[[:space:]]Network[[:space:]]for[[:space:]]Large[[:space:]]Scale[[:space:]]Multimodal[[:space:]]and[[:space:]]Multitask[[:space:]]Learning/6ec3ccd8-d1c0-4b40-a484-484c0b6d9113_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OmniViD_[[:space:]]A[[:space:]]Generative[[:space:]]Framework[[:space:]]for[[:space:]]Universal[[:space:]]Video[[:space:]]Understanding/0b01278a-16fd-4ba2-9c76-8164fcce4122_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/On[[:space:]]Exact[[:space:]]Inversion[[:space:]]of[[:space:]]DPM-Solvers/7b01ea9e-9c26-40f4-8a42-68a30eca8397_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/On[[:space:]]Scaling[[:space:]]Up[[:space:]]a[[:space:]]Multilingual[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Model/7b5bf043-2b45-45dc-9566-f37a556e6240_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/On[[:space:]]Train-Test[[:space:]]Class[[:space:]]Overlap[[:space:]]and[[:space:]]Detection[[:space:]]for[[:space:]]Image[[:space:]]Retrieval/b89b851c-4d4c-4b2f-8efd-527648548115_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/On[[:space:]]the[[:space:]]Content[[:space:]]Bias[[:space:]]in[[:space:]]Frechet[[:space:]]Video[[:space:]]Distance/4a4c6f30-191d-406c-899b-0c5772424002_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/On[[:space:]]the[[:space:]]Diversity[[:space:]]and[[:space:]]Realism[[:space:]]of[[:space:]]Distilled[[:space:]]Dataset_[[:space:]]An[[:space:]]Efficient[[:space:]]Dataset[[:space:]]Distillation[[:space:]]Paradigm/1ec91a88-d0e6-4b31-b112-41bf4dd6764d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/On[[:space:]]the[[:space:]]Estimation[[:space:]]of[[:space:]]Image-matching[[:space:]]Uncertainty[[:space:]]in[[:space:]]Visual[[:space:]]Place[[:space:]]Recognition/e1d06c78-79ff-4148-89da-90e46f277517_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/On[[:space:]]the[[:space:]]Faithfulness[[:space:]]of[[:space:]]Vision[[:space:]]Transformer[[:space:]]Explanations/e3b88a46-405b-4a47-9619-cedb4839a61e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/On[[:space:]]the[[:space:]]Road[[:space:]]to[[:space:]]Portability_[[:space:]]Compressing[[:space:]]End-to-End[[:space:]]Motion[[:space:]]Planner[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/4cb6cd1f-e4a2-4243-bb1e-282666827231_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/On[[:space:]]the[[:space:]]Robustness[[:space:]]of[[:space:]]Language[[:space:]]Guidance[[:space:]]for[[:space:]]Low-Level[[:space:]]Vision[[:space:]]Tasks_[[:space:]]Findings[[:space:]]from[[:space:]]Depth[[:space:]]Estimation/e4b04aaf-c9ce-4ec5-b3fa-2f1c4b14395c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/On[[:space:]]the[[:space:]]Robustness[[:space:]]of[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models[[:space:]]Against[[:space:]]Image[[:space:]]Adversarial[[:space:]]Attacks/ca0dc1f3-ce54-4685-9724-c0b25b19154e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/On[[:space:]]the[[:space:]]Scalability[[:space:]]of[[:space:]]Diffusion-based[[:space:]]Text-to-Image[[:space:]]Generation/d744a5a2-3844-4177-a4fa-9a82c6cf1277_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/On[[:space:]]the[[:space:]]Test-Time[[:space:]]Zero-Shot[[:space:]]Generalization[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models_[[:space:]]Do[[:space:]]We[[:space:]]Really[[:space:]]Need[[:space:]]Prompt[[:space:]]Learning_/c2b2307d-c835-4d52-86cb-0beca943abab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Once[[:space:]]for[[:space:]]Both_[[:space:]]Single[[:space:]]Stage[[:space:]]of[[:space:]]Importance[[:space:]]and[[:space:]]Sparsity[[:space:]]Search[[:space:]]for[[:space:]]Vision[[:space:]]Transformer[[:space:]]Compression/5efddd4e-48f4-4dff-aa96-567696f0f5ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/One[[:space:]]More[[:space:]]Step_[[:space:]]A[[:space:]]Versatile[[:space:]]Plug-and-Play[[:space:]]Module[[:space:]]for[[:space:]]Rectifying[[:space:]]Diffusion[[:space:]]Schedule[[:space:]]Flaws[[:space:]]and[[:space:]]Enhancing[[:space:]]Low-Frequency[[:space:]]Controls/fc73e359-1fc6-4d01-b59a-dc825cd34f8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/One[[:space:]]Prompt[[:space:]]Word[[:space:]]is[[:space:]]Enough[[:space:]]to[[:space:]]Boost[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]for[[:space:]]Pre-trained[[:space:]]Vision-Language[[:space:]]Models/9786d61b-9bf9-49d0-97e4-97534ee380a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/One-2-3-45++_[[:space:]]Fast[[:space:]]Single[[:space:]]Image[[:space:]]to[[:space:]]3D[[:space:]]Objects[[:space:]]with[[:space:]]Consistent[[:space:]]Multi-View[[:space:]]Generation[[:space:]]and[[:space:]]3D[[:space:]]Diffusion/0eacb321-de48-4de4-95a4-fa0e246ca007_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/One-Class[[:space:]]Face[[:space:]]Anti-spoofing[[:space:]]via[[:space:]]Spoof[[:space:]]Cue[[:space:]]Map-Guided[[:space:]]Feature[[:space:]]Learning/fc5aca44-d13e-472d-b745-d0a31f1ad9f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/One-Prompt[[:space:]]to[[:space:]]Segment[[:space:]]All[[:space:]]Medical[[:space:]]Images/1ffaeb66-b655-4edf-b9e5-f6e2f2164dcb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/One-Shot[[:space:]]Open[[:space:]]Affordance[[:space:]]Learning[[:space:]]with[[:space:]]Foundation[[:space:]]Models/885be3b6-b0d1-483e-bff0-460a838604b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/One-Shot[[:space:]]Structure-Aware[[:space:]]Stylized[[:space:]]Image[[:space:]]Synthesis/89d95d57-d1d7-402f-9087-2adcb706cb0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/One-dimensional[[:space:]]Adapter[[:space:]]to[[:space:]]Rule[[:space:]]Them[[:space:]]All_[[:space:]]Concepts[[:space:]]Diffusion[[:space:]]Models[[:space:]]and[[:space:]]Erasing[[:space:]]Applications/b1de4177-5ac4-4633-916c-3c97aa61a94a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/One-step[[:space:]]Diffusion[[:space:]]with[[:space:]]Distribution[[:space:]]Matching[[:space:]]Distillation/0f1a137b-4ca7-4a5c-8170-1a8d350b9981_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OneFormer3D_[[:space:]]One[[:space:]]Transformer[[:space:]]for[[:space:]]Unified[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation/65f7afe2-bc78-46a6-a898-dadb09e3878b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OneLLM_[[:space:]]One[[:space:]]Framework[[:space:]]to[[:space:]]Align[[:space:]]All[[:space:]]Modalities[[:space:]]with[[:space:]]Language/f096df06-7e53-488d-84fd-2aefc98a42eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OneTracker_[[:space:]]Unifying[[:space:]]Visual[[:space:]]Object[[:space:]]Tracking[[:space:]]with[[:space:]]Foundation[[:space:]]Models[[:space:]]and[[:space:]]Efficient[[:space:]]Tuning/54c9547d-a81d-4037-8df0-1c311761275d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Online[[:space:]]Task-Free[[:space:]]Continual[[:space:]]Generative[[:space:]]and[[:space:]]Discriminative[[:space:]]Learning[[:space:]]via[[:space:]]Dynamic[[:space:]]Cluster[[:space:]]Memory/6ca91bbc-21a5-4f11-810a-4e63f458bfb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Open[[:space:]]Vocabulary[[:space:]]Semantic[[:space:]]Scene[[:space:]]Sketch[[:space:]]Understanding/d5af3c2d-e5e9-4861-960b-2de2b31dc377_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Open-Set[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/c75bc9a8-6c4d-46b5-ba6d-eaddc7b8d535_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Open-Vocabulary[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Foundation[[:space:]]Models/88d18524-eed3-426e-a6f8-cab0015cd2d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Open-Vocabulary[[:space:]]Attention[[:space:]]Maps[[:space:]]with[[:space:]]Token[[:space:]]Optimization[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/e8d7168f-61bf-493e-9f54-48e240ff4327_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Open-Vocabulary[[:space:]]Object[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation/fe65d582-0ed8-4c67-b19c-888055a62b1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Open-Vocabulary[[:space:]]Segmentation[[:space:]]with[[:space:]]Semantic-Assisted[[:space:]]Calibration/f955e6fd-c342-42c2-b27f-8232a01bf3c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Image[[:space:]]Embedding[[:space:]]Balancing/b2c81e33-d74b-4f51-a71e-8128c2594c35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Open-Vocabulary[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/ddb12573-55fa-45d1-bbb3-1c76d13566c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Open-World[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection[[:space:]]via[[:space:]]Multi-modal[[:space:]]Prompts/6c0da766-a79f-42ef-b265-05ad68d48c8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Open-World[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]Including[[:space:]]Class[[:space:]]Similarity/39f16f42-e078-42bb-af5e-ca499bfe6f10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Open3DIS_[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation[[:space:]]with[[:space:]]2D[[:space:]]Mask[[:space:]]Guidance/e2eadb8e-fb12-479f-85d3-70983f8c2531_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Open3DSG_[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Scene[[:space:]]Graphs[[:space:]]from[[:space:]]Point[[:space:]]Clouds[[:space:]]with[[:space:]]Queryable[[:space:]]Objects[[:space:]]and[[:space:]]Open-Set[[:space:]]Relationships/285bd2dd-ab6a-4b43-8605-b5e3c8db3a11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OpenBias_[[:space:]]Open-set[[:space:]]Bias[[:space:]]Detection[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Generative[[:space:]]Models/95efdded-b561-455c-954c-9228226f0dcc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OpenEQA_[[:space:]]Embodied[[:space:]]Question[[:space:]]Answering[[:space:]]in[[:space:]]the[[:space:]]Era[[:space:]]of[[:space:]]Foundation[[:space:]]Models/5b45eb45-160d-4a8b-bf57-f68a55da7bd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OpenESS_[[:space:]]Event-based[[:space:]]Semantic[[:space:]]Scene[[:space:]]Understanding[[:space:]]with[[:space:]]Open[[:space:]]Vocabularies/ff6a6c3e-ea6b-444d-a5a3-e0f849cee722_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OpenStreetView-5M_[[:space:]]The[[:space:]]Many[[:space:]]Roads[[:space:]]to[[:space:]]Global[[:space:]]Visual[[:space:]]Geolocation/b234465c-36fb-48dd-8013-9738ba6d44b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OpticalDR_[[:space:]]A[[:space:]]Deep[[:space:]]Optical[[:space:]]Imaging[[:space:]]Model[[:space:]]for[[:space:]]Privacy-Protective[[:space:]]Depression[[:space:]]Recognition/b2a6110f-3f35-41ed-806e-0ec8403fcffc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Optimal[[:space:]]Transport[[:space:]]Aggregation[[:space:]]for[[:space:]]Visual[[:space:]]Place[[:space:]]Recognition/168dea48-7047-4a8c-8495-27806d9f796a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Optimizing[[:space:]]Diffusion[[:space:]]Noise[[:space:]]Can[[:space:]]Serve[[:space:]]As[[:space:]]Universal[[:space:]]Motion[[:space:]]Priors/d768a6cc-1582-4078-9e1b-ce3a20e193b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OrCo_[[:space:]]Towards[[:space:]]Better[[:space:]]Generalization[[:space:]]via[[:space:]]Orthogonality[[:space:]]and[[:space:]]Contrast[[:space:]]for[[:space:]]Few-Shot[[:space:]]Class-Incremental[[:space:]]Learning/3f769f32-87ca-4ea0-80cb-637eee49f2fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Orchestrate[[:space:]]Latent[[:space:]]Expertise_[[:space:]]Advancing[[:space:]]Online[[:space:]]Continual[[:space:]]Learning[[:space:]]with[[:space:]]Multi-Level[[:space:]]Supervision[[:space:]]and[[:space:]]Reverse[[:space:]]Self-Distillation/1f4a1b24-8f9f-4941-b0be-56630933c96e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/OrthCaps_[[:space:]]An[[:space:]]Orthogonal[[:space:]]CapsNet[[:space:]]with[[:space:]]Sparse[[:space:]]Attention[[:space:]]Routing[[:space:]]and[[:space:]]Pruning/33fbbf57-b0a1-4f4a-ac3a-108130bcde13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Orthogonal[[:space:]]Adaptation[[:space:]]for[[:space:]]Modular[[:space:]]Customization[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/d77a98f7-6853-42b4-87e1-365cf007933f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Osprey_[[:space:]]Pixel[[:space:]]Understanding[[:space:]]with[[:space:]]Visual[[:space:]]Instruction[[:space:]]Tuning/ef9011f7-3f2a-4206-be4f-65206f5043f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Outdoor[[:space:]]Scene[[:space:]]Extrapolation[[:space:]]with[[:space:]]Hierarchical[[:space:]]Generative[[:space:]]Cellular[[:space:]]Automata/e57581b0-f1a1-4246-9d9b-ea85036421cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Overcoming[[:space:]]Generic[[:space:]]Knowledge[[:space:]]Loss[[:space:]]with[[:space:]]Selective[[:space:]]Parameter[[:space:]]Update/6db25361-763f-4548-8ddf-61661d0b8be7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Overload_[[:space:]]Latency[[:space:]]Attacks[[:space:]]on[[:space:]]Object[[:space:]]Detection[[:space:]]for[[:space:]]Edge[[:space:]]Devices/42868373-7a4e-49b1-a60a-e8cc0d9f4bcc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PACER+_[[:space:]]On-Demand[[:space:]]Pedestrian[[:space:]]Animation[[:space:]]Controller[[:space:]]in[[:space:]]Driving[[:space:]]Scenarios/c4e28316-d058-48c7-bedb-3acab35a2481_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PAD_[[:space:]]Patch-Agnostic[[:space:]]Defense[[:space:]]against[[:space:]]Adversarial[[:space:]]Patch[[:space:]]Attacks/3cfe64d5-0102-423a-95dc-e075cc6270a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PAIR[[:space:]]Diffusion_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Multimodal[[:space:]]Object-Level[[:space:]]Image[[:space:]]Editor/9df6ae4e-70ec-4059-bf6d-101f71b8940e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PAPR[[:space:]]in[[:space:]]Motion_[[:space:]]Seamless[[:space:]]Point-level[[:space:]]3D[[:space:]]Scene[[:space:]]Interpolation/9ffc1679-d17a-422b-b6f1-50aa2b14936b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PARA-Drive_[[:space:]]Parallelized[[:space:]]Architecture[[:space:]]for[[:space:]]Real-time[[:space:]]Autonomous[[:space:]]Driving/d9f86298-be6b-47e8-b68e-1b7e703f8385_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PBWR_[[:space:]]Parametric-Building-Wireframe[[:space:]]Reconstruction[[:space:]]from[[:space:]]Aerial[[:space:]]LiDAR[[:space:]]Point[[:space:]]Clouds/75014833-17dd-454a-816f-48ba78742a82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PDF_[[:space:]]A[[:space:]]Probability-Driven[[:space:]]Framework[[:space:]]for[[:space:]]Open[[:space:]]World[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Semantic[[:space:]]Segmentation/62f04cc3-a79a-42ea-8264-e0a95b9d6c3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PEEKABOO_[[:space:]]Interactive[[:space:]]Video[[:space:]]Generation[[:space:]]via[[:space:]]Masked-Diffusion/3bfacf5f-ba4a-4e8a-b039-cc7f40446799_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PEGASUS_[[:space:]]Personalized[[:space:]]Generative[[:space:]]3D[[:space:]]Avatars[[:space:]]with[[:space:]]Composable[[:space:]]Attributes/1704349c-16eb-4d0d-83e8-12db33908eab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PELA_[[:space:]]Learning[[:space:]]Parameter-Efficient[[:space:]]Models[[:space:]]with[[:space:]]Low-Rank[[:space:]]Approximation/e8da0a8f-3c65-4ba1-9f6c-716ac560f83f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PEM_[[:space:]]Prototype-based[[:space:]]Efficient[[:space:]]MaskFormer[[:space:]]for[[:space:]]Image[[:space:]]Segmentation/ab2a1f2b-a379-4f54-bc84-6be19e39565c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PFStorer_[[:space:]]Personalized[[:space:]]Face[[:space:]]Restoration[[:space:]]and[[:space:]]Super-Resolution/4b9335e0-43be-48b0-9268-5a560199764b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PH-Net_[[:space:]]Semi-Supervised[[:space:]]Breast[[:space:]]Lesion[[:space:]]Segmentation[[:space:]]via[[:space:]]Patch-wise[[:space:]]Hardness/3d9665d1-82df-4df5-aa5a-dd307006cbc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PI3D_[[:space:]]Efficient[[:space:]]Text-to-3D[[:space:]]Generation[[:space:]]with[[:space:]]Pseudo-Image[[:space:]]Diffusion/a14eb0e2-a14b-496f-b0fd-13cbef2c5091_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PIA_[[:space:]]Your[[:space:]]Personalized[[:space:]]Image[[:space:]]Animator[[:space:]]via[[:space:]]Plug-and-Play[[:space:]]Modules[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Models/91e9a1c4-1dde-4cb2-b820-fe8d86d986c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PICTURE_[[:space:]]PhotorealistIC[[:space:]]virtual[[:space:]]Try-on[[:space:]]from[[:space:]]UnconstRained[[:space:]]dEsigns/1a700242-b3dd-423c-a94b-769dcf14c6d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PIE-NeRF_[[:space:]]Physics-based[[:space:]]Interactive[[:space:]]Elastodynamics[[:space:]]with[[:space:]]NeRF/bd20cd87-5b26-4c5b-92c6-ecebc3947d73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PIGEON_[[:space:]]Predicting[[:space:]]Image[[:space:]]Geolocations/ce3f002a-38a4-4f60-a80b-97f011b13856_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PIN_[[:space:]]Positional[[:space:]]Insert[[:space:]]Unlocks[[:space:]]Object[[:space:]]Localisation[[:space:]]Abilities[[:space:]]in[[:space:]]VLMs/ddf62a02-e622-4c7b-b1c0-c8dadcb5e06a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PKU-DyMVHumans_[[:space:]]A[[:space:]]Multi-View[[:space:]]Video[[:space:]]Benchmark[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Dynamic[[:space:]]Human[[:space:]]Modeling/71672484-d3d1-4a42-8b6f-207ab730f11b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PLACE_[[:space:]]Adaptive[[:space:]]Layout-Semantic[[:space:]]Fusion[[:space:]]for[[:space:]]Semantic[[:space:]]Image[[:space:]]Synthesis/23424dd5-fb6c-4e6a-8849-235afffb6c56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PLGSLAM_[[:space:]]Progressive[[:space:]]Neural[[:space:]]Scene[[:space:]]Represenation[[:space:]]with[[:space:]]Local[[:space:]]to[[:space:]]Global[[:space:]]Bundle[[:space:]]Adjustment/bf8d3a5a-0512-4192-b33d-c90f58b0517d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PNeRV_[[:space:]]Enhancing[[:space:]]Spatial[[:space:]]Consistency[[:space:]]via[[:space:]]Pyramidal[[:space:]]Neural[[:space:]]Representation[[:space:]]for[[:space:]]Videos/5649fd8d-6036-4b6b-a99b-b25497d0e754_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/POCE_[[:space:]]Primal[[:space:]]Policy[[:space:]]Optimization[[:space:]]with[[:space:]]Conservative[[:space:]]Estimation[[:space:]]for[[:space:]]Multi-constraint[[:space:]]Offline[[:space:]]Reinforcement[[:space:]]Learning/c99668a7-a770-40d8-8404-8f69c6588ad1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/POPDG_[[:space:]]Popular[[:space:]]3D[[:space:]]Dance[[:space:]]Generation[[:space:]]with[[:space:]]PopDanceSet/104be7b4-e1d6-428a-997b-681bb22a4f13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PRDP_[[:space:]]Proximal[[:space:]]Reward[[:space:]]Difference[[:space:]]Prediction[[:space:]]for[[:space:]]Large-Scale[[:space:]]Reward[[:space:]]Finetuning[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/fab2bbef-c22c-4645-a87f-603c609cea09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PREGO_[[:space:]]Online[[:space:]]Mistake[[:space:]]Detection[[:space:]]in[[:space:]]PRocedural[[:space:]]EGOcentric[[:space:]]Videos/0bec65bf-f5b4-47ef-a479-412eaa7b11a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PSDPM_[[:space:]]Prototype-based[[:space:]]Secondary[[:space:]]Discriminative[[:space:]]Pixels[[:space:]]Mining[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/d9057dd4-215a-4ba7-98a5-9b221115f7d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PTM-VQA_[[:space:]]Efficient[[:space:]]Video[[:space:]]Quality[[:space:]]Assessment[[:space:]]Leveraging[[:space:]]Diverse[[:space:]]PreTrained[[:space:]]Models[[:space:]]from[[:space:]]the[[:space:]]Wild/790332e5-7751-4672-87fa-a2f0752d259d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PTQ4SAM_[[:space:]]Post-Training[[:space:]]Quantization[[:space:]]for[[:space:]]Segment[[:space:]]Anything/0676cd26-f2dd-4af3-ae43-6d5b0763d357_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PTT_[[:space:]]Point-Trajectory[[:space:]]Transformer[[:space:]]for[[:space:]]Efficient[[:space:]]Temporal[[:space:]]3D[[:space:]]Object[[:space:]]Detection/d7002cbe-b55a-41dc-a049-2afd7d62667f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PaReNeRF_[[:space:]]Toward[[:space:]]Fast[[:space:]]Large-scale[[:space:]]Dynamic[[:space:]]NeRF[[:space:]]with[[:space:]]Patch-based[[:space:]]Reference/15a3ca23-8001-4802-835d-3838bf17923f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PaSCo_[[:space:]]Urban[[:space:]]3D[[:space:]]Panoptic[[:space:]]Scene[[:space:]]Completion[[:space:]]with[[:space:]]Uncertainty[[:space:]]Awareness/81c5f395-5319-41ae-b3fc-dfa9dfab79e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Paint-it_[[:space:]]Text-to-Texture[[:space:]]Synthesis[[:space:]]via[[:space:]]Deep[[:space:]]Convolutional[[:space:]]Texture[[:space:]]Map[[:space:]]Optimization[[:space:]]and[[:space:]]Physically-Based[[:space:]]Rendering/d3086159-a13f-4b03-9a23-365839af995c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Paint3D_[[:space:]]Paint[[:space:]]Anything[[:space:]]3D[[:space:]]with[[:space:]]Lighting-Less[[:space:]]Texture[[:space:]]Diffusion[[:space:]]Models/9ee3faec-c399-49ba-98ee-26e5b486c632_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PairAug_[[:space:]]What[[:space:]]Can[[:space:]]Augmented[[:space:]]Image-Text[[:space:]]Pairs[[:space:]]Do[[:space:]]for[[:space:]]Radiology_/fa98773c-5db9-42dd-aecf-5e715863d1e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PairDETR[[:space:]]_[[:space:]]Joint[[:space:]]Detection[[:space:]]and[[:space:]]Association[[:space:]]of[[:space:]]Human[[:space:]]Bodies[[:space:]]and[[:space:]]Faces/aa1a0140-0cd6-49a4-83d6-9e8a0ddf166a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Panacea_[[:space:]]Panoramic[[:space:]]and[[:space:]]Controllable[[:space:]]Video[[:space:]]Generation[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/0cdda859-65a2-42eb-b850-7396e36ad76e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Panda-70M_[[:space:]]Captioning[[:space:]]70M[[:space:]]Videos[[:space:]]with[[:space:]]Multiple[[:space:]]Cross-Modality[[:space:]]Teachers/83b79ec3-2715-488a-8afc-8aeaf3922ff4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PanoContext-Former_[[:space:]]Panoramic[[:space:]]Total[[:space:]]Scene[[:space:]]Understanding[[:space:]]with[[:space:]]a[[:space:]]Transformer/d810d840-93bd-4f6c-977d-d3f9419d3c6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PanoOcc_[[:space:]]Unified[[:space:]]Occupancy[[:space:]]Representation[[:space:]]for[[:space:]]Camera-based[[:space:]]3D[[:space:]]Panoptic[[:space:]]Segmentation/d6ab4c45-d940-4e9c-b520-4c53bac6a31e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PanoPose_[[:space:]]Self-supervised[[:space:]]Relative[[:space:]]Pose[[:space:]]Estimation[[:space:]]for[[:space:]]Panoramic[[:space:]]Images/1d055744-8a9c-4a71-ac84-cec231156942_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PanoRecon_[[:space:]]Real-Time[[:space:]]Panoptic[[:space:]]3D[[:space:]]Reconstruction[[:space:]]from[[:space:]]Monocular[[:space:]]Video/18caae1d-abc1-41ce-b5d3-9b2f3f1af3fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ParamISP_[[:space:]]Learned[[:space:]]Forward[[:space:]]and[[:space:]]Inverse[[:space:]]ISPs[[:space:]]using[[:space:]]Camera[[:space:]]Parameters/6e6f7fc1-77fc-428b-b530-f2a531a4289d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Parameter[[:space:]]Efficient[[:space:]]Fine-tuning[[:space:]]via[[:space:]]Cross[[:space:]]Block[[:space:]]Orchestration[[:space:]]for[[:space:]]Segment[[:space:]]Anything[[:space:]]Model/82da1259-2d1a-4e2d-ab51-2c862d1b9639_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Parameter[[:space:]]Efficient[[:space:]]Self-Supervised[[:space:]]Geospatial[[:space:]]Domain[[:space:]]Adaptation/72715c70-8a11-4b07-951a-4e5a4f00f069_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ParameterNet_[[:space:]]Parameters[[:space:]]Are[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]for[[:space:]]Large-scale[[:space:]]Visual[[:space:]]Pretraining[[:space:]]of[[:space:]]Mobile[[:space:]]Networks/4e5fbf25-b66e-4cc4-a91c-ffdacdc52d24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Part-aware[[:space:]]Unified[[:space:]]Representation[[:space:]]of[[:space:]]Language[[:space:]]and[[:space:]]Skeleton[[:space:]]for[[:space:]]Zero-shot[[:space:]]Action[[:space:]]Recognition/c8a82da2-6560-4954-bb91-5514d8b49f91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PartDistill_[[:space:]]3D[[:space:]]Shape[[:space:]]Part[[:space:]]Segmentation[[:space:]]by[[:space:]]Vision-Language[[:space:]]Model[[:space:]]Distillation/cb085cc0-2a23-40a7-8988-9755b94ab446_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Partial-to-Partial[[:space:]]Shape[[:space:]]Matching[[:space:]]with[[:space:]]Geometric[[:space:]]Consistency/c7118972-0a0c-463f-a1b8-213d45dde2f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Passive[[:space:]]Snapshot[[:space:]]Coded[[:space:]]Aperture[[:space:]]Dual-Pixel[[:space:]]RGB-D[[:space:]]Imaging/b85a1121-d303-449b-a325-e796908b5350_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Patch2Self2_[[:space:]]Self-supervised[[:space:]]Denoising[[:space:]]on[[:space:]]Coresets[[:space:]]via[[:space:]]Matrix[[:space:]]Sketching/14a944d4-1072-4170-8ade-226b48a51d2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PatchFusion_[[:space:]]An[[:space:]]End-to-End[[:space:]]Tile-Based[[:space:]]Framework[[:space:]]for[[:space:]]High-Resolution[[:space:]]Monocular[[:space:]]Metric[[:space:]]Depth[[:space:]]Estimation/3539a2c6-a813-45d0-9124-f7a646b9aab6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PeLK_[[:space:]]Parameter-efficient[[:space:]]Large[[:space:]]Kernel[[:space:]]ConvNets[[:space:]]with[[:space:]]Peripheral[[:space:]]Convolution/67c2b402-6798-4cd6-94d6-fc5c1aef87e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PeVL_[[:space:]]Pose-Enhanced[[:space:]]Vision-Language[[:space:]]Model[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Human[[:space:]]Action[[:space:]]Recognition/153c38b9-ed69-47f7-99f9-8977e804a1fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PeerAiD_[[:space:]]Improving[[:space:]]Adversarial[[:space:]]Distillation[[:space:]]from[[:space:]]a[[:space:]]Specialized[[:space:]]Peer[[:space:]]Tutor/e151ddfb-b973-4295-84c5-a92d2ae5cf59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PerAda_[[:space:]]Parameter-Efficient[[:space:]]Federated[[:space:]]Learning[[:space:]]Personalization[[:space:]]with[[:space:]]Generalization[[:space:]]Guarantees/16e8f105-7bb6-4b0c-9260-137545196b43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Perception-Oriented[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation[[:space:]]via[[:space:]]Asymmetric[[:space:]]Blending/8c8ae7ab-3690-497e-bdd4-45d7020dfc71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PerceptionGPT_[[:space:]]Effectively[[:space:]]Fusing[[:space:]]Visual[[:space:]]Perception[[:space:]]into[[:space:]]LLM/d1af003c-6a1f-4a7d-9e2c-97802aaffb17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Perceptual[[:space:]]Assessment[[:space:]]and[[:space:]]Optimization[[:space:]]of[[:space:]]HDR[[:space:]]Image[[:space:]]Rendering/723fc539-844c-4140-bb5f-f29bfa3b728e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Permutation[[:space:]]Equivariance[[:space:]]of[[:space:]]Transformers[[:space:]]and[[:space:]]Its[[:space:]]Applications/c937725d-c99e-48c2-a7c7-a1dde9927559_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Person[[:space:]]in[[:space:]]Place_[[:space:]]Generating[[:space:]]Associative[[:space:]]Skeleton-Guidance[[:space:]]Maps[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Image[[:space:]]Editing/dd982759-54c1-471a-a24b-edc7a0092870_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Person-in-WiFi[[:space:]]3D_[[:space:]]End-to-End[[:space:]]Multi-Person[[:space:]]3D[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]Wi-Fi/eb0cb2ad-2f96-4604-ab79-179b4df12c2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Personalized[[:space:]]Residuals[[:space:]]for[[:space:]]Concept-Driven[[:space:]]Text-to-Image[[:space:]]Generation/20d03ae6-ea8f-488a-a042-92c58d395087_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Perturbing[[:space:]]Attention[[:space:]]Gives[[:space:]]You[[:space:]]More[[:space:]]Bang[[:space:]]for[[:space:]]the[[:space:]]Buck_[[:space:]]Subtle[[:space:]]Imaging[[:space:]]Perturbations[[:space:]]That[[:space:]]Efficiently[[:space:]]Fool[[:space:]]Customized[[:space:]]Diffusion[[:space:]]Models/8588c654-930a-4383-a46a-c498701b582f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Photo-SLAM_[[:space:]]Real-time[[:space:]]Simultaneous[[:space:]]Localization[[:space:]]and[[:space:]]Photorealistic[[:space:]]Mapping[[:space:]]for[[:space:]]Monocular[[:space:]]Stereo[[:space:]]and[[:space:]]RGB-D[[:space:]]Cameras/5c42d1bc-b82d-41d9-9cfd-fc188d6b8698_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PhotoMaker_[[:space:]]Customizing[[:space:]]Realistic[[:space:]]Human[[:space:]]Photos[[:space:]]via[[:space:]]Stacked[[:space:]]ID[[:space:]]Embedding/13ab4726-01ca-48a9-a6d6-fb7b257dc058_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PhyScene_[[:space:]]Physically[[:space:]]Interactable[[:space:]]3D[[:space:]]Scene[[:space:]]Synthesis[[:space:]]for[[:space:]]Embodied[[:space:]]AI/b8c059b4-30c2-4bf2-9823-9fef6131d706_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PhysGaussian_[[:space:]]Physics-Integrated[[:space:]]3D[[:space:]]Gaussians[[:space:]]for[[:space:]]Generative[[:space:]]Dynamics/bcc76c3a-4e4d-45c5-9ce2-03ca17c58e03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PhysPT_[[:space:]]Physics-aware[[:space:]]Pretrained[[:space:]]Transformer[[:space:]]for[[:space:]]Estimating[[:space:]]Human[[:space:]]Dynamics[[:space:]]from[[:space:]]Monocular[[:space:]]Videos/fe4f9050-2cb2-4bff-a48e-12b968774543_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Physical[[:space:]]3D[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]against[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/36e0b279-7f2b-4f0f-90c7-de56a35da499_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Physical[[:space:]]Backdoor_[[:space:]]Towards[[:space:]]Temperature-based[[:space:]]Backdoor[[:space:]]Attacks[[:space:]]in[[:space:]]the[[:space:]]Physical[[:space:]]World/598d39e7-e679-4a5c-9f6f-b93da4661799_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Physical[[:space:]]Property[[:space:]]Understanding[[:space:]]from[[:space:]]Language-Embedded[[:space:]]Feature[[:space:]]Fields/a3a3d196-6a21-40dd-bd09-90b17287f1d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Physics-Aware[[:space:]]Hand-Object[[:space:]]Interaction[[:space:]]Denoising/541707c8-55e6-42f7-88f4-44533128d92f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Physics-guided[[:space:]]Shape-from-Template_[[:space:]]Monocular[[:space:]]Video[[:space:]]Perception[[:space:]]through[[:space:]]Neural[[:space:]]Surrogate[[:space:]]Models/3c1a298c-1084-446e-bb14-d2d65b0328c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Pick-or-Mix_[[:space:]]Dynamic[[:space:]]Channel[[:space:]]Sampling[[:space:]]for[[:space:]]ConvNets/4db02f48-dc1a-4959-be3d-8212c4474732_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PikeLPN_[[:space:]]Mitigating[[:space:]]Overlooked[[:space:]]Inefficiencies[[:space:]]of[[:space:]]Low-Precision[[:space:]]Neural[[:space:]]Networks/2eff09a9-ae4d-4cbe-af46-2c9a62cd4f03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Pink_[[:space:]]Unveiling[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]Referential[[:space:]]Comprehension[[:space:]]for[[:space:]]Multi-modal[[:space:]]LLMs/680ccd67-e937-48f7-b94b-b7706f4439f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Pixel-Aligned[[:space:]]Language[[:space:]]Model/2f8bdcf5-5e06-438d-b254-1861291bccc0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Pixel-level[[:space:]]Semantic[[:space:]]Correspondence[[:space:]]through[[:space:]]Layout-aware[[:space:]]Representation[[:space:]]Learning[[:space:]]and[[:space:]]Multi-scale[[:space:]]Matching[[:space:]]Integration/43537007-0f1e-46d1-899a-3ec9060b0c33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PixelLM_[[:space:]]Pixel[[:space:]]Reasoning[[:space:]]with[[:space:]]Large[[:space:]]Multimodal[[:space:]]Model/359ad9c1-4020-427c-a789-03674ae9db49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PixelRNN_[[:space:]]In-pixel[[:space:]]Recurrent[[:space:]]Neural[[:space:]]Networks[[:space:]]for[[:space:]]End-to-end-optimized[[:space:]]Perception[[:space:]]with[[:space:]]Neural[[:space:]]Sensors/b8339257-e4b0-4368-9ec1-5aab02e1d45b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PlatoNeRF_[[:space:]]3D[[:space:]]Reconstruction[[:space:]]in[[:space:]]Plato's[[:space:]]Cave[[:space:]]via[[:space:]]Single-View[[:space:]]Two-Bounce[[:space:]]Lidar/318980ce-3f23-4e63-8f78-9ffbaa8866fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Plug[[:space:]]and[[:space:]]Play[[:space:]]Active[[:space:]]Learning[[:space:]]for[[:space:]]Object[[:space:]]Detection/a3123346-66f0-4d38-9126-69be23f314ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Plug-and-Play[[:space:]]Diffusion[[:space:]]Distillation/81064853-e3e4-4cde-92f6-64c9793b82fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PoNQ_[[:space:]]a[[:space:]]Neural[[:space:]]QEM-based[[:space:]]Mesh[[:space:]]Representation/74f1b2ae-9620-4937-8540-cdfdcc5f7504_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Point[[:space:]]Cloud[[:space:]]Pre-training[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/7944f0e0-1287-4ae0-b321-7d5b2d90c6d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Point[[:space:]]Segment[[:space:]]and[[:space:]]Count_[[:space:]]A[[:space:]]Generalized[[:space:]]Framework[[:space:]]for[[:space:]]Object[[:space:]]Counting/c7f733c1-3c12-4587-b41a-48c318ce3f9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Point[[:space:]]Transformer[[:space:]]V3_[[:space:]]Simpler[[:space:]]Faster[[:space:]]Stronger/223e2fb3-90fd-4602-a2c3-12ef232ac077_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Point-VOS_[[:space:]]Pointing[[:space:]]Up[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/104e63fa-05fe-4e08-aec3-5f7d3385383a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Point2CAD_[[:space:]]Reverse[[:space:]]Engineering[[:space:]]CAD[[:space:]]Models[[:space:]]from[[:space:]]3D[[:space:]]Point[[:space:]]Clouds/85ceb9e9-956f-418f-bd9d-caa3828be13f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Point2RBox_[[:space:]]Combine[[:space:]]Knowledge[[:space:]]from[[:space:]]Synthetic[[:space:]]Visual[[:space:]]Patterns[[:space:]]for[[:space:]]End-to-end[[:space:]]Oriented[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Single[[:space:]]Point[[:space:]]Supervision/2d7af184-873e-41bf-8c67-8e4e5b66c4ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PointBeV_[[:space:]]A[[:space:]]Sparse[[:space:]]Approach[[:space:]]for[[:space:]]BeV[[:space:]]Predictions/8f374cbc-4b61-413c-a813-0286b9b8cda5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PointInfinity_[[:space:]]Resolution-Invariant[[:space:]]Point[[:space:]]Diffusion[[:space:]]Models/b0c1aa28-c8df-4d2d-bfda-b0779f54ea31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PointOBB_[[:space:]]Learning[[:space:]]Oriented[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Single[[:space:]]Point[[:space:]]Supervision/2fe1e671-7c57-44ee-8099-e0fedda3bb2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PolarMatte_[[:space:]]Fully[[:space:]]Computational[[:space:]]Ground-Truth-Quality[[:space:]]Alpha[[:space:]]Matte[[:space:]]Extraction[[:space:]]for[[:space:]]Images[[:space:]]and[[:space:]]Video[[:space:]]using[[:space:]]Polarized[[:space:]]Screen[[:space:]]Matting/d96f4ab5-5bb5-4818-aa62-6eaaad24b957_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PolarRec_[[:space:]]Improving[[:space:]]Radio[[:space:]]Interferometric[[:space:]]Data[[:space:]]Reconstruction[[:space:]]Using[[:space:]]Polar[[:space:]]Coordinates/e30a0a2f-3d83-4f10-b317-d80632d927bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Polarization[[:space:]]Wavefront[[:space:]]Lidar_[[:space:]]Learning[[:space:]]Large[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]from[[:space:]]Polarized[[:space:]]Wavefronts/a7352aac-b5e5-426f-b020-ad40d6e4a026_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Polos_[[:space:]]Multimodal[[:space:]]Metric[[:space:]]Learning[[:space:]]from[[:space:]]Human[[:space:]]Feedback[[:space:]]for[[:space:]]Image[[:space:]]Captioning/d2462477-0d68-4b8a-8db2-8fb663243e9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Poly[[:space:]]Kernel[[:space:]]Inception[[:space:]]Network[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Detection/dd3c9950-0396-4bcc-b2a8-8205acc31e8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Portrait4D_[[:space:]]Learning[[:space:]]One-Shot[[:space:]]4D[[:space:]]Head[[:space:]]Avatar[[:space:]]Synthesis[[:space:]]using[[:space:]]Synthetic[[:space:]]Data/522fd8d7-9609-40f8-b1cc-4fd5b4db3d1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PortraitBooth_[[:space:]]A[[:space:]]Versatile[[:space:]]Portrait[[:space:]]Model[[:space:]]for[[:space:]]Fast[[:space:]]Identity-preserved[[:space:]]Personalization/d3624f11-1ac5-4536-8885-f3ef48442fb2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Pose[[:space:]]Adapted[[:space:]]Shape[[:space:]]Learning[[:space:]]for[[:space:]]Large-Pose[[:space:]]Face[[:space:]]Reenactment/4c1e688d-71a2-4358-a358-d070463089b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Pose-Guided[[:space:]]Self-Training[[:space:]]with[[:space:]]Two-Stage[[:space:]]Clustering[[:space:]]for[[:space:]]Unsupervised[[:space:]]Landmark[[:space:]]Discovery/9e790ea8-ba20-4107-b3f8-3210d03bb170_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Pose-Transformed[[:space:]]Equivariant[[:space:]]Network[[:space:]]for[[:space:]]3D[[:space:]]Point[[:space:]]Trajectory[[:space:]]Prediction/9a83d229-a860-4b1e-98f2-b7e21eff439b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PoseIRM_[[:space:]]Enhance[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]on[[:space:]]Unseen[[:space:]]Camera[[:space:]]Settings[[:space:]]via[[:space:]]Invariant[[:space:]]Risk[[:space:]]Minimization/139707e3-7491-48fe-a91e-9278f13d4466_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Positive-Unlabeled[[:space:]]Learning[[:space:]]by[[:space:]]Latent[[:space:]]Group-Aware[[:space:]]Meta[[:space:]]Disambiguation/9530e631-b18d-45c4-bc36-a038c516e9eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Posterior[[:space:]]Distillation[[:space:]]Sampling/d7127ed1-2faf-4b57-a4ba-c62184d289ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PostureHMR_[[:space:]]Posture[[:space:]]Transformation[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/8afeca97-0261-435e-b593-3180087dc87e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PrPSeg_[[:space:]]Universal[[:space:]]Proposition[[:space:]]Learning[[:space:]]for[[:space:]]Panoramic[[:space:]]Renal[[:space:]]Pathology[[:space:]]Segmentation/a9fc6fc3-c351-465b-8f98-a88976d538dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Practical[[:space:]]Measurements[[:space:]]of[[:space:]]Translucent[[:space:]]Materials[[:space:]]with[[:space:]]Inter-Pixel[[:space:]]Translucency[[:space:]]Prior/e4ca6a47-465d-441a-aa00-02c232b162bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PracticalDG_[[:space:]]Perturbation[[:space:]]Distillation[[:space:]]on[[:space:]]Vision-Language[[:space:]]Models[[:space:]]for[[:space:]]Hybrid[[:space:]]Domain[[:space:]]Generalization/176e79da-1e74-4256-bcaa-521d42e27d93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Pre-trained[[:space:]]Model[[:space:]]Guided[[:space:]]Fine-Tuning[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Adversarial[[:space:]]Robustness/378c9228-d994-45cd-91fc-4f5bacb04465_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Pre-trained[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Transformers[[:space:]]Are[[:space:]]Few-Shot[[:space:]]Incremental[[:space:]]Learners/547c99bc-a324-4073-b174-e640d3d01fd4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Pre-training[[:space:]]Vision[[:space:]]Models[[:space:]]with[[:space:]]Mandelbulb[[:space:]]Variations/19f2472d-66db-490b-bb58-ece6642aa7c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PredToken_[[:space:]]Predicting[[:space:]]Unknown[[:space:]]Tokens[[:space:]]and[[:space:]]Beyond[[:space:]]with[[:space:]]Coarse-to-Fine[[:space:]]Iterative[[:space:]]Decoding/ef4f829b-d707-425e-9afc-c7ff9af6c382_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Predicated[[:space:]]Diffusion_[[:space:]]Predicate[[:space:]]Logic-Based[[:space:]]Attention[[:space:]]Guidance[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/e9d5ef3e-fad0-4a03-9998-16bb9bd60b31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Preserving[[:space:]]Fairness[[:space:]]Generalization[[:space:]]in[[:space:]]Deepfake[[:space:]]Detection/c624f7a7-a975-4d45-8db3-a8cf2bac8143_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Previously[[:space:]]on[[:space:]]...[[:space:]]From[[:space:]]Recaps[[:space:]]to[[:space:]]Story[[:space:]]Summarization/a0e90dfa-bf12-44bf-916a-374a814887e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Privacy-Preserving[[:space:]]Face[[:space:]]Recognition[[:space:]]Using[[:space:]]Trainable[[:space:]]Feature[[:space:]]Subtraction/5865e00c-a618-4169-9631-a728819d6c4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Privacy-Preserving[[:space:]]Optics[[:space:]]for[[:space:]]Enhancing[[:space:]]Protection[[:space:]]in[[:space:]]Face[[:space:]]De-Identification/d5081748-5fb1-4115-8c21-8a01779d514f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ProMark_[[:space:]]Proactive[[:space:]]Diffusion[[:space:]]Watermarking[[:space:]]for[[:space:]]Causal[[:space:]]Attribution/c6b384b0-a56d-42d9-9df9-b21b5b22f3be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ProMotion_[[:space:]]Prototypes[[:space:]]As[[:space:]]Motion[[:space:]]Learners/244ef0a3-d763-4985-bc28-4e4843bf4778_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ProS_[[:space:]]Prompting-to-simulate[[:space:]]Generalized[[:space:]]knowledge[[:space:]]for[[:space:]]Universal[[:space:]]Cross-Domain[[:space:]]Retrieval/ebd3c7b1-3edf-401d-a6b2-24967a853d7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ProTeCt_[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Taxonomic[[:space:]]Open[[:space:]]Set[[:space:]]Classification/022010f8-6dee-4cad-9d8c-528d8c6ec19f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Probabilistic[[:space:]]Sampling[[:space:]]of[[:space:]]Balanced[[:space:]]K-Means[[:space:]]using[[:space:]]Adiabatic[[:space:]]Quantum[[:space:]]Computing/e8a19710-df88-409f-b957-4f36371b1173_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Probabilistic[[:space:]]Speech-Driven[[:space:]]3D[[:space:]]Facial[[:space:]]Motion[[:space:]]Synthesis_[[:space:]]New[[:space:]]Benchmarks[[:space:]]Methods[[:space:]]and[[:space:]]Applications/0009d358-9190-459e-8345-20ee121f9ed0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Probing[[:space:]]Synergistic[[:space:]]High-Order[[:space:]]Interaction[[:space:]]in[[:space:]]Infrared[[:space:]]and[[:space:]]Visible[[:space:]]Image[[:space:]]Fusion/8aabeead-a9c1-4690-8a0e-cc0b42bad4e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Probing[[:space:]]the[[:space:]]3D[[:space:]]Awareness[[:space:]]of[[:space:]]Visual[[:space:]]Foundation[[:space:]]Models/58c4101e-04b3-4ac7-82ac-0409f781a407_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Producing[[:space:]]and[[:space:]]Leveraging[[:space:]]Online[[:space:]]Map[[:space:]]Uncertainty[[:space:]]in[[:space:]]Trajectory[[:space:]]Prediction/f2a200bf-2f6d-46c4-912b-6938d23afb8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Programmable[[:space:]]Motion[[:space:]]Generation[[:space:]]for[[:space:]]Open-Set[[:space:]]Motion[[:space:]]Control[[:space:]]Tasks/7033c1ce-7714-419c-92ed-c492f6a60167_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Progress-Aware[[:space:]]Online[[:space:]]Action[[:space:]]Segmentation[[:space:]]for[[:space:]]Egocentric[[:space:]]Procedural[[:space:]]Task[[:space:]]Videos/2595e2d9-b44d-41b6-8f19-4281d6f38d6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Progressive[[:space:]]Divide-and-Conquer[[:space:]]via[[:space:]]Subsampling[[:space:]]Decomposition[[:space:]]for[[:space:]]Accelerated[[:space:]]MRI/e36999ff-440f-4d8f-850b-a99d4038b0c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Progressive[[:space:]]Semantic-Guided[[:space:]]Vision[[:space:]]Transformer[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Learning/8837f52e-4dd8-486c-84e5-4d78ed226a21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Projecting[[:space:]]Trackable[[:space:]]Thermal[[:space:]]Patterns[[:space:]]for[[:space:]]Dynamic[[:space:]]Computer[[:space:]]Vision/8ba310b9-f072-4209-8532-c49b6edf953c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Prompt[[:space:]]Augmentation[[:space:]]for[[:space:]]Self-supervised[[:space:]]Text-guided[[:space:]]Image[[:space:]]Manipulation/9eeea513-ecf6-4fe7-b21d-1615ec14d456_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Prompt[[:space:]]Highlighter_[[:space:]]Interactive[[:space:]]Control[[:space:]]for[[:space:]]Multi-Modal[[:space:]]LLMs/532aadf8-0c5a-4519-979f-20b2b1f071b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Prompt[[:space:]]Learning[[:space:]]via[[:space:]]Meta-Regularization/61480337-6c94-45d7-918d-6c21bc4b9a92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Prompt-Driven[[:space:]]Dynamic[[:space:]]Object-Centric[[:space:]]Learning[[:space:]]for[[:space:]]Single[[:space:]]Domain[[:space:]]Generalization/fa423c43-f7fd-485b-b75b-e7888bf153ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Prompt-Driven[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation[[:space:]]with[[:space:]]Instance[[:space:]]Contrasting/43df0ca6-50ea-4646-8f21-c61917cd4f3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Prompt-Enhanced[[:space:]]Multiple[[:space:]]Instance[[:space:]]Learning[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/04e58dac-ac47-4d60-962f-abdf21f1022d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Prompt-Free[[:space:]]Diffusion_[[:space:]]Taking[[:space:]]'Text'[[:space:]]out[[:space:]]of[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/e88fe996-d235-4c11-bb8a-a9c350671f75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Prompt3D_[[:space:]]Random[[:space:]]Prompt[[:space:]]Assisted[[:space:]]Weakly-Supervised[[:space:]]3D[[:space:]]Object[[:space:]]Detection/ec237acd-bd51-4b06-a43b-c2453097f6bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PromptAD_[[:space:]]Learning[[:space:]]Prompts[[:space:]]with[[:space:]]only[[:space:]]Normal[[:space:]]Samples[[:space:]]for[[:space:]]Few-Shot[[:space:]]Anomaly[[:space:]]Detection/fa1491bc-dbd7-424d-9812-fec8079c4787_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PromptCoT_[[:space:]]Align[[:space:]]Prompt[[:space:]]Distribution[[:space:]]via[[:space:]]Adapted[[:space:]]Chain-of-Thought/c0afb0b5-6d83-4c07-9792-0db1f53f38e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/PromptKD_[[:space:]]Unsupervised[[:space:]]Prompt[[:space:]]Distillation[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/6c2b57f2-bcf9-4927-9a6a-a774195345f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Promptable[[:space:]]Behaviors_[[:space:]]Personalizing[[:space:]]Multi-Objective[[:space:]]Rewards[[:space:]]from[[:space:]]Human[[:space:]]Preferences/aa814e82-ef14-486e-bdc3-a110128694ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Prompting[[:space:]]Hard[[:space:]]or[[:space:]]Hardly[[:space:]]Prompting_[[:space:]]Prompt[[:space:]]Inversion[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/f66b9775-1159-461b-bc0e-d493bd2f94a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Prompting[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models[[:space:]]for[[:space:]]Pathology[[:space:]]Image[[:space:]]Analysis/555a28b8-dd33-4223-8240-45978e96f50a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ProxyCap_[[:space:]]Real-time[[:space:]]Monocular[[:space:]]Full-body[[:space:]]Capture[[:space:]]in[[:space:]]World[[:space:]]Space[[:space:]]via[[:space:]]Human-Centric[[:space:]]Proxy-to-Motion[[:space:]]Learning/3c31a44e-30c3-417b-af00-8d10d02afaee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Pseudo[[:space:]]Label[[:space:]]Refinery[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]on[[:space:]]Cross-dataset[[:space:]]3D[[:space:]]Object[[:space:]]Detection/93fadacd-19a7-4e1b-ae95-074dedde35fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Psychometry_[[:space:]]An[[:space:]]Omnifit[[:space:]]Model[[:space:]]for[[:space:]]Image[[:space:]]Reconstruction[[:space:]]from[[:space:]]Human[[:space:]]Brain[[:space:]]Activity/cf779e1f-55ed-473e-94cb-014181539060_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Puff-Net_[[:space:]]Efficient[[:space:]]Style[[:space:]]Transfer[[:space:]]with[[:space:]]Pure[[:space:]]Content[[:space:]]and[[:space:]]Style[[:space:]]Feature[[:space:]]Fusion[[:space:]]Network/55c8436e-1990-4fb9-9b6d-f0508033864c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Purified[[:space:]]and[[:space:]]Unified[[:space:]]Steganographic[[:space:]]Network/09232b8e-b7e4-4bc1-9664-b862886b20d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Putting[[:space:]]the[[:space:]]Object[[:space:]]Back[[:space:]]into[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/bae8acf7-a488-4cd7-b781-30d9a124f869_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Q-Instruct_[[:space:]]Improving[[:space:]]Low-level[[:space:]]Visual[[:space:]]Abilities[[:space:]]for[[:space:]]Multi-modality[[:space:]]Foundation[[:space:]]Models/773bc0f1-0e0a-4675-92a6-b0940cf1c699_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/QDFormer_[[:space:]]Towards[[:space:]]Robust[[:space:]]Audiovisual[[:space:]]Segmentation[[:space:]]in[[:space:]]Complex[[:space:]]Environments[[:space:]]with[[:space:]]Quantization-based[[:space:]]Semantic[[:space:]]Decomposition/b1ab0342-ace7-42ae-a5f1-8bceae24a0a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/QN-Mixer_[[:space:]]A[[:space:]]Quasi-Newton[[:space:]]MLP-Mixer[[:space:]]Model[[:space:]]for[[:space:]]Sparse-View[[:space:]]CT[[:space:]]Reconstruction/1c036bf8-1241-4401-8e00-c0473f9b27ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/QUADify_[[:space:]]Extracting[[:space:]]Meshes[[:space:]]with[[:space:]]Pixel-level[[:space:]]Details[[:space:]]and[[:space:]]Materials[[:space:]]from[[:space:]]Images/b5e9a83d-0011-440e-aef7-e8590409dbe4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Quantifying[[:space:]]Task[[:space:]]Priority[[:space:]]for[[:space:]]Multi-Task[[:space:]]Optimization/aceac2e6-abfc-4dbd-aac2-900921c47abc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Quantifying[[:space:]]Uncertainty[[:space:]]in[[:space:]]Motion[[:space:]]Prediction[[:space:]]with[[:space:]]Variational[[:space:]]Bayesian[[:space:]]Mixture/364580b0-4ae2-408e-bf48-5cedf11b380c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Querying[[:space:]]as[[:space:]]Prompt_[[:space:]]Parameter-Efficient[[:space:]]Learning[[:space:]]for[[:space:]]Multimodal[[:space:]]Language[[:space:]]Model/fd4eae26-5342-4ee1-a578-81b0ea4e1393_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Question[[:space:]]Aware[[:space:]]Vision[[:space:]]Transformer[[:space:]]for[[:space:]]Multimodal[[:space:]]Reasoning/77789a54-5089-434f-a5f4-e5c5938d0892_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Quilt-LLaVA_[[:space:]]Visual[[:space:]]Instruction[[:space:]]Tuning[[:space:]]by[[:space:]]Extracting[[:space:]]Localized[[:space:]]Narratives[[:space:]]from[[:space:]]Open-Source[[:space:]]Histopathology[[:space:]]Videos/21ae6c6f-2aa8-4979-a7bb-6a70e6f9e1ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/R-Cyclic[[:space:]]Diffuser_[[:space:]]Reductive[[:space:]]and[[:space:]]Cyclic[[:space:]]Latent[[:space:]]Diffusion[[:space:]]for[[:space:]]3D[[:space:]]Clothed[[:space:]]Human[[:space:]]Digitalization/74e5314a-5049-43ed-a297-8976a79b7ad0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RAM-Avatar_[[:space:]]Real-time[[:space:]]Photo-Realistic[[:space:]]Avatar[[:space:]]from[[:space:]]Monocular[[:space:]]Videos[[:space:]]with[[:space:]]Full-body[[:space:]]Control/821d07e0-da6b-42fa-a1be-a488c114c30f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RAVE_[[:space:]]Randomized[[:space:]]Noise[[:space:]]Shuffling[[:space:]]for[[:space:]]Fast[[:space:]]and[[:space:]]Consistent[[:space:]]Video[[:space:]]Editing[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/e5b65600-12fc-45eb-91b2-4031c25d288f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RCBEVDet_[[:space:]]Radar-camera[[:space:]]Fusion[[:space:]]in[[:space:]]Bird's[[:space:]]Eye[[:space:]]View[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/b0d90695-b1e4-4bf3-a8ba-60b5f270854f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RCL_[[:space:]]Reliable[[:space:]]Continual[[:space:]]Learning[[:space:]]for[[:space:]]Unified[[:space:]]Failure[[:space:]]Detection/7afef597-d579-4575-b4b4-c95833343b33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RCooper_[[:space:]]A[[:space:]]Real-world[[:space:]]Large-scale[[:space:]]Dataset[[:space:]]for[[:space:]]Roadside[[:space:]]Cooperative[[:space:]]Perception/add22b40-fcea-4fb3-bfc0-b241418da181_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/REACTO_[[:space:]]Reconstructing[[:space:]]Articulated[[:space:]]Objects[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Video/840f88e7-9cf4-4e76-87b3-f8f7ea848534_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/READ_[[:space:]]Retrieval-Enhanced[[:space:]]Asymmetric[[:space:]]Diffusion[[:space:]]for[[:space:]]Motion[[:space:]]Planning/cc4b795e-0ec0-400d-b7ec-6d86bc3b8348_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RELI11D_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Multimodal[[:space:]]Human[[:space:]]Motion[[:space:]]Dataset[[:space:]]and[[:space:]]Method/8b926fe0-3373-451a-9fee-4a6df129e21c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RGBD[[:space:]]Objects[[:space:]]in[[:space:]]the[[:space:]]Wild_[[:space:]]Scaling[[:space:]]Real-World[[:space:]]3D[[:space:]]Object[[:space:]]Learning[[:space:]]from[[:space:]]RGB-D[[:space:]]Videos/4e9d1e22-4d9c-4914-bd27-b1b0adc8a30b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RILA_[[:space:]]Reflective[[:space:]]and[[:space:]]Imaginative[[:space:]]Language[[:space:]]Agent[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Semantic[[:space:]]Audio-Visual[[:space:]]Navigation/3117d8c5-e5ee-488f-8781-fc5ca5bd4660_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RLHF-V_[[:space:]]Towards[[:space:]]Trustworthy[[:space:]]MLLMs[[:space:]]via[[:space:]]Behavior[[:space:]]Alignment[[:space:]]from[[:space:]]Fine-grained[[:space:]]Correctional[[:space:]]Human[[:space:]]Feedback/329e83ec-7b96-4e54-9403-e94bf6e14f15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RMT_[[:space:]]Retentive[[:space:]]Networks[[:space:]]Meet[[:space:]]Vision[[:space:]]Transformers/89757234-5976-4f4b-82d4-6f3a9e9bf667_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RMem_[[:space:]]Restricted[[:space:]]Memory[[:space:]]Banks[[:space:]]Improve[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/29d13535-e71a-4c51-94db-85d81bcf7a71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RNb-NeuS_[[:space:]]Reflectance[[:space:]]and[[:space:]]Normal-based[[:space:]]Multi-View[[:space:]]3D[[:space:]]Reconstruction/2cf6acdc-4a58-4ab9-b89a-2e798074e862_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RTMO_[[:space:]]Towards[[:space:]]High-Performance[[:space:]]One-Stage[[:space:]]Real-Time[[:space:]]Multi-Person[[:space:]]Pose[[:space:]]Estimation/eabfd822-e58a-48a5-bba5-669edcdb54b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RTracker_[[:space:]]Recoverable[[:space:]]Tracking[[:space:]]via[[:space:]]PN[[:space:]]Tree[[:space:]]Structured[[:space:]]Memory/5c767094-3f39-4673-ae65-29849b40ebfa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RadSimReal_[[:space:]]Bridging[[:space:]]the[[:space:]]Gap[[:space:]]Between[[:space:]]Synthetic[[:space:]]and[[:space:]]Real[[:space:]]Data[[:space:]]in[[:space:]]Radar[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]Simulation/3210e22c-5c1a-47ef-81b0-a39e3392d222_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RadarDistill_[[:space:]]Boosting[[:space:]]Radar-based[[:space:]]Object[[:space:]]Detection[[:space:]]Performance[[:space:]]via[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]from[[:space:]]LiDAR[[:space:]]Features/b4487c4a-2016-4228-8bfa-a49a52dd1c84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Random[[:space:]]Entangled[[:space:]]Tokens[[:space:]]for[[:space:]]Adversarially[[:space:]]Robust[[:space:]]Vision[[:space:]]Transformer/33cd916c-5131-4e62-ae74-d4e71de93393_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RankED_[[:space:]]Addressing[[:space:]]Imbalance[[:space:]]and[[:space:]]Uncertainty[[:space:]]in[[:space:]]Edge[[:space:]]Detection[[:space:]]Using[[:space:]]Ranking-based[[:space:]]Losses/4a3b671d-9296-4842-b55a-b4650a2e2928_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RankMatch_[[:space:]]Exploring[[:space:]]the[[:space:]]Better[[:space:]]Consistency[[:space:]]Regularization[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Semantic[[:space:]]Segmentation/92273068-2ca5-4a30-9d06-c916211ded1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Ranking[[:space:]]Distillation[[:space:]]for[[:space:]]Open-Ended[[:space:]]Video[[:space:]]Question[[:space:]]Answering[[:space:]]with[[:space:]]Insufficient[[:space:]]Labels/b1ed6a65-5f1b-4ff7-a864-51a190cb9b01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Ranni_[[:space:]]Taming[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]for[[:space:]]Accurate[[:space:]]Instruction[[:space:]]Following/ab8d83cf-be70-4893-ad49-7fcf67a93ef2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rapid[[:space:]]3D[[:space:]]Model[[:space:]]Generation[[:space:]]with[[:space:]]Intuitive[[:space:]]3D[[:space:]]Input/6d005eab-2572-4044-abd0-9d9548ad525e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rapid[[:space:]]Motor[[:space:]]Adaptation[[:space:]]for[[:space:]]Robotic[[:space:]]Manipulator[[:space:]]Arms/c44df55a-9d1b-4a35-b8e6-b6cbdb4f801a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Re-thinking[[:space:]]Data[[:space:]]Availability[[:space:]]Attacks[[:space:]]Against[[:space:]]Deep[[:space:]]Neural[[:space:]]Networks/5493e31e-4975-4ec4-8f14-ab21ce80dbca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ReCoRe_[[:space:]]Regularized[[:space:]]Contrastive[[:space:]]Representation[[:space:]]Learning[[:space:]]of[[:space:]]World[[:space:]]Model/d46bbfa8-f984-4ab6-9954-90a9f571dfaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ReGenNet_[[:space:]]Towards[[:space:]]Human[[:space:]]Action-Reaction[[:space:]]Synthesis/60a2f957-f2c1-4752-ae9d-4349d326a957_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Readout[[:space:]]Guidance_[[:space:]]Learning[[:space:]]Control[[:space:]]from[[:space:]]Diffusion[[:space:]]Features/b0467899-34c8-4470-8b85-2581c0bcdb96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Real[[:space:]]Acoustic[[:space:]]Fields_[[:space:]]An[[:space:]]Audio-Visual[[:space:]]Room[[:space:]]Acoustics[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark/ef814f50-8024-4492-8349-f5e231e34c1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Real-IAD_[[:space:]]A[[:space:]]Real-World[[:space:]]Multi-View[[:space:]]Dataset[[:space:]]for[[:space:]]Benchmarking[[:space:]]Versatile[[:space:]]Industrial[[:space:]]Anomaly[[:space:]]Detection/03cb6cbd-600f-49ea-b7f2-d2236d8a7b6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Real-Time[[:space:]]Exposure[[:space:]]Correction[[:space:]]via[[:space:]]Collaborative[[:space:]]Transformations[[:space:]]and[[:space:]]Adaptive[[:space:]]Sampling/c99c88e6-ac64-4d92-9afb-f88eacc4b006_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Real-Time[[:space:]]Neural[[:space:]]BRDF[[:space:]]with[[:space:]]Spherically[[:space:]]Distributed[[:space:]]Primitives/c15fb99c-e73c-4418-9815-bbe4c4cdd2b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Real-Time[[:space:]]Simulated[[:space:]]Avatar[[:space:]]from[[:space:]]Head-Mounted[[:space:]]Sensors/20930e82-e883-44a9-8eb6-5ec227941587_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Real-World[[:space:]]Efficient[[:space:]]Blind[[:space:]]Motion[[:space:]]Deblurring[[:space:]]via[[:space:]]Blur[[:space:]]Pixel[[:space:]]Discretization/4f66cde2-5d54-41b8-8609-667663466938_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Real-World[[:space:]]Mobile[[:space:]]Image[[:space:]]Denoising[[:space:]]Dataset[[:space:]]with[[:space:]]Efficient[[:space:]]Baselines/8f0ad7f8-47c2-4965-b669-a64bfbc0ce54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Real-time[[:space:]]3D-aware[[:space:]]Portrait[[:space:]]Video[[:space:]]Relighting/068d7d2c-69dc-42f1-b0ee-ee56c2b6ea5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Real-time[[:space:]]Acquisition[[:space:]]and[[:space:]]Reconstruction[[:space:]]of[[:space:]]Dynamic[[:space:]]Volumes[[:space:]]with[[:space:]]Neural[[:space:]]Structured[[:space:]]Illumination/3260c081-ea49-4d4b-98da-987202d3edf0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RealCustom_[[:space:]]Narrowing[[:space:]]Real[[:space:]]Text[[:space:]]Word[[:space:]]for[[:space:]]Real-Time[[:space:]]Open-Domain[[:space:]]Text-to-Image[[:space:]]Customization/28c71233-004f-4d93-8367-2b862c8edb88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RealNet_[[:space:]]A[[:space:]]Feature[[:space:]]Selection[[:space:]]Network[[:space:]]with[[:space:]]Realistic[[:space:]]Synthetic[[:space:]]Anomaly[[:space:]]for[[:space:]]Anomaly[[:space:]]Detection/ffd09980-57de-422c-8784-05dc915e2793_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Realigning[[:space:]]Confidence[[:space:]]with[[:space:]]Temporal[[:space:]]Saliency[[:space:]]Information[[:space:]]for[[:space:]]Point-Level[[:space:]]Weakly-Supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization/cd72e806-5086-45a8-8e64-f6a18fc463b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RecDiffusion_[[:space:]]Rectangling[[:space:]]for[[:space:]]Image[[:space:]]Stitching[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/500b454f-f62f-4039-adeb-b16948258c51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ReconFusion_[[:space:]]3D[[:space:]]Reconstruction[[:space:]]with[[:space:]]Diffusion[[:space:]]Priors/396bb83b-5d95-4b76-be7f-2752548a5fa6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Reconstructing[[:space:]]Hands[[:space:]]in[[:space:]]3D[[:space:]]with[[:space:]]Transformers/48cc93b2-3a5e-4dcf-ba8b-38e3f0d1a8d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Reconstruction-free[[:space:]]Cascaded[[:space:]]Adaptive[[:space:]]Compressive[[:space:]]Sensing/f7368531-9c5f-40eb-88dc-a74fa64ae3d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Referring[[:space:]]Expression[[:space:]]Counting/e5d6d175-9fc7-429b-8357-2e9f7d02260a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Referring[[:space:]]Image[[:space:]]Editing_[[:space:]]Object-level[[:space:]]Image[[:space:]]Editing[[:space:]]via[[:space:]]Referring[[:space:]]Expressions/2dca33a1-b644-4171-b2f1-b57f5e8c34bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Reg-PTQ_[[:space:]]Regression-specialized[[:space:]]Post-training[[:space:]]Quantization[[:space:]]for[[:space:]]Fully[[:space:]]Quantized[[:space:]]Object[[:space:]]Detector/5401727b-9b8c-4edb-9256-f52cc6e3460c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Region-Based[[:space:]]Representations[[:space:]]Revisited/08e42b47-b6aa-4753-b5b0-1d8f42c2fef1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RegionGPT_[[:space:]]Towards[[:space:]]Region[[:space:]]Understanding[[:space:]]Vision[[:space:]]Language[[:space:]]Model/9199a652-98cd-44cb-ab46-27cfc8e511d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RegionPLC_[[:space:]]Regional[[:space:]]Point-Language[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Open-World[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/c483a524-1a5a-471b-ac89-f15b03f29d82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Regressor-Segmenter[[:space:]]Mutual[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Crowd[[:space:]]Counting/204f8b6f-d256-43f0-b193-9a0dc6919559_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Regularized[[:space:]]Parameter[[:space:]]Uncertainty[[:space:]]for[[:space:]]Improving[[:space:]]Generalization[[:space:]]in[[:space:]]Reinforcement[[:space:]]Learning/4127fb97-5349-4e1c-a9bd-0a2fa62cf282_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Relation[[:space:]]Rectification[[:space:]]in[[:space:]]Diffusion[[:space:]]Model/5a0425dd-f260-45de-81e5-9d40edfc3465_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Relational[[:space:]]Matching[[:space:]]for[[:space:]]Weakly[[:space:]]Semi-Supervised[[:space:]]Oriented[[:space:]]Object[[:space:]]Detection/065accfa-24d2-4191-aa40-9d78f4a81e0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Relaxed[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Federated[[:space:]]Learning/c98e8bcc-7d5a-42d2-a87f-4ddfc6b37bbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Relightable[[:space:]]Gaussian[[:space:]]Codec[[:space:]]Avatars/d800aaf6-cbda-4c82-8aff-4a363614b07a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Relightable[[:space:]]and[[:space:]]Animatable[[:space:]]Neural[[:space:]]Avatar[[:space:]]from[[:space:]]Sparse-View[[:space:]]Video/9adfd7d2-2c84-4dba-96ee-12195a7d5e9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Relightful[[:space:]]Harmonization_[[:space:]]Lighting-aware[[:space:]]Portrait[[:space:]]Background[[:space:]]Replacement/de5b6700-115b-4583-a772-75725001ead5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RepAn_[[:space:]]Enhanced[[:space:]]Annealing[[:space:]]through[[:space:]]Re-parameterization/d2b4ece8-07d6-4858-94fb-ac3060a682ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RepKPU_[[:space:]]Point[[:space:]]Cloud[[:space:]]Upsampling[[:space:]]with[[:space:]]Kernel[[:space:]]Point[[:space:]]Representation[[:space:]]and[[:space:]]Deformation/4ae41549-75ac-4940-83c4-190ef51ed969_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RepViT_[[:space:]]Revisiting[[:space:]]Mobile[[:space:]]CNN[[:space:]]From[[:space:]]ViT[[:space:]]Perspective/8010525d-bbb0-4842-a3db-2bfd5079167d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Representing[[:space:]]Part-Whole[[:space:]]Hierarchies[[:space:]]in[[:space:]]Foundation[[:space:]]Models[[:space:]]by[[:space:]]Learning[[:space:]]Localizability[[:space:]]Composability[[:space:]]and[[:space:]]Decomposability[[:space:]]from[[:space:]]Anatomy[[:space:]]via[[:space:]]Self[[:space:]]Supervision/a46b20c0-50b9-4767-8ed1-5f5e9b381d3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Repurposing[[:space:]]Diffusion-Based[[:space:]]Image[[:space:]]Generators[[:space:]]for[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/48e61a8d-5974-4d40-809d-210d92e39ed5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Residual[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Models/99af0b28-01e1-4010-848e-f1882b845e48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Residual[[:space:]]Learning[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/d894c1e9-60ef-4c42-93cb-32cc9efbd330_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Resolution[[:space:]]Limit[[:space:]]of[[:space:]]Single-Photon[[:space:]]LiDAR/5ee650a4-22b0-44d5-9bd3-44a2c9eddd68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Resource-Efficient[[:space:]]Transformer[[:space:]]Pruning[[:space:]]for[[:space:]]Finetuning[[:space:]]of[[:space:]]Large[[:space:]]Models/a4188dee-aeac-4329-a110-46e6e87d64e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Restoration[[:space:]]by[[:space:]]Generation[[:space:]]with[[:space:]]Constrained[[:space:]]Priors/ea301077-8eb0-49e1-bce1-39e3d05125cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Resurrecting[[:space:]]Old[[:space:]]Classes[[:space:]]with[[:space:]]New[[:space:]]Data[[:space:]]for[[:space:]]Exemplar-Free[[:space:]]Continual[[:space:]]Learning/16b6da51-9fc8-417f-9c61-49ebb34815bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]Boundary[[:space:]]Discontinuity[[:space:]]Problem[[:space:]]for[[:space:]]Oriented[[:space:]]Object[[:space:]]Detection/c0684b48-be21-4f4c-975e-f70598d127c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Multi-Contrast[[:space:]]MRI[[:space:]]Super-Resolution/a53639a1-54de-4528-81dd-835fd09b282f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]FID_[[:space:]]Towards[[:space:]]a[[:space:]]Better[[:space:]]Evaluation[[:space:]]Metric[[:space:]]for[[:space:]]Image[[:space:]]Generation/c494a202-6835-4602-bc33-aed76ba58e9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]Few-shot[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Semantic[[:space:]]Segmentation/291bf87a-3ccd-4ef0-b671-6a187d57f090_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]Generalizable[[:space:]]Face[[:space:]]Anti-spoofing[[:space:]]via[[:space:]]Hierarchical[[:space:]]Prototype-guided[[:space:]]Distribution[[:space:]]Refinement[[:space:]]in[[:space:]]Hyperbolic[[:space:]]Space/7ef0f8d7-3ea3-4a31-bd4e-b70428c73790_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]Human[[:space:]]Motion[[:space:]]Prediction[[:space:]]with[[:space:]]Symplectic[[:space:]]Integral/ed265d55-1a7a-4f4c-9777-3a672c03104d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]Inductive[[:space:]]Biases[[:space:]]for[[:space:]]Surface[[:space:]]Normal[[:space:]]Estimation/5c730124-b70d-4a85-bc33-fc6e626e6112_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]Interactive[[:space:]]Image[[:space:]]Segmentation[[:space:]]with[[:space:]]Low[[:space:]]Latency[[:space:]]High[[:space:]]Quality[[:space:]]and[[:space:]]Diverse[[:space:]]Prompts/fdb92af5-e048-4a2b-b10e-acb79f6405b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]Multi-domain[[:space:]]Generalization[[:space:]]with[[:space:]]A[[:space:]]General[[:space:]]Learning[[:space:]]Objective/76a8407a-bfcf-4e93-88ce-78f36a58b847_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]Multi-view[[:space:]]Representation[[:space:]]Learning[[:space:]]via[[:space:]]Distilled[[:space:]]Disentangling/8ffcd189-d131-4b7a-9b30-57dec04b1ffc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]Prior[[:space:]]Information[[:space:]]Generation[[:space:]]with[[:space:]]CLIP[[:space:]]for[[:space:]]Few-Shot[[:space:]]Segmentation/2da751e7-7be1-4c31-8d47-5e3d8213029f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]Transformers[[:space:]]Pre-training[[:space:]]for[[:space:]]Multi-Spectral[[:space:]]Satellite[[:space:]]Imagery/f9c027d5-4347-4886-9133-44a1b7373495_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]the[[:space:]]Evaluation[[:space:]]Protocol[[:space:]]of[[:space:]]Domain[[:space:]]Generalization/4d90fd9e-5986-457e-a50b-098e28f6bf6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]the[[:space:]]Objectives[[:space:]]of[[:space:]]Vector-Quantized[[:space:]]Tokenizers[[:space:]]for[[:space:]]Image[[:space:]]Synthesis/c8080dad-443a-4b97-ad22-74c7fcda5d05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]the[[:space:]]Representation[[:space:]]in[[:space:]]Federated[[:space:]]Unsupervised[[:space:]]Learning[[:space:]]with[[:space:]]Non-IID[[:space:]]Data/e7146ead-968c-4dbc-9e16-a80c8c76cd36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]the[[:space:]]Spatial[[:space:]]Inconsistency[[:space:]]in[[:space:]]Classifier-Free[[:space:]]Diffusion[[:space:]]Guidance/58bc302f-25b9-4bc3-9680-638d616a772b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rethinking[[:space:]]the[[:space:]]Up-Sampling[[:space:]]Operations[[:space:]]in[[:space:]]CNN-based[[:space:]]Generative[[:space:]]Network[[:space:]]for[[:space:]]Generalizable[[:space:]]Deepfake[[:space:]]Detection/b794cfee-9b1e-4b1b-8bde-3e4b7fffe48c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Retraining-Free[[:space:]]Model[[:space:]]Quantization[[:space:]]via[[:space:]]One-Shot[[:space:]]Weight-Coupling[[:space:]]Learning/7267ee35-7fd8-476d-87d1-23f858f78ab9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Retrieval-Augmented[[:space:]]Egocentric[[:space:]]Video[[:space:]]Captioning/a246c259-a004-4692-9597-63780ed6d8a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Retrieval-Augmented[[:space:]]Embodied[[:space:]]Agents/a75d445a-1b4b-4d51-b2c0-ac2e0071050c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Retrieval-Augmented[[:space:]]Layout[[:space:]]Transformer[[:space:]]for[[:space:]]Content-Aware[[:space:]]Layout[[:space:]]Generation/c9b07a93-91b3-4717-96b1-e3efcc9a8b69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Retrieval-Augmented[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detection/58e24018-fd9c-4933-a69a-ece134887db8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Revamping[[:space:]]Federated[[:space:]]Learning[[:space:]]Security[[:space:]]from[[:space:]]a[[:space:]]Defender's[[:space:]]Perspective_[[:space:]]A[[:space:]]Unified[[:space:]]Defense[[:space:]]with[[:space:]]Homomorphic[[:space:]]Encrypted[[:space:]]Data[[:space:]]Space/3fd95dd4-4d9d-4717-8b0e-0845ff0b8f0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Revisiting[[:space:]]Adversarial[[:space:]]Training[[:space:]]Under[[:space:]]Long-Tailed[[:space:]]Distributions/8eb5302b-e002-464b-a735-9c93d797001f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Revisiting[[:space:]]Adversarial[[:space:]]Training[[:space:]]at[[:space:]]Scale/9b6affb2-0d04-46b7-a2a8-fdd833c435ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Revisiting[[:space:]]Counterfactual[[:space:]]Problems[[:space:]]in[[:space:]]Referring[[:space:]]Expression[[:space:]]Comprehension/0484480e-6a54-4e2e-8d12-2e02965e1fac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Revisiting[[:space:]]Global[[:space:]]Translation[[:space:]]Estimation[[:space:]]with[[:space:]]Feature[[:space:]]Tracks/30c88065-b0c2-45e0-a62a-eeab99e95801_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Revisiting[[:space:]]Non-Autoregressive[[:space:]]Transformers[[:space:]]for[[:space:]]Efficient[[:space:]]Image[[:space:]]Synthesis/d59fe238-c538-46fd-bef9-dce39ad7dc8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Revisiting[[:space:]]Sampson[[:space:]]Approximations[[:space:]]for[[:space:]]Geometric[[:space:]]Estimation[[:space:]]Problems/e10efca1-7dc7-4f22-8096-38c2667b5462_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Revisiting[[:space:]]Single[[:space:]]Image[[:space:]]Reflection[[:space:]]Removal[[:space:]]In[[:space:]]the[[:space:]]Wild/f79e9ba5-c8d4-4265-89e8-5451dc9f1e93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Revisiting[[:space:]]Spatial-Frequency[[:space:]]Information[[:space:]]Integration[[:space:]]from[[:space:]]a[[:space:]]Hierarchical[[:space:]]Perspective[[:space:]]for[[:space:]]Panchromatic[[:space:]]and[[:space:]]Multi-Spectral[[:space:]]Image[[:space:]]Fusion/17020f64-6d1d-4d13-b4be-02b071d59a0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Revisiting[[:space:]]the[[:space:]]Domain[[:space:]]Shift[[:space:]]and[[:space:]]Sample[[:space:]]Uncertainty[[:space:]]in[[:space:]]Multi-source[[:space:]]Active[[:space:]]Domain[[:space:]]Transfer/c6e7a8a8-b5e8-4613-a70b-e0acca787e9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rewrite[[:space:]]the[[:space:]]Stars/0894008a-85fe-4608-b5fa-f10ba4aaaa3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rich[[:space:]]Human[[:space:]]Feedback[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/3aa2ac1a-2b55-464b-95c3-e44603ea222b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RichDreamer_[[:space:]]A[[:space:]]Generalizable[[:space:]]Normal-Depth[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Detail[[:space:]]Richness[[:space:]]in[[:space:]]Text-to-3D/cf634621-9d90-4950-b541-236f726868ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Riemannian[[:space:]]Multinomial[[:space:]]Logistics[[:space:]]Regression[[:space:]]for[[:space:]]SPD[[:space:]]Neural[[:space:]]Networks/91b38fd9-c907-404d-a450-6ffb8bb5dcdc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RoDLA_[[:space:]]Benchmarking[[:space:]]the[[:space:]]Robustness[[:space:]]of[[:space:]]Document[[:space:]]Layout[[:space:]]Analysis[[:space:]]Models/7467670a-9757-4ce6-9177-a7c78aefc079_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RoHM_[[:space:]]Robust[[:space:]]Human[[:space:]]Motion[[:space:]]Reconstruction[[:space:]]via[[:space:]]Diffusion/97f54a8b-fea2-46db-abf1-12e546b1cdcb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RoMa_[[:space:]]Robust[[:space:]]Dense[[:space:]]Feature[[:space:]]Matching/68b63694-3fc0-4e0a-8567-e2d1e983291f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Robust[[:space:]]Depth[[:space:]]Enhancement[[:space:]]via[[:space:]]Polarization[[:space:]]Prompt[[:space:]]Fusion[[:space:]]Tuning/a2dbf035-771e-40e0-be7b-3b46830e8a3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Robust[[:space:]]Distillation[[:space:]]via[[:space:]]Untargeted[[:space:]]and[[:space:]]Targeted[[:space:]]Intermediate[[:space:]]Adversarial[[:space:]]Samples/42990044-1b07-404e-aac4-1da8bccf6b49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Robust[[:space:]]Emotion[[:space:]]Recognition[[:space:]]in[[:space:]]Context[[:space:]]Debiasing/1484eec4-708e-4afa-9e70-ea0fce236562_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Robust[[:space:]]Image[[:space:]]Denoising[[:space:]]through[[:space:]]Adversarial[[:space:]]Frequency[[:space:]]Mixup/9f200b90-0327-4d15-b036-58858e33b461_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Robust[[:space:]]Noisy[[:space:]]Correspondence[[:space:]]Learning[[:space:]]with[[:space:]]Equivariant[[:space:]]Similarity[[:space:]]Consistency/655d0444-a71a-4530-a051-140058d72a93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Robust[[:space:]]Overfitting[[:space:]]Does[[:space:]]Matter_[[:space:]]Test-Time[[:space:]]Adversarial[[:space:]]Purification[[:space:]]With[[:space:]]FGSM/7b217dc3-b2d1-46d9-b36f-5aca6479c024_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Robust[[:space:]]Self-calibration[[:space:]]of[[:space:]]Focal[[:space:]]Lengths[[:space:]]from[[:space:]]the[[:space:]]Fundamental[[:space:]]Matrix/803d7a66-f53b-428e-b191-3725e80860be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Robust[[:space:]]Synthetic-to-Real[[:space:]]Transfer[[:space:]]for[[:space:]]Stereo[[:space:]]Matching/290c2f11-6aff-474b-bd94-96c9e73e5ac8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/RobustSAM_[[:space:]]Segment[[:space:]]Anything[[:space:]]Robustly[[:space:]]on[[:space:]]Degraded[[:space:]]Images/8979b06b-35b5-4fb8-9c6f-7a0c82c57ad1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rolling[[:space:]]Shutter[[:space:]]Correction[[:space:]]with[[:space:]]Intermediate[[:space:]]Distortion[[:space:]]Flow[[:space:]]Estimation/e2d1615c-fc51-4c27-8eeb-53c1717e8316_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rotated[[:space:]]Multi-Scale[[:space:]]Interaction[[:space:]]Network[[:space:]]for[[:space:]]Referring[[:space:]]Remote[[:space:]]Sensing[[:space:]]Image[[:space:]]Segmentation/8c412c10-325e-4762-9d3c-d78c0b31efc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Rotation-Agnostic[[:space:]]Image[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Digital[[:space:]]Pathology/fc50c2f5-e245-4869-8d85-f2f6410fd414_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/S-DyRF_[[:space:]]Reference-Based[[:space:]]Stylized[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Dynamic[[:space:]]Scenes/6e24627c-f75e-48b2-b52b-0411553098da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/S2MAE_[[:space:]]A[[:space:]]Spatial-Spectral[[:space:]]Pretraining[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Spectral[[:space:]]Remote[[:space:]]Sensing[[:space:]]Data/cf365394-5e31-4759-b679-e83db8226d46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/S2MVTC_[[:space:]]a[[:space:]]Simple[[:space:]]yet[[:space:]]Efficient[[:space:]]Scalable[[:space:]]Multi-View[[:space:]]Tensor[[:space:]]Clustering/779590c1-9c31-432b-afd3-e70598636916_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SAFDNet_[[:space:]]A[[:space:]]Simple[[:space:]]and[[:space:]]Effective[[:space:]]Network[[:space:]]for[[:space:]]Fully[[:space:]]Sparse[[:space:]]3D[[:space:]]Object[[:space:]]Detection/4f5506fe-8736-4e2c-b27f-2e2c1ccc3b19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SAI3D_[[:space:]]Segment[[:space:]]Any[[:space:]]Instance[[:space:]]in[[:space:]]3D[[:space:]]Scenes/25612317-b346-4529-95ab-1f26a422e27e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SAM-6D_[[:space:]]Segment[[:space:]]Anything[[:space:]]Model[[:space:]]Meets[[:space:]]Zero-Shot[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/05f7f0c0-59f4-4c18-8eb6-2da2ef48d5fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SANeRF-HQ_[[:space:]]Segment[[:space:]]Anything[[:space:]]for[[:space:]]NeRF[[:space:]]in[[:space:]]High[[:space:]]Quality/dea6233c-97b4-4961-bd88-0c8c1799293f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SAOR_[[:space:]]Single-View[[:space:]]Articulated[[:space:]]Object[[:space:]]Reconstruction/b14282e5-780e-4bca-9979-425bbc9929b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SC-GS_[[:space:]]Sparse-Controlled[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Editable[[:space:]]Dynamic[[:space:]]Scenes/99e6f029-7abf-43a2-990b-dad28141df08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SC-Tune_[[:space:]]Unleashing[[:space:]]Self-Consistent[[:space:]]Referential[[:space:]]Comprehension[[:space:]]in[[:space:]]Large[[:space:]]Vision[[:space:]]Language[[:space:]]Models/2443f212-9af3-4746-8063-6308f9c03b1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SCE-MAE_[[:space:]]Selective[[:space:]]Correspondence[[:space:]]Enhancement[[:space:]]with[[:space:]]Masked[[:space:]]Autoencoder[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Landmark[[:space:]]Estimation/815d46a6-c1dc-4409-8244-3f04afe68fa4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SCEdit_[[:space:]]Efficient[[:space:]]and[[:space:]]Controllable[[:space:]]Image[[:space:]]Diffusion[[:space:]]Generation[[:space:]]via[[:space:]]Skip[[:space:]]Connection[[:space:]]Editing/0cfdf32e-c05d-45a9-accb-54065498726d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SCINeRF_[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]from[[:space:]]a[[:space:]]Snapshot[[:space:]]Compressive[[:space:]]Image/41c276c3-cfdd-461d-a725-1f20998f5a8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SCULPT_[[:space:]]Shape-Conditioned[[:space:]]Unpaired[[:space:]]Learning[[:space:]]of[[:space:]]Pose-dependent[[:space:]]Clothed[[:space:]]and[[:space:]]Textured[[:space:]]Human[[:space:]]Meshes/2065c96a-da3d-49e5-9b12-5562c59c000b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SCoFT_[[:space:]]Self-Contrastive[[:space:]]Fine-Tuning[[:space:]]for[[:space:]]Equitable[[:space:]]Image[[:space:]]Generation/6385dd1b-7ca8-44f0-a961-d9576212ffc8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SD-DiT_[[:space:]]Unleashing[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]Self-supervised[[:space:]]Discrimination[[:space:]]in[[:space:]]Diffusion[[:space:]]Transformer/e7dbeaf0-2e2f-4678-8fc1-f056872deaea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SD2Event_Self-supervised[[:space:]]Learning[[:space:]]of[[:space:]]Dynamic[[:space:]]Detectors[[:space:]]and[[:space:]]Contextual[[:space:]]Descriptors[[:space:]]for[[:space:]]Event[[:space:]]Cameras/e0508eb7-cc55-4c47-8443-c9bafa9ccce4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SD4Match_[[:space:]]Learning[[:space:]]to[[:space:]]Prompt[[:space:]]Stable[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Semantic[[:space:]]Matching/51b1e84a-6eee-45df-9c05-5243877436f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SDDGR_[[:space:]]Stable[[:space:]]Diffusion-based[[:space:]]Deep[[:space:]]Generative[[:space:]]Replay[[:space:]]for[[:space:]]Class[[:space:]]Incremental[[:space:]]Object[[:space:]]Detection/357078b1-0a34-43c9-99b4-753ae63603e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SDPose_[[:space:]]Tokenized[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Circulation-Guide[[:space:]]Self-Distillation/79398682-7f3d-4508-a597-00d1ee00ed2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SDSTrack_[[:space:]]Self-Distillation[[:space:]]Symmetric[[:space:]]Adapter[[:space:]]Learning[[:space:]]for[[:space:]]Multi-Modal[[:space:]]Visual[[:space:]]Object[[:space:]]Tracking/c8d0a0ad-b475-4fb8-8d9b-9ae4a8202f51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SEAS_[[:space:]]ShapE-Aligned[[:space:]]Supervision[[:space:]]for[[:space:]]Person[[:space:]]Re-Identification/44da8eb9-693a-4deb-953a-cd500207fbac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SED_[[:space:]]A[[:space:]]Simple[[:space:]]Encoder-Decoder[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/01a5742e-3ad0-4c78-b82b-e36354ccdf93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SEED-Bench_[[:space:]]Benchmarking[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/b5664180-61f3-44a2-8332-fdf377b671d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SFOD_[[:space:]]Spiking[[:space:]]Fusion[[:space:]]Object[[:space:]]Detector/9a8d7da7-4281-47f4-a61c-0dff0781b419_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SG-BEV_[[:space:]]Satellite-Guided[[:space:]]BEV[[:space:]]Fusion[[:space:]]for[[:space:]]Cross-View[[:space:]]Semantic[[:space:]]Segmentation/3c76a53b-db9e-4b33-bbca-2e4f311ee7fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SG-PGM_[[:space:]]Partial[[:space:]]Graph[[:space:]]Matching[[:space:]]Network[[:space:]]with[[:space:]]Semantic[[:space:]]Geometric[[:space:]]Fusion[[:space:]]for[[:space:]]3D[[:space:]]Scene[[:space:]]Graph[[:space:]]Alignment[[:space:]]and[[:space:]]Its[[:space:]]Downstream[[:space:]]Tasks/e31bccfb-817f-44d3-b01c-cee40797f056_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SHAP-EDITOR_[[:space:]]Instruction-Guided[[:space:]]Latent[[:space:]]3D[[:space:]]Editing[[:space:]]in[[:space:]]Seconds/aa3a85e1-774e-470a-b09d-48f838a4060a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SHINOBI_[[:space:]]Shape[[:space:]]and[[:space:]]Illumination[[:space:]]using[[:space:]]Neural[[:space:]]Object[[:space:]]Decomposition[[:space:]]via[[:space:]]BRDF[[:space:]]Optimization[[:space:]]In-the-wild/6d98d653-de77-4b59-ba8b-af63dc909440_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SHViT_[[:space:]]Single-Head[[:space:]]Vision[[:space:]]Transformer[[:space:]]with[[:space:]]Memory[[:space:]]Efficient[[:space:]]Macro[[:space:]]Design/5b884264-d739-451e-921c-d580b3c71c7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SHiNe_[[:space:]]Semantic[[:space:]]Hierarchy[[:space:]]Nexus[[:space:]]for[[:space:]]Open-vocabulary[[:space:]]Object[[:space:]]Detection/a0eba3e6-9f17-4533-81d0-e87e16086cf1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SI-MIL_[[:space:]]Taming[[:space:]]Deep[[:space:]]MIL[[:space:]]for[[:space:]]Self-Interpretability[[:space:]]in[[:space:]]Gigapixel[[:space:]]Histopathology/a36a559c-4d6e-451d-b37c-afc0f196cdf4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SIFU_[[:space:]]Side-view[[:space:]]Conditioned[[:space:]]Implicit[[:space:]]Function[[:space:]]for[[:space:]]Real-world[[:space:]]Usable[[:space:]]Clothed[[:space:]]Human[[:space:]]Reconstruction/1e26eda8-1c34-4a16-a92f-9fe0d1de7fe7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SIGNeRF_[[:space:]]Scene[[:space:]]Integrated[[:space:]]Generation[[:space:]]for[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/cfef7298-0269-4eb4-b529-fcdcfc58ff97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SIRA_[[:space:]]Scalable[[:space:]]Inter-frame[[:space:]]Relation[[:space:]]and[[:space:]]Association[[:space:]]for[[:space:]]Radar[[:space:]]Perception/98fd4904-05b7-4fab-b6e8-bbe97c09689a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SLICE_[[:space:]]Stabilized[[:space:]]LIME[[:space:]]for[[:space:]]Consistent[[:space:]]Explanations[[:space:]]for[[:space:]]Image[[:space:]]Classification/a284bb71-7b50-4d6d-a9c5-f4ff6a85125c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SNED_[[:space:]]Superposition[[:space:]]Network[[:space:]]Architecture[[:space:]]Search[[:space:]]for[[:space:]]Efficient[[:space:]]Video[[:space:]]Diffusion[[:space:]]Model/2c99e83b-2b75-4d09-b682-19e534aa4dd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SNI-SLAM_[[:space:]]Semantic[[:space:]]Neural[[:space:]]Implicit[[:space:]]SLAM/79a5588c-619b-4039-b591-9ab33301765a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SNIDA_[[:space:]]Unlocking[[:space:]]Few-Shot[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Non-linear[[:space:]]Semantic[[:space:]]Decoupling[[:space:]]Augmentation/dbb0f46f-05b0-41a0-9468-3a0b9cff97ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SNIFFER_[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Explainable[[:space:]]Out-of-Context[[:space:]]Misinformation[[:space:]]Detection/e4c26e7a-a633-43c1-91ad-73c195b651b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SOAC_[[:space:]]Spatio-Temporal[[:space:]]Overlap-Aware[[:space:]]Multi-Sensor[[:space:]]Calibration[[:space:]]using[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/e3dde221-a41b-4dd8-9936-e3249aae3fdb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SODA_[[:space:]]Bottleneck[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Representation[[:space:]]Learning/c1c04ec8-d9aa-4064-8e00-b73856e9a85b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SOK-Bench_[[:space:]]A[[:space:]]Situated[[:space:]]Video[[:space:]]Reasoning[[:space:]]Benchmark[[:space:]]with[[:space:]]Aligned[[:space:]]Open-World[[:space:]]Knowledge/aba38613-84ba-4811-b631-05c6301fbc4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SPAD_[[:space:]]Spatially[[:space:]]Aware[[:space:]]Multi-View[[:space:]]Diffusers/a68f1eb9-cfc4-496d-bd0d-243f5d5214f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SPECAT_[[:space:]]SPatial-spEctral[[:space:]]Cumulative-Attention[[:space:]]Transformer[[:space:]]for[[:space:]]High-Resolution[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Reconstruction/d5c518f8-4d95-4cec-8644-de6c82f28921_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SPIDeRS_[[:space:]]Structured[[:space:]]Polarization[[:space:]]for[[:space:]]Invisible[[:space:]]Depth[[:space:]]and[[:space:]]Reflectance[[:space:]]Sensing/29bb0aa4-cc25-4dec-9934-064587fb4be3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SPIN_[[:space:]]Simultaneous[[:space:]]Perception[[:space:]]Interaction[[:space:]]and[[:space:]]Navigation/4d7b547b-06a2-428b-adc1-106f86ca14bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SPOC_[[:space:]]Imitating[[:space:]]Shortest[[:space:]]Paths[[:space:]]in[[:space:]]Simulation[[:space:]]Enables[[:space:]]Effective[[:space:]]Navigation[[:space:]]and[[:space:]]Manipulation[[:space:]]in[[:space:]]the[[:space:]]Real[[:space:]]World/86eb408c-afba-42d0-9587-266d14d25c87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SPOT_[[:space:]]Self-Training[[:space:]]with[[:space:]]Patch-Order[[:space:]]Permutation[[:space:]]for[[:space:]]Object-Centric[[:space:]]Learning[[:space:]]with[[:space:]]Autoregressive[[:space:]]Transformers/d656ed20-33f2-4ef4-86f5-8ff90938a081_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SPU-PMD_[[:space:]]Self-Supervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Upsampling[[:space:]]via[[:space:]]Progressive[[:space:]]Mesh[[:space:]]Deformation/c59d7cbe-f05b-443f-80a9-f152321a400d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SRTube_[[:space:]]Video-Language[[:space:]]Pre-Training[[:space:]]with[[:space:]]Action-Centric[[:space:]]Video[[:space:]]Tube[[:space:]]Features[[:space:]]and[[:space:]]Semantic[[:space:]]Role[[:space:]]Labeling/f8f7bb02-cdaa-4020-802e-37f87318244d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SSR-Encoder_[[:space:]]Encoding[[:space:]]Selective[[:space:]]Subject[[:space:]]Representation[[:space:]]for[[:space:]]Subject-Driven[[:space:]]Generation/de08affe-322f-4f71-9793-c50e4af5935c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SUGAR_[[:space:]]Pre-training[[:space:]]3D[[:space:]]Visual[[:space:]]Representations[[:space:]]for[[:space:]]Robotics/c5d0ddd0-932c-42bf-9c2d-df61820f3548_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SURE_[[:space:]]SUrvey[[:space:]]REcipes[[:space:]]for[[:space:]]building[[:space:]]reliable[[:space:]]and[[:space:]]robust[[:space:]]deep[[:space:]]networks/5fb8784e-4c6c-4ab1-a82b-cee73fc2721a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SVDTree_[[:space:]]Semantic[[:space:]]Voxel[[:space:]]Diffusion[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Tree[[:space:]]Reconstruction/493ef5a5-6db7-4c6f-a277-758ddae89aaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SVDinsTN_[[:space:]]A[[:space:]]Tensor[[:space:]]Network[[:space:]]Paradigm[[:space:]]for[[:space:]]Efficient[[:space:]]Structure[[:space:]]Search[[:space:]]from[[:space:]]Regularized[[:space:]]Modeling[[:space:]]Perspective/9c4f1c4f-8be0-4a1f-a843-f10c4e470e45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SVGDreamer_[[:space:]]Text[[:space:]]Guided[[:space:]]SVG[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion[[:space:]]Model/e30bed44-3c2b-42b8-a80f-c6160b0fd7ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SaCo[[:space:]]Loss_[[:space:]]Sample-wise[[:space:]]Affinity[[:space:]]Consistency[[:space:]]for[[:space:]]Vision-Language[[:space:]]Pre-training/3a4460ab-4a1d-41d6-a136-48c0ad2f0080_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Salience[[:space:]]DETR_[[:space:]]Enhancing[[:space:]]Detection[[:space:]]Transformer[[:space:]]with[[:space:]]Hierarchical[[:space:]]Salience[[:space:]]Filtering[[:space:]]Refinement/47405005-e657-4fb4-94f4-34acf74c10c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Sat2Scene_[[:space:]]3D[[:space:]]Urban[[:space:]]Scene[[:space:]]Generation[[:space:]]from[[:space:]]Satellite[[:space:]]Images[[:space:]]with[[:space:]]Diffusion/7cfe908e-b9d8-4851-902a-033f108ae00a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SatSynth_[[:space:]]Augmenting[[:space:]]Image-Mask[[:space:]]Pairs[[:space:]]through[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Aerial[[:space:]]Semantic[[:space:]]Segmentation/a70a7595-e428-4e5f-9cf0-3f2ed797a7f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Scaffold-GS_[[:space:]]Structured[[:space:]]3D[[:space:]]Gaussians[[:space:]]for[[:space:]]View-Adaptive[[:space:]]Rendering/ab13b82f-eb5f-4998-a6cb-284f692ff6ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Scalable[[:space:]]3D[[:space:]]Registration[[:space:]]via[[:space:]]Truncated[[:space:]]Entry-wise[[:space:]]Absolute[[:space:]]Residuals/726a0c98-c478-4703-a37b-dc1beb1efb19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Scaled[[:space:]]Decoupled[[:space:]]Distillation/9e11444b-bb0e-4927-b30f-6e0499b9ca08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Scaling[[:space:]]Diffusion[[:space:]]Models[[:space:]]to[[:space:]]Real-World[[:space:]]3D[[:space:]]LiDAR[[:space:]]Scene[[:space:]]Completion/de507081-73d8-4d94-954f-5aa0de9e9ffc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Scaling[[:space:]]Laws[[:space:]]for[[:space:]]Data[[:space:]]Filtering--[[:space:]]Data[[:space:]]Curation[[:space:]]cannot[[:space:]]be[[:space:]]Compute[[:space:]]Agnostic/ed7b37a2-0369-4d10-94a9-6b2641b0eec8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Scaling[[:space:]]Laws[[:space:]]of[[:space:]]Synthetic[[:space:]]Images[[:space:]]for[[:space:]]Model[[:space:]]Training[[:space:]]...[[:space:]]for[[:space:]]Now/bc20ced2-0a9b-4a2d-aaef-739dcd5ed654_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Scaling[[:space:]]Up[[:space:]]Dynamic[[:space:]]Human-Scene[[:space:]]Interaction[[:space:]]Modeling/ee0a9e38-b2c8-41a0-81fc-e8c96f42f93b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Scaling[[:space:]]Up[[:space:]]Video[[:space:]]Summarization[[:space:]]Pretraining[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Models/82d7d4de-736b-4639-ac98-74a59922fda4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Scaling[[:space:]]Up[[:space:]]to[[:space:]]Excellence_[[:space:]]Practicing[[:space:]]Model[[:space:]]Scaling[[:space:]]for[[:space:]]Photo-Realistic[[:space:]]Image[[:space:]]Restoration[[:space:]]In[[:space:]]the[[:space:]]Wild/484e5aa9-078f-4df4-afd0-bcd2d6454bbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ScanFormer_[[:space:]]Referring[[:space:]]Expression[[:space:]]Comprehension[[:space:]]by[[:space:]]Iteratively[[:space:]]Scanning/668110a3-8b98-4c3e-ad5c-7dda2133fd78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Scene[[:space:]]Adaptive[[:space:]]Sparse[[:space:]]Transformer[[:space:]]for[[:space:]]Event-based[[:space:]]Object[[:space:]]Detection/4abf9392-ffc1-43c3-9565-39c608b40e6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Scene-adaptive[[:space:]]and[[:space:]]Region-aware[[:space:]]Multi-modal[[:space:]]Prompt[[:space:]]for[[:space:]]Open[[:space:]]Vocabulary[[:space:]]Object[[:space:]]Detection/4a1f1cb7-5347-4c49-85a2-587eadcaf558_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SceneFun3D_[[:space:]]Fine-Grained[[:space:]]Functionality[[:space:]]and[[:space:]]Affordance[[:space:]]Understanding[[:space:]]in[[:space:]]3D[[:space:]]Scenes/3c35f917-0cd6-4287-9aff-58692efc884b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SceneTex_[[:space:]]High-Quality[[:space:]]Texture[[:space:]]Synthesis[[:space:]]for[[:space:]]Indoor[[:space:]]Scenes[[:space:]]via[[:space:]]Diffusion[[:space:]]Priors/7fb22265-9a71-432b-b6ee-634ecd69d0e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SchurVINS_[[:space:]]Schur[[:space:]]Complement-Based[[:space:]]Lightweight[[:space:]]Visual[[:space:]]Inertial[[:space:]]Navigation[[:space:]]System/07f22dc1-5b25-47c2-8794-5c7666d781d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Score-Guided[[:space:]]Diffusion[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Recovery/eae3b19d-b825-4f80-a5e1-6645a8e0f3e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ScoreHypo_[[:space:]]Probabilistic[[:space:]]Human[[:space:]]Mesh[[:space:]]Estimation[[:space:]]with[[:space:]]Hypothesis[[:space:]]Scoring/4c3821dc-8dcb-46f2-8e40-20ddd26decec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Sculpt3D_[[:space:]]Multi-View[[:space:]]Consistent[[:space:]]Text-to-3D[[:space:]]Generation[[:space:]]with[[:space:]]Sparse[[:space:]]3D[[:space:]]Prior/1a079134-206c-44c9-91d4-ee11d6828086_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Sculpting[[:space:]]Holistic[[:space:]]3D[[:space:]]Representation[[:space:]]in[[:space:]]Contrastive[[:space:]]Language-Image-3D[[:space:]]Pre-training/fdcf0f95-0850-4ad0-ad3b-101195b0229b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SeD_[[:space:]]Semantic-Aware[[:space:]]Discriminator[[:space:]]for[[:space:]]Image[[:space:]]Super-Resolution/62a328ef-829d-4ba5-afdd-7facb15cebc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SeMoLi_[[:space:]]What[[:space:]]Moves[[:space:]]Together[[:space:]]Belongs[[:space:]]Together/a3354110-9c50-4605-aa23-82f1f596ca00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SeNM-VAE_[[:space:]]Semi-Supervised[[:space:]]Noise[[:space:]]Modeling[[:space:]]with[[:space:]]Hierarchical[[:space:]]Variational[[:space:]]Autoencoder/ad73f036-a335-48d2-9a86-e9a7c89f96d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SeaBird_[[:space:]]Segmentation[[:space:]]in[[:space:]]Bird's[[:space:]]View[[:space:]]with[[:space:]]Dice[[:space:]]Loss[[:space:]]Improves[[:space:]]Monocular[[:space:]]3D[[:space:]]Detection[[:space:]]of[[:space:]]Large[[:space:]]Objects/5d863808-8f67-401b-ab4d-ceb87e34627a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Seamless[[:space:]]Human[[:space:]]Motion[[:space:]]Composition[[:space:]]with[[:space:]]Blended[[:space:]]Positional[[:space:]]Encodings/0f46e4eb-b6a7-4582-ae3d-56955ad38693_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SecondPose_[[:space:]]SE(3)-Consistent[[:space:]]Dual-Stream[[:space:]]Feature[[:space:]]Fusion[[:space:]]for[[:space:]]Category-Level[[:space:]]Pose[[:space:]]Estimation/a96143ec-828f-4f36-ae04-b2d932b35990_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/See[[:space:]]Say[[:space:]]and[[:space:]]Segment_[[:space:]]Teaching[[:space:]]LMMs[[:space:]]to[[:space:]]Overcome[[:space:]]False[[:space:]]Premises/85f7d456-4273-4828-a410-9f476ec4cb7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SeeSR_[[:space:]]Towards[[:space:]]Semantics-Aware[[:space:]]Real-World[[:space:]]Image[[:space:]]Super-Resolution/540b33bb-6f79-4eed-b435-74ddf2e0bbf0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Seeing[[:space:]]Motion[[:space:]]at[[:space:]]Nighttime[[:space:]]with[[:space:]]an[[:space:]]Event[[:space:]]Camera/bf7e93c3-a2d7-403d-9ebe-9dbd7e5e81e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Seeing[[:space:]]Unseen_[[:space:]]Discover[[:space:]]Novel[[:space:]]Biomedical[[:space:]]Concepts[[:space:]]via[[:space:]]Geometry-Constrained[[:space:]]Probabilistic[[:space:]]Modeling/a6051eca-f3e6-47a7-b7f9-2edd3c41e95f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Seeing[[:space:]]and[[:space:]]Hearing_[[:space:]]Open-domain[[:space:]]Visual-Audio[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion[[:space:]]Latent[[:space:]]Aligners/310ac9d8-7d21-4734-af22-86886ffd82cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Seeing[[:space:]]the[[:space:]]Unseen_[[:space:]]Visual[[:space:]]Common[[:space:]]Sense[[:space:]]for[[:space:]]Semantic[[:space:]]Placement/34d6596c-22c5-411c-b7ff-2f5783198f79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Seeing[[:space:]]the[[:space:]]World[[:space:]]through[[:space:]]Your[[:space:]]Eyes/6a3ff33c-caf8-4f01-b3c9-10f1076190fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Seg2Reg_[[:space:]]Differentiable[[:space:]]2D[[:space:]]Segmentation[[:space:]]to[[:space:]]1D[[:space:]]Regression[[:space:]]Rendering[[:space:]]for[[:space:]]360[[:space:]]Room[[:space:]]Layout[[:space:]]Reconstruction/cf8f43c7-e5f4-450d-91ea-332db569c87b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Segment[[:space:]]Any[[:space:]]Event[[:space:]]Streams[[:space:]]via[[:space:]]Weighted[[:space:]]Adaptation[[:space:]]of[[:space:]]Pivotal[[:space:]]Tokens/89b5d2fb-5221-4134-9996-0d10d47f78df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Segment[[:space:]]Every[[:space:]]Out-of-Distribution[[:space:]]Object/abe08b56-ae22-4da3-bb6d-5f88d5a262ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Segment[[:space:]]and[[:space:]]Caption[[:space:]]Anything/959b817d-225c-463d-a58a-99ae6f11b04f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Selective[[:space:]]Hourglass[[:space:]]Mapping[[:space:]]for[[:space:]]Universal[[:space:]]Image[[:space:]]Restoration[[:space:]]Based[[:space:]]on[[:space:]]Diffusion[[:space:]]Model/c78838dc-b69c-4fcf-b130-147beadbf273_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Selective[[:space:]]Interpretable[[:space:]]and[[:space:]]Motion[[:space:]]Consistent[[:space:]]Privacy[[:space:]]Attribute[[:space:]]Obfuscation[[:space:]]for[[:space:]]Action[[:space:]]Recognition/a1bdfbcc-dfb6-44e6-9a88-6216e9fd5ee6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Selective[[:space:]]Nonlinearities[[:space:]]Removal[[:space:]]from[[:space:]]Digital[[:space:]]Signals/ce4a4f68-a75b-4b4e-affa-326923d0b5b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Selective-Stereo_[[:space:]]Adaptive[[:space:]]Frequency[[:space:]]Information[[:space:]]Selection[[:space:]]for[[:space:]]Stereo[[:space:]]Matching/584de1de-5f52-4125-8106-326f8d8f0441_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Selectively[[:space:]]Informative[[:space:]]Description[[:space:]]can[[:space:]]Reduce[[:space:]]Undesired[[:space:]]Embedding[[:space:]]Entanglements[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Personalization/6ad8fbbe-ab46-41fb-8d30-63234684924c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Self-Adaptive[[:space:]]Reality-Guided[[:space:]]Diffusion[[:space:]]for[[:space:]]Artifact-Free[[:space:]]Super-Resolution/c6d65069-6c73-46fe-956d-1a0443a4ed2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Self-Calibrating[[:space:]]Vicinal[[:space:]]Risk[[:space:]]Minimisation[[:space:]]for[[:space:]]Model[[:space:]]Calibration/d8d13c75-1d94-496a-b79e-a2d1b272921f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Self-Discovering[[:space:]]Interpretable[[:space:]]Diffusion[[:space:]]Latent[[:space:]]Directions[[:space:]]for[[:space:]]Responsible[[:space:]]Text-to-Image[[:space:]]Generation/3095ee86-791c-49b2-8a2e-753d19e700c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Self-Distilled[[:space:]]Masked[[:space:]]Auto-Encoders[[:space:]]are[[:space:]]Efficient[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detectors/b79b2889-9304-49a2-9631-466d8e496991_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Self-Supervised[[:space:]]Class-Agnostic[[:space:]]Motion[[:space:]]Prediction[[:space:]]with[[:space:]]Spatial[[:space:]]and[[:space:]]Temporal[[:space:]]Consistency[[:space:]]Regularizations/b067b514-00bd-444a-bc87-e7fa056dae8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Self-Supervised[[:space:]]Dual[[:space:]]Contouring/0eb29fd9-574c-47d8-9f41-53c9d4c10474_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Self-Supervised[[:space:]]Facial[[:space:]]Representation[[:space:]]Learning[[:space:]]with[[:space:]]Facial[[:space:]]Region[[:space:]]Awareness/bb83bfc2-6351-436b-90f5-921020f4fe95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Self-Supervised[[:space:]]Multi-Object[[:space:]]Tracking[[:space:]]with[[:space:]]Path[[:space:]]Consistency/126e321e-aac6-4315-8a05-8ce49f269b09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Self-Supervised[[:space:]]Representation[[:space:]]Learning[[:space:]]from[[:space:]]Arbitrary[[:space:]]Scenarios/90fe2be8-b444-4296-b490-1791873650b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Self-Training[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Improved[[:space:]]Visual[[:space:]]Program[[:space:]]Synthesis[[:space:]]With[[:space:]]Visual[[:space:]]Reinforcement/011d32ff-cfb4-4400-94b5-e42371e5459f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Self-correcting[[:space:]]LLM-controlled[[:space:]]Diffusion[[:space:]]Models/1d07a470-ae6d-42d9-bf7a-62b3e923ba50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Self-supervised[[:space:]]Debiasing[[:space:]]Using[[:space:]]Low[[:space:]]Rank[[:space:]]Regularization/9a99109e-c17a-4511-ac59-3693469d8cef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SelfOcc_[[:space:]]Self-Supervised[[:space:]]Vision-Based[[:space:]]3D[[:space:]]Occupancy[[:space:]]Prediction/322bb044-d09b-4af1-b281-2137f8b01f03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SelfPose3d_[[:space:]]Self-Supervised[[:space:]]Multi-Person[[:space:]]Multi-View[[:space:]]3d[[:space:]]Pose[[:space:]]Estimation/4659f454-3934-4652-8e58-699358f95d03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SemCity_[[:space:]]Semantic[[:space:]]Scene[[:space:]]Generation[[:space:]]with[[:space:]]Triplane[[:space:]]Diffusion/36ac8a85-7845-4237-8f1d-f2203275c4a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Semantic[[:space:]]Human[[:space:]]Mesh[[:space:]]Reconstruction[[:space:]]with[[:space:]]Textures/08bc6bd7-0008-4e31-94a7-d23c37e8682d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Semantic[[:space:]]Line[[:space:]]Combination[[:space:]]Detector/09fa8b25-63f5-44a7-8c3c-7ce99a11a2c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Separating[[:space:]]the[[:space:]]'Chirp'[[:space:]]from[[:space:]]the[[:space:]]'Chat'_[[:space:]]Self-supervised[[:space:]]Visual[[:space:]]Grounding[[:space:]]of[[:space:]]Sound[[:space:]]and[[:space:]]Language/7873c286-c4ef-4202-8df2-5bd3e7312a38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Sequential[[:space:]]Modeling[[:space:]]Enables[[:space:]]Scalable[[:space:]]Learning[[:space:]]for[[:space:]]Large[[:space:]]Vision[[:space:]]Models/262d155e-8b9e-40b4-8485-80e9c6034a6b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SfmCAD_[[:space:]]Unsupervised[[:space:]]CAD[[:space:]]Reconstruction[[:space:]]by[[:space:]]Learning[[:space:]]Sketch-based[[:space:]]Feature[[:space:]]Modeling[[:space:]]Operations/a4171579-c596-406f-93c6-62c45be3925d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Shadow[[:space:]]Generation[[:space:]]for[[:space:]]Composite[[:space:]]Image[[:space:]]Using[[:space:]]Diffusion[[:space:]]Model/d6e83b88-6d3b-4f78-b729-db99f8dc57c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Shadow-Enlightened[[:space:]]Image[[:space:]]Outpainting/50380c3f-4d94-46f3-bb7f-dbe694f86799_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Shadows[[:space:]]Don't[[:space:]]Lie[[:space:]]and[[:space:]]Lines[[:space:]]Can't[[:space:]]Bend![[:space:]]Generative[[:space:]]Models[[:space:]]don't[[:space:]]know[[:space:]]Projective[[:space:]]Geometry...for[[:space:]]now/6c6001bd-500a-4e2e-8037-41da45d3b01b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Shallow-Deep[[:space:]]Collaborative[[:space:]]Learning[[:space:]]for[[:space:]]Unsupervised[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification/93501bbe-f504-4162-9926-acaa45b1fb43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ShapeMatcher_[[:space:]]Self-Supervised[[:space:]]Joint[[:space:]]Shape[[:space:]]Canonicalization[[:space:]]Segmentation[[:space:]]Retrieval[[:space:]]and[[:space:]]Deformation/8238426b-d0c8-4a68-b942-49f91f1e7a27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ShapeWalk_[[:space:]]Compositional[[:space:]]Shape[[:space:]]Editing[[:space:]]Through[[:space:]]Language-Guided[[:space:]]Chains/571b12bc-0855-45a3-8d18-3317ec9acf46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Sharingan_[[:space:]]A[[:space:]]Transformer[[:space:]]Architecture[[:space:]]for[[:space:]]Multi-Person[[:space:]]Gaze[[:space:]]Following/df6417d4-023a-4e78-a931-75b2f738abf7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Sheared[[:space:]]Backpropagation[[:space:]]for[[:space:]]Fine-tuning[[:space:]]Foundation[[:space:]]Models/362c0568-56d1-4aa5-ab6f-aa648bbaa531_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Sherpa3D_[[:space:]]Boosting[[:space:]]High-Fidelity[[:space:]]Text-to-3D[[:space:]]Generation[[:space:]]via[[:space:]]Coarse[[:space:]]3D[[:space:]]Prior/4ac2d1ab-faf4-4d9c-949b-ff0f163e6a1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SiTH_[[:space:]]Single-view[[:space:]]Textured[[:space:]]Human[[:space:]]Reconstruction[[:space:]]with[[:space:]]Image-Conditioned[[:space:]]Diffusion/978debbe-87ee-4742-aad1-3f1d3627166b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Siamese[[:space:]]Learning[[:space:]]with[[:space:]]Joint[[:space:]]Alignment[[:space:]]and[[:space:]]Regression[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Video[[:space:]]Paragraph[[:space:]]Grounding/b2a21229-501b-4026-b12f-315fb210d141_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Sieve_[[:space:]]Multimodal[[:space:]]Dataset[[:space:]]Pruning[[:space:]]using[[:space:]]Image[[:space:]]Captioning[[:space:]]Models/6ff5f5b2-831e-43d1-8c97-18e5f39eb692_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SignGraph_[[:space:]]A[[:space:]]Sign[[:space:]]Sequence[[:space:]]is[[:space:]]Worth[[:space:]]Graphs[[:space:]]of[[:space:]]Nodes/5f8bdf16-8e1b-41c3-9a76-ee437197dbea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SimAC_[[:space:]]A[[:space:]]Simple[[:space:]]Anti-Customization[[:space:]]Method[[:space:]]for[[:space:]]Protecting[[:space:]]Face[[:space:]]Privacy[[:space:]]against[[:space:]]Text-to-Image[[:space:]]Synthesis[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/95390555-fbca-4ec8-a42e-5840969340e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SimDA_[[:space:]]Simple[[:space:]]Diffusion[[:space:]]Adapter[[:space:]]for[[:space:]]Efficient[[:space:]]Video[[:space:]]Generation/2fcde3ae-cff4-454a-8644-ea50caebb3c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Simple[[:space:]]Semantic-Aided[[:space:]]Few-Shot[[:space:]]Learning/f104bef6-8837-4adc-8ad9-f16490a614a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SinSR_[[:space:]]Diffusion-Based[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]in[[:space:]]a[[:space:]]Single[[:space:]]Step/7e640a14-4213-4e9a-b98f-0e2a0e646830_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Single[[:space:]]Domain[[:space:]]Generalization[[:space:]]for[[:space:]]Crowd[[:space:]]Counting/f370ad4c-18eb-4cb3-987b-d53d7152a8e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Single[[:space:]]Mesh[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Field[[:space:]]Latents[[:space:]]for[[:space:]]Texture[[:space:]]Generation/bf087a44-a85c-4958-a01e-c35dc0ca2939_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Single[[:space:]]View[[:space:]]Refractive[[:space:]]Index[[:space:]]Tomography[[:space:]]with[[:space:]]Neural[[:space:]]Fields/887b4199-788b-44cf-9fe5-91d50d30037c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Single-Model[[:space:]]and[[:space:]]Any-Modality[[:space:]]for[[:space:]]Video[[:space:]]Object[[:space:]]Tracking/06dd6563-2018-4142-8391-b2e6da00243a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Single-View[[:space:]]Scene[[:space:]]Point[[:space:]]Cloud[[:space:]]Human[[:space:]]Grasp[[:space:]]Generation/4153e0a8-7264-4ade-83b0-cfe42df66a8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Single-to-Dual-View[[:space:]]Adaptation[[:space:]]for[[:space:]]Egocentric[[:space:]]3D[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation/899acf81-bf1d-4090-8b2e-d789914e0b58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SingularTrajectory_[[:space:]]Universal[[:space:]]Trajectory[[:space:]]Predictor[[:space:]]Using[[:space:]]Diffusion[[:space:]]Model/2a853686-8806-4e80-a078-f0cdecbd3caf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Situational[[:space:]]Awareness[[:space:]]Matters[[:space:]]in[[:space:]]3D[[:space:]]Vision[[:space:]]Language[[:space:]]Reasoning/1ff0a4cb-e0c0-41a2-9620-42d78c2488cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Skeleton-in-Context_[[:space:]]Unified[[:space:]]Skeleton[[:space:]]Sequence[[:space:]]Modeling[[:space:]]with[[:space:]]In-Context[[:space:]]Learning/d6e51c0b-3655-41d0-b00a-3888d1dd8906_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SketchINR_[[:space:]]A[[:space:]]First[[:space:]]Look[[:space:]]into[[:space:]]Sketches[[:space:]]as[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations/ab26a978-e483-4292-89b7-9fe0134cf389_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SkillDiffuser_[[:space:]]Interpretable[[:space:]]Hierarchical[[:space:]]Planning[[:space:]]via[[:space:]]Skill[[:space:]]Abstractions[[:space:]]in[[:space:]]Diffusion-Based[[:space:]]Task[[:space:]]Execution/7ae363f2-9817-4f7b-ab1b-3e5ad622cb0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SkySense_[[:space:]]A[[:space:]]Multi-Modal[[:space:]]Remote[[:space:]]Sensing[[:space:]]Foundation[[:space:]]Model[[:space:]]Towards[[:space:]]Universal[[:space:]]Interpretation[[:space:]]for[[:space:]]Earth[[:space:]]Observation[[:space:]]Imagery/e566e7dc-1647-497b-a48b-fadeb9fef24f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SleepVST_[[:space:]]Sleep[[:space:]]Staging[[:space:]]from[[:space:]]Near-Infrared[[:space:]]Video[[:space:]]Signals[[:space:]]using[[:space:]]Pre-Trained[[:space:]]Transformers/fb8d4880-9157-4c04-a223-f42e3fd034fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Slice3D_[[:space:]]Multi-Slice[[:space:]]Occlusion-Revealing[[:space:]]Single[[:space:]]View[[:space:]]3D[[:space:]]Reconstruction/e6768b9a-cd0d-4c71-899b-0e380d81b135_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SlowFormer_[[:space:]]Adversarial[[:space:]]Attack[[:space:]]on[[:space:]]Compute[[:space:]]and[[:space:]]Energy[[:space:]]Consumption[[:space:]]of[[:space:]]Efficient[[:space:]]Vision[[:space:]]Transformers/93475bbf-f8a6-48bc-a54b-7b5297d3501e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Small[[:space:]]Scale[[:space:]]Data-Free[[:space:]]Knowledge[[:space:]]Distillation/9c1a5016-dfa0-4549-8999-c0090c1c7132_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Small[[:space:]]Steps[[:space:]]and[[:space:]]Level[[:space:]]Sets_[[:space:]]Fitting[[:space:]]Neural[[:space:]]Surface[[:space:]]Models[[:space:]]with[[:space:]]Point[[:space:]]Guidance/688fa8a6-01ab-4bec-bd21-cd592944e2e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Smart[[:space:]]Help_[[:space:]]Strategic[[:space:]]Opponent[[:space:]]Modeling[[:space:]]for[[:space:]]Proactive[[:space:]]and[[:space:]]Adaptive[[:space:]]Robot[[:space:]]Assistance[[:space:]]in[[:space:]]Households/e39ca544-b333-4b15-8b22-bd984a4f8117_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SmartEdit_[[:space:]]Exploring[[:space:]]Complex[[:space:]]Instruction-based[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/a4f5fe59-781a-4bca-bc8e-e339b6bcabf9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SmartMask_[[:space:]]Context[[:space:]]Aware[[:space:]]High-Fidelity[[:space:]]Mask[[:space:]]Generation[[:space:]]for[[:space:]]Fine-grained[[:space:]]Object[[:space:]]Insertion[[:space:]]and[[:space:]]Layout[[:space:]]Control/33eba842-4e63-4787-952f-aab8f2b01e73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SmartRefine_[[:space:]]A[[:space:]]Scenario-Adaptive[[:space:]]Refinement[[:space:]]Framework[[:space:]]for[[:space:]]Efficient[[:space:]]Motion[[:space:]]Prediction/f7e9eb16-11f4-4de8-b800-9a022b980e5e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Smooth[[:space:]]Diffusion_[[:space:]]Crafting[[:space:]]Smooth[[:space:]]Latent[[:space:]]Spaces[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/88a5ba4c-516f-44e9-b601-18a2cbb929d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SnAG_[[:space:]]Scalable[[:space:]]and[[:space:]]Accurate[[:space:]]Video[[:space:]]Grounding/4d098779-b5a1-4518-ac90-5b419d9005ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Snap[[:space:]]Video_[[:space:]]Scaled[[:space:]]Spatiotemporal[[:space:]]Transformers[[:space:]]for[[:space:]]Text-to-Video[[:space:]]Synthesis/adef51ac-64f5-47cc-b577-559e45a201d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Snapshot[[:space:]]Lidar_[[:space:]]Fourier[[:space:]]Embedding[[:space:]]of[[:space:]]Amplitude[[:space:]]and[[:space:]]Phase[[:space:]]for[[:space:]]Single-Image[[:space:]]Depth[[:space:]]Reconstruction/4828c6e5-d978-4f03-9ac3-919a10abf8a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SocialCircle_[[:space:]]Learning[[:space:]]the[[:space:]]Angle-based[[:space:]]Social[[:space:]]Interaction[[:space:]]Representation[[:space:]]for[[:space:]]Pedestrian[[:space:]]Trajectory[[:space:]]Prediction/5b88fdbf-49dd-42ca-9334-fda1ac3ea6a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SocialCounterfactuals_[[:space:]]Probing[[:space:]]and[[:space:]]Mitigating[[:space:]]Intersectional[[:space:]]Social[[:space:]]Biases[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]Counterfactual[[:space:]]Examples/e6b7bc50-d0f7-4de0-a9f0-f633baf42ce9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Soften[[:space:]]to[[:space:]]Defend_[[:space:]]Towards[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]via[[:space:]]Self-Guided[[:space:]]Label[[:space:]]Refinement/e8af8e09-8ccf-4c64-8c0b-738d9e4d84de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Solving[[:space:]]Masked[[:space:]]Jigsaw[[:space:]]Puzzles[[:space:]]with[[:space:]]Diffusion[[:space:]]Vision[[:space:]]Transformers/f9db8bd8-d9fa-40a9-84f3-75a6cea32e88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Solving[[:space:]]the[[:space:]]Catastrophic[[:space:]]Forgetting[[:space:]]Problem[[:space:]]in[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/6bc5b6dc-66ef-4cae-bd24-fa6ae8788463_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SonicVisionLM_[[:space:]]Playing[[:space:]]Sound[[:space:]]with[[:space:]]Vision[[:space:]]Language[[:space:]]Models/578ceece-2af8-4d00-b6af-84640cb38a96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SoundingActions_[[:space:]]Learning[[:space:]]How[[:space:]]Actions[[:space:]]Sound[[:space:]]from[[:space:]]Narrated[[:space:]]Egocentric[[:space:]]Videos/87979a59-b194-4c7b-8bc8-5f7988336f66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Source-Free[[:space:]]Domain[[:space:]]Adaptation[[:space:]]with[[:space:]]Frozen[[:space:]]Multimodal[[:space:]]Foundation[[:space:]]Model/031d437d-6302-4046-b01d-f59ceeac78db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Space-Time[[:space:]]Diffusion[[:space:]]Features[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Text-Driven[[:space:]]Motion[[:space:]]Transfer/2607e128-007d-45ee-9a2f-199fbf395f11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Spacetime[[:space:]]Gaussian[[:space:]]Feature[[:space:]]Splatting[[:space:]]for[[:space:]]Real-Time[[:space:]]Dynamic[[:space:]]View[[:space:]]Synthesis/fc66fcb8-31e6-496e-8bf3-fedca92ba729_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Spanning[[:space:]]Training[[:space:]]Progress_[[:space:]]Temporal[[:space:]]Dual-Depth[[:space:]]Scoring[[:space:]](TDDS)[[:space:]]for[[:space:]]Enhanced[[:space:]]Dataset[[:space:]]Pruning/aa264c21-255c-4403-aadd-d1e518b57566_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Sparse[[:space:]]Global[[:space:]]Matching[[:space:]]for[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation[[:space:]]with[[:space:]]Large[[:space:]]Motion/e8af234c-bd88-478c-b3ba-46c221c02a9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Sparse[[:space:]]Semi-DETR_[[:space:]]Sparse[[:space:]]Learnable[[:space:]]Queries[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Object[[:space:]]Detection/b3fec2a6-daae-4072-a5bd-8911b9d30c68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Sparse[[:space:]]Views[[:space:]]Near[[:space:]]Light_[[:space:]]A[[:space:]]Practical[[:space:]]Paradigm[[:space:]]for[[:space:]]Uncalibrated[[:space:]]Point-light[[:space:]]Photometric[[:space:]]Stereo/631ca577-50bf-4627-9720-eaa20e92dcf3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SparseOcc_[[:space:]]Rethinking[[:space:]]Sparse[[:space:]]Latent[[:space:]]Representation[[:space:]]for[[:space:]]Vision-Based[[:space:]]Semantic[[:space:]]Occupancy[[:space:]]Prediction/6bce05c4-f8ee-4452-886d-3f99d4f2a003_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Spatial-Aware[[:space:]]Regression[[:space:]]for[[:space:]]Keypoint[[:space:]]Localization/fbd66afd-f4f6-451c-8993-88b8e434975c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SpatialTracker_[[:space:]]Tracking[[:space:]]Any[[:space:]]2D[[:space:]]Pixels[[:space:]]in[[:space:]]3D[[:space:]]Space/fee0371c-0071-4b47-bb08-5807dbb57d30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SpatialVLM_[[:space:]]Endowing[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]Spatial[[:space:]]Reasoning[[:space:]]Capabilities/dc5db9cb-fcf5-4067-8b85-ffb92320aba8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Spatio-Temporal[[:space:]]Turbulence[[:space:]]Mitigation_[[:space:]]A[[:space:]]Translational[[:space:]]Perspective/eefae261-74a4-4bbc-a466-69dbac77972f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SpecNeRF_[[:space:]]Gaussian[[:space:]]Directional[[:space:]]Encoding[[:space:]]for[[:space:]]Specular[[:space:]]Reflections/3d459af2-f520-428c-a01a-361ee7f87043_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Spectral[[:space:]]Meets[[:space:]]Spatial_[[:space:]]Harmonising[[:space:]]3D[[:space:]]Shape[[:space:]]Matching[[:space:]]and[[:space:]]Interpolation/2d618972-8924-4059-ac0c-1df6a06ff55a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Spectral[[:space:]]and[[:space:]]Polarization[[:space:]]Vision_[[:space:]]Spectro-polarimetric[[:space:]]Real-world[[:space:]]Dataset/af12fa50-9d20-41ff-af46-6bfc69201b2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Spectrum[[:space:]]AUC[[:space:]]Difference[[:space:]](SAUCD)_[[:space:]]Human-aligned[[:space:]]3D[[:space:]]Shape[[:space:]]Evaluation/dc4ca58e-2733-4afc-93ce-8d6ac2242638_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Specularity[[:space:]]Factorization[[:space:]]for[[:space:]]Low-Light[[:space:]]Enhancement/f1d0299e-d664-4662-ab2f-1559d4f3b0a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Spherical[[:space:]]Mask_[[:space:]]Coarse-to-Fine[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Instance[[:space:]]Segmentation[[:space:]]with[[:space:]]Spherical[[:space:]]Representation/adcaad31-f6bf-4625-9641-e0b1d8451ad5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SpiderMatch_[[:space:]]3D[[:space:]]Shape[[:space:]]Matching[[:space:]]with[[:space:]]Global[[:space:]]Optimality[[:space:]]and[[:space:]]Geometric[[:space:]]Consistency/dc02bdd8-9eaf-4562-b3e4-4eb5d67e4085_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Spike-guided[[:space:]]Motion[[:space:]]Deblurring[[:space:]]with[[:space:]]Unknown[[:space:]]Modal[[:space:]]Spatiotemporal[[:space:]]Alignment/fc7703e6-e447-4073-bf40-f253b97da714_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SpikeNeRF_[[:space:]]Learning[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]from[[:space:]]Continuous[[:space:]]Spike[[:space:]]Stream/028c841a-a56f-4e19-871e-e3d2e2b3ec7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SpikingResformer_[[:space:]]Bridging[[:space:]]ResNet[[:space:]]and[[:space:]]Vision[[:space:]]Transformer[[:space:]]in[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks/02fe2792-da1c-4c04-8cc2-8f8d42352f7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Spin-UP_[[:space:]]Spin[[:space:]]Light[[:space:]]for[[:space:]]Natural[[:space:]]Light[[:space:]]Uncalibrated[[:space:]]Photometric[[:space:]]Stereo/6c1acd6f-00c1-4dc9-a5c3-6c6dd410bb2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SplaTAM_[[:space:]]Splat[[:space:]]Track[[:space:]]&[[:space:]]Map[[:space:]]3D[[:space:]]Gaussians[[:space:]]for[[:space:]]Dense[[:space:]]RGB-D[[:space:]]SLAM/177d5eaf-8449-4cd5-ba1f-ab6fb569d8d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Splatter[[:space:]]Image_[[:space:]]Ultra-Fast[[:space:]]Single-View[[:space:]]3D[[:space:]]Reconstruction/a13161ef-830a-482f-925a-3b60cfe25a4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SplattingAvatar_[[:space:]]Realistic[[:space:]]Real-Time[[:space:]]Human[[:space:]]Avatars[[:space:]]with[[:space:]]Mesh-Embedded[[:space:]]Gaussian[[:space:]]Splatting/2a923612-0df2-4faf-a040-0816baf81069_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Split[[:space:]]to[[:space:]]Merge_[[:space:]]Unifying[[:space:]]Separated[[:space:]]Modalities[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/0b95bf41-b7e1-4df0-9ad3-d873e89c86e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SportsHHI_[[:space:]]A[[:space:]]Dataset[[:space:]]for[[:space:]]Human-Human[[:space:]]Interaction[[:space:]]Detection[[:space:]]in[[:space:]]Sports[[:space:]]Videos/c0955323-ec6c-46f0-bf26-c3cb285a2d3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Stable[[:space:]]Neighbor[[:space:]]Denoising[[:space:]]for[[:space:]]Source-free[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Segmentation/06cbf727-53a9-4ece-bc0b-04416037b63c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/StableVITON_[[:space:]]Learning[[:space:]]Semantic[[:space:]]Correspondence[[:space:]]with[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Virtual[[:space:]]Try-On/d486dcf7-d514-4538-af86-ab7be3d92c46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/State[[:space:]]Space[[:space:]]Models[[:space:]]for[[:space:]]Event[[:space:]]Cameras/04cf8976-65e9-4903-adef-22c48bc5447f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Stationary[[:space:]]Representations_[[:space:]]Optimally[[:space:]]Approximating[[:space:]]Compatibility[[:space:]]and[[:space:]]Implications[[:space:]]for[[:space:]]Improved[[:space:]]Model[[:space:]]Replacements/554d31c6-935e-4439-a512-f2ee4bfee481_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Steerers_[[:space:]]A[[:space:]]Framework[[:space:]]for[[:space:]]Rotation[[:space:]]Equivariant[[:space:]]Keypoint[[:space:]]Descriptors/e8356a95-c6c3-4afe-8ccf-5c5531990c8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Steganographic[[:space:]]Passport_[[:space:]]An[[:space:]]Owner[[:space:]]and[[:space:]]User[[:space:]]Verifiable[[:space:]]Credential[[:space:]]for[[:space:]]Deep[[:space:]]Model[[:space:]]IP[[:space:]]Protection[[:space:]]Without[[:space:]]Retraining/723188cf-21a3-4072-a767-991efe3bf48f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/StegoGAN_[[:space:]]Leveraging[[:space:]]Steganography[[:space:]]for[[:space:]]Non-Bijective[[:space:]]Image-to-Image[[:space:]]Translation/6c6dbef8-429a-42af-8789-b7a3716cc2bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Step[[:space:]]Differences[[:space:]]in[[:space:]]Instructional[[:space:]]Video/ddcdbdd3-a745-4d02-aef0-ad93fd7b5b5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/StraightPCF_[[:space:]]Straight[[:space:]]Point[[:space:]]Cloud[[:space:]]Filtering/af60050a-e20e-4b9c-abbd-b5e0d81faed0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Stratified[[:space:]]Avatar[[:space:]]Generation[[:space:]]from[[:space:]]Sparse[[:space:]]Observations/efda3faf-6f32-487b-b0f2-cf234b098ea4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Streaming[[:space:]]Dense[[:space:]]Video[[:space:]]Captioning/d7bbec97-020a-4518-8c20-5703b1fc127a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/StreamingFlow_[[:space:]]Streaming[[:space:]]Occupancy[[:space:]]Forecasting[[:space:]]with[[:space:]]Asynchronous[[:space:]]Multi-modal[[:space:]]Data[[:space:]]Streams[[:space:]]via[[:space:]]Neural[[:space:]]Ordinary[[:space:]]Differential[[:space:]]Equation/d9ff698c-3801-40b5-9c86-560933f80c96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/StrokeFaceNeRF_[[:space:]]Stroke-based[[:space:]]Facial[[:space:]]Appearance[[:space:]]Editing[[:space:]]in[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field/7fe1ba34-6e33-48a7-bca3-a74b49e199ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Strong[[:space:]]Transferable[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]via[[:space:]]Ensembled[[:space:]]Asymptotically[[:space:]]Normal[[:space:]]Distribution[[:space:]]Learning/6823fc5a-81c6-4d57-a39f-933f81a96ddc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Stronger[[:space:]]Fewer[[:space:]]&[[:space:]]Superior_[[:space:]]Harnessing[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models[[:space:]]for[[:space:]]Domain[[:space:]]Generalized[[:space:]]Semantic[[:space:]]Segmentation/89aa68d1-2eb8-4184-895d-0bf0c9b627f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Structure[[:space:]]Matters_[[:space:]]Tackling[[:space:]]the[[:space:]]Semantic[[:space:]]Discrepancy[[:space:]]in[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Image[[:space:]]Inpainting/14a78d05-b6ef-4749-8f8e-25cfaadc10a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Structure-Aware[[:space:]]Sparse-View[[:space:]]X-ray[[:space:]]3D[[:space:]]Reconstruction/e0ae7df8-d3f1-46c2-a34a-52862215b941_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Structure-Guided[[:space:]]Adversarial[[:space:]]Training[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/b70a5045-dbc8-47eb-880e-4252d02e0e55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Structured[[:space:]]Gradient-based[[:space:]]Interpretations[[:space:]]via[[:space:]]Norm-Regularized[[:space:]]Adversarial[[:space:]]Training/25726c0b-b33a-4e50-8b53-f013deb7f334_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Structured[[:space:]]Model[[:space:]]Probing_[[:space:]]Empowering[[:space:]]Efficient[[:space:]]Transfer[[:space:]]Learning[[:space:]]by[[:space:]]Structured[[:space:]]Regularization/fcd62d92-907d-4f7f-bc80-b88db83abf7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/StyLitGAN_[[:space:]]Image-Based[[:space:]]Relighting[[:space:]]via[[:space:]]Latent[[:space:]]Control/37a6c7df-f82e-4da6-aa4d-10d6b8e2b178_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Style[[:space:]]Aligned[[:space:]]Image[[:space:]]Generation[[:space:]]via[[:space:]]Shared[[:space:]]Attention/0cae3168-e63f-4f71-908a-5bdeae8d8200_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Style[[:space:]]Blind[[:space:]]Domain[[:space:]]Generalized[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]via[[:space:]]Covariance[[:space:]]Alignment[[:space:]]and[[:space:]]Semantic[[:space:]]Consistence[[:space:]]Contrastive[[:space:]]Learning/1e340b0f-2934-49e8-9933-238d4fb891ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Style[[:space:]]Injection[[:space:]]in[[:space:]]Diffusion_[[:space:]]A[[:space:]]Training-free[[:space:]]Approach[[:space:]]for[[:space:]]Adapting[[:space:]]Large-scale[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Style[[:space:]]Transfer/43978830-73bd-4db5-8b8c-57888ce93c6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/StyleCineGAN_[[:space:]]Landscape[[:space:]]Cinemagraph[[:space:]]Generation[[:space:]]using[[:space:]]a[[:space:]]Pre-trained[[:space:]]StyleGAN/51ff4f8c-4976-4d05-85df-cf04bbc75017_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SuGaR_[[:space:]]Surface-Aligned[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Efficient[[:space:]]3D[[:space:]]Mesh[[:space:]]Reconstruction[[:space:]]and[[:space:]]High-Quality[[:space:]]Mesh[[:space:]]Rendering/8c37e125-7381-4a71-8f2b-015c3d8a066a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SubT-MRS[[:space:]]Dataset_[[:space:]]Pushing[[:space:]]SLAM[[:space:]]Towards[[:space:]]All-weather[[:space:]]Environments/80ad799c-9412-49c0-a322-3a5e2a9d1a8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Summarize[[:space:]]the[[:space:]]Past[[:space:]]to[[:space:]]Predict[[:space:]]the[[:space:]]Future_[[:space:]]Natural[[:space:]]Language[[:space:]]Descriptions[[:space:]]of[[:space:]]Context[[:space:]]Boost[[:space:]]Multimodal[[:space:]]Object[[:space:]]Interaction[[:space:]]Anticipation/5f051557-0067-4cd2-89f8-ff27dddafa6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Super-Resolution[[:space:]]Reconstruction[[:space:]]from[[:space:]]Bayer-Pattern[[:space:]]Spike[[:space:]]Streams/ac654108-cae7-444a-a7e8-a0faf89a682d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SuperNormal_[[:space:]]Neural[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]via[[:space:]]Multi-View[[:space:]]Normal[[:space:]]Integration/a4b33f00-d1f9-45e2-a62c-bb25956b58ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SuperPrimitive_[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]at[[:space:]]a[[:space:]]Primitive[[:space:]]Level/556fb01c-1445-4c08-b116-0426ba52ea17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SuperSVG_[[:space:]]Superpixel-based[[:space:]]Scalable[[:space:]]Vector[[:space:]]Graphics[[:space:]]Synthesis/3fa1e5c3-5890-4208-976c-19acdb43c796_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Supervised[[:space:]]Anomaly[[:space:]]Detection[[:space:]]for[[:space:]]Complex[[:space:]]Industrial[[:space:]]Images/ef358e69-3d45-4c6c-aa99-f5a5ceb11e08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Suppress[[:space:]]and[[:space:]]Rebalance_[[:space:]]Towards[[:space:]]Generalized[[:space:]]Multi-Modal[[:space:]]Face[[:space:]]Anti-Spoofing/91e475e8-d317-465c-893c-024b420e20a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SurMo_[[:space:]]Surface-based[[:space:]]4D[[:space:]]Motion[[:space:]]Modeling[[:space:]]for[[:space:]]Dynamic[[:space:]]Human[[:space:]]Rendering/6283bde0-3818-458b-8029-7eb693822789_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SurroundSDF_[[:space:]]Implicit[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding[[:space:]]Based[[:space:]]on[[:space:]]Signed[[:space:]]Distance[[:space:]]Field/213de3e8-6a97-4d16-96b2-ac03f2f0441a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SwiftBrush_[[:space:]]One-Step[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Model[[:space:]]with[[:space:]]Variational[[:space:]]Score[[:space:]]Distillation/13853f95-872e-49e0-9add-f414db12f2de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SwitchLight_[[:space:]]Co-design[[:space:]]of[[:space:]]Physics-driven[[:space:]]Architecture[[:space:]]and[[:space:]]Pre-training[[:space:]]Framework[[:space:]]for[[:space:]]Human[[:space:]]Portrait[[:space:]]Relighting/ce409cac-fd5a-4614-9e6d-87e790ca611e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Symphonize[[:space:]]3D[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion[[:space:]]with[[:space:]]Contextual[[:space:]]Instance[[:space:]]Queries/ba02978b-283d-434b-af7a-d4b604064e5e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SynFog_[[:space:]]A[[:space:]]Photo-realistic[[:space:]]Synthetic[[:space:]]Fog[[:space:]]Dataset[[:space:]]based[[:space:]]on[[:space:]]End-to-end[[:space:]]Imaging[[:space:]]Simulation[[:space:]]for[[:space:]]Advancing[[:space:]]Real-World[[:space:]]Defogging[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/30117727-713c-4f8f-a215-de682288f1dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SynSP_[[:space:]]Synergy[[:space:]]of[[:space:]]Smoothness[[:space:]]and[[:space:]]Precision[[:space:]]in[[:space:]]Pose[[:space:]]Sequences[[:space:]]Refinement/e165eedf-6d86-4e7b-90c3-ccc6312e093e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SyncMask_[[:space:]]Synchronized[[:space:]]Attentional[[:space:]]Masking[[:space:]]for[[:space:]]Fashion-centric[[:space:]]Vision-Language[[:space:]]Pretraining/92bc41a7-a993-4d23-a197-14e6dd1e0509_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/SyncTalk_[[:space:]]The[[:space:]]Devil[[:space:]]is[[:space:]]in[[:space:]]the[[:space:]]Synchronization[[:space:]]for[[:space:]]Talking[[:space:]]Head[[:space:]]Synthesis/89645a31-8829-4d8f-939e-0d17fcee95a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Synergistic[[:space:]]Global-space[[:space:]]Camera[[:space:]]and[[:space:]]Human[[:space:]]Reconstruction[[:space:]]from[[:space:]]Videos/908ee619-1331-452b-9ccb-4e5b4d116ac0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Synthesize[[:space:]]Diagnose[[:space:]]and[[:space:]]Optimize_[[:space:]]Towards[[:space:]]Fine-Grained[[:space:]]Vision-Language[[:space:]]Understanding/eca2ceba-84d3-4751-b38a-1ba7e74c92ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Synthesize[[:space:]]Step-by-Step_[[:space:]]Tools[[:space:]]Templates[[:space:]]and[[:space:]]LLMs[[:space:]]as[[:space:]]Data[[:space:]]Generators[[:space:]]for[[:space:]]Reasoning-Based[[:space:]]Chart[[:space:]]VQA/58e8331e-f572-44c9-b724-02ca5cd604a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Systematic[[:space:]]Comparison[[:space:]]of[[:space:]]Semi-supervised[[:space:]]and[[:space:]]Self-supervised[[:space:]]Learning[[:space:]]for[[:space:]]Medical[[:space:]]Image[[:space:]]Classification/a33fa80f-bfa4-45b2-878d-abb01f3d724e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/T-VSL_[[:space:]]Text-Guided[[:space:]]Visual[[:space:]]Sound[[:space:]]Source[[:space:]]Localization[[:space:]]in[[:space:]]Mixtures/2ac9042c-6511-4205-8b71-c79e7f00a075_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/T4P_[[:space:]]Test-Time[[:space:]]Training[[:space:]]of[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]via[[:space:]]Masked[[:space:]]Autoencoder[[:space:]]and[[:space:]]Actor-specific[[:space:]]Token[[:space:]]Memory/11bec0da-9290-4827-bb16-e062c2aadc7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TACO_[[:space:]]Benchmarking[[:space:]]Generalizable[[:space:]]Bimanual[[:space:]]Tool-ACtion-Object[[:space:]]Understanding/b3378a53-d4f8-49d5-95a4-af3ae029dc3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TAMM_[[:space:]]TriAdapter[[:space:]]Multi-Modal[[:space:]]Learning[[:space:]]for[[:space:]]3D[[:space:]]Shape[[:space:]]Understanding/f474f9b7-e45c-4a08-b76f-0b118ae61d7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TASeg_[[:space:]]Temporal[[:space:]]Aggregation[[:space:]]Network[[:space:]]for[[:space:]]LiDAR[[:space:]]Semantic[[:space:]]Segmentation/745b72e4-2dd6-4811-ade0-f52413f14141_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TCP_Textual-based[[:space:]]Class-aware[[:space:]]Prompt[[:space:]]tuning[[:space:]]for[[:space:]]Visual-Language[[:space:]]Model/137896e1-f0a6-4b80-849c-9a14cbb3250c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TE-TAD_[[:space:]]Towards[[:space:]]Full[[:space:]]End-to-End[[:space:]]Temporal[[:space:]]Action[[:space:]]Detection[[:space:]]via[[:space:]]Time-Aligned[[:space:]]Coordinate[[:space:]]Expression/63fde73a-d557-47b6-aee9-074fa8fa7370_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TEA_[[:space:]]Test-time[[:space:]]Energy[[:space:]]Adaptation/c8c44200-6095-44b9-b188-9639c747d9ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TFMQ-DM_[[:space:]]Temporal[[:space:]]Feature[[:space:]]Maintenance[[:space:]]Quantization[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/a91d56f7-042a-42e5-9c77-ba0dbaac2109_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/THRONE_[[:space:]]An[[:space:]]Object-based[[:space:]]Hallucination[[:space:]]Benchmark[[:space:]]for[[:space:]]the[[:space:]]Free-form[[:space:]]Generations[[:space:]]of[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/da9cf0d5-8d8b-4df8-b396-af5ff1ab4d0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TI2V-Zero_[[:space:]]Zero-Shot[[:space:]]Image[[:space:]]Conditioning[[:space:]]for[[:space:]]Text-to-Video[[:space:]]Diffusion[[:space:]]Models/17566a41-32ff-45d4-9eb1-0099d95d87b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TIGER_[[:space:]]Time-Varying[[:space:]]Denoising[[:space:]]Model[[:space:]]for[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion[[:space:]]Process/8fe5d2c1-4f78-435e-8bde-f7a4afed8604_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TIM_[[:space:]]A[[:space:]]Time[[:space:]]Interval[[:space:]]Machine[[:space:]]for[[:space:]]Audio-Visual[[:space:]]Action[[:space:]]Recognition/03af8456-92c0-4a22-a6d9-94775506b39a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TRINS_[[:space:]]Towards[[:space:]]Multimodal[[:space:]]Language[[:space:]]Models[[:space:]]that[[:space:]]Can[[:space:]]Read/527c7573-3fb6-42ad-a14f-b7b957e97243_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TRIP_[[:space:]]Temporal[[:space:]]Residual[[:space:]]Learning[[:space:]]with[[:space:]]Image[[:space:]]Noise[[:space:]]Prior[[:space:]]for[[:space:]]Image-to-Video[[:space:]]Diffusion[[:space:]]Models/51572829-1348-4002-a40a-f8702f404317_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TTA-EVF_[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Event-based[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation[[:space:]]via[[:space:]]Reliable[[:space:]]Pixel[[:space:]]and[[:space:]]Sample[[:space:]]Estimation/fa811509-bdf9-4c56-97db-551fcc2b0a3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TULIP_[[:space:]]Multi-camera[[:space:]]3D[[:space:]]Precision[[:space:]]Assessment[[:space:]]of[[:space:]]Parkinson's[[:space:]]Disease/532b2907-27ca-48aa-bd4d-3e8291e221ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TULIP_[[:space:]]Transformer[[:space:]]for[[:space:]]Upsampling[[:space:]]of[[:space:]]LiDAR[[:space:]]Point[[:space:]]Clouds/52ac8b22-207c-41ff-adcb-c9f4c82bca08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TUMTraf[[:space:]]V2X[[:space:]]Cooperative[[:space:]]Perception[[:space:]]Dataset/5f2a634b-45a2-4d92-81e8-cda538302a8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Tackling[[:space:]]the[[:space:]]Singularities[[:space:]]at[[:space:]]the[[:space:]]Endpoints[[:space:]]of[[:space:]]Time[[:space:]]Intervals[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/992f5837-40b9-47fe-b9bd-9d2e369d7fdb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Tactile-Augmented[[:space:]]Radiance[[:space:]]Fields/f9e9c348-b82c-415a-b07e-5b945a68c375_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Tailored[[:space:]]Visions_[[:space:]]Enhancing[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]with[[:space:]]Personalized[[:space:]]Prompt[[:space:]]Rewriting/7aad987a-39f2-4096-8aa1-a539bbaacb9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Taming[[:space:]]Mode[[:space:]]Collapse[[:space:]]in[[:space:]]Score[[:space:]]Distillation[[:space:]]for[[:space:]]Text-to-3D[[:space:]]Generation/695a2b86-fc93-420e-a626-d23f402233d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Taming[[:space:]]Self-Training[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detection/08e6cd2d-6b02-4147-b946-b39f423452c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Taming[[:space:]]Stable[[:space:]]Diffusion[[:space:]]for[[:space:]]Text[[:space:]]to[[:space:]]360[[:space:]]Panorama[[:space:]]Image[[:space:]]Generation/3259ff45-5fcb-4059-9736-190d9ce81cb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Taming[[:space:]]the[[:space:]]Tail[[:space:]]in[[:space:]]Class-Conditional[[:space:]]GANs_[[:space:]]Knowledge[[:space:]]Sharing[[:space:]]via[[:space:]]Unconditional[[:space:]]Training[[:space:]]at[[:space:]]Lower[[:space:]]Resolutions/301b4dc8-6f1d-458c-a509-0fe38145fc73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Targeted[[:space:]]Representation[[:space:]]Alignment[[:space:]]for[[:space:]]Open-World[[:space:]]Semi-Supervised[[:space:]]Learning/b33baf4c-8c46-4e13-a308-0177b1e2eee6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Task-Adaptive[[:space:]]Saliency[[:space:]]Guidance[[:space:]]for[[:space:]]Exemplar-free[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning/ef2d4502-c3bc-43a6-a846-89e23e77cec3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Task-Aware[[:space:]]Encoder[[:space:]]Control[[:space:]]for[[:space:]]Deep[[:space:]]Video[[:space:]]Compression/b978be40-5174-4abc-9649-d7daa2d37c89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Task-Conditioned[[:space:]]Adaptation[[:space:]]of[[:space:]]Visual[[:space:]]Features[[:space:]]in[[:space:]]Multi-Task[[:space:]]Policy[[:space:]]Learning/5ebc706d-8b34-4357-9c79-967e39bda2a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Task-Customized[[:space:]]Mixture[[:space:]]of[[:space:]]Adapters[[:space:]]for[[:space:]]General[[:space:]]Image[[:space:]]Fusion/7c9be988-491f-4562-822c-11cfddbdaa7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Task-Driven[[:space:]]Exploration_[[:space:]]Decoupling[[:space:]]and[[:space:]]Inter-Task[[:space:]]Feedback[[:space:]]for[[:space:]]Joint[[:space:]]Moment[[:space:]]Retrieval[[:space:]]and[[:space:]]Highlight[[:space:]]Detection/a6a833ff-f398-48fe-909b-70ff81bf1b94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Task-Driven[[:space:]]Wavelets[[:space:]]using[[:space:]]Constrained[[:space:]]Empirical[[:space:]]Risk[[:space:]]Minimization/d6dc5c43-4c52-4779-8c6a-d513a47a71be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Task-aligned[[:space:]]Part-aware[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]through[[:space:]]Joint[[:space:]]Object-Part[[:space:]]Representations/9dbd5f9f-08af-4d23-b3b6-162e326d318c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Task2Box_[[:space:]]Box[[:space:]]Embeddings[[:space:]]for[[:space:]]Modeling[[:space:]]Asymmetric[[:space:]]Task[[:space:]]Relationships/dd90eec4-a932-4045-a86b-cf06d264eadc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TeMO_[[:space:]]Towards[[:space:]]Text-Driven[[:space:]]3D[[:space:]]Stylization[[:space:]]for[[:space:]]Multi-Object[[:space:]]Meshes/d9ed6877-0911-45bb-b575-e2ba93e2db7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TeTriRF_[[:space:]]Temporal[[:space:]]Tri-Plane[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Efficient[[:space:]]Free-Viewpoint[[:space:]]Video/f5afc893-1338-4f24-aacc-d2b22ab50885_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Teeth-SEG_[[:space:]]An[[:space:]]Efficient[[:space:]]Instance[[:space:]]Segmentation[[:space:]]Framework[[:space:]]for[[:space:]]Orthodontic[[:space:]]Treatment[[:space:]]based[[:space:]]on[[:space:]]Multi-Scale[[:space:]]Aggregation[[:space:]]and[[:space:]]Anthropic[[:space:]]Prior[[:space:]]Knowledge/3aa2a0f8-8a35-4554-8301-990f21126e8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Telling[[:space:]]Left[[:space:]]from[[:space:]]Right_[[:space:]]Identifying[[:space:]]Geometry-Aware[[:space:]]Semantic[[:space:]]Correspondence/8db58ac9-d223-444a-8360-3584791fb40b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Template[[:space:]]Free[[:space:]]Reconstruction[[:space:]]of[[:space:]]Human-object[[:space:]]Interaction[[:space:]]with[[:space:]]Procedural[[:space:]]Interaction[[:space:]]Generation/d155ef3d-1c34-4223-9cdc-aa0f315b2112_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Temporally[[:space:]]Consistent[[:space:]]Unbalanced[[:space:]]Optimal[[:space:]]Transport[[:space:]]for[[:space:]]Unsupervised[[:space:]]Action[[:space:]]Segmentation/a838d212-14b0-4b58-80aa-c43a56860fe1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Depth[[:space:]]Completion/ff74bab9-18a7-4e07-aaa5-ad8163ffcf40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Test-Time[[:space:]]Domain[[:space:]]Generalization[[:space:]]for[[:space:]]Face[[:space:]]Anti-Spoofing/78c2d6b2-0e74-471f-8fda-d5cf6a9dcf82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Test-Time[[:space:]]Linear[[:space:]]Out-of-Distribution[[:space:]]Detection/44ff60bb-a5d5-4ba7-b2d3-4a45ce3a8c75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Test-Time[[:space:]]Zero-Shot[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization/25fb7596-ce79-4f35-b184-c426ff477242_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TetraSphere_[[:space:]]A[[:space:]]Neural[[:space:]]Descriptor[[:space:]]for[[:space:]]O(3)-Invariant[[:space:]]Point[[:space:]]Cloud[[:space:]]Analysis/3abad14e-a1cb-4445-9a59-e083a0427405_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TexOct_[[:space:]]Generating[[:space:]]Textures[[:space:]]of[[:space:]]3D[[:space:]]Models[[:space:]]with[[:space:]]Octree-based[[:space:]]Diffusion/f696ab69-32e4-4a60-813e-243600fb90c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TexTile_[[:space:]]A[[:space:]]Differentiable[[:space:]]Metric[[:space:]]for[[:space:]]Texture[[:space:]]Tileability/d0f4be86-63f8-4c05-9dcd-68da17640c4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TexVocab_[[:space:]]Texture[[:space:]]Vocabulary-conditioned[[:space:]]Human[[:space:]]Avatars/73ddc7b9-7e63-4ba4-be78-2ce01853f19e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text[[:space:]]Grouping[[:space:]]Adapter_[[:space:]]Adapting[[:space:]]Pre-trained[[:space:]]Text[[:space:]]Detector[[:space:]]for[[:space:]]Layout[[:space:]]Analysis/fa84e1e0-b625-49d2-9c9c-ada043e243e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text[[:space:]]Is[[:space:]]MASS_[[:space:]]Modeling[[:space:]]as[[:space:]]Stochastic[[:space:]]Embedding[[:space:]]for[[:space:]]Text-Video[[:space:]]Retrieval/1408fcd0-ca60-4011-8037-14d60f420de4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text[[:space:]]Prompt[[:space:]]with[[:space:]]Normality[[:space:]]Guidance[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/bcf60100-0941-4689-8fcb-a862fa8fb7f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text-Conditioned[[:space:]]Generative[[:space:]]Model[[:space:]]of[[:space:]]3D[[:space:]]Strand-based[[:space:]]Human[[:space:]]Hairstyles/10513fe0-1660-4566-aba6-8db9aa461445_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text-Driven[[:space:]]Image[[:space:]]Editing[[:space:]]via[[:space:]]Learnable[[:space:]]Regions/56deecf0-e61b-46fa-9f80-889d79ae2324_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text-Enhanced[[:space:]]Data-free[[:space:]]Approach[[:space:]]for[[:space:]]Federated[[:space:]]Class-Incremental[[:space:]]Learning/3b220467-d1e6-4c7a-94d4-df3253055a71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text-Guided[[:space:]]3D[[:space:]]Face[[:space:]]Synthesis[[:space:]]-[[:space:]]From[[:space:]]Generation[[:space:]]to[[:space:]]Editing/5cfd0ae8-665a-4459-8664-118bab2a7d0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text-Guided[[:space:]]Variational[[:space:]]Image[[:space:]]Generation[[:space:]]for[[:space:]]Industrial[[:space:]]Anomaly[[:space:]]Detection[[:space:]]and[[:space:]]Segmentation/9e5c99b5-e43b-429f-a15a-c3d64931ca08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text-IF_[[:space:]]Leveraging[[:space:]]Semantic[[:space:]]Text[[:space:]]Guidance[[:space:]]for[[:space:]]Degradation-Aware[[:space:]]and[[:space:]]Interactive[[:space:]]Image[[:space:]]Fusion/0717395c-ac6e-4ead-aecd-81d8a86f4d4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text-Image[[:space:]]Alignment[[:space:]]for[[:space:]]Diffusion-Based[[:space:]]Perception/22f667aa-1467-4844-94ad-5a820a37fffe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text-conditional[[:space:]]Attribute[[:space:]]Alignment[[:space:]]across[[:space:]]Latent[[:space:]]Spaces[[:space:]]for[[:space:]]3D[[:space:]]Controllable[[:space:]]Face[[:space:]]Image[[:space:]]Synthesis/de143cf9-4c5b-4a56-83ec-6ade7727f565_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text-guided[[:space:]]Explorable[[:space:]]Image[[:space:]]Super-resolution/2d3b0ff0-b3f2-4b6c-9506-b3289b41b2ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text-to-3D[[:space:]]Generation[[:space:]]with[[:space:]]Bidirectional[[:space:]]Diffusion[[:space:]]using[[:space:]]both[[:space:]]2D[[:space:]]and[[:space:]]3D[[:space:]]priors/b3b30090-9596-47d8-8960-c9e0604f1cd0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text-to-3D[[:space:]]using[[:space:]]Gaussian[[:space:]]Splatting/0d9d46e6-82d8-4612-a100-ef377f1dd375_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text-to-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]are[[:space:]]Great[[:space:]]Sketch-Photo[[:space:]]Matchmakers/0fe17683-84f4-418f-9875-244c12e64078_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text2HOI_[[:space:]]Text-guided[[:space:]]3D[[:space:]]Motion[[:space:]]Generation[[:space:]]for[[:space:]]Hand-Object[[:space:]]Interaction/c344b2e2-542c-481a-b39e-5e4636e61a03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text2Loc_[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Localization[[:space:]]from[[:space:]]Natural[[:space:]]Language/496b20e7-49a6-417a-85f2-e9133bbfb1a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Text2QR_[[:space:]]Harmonizing[[:space:]]Aesthetic[[:space:]]Customization[[:space:]]and[[:space:]]Scanning[[:space:]]Robustness[[:space:]]for[[:space:]]Text-Guided[[:space:]]QR[[:space:]]Code[[:space:]]Generation/23daf43e-adb8-4577-ae70-e501b1bb9df6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TextCraftor_[[:space:]]Your[[:space:]]Text[[:space:]]Encoder[[:space:]]Can[[:space:]]be[[:space:]]Image[[:space:]]Quality[[:space:]]Controller/f92e9250-3150-4399-9435-c3d8638dd43f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TextNeRF_[[:space:]]A[[:space:]]Novel[[:space:]]Scene-Text[[:space:]]Image[[:space:]]Synthesis[[:space:]]Method[[:space:]]based[[:space:]]on[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/ab5e4563-b756-4686-a12b-6f84aabd1b1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Texture-Preserving[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Virtual[[:space:]]Try-On/bb8d8ebe-5b25-42cc-a93b-7a064e1f37cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TextureDreamer_[[:space:]]Image-Guided[[:space:]]Texture[[:space:]]Synthesis[[:space:]]Through[[:space:]]Geometry-Aware[[:space:]]Diffusion/790b8c09-9c52-4b0c-83e8-5b972ed90f31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/The[[:space:]]Audio-Visual[[:space:]]Conversational[[:space:]]Graph_[[:space:]]From[[:space:]]an[[:space:]]Egocentric-Exocentric[[:space:]]Perspective/a7f6f462-bd9a-401a-99b0-a75ee2ee4aa9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/The[[:space:]]Devil[[:space:]]is[[:space:]]in[[:space:]]the[[:space:]]Details_[[:space:]]StyleFeatureEditor[[:space:]]for[[:space:]]Detail-Rich[[:space:]]StyleGAN[[:space:]]Inversion[[:space:]]and[[:space:]]High[[:space:]]Quality[[:space:]]Image[[:space:]]Editing/3c5e0a8b-535d-4898-a518-4886e69ee42d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/The[[:space:]]Devil[[:space:]]is[[:space:]]in[[:space:]]the[[:space:]]Fine-Grained[[:space:]]Details_[[:space:]]Evaluating[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detectors[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Understanding/40e835f1-e4b1-4773-827c-ca23acc4f744_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/The[[:space:]]Manga[[:space:]]Whisperer_[[:space:]]Automatically[[:space:]]Generating[[:space:]]Transcriptions[[:space:]]for[[:space:]]Comics/56a43b60-6fbe-4f5b-aa81-122982517cb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/The[[:space:]]Mirrored[[:space:]]Influence[[:space:]]Hypothesis_[[:space:]]Efficient[[:space:]]Data[[:space:]]Influence[[:space:]]Estimation[[:space:]]by[[:space:]]Harnessing[[:space:]]Forward[[:space:]]Passes/64d76524-8fcf-4abc-b053-7cdc51376dea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/The[[:space:]]More[[:space:]]You[[:space:]]See[[:space:]]in[[:space:]]2D[[:space:]]the[[:space:]]More[[:space:]]You[[:space:]]Perceive[[:space:]]in[[:space:]]3D/cc9bde4e-a67f-40c4-ba2f-7fe9f1933baf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/The[[:space:]]Neglected[[:space:]]Tails[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models/86dfac67-9c75-4fda-a8e2-089e28e9ab24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/The[[:space:]]STVchrono[[:space:]]Dataset_[[:space:]]Towards[[:space:]]Continuous[[:space:]]Change[[:space:]]Recognition[[:space:]]in[[:space:]]Time/80599102-d6e0-4028-b5a4-3cbca2dbca21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/The[[:space:]]Unreasonable[[:space:]]Effectiveness[[:space:]]of[[:space:]]Pre-Trained[[:space:]]Features[[:space:]]for[[:space:]]Camera[[:space:]]Pose[[:space:]]Refinement/217447c5-1a51-4d0e-894c-9acbfe9ac317_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Theoretically[[:space:]]Achieving[[:space:]]Continuous[[:space:]]Representation[[:space:]]of[[:space:]]Oriented[[:space:]]Bounding[[:space:]]Boxes/086c0dd8-a25e-43bb-9a6e-4faeedc799f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Think[[:space:]]Twice[[:space:]]Before[[:space:]]Selection_[[:space:]]Federated[[:space:]]Evidential[[:space:]]Active[[:space:]]Learning[[:space:]]for[[:space:]]Medical[[:space:]]Image[[:space:]]Analysis[[:space:]]with[[:space:]]Domain[[:space:]]Shifts/754b1538-5256-4f77-a634-03eee7003902_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Three[[:space:]]Pillars[[:space:]]Improving[[:space:]]Vision[[:space:]]Foundation[[:space:]]Model[[:space:]]Distillation[[:space:]]for[[:space:]]Lidar/30a3d349-5af7-4e6f-ba29-727cf30b74f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TiNO-Edit_[[:space:]]Timestep[[:space:]]and[[:space:]]Noise[[:space:]]Optimization[[:space:]]for[[:space:]]Robust[[:space:]]Diffusion-Based[[:space:]]Image[[:space:]]Editing/2cb307c9-e2a5-4b55-9280-3c0b58fd69cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Time-[[:space:]]Memory-[[:space:]]and[[:space:]]Parameter-Efficient[[:space:]]Visual[[:space:]]Adaptation/b662c765-b1d6-49f2-aee1-22575401725b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Time-Efficient[[:space:]]Light-Field[[:space:]]Acquisition[[:space:]]Using[[:space:]]Coded[[:space:]]Aperture[[:space:]]and[[:space:]]Events/caa9f08c-90f5-4a33-9513-626b0d260c42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TimeChat_[[:space:]]A[[:space:]]Time-sensitive[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Long[[:space:]]Video[[:space:]]Understanding/65fdd545-a868-4520-87b6-eb81602ef0c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ToNNO_[[:space:]]Tomographic[[:space:]]Reconstruction[[:space:]]of[[:space:]]a[[:space:]]Neural[[:space:]]Network's[[:space:]]Output[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Segmentation[[:space:]]of[[:space:]]3D[[:space:]]Medical[[:space:]]Images/d04521a0-cc99-48c4-9f59-72021aded526_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Token[[:space:]]Transformation[[:space:]]Matters_[[:space:]]Towards[[:space:]]Faithful[[:space:]]Post-hoc[[:space:]]Explanation[[:space:]]for[[:space:]]Vision[[:space:]]Transformer/ed35a9a2-b80a-4823-b600-4f656f3f403b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TokenCompose_[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]with[[:space:]]Token-level[[:space:]]Supervision/7285fcb9-ba63-4424-a5f0-bb564f5971f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TokenHMR_[[:space:]]Advancing[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery[[:space:]]with[[:space:]]a[[:space:]]Tokenized[[:space:]]Pose[[:space:]]Representation/39771545-28e9-489d-ba60-f6fe27b3c143_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ToonerGAN_[[:space:]]Reinforcing[[:space:]]GANs[[:space:]]for[[:space:]]Obfuscating[[:space:]]Automated[[:space:]]Facial[[:space:]]Indexing/70e9a913-ba04-4183-a9f3-5de395c38caa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Total[[:space:]]Selfie_[[:space:]]Generating[[:space:]]Full-Body[[:space:]]Selfies/9646b125-1bdd-43e3-8e66-a76d835ea69b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Total-Decom_[[:space:]]Decomposed[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]with[[:space:]]Minimal[[:space:]]Interaction/52631292-8321-4991-bfed-87e7d2d7ae86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Toward[[:space:]]Generalist[[:space:]]Anomaly[[:space:]]Detection[[:space:]]via[[:space:]]In-context[[:space:]]Residual[[:space:]]Learning[[:space:]]with[[:space:]]Few-shot[[:space:]]Sample[[:space:]]Prompts/0241f53b-e08c-48bf-b85b-28bb70b55d57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]3D[[:space:]]Vision[[:space:]]with[[:space:]]Low-Cost[[:space:]]Single-Photon[[:space:]]Cameras/08870c78-fed0-4c1c-bea3-892f40c73641_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Accurate[[:space:]]Post-training[[:space:]]Quantization[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/dade5ef4-b4d8-476b-a8d6-fcaea41fce4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Accurate[[:space:]]and[[:space:]]Robust[[:space:]]Architectures[[:space:]]via[[:space:]]Neural[[:space:]]Architecture[[:space:]]Search/7a054c74-76ec-4573-a03c-5275d463e7ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Automated[[:space:]]Movie[[:space:]]Trailer[[:space:]]Generation/1fa89d41-e617-4d82-83b6-82c144b17441_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Automatic[[:space:]]Power[[:space:]]Battery[[:space:]]Detection_[[:space:]]New[[:space:]]Challenge[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]and[[:space:]]Baseline/8fb182f9-128f-4927-aa5f-bec84926452c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Backward-Compatible[[:space:]]Continual[[:space:]]Learning[[:space:]]of[[:space:]]Image[[:space:]]Compression/74b3c982-6c92-4d85-be5f-ac544847ec98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Better[[:space:]]Vision-Inspired[[:space:]]Vision-Language[[:space:]]Models/abd49d48-e2ba-4cb9-a925-03bd1f254561_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]CLIP-driven[[:space:]]Language-free[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding[[:space:]]via[[:space:]]2D-3D[[:space:]]Relational[[:space:]]Enhancement[[:space:]]and[[:space:]]Consistency/bac9a679-cf7e-46ee-9e36-43a5fb36fc9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Calibrated[[:space:]]Multi-label[[:space:]]Deep[[:space:]]Neural[[:space:]]Networks/b864a370-5732-4c9a-8a45-07396e4880c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Co-Evaluation[[:space:]]of[[:space:]]Cameras[[:space:]]HDR[[:space:]]and[[:space:]]Algorithms[[:space:]]for[[:space:]]Industrial-Grade[[:space:]]6DoF[[:space:]]Pose[[:space:]]Estimation/db7b4dc1-e676-4ee1-a481-e96181f20987_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Effective[[:space:]]Usage[[:space:]]of[[:space:]]Human-Centric[[:space:]]Priors[[:space:]]in[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Text-based[[:space:]]Human[[:space:]]Image[[:space:]]Generation/0011defb-8b77-4228-b955-c563e68c36bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Efficient[[:space:]]Replay[[:space:]]in[[:space:]]Federated[[:space:]]Incremental[[:space:]]Learning/bb3a6a6c-80c7-49d7-95b4-2dda0dc68f88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Fairness-Aware[[:space:]]Adversarial[[:space:]]Learning/3084c0d3-cbcf-49cd-9968-259a3c508df7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]General[[:space:]]Robustness[[:space:]]Verification[[:space:]]of[[:space:]]MaxPool-based[[:space:]]Convolutional[[:space:]]Neural[[:space:]]Networks[[:space:]]via[[:space:]]Tightening[[:space:]]Linear[[:space:]]Approximation/3e490d9c-0529-4a45-81aa-e9c6779604bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Generalizable[[:space:]]Multi-Object[[:space:]]Tracking/569c7951-2995-400e-bd13-dea15348a6bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Generalizable[[:space:]]Tumor[[:space:]]Synthesis/524ca3b8-4957-4efa-8520-bc16a901505e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Generalizing[[:space:]]to[[:space:]]Unseen[[:space:]]Domains[[:space:]]with[[:space:]]Few[[:space:]]Labels/e55fb2c1-e40f-448c-b25a-6b2ab30e5cea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]HDR[[:space:]]and[[:space:]]HFR[[:space:]]Video[[:space:]]from[[:space:]]Rolling-Mixed-Bit[[:space:]]Spikings/09b152e9-220f-4d98-84ad-755a808ff1ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]High-fidelity[[:space:]]Artistic[[:space:]]Image[[:space:]]Vectorization[[:space:]]via[[:space:]]Texture-Encapsulated[[:space:]]Shape[[:space:]]Parameterization/cc5a6804-22ec-4bc9-989b-454b2b4fab56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Language-Driven[[:space:]]Video[[:space:]]Inpainting[[:space:]]via[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/44892258-c2fd-4875-a381-36cb388386e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Large-scale[[:space:]]3D[[:space:]]Representation[[:space:]]Learning[[:space:]]with[[:space:]]Multi-dataset[[:space:]]Point[[:space:]]Prompt[[:space:]]Training/169b8810-12da-48d8-8f1b-50283a934430_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Learning[[:space:]]a[[:space:]]Generalist[[:space:]]Model[[:space:]]for[[:space:]]Embodied[[:space:]]Navigation/c74dddc3-29c3-4944-bf6f-86cfb9668f71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Memorization-Free[[:space:]]Diffusion[[:space:]]Models/7d5eed98-f2f8-44f9-a87b-0e61db88cf9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Modern[[:space:]]Image[[:space:]]Manipulation[[:space:]]Localization_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]and[[:space:]]Novel[[:space:]]Methods/a71e0d7b-32cb-4bcb-9bc8-91174ff8ec7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]More[[:space:]]Accurate[[:space:]]Diffusion[[:space:]]Model[[:space:]]Acceleration[[:space:]]with[[:space:]]A[[:space:]]Timestep[[:space:]]Tuner/9f995471-4494-42e5-9d77-87ab033f6b8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]More[[:space:]]Unified[[:space:]]In-context[[:space:]]Visual[[:space:]]Understanding/70266367-585a-463c-a356-ff1367e593cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Progressive[[:space:]]Multi-Frequency[[:space:]]Representation[[:space:]]for[[:space:]]Image[[:space:]]Warping/de6898a9-404c-436c-8603-635d013aa16a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Real-World[[:space:]]HDR[[:space:]]Video[[:space:]]Reconstruction_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]and[[:space:]]A[[:space:]]Two-Stage[[:space:]]Alignment[[:space:]]Network/2e56d1f1-279c-49a4-9e69-e2acdc995377_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Realistic[[:space:]]Scene[[:space:]]Generation[[:space:]]with[[:space:]]LiDAR[[:space:]]Diffusion[[:space:]]Models/d8e671e6-51a7-4894-bfa7-9c2030e60478_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Robust[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]LiDAR[[:space:]]and[[:space:]]4D[[:space:]]Radar[[:space:]]Fusion[[:space:]]in[[:space:]]Various[[:space:]]Weather[[:space:]]Conditions/ba348142-930f-4907-9a74-13eabbb6ee87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Robust[[:space:]]3D[[:space:]]Pose[[:space:]]Transfer[[:space:]]with[[:space:]]Adversarial[[:space:]]Learning/5cf5f091-0a4b-4124-b039-76fd65b97fd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Robust[[:space:]]Event-guided[[:space:]]Low-Light[[:space:]]Image[[:space:]]Enhancement_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Real-World[[:space:]]Event-Image[[:space:]]Dataset[[:space:]]and[[:space:]]Novel[[:space:]]Approach/2df51050-cd07-4a7e-81f7-f34bc968893b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Robust[[:space:]]Learning[[:space:]]to[[:space:]]Optimize[[:space:]]with[[:space:]]Theoretical[[:space:]]Guarantees/3a42d29b-f2d3-43f4-aa8a-7a5bdcd7e480_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Scalable[[:space:]]3D[[:space:]]Anomaly[[:space:]]Detection[[:space:]]and[[:space:]]Localization_[[:space:]]A[[:space:]]Benchmark[[:space:]]via[[:space:]]3D[[:space:]]Anomaly[[:space:]]Synthesis[[:space:]]and[[:space:]]A[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]Network/ca90097c-8c70-4290-ab2d-8fc664c27dfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Surveillance[[:space:]]Video-and-Language[[:space:]]Understanding_[[:space:]]New[[:space:]]Dataset[[:space:]]Baselines[[:space:]]and[[:space:]]Challenges/0284743e-f60d-4950-8608-db2d06fbd223_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Text-guided[[:space:]]3D[[:space:]]Scene[[:space:]]Composition/3bf490cf-98df-41a1-9296-56fd8fa5c186_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Transferable[[:space:]]Targeted[[:space:]]3D[[:space:]]Adversarial[[:space:]]Attack[[:space:]]in[[:space:]]the[[:space:]]Physical[[:space:]]World/4ca6eb7e-4601-4426-a03e-76ac78cf9273_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Understanding[[:space:]]Cross[[:space:]]and[[:space:]]Self-Attention[[:space:]]in[[:space:]]Stable[[:space:]]Diffusion[[:space:]]for[[:space:]]Text-Guided[[:space:]]Image[[:space:]]Editing/7f15f446-d3c6-4102-890c-26e8ef7e15e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Understanding[[:space:]]and[[:space:]]Improving[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]of[[:space:]]Vision[[:space:]]Transformers/cc01d971-439b-4657-93f0-3693e9c04c8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]Variable[[:space:]]and[[:space:]]Coordinated[[:space:]]Holistic[[:space:]]Co-Speech[[:space:]]Motion[[:space:]]Generation/3f5ab241-bf07-4f57-9ac9-e2c396f6caea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]a[[:space:]]Perceptual[[:space:]]Evaluation[[:space:]]Framework[[:space:]]for[[:space:]]Lighting[[:space:]]Estimation/6bd613d9-d64b-47fe-801e-5ea623485975_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]a[[:space:]]Simultaneous[[:space:]]and[[:space:]]Granular[[:space:]]Identity-Expression[[:space:]]Control[[:space:]]in[[:space:]]Personalized[[:space:]]Face[[:space:]]Generation/ef35ee72-946d-4566-bc04-839c97ba94f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Towards[[:space:]]the[[:space:]]Uncharted_[[:space:]]Density-Descending[[:space:]]Feature[[:space:]]Perturbation[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Semantic[[:space:]]Segmentation/ac475561-048f-43d3-a92f-b7eb58baa4f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Traceable[[:space:]]Federated[[:space:]]Continual[[:space:]]Learning/5f927d5a-3694-4746-b3f1-40498f421c3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Traffic[[:space:]]Scene[[:space:]]Parsing[[:space:]]through[[:space:]]the[[:space:]]TSP6K[[:space:]]Dataset/dda2def5-9ca2-41ec-9dbb-0e534fa7ed9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Training[[:space:]]Diffusion[[:space:]]Models[[:space:]]Towards[[:space:]]Diverse[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Reinforcement[[:space:]]Learning/cd040198-6d97-4201-b0f8-493b507296ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Training[[:space:]]Generative[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]Models[[:space:]]by[[:space:]]Wavelet-Domain[[:space:]]Losses[[:space:]]Enables[[:space:]]Better[[:space:]]Control[[:space:]]of[[:space:]]Artifacts/3738de51-5ddc-4d0e-8674-020af097e81f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Training[[:space:]]Like[[:space:]]a[[:space:]]Medical[[:space:]]Resident_[[:space:]]Context-Prior[[:space:]]Learning[[:space:]]Toward[[:space:]]Universal[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/b7bfb25e-a5eb-4084-a48e-d8d30ad63be8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Training[[:space:]]Vision[[:space:]]Transformers[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation/2b3d8b4c-2a07-413f-896a-ad8ca3463289_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Training-Free[[:space:]]Open-Vocabulary[[:space:]]Segmentation[[:space:]]with[[:space:]]Offline[[:space:]]Diffusion-Augmented[[:space:]]Prototype[[:space:]]Generation/18ccf3d7-544d-4617-a3fe-92abe39c0ea7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Training-Free[[:space:]]Pretrained[[:space:]]Model[[:space:]]Merging/4c95b517-09ab-4eca-9bb9-2dc8a7d90c28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TransLoc4D_[[:space:]]Transformer-based[[:space:]]4D[[:space:]]Radar[[:space:]]Place[[:space:]]Recognition/fdc4ab9f-0b82-4dab-a93a-7c916b4aaa73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TransNeXt_[[:space:]]Robust[[:space:]]Foveal[[:space:]]Visual[[:space:]]Perception[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/d540c455-5ac4-4f9b-b9c1-a1a77d4db23c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Transcending[[:space:]]Forgery[[:space:]]Specificity[[:space:]]with[[:space:]]Latent[[:space:]]Space[[:space:]]Augmentation[[:space:]]for[[:space:]]Generalizable[[:space:]]Deepfake[[:space:]]Detection/5a1b86c5-1aec-44cc-a768-22aeb3c488f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Transcending[[:space:]]the[[:space:]]Limit[[:space:]]of[[:space:]]Local[[:space:]]Window_[[:space:]]Advanced[[:space:]]Super-Resolution[[:space:]]Transformer[[:space:]]with[[:space:]]Adaptive[[:space:]]Token[[:space:]]Dictionary/390926ab-4318-47ac-acc2-a772a561c7fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Transcriptomics-guided[[:space:]]Slide[[:space:]]Representation[[:space:]]Learning[[:space:]]in[[:space:]]Computational[[:space:]]Pathology/8f99541a-e04d-41e4-a652-7cee09af0b56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Transductive[[:space:]]Zero-Shot[[:space:]]and[[:space:]]Few-Shot[[:space:]]CLIP/87014371-2b40-48e4-85bb-595bf0e65c5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Transfer[[:space:]]CLIP[[:space:]]for[[:space:]]Generalizable[[:space:]]Image[[:space:]]Denoising/eaf50ab1-e4b8-4759-ab42-2d059746ee59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Transferable[[:space:]]Structural[[:space:]]Sparse[[:space:]]Adversarial[[:space:]]Attack[[:space:]]Via[[:space:]]Exact[[:space:]]Group[[:space:]]Sparsity[[:space:]]Training/1f6c96cd-34ea-49e5-8a97-00e3ab43fae5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Transferable[[:space:]]and[[:space:]]Principled[[:space:]]Efficiency[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Segmentation/a2a92b4e-0e5b-4469-a60d-fb6f9ef1005c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Tri-Modal[[:space:]]Motion[[:space:]]Retrieval[[:space:]]by[[:space:]]Learning[[:space:]]a[[:space:]]Joint[[:space:]]Embedding[[:space:]]Space/9a1db899-954d-4829-8ca2-b2ec91608dd1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Tri-Perspective[[:space:]]View[[:space:]]Decomposition[[:space:]]for[[:space:]]Geometry-Aware[[:space:]]Depth[[:space:]]Completion/2a306b69-40d9-408c-9975-e38d8af7da2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Triplane[[:space:]]Meets[[:space:]]Gaussian[[:space:]]Splatting_[[:space:]]Fast[[:space:]]and[[:space:]]Generalizable[[:space:]]Single-View[[:space:]]3D[[:space:]]Reconstruction[[:space:]]with[[:space:]]Transformers/39e71c8f-3ea2-4f6d-8eb4-649df6214416_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Troika_[[:space:]]Multi-Path[[:space:]]Cross-Modal[[:space:]]Traction[[:space:]]for[[:space:]]Compositional[[:space:]]Zero-Shot[[:space:]]Learning/49a1785f-304e-4553-b7d4-f0f460cfda9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Tumor[[:space:]]Micro-environment[[:space:]]Interactions[[:space:]]Guided[[:space:]]Graph[[:space:]]Learning[[:space:]]for[[:space:]]Survival[[:space:]]Analysis[[:space:]]of[[:space:]]Human[[:space:]]Cancers[[:space:]]from[[:space:]]Whole-slide[[:space:]]Pathological[[:space:]]Images/659ef302-e6fe-4a53-8ffe-734b71fb2cc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Tune-An-Ellipse_[[:space:]]CLIP[[:space:]]Has[[:space:]]Potential[[:space:]]to[[:space:]]Find[[:space:]]What[[:space:]]You[[:space:]]Want/46ae7dc6-b313-454f-988d-a96f6627240a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Tuning[[:space:]]Stable[[:space:]]Rank[[:space:]]Shrinkage_[[:space:]]Aiming[[:space:]]at[[:space:]]the[[:space:]]Overlooked[[:space:]]Structural[[:space:]]Risk[[:space:]]in[[:space:]]Fine-tuning/c90a1a3d-3e81-47b2-b7ce-bb3ce1716d0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Turb-Seg-Res_[[:space:]]A[[:space:]]Segment-then-Restore[[:space:]]Pipeline[[:space:]]for[[:space:]]Dynamic[[:space:]]Videos[[:space:]]with[[:space:]]Atmospheric[[:space:]]Turbulence/3c042daa-8743-49f7-8d06-1fa252593a9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TurboSL_[[:space:]]Dense[[:space:]]Accurate[[:space:]]and[[:space:]]Fast[[:space:]]3D[[:space:]]by[[:space:]]Neural[[:space:]]Inverse[[:space:]]Structured[[:space:]]Light/201a4818-9f41-488b-91b6-f44cd4367e6c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/TutteNet_[[:space:]]Injective[[:space:]]3D[[:space:]]Deformations[[:space:]]by[[:space:]]Composition[[:space:]]of[[:space:]]2D[[:space:]]Mesh[[:space:]]Deformations/171684f5-4564-4856-90a0-39940ea72d26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Tyche_[[:space:]]Stochastic[[:space:]]In-Context[[:space:]]Learning[[:space:]]for[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/5678ce7f-c335-45ae-8836-2e088b76f1be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/U-VAP_[[:space:]]User-specified[[:space:]]Visual[[:space:]]Appearance[[:space:]]Personalization[[:space:]]via[[:space:]]Decoupled[[:space:]]Self[[:space:]]Augmentation/1eac2f90-6cbb-4235-b87c-db32b64a1087_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UDiFF_[[:space:]]Generating[[:space:]]Conditional[[:space:]]Unsigned[[:space:]]Distance[[:space:]]Fields[[:space:]]with[[:space:]]Optimal[[:space:]]Wavelet[[:space:]]Diffusion/9987b407-9218-49df-a963-1a180e537d6b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UFC-Net_[[:space:]]Unrolling[[:space:]]Fixed-point[[:space:]]Continuous[[:space:]]Network[[:space:]]for[[:space:]]Deep[[:space:]]Compressive[[:space:]]Sensing/189aa769-4c51-4d78-8325-66da7ea04e07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UFOGen_[[:space:]]You[[:space:]]Forward[[:space:]]Once[[:space:]]Large[[:space:]]Scale[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]via[[:space:]]Diffusion[[:space:]]GANs/7376c247-c308-463f-91be-e0bd00db162b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UFORecon_[[:space:]]Generalizable[[:space:]]Sparse-View[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]from[[:space:]]Arbitrary[[:space:]]and[[:space:]]Unfavorable[[:space:]]Sets/a3611619-17bd-48a7-88a9-a732beffde9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UFineBench_[[:space:]]Towards[[:space:]]Text-based[[:space:]]Person[[:space:]]Retrieval[[:space:]]with[[:space:]]Ultra-fine[[:space:]]Granularity/16ef4b39-0850-464e-80b9-7b082912587f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ULIP-2_[[:space:]]Towards[[:space:]]Scalable[[:space:]]Multimodal[[:space:]]Pre-training[[:space:]]for[[:space:]]3D[[:space:]]Understanding/d99a25fc-3fce-431b-95f0-08b191dae28b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/URHand_[[:space:]]Universal[[:space:]]Relightable[[:space:]]Hands/392d1d56-0684-44db-b271-929ebc0dece0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/USE_[[:space:]]Universal[[:space:]]Segment[[:space:]]Embeddings[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Image[[:space:]]Segmentation/22025dc1-7043-49ad-ad06-5d56dd390204_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UV-IDM_[[:space:]]Identity-Conditioned[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Face[[:space:]]UV-Texture[[:space:]]Generation/6bfbc35e-fdfe-42dd-8242-bbd15ee14d93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UVEB_[[:space:]]A[[:space:]]Large-scale[[:space:]]Benchmark[[:space:]]and[[:space:]]Baseline[[:space:]]Towards[[:space:]]Real-World[[:space:]]Underwater[[:space:]]Video[[:space:]]Enhancement/03a8f120-6795-4a04-bdca-749e8e90db1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UltrAvatar_[[:space:]]A[[:space:]]Realistic[[:space:]]Animatable[[:space:]]3D[[:space:]]Avatar[[:space:]]Diffusion[[:space:]]Model[[:space:]]with[[:space:]]Authenticity[[:space:]]Guided[[:space:]]Textures/2bdd41de-4b99-48e0-b3d6-edc4a6af93d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UnO_[[:space:]]Unsupervised[[:space:]]Occupancy[[:space:]]Fields[[:space:]]for[[:space:]]Perception[[:space:]]and[[:space:]]Forecasting/b4327c20-ff9c-4048-9751-821018fbe5c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UnSAMFlow_[[:space:]]Unsupervised[[:space:]]Optical[[:space:]]Flow[[:space:]]Guided[[:space:]]by[[:space:]]Segment[[:space:]]Anything[[:space:]]Model/4183af4e-16a8-43ba-b470-37c5f3c02313_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UnScene3D_[[:space:]]Unsupervised[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation[[:space:]]for[[:space:]]Indoor[[:space:]]Scenes/3e761d39-4248-43da-9d07-08fbe5e6ac89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unbiased[[:space:]]Estimator[[:space:]]for[[:space:]]Distorted[[:space:]]Conics[[:space:]]in[[:space:]]Camera[[:space:]]Calibration/dad42e0c-fedf-4b2a-920e-db0f6c3a68f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unbiased[[:space:]]Faster[[:space:]]R-CNN[[:space:]]for[[:space:]]Single-source[[:space:]]Domain[[:space:]]Generalized[[:space:]]Object[[:space:]]Detection/6afb7836-796a-4d2a-b167-ea9a9f2927eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Uncertainty[[:space:]]Visualization[[:space:]]via[[:space:]]Low-Dimensional[[:space:]]Posterior[[:space:]]Projections/7bde1d6f-b4d8-4449-999c-1ce9d06d3c2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Uncertainty-Aware[[:space:]]Source-Free[[:space:]]Adaptive[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]with[[:space:]]Wavelet[[:space:]]Augmentation[[:space:]]Transformer/bd523033-0e15-43ac-bf7a-d7589df4b3a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Uncertainty-Guided[[:space:]]Never-Ending[[:space:]]Learning[[:space:]]to[[:space:]]Drive/4aa34121-12f3-4a5e-8d56-4197483b031c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Uncertainty-aware[[:space:]]Action[[:space:]]Decoupling[[:space:]]Transformer[[:space:]]for[[:space:]]Action[[:space:]]Anticipation/1e914fe3-c5ea-4f2f-ac83-a42005517ba2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Uncovering[[:space:]]What[[:space:]]Why[[:space:]]and[[:space:]]How_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Benchmark[[:space:]]for[[:space:]]Causation[[:space:]]Understanding[[:space:]]of[[:space:]]Video[[:space:]]Anomaly/8d3bd149-4866-4c7e-8a8d-b0563fea3339_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Understanding[[:space:]]Video[[:space:]]Transformers[[:space:]]via[[:space:]]Universal[[:space:]]Concept[[:space:]]Discovery/61573633-d6ff-47e9-8ab8-495cb916d006_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Understanding[[:space:]]and[[:space:]]Improving[[:space:]]Source-free[[:space:]]Domain[[:space:]]Adaptation[[:space:]]from[[:space:]]a[[:space:]]Theoretical[[:space:]]Perspective/261e8253-2667-4e03-ae25-9e617fd0ac56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unexplored[[:space:]]Faces[[:space:]]of[[:space:]]Robustness[[:space:]]and[[:space:]]Out-of-Distribution_[[:space:]]Covariate[[:space:]]Shifts[[:space:]]in[[:space:]]Environment[[:space:]]and[[:space:]]Sensor[[:space:]]Domains/357d3542-01a0-43a8-82f6-3628eaa278fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Ungeneralizable[[:space:]]Examples/2dc654c4-1039-4579-ad1a-3e9e824887f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UniBind_[[:space:]]LLM-Augmented[[:space:]]Unified[[:space:]]and[[:space:]]Balanced[[:space:]]Representation[[:space:]]Space[[:space:]]to[[:space:]]Bind[[:space:]]Them[[:space:]]All/a88745ee-4af1-4a95-9de2-24e990e5d9c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UniDepth_[[:space:]]Universal[[:space:]]Monocular[[:space:]]Metric[[:space:]]Depth[[:space:]]Estimation/8d7d3d75-b65e-4a58-bf63-0841425e5e1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UniGS_[[:space:]]Unified[[:space:]]Representation[[:space:]]for[[:space:]]Image[[:space:]]Generation[[:space:]]and[[:space:]]Segmentation/d0457095-88ad-4ed2-ba44-2e5e3637b855_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UniGarmentManip_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Category-Level[[:space:]]Garment[[:space:]]Manipulation[[:space:]]via[[:space:]]Dense[[:space:]]Visual[[:space:]]Correspondence/89d34176-8bdd-4f9c-8ec6-dabc92c23713_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UniHuman_[[:space:]]A[[:space:]]Unified[[:space:]]Model[[:space:]]For[[:space:]]Editing[[:space:]]Human[[:space:]]Images[[:space:]]in[[:space:]]the[[:space:]]Wild/aa5a6a3c-43d1-480d-8d08-ec0c10fe8f5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UniMODE_[[:space:]]Unified[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection/590bd3ec-d564-41bb-9cc2-3017e9fe1be4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UniMix_[[:space:]]Towards[[:space:]]Domain[[:space:]]Adaptive[[:space:]]and[[:space:]]Generalizable[[:space:]]LiDAR[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]in[[:space:]]Adverse[[:space:]]Weather/8f1080de-63f3-4cc4-895f-b20fe0a76a8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UniPAD_[[:space:]]A[[:space:]]Universal[[:space:]]Pre-training[[:space:]]Paradigm[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/5d5791c9-388c-45da-b0a9-0f8ac81d4eed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UniPTS_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Proficient[[:space:]]Post-Training[[:space:]]Sparsity/d2985861-4a0e-450f-9d43-973da89f0537_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UniPT_[[:space:]]Universal[[:space:]]Parallel[[:space:]]Tuning[[:space:]]for[[:space:]]Transfer[[:space:]]Learning[[:space:]]with[[:space:]]Efficient[[:space:]]Parameter[[:space:]]and[[:space:]]Memory/3cb7255b-0fd3-4557-9efe-e35463258eae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UniRepLKNet_[[:space:]]A[[:space:]]Universal[[:space:]]Perception[[:space:]]Large-Kernel[[:space:]]ConvNet[[:space:]]for[[:space:]]Audio[[:space:]]Video[[:space:]]Point[[:space:]]Cloud[[:space:]]Time-Series[[:space:]]and[[:space:]]Image[[:space:]]Recognition/a1844c0e-581e-46dc-aac5-fbcaaf39b2de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UniVS_[[:space:]]Unified[[:space:]]and[[:space:]]Universal[[:space:]]Video[[:space:]]Segmentation[[:space:]]with[[:space:]]Prompts[[:space:]]as[[:space:]]Queries/82f82719-d8c2-450e-a76e-a49750b4f372_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unified[[:space:]]Entropy[[:space:]]Optimization[[:space:]]for[[:space:]]Open-Set[[:space:]]Test-Time[[:space:]]Adaptation/0afbbd0f-78e3-4114-88ec-421304ec7f2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unified[[:space:]]Language-driven[[:space:]]Zero-shot[[:space:]]Domain[[:space:]]Adaptation/9b9242e3-7d1e-48c1-83b7-9d9d60dad0d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unified-IO[[:space:]]2_[[:space:]]Scaling[[:space:]]Autoregressive[[:space:]]Multimodal[[:space:]]Models[[:space:]]with[[:space:]]Vision[[:space:]]Language[[:space:]]Audio[[:space:]]and[[:space:]]Action/6d3fc5e0-8b71-4ea2-955f-cfe27f84d10f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unifying[[:space:]]Automatic[[:space:]]and[[:space:]]Interactive[[:space:]]Matting[[:space:]]with[[:space:]]Pretrained[[:space:]]ViTs/bd988ad1-b8d4-48ee-972a-da7ca8571a55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unifying[[:space:]]Correspondence[[:space:]]Pose[[:space:]]and[[:space:]]NeRF[[:space:]]for[[:space:]]Generalized[[:space:]]Pose-Free[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/e1103081-8cc9-4bba-bf35-8cf726fe0c82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unifying[[:space:]]Top-down[[:space:]]and[[:space:]]Bottom-up[[:space:]]Scanpath[[:space:]]Prediction[[:space:]]Using[[:space:]]Transformers/4f3692b4-b112-4c08-ba9a-2f36fb743bd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/UnionFormer_[[:space:]]Unified-Learning[[:space:]]Transformer[[:space:]]with[[:space:]]Multi-View[[:space:]]Representation[[:space:]]for[[:space:]]Image[[:space:]]Manipulation[[:space:]]Detection[[:space:]]and[[:space:]]Localization/5ef09fc0-0335-496d-aad7-a5c4f99be369_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Universal[[:space:]]Novelty[[:space:]]Detection[[:space:]]Through[[:space:]]Adaptive[[:space:]]Contrastive[[:space:]]Learning/b0ab81ba-4899-418b-a884-078e483078e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Universal[[:space:]]Robustness[[:space:]]via[[:space:]]Median[[:space:]]Randomized[[:space:]]Smoothing[[:space:]]for[[:space:]]Real-World[[:space:]]Super-Resolution/e2d4cb6b-3a69-48fe-82d6-33bc784fee76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Universal[[:space:]]Segmentation[[:space:]]at[[:space:]]Arbitrary[[:space:]]Granularity[[:space:]]with[[:space:]]Language[[:space:]]Instruction/ce064b13-7e7f-46d5-bd1a-128eb726a222_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Universal[[:space:]]Semi-Supervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]by[[:space:]]Mitigating[[:space:]]Common-Class[[:space:]]Bias/dbb2f9ec-31ed-4372-a511-c26ea3cc7701_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unknown[[:space:]]Prompt[[:space:]]the[[:space:]]only[[:space:]]Lacuna_[[:space:]]Unveiling[[:space:]]CLIP's[[:space:]]Potential[[:space:]]for[[:space:]]Open[[:space:]]Domain[[:space:]]Generalization/743c1b94-3132-4682-927f-fa8ba0b0e67f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unleashing[[:space:]]Channel[[:space:]]Potential_[[:space:]]Space-Frequency[[:space:]]Selection[[:space:]]Convolution[[:space:]]for[[:space:]]SAR[[:space:]]Object[[:space:]]Detection/48b414e5-8fa3-4981-89d2-e1dc18efde8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unleashing[[:space:]]Network[[:space:]]Potentials[[:space:]]for[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion/fb31f85b-8134-45b7-af6e-f9c7b9c72ffc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unleashing[[:space:]]Unlabeled[[:space:]]Data_[[:space:]]A[[:space:]]Paradigm[[:space:]]for[[:space:]]Cross-View[[:space:]]Geo-Localization/a294a1f9-a695-46b7-9e2b-607f6b52a577_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unleashing[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]SAM[[:space:]]for[[:space:]]Medical[[:space:]]Adaptation[[:space:]]via[[:space:]]Hierarchical[[:space:]]Decoding/bbaf5bae-9b3e-4faf-a1cc-954c02312086_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unlocking[[:space:]]Pre-trained[[:space:]]Image[[:space:]]Backbones[[:space:]]for[[:space:]]Semantic[[:space:]]Image[[:space:]]Synthesis/4ac46fe0-6e02-477c-b2c1-461ade0973bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unlocking[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Pre-trained[[:space:]]Vision[[:space:]]Transformers[[:space:]]for[[:space:]]Few-Shot[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]through[[:space:]]Relationship[[:space:]]Descriptors/e94cac75-6ec7-4302-90b5-969b1f8f981e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unlocking[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Prompt-Tuning[[:space:]]in[[:space:]]Bridging[[:space:]]Generalized[[:space:]]and[[:space:]]Personalized[[:space:]]Federated[[:space:]]Learning/ea9edf79-ec18-40b2-8464-e9c33ab1758f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unmixing[[:space:]]Before[[:space:]]Fusion_[[:space:]]A[[:space:]]Generalized[[:space:]]Paradigm[[:space:]]for[[:space:]]Multi-Source-based[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Synthesis/9e5c429d-6068-4bcb-8001-4a8c329c75ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unmixing[[:space:]]Diffusion[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Denoising/797682c3-bf3d-4c04-99c3-f9f248ca0b5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unraveling[[:space:]]Instance[[:space:]]Associations_[[:space:]]A[[:space:]]Closer[[:space:]]Look[[:space:]]for[[:space:]]Audio-Visual[[:space:]]Segmentation/bbcd7ad3-823f-4a3e-8587-9fa89b871675_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsegment[[:space:]]Anything[[:space:]]by[[:space:]]Simulating[[:space:]]Deformation/e8c8ca9c-763d-45fa-8f3b-d176b29e79d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsigned[[:space:]]Orthogonal[[:space:]]Distance[[:space:]]Fields_[[:space:]]An[[:space:]]Accurate[[:space:]]Neural[[:space:]]Implicit[[:space:]]Representation[[:space:]]for[[:space:]]Diverse[[:space:]]3D[[:space:]]Shapes/7478ca21-59b6-4d95-b955-47a5490cfeda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]3D[[:space:]]Structure[[:space:]]Inference[[:space:]]from[[:space:]]Category-Specific[[:space:]]Image[[:space:]]Collections/9e81eb55-2017-47d1-ab50-718fc069c773_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]Blind[[:space:]]Image[[:space:]]Deblurring[[:space:]]Based[[:space:]]on[[:space:]]Self-Enhancement/ae508369-dbac-472c-9eda-522b8f70eff5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]Deep[[:space:]]Unrolling[[:space:]]Networks[[:space:]]for[[:space:]]Phase[[:space:]]Unwrapping/a918881b-c4ff-499e-bbcd-69ac8904aaf8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]Feature[[:space:]]Learning[[:space:]]with[[:space:]]Emergent[[:space:]]Data-Driven[[:space:]]Prototypicality/b2742b78-cc7b-4e1c-b277-447debc213bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]Gaze[[:space:]]Representation[[:space:]]Learning[[:space:]]from[[:space:]]Multi-view[[:space:]]Face[[:space:]]Images/c85bb360-36d7-45dd-b02d-b8b6c1267338_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]Keypoints[[:space:]]from[[:space:]]Pretrained[[:space:]]Diffusion[[:space:]]Models/4645df0d-9e97-449d-a096-f3522e36e6da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]Learning[[:space:]]of[[:space:]]Category-Level[[:space:]]3D[[:space:]]Pose[[:space:]]from[[:space:]]Object-Centric[[:space:]]Videos/3103a10e-d4f1-46bf-b544-2a9a7eded8b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]Occupancy[[:space:]]Learning[[:space:]]from[[:space:]]Sparse[[:space:]]Point[[:space:]]Cloud/421a640a-85ac-4767-9ea6-13d292c25e4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]Salient[[:space:]]Instance[[:space:]]Detection/d3668a6f-65c6-4577-bc44-ffa98f5eddad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]Through[[:space:]]Depth-Guided[[:space:]]Feature[[:space:]]Correlation[[:space:]]and[[:space:]]Sampling/490ac818-ae06-4da1-a305-2db07efc0513_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]Template-assisted[[:space:]]Point[[:space:]]Cloud[[:space:]]Shape[[:space:]]Correspondence[[:space:]]Network/812a1d24-1f0f-400f-8c84-da6720a2a353_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]Universal[[:space:]]Image[[:space:]]Segmentation/f08d4016-acc4-41d7-91e7-a46018d06d6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unsupervised[[:space:]]Video[[:space:]]Domain[[:space:]]Adaptation[[:space:]]with[[:space:]]Masked[[:space:]]Pre-Training[[:space:]]and[[:space:]]Collaborative[[:space:]]Self-Training/1edbe2de-8bfe-4c07-9630-27dfd62c6876_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unveiling[[:space:]]Parts[[:space:]]Beyond[[:space:]]Objects_[[:space:]]Towards[[:space:]]Finer-Granularity[[:space:]]Referring[[:space:]]Expression[[:space:]]Segmentation/0ef1dc22-21b7-474a-a254-d6f2d4943d35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unveiling[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]Audio-Visual[[:space:]]Early[[:space:]]Fusion[[:space:]]Transformers[[:space:]]with[[:space:]]Dense[[:space:]]Interactions[[:space:]]through[[:space:]]Masked[[:space:]]Modeling/6a06bd12-56a6-4c87-8823-cfa2f03902c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Unveiling[[:space:]]the[[:space:]]Unknown_[[:space:]]Unleashing[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]Unknown[[:space:]]to[[:space:]]Known[[:space:]]in[[:space:]]Open-Set[[:space:]]Source-Free[[:space:]]Domain[[:space:]]Adaptation/496a9fbb-032a-4d75-b940-61e6c2422590_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Upscale-A-Video_[[:space:]]Temporal-Consistent[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Real-World[[:space:]]Video[[:space:]]Super-Resolution/415b5691-a710-48d9-b0ac-0daaf56a288f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Using[[:space:]]Human[[:space:]]Feedback[[:space:]]to[[:space:]]Fine-tune[[:space:]]Diffusion[[:space:]]Models[[:space:]]without[[:space:]]Any[[:space:]]Reward[[:space:]]Model/df1c01a2-2342-4d01-85c2-b06c392b4ad2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Utility-Fairness[[:space:]]Trade-Offs[[:space:]]and[[:space:]]How[[:space:]]to[[:space:]]Find[[:space:]]Them/dffbbb65-0327-47d2-9043-f34845ca6d7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VA3_[[:space:]]Virtually[[:space:]]Assured[[:space:]]Amplification[[:space:]]Attack[[:space:]]on[[:space:]]Probabilistic[[:space:]]Copyright[[:space:]]Protection[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generative[[:space:]]Models/294458f2-3d11-4022-9469-5c90ec662dfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VAREN_[[:space:]]Very[[:space:]]Accurate[[:space:]]and[[:space:]]Realistic[[:space:]]Equine[[:space:]]Network/0f31a4f3-cf8b-45be-b24e-a0b28f1cf54f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VBench_[[:space:]]Comprehensive[[:space:]]Benchmark[[:space:]]Suite[[:space:]]for[[:space:]]Video[[:space:]]Generative[[:space:]]Models/852c042a-18a2-4ced-8d57-cde6f2ed2650_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VCoder_[[:space:]]Versatile[[:space:]]Vision[[:space:]]Encoders[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/817f63e6-6ca6-4fee-9f1a-257768563a6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VGGSfM_[[:space:]]Visual[[:space:]]Geometry[[:space:]]Grounded[[:space:]]Deep[[:space:]]Structure[[:space:]]From[[:space:]]Motion/d2ee9e08-83a5-42e7-ad6b-08e958fae9fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VILA_[[:space:]]On[[:space:]]Pre-training[[:space:]]for[[:space:]]Visual[[:space:]]Language[[:space:]]Models/f648f3d5-06a0-470c-8e61-c14e811bec54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VINECS_[[:space:]]Video-based[[:space:]]Neural[[:space:]]Character[[:space:]]Skinning/14057e08-ebbd-4725-9942-632446ff87e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VISTA-LLAMA_[[:space:]]Reducing[[:space:]]Hallucination[[:space:]]in[[:space:]]Video[[:space:]]Language[[:space:]]Models[[:space:]]via[[:space:]]Equal[[:space:]]Distance[[:space:]]to[[:space:]]Visual[[:space:]]Tokens/7b3a6ad4-efd5-421f-99e3-ea5d44dea958_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VLP_[[:space:]]Vision[[:space:]]Language[[:space:]]Planning[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/16cda83b-9dea-4aab-a622-217c5537ce0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VMC_[[:space:]]Video[[:space:]]Motion[[:space:]]Customization[[:space:]]using[[:space:]]Temporal[[:space:]]Attention[[:space:]]Adaption[[:space:]]for[[:space:]]Text-to-Video[[:space:]]Diffusion[[:space:]]Models/ffbb1e4a-8f04-44a4-a263-284d8b8f46c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VMINer_[[:space:]]Versatile[[:space:]]Multi-view[[:space:]]Inverse[[:space:]]Rendering[[:space:]]with[[:space:]]Near-[[:space:]]and[[:space:]]Far-field[[:space:]]Light[[:space:]]Sources/e2a0f876-3b69-4ca2-9f34-bbb92417c831_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VOODOO[[:space:]]3D_[[:space:]]Volumetric[[:space:]]Portrait[[:space:]]Disentanglement[[:space:]]For[[:space:]]One-Shot[[:space:]]3D[[:space:]]Head[[:space:]]Reenactment/232fb53c-e6e5-4c0f-99d0-a2b74f05feaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VP3D_[[:space:]]Unleashing[[:space:]]2D[[:space:]]Visual[[:space:]]Prompt[[:space:]]for[[:space:]]Text-to-3D[[:space:]]Generation/d6596473-eeb8-454d-940e-3cbe50e191d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VRP-SAM_[[:space:]]SAM[[:space:]]with[[:space:]]Visual[[:space:]]Reference[[:space:]]Prompt/7f3ecbdb-6106-4e8e-86ff-9c2733f56191_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VRetouchEr_[[:space:]]Learning[[:space:]]Cross-frame[[:space:]]Feature[[:space:]]Interdependence[[:space:]]with[[:space:]]Imperfection[[:space:]]Flow[[:space:]]for[[:space:]]Face[[:space:]]Retouching[[:space:]]in[[:space:]]Videos/8a896d01-f26d-44df-a67e-710260f2b006_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VSCode_[[:space:]]General[[:space:]]Visual[[:space:]]Salient[[:space:]]and[[:space:]]Camouflaged[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]2D[[:space:]]Prompt[[:space:]]Learning/745b90c6-8ea5-4303-a0f1-3914d657b814_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VSRD_[[:space:]]Instance-Aware[[:space:]]Volumetric[[:space:]]Silhouette[[:space:]]Rendering[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]3D[[:space:]]Object[[:space:]]Detection/af94e918-61b9-42a3-a56a-ca638b10a2e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VS_[[:space:]]Reconstructing[[:space:]]Clothed[[:space:]]3D[[:space:]]Human[[:space:]]from[[:space:]]Single[[:space:]]Image[[:space:]]via[[:space:]]Vertex[[:space:]]Shift/1d0a8a04-2343-4d6f-871f-2ebf0c5dbdc7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VTQA_[[:space:]]Visual[[:space:]]Text[[:space:]]Question[[:space:]]Answering[[:space:]]via[[:space:]]Entity[[:space:]]Alignment[[:space:]]and[[:space:]]Cross-Media[[:space:]]Reasoning/e4ec05ff-3fa1-4ffb-b826-5f31038b1d51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VTimeLLM_[[:space:]]Empower[[:space:]]LLM[[:space:]]to[[:space:]]Grasp[[:space:]]Video[[:space:]]Moments/93551b6a-586f-4f31-83ad-4d86b76784d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/V__[[:space:]]Guided[[:space:]]Visual[[:space:]]Search[[:space:]]as[[:space:]]a[[:space:]]Core[[:space:]]Mechanism[[:space:]]in[[:space:]]Multimodal[[:space:]]LLMs/da9d620a-ef32-4c5e-b349-8d066a704f0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Validating[[:space:]]Privacy-Preserving[[:space:]]Face[[:space:]]Recognition[[:space:]]under[[:space:]]a[[:space:]]Minimum[[:space:]]Assumption/9139a567-fbb9-4834-bbbf-7d18ccf906f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Vanishing-Point-Guided[[:space:]]Video[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]of[[:space:]]Driving[[:space:]]Scenes/c95c8cd9-ea0a-498e-afad-6c04356cea08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VastGaussian_[[:space:]]Vast[[:space:]]3D[[:space:]]Gaussians[[:space:]]for[[:space:]]Large[[:space:]]Scene[[:space:]]Reconstruction/dae866b8-bba9-4088-8662-ebfdb451138c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VecFusion_[[:space:]]Vector[[:space:]]Font[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion/c5f5c319-ebd3-42c5-86d7-2ff36fd3c349_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Vector[[:space:]]Graphics[[:space:]]Generation[[:space:]]via[[:space:]]Mutually[[:space:]]Impulsed[[:space:]]Dual-domain[[:space:]]Diffusion/dad89f64-ff9a-4db0-b8ae-8201e4fb21a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Versatile[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation[[:space:]]Learned[[:space:]]from[[:space:]]Multi-Source[[:space:]]Datasets[[:space:]]via[[:space:]]Model[[:space:]]Self-Disambiguation/c3e7e066-94ab-4dd4-a693-b92631ccd737_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Versatile[[:space:]]Navigation[[:space:]]Under[[:space:]]Partial[[:space:]]Observability[[:space:]]via[[:space:]]Value-guided[[:space:]]Diffusion[[:space:]]Policy/dec738eb-2199-49f4-ba41-b8bdbd7e0fa4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ViLa-MIL_[[:space:]]Dual-scale[[:space:]]Vision-Language[[:space:]]Multiple[[:space:]]Instance[[:space:]]Learning[[:space:]]for[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Classification/85372ebf-31af-4d52-a240-dcdd7001d47f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ViP-LLaVA_[[:space:]]Making[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models[[:space:]]Understand[[:space:]]Arbitrary[[:space:]]Visual[[:space:]]Prompts/4764e32b-b5e0-4cec-b885-ce28665cbc4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ViT-CoMer_[[:space:]]Vision[[:space:]]Transformer[[:space:]]with[[:space:]]Convolutional[[:space:]]Multi-scale[[:space:]]Feature[[:space:]]Interaction[[:space:]]for[[:space:]]Dense[[:space:]]Predictions/2a3323b9-7324-4cd3-aae1-ea2eb71358b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ViT-Lens_[[:space:]]Towards[[:space:]]Omni-modal[[:space:]]Representations/f3bd0731-24e0-43f0-9476-94f88d4b11b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ViTamin_[[:space:]]Designing[[:space:]]Scalable[[:space:]]Vision[[:space:]]Models[[:space:]]in[[:space:]]the[[:space:]]Vision-Language[[:space:]]Era/ca2d983d-6b19-4d87-995a-d003078f34a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ViVid-1-to-3_[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/d4609e1c-e930-4647-bcd0-e97dd36ab5d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VicTR_[[:space:]]Video-conditioned[[:space:]]Text[[:space:]]Representations[[:space:]]for[[:space:]]Activity[[:space:]]Recognition/29323e50-5f81-4d07-b22d-e4ee0037aaef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VidLA_[[:space:]]Video-Language[[:space:]]Alignment[[:space:]]at[[:space:]]Scale/fc43318d-405d-4040-b280-5d6a6f3f19c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VidToMe_[[:space:]]Video[[:space:]]Token[[:space:]]Merging[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Video[[:space:]]Editing/fb29b626-2268-4dba-853b-90322a5c4cdf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Video[[:space:]]Frame[[:space:]]Interpolation[[:space:]]via[[:space:]]Direct[[:space:]]Synthesis[[:space:]]with[[:space:]]the[[:space:]]Event-based[[:space:]]Reference/2d993f24-453b-4589-9f64-71049d897aff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Video[[:space:]]Harmonization[[:space:]]with[[:space:]]Triplet[[:space:]]Spatio-Temporal[[:space:]]Variation[[:space:]]Patterns/5c610f72-b14a-4ec5-88f7-a1cf9a33dda3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Video[[:space:]]Interpolation[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/fd333655-848a-4ba1-8c05-96c4cf9bfdad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Video[[:space:]]Prediction[[:space:]]by[[:space:]]Modeling[[:space:]]Videos[[:space:]]as[[:space:]]Continuous[[:space:]]Multi-Dimensional[[:space:]]Processes/ccc42f26-9a12-4018-9b85-3006f577d2a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Video[[:space:]]ReCap_[[:space:]]Recursive[[:space:]]Captioning[[:space:]]of[[:space:]]Hour-Long[[:space:]]Videos/34833510-7e3f-46a3-8e63-d31ed9f4ff24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Video[[:space:]]Recognition[[:space:]]in[[:space:]]Portrait[[:space:]]Mode/1430817f-2fb2-4e8e-aaf4-dc2cf17e83e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Video[[:space:]]Super-Resolution[[:space:]]Transformer[[:space:]]with[[:space:]]Masked[[:space:]]Inter&Intra-Frame[[:space:]]Attention/1e69759c-d8e8-49da-aff4-8c9d96153686_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Video-Based[[:space:]]Human[[:space:]]Pose[[:space:]]Regression[[:space:]]via[[:space:]]Decoupled[[:space:]]Space-Time[[:space:]]Aggregation/f3738d37-ef8c-4352-b1e1-4943413c7dec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Video-P2P_[[:space:]]Video[[:space:]]Editing[[:space:]]with[[:space:]]Cross-attention[[:space:]]Control/81ad8533-a31f-4ffc-839e-f07d4869e500_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Video2Game_[[:space:]]Real-time[[:space:]]Interactive[[:space:]]Realistic[[:space:]]and[[:space:]]Browser-Compatible[[:space:]]Environment[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Video/5bc42c40-19c8-4162-b372-460cfcad74f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VideoBooth_[[:space:]]Diffusion-based[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Image[[:space:]]Prompts/f4a1b767-87c7-478e-8bb4-526b79e30b63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VideoCon_[[:space:]]Robust[[:space:]]Video-Language[[:space:]]Alignment[[:space:]]via[[:space:]]Contrast[[:space:]]Captions/7aaa3f92-fb68-4a9b-809d-b79e89d5cc93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VideoCrafter2_[[:space:]]Overcoming[[:space:]]Data[[:space:]]Limitations[[:space:]]for[[:space:]]High-Quality[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/ab56d7df-62cd-4c60-8578-721928a2487b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VideoCutLER_[[:space:]]Surprisingly[[:space:]]Simple[[:space:]]Unsupervised[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation/d5335ce2-de3e-4c31-a9e0-97b9c027eebd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VideoGrounding-DINO_[[:space:]]Towards[[:space:]]Open-Vocabulary[[:space:]]Spatio-Temporal[[:space:]]Video[[:space:]]Grounding/8669738c-7c6c-4e4e-b34d-5818146c2193_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VideoLLM-online_[[:space:]]Online[[:space:]]Video[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Streaming[[:space:]]Video/c25a98ca-87a4-45af-9e10-acd055b937ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VideoMAC_[[:space:]]Video[[:space:]]Masked[[:space:]]Autoencoders[[:space:]]Meet[[:space:]]ConvNets/d723e353-3b17-4db6-af6c-92f3b43a3ed7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VideoRF_[[:space:]]Rendering[[:space:]]Dynamic[[:space:]]Radiance[[:space:]]Fields[[:space:]]as[[:space:]]2D[[:space:]]Feature[[:space:]]Video[[:space:]]Streams/12d72448-e14e-4deb-b4a1-e20c3934e649_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VideoSwap_[[:space:]]Customized[[:space:]]Video[[:space:]]Subject[[:space:]]Swapping[[:space:]]with[[:space:]]Interactive[[:space:]]Semantic[[:space:]]Point[[:space:]]Correspondence/f25f0e8f-6c9a-4388-bda9-449fe7becf31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/View[[:space:]]From[[:space:]]Above_[[:space:]]Orthogonal-View[[:space:]]aware[[:space:]]Cross-view[[:space:]]Localization/7abae4da-a8d5-4e84-bac8-8b520ca7f32d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/View-Category[[:space:]]Interactive[[:space:]]Sharing[[:space:]]Transformer[[:space:]]for[[:space:]]Incomplete[[:space:]]Multi-View[[:space:]]Multi-Label[[:space:]]Learning/18ba094d-30eb-4383-97ef-520cbe2b100a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/View-decoupled[[:space:]]Transformer[[:space:]]for[[:space:]]Person[[:space:]]Re-identification[[:space:]]under[[:space:]]Aerial-ground[[:space:]]Camera[[:space:]]Network/66ee0e69-7b19-441f-b54a-195f4cb4500a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ViewDiff_[[:space:]]3D-Consistent[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Text-to-Image[[:space:]]Models/ca4acaac-6100-4888-9856-8b0c1fe30b3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ViewFusion_[[:space:]]Towards[[:space:]]Multi-View[[:space:]]Consistency[[:space:]]via[[:space:]]Interpolated[[:space:]]Denoising/cae62f84-c32b-4a99-a5f6-784c0ce32aad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Viewpoint-Aware[[:space:]]Visual[[:space:]]Grounding[[:space:]]in[[:space:]]3D[[:space:]]Scenes/8d04b1f0-f2a1-4a61-aa7a-713e99eb821c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Virtual[[:space:]]Immunohistochemistry[[:space:]]Staining[[:space:]]for[[:space:]]Histological[[:space:]]Images[[:space:]]Assisted[[:space:]]by[[:space:]]Weakly-supervised[[:space:]]Learning/0012216a-0bd1-4dfe-bd44-a5f2cb587723_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Vision-and-Language[[:space:]]Navigation[[:space:]]via[[:space:]]Causal[[:space:]]Learning/8ed24918-937e-4d68-a4ff-fac9ff1bc6a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Visual[[:space:]]Anagrams_[[:space:]]Generating[[:space:]]Multi-View[[:space:]]Optical[[:space:]]Illusions[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/f2a047a9-fca8-496c-9ac0-94e4fff7168a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Visual[[:space:]]Concept[[:space:]]Connectome[[:space:]](VCC)_[[:space:]]Open[[:space:]]World[[:space:]]Concept[[:space:]]Discovery[[:space:]]and[[:space:]]their[[:space:]]Interlayer[[:space:]]Connections[[:space:]]in[[:space:]]Deep[[:space:]]Models/1a616d0d-3730-44ab-a630-ecdfca09f3c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Visual[[:space:]]Delta[[:space:]]Generator[[:space:]]with[[:space:]]Large[[:space:]]Multi-modal[[:space:]]Models[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/0e65e087-fb63-4098-9579-6b8185c1ed79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Visual[[:space:]]Fact[[:space:]]Checker_[[:space:]]Enabling[[:space:]]High-Fidelity[[:space:]]Detailed[[:space:]]Caption[[:space:]]Generation/41baf37b-e81b-4a82-a385-a78a616c7868_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Visual[[:space:]]In-Context[[:space:]]Prompting/ecd3f7ab-166e-4fd1-8364-6864de752bd7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Visual[[:space:]]Layout[[:space:]]Composer_[[:space:]]Image-Vector[[:space:]]Dual[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Design[[:space:]]Layout[[:space:]]Generation/75d905c0-508c-4fbc-a17b-7ac7fef387c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Visual[[:space:]]Objectification[[:space:]]in[[:space:]]Films_[[:space:]]Towards[[:space:]]a[[:space:]]New[[:space:]]AI[[:space:]]Task[[:space:]]for[[:space:]]Video[[:space:]]Interpretation/410cda22-2c59-4ec1-949f-7a68b660e50b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Visual[[:space:]]Point[[:space:]]Cloud[[:space:]]Forecasting[[:space:]]enables[[:space:]]Scalable[[:space:]]Autonomous[[:space:]]Driving/91fee4e7-5b31-4963-9750-c742a8cbf742_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Visual[[:space:]]Program[[:space:]]Distillation_[[:space:]]Distilling[[:space:]]Tools[[:space:]]and[[:space:]]Programmatic[[:space:]]Reasoning[[:space:]]into[[:space:]]Vision-Language[[:space:]]Models/d3d264df-7a0f-49c2-be47-325ff84855b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Visual[[:space:]]Programming[[:space:]]for[[:space:]]Zero-shot[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding/7da4d511-ff1b-4061-9760-a8f55231549a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Visual[[:space:]]Prompting[[:space:]]for[[:space:]]Generalized[[:space:]]Few-shot[[:space:]]Segmentation_[[:space:]]A[[:space:]]Multi-scale[[:space:]]Approach/d4543bc4-aa84-4173-9885-2737fde4ca07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Visual-Augmented[[:space:]]Dynamic[[:space:]]Semantic[[:space:]]Prototype[[:space:]]for[[:space:]]Generative[[:space:]]Zero-Shot[[:space:]]Learning/7c08a69a-09ad-40db-a963-6aaf195fbfa6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VkD_[[:space:]]Improving[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]using[[:space:]]Orthogonal[[:space:]]Projections/260ea0d3-c32f-4988-98e9-ada076a48353_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Vlogger_[[:space:]]Make[[:space:]]Your[[:space:]]Dream[[:space:]]A[[:space:]]Vlog/3067237f-4763-4f29-bd3f-1f0c7a112ee8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/VoCo_[[:space:]]A[[:space:]]Simple-yet-Effective[[:space:]]Volume[[:space:]]Contrastive[[:space:]]Learning[[:space:]]Framework[[:space:]]for[[:space:]]3D[[:space:]]Medical[[:space:]]Image[[:space:]]Analysis/8902ed6b-c8cb-4bc7-a572-045797bd1786_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Volumetric[[:space:]]Environment[[:space:]]Representation[[:space:]]for[[:space:]]Vision-Language[[:space:]]Navigation/6562b0ae-fedd-4e65-b84a-d2cbb4f40727_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/WALT3D_[[:space:]]Generating[[:space:]]Realistic[[:space:]]Training[[:space:]]Data[[:space:]]from[[:space:]]Time-Lapse[[:space:]]Imagery[[:space:]]for[[:space:]]Reconstructing[[:space:]]Dynamic[[:space:]]Objects[[:space:]]Under[[:space:]]Occlusion/048cb13a-fa04-482f-97a0-4548a36785ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/WANDR_[[:space:]]Intention-guided[[:space:]]Human[[:space:]]Motion[[:space:]]Generation/f6756573-8e2b-41a5-9d35-381ea4ebb252_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/WHAM_[[:space:]]Reconstructing[[:space:]]World-grounded[[:space:]]Humans[[:space:]]with[[:space:]]Accurate[[:space:]]3D[[:space:]]Motion/63a7b8f6-97aa-4651-a01d-91f5ef8cd504_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/WOUAF_[[:space:]]Weight[[:space:]]Modulation[[:space:]]for[[:space:]]User[[:space:]]Attribution[[:space:]]and[[:space:]]Fingerprinting[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/92d9125f-3522-4960-9c33-6d71a67f7675_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/WWW_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Explaining[[:space:]]What[[:space:]]Where[[:space:]]and[[:space:]]Why[[:space:]]of[[:space:]]Neural[[:space:]]Networks[[:space:]]by[[:space:]]Interpretation[[:space:]]of[[:space:]]Neuron[[:space:]]Concepts/81e2b3ad-69ba-401e-815e-45595c76f065_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/WateRF_[[:space:]]Robust[[:space:]]Watermarks[[:space:]]in[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Protection[[:space:]]of[[:space:]]Copyrights/36a5a6f2-e177-4dfb-ad96-1aaeb75afb72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Watermark-embedded[[:space:]]Adversarial[[:space:]]Examples[[:space:]]for[[:space:]]Copyright[[:space:]]Protection[[:space:]]against[[:space:]]Diffusion[[:space:]]Models/9f5ff66f-0862-4163-9fbc-9a609665d28d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/WaveFace_[[:space:]]Authentic[[:space:]]Face[[:space:]]Restoration[[:space:]]with[[:space:]]Efficient[[:space:]]Frequency[[:space:]]Recovery/8a8efe74-fca8-4300-90b6-5878a71afa4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/WaveMo_[[:space:]]Learning[[:space:]]Wavefront[[:space:]]Modulations[[:space:]]to[[:space:]]See[[:space:]]Through[[:space:]]Scattering/7f09b732-d1d1-4da5-806d-7ae1886c2551_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Wavelet-based[[:space:]]Fourier[[:space:]]Information[[:space:]]Interaction[[:space:]]with[[:space:]]Frequency[[:space:]]Diffusion[[:space:]]Adjustment[[:space:]]for[[:space:]]Underwater[[:space:]]Image[[:space:]]Restoration/f627431b-e7bf-434f-9f57-1c74972ca141_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Weak-to-Strong[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]X-Ray[[:space:]]Distillation/ecb5d87e-b1e5-43e0-b246-7e30c3a3bbcd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Weakly[[:space:]]Misalignment-free[[:space:]]Adaptive[[:space:]]Feature[[:space:]]Alignment[[:space:]]for[[:space:]]UAVs-based[[:space:]]Multimodal[[:space:]]Object[[:space:]]Detection/7c6f4239-84d8-42d9-a2a3-2ef9f81fe732_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Weakly[[:space:]]Supervised[[:space:]]Monocular[[:space:]]3D[[:space:]]Detection[[:space:]]with[[:space:]]a[[:space:]]Single-View[[:space:]]Image/e8def396-7aa9-4446-a151-70f8d14a292d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Weakly[[:space:]]Supervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]via[[:space:]]Artificial[[:space:]]Oracle/310f76a4-fe34-49c0-9b81-fed65524e491_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Weakly[[:space:]]Supervised[[:space:]]Video[[:space:]]Individual[[:space:]]Counting/357e37d5-1951-48de-9fbc-83246528162f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Weakly-Supervised[[:space:]]Audio-Visual[[:space:]]Video[[:space:]]Parsing[[:space:]]with[[:space:]]Prototype-based[[:space:]]Pseudo-Labeling/9ca95c3c-3225-4d8b-a7dd-d7e7b707e693_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Weakly-Supervised[[:space:]]Emotion[[:space:]]Transition[[:space:]]Learning[[:space:]]for[[:space:]]Diverse[[:space:]]3D[[:space:]]Co-speech[[:space:]]Gesture[[:space:]]Generation/82233940-816a-43f0-81c1-3331488ee413_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/What[[:space:]]Do[[:space:]]You[[:space:]]See[[:space:]]in[[:space:]]Vehicle_[[:space:]]Comprehensive[[:space:]]Vision[[:space:]]Solution[[:space:]]for[[:space:]]In-Vehicle[[:space:]]Gaze[[:space:]]Estimation/d9745d3f-8ba4-4f6a-bd07-c7535881512b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/What[[:space:]]How[[:space:]]and[[:space:]]When[[:space:]]Should[[:space:]]Object[[:space:]]Detectors[[:space:]]Update[[:space:]]in[[:space:]]Continually[[:space:]]Changing[[:space:]]Test[[:space:]]Domains_/d778613f-28c0-4d7a-80d6-b7c15f243643_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/What[[:space:]]If[[:space:]]the[[:space:]]TV[[:space:]]Was[[:space:]]Off_[[:space:]]Examining[[:space:]]Counterfactual[[:space:]]Reasoning[[:space:]]Abilities[[:space:]]of[[:space:]]Multi-modal[[:space:]]Language[[:space:]]Models/9f3da693-f7a9-4a65-84ec-f6f22431f6d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/What[[:space:]]Sketch[[:space:]]Explainability[[:space:]]Really[[:space:]]Means[[:space:]]for[[:space:]]Downstream[[:space:]]Tasks_/22ec61cf-a0b2-4071-b86f-9e1362f3069e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/What[[:space:]]When[[:space:]]and[[:space:]]Where_[[:space:]]Self-Supervised[[:space:]]Spatio-Temporal[[:space:]]Grounding[[:space:]]in[[:space:]]Untrimmed[[:space:]]Multi-Action[[:space:]]Videos[[:space:]]from[[:space:]]Narrated[[:space:]]Instructions/59de70a6-beec-4552-9d70-f207c0df32db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/What[[:space:]]You[[:space:]]See[[:space:]]is[[:space:]]What[[:space:]]You[[:space:]]GAN_[[:space:]]Rendering[[:space:]]Every[[:space:]]Pixel[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Geometry[[:space:]]in[[:space:]]3D[[:space:]]GANs/2c134100-ed06-4b65-8341-cd48f2b74244_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/When[[:space:]]StyleGAN[[:space:]]Meets[[:space:]]Stable[[:space:]]Diffusion_[[:space:]]a[[:space:]]W+[[:space:]]Adapter[[:space:]]for[[:space:]]Personalized[[:space:]]Image[[:space:]]Generation/2ed92abb-5de3-46c9-ba95-16327031ae50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/When[[:space:]]Visual[[:space:]]Grounding[[:space:]]Meets[[:space:]]Gigapixel-level[[:space:]]Large-scale[[:space:]]Scenes_[[:space:]]Benchmark[[:space:]]and[[:space:]]Approach/69fa19e8-30d0-4373-b41c-ba0437416a98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Why[[:space:]]Not[[:space:]]Use[[:space:]]Your[[:space:]]Textbook_[[:space:]]Knowledge-Enhanced[[:space:]]Procedure[[:space:]]Planning[[:space:]]of[[:space:]]Instructional[[:space:]]Videos/f83cbc4e-8548-48f7-bd06-1fc89d0517b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/WildlifeMapper_[[:space:]]Aerial[[:space:]]Image[[:space:]]Analysis[[:space:]]for[[:space:]]Multi-Species[[:space:]]Detection[[:space:]]and[[:space:]]Identification/75c6b881-4935-4fc9-8673-9dc09a1854b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/WinSyn_[[:space:]]_[[:space:]]A[[:space:]]High[[:space:]]Resolution[[:space:]]Testbed[[:space:]]for[[:space:]]Synthetic[[:space:]]Data/dce1c1ee-50a6-4a10-b211-db6a6195eecc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Wired[[:space:]]Perspectives_[[:space:]]Multi-View[[:space:]]Wire[[:space:]]Art[[:space:]]Embraces[[:space:]]Generative[[:space:]]AI/2a286243-d96a-406c-bb14-a19955ff2a60_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Wonder3D_[[:space:]]Single[[:space:]]Image[[:space:]]to[[:space:]]3D[[:space:]]using[[:space:]]Cross-Domain[[:space:]]Diffusion/736178a6-38c9-4e35-a376-b88b1b09495e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/WonderJourney_[[:space:]]Going[[:space:]]from[[:space:]]Anywhere[[:space:]]to[[:space:]]Everywhere/d25f0803-ff6d-4fc4-a3ae-eff93a12eda4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/WorDepth_[[:space:]]Variational[[:space:]]Language[[:space:]]Prior[[:space:]]for[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/6e3a96b9-e9a3-4fa6-bc57-71444d80cfa8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Would[[:space:]]Deep[[:space:]]Generative[[:space:]]Models[[:space:]]Amplify[[:space:]]Bias[[:space:]]in[[:space:]]Future[[:space:]]Models_/2c6cf8dd-91a5-4e41-a6f6-998a2456780f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/X-3D_[[:space:]]Explicit[[:space:]]3D[[:space:]]Structure[[:space:]]Modeling[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Recognition/270579d6-b008-4b7d-affb-d28fb22d6d5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/X-Adapter_[[:space:]]Adding[[:space:]]Universal[[:space:]]Compatibility[[:space:]]of[[:space:]]Plugins[[:space:]]for[[:space:]]Upgraded[[:space:]]Diffusion[[:space:]]Model/fec7dee8-4271-40e1-a35e-f358f2749365_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/X-MIC_[[:space:]]Cross-Modal[[:space:]]Instance[[:space:]]Conditioning[[:space:]]for[[:space:]]Egocentric[[:space:]]Action[[:space:]]Generalization/ae0ecb40-b12e-4ca4-b3c4-7741f7586ab6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/XCube_[[:space:]]Large-Scale[[:space:]]3D[[:space:]]Generative[[:space:]]Modeling[[:space:]]using[[:space:]]Sparse[[:space:]]Voxel[[:space:]]Hierarchies/53210037-d63b-4b3b-af2c-01f324e7439d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/XFeat_[[:space:]]Accelerated[[:space:]]Features[[:space:]]for[[:space:]]Lightweight[[:space:]]Image[[:space:]]Matching/816a5890-ef40-4756-b581-f63df9243d92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/XFibrosis_[[:space:]]Explicit[[:space:]]Vessel-Fiber[[:space:]]Modeling[[:space:]]for[[:space:]]Fibrosis[[:space:]]Staging[[:space:]]from[[:space:]]Liver[[:space:]]Pathology[[:space:]]Images/b87a72ab-13bf-4696-a4b0-9acbf31582a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/XScale-NVS_[[:space:]]Cross-Scale[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]Hash[[:space:]]Featurized[[:space:]]Manifold/4a3d08c6-ff96-49c2-9291-c654cc37b7fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/YOLO-World_[[:space:]]Real-Time[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detection/15948506-ac19-4b38-ae4b-e409c3ad8cfb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/YolOOD_[[:space:]]Utilizing[[:space:]]Object[[:space:]]Detection[[:space:]]Concepts[[:space:]]for[[:space:]]Multi-Label[[:space:]]Out-of-Distribution[[:space:]]Detection/a27ac8a5-b874-4820-82b3-e4da7838f3df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/You[[:space:]]Only[[:space:]]Need[[:space:]]Less[[:space:]]Attention[[:space:]]at[[:space:]]Each[[:space:]]Stage[[:space:]]in[[:space:]]Vision[[:space:]]Transformers/d4917183-09e8-4a93-a190-f7b023cbd63c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/You'll[[:space:]]Never[[:space:]]Walk[[:space:]]Alone_[[:space:]]A[[:space:]]Sketch[[:space:]]and[[:space:]]Text[[:space:]]Duet[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Image[[:space:]]Retrieval/1752c138-7203-4c24-b026-11ea262763d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Your[[:space:]]Image[[:space:]]is[[:space:]]My[[:space:]]Video_[[:space:]]Reshaping[[:space:]]the[[:space:]]Receptive[[:space:]]Field[[:space:]]via[[:space:]]Image-To-Video[[:space:]]Differentiable[[:space:]]AutoAugmentation[[:space:]]and[[:space:]]Fusion/172e0d2e-c491-4fb3-b948-bd0ad16bb15f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Your[[:space:]]Student[[:space:]]is[[:space:]]Better[[:space:]]Than[[:space:]]Expected_[[:space:]]Adaptive[[:space:]]Teacher-Student[[:space:]]Collaboration[[:space:]]for[[:space:]]Text-Conditional[[:space:]]Diffusion[[:space:]]Models/073fa4b9-1533-4c93-9378-2fb3f5e2e584_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Your[[:space:]]Transferability[[:space:]]Barrier[[:space:]]is[[:space:]]Fragile_[[:space:]]Free-Lunch[[:space:]]for[[:space:]]Transferring[[:space:]]the[[:space:]]Non-Transferable[[:space:]]Learning/6c79646e-1839-4c23-9f2c-a754bd23c52a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ZERO-IG_[[:space:]]Zero-Shot[[:space:]]Illumination-Guided[[:space:]]Joint[[:space:]]Denoising[[:space:]]and[[:space:]]Adaptive[[:space:]]Enhancement[[:space:]]for[[:space:]]Low-Light[[:space:]]Images/fe0053b6-5953-458a-a584-815b7a7802d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ZONE_[[:space:]]Zero-Shot[[:space:]]Instruction-Guided[[:space:]]Local[[:space:]]Editing/1c622694-08e2-4753-8dca-d4c2be33342a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Z__[[:space:]]Zero-shot[[:space:]]Style[[:space:]]Transfer[[:space:]]via[[:space:]]Attention[[:space:]]Reweighting/600de97e-c638-463e-8e20-354165caf630_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ZePT_[[:space:]]Zero-Shot[[:space:]]Pan-Tumor[[:space:]]Segmentation[[:space:]]via[[:space:]]Query-Disentangling[[:space:]]and[[:space:]]Self-Prompting/c94c42d9-224a-494d-baa3-5bd94af96d4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Zero-Painter_[[:space:]]Training-Free[[:space:]]Layout[[:space:]]Control[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Synthesis/64e10689-61e0-4e7d-a753-d0641c758d29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Zero-Reference[[:space:]]Low-Light[[:space:]]Enhancement[[:space:]]via[[:space:]]Physical[[:space:]]Quadruple[[:space:]]Priors/c6733bf2-9a31-4fbc-9cfc-5083de193790_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Zero-Shot[[:space:]]Structure-Preserving[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]High[[:space:]]Dynamic[[:space:]]Range[[:space:]]Tone[[:space:]]Mapping/27347abd-0e6e-4d5f-b927-cc6a4dfeabad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Zero-TPrune_[[:space:]]Zero-Shot[[:space:]]Token[[:space:]]Pruning[[:space:]]through[[:space:]]Leveraging[[:space:]]of[[:space:]]the[[:space:]]Attention[[:space:]]Graph[[:space:]]in[[:space:]]Pre-Trained[[:space:]]Transformers/cad006cb-71e1-4787-82dc-67186ad24ee3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Zero-shot[[:space:]]Referring[[:space:]]Expression[[:space:]]Comprehension[[:space:]]via[[:space:]]Structural[[:space:]]Similarity[[:space:]]Between[[:space:]]Images[[:space:]]and[[:space:]]Captions/c2e74620-bcd9-4076-922f-31b3012b4f77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ZeroNVS_[[:space:]]Zero-Shot[[:space:]]360-Degree[[:space:]]View[[:space:]]Synthesis[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/99a5e0a8-3947-4a04-8666-b90370b52b3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ZeroRF_[[:space:]]Fast[[:space:]]Sparse[[:space:]]View[[:space:]]360deg[[:space:]]Reconstruction[[:space:]]with[[:space:]]Zero[[:space:]]Pretraining/071d7685-0262-4e09-a6b8-183dc1dfd424_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/ZeroShape_[[:space:]]Regression-based[[:space:]]Zero-shot[[:space:]]Shape[[:space:]]Reconstruction/f3394027-3aa8-4a0e-8e4e-7e97bd1eae78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/eTraM_[[:space:]]Event-based[[:space:]]Traffic[[:space:]]Monitoring[[:space:]]Dataset/0d9c17c7-80f8-46c4-a8f5-88cbd231d7ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/iKUN_[[:space:]]Speak[[:space:]]to[[:space:]]Trackers[[:space:]]without[[:space:]]Retraining/b2c65ad9-c18e-4b68-932a-9bb410435342_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/iToF-flow-based[[:space:]]High[[:space:]]Frame[[:space:]]Rate[[:space:]]Depth[[:space:]]Imaging/2a38c499-ed6b-4898-9679-13175a7ffb47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/mPLUG-Owl2_[[:space:]]Revolutionizing[[:space:]]Multi-modal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]with[[:space:]]Modality[[:space:]]Collaboration/92d91f80-3085-4c9d-a9ed-3ec589bdd42d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/pix2gestalt_[[:space:]]Amodal[[:space:]]Segmentation[[:space:]]by[[:space:]]Synthesizing[[:space:]]Wholes/d85b1e03-3389-4732-b9c5-9b1267816128_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/pixelSplat_[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splats[[:space:]]from[[:space:]]Image[[:space:]]Pairs[[:space:]]for[[:space:]]Scalable[[:space:]]Generalizable[[:space:]]3D[[:space:]]Reconstruction/ecc56ffa-c00f-4d66-a386-40e83747dac3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/vid-TLDR_[[:space:]]Training[[:space:]]Free[[:space:]]Token[[:space:]]Merging[[:space:]]for[[:space:]]Light-weight[[:space:]]Video[[:space:]]Transformer/bcf90d09-c130-4a92-9743-0c3654ba1208_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Semantic[[:space:]]Shield_[[:space:]]Defending[[:space:]]Vision-Language[[:space:]]Models[[:space:]]Against[[:space:]]Backdooring[[:space:]]and[[:space:]]Poisoning[[:space:]]via[[:space:]]Fine-grained[[:space:]]Knowledge[[:space:]]Alignment/811b225c-f7ba-4d5f-980e-3c936fd65339_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Semantic-Aware[[:space:]]Multi-Label[[:space:]]Adversarial[[:space:]]Attacks/a7635c9b-bbb0-4721-869c-8ded1f1d58af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Semantic-aware[[:space:]]SAM[[:space:]]for[[:space:]]Point-Prompted[[:space:]]Instance[[:space:]]Segmentation/98ef3dac-add6-4e97-bd9e-baeff12acffa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Semantically-Shifted[[:space:]]Incremental[[:space:]]Adapter-Tuning[[:space:]]is[[:space:]]A[[:space:]]Continual[[:space:]]ViTransformer/3e66199c-eb4c-4d3b-89c8-b97d56ae08e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Semantics[[:space:]]Distortion[[:space:]]and[[:space:]]Style[[:space:]]Matter_[[:space:]]Towards[[:space:]]Source-free[[:space:]]UDA[[:space:]]for[[:space:]]Panoramic[[:space:]]Segmentation/fc7b011a-8b94-4431-a0e3-dcc09545c286_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Semantics-aware[[:space:]]Motion[[:space:]]Retargeting[[:space:]]with[[:space:]]Vision-Language[[:space:]]Models/94f0a721-55ce-4d79-a7e3-1129e082a51d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2024/Separate[[:space:]]and[[:space:]]Conquer_[[:space:]]Decoupling[[:space:]]Co-occurrence[[:space:]]via[[:space:]]Decomposition[[:space:]]and[[:space:]]Representation[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/11809eba-7acc-4ecd-8990-e41f9013a9db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DroneSplat_[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Robust[[:space:]]3D[[:space:]]Reconstruction[[:space:]]from[[:space:]]In-the-Wild[[:space:]]Drone[[:space:]]Imagery/f4bb2730-8b51-4763-bc92-d8e30f5ac837_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DropGaussian_[[:space:]]Structural[[:space:]]Regularization[[:space:]]for[[:space:]]Sparse-view[[:space:]]Gaussian[[:space:]]Splatting/d54606e8-0565-4b7a-8755-24626c22d0de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DropoutGS_[[:space:]]Dropping[[:space:]]Out[[:space:]]Gaussians[[:space:]]for[[:space:]]Better[[:space:]]Sparse-view[[:space:]]Rendering/6430309a-4b53-46c7-842b-046e4f57064c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual[[:space:]]Consolidation[[:space:]]for[[:space:]]Pre-Trained[[:space:]]Model-Based[[:space:]]Domain-Incremental[[:space:]]Learning/81ab03ae-58d9-44c3-bbcd-bc4b8291835f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual[[:space:]]Diffusion[[:space:]]for[[:space:]]Unified[[:space:]]Image[[:space:]]Generation[[:space:]]and[[:space:]]Understanding/6c67f231-518e-4105-8c37-80036fc6ba48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual[[:space:]]Energy-Based[[:space:]]Model[[:space:]]with[[:space:]]Open-World[[:space:]]Uncertainty[[:space:]]Estimation[[:space:]]for[[:space:]]Out-of-distribution[[:space:]]Detection/3a1741c3-d23d-47e7-b209-5f3583ea54da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual[[:space:]]Exposure[[:space:]]Stereo[[:space:]]for[[:space:]]Extended[[:space:]]Dynamic[[:space:]]Range[[:space:]]3D[[:space:]]Imaging/bdd34ba1-9aaf-4515-be7c-34458b797d5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual[[:space:]]Focus-Attention[[:space:]]Transformer[[:space:]]for[[:space:]]Robust[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/1504f286-9f4e-40fc-a50e-031be91845bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual[[:space:]]Prompting[[:space:]]Image[[:space:]]Restoration[[:space:]]with[[:space:]]Diffusion[[:space:]]Transformers/0167513a-f2bb-4815-986e-556fa6d10ed2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual[[:space:]]Semantic[[:space:]]Guidance[[:space:]]for[[:space:]]Open[[:space:]]Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/08ea4716-56f0-4e2a-ab63-dfc1f7d409d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual-Agent[[:space:]]Optimization[[:space:]]framework[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Few-Shot[[:space:]]Segmentation/2dde1188-8599-4a43-867d-d400f59ab0cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual-Granularity[[:space:]]Semantic[[:space:]]Guided[[:space:]]Sparse[[:space:]]Routing[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]General[[:space:]]Pansharpening/6f8e11ff-cb5a-4f09-a665-4dc98202f12c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual-Interrelated[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Few-Shot[[:space:]]Anomaly[[:space:]]Image[[:space:]]Generation/48fd8ec3-7139-42a5-8004-16737329b070_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual-view[[:space:]]X-ray[[:space:]]Detection_[[:space:]]Can[[:space:]]AI[[:space:]]Detect[[:space:]]Prohibited[[:space:]]Items[[:space:]]from[[:space:]]Dual-view[[:space:]]X-ray[[:space:]]Images[[:space:]]like[[:space:]]Humans_/31a48103-8443-49ef-94bc-e135e6ffd9b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DualPM_[[:space:]]Dual[[:space:]]Posed-Canonical[[:space:]]Point[[:space:]]Maps[[:space:]]for[[:space:]]3D[[:space:]]Shape[[:space:]]and[[:space:]]Pose[[:space:]]Reconstruction/126f70f5-3c25-4658-96da-89dd74c570b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DualTalk_[[:space:]]Dual-Speaker[[:space:]]Interaction[[:space:]]for[[:space:]]3D[[:space:]]Talking[[:space:]]Head[[:space:]]Conversations/1709bc2c-fbb2-4d37-9e23-6f06b89da286_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DyCON_[[:space:]]Dynamic[[:space:]]Uncertainty-aware[[:space:]]Consistency[[:space:]]and[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/b1207486-0534-4ae7-945d-8473b95cc3df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DyCoke_[[:space:]]Dynamic[[:space:]]Compression[[:space:]]of[[:space:]]Tokens[[:space:]]for[[:space:]]Fast[[:space:]]Video[[:space:]]Large[[:space:]]Language[[:space:]]Models/e008b780-79b0-4745-a3d1-8043a7c7b22a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DyFo_[[:space:]]A[[:space:]]Training-Free[[:space:]]Dynamic[[:space:]]Focus[[:space:]]Visual[[:space:]]Search[[:space:]]for[[:space:]]Enhancing[[:space:]]LMMs[[:space:]]in[[:space:]]Fine-Grained[[:space:]]Visual[[:space:]]Understanding/26a78ede-25a4-464c-91bc-a2e676c7b753_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DyMO_[[:space:]]Training-Free[[:space:]]Diffusion[[:space:]]Model[[:space:]]Alignment[[:space:]]with[[:space:]]Dynamic[[:space:]]Multi-Objective[[:space:]]Scheduling/53aa967e-00d3-4f7d-b70f-cfc84e05b64d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dyn-HaMR_[[:space:]]Recovering[[:space:]]4D[[:space:]]Interacting[[:space:]]Hand[[:space:]]Motion[[:space:]]from[[:space:]]a[[:space:]]Dynamic[[:space:]]Camera/80a0037d-d380-4248-8e04-b253867efda2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DynFocus_[[:space:]]Dynamic[[:space:]]Cooperative[[:space:]]Network[[:space:]]Empowers[[:space:]]LLMs[[:space:]]with[[:space:]]Video[[:space:]]Understanding/f57b2015-b4e9-4361-b00b-d1b4df4e0887_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DynPose_[[:space:]]Largely[[:space:]]Improving[[:space:]]the[[:space:]]Efficiency[[:space:]]of[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]by[[:space:]]a[[:space:]]Simple[[:space:]]Dynamic[[:space:]]Framework/5025e1f0-767c-48e3-a0f0-afe4b2e4a49e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DynRefer_[[:space:]]Delving[[:space:]]into[[:space:]]Region-level[[:space:]]Multimodal[[:space:]]Tasks[[:space:]]via[[:space:]]Dynamic[[:space:]]Resolution/34bf4203-a76a-4ec7-8239-378e6f6961ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DynScene_[[:space:]]Scalable[[:space:]]Generation[[:space:]]of[[:space:]]Dynamic[[:space:]]Robotic[[:space:]]Manipulation[[:space:]]Scenes[[:space:]]for[[:space:]]Embodied[[:space:]]AI/3f94df41-5770-4b41-a4f8-992d53b897aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DynaMoDe-NeRF_[[:space:]]Motion-aware[[:space:]]Deblurring[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field[[:space:]]for[[:space:]]Dynamic[[:space:]]Scenes/0973a7bb-83b0-44f5-b3f4-4770e05aeff7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Camera[[:space:]]Poses[[:space:]]and[[:space:]]Where[[:space:]]to[[:space:]]Find[[:space:]]Them/615bdf01-2f59-42e5-9c38-0faf0a49eaab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Content[[:space:]]Prediction[[:space:]]with[[:space:]]Motion-aware[[:space:]]Priors[[:space:]]for[[:space:]]Blind[[:space:]]Face[[:space:]]Video[[:space:]]Restoration/1f73aad6-035c-41e0-b4da-3331c16ddf1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Derivation[[:space:]]and[[:space:]]Elimination_[[:space:]]Audio[[:space:]]Visual[[:space:]]Segmentation[[:space:]]with[[:space:]]Enhanced[[:space:]]Audio[[:space:]]Semantics/a1bf7557-92e7-4edc-91c2-9633c3f4fc1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Group[[:space:]]Normalization_[[:space:]]Spatio-Temporal[[:space:]]Adaptation[[:space:]]to[[:space:]]Evolving[[:space:]]Data[[:space:]]Statistics/927b8e03-a564-43aa-886a-83946f0d90a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Integration[[:space:]]of[[:space:]]Task-Specific[[:space:]]Adapters[[:space:]]for[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning/d44d80f8-783b-4d73-bbc1-692b400afb75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Motion[[:space:]]Blending[[:space:]]for[[:space:]]Versatile[[:space:]]Motion[[:space:]]Editing/8338cd38-d1bc-44d4-81c3-ce8df03f3737_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Neural[[:space:]]Surfaces[[:space:]]for[[:space:]]Elastic[[:space:]]4D[[:space:]]Shape[[:space:]]Representation[[:space:]]and[[:space:]]Analysis/4ac71e2c-752a-4eb1-963b-6fc59e4156d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Pseudo[[:space:]]Labeling[[:space:]]via[[:space:]]Gradient[[:space:]]Cutting[[:space:]]for[[:space:]]High-Low[[:space:]]Entropy[[:space:]]Exploration/16b317ce-6e6f-4603-9593-81dae1882306_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Stereotype[[:space:]]Theory[[:space:]]Induced[[:space:]]Micro-expression[[:space:]]Recognition[[:space:]]with[[:space:]]Oriented[[:space:]]Deformation/330ddcc6-97ad-4cb2-9179-a9dda253e04b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Updates[[:space:]]for[[:space:]]Language[[:space:]]Adaptation[[:space:]]in[[:space:]]Visual-Language[[:space:]]Tracking/224e6c45-9bd9-4e27-bc19-559db94ae903_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DynamicScaler_[[:space:]]Seamless[[:space:]]and[[:space:]]Scalable[[:space:]]Video[[:space:]]Generation[[:space:]]for[[:space:]]Panoramic[[:space:]]Scenes/4811ebd5-bfde-4dc5-a338-798b56c04f8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EAP-GS_[[:space:]]Efficient[[:space:]]Augmentation[[:space:]]of[[:space:]]Pointcloud[[:space:]]for[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]in[[:space:]]Few-shot[[:space:]]Scene[[:space:]]Reconstruction/b763ea28-e188-47fe-8d63-d0669fcac57a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EASEMVC_Efficient[[:space:]]Dual[[:space:]]Selection[[:space:]]Mechanism[[:space:]]for[[:space:]]Deep[[:space:]]Multi-View[[:space:]]Clustering/4bf3234e-08c2-4c5a-9a94-3a1ebb84c03c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EBS-EKF_[[:space:]]Accurate[[:space:]]and[[:space:]]High[[:space:]]Frequency[[:space:]]Event-based[[:space:]]Star[[:space:]]Tracking/08b2d67e-fefe-4f61-a1de-38b8b6bd9092_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ECBench_[[:space:]]Can[[:space:]]Multi-modal[[:space:]]Foundation[[:space:]]Models[[:space:]]Understand[[:space:]]the[[:space:]]Egocentric[[:space:]]World_[[:space:]]A[[:space:]]Holistic[[:space:]]Embodied[[:space:]]Cognition[[:space:]]Benchmark/821dbf45-2126-4033-a280-65d8b022e939_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ECVC_[[:space:]]Exploiting[[:space:]]Non-Local[[:space:]]Correlations[[:space:]]in[[:space:]]Multiple[[:space:]]Frames[[:space:]]for[[:space:]]Contextual[[:space:]]Video[[:space:]]Compression/60441634-6ccc-4e04-8248-d8328047d5fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EDCFlow_[[:space:]]Exploring[[:space:]]Temporally[[:space:]]Dense[[:space:]]Difference[[:space:]]Maps[[:space:]]for[[:space:]]Event-based[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation/e8620442-de74-4f18-8ca9-ceef8a8448a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EDEN_[[:space:]]Enhanced[[:space:]]Diffusion[[:space:]]for[[:space:]]High-quality[[:space:]]Large-motion[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation/a6499f2b-81b5-4135-9546-59cc145d72cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EDM_[[:space:]]Equirectangular[[:space:]]Projection-Oriented[[:space:]]Dense[[:space:]]Kernelized[[:space:]]Feature[[:space:]]Matching/f5974cc9-c152-4230-82a7-5d78a803dd23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EEE-Bench_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Multimodal[[:space:]]Electrical[[:space:]]And[[:space:]]Electronics[[:space:]]Engineering[[:space:]]Benchmark/9fcd72d8-d952-40de-91d0-e2b4f2ce197a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EIDT-V_[[:space:]]Exploiting[[:space:]]Intersections[[:space:]]in[[:space:]]Diffusion[[:space:]]Trajectories[[:space:]]for[[:space:]]Model-Agnostic,[[:space:]]Zero-Shot,[[:space:]]Training-Free[[:space:]]Text-to-Video[[:space:]]Generation/840bfa2e-848d-49f4-8051-efae8bec43f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EMOE_[[:space:]]Modality-Specific[[:space:]]Enhanced[[:space:]]Dynamic[[:space:]]Emotion[[:space:]]Experts/eb0fe62b-4c37-4cb6-a97a-6d88a1905db6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EMOVA_[[:space:]]Empowering[[:space:]]Language[[:space:]]Models[[:space:]]to[[:space:]]See,[[:space:]]Hear[[:space:]]and[[:space:]]Speak[[:space:]]with[[:space:]]Vivid[[:space:]]Emotions/e9790db3-2b6b-4589-96ad-63d5dee63ff7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ERUPT_[[:space:]]Efficient[[:space:]]Rendering[[:space:]]with[[:space:]]Unposed[[:space:]]Patch[[:space:]]Transformer/cc46ae2a-87a1-49a2-aee2-0722824f9c35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ESCAPE_[[:space:]]Equivariant[[:space:]]Shape[[:space:]]Completion[[:space:]]via[[:space:]]Anchor[[:space:]]Point[[:space:]]Encoding/05f9ffbb-79be-4252-856b-bcbb959e7901_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ESC_[[:space:]]Erasing[[:space:]]Space[[:space:]]Concept[[:space:]]for[[:space:]]Knowledge[[:space:]]Deletion/b4376d51-9ce2-4f36-8582-4ccc31e87b92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ETAP_[[:space:]]Event-based[[:space:]]Tracking[[:space:]]of[[:space:]]Any[[:space:]]Point/23d5da3e-d7a8-48a4-a691-b1c9580f0a28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EVOS_[[:space:]]Efficient[[:space:]]Implicit[[:space:]]Neural[[:space:]]Training[[:space:]]via[[:space:]]EVOlutionary[[:space:]]Selector/8dc950d7-ef35-4b48-8530-89be0c289ca4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EVPGS_[[:space:]]Enhanced[[:space:]]View[[:space:]]Prior[[:space:]]Guidance[[:space:]]for[[:space:]]Splatting-based[[:space:]]Extrapolated[[:space:]]View[[:space:]]Synthesis/4eb71053-a0fc-4555-9dba-f69f6afcfe0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EVolSplat_[[:space:]]Efficient[[:space:]]Volume-based[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Urban[[:space:]]View[[:space:]]Synthesis/879718c6-b66d-4d4a-bb41-803f55692ffb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EZSR_[[:space:]]Event-based[[:space:]]Zero-Shot[[:space:]]Recognition/0639cc60-53e6-4fe1-98d2-936869092ca0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Early-Bird[[:space:]]Diffusion_[[:space:]]Investigating[[:space:]]and[[:space:]]Leveraging[[:space:]]Timestep-Aware[[:space:]]Early-Bird[[:space:]]Tickets[[:space:]]in[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Efficient[[:space:]]Training/83fc954b-bea1-4411-92c8-17797037d386_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EarthDial_[[:space:]]Turning[[:space:]]Multi-sensory[[:space:]]Earth[[:space:]]Observations[[:space:]]to[[:space:]]Interactive[[:space:]]Dialogues/59e69f64-1c75-4d77-a14f-e0dce6e2150a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Easy-editable[[:space:]]Image[[:space:]]Vectorization[[:space:]]with[[:space:]]Multi-layer[[:space:]]Multi-scale[[:space:]]Distributed[[:space:]]Visual[[:space:]]Feature[[:space:]]Embedding/cda023af-c65e-47eb-ac99-1c4cce4fb016_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EasyCraft_[[:space:]]A[[:space:]]Robust[[:space:]]and[[:space:]]Efficient[[:space:]]Framework[[:space:]]for[[:space:]]Automatic[[:space:]]Avatar[[:space:]]Crafting/303f4075-3c55-47c5-a507-856a75818043_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EasyHOI_[[:space:]]Unleashing[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]Large[[:space:]]Models[[:space:]]for[[:space:]]Reconstructing[[:space:]]Hand-Object[[:space:]]Interactions[[:space:]]in[[:space:]]the[[:space:]]Wild/8c622e8a-33b6-400d-9aee-8eb774d315a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EchoMatch_[[:space:]]Partial-to-Partial[[:space:]]Shape[[:space:]]Matching[[:space:]]via[[:space:]]Correspondence[[:space:]]Reflection/7a3c48f7-95c7-446b-9823-5172d80222c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EchoMimicV2_[[:space:]]Towards[[:space:]]Striking,[[:space:]]Simplified,[[:space:]]and[[:space:]]Semi-Body[[:space:]]Human[[:space:]]Animation/288a5554-9d81-4d12-bbd8-a666ffaecfb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Few-Shot[[:space:]]Class-Incremental[[:space:]]Learning[[:space:]]via[[:space:]]Training-Free[[:space:]]Bi-Level[[:space:]]Modality[[:space:]]Calibration/dbbdd019-448e-4ab4-b391-57c7f46b59cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Online[[:space:]]Continual[[:space:]]Learning[[:space:]]with[[:space:]]Plug-and-Play[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]and[[:space:]]Class-Conditional[[:space:]]Mixture[[:space:]]of[[:space:]]Discretization/3a332709-76df-43e2-b1a6-2e585b16d645_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Privacy-Utility[[:space:]]Trade-offs[[:space:]]to[[:space:]]Mitigate[[:space:]]Memorization[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/0d123934-7ce7-45d9-ad3f-707ece8265c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]SAM[[:space:]]with[[:space:]]Efficient[[:space:]]Prompting[[:space:]]and[[:space:]]Preference[[:space:]]Optimization[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/246f03d7-fa39-4c5a-a7fe-2aa5a7ce4b40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Testing-Time[[:space:]]Robustness[[:space:]]for[[:space:]]Trusted[[:space:]]Multi-View[[:space:]]Classification[[:space:]]in[[:space:]]the[[:space:]]Wild/5a9084f0-d242-4d6e-a2eb-6959dbb6b0a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Video-LLM[[:space:]]Reasoning[[:space:]]via[[:space:]]Agent-of-Thoughts[[:space:]]Distillation/9af6fd66-0efe-4a13-87b2-86d45a972105_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Virtual[[:space:]]Try-On[[:space:]]with[[:space:]]Synthetic[[:space:]]Pairs[[:space:]]and[[:space:]]Error-Aware[[:space:]]Noise[[:space:]]Scheduling/05159407-58d0-4826-9db8-7d0aa541ed98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Vision-Language[[:space:]]Compositional[[:space:]]Understanding[[:space:]]with[[:space:]]Multimodal[[:space:]]Synthetic[[:space:]]Data/653d9e05-cace-4615-bcbc-a402fd9e9c3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EnliveningGS_[[:space:]]Active[[:space:]]Locomotion[[:space:]]of[[:space:]]3DGS/678bd279-c773-4d3f-9e3f-a7dfe7dbc9fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EntityErasure_[[:space:]]Erasing[[:space:]]Entity[[:space:]]Cleanly[[:space:]]via[[:space:]]Amodal[[:space:]]Entity[[:space:]]Segmentation[[:space:]]and[[:space:]]Completion/841507d5-5b60-4a74-9926-0461296e0a84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EntitySAM_[[:space:]]Segment[[:space:]]Everything[[:space:]]in[[:space:]]Video/cb4f400f-a69b-43ea-b233-f7f126e62f4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EntropyMark_[[:space:]]Towards[[:space:]]More[[:space:]]Harmless[[:space:]]Backdoor[[:space:]]Watermark[[:space:]]via[[:space:]]Entropy-based[[:space:]]Constraint[[:space:]]for[[:space:]]Open-source[[:space:]]Dataset[[:space:]]Copyright[[:space:]]Protection/79eea005-2d9c-4999-9dbc-01e224e70254_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EnvGS_[[:space:]]Modeling[[:space:]]View-Dependent[[:space:]]Appearance[[:space:]]with[[:space:]]Environment[[:space:]]Gaussian/9aaa52c4-5f11-4ae5-bb15-8f551966ee84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EnvPoser_[[:space:]]Environment-aware[[:space:]]Realistic[[:space:]]Human[[:space:]]Motion[[:space:]]Estimation[[:space:]]from[[:space:]]Sparse[[:space:]]Observations[[:space:]]with[[:space:]]Uncertainty[[:space:]]Modeling/ca9b40de-cd3a-44e8-85b9-dec7c06efbee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EquiPose_[[:space:]]Exploiting[[:space:]]Permutation[[:space:]]Equivariance[[:space:]]for[[:space:]]Relative[[:space:]]Camera[[:space:]]Pose[[:space:]]Estimation/6aabd5e5-4c6d-47d9-bb8c-4cbc4163ed65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Erase[[:space:]]Diffusion_[[:space:]]Empowering[[:space:]]Object[[:space:]]Removal[[:space:]]Through[[:space:]]Calibrating[[:space:]]Diffusion[[:space:]]Pathways/eb77825c-6cbb-4c39-948d-52efc39ace4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Erasing[[:space:]]Undesirable[[:space:]]Influence[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/cfcabb09-71fe-461a-a521-c3d740140c91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Escaping[[:space:]]Plato's[[:space:]]Cave_[[:space:]]Towards[[:space:]]the[[:space:]]Alignment[[:space:]]of[[:space:]]3D[[:space:]]and[[:space:]]Text[[:space:]]Latent[[:space:]]Spaces/85c228c2-88b3-4b42-91cd-0b9c6fff9635_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Estimating[[:space:]]Body[[:space:]]and[[:space:]]Hand[[:space:]]Motion[[:space:]]in[[:space:]]an[[:space:]]Ego-sensed[[:space:]]World/71e1babb-034c-40be-8709-39f13bf4244a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ev-3DOD_[[:space:]]Pushing[[:space:]]the[[:space:]]Temporal[[:space:]]Boundaries[[:space:]]of[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Event[[:space:]]Cameras/e0858502-8bdd-48cc-ae6e-4de96b374d49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EvEnhancer_[[:space:]]Empowering[[:space:]]Effectiveness,[[:space:]]Efficiency[[:space:]]and[[:space:]]Generalizability[[:space:]]for[[:space:]]Continuous[[:space:]]Space-Time[[:space:]]Video[[:space:]]Super-Resolution[[:space:]]with[[:space:]]Events/84be2ed9-42d5-47be-92a0-ca505e0f2685_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EvOcc_[[:space:]]Accurate[[:space:]]Semantic[[:space:]]Occupancy[[:space:]]for[[:space:]]Automated[[:space:]]Driving[[:space:]]Using[[:space:]]Evidence[[:space:]]Theory/7c502132-611b-4140-a7e5-11680d62e1b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Eval3D_[[:space:]]Interpretable[[:space:]]and[[:space:]]Fine-grained[[:space:]]Evaluation[[:space:]]for[[:space:]]3D[[:space:]]Generation/ec87f8a0-6f47-4f99-9af3-0e617994c3ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Evaluating[[:space:]]Model[[:space:]]Perception[[:space:]]of[[:space:]]Color[[:space:]]Illusions[[:space:]]in[[:space:]]Photorealistic[[:space:]]Scenes/9bb16cc9-5b88-4a24-a8e1-499494d36b2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Evaluating[[:space:]]Vision-Language[[:space:]]Models[[:space:]]as[[:space:]]Evaluators[[:space:]]in[[:space:]]Path[[:space:]]Planning/ea2b9d98-7774-421e-b29a-893befcc86ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Event[[:space:]]Ellipsometer_[[:space:]]Event-based[[:space:]]Mueller-Matrix[[:space:]]Video[[:space:]]Imaging/00131384-06b2-4638-a55f-16332464dd84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Event[[:space:]]Fields_[[:space:]]Capturing[[:space:]]Light[[:space:]]Fields[[:space:]]at[[:space:]]High[[:space:]]Speed,[[:space:]]Resolution,[[:space:]]and[[:space:]]Dynamic[[:space:]]Range/b83374cf-6c00-4bd9-b132-43495f3f9c2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Event-Equalized[[:space:]]Dense[[:space:]]Video[[:space:]]Captioning/fd5415a0-125d-45cf-9033-a51e081e5f8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Event-based[[:space:]]Video[[:space:]]Super-Resolution[[:space:]]via[[:space:]]State[[:space:]]Space[[:space:]]Models/c8e5af22-690b-462c-ba33-5d416cd06ebe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EventFly_[[:space:]]Event[[:space:]]Camera[[:space:]]Perception[[:space:]]from[[:space:]]Ground[[:space:]]to[[:space:]]the[[:space:]]Sky/ecab4823-da56-483f-8321-47b8a1d3b698_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EventGPT_[[:space:]]Event[[:space:]]Stream[[:space:]]Understanding[[:space:]]with[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/bb8d5233-3517-4797-9367-5fe773bd874a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EventPSR_[[:space:]]Surface[[:space:]]Normal[[:space:]]and[[:space:]]Reflectance[[:space:]]Estimation[[:space:]]from[[:space:]]Photometric[[:space:]]Stereo[[:space:]]Using[[:space:]]an[[:space:]]Event[[:space:]]Camera/133da766-89f1-4c63-bf52-44b0090134e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EventSplat_[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Moving[[:space:]]Event[[:space:]]Cameras[[:space:]]for[[:space:]]Real-time[[:space:]]Rendering/5e6f3efa-e5cc-4708-8b09-713b537fa1a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Every[[:space:]]SAM[[:space:]]Drop[[:space:]]Counts_[[:space:]]Embracing[[:space:]]Semantic[[:space:]]Priors[[:space:]]for[[:space:]]Multi-Modality[[:space:]]Image[[:space:]]Fusion[[:space:]]and[[:space:]]Beyond/a7813684-e289-4dec-92c0-ee63fc512deb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Everything[[:space:]]to[[:space:]]the[[:space:]]Synthetic_[[:space:]]Diffusion-driven[[:space:]]Test-time[[:space:]]Adaptation[[:space:]]via[[:space:]]Synthetic-Domain[[:space:]]Alignment/fac30b3d-4109-4df5-93e5-c2263ebbddec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Evolving[[:space:]]High-Quality[[:space:]]Rendering[[:space:]]and[[:space:]]Reconstruction[[:space:]]in[[:space:]]a[[:space:]]Unified[[:space:]]Framework[[:space:]]with[[:space:]]Contribution-Adaptive[[:space:]]Regularization/00b4d7e3-bd83-465f-afc0-737dc88f6946_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exact_[[:space:]]Exploring[[:space:]]Space-Time[[:space:]]Perceptive[[:space:]]Clues[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Satellite[[:space:]]Image[[:space:]]Time[[:space:]]Series[[:space:]]Semantic[[:space:]]Segmentation/e1d2758f-cf15-4719-87c6-3d5649f01379_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ExpertAF_[[:space:]]Expert[[:space:]]Actionable[[:space:]]Feedback[[:space:]]from[[:space:]]Video/ccac2a98-a341-4697-825f-e838745145a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Explainable[[:space:]]Saliency_[[:space:]]Articulating[[:space:]]Reasoning[[:space:]]with[[:space:]]Contextual[[:space:]]Prioritization/5a9d4569-3627-48db-87ca-f6f74c038d6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Explaining[[:space:]]Domain[[:space:]]Shifts[[:space:]]in[[:space:]]Language_[[:space:]]Concept[[:space:]]Erasing[[:space:]]for[[:space:]]Interpretable[[:space:]]Image[[:space:]]Classification/5b5f023c-5e83-4e83-95d9-ae7c139631a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Explaining[[:space:]]in[[:space:]]Diffusion_[[:space:]]Explaining[[:space:]]a[[:space:]]Classifier[[:space:]]with[[:space:]]Diffusion[[:space:]]Semantics/e12e2fee-c2c0-4bbc-99f8-f5d2cb6dbe06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Explicit[[:space:]]Depth-Aware[[:space:]]Blurry[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation[[:space:]]Guided[[:space:]]by[[:space:]]Differential[[:space:]]Curves/335d79cd-ff57-4d73-aab7-e080ed86c5a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploiting[[:space:]]Deblurring[[:space:]]Networks[[:space:]]for[[:space:]]Radiance[[:space:]]Fields/2b1465fe-6939-4acc-92a7-0a9426b9c956_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploiting[[:space:]]Temporal[[:space:]]State[[:space:]]Space[[:space:]]Sharing[[:space:]]for[[:space:]]Video[[:space:]]Semantic[[:space:]]Segmentation/205ea7bc-4216-405a-b9ee-ef9433534ad4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploration-Driven[[:space:]]Generative[[:space:]]Interactive[[:space:]]Environments/4c28c31d-7f28-4419-9de8-e3891fd9cbcc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]CLIP's[[:space:]]Dense[[:space:]]Knowledge[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/051b2c9f-5a13-4131-b880-e276059baaff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Contextual[[:space:]]Attribute[[:space:]]Density[[:space:]]in[[:space:]]Referring[[:space:]]Expression[[:space:]]Counting/f78a8ba2-f99c-43c8-aadb-f099a2092f51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Historical[[:space:]]Information[[:space:]]for[[:space:]]RGBE[[:space:]]Visual[[:space:]]Tracking[[:space:]]with[[:space:]]Mamba/36e76161-ac87-43e3-905e-9c2eccb37ec2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Intrinsic[[:space:]]Normal[[:space:]]Prototypes[[:space:]]within[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]for[[:space:]]Universal[[:space:]]Anomaly[[:space:]]Detection/8e301c2a-3c32-4f9e-8f87-33596952eadb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Scene[[:space:]]Affinity[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]LiDAR[[:space:]]Semantic[[:space:]]Segmentation/8e486940-e51f-442d-a6a0-a73d2630afa0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Semantic[[:space:]]Feature[[:space:]]Discrimination[[:space:]]for[[:space:]]Perceptual[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]and[[:space:]]Opinion-Unaware[[:space:]]No-Reference[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment/cde8052d-f8eb-472e-b000-fd4d976b6a20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Simple[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/495afadb-de9a-431f-ba2d-9db6ebb6d17d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Sparse[[:space:]]MoE[[:space:]]in[[:space:]]GANs[[:space:]]for[[:space:]]Text-conditioned[[:space:]]Image[[:space:]]Synthesis/7078b157-7879-47cb-991a-09f81047db7a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Temporally-Aware[[:space:]]Features[[:space:]]for[[:space:]]Point[[:space:]]Tracking/ac98a5e9-90fa-4012-89c2-06839140a6cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Timeline[[:space:]]Control[[:space:]]for[[:space:]]Facial[[:space:]]Motion[[:space:]]Generation/65b84757-8067-4189-a894-056e4be1d124_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Visual[[:space:]]Vulnerabilities[[:space:]]via[[:space:]]Multi-Loss[[:space:]]Adversarial[[:space:]]Search[[:space:]]for[[:space:]]Jailbreaking[[:space:]]Vision-Language[[:space:]]Models/0fb20cd3-12d6-4733-afec-1c9df735d84b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]the[[:space:]]Deep[[:space:]]Fusion[[:space:]]of[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]and[[:space:]]Diffusion[[:space:]]Transformers[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Synthesis/3434d4fb-2873-49b0-8ac0-297681602f31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exposure-slot_[[:space:]]Exposure-centric[[:space:]]Representations[[:space:]]Learning[[:space:]]with[[:space:]]Slot-in-Slot[[:space:]]Attention[[:space:]]for[[:space:]]Region-aware[[:space:]]Exposure[[:space:]]Correction/83f68ecd-43a7-49b5-b8e2-1ddf3578f781_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Extrapolating[[:space:]]and[[:space:]]Decoupling[[:space:]]Image-to-Video[[:space:]]Generation[[:space:]]Models_[[:space:]]Motion[[:space:]]Modeling[[:space:]]is[[:space:]]Easier[[:space:]]Than[[:space:]]You[[:space:]]Think/34056abe-9c83-4a9a-ac86-67d3126b13e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Extreme[[:space:]]Rotation[[:space:]]Estimation[[:space:]]in[[:space:]]the[[:space:]]Wild/1bd2da5f-45ff-4055-bc43-55a1a6080681_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/F-LMM_[[:space:]]Grounding[[:space:]]Frozen[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models/6e34bb31-55f8-4e82-8db1-d754fa78afe2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FADA_[[:space:]]Fast[[:space:]]Diffusion[[:space:]]Avatar[[:space:]]Synthesis[[:space:]]with[[:space:]]Mixed-Supervised[[:space:]]Multi-CFG[[:space:]]Distillation/eae624fb-184e-49bf-afbc-d98a98aeb9b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FADE_[[:space:]]Frequency-Aware[[:space:]]Diffusion[[:space:]]Model[[:space:]]Factorization[[:space:]]for[[:space:]]Video[[:space:]]Editing/37abe74e-aca7-48e5-911c-4b7009fcf3ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FALCON_[[:space:]]Fairness[[:space:]]Learning[[:space:]]via[[:space:]]Contrastive[[:space:]]Attention[[:space:]]Approach[[:space:]]to[[:space:]]Continual[[:space:]]Semantic[[:space:]]Scene[[:space:]]Understanding/2e34452d-f7d4-4d46-8b4f-8391c1e12974_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HoVLE_[[:space:]]Unleashing[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]Monolithic[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]Holistic[[:space:]]Vision-Language[[:space:]]Embedding/126721ba-0c45-4655-8cc5-9017b39955e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Holmes-VAU_[[:space:]]Towards[[:space:]]Long-term[[:space:]]Video[[:space:]]Anomaly[[:space:]]Understanding[[:space:]]at[[:space:]]Any[[:space:]]Granularity/8b3a9750-2575-4fe7-bb78-a5c96dc5aac4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HomoGen_[[:space:]]Enhanced[[:space:]]Video[[:space:]]Inpainting[[:space:]]via[[:space:]]Homography[[:space:]]Propagation[[:space:]]and[[:space:]]Diffusion/c0acdc17-6ead-4f9d-8631-4f1535650ac6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Homogeneous[[:space:]]Dynamics[[:space:]]Space[[:space:]]for[[:space:]]Heterogeneous[[:space:]]Humans/a999c957-db24-46d0-83c3-e6da7809465e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Horizon-GS_[[:space:]]Unified[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Large-Scale[[:space:]]Aerial-to-Ground[[:space:]]Scenes/65e9755f-fe8e-4a32-88df-f20ee7cf9c9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HotSpot_[[:space:]]Signed[[:space:]]Distance[[:space:]]Function[[:space:]]Optimization[[:space:]]with[[:space:]]an[[:space:]]Asymptotically[[:space:]]Sufficient[[:space:]]Condition/733f46a7-f78c-4e53-bba4-ead7eb705bb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/How[[:space:]]Do[[:space:]]I[[:space:]]Do[[:space:]]That_[[:space:]]Synthesizing[[:space:]]3D[[:space:]]Hand[[:space:]]Motion[[:space:]]and[[:space:]]Contacts[[:space:]]for[[:space:]]Everyday[[:space:]]Interactions/4f85bf33-0c00-4386-bb62-720463efe772_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/How[[:space:]]to[[:space:]]Merge[[:space:]]Your[[:space:]]Multimodal[[:space:]]Models[[:space:]]Over[[:space:]]Time_/76e2370a-2b50-436a-9bdc-9b8af8c8e7ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HuMoCon_[[:space:]]Concept[[:space:]]Discovery[[:space:]]for[[:space:]]Human[[:space:]]Motion[[:space:]]Understanding/b5dd425e-4f3e-4b2d-9c88-36e4f0bafd85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HuPerFlow_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Benchmark[[:space:]]for[[:space:]]Human[[:space:]]vs.[[:space:]]Machine[[:space:]]Motion[[:space:]]Estimation[[:space:]]Comparison/5a790d2d-155e-42dc-bc07-eb0a483c9c11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Human[[:space:]]Motion[[:space:]]Instruction[[:space:]]Tuning/8c08c8b0-9aae-431a-b68e-699a5ddd5c32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Human-centered[[:space:]]Interactive[[:space:]]Learning[[:space:]]via[[:space:]]MLLMs[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Person[[:space:]]Re-identification/5cc75be5-e122-4a4f-ba40-7450ab436d6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HumanDreamer_[[:space:]]Generating[[:space:]]Controllable[[:space:]]Human-Motion[[:space:]]Videos[[:space:]]via[[:space:]]Decoupled[[:space:]]Generation/8f919b44-e98e-44f2-a774-268fc39345bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HumanMM_[[:space:]]Global[[:space:]]Human[[:space:]]Motion[[:space:]]Recovery[[:space:]]from[[:space:]]Multi-shot[[:space:]]Videos/8ae05aea-0260-4aa4-bbb0-352a7f188bbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HumanRig_[[:space:]]Learning[[:space:]]Automatic[[:space:]]Rigging[[:space:]]for[[:space:]]Humanoid[[:space:]]Character[[:space:]]in[[:space:]]a[[:space:]]Large[[:space:]]Scale[[:space:]]Dataset/a5721584-3cc2-4564-a9c0-b5eefdb73fbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HunyuanPortrait_[[:space:]]Implicit[[:space:]]Condition[[:space:]]Control[[:space:]]for[[:space:]]Enhanced[[:space:]]Portrait[[:space:]]Animation/a6cc1e88-c49b-4877-897d-6a6632aadecd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hybrid[[:space:]]Concept[[:space:]]Bottleneck[[:space:]]Models/7a729be8-7e97-4b64-9b6a-2c339ff64815_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hybrid[[:space:]]Global-Local[[:space:]]Representation[[:space:]]with[[:space:]]Augmented[[:space:]]Spatial[[:space:]]Guidance[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation/9c88ce4f-9647-4320-bdd2-f6fc8a88e98e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hybrid[[:space:]]Reciprocal[[:space:]]Transformer[[:space:]]with[[:space:]]Triplet[[:space:]]Feature[[:space:]]Alignment[[:space:]]for[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/11e7a635-4306-4d78-8190-6be9b1f3acac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hybrid-Level[[:space:]]Instruction[[:space:]]Injection[[:space:]]for[[:space:]]Video[[:space:]]Token[[:space:]]Compression[[:space:]]in[[:space:]]Multi-modal[[:space:]]Large[[:space:]]Language[[:space:]]Models/a0ce533d-bc87-47ce-aab6-707ed1ae09bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HybridGS_[[:space:]]Decoupling[[:space:]]Transients[[:space:]]and[[:space:]]Statics[[:space:]]with[[:space:]]2D[[:space:]]and[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/347531dd-b02e-491d-be81-0620c6e2bfcd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HybridMQA_[[:space:]]Exploring[[:space:]]Geometry-Texture[[:space:]]Interactions[[:space:]]for[[:space:]]Colored[[:space:]]Mesh[[:space:]]Quality[[:space:]]Assessment/a45af5e0-e0d7-4402-9367-2386d583248b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HyperFree_[[:space:]]A[[:space:]]Channel-adaptive[[:space:]]and[[:space:]]Tuning-free[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Hyperspectral[[:space:]]Remote[[:space:]]Sensing[[:space:]]Imagery/9236da20-05f2-4327-b226-ca754d2dead5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HyperGLM_[[:space:]]HyperGraph[[:space:]]for[[:space:]]Video[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]and[[:space:]]Anticipation/07619f38-6a65-46d1-ae2a-3885ad0e2f6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HyperGS_[[:space:]]Hyperspectral[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/8e6d9c3d-45db-41da-98c2-e1299df7de27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HyperLoRA_[[:space:]]Parameter-Efficient[[:space:]]Adaptive[[:space:]]Generation[[:space:]]for[[:space:]]Portrait[[:space:]]Synthesis/0dcd811d-bb8b-4fcc-9377-9711c569d3e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HyperNVD_[[:space:]]Accelerating[[:space:]]Neural[[:space:]]Video[[:space:]]Decomposition[[:space:]]via[[:space:]]Hypernetworks/f624369a-7a34-4be6-8ddf-7450e2fca80c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HyperNet[[:space:]]Fields_[[:space:]]Efficiently[[:space:]]Training[[:space:]]Hypernetworks[[:space:]]without[[:space:]]Ground[[:space:]]Truth[[:space:]]by[[:space:]]Learning[[:space:]]Weight[[:space:]]Trajectories/68fd503f-6b57-47fa-8aca-593407ee13b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HyperPose_[[:space:]]Hypernetwork-Infused[[:space:]]Camera[[:space:]]Pose[[:space:]]Localization[[:space:]]and[[:space:]]an[[:space:]]Extended[[:space:]]Cambridge[[:space:]]Landmarks[[:space:]]Dataset/dd4d95b8-93a1-4dc9-acfe-357b62a866ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HyperSeg_[[:space:]]Hybrid[[:space:]]Segmentation[[:space:]]Assistant[[:space:]]with[[:space:]]Fine-grained[[:space:]]Visual[[:space:]]Perceiver/d3d1df0c-7574-4e1e-b447-81df1bf28440_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hyperbolic[[:space:]]Category[[:space:]]Discovery/103e7fd4-0c6b-413e-8aae-04fa73b4770b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hyperbolic[[:space:]]Safety-Aware[[:space:]]Vision-Language[[:space:]]Models/5ca9a6d6-45f1-4b74-b091-1c79b0b891b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hyperbolic[[:space:]]Uncertainty-Aware[[:space:]]Few-Shot[[:space:]]Incremental[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation/1247e066-a517-4ffe-b97e-a9bff43b89c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hyperdimensional[[:space:]]Uncertainty[[:space:]]Quantification[[:space:]]for[[:space:]]Multimodal[[:space:]]Uncertainty[[:space:]]Fusion[[:space:]]in[[:space:]]Autonomous[[:space:]]Vehicles[[:space:]]Perception/8d8555dd-a579-45c4-ad75-fee4bbfad5af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hypergraph[[:space:]]Vision[[:space:]]Transformers_[[:space:]]Images[[:space:]]are[[:space:]]More[[:space:]]than[[:space:]]Nodes,[[:space:]]More[[:space:]]than[[:space:]]Edges/b3c66b72-bf5c-47e0-a296-e834f5e49c0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hyperspectral[[:space:]]Pansharpening[[:space:]]via[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Iteratively[[:space:]]Zero-Shot[[:space:]]Guidance/f79e03e7-ea12-4e70-879c-79258b8c5e82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/I2VGuard_[[:space:]]Safeguarding[[:space:]]Images[[:space:]]against[[:space:]]Misuse[[:space:]]in[[:space:]]Diffusion-based[[:space:]]Image-to-Video[[:space:]]Models/612ebd29-cda5-46e3-b035-7c7393e6f819_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IAAO_[[:space:]]Interactive[[:space:]]Affordance[[:space:]]Learning[[:space:]]for[[:space:]]Articulated[[:space:]]Objects[[:space:]]in[[:space:]]3D[[:space:]]Environments/c94d31b4-1877-4bc2-8b0f-9d4d8e45806a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ICE_[[:space:]]Intrinsic[[:space:]]Concept[[:space:]]Extraction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]via[[:space:]]Diffusion[[:space:]]Models/e9edff49-034e-4138-82ed-91d11ac96969_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ICP_[[:space:]]Immediate[[:space:]]Compensation[[:space:]]Pruning[[:space:]]for[[:space:]]Mid-to-high[[:space:]]Sparsity/9ac0c560-9a5b-4372-9074-16698b23f1e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ICT_[[:space:]]Image-Object[[:space:]]Cross-Level[[:space:]]Trusted[[:space:]]Intervention[[:space:]]for[[:space:]]Mitigating[[:space:]]Object[[:space:]]Hallucination[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/a90828a4-8b1b-4072-b5af-bed317d47267_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ID-Patch_[[:space:]]Robust[[:space:]]ID[[:space:]]Association[[:space:]]for[[:space:]]Group[[:space:]]Photo[[:space:]]Personalization/a37d3ef3-36c1-46f2-886c-58bf6589bcc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IDEA-Bench_[[:space:]]How[[:space:]]Far[[:space:]]are[[:space:]]Generative[[:space:]]Models[[:space:]]from[[:space:]]Professional[[:space:]]Designing_/970924c3-54e6-4027-b252-42ee55ce3b43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IDEA_[[:space:]]Inverted[[:space:]]Text[[:space:]]with[[:space:]]Cooperative[[:space:]]Deformable[[:space:]]Aggregation[[:space:]]for[[:space:]]Multi-modal[[:space:]]Object[[:space:]]Re-Identification/b7176393-d8d0-4ced-ba0c-43acd09bc650_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IDOL_[[:space:]]Instant[[:space:]]Photorealistic[[:space:]]3D[[:space:]]Human[[:space:]]Creation[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/c28e8ffb-3109-4ced-b9e1-553dbeace189_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IDProtector_[[:space:]]An[[:space:]]Adversarial[[:space:]]Noise[[:space:]]Encoder[[:space:]]to[[:space:]]Protect[[:space:]]Against[[:space:]]ID-Preserving[[:space:]]Image[[:space:]]Generation/cbe88f8c-13ab-4200-b1b9-bb453b62ad1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ILIAS_[[:space:]]Instance-Level[[:space:]]Image[[:space:]]retrieval[[:space:]]At[[:space:]]Scale/bd262cec-535c-47b6-b08f-5430c3f91bd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IM-Portrait_[[:space:]]Learning[[:space:]]3D-aware[[:space:]]Video[[:space:]]Diffusion[[:space:]]for[[:space:]]Photorealistic[[:space:]]Talking[[:space:]]Heads[[:space:]]from[[:space:]]Monocular[[:space:]]VideosC/d89a7508-cd12-421c-9424-93ad28adceb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IM-Zero_[[:space:]]Instance-level[[:space:]]Motion[[:space:]]Controllable[[:space:]]Video[[:space:]]Generation[[:space:]]in[[:space:]]a[[:space:]]Zero-shot[[:space:]]Manner/a669fc0a-17c9-4001-97e0-7737b0fc24de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IMFine_[[:space:]]3D[[:space:]]Inpainting[[:space:]]via[[:space:]]Geometry-guided[[:space:]]Multi-view[[:space:]]Refinement/0becb03e-65d1-4a50-a09c-58afb744258c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/INFP_[[:space:]]Audio-Driven[[:space:]]Interactive[[:space:]]Head[[:space:]]Generation[[:space:]]in[[:space:]]Dyadic[[:space:]]Conversations/4440ddb1-10e9-45a7-942b-205ac25fbaaa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IRGS_[[:space:]]Inter-Reflective[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]2D[[:space:]]Gaussian[[:space:]]Ray[[:space:]]Tracing/ff8cd318-c293-446d-b3d4-b22f9fc186ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IRIS_[[:space:]]Inverse[[:space:]]Rendering[[:space:]]of[[:space:]]Indoor[[:space:]]Scenes[[:space:]]from[[:space:]]Low[[:space:]]Dynamic[[:space:]]Range[[:space:]]Images/53c93036-c714-4f00-9960-2f72fce37288_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ITA-MDT_[[:space:]]Image-Timestep-Adaptive[[:space:]]Masked[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]Framework[[:space:]]for[[:space:]]Image-Based[[:space:]]Virtual[[:space:]]Try-On/e6b13099-74a4-4157-8bc0-2e7511ad889f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IceDiff_[[:space:]]High[[:space:]]Resolution[[:space:]]and[[:space:]]High-Quality[[:space:]]Arctic[[:space:]]Sea[[:space:]]Ice[[:space:]]Forecasting[[:space:]]with[[:space:]]Generative[[:space:]]Diffusion[[:space:]]Prior/3f2f5e81-8b45-4fda-b37e-2ea2e3413d80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Identifying[[:space:]]and[[:space:]]Mitigating[[:space:]]Position[[:space:]]Bias[[:space:]]of[[:space:]]Multi-image[[:space:]]Vision-Language[[:space:]]Models/17ea1e19-b848-462e-8296-051308ed2dd4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Identifying[[:space:]]and[[:space:]]Mitigating[[:space:]]Spurious[[:space:]]Correlation[[:space:]]in[[:space:]]Multi-Task[[:space:]]Learning/67faa86d-7895-441b-b372-45b9fee04763_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Identity-Clothing[[:space:]]Similarity[[:space:]]Modeling[[:space:]]for[[:space:]]Unsupervised[[:space:]]Clothing[[:space:]]Change[[:space:]]Person[[:space:]]Re-Identification/42258662-090f-4600-a2a7-5bf4fc16321d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Identity-Preserving[[:space:]]Text-to-Video[[:space:]]Generation[[:space:]]by[[:space:]]Frequency[[:space:]]Decomposition/16b4d9e1-d6f9-4df5-9c6a-3949d63cad2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Identity-preserving[[:space:]]Distillation[[:space:]]Sampling[[:space:]]by[[:space:]]Fixed-Point[[:space:]]Iterator/8dc89aa9-e649-490d-ba0b-e87a936deb98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Illumination[[:space:]]Spectrum[[:space:]]Estimation[[:space:]]for[[:space:]]Multispectral[[:space:]]Images[[:space:]]via[[:space:]]Surface[[:space:]]Reflectance[[:space:]]Modeling[[:space:]]and[[:space:]]Spatial-Spectral[[:space:]]Feature[[:space:]]Generation/afff26f0-40a0-4b11-8350-630ae1c85fad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ImViD_[[:space:]]Immersive[[:space:]]Volumetric[[:space:]]Videos[[:space:]]for[[:space:]]Enhanced[[:space:]]VR[[:space:]]Engagement/6da376dc-6436-436c-b26c-23ba201328ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Image[[:space:]]Generation[[:space:]]Diversity[[:space:]]Issues[[:space:]]and[[:space:]]How[[:space:]]to[[:space:]]Tame[[:space:]]Them/7493d326-112b-4a6d-8c2c-7e4449ad1384_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Image[[:space:]]Over[[:space:]]Text_[[:space:]]Transforming[[:space:]]Formula[[:space:]]Recognition[[:space:]]Evaluation[[:space:]]with[[:space:]]Character[[:space:]]Detection[[:space:]]Matching/e8290078-d6a1-49c9-b28d-cb4de981b7cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MultiMorph_[[:space:]]On-demand[[:space:]]Atlas[[:space:]]Construction/835f5fd5-22f7-4738-971d-e141bd5ed0b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MultiVENT[[:space:]]2.0_[[:space:]]A[[:space:]]Massive[[:space:]]Multilingual[[:space:]]Benchmark[[:space:]]for[[:space:]]Event-Centric[[:space:]]Video[[:space:]]Retrieval/9a33fad7-0631-4be8-9fe9-44e2f2811359_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multimodal[[:space:]]Autoregressive[[:space:]]Pre-training[[:space:]]of[[:space:]]Large[[:space:]]Vision[[:space:]]Encoders/bd04863f-fe58-46af-911f-4262f00ee080_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MultimodalStudio_[[:space:]]A[[:space:]]Heterogeneous[[:space:]]Sensor[[:space:]]Dataset[[:space:]]and[[:space:]]Framework[[:space:]]for[[:space:]]Neural[[:space:]]Rendering[[:space:]]across[[:space:]]Multiple[[:space:]]Imaging[[:space:]]Modalities/1ee13744-0c4e-4510-86cf-8756a6bb7ece_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multiple[[:space:]]Object[[:space:]]Tracking[[:space:]]as[[:space:]]ID[[:space:]]Prediction/a72861e2-8977-4616-abd9-6fe8e380f09b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multirate[[:space:]]Neural[[:space:]]Image[[:space:]]Compression[[:space:]]with[[:space:]]Adaptive[[:space:]]Lattice[[:space:]]Vector[[:space:]]Quantization/a762ceac-a933-46dc-909c-a5ef12fea627_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multitwine_[[:space:]]Multi-Object[[:space:]]Compositing[[:space:]]with[[:space:]]Text[[:space:]]and[[:space:]]Layout[[:space:]]Control/721943ec-00c9-44be-a7d4-e8a4c4837af5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NADER_[[:space:]]Neural[[:space:]]Architecture[[:space:]]Design[[:space:]]via[[:space:]]Multi-Agent[[:space:]]Collaboration/ce8827da-787f-49a5-9f7f-26a77a631708_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NLPrompt_[[:space:]]Noise-Label[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/350b5e17-deda-4b15-885e-4218a3f1d9be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NN-Former_[[:space:]]Rethinking[[:space:]]Graph[[:space:]]Structure[[:space:]]in[[:space:]]Neural[[:space:]]Architecture[[:space:]]Representation/c7820583-1345-4c6c-bd8f-487b62090764_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NSD-Imagery_[[:space:]]A[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]for[[:space:]]Extending[[:space:]]fMRI[[:space:]]Vision[[:space:]]Decoding[[:space:]]Methods[[:space:]]to[[:space:]]Mental[[:space:]]Imagery/3e0e4b40-1cae-42fd-bfe0-2df7ed05ef31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NTClick_[[:space:]]Achieving[[:space:]]Precise[[:space:]]Interactive[[:space:]]Segmentation[[:space:]]With[[:space:]]Noise-tolerant[[:space:]]Clicks/6af3af96-9584-4019-b8ec-66519b8feede_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NTR-Gaussian_[[:space:]]Nighttime[[:space:]]Dynamic[[:space:]]Thermal[[:space:]]Reconstruction[[:space:]]with[[:space:]]4D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Based[[:space:]]on[[:space:]]Thermodynamics/7d1be50c-0d43-4ce3-94dc-2fb383e95f0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NVComposer_[[:space:]]Boosting[[:space:]]Generative[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]Multiple[[:space:]]Sparse[[:space:]]and[[:space:]]Unposed[[:space:]]Images/244660d5-a0b9-4598-b7a2-27194551391d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NVILA_[[:space:]]Efficient[[:space:]]Frontier[[:space:]]Visual[[:space:]]Language[[:space:]]Models/c9535793-1425-420b-a91c-d7c12dc6c285_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Narrating[[:space:]]the[[:space:]]Video_[[:space:]]Boosting[[:space:]]Text-Video[[:space:]]Retrieval[[:space:]]via[[:space:]]Comprehensive[[:space:]]Utilization[[:space:]]of[[:space:]]Frame-Level[[:space:]]Captions/0ada88c3-72aa-45d0-ba21-7249082574ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Navigating[[:space:]]Image[[:space:]]Restoration[[:space:]]with[[:space:]]VAR's[[:space:]]Distribution[[:space:]]Alignment[[:space:]]Prior/357706c2-a9de-4a5e-99b1-d7de185447a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Navigating[[:space:]]the[[:space:]]Unseen_[[:space:]]Zero-shot[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]via[[:space:]]Capsule-Based[[:space:]]Equivariant[[:space:]]Features/bb5af2ab-9f34-48a9-80da-3f64ea17d26d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Navigation[[:space:]]World[[:space:]]Models/1d8b79db-6b0a-4b1e-ab3c-29578159b0af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NeISF++_[[:space:]]Neural[[:space:]]Incident[[:space:]]Stokes[[:space:]]Field[[:space:]]for[[:space:]]Polarized[[:space:]]Inverse[[:space:]]Rendering[[:space:]]of[[:space:]]Conductors[[:space:]]and[[:space:]]Dielectrics/1cd7366d-ab0e-4bb1-9662-17926cde5f67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NeRFPrior_[[:space:]]Learning[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field[[:space:]]as[[:space:]]a[[:space:]]Prior[[:space:]]for[[:space:]]Indoor[[:space:]]Scene[[:space:]]Reconstruction/47f0c3fb-83fc-4500-935a-f7346bce47ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Nearly[[:space:]]Zero-Cost[[:space:]]Protection[[:space:]]Against[[:space:]]Mimicry[[:space:]]by[[:space:]]Personalized[[:space:]]Diffusion[[:space:]]Models/3161eef3-d869-477e-8e91-7c8e0c4d13c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NeighborRetr_[[:space:]]Balancing[[:space:]]Hub[[:space:]]Centrality[[:space:]]in[[:space:]]Cross-Modal[[:space:]]Retrieval/811eac68-77ba-4a5d-9f92-d74d2b733947_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Nested[[:space:]]Diffusion[[:space:]]Models[[:space:]]Using[[:space:]]Hierarchical[[:space:]]Latent[[:space:]]Priors/56d79977-51d8-4658-8026-8d4bb4b804ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neural[[:space:]]Hierarchical[[:space:]]Decomposition[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Plant[[:space:]]Modeling/a3b9505e-68ef-4522-a53d-301939ad9e1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neural[[:space:]]Inverse[[:space:]]Rendering[[:space:]]from[[:space:]]Propagating[[:space:]]Light/b673dcdb-f055-4a8f-a4ae-e53623c0d890_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neural[[:space:]]LightRig_[[:space:]]Unlocking[[:space:]]Accurate[[:space:]]Object[[:space:]]Normal[[:space:]]and[[:space:]]Material[[:space:]]Estimation[[:space:]]with[[:space:]]Multi-Light[[:space:]]Diffusion/94ba4345-34b1-42d2-8ae9-1c7a26f89b8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neural[[:space:]]Motion[[:space:]]Simulator[[:space:]]Pushing[[:space:]]the[[:space:]]Limit[[:space:]]of[[:space:]]World[[:space:]]Models[[:space:]]in[[:space:]]Reinforcement[[:space:]]Learning/9dffe2f7-7566-4879-8a99-78a5e950d1d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neural[[:space:]]Video[[:space:]]Compression[[:space:]]with[[:space:]]Context[[:space:]]Modulation/68ebd7f3-0bd4-4895-a5bc-f4b5d8506935_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neuro-3D_[[:space:]]Towards[[:space:]]3D[[:space:]]Visual[[:space:]]Decoding[[:space:]]from[[:space:]]EEG[[:space:]]Signals/c64ff789-e05d-48c0-86cb-c1c8e9129eb8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neuro-Symbolic[[:space:]]Evaluation[[:space:]]of[[:space:]]Text-to-Video[[:space:]]Models[[:space:]]using[[:space:]]Formal[[:space:]]Verification/d3304ca8-5247-4485-880b-12e227bde66e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neuron_[[:space:]]Learning[[:space:]]Context-Aware[[:space:]]Evolving[[:space:]]Representations[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Skeleton[[:space:]]Action[[:space:]]Recognition/cc8eef47-4ad9-4bac-8642-1b4f3dd721b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NexusGS_[[:space:]]Sparse[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]Epipolar[[:space:]]Depth[[:space:]]Priors[[:space:]]in[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/04eacff3-9df6-4753-bcf4-0c16ed6b954d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NightAdapter_[[:space:]]Learning[[:space:]]a[[:space:]]Frequency[[:space:]]Adapter[[:space:]]for[[:space:]]Generalizable[[:space:]]Night-time[[:space:]]Scene[[:space:]]Segmentation/62fc540b-c931-426a-9335-f9c1fbd1f478_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NitroFusion_[[:space:]]High-Fidelity[[:space:]]Single-Step[[:space:]]Diffusion[[:space:]]through[[:space:]]Dynamic[[:space:]]Adversarial[[:space:]]Training/e996a680-5b32-45a3-8e03-95ee7a8350c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/No[[:space:]]Pains,[[:space:]]More[[:space:]]Gains_[[:space:]]Recycling[[:space:]]Sub-Salient[[:space:]]Patches[[:space:]]for[[:space:]]Efficient[[:space:]]High-Resolution[[:space:]]Image[[:space:]]Recognition/3c65c013-ed70-4936-a74b-821d69c0f162_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/No[[:space:]]Thing,[[:space:]]Nothing_[[:space:]]Highlighting[[:space:]]Safety-Critical[[:space:]]Classes[[:space:]]for[[:space:]]Robust[[:space:]]LiDAR[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]in[[:space:]]Adverse[[:space:]]Weather/19c9259f-be87-45c4-8c08-791f7f73ece6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NoPain_[[:space:]]No-box[[:space:]]Point[[:space:]]Cloud[[:space:]]Attack[[:space:]]via[[:space:]]Optimal[[:space:]]Transport[[:space:]]Singular[[:space:]]Boundary/12a4bef7-a869-4e6b-80a8-0f37ecd7a10f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NoT_[[:space:]]Federated[[:space:]]Unlearning[[:space:]]via[[:space:]]Weight[[:space:]]Negation/613bfbbe-c215-492b-8f03-02b76f910af6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Noise[[:space:]]Calibration[[:space:]]and[[:space:]]Spatial-Frequency[[:space:]]Interactive[[:space:]]Network[[:space:]]for[[:space:]]STEM[[:space:]]Image[[:space:]]Enhancement/346c8d6e-bd40-4a50-929d-82a10aabc4c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Noise[[:space:]]Diffusion[[:space:]]for[[:space:]]Enhancing[[:space:]]Semantic[[:space:]]Faithfulness[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Synthesis/00980f73-7bcb-4c8e-b086-7e4b2c1734ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Noise[[:space:]]Modeling[[:space:]]in[[:space:]]One[[:space:]]Hour_[[:space:]]Minimizing[[:space:]]Preparation[[:space:]]Efforts[[:space:]]for[[:space:]]Self-supervised[[:space:]]Low-Light[[:space:]]RAW[[:space:]]Image[[:space:]]Denoising/2558657f-b3f0-4cb2-b34a-ef535f51ed9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Noise-Consistent[[:space:]]Siamese-Diffusion[[:space:]]for[[:space:]]Medical[[:space:]]Image[[:space:]]Synthesis[[:space:]]and[[:space:]]Segmentation/19dcbc41-2b5e-410b-ab71-1dc927bf707e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Noise-Resistant[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection[[:space:]]via[[:space:]]RGB[[:space:]]Error-Guided[[:space:]]Multiscale[[:space:]]Predictive[[:space:]]Coding[[:space:]]and[[:space:]]Dynamic[[:space:]]Memory/8f352e24-f818-44e2-9bcb-4e4f8fb2162e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NoiseCtrl_[[:space:]]A[[:space:]]Sampling-Algorithm-Agnostic[[:space:]]Conditional[[:space:]]Generation[[:space:]]Method[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/2cd0c24c-92a0-49fb-a59a-aa612972886a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Non-Natural[[:space:]]Image[[:space:]]Understanding[[:space:]]with[[:space:]]Advancing[[:space:]]Frequency-based[[:space:]]Vision[[:space:]]Encoders/add37f6f-bd11-41ee-99be-4cafb8ae44f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Nonisotropic[[:space:]]Gaussian[[:space:]]Diffusion[[:space:]]for[[:space:]]Realistic[[:space:]]3D[[:space:]]Human[[:space:]]Motion[[:space:]]Prediction/dd635575-1d09-4d87-b43b-607bd556bf0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Not[[:space:]]All[[:space:]]Parameters[[:space:]]Matter_[[:space:]]Masking[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Enhancing[[:space:]]Generation[[:space:]]Ability/228fe406-55b0-4580-970a-c961ec1cdf30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Not[[:space:]]Just[[:space:]]Text_[[:space:]]Uncovering[[:space:]]Vision[[:space:]]Modality[[:space:]]Typographic[[:space:]]Threats[[:space:]]in[[:space:]]Image[[:space:]]Generation[[:space:]]Models/9edc9c02-4343-4634-8e07-49d1ae0fca3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Not[[:space:]]Only[[:space:]]Text_[[:space:]]Exploring[[:space:]]Compositionality[[:space:]]of[[:space:]]Visual[[:space:]]Representations[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models/8a719bbd-e2e4-4d4e-bf35-17fbcd18669e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Notes-guided[[:space:]]MLLM[[:space:]]Reasoning_[[:space:]]Enhancing[[:space:]]MLLM[[:space:]]with[[:space:]]Knowledge[[:space:]]and[[:space:]]Visual[[:space:]]Notes[[:space:]]for[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/879e6bbe-02d1-4dee-929c-feb3a9739c3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]Pixel-Space[[:space:]]Diffusion[[:space:]]Models/c226fafe-323c-45d8-ab96-898798237def_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Nullu_[[:space:]]Mitigating[[:space:]]Object[[:space:]]Hallucinations[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]via[[:space:]]HalluSpace[[:space:]]Projection/6c852f31-7de2-4cc5-8855-c9a9c0bf3ac1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Number[[:space:]]it_[[:space:]]Temporal[[:space:]]Grounding[[:space:]]Videos[[:space:]]like[[:space:]]Flipping[[:space:]]Manga/92fd3b25-5527-456c-8467-38b74d494405_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/O-TPT_[[:space:]]Orthogonality[[:space:]]Constraints[[:space:]]for[[:space:]]Calibrating[[:space:]]Test-time[[:space:]]Prompt[[:space:]]Tuning[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models/85228d85-f642-43fe-9b5e-7c09729c20a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OCRT_[[:space:]]Boosting[[:space:]]Foundation[[:space:]]Models[[:space:]]in[[:space:]]the[[:space:]]Open[[:space:]]World[[:space:]]with[[:space:]]Object-Concept-Relation[[:space:]]Triad/dc1cac48-f732-4e21-8f2b-4f11d664c810_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ODA-GAN_[[:space:]]Orthogonal[[:space:]]Decoupling[[:space:]]Alignment[[:space:]]GAN[[:space:]]Assisted[[:space:]]by[[:space:]]Weakly-supervised[[:space:]]Learning[[:space:]]for[[:space:]]Virtual[[:space:]]Immunohistochemistry[[:space:]]Staining/0dbc0957-62ff-4048-97b0-a7839e5d1337_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ODE_[[:space:]]Open-Set[[:space:]]Evaluation[[:space:]]of[[:space:]]Hallucinations[[:space:]]in[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/0ba219d5-a22a-4699-9e2c-92b4f6229786_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ODHSR_[[:space:]]Online[[:space:]]Dense[[:space:]]3D[[:space:]]Reconstruction[[:space:]]of[[:space:]]Humans[[:space:]]and[[:space:]]Scenes[[:space:]]from[[:space:]]Monocular[[:space:]]Videos/0c26c7c3-d9ef-4934-8350-e80ccfdc1e8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OFER_[[:space:]]Occluded[[:space:]]Face[[:space:]]Expression[[:space:]]Reconstruction/749e93dd-85de-43c9-8c38-d018fdf4ef57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ONDA-Pose_[[:space:]]Occlusion-Aware[[:space:]]Neural[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Self-Supervised[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/aefb56f0-91e4-46ac-a4fa-7e728ca9e81f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OODD_[[:space:]]Test-time[[:space:]]Out-of-Distribution[[:space:]]Detection[[:space:]]with[[:space:]]Dynamic[[:space:]]Dictionary/84051348-6010-4cb7-8043-9c01f814eb5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OPTICAL_[[:space:]]Leveraging[[:space:]]Optimal[[:space:]]Transport[[:space:]]for[[:space:]]Contribution[[:space:]]Allocation[[:space:]]in[[:space:]]Dataset[[:space:]]Distillation/f598bf32-90c4-4d68-a12b-48959f71e05d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ORIDa_[[:space:]]Object-centric[[:space:]]Real-world[[:space:]]Image[[:space:]]Composition[[:space:]]Dataset/c7c3e8f5-0a27-4e96-b04f-3dddde0f76be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SegAgent_[[:space:]]Exploring[[:space:]]Pixel[[:space:]]Understanding[[:space:]]Capabilities[[:space:]]in[[:space:]]MLLMs[[:space:]]by[[:space:]]Imitating[[:space:]]Human[[:space:]]Annotator[[:space:]]Trajectories/b6444176-c113-4ffe-9cd2-8898c483f069_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SegEarth-OV_[[:space:]]Towards[[:space:]]Training-Free[[:space:]]Open-Vocabulary[[:space:]]Segmentation[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Images/e16e969d-226c-408e-ac45-2cd707494957_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SegMAN_[[:space:]]Omni-scale[[:space:]]Context[[:space:]]Modeling[[:space:]]with[[:space:]]State[[:space:]]Space[[:space:]]Models[[:space:]]and[[:space:]]Local[[:space:]]Attention[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/e4f2e2d4-e859-483d-9efe-a289f9871941_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Segment[[:space:]]Any[[:space:]]Motion[[:space:]]in[[:space:]]Videos/25801bc1-cb6e-4673-83cb-3317052d8a3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Segment[[:space:]]Any-Quality[[:space:]]Images[[:space:]]with[[:space:]]Generative[[:space:]]Latent[[:space:]]Space[[:space:]]Enhancement/1afb4c97-64af-47c3-b3be-9332e1082a6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Segment[[:space:]]Anything,[[:space:]]Even[[:space:]]Occluded/91f2ed4a-96ac-4e81-9860-50746edbae3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Segment[[:space:]]This[[:space:]]Thing_[[:space:]]Foveated[[:space:]]Tokenization[[:space:]]for[[:space:]]Efficient[[:space:]]Point-Prompted[[:space:]]Segmentation/f4675c76-12fd-46eb-8463-021215e9a2e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Segmenting[[:space:]]Maxillofacial[[:space:]]Structures[[:space:]]in[[:space:]]CBCT[[:space:]]Volumes/9fcc46a3-6daf-45d8-9a08-aa8a9a72949e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Cross[[:space:]]Diffusion[[:space:]]Guidance[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Synthesis[[:space:]]of[[:space:]]Similar[[:space:]]Subjects/83e0a95c-7140-478a-a40d-1c8962e615cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Evolving[[:space:]]Visual[[:space:]]Concept[[:space:]]Library[[:space:]]using[[:space:]]Vision-Language[[:space:]]Critics/62355ea3-56f2-42bd-a178-b9a4adbdf932_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Expansion[[:space:]]of[[:space:]]Pre-trained[[:space:]]Models[[:space:]]with[[:space:]]Mixture[[:space:]]of[[:space:]]Adapters[[:space:]]for[[:space:]]Continual[[:space:]]Learning/8dc52086-492d-48e3-b21b-9cfb917c190e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Learning[[:space:]]Hyperspectral[[:space:]]and[[:space:]]Multispectral[[:space:]]Image[[:space:]]Fusion[[:space:]]via[[:space:]]Adaptive[[:space:]]Residual[[:space:]]Guided[[:space:]]Subspace[[:space:]]Diffusion[[:space:]]Model/1ec491d6-a0a0-457b-bfa7-480f4910c1c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Supervised[[:space:]]Cross-View[[:space:]]Correspondence[[:space:]]with[[:space:]]Predictive[[:space:]]Cycle[[:space:]]Consistency/04de2c6b-9723-4b48-a712-82d3adfbf21e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Supervised[[:space:]]Large[[:space:]]Scale[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion[[:space:]]for[[:space:]]Archaeological[[:space:]]Site[[:space:]]Restoration/3fdda516-00e3-4291-8037-b72ad987992c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Supervised[[:space:]]Learning[[:space:]]for[[:space:]]Color[[:space:]]Spike[[:space:]]Camera[[:space:]]Reconstruction/48099590-89bf-4222-a359-563bb17451be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Supervised[[:space:]]Spatial[[:space:]]Correspondence[[:space:]]Across[[:space:]]Modalities/67b3acbf-55e8-459a-a4da-ba8ff6ef0c41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-supervised[[:space:]]ControlNet[[:space:]]with[[:space:]]Spatio-Temporal[[:space:]]Mamba[[:space:]]for[[:space:]]Real-world[[:space:]]Video[[:space:]]Super-resolution/17058328-37bf-4b7c-b282-396fbd60be04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SelfSplat_[[:space:]]Pose-Free[[:space:]]and[[:space:]]3D[[:space:]]Prior-Free[[:space:]]Generalizable[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/bbd53a21-f039-4031-a580-5b3f999b5a9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SemAlign3D_[[:space:]]Semantic[[:space:]]Correspondence[[:space:]]between[[:space:]]RGB-Images[[:space:]]through[[:space:]]Aligning[[:space:]]3D[[:space:]]Object-Class[[:space:]]Representations/5cbc6a72-c5f6-4ba9-8961-eb87b27e63b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SemGeoMo_[[:space:]]Dynamic[[:space:]]Contextual[[:space:]]Human[[:space:]]Motion[[:space:]]Generation[[:space:]]with[[:space:]]Semantic[[:space:]]and[[:space:]]Geometric[[:space:]]Guidance/9a2d022e-5c87-49e4-b542-4c957d3bcf16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semantic[[:space:]]Library[[:space:]]Adaptation_[[:space:]]LoRA[[:space:]]Retrieval[[:space:]]and[[:space:]]Fusion[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/9ff37ee2-adc0-47f0-9f75-7ff2202be291_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semantic[[:space:]]and[[:space:]]Expressive[[:space:]]Variations[[:space:]]in[[:space:]]Image[[:space:]]Captions[[:space:]]Across[[:space:]]Languages/465c8cd8-d1c4-4df0-a3b2-1dbb5fb3f3bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semantic[[:space:]]and[[:space:]]Sequential[[:space:]]Alignment[[:space:]]for[[:space:]]Referring[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/0f152076-9d57-482c-a9cc-897a0bc8eb28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semantic-guided[[:space:]]Cross-Modal[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Skeleton-based[[:space:]]Zero-shot[[:space:]]Action[[:space:]]Recognition/433d7f52-bf59-406e-8aa1-aae038005c90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SemanticDraw_[[:space:]]Towards[[:space:]]Real-Time[[:space:]]Interactive[[:space:]]Content[[:space:]]Creation[[:space:]]from[[:space:]]Image[[:space:]]Diffusion[[:space:]]Models/42ac13af-fc7a-4c50-8fb5-d598ec99a242_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semi-Supervised[[:space:]]State-Space[[:space:]]Model[[:space:]]with[[:space:]]Dynamic[[:space:]]Stacking[[:space:]]Filter[[:space:]]for[[:space:]]Real-World[[:space:]]Video[[:space:]]Deraining/cb0aa053-c5a4-4837-8b51-1c448c661a37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SemiDAViL_[[:space:]]Semi-supervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]with[[:space:]]Vision-Language[[:space:]]Guidance[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/bef13b7f-93f6-4bb3-a82f-a8785de2a581_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SemiETS_[[:space:]]Integrating[[:space:]]Spatial[[:space:]]and[[:space:]]Content[[:space:]]Consistencies[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]End-to-end[[:space:]]Text[[:space:]]Spotting/93ac07f9-5c5a-4084-8709-6f1dc8a9bd3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sensitivity-Aware[[:space:]]Efficient[[:space:]]Fine-Tuning[[:space:]]via[[:space:]]Compact[[:space:]]Dynamic-Rank[[:space:]]Adaptation/a68a581f-a077-45a9-9839-521859c5961e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Separation[[:space:]]of[[:space:]]Powers_[[:space:]]On[[:space:]]Segregating[[:space:]]Knowledge[[:space:]]from[[:space:]]Observation[[:space:]]in[[:space:]]LLM-enabled[[:space:]]Knowledge-based[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/b0b27d1a-186c-41cf-a28c-523d8082bf6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seq2Time_[[:space:]]Sequential[[:space:]]Knowledge[[:space:]]Transfer[[:space:]]for[[:space:]]Video[[:space:]]LLM[[:space:]]Temporal[[:space:]]Grounding/767e680b-2744-4105-a84b-e4f95953240e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SeqAfford_[[:space:]]Sequential[[:space:]]3D[[:space:]]Affordance[[:space:]]Reasoning[[:space:]]via[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model/a3234d97-d927-4684-915f-578f426ea470_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SeqMvRL_[[:space:]]A[[:space:]]Sequential[[:space:]]Fusion[[:space:]]Framework[[:space:]]for[[:space:]]Multi-view[[:space:]]Representation[[:space:]]Learning/1514ab4f-633b-4f04-8a2e-1c9625f283df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SerialGen_[[:space:]]Personalized[[:space:]]Image[[:space:]]Generation[[:space:]]by[[:space:]]First[[:space:]]Standardization[[:space:]]Then[[:space:]]Personalization/4db92e6e-13aa-4914-acf1-41ac9de88e3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SeriesBench_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Narrative-Driven[[:space:]]Drama[[:space:]]Series[[:space:]]Understanding/a7b4ece0-fc55-49e5-bbf2-188de74cecca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seurat_[[:space:]]From[[:space:]]Moving[[:space:]]Points[[:space:]]to[[:space:]]Depth/7b500a33-6c02-4d6d-8fe4-184de0c5a7f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SfM-Free[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]via[[:space:]]Hierarchical[[:space:]]Training/7a80aa8f-8f76-44f7-8c3e-53c972bd2f84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Shading[[:space:]]Meets[[:space:]]Motion_[[:space:]]Self-supervised[[:space:]]Indoor[[:space:]]3D[[:space:]]Reconstruction[[:space:]]Via[[:space:]]Simultaneous[[:space:]]Shape-from-Shading[[:space:]]and[[:space:]]Structure-from-Motion/86916c80-f47e-47cb-9eef-47dac43d7bb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Shadow[[:space:]]Generation[[:space:]]Using[[:space:]]Diffusion[[:space:]]Model[[:space:]]with[[:space:]]Geometry[[:space:]]Prior/895c3c2b-6650-4488-847e-3ab245db27e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Shape[[:space:]]Abstraction[[:space:]]via[[:space:]]Marching[[:space:]]Differentiable[[:space:]]Support[[:space:]]Functions/db972709-2487-47f2-8ee0-97e1d23e77bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Shape[[:space:]]My[[:space:]]Moves_[[:space:]]Text-Driven[[:space:]]Shape-Aware[[:space:]]Synthesis[[:space:]]of[[:space:]]Human[[:space:]]Motions/38963ce2-7528-4468-bb05-644eea549fbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Shape[[:space:]]and[[:space:]]Texture_[[:space:]]What[[:space:]]Influences[[:space:]]Reliable[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation_/3e3de2dd-3f94-482d-943a-5ecc9a81029e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ShapeShifter_[[:space:]]3D[[:space:]]Variations[[:space:]]Using[[:space:]]Multiscale[[:space:]]and[[:space:]]Sparse[[:space:]]Point-Voxel[[:space:]]Diffusion/9bca979d-6491-4ddd-a007-065d6844c69e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ShapeWords_[[:space:]]Guiding[[:space:]]Text-to-Image[[:space:]]Synthesis[[:space:]]with[[:space:]]3D[[:space:]]Shape-Aware[[:space:]]Prompts/662848fa-f27a-44b0-aab6-66ad47c58ea6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sharp-It_[[:space:]]A[[:space:]]Multi-view[[:space:]]to[[:space:]]Multi-view[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]3D[[:space:]]Synthesis[[:space:]]and[[:space:]]Manipulation/64527bab-6930-4ef4-8416-dc96246edd0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SharpDepth_[[:space:]]Sharpening[[:space:]]Metric[[:space:]]Depth[[:space:]]Predictions[[:space:]]Using[[:space:]]Diffusion[[:space:]]Distillation/4d7ee1b6-0c76-4b25-82c5-ad3cd2dab00e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Shift[[:space:]]the[[:space:]]Lens_[[:space:]]Environment-Aware[[:space:]]Unsupervised[[:space:]]Camouflaged[[:space:]]Object[[:space:]]Detection/6231c8d2-7f2e-415a-9969-13e7f045bb16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ShiftwiseConv_[[:space:]]Small[[:space:]]Convolutional[[:space:]]Kernel[[:space:]]with[[:space:]]Large[[:space:]]Kernel[[:space:]]Effect/a38e867c-5c5d-41e7-90eb-6e678974e0a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Shining[[:space:]]Yourself_[[:space:]]High-Fidelity[[:space:]]Ornaments[[:space:]]Virtual[[:space:]]Try-on[[:space:]]with[[:space:]]Diffusion[[:space:]]Model/1ef5bf08-a8fb-4263-a952-98303f0ec85f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ShotAdapter_[[:space:]]Text-to-Multi-Shot[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/e010292b-85e5-4ad6-b5ff-fd493535422e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Show[[:space:]]and[[:space:]]Segment_[[:space:]]Universal[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation[[:space:]]via[[:space:]]In-Context[[:space:]]Learning/e0ad8291-f3c4-4d3c-9431-5230f4f74cc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Show[[:space:]]and[[:space:]]Tell_[[:space:]]Visually[[:space:]]Explainable[[:space:]]Deep[[:space:]]Neural[[:space:]]Nets[[:space:]]via[[:space:]]Spatially-Aware[[:space:]]Concept[[:space:]]Bottleneck[[:space:]]Models/8412533f-67db-465a-8494-1516f366251f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ShowHowTo_[[:space:]]Generating[[:space:]]Scene-Conditioned[[:space:]]Step-by-Step[[:space:]]Visual[[:space:]]Instructions/61e795ea-4574-473e-b3fc-6f3535dfb4bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ShowMak3r_[[:space:]]Compositional[[:space:]]TV[[:space:]]Show[[:space:]]Reconstruction/1c889638-ab1b-4a02-af5b-597eb1c7a5fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ShowUI_[[:space:]]One[[:space:]]Vision-Language-Action[[:space:]]Model[[:space:]]for[[:space:]]GUI[[:space:]]Visual[[:space:]]Agent/54b990c7-51ee-410a-a2ff-63f58f87fc89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Silence[[:space:]]is[[:space:]]Golden_[[:space:]]Leveraging[[:space:]]Adversarial[[:space:]]Examples[[:space:]]to[[:space:]]Nullify[[:space:]]Audio[[:space:]]Control[[:space:]]in[[:space:]]LDM-based[[:space:]]Talking-Head[[:space:]]Generation/5f0bd45b-46e7-4a6d-88ca-50698109492a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Silent[[:space:]]Branding[[:space:]]Attack_[[:space:]]Trigger-free[[:space:]]Data[[:space:]]Poisoning[[:space:]]Attack[[:space:]]on[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/93769aac-6d09-4da8-8358-002554f96486_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sim-to-Real[[:space:]]Causal[[:space:]]Transfer_[[:space:]]A[[:space:]]Metric[[:space:]]Learning[[:space:]]Approach[[:space:]]to[[:space:]]Causally-Aware[[:space:]]Interaction[[:space:]]Representations/a4db3929-4f42-4825-9fca-4090c03bc6f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SimAvatar_[[:space:]]Simulation-Ready[[:space:]]Avatars[[:space:]]with[[:space:]]Layered[[:space:]]Hair[[:space:]]and[[:space:]]Clothing/d9e30bdb-1cd5-4099-8e86-d18344a3f5c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SimLTD_[[:space:]]Simple[[:space:]]Supervised[[:space:]]and[[:space:]]Semi-Supervised[[:space:]]Long-Tailed[[:space:]]Object[[:space:]]Detection/ffdb5b83-0c2a-472c-9852-6c5f68369032_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SimLingo_[[:space:]]Vision-Only[[:space:]]Closed-Loop[[:space:]]Autonomous[[:space:]]Driving[[:space:]]with[[:space:]]Language-Action[[:space:]]Alignment/4ff21147-366e-4a55-a1d6-2eb9d6bba6ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SimMotionEdit_[[:space:]]Text-Based[[:space:]]Human[[:space:]]Motion[[:space:]]Editing[[:space:]]with[[:space:]]Motion[[:space:]]Similarity[[:space:]]Prediction/35d364b8-3b24-4d25-b0d5-f4f2de5e4e95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SimVS_[[:space:]]Simulating[[:space:]]World[[:space:]]Inconsistencies[[:space:]]for[[:space:]]Robust[[:space:]]View[[:space:]]Synthesis/dfa0e139-5a4d-4319-9718-259db56b3f39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Similarity-Guided[[:space:]]Layer-Adaptive[[:space:]]Vision[[:space:]]Transformer[[:space:]]for[[:space:]]UAV[[:space:]]Tracking/c334a2a6-9747-41af-b8b6-5b13b27ea619_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/(ML)$^2$P-Encoder_[[:space:]]On[[:space:]]Exploration[[:space:]]of[[:space:]]Channel-Class[[:space:]]Correlation[[:space:]]for[[:space:]]Multi-Label[[:space:]]Zero-Shot[[:space:]]Learning/2f5d72e4-31bc-4c21-9948-28d1063a50fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/1%[[:space:]]VS[[:space:]]100%_[[:space:]]Parameter-Efficient[[:space:]]Low[[:space:]]Rank[[:space:]]Adapter[[:space:]]for[[:space:]]Dense[[:space:]]Predictions/3b75c6c9-33bc-4e41-9df3-2e14ac85ef59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/1000[[:space:]]FPS[[:space:]]HDR[[:space:]]Video[[:space:]]With[[:space:]]a[[:space:]]Spike-RGB[[:space:]]Hybrid[[:space:]]Camera/1c93f555-c37f-43ed-866a-0e7c5d4458e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/2PCNet_[[:space:]]Two-Phase[[:space:]]Consistency[[:space:]]Training[[:space:]]for[[:space:]]Day-to-Night[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detection/818b1ea7-c7c2-488e-9c91-78c9a94fffa2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Cinemagraphy[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Image/822e0c52-d8c7-4a4e-8a84-1a2d57dbe08f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Concept[[:space:]]Learning[[:space:]]and[[:space:]]Reasoning[[:space:]]From[[:space:]]Multi-View[[:space:]]Images/6720ecfb-203e-4307-9b9b-8d1051d4343b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]GAN[[:space:]]Inversion[[:space:]]With[[:space:]]Facial[[:space:]]Symmetry[[:space:]]Prior/02a489c6-c89c-4dc3-afcb-600bfa013373_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Highlighter_[[:space:]]Localizing[[:space:]]Regions[[:space:]]on[[:space:]]3D[[:space:]]Shapes[[:space:]]via[[:space:]]Text[[:space:]]Descriptions/40cb675d-902c-46da-982e-90a4332ad0f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Human[[:space:]]Keypoints[[:space:]]Estimation[[:space:]]From[[:space:]]Point[[:space:]]Clouds[[:space:]]in[[:space:]]the[[:space:]]Wild[[:space:]]Without[[:space:]]Human[[:space:]]Labels/833a9b3e-a176-4092-b5fd-3122723612f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Human[[:space:]]Mesh[[:space:]]Estimation[[:space:]]From[[:space:]]Virtual[[:space:]]Markers/067f420e-7fdc-4668-8983-b6715ae47be7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]With[[:space:]]Spatio-Temporal[[:space:]]Criss-Cross[[:space:]]Attention/54678f96-220e-4220-837c-0b75958caa1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Intuitive[[:space:]]Physics/23a54e7d-fed1-435b-b507-df1bdee18df4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Line[[:space:]]Mapping[[:space:]]Revisited/6d931762-d036-45d2-bafa-8ad88d81ad10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Neural[[:space:]]Field[[:space:]]Generation[[:space:]]Using[[:space:]]Triplane[[:space:]]Diffusion/9d99632a-6c66-4f96-953f-d0f7ffc4caf8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Registration[[:space:]]With[[:space:]]Maximal[[:space:]]Cliques/6c9eb542-01ea-4edb-baf1-31469bcf7e1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]in[[:space:]]the[[:space:]]Wild_[[:space:]]Learning[[:space:]]Generalized[[:space:]]Models[[:space:]]for[[:space:]]Adverse-Condition[[:space:]]Point[[:space:]]Clouds/a6bb8bb5-8301-40cc-afda-a77312b4139d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Shape[[:space:]]Reconstruction[[:space:]]of[[:space:]]Semi-Transparent[[:space:]]Worms/541a37a3-ad08-4ec0-acf7-4ca83662c9c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Spatial[[:space:]]Multimodal[[:space:]]Knowledge[[:space:]]Accumulation[[:space:]]for[[:space:]]Scene[[:space:]]Graph[[:space:]]Prediction[[:space:]]in[[:space:]]Point[[:space:]]Cloud/8b57cee0-fdf2-4526-9ea1-36db5e008e92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Video[[:space:]]Loops[[:space:]]From[[:space:]]Asynchronous[[:space:]]Input/7bb72ce9-0dd3-422a-99d9-0bd1bcda48bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Video[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]Learnable[[:space:]]Object-Centric[[:space:]]Global[[:space:]]Optimization/2347d966-1e20-4c7d-aef9-82586306a3eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D-Aware[[:space:]]Conditional[[:space:]]Image[[:space:]]Synthesis/b9625555-02d4-4da7-b507-7cd64cc67a00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D-Aware[[:space:]]Face[[:space:]]Swapping/66d1bee4-1a69-4f6f-8a65-3f5202fddfc5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D-Aware[[:space:]]Facial[[:space:]]Landmark[[:space:]]Detection[[:space:]]via[[:space:]]Multi-View[[:space:]]Consistent[[:space:]]Training[[:space:]]on[[:space:]]Synthetic[[:space:]]Data/4aaf53b5-ffe9-4822-bbbc-9f293082f284_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D-Aware[[:space:]]Multi-Class[[:space:]]Image-to-Image[[:space:]]Translation[[:space:]]With[[:space:]]NeRFs/38da797f-7f59-48cd-af34-af72487f73d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D-Aware[[:space:]]Object[[:space:]]Goal[[:space:]]Navigation[[:space:]]via[[:space:]]Simultaneous[[:space:]]Exploration[[:space:]]and[[:space:]]Identification/e3176243-c1cd-415f-8bca-116983524509_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D-POP[[:space:]]-[[:space:]]An[[:space:]]Automated[[:space:]]Annotation[[:space:]]Approach[[:space:]]to[[:space:]]Facilitate[[:space:]]Markerless[[:space:]]2D-3D[[:space:]]Tracking[[:space:]]of[[:space:]]Freely[[:space:]]Moving[[:space:]]Birds[[:space:]]With[[:space:]]Marker-Based[[:space:]]Motion[[:space:]]Capture/5371de19-661e-4e67-a303-36ffc7847ea6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3DAvatarGAN_[[:space:]]Bridging[[:space:]]Domains[[:space:]]for[[:space:]]Personalized[[:space:]]Editable[[:space:]]Avatars/ddf7c6ad-f988-4a54-8cf6-7aff7d8dd81c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3Mformer_[[:space:]]Multi-Order[[:space:]]Multi-Mode[[:space:]]Transformer[[:space:]]for[[:space:]]Skeletal[[:space:]]Action[[:space:]]Recognition/59904744-5656-40cd-af70-98473e4f87a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Bag-of-Prototypes[[:space:]]Representation[[:space:]]for[[:space:]]Dataset-Level[[:space:]]Applications/f45f628e-fe49-4cb9-b5bd-808953724624_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Characteristic[[:space:]]Function-Based[[:space:]]Method[[:space:]]for[[:space:]]Bottom-Up[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/1484ca20-37b6-4284-8188-8a19d046c61f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Data-Based[[:space:]]Perspective[[:space:]]on[[:space:]]Transfer[[:space:]]Learning/b077d70d-8608-4443-a4ce-0c29fda55f28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Dynamic[[:space:]]Multi-Scale[[:space:]]Voxel[[:space:]]Flow[[:space:]]Network[[:space:]]for[[:space:]]Video[[:space:]]Prediction/932e5c1f-279d-4c41-943b-431182e5f76a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]General[[:space:]]Regret[[:space:]]Bound[[:space:]]of[[:space:]]Preconditioned[[:space:]]Gradient[[:space:]]Method[[:space:]]for[[:space:]]DNN[[:space:]]Training/a806573e-912a-4e15-8891-1f914fce477d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Generalized[[:space:]]Framework[[:space:]]for[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation/1a9f411f-250e-4299-a2f7-53a15f5210ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Hierarchical[[:space:]]Representation[[:space:]]Network[[:space:]]for[[:space:]]Accurate[[:space:]]and[[:space:]]Detailed[[:space:]]Face[[:space:]]Reconstruction[[:space:]]From[[:space:]]In-the-Wild[[:space:]]Images/e9a1eb39-7ce0-41ae-94da-5550d2c42a24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Large-Scale[[:space:]]Homography[[:space:]]Benchmark/e3b07a31-6172-4d9c-be61-9e165adb34c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Large-Scale[[:space:]]Robustness[[:space:]]Analysis[[:space:]]of[[:space:]]Video[[:space:]]Action[[:space:]]Recognition[[:space:]]Models/c14579d3-229a-4b31-8be1-4bdbdda8432b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Light[[:space:]]Touch[[:space:]]Approach[[:space:]]to[[:space:]]Teaching[[:space:]]Transformers[[:space:]]Multi-View[[:space:]]Geometry/987fa748-4355-4280-aa9d-f468a299dbbf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Light[[:space:]]Weight[[:space:]]Model[[:space:]]for[[:space:]]Active[[:space:]]Speaker[[:space:]]Detection/9ad8cf23-1314-4700-8505-aa44ebeac51c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Loopback[[:space:]]Network[[:space:]]for[[:space:]]Explainable[[:space:]]Microvascular[[:space:]]Invasion[[:space:]]Classification/fb77d6ec-7fb5-4ffa-a01f-29a9f56320e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Meta-Learning[[:space:]]Approach[[:space:]]to[[:space:]]Predicting[[:space:]]Performance[[:space:]]and[[:space:]]Data[[:space:]]Requirements/02ed2813-1f53-4474-afa7-1c7c5c3b0181_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]New[[:space:]]Benchmark_[[:space:]]On[[:space:]]the[[:space:]]Utility[[:space:]]of[[:space:]]Synthetic[[:space:]]Data[[:space:]]With[[:space:]]Blender[[:space:]]for[[:space:]]Bare[[:space:]]Supervised[[:space:]]Learning[[:space:]]and[[:space:]]Downstream[[:space:]]Domain[[:space:]]Adaptation/6a4c04bb-85f7-4652-a1ed-07fb059d04cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]New[[:space:]]Comprehensive[[:space:]]Benchmark[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection[[:space:]]and[[:space:]]Anticipation/65fac8b2-2580-4ed9-bf55-f97afee2c392_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]New[[:space:]]Dataset[[:space:]]Based[[:space:]]on[[:space:]]Images[[:space:]]Taken[[:space:]]by[[:space:]]Blind[[:space:]]People[[:space:]]for[[:space:]]Testing[[:space:]]the[[:space:]]Robustness[[:space:]]of[[:space:]]Image[[:space:]]Classification[[:space:]]Models[[:space:]]Trained[[:space:]]for[[:space:]]ImageNet[[:space:]]Categories/9dd60dd9-ba90-4a41-ae85-9e3260b35ac0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]New[[:space:]]Path_[[:space:]]Scaling[[:space:]]Vision-and-Language[[:space:]]Navigation[[:space:]]With[[:space:]]Synthetic[[:space:]]Instructions[[:space:]]and[[:space:]]Imitation[[:space:]]Learning/57be1fed-990b-43fc-a088-d15443fca14c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Practical[[:space:]]Stereo[[:space:]]Depth[[:space:]]System[[:space:]]for[[:space:]]Smart[[:space:]]Glasses/d5f455f0-b537-42a4-9205-ab089ce13637_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Practical[[:space:]]Upper[[:space:]]Bound[[:space:]]for[[:space:]]the[[:space:]]Worst-Case[[:space:]]Attribution[[:space:]]Deviations/eadde5c3-88cf-4c1c-9173-a95a048d8da6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Probabilistic[[:space:]]Attention[[:space:]]Model[[:space:]]With[[:space:]]Occlusion-Aware[[:space:]]Texture[[:space:]]Regression[[:space:]]for[[:space:]]3D[[:space:]]Hand[[:space:]]Reconstruction[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]RGB[[:space:]]Image/2c635c94-49b2-49e1-9aeb-435915a0cad3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Probabilistic[[:space:]]Framework[[:space:]]for[[:space:]]Lifelong[[:space:]]Test-Time[[:space:]]Adaptation/ee89d8f8-e027-4a77-90fc-b4cc9e5a2c22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Rotation-Translation-Decoupled[[:space:]]Solution[[:space:]]for[[:space:]]Robust[[:space:]]and[[:space:]]Efficient[[:space:]]Visual-Inertial[[:space:]]Initialization/a75843d0-b622-47d1-8c2b-838fc82ded84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Simple[[:space:]]Baseline[[:space:]]for[[:space:]]Video[[:space:]]Restoration[[:space:]]With[[:space:]]Grouped[[:space:]]Spatial-Temporal[[:space:]]Shift/719c451c-5b96-4803-a991-1e62fe88accb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Simple[[:space:]]Framework[[:space:]]for[[:space:]]Text-Supervised[[:space:]]Semantic[[:space:]]Segmentation/29325981-aecc-414c-805b-cde6a49719c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Soma[[:space:]]Segmentation[[:space:]]Benchmark[[:space:]]in[[:space:]]Full[[:space:]]Adult[[:space:]]Fly[[:space:]]Brain/87162a49-daf4-4f2b-b1d3-f856ad56a7de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Strong[[:space:]]Baseline[[:space:]]for[[:space:]]Generalized[[:space:]]Few-Shot[[:space:]]Semantic[[:space:]]Segmentation/6499e360-97d8-492b-b59c-c2df61403fc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Unified[[:space:]]HDR[[:space:]]Imaging[[:space:]]Method[[:space:]]With[[:space:]]Pixel[[:space:]]and[[:space:]]Patch[[:space:]]Level/e868fbca-1293-4c58-b639-3c2b1db0a421_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Unified[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]Framework[[:space:]]for[[:space:]]Deep[[:space:]]Directed[[:space:]]Graphical[[:space:]]Models/b64196ea-df79-4b42-9e7a-c72f0b7a6d53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Unified[[:space:]]Pyramid[[:space:]]Recurrent[[:space:]]Network[[:space:]]for[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation/02ffd161-2ea1-43df-bd4e-2082029f05e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Unified[[:space:]]Spatial-Angular[[:space:]]Structured[[:space:]]Light[[:space:]]for[[:space:]]Single-View[[:space:]]Acquisition[[:space:]]of[[:space:]]Shape[[:space:]]and[[:space:]]Reflectance/a179941c-ec9c-408c-8e26-6fc194709e15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Whac-a-Mole[[:space:]]Dilemma_[[:space:]]Shortcuts[[:space:]]Come[[:space:]]in[[:space:]]Multiples[[:space:]]Where[[:space:]]Mitigating[[:space:]]One[[:space:]]Amplifies[[:space:]]Others/29b75eb3-d507-495f-b874-75df578c6a43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A-Cap_[[:space:]]Anticipation[[:space:]]Captioning[[:space:]]With[[:space:]]Commonsense[[:space:]]Knowledge/0e9d14f2-ed8b-47a0-933f-6fa45b08d07e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A-La-Carte[[:space:]]Prompt[[:space:]]Tuning[[:space:]](APT)_[[:space:]]Combining[[:space:]]Distinct[[:space:]]Data[[:space:]]via[[:space:]]Composable[[:space:]]Prompting/585e87c3-b2f4-4a4c-91df-2e593fcf5789_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A2J-Transformer_[[:space:]]Anchor-to-Joint[[:space:]]Transformer[[:space:]]Network[[:space:]]for[[:space:]]3D[[:space:]]Interacting[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]RGB[[:space:]]Image/ab223230-b52b-491a-b054-0fe645339363_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ABCD_[[:space:]]Arbitrary[[:space:]]Bitwise[[:space:]]Coefficient[[:space:]]for[[:space:]]De-Quantization/74de187f-bb6a-4967-8de8-7245e6e6dc31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ABLE-NeRF_[[:space:]]Attention-Based[[:space:]]Rendering[[:space:]]With[[:space:]]Learnable[[:space:]]Embeddings[[:space:]]for[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field/b8595b73-1c79-4c16-b7bb-6fddbb1f1ffa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ACL-SPC_[[:space:]]Adaptive[[:space:]]Closed-Loop[[:space:]]System[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion/6240ac72-711b-45ce-a021-754b5e04e83c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ACR_[[:space:]]Attention[[:space:]]Collaboration-Based[[:space:]]Regressor[[:space:]]for[[:space:]]Arbitrary[[:space:]]Two-Hand[[:space:]]Reconstruction/7ab86863-28c3-4d52-876a-d70c2e93db98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ACSeg_[[:space:]]Adaptive[[:space:]]Conceptualization[[:space:]]for[[:space:]]Unsupervised[[:space:]]Semantic[[:space:]]Segmentation/396ae74f-58da-4076-862e-ba4411971380_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AGAIN_[[:space:]]Adversarial[[:space:]]Training[[:space:]]With[[:space:]]Attribution[[:space:]]Span[[:space:]]Enlargement[[:space:]]and[[:space:]]Hybrid[[:space:]]Feature[[:space:]]Fusion/61e731ad-8ce0-4171-8073-8c49cff5bdd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ALOFT_[[:space:]]A[[:space:]]Lightweight[[:space:]]MLP-Like[[:space:]]Architecture[[:space:]]With[[:space:]]Dynamic[[:space:]]Low-Frequency[[:space:]]Transform[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/7e284a78-ed9b-4710-a528-83ac8d95a442_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ALSO_[[:space:]]Automotive[[:space:]]Lidar[[:space:]]Self-Supervision[[:space:]]by[[:space:]]Occupancy[[:space:]]Estimation/fa456ebb-15d7-4bc9-8d53-28160b37e988_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ALTO_[[:space:]]Alternating[[:space:]]Latent[[:space:]]Topologies[[:space:]]for[[:space:]]Implicit[[:space:]]3D[[:space:]]Reconstruction/f0742f4d-d7c2-47dd-9e91-bef7b6fda552_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AMT_[[:space:]]All-Pairs[[:space:]]Multi-Field[[:space:]]Transforms[[:space:]]for[[:space:]]Efficient[[:space:]]Frame[[:space:]]Interpolation/c8f12330-6f3c-4016-b9ea-39cbcd141ff4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ANetQA_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Benchmark[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Compositional[[:space:]]Reasoning[[:space:]]Over[[:space:]]Untrimmed[[:space:]]Videos/4120dd38-be13-46db-a886-ac659b1e8dcb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ARCTIC_[[:space:]]A[[:space:]]Dataset[[:space:]]for[[:space:]]Dexterous[[:space:]]Bimanual[[:space:]]Hand-Object[[:space:]]Manipulation/203a7ac7-af1a-43b1-af6e-19724dea8ee4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ARKitTrack_[[:space:]]A[[:space:]]New[[:space:]]Diverse[[:space:]]Dataset[[:space:]]for[[:space:]]Tracking[[:space:]]Using[[:space:]]Mobile[[:space:]]RGB-D[[:space:]]Data/aff10354-4e0b-479b-bec7-99e1bb5614ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ARO-Net_[[:space:]]Learning[[:space:]]Implicit[[:space:]]Fields[[:space:]]From[[:space:]]Anchored[[:space:]]Radial[[:space:]]Observations/0424e15f-bc6d-4c85-8c30-4cff9510d066_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ASPnet_[[:space:]]Action[[:space:]]Segmentation[[:space:]]With[[:space:]]Shared-Private[[:space:]]Representation[[:space:]]of[[:space:]]Multiple[[:space:]]Data[[:space:]]Sources/1cc127a4-3932-4274-8888-e5f95695b1a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AShapeFormer_[[:space:]]Semantics-Guided[[:space:]]Object-Level[[:space:]]Active[[:space:]]Shape[[:space:]]Encoding[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Transformers/8adfb62d-3f8c-411f-a44f-2801eba7452a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AUNet_[[:space:]]Learning[[:space:]]Relations[[:space:]]Between[[:space:]]Action[[:space:]]Units[[:space:]]for[[:space:]]Face[[:space:]]Forgery[[:space:]]Detection/62de5785-4986-424f-a3f4-6b8eecec90b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AVFace_[[:space:]]Towards[[:space:]]Detailed[[:space:]]Audio-Visual[[:space:]]4D[[:space:]]Face[[:space:]]Reconstruction/0b7d6d29-4bac-46ef-b05d-8e479c92824a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AVFormer_[[:space:]]Injecting[[:space:]]Vision[[:space:]]Into[[:space:]]Frozen[[:space:]]Speech[[:space:]]Models[[:space:]]for[[:space:]]Zero-Shot[[:space:]]AV-ASR/07c001fc-8cba-4cb6-9dbb-389099481643_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Abstract[[:space:]]Visual[[:space:]]Reasoning_[[:space:]]An[[:space:]]Algebraic[[:space:]]Approach[[:space:]]for[[:space:]]Solving[[:space:]]Raven's[[:space:]]Progressive[[:space:]]Matrices/0373c754-2d12-4d52-95c7-16e20ad62c3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AccelIR_[[:space:]]Task-Aware[[:space:]]Image[[:space:]]Compression[[:space:]]for[[:space:]]Accelerating[[:space:]]Neural[[:space:]]Restoration/3ee3bce8-b085-4d08-be6f-5f7d348c3f3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Accelerated[[:space:]]Coordinate[[:space:]]Encoding_[[:space:]]Learning[[:space:]]to[[:space:]]Relocalize[[:space:]]in[[:space:]]Minutes[[:space:]]Using[[:space:]]RGB[[:space:]]and[[:space:]]Poses/49024a39-e689-4fa1-87f4-7de949651fed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Accelerating[[:space:]]Dataset[[:space:]]Distillation[[:space:]]via[[:space:]]Model[[:space:]]Augmentation/38ccc529-8263-4fb5-b4f1-49f67321c951_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Accelerating[[:space:]]Vision-Language[[:space:]]Pretraining[[:space:]]With[[:space:]]Free[[:space:]]Language[[:space:]]Modeling/e514fec0-dcec-43a5-8ee8-992c5bb1bf30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Accidental[[:space:]]Light[[:space:]]Probes/dc85ac92-a253-4d51-a800-2fbce9d162e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Achieving[[:space:]]a[[:space:]]Better[[:space:]]Stability-Plasticity[[:space:]]Trade-Off[[:space:]]via[[:space:]]Auxiliary[[:space:]]Networks[[:space:]]in[[:space:]]Continual[[:space:]]Learning/4a5fef63-1ac1-42e6-afc9-6401add047f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ActMAD_[[:space:]]Activation[[:space:]]Matching[[:space:]]To[[:space:]]Align[[:space:]]Distributions[[:space:]]for[[:space:]]Test-Time-Training/49da5ef2-da32-47e8-9d0f-d937891d1f91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Actionlet-Dependent[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Unsupervised[[:space:]]Skeleton-Based[[:space:]]Action[[:space:]]Recognition/ac7f0075-f8cc-4153-9f3c-8dc3b241e3bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Activating[[:space:]]More[[:space:]]Pixels[[:space:]]in[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]Transformer/082e91c9-1a85-416b-8411-efe827873c6c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Active[[:space:]]Exploration[[:space:]]of[[:space:]]Multimodal[[:space:]]Complementarity[[:space:]]for[[:space:]]Few-Shot[[:space:]]Action[[:space:]]Recognition/f1242791-3dba-4357-80ab-fba6a589b2ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Active[[:space:]]Finetuning_[[:space:]]Exploiting[[:space:]]Annotation[[:space:]]Budget[[:space:]]in[[:space:]]the[[:space:]]Pretraining-Finetuning[[:space:]]Paradigm/535ff76b-59f6-4bb4-af59-4b5da385d910_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AdaMAE_[[:space:]]Adaptive[[:space:]]Masking[[:space:]]for[[:space:]]Efficient[[:space:]]Spatiotemporal[[:space:]]Learning[[:space:]]With[[:space:]]Masked[[:space:]]Autoencoders/9fc47e4f-8bde-4d26-90b5-f77c167b6b0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AdamsFormer[[:space:]]for[[:space:]]Spatial[[:space:]]Action[[:space:]]Localization[[:space:]]in[[:space:]]the[[:space:]]Future/f6db17d7-3256-46b5-a1a3-6fac25182ee0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adapting[[:space:]]Shortcut[[:space:]]With[[:space:]]Normalizing[[:space:]]Flow_[[:space:]]An[[:space:]]Efficient[[:space:]]Tuning[[:space:]]Framework[[:space:]]for[[:space:]]Visual[[:space:]]Recognition/1b82636b-027a-45dd-a3f9-d2187e5186fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Annealing[[:space:]]for[[:space:]]Robust[[:space:]]Geometric[[:space:]]Estimation/aaa82561-2f75-4a96-a320-e7e09040c736_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Assignment[[:space:]]for[[:space:]]Geometry[[:space:]]Aware[[:space:]]Local[[:space:]]Feature[[:space:]]Matching/5f22d870-9943-42d5-861a-8c748175002c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Channel[[:space:]]Sparsity[[:space:]]for[[:space:]]Federated[[:space:]]Learning[[:space:]]Under[[:space:]]System[[:space:]]Heterogeneity/4398d2a1-da3f-4ade-af8d-4d85fe001b54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Data-Free[[:space:]]Quantization/27e6eea5-8d81-44fb-8994-f7511d222596_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Global[[:space:]]Decay[[:space:]]Process[[:space:]]for[[:space:]]Event[[:space:]]Cameras/7f549841-7f2c-4448-878d-0f7a04df3040_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Graph[[:space:]]Convolutional[[:space:]]Subspace[[:space:]]Clustering/f34b945c-912a-42b9-849e-e879bf92e34c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Human[[:space:]]Matting[[:space:]]for[[:space:]]Dynamic[[:space:]]Videos/61fc4a26-2b88-4ad8-beec-6a5efbb3b94c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Patch[[:space:]]Deformation[[:space:]]for[[:space:]]Textureless-Resilient[[:space:]]Multi-View[[:space:]]Stereo/3bcda38e-de0f-440a-8bc6-7d64bdf2f5d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Plasticity[[:space:]]Improvement[[:space:]]for[[:space:]]Continual[[:space:]]Learning/0c08555a-9c87-408a-b2cb-7722d9537bec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Sparse[[:space:]]Convolutional[[:space:]]Networks[[:space:]]With[[:space:]]Global[[:space:]]Context[[:space:]]Enhancement[[:space:]]for[[:space:]]Faster[[:space:]]Object[[:space:]]Detection[[:space:]]on[[:space:]]Drone[[:space:]]Images/986d4984-3be6-49bb-b124-1857679d14f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Sparse[[:space:]]Pairwise[[:space:]]Loss[[:space:]]for[[:space:]]Object[[:space:]]Re-Identification/521e7dfb-fb3f-4f1d-9abc-4d597b34f91d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Spot-Guided[[:space:]]Transformer[[:space:]]for[[:space:]]Consistent[[:space:]]Local[[:space:]]Feature[[:space:]]Matching/042f31dd-50c9-4d61-9a4e-3b9c1c2035fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Zone-Aware[[:space:]]Hierarchical[[:space:]]Planner[[:space:]]for[[:space:]]Vision-Language[[:space:]]Navigation/f7036780-2e83-4485-8390-9ef34bf651b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AdaptiveMix_[[:space:]]Improving[[:space:]]GAN[[:space:]]Training[[:space:]]via[[:space:]]Feature[[:space:]]Space[[:space:]]Shrinkage/1735bd44-57c2-4401-ae49-87b869a7308d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adjustment[[:space:]]and[[:space:]]Alignment[[:space:]]for[[:space:]]Unbiased[[:space:]]Open[[:space:]]Set[[:space:]]Domain[[:space:]]Adaptation/d58b66e5-8387-4825-9d7a-2511ad30c4bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Advancing[[:space:]]Visual[[:space:]]Grounding[[:space:]]With[[:space:]]Scene[[:space:]]Knowledge_[[:space:]]Benchmark[[:space:]]and[[:space:]]Method/5ba325dd-c039-4f89-893e-790c3d56245a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adversarial[[:space:]]Counterfactual[[:space:]]Visual[[:space:]]Explanations/82ff8cee-91e5-4e11-9414-4a91bc5c5add_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adversarial[[:space:]]Normalization_[[:space:]]I[[:space:]]Can[[:space:]]Visualize[[:space:]]Everything[[:space:]](ICE)/1559ab12-3936-4912-b9ce-76fff24fcfcd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adversarial[[:space:]]Robustness[[:space:]]via[[:space:]]Random[[:space:]]Projection[[:space:]]Filters/532fe391-797b-44a4-b31e-45b97a89b5ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adversarially[[:space:]]Masking[[:space:]]Synthetic[[:space:]]To[[:space:]]Mimic[[:space:]]Real_[[:space:]]Adaptive[[:space:]]Noise[[:space:]]Injection[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation[[:space:]]Adaptation/ebcac971-0782-4625-be69-0efc0da1b530_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adversarially[[:space:]]Robust[[:space:]]Neural[[:space:]]Architecture[[:space:]]Search[[:space:]]for[[:space:]]Graph[[:space:]]Neural[[:space:]]Networks/85115071-6108-4823-9ad1-fed37100676f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AeDet_[[:space:]]Azimuth-Invariant[[:space:]]Multi-View[[:space:]]3D[[:space:]]Object[[:space:]]Detection/05cea7b0-c5b4-4fcf-8500-7400b5888637_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Affection_[[:space:]]Learning[[:space:]]Affective[[:space:]]Explanations[[:space:]]for[[:space:]]Real-World[[:space:]]Visual[[:space:]]Data/dc925056-b09b-46da-b8af-41c3fa946b76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Affordance[[:space:]]Diffusion_[[:space:]]Synthesizing[[:space:]]Hand-Object[[:space:]]Interactions/4c60df3f-aeb1-4d0d-ba8d-0055b1df65fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Affordance[[:space:]]Grounding[[:space:]]From[[:space:]]Demonstration[[:space:]]Video[[:space:]]To[[:space:]]Target[[:space:]]Image/2a72433b-fe7e-4ae6-a78a-2a1095666fed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Affordances[[:space:]]From[[:space:]]Human[[:space:]]Videos[[:space:]]as[[:space:]]a[[:space:]]Versatile[[:space:]]Representation[[:space:]]for[[:space:]]Robotics/eefe6dc4-ce94-4f79-af0e-ee1637524c42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Alias-Free[[:space:]]Convnets_[[:space:]]Fractional[[:space:]]Shift[[:space:]]Invariance[[:space:]]via[[:space:]]Polynomial[[:space:]]Activations/baf18f57-29e8-475e-ba02-cf7151a6fd54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AligNeRF_[[:space:]]High-Fidelity[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]via[[:space:]]Alignment-Aware[[:space:]]Training/5671aa90-954d-46d9-85fb-a274595998ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Align[[:space:]]Your[[:space:]]Latents_[[:space:]]High-Resolution[[:space:]]Video[[:space:]]Synthesis[[:space:]]With[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models/36e51489-12ef-4874-9d60-949c0dddcd56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Align[[:space:]]and[[:space:]]Attend_[[:space:]]Multimodal[[:space:]]Summarization[[:space:]]With[[:space:]]Dual[[:space:]]Contrastive[[:space:]]Losses/31df57f8-5eb4-4f2a-b3c4-00cdb3c4abc7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Aligning[[:space:]]Bag[[:space:]]of[[:space:]]Regions[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detection/f3cbb3f4-e920-4980-9069-6dcfa0432e20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Aligning[[:space:]]Step-by-Step[[:space:]]Instructional[[:space:]]Diagrams[[:space:]]to[[:space:]]Video[[:space:]]Demonstrations/ff1fa0c2-61c5-472a-ae5e-585cec204eab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/All[[:space:]]Are[[:space:]]Worth[[:space:]]Words_[[:space:]]A[[:space:]]ViT[[:space:]]Backbone[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/93f4696b-665c-4e18-9379-42b21e05208c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/All[[:space:]]in[[:space:]]One_[[:space:]]Exploring[[:space:]]Unified[[:space:]]Video-Language[[:space:]]Pre-Training/bbdc05c2-ce87-43d0-be23-fd36a2dd6db5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/All-in-Focus[[:space:]]Imaging[[:space:]]From[[:space:]]Event[[:space:]]Focal[[:space:]]Stack/c479431e-bfb0-4703-ba5f-65bdeeff1ab8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/All-in-One[[:space:]]Image[[:space:]]Restoration[[:space:]]for[[:space:]]Unknown[[:space:]]Degradations[[:space:]]Using[[:space:]]Adaptive[[:space:]]Discriminative[[:space:]]Filters[[:space:]]for[[:space:]]Specific[[:space:]]Degradations/8f4fab61-7bec-49d1-8f2d-07269bbdba1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AltFreezing[[:space:]]for[[:space:]]More[[:space:]]General[[:space:]]Video[[:space:]]Face[[:space:]]Forgery[[:space:]]Detection/6da47c05-4ebf-47d8-9aab-a99ddd2e2875_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ambiguity-Resistant[[:space:]]Semi-Supervised[[:space:]]Learning[[:space:]]for[[:space:]]Dense[[:space:]]Object[[:space:]]Detection/b75f6046-3329-4755-aa33-7d5c3f1284e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ambiguous[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation[[:space:]]Using[[:space:]]Diffusion[[:space:]]Models/aaa01da2-687c-4ced-98ed-311d51a1d6b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/An[[:space:]]Actor-Centric[[:space:]]Causality[[:space:]]Graph[[:space:]]for[[:space:]]Asynchronous[[:space:]]Temporal[[:space:]]Inference[[:space:]]in[[:space:]]Group[[:space:]]Activity/88a96b10-3f89-4fa5-a31e-4aa550e3edce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/An[[:space:]]Empirical[[:space:]]Study[[:space:]]of[[:space:]]End-to-End[[:space:]]Video-Language[[:space:]]Transformers[[:space:]]With[[:space:]]Masked[[:space:]]Visual[[:space:]]Modeling/3ed82dc7-e45c-469c-bb5a-140c3ed692b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/An[[:space:]]Erudite[[:space:]]Fine-Grained[[:space:]]Visual[[:space:]]Classification[[:space:]]Model/427adf1c-5ce3-4fe2-ad51-8091be4f3c2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/An[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment[[:space:]]Dataset[[:space:]]for[[:space:]]Portraits/4d629a97-5924-456e-b0bf-45e70d6491fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/An[[:space:]]In-Depth[[:space:]]Exploration[[:space:]]of[[:space:]]Person[[:space:]]Re-Identification[[:space:]]and[[:space:]]Gait[[:space:]]Recognition[[:space:]]in[[:space:]]Cloth-Changing[[:space:]]Conditions/d92839ff-b85f-41d5-8e16-b8c7ae242cee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Analyzing[[:space:]]Physical[[:space:]]Impacts[[:space:]]Using[[:space:]]Transient[[:space:]]Surface[[:space:]]Wave[[:space:]]Imaging/7ab97a36-0f0a-4d46-8bea-1d4f23ed6381_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Analyzing[[:space:]]and[[:space:]]Diagnosing[[:space:]]Pose[[:space:]]Estimation[[:space:]]With[[:space:]]Attributions/b0fd78b7-27af-4b89-87e1-e5056a22f003_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Anchor3DLane_[[:space:]]Learning[[:space:]]To[[:space:]]Regress[[:space:]]3D[[:space:]]Anchors[[:space:]]for[[:space:]]Monocular[[:space:]]3D[[:space:]]Lane[[:space:]]Detection/2d2702a4-6454-454d-96a6-5f250100d84e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AnchorFormer_[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion[[:space:]]From[[:space:]]Discriminative[[:space:]]Nodes/4a72a000-5fd1-4ca1-99a7-6b816e8c571a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Angelic[[:space:]]Patches[[:space:]]for[[:space:]]Improving[[:space:]]Third-Party[[:space:]]Object[[:space:]]Detector[[:space:]]Performance/4769dd88-2368-400d-aa8c-9142e5332b9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Annealing-Based[[:space:]]Label-Transfer[[:space:]]Learning[[:space:]]for[[:space:]]Open[[:space:]]World[[:space:]]Object[[:space:]]Detection/ce220ee9-ae17-44c5-9491-4642f681550c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AnyFlow_[[:space:]]Arbitrary[[:space:]]Scale[[:space:]]Optical[[:space:]]Flow[[:space:]]With[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representation/5b5008ca-4652-483e-a296-f0ad3554de89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Architectural[[:space:]]Backdoors[[:space:]]in[[:space:]]Neural[[:space:]]Networks/299b826f-cd83-4c87-96a7-44f4156f5b34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Architecture,[[:space:]]Dataset[[:space:]]and[[:space:]]Model-Scale[[:space:]]Agnostic[[:space:]]Data-Free[[:space:]]Meta-Learning/1065addd-b31e-4e37-bc88-7b257f35e63f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Are[[:space:]]Binary[[:space:]]Annotations[[:space:]]Sufficient_[[:space:]]Video[[:space:]]Moment[[:space:]]Retrieval[[:space:]]via[[:space:]]Hierarchical[[:space:]]Uncertainty-Based[[:space:]]Active[[:space:]]Learning/0eb02ea1-c23c-4ac6-bd57-65cb9e7b1a4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Are[[:space:]]Data-Driven[[:space:]]Explanations[[:space:]]Robust[[:space:]]Against[[:space:]]Out-of-Distribution[[:space:]]Data_/cb3fd91d-399b-49d0-baae-bec3571ac48a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Are[[:space:]]Deep[[:space:]]Neural[[:space:]]Networks[[:space:]]SMARTer[[:space:]]Than[[:space:]]Second[[:space:]]Graders_/df57b28f-9242-4513-9dff-8bb62891e5b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Are[[:space:]]We[[:space:]]Ready[[:space:]]for[[:space:]]Vision-Centric[[:space:]]Driving[[:space:]]Streaming[[:space:]]Perception_[[:space:]]The[[:space:]]ASAP[[:space:]]Benchmark/a4f42a10-3024-46f5-b96b-be3c5a994b7a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AssemblyHands_[[:space:]]Towards[[:space:]]Egocentric[[:space:]]Activity[[:space:]]Understanding[[:space:]]via[[:space:]]3D[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation/8c217c4a-fb55-4aaf-a4cc-605f2503ceec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AstroNet_[[:space:]]When[[:space:]]Astrocyte[[:space:]]Meets[[:space:]]Artificial[[:space:]]Neural[[:space:]]Network/419a37a1-126e-4169-bfb2-71d361673e04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AsyFOD_[[:space:]]An[[:space:]]Asymmetric[[:space:]]Adaptation[[:space:]]Paradigm[[:space:]]for[[:space:]]Few-Shot[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detection/fb92a939-cb5b-4385-89d0-0b06cf3ef060_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Asymmetric[[:space:]]Feature[[:space:]]Fusion[[:space:]]for[[:space:]]Image[[:space:]]Retrieval/6a0b96fb-63b5-44fc-ac06-c069a37d4ff1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Attention-Based[[:space:]]Point[[:space:]]Cloud[[:space:]]Edge[[:space:]]Sampling/2fc0394d-b689-46d1-88e3-429be718d4c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AttentionShift_[[:space:]]Iteratively[[:space:]]Estimated[[:space:]]Part-Based[[:space:]]Attention[[:space:]]Map[[:space:]]for[[:space:]]Pointly[[:space:]]Supervised[[:space:]]Instance[[:space:]]Segmentation/40cc6988-e09e-472d-a1b0-8a2eb070b622_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AttriCLIP_[[:space:]]A[[:space:]]Non-Incremental[[:space:]]Learner[[:space:]]for[[:space:]]Incremental[[:space:]]Knowledge[[:space:]]Learning/e241f2cb-19e9-4b95-8eb2-0c18410b9b4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Attribute-Preserving[[:space:]]Face[[:space:]]Dataset[[:space:]]Anonymization[[:space:]]via[[:space:]]Latent[[:space:]]Code[[:space:]]Optimization/93e9672d-b548-42e5-b635-3b4b6f7ec320_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Audio-Visual[[:space:]]Grouping[[:space:]]Network[[:space:]]for[[:space:]]Sound[[:space:]]Localization[[:space:]]From[[:space:]]Mixtures/2877481c-4ab3-4715-945d-0f77ed894aaa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Augmentation[[:space:]]Matters_[[:space:]]A[[:space:]]Simple-Yet-Effective[[:space:]]Approach[[:space:]]to[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation/25e7c5f6-7494-4d44-be01-a322effbcb17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Auto-CARD_[[:space:]]Efficient[[:space:]]and[[:space:]]Robust[[:space:]]Codec[[:space:]]Avatar[[:space:]]Driving[[:space:]]for[[:space:]]Real-Time[[:space:]]Mobile[[:space:]]Telepresence/b86ea637-2598-4b4b-8848-c12429d59300_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AutoAD_[[:space:]]Movie[[:space:]]Description[[:space:]]in[[:space:]]Context/011458d2-abe0-4152-afe8-6d1de29478cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AutoFocusFormer_[[:space:]]Image[[:space:]]Segmentation[[:space:]]off[[:space:]]the[[:space:]]Grid/61ea6171-e7c4-46e3-8997-63520cc1398f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AutoLabel_[[:space:]]CLIP-Based[[:space:]]Framework[[:space:]]for[[:space:]]Open-Set[[:space:]]Video[[:space:]]Domain[[:space:]]Adaptation/c65aa4be-9695-4b69-85c0-1d3d5eb93bc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AutoRecon_[[:space:]]Automated[[:space:]]3D[[:space:]]Object[[:space:]]Discovery[[:space:]]and[[:space:]]Reconstruction/4b32ddb8-76d4-4c21-88bf-a89d711c0666_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Automatic[[:space:]]High[[:space:]]Resolution[[:space:]]Wire[[:space:]]Segmentation[[:space:]]and[[:space:]]Removal/464dabd8-ebe8-4fc7-b9bc-d265cb1f0ed7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Autonomous[[:space:]]Manipulation[[:space:]]Learning[[:space:]]for[[:space:]]Similar[[:space:]]Deformable[[:space:]]Objects[[:space:]]via[[:space:]]Only[[:space:]]One[[:space:]]Demonstration/7bba1dd6-2ecb-4e95-8642-457358fe74fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Autoregressive[[:space:]]Visual[[:space:]]Tracking/ef5b2ed4-22f8-4b05-9411-f2c8c72b8d97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Avatars[[:space:]]Grow[[:space:]]Legs_[[:space:]]Generating[[:space:]]Smooth[[:space:]]Human[[:space:]]Motion[[:space:]]From[[:space:]]Sparse[[:space:]]Tracking[[:space:]]Inputs[[:space:]]With[[:space:]]Diffusion[[:space:]]Model/a85d7e9a-53af-494a-bbd5-9d0ff401f9bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Azimuth[[:space:]]Super-Resolution[[:space:]]for[[:space:]]FMCW[[:space:]]Radar[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/b5f129c6-4554-42b3-9823-fadc9fb76a0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/B-Spline[[:space:]]Texture[[:space:]]Coefficients[[:space:]]Estimator[[:space:]]for[[:space:]]Screen[[:space:]]Content[[:space:]]Image[[:space:]]Super-Resolution/a4db501d-1eed-482e-99c4-8685225cdd7a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BAAM_[[:space:]]Monocular[[:space:]]3D[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Reconstruction[[:space:]]With[[:space:]]Bi-Contextual[[:space:]]Attention[[:space:]]Module[[:space:]]and[[:space:]]Attention-Guided[[:space:]]Modeling/6a7e4fbc-ebfd-4c86-8505-4c0cbf9f00cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BAD-NeRF_[[:space:]]Bundle[[:space:]]Adjusted[[:space:]]Deblur[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/a806cb90-c5e2-4fd8-b44c-b3448615c403_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BAEFormer_[[:space:]]Bi-Directional[[:space:]]and[[:space:]]Early[[:space:]]Interaction[[:space:]]Transformers[[:space:]]for[[:space:]]Bird's[[:space:]]Eye[[:space:]]View[[:space:]]Semantic[[:space:]]Segmentation/503ad3e9-7ebc-4243-a4e9-a500e02db1e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BASiS_[[:space:]]Batch[[:space:]]Aligned[[:space:]]Spectral[[:space:]]Embedding[[:space:]]Space/565f8bdf-9da3-467a-be7d-24db8abfb75f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BBDM_[[:space:]]Image-to-Image[[:space:]]Translation[[:space:]]With[[:space:]]Brownian[[:space:]]Bridge[[:space:]]Diffusion[[:space:]]Models/89d87aff-6464-4483-b0f9-258809929ad5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BEDLAM_[[:space:]]A[[:space:]]Synthetic[[:space:]]Dataset[[:space:]]of[[:space:]]Bodies[[:space:]]Exhibiting[[:space:]]Detailed[[:space:]]Lifelike[[:space:]]Animated[[:space:]]Motion/be72b6c2-3748-43ce-a3c9-631ac8633e13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BEV-Guided[[:space:]]Multi-Modality[[:space:]]Fusion[[:space:]]for[[:space:]]Driving[[:space:]]Perception/dbfd9f80-9f02-4c0b-b516-35088297a83b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BEV-LaneDet_[[:space:]]An[[:space:]]Efficient[[:space:]]3D[[:space:]]Lane[[:space:]]Detection[[:space:]]Based[[:space:]]on[[:space:]]Virtual[[:space:]]Camera[[:space:]]via[[:space:]]Key-Points/11522d31-c0fa-4fcf-b9e5-1f9d3afe8599_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BEV-SAN_[[:space:]]Accurate[[:space:]]BEV[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Slice[[:space:]]Attention[[:space:]]Networks/a07a3ebe-432e-4827-bdfb-68ff6b9a4a4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BEV@DC_[[:space:]]Bird's-Eye[[:space:]]View[[:space:]]Assisted[[:space:]]Training[[:space:]]for[[:space:]]Depth[[:space:]]Completion/312f4c96-1d4c-4d55-a8f3-641552a1bb07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BEVFormer[[:space:]]v2_[[:space:]]Adapting[[:space:]]Modern[[:space:]]Image[[:space:]]Backbones[[:space:]]to[[:space:]]Bird's-Eye-View[[:space:]]Recognition[[:space:]]via[[:space:]]Perspective[[:space:]]Supervision/1f4134b1-4078-4bf3-b723-133aeb149dcb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BEVHeight_[[:space:]]A[[:space:]]Robust[[:space:]]Framework[[:space:]]for[[:space:]]Vision-Based[[:space:]]Roadside[[:space:]]3D[[:space:]]Object[[:space:]]Detection/c5222895-184d-4bb8-8744-b960f1806460_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BITE_[[:space:]]Beyond[[:space:]]Priors[[:space:]]for[[:space:]]Improved[[:space:]]Three-D[[:space:]]Dog[[:space:]]Pose[[:space:]]Estimation/dfa5ac30-f058-4f9a-a334-464fce58a602_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BKinD-3D_[[:space:]]Self-Supervised[[:space:]]3D[[:space:]]Keypoint[[:space:]]Discovery[[:space:]]From[[:space:]]Multi-View[[:space:]]Videos/d46e0a86-7025-43f6-9dcd-5c85c3dd9a1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BUFFER_[[:space:]]Balancing[[:space:]]Accuracy,[[:space:]]Efficiency,[[:space:]]and[[:space:]]Generalizability[[:space:]]in[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/21a39bc1-0ef1-46e2-a32f-bea71a2cc511_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BUOL_[[:space:]]A[[:space:]]Bottom-Up[[:space:]]Framework[[:space:]]With[[:space:]]Occupancy-Aware[[:space:]]Lifting[[:space:]]for[[:space:]]Panoptic[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Image/2d33e805-bfdb-42e1-93db-ca0142ed7b3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Back[[:space:]]to[[:space:]]the[[:space:]]Source_[[:space:]]Diffusion-Driven[[:space:]]Adaptation[[:space:]]To[[:space:]]Test-Time[[:space:]]Corruption/e2064909-f2da-4931-a7c1-7557b6264505_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Backdoor[[:space:]]Attacks[[:space:]]Against[[:space:]]Deep[[:space:]]Image[[:space:]]Compression[[:space:]]via[[:space:]]Adaptive[[:space:]]Frequency[[:space:]]Trigger/dfd19d8b-e576-4436-92af-2d0513f292d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Backdoor[[:space:]]Cleansing[[:space:]]With[[:space:]]Unlabeled[[:space:]]Data/e6f74d09-f2c6-4617-96e7-8a5a486dfc13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Backdoor[[:space:]]Defense[[:space:]]via[[:space:]]Adaptively[[:space:]]Splitting[[:space:]]Poisoned[[:space:]]Dataset/ac7885e2-f435-4f69-90cb-fcb2ace6cd10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Backdoor[[:space:]]Defense[[:space:]]via[[:space:]]Deconfounded[[:space:]]Representation[[:space:]]Learning/748b9edf-bf9c-480e-9086-40ebfc2827e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Balanced[[:space:]]Energy[[:space:]]Regularization[[:space:]]Loss[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection/d50a6a54-13b1-4aa0-a6c9-d1be24c46a62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Balanced[[:space:]]Product[[:space:]]of[[:space:]]Calibrated[[:space:]]Experts[[:space:]]for[[:space:]]Long-Tailed[[:space:]]Recognition/2cee0887-cfe5-4a87-98ad-70a903543378_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Balanced[[:space:]]Spherical[[:space:]]Grid[[:space:]]for[[:space:]]Egocentric[[:space:]]View[[:space:]]Synthesis/0a20c573-7897-4f9c-9610-588567f74304_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Balancing[[:space:]]Logit[[:space:]]Variation[[:space:]]for[[:space:]]Long-Tailed[[:space:]]Semantic[[:space:]]Segmentation/6fc96245-ac01-4ca9-8a69-d2965b1f132d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Batch[[:space:]]Model[[:space:]]Consolidation_[[:space:]]A[[:space:]]Multi-Task[[:space:]]Model[[:space:]]Consolidation[[:space:]]Framework/740e5a68-48b8-4b34-9790-c92625654c4c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bayesian[[:space:]]Posterior[[:space:]]Approximation[[:space:]]With[[:space:]]Stochastic[[:space:]]Ensembles/ed10b98e-3a59-4ec7-9093-b69d64baf92b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Behavioral[[:space:]]Analysis[[:space:]]of[[:space:]]Vision-and-Language[[:space:]]Navigation[[:space:]]Agents/2f4996d1-1ca5-43aa-b480-d03926d34525_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Behind[[:space:]]the[[:space:]]Scenes_[[:space:]]Density[[:space:]]Fields[[:space:]]for[[:space:]]Single[[:space:]]View[[:space:]]Reconstruction/bd168156-a458-4b2e-9ebc-9d0ffa38ba35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Being[[:space:]]Comes[[:space:]]From[[:space:]]Not-Being_[[:space:]]Open-Vocabulary[[:space:]]Text-to-Motion[[:space:]]Generation[[:space:]]With[[:space:]]Wordless[[:space:]]Training/dc07f486-0202-4f88-95cc-640e803e4050_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Benchmarking[[:space:]]Robustness[[:space:]]of[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]to[[:space:]]Common[[:space:]]Corruptions/e67ba5d3-9ea6-4a76-b4c3-08ee53b6c4ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Benchmarking[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]on[[:space:]]Diverse[[:space:]]Pathology[[:space:]]Datasets/2d6f4cb4-4eee-4864-95a8-d27b039aeb87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Best[[:space:]]of[[:space:]]Both[[:space:]]Worlds_[[:space:]]Multimodal[[:space:]]Contrastive[[:space:]]Learning[[:space:]]With[[:space:]]Tabular[[:space:]]and[[:space:]]Imaging[[:space:]]Data/839009bf-5700-4375-b350-993f7ad7b833_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Better[[:space:]]'CMOS'[[:space:]]Produces[[:space:]]Clearer[[:space:]]Images_[[:space:]]Learning[[:space:]]Space-Variant[[:space:]]Blur[[:space:]]Estimation[[:space:]]for[[:space:]]Blind[[:space:]]Image[[:space:]]Super-Resolution/860a1969-84a5-475c-830a-9bec5c337fc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Beyond[[:space:]]Appearance_[[:space:]]A[[:space:]]Semantic[[:space:]]Controllable[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]Framework[[:space:]]for[[:space:]]Human-Centric[[:space:]]Visual[[:space:]]Tasks/ce1149fb-258e-4437-bebd-dd2ab4e2f7e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Beyond[[:space:]]Attentive[[:space:]]Tokens_[[:space:]]Incorporating[[:space:]]Token[[:space:]]Importance[[:space:]]and[[:space:]]Diversity[[:space:]]for[[:space:]]Efficient[[:space:]]Vision[[:space:]]Transformers/b1aa69cd-f2f3-4abd-b94a-05ef53119b15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Beyond[[:space:]]mAP_[[:space:]]Towards[[:space:]]Better[[:space:]]Evaluation[[:space:]]of[[:space:]]Instance[[:space:]]Segmentation/71f94499-ffb3-426d-8f17-ddebe8102777_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bi-Directional[[:space:]]Distribution[[:space:]]Alignment[[:space:]]for[[:space:]]Transductive[[:space:]]Zero-Shot[[:space:]]Learning/97bb29e4-6f90-4796-8a16-5cb404ef8a77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bi-Directional[[:space:]]Feature[[:space:]]Fusion[[:space:]]Generative[[:space:]]Adversarial[[:space:]]Network[[:space:]]for[[:space:]]Ultra-High[[:space:]]Resolution[[:space:]]Pathological[[:space:]]Image[[:space:]]Virtual[[:space:]]Re-Staining/f2998ddc-bd5c-4ded-883c-844bce3ae61b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bi-LRFusion_[[:space:]]Bi-Directional[[:space:]]LiDAR-Radar[[:space:]]Fusion[[:space:]]for[[:space:]]3D[[:space:]]Dynamic[[:space:]]Object[[:space:]]Detection/dd3f2e7b-adc6-4d1f-8f2d-c239adbd1eb8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bi-Level[[:space:]]Meta-Learning[[:space:]]for[[:space:]]Few-Shot[[:space:]]Domain[[:space:]]Generalization/e9a12147-3803-457c-b21f-5fd45ab6ecd3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bi3D_[[:space:]]Bi-Domain[[:space:]]Active[[:space:]]Learning[[:space:]]for[[:space:]]Cross-Domain[[:space:]]3D[[:space:]]Object[[:space:]]Detection/c0652f32-3ce0-47cd-aeb5-4963ea5ba2a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BiCro_[[:space:]]Noisy[[:space:]]Correspondence[[:space:]]Rectification[[:space:]]for[[:space:]]Multi-Modality[[:space:]]Data[[:space:]]via[[:space:]]Bi-Directional[[:space:]]Cross-Modal[[:space:]]Similarity[[:space:]]Consistency/b7e147f8-7f4a-4e55-afe3-0b24c69024b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BiFormer_[[:space:]]Learning[[:space:]]Bilateral[[:space:]]Motion[[:space:]]Estimation[[:space:]]via[[:space:]]Bilateral[[:space:]]Transformer[[:space:]]for[[:space:]]4K[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation/d64ddd23-6dd9-4653-adc6-a839ec93584b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BiFormer_[[:space:]]Vision[[:space:]]Transformer[[:space:]]With[[:space:]]Bi-Level[[:space:]]Routing[[:space:]]Attention/98515e7a-7010-4ec0-9250-606bc6e69c51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bias[[:space:]]Mimicking_[[:space:]]A[[:space:]]Simple[[:space:]]Sampling[[:space:]]Approach[[:space:]]for[[:space:]]Bias[[:space:]]Mitigation/bf04d582-570b-4408-918e-fbf3f1361d4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bias[[:space:]]in[[:space:]]Pruned[[:space:]]Vision[[:space:]]Models_[[:space:]]In-Depth[[:space:]]Analysis[[:space:]]and[[:space:]]Countermeasures/6cbe77eb-9502-4808-8812-3900c8414176_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bias-Eliminating[[:space:]]Augmentation[[:space:]]Learning[[:space:]]for[[:space:]]Debiased[[:space:]]Federated[[:space:]]Learning/2b44171b-a9cf-4d52-afed-913fddd7466d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BiasAdv_[[:space:]]Bias-Adversarial[[:space:]]Augmentation[[:space:]]for[[:space:]]Model[[:space:]]Debiasing/deb908be-d968-489d-be85-13e06ec7de6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BiasBed[[:space:]]-[[:space:]]Rigorous[[:space:]]Texture[[:space:]]Bias[[:space:]]Evaluation/b30f0710-9624-4c7d-9dbd-ca404a8832cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bidirectional[[:space:]]Copy-Paste[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/bf919a33-7357-4875-b015-b5c6e529f76f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bidirectional[[:space:]]Cross-Modal[[:space:]]Knowledge[[:space:]]Exploration[[:space:]]for[[:space:]]Video[[:space:]]Recognition[[:space:]]With[[:space:]]Pre-Trained[[:space:]]Vision-Language[[:space:]]Models/836cef1b-823e-4d94-a1b6-c7fc1dac3725_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bilateral[[:space:]]Memory[[:space:]]Consolidation[[:space:]]for[[:space:]]Continual[[:space:]]Learning/e0dec229-2bee-49d7-8316-358552f6e848_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Binarizing[[:space:]]Sparse[[:space:]]Convolutional[[:space:]]Networks[[:space:]]for[[:space:]]Efficient[[:space:]]Point[[:space:]]Cloud[[:space:]]Analysis/8eccbfd9-fadd-48ac-88cf-1cf861e9db76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Binary[[:space:]]Latent[[:space:]]Diffusion/c2b3a483-8ce5-4a4b-b61c-b5ec826f657b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BioNet_[[:space:]]A[[:space:]]Biologically-Inspired[[:space:]]Network[[:space:]]for[[:space:]]Face[[:space:]]Recognition/dd138417-3bdb-499e-8901-0e96b61ff9d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Biomechanics-Guided[[:space:]]Facial[[:space:]]Action[[:space:]]Unit[[:space:]]Detection[[:space:]]Through[[:space:]]Force[[:space:]]Modeling/60c3f3ea-a133-4fc5-9529-cd39a5efb948_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bit-Shrinking_[[:space:]]Limiting[[:space:]]Instantaneous[[:space:]]Sharpness[[:space:]]for[[:space:]]Improving[[:space:]]Post-Training[[:space:]]Quantization/de9524a3-dc5a-40fb-9479-036ae731c09b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bitstream-Corrupted[[:space:]]JPEG[[:space:]]Images[[:space:]]Are[[:space:]]Restorable_[[:space:]]Two-Stage[[:space:]]Compensation[[:space:]]and[[:space:]]Alignment[[:space:]]Framework[[:space:]]for[[:space:]]Image[[:space:]]Restoration/42bc4dd1-bda7-47f2-b32b-1d1f9328b0f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Black-Box[[:space:]]Sparse[[:space:]]Adversarial[[:space:]]Attack[[:space:]]via[[:space:]]Multi-Objective[[:space:]]Optimisation/f6846207-5b2c-4caa-9228-f839a8eb3efb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BlackVIP_[[:space:]]Black-Box[[:space:]]Visual[[:space:]]Prompting[[:space:]]for[[:space:]]Robust[[:space:]]Transfer[[:space:]]Learning/ed2e9801-cd6e-47c3-ad5e-52faea019619_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Blemish-Aware[[:space:]]and[[:space:]]Progressive[[:space:]]Face[[:space:]]Retouching[[:space:]]With[[:space:]]Limited[[:space:]]Paired[[:space:]]Data/e5d0a1c4-90ea-4383-8aff-70734a098f7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BlendFields_[[:space:]]Few-Shot[[:space:]]Example-Driven[[:space:]]Facial[[:space:]]Modeling/be6a7937-d3a9-4acf-aa11-76486c96a44e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Blind[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment[[:space:]]via[[:space:]]Vision-Language[[:space:]]Correspondence_[[:space:]]A[[:space:]]Multitask[[:space:]]Learning[[:space:]]Perspective/278cb47e-46d5-4f57-94e9-354c873f7ab8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Blind[[:space:]]Video[[:space:]]Deflickering[[:space:]]by[[:space:]]Neural[[:space:]]Filtering[[:space:]]With[[:space:]]a[[:space:]]Flawed[[:space:]]Atlas/865f391c-59e0-4d9f-8399-ac34a0fcca6c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Block[[:space:]]Selection[[:space:]]Method[[:space:]]for[[:space:]]Using[[:space:]]Feature[[:space:]]Norm[[:space:]]in[[:space:]]Out-of-Distribution[[:space:]]Detection/25caa4c2-3e98-4e6b-b64a-6e6997e1e7e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Blowing[[:space:]]in[[:space:]]the[[:space:]]Wind_[[:space:]]CycleNet[[:space:]]for[[:space:]]Human[[:space:]]Cinemagraphs[[:space:]]From[[:space:]]Still[[:space:]]Images/dc29955e-dac2-4aae-aaf8-800e19cf8412_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Blur[[:space:]]Interpolation[[:space:]]Transformer[[:space:]]for[[:space:]]Real-World[[:space:]]Motion[[:space:]]From[[:space:]]Blur/ef4b3596-6c29-4306-88e6-57961948f3c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boost[[:space:]]Vision[[:space:]]Transformer[[:space:]]With[[:space:]]GPU-Friendly[[:space:]]Sparsity[[:space:]]and[[:space:]]Quantization/8ce9d2a7-cfbb-4a60-9aaf-fd7d90bdfb3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Accuracy[[:space:]]and[[:space:]]Robustness[[:space:]]of[[:space:]]Student[[:space:]]Models[[:space:]]via[[:space:]]Adaptive[[:space:]]Adversarial[[:space:]]Distillation/3d34d594-290c-4b50-85bb-d5de7b089346_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Detection[[:space:]]in[[:space:]]Crowd[[:space:]]Analysis[[:space:]]via[[:space:]]Underutilized[[:space:]]Output[[:space:]]Features/4b5807e4-133e-4163-bfb3-5c9ea4055d48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Low-Data[[:space:]]Instance[[:space:]]Segmentation[[:space:]]by[[:space:]]Unsupervised[[:space:]]Pre-Training[[:space:]]With[[:space:]]Saliency[[:space:]]Prompt/e6591e13-bda4-44dd-b426-50bb49a38d5e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Semi-Supervised[[:space:]]Learning[[:space:]]by[[:space:]]Exploiting[[:space:]]All[[:space:]]Unlabeled[[:space:]]Data/17f31ffa-7421-45e3-8bce-f085dde98823_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Transductive[[:space:]]Few-Shot[[:space:]]Fine-Tuning[[:space:]]With[[:space:]]Margin-Based[[:space:]]Uncertainty[[:space:]]Weighting[[:space:]]and[[:space:]]Probability[[:space:]]Regularization/dc915ce2-1eda-4cc1-a26c-a00973492a24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Verified[[:space:]]Training[[:space:]]for[[:space:]]Robust[[:space:]]Image[[:space:]]Classifications[[:space:]]via[[:space:]]Abstraction/4f8d7321-be83-48c3-abba-39230357c7d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation[[:space:]]via[[:space:]]Space-Time[[:space:]]Correspondence[[:space:]]Learning/49aba0a9-1562-4f4c-b4f5-e730e8322046_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Weakly-Supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization[[:space:]]With[[:space:]]Text[[:space:]]Information/48634870-26df-4c80-9677-d08e367d6911_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bootstrap[[:space:]]Your[[:space:]]Own[[:space:]]Prior_[[:space:]]Towards[[:space:]]Distribution-Agnostic[[:space:]]Novel[[:space:]]Class[[:space:]]Discovery/10137e9f-f8c4-4bd0-b99d-6c77616d4219_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bootstrapping[[:space:]]Objectness[[:space:]]From[[:space:]]Videos[[:space:]]by[[:space:]]Relaxed[[:space:]]Common[[:space:]]Fate[[:space:]]and[[:space:]]Visual[[:space:]]Grouping/88d89fef-eb64-45c7-9f62-5506ff6dd163_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Both[[:space:]]Style[[:space:]]and[[:space:]]Distortion[[:space:]]Matter_[[:space:]]Dual-Path[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Panoramic[[:space:]]Semantic[[:space:]]Segmentation/89bb9107-d3b7-4237-9808-416c50cc99f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boundary[[:space:]]Unlearning_[[:space:]]Rapid[[:space:]]Forgetting[[:space:]]of[[:space:]]Deep[[:space:]]Networks[[:space:]]via[[:space:]]Shifting[[:space:]]the[[:space:]]Decision[[:space:]]Boundary/08915d68-0d50-40d5-b8b3-63fededaac68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boundary-Aware[[:space:]]Backward-Compatible[[:space:]]Representation[[:space:]]via[[:space:]]Adversarial[[:space:]]Learning[[:space:]]in[[:space:]]Image[[:space:]]Retrieval/aeaa696a-195f-4fb6-bc68-1ed42f82bd4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boundary-Enhanced[[:space:]]Co-Training[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/9cf607e0-5857-4eed-a8a0-562ce1572890_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Box-Level[[:space:]]Active[[:space:]]Detection/2254bd04-9bd9-47b1-8293-48c724c20285_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BoxTeacher_[[:space:]]Exploring[[:space:]]High-Quality[[:space:]]Pseudo[[:space:]]Labels[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Instance[[:space:]]Segmentation/dec6fe6c-7740-461f-8e40-4edd95ee4cc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Breaching[[:space:]]FedMD_[[:space:]]Image[[:space:]]Recovery[[:space:]]via[[:space:]]Paired-Logits[[:space:]]Inversion[[:space:]]Attack/953cdf94-182e-438a-b8a3-c8314b5eba1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Breaking[[:space:]]the[[:space:]]'Object'[[:space:]]in[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/925f8052-4a7e-48a2-97a5-0cf46c4817c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bridging[[:space:]]Precision[[:space:]]and[[:space:]]Confidence_[[:space:]]A[[:space:]]Train-Time[[:space:]]Loss[[:space:]]for[[:space:]]Calibrating[[:space:]]Object[[:space:]]Detection/7d50f944-6bff-47ca-81a7-21f4127a55f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bridging[[:space:]]Search[[:space:]]Region[[:space:]]Interaction[[:space:]]With[[:space:]]Template[[:space:]]for[[:space:]]RGB-T[[:space:]]Tracking/15c16d1f-b3d4-43e0-8709-824745cc14c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bridging[[:space:]]the[[:space:]]Gap[[:space:]]Between[[:space:]]Model[[:space:]]Explanations[[:space:]]in[[:space:]]Partially[[:space:]]Annotated[[:space:]]Multi-Label[[:space:]]Classification/22a5e23e-8e25-4bd7-9200-71f3916bdea5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bringing[[:space:]]Inputs[[:space:]]to[[:space:]]Shared[[:space:]]Domains[[:space:]]for[[:space:]]3D[[:space:]]Interacting[[:space:]]Hands[[:space:]]Recovery[[:space:]]in[[:space:]]the[[:space:]]Wild/16b9e757-e360-42a1-a02f-2703c36248f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Building[[:space:]]Rearticulable[[:space:]]Models[[:space:]]for[[:space:]]Arbitrary[[:space:]]3D[[:space:]]Objects[[:space:]]From[[:space:]]4D[[:space:]]Point[[:space:]]Clouds/37577af3-d3ba-471b-8210-6bb12959f1b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BundleSDF_[[:space:]]Neural[[:space:]]6-DoF[[:space:]]Tracking[[:space:]]and[[:space:]]3D[[:space:]]Reconstruction[[:space:]]of[[:space:]]Unknown[[:space:]]Objects/afd4dbac-61f8-4b21-9472-f8dc2398804f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Burstormer_[[:space:]]Burst[[:space:]]Image[[:space:]]Restoration[[:space:]]and[[:space:]]Enhancement[[:space:]]Transformer/510b8ae2-2933-4c60-adc1-11383e9f3785_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/C-SFDA_[[:space:]]A[[:space:]]Curriculum[[:space:]]Learning[[:space:]]Aided[[:space:]]Self-Training[[:space:]]Framework[[:space:]]for[[:space:]]Efficient[[:space:]]Source[[:space:]]Free[[:space:]]Domain[[:space:]]Adaptation/e056e06e-157a-46d6-b867-7616a9c15968_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CABM_[[:space:]]Content-Aware[[:space:]]Bit[[:space:]]Mapping[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]Network[[:space:]]With[[:space:]]Large[[:space:]]Input/12a17478-d96f-4f11-929c-92eec393d2ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CAMS_[[:space:]]CAnonicalized[[:space:]]Manipulation[[:space:]]Spaces[[:space:]]for[[:space:]]Category-Level[[:space:]]Functional[[:space:]]Hand-Object[[:space:]]Manipulation[[:space:]]Synthesis/f8c63aed-ea30-45f5-9774-5469caceb3f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CAP-VSTNet_[[:space:]]Content[[:space:]]Affinity[[:space:]]Preserved[[:space:]]Versatile[[:space:]]Style[[:space:]]Transfer/e154969a-307d-41b0-b215-0986c9541add_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CAPE_[[:space:]]Camera[[:space:]]View[[:space:]]Position[[:space:]]Embedding[[:space:]]for[[:space:]]Multi-View[[:space:]]3D[[:space:]]Object[[:space:]]Detection/024b0ccb-ae5d-4ebb-8c8f-af21c3f79bf5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CAP_[[:space:]]Robust[[:space:]]Point[[:space:]]Cloud[[:space:]]Classification[[:space:]]via[[:space:]]Semantic[[:space:]]and[[:space:]]Structural[[:space:]]Modeling/7ac09030-f12f-4e8a-be0b-cb49a50d5561_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CARTO_[[:space:]]Category[[:space:]]and[[:space:]]Joint[[:space:]]Agnostic[[:space:]]Reconstruction[[:space:]]of[[:space:]]ARTiculated[[:space:]]Objects/c8ca292e-2adb-4b1c-851a-6416d7ad31ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CASP-Net_[[:space:]]Rethinking[[:space:]]Video[[:space:]]Saliency[[:space:]]Prediction[[:space:]]From[[:space:]]an[[:space:]]Audio-Visual[[:space:]]Consistency[[:space:]]Perceptual[[:space:]]Perspective/7e2afa9f-6e88-4f98-9128-a5e863bc206a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CAT_[[:space:]]LoCalization[[:space:]]and[[:space:]]IdentificAtion[[:space:]]Cascade[[:space:]]Detection[[:space:]]Transformer[[:space:]]for[[:space:]]Open-World[[:space:]]Object[[:space:]]Detection/05f8d209-763c-4cd4-831c-277edb8e958b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CCuantuMM_[[:space:]]Cycle-Consistent[[:space:]]Quantum-Hybrid[[:space:]]Matching[[:space:]]of[[:space:]]Multiple[[:space:]]Shapes/bbd609b4-630e-4fd2-8fcd-332ed43c8040_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CDDFuse_[[:space:]]Correlation-Driven[[:space:]]Dual-Branch[[:space:]]Feature[[:space:]]Decomposition[[:space:]]for[[:space:]]Multi-Modality[[:space:]]Image[[:space:]]Fusion/fd3107ed-7c9c-4a48-ad97-0da2d3a477af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CF-Font_[[:space:]]Content[[:space:]]Fusion[[:space:]]for[[:space:]]Few-Shot[[:space:]]Font[[:space:]]Generation/395d9918-af5c-48a1-ad32-1f39eed23fbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CFA_[[:space:]]Class-Wise[[:space:]]Calibrated[[:space:]]Fair[[:space:]]Adversarial[[:space:]]Training/bb4d7a9a-b4b4-4303-be00-1db5d6c18d16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CHMATCH_[[:space:]]Contrastive[[:space:]]Hierarchical[[:space:]]Matching[[:space:]]and[[:space:]]Robust[[:space:]]Adaptive[[:space:]]Threshold[[:space:]]Boosted[[:space:]]Semi-Supervised[[:space:]]Learning/a37ce418-2547-49c1-9988-9610ba014f00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CIGAR_[[:space:]]Cross-Modality[[:space:]]Graph[[:space:]]Reasoning[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detection/b334ffc7-7787-4914-b5d5-cfdc541dcdb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CIMI4D_[[:space:]]A[[:space:]]Large[[:space:]]Multimodal[[:space:]]Climbing[[:space:]]Motion[[:space:]]Dataset[[:space:]]Under[[:space:]]Human-Scene[[:space:]]Interactions/9690b548-8d50-4b7a-a21d-b2d4614486a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CIRCLE_[[:space:]]Capture[[:space:]]in[[:space:]]Rich[[:space:]]Contextual[[:space:]]Environments/7cbfda4e-5ce1-47f8-8e06-90a3de7278b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLAMP_[[:space:]]Prompt-Based[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Connecting[[:space:]]Language[[:space:]]and[[:space:]]Animal[[:space:]]Pose/8ae09e2c-7a8c-4482-9aec-5625c4c90e8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIP[[:space:]]Is[[:space:]]Also[[:space:]]an[[:space:]]Efficient[[:space:]]Segmenter_[[:space:]]A[[:space:]]Text-Driven[[:space:]]Approach[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/cffa99de-dd8c-4a70-9b1f-5f8553581a2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIP[[:space:]]for[[:space:]]All[[:space:]]Things[[:space:]]Zero-Shot[[:space:]]Sketch-Based[[:space:]]Image[[:space:]]Retrieval,[[:space:]]Fine-Grained[[:space:]]or[[:space:]]Not/e8e0cef4-6aa4-43b7-996e-85dcc9fbe77b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIP[[:space:]]the[[:space:]]Gap_[[:space:]]A[[:space:]]Single[[:space:]]Domain[[:space:]]Generalization[[:space:]]Approach[[:space:]]for[[:space:]]Object[[:space:]]Detection/8bfd13ef-ca5c-4f47-a9f9-d8d7eecb88de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIP-S4_[[:space:]]Language-Guided[[:space:]]Self-Supervised[[:space:]]Semantic[[:space:]]Segmentation/c24eb4b4-f000-47c7-8244-3296ba8e2cac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIP-Sculptor_[[:space:]]Zero-Shot[[:space:]]Generation[[:space:]]of[[:space:]]High-Fidelity[[:space:]]and[[:space:]]Diverse[[:space:]]Shapes[[:space:]]From[[:space:]]Natural[[:space:]]Language/e64fac6d-0780-4018-a6c0-4f909bfb4337_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIP2Protect_[[:space:]]Protecting[[:space:]]Facial[[:space:]]Privacy[[:space:]]Using[[:space:]]Text-Guided[[:space:]]Makeup[[:space:]]via[[:space:]]Adversarial[[:space:]]Latent[[:space:]]Search/d379e97e-b7a0-42bb-a735-a3de89cdb054_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIP2Scene_[[:space:]]Towards[[:space:]]Label-Efficient[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding[[:space:]]by[[:space:]]CLIP/58dade4c-9eec-4c3a-968d-a582ab672d68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIP2_[[:space:]]Contrastive[[:space:]]Language-Image-Point[[:space:]]Pretraining[[:space:]]From[[:space:]]Real-World[[:space:]]Point[[:space:]]Cloud[[:space:]]Data/b3efe108-020c-4b4b-bc30-fbdcfe854136_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIPPING_[[:space:]]Distilling[[:space:]]CLIP-Based[[:space:]]Models[[:space:]]With[[:space:]]a[[:space:]]Student[[:space:]]Base[[:space:]]for[[:space:]]Video-Language[[:space:]]Retrieval/1eb2010b-78e6-492d-9404-5d49a6b93ef0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIPPO_[[:space:]]Image-and-Language[[:space:]]Understanding[[:space:]]From[[:space:]]Pixels[[:space:]]Only/6a9a5c21-30d9-4b55-b8d0-e491b4946877_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLOTH4D_[[:space:]]A[[:space:]]Dataset[[:space:]]for[[:space:]]Clothed[[:space:]]Human[[:space:]]Reconstruction/b8c1cf88-b3e6-48d2-932c-20d8bacc61eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CNVid-3.5M_[[:space:]]Build,[[:space:]]Filter,[[:space:]]and[[:space:]]Pre-Train[[:space:]]the[[:space:]]Large-Scale[[:space:]]Public[[:space:]]Chinese[[:space:]]Video-Text[[:space:]]Dataset/3819ec34-d037-4584-a0be-1f25bc0cb37d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CODA-Prompt_[[:space:]]COntinual[[:space:]]Decomposed[[:space:]]Attention-Based[[:space:]]Prompting[[:space:]]for[[:space:]]Rehearsal-Free[[:space:]]Continual[[:space:]]Learning/44360d49-0340-4846-917d-a266d277946e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CORA_[[:space:]]Adapting[[:space:]]CLIP[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Detection[[:space:]]With[[:space:]]Region[[:space:]]Prompting[[:space:]]and[[:space:]]Anchor[[:space:]]Pre-Matching/e257818b-1aea-4bd1-9288-7c6a81195cd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/COT_[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]With[[:space:]]Clustering[[:space:]]and[[:space:]]Optimal[[:space:]]Transport/53ca0802-c684-4d48-b297-7731efdaab53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CP3_[[:space:]]Channel[[:space:]]Pruning[[:space:]]Plug-In[[:space:]]for[[:space:]]Point-Based[[:space:]]Networks/f7e54c81-931a-42ba-b8dd-1b55f7342f80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CR-FIQA_[[:space:]]Face[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment[[:space:]]by[[:space:]]Learning[[:space:]]Sample[[:space:]]Relative[[:space:]]Classifiability/2ae421c8-bef3-40d8-a9e9-022734d8a26e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CRAFT_[[:space:]]Concept[[:space:]]Recursive[[:space:]]Activation[[:space:]]FacTorization[[:space:]]for[[:space:]]Explainability/6e462b05-7931-4758-8402-5b3e432aab49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CREPE_[[:space:]]Can[[:space:]]Vision-Language[[:space:]]Foundation[[:space:]]Models[[:space:]]Reason[[:space:]]Compositionally_/6b09bf56-aeff-4b97-8111-a915e425b709_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CUDA_[[:space:]]Convolution-Based[[:space:]]Unlearnable[[:space:]]Datasets/e0179aca-27a9-460d-b895-5e7378133f26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CUF_[[:space:]]Continuous[[:space:]]Upsampling[[:space:]]Filters/5b0cd1d3-2b2b-407b-9bec-e263533bd51d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CVT-SLR_[[:space:]]Contrastive[[:space:]]Visual-Textual[[:space:]]Transformation[[:space:]]for[[:space:]]Sign[[:space:]]Language[[:space:]]Recognition[[:space:]]With[[:space:]]Variational[[:space:]]Alignment/012722bf-9201-40ff-b0cf-52287586b76f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CXTrack_[[:space:]]Improving[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Tracking[[:space:]]With[[:space:]]Contextual[[:space:]]Information/c7b22be7-4a5c-4780-8a5a-1db50c1d60af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CaPriDe[[:space:]]Learning_[[:space:]]Confidential[[:space:]]and[[:space:]]Private[[:space:]]Decentralized[[:space:]]Learning[[:space:]]Based[[:space:]]on[[:space:]]Encryption-Friendly[[:space:]]Distillation[[:space:]]Loss/ca732fe2-4161-4d55-be24-51fc46dde867_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CafeBoost_[[:space:]]Causal[[:space:]]Feature[[:space:]]Boost[[:space:]]To[[:space:]]Eliminate[[:space:]]Task-Induced[[:space:]]Bias[[:space:]]for[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning/6f2564b9-d9c3-4c4f-b871-b0003952027b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Camouflaged[[:space:]]Instance[[:space:]]Segmentation[[:space:]]via[[:space:]]Explicit[[:space:]]De-Camouflaging/9acac0a8-5352-4eef-91f5-74adb1a1e4b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Camouflaged[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]Feature[[:space:]]Decomposition[[:space:]]and[[:space:]]Edge[[:space:]]Reconstruction/69e6313d-2f4d-49ec-aeb1-e472deb81070_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Can't[[:space:]]Steal_[[:space:]]Cont-Steal![[:space:]]Contrastive[[:space:]]Stealing[[:space:]]Attacks[[:space:]]Against[[:space:]]Image[[:space:]]Encoders/d755a24e-203a-45cb-a84e-020f7fb471b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Canonical[[:space:]]Fields_[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]of[[:space:]]Pose-Canonicalized[[:space:]]Neural[[:space:]]Fields/54da6270-9672-4d8f-9893-3339952d04d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cap4Video_[[:space:]]What[[:space:]]Can[[:space:]]Auxiliary[[:space:]]Captions[[:space:]]Do[[:space:]]for[[:space:]]Text-Video[[:space:]]Retrieval_/926c868d-a104-48e5-bcb0-c1285d26ad5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CapDet_[[:space:]]Unifying[[:space:]]Dense[[:space:]]Captioning[[:space:]]and[[:space:]]Open-World[[:space:]]Detection[[:space:]]Pretraining/c3eddfd4-ae1f-40f7-a7cf-96c54f1fea7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cascade[[:space:]]Evidential[[:space:]]Learning[[:space:]]for[[:space:]]Open-World[[:space:]]Weakly-Supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization/3c2843a4-d81f-4df7-8b2c-5c29eaeafad1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cascaded[[:space:]]Local[[:space:]]Implicit[[:space:]]Transformer[[:space:]]for[[:space:]]Arbitrary-Scale[[:space:]]Super-Resolution/ae047e51-87a0-4eec-807f-144baf2a48b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Castling-ViT_[[:space:]]Compressing[[:space:]]Self-Attention[[:space:]]via[[:space:]]Switching[[:space:]]Towards[[:space:]]Linear-Angular[[:space:]]Attention[[:space:]]at[[:space:]]Vision[[:space:]]Transformer[[:space:]]Inference/d74c49bb-effe-4caf-8858-0066e120e08d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Catch[[:space:]]Missing[[:space:]]Details_[[:space:]]Image[[:space:]]Reconstruction[[:space:]]With[[:space:]]Frequency[[:space:]]Augmented[[:space:]]Variational[[:space:]]Autoencoder/0a253bb0-c0db-46ba-9d61-8be0d603b311_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Category[[:space:]]Query[[:space:]]Learning[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Classification/c48c8f3f-b487-4c24-815b-6385a57d02e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Causally-Aware[[:space:]]Intraoperative[[:space:]]Imputation[[:space:]]for[[:space:]]Overall[[:space:]]Survival[[:space:]]Time[[:space:]]Prediction/58789d3b-4b75-4c2b-ad82-075b8268b238_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CelebV-Text_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Facial[[:space:]]Text-Video[[:space:]]Dataset/b3bc8119-1bc3-419d-89b1-046a6b2d92d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Center[[:space:]]Focusing[[:space:]]Network[[:space:]]for[[:space:]]Real-Time[[:space:]]LiDAR[[:space:]]Panoptic[[:space:]]Segmentation/e749a0dd-41b1-4b7d-9325-ea72d56e45cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Change-Aware[[:space:]]Sampling[[:space:]]and[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Satellite[[:space:]]Images/ef28aefe-c0fe-45ce-8981-6fc9107e6b77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Chat2Map_[[:space:]]Efficient[[:space:]]Scene[[:space:]]Mapping[[:space:]]From[[:space:]]Multi-Ego[[:space:]]Conversations/9562f06a-6203-4663-996c-df2336adbe79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CiCo_[[:space:]]Domain-Aware[[:space:]]Sign[[:space:]]Language[[:space:]]Retrieval[[:space:]]via[[:space:]]Cross-Lingual[[:space:]]Contrastive[[:space:]]Learning/047059b8-c0b2-4bcc-9546-ab9b6e4bc0e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CiaoSR_[[:space:]]Continuous[[:space:]]Implicit[[:space:]]Attention-in-Attention[[:space:]]Network[[:space:]]for[[:space:]]Arbitrary-Scale[[:space:]]Image[[:space:]]Super-Resolution/dcf45018-bd80-4a92-ad40-a481bde62c68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class[[:space:]]Adaptive[[:space:]]Network[[:space:]]Calibration/d3a59c4a-62e4-447c-b052-c6b11b8ed279_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class[[:space:]]Attention[[:space:]]Transfer[[:space:]]Based[[:space:]]Knowledge[[:space:]]Distillation/a68b12fe-6c62-4a74-9758-0370df04b2ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class[[:space:]]Balanced[[:space:]]Adaptive[[:space:]]Pseudo[[:space:]]Labeling[[:space:]]for[[:space:]]Federated[[:space:]]Semi-Supervised[[:space:]]Learning/27b4466b-7a94-415b-9a3c-c8f4f39a02f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class[[:space:]]Prototypes[[:space:]]Based[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Classifying[[:space:]]Multi-Label[[:space:]]and[[:space:]]Fine-Grained[[:space:]]Educational[[:space:]]Videos/bbbb89aa-912d-4dd5-8424-10e3f0db2ed4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class[[:space:]]Relationship[[:space:]]Embedded[[:space:]]Learning[[:space:]]for[[:space:]]Source-Free[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/a0503f36-a779-42c8-98b9-ccfd34b3ed5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class-Balancing[[:space:]]Diffusion[[:space:]]Models/eae79914-520a-4687-a1d4-1e04f53e44f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class-Conditional[[:space:]]Sharpness-Aware[[:space:]]Minimization[[:space:]]for[[:space:]]Deep[[:space:]]Long-Tailed[[:space:]]Recognition/5537ac60-2d54-48e8-9091-a4f2d7a57e6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class-Incremental[[:space:]]Exemplar[[:space:]]Compression[[:space:]]for[[:space:]]Class-Incremental[[:space:]]Learning/47061da4-a9f4-49a6-b036-8d247f22c620_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CloSET_[[:space:]]Modeling[[:space:]]Clothed[[:space:]]Humans[[:space:]]on[[:space:]]Continuous[[:space:]]Surface[[:space:]]With[[:space:]]Explicit[[:space:]]Template[[:space:]]Decomposition/4725afdd-f941-4b99-a289-05a751ed5c4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Clothed[[:space:]]Human[[:space:]]Performance[[:space:]]Capture[[:space:]]With[[:space:]]a[[:space:]]Double-Layer[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/edeeca28-ac72-40e3-9181-9737d0c908f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cloud-Device[[:space:]]Collaborative[[:space:]]Adaptation[[:space:]]to[[:space:]]Continual[[:space:]]Changing[[:space:]]Environments[[:space:]]in[[:space:]]the[[:space:]]Real-World/071f2817-bbaa-4424-acc0-27135344a577_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Clover_[[:space:]]Towards[[:space:]]a[[:space:]]Unified[[:space:]]Video-Language[[:space:]]Alignment[[:space:]]and[[:space:]]Fusion[[:space:]]Model/eddf9a27-0fae-4139-acd3-5b79001d8460_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Co-SLAM_[[:space:]]Joint[[:space:]]Coordinate[[:space:]]and[[:space:]]Sparse[[:space:]]Parametric[[:space:]]Encodings[[:space:]]for[[:space:]]Neural[[:space:]]Real-Time[[:space:]]SLAM/98d936fc-c445-4925-9f16-8d4e0f8fc517_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Co-Salient[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]Uncertainty-Aware[[:space:]]Group[[:space:]]Exchange-Masking/9cd15a5b-7ef1-4ba9-a24c-48575755f531_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Co-Speech[[:space:]]Gesture[[:space:]]Synthesis[[:space:]]by[[:space:]]Reinforcement[[:space:]]Learning[[:space:]]With[[:space:]]Contrastive[[:space:]]Pre-Trained[[:space:]]Rewards/6b1eace7-6844-4cc2-b39b-76192850eed9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Co-Training[[:space:]]2L[[:space:]]Submodels[[:space:]]for[[:space:]]Visual[[:space:]]Recognition/51b2445a-3487-4e2d-824b-ee9b871d1274_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CoMFormer_[[:space:]]Continual[[:space:]]Learning[[:space:]]in[[:space:]]Semantic[[:space:]]and[[:space:]]Panoptic[[:space:]]Segmentation/9493268b-7f35-48c3-8aa1-39be491edc53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CoWs[[:space:]]on[[:space:]]Pasture_[[:space:]]Baselines[[:space:]]and[[:space:]]Benchmarks[[:space:]]for[[:space:]]Language-Driven[[:space:]]Zero-Shot[[:space:]]Object[[:space:]]Navigation/a31a35f2-4957-4777-92dc-31eae2e5deb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Coaching[[:space:]]a[[:space:]]Teachable[[:space:]]Student/67002b28-2cb3-47c6-8f0a-06c8bb56b862_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CodeTalker_[[:space:]]Speech-Driven[[:space:]]3D[[:space:]]Facial[[:space:]]Animation[[:space:]]With[[:space:]]Discrete[[:space:]]Motion[[:space:]]Prior/103a044f-c0a4-40fe-af3f-96d0d65c126f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Collaboration[[:space:]]Helps[[:space:]]Camera[[:space:]]Overtake[[:space:]]LiDAR[[:space:]]in[[:space:]]3D[[:space:]]Detection/0d6cf930-2724-4762-9022-d5c25917e98a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Collaborative[[:space:]]Diffusion[[:space:]]for[[:space:]]Multi-Modal[[:space:]]Face[[:space:]]Generation[[:space:]]and[[:space:]]Editing/3b7ffa39-2244-45b9-9833-f768a1b5b8c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Collaborative[[:space:]]Noisy[[:space:]]Label[[:space:]]Cleaner_[[:space:]]Learning[[:space:]]Scene-Aware[[:space:]]Trailers[[:space:]]for[[:space:]]Multi-Modal[[:space:]]Highlight[[:space:]]Detection[[:space:]]in[[:space:]]Movies/b5b9b0f2-713f-4176-89e7-acec845e8cbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Collaborative[[:space:]]Static[[:space:]]and[[:space:]]Dynamic[[:space:]]Vision-Language[[:space:]]Streams[[:space:]]for[[:space:]]Spatio-Temporal[[:space:]]Video[[:space:]]Grounding/40d01bed-f6c0-4114-9925-d43e564f9389_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Collecting[[:space:]]Cross-Modal[[:space:]]Presence-Absence[[:space:]]Evidence[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Audio-Visual[[:space:]]Event[[:space:]]Perception/022a8b1e-0630-4afb-84c7-1eee9f21c9fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Color[[:space:]]Backdoor_[[:space:]]A[[:space:]]Robust[[:space:]]Poisoning[[:space:]]Attack[[:space:]]in[[:space:]]Color[[:space:]]Space/62d05799-8539-4626-86c2-09c14f52825e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Combining[[:space:]]Implicit-Explicit[[:space:]]View[[:space:]]Correlation[[:space:]]for[[:space:]]Light[[:space:]]Field[[:space:]]Semantic[[:space:]]Segmentation/d513a3d0-08ea-40e8-886e-4ab6e2f71d8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Command-Driven[[:space:]]Articulated[[:space:]]Object[[:space:]]Understanding[[:space:]]and[[:space:]]Manipulation/319777ad-1631-450a-bc10-c16b3c7f113d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Common[[:space:]]Pets[[:space:]]in[[:space:]]3D_[[:space:]]Dynamic[[:space:]]New-View[[:space:]]Synthesis[[:space:]]of[[:space:]]Real-Life[[:space:]]Deformable[[:space:]]Categories/15b2959f-a8de-4f7b-8400-8cdcdef2571d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Compacting[[:space:]]Binary[[:space:]]Neural[[:space:]]Networks[[:space:]]by[[:space:]]Sparse[[:space:]]Kernel[[:space:]]Selection/4763ecc8-2828-438e-906c-559a33868956_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Complementary[[:space:]]Intrinsics[[:space:]]From[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]and[[:space:]]CNNs[[:space:]]for[[:space:]]Outdoor[[:space:]]Scene[[:space:]]Relighting/8b52d905-8ef6-4b0b-8ba0-6cab2f21d3ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Complete[[:space:]]3D[[:space:]]Human[[:space:]]Reconstruction[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Incomplete[[:space:]]Image/2b9adf9d-41b3-4940-bb4b-4ff3bb85f433_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Complete-to-Partial[[:space:]]4D[[:space:]]Distillation[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Sequence[[:space:]]Representation[[:space:]]Learning/f1692118-f65f-4030-b804-1d3ea605bd91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CompletionFormer_[[:space:]]Depth[[:space:]]Completion[[:space:]]With[[:space:]]Convolutions[[:space:]]and[[:space:]]Vision[[:space:]]Transformers/ae6b612c-7ee6-4fbf-933e-6d7640ee9888_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Complexity-Guided[[:space:]]Slimmable[[:space:]]Decoder[[:space:]]for[[:space:]]Efficient[[:space:]]Deep[[:space:]]Video[[:space:]]Compression/0a7d7877-3a01-45ee-bc82-a07bd47abb2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Compositor_[[:space:]]Bottom-Up[[:space:]]Clustering[[:space:]]and[[:space:]]Compositing[[:space:]]for[[:space:]]Robust[[:space:]]Part[[:space:]]and[[:space:]]Object[[:space:]]Segmentation/246f8014-3459-478e-a17d-16df8cd3e5f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Comprehensive[[:space:]]and[[:space:]]Delicate_[[:space:]]An[[:space:]]Efficient[[:space:]]Transformer[[:space:]]for[[:space:]]Image[[:space:]]Restoration/42f6cb6a-ed90-44fd-9d20-014af0585d30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Compressing[[:space:]]Volumetric[[:space:]]Radiance[[:space:]]Fields[[:space:]]to[[:space:]]1[[:space:]]MB/4032bc9b-ef5d-493d-91d0-15efee67648a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Compression-Aware[[:space:]]Video[[:space:]]Super-Resolution/623e6a99-6f61-4ce0-a290-bc0841f9d1fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Computational[[:space:]]Flash[[:space:]]Photography[[:space:]]Through[[:space:]]Intrinsics/3802409a-3baf-4ebf-a7b3-df2ba0495df6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Computationally[[:space:]]Budgeted[[:space:]]Continual[[:space:]]Learning_[[:space:]]What[[:space:]]Does[[:space:]]Matter_/ec287ab4-34a7-45dd-9d26-825f625733ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ConQueR_[[:space:]]Query[[:space:]]Contrast[[:space:]]Voxel-DETR[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/ebf68d50-f74a-48a1-a8f1-f400b77d65a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ConStruct-VL_[[:space:]]Data-Free[[:space:]]Continual[[:space:]]Structured[[:space:]]VL[[:space:]]Concepts[[:space:]]Learning/2266eb39-2a2e-4da9-be15-65d26e6f2079_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ConZIC_[[:space:]]Controllable[[:space:]]Zero-Shot[[:space:]]Image[[:space:]]Captioning[[:space:]]by[[:space:]]Sampling-Based[[:space:]]Polishing/e49825b4-5a2e-4d7b-a02b-37c734f4fde5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Conditional[[:space:]]Generation[[:space:]]of[[:space:]]Audio[[:space:]]From[[:space:]]Video[[:space:]]via[[:space:]]Foley[[:space:]]Analogies/58c6106f-ab5d-4041-8575-dfda2e59d3b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Conditional[[:space:]]Image-to-Video[[:space:]]Generation[[:space:]]With[[:space:]]Latent[[:space:]]Flow[[:space:]]Diffusion[[:space:]]Models/17ac2d28-81aa-4886-863e-9b7d19c8cbd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Conditional[[:space:]]Text[[:space:]]Image[[:space:]]Generation[[:space:]]With[[:space:]]Diffusion[[:space:]]Models/6082ef8f-5f8c-4be4-99df-b3a4b2f5e0f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Confidence-Aware[[:space:]]Personalized[[:space:]]Federated[[:space:]]Learning[[:space:]]via[[:space:]]Variational[[:space:]]Expectation[[:space:]]Maximization/c80b56f8-0c00-4ea6-8b92-e999dba34957_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Conflict-Based[[:space:]]Cross-View[[:space:]]Consistency[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation/bfb4f340-3285-41a0-9877-2e182f201399_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Conjugate[[:space:]]Product[[:space:]]Graphs[[:space:]]for[[:space:]]Globally[[:space:]]Optimal[[:space:]]2D-3D[[:space:]]Shape[[:space:]]Matching/2fc07260-9db8-4542-91b3-78f227d921b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Connecting[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]With[[:space:]]Video[[:space:]]Localized[[:space:]]Narratives/93d2a33f-5ccd-4d44-aecb-15bd8ae7e93c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Connecting[[:space:]]the[[:space:]]Dots_[[:space:]]Floorplan[[:space:]]Reconstruction[[:space:]]Using[[:space:]]Two-Level[[:space:]]Queries/1e844eba-8348-4672-94ed-fe6e3a253087_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Consistent[[:space:]]Direct[[:space:]]Time-of-Flight[[:space:]]Video[[:space:]]Depth[[:space:]]Super-Resolution/81373ef4-5f0f-4714-84bf-da2386f268a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Consistent[[:space:]]View[[:space:]]Synthesis[[:space:]]With[[:space:]]Pose-Guided[[:space:]]Diffusion[[:space:]]Models/50561765-41b3-4ba4-8b44-a810fdd13e43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Consistent-Teacher_[[:space:]]Towards[[:space:]]Reducing[[:space:]]Inconsistent[[:space:]]Pseudo-Targets[[:space:]]in[[:space:]]Semi-Supervised[[:space:]]Object[[:space:]]Detection/bc246ba1-e0ca-4477-8e4e-d27a2b35730a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Constrained[[:space:]]Evolutionary[[:space:]]Diffusion[[:space:]]Filter[[:space:]]for[[:space:]]Monocular[[:space:]]Endoscope[[:space:]]Tracking/b7859401-56d0-4c10-a403-b6ffaa4e47fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Constructing[[:space:]]Deep[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks[[:space:]]From[[:space:]]Artificial[[:space:]]Neural[[:space:]]Networks[[:space:]]With[[:space:]]Knowledge[[:space:]]Distillation/3ad2f5dd-c585-4033-9801-c152ace687ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Content-Aware[[:space:]]Token[[:space:]]Sharing[[:space:]]for[[:space:]]Efficient[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]With[[:space:]]Vision[[:space:]]Transformers/bc02af7d-f5fc-44b2-8ca4-50f83c076d9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Context[[:space:]]De-Confounded[[:space:]]Emotion[[:space:]]Recognition/6680b350-fdb0-4750-b384-4b02bc7fcbf3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Context-Aware[[:space:]]Alignment[[:space:]]and[[:space:]]Mutual[[:space:]]Masking[[:space:]]for[[:space:]]3D-Language[[:space:]]Pre-Training/088bb57f-a7c8-403c-8b7b-eb87a994cc22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Context-Aware[[:space:]]Pretraining[[:space:]]for[[:space:]]Efficient[[:space:]]Blind[[:space:]]Image[[:space:]]Decomposition/5deff675-7cda-4623-a77d-dbea589ddcb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Context-Aware[[:space:]]Relative[[:space:]]Object[[:space:]]Queries[[:space:]]To[[:space:]]Unify[[:space:]]Video[[:space:]]Instance[[:space:]]and[[:space:]]Panoptic[[:space:]]Segmentation/233a48c6-aa04-420c-9fa1-4f6d9e1140af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Context-Based[[:space:]]Trit-Plane[[:space:]]Coding[[:space:]]for[[:space:]]Progressive[[:space:]]Image[[:space:]]Compression/128ebb00-7d32-458a-a0fc-7746be58f0c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Continual[[:space:]]Detection[[:space:]]Transformer[[:space:]]for[[:space:]]Incremental[[:space:]]Object[[:space:]]Detection/80582a1c-a3d4-46b4-856c-2fdad62db871_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Continual[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]With[[:space:]]Automatic[[:space:]]Memory[[:space:]]Sample[[:space:]]Selection/7a8bdb78-9db1-4978-a982-14118e69f63b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Continuous[[:space:]]Intermediate[[:space:]]Token[[:space:]]Learning[[:space:]]With[[:space:]]Implicit[[:space:]]Motion[[:space:]]Manifold[[:space:]]for[[:space:]]Keyframe[[:space:]]Based[[:space:]]Motion[[:space:]]Interpolation/3ea584ac-5747-4cb6-8c79-cbc56ecf7df9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Continuous[[:space:]]Landmark[[:space:]]Detection[[:space:]]With[[:space:]]3D[[:space:]]Queries/0fec2f65-8a37-4e3c-8fcb-a0b525d86c89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Continuous[[:space:]]Pseudo-Label[[:space:]]Rectified[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]With[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations/041902b3-8bf2-42c9-8112-6e4a92bcf54c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Continuous[[:space:]]Sign[[:space:]]Language[[:space:]]Recognition[[:space:]]With[[:space:]]Correlation[[:space:]]Network/130bbfa9-7220-4fe2-83cf-60effd03bfcb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ContraNeRF_[[:space:]]Generalizable[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Synthetic-to-Real[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]via[[:space:]]Contrastive[[:space:]]Learning/0341d3e9-cde4-4914-997a-82fd2e339d88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Contrastive[[:space:]]Grouping[[:space:]]With[[:space:]]Transformer[[:space:]]for[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation/add5a48c-b730-4d71-a865-e61a86668414_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Contrastive[[:space:]]Mean[[:space:]]Teacher[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detectors/cbf1acf1-9432-4b4f-9b13-5cf00945c57f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Contrastive[[:space:]]Semi-Supervised[[:space:]]Learning[[:space:]]for[[:space:]]Underwater[[:space:]]Image[[:space:]]Restoration[[:space:]]via[[:space:]]Reliable[[:space:]]Bank/46bfb6fa-6f6f-45e6-88ec-9f6deeaa3512_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Controllable[[:space:]]Light[[:space:]]Diffusion[[:space:]]for[[:space:]]Portraits/3b924d95-7ac0-4749-a53f-1b0a7cc18b8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Controllable[[:space:]]Mesh[[:space:]]Generation[[:space:]]Through[[:space:]]Sparse[[:space:]]Latent[[:space:]]Point[[:space:]]Diffusion[[:space:]]Models/13a42593-20d1-4049-bc84-57bc016a50ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ConvNeXt[[:space:]]V2_[[:space:]]Co-Designing[[:space:]]and[[:space:]]Scaling[[:space:]]ConvNets[[:space:]]With[[:space:]]Masked[[:space:]]Autoencoders/ef7385b8-2a71-40e2-88b7-3cd1fabf9655_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cooperation[[:space:]]or[[:space:]]Competition_[[:space:]]Avoiding[[:space:]]Player[[:space:]]Domination[[:space:]]for[[:space:]]Multi-Target[[:space:]]Robustness[[:space:]]via[[:space:]]Adaptive[[:space:]]Budgets/3728daa4-7fcb-4874-bde0-3dc32ce3bb0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CoralStyleCLIP_[[:space:]]Co-Optimized[[:space:]]Region[[:space:]]and[[:space:]]Layer[[:space:]]Selection[[:space:]]for[[:space:]]Image[[:space:]]Editing/ebbb9d7d-8954-44cd-abb0-c38c1ad2e55b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Coreset[[:space:]]Sampling[[:space:]]From[[:space:]]Open-Set[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Self-Supervised[[:space:]]Learning/9644741f-1429-44da-8b6c-2fd5fc0e6150_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Correlational[[:space:]]Image[[:space:]]Modeling[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Visual[[:space:]]Pre-Training/ac6472b7-347f-49f7-9205-64334ffd2f97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Correspondence[[:space:]]Transformers[[:space:]]With[[:space:]]Asymmetric[[:space:]]Feature[[:space:]]Learning[[:space:]]and[[:space:]]Matching[[:space:]]Flow[[:space:]]Super-Resolution/b0b72061-7b7b-42b0-bf76-e754faa219f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CrOC_[[:space:]]Cross-View[[:space:]]Online[[:space:]]Clustering[[:space:]]for[[:space:]]Dense[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/58373ecb-4c2d-4172-ba3e-6c2d19a18048_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Critical[[:space:]]Learning[[:space:]]Periods[[:space:]]for[[:space:]]Multisensory[[:space:]]Integration[[:space:]]in[[:space:]]Deep[[:space:]]Networks/a959b2ed-3852-47fe-ba37-5eeecafa5889_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-Domain[[:space:]]3D[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation[[:space:]]With[[:space:]]Dual[[:space:]]Modalities/6f492107-d777-4d16-a9d4-579ae33cab3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-Domain[[:space:]]Image[[:space:]]Captioning[[:space:]]With[[:space:]]Discriminative[[:space:]]Finetuning/2033e2de-1df3-49e3-b81c-419019511a16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-GAN[[:space:]]Auditing_[[:space:]]Unsupervised[[:space:]]Identification[[:space:]]of[[:space:]]Attribute[[:space:]]Level[[:space:]]Similarities[[:space:]]and[[:space:]]Differences[[:space:]]Between[[:space:]]Pretrained[[:space:]]Generative[[:space:]]Models/6880bb25-7b17-4783-bff9-492b87df1440_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-Guided[[:space:]]Optimization[[:space:]]of[[:space:]]Radiance[[:space:]]Fields[[:space:]]With[[:space:]]Multi-View[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]for[[:space:]]High-Resolution[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/c6dff55c-469e-4773-99a8-7e82467ee2c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-Image-Attention[[:space:]]for[[:space:]]Conditional[[:space:]]Embeddings[[:space:]]in[[:space:]]Deep[[:space:]]Metric[[:space:]]Learning/8ac04ea3-84df-4f15-bf11-cde8d966a0f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-Modal[[:space:]]Implicit[[:space:]]Relation[[:space:]]Reasoning[[:space:]]and[[:space:]]Aligning[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Person[[:space:]]Retrieval/b887fe8e-cbd1-4a56-bf1d-af392cf4d0f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Crossing[[:space:]]the[[:space:]]Gap_[[:space:]]Domain[[:space:]]Generalization[[:space:]]for[[:space:]]Image[[:space:]]Captioning/76a9a70a-6b63-4556-b3f5-55c2aa598674_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Crowd3D_[[:space:]]Towards[[:space:]]Hundreds[[:space:]]of[[:space:]]People[[:space:]]Reconstruction[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Image/4c85e22c-9b87-4239-87aa-ce176821b21e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CrowdCLIP_[[:space:]]Unsupervised[[:space:]]Crowd[[:space:]]Counting[[:space:]]via[[:space:]]Vision-Language[[:space:]]Model/271daa69-d383-4ac7-b68a-6416af6bdc75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Curricular[[:space:]]Contrastive[[:space:]]Regularization[[:space:]]for[[:space:]]Physics-Aware[[:space:]]Single[[:space:]]Image[[:space:]]Dehazing/8b57d2bd-aa23-4588-9af2-f32474b95abc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Curricular[[:space:]]Object[[:space:]]Manipulation[[:space:]]in[[:space:]]LiDAR-Based[[:space:]]Object[[:space:]]Detection/cfdff6cf-02e9-45b5-967a-c9640594f9bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Curvature-Balanced[[:space:]]Feature[[:space:]]Manifold[[:space:]]Learning[[:space:]]for[[:space:]]Long-Tailed[[:space:]]Classification/da111f4a-962c-40ec-ae1e-042fc250a546_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cut[[:space:]]and[[:space:]]Learn[[:space:]]for[[:space:]]Unsupervised[[:space:]]Object[[:space:]]Detection[[:space:]]and[[:space:]]Instance[[:space:]]Segmentation/5ff5b645-e93f-456c-ad19-661d72bfb105_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CutMIB_[[:space:]]Boosting[[:space:]]Light[[:space:]]Field[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Multi-View[[:space:]]Image[[:space:]]Blending/3c6b3cee-77fa-45b7-94c7-df11db43130d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/D2Former_[[:space:]]Jointly[[:space:]]Learning[[:space:]]Hierarchical[[:space:]]Detectors[[:space:]]and[[:space:]]Contextual[[:space:]]Descriptors[[:space:]]via[[:space:]]Agent-Based[[:space:]]Transformers/7bf16cf2-cfec-4a8f-a3d9-b3c13df7561c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DA[[:space:]]Wand_[[:space:]]Distortion-Aware[[:space:]]Selection[[:space:]]Using[[:space:]]Neural[[:space:]]Mesh[[:space:]]Parameterization/f203d616-c279-4bc3-b0f8-b6ba742868fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DA-DETR_[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Detection[[:space:]]Transformer[[:space:]]With[[:space:]]Information[[:space:]]Fusion/d78078b6-9a8a-4c0e-b894-fdc422cfb8e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DAA_[[:space:]]A[[:space:]]Delta[[:space:]]Age[[:space:]]AdaIN[[:space:]]Operation[[:space:]]for[[:space:]]Age[[:space:]]Estimation[[:space:]]via[[:space:]]Binary[[:space:]]Code[[:space:]]Transformer/150b22e1-97b6-46ea-80a6-1f0e0ed15a9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DANI-Net_[[:space:]]Uncalibrated[[:space:]]Photometric[[:space:]]Stereo[[:space:]]by[[:space:]]Differentiable[[:space:]]Shadow[[:space:]]Handling,[[:space:]]Anisotropic[[:space:]]Reflectance[[:space:]]Modeling,[[:space:]]and[[:space:]]Neural[[:space:]]Inverse[[:space:]]Rendering/26a7ddc6-0bdc-4acb-8c0d-516aa1ca629f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DARE-GRAM_[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]Regression[[:space:]]by[[:space:]]Aligning[[:space:]]Inverse[[:space:]]Gram[[:space:]]Matrices/3ba531b8-bd84-4138-b3f6-0acf8645a0f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DART_[[:space:]]Diversify-Aggregate-Repeat[[:space:]]Training[[:space:]]Improves[[:space:]]Generalization[[:space:]]of[[:space:]]Neural[[:space:]]Networks/5f0541a3-9bae-4171-bd6b-429d470b4242_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DATE_[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Product[[:space:]]Seeker[[:space:]]for[[:space:]]E-Commerce/4260a7e1-e0c3-464e-b18c-1e8c4b5d4842_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DATID-3D_[[:space:]]Diversity-Preserved[[:space:]]Domain[[:space:]]Adaptation[[:space:]]Using[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]for[[:space:]]3D[[:space:]]Generative[[:space:]]Model/eea4efc1-2cae-4eb2-8560-34d2cc206b43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DBARF_[[:space:]]Deep[[:space:]]Bundle-Adjusting[[:space:]]Generalizable[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/b509cd25-8763-4a88-913e-7d6aa1685484_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DC2_[[:space:]]Dual-Camera[[:space:]]Defocus[[:space:]]Control[[:space:]]by[[:space:]]Learning[[:space:]]To[[:space:]]Refocus/803db2ee-9105-41e9-ac3d-4ef3a8d1a92d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DCFace_[[:space:]]Synthetic[[:space:]]Face[[:space:]]Generation[[:space:]]With[[:space:]]Dual[[:space:]]Condition[[:space:]]Diffusion[[:space:]]Model/634036d9-0152-443e-a7e7-5e9ea744a97d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DETR[[:space:]]With[[:space:]]Additional[[:space:]]Global[[:space:]]Aggregation[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Object[[:space:]]Detection/bc76d200-205f-4657-a037-6e3e1db05d65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DETRs[[:space:]]With[[:space:]]Hybrid[[:space:]]Matching/9fd67dc8-83e4-4b0d-854f-0fcc80e24bc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DF-Platter_[[:space:]]Multi-Face[[:space:]]Heterogeneous[[:space:]]Deepfake[[:space:]]Dataset/5905eee6-9181-423e-8f35-223ea7de8af5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DIFu_[[:space:]]Depth-Guided[[:space:]]Implicit[[:space:]]Function[[:space:]]for[[:space:]]Clothed[[:space:]]Human[[:space:]]Reconstruction/7ca11d3f-298e-4eb8-a0ae-9305195c1207_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DINER_[[:space:]]Depth-Aware[[:space:]]Image-Based[[:space:]]NEural[[:space:]]Radiance[[:space:]]Fields/851ff98c-62c9-4318-87d4-e846a42111dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DINER_[[:space:]]Disorder-Invariant[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representation/5bcd9fe7-f0b1-4450-a351-677c4303924c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DINN360_[[:space:]]Deformable[[:space:]]Invertible[[:space:]]Neural[[:space:]]Network[[:space:]]for[[:space:]]Latitude-Aware[[:space:]]360deg[[:space:]]Image[[:space:]]Rescaling/34f5e1c1-a8db-4ab0-a089-dc2abac96492_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DIP_[[:space:]]Dual[[:space:]]Incongruity[[:space:]]Perceiving[[:space:]]Network[[:space:]]for[[:space:]]Sarcasm[[:space:]]Detection/862bbd31-c149-413f-a4fd-03bfd085c50c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DISC_[[:space:]]Learning[[:space:]]From[[:space:]]Noisy[[:space:]]Labels[[:space:]]via[[:space:]]Dynamic[[:space:]]Instance-Specific[[:space:]]Selection[[:space:]]and[[:space:]]Correction/fa79af3d-63e7-4bb4-9d74-14aba33a31ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DKM_[[:space:]]Dense[[:space:]]Kernelized[[:space:]]Feature[[:space:]]Matching[[:space:]]for[[:space:]]Geometry[[:space:]]Estimation/aaa31bf5-6d6b-40db-86c1-b325dce34358_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DKT_[[:space:]]Diverse[[:space:]]Knowledge[[:space:]]Transfer[[:space:]]Transformer[[:space:]]for[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning/e0db879e-dad0-41ce-a566-41f8e9fb6787_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DLBD_[[:space:]]A[[:space:]]Self-Supervised[[:space:]]Direct-Learned[[:space:]]Binary[[:space:]]Descriptor/608a76c0-6aee-4cd2-ad67-c2b3ca6dfb0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DNF_[[:space:]]Decouple[[:space:]]and[[:space:]]Feedback[[:space:]]Network[[:space:]]for[[:space:]]Seeing[[:space:]]in[[:space:]]the[[:space:]]Dark/b896c5c2-6081-498a-8c72-28b256305573_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DNeRV_[[:space:]]Modeling[[:space:]]Inherent[[:space:]]Dynamics[[:space:]]via[[:space:]]Difference[[:space:]]Neural[[:space:]]Representation[[:space:]]for[[:space:]]Videos/01ebd8d9-b069-4d63-aa91-6fb8806e7bcf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DP-NeRF_[[:space:]]Deblurred[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field[[:space:]]With[[:space:]]Physical[[:space:]]Scene[[:space:]]Priors/6b27db0e-c895-467a-9d6e-b280d4ab531a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DPE_[[:space:]]Disentanglement[[:space:]]of[[:space:]]Pose[[:space:]]and[[:space:]]Expression[[:space:]]for[[:space:]]General[[:space:]]Video[[:space:]]Portrait[[:space:]]Editing/11110830-ad1a-43bd-ab77-c99c43a5f67a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DPF_[[:space:]]Learning[[:space:]]Dense[[:space:]]Prediction[[:space:]]Fields[[:space:]]With[[:space:]]Weak[[:space:]]Supervision/5fcedabf-7dcd-4447-bf40-e802b4dbc715_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DR2_[[:space:]]Diffusion-Based[[:space:]]Robust[[:space:]]Degradation[[:space:]]Remover[[:space:]]for[[:space:]]Blind[[:space:]]Face[[:space:]]Restoration/6fd505d8-8c76-4f80-922e-6bb8f6c01a9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DSFNet_[[:space:]]Dual[[:space:]]Space[[:space:]]Fusion[[:space:]]Network[[:space:]]for[[:space:]]Occlusion-Robust[[:space:]]3D[[:space:]]Dense[[:space:]]Face[[:space:]]Alignment/577078f6-30d2-449c-8058-64e1e39edc55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DSVT_[[:space:]]Dynamic[[:space:]]Sparse[[:space:]]Voxel[[:space:]]Transformer[[:space:]]With[[:space:]]Rotated[[:space:]]Sets/6579b368-6cdd-40bc-93c3-e08e14f16ea2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DaFKD_[[:space:]]Domain-Aware[[:space:]]Federated[[:space:]]Knowledge[[:space:]]Distillation/7f9ee5fd-a1b5-4587-b184-9b018b3c3aea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DartBlur_[[:space:]]Privacy[[:space:]]Preservation[[:space:]]With[[:space:]]Detection[[:space:]]Artifact[[:space:]]Suppression/15722f2d-e6d0-4b1d-a59c-f93bc399cb5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Data-Driven[[:space:]]Feature[[:space:]]Tracking[[:space:]]for[[:space:]]Event[[:space:]]Cameras/01cd21f8-5e5e-485a-83c3-59a5f05e5e29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Data-Efficient[[:space:]]Large[[:space:]]Scale[[:space:]]Place[[:space:]]Recognition[[:space:]]With[[:space:]]Graded[[:space:]]Similarity[[:space:]]Supervision/26da24b2-8628-48d9-b6a0-e966fd91011d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Data-Free[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]via[[:space:]]Feature[[:space:]]Exchange[[:space:]]and[[:space:]]Activation[[:space:]]Region[[:space:]]Constraint/070ac438-1887-4d45-ab53-c3f194c60c4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Data-Free[[:space:]]Sketch-Based[[:space:]]Image[[:space:]]Retrieval/307dfda0-ee47-43aa-b226-c3c52ef3241f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeAR_[[:space:]]Debiasing[[:space:]]Vision-Language[[:space:]]Models[[:space:]]With[[:space:]]Additive[[:space:]]Residuals/1c1934a9-a2f5-4086-b803-f61ed1d65d52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeCo_[[:space:]]Decomposition[[:space:]]and[[:space:]]Reconstruction[[:space:]]for[[:space:]]Compositional[[:space:]]Temporal[[:space:]]Grounding[[:space:]]via[[:space:]]Coarse-To-Fine[[:space:]]Contrastive[[:space:]]Ranking/a08065dd-1f03-4d1f-b72e-a65dd5ae3677_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeFeeNet_[[:space:]]Consecutive[[:space:]]3D[[:space:]]Human[[:space:]]Motion[[:space:]]Prediction[[:space:]]With[[:space:]]Deviation[[:space:]]Feedback/d4550b77-8ce1-4b71-bfc4-cb9895e53200_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeGPR_[[:space:]]Deep[[:space:]]Guided[[:space:]]Posterior[[:space:]]Regularization[[:space:]]for[[:space:]]Multi-Class[[:space:]]Cell[[:space:]]Detection[[:space:]]and[[:space:]]Counting/f75e307d-3706-4130-b4cf-94f1c7a42ec0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeSTSeg_[[:space:]]Segmentation[[:space:]]Guided[[:space:]]Denoising[[:space:]]Student-Teacher[[:space:]]for[[:space:]]Anomaly[[:space:]]Detection/27a06ba2-4c3b-4336-bf66-e6a67c1ea3c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dealing[[:space:]]With[[:space:]]Cross-Task[[:space:]]Class[[:space:]]Discrimination[[:space:]]in[[:space:]]Online[[:space:]]Continual[[:space:]]Learning/dbfd6e43-a4a9-447f-8c86-380a961e27b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decentralized[[:space:]]Learning[[:space:]]With[[:space:]]Multi-Headed[[:space:]]Distillation/efa1f2dd-0a8b-4150-b5b3-031d70469dcb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decompose[[:space:]]More[[:space:]]and[[:space:]]Aggregate[[:space:]]Better_[[:space:]]Two[[:space:]]Closer[[:space:]]Looks[[:space:]]at[[:space:]]Frequency[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Human[[:space:]]Motion[[:space:]]Prediction/f6564538-6c19-4f0e-9d17-104719e0e46b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decompose,[[:space:]]Adjust,[[:space:]]Compose_[[:space:]]Effective[[:space:]]Normalization[[:space:]]by[[:space:]]Playing[[:space:]]With[[:space:]]Frequency[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/a24fce25-9487-4aca-b78f-56a7e413c4df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decomposed[[:space:]]Cross-Modal[[:space:]]Distillation[[:space:]]for[[:space:]]RGB-Based[[:space:]]Temporal[[:space:]]Action[[:space:]]Detection/274297e7-f082-419d-aeff-028480fdf2dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decomposed[[:space:]]Soft[[:space:]]Prompt[[:space:]]Guided[[:space:]]Fusion[[:space:]]Enhancing[[:space:]]for[[:space:]]Compositional[[:space:]]Zero-Shot[[:space:]]Learning/a6a340bb-ce69-4a90-a726-80004f2b35c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decoupled[[:space:]]Multimodal[[:space:]]Distilling[[:space:]]for[[:space:]]Emotion[[:space:]]Recognition/3aad95da-bab3-40f9-b208-54d60ba66017_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decoupled[[:space:]]Semantic[[:space:]]Prototypes[[:space:]]Enable[[:space:]]Learning[[:space:]]From[[:space:]]Diverse[[:space:]]Annotation[[:space:]]Types[[:space:]]for[[:space:]]Semi-Weakly[[:space:]]Segmentation[[:space:]]in[[:space:]]Expert-Driven[[:space:]]Domains/fc3418d8-b837-4b2e-9bc8-cb99b8c5609c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decoupling[[:space:]]Human[[:space:]]and[[:space:]]Camera[[:space:]]Motion[[:space:]]From[[:space:]]Videos[[:space:]]in[[:space:]]the[[:space:]]Wild/817b5151-a71b-41b9-ad35-4087d380a835_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decoupling[[:space:]]Learning[[:space:]]and[[:space:]]Remembering_[[:space:]]A[[:space:]]Bilevel[[:space:]]Memory[[:space:]]Framework[[:space:]]With[[:space:]]Knowledge[[:space:]]Projection[[:space:]]for[[:space:]]Task-Incremental[[:space:]]Learning/e0168fa3-9e08-43fd-9e8a-4e9b92af305e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decoupling[[:space:]]MaxLogit[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection/3a0f010e-75ec-4fc8-88de-6da4ecf1951a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decoupling-and-Aggregating[[:space:]]for[[:space:]]Image[[:space:]]Exposure[[:space:]]Correction/f95ce198-d05a-48ae-be50-84b6536a0f4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Arbitrary-Scale[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Scale-Equivariance[[:space:]]Pursuit/d97fa7e4-e7ba-462b-b0eb-25ed504adc96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Curvilinear[[:space:]]Editing_[[:space:]]Commutative[[:space:]]and[[:space:]]Nonlinear[[:space:]]Image[[:space:]]Manipulation[[:space:]]for[[:space:]]Pretrained[[:space:]]Deep[[:space:]]Generative[[:space:]]Model/b0ed9f72-44b0-4409-8c53-069e9e7f8de9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Depth[[:space:]]Estimation[[:space:]]From[[:space:]]Thermal[[:space:]]Image/478e5e84-c41c-449d-815d-c9f040a73244_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Deterministic[[:space:]]Uncertainty_[[:space:]]A[[:space:]]New[[:space:]]Simple[[:space:]]Baseline/ac7b2945-312f-45fe-98fd-bb79f95093e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Discriminative[[:space:]]Spatial[[:space:]]and[[:space:]]Temporal[[:space:]]Network[[:space:]]for[[:space:]]Efficient[[:space:]]Video[[:space:]]Deblurring/0b0f8173-99de-4106-9e0e-040544846fdd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Dive[[:space:]]Into[[:space:]]Gradients_[[:space:]]Better[[:space:]]Optimization[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]Gradient-Corrected[[:space:]]IoU[[:space:]]Supervision/06d63865-9340-4b6b-bdca-b03432a172fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Factorized[[:space:]]Metric[[:space:]]Learning/1af216b7-b2f3-4949-afe0-ef70f1b5b33e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Fair[[:space:]]Clustering[[:space:]]via[[:space:]]Maximizing[[:space:]]and[[:space:]]Minimizing[[:space:]]Mutual[[:space:]]Information_[[:space:]]Theory,[[:space:]]Algorithm[[:space:]]and[[:space:]]Metric/004dfc3b-0dd8-479b-9032-3151769a8e63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Frequency[[:space:]]Filtering[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/dbd89e67-2bea-4e2e-b53d-33b90e680ce5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Graph[[:space:]]Reprogramming/c69b2d53-4706-4d86-9c60-77f91eb87405_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Graph-Based[[:space:]]Spatial[[:space:]]Consistency[[:space:]]for[[:space:]]Robust[[:space:]]Non-Rigid[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/92aee350-5b8e-4f12-9ae0-342503963d95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Hashing[[:space:]]With[[:space:]]Minimal-Distance-Separated[[:space:]]Hash[[:space:]]Centers/fc26e49d-0079-48af-b0f6-3bde53f32114_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Incomplete[[:space:]]Multi-View[[:space:]]Clustering[[:space:]]With[[:space:]]Cross-View[[:space:]]Partial[[:space:]]Sample[[:space:]]and[[:space:]]Prototype[[:space:]]Alignment/4818e8ea-a256-4b49-ae8f-864ce56b7d32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Learning[[:space:]]of[[:space:]]Partial[[:space:]]Graph[[:space:]]Matching[[:space:]]via[[:space:]]Differentiable[[:space:]]Top-K/ba57ad56-8faf-4d37-9f7c-e4bf5b0b01d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Polarization[[:space:]]Reconstruction[[:space:]]With[[:space:]]PDAVIS[[:space:]]Events/19ed6ec8-72a4-49e1-805a-377dcf7ad6a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Random[[:space:]]Projector_[[:space:]]Accelerated[[:space:]]Deep[[:space:]]Image[[:space:]]Prior/686a76c4-7c6d-400b-a2cb-96c70384d52f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Semi-Supervised[[:space:]]Metric[[:space:]]Learning[[:space:]]With[[:space:]]Mixed[[:space:]]Label[[:space:]]Propagation/5e0cf7c0-45a5-4376-88cc-1d82c5368f6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Stereo[[:space:]]Video[[:space:]]Inpainting/c6815317-038d-4ca6-aa15-4abbb7cdedfe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeepLSD_[[:space:]]Line[[:space:]]Segment[[:space:]]Detection[[:space:]]and[[:space:]]Refinement[[:space:]]With[[:space:]]Deep[[:space:]]Image[[:space:]]Gradients/9eb91c73-f034-4152-a9f2-af586118382c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeepMAD_[[:space:]]Mathematical[[:space:]]Architecture[[:space:]]Design[[:space:]]for[[:space:]]Deep[[:space:]]Convolutional[[:space:]]Neural[[:space:]]Network/f16305f5-fd58-4e17-9951-c56843a2d399_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeepMapping2_[[:space:]]Self-Supervised[[:space:]]Large-Scale[[:space:]]LiDAR[[:space:]]Map[[:space:]]Optimization/2e2f0be3-238a-4b04-a1e1-354cf346588f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeepSolo_[[:space:]]Let[[:space:]]Transformer[[:space:]]Decoder[[:space:]]With[[:space:]]Explicit[[:space:]]Points[[:space:]]Solo[[:space:]]for[[:space:]]Text[[:space:]]Spotting/f439817a-c7fd-41ab-bd68-295cc992144d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeepVecFont-v2_[[:space:]]Exploiting[[:space:]]Transformers[[:space:]]To[[:space:]]Synthesize[[:space:]]Vector[[:space:]]Fonts[[:space:]]With[[:space:]]Higher[[:space:]]Quality/a3831625-8663-4e9d-98be-64804c498244_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Defending[[:space:]]Against[[:space:]]Patch-Based[[:space:]]Backdoor[[:space:]]Attacks[[:space:]]on[[:space:]]Self-Supervised[[:space:]]Learning/d96c2eda-3033-429c-a236-c3b02e39c7cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Defining[[:space:]]and[[:space:]]Quantifying[[:space:]]the[[:space:]]Emergence[[:space:]]of[[:space:]]Sparse[[:space:]]Concepts[[:space:]]in[[:space:]]DNNs/43fdd746-207d-4592-ba30-0729a4d78baf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deformable[[:space:]]Mesh[[:space:]]Transformer[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/160ccad0-305c-4b71-903e-a00df6535a0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DegAE_[[:space:]]A[[:space:]]New[[:space:]]Pretraining[[:space:]]Paradigm[[:space:]]for[[:space:]]Low-Level[[:space:]]Vision/8346633f-835f-44d6-a110-ee4cea339a46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DejaVu_[[:space:]]Conditional[[:space:]]Regenerative[[:space:]]Learning[[:space:]]To[[:space:]]Enhance[[:space:]]Dense[[:space:]]Prediction/fa352b23-d566-4cee-8df4-088354eea4df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Delivering[[:space:]]Arbitrary-Modal[[:space:]]Semantic[[:space:]]Segmentation/58971b7c-39ae-4e57-b7d8-5e9fa0664841_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Delving[[:space:]]Into[[:space:]]Discrete[[:space:]]Normalizing[[:space:]]Flows[[:space:]]on[[:space:]]SO(3)[[:space:]]Manifold[[:space:]]for[[:space:]]Probabilistic[[:space:]]Rotation[[:space:]]Modeling/a32e70f7-9038-4a73-aee4-2fafd7deed86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Delving[[:space:]]Into[[:space:]]Shape-Aware[[:space:]]Zero-Shot[[:space:]]Semantic[[:space:]]Segmentation/88133c03-747f-4c6e-8dcb-cbdefd84d34d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Delving[[:space:]]StyleGAN[[:space:]]Inversion[[:space:]]for[[:space:]]Image[[:space:]]Editing_[[:space:]]A[[:space:]]Foundation[[:space:]]Latent[[:space:]]Space[[:space:]]Viewpoint/20f1e296-a868-4de6-ae0c-80f72f9b4822_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Demystifying[[:space:]]Causal[[:space:]]Features[[:space:]]on[[:space:]]Adversarial[[:space:]]Examples[[:space:]]and[[:space:]]Causal[[:space:]]Inoculation[[:space:]]for[[:space:]]Robust[[:space:]]Network[[:space:]]by[[:space:]]Adversarial[[:space:]]Instrumental[[:space:]]Variable[[:space:]]Regression/d1097a42-cb19-4cc0-867b-d032b15c0e6b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dense[[:space:]]Distinct[[:space:]]Query[[:space:]]for[[:space:]]End-to-End[[:space:]]Object[[:space:]]Detection/581a336c-6e10-4dad-a028-e3321933637a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dense[[:space:]]Network[[:space:]]Expansion[[:space:]]for[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning/f54be291-1d51-4460-965b-c1559f94196b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dense-Localizing[[:space:]]Audio-Visual[[:space:]]Events[[:space:]]in[[:space:]]Untrimmed[[:space:]]Videos_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Benchmark[[:space:]]and[[:space:]]Baseline/102bbb59-4b01-41a2-8db4-885f68f6caf1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Density-Insensitive[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaption[[:space:]]on[[:space:]]3D[[:space:]]Object[[:space:]]Detection/ae908082-ec25-46a5-b4ce-f50d072a46ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DepGraph_[[:space:]]Towards[[:space:]]Any[[:space:]]Structural[[:space:]]Pruning/ccf5b5af-0347-4f1f-80b6-475758ef8334_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Depth[[:space:]]Estimation[[:space:]]From[[:space:]]Camera[[:space:]]Image[[:space:]]and[[:space:]]mmWave[[:space:]]Radar[[:space:]]Point[[:space:]]Cloud/85ca7564-cfc7-442d-813f-7e0ee5c44630_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Depth[[:space:]]Estimation[[:space:]]From[[:space:]]Indoor[[:space:]]Panoramas[[:space:]]With[[:space:]]Neural[[:space:]]Scene[[:space:]]Representation/89f7fa53-2683-43b8-8687-2ff686bba895_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DetCLIPv2_[[:space:]]Scalable[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detection[[:space:]]Pre-Training[[:space:]]via[[:space:]]Word-Region[[:space:]]Alignment/8830e603-5a18-414d-8bde-0562da353152_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Detecting[[:space:]]Backdoors[[:space:]]During[[:space:]]the[[:space:]]Inference[[:space:]]Stage[[:space:]]Based[[:space:]]on[[:space:]]Corruption[[:space:]]Robustness[[:space:]]Consistency/ebf2eaf8-c0f9-4106-baa7-5025ab05a6b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Detecting[[:space:]]Backdoors[[:space:]]in[[:space:]]Pre-Trained[[:space:]]Encoders/61462042-5f91-42f2-b77f-7ce707c00ba8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Detecting[[:space:]]Everything[[:space:]]in[[:space:]]the[[:space:]]Open[[:space:]]World_[[:space:]]Towards[[:space:]]Universal[[:space:]]Object[[:space:]]Detection/50c70bfa-8326-4264-b3f7-f8a12b15b504_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Detecting[[:space:]]Human-Object[[:space:]]Contact[[:space:]]in[[:space:]]Images/2135aaff-208f-4e99-bf72-f068c150352c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Detecting[[:space:]]and[[:space:]]Grounding[[:space:]]Multi-Modal[[:space:]]Media[[:space:]]Manipulation/b3d06fe4-293a-40de-9382-da6c735f0ce5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Detection[[:space:]]Hub_[[:space:]]Unifying[[:space:]]Object[[:space:]]Detection[[:space:]]Datasets[[:space:]]via[[:space:]]Query[[:space:]]Adaptation[[:space:]]on[[:space:]]Language[[:space:]]Embedding/11542ae7-f105-4e55-b7e4-cb9fb031f9d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Detection[[:space:]]of[[:space:]]Out-of-Distribution[[:space:]]Samples[[:space:]]Using[[:space:]]Binary[[:space:]]Neuron[[:space:]]Activation[[:space:]]Patterns/3c29d982-f962-4a7a-86e1-ee96a745f021_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Devil[[:space:]]Is[[:space:]]in[[:space:]]the[[:space:]]Queries_[[:space:]]Advancing[[:space:]]Mask[[:space:]]Transformers[[:space:]]for[[:space:]]Real-World[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation[[:space:]]and[[:space:]]Out-of-Distribution[[:space:]]Localization/5443a5cf-cf65-4882-bae8-0855677bf031_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Devil's[[:space:]]on[[:space:]]the[[:space:]]Edges_[[:space:]]Selective[[:space:]]Quad[[:space:]]Attention[[:space:]]for[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/3f8c6ff2-828d-4edb-be1e-82c65ae48a28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DexArt_[[:space:]]Benchmarking[[:space:]]Generalizable[[:space:]]Dexterous[[:space:]]Manipulation[[:space:]]With[[:space:]]Articulated[[:space:]]Objects/5c78a0ad-90f0-4f81-8ea1-af9b7b52d3f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiGA_[[:space:]]Distil[[:space:]]To[[:space:]]Generalize[[:space:]]and[[:space:]]Then[[:space:]]Adapt[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Semantic[[:space:]]Segmentation/662e842f-d32d-4d85-9018-433e30f2085f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiGeo_[[:space:]]Discriminative[[:space:]]Geometry-Aware[[:space:]]Learning[[:space:]]for[[:space:]]Generalized[[:space:]]Few-Shot[[:space:]]Object[[:space:]]Detection/7c36d462-198b-4ccb-93e3-551958d1821c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffCollage_[[:space:]]Parallel[[:space:]]Generation[[:space:]]of[[:space:]]Large[[:space:]]Content[[:space:]]With[[:space:]]Diffusion[[:space:]]Models/4d97d691-9146-4974-9194-49a946ffdb56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffPose_[[:space:]]Toward[[:space:]]More[[:space:]]Reliable[[:space:]]3D[[:space:]]Pose[[:space:]]Estimation/1102242c-7632-448d-a0ab-f38d27fdb87d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffRF_[[:space:]]Rendering-Guided[[:space:]]3D[[:space:]]Radiance[[:space:]]Field[[:space:]]Diffusion/d6680066-9500-4c71-925e-f3a4aaa88c96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffSwap_[[:space:]]High-Fidelity[[:space:]]and[[:space:]]Controllable[[:space:]]Face[[:space:]]Swapping[[:space:]]via[[:space:]]3D-Aware[[:space:]]Masked[[:space:]]Diffusion/4da170b2-c661-42ae-900f-57e410fd3937_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffTalk_[[:space:]]Crafting[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Generalized[[:space:]]Audio-Driven[[:space:]]Portraits[[:space:]]Animation/088ded45-4910-4733-860e-0eeb0cc8ad2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Differentiable[[:space:]]Architecture[[:space:]]Search[[:space:]]With[[:space:]]Random[[:space:]]Features/ef268532-2d3b-43ac-bd68-8c3ead76aa8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Differentiable[[:space:]]Shadow[[:space:]]Mapping[[:space:]]for[[:space:]]Efficient[[:space:]]Inverse[[:space:]]Graphics/dba25f21-3293-4979-9821-c762d5844bd7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Difficulty-Based[[:space:]]Sampling[[:space:]]for[[:space:]]Debiased[[:space:]]Contrastive[[:space:]]Representation[[:space:]]Learning/a10d59a8-6ebd-46d6-aac0-01b6920f4651_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffusioNeRF_[[:space:]]Regularizing[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]With[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Models/6d96060c-8b9f-4584-adf5-3e50b226be17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion[[:space:]]Art[[:space:]]or[[:space:]]Digital[[:space:]]Forgery_[[:space:]]Investigating[[:space:]]Data[[:space:]]Replication[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/2e6a2418-d945-4394-9711-fc13fa4564dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion[[:space:]]Probabilistic[[:space:]]Model[[:space:]]Made[[:space:]]Slim/bf8212f7-833a-4c9d-9c61-a33889819e89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion[[:space:]]Video[[:space:]]Autoencoders_[[:space:]]Toward[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Face[[:space:]]Video[[:space:]]Editing[[:space:]]via[[:space:]]Disentangled[[:space:]]Video[[:space:]]Encoding/9c72d0e6-e653-459a-87e1-dce2e1570589_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion-Based[[:space:]]Generation,[[:space:]]Optimization,[[:space:]]and[[:space:]]Planning[[:space:]]in[[:space:]]3D[[:space:]]Scenes/2c0ff292-024d-4fa0-9487-b60809586e2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion-Based[[:space:]]Signed[[:space:]]Distance[[:space:]]Fields[[:space:]]for[[:space:]]3D[[:space:]]Shape[[:space:]]Generation/959cbb8a-2f33-407b-925b-14af99929d47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion-SDF_[[:space:]]Text-To-Shape[[:space:]]via[[:space:]]Voxelized[[:space:]]Diffusion/a08ad0f7-f625-4886-ac27-deeab3818eeb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffusionRig_[[:space:]]Learning[[:space:]]Personalized[[:space:]]Priors[[:space:]]for[[:space:]]Facial[[:space:]]Appearance[[:space:]]Editing/d7adefe3-24be-45f5-8e46-305721be8441_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dimensionality-Varying[[:space:]]Diffusion[[:space:]]Process/6d385449-d42c-4da7-84b2-17c8988b13c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dionysus_[[:space:]]Recovering[[:space:]]Scene[[:space:]]Structures[[:space:]]by[[:space:]]Dividing[[:space:]]Into[[:space:]]Semantic[[:space:]]Pieces/13315114-2bd9-46a3-8750-70a0a8a52fb9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Directional[[:space:]]Connectivity-Based[[:space:]]Segmentation[[:space:]]of[[:space:]]Medical[[:space:]]Images/b9c53750-6c7d-4628-8a5d-c4d2ad764a83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DisCo-CLIP_[[:space:]]A[[:space:]]Distributed[[:space:]]Contrastive[[:space:]]Loss[[:space:]]for[[:space:]]Memory[[:space:]]Efficient[[:space:]]CLIP[[:space:]]Training/e80e2080-33a9-4c63-8bae-91e19a54d7b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DisCoScene_[[:space:]]Spatially[[:space:]]Disentangled[[:space:]]Generative[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Controllable[[:space:]]3D-Aware[[:space:]]Scene[[:space:]]Synthesis/bdaa5712-8b26-43d5-9ca1-465a14b26701_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DisWOT_[[:space:]]Student[[:space:]]Architecture[[:space:]]Search[[:space:]]for[[:space:]]Distillation[[:space:]]WithOut[[:space:]]Training/cc4b4c46-ca23-4c2e-9650-90533edd035d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Discovering[[:space:]]the[[:space:]]Real[[:space:]]Association_[[:space:]]Multimodal[[:space:]]Causal[[:space:]]Reasoning[[:space:]]in[[:space:]]Video[[:space:]]Question[[:space:]]Answering/f46331e4-a9e1-4fa6-843a-36c1e6734399_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Discrete[[:space:]]Point-Wise[[:space:]]Attack[[:space:]]Is[[:space:]]Not[[:space:]]Enough_[[:space:]]Generalized[[:space:]]Manifold[[:space:]]Adversarial[[:space:]]Attack[[:space:]]for[[:space:]]Face[[:space:]]Recognition/dc2b0e92-b412-418d-ab5e-fff033c9b7f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Discriminating[[:space:]]Known[[:space:]]From[[:space:]]Unknown[[:space:]]Objects[[:space:]]via[[:space:]]Structure-Enhanced[[:space:]]Recurrent[[:space:]]Variational[[:space:]]AutoEncoder/7f53095f-c4ed-4448-89ff-3db316832dfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Discriminative[[:space:]]Co-Saliency[[:space:]]and[[:space:]]Background[[:space:]]Mining[[:space:]]Transformer[[:space:]]for[[:space:]]Co-Salient[[:space:]]Object[[:space:]]Detection/1ba243a9-552c-4f70-ae66-3201d18a4d66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Discriminator-Cooperated[[:space:]]Feature[[:space:]]Map[[:space:]]Distillation[[:space:]]for[[:space:]]GAN[[:space:]]Compression/e9448640-3029-45d7-94ad-42638a17674c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Disentangled[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Unsupervised[[:space:]]Neural[[:space:]]Quantization/06ed76f2-80a7-4ca7-b9b2-5c10f4249fdb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Disentangling[[:space:]]Orthogonal[[:space:]]Planes[[:space:]]for[[:space:]]Indoor[[:space:]]Panoramic[[:space:]]Room[[:space:]]Layout[[:space:]]Estimation[[:space:]]With[[:space:]]Cross-Scale[[:space:]]Distortion[[:space:]]Awareness/13fac3e7-4155-4a84-8381-fe68c09139a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Disentangling[[:space:]]Writer[[:space:]]and[[:space:]]Character[[:space:]]Styles[[:space:]]for[[:space:]]Handwriting[[:space:]]Generation/39ec325e-e585-4646-a631-108d72fa1121_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DistilPose_[[:space:]]Tokenized[[:space:]]Pose[[:space:]]Regression[[:space:]]With[[:space:]]Heatmap[[:space:]]Distillation/3fcfeb96-13c3-4c6a-a012-9770732b455e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distilling[[:space:]]Cross-Temporal[[:space:]]Contexts[[:space:]]for[[:space:]]Continuous[[:space:]]Sign[[:space:]]Language[[:space:]]Recognition/67f2d609-710a-4aac-8b1f-1eb63e00c70c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distilling[[:space:]]Focal[[:space:]]Knowledge[[:space:]]From[[:space:]]Imperfect[[:space:]]Expert[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/26d504d2-cd24-4e73-99c4-214e67e817bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distilling[[:space:]]Neural[[:space:]]Fields[[:space:]]for[[:space:]]Real-Time[[:space:]]Articulated[[:space:]]Shape[[:space:]]Reconstruction/b3e477ee-24cc-4767-adae-b678b4d54a08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distilling[[:space:]]Self-Supervised[[:space:]]Vision[[:space:]]Transformers[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Few-Shot[[:space:]]Classification[[:space:]]&[[:space:]]Segmentation/e6a6b9de-f70e-4a6b-bd84-70683a76add4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distilling[[:space:]]Vision-Language[[:space:]]Pre-Training[[:space:]]To[[:space:]]Collaborate[[:space:]]With[[:space:]]Weakly-Supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization/01d592ea-9922-4de7-a537-4d96044a494d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DistractFlow_[[:space:]]Improving[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation[[:space:]]via[[:space:]]Realistic[[:space:]]Distractions[[:space:]]and[[:space:]]Pseudo-Labeling/71dcd135-f623-44d3-88eb-6335db05d14d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distribution[[:space:]]Shift[[:space:]]Inversion[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Prediction/2bfae0da-d459-4405-8157-11be349e6b0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DivClust_[[:space:]]Controlling[[:space:]]Diversity[[:space:]]in[[:space:]]Deep[[:space:]]Clustering/66ea5bda-5daa-4792-a2cb-68e8650cf8e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diverse[[:space:]]3D[[:space:]]Hand[[:space:]]Gesture[[:space:]]Prediction[[:space:]]From[[:space:]]Body[[:space:]]Dynamics[[:space:]]by[[:space:]]Bilateral[[:space:]]Hand[[:space:]]Disentanglement/d0994cbf-d27b-45f6-a350-1ef9d2d99ed2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diverse[[:space:]]Embedding[[:space:]]Expansion[[:space:]]Network[[:space:]]and[[:space:]]Low-Light[[:space:]]Cross-Modality[[:space:]]Benchmark[[:space:]]for[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification/fe5b9645-eedb-4dae-90a3-37c29c7fb846_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diversity-Aware[[:space:]]Meta[[:space:]]Visual[[:space:]]Prompting/ab39c456-2bd7-4413-8ddf-e238b3250192_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diversity-Measurable[[:space:]]Anomaly[[:space:]]Detection/43dbf769-f361-451d-b986-1893419fc738_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Divide[[:space:]]and[[:space:]]Adapt_[[:space:]]Active[[:space:]]Domain[[:space:]]Adaptation[[:space:]]via[[:space:]]Customized[[:space:]]Learning/86a989bf-4e7d-4c1f-bfb3-37887b13a20b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Divide[[:space:]]and[[:space:]]Conquer_[[:space:]]Answering[[:space:]]Questions[[:space:]]With[[:space:]]Object[[:space:]]Factorization[[:space:]]and[[:space:]]Compositional[[:space:]]Reasoning/9361ff3c-8aee-49ab-8084-0a759e885614_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DoNet_[[:space:]]Deep[[:space:]]De-Overlapping[[:space:]]Network[[:space:]]for[[:space:]]Cytology[[:space:]]Instance[[:space:]]Segmentation/6ce3e51d-a82a-47f1-b18c-a1b3922a9d7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Document[[:space:]]Image[[:space:]]Shadow[[:space:]]Removal[[:space:]]Guided[[:space:]]by[[:space:]]Color-Aware[[:space:]]Background/488a81e8-9aa8-4e58-aac0-74e15ffd5e40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Domain[[:space:]]Expansion[[:space:]]of[[:space:]]Image[[:space:]]Generators/d894b4ee-e2ae-449a-85b9-1c0622cd67b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Domain[[:space:]]Generalized[[:space:]]Stereo[[:space:]]Matching[[:space:]]via[[:space:]]Hierarchical[[:space:]]Visual[[:space:]]Transformation/9850d655-e7cd-45f5-81bc-ed4d96ec4425_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Don't[[:space:]]Lie[[:space:]]to[[:space:]]Me![[:space:]]Robust[[:space:]]and[[:space:]]Efficient[[:space:]]Explainability[[:space:]]With[[:space:]]Verified[[:space:]]Perturbation[[:space:]]Analysis/b65bbb01-b52c-43c2-b278-f068bfd77b29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Doubly[[:space:]]Right[[:space:]]Object[[:space:]]Recognition_[[:space:]]A[[:space:]]Why[[:space:]]Prompt[[:space:]]for[[:space:]]Visual[[:space:]]Rationales/3e3819d8-6dd5-452f-ae03-010df00b34bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DrapeNet_[[:space:]]Garment[[:space:]]Generation[[:space:]]and[[:space:]]Self-Supervised[[:space:]]Draping/bc8ad2b1-5bdc-4117-96e0-b09e43677d93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dream3D_[[:space:]]Zero-Shot[[:space:]]Text-to-3D[[:space:]]Synthesis[[:space:]]Using[[:space:]]3D[[:space:]]Shape[[:space:]]Prior[[:space:]]and[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/5d7485a5-1935-43d5-ba5e-14f7126fbb06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DreamBooth_[[:space:]]Fine[[:space:]]Tuning[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Subject-Driven[[:space:]]Generation/88ef8e93-ceab-4a73-b104-1d6bfb6b0d91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DropKey[[:space:]]for[[:space:]]Vision[[:space:]]Transformer/539b4cee-3610-4b35-8990-86b8994c51f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DropMAE_[[:space:]]Masked[[:space:]]Autoencoders[[:space:]]With[[:space:]]Spatial-Attention[[:space:]]Dropout[[:space:]]for[[:space:]]Tracking[[:space:]]Tasks/63a61ae0-513f-436f-959c-a8f318e90bde_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dual[[:space:]]Alignment[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Video-Text[[:space:]]Retrieval/8b3047df-13e0-4f30-a21d-3acd07314d19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dual-Bridging[[:space:]]With[[:space:]]Adversarial[[:space:]]Noise[[:space:]]Generation[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]rPPG[[:space:]]Estimation/d848ec4a-af95-428f-9e60-c6e80b3b4e6b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dual-Path[[:space:]]Adaptation[[:space:]]From[[:space:]]Image[[:space:]]to[[:space:]]Video[[:space:]]Transformers/58cf23aa-7524-4b95-9291-20d982a61012_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DualRefine_[[:space:]]Self-Supervised[[:space:]]Depth[[:space:]]and[[:space:]]Pose[[:space:]]Estimation[[:space:]]Through[[:space:]]Iterative[[:space:]]Epipolar[[:space:]]Sampling[[:space:]]and[[:space:]]Refinement[[:space:]]Toward[[:space:]]Equilibrium/afa982aa-a8cf-442d-886c-70412ae83518_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DualRel_[[:space:]]Semi-Supervised[[:space:]]Mitochondria[[:space:]]Segmentation[[:space:]]From[[:space:]]a[[:space:]]Prototype[[:space:]]Perspective/ffadc065-1edc-42b2-a626-096c78b5c896_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DualVector_[[:space:]]Unsupervised[[:space:]]Vector[[:space:]]Font[[:space:]]Synthesis[[:space:]]With[[:space:]]Dual-Part[[:space:]]Representation/64299b13-1bde-4ab7-949f-d03cb4326255_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DyLiN_[[:space:]]Making[[:space:]]Light[[:space:]]Field[[:space:]]Networks[[:space:]]Dynamic/52436ab0-0f6f-44ce-bee8-32ea0aeba3bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DyNCA_[[:space:]]Real-Time[[:space:]]Dynamic[[:space:]]Texture[[:space:]]Synthesis[[:space:]]Using[[:space:]]Neural[[:space:]]Cellular[[:space:]]Automata/786b41ed-e991-424e-9c74-c623b61a6162_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DynIBaR_[[:space:]]Neural[[:space:]]Dynamic[[:space:]]Image-Based[[:space:]]Rendering/b5d1340a-3f6f-462f-9df3-03c14583f216_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DynaFed_[[:space:]]Tackling[[:space:]]Client[[:space:]]Data[[:space:]]Heterogeneity[[:space:]]With[[:space:]]Global[[:space:]]Dynamics/ba56f7cd-e332-49dd-9060-cbf5c82b8082_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DynaMask_[[:space:]]Dynamic[[:space:]]Mask[[:space:]]Selection[[:space:]]for[[:space:]]Instance[[:space:]]Segmentation/6de3482f-3d80-4aaa-a48e-484855bf3aa9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Aggregated[[:space:]]Network[[:space:]]for[[:space:]]Gait[[:space:]]Recognition/3a3ce788-5b5c-498f-8336-10d8bcc8bede_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Coarse-To-Fine[[:space:]]Learning[[:space:]]for[[:space:]]Oriented[[:space:]]Tiny[[:space:]]Object[[:space:]]Detection/e9804a4f-6933-4148-ae33-ea1a17587f5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Conceptional[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/40efeb24-3451-4def-8441-e92b604d3421_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Focus-Aware[[:space:]]Positional[[:space:]]Queries[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/0da0c2f9-7cdb-4694-a9e6-89d455ec6130_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Generative[[:space:]]Targeted[[:space:]]Attacks[[:space:]]With[[:space:]]Pattern[[:space:]]Injection/03048e96-e108-4f0d-8eee-8c14ea87df56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Graph[[:space:]]Enhanced[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Chest[[:space:]]X-Ray[[:space:]]Report[[:space:]]Generation/7bbdab16-5f4c-46f9-abdb-480a53791847_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Graph[[:space:]]Learning[[:space:]]With[[:space:]]Content-Guided[[:space:]]Spatial-Frequency[[:space:]]Relation[[:space:]]Reasoning[[:space:]]for[[:space:]]Deepfake[[:space:]]Detection/0db433fe-2bb5-4d32-8dad-d6501909824f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Inference[[:space:]]With[[:space:]]Grounding[[:space:]]Based[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Models/605f806c-dc72-4654-be1c-1d10020323b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Neural[[:space:]]Network[[:space:]]for[[:space:]]Multi-Task[[:space:]]Learning[[:space:]]Searching[[:space:]]Across[[:space:]]Diverse[[:space:]]Network[[:space:]]Topologies/08f871aa-0255-4d39-aee9-94ac7465f9f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DynamicDet_[[:space:]]A[[:space:]]Unified[[:space:]]Dynamic[[:space:]]Architecture[[:space:]]for[[:space:]]Object[[:space:]]Detection/a2c8ee28-6898-4bb7-b856-836c2ee0fe1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DynamicStereo_[[:space:]]Consistent[[:space:]]Dynamic[[:space:]]Depth[[:space:]]From[[:space:]]Stereo[[:space:]]Videos/f0255026-ccd7-4c09-9102-f9355b92857a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamically[[:space:]]Instance-Guided[[:space:]]Adaptation_[[:space:]]A[[:space:]]Backward-Free[[:space:]]Approach[[:space:]]for[[:space:]]Test-Time[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Semantic[[:space:]]Segmentation/8fab8469-0a5d-4f13-bb47-6165f5c91b04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/E2PN_[[:space:]]Efficient[[:space:]]SE(3)-Equivariant[[:space:]]Point[[:space:]]Network/8cfd95ec-52b9-4e92-8a17-84d6d71e6730_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EC2_[[:space:]]Emergent[[:space:]]Communication[[:space:]]for[[:space:]]Embodied[[:space:]]Control/1439d777-9b35-45c1-8de3-484abb88a29b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ECON_[[:space:]]Explicit[[:space:]]Clothed[[:space:]]Humans[[:space:]]Optimized[[:space:]]via[[:space:]]Normal[[:space:]]Integration/3cdd03c7-8c50-4311-9381-5af32afe63c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EDA_[[:space:]]Explicit[[:space:]]Text-Decoupling[[:space:]]and[[:space:]]Dense[[:space:]]Alignment[[:space:]]for[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding/3140b022-de45-49be-b6c5-fb6d9f5aae53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EDGE_[[:space:]]Editable[[:space:]]Dance[[:space:]]Generation[[:space:]]From[[:space:]]Music/97e29d71-2a36-4f04-a26f-23105d30e787_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EDICT_[[:space:]]Exact[[:space:]]Diffusion[[:space:]]Inversion[[:space:]]via[[:space:]]Coupled[[:space:]]Transformations/5ac781e1-0cfa-4238-8cf2-ceb66647f068_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EFEM_[[:space:]]Equivariant[[:space:]]Neural[[:space:]]Field[[:space:]]Expectation[[:space:]]Maximization[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Segmentation[[:space:]]Without[[:space:]]Scene[[:space:]]Supervision/3d7e3a10-9770-4e87-8f5a-9ec61b4fbebc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EMT-NAS_Transferring[[:space:]]Architectural[[:space:]]Knowledge[[:space:]]Between[[:space:]]Tasks[[:space:]]From[[:space:]]Different[[:space:]]Datasets/f54bf63e-2702-4fdf-815c-53bb8912c6fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ERM-KTP_[[:space:]]Knowledge-Level[[:space:]]Machine[[:space:]]Unlearning[[:space:]]via[[:space:]]Knowledge[[:space:]]Transfer/ab5287df-b899-4bb9-ae6b-5d2f41d1b6fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ERNIE-ViLG[[:space:]]2.0_[[:space:]]Improving[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Model[[:space:]]With[[:space:]]Knowledge-Enhanced[[:space:]]Mixture-of-Denoising-Experts/4f139481-82b6-4525-8567-38b399fee431_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ESLAM_[[:space:]]Efficient[[:space:]]Dense[[:space:]]SLAM[[:space:]]System[[:space:]]Based[[:space:]]on[[:space:]]Hybrid[[:space:]]Representation[[:space:]]of[[:space:]]Signed[[:space:]]Distance[[:space:]]Fields/d4f0d86a-9d2c-4b36-b0b6-8aa8c847e904_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EVAL_[[:space:]]Explainable[[:space:]]Video[[:space:]]Anomaly[[:space:]]Localization/396d953c-501f-406c-be61-8e06127066d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EVA_[[:space:]]Exploring[[:space:]]the[[:space:]]Limits[[:space:]]of[[:space:]]Masked[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning[[:space:]]at[[:space:]]Scale/c4046a48-182c-4cee-86a4-65a7035e3e1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EXCALIBUR_[[:space:]]Encouraging[[:space:]]and[[:space:]]Evaluating[[:space:]]Embodied[[:space:]]Exploration/c84853fe-e5d3-4737-8cec-5e8fb595ae84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EXIF[[:space:]]As[[:space:]]Language_[[:space:]]Learning[[:space:]]Cross-Modal[[:space:]]Associations[[:space:]]Between[[:space:]]Images[[:space:]]and[[:space:]]Camera[[:space:]]Metadata/dfc40f9a-79af-47cc-ac0b-335308cd3774_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EcoTTA_[[:space:]]Memory-Efficient[[:space:]]Continual[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]via[[:space:]]Self-Distilled[[:space:]]Regularization/62c54281-c979-438c-bdac-9fc9ff81d220_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Edge-Aware[[:space:]]Regional[[:space:]]Message[[:space:]]Passing[[:space:]]Controller[[:space:]]for[[:space:]]Image[[:space:]]Forgery[[:space:]]Localization/c5c52750-0529-4615-bd1f-1262f32b5c87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Edges[[:space:]]to[[:space:]]Shapes[[:space:]]to[[:space:]]Concepts_[[:space:]]Adversarial[[:space:]]Augmentation[[:space:]]for[[:space:]]Robust[[:space:]]Vision/d0d50664-1fef-4fb1-8679-7306724f6c4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EditableNeRF_[[:space:]]Editing[[:space:]]Topologically[[:space:]]Varying[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]by[[:space:]]Key[[:space:]]Points/891f5efd-20c5-4d2e-bc9c-a2da74e10197_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Effective[[:space:]]Ambiguity[[:space:]]Attack[[:space:]]Against[[:space:]]Passport-Based[[:space:]]DNN[[:space:]]Intellectual[[:space:]]Property[[:space:]]Protection[[:space:]]Schemes[[:space:]]Through[[:space:]]Fully[[:space:]]Connected[[:space:]]Layer[[:space:]]Substitution/e387fd7f-d2cf-442f-899c-5257e75d504c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Frequency[[:space:]]Domain-Based[[:space:]]Transformers[[:space:]]for[[:space:]]High-Quality[[:space:]]Image[[:space:]]Deblurring/c856a1da-867e-42ea-84a9-c14d1b12f290_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Hierarchical[[:space:]]Entropy[[:space:]]Model[[:space:]]for[[:space:]]Learned[[:space:]]Point[[:space:]]Cloud[[:space:]]Compression/cb22330c-c1cc-443f-b5db-982a0288da0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Loss[[:space:]]Function[[:space:]]by[[:space:]]Minimizing[[:space:]]the[[:space:]]Detrimental[[:space:]]Effect[[:space:]]of[[:space:]]Floating-Point[[:space:]]Errors[[:space:]]on[[:space:]]Gradient-Based[[:space:]]Attacks/e43d26b1-1945-4f4e-bd26-fe5c4d18f332_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Map[[:space:]]Sparsification[[:space:]]Based[[:space:]]on[[:space:]]2D[[:space:]]and[[:space:]]3D[[:space:]]Discretized[[:space:]]Grids/4f0f1ce5-d45d-4b52-a587-2da996f0a8b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Mask[[:space:]]Correction[[:space:]]for[[:space:]]Click-Based[[:space:]]Interactive[[:space:]]Image[[:space:]]Segmentation/0a4cc1dd-64c3-44d2-8dfa-4f17be83df54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Movie[[:space:]]Scene[[:space:]]Detection[[:space:]]Using[[:space:]]State-Space[[:space:]]Transformers/b626675b-8b28-4fca-aa4b-1742033f43ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Multimodal[[:space:]]Fusion[[:space:]]via[[:space:]]Interactive[[:space:]]Prompting/9f845069-d99b-4851-8fff-d097c545215b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]On-Device[[:space:]]Training[[:space:]]via[[:space:]]Gradient[[:space:]]Filtering/3b76baa7-39c6-473d-83e4-4a2631e20341_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]RGB-T[[:space:]]Tracking[[:space:]]via[[:space:]]Cross-Modality[[:space:]]Distillation/cf5e82ff-3a69-46ad-ac66-64191f806cdf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Robust[[:space:]]Principal[[:space:]]Component[[:space:]]Analysis[[:space:]]via[[:space:]]Block[[:space:]]Krylov[[:space:]]Iteration[[:space:]]and[[:space:]]CUR[[:space:]]Decomposition/07ab4d04-cd27-49d8-a510-e7e724ff7dc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Scale-Invariant[[:space:]]Generator[[:space:]]With[[:space:]]Column-Row[[:space:]]Entangled[[:space:]]Pixel[[:space:]]Synthesis/e82bfa76-118c-45c9-a6c3-50265e9de1a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Second-Order[[:space:]]Plane[[:space:]]Adjustment/0ca972fe-43f5-4fa8-bf07-17ebcd5d268a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]by[[:space:]]Altering[[:space:]]Resolutions[[:space:]]for[[:space:]]Compressed[[:space:]]Videos/7cf71203-1b67-4ff6-a186-04d01a4ea3fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Verification[[:space:]]of[[:space:]]Neural[[:space:]]Networks[[:space:]]Against[[:space:]]LVM-Based[[:space:]]Specifications/f017e278-be6a-4fc6-a7c8-64ae1c47c99f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]View[[:space:]]Synthesis[[:space:]]and[[:space:]]3D-Based[[:space:]]Multi-Frame[[:space:]]Denoising[[:space:]]With[[:space:]]Multiplane[[:space:]]Feature[[:space:]]Representations/b6c4c560-21c7-4cb1-aa8c-86a25b81c853_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]and[[:space:]]Explicit[[:space:]]Modelling[[:space:]]of[[:space:]]Image[[:space:]]Hierarchies[[:space:]]for[[:space:]]Image[[:space:]]Restoration/61b7e2ab-0381-45d6-b5b0-549e3c623e77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EfficientSCI_[[:space:]]Densely[[:space:]]Connected[[:space:]]Network[[:space:]]With[[:space:]]Space-Time[[:space:]]Factorization[[:space:]]for[[:space:]]Large-Scale[[:space:]]Video[[:space:]]Snapshot[[:space:]]Compressive[[:space:]]Imaging/ba9925d6-1c9f-468a-bc9d-dc4b6c4a5ca6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EfficientViT_[[:space:]]Memory[[:space:]]Efficient[[:space:]]Vision[[:space:]]Transformer[[:space:]]With[[:space:]]Cascaded[[:space:]]Group[[:space:]]Attention/351f9dba-430a-462e-86ab-2879d7f983dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ego-Body[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Ego-Head[[:space:]]Pose[[:space:]]Estimation/efdf65a4-1b22-494a-ba23-4617e8cb1fe5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Egocentric[[:space:]]Audio-Visual[[:space:]]Object[[:space:]]Localization/ba22c8dd-0e4e-4696-a837-4a7f7a179a45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Egocentric[[:space:]]Auditory[[:space:]]Attention[[:space:]]Localization[[:space:]]in[[:space:]]Conversations/89830ac3-80bf-409b-844e-adf1daa6e387_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Egocentric[[:space:]]Video[[:space:]]Task[[:space:]]Translation/4f7e3293-877b-4c9b-8bc7-d95d33f4f800_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Elastic[[:space:]]Aggregation[[:space:]]for[[:space:]]Federated[[:space:]]Optimization/aa6cbfa9-0a85-4695-927f-135bdd87fa31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/End-to-End[[:space:]]3D[[:space:]]Dense[[:space:]]Captioning[[:space:]]With[[:space:]]Vote2Cap-DETR/b3b0c666-2a98-4d91-907b-4e3009df208c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/End-to-End[[:space:]]Vectorized[[:space:]]HD-Map[[:space:]]Construction[[:space:]]With[[:space:]]Piecewise[[:space:]]Bezier[[:space:]]Curve/798a815c-ad48-46e8-9f91-80fae212fa9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/End-to-End[[:space:]]Video[[:space:]]Matting[[:space:]]With[[:space:]]Trimap[[:space:]]Propagation/5671e405-73cb-4243-be20-91144c26868b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Endpoints[[:space:]]Weight[[:space:]]Fusion[[:space:]]for[[:space:]]Class[[:space:]]Incremental[[:space:]]Semantic[[:space:]]Segmentation/ff7be89f-f21f-4256-a5c9-a98275c1a5e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Energy-Efficient[[:space:]]Adaptive[[:space:]]3D[[:space:]]Sensing/84eaed71-d9fd-4068-b870-f6dc2370efb9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhanced[[:space:]]Multimodal[[:space:]]Representation[[:space:]]Learning[[:space:]]With[[:space:]]Cross-Modal[[:space:]]KD/550b12f3-e75d-4ee0-8736-cf2526b0aff9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhanced[[:space:]]Stable[[:space:]]View[[:space:]]Synthesis/0a0f1a7f-c68c-47e4-9fa7-95cd72aa0e19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhanced[[:space:]]Training[[:space:]]of[[:space:]]Query-Based[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Selective[[:space:]]Query[[:space:]]Recollection/514d250b-ea10-4630-9256-9e649ac9271f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhancing[[:space:]]Deformable[[:space:]]Local[[:space:]]Features[[:space:]]by[[:space:]]Jointly[[:space:]]Learning[[:space:]]To[[:space:]]Detect[[:space:]]and[[:space:]]Describe[[:space:]]Keypoints/16ef7ac5-e127-4979-a2b4-14363013dd09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhancing[[:space:]]Multiple[[:space:]]Reliability[[:space:]]Measures[[:space:]]via[[:space:]]Nuisance-Extended[[:space:]]Information[[:space:]]Bottleneck/13a10611-b5a4-413b-b3b8-4101508c9783_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhancing[[:space:]]the[[:space:]]Self-Universality[[:space:]]for[[:space:]]Transferable[[:space:]]Targeted[[:space:]]Attacks/4a2df1aa-152c-427f-bd2e-28ba0bd1cd5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enlarging[[:space:]]Instance-Specific[[:space:]]and[[:space:]]Class-Specific[[:space:]]Information[[:space:]]for[[:space:]]Open-Set[[:space:]]Action[[:space:]]Recognition/a1eb877e-c60e-4804-81d3-1cd98ec238d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ensemble-Based[[:space:]]Blackbox[[:space:]]Attacks[[:space:]]on[[:space:]]Dense[[:space:]]Prediction/1f7b1ddd-65c3-47e9-8398-1b94d1c9d1e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EqMotion_[[:space:]]Equivariant[[:space:]]Multi-Agent[[:space:]]Motion[[:space:]]Prediction[[:space:]]With[[:space:]]Invariant[[:space:]]Interaction[[:space:]]Reasoning/f8c65856-4ee9-403e-8f18-e949bafd8ca6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Equiangular[[:space:]]Basis[[:space:]]Vectors/d0d8b892-8a3a-423d-9839-0349feda6f33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Equivalent[[:space:]]Transformation[[:space:]]and[[:space:]]Dual[[:space:]]Stream[[:space:]]Network[[:space:]]Construction[[:space:]]for[[:space:]]Mobile[[:space:]]Image[[:space:]]Super-Resolution/dacaf17d-de1f-48bb-830c-55e8de8a8f85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EvShutter_[[:space:]]Transforming[[:space:]]Events[[:space:]]for[[:space:]]Unconstrained[[:space:]]Rolling[[:space:]]Shutter[[:space:]]Correction/65a66f35-8745-4d12-9133-82f4227811f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Evading[[:space:]]DeepFake[[:space:]]Detectors[[:space:]]via[[:space:]]Adversarial[[:space:]]Statistical[[:space:]]Consistency/d203d2a8-41cc-441c-9ac9-d1d80926ead3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Evading[[:space:]]Forensic[[:space:]]Classifiers[[:space:]]With[[:space:]]Attribute-Conditioned[[:space:]]Adversarial[[:space:]]Faces/b3fbbdae-4502-4dcf-aaab-268d1d919ba6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Event-Based[[:space:]]Blurry[[:space:]]Frame[[:space:]]Interpolation[[:space:]]Under[[:space:]]Blind[[:space:]]Exposure/d11f5758-897e-4530-8e74-aaf511b87d4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Event-Based[[:space:]]Frame[[:space:]]Interpolation[[:space:]]With[[:space:]]Ad-Hoc[[:space:]]Deblurring/c5cb82dd-f2db-492a-ab2f-38c109373b96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Event-Based[[:space:]]Shape[[:space:]]From[[:space:]]Polarization/37883796-c146-454b-8057-2e9dd9252133_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Event-Based[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation[[:space:]]With[[:space:]]Cross-Modal[[:space:]]Asymmetric[[:space:]]Bidirectional[[:space:]]Motion[[:space:]]Fields/bce1dadf-d971-4ac1-8dff-f94d918ce810_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Event-Guided[[:space:]]Person[[:space:]]Re-Identification[[:space:]]via[[:space:]]Sparse-Dense[[:space:]]Complementary[[:space:]]Learning/17184c2c-46ca-4056-8eec-804edf8d23e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EventNeRF_[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Colour[[:space:]]Event[[:space:]]Camera/702b1723-abc0-4483-a2e2-952b47e654c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Evolved[[:space:]]Part[[:space:]]Masking[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Learning/a6d04888-a63e-4dda-94fb-a5eb458c603b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exact-NeRF_[[:space:]]An[[:space:]]Exploration[[:space:]]of[[:space:]]a[[:space:]]Precise[[:space:]]Volumetric[[:space:]]Parameterization[[:space:]]for[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/a6bb31c8-923b-4125-b04d-e8785c6598e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Executing[[:space:]]Your[[:space:]]Commands[[:space:]]via[[:space:]]Motion[[:space:]]Diffusion[[:space:]]in[[:space:]]Latent[[:space:]]Space/3e84fb6e-c61b-4849-a51b-908dd51f5773_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exemplar-FreeSOLO_[[:space:]]Enhancing[[:space:]]Unsupervised[[:space:]]Instance[[:space:]]Segmentation[[:space:]]With[[:space:]]Exemplars/d91644c2-4a89-46af-ba4b-1b12deab727f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Explaining[[:space:]]Image[[:space:]]Classifiers[[:space:]]With[[:space:]]Multiscale[[:space:]]Directional[[:space:]]Image[[:space:]]Representation/1ffb1caf-8819-40ed-8b03-0529e978075d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Explicit[[:space:]]Boundary[[:space:]]Guided[[:space:]]Semi-Push-Pull[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Supervised[[:space:]]Anomaly[[:space:]]Detection/07a2ee18-ee0c-48c8-8450-5f06b3a40d2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Explicit[[:space:]]Visual[[:space:]]Prompting[[:space:]]for[[:space:]]Low-Level[[:space:]]Structure[[:space:]]Segmentations/a47720cf-de15-4ad0-b570-23431739c716_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploiting[[:space:]]Completeness[[:space:]]and[[:space:]]Uncertainty[[:space:]]of[[:space:]]Pseudo[[:space:]]Labels[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/36f1d3f3-a623-4891-81fc-3166e480e9bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploiting[[:space:]]Unlabelled[[:space:]]Photos[[:space:]]for[[:space:]]Stronger[[:space:]]Fine-Grained[[:space:]]SBIR/9a1a3798-66eb-4d67-b011-0667fd5c9826_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Data[[:space:]]Geometry[[:space:]]for[[:space:]]Continual[[:space:]]Learning/89f350b6-db3f-4563-b40c-3570173cbf28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Discontinuity[[:space:]]for[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation/888013f1-329c-451e-994c-4a3ccff797e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Incompatible[[:space:]]Knowledge[[:space:]]Transfer[[:space:]]in[[:space:]]Few-Shot[[:space:]]Image[[:space:]]Generation/6fbda8cb-f270-42c7-afcf-872eae14235e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Intra-Class[[:space:]]Variation[[:space:]]Factors[[:space:]]With[[:space:]]Learnable[[:space:]]Cluster[[:space:]]Prompts[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Image[[:space:]]Synthesis/f1bd89fa-73e5-41b5-b953-4b0c8dab1acb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Motion[[:space:]]Ambiguity[[:space:]]and[[:space:]]Alignment[[:space:]]for[[:space:]]High-Quality[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation/af6698fd-2606-41c2-aa9f-8bb5f655c72d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Structured[[:space:]]Semantic[[:space:]]Prior[[:space:]]for[[:space:]]Multi[[:space:]]Label[[:space:]]Recognition[[:space:]]With[[:space:]]Incomplete[[:space:]]Labels/dff9c5cd-c678-47f1-96db-064bc282ec75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]and[[:space:]]Exploiting[[:space:]]Uncertainty[[:space:]]for[[:space:]]Incomplete[[:space:]]Multi-View[[:space:]]Classification/1d5de150-5d9a-4084-8e1e-b024dd291af4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]and[[:space:]]Utilizing[[:space:]]Pattern[[:space:]]Imbalance/f11fac6b-6df4-4247-a137-3f4d0c4989fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]the[[:space:]]Effect[[:space:]]of[[:space:]]Primitives[[:space:]]for[[:space:]]Compositional[[:space:]]Generalization[[:space:]]in[[:space:]]Vision-and-Language/5997d4bb-9f63-4a2e-a39f-069171b28f09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]the[[:space:]]Relationship[[:space:]]Between[[:space:]]Architectural[[:space:]]Design[[:space:]]and[[:space:]]Adversarially[[:space:]]Robust[[:space:]]Generalization/c9480bfe-0710-40be-aec8-787d89451289_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Extracting[[:space:]]Class[[:space:]]Activation[[:space:]]Maps[[:space:]]From[[:space:]]Non-Discriminative[[:space:]]Features[[:space:]]As[[:space:]]Well/7d008d19-1e3e-496d-bbcf-9d25c0e00f30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Extracting[[:space:]]Motion[[:space:]]and[[:space:]]Appearance[[:space:]]via[[:space:]]Inter-Frame[[:space:]]Attention[[:space:]]for[[:space:]]Efficient[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation/253767af-68d4-4c74-9940-b662c62c6d63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/F2-NeRF_[[:space:]]Fast[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field[[:space:]]Training[[:space:]]With[[:space:]]Free[[:space:]]Camera[[:space:]]Trajectories/9d39c962-f2b8-4739-b36c-ab68dff18df6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FAC_[[:space:]]3D[[:space:]]Representation[[:space:]]Learning[[:space:]]via[[:space:]]Foreground[[:space:]]Aware[[:space:]]Feature[[:space:]]Contrast/c37af97d-8221-4c8e-b1d2-7f1386ddb6ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FAME-ViL_[[:space:]]Multi-Tasking[[:space:]]Vision-Language[[:space:]]Model[[:space:]]for[[:space:]]Heterogeneous[[:space:]]Fashion[[:space:]]Tasks/faff3e9d-4bae-4bfd-8685-d8e76a7b038e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FCC_[[:space:]]Feature[[:space:]]Clusters[[:space:]]Compression[[:space:]]for[[:space:]]Long-Tailed[[:space:]]Visual[[:space:]]Recognition/92f8206f-6584-4ab1-950b-52d5ab7ae16a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FEND_[[:space:]]A[[:space:]]Future[[:space:]]Enhanced[[:space:]]Distribution-Aware[[:space:]]Contrastive[[:space:]]Learning[[:space:]]Framework[[:space:]]for[[:space:]]Long-Tail[[:space:]]Trajectory[[:space:]]Prediction/ae6846ba-4334-444b-b33e-0059555944a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FFCV_[[:space:]]Accelerating[[:space:]]Training[[:space:]]by[[:space:]]Removing[[:space:]]Data[[:space:]]Bottlenecks/ead6e9a7-8dba-4893-b439-95242f8113d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FFF_[[:space:]]Fragment-Guided[[:space:]]Flexible[[:space:]]Fitting[[:space:]]for[[:space:]]Building[[:space:]]Complete[[:space:]]Protein[[:space:]]Structures/4eeef39d-e975-4654-9648-1c62abe7fa74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FFHQ-UV_[[:space:]]Normalized[[:space:]]Facial[[:space:]]UV-Texture[[:space:]]Dataset[[:space:]]for[[:space:]]3D[[:space:]]Face[[:space:]]Reconstruction/5d8bbd51-51d6-419c-9391-62a535ebbc83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FIANCEE_[[:space:]]Faster[[:space:]]Inference[[:space:]]of[[:space:]]Adversarial[[:space:]]Networks[[:space:]]via[[:space:]]Conditional[[:space:]]Early[[:space:]]Exits/e57d1a4b-6629-47c1-a93c-5f3290be15d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FJMP_[[:space:]]Factorized[[:space:]]Joint[[:space:]]Multi-Agent[[:space:]]Motion[[:space:]]Prediction[[:space:]]Over[[:space:]]Learned[[:space:]]Directed[[:space:]]Acyclic[[:space:]]Interaction[[:space:]]Graphs/70d94c18-b8a8-4d6f-97e1-a54154fa4279_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FLAG3D_[[:space:]]A[[:space:]]3D[[:space:]]Fitness[[:space:]]Activity[[:space:]]Dataset[[:space:]]With[[:space:]]Language[[:space:]]Instruction/3a069877-f70b-4235-9e65-8f281574b2d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FLEX_[[:space:]]Full-Body[[:space:]]Grasping[[:space:]]Without[[:space:]]Full-Body[[:space:]]Grasps/4d101095-349c-4bf1-8ef8-e755998691a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FREDOM_[[:space:]]Fairness[[:space:]]Domain[[:space:]]Adaptation[[:space:]]Approach[[:space:]]to[[:space:]]Semantic[[:space:]]Scene[[:space:]]Understanding/f34737cb-ecff-4970-8244-2013ddc592e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FaceLit_[[:space:]]Neural[[:space:]]3D[[:space:]]Relightable[[:space:]]Faces/bff6f074-10c4-4a1f-a979-ba93e1ca7d5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fair[[:space:]]Federated[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation[[:space:]]via[[:space:]]Client[[:space:]]Contribution[[:space:]]Estimation/462e38c8-3ab9-4a58-af40-27d62abd4e24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fair[[:space:]]Scratch[[:space:]]Tickets_[[:space:]]Finding[[:space:]]Fair[[:space:]]Sparse[[:space:]]Networks[[:space:]]Without[[:space:]]Weight[[:space:]]Training/9546022a-fbf0-4d25-bd57-1adc4e6b4183_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fake[[:space:]]It[[:space:]]Till[[:space:]]You[[:space:]]Make[[:space:]]It_[[:space:]]Learning[[:space:]]Transferable[[:space:]]Representations[[:space:]]From[[:space:]]Synthetic[[:space:]]ImageNet[[:space:]]Clones/347f1126-e010-42d7-8702-f477e7d083cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fantastic[[:space:]]Breaks_[[:space:]]A[[:space:]]Dataset[[:space:]]of[[:space:]]Paired[[:space:]]3D[[:space:]]Scans[[:space:]]of[[:space:]]Real-World[[:space:]]Broken[[:space:]]Objects[[:space:]]and[[:space:]]Their[[:space:]]Complete[[:space:]]Counterparts/07edf794-3189-4ce7-89c5-4394c00b209c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FashionSAP_[[:space:]]Symbols[[:space:]]and[[:space:]]Attributes[[:space:]]Prompt[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Fashion[[:space:]]Vision-Language[[:space:]]Pre-Training/0fe80008-1956-4ff4-af63-00700e8ef407_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fast[[:space:]]Contextual[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]With[[:space:]]Unbiased[[:space:]]Context[[:space:]]Augmentation/7b0515ec-f1c1-47d1-a37b-686cbf80e16c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fast[[:space:]]Monocular[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]With[[:space:]]Global-Sparse[[:space:]]Local-Dense[[:space:]]Grids/1724a28c-e4e9-4282-82b8-9e68fedba837_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fast[[:space:]]Point[[:space:]]Cloud[[:space:]]Generation[[:space:]]With[[:space:]]Straight[[:space:]]Flows/c45cc7bd-2461-48c7-9730-b1f1045fb128_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FastInst_[[:space:]]A[[:space:]]Simple[[:space:]]Query-Based[[:space:]]Model[[:space:]]for[[:space:]]Real-Time[[:space:]]Instance[[:space:]]Segmentation/740ea800-bc01-4ff4-94eb-f36e1b6485aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FeatER_[[:space:]]An[[:space:]]Efficient[[:space:]]Network[[:space:]]for[[:space:]]Human[[:space:]]Reconstruction[[:space:]]via[[:space:]]Feature[[:space:]]Map-Based[[:space:]]TransformER/0dfbc62e-bdb0-4fcc-9316-9f8bc084fbd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Feature[[:space:]]Aggregated[[:space:]]Queries[[:space:]]for[[:space:]]Transformer-Based[[:space:]]Video[[:space:]]Object[[:space:]]Detectors/cc79f9d7-e4b3-40fb-937c-e8548686d240_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Feature[[:space:]]Alignment[[:space:]]and[[:space:]]Uniformity[[:space:]]for[[:space:]]Test[[:space:]]Time[[:space:]]Adaptation/ffd8296f-b4fd-4b7c-bfca-0537fe33ffd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Feature[[:space:]]Representation[[:space:]]Learning[[:space:]]With[[:space:]]Adaptive[[:space:]]Displacement[[:space:]]Generation[[:space:]]and[[:space:]]Transformer[[:space:]]Fusion[[:space:]]for[[:space:]]Micro-Expression[[:space:]]Recognition/385a99de-9dc7-49fa-9247-706c08360dba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Feature[[:space:]]Separation[[:space:]]and[[:space:]]Recalibration[[:space:]]for[[:space:]]Adversarial[[:space:]]Robustness/9a91f81f-b8d3-455b-aee6-d22bf69e2585_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Feature[[:space:]]Shrinkage[[:space:]]Pyramid[[:space:]]for[[:space:]]Camouflaged[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]Transformers/f96e12d5-05c2-473d-aaba-ea8971b963d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FeatureBooster_[[:space:]]Boosting[[:space:]]Feature[[:space:]]Descriptors[[:space:]]With[[:space:]]a[[:space:]]Lightweight[[:space:]]Neural[[:space:]]Network/ce68f370-be02-4a4b-b28c-cb1a5fc217e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FedDM_[[:space:]]Iterative[[:space:]]Distribution[[:space:]]Matching[[:space:]]for[[:space:]]Communication-Efficient[[:space:]]Federated[[:space:]]Learning/e837dbf5-0dbb-458e-a35b-2ca7fee0f6b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FedSeg_[[:space:]]Class-Heterogeneous[[:space:]]Federated[[:space:]]Learning[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/850db7e1-d8bb-4d6c-98a5-04d6426f0233_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Federated[[:space:]]Domain[[:space:]]Generalization[[:space:]]With[[:space:]]Generalization[[:space:]]Adjustment/3e7ce20b-739c-44d1-9037-7391ed34fc30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Federated[[:space:]]Incremental[[:space:]]Semantic[[:space:]]Segmentation/0169dc0d-17c0-407d-9e37-70aad7d7dd0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Federated[[:space:]]Learning[[:space:]]With[[:space:]]Data-Agnostic[[:space:]]Distribution[[:space:]]Fusion/4029ce00-63fd-41e7-ad21-86b7639fa400_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Few-Shot[[:space:]]Class-Incremental[[:space:]]Learning[[:space:]]via[[:space:]]Class-Aware[[:space:]]Bilateral[[:space:]]Distillation/fbd253e9-d093-4829-a53f-1b789ec14e31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Few-Shot[[:space:]]Geometry-Aware[[:space:]]Keypoint[[:space:]]Localization/74f89256-ee7e-4f5d-9f08-0f6886280aef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Few-Shot[[:space:]]Learning[[:space:]]With[[:space:]]Visual[[:space:]]Distribution[[:space:]]Calibration[[:space:]]and[[:space:]]Cross-Modal[[:space:]]Distribution[[:space:]]Alignment/f084ef48-dc4f-4f37-8531-8a65c34505bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Few-Shot[[:space:]]Non-Line-of-Sight[[:space:]]Imaging[[:space:]]With[[:space:]]Signal-Surface[[:space:]]Collaborative[[:space:]]Regularization/dcbae509-9f7c-4fbd-8f9c-876ce3c70a14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Few-Shot[[:space:]]Referring[[:space:]]Relationships[[:space:]]in[[:space:]]Videos/d0d1bd46-db5b-4183-9eee-508655d41a49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Few-Shot[[:space:]]Semantic[[:space:]]Image[[:space:]]Synthesis[[:space:]]With[[:space:]]Class[[:space:]]Affinity[[:space:]]Transfer/afa0d173-996f-41d6-9889-c56c2711cce1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Filtering,[[:space:]]Distillation,[[:space:]]and[[:space:]]Hard[[:space:]]Negatives[[:space:]]for[[:space:]]Vision-Language[[:space:]]Pre-Training/06a92d63-97e4-4250-8e20-b8fd91b1e45d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Finding[[:space:]]Geometric[[:space:]]Models[[:space:]]by[[:space:]]Clustering[[:space:]]in[[:space:]]the[[:space:]]Consensus[[:space:]]Space/a7db908c-f937-4c6e-b75f-8e4505db4e2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fine-Grained[[:space:]]Audible[[:space:]]Video[[:space:]]Description/5044d6f7-bd07-46a8-a635-39cdb6d63928_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fine-Grained[[:space:]]Classification[[:space:]]With[[:space:]]Noisy[[:space:]]Labels/beddf9fe-f24d-4e55-a8fc-a59c77958a1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fine-Grained[[:space:]]Face[[:space:]]Swapping[[:space:]]via[[:space:]]Regional[[:space:]]GAN[[:space:]]Inversion/c899be37-60d1-45ad-93dc-5c5a74def8ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fine-Grained[[:space:]]Image-Text[[:space:]]Matching[[:space:]]by[[:space:]]Cross-Modal[[:space:]]Hard[[:space:]]Aligning[[:space:]]Network/8447f60f-b171-4599-bfb2-83c3ae11d6f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fine-Tuned[[:space:]]CLIP[[:space:]]Models[[:space:]]Are[[:space:]]Efficient[[:space:]]Video[[:space:]]Learners/676ef67e-cefc-47de-bf40-4f236f976401_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Finetune[[:space:]]Like[[:space:]]You[[:space:]]Pretrain_[[:space:]]Improved[[:space:]]Finetuning[[:space:]]of[[:space:]]Zero-Shot[[:space:]]Vision[[:space:]]Models/39bc201f-0c36-4517-920a-5e83e0952c0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FitMe_[[:space:]]Deep[[:space:]]Photorealistic[[:space:]]3D[[:space:]]Morphable[[:space:]]Model[[:space:]]Avatars/88e39874-5edb-4836-9636-7e4302d613b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fix[[:space:]]the[[:space:]]Noise_[[:space:]]Disentangling[[:space:]]Source[[:space:]]Feature[[:space:]]for[[:space:]]Controllable[[:space:]]Domain[[:space:]]Translation/6a5f48f9-e874-430d-8ed2-4e3756343286_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FlatFormer_[[:space:]]Flattened[[:space:]]Window[[:space:]]Attention[[:space:]]for[[:space:]]Efficient[[:space:]]Point[[:space:]]Cloud[[:space:]]Transformer/19aaa6a8-b9d8-40d7-976e-4445dd88b4b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FlexNeRF_[[:space:]]Photorealistic[[:space:]]Free-Viewpoint[[:space:]]Rendering[[:space:]]of[[:space:]]Moving[[:space:]]Humans[[:space:]]From[[:space:]]Sparse[[:space:]]Views/3be5ee0e-5ace-4712-9861-c18c66b43375_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FlexiViT_[[:space:]]One[[:space:]]Model[[:space:]]for[[:space:]]All[[:space:]]Patch[[:space:]]Sizes/dfbd44ff-6aa4-4b7c-932e-1b8321e8eee6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Flexible-Cm[[:space:]]GAN_[[:space:]]Towards[[:space:]]Precise[[:space:]]3D[[:space:]]Dose[[:space:]]Prediction[[:space:]]in[[:space:]]Radiotherapy/56b8d404-5ba4-475e-b5c2-e0c97be9a1d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Flow[[:space:]]Supervision[[:space:]]for[[:space:]]Deformable[[:space:]]NeRF/d4285961-1cfa-4057-be17-2a9759adb25b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FlowFormer++_[[:space:]]Masked[[:space:]]Cost[[:space:]]Volume[[:space:]]Autoencoding[[:space:]]for[[:space:]]Pretraining[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation/39c0b169-2e1c-4934-aa41-794fcf0dc6dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FlowGrad_[[:space:]]Controlling[[:space:]]the[[:space:]]Output[[:space:]]of[[:space:]]Generative[[:space:]]ODEs[[:space:]]With[[:space:]]Gradients/eab69965-6b5e-4c10-ac2c-92a12e805fc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Focus[[:space:]]on[[:space:]]Details_[[:space:]]Online[[:space:]]Multi-Object[[:space:]]Tracking[[:space:]]With[[:space:]]Diverse[[:space:]]Fine-Grained[[:space:]]Representation/1e48dbf4-ddcb-428b-99cb-191c2f384b09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Focused[[:space:]]and[[:space:]]Collaborative[[:space:]]Feedback[[:space:]]Integration[[:space:]]for[[:space:]]Interactive[[:space:]]Image[[:space:]]Segmentation/35323109-af71-470a-8378-8f5ccd339e4c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Foundation[[:space:]]Model[[:space:]]Drives[[:space:]]Weakly[[:space:]]Incremental[[:space:]]Learning[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/f81b1078-d0f7-4ea3-bc47-db5f45467809_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Four-View[[:space:]]Geometry[[:space:]]With[[:space:]]Unknown[[:space:]]Radial[[:space:]]Distortion/e23bf721-4aa5-43d7-803d-be1bff6923ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Frame[[:space:]]Flexible[[:space:]]Network/ab8eae57-aff5-49e9-b16d-c02a2789ef65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Frame[[:space:]]Interpolation[[:space:]]Transformer[[:space:]]and[[:space:]]Uncertainty[[:space:]]Guidance/a6a77d3c-19bb-489d-b468-9be4b07a583a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Frame-Event[[:space:]]Alignment[[:space:]]and[[:space:]]Fusion[[:space:]]Network[[:space:]]for[[:space:]]High[[:space:]]Frame[[:space:]]Rate[[:space:]]Tracking/5e7e1191-8301-44a8-88e8-0c3aefe56c83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FreeNeRF_[[:space:]]Improving[[:space:]]Few-Shot[[:space:]]Neural[[:space:]]Rendering[[:space:]]With[[:space:]]Free[[:space:]]Frequency[[:space:]]Regularization/d4c012b6-0147-4709-848a-94ef1b24cad2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FreeSeg_[[:space:]]Unified,[[:space:]]Universal[[:space:]]and[[:space:]]Open-Vocabulary[[:space:]]Image[[:space:]]Segmentation/be10e80a-a913-48be-8a41-ea96a6758972_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Freestyle[[:space:]]Layout-to-Image[[:space:]]Synthesis/1f12f34e-29a1-4038-803d-f942a5783d53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Frequency-Modulated[[:space:]]Point[[:space:]]Cloud[[:space:]]Rendering[[:space:]]With[[:space:]]Easy[[:space:]]Editing/f32cc457-c972-4574-ad88-5a1d55e76f0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fresnel[[:space:]]Microfacet[[:space:]]BRDF_[[:space:]]Unification[[:space:]]of[[:space:]]Polari-Radiometric[[:space:]]Surface-Body[[:space:]]Reflection/ded4c430-3d14-45d2-847c-e93f5ff8d012_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/From[[:space:]]Images[[:space:]]to[[:space:]]Textual[[:space:]]Prompts_[[:space:]]Zero-Shot[[:space:]]Visual[[:space:]]Question[[:space:]]Answering[[:space:]]With[[:space:]]Frozen[[:space:]]Large[[:space:]]Language[[:space:]]Models/92a87d35-b9f4-47c2-a77d-a57a569d1193_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/From[[:space:]]Node[[:space:]]Interaction[[:space:]]To[[:space:]]Hop[[:space:]]Interaction_[[:space:]]New[[:space:]]Effective[[:space:]]and[[:space:]]Scalable[[:space:]]Graph[[:space:]]Learning[[:space:]]Paradigm/f58209a7-0816-4c18-9a44-a49f5b4b97f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Frustratingly[[:space:]]Easy[[:space:]]Regularization[[:space:]]on[[:space:]]Representation[[:space:]]Can[[:space:]]Boost[[:space:]]Deep[[:space:]]Reinforcement[[:space:]]Learning/3341557f-8726-40d5-8408-2530b9a0afb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FrustumFormer_[[:space:]]Adaptive[[:space:]]Instance-Aware[[:space:]]Resampling[[:space:]]for[[:space:]]Multi-View[[:space:]]3D[[:space:]]Detection/2443d233-957b-4c13-86d9-51a182d54142_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Full[[:space:]]or[[:space:]]Weak[[:space:]]Annotations_[[:space:]]An[[:space:]]Adaptive[[:space:]]Strategy[[:space:]]for[[:space:]]Budget-Constrained[[:space:]]Annotation[[:space:]]Campaigns/f8f663b0-ce74-40a3-8b09-5eab626fcf63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fully[[:space:]]Self-Supervised[[:space:]]Depth[[:space:]]Estimation[[:space:]]From[[:space:]]Defocus[[:space:]]Clue/473b72af-c06f-4407-a75e-b99ec6df9cd1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fusing[[:space:]]Pre-Trained[[:space:]]Language[[:space:]]Models[[:space:]]With[[:space:]]Multimodal[[:space:]]Prompts[[:space:]]Through[[:space:]]Reinforcement[[:space:]]Learning/37da4808-71db-49c5-8871-b49921d65181_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fuzzy[[:space:]]Positive[[:space:]]Learning[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation/79f0f096-a82d-40e1-89b4-b67ce26de0e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/G-MSM_[[:space:]]Unsupervised[[:space:]]Multi-Shape[[:space:]]Matching[[:space:]]With[[:space:]]Graph-Based[[:space:]]Affinity[[:space:]]Priors/6fe34fa3-df34-46d2-8fa7-50ce5abc6159_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GALIP_[[:space:]]Generative[[:space:]]Adversarial[[:space:]]CLIPs[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Synthesis/d3291db7-b6a3-4233-b667-95991752e84c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GANHead_[[:space:]]Towards[[:space:]]Generative[[:space:]]Animatable[[:space:]]Neural[[:space:]]Head[[:space:]]Avatars/f689609a-3b35-4486-bba1-18d6217c13ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GANmouflage_[[:space:]]3D[[:space:]]Object[[:space:]]Nondetection[[:space:]]With[[:space:]]Texture[[:space:]]Fields/3b8567e6-b130-4bf1-ad56-f6860f916898_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GAPartNet_[[:space:]]Cross-Category[[:space:]]Domain-Generalizable[[:space:]]Object[[:space:]]Perception[[:space:]]and[[:space:]]Manipulation[[:space:]]via[[:space:]]Generalizable[[:space:]]and[[:space:]]Actionable[[:space:]]Parts/f30c4b95-2314-49d4-bbb8-8f5dc03c4a53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GCFAgg_[[:space:]]Global[[:space:]]and[[:space:]]Cross-View[[:space:]]Feature[[:space:]]Aggregation[[:space:]]for[[:space:]]Multi-View[[:space:]]Clustering/f1ffb7ba-6355-4a27-b5f0-5bc41a1376a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GD-MAE_[[:space:]]Generative[[:space:]]Decoder[[:space:]]for[[:space:]]MAE[[:space:]]Pre-Training[[:space:]]on[[:space:]]LiDAR[[:space:]]Point[[:space:]]Clouds/23d7fb77-417d-4a86-89d6-3875e98e2469_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GEN_[[:space:]]Pushing[[:space:]]the[[:space:]]Limits[[:space:]]of[[:space:]]Softmax-Based[[:space:]]Out-of-Distribution[[:space:]]Detection/001b88ed-2ac8-4fc3-991f-9d6c4d1308c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GFIE_[[:space:]]A[[:space:]]Dataset[[:space:]]and[[:space:]]Baseline[[:space:]]for[[:space:]]Gaze-Following[[:space:]]From[[:space:]]2D[[:space:]]to[[:space:]]3D[[:space:]]in[[:space:]]Indoor[[:space:]]Environments/603cdcbc-7711-4a4d-bf7c-7f3f35fedc34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GFPose_[[:space:]]Learning[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Prior[[:space:]]With[[:space:]]Gradient[[:space:]]Fields/b6dac27d-bd8f-4898-87ef-d484ae4db4ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GINA-3D_[[:space:]]Learning[[:space:]]To[[:space:]]Generate[[:space:]]Implicit[[:space:]]Neural[[:space:]]Assets[[:space:]]in[[:space:]]the[[:space:]]Wild/6af78c34-2888-472a-b1a8-571a9ea305e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GIVL_[[:space:]]Improving[[:space:]]Geographical[[:space:]]Inclusivity[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models[[:space:]]With[[:space:]]Pre-Training[[:space:]]Methods/96cabc56-4e99-420e-a869-6caa27823e00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GKEAL_[[:space:]]Gaussian[[:space:]]Kernel[[:space:]]Embedded[[:space:]]Analytic[[:space:]]Learning[[:space:]]for[[:space:]]Few-Shot[[:space:]]Class[[:space:]]Incremental[[:space:]]Task/7bc67700-534f-4643-b8ee-2f5ff62fbe0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GLIGEN_[[:space:]]Open-Set[[:space:]]Grounded[[:space:]]Text-to-Image[[:space:]]Generation/46d95833-da20-4df2-8222-9407fb23ba0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GLeaD_[[:space:]]Improving[[:space:]]GANs[[:space:]]With[[:space:]]a[[:space:]]Generator-Leading[[:space:]]Task/6f829ad5-1705-4738-9f27-3ef692978fc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GM-NeRF_[[:space:]]Learning[[:space:]]Generalizable[[:space:]]Model-Based[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]From[[:space:]]Multi-View[[:space:]]Images/b2f4aa82-a4ec-4a99-99e7-e82549cce122_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GP-VTON_[[:space:]]Towards[[:space:]]General[[:space:]]Purpose[[:space:]]Virtual[[:space:]]Try-On[[:space:]]via[[:space:]]Collaborative[[:space:]]Local-Flow[[:space:]]Global-Parsing[[:space:]]Learning/03bb5814-18a4-42b7-9cc9-da3c7e5e77cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GRES_[[:space:]]Generalized[[:space:]]Referring[[:space:]]Expression[[:space:]]Segmentation/4f99d13e-41bc-4102-9aed-41e84724b477_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GaitGCI_[[:space:]]Generative[[:space:]]Counterfactual[[:space:]]Intervention[[:space:]]for[[:space:]]Gait[[:space:]]Recognition/e921b924-f6c8-4f9a-80f5-a172aeb5ff04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Galactic_[[:space:]]Scaling[[:space:]]End-to-End[[:space:]]Reinforcement[[:space:]]Learning[[:space:]]for[[:space:]]Rearrangement[[:space:]]at[[:space:]]100k[[:space:]]Steps-per-Second/b752f925-9394-404d-b54a-6dcafeb71c4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GamutMLP_[[:space:]]A[[:space:]]Lightweight[[:space:]]MLP[[:space:]]for[[:space:]]Color[[:space:]]Loss[[:space:]]Recovery/f293da40-a174-49c6-95bf-689b226934e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GarmentTracking_[[:space:]]Category-Level[[:space:]]Garment[[:space:]]Pose[[:space:]]Tracking/95caa075-2422-4cc1-9bf0-b32452a1fee1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gated[[:space:]]Multi-Resolution[[:space:]]Transfer[[:space:]]Network[[:space:]]for[[:space:]]Burst[[:space:]]Restoration[[:space:]]and[[:space:]]Enhancement/c09f27b3-f5e5-42ca-a645-0d86d46afc10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gated[[:space:]]Stereo_[[:space:]]Joint[[:space:]]Depth[[:space:]]Estimation[[:space:]]From[[:space:]]Gated[[:space:]]and[[:space:]]Wide-Baseline[[:space:]]Active[[:space:]]Stereo[[:space:]]Cues/5d98fa5b-9812-4cde-b2ed-c97d64ffbf5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gaussian[[:space:]]Label[[:space:]]Distribution[[:space:]]Learning[[:space:]]for[[:space:]]Spherical[[:space:]]Image[[:space:]]Object[[:space:]]Detection/577cc560-9e49-4b3e-924d-ef835d876f19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GazeNeRF_[[:space:]]3D-Aware[[:space:]]Gaze[[:space:]]Redirection[[:space:]]With[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/29eaaae8-61c2-488b-bbb0-fdfc51fad88e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gazeformer_[[:space:]]Scalable,[[:space:]]Effective[[:space:]]and[[:space:]]Fast[[:space:]]Prediction[[:space:]]of[[:space:]]Goal-Directed[[:space:]]Human[[:space:]]Attention/1fbf7ac7-ca7e-47ee-87de-1337121749f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GeneCIS_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]General[[:space:]]Conditional[[:space:]]Image[[:space:]]Similarity/b8733927-92fb-4c1c-a651-0bbf3a75a237_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalist_[[:space:]]Decoupling[[:space:]]Natural[[:space:]]and[[:space:]]Robust[[:space:]]Generalization/0d84e267-f0a4-4706-9890-737b39627f4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalizable[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations[[:space:]]via[[:space:]]Instance[[:space:]]Pattern[[:space:]]Composers/0d1e96dd-ca74-47c5-869d-cf4d5f8c48b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalizable[[:space:]]Local[[:space:]]Feature[[:space:]]Pre-Training[[:space:]]for[[:space:]]Deformable[[:space:]]Shape[[:space:]]Analysis/82166b12-5ff8-4344-b2e2-e450ee4da6dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalization[[:space:]]Matters_[[:space:]]Loss[[:space:]]Minima[[:space:]]Flattening[[:space:]]via[[:space:]]Parameter[[:space:]]Hybridization[[:space:]]for[[:space:]]Efficient[[:space:]]Online[[:space:]]Knowledge[[:space:]]Distillation/88c4cc3c-1fa2-4700-a5e4-50a1a9b8583f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalized[[:space:]]Decoding[[:space:]]for[[:space:]]Pixel,[[:space:]]Image,[[:space:]]and[[:space:]]Language/8664b015-aec5-4cf4-b802-aafad5d4566e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalized[[:space:]]Deep[[:space:]]3D[[:space:]]Shape[[:space:]]Prior[[:space:]]via[[:space:]]Part-Discretized[[:space:]]Diffusion[[:space:]]Process/041e4223-13e8-4ef9-96bf-d34cd49799b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalized[[:space:]]Relation[[:space:]]Modeling[[:space:]]for[[:space:]]Transformer[[:space:]]Tracking/ef24612a-e0fc-43f8-a2ef-726012936ccc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalized[[:space:]]UAV[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Frequency[[:space:]]Domain[[:space:]]Disentanglement/93aca1cf-171d-4917-b030-aff32a4a198f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalizing[[:space:]]Dataset[[:space:]]Distillation[[:space:]]via[[:space:]]Deep[[:space:]]Generative[[:space:]]Prior/b5678fb3-5434-4cce-8d46-61766bacd1b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generating[[:space:]]Aligned[[:space:]]Pseudo-Supervision[[:space:]]From[[:space:]]Non-Aligned[[:space:]]Data[[:space:]]for[[:space:]]Image[[:space:]]Restoration[[:space:]]in[[:space:]]Under-Display[[:space:]]Camera/727805ad-1c9c-4052-8065-131c85105a71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generating[[:space:]]Anomalies[[:space:]]for[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection[[:space:]]With[[:space:]]Prompt-Based[[:space:]]Feature[[:space:]]Mapping/41fd9461-ac71-49b4-8511-8ee94003fa1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generating[[:space:]]Features[[:space:]]With[[:space:]]Increased[[:space:]]Crop-Related[[:space:]]Diversity[[:space:]]for[[:space:]]Few-Shot[[:space:]]Object[[:space:]]Detection/3f152bdb-f3ee-4bdd-b0b0-37e362aba7c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generating[[:space:]]Holistic[[:space:]]3D[[:space:]]Human[[:space:]]Motion[[:space:]]From[[:space:]]Speech/1a41415a-ba2d-4062-b79c-023c1d8a0877_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generating[[:space:]]Human[[:space:]]Motion[[:space:]]From[[:space:]]Textual[[:space:]]Descriptions[[:space:]]With[[:space:]]Discrete[[:space:]]Representations/05522972-d9e1-47bd-81e4-8c07a32b96fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generating[[:space:]]Part-Aware[[:space:]]Editable[[:space:]]3D[[:space:]]Shapes[[:space:]]Without[[:space:]]3D[[:space:]]Supervision/b269a9a8-ce6e-4189-89c2-6a5ee04ca326_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generative[[:space:]]Bias[[:space:]]for[[:space:]]Robust[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/5916735e-c6c3-4eeb-bd43-8ead2be0f784_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generative[[:space:]]Diffusion[[:space:]]Prior[[:space:]]for[[:space:]]Unified[[:space:]]Image[[:space:]]Restoration[[:space:]]and[[:space:]]Enhancement/bfa3600f-560e-4a52-b7d5-b1336d73de16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generative[[:space:]]Semantic[[:space:]]Segmentation/073bcf72-97eb-4ed1-81d6-90210f72fda4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generic-to-Specific[[:space:]]Distillation[[:space:]]of[[:space:]]Masked[[:space:]]Autoencoders/7140a85a-1b69-4bb4-8275-5e013099142d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Genie_[[:space:]]Show[[:space:]]Me[[:space:]]the[[:space:]]Data[[:space:]]for[[:space:]]Quantization/456005a8-6617-4d76-a3f5-0e87783c7cfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GeoLayoutLM_[[:space:]]Geometric[[:space:]]Pre-Training[[:space:]]for[[:space:]]Visual[[:space:]]Information[[:space:]]Extraction/63c5823e-6506-44c9-8edf-0374742538e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GeoMAE_[[:space:]]Masked[[:space:]]Geometric[[:space:]]Target[[:space:]]Prediction[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Pre-Training/8f252a29-319b-4d51-a0df-a0aabc3ead76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GeoMVSNet_[[:space:]]Learning[[:space:]]Multi-View[[:space:]]Stereo[[:space:]]With[[:space:]]Geometry[[:space:]]Perception/92fabf3b-2112-49c0-96cb-303e481d0a3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GeoNet_[[:space:]]Benchmarking[[:space:]]Unsupervised[[:space:]]Adaptation[[:space:]]Across[[:space:]]Geographies/33bd72eb-a7ef-4893-a191-49fb0b5194c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GeoVLN_[[:space:]]Learning[[:space:]]Geometry-Enhanced[[:space:]]Visual[[:space:]]Representation[[:space:]]With[[:space:]]Slot[[:space:]]Attention[[:space:]]for[[:space:]]Vision-and-Language[[:space:]]Navigation/01a2c1e9-e154-435d-9e81-421a75c39887_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Geometric[[:space:]]Visual[[:space:]]Similarity[[:space:]]Learning[[:space:]]in[[:space:]]3D[[:space:]]Medical[[:space:]]Image[[:space:]]Self-Supervised[[:space:]]Pre-Training/d7cd70b7-0ba7-4819-9ce9-395ca6418b4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Geometry[[:space:]]and[[:space:]]Uncertainty-Aware[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Class-Incremental[[:space:]]Semantic[[:space:]]Segmentation/8c9fd5eb-45d5-469e-9d12-a663d955573d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GlassesGAN_[[:space:]]Eyewear[[:space:]]Personalization[[:space:]]Using[[:space:]]Synthetic[[:space:]]Appearance[[:space:]]Discovery[[:space:]]and[[:space:]]Targeted[[:space:]]Subspace[[:space:]]Modeling/f0af491f-057d-4f92-af4d-853638df19c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Global[[:space:]]Vision[[:space:]]Transformer[[:space:]]Pruning[[:space:]]With[[:space:]]Hessian-Aware[[:space:]]Saliency/3d105b5e-5e9c-4887-93a3-58dd5db39522_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Global[[:space:]]and[[:space:]]Local[[:space:]]Mixture[[:space:]]Consistency[[:space:]]Cumulative[[:space:]]Learning[[:space:]]for[[:space:]]Long-Tailed[[:space:]]Visual[[:space:]]Recognitions/a01a79e2-23cb-44c6-abaa-ab8e8f28106a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Global-to-Local[[:space:]]Modeling[[:space:]]for[[:space:]]Video-Based[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Estimation/cf1d20bd-b13b-4510-a7f4-aa216748988b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Glocal[[:space:]]Energy-Based[[:space:]]Learning[[:space:]]for[[:space:]]Few-Shot[[:space:]]Open-Set[[:space:]]Recognition/08e60982-dd59-47ff-9966-35a586447e3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gloss[[:space:]]Attention[[:space:]]for[[:space:]]Gloss-Free[[:space:]]Sign[[:space:]]Language[[:space:]]Translation/a7e9cb88-72d6-4180-bd23-0f64ab47c5a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Good[[:space:]]Is[[:space:]]Bad_[[:space:]]Causality[[:space:]]Inspired[[:space:]]Cloth-Debiasing[[:space:]]for[[:space:]]Cloth-Changing[[:space:]]Person[[:space:]]Re-Identification/b3294907-09bb-4bb9-8ab7-4a1734cfaf03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GraVoS_[[:space:]]Voxel[[:space:]]Selection[[:space:]]for[[:space:]]3D[[:space:]]Point-Cloud[[:space:]]Detection/c98996da-6504-4514-88b1-e4d860f88998_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Grad-PU_[[:space:]]Arbitrary-Scale[[:space:]]Point[[:space:]]Cloud[[:space:]]Upsampling[[:space:]]via[[:space:]]Gradient[[:space:]]Descent[[:space:]]With[[:space:]]Learned[[:space:]]Distance[[:space:]]Functions/89782297-25c6-43b9-973b-c7d4d6ec26df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GradICON_[[:space:]]Approximate[[:space:]]Diffeomorphisms[[:space:]]via[[:space:]]Gradient[[:space:]]Inverse[[:space:]]Consistency/b57515fb-e670-4d72-8a7f-25fbb8812a1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GradMA_[[:space:]]A[[:space:]]Gradient-Memory-Based[[:space:]]Accelerated[[:space:]]Federated[[:space:]]Learning[[:space:]]With[[:space:]]Alleviated[[:space:]]Catastrophic[[:space:]]Forgetting/aef7d684-7a57-4e70-bdf0-7f08da932355_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gradient[[:space:]]Norm[[:space:]]Aware[[:space:]]Minimization[[:space:]]Seeks[[:space:]]First-Order[[:space:]]Flatness[[:space:]]and[[:space:]]Improves[[:space:]]Generalization/49b4f237-d3bc-43f5-a6af-fedbff859b5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gradient-Based[[:space:]]Uncertainty[[:space:]]Attribution[[:space:]]for[[:space:]]Explainable[[:space:]]Bayesian[[:space:]]Deep[[:space:]]Learning/d8decc43-10a7-4b9d-bc17-4d7780e58665_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Graph[[:space:]]Representation[[:space:]]for[[:space:]]Order-Aware[[:space:]]Visual[[:space:]]Transformation/f902fe1d-bc38-48b7-8541-6cebc85378d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Graph[[:space:]]Transformer[[:space:]]GANs[[:space:]]for[[:space:]]Graph-Constrained[[:space:]]House[[:space:]]Generation/c23313a4-8a49-4c0d-ad02-8ed18c219f16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Graphics[[:space:]]Capsule_[[:space:]]Learning[[:space:]]Hierarchical[[:space:]]3D[[:space:]]Face[[:space:]]Representations[[:space:]]From[[:space:]]2D[[:space:]]Images/0b33e474-5d91-4e73-ac61-7d49e2c2429a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Grid-Guided[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Large[[:space:]]Urban[[:space:]]Scenes/d95c3201-8189-457a-81f1-1bbe76c7aa53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ground-Truth[[:space:]]Free[[:space:]]Meta-Learning[[:space:]]for[[:space:]]Deep[[:space:]]Compressive[[:space:]]Sampling/64524939-3424-440f-a717-5ca20a6e12d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Grounding[[:space:]]Counterfactual[[:space:]]Explanation[[:space:]]of[[:space:]]Image[[:space:]]Classifiers[[:space:]]to[[:space:]]Textual[[:space:]]Concept[[:space:]]Space/21ad90d3-08e3-46ea-b30d-13a44a20632a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GrowSP_[[:space:]]Unsupervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]of[[:space:]]3D[[:space:]]Point[[:space:]]Clouds/80056438-954c-4d80-b162-01e3c73bdb5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Guided[[:space:]]Depth[[:space:]]Super-Resolution[[:space:]]by[[:space:]]Deep[[:space:]]Anisotropic[[:space:]]Diffusion/c8c32274-76fc-4b32-99bd-1b65cf8abd86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Guided[[:space:]]Recommendation[[:space:]]for[[:space:]]Model[[:space:]]Fine-Tuning/075a8856-68a4-4981-8a4b-d3d126bd8598_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Guiding[[:space:]]Pseudo-Labels[[:space:]]With[[:space:]]Uncertainty[[:space:]]Estimation[[:space:]]for[[:space:]]Source-Free[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/6b91d979-534c-43ca-90a6-4774f42f1db3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/H2ONet_[[:space:]]Hand-Occlusion-and-Orientation-Aware[[:space:]]Network[[:space:]]for[[:space:]]Real-Time[[:space:]]3D[[:space:]]Hand[[:space:]]Mesh[[:space:]]Reconstruction/23222901-0448-41d0-b7f8-f543d0a999d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HAAV_[[:space:]]Hierarchical[[:space:]]Aggregation[[:space:]]of[[:space:]]Augmented[[:space:]]Views[[:space:]]for[[:space:]]Image[[:space:]]Captioning/be695692-599f-4fb2-86f3-9569c17062bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HARP_[[:space:]]Personalized[[:space:]]Hand[[:space:]]Reconstruction[[:space:]]From[[:space:]]a[[:space:]]Monocular[[:space:]]RGB[[:space:]]Video/6a70724a-a311-47e2-bbb2-e76eb3e9002f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HDR[[:space:]]Imaging[[:space:]]With[[:space:]]Spatially[[:space:]]Varying[[:space:]]Signal-to-Noise[[:space:]]Ratios/a34a8639-1e50-46b1-b571-108f0ba0ad6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HGFormer_[[:space:]]Hierarchical[[:space:]]Grouping[[:space:]]Transformer[[:space:]]for[[:space:]]Domain[[:space:]]Generalized[[:space:]]Semantic[[:space:]]Segmentation/ebad4766-5acc-4e4d-b66c-7087814fa72f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HGNet_[[:space:]]Learning[[:space:]]Hierarchical[[:space:]]Geometry[[:space:]]From[[:space:]]Points,[[:space:]]Edges,[[:space:]]and[[:space:]]Surfaces/dc471933-fb33-4e09-9b5d-54c05533f543_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HIER_[[:space:]]Metric[[:space:]]Learning[[:space:]]Beyond[[:space:]]Class[[:space:]]Labels[[:space:]]via[[:space:]]Hierarchical[[:space:]]Regularization/858027b5-9808-4694-aa92-75d3fb42beed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HNeRV_[[:space:]]A[[:space:]]Hybrid[[:space:]]Neural[[:space:]]Representation[[:space:]]for[[:space:]]Videos/b3f47861-072c-45a5-a145-ed433f6efc79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HOICLIP_[[:space:]]Efficient[[:space:]]Knowledge[[:space:]]Transfer[[:space:]]for[[:space:]]HOI[[:space:]]Detection[[:space:]]With[[:space:]]Vision-Language[[:space:]]Models/4dae0366-9960-402d-8e49-a116e85e97d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HOLODIFFUSION_[[:space:]]Training[[:space:]]a[[:space:]]3D[[:space:]]Diffusion[[:space:]]Model[[:space:]]Using[[:space:]]2D[[:space:]]Images/b840be1a-583b-49dd-a0bd-ac1ee8c2a22a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HOOD_[[:space:]]Hierarchical[[:space:]]Graphs[[:space:]]for[[:space:]]Generalized[[:space:]]Modelling[[:space:]]of[[:space:]]Clothing[[:space:]]Dynamics/6e4eac6a-b406-4300-8039-855c2bb63036_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HOTNAS_[[:space:]]Hierarchical[[:space:]]Optimal[[:space:]]Transport[[:space:]]for[[:space:]]Neural[[:space:]]Architecture[[:space:]]Search/f65c4ad2-113f-43c8-872d-7983e4910b7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HRDFuse_[[:space:]]Monocular[[:space:]]360deg[[:space:]]Depth[[:space:]]Estimation[[:space:]]by[[:space:]]Collaboratively[[:space:]]Learning[[:space:]]Holistic-With-Regional[[:space:]]Depth[[:space:]]Distributions/3e37ac00-5a26-4085-83ed-142b806c53db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HS-Pose_[[:space:]]Hybrid[[:space:]]Scope[[:space:]]Feature[[:space:]]Extraction[[:space:]]for[[:space:]]Category-Level[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/c85bd2f4-d709-4012-ab16-1fea8b75c815_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HaLP_[[:space:]]Hallucinating[[:space:]]Latent[[:space:]]Positives[[:space:]]for[[:space:]]Skeleton-Based[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]of[[:space:]]Actions/c84e81f2-1a91-4a9d-9842-2e6c9107c706_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Habitat-Matterport[[:space:]]3D[[:space:]]Semantics[[:space:]]Dataset/74f6e904-ded5-471d-98a0-8e689490b2d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HairStep_[[:space:]]Transfer[[:space:]]Synthetic[[:space:]]to[[:space:]]Real[[:space:]]Using[[:space:]]Strand[[:space:]]and[[:space:]]Depth[[:space:]]Maps[[:space:]]for[[:space:]]Single-View[[:space:]]3D[[:space:]]Hair[[:space:]]Modeling/23a02b3e-1bfd-4921-900e-54413690cada_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ham2Pose_[[:space:]]Animating[[:space:]]Sign[[:space:]]Language[[:space:]]Notation[[:space:]]Into[[:space:]]Pose[[:space:]]Sequences/09b772eb-9ae5-4eaf-b19b-10b985b4eac3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hand[[:space:]]Avatar_[[:space:]]Free-Pose[[:space:]]Hand[[:space:]]Animation[[:space:]]and[[:space:]]Rendering[[:space:]]From[[:space:]]Monocular[[:space:]]Video/3757f991-4d12-4f25-a3d4-2a969ceef018_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HandNeRF_[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Animatable[[:space:]]Interacting[[:space:]]Hands/f003dc5f-24be-45b1-af70-c31e73601e62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HandsOff_[[:space:]]Labeled[[:space:]]Dataset[[:space:]]Generation[[:space:]]With[[:space:]]No[[:space:]]Additional[[:space:]]Human[[:space:]]Annotations/a2cd654b-8395-48de-aa8a-5ddf3f965f37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Handwritten[[:space:]]Text[[:space:]]Generation[[:space:]]From[[:space:]]Visual[[:space:]]Archetypes/2ce7ce58-c05c-4352-821b-701be0be36b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Handy_[[:space:]]Towards[[:space:]]a[[:space:]]High[[:space:]]Fidelity[[:space:]]3D[[:space:]]Hand[[:space:]]Shape[[:space:]]and[[:space:]]Appearance[[:space:]]Model/e9764bc2-8245-4035-9645-48c237e5d238_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hard[[:space:]]Patches[[:space:]]Mining[[:space:]]for[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling/2ad7317d-49b9-44a3-a806-d318b8e167de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hard[[:space:]]Sample[[:space:]]Matters[[:space:]]a[[:space:]]Lot[[:space:]]in[[:space:]]Zero-Shot[[:space:]]Quantization/36b7504a-58f3-4796-8b64-aab8fc97fc18_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Harmonious[[:space:]]Feature[[:space:]]Learning[[:space:]]for[[:space:]]Interactive[[:space:]]Hand-Object[[:space:]]Pose[[:space:]]Estimation/6de78cc8-6cec-4928-9d0c-a08dca9e504e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Harmonious[[:space:]]Teacher[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Object[[:space:]]Detection/b29b9bb8-76ae-49a8-9b29-693b83c4d688_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Heat[[:space:]]Diffusion[[:space:]]Based[[:space:]]Multi-Scale[[:space:]]and[[:space:]]Geometric[[:space:]]Structure-Aware[[:space:]]Transformer[[:space:]]for[[:space:]]Mesh[[:space:]]Segmentation/9ae3f659-2da7-427b-95a3-d1be627c0062_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HelixSurf_[[:space:]]A[[:space:]]Robust[[:space:]]and[[:space:]]Efficient[[:space:]]Neural[[:space:]]Implicit[[:space:]]Surface[[:space:]]Learning[[:space:]]of[[:space:]]Indoor[[:space:]]Scenes[[:space:]]With[[:space:]]Iterative[[:space:]]Intertwined[[:space:]]Regularization/8612a135-9bd3-4ddd-b296-b5d7dcc74005_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Heterogeneous[[:space:]]Continual[[:space:]]Learning/33e2cd22-b8cb-45ea-ac3f-20c535a2c84e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HexPlane_[[:space:]]A[[:space:]]Fast[[:space:]]Representation[[:space:]]for[[:space:]]Dynamic[[:space:]]Scenes/7f352709-b49b-4e2d-93ad-f927aa9aff95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hi-LASSIE_[[:space:]]High-Fidelity[[:space:]]Articulated[[:space:]]Shape[[:space:]]and[[:space:]]Skeleton[[:space:]]Discovery[[:space:]]From[[:space:]]Sparse[[:space:]]Image[[:space:]]Ensemble/af0840d5-209b-48d5-80aa-d50bc4dc2ecb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hi4D_[[:space:]]4D[[:space:]]Instance[[:space:]]Segmentation[[:space:]]of[[:space:]]Close[[:space:]]Human[[:space:]]Interaction/143259ff-ca17-4092-a12a-14d488dc882e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hidden[[:space:]]Gems_[[:space:]]4D[[:space:]]Radar[[:space:]]Scene[[:space:]]Flow[[:space:]]Learning[[:space:]]Using[[:space:]]Cross-Modal[[:space:]]Supervision/37f535df-cdf7-43b3-80e5-cbfd70b4d964_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HierVL_[[:space:]]Learning[[:space:]]Hierarchical[[:space:]]Video-Language[[:space:]]Embeddings/e12cff7e-b4c8-4706-b18f-ed510a1ce232_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]B-Frame[[:space:]]Video[[:space:]]Coding[[:space:]]Using[[:space:]]Two-Layer[[:space:]]CANF[[:space:]]Without[[:space:]]Motion[[:space:]]Coding/c6db7b0b-0000-43c5-9ad2-de3793a02a60_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Dense[[:space:]]Correlation[[:space:]]Distillation[[:space:]]for[[:space:]]Few-Shot[[:space:]]Segmentation/456f5492-17eb-4611-a55b-27fa870c23d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Discriminative[[:space:]]Learning[[:space:]]Improves[[:space:]]Visual[[:space:]]Representations[[:space:]]of[[:space:]]Biomedical[[:space:]]Microscopy/0b959d4d-6c5f-4e4b-b67f-6aa245ff9947_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Fine-Grained[[:space:]]Image[[:space:]]Forgery[[:space:]]Detection[[:space:]]and[[:space:]]Localization/e1ec5777-2da7-4ba7-9987-0f0e987e0fe1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Neural[[:space:]]Memory[[:space:]]Network[[:space:]]for[[:space:]]Low[[:space:]]Latency[[:space:]]Event[[:space:]]Processing/8eaa0854-28ec-4048-bacc-7908ba2e8a39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Multi-Task[[:space:]]Learning/9b83122e-7df6-4755-9080-045132a86637_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Semantic[[:space:]]Contrast[[:space:]]for[[:space:]]Scene-Aware[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/e32e9c9d-f444-4f7c-95f5-b5e984a20b62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Semantic[[:space:]]Correspondence[[:space:]]Networks[[:space:]]for[[:space:]]Video[[:space:]]Paragraph[[:space:]]Grounding/317ca47c-a82b-4b18-8771-116b604103c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Supervision[[:space:]]and[[:space:]]Shuffle[[:space:]]Data[[:space:]]Augmentation[[:space:]]for[[:space:]]3D[[:space:]]Semi-Supervised[[:space:]]Object[[:space:]]Detection/493ad7c1-8355-4d08-80b3-578bfbf31de8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Temporal[[:space:]]Transformer[[:space:]]for[[:space:]]3D[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation[[:space:]]and[[:space:]]Action[[:space:]]Recognition[[:space:]]From[[:space:]]Egocentric[[:space:]]RGB[[:space:]]Videos/f04bd8f4-61fd-4011-88ce-0dde9eafa207_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Video-Moment[[:space:]]Retrieval[[:space:]]and[[:space:]]Step-Captioning/a738c1d6-a786-4056-b7d7-075155803c7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High[[:space:]]Fidelity[[:space:]]3D[[:space:]]Hand[[:space:]]Shape[[:space:]]Reconstruction[[:space:]]via[[:space:]]Scalable[[:space:]]Graph[[:space:]]Frequency[[:space:]]Decomposition/5a590590-ec40-4b02-82e4-a6e040412cff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Fidelity[[:space:]]3D[[:space:]]Face[[:space:]]Generation[[:space:]]From[[:space:]]Natural[[:space:]]Language[[:space:]]Descriptions/221bc20e-9f22-4f39-b9c6-3347a161adaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Fidelity[[:space:]]3D[[:space:]]GAN[[:space:]]Inversion[[:space:]]by[[:space:]]Pseudo-Multi-View[[:space:]]Optimization/b41618d3-100c-44af-b4bf-62d68858fc65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Fidelity[[:space:]]3D[[:space:]]Human[[:space:]]Digitization[[:space:]]From[[:space:]]Single[[:space:]]2K[[:space:]]Resolution[[:space:]]Images/7fcc5c91-05d2-4976-9a74-91c4624a0587_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Fidelity[[:space:]]Clothed[[:space:]]Avatar[[:space:]]Reconstruction[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Image/edf46f1c-1cbd-4fed-b47b-e63ec9f4e3f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Fidelity[[:space:]]Event-Radiance[[:space:]]Recovery[[:space:]]via[[:space:]]Transient[[:space:]]Event[[:space:]]Frequency/a5c13768-b95e-4541-9dc8-b9f3f13ac1ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Fidelity[[:space:]]Facial[[:space:]]Avatar[[:space:]]Reconstruction[[:space:]]From[[:space:]]Monocular[[:space:]]Video[[:space:]]With[[:space:]]Generative[[:space:]]Priors/7dd7e4bf-7109-441a-b1a6-5c2de3405517_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Fidelity[[:space:]]Generalized[[:space:]]Emotional[[:space:]]Talking[[:space:]]Face[[:space:]]Generation[[:space:]]With[[:space:]]Multi-Modal[[:space:]]Emotion[[:space:]]Space[[:space:]]Learning/9029c570-d1ac-421e-bf1e-2f5f989e8775_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Fidelity[[:space:]]Guided[[:space:]]Image[[:space:]]Synthesis[[:space:]]With[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models/fa157b34-3a8f-41e4-8009-4dc55886661c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Fidelity[[:space:]]and[[:space:]]Freely[[:space:]]Controllable[[:space:]]Talking[[:space:]]Head[[:space:]]Video[[:space:]]Generation/17ca05fa-7c4c-4a5c-aabb-ca2a4ae5ee0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Frequency[[:space:]]Stereo[[:space:]]Matching[[:space:]]Network/8daeb755-feba-464f-861a-07dcceadb29a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Res[[:space:]]Facial[[:space:]]Appearance[[:space:]]Capture[[:space:]]From[[:space:]]Polarized[[:space:]]Smartphone[[:space:]]Images/404dd195-909b-4cc5-892f-af8d389cf14c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Resolution[[:space:]]Image[[:space:]]Reconstruction[[:space:]]With[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models[[:space:]]From[[:space:]]Human[[:space:]]Brain[[:space:]]Activity/7e5297ac-5036-48f8-9c28-3dde672f4b29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Highly[[:space:]]Confident[[:space:]]Local[[:space:]]Structure[[:space:]]Based[[:space:]]Consensus[[:space:]]Graph[[:space:]]Learning[[:space:]]for[[:space:]]Incomplete[[:space:]]Multi-View[[:space:]]Clustering/46969f18-0f4a-4a19-b481-f87ccd56d6a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hint-Aug_[[:space:]]Drawing[[:space:]]Hints[[:space:]]From[[:space:]]Foundation[[:space:]]Vision[[:space:]]Transformers[[:space:]]Towards[[:space:]]Boosted[[:space:]]Few-Shot[[:space:]]Parameter-Efficient[[:space:]]Tuning/a3764503-118e-455a-96b8-d25467d3741e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Histopathology[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Analysis[[:space:]]With[[:space:]]Heterogeneous[[:space:]]Graph[[:space:]]Representation[[:space:]]Learning/6f12123e-d9f5-4165-8eb4-dd7fc33ab7d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HouseDiffusion_[[:space:]]Vector[[:space:]]Floorplan[[:space:]]Generation[[:space:]]via[[:space:]]a[[:space:]]Diffusion[[:space:]]Model[[:space:]]With[[:space:]]Discrete[[:space:]]and[[:space:]]Continuous[[:space:]]Denoising/6d4aecb2-b581-46ec-9631-5b1dce6c1d67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/How[[:space:]]Can[[:space:]]Objects[[:space:]]Help[[:space:]]Action[[:space:]]Recognition_/726ae121-68dc-4c96-a056-2bcedd52cfd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/How[[:space:]]To[[:space:]]Prevent[[:space:]]the[[:space:]]Continuous[[:space:]]Damage[[:space:]]of[[:space:]]Noises[[:space:]]To[[:space:]]Model[[:space:]]Training_/e9f2cf77-df06-41f1-b222-853f974f647d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/How[[:space:]]To[[:space:]]Prevent[[:space:]]the[[:space:]]Poor[[:space:]]Performance[[:space:]]Clients[[:space:]]for[[:space:]]Personalized[[:space:]]Federated[[:space:]]Learning_/62406dbb-5ede-484f-96dc-2ad551e02862_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/How[[:space:]]You[[:space:]]Feelin'_[[:space:]]Learning[[:space:]]Emotions[[:space:]]and[[:space:]]Mental[[:space:]]States[[:space:]]in[[:space:]]Movie[[:space:]]Scenes/5d702412-d557-4075-8791-65f42afd3c14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/How[[:space:]]to[[:space:]]Backdoor[[:space:]]Diffusion[[:space:]]Models_/ce8e3795-2464-47fe-a04a-b70cd833b12c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HuManiFlow_[[:space:]]Ancestor-Conditioned[[:space:]]Normalising[[:space:]]Flows[[:space:]]on[[:space:]]SO(3)[[:space:]]Manifolds[[:space:]]for[[:space:]]Human[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Distribution[[:space:]]Estimation/a6b3b142-4338-445f-8b3d-5b82447982e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hubs[[:space:]]and[[:space:]]Hyperspheres_[[:space:]]Reducing[[:space:]]Hubness[[:space:]]and[[:space:]]Improving[[:space:]]Transductive[[:space:]]Few-Shot[[:space:]]Learning[[:space:]]With[[:space:]]Hyperspherical[[:space:]]Embeddings/ecdd0353-f895-4695-b19a-6dacd597a180_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Human[[:space:]]Body[[:space:]]Shape[[:space:]]Completion[[:space:]]With[[:space:]]Implicit[[:space:]]Shape[[:space:]]and[[:space:]]Flow[[:space:]]Learning/6439876b-9981-4006-8266-c65ed29d5889_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Human[[:space:]]Guided[[:space:]]Ground-Truth[[:space:]]Generation[[:space:]]for[[:space:]]Realistic[[:space:]]Image[[:space:]]Super-Resolution/3790a511-54be-4720-a8d5-ceac9d963af1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Human[[:space:]]Pose[[:space:]]As[[:space:]]Compositional[[:space:]]Tokens/aa759f39-ae1e-41a3-a037-89ca900f8dee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]in[[:space:]]Extremely[[:space:]]Low-Light[[:space:]]Conditions/4263da48-7ba1-4189-ab3d-59469cf96f74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Human-Art_[[:space:]]A[[:space:]]Versatile[[:space:]]Human-Centric[[:space:]]Dataset[[:space:]]Bridging[[:space:]]Natural[[:space:]]and[[:space:]]Artificial[[:space:]]Scenes/84a1992c-6878-46b7-94e0-5499ebed3235_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HumanBench_[[:space:]]Towards[[:space:]]General[[:space:]]Human-Centric[[:space:]]Perception[[:space:]]With[[:space:]]Projector[[:space:]]Assisted[[:space:]]Pretraining/1a0cb459-1a3a-46da-99ae-f3721a39309c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HumanGen_[[:space:]]Generating[[:space:]]Human[[:space:]]Radiance[[:space:]]Fields[[:space:]]With[[:space:]]Explicit[[:space:]]Priors/69464b5e-91c0-4050-adb1-f5fa7b785cf9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Humans[[:space:]]As[[:space:]]Light[[:space:]]Bulbs_[[:space:]]3D[[:space:]]Human[[:space:]]Reconstruction[[:space:]]From[[:space:]]Thermal[[:space:]]Reflection/bae8edbe-3131-4b22-b081-c789d02fab37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hunting[[:space:]]Sparsity_[[:space:]]Density-Guided[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation/8c29d191-10a8-459e-a148-be59ee91c043_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hybrid[[:space:]]Active[[:space:]]Learning[[:space:]]via[[:space:]]Deep[[:space:]]Clustering[[:space:]]for[[:space:]]Video[[:space:]]Action[[:space:]]Detection/a8a112e6-a4fe-4a0b-a6ad-f7af3ac395cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hybrid[[:space:]]Neural[[:space:]]Rendering[[:space:]]for[[:space:]]Large-Scale[[:space:]]Scenes[[:space:]]With[[:space:]]Motion[[:space:]]Blur/51b16ea6-4291-481a-8498-5418b8030b46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HypLiLoc_[[:space:]]Towards[[:space:]]Effective[[:space:]]LiDAR[[:space:]]Pose[[:space:]]Regression[[:space:]]With[[:space:]]Hyperbolic[[:space:]]Fusion/2db3674a-486f-43a8-8b66-44511592eeb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HyperCUT_[[:space:]]Video[[:space:]]Sequence[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Blurry[[:space:]]Image[[:space:]]Using[[:space:]]Unsupervised[[:space:]]Ordering/18e04229-5465-4907-a12f-843622abd9ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HyperMatch_[[:space:]]Noise-Tolerant[[:space:]]Semi-Supervised[[:space:]]Learning[[:space:]]via[[:space:]]Relaxed[[:space:]]Contrastive[[:space:]]Constraint/ecf81ad9-b8ab-4223-ac65-5d2053f580ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HyperReel_[[:space:]]High-Fidelity[[:space:]]6-DoF[[:space:]]Video[[:space:]]With[[:space:]]Ray-Conditioned[[:space:]]Sampling/e2850493-e7aa-4f0b-bde7-8e913cbbd3dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hyperbolic[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Visual[[:space:]]Representations[[:space:]]Beyond[[:space:]]Objects/54b10707-34b7-45d6-a834-f66035d29499_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hyperspherical[[:space:]]Embedding[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion/56ff3a9e-20dd-4b07-9d66-cf724eeaec63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/I2-SDF_[[:space:]]Intrinsic[[:space:]]Indoor[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]and[[:space:]]Editing[[:space:]]via[[:space:]]Raytracing[[:space:]]in[[:space:]]Neural[[:space:]]SDFs/5de54e7a-4d2b-4c06-8e96-13c21af21603_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/I2MVFormer_[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]Generated[[:space:]]Multi-View[[:space:]]Document[[:space:]]Supervision[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Image[[:space:]]Classification/fc9bc043-3ba0-40d8-8097-ac09846fea31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IDGI_[[:space:]]A[[:space:]]Framework[[:space:]]To[[:space:]]Eliminate[[:space:]]Explanation[[:space:]]Noise[[:space:]]From[[:space:]]Integrated[[:space:]]Gradients/cb11e2a9-b2ee-44b7-99a6-1ef3bc953ec3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IFSeg_[[:space:]]Image-Free[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]via[[:space:]]Vision-Language[[:space:]]Model/10342523-2c3e-4e4a-8ce5-7b0943e2b286_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IMP_[[:space:]]Iterative[[:space:]]Matching[[:space:]]and[[:space:]]Pose[[:space:]]Estimation[[:space:]]With[[:space:]]Adaptive[[:space:]]Pooling/52dd14c2-dcc5-48a9-a18e-5226c4171479_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IPCC-TP_[[:space:]]Utilizing[[:space:]]Incremental[[:space:]]Pearson[[:space:]]Correlation[[:space:]]Coefficient[[:space:]]for[[:space:]]Joint[[:space:]]Multi-Agent[[:space:]]Trajectory[[:space:]]Prediction/fe316754-ed7b-4ccb-9bb9-dce5f41ea3ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IS-GGT_[[:space:]]Iterative[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]With[[:space:]]Generative[[:space:]]Transformers/46645ee6-9511-4812-ad1e-91a0b124dfde_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ISBNet_[[:space:]]A[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Instance[[:space:]]Segmentation[[:space:]]Network[[:space:]]With[[:space:]]Instance-Aware[[:space:]]Sampling[[:space:]]and[[:space:]]Box-Aware[[:space:]]Dynamic[[:space:]]Convolution/b7d032a2-4153-4d10-8654-e5ba9dd6b9b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Identity-Preserving[[:space:]]Talking[[:space:]]Face[[:space:]]Generation[[:space:]]With[[:space:]]Landmark[[:space:]]and[[:space:]]Appearance[[:space:]]Priors/e60e94c3-78c5-44c1-80fa-7fb97af7142c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Im2Hands_[[:space:]]Learning[[:space:]]Attentive[[:space:]]Implicit[[:space:]]Representation[[:space:]]of[[:space:]]Interacting[[:space:]]Two-Hand[[:space:]]Shapes/df9f3320-e5e1-41b1-bc67-966a216589c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Image[[:space:]]Cropping[[:space:]]With[[:space:]]Spatial-Aware[[:space:]]Feature[[:space:]]and[[:space:]]Rank[[:space:]]Consistency/aec9641b-baa6-4acd-9798-245f436ba266_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Image[[:space:]]Quality-Aware[[:space:]]Diagnosis[[:space:]]via[[:space:]]Meta-Knowledge[[:space:]]Co-Embedding/54a81137-c9c6-477e-8115-f39c919df19c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Image[[:space:]]Super-Resolution[[:space:]]Using[[:space:]]T-Tetromino[[:space:]]Pixels/7e878c58-c847-431e-8e56-a58b2eb703e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Image[[:space:]]as[[:space:]]a[[:space:]]Foreign[[:space:]]Language_[[:space:]]BEiT[[:space:]]Pretraining[[:space:]]for[[:space:]]Vision[[:space:]]and[[:space:]]Vision-Language[[:space:]]Tasks/1d9e8d09-fde4-41b2-87f0-86c54c9ece47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ImageBind_[[:space:]]One[[:space:]]Embedding[[:space:]]Space[[:space:]]To[[:space:]]Bind[[:space:]]Them[[:space:]]All/1bbbd773-5bf3-4684-9a07-908da7761a18_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ImageNet-E_[[:space:]]Benchmarking[[:space:]]Neural[[:space:]]Network[[:space:]]Robustness[[:space:]]via[[:space:]]Attribute[[:space:]]Editing/172f3027-d7b0-4c5b-aa43-f80b62871657_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Imagen[[:space:]]Editor[[:space:]]and[[:space:]]EditBench_[[:space:]]Advancing[[:space:]]and[[:space:]]Evaluating[[:space:]]Text-Guided[[:space:]]Image[[:space:]]Inpainting/156efcc5-b9b3-4083-bcfd-2e81068c27b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Images[[:space:]]Speak[[:space:]]in[[:space:]]Images_[[:space:]]A[[:space:]]Generalist[[:space:]]Painter[[:space:]]for[[:space:]]In-Context[[:space:]]Visual[[:space:]]Learning/7d9197bb-465d-4182-9e13-6a1b6aae9814_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Imagic_[[:space:]]Text-Based[[:space:]]Real[[:space:]]Image[[:space:]]Editing[[:space:]]With[[:space:]]Diffusion[[:space:]]Models/e2f1a0e0-e454-40ad-9a24-9432a408789e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Imitation[[:space:]]Learning[[:space:]]As[[:space:]]State[[:space:]]Matching[[:space:]]via[[:space:]]Differentiable[[:space:]]Physics/bb1a2983-8570-4172-9bc5-5e7066d992de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Implicit[[:space:]]3D[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery[[:space:]]Using[[:space:]]Consistency[[:space:]]With[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]From[[:space:]]Unseen-View/02a74418-bde6-4175-a0ac-eacc4ed96ada_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Implicit[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Continuous[[:space:]]Super-Resolution/018c04b8-6295-4cf0-8b07-1d11f7376007_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Implicit[[:space:]]Identity[[:space:]]Driven[[:space:]]Deepfake[[:space:]]Face[[:space:]]Swapping[[:space:]]Detection/f6978cae-20aa-44fc-8088-7dbc2be8447c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Implicit[[:space:]]Identity[[:space:]]Leakage_[[:space:]]The[[:space:]]Stumbling[[:space:]]Block[[:space:]]to[[:space:]]Improving[[:space:]]Deepfake[[:space:]]Detection[[:space:]]Generalization/eb6d2bc8-db9f-4ab4-9f5b-cf000785f758_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Implicit[[:space:]]Neural[[:space:]]Head[[:space:]]Synthesis[[:space:]]via[[:space:]]Controllable[[:space:]]Local[[:space:]]Deformation[[:space:]]Fields/e6cdbeea-eae3-44e3-9a0b-ade835b7ccd9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Implicit[[:space:]]Occupancy[[:space:]]Flow[[:space:]]Fields[[:space:]]for[[:space:]]Perception[[:space:]]and[[:space:]]Prediction[[:space:]]in[[:space:]]Self-Driving/1888efa8-e52c-4630-8394-f78b32a580ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Implicit[[:space:]]Surface[[:space:]]Contrastive[[:space:]]Clustering[[:space:]]for[[:space:]]LiDAR[[:space:]]Point[[:space:]]Clouds/c7bb8f6e-ceb9-4aaf-ac6d-78156b5f35a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Implicit[[:space:]]View-Time[[:space:]]Interpolation[[:space:]]of[[:space:]]Stereo[[:space:]]Videos[[:space:]]Using[[:space:]]Multi-Plane[[:space:]]Disparities[[:space:]]and[[:space:]]Non-Uniform[[:space:]]Coordinates/7c0e378e-4ac6-481e-8cc0-e33e342ed0cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improved[[:space:]]Distribution[[:space:]]Matching[[:space:]]for[[:space:]]Dataset[[:space:]]Condensation/9cfe6898-90d0-4352-b4d6-4c7eea2344b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improved[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/54da48e3-4e44-4021-91e3-1b5df0aea393_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Commonsense[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models[[:space:]]via[[:space:]]Knowledge[[:space:]]Graph[[:space:]]Riddles/212e9be8-70c0-4eeb-adf2-5094d71a6803_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Cross-Modal[[:space:]]Retrieval[[:space:]]With[[:space:]]Set[[:space:]]of[[:space:]]Diverse[[:space:]]Embeddings/4d6eaca3-6d5d-4391-8f25-19f25013f0ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Fairness[[:space:]]in[[:space:]]Facial[[:space:]]Albedo[[:space:]]Estimation[[:space:]]via[[:space:]]Visual-Textual[[:space:]]Cues/4608f575-e5be-48a5-be89-eebc94bd94ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Generalization[[:space:]]With[[:space:]]Domain[[:space:]]Convex[[:space:]]Game/7f80ea84-50ea-4ee4-b212-3c18894bc2b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Generalization[[:space:]]of[[:space:]]Meta-Learning[[:space:]]With[[:space:]]Inverted[[:space:]]Regularization[[:space:]]at[[:space:]]Inner-Level/86a82c8b-6779-4dc2-bce7-f25ac2eb6d6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Graph[[:space:]]Representation[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation[[:space:]]via[[:space:]]Attentive[[:space:]]Filtering/d56b4a8c-bc3f-49f5-9828-2c99713e2b55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Image[[:space:]]Recognition[[:space:]]by[[:space:]]Retrieving[[:space:]]From[[:space:]]Web-Scale[[:space:]]Image-Text[[:space:]]Data/0ded306f-5523-452c-981c-e3cc280337e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Robust[[:space:]]Generalization[[:space:]]by[[:space:]]Direct[[:space:]]PAC-Bayesian[[:space:]]Bound[[:space:]]Minimization/70018485-afca-4088-9d2c-67c72db458c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Robustness[[:space:]]of[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]to[[:space:]]Motion-Blur[[:space:]]Using[[:space:]]Class-Centric[[:space:]]Augmentation/a282fd27-0ee0-4d2d-9a3a-127af7ba9239_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Robustness[[:space:]]of[[:space:]]Vision[[:space:]]Transformers[[:space:]]by[[:space:]]Reducing[[:space:]]Sensitivity[[:space:]]To[[:space:]]Patch[[:space:]]Corruptions/3a77ede9-54fb-4078-9559-939c79433d57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Selective[[:space:]]Visual[[:space:]]Question[[:space:]]Answering[[:space:]]by[[:space:]]Learning[[:space:]]From[[:space:]]Your[[:space:]]Peers/50c7f16f-45ec-4252-a9b2-8e06d544fb8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Table[[:space:]]Structure[[:space:]]Recognition[[:space:]]With[[:space:]]Visual-Alignment[[:space:]]Sequential[[:space:]]Coordinate[[:space:]]Modeling/fc2265f5-9ab2-481d-87b6-0731e4a92992_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Vision-and-Language[[:space:]]Navigation[[:space:]]by[[:space:]]Generating[[:space:]]Future-View[[:space:]]Image[[:space:]]Semantics/be8a6192-e0f9-4c27-a6f0-1af17d601b0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Visual[[:space:]]Grounding[[:space:]]by[[:space:]]Encouraging[[:space:]]Consistent[[:space:]]Gradient-Based[[:space:]]Explanations/4b8f00a7-285b-4329-86a5-25d33454c1e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning[[:space:]]Through[[:space:]]Perceptual[[:space:]]Understanding/ebe1374c-d283-4296-bc5d-a86eceab16ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization[[:space:]]by[[:space:]]Bridging[[:space:]]Train-Test[[:space:]]Gap[[:space:]]in[[:space:]]Pseudo[[:space:]]Labels/002ec960-a59a-4f43-b5ba-c2d654be20fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Zero-Shot[[:space:]]Generalization[[:space:]]and[[:space:]]Robustness[[:space:]]of[[:space:]]Multi-Modal[[:space:]]Models/a40cb1ef-03ce-4c5d-b241-8cda4bc6eddd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]the[[:space:]]Transferability[[:space:]]of[[:space:]]Adversarial[[:space:]]Samples[[:space:]]by[[:space:]]Path-Augmented[[:space:]]Method/3cee8eb6-20d0-4b99-abc3-8a5c4cade606_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/In-Hand[[:space:]]3D[[:space:]]Object[[:space:]]Scanning[[:space:]]From[[:space:]]an[[:space:]]RGB[[:space:]]Sequence/da2913eb-6794-4825-9c91-f0287b6877a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Incremental[[:space:]]3D[[:space:]]Semantic[[:space:]]Scene[[:space:]]Graph[[:space:]]Prediction[[:space:]]From[[:space:]]RGB[[:space:]]Sequences/99c76b76-968d-4e22-9108-5e25f532e232_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Incrementer_[[:space:]]Transformer[[:space:]]for[[:space:]]Class-Incremental[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]With[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]Focusing[[:space:]]on[[:space:]]Old[[:space:]]Class/0f64f779-de99-41f0-9cc8-bb390a9f705e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Independent[[:space:]]Component[[:space:]]Alignment[[:space:]]for[[:space:]]Multi-Task[[:space:]]Learning/7ff07d03-a610-4a47-af90-6efb8e17ed9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Indescribable[[:space:]]Multi-Modal[[:space:]]Spatial[[:space:]]Evaluator/fc5f01df-5dea-4e88-b9bf-4aaccd6d7caf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Indiscernible[[:space:]]Object[[:space:]]Counting[[:space:]]in[[:space:]]Underwater[[:space:]]Scenes/3b92dac7-330b-4774-b446-fe6ba521a12d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Inferring[[:space:]]and[[:space:]]Leveraging[[:space:]]Parts[[:space:]]From[[:space:]]Object[[:space:]]Shape[[:space:]]for[[:space:]]Improving[[:space:]]Semantic[[:space:]]Image[[:space:]]Synthesis/8dc46856-75dc-4bd2-ae09-ea8ae5b210ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Infinite[[:space:]]Photorealistic[[:space:]]Worlds[[:space:]]Using[[:space:]]Procedural[[:space:]]Generation/8d1e34f1-c712-47ba-846d-53e57ca041f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ingredient-Oriented[[:space:]]Multi-Degradation[[:space:]]Learning[[:space:]]for[[:space:]]Image[[:space:]]Restoration/c5030ec9-d41c-4c34-aa34-bb0732f75de7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Initialization[[:space:]]Noise[[:space:]]in[[:space:]]Image[[:space:]]Gradients[[:space:]]and[[:space:]]Saliency[[:space:]]Maps/afb065b7-6465-442f-afc5-798194dd4075_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/InstMove_[[:space:]]Instance[[:space:]]Motion[[:space:]]for[[:space:]]Object-Centric[[:space:]]Video[[:space:]]Segmentation/b122901d-d995-4000-ab46-e44cc4a86152_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Instance[[:space:]]Relation[[:space:]]Graph[[:space:]]Guided[[:space:]]Source-Free[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detection/f2689d5f-44fd-4337-a255-1e71e9f925cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Instance-Aware[[:space:]]Domain[[:space:]]Generalization[[:space:]]for[[:space:]]Face[[:space:]]Anti-Spoofing/8cfb2ae5-fa9d-4955-8e2f-7aa50f0ed16f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Instance-Specific[[:space:]]and[[:space:]]Model-Adaptive[[:space:]]Supervision[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation/d0e728fa-f969-4c73-bfac-4ea31ebb865f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Instant[[:space:]]Domain[[:space:]]Augmentation[[:space:]]for[[:space:]]LiDAR[[:space:]]Semantic[[:space:]]Segmentation/4cf0ddb6-ca80-41d1-8e06-68936f0c30db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Instant[[:space:]]Multi-View[[:space:]]Head[[:space:]]Capture[[:space:]]Through[[:space:]]Learnable[[:space:]]Registration/d74edd43-454c-4f5f-a725-65069c15f0f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Instant[[:space:]]Volumetric[[:space:]]Head[[:space:]]Avatars/703ce35c-0736-4559-b091-70cf5d3b19cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Instant-NVR_[[:space:]]Instant[[:space:]]Neural[[:space:]]Volumetric[[:space:]]Rendering[[:space:]]for[[:space:]]Human-Object[[:space:]]Interactions[[:space:]]From[[:space:]]Monocular[[:space:]]RGBD[[:space:]]Stream/7124606c-b0ab-4490-8f6b-13bc6a1e0190_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/InstantAvatar_[[:space:]]Learning[[:space:]]Avatars[[:space:]]From[[:space:]]Monocular[[:space:]]Video[[:space:]]in[[:space:]]60[[:space:]]Seconds/ab92da4b-faa3-423d-a600-fd46be2cf1e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/InstructPix2Pix_[[:space:]]Learning[[:space:]]To[[:space:]]Follow[[:space:]]Image[[:space:]]Editing[[:space:]]Instructions/87306804-77d1-461c-be15-ffbc2cee9ee3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Integral[[:space:]]Neural[[:space:]]Networks/b5efbf27-bfb4-4b26-9456-3e0066637c4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Integrally[[:space:]]Pre-Trained[[:space:]]Transformer[[:space:]]Pyramid[[:space:]]Networks/384c1998-69b1-443b-b43d-9f064d8c6bde_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Interactive[[:space:]]Cartoonization[[:space:]]With[[:space:]]Controllable[[:space:]]Perceptual[[:space:]]Factors/98f99240-08fa-4416-a391-89289cbee820_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Interactive[[:space:]]Segmentation[[:space:]]As[[:space:]]Gaussion[[:space:]]Process[[:space:]]Classification/977fb06b-4756-4c5a-8808-57d27a036221_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Interactive[[:space:]]Segmentation[[:space:]]of[[:space:]]Radiance[[:space:]]Fields/f779c377-ef28-4473-ab0c-e6e3f68475e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Interactive[[:space:]]and[[:space:]]Explainable[[:space:]]Region-Guided[[:space:]]Radiology[[:space:]]Report[[:space:]]Generation/e36c846b-7654-4609-bd0c-09379a7f03c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/InternImage_[[:space:]]Exploring[[:space:]]Large-Scale[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models[[:space:]]With[[:space:]]Deformable[[:space:]]Convolutions/837302ce-f48f-46e6-b376-150cb63c5474_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Interventional[[:space:]]Bag[[:space:]]Multi-Instance[[:space:]]Learning[[:space:]]on[[:space:]]Whole-Slide[[:space:]]Pathological[[:space:]]Images/dedfddb7-4659-4105-af67-a6b38e9f3f73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Intrinsic[[:space:]]Physical[[:space:]]Concepts[[:space:]]Discovery[[:space:]]With[[:space:]]Object-Centric[[:space:]]Predictive[[:space:]]Models/973d2848-95a7-4066-90ea-aaa311d539e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Introducing[[:space:]]Competition[[:space:]]To[[:space:]]Boost[[:space:]]the[[:space:]]Transferability[[:space:]]of[[:space:]]Targeted[[:space:]]Adversarial[[:space:]]Examples[[:space:]]Through[[:space:]]Clean[[:space:]]Feature[[:space:]]Mixup/1c865984-1560-4a01-bcc7-f1d4ec6fae0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Inverse[[:space:]]Rendering[[:space:]]of[[:space:]]Translucent[[:space:]]Objects[[:space:]]Using[[:space:]]Physical[[:space:]]and[[:space:]]Neural[[:space:]]Renderers/9dc01f30-f6e4-4cae-9a01-cc94073ffba3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Inversion-Based[[:space:]]Style[[:space:]]Transfer[[:space:]]With[[:space:]]Diffusion[[:space:]]Models/58c0a983-36b0-4d08-a373-2f1dbcab7480_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Invertible[[:space:]]Neural[[:space:]]Skinning/4c4d3853-bd1a-4f63-a8b9-8b515d9a095f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Inverting[[:space:]]the[[:space:]]Imaging[[:space:]]Process[[:space:]]by[[:space:]]Learning[[:space:]]an[[:space:]]Implicit[[:space:]]Camera[[:space:]]Model/6baeb3a6-48b8-4497-90fc-33a7851050d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Is[[:space:]]BERT[[:space:]]Blind_[[:space:]]Exploring[[:space:]]the[[:space:]]Effect[[:space:]]of[[:space:]]Vision-and-Language[[:space:]]Pretraining[[:space:]]on[[:space:]]Visual[[:space:]]Language[[:space:]]Understanding/1fdb6d1e-b4c3-4373-801c-511c79e049ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Iterative[[:space:]]Geometry[[:space:]]Encoding[[:space:]]Volume[[:space:]]for[[:space:]]Stereo[[:space:]]Matching/f69a64e4-01c8-4e22-bcb3-002953a948fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Iterative[[:space:]]Next[[:space:]]Boundary[[:space:]]Detection[[:space:]]for[[:space:]]Instance[[:space:]]Segmentation[[:space:]]of[[:space:]]Tree[[:space:]]Rings[[:space:]]in[[:space:]]Microscopy[[:space:]]Images[[:space:]]of[[:space:]]Shrub[[:space:]]Cross[[:space:]]Sections/9d2bfc37-8fa1-4682-b121-78ffd9b4d896_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Iterative[[:space:]]Proposal[[:space:]]Refinement[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Video[[:space:]]Grounding/82acf587-bd53-4151-92a2-25d13d003a51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Iterative[[:space:]]Vision-and-Language[[:space:]]Navigation/cbd9e383-8018-4585-b7d2-0b5f9d45d619_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IterativePFN_[[:space:]]True[[:space:]]Iterative[[:space:]]Point[[:space:]]Cloud[[:space:]]Filtering/9c78fc39-01a6-4567-a0ab-839e731decd3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/JAWS_[[:space:]]Just[[:space:]]a[[:space:]]Wild[[:space:]]Shot[[:space:]]for[[:space:]]Cinematic[[:space:]]Transfer[[:space:]]in[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/1ba6e2ae-8328-4100-b7a0-531e2281fa76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/JRDB-Pose_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]for[[:space:]]Multi-Person[[:space:]]Pose[[:space:]]Estimation[[:space:]]and[[:space:]]Tracking/abc08de0-7a29-4c5d-b910-39d934620534_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/JacobiNeRF_[[:space:]]NeRF[[:space:]]Shaping[[:space:]]With[[:space:]]Mutual[[:space:]]Information[[:space:]]Gradients/49fd79ab-c98a-42dc-ae0a-3b199d4c130d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Jedi_[[:space:]]Entropy-Based[[:space:]]Localization[[:space:]]and[[:space:]]Removal[[:space:]]of[[:space:]]Adversarial[[:space:]]Patches/b5e586be-6a47-4d75-87b4-e60015fc48fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Joint[[:space:]]Appearance[[:space:]]and[[:space:]]Motion[[:space:]]Learning[[:space:]]for[[:space:]]Efficient[[:space:]]Rolling[[:space:]]Shutter[[:space:]]Correction/b278f23e-30e0-416e-a63b-4d887dba0c9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Joint[[:space:]]HDR[[:space:]]Denoising[[:space:]]and[[:space:]]Fusion_[[:space:]]A[[:space:]]Real-World[[:space:]]Mobile[[:space:]]HDR[[:space:]]Image[[:space:]]Dataset/cccac156-2014-45f9-8c9a-54fc0f112437_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Joint[[:space:]]Token[[:space:]]Pruning[[:space:]]and[[:space:]]Squeezing[[:space:]]Towards[[:space:]]More[[:space:]]Aggressive[[:space:]]Compression[[:space:]]of[[:space:]]Vision[[:space:]]Transformers/aec0f9db-74ca-4e91-a256-a370732ec7db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Joint[[:space:]]Video[[:space:]]Multi-Frame[[:space:]]Interpolation[[:space:]]and[[:space:]]Deblurring[[:space:]]Under[[:space:]]Unknown[[:space:]]Exposure[[:space:]]Time/d8c97a72-72fc-4f1b-9322-f73f39d769f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Joint[[:space:]]Visual[[:space:]]Grounding[[:space:]]and[[:space:]]Tracking[[:space:]]With[[:space:]]Natural[[:space:]]Language[[:space:]]Specification/4eb040cf-4bee-48aa-9c5a-a8d9445557da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/K-Planes_[[:space:]]Explicit[[:space:]]Radiance[[:space:]]Fields[[:space:]]in[[:space:]]Space,[[:space:]]Time,[[:space:]]and[[:space:]]Appearance/dbd83f59-78bf-46ee-9d6d-e80cec11d754_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/K3DN_[[:space:]]Disparity-Aware[[:space:]]Kernel[[:space:]]Estimation[[:space:]]for[[:space:]]Dual-Pixel[[:space:]]Defocus[[:space:]]Deblurring/dd60ffb2-bd49-488b-9d4e-c5511e5a7132_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/KD-DLGAN_[[:space:]]Data[[:space:]]Limited[[:space:]]Image[[:space:]]Generation[[:space:]]via[[:space:]]Knowledge[[:space:]]Distillation/2965d860-c9a1-447a-99a0-07f0e7bf1f9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/KERM_[[:space:]]Knowledge[[:space:]]Enhanced[[:space:]]Reasoning[[:space:]]for[[:space:]]Vision-and-Language[[:space:]]Navigation/150d2327-2a6d-4b82-ad9b-bc44e58eaf56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Kernel[[:space:]]Aware[[:space:]]Resampler/6e202edb-dae4-468e-9b5b-7256547c96e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/KiUT_[[:space:]]Knowledge-Injected[[:space:]]U-Transformer[[:space:]]for[[:space:]]Radiology[[:space:]]Report[[:space:]]Generation/2032ee4e-f798-4e8b-af71-eef5e09e4f68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Knowledge[[:space:]]Combination[[:space:]]To[[:space:]]Learn[[:space:]]Rotated[[:space:]]Detection[[:space:]]Without[[:space:]]Rotated[[:space:]]Annotation/b1d5a4e1-32dc-4a83-96d8-906a1696d485_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation[[:space:]]by[[:space:]]Aligning[[:space:]]Distributions[[:space:]]of[[:space:]]Local[[:space:]]Predictions/a3b91ea3-fb8e-4431-b2eb-5016cba18c15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/L-CoIns_[[:space:]]Language-Based[[:space:]]Colorization[[:space:]]With[[:space:]]Instance[[:space:]]Awareness/217ce38a-9f25-4f9e-ac73-0e133df62e0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LANA_[[:space:]]A[[:space:]]Language-Capable[[:space:]]Navigator[[:space:]]for[[:space:]]Instruction[[:space:]]Following[[:space:]]and[[:space:]]Generation/49634632-c82f-4508-9b73-06d8907c3900_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LANIT_[[:space:]]Language-Driven[[:space:]]Image-to-Image[[:space:]]Translation[[:space:]]for[[:space:]]Unlabeled[[:space:]]Data/58f66285-1e2f-4bd3-8b4e-02783e97cc00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LASP_[[:space:]]Text-to-Text[[:space:]]Optimization[[:space:]]for[[:space:]]Language-Aware[[:space:]]Soft[[:space:]]Prompting[[:space:]]of[[:space:]]Vision[[:space:]]&[[:space:]]Language[[:space:]]Models/fd242174-dc17-4129-bfc3-d467e32f1332_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LAVENDER_[[:space:]]Unifying[[:space:]]Video-Language[[:space:]]Understanding[[:space:]]As[[:space:]]Masked[[:space:]]Language[[:space:]]Modeling/4b1bc9f7-8f79-4b6a-a32b-02935205c197_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LEGO-Net_[[:space:]]Learning[[:space:]]Regular[[:space:]]Rearrangements[[:space:]]of[[:space:]]Objects[[:space:]]in[[:space:]]Rooms/5fe8cc93-fbb1-4458-b609-69e7e159e2bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LEMaRT_[[:space:]]Label-Efficient[[:space:]]Masked[[:space:]]Region[[:space:]]Transform[[:space:]]for[[:space:]]Image[[:space:]]Harmonization/671cb390-5998-4ed7-a320-53fa25cb507c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LG-BPN_[[:space:]]Local[[:space:]]and[[:space:]]Global[[:space:]]Blind-Patch[[:space:]]Network[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Real-World[[:space:]]Denoising/b106948a-bc0f-4862-b17d-1b95904390ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LINe_[[:space:]]Out-of-Distribution[[:space:]]Detection[[:space:]]by[[:space:]]Leveraging[[:space:]]Important[[:space:]]Neurons/7c02f584-1638-4590-ac47-479482d49ca3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LOCATE_[[:space:]]Localize[[:space:]]and[[:space:]]Transfer[[:space:]]Object[[:space:]]Parts[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Affordance[[:space:]]Grounding/c02a6207-738d-48dd-beea-456bd084b930_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LOGO_[[:space:]]A[[:space:]]Long-Form[[:space:]]Video[[:space:]]Dataset[[:space:]]for[[:space:]]Group[[:space:]]Action[[:space:]]Quality[[:space:]]Assessment/a2174fdf-969f-47ee-bc14-05e84961929f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LP-DIF_[[:space:]]Learning[[:space:]]Local[[:space:]]Pattern-Specific[[:space:]]Deep[[:space:]]Implicit[[:space:]]Function[[:space:]]for[[:space:]]3D[[:space:]]Objects[[:space:]]and[[:space:]]Scenes/e23a5c43-ebb5-406b-9d78-64277af6a197_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LSTFE-Net_Long[[:space:]]Short-Term[[:space:]]Feature[[:space:]]Enhancement[[:space:]]Network[[:space:]]for[[:space:]]Video[[:space:]]Small[[:space:]]Object[[:space:]]Detection/f85fdbe6-5d06-483d-a4fa-dfa64b2fcc9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LVQAC_[[:space:]]Lattice[[:space:]]Vector[[:space:]]Quantization[[:space:]]Coupled[[:space:]]With[[:space:]]Spatially[[:space:]]Adaptive[[:space:]]Companding[[:space:]]for[[:space:]]Efficient[[:space:]]Learned[[:space:]]Image[[:space:]]Compression/7db2f3f4-b386-46ae-b0d0-6d14527e76a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Label[[:space:]]Information[[:space:]]Bottleneck[[:space:]]for[[:space:]]Label[[:space:]]Enhancement/17915eba-e6e6-4f83-bad7-de11a7a573e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Label-Free[[:space:]]Liver[[:space:]]Tumor[[:space:]]Segmentation/13c17228-4f2d-4f7f-be2f-fb50e22321a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Language[[:space:]]Adaptive[[:space:]]Weight[[:space:]]Generation[[:space:]]for[[:space:]]Multi-Task[[:space:]]Visual[[:space:]]Grounding/09bf5769-e1b7-4520-b426-6463f773ed2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Language[[:space:]]in[[:space:]]a[[:space:]]Bottle_[[:space:]]Language[[:space:]]Model[[:space:]]Guided[[:space:]]Concept[[:space:]]Bottlenecks[[:space:]]for[[:space:]]Interpretable[[:space:]]Image[[:space:]]Classification/b8f072ca-77d7-4ea7-aa8e-47b49a1b8d4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Language-Guided[[:space:]]Audio-Visual[[:space:]]Source[[:space:]]Separation[[:space:]]via[[:space:]]Trimodal[[:space:]]Consistency/8ed77a3b-731a-4b06-bbf1-882c057c335a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Language-Guided[[:space:]]Music[[:space:]]Recommendation[[:space:]]for[[:space:]]Video[[:space:]]via[[:space:]]Prompt[[:space:]]Analogies/c0941191-f1a3-4b4b-84fc-4169036b56b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Large-Capacity[[:space:]]and[[:space:]]Flexible[[:space:]]Video[[:space:]]Steganography[[:space:]]via[[:space:]]Invertible[[:space:]]Neural[[:space:]]Network/e548c867-6b1a-4988-b551-f9862e60583c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Large-Scale[[:space:]]Training[[:space:]]Data[[:space:]]Search[[:space:]]for[[:space:]]Object[[:space:]]Re-Identification/ba5e79f5-4095-4a3d-aee6-38393b11d227_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LargeKernel3D_[[:space:]]Scaling[[:space:]]Up[[:space:]]Kernels[[:space:]]in[[:space:]]3D[[:space:]]Sparse[[:space:]]CNNs/55787f59-b334-48a8-9764-98d4fb45ca63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LaserMix[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]LiDAR[[:space:]]Semantic[[:space:]]Segmentation/0dfb46c8-457c-4c6b-aa9f-3e7c68a2add5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Latency[[:space:]]Matters_[[:space:]]Real-Time[[:space:]]Action[[:space:]]Forecasting[[:space:]]Transformer/12d94c79-6b2e-4488-a967-716930fdb6d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Latent-NeRF[[:space:]]for[[:space:]]Shape-Guided[[:space:]]Generation[[:space:]]of[[:space:]]3D[[:space:]]Shapes[[:space:]]and[[:space:]]Textures/ca0b263a-3321-4fb3-b039-4fe859baac3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Layout-Based[[:space:]]Causal[[:space:]]Inference[[:space:]]for[[:space:]]Object[[:space:]]Navigation/18995e77-a0be-4475-8773-41d7233fd2aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LayoutDM_[[:space:]]Discrete[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Controllable[[:space:]]Layout[[:space:]]Generation/77614c7d-0145-49cc-9bc8-8f5b07a869f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LayoutDM_[[:space:]]Transformer-Based[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Layout[[:space:]]Generation/5a042bec-4143-4aa8-b201-8f11f7b47f24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LayoutDiffusion_[[:space:]]Controllable[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Layout-to-Image[[:space:]]Generation/b7471e7f-ee4c-4b84-b778-22769e778b1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LayoutFormer++_[[:space:]]Conditional[[:space:]]Graphic[[:space:]]Layout[[:space:]]Generation[[:space:]]via[[:space:]]Constraint[[:space:]]Serialization[[:space:]]and[[:space:]]Decoding[[:space:]]Space[[:space:]]Restriction/b766a11c-0ea4-4e2b-868e-229f60bdc47b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Leapfrog[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Stochastic[[:space:]]Trajectory[[:space:]]Prediction/0cea661d-f5ad-40d3-86cc-3b1fbed56fa4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learnable[[:space:]]Skeleton-Aware[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Sampling/8f02b705-b134-4074-ba5a-caacbe000bbf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learned[[:space:]]Image[[:space:]]Compression[[:space:]]With[[:space:]]Mixed[[:space:]]Transformer-CNN[[:space:]]Architectures/7db49ad4-b6a9-4b40-bd7f-3899bfab9dd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learned[[:space:]]Two-Plane[[:space:]]Perspective[[:space:]]Prior[[:space:]]Based[[:space:]]Image[[:space:]]Resampling[[:space:]]for[[:space:]]Efficient[[:space:]]Object[[:space:]]Detection/875b7e38-656d-435c-a262-1765e6ab04d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]3D[[:space:]]Representations[[:space:]]From[[:space:]]2D[[:space:]]Pre-Trained[[:space:]]Models[[:space:]]via[[:space:]]Image-to-Point[[:space:]]Masked[[:space:]]Autoencoders/89a76a16-b222-401c-8388-b5d09eb2345e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]3D[[:space:]]Scene[[:space:]]Priors[[:space:]]With[[:space:]]2D[[:space:]]Supervision/f5053186-7d13-4536-aa34-2cf3b485f9c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]3D-Aware[[:space:]]Image[[:space:]]Synthesis[[:space:]]With[[:space:]]Unknown[[:space:]]Pose[[:space:]]Distribution/d83e531d-6687-4f90-82b7-f461cb8780cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Accurate[[:space:]]3D[[:space:]]Shape[[:space:]]Based[[:space:]]on[[:space:]]Stereo[[:space:]]Polarimetric[[:space:]]Imaging/44df70c3-8a85-4f37-a0ff-da1913972e67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Action[[:space:]]Changes[[:space:]]by[[:space:]]Measuring[[:space:]]Verb-Adverb[[:space:]]Textual[[:space:]]Relationships/bf106286-0480-42c9-8e91-7fe0ead5cbe8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Adaptive[[:space:]]Dense[[:space:]]Event[[:space:]]Stereo[[:space:]]From[[:space:]]the[[:space:]]Image[[:space:]]Domain/4f451359-7481-4ea3-a57c-f3f3dbdd863e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Analytical[[:space:]]Posterior[[:space:]]Probability[[:space:]]for[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/3ccf2427-6060-4a89-9b0c-48648bd3064c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Anchor[[:space:]]Transformations[[:space:]]for[[:space:]]3D[[:space:]]Garment[[:space:]]Animation/172fc2a1-00b1-413a-bcb3-423ae7376cbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Articulated[[:space:]]Shape[[:space:]]With[[:space:]]Keypoint[[:space:]]Pseudo-Labels[[:space:]]From[[:space:]]Web[[:space:]]Images/e9a40181-29f2-4fab-842a-0bf24d5ccd9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Attention[[:space:]]As[[:space:]]Disentangler[[:space:]]for[[:space:]]Compositional[[:space:]]Zero-Shot[[:space:]]Learning/a88cae1a-24ee-4f75-ab78-cc4922b86fa8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Attribute[[:space:]]and[[:space:]]Class-Specific[[:space:]]Representation[[:space:]]Duet[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Fashion[[:space:]]Analysis/8a754c43-06a8-48b2-bbea-8eaf40d7cf09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Audio-Visual[[:space:]]Source[[:space:]]Localization[[:space:]]via[[:space:]]False[[:space:]]Negative[[:space:]]Aware[[:space:]]Contrastive[[:space:]]Learning/e5b67298-ac49-4b7c-9b93-510776d56f4c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Bottleneck[[:space:]]Concepts[[:space:]]in[[:space:]]Image[[:space:]]Classification/c0f445a3-9476-4f99-8d03-749bf8f8a370_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Common[[:space:]]Rationale[[:space:]]To[[:space:]]Improve[[:space:]]Self-Supervised[[:space:]]Representation[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Visual[[:space:]]Recognition[[:space:]]Problems/f542c58d-1024-4f6c-81dd-99d6fba543cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Compact[[:space:]]Representations[[:space:]]for[[:space:]]LiDAR[[:space:]]Completion[[:space:]]and[[:space:]]Generation/99e8a6b8-3456-496c-a230-7faca7e11fdf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Conditional[[:space:]]Attributes[[:space:]]for[[:space:]]Compositional[[:space:]]Zero-Shot[[:space:]]Learning/c4434e76-0cd3-4a7c-8473-28f0e15837c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Correspondence[[:space:]]Uncertainty[[:space:]]via[[:space:]]Differentiable[[:space:]]Nonlinear[[:space:]]Least[[:space:]]Squares/8f30fa60-e502-4ee5-838a-d9c7713dcf6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Customized[[:space:]]Visual[[:space:]]Models[[:space:]]With[[:space:]]Retrieval-Augmented[[:space:]]Knowledge/5534a4dd-73bd-49cc-aaa8-1125042ab403_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Debiased[[:space:]]Representations[[:space:]]via[[:space:]]Conditional[[:space:]]Attribute[[:space:]]Interpolation/35e0c901-be24-4ae3-9fbb-72a8952235b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Decorrelated[[:space:]]Representations[[:space:]]Efficiently[[:space:]]Using[[:space:]]Fast[[:space:]]Fourier[[:space:]]Transform/2745a979-d548-4430-88ec-679e29ec0301_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Detailed[[:space:]]Radiance[[:space:]]Manifolds[[:space:]]for[[:space:]]High-Fidelity[[:space:]]and[[:space:]]3D-Consistent[[:space:]]Portrait[[:space:]]Synthesis[[:space:]]From[[:space:]]Monocular[[:space:]]Image/ee76865a-ff6b-45fe-83d5-c6e1d7d201e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Discriminative[[:space:]]Representations[[:space:]]for[[:space:]]Skeleton[[:space:]]Based[[:space:]]Action[[:space:]]Recognition/d7b634f0-e078-4d7a-89da-58619960a3b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Distortion[[:space:]]Invariant[[:space:]]Representation[[:space:]]for[[:space:]]Image[[:space:]]Restoration[[:space:]]From[[:space:]]a[[:space:]]Causality[[:space:]]Perspective/bf96eb62-eb0a-4620-9268-36a8377665e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Dynamic[[:space:]]Style[[:space:]]Kernels[[:space:]]for[[:space:]]Artistic[[:space:]]Style[[:space:]]Transfer/d30e13bb-a2dd-4b9a-b639-90ba30ba2cc5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Emotion[[:space:]]Representations[[:space:]]From[[:space:]]Verbal[[:space:]]and[[:space:]]Nonverbal[[:space:]]Communication/73b56492-ee9f-4305-8379-ca00142aa516_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Event[[:space:]]Guided[[:space:]]High[[:space:]]Dynamic[[:space:]]Range[[:space:]]Video[[:space:]]Reconstruction/33cfc19b-aa34-4b73-9613-d3bfaa3bfbaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Expressive[[:space:]]Prompting[[:space:]]With[[:space:]]Residuals[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/4b3a492c-f313-415e-8ace-beb325209ff5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Federated[[:space:]]Visual[[:space:]]Prompt[[:space:]]in[[:space:]]Null[[:space:]]Space[[:space:]]for[[:space:]]MRI[[:space:]]Reconstruction/20f20c2b-43db-4f5b-b2e9-0c30622bc0bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]From[[:space:]]Noisy[[:space:]]Labels[[:space:]]With[[:space:]]Decoupled[[:space:]]Meta[[:space:]]Label[[:space:]]Purifier/39f5cc23-28da-4792-9a48-41c37b920aca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]From[[:space:]]Unique[[:space:]]Perspectives_[[:space:]]User-Aware[[:space:]]Saliency[[:space:]]Modeling/f7bfd1c8-4126-4d0a-afcb-ecafb59d0e63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Generative[[:space:]]Structure[[:space:]]Prior[[:space:]]for[[:space:]]Blind[[:space:]]Text[[:space:]]Image[[:space:]]Super-Resolution/be24eaf6-c383-4655-bacc-0e5ade4d218b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Geometric-Aware[[:space:]]Properties[[:space:]]in[[:space:]]2D[[:space:]]Representation[[:space:]]Using[[:space:]]Lightweight[[:space:]]CAD[[:space:]]Models,[[:space:]]or[[:space:]]Zero[[:space:]]Real[[:space:]]3D[[:space:]]Pairs/17e7e6f9-e779-4537-89d2-408c0b6a8259_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Geometry-Aware[[:space:]]Representations[[:space:]]by[[:space:]]Sketching/f80a6701-d163-4c1c-a5df-2e5aa49a37e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery[[:space:]]in[[:space:]]3D[[:space:]]Scenes/76a14e05-6adf-4600-8709-b0e5b9fb7b40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Human-to-Robot[[:space:]]Handovers[[:space:]]From[[:space:]]Point[[:space:]]Clouds/336c5a3d-faa8-45f9-8aaa-bb35d62132fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Imbalanced[[:space:]]Data[[:space:]]With[[:space:]]Vision[[:space:]]Transformers/2c5c04d7-db27-4f72-9591-822af7466068_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Instance-Level[[:space:]]Representation[[:space:]]for[[:space:]]Large-Scale[[:space:]]Multi-Modal[[:space:]]Pretraining[[:space:]]in[[:space:]]E-Commerce/88606539-5034-4f6c-aa8b-089f9230525d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Joint[[:space:]]Latent[[:space:]]Space[[:space:]]EBM[[:space:]]Prior[[:space:]]Model[[:space:]]for[[:space:]]Multi-Layer[[:space:]]Generator/cb972244-96fa-449d-98c3-61aa444fc92d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Locally[[:space:]]Editable[[:space:]]Virtual[[:space:]]Humans/81a768fa-0f17-4406-9706-ef4fe06ddff4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Multi-Modal[[:space:]]Class-Specific[[:space:]]Tokens[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Dense[[:space:]]Object[[:space:]]Localization/56c83ae8-f628-4d2b-be75-d685154654de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Neural[[:space:]]Duplex[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Real-Time[[:space:]]View[[:space:]]Synthesis/0aecd6d6-eb8c-49cd-ab29-e7461ef9c969_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Neural[[:space:]]Parametric[[:space:]]Head[[:space:]]Models/062fa4bb-0e36-4eeb-adb0-d6de77f80c20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Neural[[:space:]]Proto-Face[[:space:]]Field[[:space:]]for[[:space:]]Disentangled[[:space:]]3D[[:space:]]Face[[:space:]]Modeling[[:space:]]in[[:space:]]the[[:space:]]Wild/5ebd965c-2fcd-48b5-82dd-8be4b32c7eaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Neural[[:space:]]Volumetric[[:space:]]Representations[[:space:]]of[[:space:]]Dynamic[[:space:]]Humans[[:space:]]in[[:space:]]Minutes/4b95dcfc-29a4-4f2f-b770-89d83ba2b62f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]Models[[:space:]]From[[:space:]]Natural[[:space:]]Language[[:space:]]Supervision/c64e9aaf-56b9-414a-803b-d9922f77f018_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Optical[[:space:]]Expansion[[:space:]]From[[:space:]]Scale[[:space:]]Matching/fadba611-fc8a-473b-9570-70ac81007770_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Orthogonal[[:space:]]Prototypes[[:space:]]for[[:space:]]Generalized[[:space:]]Few-Shot[[:space:]]Semantic[[:space:]]Segmentation/f55ad462-d47a-4afc-95bc-6a55486b79b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Partial[[:space:]]Correlation[[:space:]]Based[[:space:]]Deep[[:space:]]Visual[[:space:]]Representation[[:space:]]for[[:space:]]Image[[:space:]]Classification/87245566-23af-421b-80ad-1a8c0309df53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Personalized[[:space:]]High[[:space:]]Quality[[:space:]]Volumetric[[:space:]]Head[[:space:]]Avatars[[:space:]]From[[:space:]]Monocular[[:space:]]RGB[[:space:]]Videos/f99f5e12-9192-46a6-b4e3-a025ac8e0105_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Procedure-Aware[[:space:]]Video[[:space:]]Representation[[:space:]]From[[:space:]]Instructional[[:space:]]Videos[[:space:]]and[[:space:]]Their[[:space:]]Narrations/471fce2b-475c-4777-b3f3-36131b107c8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Rotation-Equivariant[[:space:]]Features[[:space:]]for[[:space:]]Visual[[:space:]]Correspondence/3ac6cb22-128e-410d-8dc0-595aae3be692_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Sample[[:space:]]Relationship[[:space:]]for[[:space:]]Exposure[[:space:]]Correction/328dc83f-b0c4-4ce9-90e2-03cb25bedb51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Semantic[[:space:]]Relationship[[:space:]]Among[[:space:]]Instances[[:space:]]for[[:space:]]Image-Text[[:space:]]Matching/5b93c863-5d73-452f-a25d-f402de582d22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Semantic-Aware[[:space:]]Disentangled[[:space:]]Representation[[:space:]]for[[:space:]]Flexible[[:space:]]3D[[:space:]]Human[[:space:]]Body[[:space:]]Editing/ba9dabaf-077e-4e19-9ae1-c1f2c2a30c7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Semantic-Aware[[:space:]]Knowledge[[:space:]]Guidance[[:space:]]for[[:space:]]Low-Light[[:space:]]Image[[:space:]]Enhancement/23000cae-300d-4cde-a471-90bcf65ddfb8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Situation[[:space:]]Hyper-Graphs[[:space:]]for[[:space:]]Video[[:space:]]Question[[:space:]]Answering/cab54a6b-1009-4dbc-b1ca-0ed1a0963ee9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Spatial-Temporal[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations[[:space:]]for[[:space:]]Event-Guided[[:space:]]Video[[:space:]]Super-Resolution/7f46a567-7a9f-43dd-9afb-c2991cdfbe26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Steerable[[:space:]]Function[[:space:]]for[[:space:]]Efficient[[:space:]]Image[[:space:]]Resampling/067ca41f-5208-49cd-b2f0-82f2850ff9d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Detect[[:space:]]Mirrors[[:space:]]From[[:space:]]Videos[[:space:]]via[[:space:]]Dual[[:space:]]Correspondences/1cca248f-f36a-45a2-94fd-cecf65153065_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Detect[[:space:]]and[[:space:]]Segment[[:space:]]for[[:space:]]Open[[:space:]]Vocabulary[[:space:]]Object[[:space:]]Detection/5ac75f03-1745-4705-ac4d-6c3f7fb370ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Dub[[:space:]]Movies[[:space:]]via[[:space:]]Hierarchical[[:space:]]Prosody[[:space:]]Models/a85158d4-00c7-4fae-9712-36af3c2f8ea8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Exploit[[:space:]]Temporal[[:space:]]Structure[[:space:]]for[[:space:]]Biomedical[[:space:]]Vision-Language[[:space:]]Processing/fc7c021d-c682-4b4a-b7f3-6c9ca632109a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Exploit[[:space:]]the[[:space:]]Sequence-Specific[[:space:]]Prior[[:space:]]Knowledge[[:space:]]for[[:space:]]Image[[:space:]]Processing[[:space:]]Pipelines[[:space:]]Optimization/5043327e-5e28-4a8f-8c57-7b23c08760b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Fuse[[:space:]]Monocular[[:space:]]and[[:space:]]Multi-View[[:space:]]Cues[[:space:]]for[[:space:]]Multi-Frame[[:space:]]Depth[[:space:]]Estimation[[:space:]]in[[:space:]]Dynamic[[:space:]]Scenes/62c5b343-b419-485f-9576-4b167b86494f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Generate[[:space:]]Image[[:space:]]Embeddings[[:space:]]With[[:space:]]User-Level[[:space:]]Differential[[:space:]]Privacy/35eb4863-9483-4c5e-b3f4-41696055c435_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Generate[[:space:]]Language-Supervised[[:space:]]and[[:space:]]Open-Vocabulary[[:space:]]Scene[[:space:]]Graph[[:space:]]Using[[:space:]]Pre-Trained[[:space:]]Visual-Semantic[[:space:]]Space/b669970d-e1b1-4a7a-a8da-18909a4dbc7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Generate[[:space:]]Text-Grounded[[:space:]]Mask[[:space:]]for[[:space:]]Open-World[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]From[[:space:]]Only[[:space:]]Image-Text[[:space:]]Pairs/c9bc34a3-cd10-43e4-9d6d-81c2b8ccce5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Measure[[:space:]]the[[:space:]]Point[[:space:]]Cloud[[:space:]]Reconstruction[[:space:]]Loss[[:space:]]in[[:space:]]a[[:space:]]Representation[[:space:]]Space/0855f478-27d1-47de-b23b-7a901bf12cf2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Name[[:space:]]Classes[[:space:]]for[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Models/784488d3-1965-4a7e-8402-437c5a3d1158_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Predict[[:space:]]Scene-Level[[:space:]]Implicit[[:space:]]3D[[:space:]]From[[:space:]]Posed[[:space:]]RGBD[[:space:]]Data/9e708598-d8b4-40c1-87d1-c7aee00317d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Render[[:space:]]Novel[[:space:]]Views[[:space:]]From[[:space:]]Wide-Baseline[[:space:]]Stereo[[:space:]]Pairs/81b4fe5b-6c46-4bd1-8b70-87cb209929c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Retain[[:space:]]While[[:space:]]Acquiring_[[:space:]]Combating[[:space:]]Distribution-Shift[[:space:]]in[[:space:]]Adversarial[[:space:]]Data-Free[[:space:]]Knowledge[[:space:]]Distillation/b93957f9-1257-4220-880f-a2d772770ee6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Segment[[:space:]]Every[[:space:]]Referring[[:space:]]Object[[:space:]]Point[[:space:]]by[[:space:]]Point/3793df0f-b6b3-448f-85d5-cdd2cf084f15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]To[[:space:]]Zoom[[:space:]]and[[:space:]]Unzoom/87db51f5-d1a6-44d6-b2f5-c2462a285975_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Transferable[[:space:]]Spatiotemporal[[:space:]]Representations[[:space:]]From[[:space:]]Natural[[:space:]]Script[[:space:]]Knowledge/0f53488e-9219-482a-95c2-382c51f54644_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Transformation-Predictive[[:space:]]Representations[[:space:]]for[[:space:]]Detection[[:space:]]and[[:space:]]Description[[:space:]]of[[:space:]]Local[[:space:]]Features/aa5b05da-b5ee-4eea-ae47-f2c476393ca4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Transformations[[:space:]]To[[:space:]]Reduce[[:space:]]the[[:space:]]Geometric[[:space:]]Shift[[:space:]]in[[:space:]]Object[[:space:]]Detection/69fb3432-f3b6-48da-9de8-e985d9696176_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Video[[:space:]]Representations[[:space:]]From[[:space:]]Large[[:space:]]Language[[:space:]]Models/c00bc946-1a6e-40fd-9b0d-ec2826e3ea29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Visibility[[:space:]]Field[[:space:]]for[[:space:]]Detailed[[:space:]]3D[[:space:]]Human[[:space:]]Reconstruction[[:space:]]and[[:space:]]Relighting/ad55c0fb-3962-4c84-be1b-ade77062812f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Visual[[:space:]]Representations[[:space:]]via[[:space:]]Language-Guided[[:space:]]Sampling/c0b46ea1-1183-4a23-a0d4-68578b0842d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Weather-General[[:space:]]and[[:space:]]Weather-Specific[[:space:]]Features[[:space:]]for[[:space:]]Image[[:space:]]Restoration[[:space:]]Under[[:space:]]Multiple[[:space:]]Adverse[[:space:]]Weather[[:space:]]Conditions/af74b836-ff73-4675-8c73-1cc8c534f9c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]With[[:space:]]Fantasy_[[:space:]]Semantic-Aware[[:space:]]Virtual[[:space:]]Contrastive[[:space:]]Constraint[[:space:]]for[[:space:]]Few-Shot[[:space:]]Class-Incremental[[:space:]]Learning/5de50010-4b67-476f-aff9-6ea2c799f91f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]With[[:space:]]Noisy[[:space:]]Labels[[:space:]]via[[:space:]]Self-Supervised[[:space:]]Adversarial[[:space:]]Noisy[[:space:]]Masking/fee95f26-a8a8-4c1d-9f44-c06f33970b8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]a[[:space:]]3D[[:space:]]Morphable[[:space:]]Face[[:space:]]Reflectance[[:space:]]Model[[:space:]]From[[:space:]]Low-Cost[[:space:]]Data/bfac72ff-9a91-49c0-91b2-40f3beec823c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]a[[:space:]]Deep[[:space:]]Color[[:space:]]Difference[[:space:]]Metric[[:space:]]for[[:space:]]Photographic[[:space:]]Images/dbc258de-351a-4a82-83a7-8b31f7a409d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]a[[:space:]]Depth[[:space:]]Covariance[[:space:]]Function/ce47874e-679a-4e09-bfc4-c70c4939d294_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]a[[:space:]]Practical[[:space:]]SDR-to-HDRTV[[:space:]]Up-Conversion[[:space:]]Using[[:space:]]New[[:space:]]Dataset[[:space:]]and[[:space:]]Degradation[[:space:]]Models/ed51a5a5-a3b7-4784-b738-9ddfdcaf0202_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]a[[:space:]]Simple[[:space:]]Low-Light[[:space:]]Image[[:space:]]Enhancer[[:space:]]From[[:space:]]Paired[[:space:]]Low-Light[[:space:]]Instances/443449b8-5cd8-436a-ab74-6d4bb5de9279_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]a[[:space:]]Sparse[[:space:]]Transformer[[:space:]]Network[[:space:]]for[[:space:]]Effective[[:space:]]Image[[:space:]]Deraining/f486121e-86e3-4a45-bbf0-b8f552e4c49f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]and[[:space:]]Aggregating[[:space:]]Lane[[:space:]]Graphs[[:space:]]for[[:space:]]Urban[[:space:]]Automated[[:space:]]Driving/71072545-68f3-478a-8f5b-1abe54803639_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]on[[:space:]]Gradients_[[:space:]]Generalized[[:space:]]Artifacts[[:space:]]Representation[[:space:]]for[[:space:]]GAN-Generated[[:space:]]Images[[:space:]]Detection/2c415e18-250d-42c6-9ecf-17946adf591a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]the[[:space:]]Distribution[[:space:]]of[[:space:]]Errors[[:space:]]in[[:space:]]Stereo[[:space:]]Matching[[:space:]]for[[:space:]]Joint[[:space:]]Disparity[[:space:]]and[[:space:]]Uncertainty[[:space:]]Estimation/52ba4497-4417-4913-a633-6839eab97318_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Less[[:space:]]Is[[:space:]]More_[[:space:]]Reducing[[:space:]]Task[[:space:]]and[[:space:]]Model[[:space:]]Complexity[[:space:]]for[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Semantic[[:space:]]Segmentation/d3f4ed9a-b980-4e20-ba23-ec86d6475569_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Level-S$^2$fM_[[:space:]]Structure[[:space:]]From[[:space:]]Motion[[:space:]]on[[:space:]]Neural[[:space:]]Level[[:space:]]Set[[:space:]]of[[:space:]]Implicit[[:space:]]Surfaces/b319b286-2e10-4d79-8a32-5ac089c156b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Leverage[[:space:]]Interactive[[:space:]]Affinity[[:space:]]for[[:space:]]Affordance[[:space:]]Learning/65bdc10c-8ced-4506-a6bb-43e12b06ea29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Leveraging[[:space:]]Hidden[[:space:]]Positives[[:space:]]for[[:space:]]Unsupervised[[:space:]]Semantic[[:space:]]Segmentation/34b6701a-58b4-43b7-90c9-cd9aa3bd0efc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Leveraging[[:space:]]Inter-Rater[[:space:]]Agreement[[:space:]]for[[:space:]]Classification[[:space:]]in[[:space:]]the[[:space:]]Presence[[:space:]]of[[:space:]]Noisy[[:space:]]Labels/5cf56994-80ce-462c-9dfe-e6178b31a06f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Leveraging[[:space:]]Temporal[[:space:]]Context[[:space:]]in[[:space:]]Low[[:space:]]Representational[[:space:]]Power[[:space:]]Regimes/23cf52f5-5ab9-4d57-929a-b5407b0b21e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Leveraging[[:space:]]per[[:space:]]Image-Token[[:space:]]Consistency[[:space:]]for[[:space:]]Vision-Language[[:space:]]Pre-Training/59c7554d-450f-4872-aa6d-2245cedbd11c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LiDAR-in-the-Loop[[:space:]]Hyperparameter[[:space:]]Optimization/9dfdd95b-2191-44f3-8949-8eb946acc7a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LiDAR2Map_[[:space:]]In[[:space:]]Defense[[:space:]]of[[:space:]]LiDAR-Based[[:space:]]Semantic[[:space:]]Map[[:space:]]Construction[[:space:]]Using[[:space:]]Online[[:space:]]Camera[[:space:]]Distillation/dc5083b3-05fb-4540-b18b-1e5730aae07a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LidarGait_[[:space:]]Benchmarking[[:space:]]3D[[:space:]]Gait[[:space:]]Recognition[[:space:]]With[[:space:]]Point[[:space:]]Clouds/e8d2a9c3-d2ea-4610-932d-6ca5fcc5955e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Lift3D_[[:space:]]Synthesize[[:space:]]3D[[:space:]]Training[[:space:]]Data[[:space:]]by[[:space:]]Lifting[[:space:]]2D[[:space:]]GAN[[:space:]]to[[:space:]]3D[[:space:]]Generative[[:space:]]Radiance[[:space:]]Field/62c3a436-bf26-4084-841d-44671486e8c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Light[[:space:]]Source[[:space:]]Separation[[:space:]]and[[:space:]]Intrinsic[[:space:]]Image[[:space:]]Decomposition[[:space:]]Under[[:space:]]AC[[:space:]]Illumination/4859f9ac-2424-4bdc-90fe-7ade4fe9025d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LightPainter_[[:space:]]Interactive[[:space:]]Portrait[[:space:]]Relighting[[:space:]]With[[:space:]]Freehand[[:space:]]Scribble/2d73af7a-074b-40d7-8aa8-f60d3e2e01fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LightedDepth_[[:space:]]Video[[:space:]]Depth[[:space:]]Estimation[[:space:]]in[[:space:]]Light[[:space:]]of[[:space:]]Limited[[:space:]]Inference[[:space:]]View[[:space:]]Angles/2f557e6f-6fa9-4c32-92d0-1b36cd7e14e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LinK_[[:space:]]Linear[[:space:]]Kernel[[:space:]]for[[:space:]]LiDAR-Based[[:space:]]3D[[:space:]]Perception/dcfdbcd3-5ebe-4867-b43b-d32b4dc91a0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Linking[[:space:]]Garment[[:space:]]With[[:space:]]Person[[:space:]]via[[:space:]]Semantically[[:space:]]Associated[[:space:]]Landmarks[[:space:]]for[[:space:]]Virtual[[:space:]]Try-On/7cf03ab8-2202-4c8b-b8c7-50c76841db33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LipFormer_[[:space:]]High-Fidelity[[:space:]]and[[:space:]]Generalizable[[:space:]]Talking[[:space:]]Face[[:space:]]Generation[[:space:]]With[[:space:]]a[[:space:]]Pre-Learned[[:space:]]Facial[[:space:]]Codebook/006b5d23-7d18-417a-b46d-d8cc19041743_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Listening[[:space:]]Human[[:space:]]Behavior_[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]With[[:space:]]Acoustic[[:space:]]Signals/24c11382-f016-4f61-8dc2-0550c444ff9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Lite[[:space:]]DETR_[[:space:]]An[[:space:]]Interleaved[[:space:]]Multi-Scale[[:space:]]Encoder[[:space:]]for[[:space:]]Efficient[[:space:]]DETR/48c32d17-70d7-4859-bd15-5be9acbfceb9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Lite-Mono_[[:space:]]A[[:space:]]Lightweight[[:space:]]CNN[[:space:]]and[[:space:]]Transformer[[:space:]]Architecture[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/d7829722-945b-4a4e-9115-33dc83ee2e43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LoGoNet_[[:space:]]Towards[[:space:]]Accurate[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]Local-to-Global[[:space:]]Cross-Modal[[:space:]]Fusion/0a164324-cbe5-4662-9833-22b46bd22948_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Local[[:space:]]3D[[:space:]]Editing[[:space:]]via[[:space:]]3D[[:space:]]Distillation[[:space:]]of[[:space:]]CLIP[[:space:]]Knowledge/14ceb3de-1fc3-418b-b44f-0f32a39ccb88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Local[[:space:]]Connectivity-Based[[:space:]]Density[[:space:]]Estimation[[:space:]]for[[:space:]]Face[[:space:]]Clustering/cfb77285-e10d-4cb2-a1ca-424ed0bdce02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Local[[:space:]]Implicit[[:space:]]Normalizing[[:space:]]Flow[[:space:]]for[[:space:]]Arbitrary-Scale[[:space:]]Image[[:space:]]Super-Resolution/03a28b0e-03c4-443a-916c-0b16522ffb14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Local[[:space:]]Implicit[[:space:]]Ray[[:space:]]Function[[:space:]]for[[:space:]]Generalizable[[:space:]]Radiance[[:space:]]Field[[:space:]]Representation/ecff6dbb-1a06-47a1-9ed3-92d44aef1359_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Local-Guided[[:space:]]Global_[[:space:]]Paired[[:space:]]Similarity[[:space:]]Representation[[:space:]]for[[:space:]]Visual[[:space:]]Reinforcement[[:space:]]Learning/c196012a-6e50-4427-9b83-392f68650d3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Local-to-Global[[:space:]]Registration[[:space:]]for[[:space:]]Bundle-Adjusting[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/28e50302-1d15-478c-86c1-f03eebbc2df2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Localized[[:space:]]Semantic[[:space:]]Feature[[:space:]]Mixers[[:space:]]for[[:space:]]Efficient[[:space:]]Pedestrian[[:space:]]Detection[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/6d7dd383-3474-4610-9590-e1bec21111d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Logical[[:space:]]Consistency[[:space:]]and[[:space:]]Greater[[:space:]]Descriptive[[:space:]]Power[[:space:]]for[[:space:]]Facial[[:space:]]Hair[[:space:]]Attribute[[:space:]]Learning/7dc77238-0bdf-4241-b215-099edd10be85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Logical[[:space:]]Implications[[:space:]]for[[:space:]]Visual[[:space:]]Question[[:space:]]Answering[[:space:]]Consistency/3f234ef3-c24b-4e4a-bc41-6849b4132afa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Long[[:space:]]Range[[:space:]]Pooling[[:space:]]for[[:space:]]3D[[:space:]]Large-Scale[[:space:]]Scene[[:space:]]Understanding/16c3aedb-4786-4da0-9c4c-98bf73382fb9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Long-Tailed[[:space:]]Visual[[:space:]]Recognition[[:space:]]via[[:space:]]Self-Heterogeneous[[:space:]]Integration[[:space:]]With[[:space:]]Knowledge[[:space:]]Excavation/ae708f50-646d-4bcf-a290-e6336e3f31a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Long-Term[[:space:]]Visual[[:space:]]Localization[[:space:]]With[[:space:]]Mobile[[:space:]]Sensors/99c61fdb-0362-46b8-9641-2d30f82a7634_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Look[[:space:]]Around[[:space:]]for[[:space:]]Anomalies_[[:space:]]Weakly-Supervised[[:space:]]Anomaly[[:space:]]Detection[[:space:]]via[[:space:]]Context-Motion[[:space:]]Relational[[:space:]]Learning/2c36ab48-e72e-4af5-afc3-8d009f8992ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Look[[:space:]]Before[[:space:]]You[[:space:]]Match_[[:space:]]Instance[[:space:]]Understanding[[:space:]]Matters[[:space:]]in[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/50ce8037-744b-4f22-b531-9415db38fcd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Look,[[:space:]]Radiate,[[:space:]]and[[:space:]]Learn_[[:space:]]Self-Supervised[[:space:]]Localisation[[:space:]]via[[:space:]]Radio-Visual[[:space:]]Correspondence/8573a937-76d6-4072-9210-661d285135a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Lookahead[[:space:]]Diffusion[[:space:]]Probabilistic[[:space:]]Models[[:space:]]for[[:space:]]Refining[[:space:]]Mean[[:space:]]Estimation/07aaacef-90c8-4d92-ac23-d1b2a95181ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Looking[[:space:]]Through[[:space:]]the[[:space:]]Glass_[[:space:]]Neural[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]Against[[:space:]]High[[:space:]]Specular[[:space:]]Reflections/4cb9c368-7254-4001-9e7a-089c667374e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Low-Light[[:space:]]Image[[:space:]]Enhancement[[:space:]]via[[:space:]]Structure[[:space:]]Modeling[[:space:]]and[[:space:]]Guidance/5f3d1b69-d3e5-461a-be7c-06eefb9b4857_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/M6Doc_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Multi-Format,[[:space:]]Multi-Type,[[:space:]]Multi-Layout,[[:space:]]Multi-Language,[[:space:]]Multi-Annotation[[:space:]]Category[[:space:]]Dataset[[:space:]]for[[:space:]]Modern[[:space:]]Document[[:space:]]Layout[[:space:]]Analysis/243b2306-5e4c-4795-abad-10d4c8650f29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MACARONS_[[:space:]]Mapping[[:space:]]and[[:space:]]Coverage[[:space:]]Anticipation[[:space:]]With[[:space:]]RGB[[:space:]]Online[[:space:]]Self-Supervision/e804255e-4435-4902-a22b-a8da7dcef108_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAESTER_[[:space:]]Masked[[:space:]]Autoencoder[[:space:]]Guided[[:space:]]Segmentation[[:space:]]at[[:space:]]Pixel[[:space:]]Resolution[[:space:]]for[[:space:]]Accurate,[[:space:]]Self-Supervised[[:space:]]Subcellular[[:space:]]Structure[[:space:]]Recognition/de9d04d5-47fc-4e71-90a8-ac22c0c921b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAGE_[[:space:]]MAsked[[:space:]]Generative[[:space:]]Encoder[[:space:]]To[[:space:]]Unify[[:space:]]Representation[[:space:]]Learning[[:space:]]and[[:space:]]Image[[:space:]]Synthesis/fcc7133d-e1b6-459f-b8da-52300c8f0f85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAGVIT_[[:space:]]Masked[[:space:]]Generative[[:space:]]Video[[:space:]]Transformer/eeb0e19d-1909-45be-b67f-930b30dd9d8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAGVLT_[[:space:]]Masked[[:space:]]Generative[[:space:]]Vision-and-Language[[:space:]]Transformer/9d17665d-66a1-4054-990a-33054e3b5be2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAIR_[[:space:]]Multi-View[[:space:]]Attention[[:space:]]Inverse[[:space:]]Rendering[[:space:]]With[[:space:]]3D[[:space:]]Spatially-Varying[[:space:]]Lighting[[:space:]]Estimation/fa6c83b8-2e4c-4379-9dbd-afedd0980164_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAP_[[:space:]]Multimodal[[:space:]]Uncertainty-Aware[[:space:]]Vision-Language[[:space:]]Pre-Training[[:space:]]Model/cc12aae9-6e59-4ab7-8656-d69ff4fc7708_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MARLIN_[[:space:]]Masked[[:space:]]Autoencoder[[:space:]]for[[:space:]]Facial[[:space:]]Video[[:space:]]Representation[[:space:]]LearnINg/9542447b-b4c8-4652-a95e-bd212439a27e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MCF_[[:space:]]Mutual[[:space:]]Correction[[:space:]]Framework[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/7cd1d28d-51a3-4703-af93-4761abebdbca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MD-VQA_[[:space:]]Multi-Dimensional[[:space:]]Quality[[:space:]]Assessment[[:space:]]for[[:space:]]UGC[[:space:]]Live[[:space:]]Videos/ce676813-7d3b-4a37-805f-24da5240d5cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MDL-NAS_[[:space:]]A[[:space:]]Joint[[:space:]]Multi-Domain[[:space:]]Learning[[:space:]]Framework[[:space:]]for[[:space:]]Vision[[:space:]]Transformer/f1e329d8-7121-4774-b94d-1fc2cc61b24e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MDQE_[[:space:]]Mining[[:space:]]Discriminative[[:space:]]Query[[:space:]]Embeddings[[:space:]]To[[:space:]]Segment[[:space:]]Occluded[[:space:]]Instances[[:space:]]on[[:space:]]Challenging[[:space:]]Videos/cd3c7b65-cc95-47b1-a519-363475ffb654_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MED-VT_[[:space:]]Multiscale[[:space:]]Encoder-Decoder[[:space:]]Video[[:space:]]Transformer[[:space:]]With[[:space:]]Application[[:space:]]To[[:space:]]Object[[:space:]]Segmentation/2365852c-0f8d-4ecc-8b91-dc677ced66a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MEDIC_[[:space:]]Remove[[:space:]]Model[[:space:]]Backdoors[[:space:]]via[[:space:]]Importance[[:space:]]Driven[[:space:]]Cloning/a46d8363-6eec-4b3b-932c-20c18d5c66b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MEGANE_[[:space:]]Morphable[[:space:]]Eyeglass[[:space:]]and[[:space:]]Avatar[[:space:]]Network/c06e7345-7eb9-4f48-9526-2e3f6c9190c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MELTR_[[:space:]]Meta[[:space:]]Loss[[:space:]]Transformer[[:space:]]for[[:space:]]Learning[[:space:]]To[[:space:]]Fine-Tune[[:space:]]Video[[:space:]]Foundation[[:space:]]Models/5c2c9f06-76a9-452f-b64c-33e7a9622919_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/METransformer_[[:space:]]Radiology[[:space:]]Report[[:space:]]Generation[[:space:]]by[[:space:]]Transformer[[:space:]]With[[:space:]]Multiple[[:space:]]Learnable[[:space:]]Expert[[:space:]]Tokens/48cb8e46-00e8-4285-bad0-eda4ae42b291_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MHPL_[[:space:]]Minimum[[:space:]]Happy[[:space:]]Points[[:space:]]Learning[[:space:]]for[[:space:]]Active[[:space:]]Source[[:space:]]Free[[:space:]]Domain[[:space:]]Adaptation/9b9def4c-aa1a-4b72-85cf-67a24e6d6c0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MIANet_[[:space:]]Aggregating[[:space:]]Unbiased[[:space:]]Instance[[:space:]]and[[:space:]]General[[:space:]]Information[[:space:]]for[[:space:]]Few-Shot[[:space:]]Semantic[[:space:]]Segmentation/0c9beb78-0d59-4d4c-8392-ae1e7faa505f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MIC_[[:space:]]Masked[[:space:]]Image[[:space:]]Consistency[[:space:]]for[[:space:]]Context-Enhanced[[:space:]]Domain[[:space:]]Adaptation/dc0eda45-7f27-4991-be9d-0b19d94991ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MIME_[[:space:]]Human-Aware[[:space:]]3D[[:space:]]Scene[[:space:]]Generation/f0622dfe-a182-4d59-a177-4317f57292aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MISC210K_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]for[[:space:]]Multi-Instance[[:space:]]Semantic[[:space:]]Correspondence/569a5890-0909-4273-a49b-929a202c03a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MIST_[[:space:]]Multi-Modal[[:space:]]Iterative[[:space:]]Spatial-Temporal[[:space:]]Transformer[[:space:]]for[[:space:]]Long-Form[[:space:]]Video[[:space:]]Question[[:space:]]Answering/3f268876-d907-4f69-afd0-585a30dd26a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MM-3DScene_[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding[[:space:]]by[[:space:]]Customizing[[:space:]]Masked[[:space:]]Modeling[[:space:]]With[[:space:]]Informative-Preserved[[:space:]]Reconstruction[[:space:]]and[[:space:]]Self-Distilled[[:space:]]Consistency/18a8ed75-2792-4da0-b6ea-5a12ad6939d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MM-Diffusion_[[:space:]]Learning[[:space:]]Multi-Modal[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Joint[[:space:]]Audio[[:space:]]and[[:space:]]Video[[:space:]]Generation/45235ffe-76fb-41b2-b8fe-bf069d64ba8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MMANet_[[:space:]]Margin-Aware[[:space:]]Distillation[[:space:]]and[[:space:]]Modality-Aware[[:space:]]Regularization[[:space:]]for[[:space:]]Incomplete[[:space:]]Multimodal[[:space:]]Learning/9dd2ad9c-780a-4e2c-bf7b-cf91fcd27bae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MMG-Ego4D_[[:space:]]Multimodal[[:space:]]Generalization[[:space:]]in[[:space:]]Egocentric[[:space:]]Action[[:space:]]Recognition/2a7dcba0-be57-4227-9045-1486356ca14f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MMVC_[[:space:]]Learned[[:space:]]Multi-Mode[[:space:]]Video[[:space:]]Compression[[:space:]]With[[:space:]]Block-Based[[:space:]]Prediction[[:space:]]Mode[[:space:]]Selection[[:space:]]and[[:space:]]Density-Adaptive[[:space:]]Entropy[[:space:]]Coding/30a412ca-a385-49e9-91a5-2169d9e47266_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MOSO_[[:space:]]Decomposing[[:space:]]MOtion,[[:space:]]Scene[[:space:]]and[[:space:]]Object[[:space:]]for[[:space:]]Video[[:space:]]Prediction/04dcb031-2375-47c8-9f5e-23c47e616d8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MOTRv2_[[:space:]]Bootstrapping[[:space:]]End-to-End[[:space:]]Multi-Object[[:space:]]Tracking[[:space:]]by[[:space:]]Pretrained[[:space:]]Object[[:space:]]Detectors/02a94155-5f1f-4c5a-b31e-6e19a4b86117_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MOT_[[:space:]]Masked[[:space:]]Optimal[[:space:]]Transport[[:space:]]for[[:space:]]Partial[[:space:]]Domain[[:space:]]Adaptation/fe7a584b-6d50-46aa-831e-cebc57fd2083_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MOVES_[[:space:]]Manipulated[[:space:]]Objects[[:space:]]in[[:space:]]Video[[:space:]]Enable[[:space:]]Segmentation/dfff92a5-3ca1-43fe-8e97-b9057b1c72e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MP-Former_[[:space:]]Mask-Piloted[[:space:]]Transformer[[:space:]]for[[:space:]]Image[[:space:]]Segmentation/c436cdaa-0279-4efb-be23-f53eeb77c336_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MSF_[[:space:]]Motion-Guided[[:space:]]Sequential[[:space:]]Fusion[[:space:]]for[[:space:]]Efficient[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]From[[:space:]]Point[[:space:]]Cloud[[:space:]]Sequences/e0f58d96-cf61-42b5-9006-0e23915bae11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MSINet_[[:space:]]Twins[[:space:]]Contrastive[[:space:]]Search[[:space:]]of[[:space:]]Multi-Scale[[:space:]]Interaction[[:space:]]for[[:space:]]Object[[:space:]]ReID/5b190a6d-e663-4703-8784-f68b7d4a2470_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MSMDFusion_[[:space:]]Fusing[[:space:]]LiDAR[[:space:]]and[[:space:]]Camera[[:space:]]at[[:space:]]Multiple[[:space:]]Scales[[:space:]]With[[:space:]]Multi-Depth[[:space:]]Seeds[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/2069ca18-1048-442b-b642-22315aec0f29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MSeg3D_[[:space:]]Multi-Modal[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/260fb1ca-a74f-4680-81e2-e4e9fb748eda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MV-JAR_[[:space:]]Masked[[:space:]]Voxel[[:space:]]Jigsaw[[:space:]]and[[:space:]]Reconstruction[[:space:]]for[[:space:]]LiDAR-Based[[:space:]]Self-Supervised[[:space:]]Pre-Training/12c44818-ff06-4c0a-af41-417eda66e48c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MVImgNet_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]of[[:space:]]Multi-View[[:space:]]Images/3e031e8b-eec6-45a0-a28e-e6489370a07d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MaLP_[[:space:]]Manipulation[[:space:]]Localization[[:space:]]Using[[:space:]]a[[:space:]]Proactive[[:space:]]Scheme/f5c5562b-3811-4dcc-b7bd-6a7a14d3f217_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MaPLe_[[:space:]]Multi-Modal[[:space:]]Prompt[[:space:]]Learning/9c2b1d58-6fd3-4d3d-b52c-e47be34ef7bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Magic3D_[[:space:]]High-Resolution[[:space:]]Text-to-3D[[:space:]]Content[[:space:]]Creation/1e2b4976-fbca-4313-8151-9395cc1c1413_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MagicNet_[[:space:]]Semi-Supervised[[:space:]]Multi-Organ[[:space:]]Segmentation[[:space:]]via[[:space:]]Magic-Cube[[:space:]]Partition[[:space:]]and[[:space:]]Recovery/10c556e7-54b4-4db4-bc9b-113fcac4fcbd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MagicPony_[[:space:]]Learning[[:space:]]Articulated[[:space:]]3D[[:space:]]Animals[[:space:]]in[[:space:]]the[[:space:]]Wild/e6926ba7-5fe0-4c65-9f5a-38d28ecebe6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Make[[:space:]]Landscape[[:space:]]Flatter[[:space:]]in[[:space:]]Differentially[[:space:]]Private[[:space:]]Federated[[:space:]]Learning/9c1523d7-4b35-4c98-84fe-da971b2c5a48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Make-a-Story_[[:space:]]Visual[[:space:]]Memory[[:space:]]Conditioned[[:space:]]Consistent[[:space:]]Story[[:space:]]Generation/3ce6b598-c9ba-4565-b293-c61066c7677c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Making[[:space:]]Vision[[:space:]]Transformers[[:space:]]Efficient[[:space:]]From[[:space:]]a[[:space:]]Token[[:space:]]Sparsification[[:space:]]View/0f00c5ea-237b-435f-ba19-f0ae533b96fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MammalNet_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Video[[:space:]]Benchmark[[:space:]]for[[:space:]]Mammal[[:space:]]Recognition[[:space:]]and[[:space:]]Behavior[[:space:]]Understanding/14f4b617-eec8-453d-832d-c91f7b0aa424_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Manipulating[[:space:]]Transfer[[:space:]]Learning[[:space:]]for[[:space:]]Property[[:space:]]Inference/30a4d03f-7b0e-4005-b9ca-7776bedd6658_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mapping[[:space:]]Degeneration[[:space:]]Meets[[:space:]]Label[[:space:]]Evolution_[[:space:]]Learning[[:space:]]Infrared[[:space:]]Small[[:space:]]Target[[:space:]]Detection[[:space:]]With[[:space:]]Single[[:space:]]Point[[:space:]]Supervision/9bbb805e-162c-44dd-853a-dcb209359b00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MarS3D_[[:space:]]A[[:space:]]Plug-and-Play[[:space:]]Motion-Aware[[:space:]]Model[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]on[[:space:]]Multi-Scan[[:space:]]3D[[:space:]]Point[[:space:]]Clouds/a0a4e92d-730a-4b0b-bb00-036e1af8bb9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Marching-Primitives_[[:space:]]Shape[[:space:]]Abstraction[[:space:]]From[[:space:]]Signed[[:space:]]Distance[[:space:]]Function/c067a9c9-d8c5-4894-a294-6c3c0b3f3bd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MarginMatch_[[:space:]]Improving[[:space:]]Semi-Supervised[[:space:]]Learning[[:space:]]with[[:space:]]Pseudo-Margins/1e527083-57e4-4db8-89d4-269dfeae517d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Markerless[[:space:]]Camera-to-Robot[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Self-Supervised[[:space:]]Sim-to-Real[[:space:]]Transfer/7c4fd6ff-03a8-40b6-8677-90772bb2a92c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mask[[:space:]]DINO_[[:space:]]Towards[[:space:]]a[[:space:]]Unified[[:space:]]Transformer-Based[[:space:]]Framework[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]and[[:space:]]Segmentation/df329ef8-a799-4ebf-906c-9403b6e58fb2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mask-Free[[:space:]]OVIS_[[:space:]]Open-Vocabulary[[:space:]]Instance[[:space:]]Segmentation[[:space:]]Without[[:space:]]Manual[[:space:]]Mask[[:space:]]Annotations/d8a0bc48-f3dc-4cd8-b893-b6a4dd0a47ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mask-Free[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation/91f920d3-e27b-4553-9ceb-9dfd929137de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mask-Guided[[:space:]]Matting[[:space:]]in[[:space:]]the[[:space:]]Wild/09ad983f-3200-4b62-8566-4c22c8b72f84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mask3D_[[:space:]]Pre-Training[[:space:]]2D[[:space:]]Vision[[:space:]]Transformers[[:space:]]by[[:space:]]Learning[[:space:]]Masked[[:space:]]3D[[:space:]]Priors/2e3186da-df36-4393-adc4-0fa05e22b0e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MaskCLIP_[[:space:]]Masked[[:space:]]Self-Distillation[[:space:]]Advances[[:space:]]Contrastive[[:space:]]Language-Image[[:space:]]Pretraining/903de9b2-1bcc-4b06-bff8-9b57aa0823cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MaskCon_[[:space:]]Masked[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Coarse-Labelled[[:space:]]Dataset/4e73c1aa-93be-4570-b85f-4e3d7202a290_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MaskSketch_[[:space:]]Unpaired[[:space:]]Structure-Guided[[:space:]]Masked[[:space:]]Image[[:space:]]Generation/73d92261-ba42-4e3c-9334-1f86ad52f64d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Auto-Encoders[[:space:]]Meet[[:space:]]Generative[[:space:]]Adversarial[[:space:]]Networks[[:space:]]and[[:space:]]Beyond/ba6bf709-c33b-4412-be5d-32997f46a850_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Autoencoders[[:space:]]Enable[[:space:]]Efficient[[:space:]]Knowledge[[:space:]]Distillers/37d75dfa-064e-4dc6-9b5a-a491cac2165f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Autoencoding[[:space:]]Does[[:space:]]Not[[:space:]]Help[[:space:]]Natural[[:space:]]Language[[:space:]]Supervision[[:space:]]at[[:space:]]Scale/3ac61577-bf5b-4f86-80cd-9aa136bbba7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Image[[:space:]]Modeling[[:space:]]With[[:space:]]Local[[:space:]]Multi-Scale[[:space:]]Reconstruction/cc2ef537-2f02-42c5-9105-77e1170379e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Image[[:space:]]Training[[:space:]]for[[:space:]]Generalizable[[:space:]]Deep[[:space:]]Image[[:space:]]Denoising/1b515262-1a18-4a38-b492-e80c892d4e84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Images[[:space:]]Are[[:space:]]Counterfactual[[:space:]]Samples[[:space:]]for[[:space:]]Robust[[:space:]]Fine-Tuning/7801f6b3-3ee6-46d7-a157-f81dc90da6f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Jigsaw[[:space:]]Puzzle_[[:space:]]A[[:space:]]Versatile[[:space:]]Position[[:space:]]Embedding[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/e08baece-e2f2-4620-85a3-120ecb725f4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Motion[[:space:]]Encoding[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Video[[:space:]]Representation[[:space:]]Learning/ffb96eff-92ef-423e-9c85-df5684067148_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Domain[[:space:]]Generalized[[:space:]]Stereo[[:space:]]Matching/33890d9e-3bb5-4cca-822b-e87656eea803_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Scene[[:space:]]Contrast_[[:space:]]A[[:space:]]Scalable[[:space:]]Framework[[:space:]]for[[:space:]]Unsupervised[[:space:]]3D[[:space:]]Representation[[:space:]]Learning/5c72a7be-cb6e-40e5-904d-21114e3e1daa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Video[[:space:]]Distillation_[[:space:]]Rethinking[[:space:]]Masked[[:space:]]Feature[[:space:]]Modeling[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Video[[:space:]]Representation[[:space:]]Learning/5401dde5-fc89-4f9e-a63d-583f6d1c59d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Wavelet[[:space:]]Representation[[:space:]]for[[:space:]]Compact[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/45d0202b-5376-4366-b728-470ed1f9b25b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]and[[:space:]]Adaptive[[:space:]]Transformer[[:space:]]for[[:space:]]Exemplar[[:space:]]Based[[:space:]]Image[[:space:]]Translation/122f9675-10c2-4e31-827b-15924283ce1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Master_[[:space:]]Meta[[:space:]]Style[[:space:]]Transformer[[:space:]]for[[:space:]]Controllable[[:space:]]Zero-Shot[[:space:]]and[[:space:]]Few-Shot[[:space:]]Artistic[[:space:]]Style[[:space:]]Transfer/936c6f6e-56a4-479a-88a9-b546f91bc886_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Matching[[:space:]]Is[[:space:]]Not[[:space:]]Enough_[[:space:]]A[[:space:]]Two-Stage[[:space:]]Framework[[:space:]]for[[:space:]]Category-Agnostic[[:space:]]Pose[[:space:]]Estimation/dd58280d-b5a5-4fde-a3e8-4c9ec2674829_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MeMaHand_[[:space:]]Exploiting[[:space:]]Mesh-Mano[[:space:]]Interaction[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Two-Hand[[:space:]]Reconstruction/867425da-f5a0-40aa-92d2-6339df5b377e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Megahertz[[:space:]]Light[[:space:]]Steering[[:space:]]Without[[:space:]]Moving[[:space:]]Parts/dc876064-04d9-4c6f-9ea4-3a0fbd5711b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Memory-Friendly[[:space:]]Scalable[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Rewinding[[:space:]]Lottery[[:space:]]Ticket[[:space:]]Hypothesis/d7f650f5-0e90-4e4e-9e3b-372fb6561fe3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Meta[[:space:]]Architecture[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Analysis/ef212f38-9097-46a0-9adc-735684e16116_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Meta[[:space:]]Compositional[[:space:]]Referring[[:space:]]Expression[[:space:]]Segmentation/9f238d16-af3f-474e-8b36-772f72bf116f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Meta[[:space:]]Omnium_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]General-Purpose[[:space:]]Learning-To-Learn/6a3d596b-7e86-4a45-b2b0-ce1ada0b267e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Meta-Causal[[:space:]]Learning[[:space:]]for[[:space:]]Single[[:space:]]Domain[[:space:]]Generalization/c68d967a-e5fc-4d94-a06a-72b59f93a733_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Meta-Explore_[[:space:]]Exploratory[[:space:]]Hierarchical[[:space:]]Vision-and-Language[[:space:]]Navigation[[:space:]]Using[[:space:]]Scene[[:space:]]Object[[:space:]]Spectrum[[:space:]]Grounding/50db118f-41a0-4ca2-be57-c2ebbcb2698f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Meta-Learning[[:space:]]With[[:space:]]a[[:space:]]Geometry-Adaptive[[:space:]]Preconditioner/79e00a9d-10a7-492d-9f6c-4be9854d55e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Meta-Personalizing[[:space:]]Vision-Language[[:space:]]Models[[:space:]]To[[:space:]]Find[[:space:]]Named[[:space:]]Instances[[:space:]]in[[:space:]]Video/7dfac407-7dc4-48cc-9bb7-f23ca54c68cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Meta-Tuning[[:space:]]Loss[[:space:]]Functions[[:space:]]and[[:space:]]Data[[:space:]]Augmentation[[:space:]]for[[:space:]]Few-Shot[[:space:]]Object[[:space:]]Detection/985b2823-7f93-4fba-8afc-b7a84ab316d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MetaCLUE_[[:space:]]Towards[[:space:]]Comprehensive[[:space:]]Visual[[:space:]]Metaphors[[:space:]]Research/5a097123-23c7-455f-934d-5ab8f12c4f80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MetaFusion_[[:space:]]Infrared[[:space:]]and[[:space:]]Visible[[:space:]]Image[[:space:]]Fusion[[:space:]]via[[:space:]]Meta-Feature[[:space:]]Embedding[[:space:]]From[[:space:]]Object[[:space:]]Detection/70dfb2c3-4648-4819-a764-7cd79a227474_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MetaMix_[[:space:]]Towards[[:space:]]Corruption-Robust[[:space:]]Continual[[:space:]]Learning[[:space:]]With[[:space:]]Temporally[[:space:]]Self-Adaptive[[:space:]]Data[[:space:]]Transformation/350ee3f5-281e-480f-bce3-b1c6caa8838a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MetaPortrait_[[:space:]]Identity-Preserving[[:space:]]Talking[[:space:]]Head[[:space:]]Generation[[:space:]]With[[:space:]]Fast[[:space:]]Personalized[[:space:]]Adaptation/23766ee0-1d0b-472f-8959-8f7e09df80cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MetaViewer_[[:space:]]Towards[[:space:]]a[[:space:]]Unified[[:space:]]Multi-View[[:space:]]Representation/cfe9d4c1-69d9-4f8e-8e6b-a1ece2428418_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Metadata-Based[[:space:]]RAW[[:space:]]Reconstruction[[:space:]]via[[:space:]]Implicit[[:space:]]Neural[[:space:]]Functions/c16d5ffd-61d0-4f71-a138-3e75fa3b8833_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MethaneMapper_[[:space:]]Spectral[[:space:]]Absorption[[:space:]]Aware[[:space:]]Hyperspectral[[:space:]]Transformer[[:space:]]for[[:space:]]Methane[[:space:]]Detection/380575b8-d276-40f4-954b-81a84c333f47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Micron-BERT_[[:space:]]BERT-Based[[:space:]]Facial[[:space:]]Micro-Expression[[:space:]]Recognition/16369e30-f89d-439b-a563-ebb4fc8aacfa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mind[[:space:]]the[[:space:]]Label[[:space:]]Shift[[:space:]]of[[:space:]]Augmentation-Based[[:space:]]Graph[[:space:]]OOD[[:space:]]Generalization/9fd83135-3b8d-4d12-9f4d-d39840145876_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Minimizing[[:space:]]Maximum[[:space:]]Model[[:space:]]Discrepancy[[:space:]]for[[:space:]]Transferable[[:space:]]Black-Box[[:space:]]Targeted[[:space:]]Attacks/d35c7c64-0af3-4ce4-857b-9bec5d6b71ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Minimizing[[:space:]]the[[:space:]]Accumulated[[:space:]]Trajectory[[:space:]]Error[[:space:]]To[[:space:]]Improve[[:space:]]Dataset[[:space:]]Distillation/f8cad635-6f84-4768-b33b-6faa878236ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mitigating[[:space:]]Task[[:space:]]Interference[[:space:]]in[[:space:]]Multi-Task[[:space:]]Learning[[:space:]]via[[:space:]]Explicit[[:space:]]Task[[:space:]]Routing[[:space:]]With[[:space:]]Non-Learnable[[:space:]]Primitives/c83e3fe5-c0bf-407d-87af-1910c2cb2f8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MixMAE_[[:space:]]Mixed[[:space:]]and[[:space:]]Masked[[:space:]]Autoencoder[[:space:]]for[[:space:]]Efficient[[:space:]]Pretraining[[:space:]]of[[:space:]]Hierarchical[[:space:]]Vision[[:space:]]Transformers/faa845fc-9ad9-4cbb-ac71-3f7d519f4d59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MixNeRF_[[:space:]]Modeling[[:space:]]a[[:space:]]Ray[[:space:]]With[[:space:]]Mixture[[:space:]]Density[[:space:]]for[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]From[[:space:]]Sparse[[:space:]]Inputs/43063f53-56c4-4f6f-b25a-6fedb8fe9dfe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MixPHM_[[:space:]]Redundancy-Aware[[:space:]]Parameter-Efficient[[:space:]]Tuning[[:space:]]for[[:space:]]Low-Resource[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/a1be5a15-4e93-4b03-a21f-0168815253ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MixSim_[[:space:]]A[[:space:]]Hierarchical[[:space:]]Framework[[:space:]]for[[:space:]]Mixed[[:space:]]Reality[[:space:]]Traffic[[:space:]]Simulation/d54a3dd5-eb91-4268-be8a-c81e82ad12f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MixTeacher_[[:space:]]Mining[[:space:]]Promising[[:space:]]Labels[[:space:]]With[[:space:]]Mixed[[:space:]]Scale[[:space:]]Teacher[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Object[[:space:]]Detection/76d5d35b-dc0a-40d8-91ec-f27012fce152_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mixed[[:space:]]Autoencoder[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/2fc0db2e-9920-4773-847b-bfa08e70e98f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MoDAR_[[:space:]]Using[[:space:]]Motion[[:space:]]Forecasting[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]Point[[:space:]]Cloud[[:space:]]Sequences/901b7681-c02f-4d98-96d7-a228e4367f90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MoDi_[[:space:]]Unconditional[[:space:]]Motion[[:space:]]Synthesis[[:space:]]From[[:space:]]Diverse[[:space:]]Data/4e037114-7258-4d7b-b9d2-6b5c736d26bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MoLo_[[:space:]]Motion-Augmented[[:space:]]Long-Short[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Few-Shot[[:space:]]Action[[:space:]]Recognition/810f2ad4-9ab9-498b-91a9-75e3dbee4444_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MoStGAN-V_[[:space:]]Video[[:space:]]Generation[[:space:]]With[[:space:]]Temporal[[:space:]]Motion[[:space:]]Styles/1d7ac8bc-34a4-4d16-850a-c0aea3c251d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mobile[[:space:]]User[[:space:]]Interface[[:space:]]Element[[:space:]]Detection[[:space:]]via[[:space:]]Adaptively[[:space:]]Prompt[[:space:]]Tuning/1a65753a-d37a-4792-9385-671bd734388f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MobileBrick_[[:space:]]Building[[:space:]]LEGO[[:space:]]for[[:space:]]3D[[:space:]]Reconstruction[[:space:]]on[[:space:]]Mobile[[:space:]]Devices/4c906a9e-19b5-498b-858f-c243be454add_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MobileNeRF_[[:space:]]Exploiting[[:space:]]the[[:space:]]Polygon[[:space:]]Rasterization[[:space:]]Pipeline[[:space:]]for[[:space:]]Efficient[[:space:]]Neural[[:space:]]Field[[:space:]]Rendering[[:space:]]on[[:space:]]Mobile[[:space:]]Architectures/d2803ad7-678d-49e5-bea4-de405074e019_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MobileOne_[[:space:]]An[[:space:]]Improved[[:space:]]One[[:space:]]Millisecond[[:space:]]Mobile[[:space:]]Backbone/c5f43de1-aa2f-4925-95d1-5460a5c1f6bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MobileVOS_[[:space:]]Real-Time[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation[[:space:]]Contrastive[[:space:]]Learning[[:space:]]Meets[[:space:]]Knowledge[[:space:]]Distillation/12c83345-4d98-460f-b67a-232a84dd4b0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mod-Squad_[[:space:]]Designing[[:space:]]Mixtures[[:space:]]of[[:space:]]Experts[[:space:]]As[[:space:]]Modular[[:space:]]Multi-Task[[:space:]]Learners/3eed4993-52e5-4bef-98ce-6794c7fd0b80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Modality-Agnostic[[:space:]]Debiasing[[:space:]]for[[:space:]]Single[[:space:]]Domain[[:space:]]Generalization/a78671a0-4344-4ef5-aa59-4929280d49ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Modality-Invariant[[:space:]]Visual[[:space:]]Odometry[[:space:]]for[[:space:]]Embodied[[:space:]]Vision/f79fa6c3-5616-4488-928f-13a43c5d5a8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Model[[:space:]]Barrier_[[:space:]]A[[:space:]]Compact[[:space:]]Un-Transferable[[:space:]]Isolation[[:space:]]Domain[[:space:]]for[[:space:]]Model[[:space:]]Intellectual[[:space:]]Property[[:space:]]Protection/d8b06487-f1f6-41fc-ab87-eeb4798f3109_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Model-Agnostic[[:space:]]Gender[[:space:]]Debiased[[:space:]]Image[[:space:]]Captioning/8e2c1ef8-ad50-465a-a9b9-aec3c11fb321_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Modeling[[:space:]]Entities[[:space:]]As[[:space:]]Semantic[[:space:]]Points[[:space:]]for[[:space:]]Visual[[:space:]]Information[[:space:]]Extraction[[:space:]]in[[:space:]]the[[:space:]]Wild/0bade557-0573-4cba-acfc-75430fd5f37d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Modeling[[:space:]]Inter-Class[[:space:]]and[[:space:]]Intra-Class[[:space:]]Constraints[[:space:]]in[[:space:]]Novel[[:space:]]Class[[:space:]]Discovery/a18fce0a-6ee0-4849-8471-353e6b8417fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Modeling[[:space:]]Video[[:space:]]As[[:space:]]Stochastic[[:space:]]Processes[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Video[[:space:]]Representation[[:space:]]Learning/c900e2bc-1224-4245-9722-d5efeb434cab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Modeling[[:space:]]the[[:space:]]Distributional[[:space:]]Uncertainty[[:space:]]for[[:space:]]Salient[[:space:]]Object[[:space:]]Detection[[:space:]]Models/410c93a3-6889-4e98-8819-64c45faada97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Modernizing[[:space:]]Old[[:space:]]Photos[[:space:]]Using[[:space:]]Multiple[[:space:]]References[[:space:]]via[[:space:]]Photorealistic[[:space:]]Style[[:space:]]Transfer/485d6153-8bc6-46df-a010-01f30c4b56d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Modular[[:space:]]Memorability_[[:space:]]Tiered[[:space:]]Representations[[:space:]]for[[:space:]]Video[[:space:]]Memorability[[:space:]]Prediction/dd74aea6-47ff-4b62-b3b1-dc2f9a7a2169_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mofusion_[[:space:]]A[[:space:]]Framework[[:space:]]for[[:space:]]Denoising-Diffusion-Based[[:space:]]Motion[[:space:]]Synthesis/a5ae0eb6-04c2-4957-9d60-ecd5d6ec4033_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MonoATT_[[:space:]]Online[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]Adaptive[[:space:]]Token[[:space:]]Transformer/9a1241a6-5ad6-40d5-851b-a7ce50f5dd35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MonoHuman_[[:space:]]Animatable[[:space:]]Human[[:space:]]Neural[[:space:]]Field[[:space:]]From[[:space:]]Monocular[[:space:]]Video/3334686a-cc4f-42f2-8795-a1a74d3d2bab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Motion[[:space:]]Information[[:space:]]Propagation[[:space:]]for[[:space:]]Neural[[:space:]]Video[[:space:]]Compression/c19451be-f16c-44b4-8812-0a4812da55fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MotionDiffuser_[[:space:]]Controllable[[:space:]]Multi-Agent[[:space:]]Motion[[:space:]]Prediction[[:space:]]Using[[:space:]]Diffusion/9268b3fe-9b2b-4785-a7d8-9fae9c2e0c87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MotionTrack_[[:space:]]Learning[[:space:]]Robust[[:space:]]Short-Term[[:space:]]and[[:space:]]Long-Term[[:space:]]Motions[[:space:]]for[[:space:]]Multi-Object[[:space:]]Tracking/a0f6573c-c3a5-4dfa-a5b7-81e0e5cdc4f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Movies2Scenes_[[:space:]]Using[[:space:]]Movie[[:space:]]Metadata[[:space:]]To[[:space:]]Learn[[:space:]]Scene[[:space:]]Representation/34abf3c1-aef3-4fa4-bf45-1d24af2536a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi[[:space:]]Domain[[:space:]]Learning[[:space:]]for[[:space:]]Motion[[:space:]]Magnification/6c196959-5a97-459b-8126-07e46d0c53d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Agent[[:space:]]Automated[[:space:]]Machine[[:space:]]Learning/e299a0c1-80ef-4cd2-9ad4-3fcfb42efb53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Centroid[[:space:]]Task[[:space:]]Descriptor[[:space:]]for[[:space:]]Dynamic[[:space:]]Class[[:space:]]Incremental[[:space:]]Inference/05f8e16c-d71b-480a-ba65-977ee2107e1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Concept[[:space:]]Customization[[:space:]]of[[:space:]]Text-to-Image[[:space:]]Diffusion/f4751b29-9fcb-4d57-acec-d197765e7245_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Granularity[[:space:]]Archaeological[[:space:]]Dating[[:space:]]of[[:space:]]Chinese[[:space:]]Bronze[[:space:]]Dings[[:space:]]Based[[:space:]]on[[:space:]]a[[:space:]]Knowledge-Guided[[:space:]]Relation[[:space:]]Graph/25fd5e74-e28a-40ad-abde-cf66f98a1d8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Label[[:space:]]Compound[[:space:]]Expression[[:space:]]Recognition_[[:space:]]C-EXPR[[:space:]]Database[[:space:]]&[[:space:]]Network/54e12a89-b184-41fe-9713-db979abd3d50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Level[[:space:]]Logit[[:space:]]Distillation/01417156-a97e-47ec-8991-13f4920cf549_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Modal[[:space:]]Gait[[:space:]]Recognition[[:space:]]via[[:space:]]Effective[[:space:]]Spatial-Temporal[[:space:]]Feature[[:space:]]Fusion/039fe3f9-83b4-466e-bbe8-2aea463cce17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Modal[[:space:]]Learning[[:space:]]With[[:space:]]Missing[[:space:]]Modality[[:space:]]via[[:space:]]Shared-Specific[[:space:]]Feature[[:space:]]Modelling/c2b5afaa-e73c-4896-be55-078fde8f06f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Modal[[:space:]]Representation[[:space:]]Learning[[:space:]]With[[:space:]]Text-Driven[[:space:]]Soft[[:space:]]Masks/228c7cbb-0250-4bed-9b71-b5e2b400cfaa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Mode[[:space:]]Online[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/73820773-2789-4b46-bb0d-c660161c3d3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Object[[:space:]]Manipulation[[:space:]]via[[:space:]]Object-Centric[[:space:]]Neural[[:space:]]Scattering[[:space:]]Functions/47f1d9ec-c767-4066-9e36-86f0ecfadd4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Realism[[:space:]]Image[[:space:]]Compression[[:space:]]With[[:space:]]a[[:space:]]Conditional[[:space:]]Generator/fc1e60d6-79f3-4024-9008-419acb2a50e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Sensor[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]for[[:space:]]Multi-View[[:space:]]3D[[:space:]]Reconstruction/32541047-6dff-40d7-a605-294fc1c164db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Space[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/d6ee3cca-2ae2-44bc-84c4-d903fadd6298_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-View[[:space:]]Adversarial[[:space:]]Discriminator_[[:space:]]Mine[[:space:]]the[[:space:]]Non-Causal[[:space:]]Factors[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]Unseen[[:space:]]Domains/7988619d-7e34-471c-8a2e-de3e2a54d2c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-View[[:space:]]Azimuth[[:space:]]Stereo[[:space:]]via[[:space:]]Tangent[[:space:]]Space[[:space:]]Consistency/f353421b-e24f-403d-a6ab-67d4e166418c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-View[[:space:]]Inverse[[:space:]]Rendering[[:space:]]for[[:space:]]Large-Scale[[:space:]]Real-World[[:space:]]Indoor[[:space:]]Scenes/0c46d057-fe95-4d0e-a838-f2b85040e311_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-View[[:space:]]Reconstruction[[:space:]]Using[[:space:]]Signed[[:space:]]Ray[[:space:]]Distance[[:space:]]Functions[[:space:]](SRDF)/6d8d1bd4-ca3e-49ee-8c91-de7ccf2c6b81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-View[[:space:]]Stereo[[:space:]]Representation[[:space:]]Revist_[[:space:]]Region-Aware[[:space:]]MVSNet/7a2b739a-2889-4297-8b13-eb16c6857b7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multiclass[[:space:]]Confidence[[:space:]]and[[:space:]]Localization[[:space:]]Calibration[[:space:]]for[[:space:]]Object[[:space:]]Detection/a0491f1d-1dcd-4a32-83da-cca1bef96b7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multilateral[[:space:]]Semantic[[:space:]]Relations[[:space:]]Modeling[[:space:]]for[[:space:]]Image[[:space:]]Text[[:space:]]Retrieval/4aaa5108-1d4f-4f49-b75b-5fc1b5511234_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multimodal[[:space:]]Industrial[[:space:]]Anomaly[[:space:]]Detection[[:space:]]via[[:space:]]Hybrid[[:space:]]Fusion/e6c82a99-788c-4e74-b48c-6ec0a9dff377_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multimodal[[:space:]]Prompting[[:space:]]With[[:space:]]Missing[[:space:]]Modalities[[:space:]]for[[:space:]]Visual[[:space:]]Recognition/a274be95-b5b7-41be-8472-638d03a3aa8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multimodality[[:space:]]Helps[[:space:]]Unimodality_[[:space:]]Cross-Modal[[:space:]]Few-Shot[[:space:]]Learning[[:space:]]With[[:space:]]Multimodal[[:space:]]Models/4db324d9-73f1-47a6-825b-02b4fec80448_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multiple[[:space:]]Instance[[:space:]]Learning[[:space:]]via[[:space:]]Iterative[[:space:]]Self-Paced[[:space:]]Supervised[[:space:]]Contrastive[[:space:]]Learning/48ecbc0d-6989-429f-9e0f-6bc76f06eb40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multiplicative[[:space:]]Fourier[[:space:]]Level[[:space:]]of[[:space:]]Detail/826958a1-cef9-4792-8192-a822ebb74e6b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multiscale[[:space:]]Tensor[[:space:]]Decomposition[[:space:]]and[[:space:]]Rendering[[:space:]]Equation[[:space:]]Encoding[[:space:]]for[[:space:]]View[[:space:]]Synthesis/4b0603b1-4600-4802-8803-24b8139f77ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multispectral[[:space:]]Video[[:space:]]Semantic[[:space:]]Segmentation_[[:space:]]A[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]and[[:space:]]Baseline/ab38836f-d94d-4560-8aaa-cd7c46409ebe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multivariate,[[:space:]]Multi-Frequency[[:space:]]and[[:space:]]Multimodal_[[:space:]]Rethinking[[:space:]]Graph[[:space:]]Neural[[:space:]]Networks[[:space:]]for[[:space:]]Emotion[[:space:]]Recognition[[:space:]]in[[:space:]]Conversation/73738f08-d244-4d2b-9a68-dbdaa85fbb4c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multiview[[:space:]]Compressive[[:space:]]Coding[[:space:]]for[[:space:]]3D[[:space:]]Reconstruction/fd4bc345-bef1-43bd-a692-f20f6ee160ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Music-Driven[[:space:]]Group[[:space:]]Choreography/3f6076c4-95b0-4be1-8ca0-d1d71d6191ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mutual[[:space:]]Information-Based[[:space:]]Temporal[[:space:]]Difference[[:space:]]Learning[[:space:]]for[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]in[[:space:]]Video/65582bdc-0680-402c-8d60-9737417221ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/N-Gram[[:space:]]in[[:space:]]Swin[[:space:]]Transformers[[:space:]]for[[:space:]]Efficient[[:space:]]Lightweight[[:space:]]Image[[:space:]]Super-Resolution/005a39bc-7994-40b4-8f91-c10763957902_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NAR-Former_[[:space:]]Neural[[:space:]]Architecture[[:space:]]Representation[[:space:]]Learning[[:space:]]Towards[[:space:]]Holistic[[:space:]]Attributes[[:space:]]Prediction/d027e0f8-7b89-485d-90b7-ea71189bf867_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NEF_[[:space:]]Neural[[:space:]]Edge[[:space:]]Fields[[:space:]]for[[:space:]]3D[[:space:]]Parametric[[:space:]]Curve[[:space:]]Reconstruction[[:space:]]From[[:space:]]Multi-View[[:space:]]Images/5cb90fee-934f-4ef1-aa1e-e27eff2cc027_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NICO++_[[:space:]]Towards[[:space:]]Better[[:space:]]Benchmarking[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/05710cf2-fc54-4fab-8e89-5a8049c6fbde_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NIFF_[[:space:]]Alleviating[[:space:]]Forgetting[[:space:]]in[[:space:]]Generalized[[:space:]]Few-Shot[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Neural[[:space:]]Instance[[:space:]]Feature[[:space:]]Forging/55589295-c081-4977-aad8-fb86ff1e6b7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NIKI_[[:space:]]Neural[[:space:]]Inverse[[:space:]]Kinematics[[:space:]]With[[:space:]]Invertible[[:space:]]Neural[[:space:]]Networks[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Estimation/ffc78224-cdd4-4609-ae1b-c2e32d0d1e37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NIPQ_[[:space:]]Noise[[:space:]]Proxy-Based[[:space:]]Integrated[[:space:]]Pseudo-Quantization/1de02070-ad66-4222-8921-bef6bd7ede01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NIRVANA_[[:space:]]Neural[[:space:]]Implicit[[:space:]]Representations[[:space:]]of[[:space:]]Videos[[:space:]]With[[:space:]]Adaptive[[:space:]]Networks[[:space:]]and[[:space:]]Autoregressive[[:space:]]Patch-Wise[[:space:]]Modeling/5e35731f-8ad1-4a3e-9cb9-35eaca53dae1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NLOST_[[:space:]]Non-Line-of-Sight[[:space:]]Imaging[[:space:]]With[[:space:]]Transformer/7a921a85-5f46-4297-b05a-31e6bbebb444_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NS3D_[[:space:]]Neuro-Symbolic[[:space:]]Grounding[[:space:]]of[[:space:]]3D[[:space:]]Objects[[:space:]]and[[:space:]]Relations/52963815-a1a7-4afc-b77d-3302415af208_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NULL-Text[[:space:]]Inversion[[:space:]]for[[:space:]]Editing[[:space:]]Real[[:space:]]Images[[:space:]]Using[[:space:]]Guided[[:space:]]Diffusion[[:space:]]Models/6caae691-3671-4850-988c-58151969494f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NUWA-LIP_[[:space:]]Language-Guided[[:space:]]Image[[:space:]]Inpainting[[:space:]]With[[:space:]]Defect-Free[[:space:]]VQGAN/7eddf61f-e167-4a12-adec-e5ebd2ecc98e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NVTC_[[:space:]]Nonlinear[[:space:]]Vector[[:space:]]Transform[[:space:]]Coding/d6b3be84-f928-4746-bfe7-e6665d8237b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NaQ_[[:space:]]Leveraging[[:space:]]Narrations[[:space:]]As[[:space:]]Queries[[:space:]]To[[:space:]]Supervise[[:space:]]Episodic[[:space:]]Memory/e02412a6-f8ef-4ee9-aeb7-c9e1b2b5298f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Natural[[:space:]]Language-Assisted[[:space:]]Sign[[:space:]]Language[[:space:]]Recognition/fb909238-053f-4de8-8c66-660559012245_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeAT_[[:space:]]Learning[[:space:]]Neural[[:space:]]Implicit[[:space:]]Surfaces[[:space:]]With[[:space:]]Arbitrary[[:space:]]Topologies[[:space:]]From[[:space:]]Multi-View[[:space:]]Images/a4955c11-14f8-49f0-ac2f-881468eb66be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeFII_[[:space:]]Inverse[[:space:]]Rendering[[:space:]]for[[:space:]]Reflectance[[:space:]]Decomposition[[:space:]]With[[:space:]]Near-Field[[:space:]]Indirect[[:space:]]Illumination/714236d9-7470-4b6e-933f-abb18412b311_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeMo_[[:space:]]Learning[[:space:]]3D[[:space:]]Neural[[:space:]]Motion[[:space:]]Fields[[:space:]]From[[:space:]]Multiple[[:space:]]Video[[:space:]]Instances[[:space:]]of[[:space:]]the[[:space:]]Same[[:space:]]Action/9a856238-13b3-43e7-a03f-f344f4eb01c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRDi_[[:space:]]Single-View[[:space:]]NeRF[[:space:]]Synthesis[[:space:]]With[[:space:]]Language-Guided[[:space:]]Diffusion[[:space:]]As[[:space:]]General[[:space:]]Image[[:space:]]Priors/92e7449b-7ab9-4887-bffe-d5d0da8e3001_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRF[[:space:]]in[[:space:]]the[[:space:]]Palm[[:space:]]of[[:space:]]Your[[:space:]]Hand_[[:space:]]Corrective[[:space:]]Augmentation[[:space:]]for[[:space:]]Robotics[[:space:]]via[[:space:]]Novel-View[[:space:]]Synthesis/4a35e7d6-74d3-449a-a05e-03a6845f1f61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRF-DS_[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Dynamic[[:space:]]Specular[[:space:]]Objects/98a4a76c-ad86-4b8a-8f52-440da1898c74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRF-RPN_[[:space:]]A[[:space:]]General[[:space:]]Framework[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]NeRFs/79c44f94-b925-4ed7-88d2-f9ff7cda9eed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRF-Supervised[[:space:]]Deep[[:space:]]Stereo/61c79f8e-9d95-4da9-bf65-b420d18a115f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRFInvertor_[[:space:]]High[[:space:]]Fidelity[[:space:]]NeRF-GAN[[:space:]]Inversion[[:space:]]for[[:space:]]Single-Shot[[:space:]]Real[[:space:]]Image[[:space:]]Animation/0379d115-8546-4310-beee-7950377bd8c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRFLight_[[:space:]]Fast[[:space:]]and[[:space:]]Light[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]Using[[:space:]]a[[:space:]]Shared[[:space:]]Feature[[:space:]]Grid/4e85eead-8ddc-4b8d-a3c3-748f68b6f0be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRFLix_[[:space:]]High-Quality[[:space:]]Neural[[:space:]]View[[:space:]]Synthesis[[:space:]]by[[:space:]]Learning[[:space:]]a[[:space:]]Degradation-Driven[[:space:]]Inter-Viewpoint[[:space:]]MiXer/bf353a57-f1ac-408c-97ad-362f62f7e8ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRFVS_[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Free[[:space:]]View[[:space:]]Synthesis[[:space:]]via[[:space:]]Geometry[[:space:]]Scaffolds/7d7f552d-da1a-4c2b-9a99-b1e15f03172a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeUDF_[[:space:]]Leaning[[:space:]]Neural[[:space:]]Unsigned[[:space:]]Distance[[:space:]]Fields[[:space:]]With[[:space:]]Volume[[:space:]]Rendering/001e4de4-a453-4acd-a0ed-17fa830fbbf5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neighborhood[[:space:]]Attention[[:space:]]Transformer/5ba68135-05da-4bc9-9e69-981fe0395bd7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NerVE_[[:space:]]Neural[[:space:]]Volumetric[[:space:]]Edges[[:space:]]for[[:space:]]Parametric[[:space:]]Curve[[:space:]]Extraction[[:space:]]From[[:space:]]Point[[:space:]]Cloud/87460004-df50-4013-aa11-4cc4f90548da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Nerflets_[[:space:]]Local[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Efficient[[:space:]]Structure-Aware[[:space:]]3D[[:space:]]Scene[[:space:]]Representation[[:space:]]From[[:space:]]2D[[:space:]]Supervision/cb6607d1-ed9a-4477-9895-3412bd4004e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Network[[:space:]]Expansion[[:space:]]for[[:space:]]Practical[[:space:]]Training[[:space:]]Acceleration/d44829b9-d472-4375-a542-bc08d47a70ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Network-Free,[[:space:]]Unsupervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]With[[:space:]]Synthetic[[:space:]]Images/0588fa90-989d-4c4d-9658-c6b45f6ec3d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeuDA_[[:space:]]Neural[[:space:]]Deformable[[:space:]]Anchor[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Implicit[[:space:]]Surface[[:space:]]Reconstruction/88de0a46-6fef-4fc8-bc71-c1be8d1b0fa2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeuFace_[[:space:]]Realistic[[:space:]]3D[[:space:]]Neural[[:space:]]Face[[:space:]]Rendering[[:space:]]From[[:space:]]Multi-View[[:space:]]Images/2a5a4919-a351-4576-8a4d-0ca8577a0a2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeuMap_[[:space:]]Neural[[:space:]]Coordinate[[:space:]]Mapping[[:space:]]by[[:space:]]Auto-Transdecoder[[:space:]]for[[:space:]]Camera[[:space:]]Localization/37d3338b-b1de-4961-af25-2bb1739598f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeuWigs_[[:space:]]A[[:space:]]Neural[[:space:]]Dynamic[[:space:]]Model[[:space:]]for[[:space:]]Volumetric[[:space:]]Hair[[:space:]]Capture[[:space:]]and[[:space:]]Animation/a1a05297-f472-4065-ba16-5b8e907bd71e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neumann[[:space:]]Network[[:space:]]With[[:space:]]Recursive[[:space:]]Kernels[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Defocus[[:space:]]Deblurring/5d36d512-e4f0-435d-9ba9-19ca89c48151_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeurOCS_[[:space:]]Neural[[:space:]]NOCS[[:space:]]Supervision[[:space:]]for[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Localization/680912d7-dcd0-48ce-9376-3b0534722447_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Congealing_[[:space:]]Aligning[[:space:]]Images[[:space:]]to[[:space:]]a[[:space:]]Joint[[:space:]]Semantic[[:space:]]Atlas/f69750de-61a8-4a7a-ae74-33e7c73aae14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Dependencies[[:space:]]Emerging[[:space:]]From[[:space:]]Learning[[:space:]]Massive[[:space:]]Categories/b43c8b00-651c-4146-b6e8-0b6da5efa8fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Fields[[:space:]]Meet[[:space:]]Explicit[[:space:]]Geometric[[:space:]]Representations[[:space:]]for[[:space:]]Inverse[[:space:]]Rendering[[:space:]]of[[:space:]]Urban[[:space:]]Scenes/f9f402a4-00fb-4fb5-b91f-a9ededac6de9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Fourier[[:space:]]Filter[[:space:]]Bank/8e28aa5e-8aee-4afd-b662-d8eef15b557a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Intrinsic[[:space:]]Embedding[[:space:]]for[[:space:]]Non-Rigid[[:space:]]Point[[:space:]]Cloud[[:space:]]Matching/5fee3d70-ffc5-4238-8bc6-54203175c488_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Kaleidoscopic[[:space:]]Space[[:space:]]Sculpting/1533656c-4675-43d2-b7d9-18dca04a2292_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Kernel[[:space:]]Surface[[:space:]]Reconstruction/6cff62eb-041c-400d-a046-678e07c34c6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Koopman[[:space:]]Pooling_[[:space:]]Control-Inspired[[:space:]]Temporal[[:space:]]Dynamics[[:space:]]Encoding[[:space:]]for[[:space:]]Skeleton-Based[[:space:]]Action[[:space:]]Recognition/e12e7026-a989-40c1-b7bd-f6f4fed9c983_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Lens[[:space:]]Modeling/195d6c6a-f6ff-409a-9b8e-8fec467bebeb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Map[[:space:]]Prior[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/a887fa06-46b3-410e-829e-f80e8bd82a93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Part[[:space:]]Priors_[[:space:]]Learning[[:space:]]To[[:space:]]Optimize[[:space:]]Part-Based[[:space:]]Object[[:space:]]Completion[[:space:]]in[[:space:]]RGB-D[[:space:]]Scans/8fce0d09-0843-4fdc-895e-b86ebe5fcfbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Pixel[[:space:]]Composition[[:space:]]for[[:space:]]3D-4D[[:space:]]View[[:space:]]Synthesis[[:space:]]From[[:space:]]Multi-Views/5aed6554-5d22-4ba0-b042-2a172c2e54da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Preset[[:space:]]for[[:space:]]Color[[:space:]]Style[[:space:]]Transfer/321eae1b-b244-4891-a620-fce28ce02a23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Rate[[:space:]]Estimator[[:space:]]and[[:space:]]Unsupervised[[:space:]]Learning[[:space:]]for[[:space:]]Efficient[[:space:]]Distributed[[:space:]]Image[[:space:]]Analytics[[:space:]]in[[:space:]]Split-DNN[[:space:]]Models/676a0d64-c310-4b48-be40-69566dc6f2c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Residual[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Streamably[[:space:]]Free-Viewpoint[[:space:]]Videos/f5f049fc-f894-4251-8377-fce38192f584_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Scene[[:space:]]Chronology/ab0ddc3b-0e66-489f-81ac-18b868a8dcf5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Texture[[:space:]]Synthesis[[:space:]]With[[:space:]]Guided[[:space:]]Correspondence/0d395694-dab5-4c6a-8e55-9527936bde58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Transformation[[:space:]]Fields[[:space:]]for[[:space:]]Arbitrary-Styled[[:space:]]Font[[:space:]]Generation/b7d79e74-13fc-423c-9b96-f2beae254d5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Vector[[:space:]]Fields_[[:space:]]Implicit[[:space:]]Representation[[:space:]]by[[:space:]]Explicit[[:space:]]Learning/b84bdac4-dddd-4b22-a197-aea99de19635_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Video[[:space:]]Compression[[:space:]]With[[:space:]]Diverse[[:space:]]Contexts/b731bba7-3b96-4f19-a574-8047f59049bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Volumetric[[:space:]]Memory[[:space:]]for[[:space:]]Visual[[:space:]]Locomotion[[:space:]]Control/dad36583-d1e2-4d95-a7b6-551afd512667_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Voting[[:space:]]Field[[:space:]]for[[:space:]]Camera-Space[[:space:]]3D[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation/96d05801-c2f5-4205-ac8b-2fcd00428ae6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeuralDome_[[:space:]]A[[:space:]]Neural[[:space:]]Modeling[[:space:]]Pipeline[[:space:]]on[[:space:]]Multi-View[[:space:]]Human-Object[[:space:]]Interactions/fa0c20cd-9328-405b-b73e-2045fdfa714d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeuralEditor_[[:space:]]Editing[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]via[[:space:]]Manipulating[[:space:]]Point[[:space:]]Clouds/70b7a137-ced8-4460-839e-a2dd484a031c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeuralField-LDM_[[:space:]]Scene[[:space:]]Generation[[:space:]]With[[:space:]]Hierarchical[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models/81bfb70e-1652-4684-aa6c-4729296b6ed4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeuralLift-360_[[:space:]]Lifting[[:space:]]an[[:space:]]In-the-Wild[[:space:]]2D[[:space:]]Photo[[:space:]]to[[:space:]]a[[:space:]]3D[[:space:]]Object[[:space:]]With[[:space:]]360deg[[:space:]]Views/90dd78db-4103-415c-8ddc-ba0f83221385_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeuralPCI_[[:space:]]Spatio-Temporal[[:space:]]Neural[[:space:]]Field[[:space:]]for[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Multi-Frame[[:space:]]Non-Linear[[:space:]]Interpolation/865808e9-fa47-4657-baf0-084c6ecae997_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeuralUDF_[[:space:]]Learning[[:space:]]Unsigned[[:space:]]Distance[[:space:]]Fields[[:space:]]for[[:space:]]Multi-View[[:space:]]Reconstruction[[:space:]]of[[:space:]]Surfaces[[:space:]]With[[:space:]]Arbitrary[[:space:]]Topologies/cda2b21c-85ea-47d7-952f-4f7b2568e13a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neuralangelo_[[:space:]]High-Fidelity[[:space:]]Neural[[:space:]]Surface[[:space:]]Reconstruction/55f0b247-41a3-45d9-bec0-4490ab23bed9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neuralizer_[[:space:]]General[[:space:]]Neuroimage[[:space:]]Analysis[[:space:]]Without[[:space:]]Re-Training/eee09609-bc86-4647-98dc-269f47c8002f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neuro-Modulated[[:space:]]Hebbian[[:space:]]Learning[[:space:]]for[[:space:]]Fully[[:space:]]Test-Time[[:space:]]Adaptation/28e8edf6-185c-4dae-9bfc-b17f1c2f7606_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neuron[[:space:]]Structure[[:space:]]Modeling[[:space:]]for[[:space:]]Generalizable[[:space:]]Remote[[:space:]]Physiological[[:space:]]Measurement/0c500895-2bd8-4cea-80ae-7bc4fd9b8ba5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NewsNet_[[:space:]]A[[:space:]]Novel[[:space:]]Dataset[[:space:]]for[[:space:]]Hierarchical[[:space:]]Temporal[[:space:]]Segmentation/d2756edd-2b85-4c25-85cd-99de3bdf0d9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Next3D_[[:space:]]Generative[[:space:]]Neural[[:space:]]Texture[[:space:]]Rasterization[[:space:]]for[[:space:]]3D-Aware[[:space:]]Head[[:space:]]Avatars/7b615aeb-a836-44b1-beba-de01bf0423b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Nighttime[[:space:]]Smartphone[[:space:]]Reflective[[:space:]]Flare[[:space:]]Removal[[:space:]]Using[[:space:]]Optical[[:space:]]Center[[:space:]]Symmetry[[:space:]]Prior/d34edf55-17a4-4aa7-b0c0-de157905c27a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/No[[:space:]]One[[:space:]]Left[[:space:]]Behind_[[:space:]]Improving[[:space:]]the[[:space:]]Worst[[:space:]]Categories[[:space:]]in[[:space:]]Long-Tailed[[:space:]]Learning/aa8d6d83-fa42-452c-97e5-1d8dfd5a2154_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NoPe-NeRF_[[:space:]]Optimising[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field[[:space:]]With[[:space:]]No[[:space:]]Pose[[:space:]]Prior/2260561b-6c24-4454-9ee0-0e88425e374a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Noisy[[:space:]]Correspondence[[:space:]]Learning[[:space:]]With[[:space:]]Meta[[:space:]]Similarity[[:space:]]Correction/0167b343-6cc5-4960-bab8-e11934c107f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NoisyQuant_[[:space:]]Noisy[[:space:]]Bias-Enhanced[[:space:]]Post-Training[[:space:]]Activation[[:space:]]Quantization[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/68808ca4-cf24-4585-b35b-3e2acef953ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NoisyTwins_[[:space:]]Class-Consistent[[:space:]]and[[:space:]]Diverse[[:space:]]Image[[:space:]]Generation[[:space:]]Through[[:space:]]StyleGANs/c323d83e-f322-4c02-8ecc-1d0c86cf0f70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Non-Contrastive[[:space:]]Learning[[:space:]]Meets[[:space:]]Language-Image[[:space:]]Pre-Training/fe567c45-5ef2-4da5-8116-0c22fc6e398a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Non-Contrastive[[:space:]]Unsupervised[[:space:]]Learning[[:space:]]of[[:space:]]Physiological[[:space:]]Signals[[:space:]]From[[:space:]]Video/f2bb9236-2bf0-41f5-a980-2c135b55dca1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Non-Line-of-Sight[[:space:]]Imaging[[:space:]]With[[:space:]]Signal[[:space:]]Superresolution[[:space:]]Network/224cfcfa-2657-4e74-b370-35626fe2670d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Normal-Guided[[:space:]]Garment[[:space:]]UV[[:space:]]Prediction[[:space:]]for[[:space:]]Human[[:space:]]Re-Texturing/6ef67559-1f18-4ef9-b45a-a009b6d372b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Normalizing[[:space:]]Flow[[:space:]]Based[[:space:]]Feature[[:space:]]Synthesis[[:space:]]for[[:space:]]Outlier-Aware[[:space:]]Object[[:space:]]Detection/44811963-183d-45a5-8a3b-7dd7378c9155_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Not[[:space:]]All[[:space:]]Image[[:space:]]Regions[[:space:]]Matter_[[:space:]]Masked[[:space:]]Vector[[:space:]]Quantization[[:space:]]for[[:space:]]Autoregressive[[:space:]]Image[[:space:]]Generation/2a88178e-8ce2-4c0f-b3f1-bb99f0452012_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Novel[[:space:]]Class[[:space:]]Discovery[[:space:]]for[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Semantic[[:space:]]Segmentation/b9953b69-6d05-422e-9f85-28806509ef58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Novel-View[[:space:]]Acoustic[[:space:]]Synthesis/2dea3110-7f39-4b1f-9a6f-a10e0ea0bd92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OCELOT_[[:space:]]Overlapped[[:space:]]Cell[[:space:]]on[[:space:]]Tissue[[:space:]]Dataset[[:space:]]for[[:space:]]Histopathology/052d2272-757d-4e0e-b0c0-b44903da46a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OCTET_[[:space:]]Object-Aware[[:space:]]Counterfactual[[:space:]]Explanations/03950ee8-92a4-474c-99c2-49607c7368f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OPE-SR_[[:space:]]Orthogonal[[:space:]]Position[[:space:]]Encoding[[:space:]]for[[:space:]]Designing[[:space:]]a[[:space:]]Parameter-Free[[:space:]]Upsampling[[:space:]]Module[[:space:]]in[[:space:]]Arbitrary-Scale[[:space:]]Image[[:space:]]Super-Resolution/538d0672-86b3-4cf7-ab0d-b9344b19df37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ORCa_[[:space:]]Glossy[[:space:]]Objects[[:space:]]As[[:space:]]Radiance-Field[[:space:]]Cameras/da15cd79-f9cc-4796-854b-9e2af61a06e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OReX_[[:space:]]Object[[:space:]]Reconstruction[[:space:]]From[[:space:]]Planar[[:space:]]Cross-Sections[[:space:]]Using[[:space:]]Neural[[:space:]]Fields/748d73f6-6f8e-4794-a6a8-f014dea611bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OSAN_[[:space:]]A[[:space:]]One-Stage[[:space:]]Alignment[[:space:]]Network[[:space:]]To[[:space:]]Unify[[:space:]]Multimodal[[:space:]]Alignment[[:space:]]and[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/4429e1d5-6fb8-40a5-9804-367d3c0ae901_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OSRT_[[:space:]]Omnidirectional[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]With[[:space:]]Distortion-Aware[[:space:]]Transformer/b6aea18d-97cd-464e-a4ad-626c19fdc060_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OT-Filter_[[:space:]]An[[:space:]]Optimal[[:space:]]Transport[[:space:]]Filter[[:space:]]for[[:space:]]Learning[[:space:]]With[[:space:]]Noisy[[:space:]]Labels/69b93138-511e-410f-9d28-563936aa9d2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OTAvatar_[[:space:]]One-Shot[[:space:]]Talking[[:space:]]Face[[:space:]]Avatar[[:space:]]With[[:space:]]Controllable[[:space:]]Tri-Plane[[:space:]]Rendering/aea7643c-93a9-4960-bd4a-4385733b4c38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OVTrack_[[:space:]]Open-Vocabulary[[:space:]]Multiple[[:space:]]Object[[:space:]]Tracking/092f1f8a-50aa-41bc-8e59-a89dde3c1cc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Objaverse_[[:space:]]A[[:space:]]Universe[[:space:]]of[[:space:]]Annotated[[:space:]]3D[[:space:]]Objects/ea3a0563-acc5-4afc-aad6-362a9401875b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Object[[:space:]]Detection[[:space:]]With[[:space:]]Self-Supervised[[:space:]]Scene[[:space:]]Adaptation/b774b6cf-eacd-4660-845f-f32b5ec4dc83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Object[[:space:]]Discovery[[:space:]]From[[:space:]]Motion-Guided[[:space:]]Tokens/cce05c92-f0c6-4fb5-8422-8cb2ccb76230_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Object[[:space:]]Pop-Up_[[:space:]]Can[[:space:]]We[[:space:]]Infer[[:space:]]3D[[:space:]]Objects[[:space:]]and[[:space:]]Their[[:space:]]Poses[[:space:]]From[[:space:]]Human[[:space:]]Interactions[[:space:]]Alone_/965b59c8-d0c6-413b-852b-108c8ff0bbe7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]With[[:space:]]Statistical[[:space:]]Guarantees_[[:space:]]Conformal[[:space:]]Keypoint[[:space:]]Detection[[:space:]]and[[:space:]]Geometric[[:space:]]Uncertainty[[:space:]]Propagation/a86330a1-e4d7-4fd5-a6ce-7c82335913b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Object-Aware[[:space:]]Distillation[[:space:]]Pyramid[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detection/b1741226-bc11-487a-861d-b993de9bcc68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Object-Goal[[:space:]]Visual[[:space:]]Navigation[[:space:]]via[[:space:]]Effective[[:space:]]Exploration[[:space:]]of[[:space:]]Relations[[:space:]]Among[[:space:]]Historical[[:space:]]Navigation[[:space:]]States/02126feb-2cfd-431e-b8b7-664314fe1580_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ObjectMatch_[[:space:]]Robust[[:space:]]Registration[[:space:]]Using[[:space:]]Canonical[[:space:]]Object[[:space:]]Correspondences/9d0104e4-40f2-4a9d-bf9d-8065e49cc737_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ObjectStitch_[[:space:]]Object[[:space:]]Compositing[[:space:]]With[[:space:]]Diffusion[[:space:]]Model/69193070-f3e5-4dd6-8743-b15c03b3e677_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Observation-Centric[[:space:]]SORT_[[:space:]]Rethinking[[:space:]]SORT[[:space:]]for[[:space:]]Robust[[:space:]]Multi-Object[[:space:]]Tracking/b4c4cc9d-52d9-4722-825a-050e1d27cf7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OcTr_[[:space:]]Octree-Based[[:space:]]Transformer[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/b6928c01-d853-4d0d-9874-89b3af0f167a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Occlusion-Free[[:space:]]Scene[[:space:]]Recovery[[:space:]]via[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/223bdb66-a260-46a1-86eb-891a595d848a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Octree[[:space:]]Guided[[:space:]]Unoriented[[:space:]]Surface[[:space:]]Reconstruction/0db48127-a127-468b-be90-b27ec5813046_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Omni[[:space:]]Aggregation[[:space:]]Networks[[:space:]]for[[:space:]]Lightweight[[:space:]]Image[[:space:]]Super-Resolution/39175e81-b86e-49a9-967e-e590c3dbdaa0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Omni3D_[[:space:]]A[[:space:]]Large[[:space:]]Benchmark[[:space:]]and[[:space:]]Model[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]the[[:space:]]Wild/2efbd142-17ac-4a1f-b83f-1ae025c49292_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OmniAL_[[:space:]]A[[:space:]]Unified[[:space:]]CNN[[:space:]]Framework[[:space:]]for[[:space:]]Unsupervised[[:space:]]Anomaly[[:space:]]Localization/390b7405-6295-46ac-87cb-a78d4a95dd63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OmniAvatar_[[:space:]]Geometry-Guided[[:space:]]Controllable[[:space:]]3D[[:space:]]Head[[:space:]]Synthesis/13b82072-fc3f-4f4d-8e9e-44b54de26be7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OmniCity_[[:space:]]Omnipotent[[:space:]]City[[:space:]]Understanding[[:space:]]With[[:space:]]Multi-Level[[:space:]]and[[:space:]]Multi-View[[:space:]]Images/1ecc869d-a020-42e0-bc1d-485949718b11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OmniMAE_[[:space:]]Single[[:space:]]Model[[:space:]]Masked[[:space:]]Pretraining[[:space:]]on[[:space:]]Images[[:space:]]and[[:space:]]Videos/2fdd9b20-eaac-4038-a247-56813890042d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OmniObject3D_[[:space:]]Large-Vocabulary[[:space:]]3D[[:space:]]Object[[:space:]]Dataset[[:space:]]for[[:space:]]Realistic[[:space:]]Perception,[[:space:]]Reconstruction[[:space:]]and[[:space:]]Generation/243023a1-e64c-4bdc-a9c5-c30584f803ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OmniVidar_[[:space:]]Omnidirectional[[:space:]]Depth[[:space:]]Estimation[[:space:]]From[[:space:]]Multi-Fisheye[[:space:]]Images/e34f49e7-aba6-408f-9647-b38f210ab7c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Omnimatte3D_[[:space:]]Associating[[:space:]]Objects[[:space:]]and[[:space:]]Their[[:space:]]Effects[[:space:]]in[[:space:]]Unconstrained[[:space:]]Monocular[[:space:]]Video/5700f2ea-1d90-4ebb-a4e2-33ec303e3c7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]Calibrating[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]Models_[[:space:]]Analyses[[:space:]]and[[:space:]]an[[:space:]]Algorithm/f35086d8-d1cb-49d4-972d-c5bbdb5412ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]Data[[:space:]]Scaling[[:space:]]in[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling/23a2cca2-7ab3-48ca-b93b-e899d7bc87f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]Distillation[[:space:]]of[[:space:]]Guided[[:space:]]Diffusion[[:space:]]Models/2a7f0724-3ab3-4aec-b0e2-13e799a34131_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]the[[:space:]]Convergence[[:space:]]of[[:space:]]IRLS[[:space:]]and[[:space:]]Its[[:space:]]Variants[[:space:]]in[[:space:]]Outlier-Robust[[:space:]]Estimation/2f3a5413-c28c-4fa1-ba13-877b20ad60c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]the[[:space:]]Difficulty[[:space:]]of[[:space:]]Unpaired[[:space:]]Infrared-to-Visible[[:space:]]Video[[:space:]]Translation_[[:space:]]Fine-Grained[[:space:]]Content-Rich[[:space:]]Patches[[:space:]]Transfer/4498f9f9-cd08-4ca4-bf35-0052cad157ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]the[[:space:]]Effects[[:space:]]of[[:space:]]Self-Supervision[[:space:]]and[[:space:]]Contrastive[[:space:]]Alignment[[:space:]]in[[:space:]]Deep[[:space:]]Multi-View[[:space:]]Clustering/b6534bc7-2c30-499e-9218-09c70044ce02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]the[[:space:]]Importance[[:space:]]of[[:space:]]Accurate[[:space:]]Geometry[[:space:]]Data[[:space:]]for[[:space:]]Dense[[:space:]]3D[[:space:]]Vision[[:space:]]Tasks/9385a813-438e-43da-88e0-d0cb933ec7e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]the[[:space:]]Pitfall[[:space:]]of[[:space:]]Mixup[[:space:]]for[[:space:]]Uncertainty[[:space:]]Calibration/e280d37f-5a1f-464c-93ba-b8b7f813b65b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]the[[:space:]]Stability-Plasticity[[:space:]]Dilemma[[:space:]]of[[:space:]]Class-Incremental[[:space:]]Learning/d81757e3-3fe0-4e2f-aa09-8471922c1a8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On-the-Fly[[:space:]]Category[[:space:]]Discovery/63501cf6-0f6c-4c00-abdf-f3132cbea43e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/One-Shot[[:space:]]High-Fidelity[[:space:]]Talking-Head[[:space:]]Synthesis[[:space:]]With[[:space:]]Deformable[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field/dca5b187-6566-4ddd-b058-7757a1938bb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/One-Shot[[:space:]]Model[[:space:]]for[[:space:]]Mixed-Precision[[:space:]]Quantization/04de4d55-6d18-4035-b6a6-f73f860594ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/One-Stage[[:space:]]3D[[:space:]]Whole-Body[[:space:]]Mesh[[:space:]]Recovery[[:space:]]With[[:space:]]Component[[:space:]]Aware[[:space:]]Transformer/0d45c9c9-79d5-4cae-8f9e-cd8c7b54f3a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/One-to-Few[[:space:]]Label[[:space:]]Assignment[[:space:]]for[[:space:]]End-to-End[[:space:]]Dense[[:space:]]Detection/2ef6d6d9-255d-465f-af82-e62703f947f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OneFormer_[[:space:]]One[[:space:]]Transformer[[:space:]]To[[:space:]]Rule[[:space:]]Universal[[:space:]]Image[[:space:]]Segmentation/50f3fd55-8486-448a-af4a-e7ac51505bec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open[[:space:]]Set[[:space:]]Action[[:space:]]Recognition[[:space:]]via[[:space:]]Multi-Label[[:space:]]Evidential[[:space:]]Learning/522e5d22-dd2d-4255-b5ab-22d4a0818916_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open[[:space:]]Vocabulary[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]With[[:space:]]Patch[[:space:]]Aligned[[:space:]]Contrastive[[:space:]]Learning/36f8c170-4d1c-4ebf-9970-85e1b3305659_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-Category[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Pre-Training[[:space:]]via[[:space:]]Language[[:space:]]Modeling[[:space:]]Framework/d0c5f672-1696-43a7-a043-eb9423188bf0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-Set[[:space:]]Fine-Grained[[:space:]]Retrieval[[:space:]]via[[:space:]]Prompting[[:space:]]Vision-Language[[:space:]]Evaluator/262de221-4818-435d-98ce-ef9089be5006_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-Set[[:space:]]Likelihood[[:space:]]Maximization[[:space:]]for[[:space:]]Few-Shot[[:space:]]Learning/b08f27c2-44e5-4d37-a025-26dfd767010a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-Set[[:space:]]Representation[[:space:]]Learning[[:space:]]Through[[:space:]]Combinatorial[[:space:]]Embedding/7e317ef4-9fec-4f8b-a34e-c58d0714935d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-Set[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]for[[:space:]]Point[[:space:]]Clouds[[:space:]]via[[:space:]]Adversarial[[:space:]]Prototype[[:space:]]Framework/6e5fad9f-ce24-4c4b-bbe6-310c0bb1d96f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-Vocabulary[[:space:]]Attribute[[:space:]]Detection/9484a54c-5530-4755-8507-4b6348cd5379_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-Vocabulary[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]With[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/40f7143e-a359-4969-9fac-2ace141f4c73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-Vocabulary[[:space:]]Point-Cloud[[:space:]]Object[[:space:]]Detection[[:space:]]Without[[:space:]]3D[[:space:]]Annotation/7ef974c3-c0f1-4e90-aba6-1e8926c0de50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]With[[:space:]]Mask-Adapted[[:space:]]CLIP/8dc066a9-e2ba-48bb-b28d-8ee57e0cadd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-World[[:space:]]Multi-Task[[:space:]]Control[[:space:]]Through[[:space:]]Goal-Aware[[:space:]]Representation[[:space:]]Learning[[:space:]]and[[:space:]]Adaptive[[:space:]]Horizon[[:space:]]Prediction/8f63ae6d-1180-4e93-b249-5af48ed53f1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OpenGait_[[:space:]]Revisiting[[:space:]]Gait[[:space:]]Recognition[[:space:]]Towards[[:space:]]Better[[:space:]]Practicality/8bf8f8d1-986c-497b-96e1-9873a12ad647_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OpenMix_[[:space:]]Exploring[[:space:]]Outlier[[:space:]]Samples[[:space:]]for[[:space:]]Misclassification[[:space:]]Detection/e28ffe43-ab1d-41eb-b4f3-d52dd976dd0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OpenScene_[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding[[:space:]]With[[:space:]]Open[[:space:]]Vocabularies/e7b60cb8-b8c6-4ecc-b036-3f5a1154cd17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Optimal[[:space:]]Proposal[[:space:]]Learning[[:space:]]for[[:space:]]Deployable[[:space:]]End-to-End[[:space:]]Pedestrian[[:space:]]Detection/1c127ef3-81d4-4d78-8979-f3005be6c571_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Optimal[[:space:]]Transport[[:space:]]Minimization_[[:space:]]Crowd[[:space:]]Localization[[:space:]]on[[:space:]]Density[[:space:]]Maps[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Counting/3fc862a3-6173-44ce-be3b-e189c9e14e9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Optimization-Inspired[[:space:]]Cross-Attention[[:space:]]Transformer[[:space:]]for[[:space:]]Compressive[[:space:]]Sensing/5dd1edae-ef03-4e93-962d-78e4a01ea3ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OrienterNet_[[:space:]]Visual[[:space:]]Localization[[:space:]]in[[:space:]]2D[[:space:]]Public[[:space:]]Maps[[:space:]]With[[:space:]]Neural[[:space:]]Matching/dc8e9f39-487e-487a-a21e-693a181dc802_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Orthogonal[[:space:]]Annotation[[:space:]]Benefits[[:space:]]Barely-Supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/5318e236-a4fa-4589-8c36-f0847cdaf022_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Out-of-Candidate[[:space:]]Rectification[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/f193053e-9323-4b58-84df-7ef844a6846c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Out-of-Distributed[[:space:]]Semantic[[:space:]]Pruning[[:space:]]for[[:space:]]Robust[[:space:]]Semi-Supervised[[:space:]]Learning/79fa8695-6d63-40ac-bd3c-8f41526ffd50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OvarNet_[[:space:]]Towards[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Attribute[[:space:]]Recognition/5beda223-6da6-4d91-a406-f4b99ebe17a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Overcoming[[:space:]]the[[:space:]]Trade-Off[[:space:]]Between[[:space:]]Accuracy[[:space:]]and[[:space:]]Plausibility[[:space:]]in[[:space:]]3D[[:space:]]Hand[[:space:]]Shape[[:space:]]Reconstruction/35101d46-7682-4ab3-81cc-a9fe76586bb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Overlooked[[:space:]]Factors[[:space:]]in[[:space:]]Concept-Based[[:space:]]Explanations_[[:space:]]Dataset[[:space:]]Choice,[[:space:]]Concept[[:space:]]Learnability,[[:space:]]and[[:space:]]Human[[:space:]]Capability/06f1b02b-8ac5-42c3-95d5-c871f263a132_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PA&DA_[[:space:]]Jointly[[:space:]]Sampling[[:space:]]Path[[:space:]]and[[:space:]]Data[[:space:]]for[[:space:]]Consistent[[:space:]]NAS/e5045fd9-c441-40f6-86d3-085411071d4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PACO_[[:space:]]Parts[[:space:]]and[[:space:]]Attributes[[:space:]]of[[:space:]]Common[[:space:]]Objects/0dc7ddd1-bc89-48c6-bdd3-73059ab3d234_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PATS_[[:space:]]Patch[[:space:]]Area[[:space:]]Transportation[[:space:]]With[[:space:]]Subdivision[[:space:]]for[[:space:]]Local[[:space:]]Feature[[:space:]]Matching/fedd6566-132b-4b72-a834-feebfb14034d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PAniC-3D_[[:space:]]Stylized[[:space:]]Single-View[[:space:]]3D[[:space:]]Reconstruction[[:space:]]From[[:space:]]Portraits[[:space:]]of[[:space:]]Anime[[:space:]]Characters/e21e8055-3906-48b4-b8f9-7fe74b8d3e35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PC2_[[:space:]]Projection-Conditioned[[:space:]]Point[[:space:]]Cloud[[:space:]]Diffusion[[:space:]]for[[:space:]]Single-Image[[:space:]]3D[[:space:]]Reconstruction/5231daab-bcac-4e08-bf04-6630c20f67a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PCR_[[:space:]]Proxy-Based[[:space:]]Contrastive[[:space:]]Replay[[:space:]]for[[:space:]]Online[[:space:]]Class-Incremental[[:space:]]Continual[[:space:]]Learning/8f9b2a24-7fd4-4a49-98e5-3863158fa655_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PCT-Net_[[:space:]]Full[[:space:]]Resolution[[:space:]]Image[[:space:]]Harmonization[[:space:]]Using[[:space:]]Pixel-Wise[[:space:]]Color[[:space:]]Transformations/c79aaad0-aa43-4f16-a5e4-de68018f5b69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PD-Quant_[[:space:]]Post-Training[[:space:]]Quantization[[:space:]]Based[[:space:]]on[[:space:]]Prediction[[:space:]]Difference[[:space:]]Metric/c921cb99-20d4-45bc-a074-931925f5f5cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PDPP_Projected[[:space:]]Diffusion[[:space:]]for[[:space:]]Procedure[[:space:]]Planning[[:space:]]in[[:space:]]Instructional[[:space:]]Videos/931434da-984d-4006-b6e4-93085b6982a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PEAL_[[:space:]]Prior-Embedded[[:space:]]Explicit[[:space:]]Attention[[:space:]]Learning[[:space:]]for[[:space:]]Low-Overlap[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/2081f971-3e70-4df1-8c0e-535dfc38c816_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PEFAT_[[:space:]]Boosting[[:space:]]Semi-Supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Classification[[:space:]]via[[:space:]]Pseudo-Loss[[:space:]]Estimation[[:space:]]and[[:space:]]Feature[[:space:]]Adversarial[[:space:]]Training/75db3e4b-7f26-4a28-9c46-9091cdf89315_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PET-NeuS_[[:space:]]Positional[[:space:]]Encoding[[:space:]]Tri-Planes[[:space:]]for[[:space:]]Neural[[:space:]]Surfaces/6961e7c8-2e0e-45b3-927d-8d214f173e0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PHA_[[:space:]]Patch-Wise[[:space:]]High-Frequency[[:space:]]Augmentation[[:space:]]for[[:space:]]Transformer-Based[[:space:]]Person[[:space:]]Re-Identification/b9915c4a-4b1f-4dc6-addb-6aeb80d1b160_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PIDNet_[[:space:]]A[[:space:]]Real-Time[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]Network[[:space:]]Inspired[[:space:]]by[[:space:]]PID[[:space:]]Controllers/1a325660-98a5-4b57-91ac-5cc192d7b52c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PIP-Net_[[:space:]]Patch-Based[[:space:]]Intuitive[[:space:]]Prototypes[[:space:]]for[[:space:]]Interpretable[[:space:]]Image[[:space:]]Classification/a3776a97-d149-4ce3-90f4-8ae0bc9a8049_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PIRLNav_[[:space:]]Pretraining[[:space:]]With[[:space:]]Imitation[[:space:]]and[[:space:]]RL[[:space:]]Finetuning[[:space:]]for[[:space:]]ObjectNav/73aaeb17-db8a-4131-bdc2-ebcd94875ab3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PIVOT_[[:space:]]Prompting[[:space:]]for[[:space:]]Video[[:space:]]Continual[[:space:]]Learning/a47f83dc-0ca1-41dc-a61e-f1e0406f3358_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PLA_[[:space:]]Language-Driven[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/3697fc23-2ee9-439f-a0bb-9cd908563b52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PLIKS_[[:space:]]A[[:space:]]Pseudo-Linear[[:space:]]Inverse[[:space:]]Kinematic[[:space:]]Solver[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Body[[:space:]]Estimation/4c41dbaa-ba4a-4466-bc0d-792ce54d8be0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PMR_[[:space:]]Prototypical[[:space:]]Modal[[:space:]]Rebalance[[:space:]]for[[:space:]]Multimodal[[:space:]]Learning/69fc37b0-4399-4abb-8882-39dbd5dec440_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PMatch_[[:space:]]Paired[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling[[:space:]]for[[:space:]]Dense[[:space:]]Geometric[[:space:]]Matching/2d09dd24-cae4-43fc-9036-5871049bacf3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/POEM_[[:space:]]Reconstructing[[:space:]]Hand[[:space:]]in[[:space:]]a[[:space:]]Point[[:space:]]Embedded[[:space:]]Multi-View[[:space:]]Stereo/15d6d97d-a9dc-497b-af2a-06e731b779c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/POTTER_[[:space:]]Pooling[[:space:]]Attention[[:space:]]Transformer[[:space:]]for[[:space:]]Efficient[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/4a5cc1da-acde-47b7-84b4-955e233ee46a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PREIM3D_[[:space:]]3D[[:space:]]Consistent[[:space:]]Precise[[:space:]]Image[[:space:]]Attribute[[:space:]]Editing[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Image/10f5e474-3c78-4c55-8560-a24c8bd5f378_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PRISE_[[:space:]]Demystifying[[:space:]]Deep[[:space:]]Lucas-Kanade[[:space:]]With[[:space:]]Strongly[[:space:]]Star-Convex[[:space:]]Constraints[[:space:]]for[[:space:]]Multimodel[[:space:]]Image[[:space:]]Alignment/cc485109-7936-414e-b9d4-875f402facd4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PROB_[[:space:]]Probabilistic[[:space:]]Objectness[[:space:]]for[[:space:]]Open[[:space:]]World[[:space:]]Object[[:space:]]Detection/77a5758d-816b-4300-9e34-8b9157a353a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PSVT_[[:space:]]End-to-End[[:space:]]Multi-Person[[:space:]]3D[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Estimation[[:space:]]With[[:space:]]Progressive[[:space:]]Video[[:space:]]Transformers/fde0c406-e0ec-44bf-820d-6f3cd719db3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PVO_[[:space:]]Panoptic[[:space:]]Visual[[:space:]]Odometry/185b22c6-80e9-4941-a6e1-f2e34ed60e6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PVT-SSD_[[:space:]]Single-Stage[[:space:]]3D[[:space:]]Object[[:space:]]Detector[[:space:]]With[[:space:]]Point-Voxel[[:space:]]Transformer/3e0bf9f4-2867-4a16-8589-82696ad035cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PaCa-ViT_[[:space:]]Learning[[:space:]]Patch-to-Cluster[[:space:]]Attention[[:space:]]in[[:space:]]Vision[[:space:]]Transformers/35be9767-98c5-4408-9a22-513650f76547_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Paint[[:space:]]by[[:space:]]Example_[[:space:]]Exemplar-Based[[:space:]]Image[[:space:]]Editing[[:space:]]With[[:space:]]Diffusion[[:space:]]Models/48567a58-6a88-4a08-878b-9a4ac5dffe1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Painting[[:space:]]3D[[:space:]]Nature[[:space:]]in[[:space:]]2D_[[:space:]]View[[:space:]]Synthesis[[:space:]]of[[:space:]]Natural[[:space:]]Scenes[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Semantic[[:space:]]Mask/45110ca0-d2cb-40cb-aff4-49873e0d5af8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Paired-Point[[:space:]]Lifting[[:space:]]for[[:space:]]Enhanced[[:space:]]Privacy-Preserving[[:space:]]Visual[[:space:]]Localization/15df0ee8-2b51-4a6f-ae53-ee0a888b663e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PaletteNeRF_[[:space:]]Palette-Based[[:space:]]Appearance[[:space:]]Editing[[:space:]]of[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/3ad1cf6f-9d28-4a28-8311-c8a82da85967_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PanelNet_[[:space:]]Understanding[[:space:]]360[[:space:]]Indoor[[:space:]]Environment[[:space:]]via[[:space:]]Panel[[:space:]]Representation/1e91e113-dab2-4a0c-8113-1260fc9b2dc8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PanoHead_[[:space:]]Geometry-Aware[[:space:]]3D[[:space:]]Full-Head[[:space:]]Synthesis[[:space:]]in[[:space:]]360deg/bf2b5eb1-5c80-4f69-8e1b-84a96fcf4e32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PanoSwin_[[:space:]]A[[:space:]]Pano-Style[[:space:]]Swin[[:space:]]Transformer[[:space:]]for[[:space:]]Panorama[[:space:]]Understanding/d5f43da6-7dc9-4fee-ba11-3185e4ebca4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Panoptic[[:space:]]Compositional[[:space:]]Feature[[:space:]]Field[[:space:]]for[[:space:]]Editable[[:space:]]Scene[[:space:]]Rendering[[:space:]]With[[:space:]]Network-Inferred[[:space:]]Labels[[:space:]]via[[:space:]]Metric[[:space:]]Learning/3992d976-5d00-4d8f-a353-c59dc51dcc78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Panoptic[[:space:]]Lifting[[:space:]]for[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding[[:space:]]With[[:space:]]Neural[[:space:]]Fields/028caa13-6237-4e4c-b2c3-e46e4758f30d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Panoptic[[:space:]]Video[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/fd47e6dd-d308-455e-afd5-dfc6b9ad7ea8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Parallel[[:space:]]Diffusion[[:space:]]Models[[:space:]]of[[:space:]]Operator[[:space:]]and[[:space:]]Image[[:space:]]for[[:space:]]Blind[[:space:]]Inverse[[:space:]]Problems/d11c4f04-1833-48fe-9f5b-75e3c48d9c85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Parameter[[:space:]]Efficient[[:space:]]Local[[:space:]]Implicit[[:space:]]Image[[:space:]]Function[[:space:]]Network[[:space:]]for[[:space:]]Face[[:space:]]Segmentation/62ab1c50-2939-4fa0-8b69-cd25bbfa08d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Parametric[[:space:]]Implicit[[:space:]]Face[[:space:]]Representation[[:space:]]for[[:space:]]Audio-Driven[[:space:]]Facial[[:space:]]Reenactment/50f30e7f-4864-4a1f-8824-ae36bd12624b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PartDistillation_[[:space:]]Learning[[:space:]]Parts[[:space:]]From[[:space:]]Instance[[:space:]]Segmentation/8d65fabb-b809-4c7d-a23a-7918b4cb9bac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PartManip_[[:space:]]Learning[[:space:]]Cross-Category[[:space:]]Generalizable[[:space:]]Part[[:space:]]Manipulation[[:space:]]Policy[[:space:]]From[[:space:]]Point[[:space:]]Cloud[[:space:]]Observations/794963e4-c2f3-48fc-a3b9-235cb1439cd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PartMix_[[:space:]]Regularization[[:space:]]Strategy[[:space:]]To[[:space:]]Learn[[:space:]]Part[[:space:]]Discovery[[:space:]]for[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification/27b8b1ab-7a00-4dae-bc2f-554c72c9c73f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PartSLIP_[[:space:]]Low-Shot[[:space:]]Part[[:space:]]Segmentation[[:space:]]for[[:space:]]3D[[:space:]]Point[[:space:]]Clouds[[:space:]]via[[:space:]]Pretrained[[:space:]]Image-Language[[:space:]]Models/4d2e467f-5c01-4679-a823-f30f7928c9bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Partial[[:space:]]Network[[:space:]]Cloning/98669002-b035-406c-aad3-e6c37c801f42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Parts2Words_[[:space:]]Learning[[:space:]]Joint[[:space:]]Embedding[[:space:]]of[[:space:]]Point[[:space:]]Clouds[[:space:]]and[[:space:]]Texts[[:space:]]by[[:space:]]Bidirectional[[:space:]]Matching[[:space:]]Between[[:space:]]Parts[[:space:]]and[[:space:]]Words/5ee0227b-151f-49da-a01d-99b327d24af5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Passive[[:space:]]Micron-Scale[[:space:]]Time-of-Flight[[:space:]]With[[:space:]]Sunlight[[:space:]]Interferometry/a55c2f92-6092-4982-9cc5-a6760ffdb739_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Patch-Based[[:space:]]3D[[:space:]]Natural[[:space:]]Scene[[:space:]]Generation[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Example/157a2266-a52a-4785-b4bf-708f97968abe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Patch-Craft[[:space:]]Self-Supervised[[:space:]]Training[[:space:]]for[[:space:]]Correlated[[:space:]]Image[[:space:]]Denoising/724cc1a8-79e5-431a-9442-c0db988857dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Patch-Mix[[:space:]]Transformer[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation_[[:space:]]A[[:space:]]Game[[:space:]]Perspective/eab73d0d-b1fe-48a1-b7bb-c43d49e913e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PeakConv_[[:space:]]Learning[[:space:]]Peak[[:space:]]Receptive[[:space:]]Field[[:space:]]for[[:space:]]Radar[[:space:]]Semantic[[:space:]]Segmentation/1aa7ae5e-2182-46af-b81c-177e4d9d5a1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Perception[[:space:]]and[[:space:]]Semantic[[:space:]]Aware[[:space:]]Regularization[[:space:]]for[[:space:]]Sequential[[:space:]]Confidence[[:space:]]Calibration/cab9acb7-a200-4e5a-8159-2f0b479f47f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Perception-Oriented[[:space:]]Single[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]Using[[:space:]]Optimal[[:space:]]Objective[[:space:]]Estimation/addf3c94-8beb-41bf-8735-8295a8296b2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PermutoSDF_[[:space:]]Fast[[:space:]]Multi-View[[:space:]]Reconstruction[[:space:]]With[[:space:]]Implicit[[:space:]]Surfaces[[:space:]]Using[[:space:]]Permutohedral[[:space:]]Lattices/6fa30115-4f33-4798-b633-6944f2083f52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Persistent[[:space:]]Nature_[[:space:]]A[[:space:]]Generative[[:space:]]Model[[:space:]]of[[:space:]]Unbounded[[:space:]]3D[[:space:]]Worlds/01709317-ca4c-4e26-b6ea-7d92c95d2a87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Person[[:space:]]Image[[:space:]]Synthesis[[:space:]]via[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Model/b64e2f00-bc44-4a43-9bf8-8e7669c52fc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PersonNeRF_[[:space:]]Personalized[[:space:]]Reconstruction[[:space:]]From[[:space:]]Photo[[:space:]]Collections/c8e6849f-e10f-4758-b188-be618d8fb4f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Perspective[[:space:]]Fields[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Camera[[:space:]]Calibration/945b2285-d735-4ef1-83e1-4ae4639a22a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Phase-Shifting[[:space:]]Coder_[[:space:]]Predicting[[:space:]]Accurate[[:space:]]Orientation[[:space:]]in[[:space:]]Oriented[[:space:]]Object[[:space:]]Detection/93b89376-9c36-407b-9f0d-5f9a94846130_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Phone2Proc_[[:space:]]Bringing[[:space:]]Robust[[:space:]]Robots[[:space:]]Into[[:space:]]Our[[:space:]]Chaotic[[:space:]]World/5ea4913d-8965-4e55-ba5c-ca13f29397e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Photo[[:space:]]Pre-Training,[[:space:]]but[[:space:]]for[[:space:]]Sketch/243db7cf-846e-4a13-99db-3013d7f7638f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Physical-World[[:space:]]Optical[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]on[[:space:]]3D[[:space:]]Face[[:space:]]Recognition/75b9d5ba-4a00-4c91-beef-fe5dd9d24ac2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Physically[[:space:]]Adversarial[[:space:]]Infrared[[:space:]]Patches[[:space:]]With[[:space:]]Learnable[[:space:]]Shapes[[:space:]]and[[:space:]]Locations/6936181a-dfb3-4ba2-9b81-9db1892b7af4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Physically[[:space:]]Realizable[[:space:]]Natural-Looking[[:space:]]Clothing[[:space:]]Textures[[:space:]]Evade[[:space:]]Person[[:space:]]Detectors[[:space:]]via[[:space:]]3D[[:space:]]Modeling/9f3a690f-4991-40ce-8929-7df47acdfb4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Physics-Driven[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Impact[[:space:]]Sound[[:space:]]Synthesis[[:space:]]From[[:space:]]Videos/6f02946d-e13d-4e8f-8fb7-e8cd09de6cf8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Physics-Guided[[:space:]]ISO-Dependent[[:space:]]Sensor[[:space:]]Noise[[:space:]]Modeling[[:space:]]for[[:space:]]Extreme[[:space:]]Low-Light[[:space:]]Photography/e38ce0f4-4d32-4cb4-a448-5ed5a8056df9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PiMAE_[[:space:]]Point[[:space:]]Cloud[[:space:]]and[[:space:]]Image[[:space:]]Interactive[[:space:]]Masked[[:space:]]Autoencoders[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/2c8a5e6c-95bf-4176-80ee-a07d2f8b4fb8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pic2Word_[[:space:]]Mapping[[:space:]]Pictures[[:space:]]to[[:space:]]Words[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/287d6a10-4b00-4b13-abc0-3ed6e0a2172c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Picture[[:space:]]That[[:space:]]Sketch_[[:space:]]Photorealistic[[:space:]]Image[[:space:]]Generation[[:space:]]From[[:space:]]Abstract[[:space:]]Sketches/d68b3067-d946-4554-87db-6426cf95dd14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PillarNeXt_[[:space:]]Rethinking[[:space:]]Network[[:space:]]Designs[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]LiDAR[[:space:]]Point[[:space:]]Clouds/bc8d1609-268c-4879-b7d9-bc7853c665e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PivoTAL_[[:space:]]Prior-Driven[[:space:]]Supervision[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization/211c5bad-3dce-4689-a751-c72a60c77ef5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pix2map_[[:space:]]Cross-Modal[[:space:]]Retrieval[[:space:]]for[[:space:]]Inferring[[:space:]]Street[[:space:]]Maps[[:space:]]From[[:space:]]Images/121ff5e7-8b65-4fc1-b4c6-004f69b1b423_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PixHt-Lab_[[:space:]]Pixel[[:space:]]Height[[:space:]]Based[[:space:]]Light[[:space:]]Effect[[:space:]]Generation[[:space:]]for[[:space:]]Image[[:space:]]Compositing/eae5957f-225b-474e-96a7-63c36acd1cb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pixels,[[:space:]]Regions,[[:space:]]and[[:space:]]Objects_[[:space:]]Multiple[[:space:]]Enhancement[[:space:]]for[[:space:]]Salient[[:space:]]Object[[:space:]]Detection/1199c39e-4cf6-41e0-89b0-fd06a1d2fe05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PlaneDepth_[[:space:]]Self-Supervised[[:space:]]Depth[[:space:]]Estimation[[:space:]]via[[:space:]]Orthogonal[[:space:]]Planes/dc773b73-cd94-4cb0-b4ee-323aaa9d04b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Planning-Oriented[[:space:]]Autonomous[[:space:]]Driving/fa37ccdb-dc2f-40f5-8e7f-f2ace06d9f28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Plateau-Reduced[[:space:]]Differentiable[[:space:]]Path[[:space:]]Tracing/b2be5718-0e2c-4d50-af35-eb158162d83f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PlenVDB_[[:space:]]Memory[[:space:]]Efficient[[:space:]]VDB-Based[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Fast[[:space:]]Training[[:space:]]and[[:space:]]Rendering/975eb6e0-ee90-406d-98d7-f5ffe566520b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Plug-and-Play[[:space:]]Diffusion[[:space:]]Features[[:space:]]for[[:space:]]Text-Driven[[:space:]]Image-to-Image[[:space:]]Translation/0e4e4e16-e745-45c1-893c-29ea331e04ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Point[[:space:]]Cloud[[:space:]]Forecasting[[:space:]]as[[:space:]]a[[:space:]]Proxy[[:space:]]for[[:space:]]4D[[:space:]]Occupancy[[:space:]]Forecasting/7e32aed7-0d97-4de8-9ede-562ff24d9366_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Point2Pix_[[:space:]]Photo-Realistic[[:space:]]Point[[:space:]]Cloud[[:space:]]Rendering[[:space:]]via[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/4b094faa-b5e5-466d-847b-fd9bdcef51ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PointAvatar_[[:space:]]Deformable[[:space:]]Point-Based[[:space:]]Head[[:space:]]Avatars[[:space:]]From[[:space:]]Videos/2531e3ca-068a-4682-ab95-60e5d3a889f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PointCMP_[[:space:]]Contrastive[[:space:]]Mask[[:space:]]Prediction[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]on[[:space:]]Point[[:space:]]Cloud[[:space:]]Videos/9791a9d8-865e-4435-93ce-52b4c610fd5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PointCert_[[:space:]]Point[[:space:]]Cloud[[:space:]]Classification[[:space:]]With[[:space:]]Deterministic[[:space:]]Certified[[:space:]]Robustness[[:space:]]Guarantees/caa8db9e-68a2-4ccb-959e-7eb3cc117818_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PointClustering_[[:space:]]Unsupervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Pre-Training[[:space:]]Using[[:space:]]Transformation[[:space:]]Invariance[[:space:]]in[[:space:]]Clustering/a3c3387a-aac8-4244-94ba-b0ebcc35d4a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PointConvFormer_[[:space:]]Revenge[[:space:]]of[[:space:]]the[[:space:]]Point-Based[[:space:]]Convolution/649eeb8e-2d29-42d6-bf01-240c5f846dcd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PointDistiller_[[:space:]]Structured[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]Towards[[:space:]]Efficient[[:space:]]and[[:space:]]Compact[[:space:]]3D[[:space:]]Detection/9e77d6be-5d33-4258-b9d0-9d5799780858_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PointListNet_[[:space:]]Deep[[:space:]]Learning[[:space:]]on[[:space:]]3D[[:space:]]Point[[:space:]]Lists/10ea8b16-a21d-4883-b433-4f6adcc17767_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PointVector_[[:space:]]A[[:space:]]Vector[[:space:]]Representation[[:space:]]in[[:space:]]Point[[:space:]]Cloud[[:space:]]Analysis/b22afba3-c6f3-4292-9369-9ea8847f12ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pointersect_[[:space:]]Neural[[:space:]]Rendering[[:space:]]With[[:space:]]Cloud-Ray[[:space:]]Intersection/6ae832b8-abdc-4f5f-a114-274e9ca71c31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Polarimetric[[:space:]]iToF_[[:space:]]Measuring[[:space:]]High-Fidelity[[:space:]]Depth[[:space:]]Through[[:space:]]Scattering[[:space:]]Media/9d79b9e0-7a78-4fb5-9dbe-dbdac0a3f583_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Polarized[[:space:]]Color[[:space:]]Image[[:space:]]Denoising/c4dbf033-6567-4ca7-9b94-1fd7b33c26fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Policy[[:space:]]Adaptation[[:space:]]From[[:space:]]Foundation[[:space:]]Model[[:space:]]Feedback/4e55768b-e76a-4a27-9a4f-1074e881a62d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Poly-PC_[[:space:]]A[[:space:]]Polyhedral[[:space:]]Network[[:space:]]for[[:space:]]Multiple[[:space:]]Point[[:space:]]Cloud[[:space:]]Tasks[[:space:]]at[[:space:]]Once/a5a68571-985e-429c-a9e3-ed974da7c66f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PolyFormer_[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation[[:space:]]As[[:space:]]Sequential[[:space:]]Polygon[[:space:]]Generation/05f24050-1f92-432d-acaf-1892a7228be8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Polynomial[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations[[:space:]]for[[:space:]]Large[[:space:]]Diverse[[:space:]]Datasets/3d06d014-a3c0-4159-92f3-1d4321283e93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pose[[:space:]]Synchronization[[:space:]]Under[[:space:]]Multiple[[:space:]]Pair-Wise[[:space:]]Relative[[:space:]]Poses/1bf0ff8b-8639-4c19-9d91-7085534a2e2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pose-Disentangled[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Facial[[:space:]]Representation/b37ab862-2e30-48b3-be8a-6e9e3ac19dc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PoseExaminer_[[:space:]]Automated[[:space:]]Testing[[:space:]]of[[:space:]]Out-of-Distribution[[:space:]]Robustness[[:space:]]in[[:space:]]Human[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Estimation/3ad8ff0c-1959-4d81-800a-1fc3758aa0bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PoseFormerV2_[[:space:]]Exploring[[:space:]]Frequency[[:space:]]Domain[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Robust[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/41fc4754-4a4c-42f9-848d-147c6f588937_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Position-Guided[[:space:]]Text[[:space:]]Prompt[[:space:]]for[[:space:]]Vision-Language[[:space:]]Pre-Training/f86e5e20-11e4-4504-ad13-eb07a2148f11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Positive-Augmented[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Captioning[[:space:]]Evaluation/c453bca7-5123-4a80-bd9c-6ea48d3ddf8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Post-Processing[[:space:]]Temporal[[:space:]]Action[[:space:]]Detection/f0abc963-f728-4ebf-bd06-e3ca1176bc55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Post-Training[[:space:]]Quantization[[:space:]]on[[:space:]]Diffusion[[:space:]]Models/1add47f7-11b3-4d9e-927d-644c403c9ae9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PosterLayout_[[:space:]]A[[:space:]]New[[:space:]]Benchmark[[:space:]]and[[:space:]]Approach[[:space:]]for[[:space:]]Content-Aware[[:space:]]Visual-Textual[[:space:]]Presentation[[:space:]]Layout/1bbab9e5-6496-4afd-ab8d-2fe3db696603_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Power[[:space:]]Bundle[[:space:]]Adjustment[[:space:]]for[[:space:]]Large-Scale[[:space:]]3D[[:space:]]Reconstruction/c71116d8-fa59-4091-bf2b-f096826b999f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Practical[[:space:]]Network[[:space:]]Acceleration[[:space:]]With[[:space:]]Tiny[[:space:]]Sets/4100d0b1-d54b-4220-97a7-9bc0c52e26e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prefix[[:space:]]Conditioning[[:space:]]Unifies[[:space:]]Language[[:space:]]and[[:space:]]Label[[:space:]]Supervision/911f163d-f064-49a8-9f9b-5231e1ffa2a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Preserving[[:space:]]Linear[[:space:]]Separability[[:space:]]in[[:space:]]Continual[[:space:]]Learning[[:space:]]by[[:space:]]Backward[[:space:]]Feature[[:space:]]Projection/35692ff2-9c0e-4dfe-b0e1-4c021dbbcb00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Primitive[[:space:]]Generation[[:space:]]and[[:space:]]Semantic-Related[[:space:]]Alignment[[:space:]]for[[:space:]]Universal[[:space:]]Zero-Shot[[:space:]]Segmentation/4ad15e32-832f-4cbc-87ec-7e3424a7eaca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Principles[[:space:]]of[[:space:]]Forgetting[[:space:]]in[[:space:]]Domain-Incremental[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]in[[:space:]]Adverse[[:space:]]Weather[[:space:]]Conditions/825292b4-61a7-4563-ae57-a62f3b64b7c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Privacy-Preserving[[:space:]]Adversarial[[:space:]]Facial[[:space:]]Features/f9e7c149-07cb-4ff2-810a-837e89437037_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Privacy-Preserving[[:space:]]Representations[[:space:]]Are[[:space:]]Not[[:space:]]Enough_[[:space:]]Recovering[[:space:]]Scene[[:space:]]Content[[:space:]]From[[:space:]]Camera[[:space:]]Poses/4a638a35-3349-4a53-9df4-12db8d7be1a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Private[[:space:]]Image[[:space:]]Generation[[:space:]]With[[:space:]]Dual-Purpose[[:space:]]Auxiliary[[:space:]]Classifier/8383cb78-129e-48cc-bc43-9fa1b3d72785_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ProD_[[:space:]]Prompting-To-Disentangle[[:space:]]Domain[[:space:]]Knowledge[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Few-Shot[[:space:]]Image[[:space:]]Classification/a199807a-4f84-4f63-894f-87db0be6f38f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ProTeGe_[[:space:]]Untrimmed[[:space:]]Pretraining[[:space:]]for[[:space:]]Video[[:space:]]Temporal[[:space:]]Grounding[[:space:]]by[[:space:]]Video[[:space:]]Temporal[[:space:]]Grounding/c77ec4ac-0886-4321-a4b7-c8971fe56d08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Probabilistic[[:space:]]Debiasing[[:space:]]of[[:space:]]Scene[[:space:]]Graphs/42d35107-209f-4441-b471-bcf1acb2ddca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Probabilistic[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]of[[:space:]]Face[[:space:]]Ensembles/2b97c9ca-8d28-4683-903e-4d6e4a6d60d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Probabilistic[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Dense[[:space:]]Prediction/fba12666-9d2d-44ec-a5b5-46772d22449c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Probability-Based[[:space:]]Global[[:space:]]Cross-Modal[[:space:]]Upsampling[[:space:]]for[[:space:]]Pansharpening/6432263c-91f8-4f29-a599-18d33a7b198d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Probing[[:space:]]Neural[[:space:]]Representations[[:space:]]of[[:space:]]Scene[[:space:]]Perception[[:space:]]in[[:space:]]a[[:space:]]Hippocampally[[:space:]]Dependent[[:space:]]Task[[:space:]]Using[[:space:]]Artificial[[:space:]]Neural[[:space:]]Networks/7cfcea70-6c96-4ace-a5eb-18ce21121f87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Probing[[:space:]]Sentiment-Oriented[[:space:]]Pre-Training[[:space:]]Inspired[[:space:]]by[[:space:]]Human[[:space:]]Sentiment[[:space:]]Perception[[:space:]]Mechanism/bf938ded-adac-47d5-84c7-8062db189fc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Procedure-Aware[[:space:]]Pretraining[[:space:]]for[[:space:]]Instructional[[:space:]]Video[[:space:]]Understanding/b436c435-1475-4ee1-9d3f-3f0f22aaccb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Progressive[[:space:]]Backdoor[[:space:]]Erasing[[:space:]]via[[:space:]]Connecting[[:space:]]Backdoor[[:space:]]and[[:space:]]Adversarial[[:space:]]Attacks/37e60e53-4829-453b-91ef-ba899ea61c68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Progressive[[:space:]]Disentangled[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Controllable[[:space:]]Talking[[:space:]]Head[[:space:]]Synthesis/ceb18cff-405d-498a-88f0-9ea8fb0fc28b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Progressive[[:space:]]Neighbor[[:space:]]Consistency[[:space:]]Mining[[:space:]]for[[:space:]]Correspondence[[:space:]]Pruning/62a3d7ee-6533-4952-8c47-4dcc73bec882_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Progressive[[:space:]]Open[[:space:]]Space[[:space:]]Expansion[[:space:]]for[[:space:]]Open-Set[[:space:]]Model[[:space:]]Attribution/05749c20-48f0-46f9-b026-bbf29ab4e9c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Progressive[[:space:]]Random[[:space:]]Convolutions[[:space:]]for[[:space:]]Single[[:space:]]Domain[[:space:]]Generalization/c1def00f-daf9-42b1-9ab7-045a6684f5da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Progressive[[:space:]]Semantic-Visual[[:space:]]Mutual[[:space:]]Adaption[[:space:]]for[[:space:]]Generalized[[:space:]]Zero-Shot[[:space:]]Learning/33d73b96-4ecf-41bf-8a6c-d33c90a496e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Progressive[[:space:]]Spatio-Temporal[[:space:]]Alignment[[:space:]]for[[:space:]]Efficient[[:space:]]Event-Based[[:space:]]Motion[[:space:]]Estimation/a03a8307-dfb5-4033-ad22-e5d7229cff0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Progressive[[:space:]]Transformation[[:space:]]Learning[[:space:]]for[[:space:]]Leveraging[[:space:]]Virtual[[:space:]]Images[[:space:]]in[[:space:]]Training/0974526a-d141-4d50-b9b4-9abb2ecc2505_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Progressively[[:space:]]Optimized[[:space:]]Local[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Robust[[:space:]]View[[:space:]]Synthesis/76c4aaa4-c56a-4b40-a1b3-1d3f0b70c336_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Promoting[[:space:]]Semantic[[:space:]]Connectivity_[[:space:]]Dual[[:space:]]Nearest[[:space:]]Neighbors[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Generalization/0fc9d943-fb46-41da-9b01-7002b4821fd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prompt,[[:space:]]Generate,[[:space:]]Then[[:space:]]Cache_[[:space:]]Cascade[[:space:]]of[[:space:]]Foundation[[:space:]]Models[[:space:]]Makes[[:space:]]Strong[[:space:]]Few-Shot[[:space:]]Learners/34933cc6-e0db-44c7-ba7d-c07eac395b5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prompt-Guided[[:space:]]Zero-Shot[[:space:]]Anomaly[[:space:]]Action[[:space:]]Recognition[[:space:]]Using[[:space:]]Pretrained[[:space:]]Deep[[:space:]]Skeleton[[:space:]]Features/ffddf195-4612-4a09-a6b4-c79b7cba79ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PromptCAL_[[:space:]]Contrastive[[:space:]]Affinity[[:space:]]Learning[[:space:]]via[[:space:]]Auxiliary[[:space:]]Prompts[[:space:]]for[[:space:]]Generalized[[:space:]]Novel[[:space:]]Category[[:space:]]Discovery/a0b56506-f91e-4c1f-9a2c-b81d2fba90e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prompting[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]With[[:space:]]Answer[[:space:]]Heuristics[[:space:]]for[[:space:]]Knowledge-Based[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/52e4dea4-060d-4709-b23d-80023fc0b4d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Propagate[[:space:]]and[[:space:]]Calibrate_[[:space:]]Real-Time[[:space:]]Passive[[:space:]]Non-Line-of-Sight[[:space:]]Tracking/cf490751-8db2-414a-9fa0-0baebad808a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ProphNet_[[:space:]]Efficient[[:space:]]Agent-Centric[[:space:]]Motion[[:space:]]Forecasting[[:space:]]With[[:space:]]Anchor-Informed[[:space:]]Proposals/998fdcb9-43ea-4bd1-b86d-be211eccd0f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Proposal-Based[[:space:]]Multiple[[:space:]]Instance[[:space:]]Learning[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization/404dd7be-4845-4b53-a4ff-005f33e7c16e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ProtoCon_[[:space:]]Pseudo-Label[[:space:]]Refinement[[:space:]]via[[:space:]]Online[[:space:]]Clustering[[:space:]]and[[:space:]]Prototypical[[:space:]]Consistency[[:space:]]for[[:space:]]Efficient[[:space:]]Semi-Supervised[[:space:]]Learning/66235807-534b-4594-89ca-7853a11b6c49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prototype-Based[[:space:]]Embedding[[:space:]]Network[[:space:]]for[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/2d6cf2dd-b1ed-41e7-b2b4-55a87a371f26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prototypical[[:space:]]Residual[[:space:]]Networks[[:space:]]for[[:space:]]Anomaly[[:space:]]Detection[[:space:]]and[[:space:]]Localization/cd9babc3-5191-478f-85ef-9cd2b6600b9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Proximal[[:space:]]Splitting[[:space:]]Adversarial[[:space:]]Attack[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/fd9f7619-b3db-4e8c-8f44-83dcd5d58ace_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ProxyFormer_[[:space:]]Proxy[[:space:]]Alignment[[:space:]]Assisted[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion[[:space:]]With[[:space:]]Missing[[:space:]]Part[[:space:]]Sensitive[[:space:]]Transformer/0e78b0e6-ec55-4264-b773-da1fb2a3e9db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pruning[[:space:]]Parameterization[[:space:]]With[[:space:]]Bi-Level[[:space:]]Optimization[[:space:]]for[[:space:]]Efficient[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]on[[:space:]]the[[:space:]]Edge/cca6cc27-1a84-41ca-a8b9-554d65aaa66d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pseudo-Label[[:space:]]Guided[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/5925d7ac-3859-49a5-af0c-db774b1a6bc5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Putting[[:space:]]People[[:space:]]in[[:space:]]Their[[:space:]]Place_[[:space:]]Affordance-Aware[[:space:]]Human[[:space:]]Insertion[[:space:]]Into[[:space:]]Scenes/1adb80a3-cbb6-4ec7-bf53-f805dd8cd4ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PyPose_[[:space:]]A[[:space:]]Library[[:space:]]for[[:space:]]Robot[[:space:]]Learning[[:space:]]With[[:space:]]Physics-Based[[:space:]]Optimization/8ab4278c-d90b-4d90-8c18-c4270d9b669e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PyramidFlow_[[:space:]]High-Resolution[[:space:]]Defect[[:space:]]Contrastive[[:space:]]Localization[[:space:]]Using[[:space:]]Pyramid[[:space:]]Normalizing[[:space:]]Flow/53c7bbfc-6160-4d44-8847-ae5a0e4a3ce1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Q-DETR_[[:space:]]An[[:space:]]Efficient[[:space:]]Low-Bit[[:space:]]Quantized[[:space:]]Detection[[:space:]]Transformer/758f64fe-cfe8-4994-bd86-745fec52e6d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/QPGesture_[[:space:]]Quantization-Based[[:space:]]and[[:space:]]Phase-Guided[[:space:]]Motion[[:space:]]Matching[[:space:]]for[[:space:]]Natural[[:space:]]Speech-Driven[[:space:]]Gesture[[:space:]]Generation/dedf7a8a-5cf9-4beb-aa1f-59750898d75f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Q_[[:space:]]How[[:space:]]To[[:space:]]Specialize[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]to[[:space:]]Data-Scarce[[:space:]]VQA[[:space:]]Tasks_[[:space:]]A_[[:space:]]Self-Train[[:space:]]on[[:space:]]Unlabeled[[:space:]]Images!/cca6ed83-7faa-4a3c-a2de-be98beed4aed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Quality-Aware[[:space:]]Pre-Trained[[:space:]]Models[[:space:]]for[[:space:]]Blind[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment/8d41dca5-5365-455a-824a-8a417d4e7839_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/QuantArt_[[:space:]]Quantizing[[:space:]]Image[[:space:]]Style[[:space:]]Transfer[[:space:]]Towards[[:space:]]High[[:space:]]Visual[[:space:]]Fidelity/4935fc7e-77d7-46b7-8112-ea03d0af8f8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Quantitative[[:space:]]Manipulation[[:space:]]of[[:space:]]Custom[[:space:]]Attributes[[:space:]]on[[:space:]]3D-Aware[[:space:]]Image[[:space:]]Synthesis/79420605-e587-4a04-9757-13abf9958858_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Quantum[[:space:]]Multi-Model[[:space:]]Fitting/aabc5258-c568-488e-9b55-06bc7747cb81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Quantum-Inspired[[:space:]]Spectral-Spatial[[:space:]]Pyramid[[:space:]]Network[[:space:]]for[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Classification/f89e1a74-57a2-4551-b632-b3352ab84a1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Query-Centric[[:space:]]Trajectory[[:space:]]Prediction/c4e71a3e-e61f-4cf4-92ef-2463bb7fd739_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Query-Dependent[[:space:]]Video[[:space:]]Representation[[:space:]]for[[:space:]]Moment[[:space:]]Retrieval[[:space:]]and[[:space:]]Highlight[[:space:]]Detection/244f80b9-7720-4627-a98e-406a817fbf7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/R2Former_[[:space:]]Unified[[:space:]]Retrieval[[:space:]]and[[:space:]]Reranking[[:space:]]Transformer[[:space:]]for[[:space:]]Place[[:space:]]Recognition/2ca4068a-0a96-471c-b512-c2d7b65af9b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RA-CLIP_[[:space:]]Retrieval[[:space:]]Augmented[[:space:]]Contrastive[[:space:]]Language-Image[[:space:]]Pre-Training/459b769d-d1fe-4e25-9ca1-08382651f316_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/REC-MV_[[:space:]]REconstructing[[:space:]]3D[[:space:]]Dynamic[[:space:]]Cloth[[:space:]]From[[:space:]]Monocular[[:space:]]Videos/dc096b27-b9fa-424b-94dc-3477e0e0fc0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/REVEAL_[[:space:]]Retrieval-Augmented[[:space:]]Visual-Language[[:space:]]Pre-Training[[:space:]]With[[:space:]]Multi-Source[[:space:]]Multimodal[[:space:]]Knowledge[[:space:]]Memory/ec28663c-769a-4a6f-bc32-f0643a00e7d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RGB[[:space:]]No[[:space:]]More_[[:space:]]Minimally-Decoded[[:space:]]JPEG[[:space:]]Vision[[:space:]]Transformers/2da81411-0f32-4a64-8564-8877ef6902a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RGBD2_[[:space:]]Generative[[:space:]]Scene[[:space:]]Synthesis[[:space:]]via[[:space:]]Incremental[[:space:]]View[[:space:]]Inpainting[[:space:]]Using[[:space:]]RGBD[[:space:]]Diffusion[[:space:]]Models/a2d21046-a37a-4e32-b58b-aa13bd3ca238_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RIATIG_[[:space:]]Reliable[[:space:]]and[[:space:]]Imperceptible[[:space:]]Adversarial[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]With[[:space:]]Natural[[:space:]]Prompts/caa635c4-0250-4102-a574-c4d125a4109d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RIAV-MVS_[[:space:]]Recurrent-Indexing[[:space:]]an[[:space:]]Asymmetric[[:space:]]Volume[[:space:]]for[[:space:]]Multi-View[[:space:]]Stereo/04aab7d4-b43b-496c-94a4-596185183726_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RIDCP_[[:space:]]Revitalizing[[:space:]]Real[[:space:]]Image[[:space:]]Dehazing[[:space:]]via[[:space:]]High-Quality[[:space:]]Codebook[[:space:]]Priors/05659269-f519-4ff3-9c15-95d37cd713bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RIFormer_[[:space:]]Keep[[:space:]]Your[[:space:]]Vision[[:space:]]Backbone[[:space:]]Effective[[:space:]]but[[:space:]]Removing[[:space:]]Token[[:space:]]Mixer/5e2f8802-cd09-4641-a7fa-0692dcfb5f55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RILS_[[:space:]]Masked[[:space:]]Visual[[:space:]]Reconstruction[[:space:]]in[[:space:]]Language[[:space:]]Semantic[[:space:]]Space/74117c37-24aa-429a-8711-0623ef6a8c64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RMLVQA_[[:space:]]A[[:space:]]Margin[[:space:]]Loss[[:space:]]Approach[[:space:]]for[[:space:]]Visual[[:space:]]Question[[:space:]]Answering[[:space:]]With[[:space:]]Language[[:space:]]Biases/211c42c4-defc-46bd-b55e-243b99c5b0bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RODIN_[[:space:]]A[[:space:]]Generative[[:space:]]Model[[:space:]]for[[:space:]]Sculpting[[:space:]]3D[[:space:]]Digital[[:space:]]Avatars[[:space:]]Using[[:space:]]Diffusion/39108314-e784-42b7-854a-19be46573cc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RONO_[[:space:]]Robust[[:space:]]Discriminative[[:space:]]Learning[[:space:]]With[[:space:]]Noisy[[:space:]]Labels[[:space:]]for[[:space:]]2D-3D[[:space:]]Cross-Modal[[:space:]]Retrieval/27b72c7c-b71d-416b-9de2-a59860d5ddfe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RUST_[[:space:]]Latent[[:space:]]Neural[[:space:]]Scene[[:space:]]Representations[[:space:]]From[[:space:]]Unposed[[:space:]]Imagery/07e323fd-0ee1-490f-90cd-f65d4d015578_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RWSC-Fusion_[[:space:]]Region-Wise[[:space:]]Style-Controlled[[:space:]]Fusion[[:space:]]Network[[:space:]]for[[:space:]]the[[:space:]]Prohibited[[:space:]]X-Ray[[:space:]]Security[[:space:]]Image[[:space:]]Synthesis/4229304a-c9a9-4e95-92bc-4ee682221af2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RaBit_[[:space:]]Parametric[[:space:]]Modeling[[:space:]]of[[:space:]]3D[[:space:]]Biped[[:space:]]Cartoon[[:space:]]Characters[[:space:]]With[[:space:]]a[[:space:]]Topological-Consistent[[:space:]]Dataset/1e9c4a16-13c2-47a3-ab7f-6611c669588a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Randomized[[:space:]]Adversarial[[:space:]]Training[[:space:]]via[[:space:]]Taylor[[:space:]]Expansion/67c6de64-e0fe-4f60-a88b-da38f54fe451_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Range-Nullspace[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation[[:space:]]With[[:space:]]Focalized[[:space:]]Motion[[:space:]]Estimation/fb8c4ec0-3368-4408-98c9-491162515b01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RangeViT_[[:space:]]Towards[[:space:]]Vision[[:space:]]Transformers[[:space:]]for[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/94b91fe6-a01e-4a69-8957-2c035de25886_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RankMix_[[:space:]]Data[[:space:]]Augmentation[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Learning[[:space:]]of[[:space:]]Classifying[[:space:]]Whole[[:space:]]Slide[[:space:]]Images[[:space:]]With[[:space:]]Diverse[[:space:]]Sizes[[:space:]]and[[:space:]]Imbalanced[[:space:]]Categories/9892a862-63a2-48f4-8e5a-1651be4bb51a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ranking[[:space:]]Regularization[[:space:]]for[[:space:]]Critical[[:space:]]Rare[[:space:]]Classes_[[:space:]]Minimizing[[:space:]]False[[:space:]]Positives[[:space:]]at[[:space:]]a[[:space:]]High[[:space:]]True[[:space:]]Positive[[:space:]]Rate/b62e2938-5b12-4774-98d6-2d9d2cd9fe90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rate[[:space:]]Gradient[[:space:]]Approximation[[:space:]]Attack[[:space:]]Threats[[:space:]]Deep[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks/5607e0f0-af41-4d3b-ab58-e27bce5086e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Raw[[:space:]]Image[[:space:]]Reconstruction[[:space:]]With[[:space:]]Learned[[:space:]]Compact[[:space:]]Metadata/8175d738-aa28-456a-99d1-d591bc04940f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rawgment_[[:space:]]Noise-Accounted[[:space:]]RAW[[:space:]]Augmentation[[:space:]]Enables[[:space:]]Recognition[[:space:]]in[[:space:]]a[[:space:]]Wide[[:space:]]Variety[[:space:]]of[[:space:]]Environments/98100afc-c86e-439b-98cd-4768d6e60a6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Re-Basin[[:space:]]via[[:space:]]Implicit[[:space:]]Sinkhorn[[:space:]]Differentiation/dc7a818b-aa01-4c23-bb8a-90b1d6f90212_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Re-GAN_[[:space:]]Data-Efficient[[:space:]]GANs[[:space:]]Training[[:space:]]via[[:space:]]Architectural[[:space:]]Reconfiguration/33c6c6a6-066b-4807-b712-d7b820960347_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Re-IQA_[[:space:]]Unsupervised[[:space:]]Learning[[:space:]]for[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment[[:space:]]in[[:space:]]the[[:space:]]Wild/90946db8-a492-4f97-b0fd-e3875b8d588a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Re-Thinking[[:space:]]Federated[[:space:]]Active[[:space:]]Learning[[:space:]]Based[[:space:]]on[[:space:]]Inter-Class[[:space:]]Diversity/f3aeecdc-b140-4a20-9d11-a2f96511d465_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Re-Thinking[[:space:]]Model[[:space:]]Inversion[[:space:]]Attacks[[:space:]]Against[[:space:]]Deep[[:space:]]Neural[[:space:]]Networks/b89c7e51-fdb6-461b-acc1-402a12f4a254_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Re2TAL_[[:space:]]Rewiring[[:space:]]Pretrained[[:space:]]Video[[:space:]]Backbones[[:space:]]for[[:space:]]Reversible[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization/a7a6ab97-064a-45bd-98df-b72107741eea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ReCo_[[:space:]]Region-Controlled[[:space:]]Text-to-Image[[:space:]]Generation/7a126cec-3dba-421c-a0e8-e03b6864db48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ReDirTrans_[[:space:]]Latent-to-Latent[[:space:]]Translation[[:space:]]for[[:space:]]Gaze[[:space:]]and[[:space:]]Head[[:space:]]Redirection/743e5765-daf6-4f6d-bed7-fe1d28d26c28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ReLight[[:space:]]My[[:space:]]NeRF_[[:space:]]A[[:space:]]Dataset[[:space:]]for[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]and[[:space:]]Relighting[[:space:]]of[[:space:]]Real[[:space:]]World[[:space:]]Objects/17ade2bb-9ad4-48d5-b13f-aa219a03e831_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ReVISE_[[:space:]]Self-Supervised[[:space:]]Speech[[:space:]]Resynthesis[[:space:]]With[[:space:]]Visual[[:space:]]Input[[:space:]]for[[:space:]]Universal[[:space:]]and[[:space:]]Generalized[[:space:]]Speech[[:space:]]Regeneration/4c656091-f78e-4996-8f21-a624cfc397a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Real-Time[[:space:]]6K[[:space:]]Image[[:space:]]Rescaling[[:space:]]With[[:space:]]Rate-Distortion[[:space:]]Optimization/cfbc14d6-4e76-42f6-ba1a-9d45e253283f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Real-Time[[:space:]]Controllable[[:space:]]Denoising[[:space:]]for[[:space:]]Image[[:space:]]and[[:space:]]Video/64f70f4f-f112-41dc-9862-a65c1c758802_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Real-Time[[:space:]]Evaluation[[:space:]]in[[:space:]]Online[[:space:]]Continual[[:space:]]Learning_[[:space:]]A[[:space:]]New[[:space:]]Hope/90670a33-6068-465e-a4fd-294478f9cef7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Real-Time[[:space:]]Multi-Person[[:space:]]Eyeblink[[:space:]]Detection[[:space:]]in[[:space:]]the[[:space:]]Wild[[:space:]]for[[:space:]]Untrimmed[[:space:]]Video/ae1d0533-ce28-4e72-a6ef-27b8f2c1f68e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Real-Time[[:space:]]Neural[[:space:]]Light[[:space:]]Field[[:space:]]on[[:space:]]Mobile[[:space:]]Devices/7cb4e48b-56d2-4da2-9743-cd9d9c1bc604_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RealFusion_[[:space:]]360deg[[:space:]]Reconstruction[[:space:]]of[[:space:]]Any[[:space:]]Object[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Image/cb32ed0a-6d89-4ef5-bddb-3a82c64b06eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RealImpact_[[:space:]]A[[:space:]]Dataset[[:space:]]of[[:space:]]Impact[[:space:]]Sound[[:space:]]Fields[[:space:]]for[[:space:]]Real[[:space:]]Objects/c9f7d18c-491d-4679-a453-2aa53d9cacc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Realistic[[:space:]]Saliency[[:space:]]Guided[[:space:]]Image[[:space:]]Enhancement/8e3ac21f-6601-42a5-a41c-6bcc6930276d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ReasonNet_[[:space:]]End-to-End[[:space:]]Driving[[:space:]]With[[:space:]]Temporal[[:space:]]and[[:space:]]Global[[:space:]]Reasoning/1fe839d8-63ad-45a1-a78b-7253a9b6afed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rebalancing[[:space:]]Batch[[:space:]]Normalization[[:space:]]for[[:space:]]Exemplar-Based[[:space:]]Class-Incremental[[:space:]]Learning/5041066b-589d-4c40-92f0-14b7543c2197_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Recognizability[[:space:]]Embedding[[:space:]]Enhancement[[:space:]]for[[:space:]]Very[[:space:]]Low-Resolution[[:space:]]Face[[:space:]]Recognition[[:space:]]and[[:space:]]Quality[[:space:]]Estimation/252ce840-97f6-49d3-b68f-e8e81b6429b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Recognizing[[:space:]]Rigid[[:space:]]Patterns[[:space:]]of[[:space:]]Unlabeled[[:space:]]Point[[:space:]]Clouds[[:space:]]by[[:space:]]Complete[[:space:]]and[[:space:]]Continuous[[:space:]]Isometry[[:space:]]Invariants[[:space:]]With[[:space:]]No[[:space:]]False[[:space:]]Negatives[[:space:]]and[[:space:]]No[[:space:]]False[[:space:]]Positives/e95e50bd-aaeb-4127-84a7-6d0f727b579f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reconstructing[[:space:]]Animatable[[:space:]]Categories[[:space:]]From[[:space:]]Videos/152172cb-1c63-4d7d-b307-3d80c30f42e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reconstructing[[:space:]]Signing[[:space:]]Avatars[[:space:]]From[[:space:]]Video[[:space:]]Using[[:space:]]Linguistic[[:space:]]Priors/f36e0963-996f-4cc3-af93-ef4d550fb47d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Recovering[[:space:]]3D[[:space:]]Hand[[:space:]]Mesh[[:space:]]Sequence[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Blurry[[:space:]]Image_[[:space:]]A[[:space:]]New[[:space:]]Dataset[[:space:]]and[[:space:]]Temporal[[:space:]]Unfolding/b7a63c82-2ead-41e3-b7d7-3cbeee09de5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Recurrence[[:space:]]Without[[:space:]]Recurrence_[[:space:]]Stable[[:space:]]Video[[:space:]]Landmark[[:space:]]Detection[[:space:]]With[[:space:]]Deep[[:space:]]Equilibrium[[:space:]]Models/75fbb2dc-eb4e-427c-bdb2-0594b83943c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Recurrent[[:space:]]Homography[[:space:]]Estimation[[:space:]]Using[[:space:]]Homography-Guided[[:space:]]Image[[:space:]]Warping[[:space:]]and[[:space:]]Focus[[:space:]]Transformer/b2d36598-87bb-4351-86c5-8b1cc41ad2e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Recurrent[[:space:]]Vision[[:space:]]Transformers[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]Event[[:space:]]Cameras/6fcdd9e3-da92-40b0-9e0a-068ff912bee6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reducing[[:space:]]the[[:space:]]Label[[:space:]]Bias[[:space:]]for[[:space:]]Timestamp[[:space:]]Supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Segmentation/986159f2-45a0-4edb-a4fb-548f08a70e7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ref-NPR_[[:space:]]Reference-Based[[:space:]]Non-Photorealistic[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Controllable[[:space:]]Scene[[:space:]]Stylization/c04ee8e8-49a4-4d83-bf71-ecc9b04aa76d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RefCLIP_[[:space:]]A[[:space:]]Universal[[:space:]]Teacher[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Referring[[:space:]]Expression[[:space:]]Comprehension/253a7eb5-c7b6-49fc-afc9-2e6811ab4226_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RefSR-NeRF_[[:space:]]Towards[[:space:]]High[[:space:]]Fidelity[[:space:]]and[[:space:]]Super[[:space:]]Resolution[[:space:]]View[[:space:]]Synthesis/322e0316-e942-4e0b-b791-bc4ea8d2687d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RefTeacher_[[:space:]]A[[:space:]]Strong[[:space:]]Baseline[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Referring[[:space:]]Expression[[:space:]]Comprehension/6dde83b9-4b57-48ed-afaf-f9e793553791_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Referring[[:space:]]Image[[:space:]]Matting/493302cf-0d67-4381-8bd8-a4db01f54348_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Referring[[:space:]]Multi-Object[[:space:]]Tracking/bdefcdeb-2794-4c06-83c0-50cbc928a6f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Region-Aware[[:space:]]Pretraining[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]Vision[[:space:]]Transformers/9b5f5230-8834-4e27-acff-5c487c7cb759_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Regularization[[:space:]]of[[:space:]]Polynomial[[:space:]]Networks[[:space:]]for[[:space:]]Image[[:space:]]Recognition/5d0c0110-c6a9-4ad7-9427-c47a82da5015_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Regularize[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representation[[:space:]]by[[:space:]]Itself/ffc63975-6ef9-4108-bbd2-b598701104a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Regularized[[:space:]]Vector[[:space:]]Quantization[[:space:]]for[[:space:]]Tokenized[[:space:]]Image[[:space:]]Synthesis/125660f9-839f-4555-beee-6c3da1c16776_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Regularizing[[:space:]]Second-Order[[:space:]]Influences[[:space:]]for[[:space:]]Continual[[:space:]]Learning/84328c03-d9ce-42d3-ad77-eace225839f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reinforcement[[:space:]]Learning-Based[[:space:]]Black-Box[[:space:]]Model[[:space:]]Inversion[[:space:]]Attacks/ecefb72a-e6ac-45a7-abf7-97524d6d4ffd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Relational[[:space:]]Context[[:space:]]Learning[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection/7bddb9c0-8fe4-44cd-a0c6-28bfbe306c52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Relational[[:space:]]Space-Time[[:space:]]Query[[:space:]]in[[:space:]]Long-Form[[:space:]]Videos/9909db19-7af3-468c-9223-98fba1f23f01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reliability[[:space:]]in[[:space:]]Semantic[[:space:]]Segmentation_[[:space:]]Are[[:space:]]We[[:space:]]on[[:space:]]the[[:space:]]Right[[:space:]]Track_/6499a1a1-d25a-44af-bf55-d0cd749f522f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reliable[[:space:]]and[[:space:]]Interpretable[[:space:]]Personalized[[:space:]]Federated[[:space:]]Learning/36094269-58c9-4cc4-9406-059d2dc8cd11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Relightable[[:space:]]Neural[[:space:]]Human[[:space:]]Assets[[:space:]]From[[:space:]]Multi-View[[:space:]]Gradient[[:space:]]Illuminations/7da77560-1865-44eb-a571-c0340f5f5c4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RelightableHands_[[:space:]]Efficient[[:space:]]Neural[[:space:]]Relighting[[:space:]]of[[:space:]]Articulated[[:space:]]Hand[[:space:]]Models/e75e4c06-6e5e-41de-a35d-2589321d86e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Removing[[:space:]]Objects[[:space:]]From[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/f0b147cf-ec66-4dab-98e4-cb2e725ef33a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RenderDiffusion_[[:space:]]Image[[:space:]]Diffusion[[:space:]]for[[:space:]]3D[[:space:]]Reconstruction,[[:space:]]Inpainting[[:space:]]and[[:space:]]Generation/fc8c13d2-a422-428d-a582-4ef42d40e56d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Renderable[[:space:]]Neural[[:space:]]Radiance[[:space:]]Map[[:space:]]for[[:space:]]Visual[[:space:]]Navigation/96d7fb59-b61a-495f-915e-2faf1f51777f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RepMode_[[:space:]]Learning[[:space:]]to[[:space:]]Re-Parameterize[[:space:]]Diverse[[:space:]]Experts[[:space:]]for[[:space:]]Subcellular[[:space:]]Structure[[:space:]]Prediction/4a9bc772-6d78-41a9-95e9-ead31161ea40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Representation[[:space:]]Learning[[:space:]]for[[:space:]]Visual[[:space:]]Object[[:space:]]Tracking[[:space:]]by[[:space:]]Masked[[:space:]]Appearance[[:space:]]Transfer/53c7f367-9678-4633-9aa2-17845fa77669_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Representing[[:space:]]Volumetric[[:space:]]Videos[[:space:]]As[[:space:]]Dynamic[[:space:]]MLP[[:space:]]Maps/fd329f8f-a686-4343-92df-04c995e2bfc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reproducible[[:space:]]Scaling[[:space:]]Laws[[:space:]]for[[:space:]]Contrastive[[:space:]]Language-Image[[:space:]]Learning/55cd3801-c733-4aa1-b5e7-a8296cc16ed2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ResFormer_[[:space:]]Scaling[[:space:]]ViTs[[:space:]]With[[:space:]]Multi-Resolution[[:space:]]Training/8d72be06-8e52-4da0-a5b1-9bcf7daeba1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Residual[[:space:]]Degradation[[:space:]]Learning[[:space:]]Unfolding[[:space:]]Framework[[:space:]]With[[:space:]]Mixing[[:space:]]Priors[[:space:]]Across[[:space:]]Spectral[[:space:]]and[[:space:]]Spatial[[:space:]]for[[:space:]]Compressive[[:space:]]Spectral[[:space:]]Imaging/a0938212-00d4-442d-911e-687d17f3420b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Resource-Efficient[[:space:]]RGBD[[:space:]]Aerial[[:space:]]Tracking/13b9c515-365c-408f-8c69-27e31d15b262_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Restoration[[:space:]]of[[:space:]]Hand-Drawn[[:space:]]Architectural[[:space:]]Drawings[[:space:]]Using[[:space:]]Latent[[:space:]]Space[[:space:]]Mapping[[:space:]]With[[:space:]]Degradation[[:space:]]Generator/59f790f0-c2fc-4075-b8be-b0c3138c9563_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Domain[[:space:]]Generalization[[:space:]]for[[:space:]]Face[[:space:]]Anti-Spoofing_[[:space:]]Separability[[:space:]]and[[:space:]]Alignment/1a5bf716-14d6-4bae-87a8-69a63162f63a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Feature-Based[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Face[[:space:]]Recognition/9d6cc8ff-97fc-494a-942f-25d90713ac90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Federated[[:space:]]Learning[[:space:]]With[[:space:]]Domain[[:space:]]Shift_[[:space:]]A[[:space:]]Prototype[[:space:]]View/fa8c5828-91e4-4b50-8915-da59ea10e8e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Few-Shot[[:space:]]Medical[[:space:]]Segmentation_[[:space:]]A[[:space:]]Vector[[:space:]]Quantization[[:space:]]View/46939912-025e-40ca-8869-77270629f6ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Gradient[[:space:]]Projection[[:space:]]Continual[[:space:]]Learning_[[:space:]]Stability[[:space:]]_[[:space:]]Plasticity[[:space:]]Feature[[:space:]]Space[[:space:]]Decoupling/da85e080-fba3-4de1-9043-9ec610fcb116_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Image[[:space:]]Super[[:space:]]Resolution[[:space:]]From[[:space:]]Long-Tailed[[:space:]]Distribution[[:space:]]Learning[[:space:]]Perspective/e9ed0720-58d1-4117-a3b2-470e31fae34e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Optical[[:space:]]Flow[[:space:]]From[[:space:]]Geometric[[:space:]]Matching[[:space:]]Consistent[[:space:]]Perspective/1fe97f27-3285-4f9c-8395-902b00f51ecb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Out-of-Distribution[[:space:]](OOD)[[:space:]]Detection_[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling[[:space:]]Is[[:space:]]All[[:space:]]You[[:space:]]Need/a01c9c7a-47b3-4e4d-ae9b-cdfdb11ecdb3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Video[[:space:]]ViTs_[[:space:]]Sparse[[:space:]]Video[[:space:]]Tubes[[:space:]]for[[:space:]]Joint[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Learning/35a8008f-803e-4992-a096-34962783b29d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]the[[:space:]]Approximation[[:space:]]Error[[:space:]]in[[:space:]]3D[[:space:]]Surface[[:space:]]Fitting[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Normal[[:space:]]Estimation/5a900b34-e66a-4c13-a5e8-b711b00f6819_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]the[[:space:]]Correlation[[:space:]]in[[:space:]]Few-Shot[[:space:]]Segmentation_[[:space:]]A[[:space:]]Buoys[[:space:]]View/4eefa9a7-70d0-4e9f-948d-7896e65ee884_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]the[[:space:]]Learning[[:space:]]Paradigm[[:space:]]for[[:space:]]Dynamic[[:space:]]Facial[[:space:]]Expression[[:space:]]Recognition/6cf474ad-10c4-4fad-8c7d-8e8811042fa2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revealing[[:space:]]the[[:space:]]Dark[[:space:]]Secrets[[:space:]]of[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling/d250ca74-0412-4902-a2eb-9cd0db128a15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Multimodal[[:space:]]Representation[[:space:]]in[[:space:]]Contrastive[[:space:]]Learning_[[:space:]]From[[:space:]]Patch[[:space:]]and[[:space:]]Token[[:space:]]Embeddings[[:space:]]to[[:space:]]Finite[[:space:]]Discrete[[:space:]]Tokens/0f3e55c7-8a8d-44aa-8f68-c7119a01cfaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Prototypical[[:space:]]Network[[:space:]]for[[:space:]]Cross[[:space:]]Domain[[:space:]]Few-Shot[[:space:]]Learning/ae638f98-f465-49c4-94f1-dcbac6f3c3b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Residual[[:space:]]Networks[[:space:]]for[[:space:]]Adversarial[[:space:]]Robustness/ecf01e55-2328-4452-92e8-241959338ee6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Reverse[[:space:]]Distillation[[:space:]]for[[:space:]]Anomaly[[:space:]]Detection/2aab2945-0b72-4431-aa0f-b240376f0e18_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Rolling[[:space:]]Shutter[[:space:]]Bundle[[:space:]]Adjustment_[[:space:]]Toward[[:space:]]Accurate[[:space:]]and[[:space:]]Fast[[:space:]]Solution/867a8358-af5b-4c86-a04c-653d22e8eea0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Rotation[[:space:]]Averaging_[[:space:]]Uncertainties[[:space:]]and[[:space:]]Robust[[:space:]]Losses/8bc74694-cc18-4136-9f7a-320ca0767dc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Self-Similarity_[[:space:]]Structural[[:space:]]Embedding[[:space:]]for[[:space:]]Image[[:space:]]Retrieval/10846c03-37d5-4e54-a2ff-ad9d2c8bf39b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Temporal[[:space:]]Modeling[[:space:]]for[[:space:]]CLIP-Based[[:space:]]Image-to-Video[[:space:]]Knowledge[[:space:]]Transferring/015a515d-32ed-4995-879a-8746a567ad72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Weak-to-Strong[[:space:]]Consistency[[:space:]]in[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation/894a4b65-87ae-4357-8b17-908ac9adc507_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]the[[:space:]]P3P[[:space:]]Problem/3c7c0744-6bca-4297-965d-15ff597a0250_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]the[[:space:]]Stack-Based[[:space:]]Inverse[[:space:]]Tone[[:space:]]Mapping/e364db15-c88a-4840-8ba3-a6fd39714cd0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rigidity-Aware[[:space:]]Detection[[:space:]]for[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/25287fba-246f-44b6-9233-6aa4efe490cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robot[[:space:]]Structure[[:space:]]Prior[[:space:]]Guided[[:space:]]Temporal[[:space:]]Attention[[:space:]]for[[:space:]]Camera-to-Robot[[:space:]]Pose[[:space:]]Estimation[[:space:]]From[[:space:]]Image[[:space:]]Sequence/d1059c34-7695-408a-b23f-8642d8c0068b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]3D[[:space:]]Shape[[:space:]]Classification[[:space:]]via[[:space:]]Non-Local[[:space:]]Graph[[:space:]]Attention[[:space:]]Network/fbd17e7e-5cac-4e0d-ab9b-537ed826d5b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Dynamic[[:space:]]Radiance[[:space:]]Fields/d9bc5d06-b3a5-4464-8f03-e0c34a90f3b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Generalization[[:space:]]Against[[:space:]]Photon-Limited[[:space:]]Corruptions[[:space:]]via[[:space:]]Worst-Case[[:space:]]Sharpness[[:space:]]Minimization/cecef2e6-239c-4022-847c-e977ae0db8c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Mean[[:space:]]Teacher[[:space:]]for[[:space:]]Continual[[:space:]]and[[:space:]]Gradual[[:space:]]Test-Time[[:space:]]Adaptation/fbb3fd05-2972-41c9-a8f9-7710d5bd2830_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Model-Based[[:space:]]Face[[:space:]]Reconstruction[[:space:]]Through[[:space:]]Weakly-Supervised[[:space:]]Outlier[[:space:]]Segmentation/517bb6d0-df52-4672-9e64-97429eaf1fa3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Multiview[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration[[:space:]]With[[:space:]]Reliable[[:space:]]Pose[[:space:]]Graph[[:space:]]Initialization[[:space:]]and[[:space:]]History[[:space:]]Reweighting/3de61b08-df6f-4aaa-8ad5-5f44dc92514d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Outlier[[:space:]]Rejection[[:space:]]for[[:space:]]3D[[:space:]]Registration[[:space:]]With[[:space:]]Variational[[:space:]]Bayes/65a0dc15-9eca-45e4-8931-aeda13fc78f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Single[[:space:]]Image[[:space:]]Reflection[[:space:]]Removal[[:space:]]Against[[:space:]]Adversarial[[:space:]]Attacks/9967a3ae-116f-48d0-8a30-d1102ebae7fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]in[[:space:]]Dynamic[[:space:]]Scenarios/18d0ec2a-ba7b-4e7c-8b52-b26b805e62f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Unsupervised[[:space:]]StyleGAN[[:space:]]Image[[:space:]]Restoration/22cf3860-c0a9-4cc3-ae76-e2aceeacd7e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]and[[:space:]]Scalable[[:space:]]Gaussian[[:space:]]Process[[:space:]]Regression[[:space:]]and[[:space:]]Its[[:space:]]Applications/def98ceb-97d5-4394-a2b2-443884eb66dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RobustNeRF_[[:space:]]Ignoring[[:space:]]Distractors[[:space:]]With[[:space:]]Robust[[:space:]]Losses/2f329418-58fe-4ad7-b462-6d0d0ce341f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Role[[:space:]]of[[:space:]]Transients[[:space:]]in[[:space:]]Two-Bounce[[:space:]]Non-Line-of-Sight[[:space:]]Imaging/dd8b3d32-1f20-46bb-9a36-d731b9d46ed2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rotation-Invariant[[:space:]]Transformer[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Matching/424bbd8d-fe04-4142-91bc-ea8eb4c92ccb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Run,[[:space:]]Don't[[:space:]]Walk_[[:space:]]Chasing[[:space:]]Higher[[:space:]]FLOPS[[:space:]]for[[:space:]]Faster[[:space:]]Neural[[:space:]]Networks/6a2198d1-cecd-4fd5-b373-fd2aeb690abb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/S3C_[[:space:]]Semi-Supervised[[:space:]]VQA[[:space:]]Natural[[:space:]]Language[[:space:]]Explanation[[:space:]]via[[:space:]]Self-Critical[[:space:]]Learning/6ead34e8-75d6-4d08-a4ab-1da1402619c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SAP-DETR_[[:space:]]Bridging[[:space:]]the[[:space:]]Gap[[:space:]]Between[[:space:]]Salient[[:space:]]Points[[:space:]]and[[:space:]]Queries-Based[[:space:]]Transformer[[:space:]]Detector[[:space:]]for[[:space:]]Fast[[:space:]]Model[[:space:]]Convergency/909bbbe4-4e26-472f-ab37-1d60d8bab4b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SCADE_[[:space:]]NeRFs[[:space:]]from[[:space:]]Space[[:space:]]Carving[[:space:]]With[[:space:]]Ambiguity-Aware[[:space:]]Depth[[:space:]]Estimates/cb201944-e311-4584-9847-096959c3a596_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SCConv_[[:space:]]Spatial[[:space:]]and[[:space:]]Channel[[:space:]]Reconstruction[[:space:]]Convolution[[:space:]]for[[:space:]]Feature[[:space:]]Redundancy/cde54b77-44c6-4dda-8761-9969150d4bec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SCOOP_[[:space:]]Self-Supervised[[:space:]]Correspondence[[:space:]]and[[:space:]]Optimization-Based[[:space:]]Scene[[:space:]]Flow/2d0fc651-4e0b-4b27-8a33-61605dae5a3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SCOTCH[[:space:]]and[[:space:]]SODA_[[:space:]]A[[:space:]]Transformer[[:space:]]Video[[:space:]]Shadow[[:space:]]Detection[[:space:]]Framework/db4a242e-46af-47cc-9782-646b5ef8b2a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SCPNet_[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion[[:space:]]on[[:space:]]Point[[:space:]]Cloud/a1134f69-6375-42b9-9a92-868bfcf809a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SCoDA_[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Shape[[:space:]]Completion[[:space:]]for[[:space:]]Real[[:space:]]Scans/a2de3a81-2619-4975-9f69-e49ad0b14643_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SDC-UDA_[[:space:]]Volumetric[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]Framework[[:space:]]for[[:space:]]Slice-Direction[[:space:]]Continuous[[:space:]]Cross-Modality[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/20148cfd-bd1f-4726-9d6d-b13924bbcc1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SDFusion_[[:space:]]Multimodal[[:space:]]3D[[:space:]]Shape[[:space:]]Completion,[[:space:]]Reconstruction,[[:space:]]and[[:space:]]Generation/f29ee3dd-8b63-4289-b6a4-d6e103a0e949_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SE-ORNet_[[:space:]]Self-Ensembling[[:space:]]Orientation-Aware[[:space:]]Network[[:space:]]for[[:space:]]Unsupervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Shape[[:space:]]Correspondence/94ea60f5-77d0-47dc-8f2c-3fe79d2f954a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SECAD-Net_[[:space:]]Self-Supervised[[:space:]]CAD[[:space:]]Reconstruction[[:space:]]by[[:space:]]Learning[[:space:]]Sketch-Extrude[[:space:]]Operations/94d1495e-0d51-4b2c-8690-ac04468c5645_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SFD2_[[:space:]]Semantic-Guided[[:space:]]Feature[[:space:]]Detection[[:space:]]and[[:space:]]Description/6efa713f-1ae0-4651-bc92-c61912feb28d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SGLoc_[[:space:]]Scene[[:space:]]Geometry[[:space:]]Encoding[[:space:]]for[[:space:]]Outdoor[[:space:]]LiDAR[[:space:]]Localization/56c4bde5-7fa2-4496-a50a-885b12091caa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SHS-Net_[[:space:]]Learning[[:space:]]Signed[[:space:]]Hyper[[:space:]]Surfaces[[:space:]]for[[:space:]]Oriented[[:space:]]Normal[[:space:]]Estimation[[:space:]]of[[:space:]]Point[[:space:]]Clouds/b88f2048-c087-4b9c-9601-f88e3009b262_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SIEDOB_[[:space:]]Semantic[[:space:]]Image[[:space:]]Editing[[:space:]]by[[:space:]]Disentangling[[:space:]]Object[[:space:]]and[[:space:]]Background/a369b37c-8e06-4d08-a268-de9d54816ab4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SIM_[[:space:]]Semantic-Aware[[:space:]]Instance[[:space:]]Mask[[:space:]]Generation[[:space:]]for[[:space:]]Box-Supervised[[:space:]]Instance[[:space:]]Segmentation/884f8e5b-a009-4236-8afe-619abfb36ed7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SINE_[[:space:]]SINgle[[:space:]]Image[[:space:]]Editing[[:space:]]With[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/9e4aea54-e956-45c7-8e0d-e18b2b0dde98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SINE_[[:space:]]Semantic-Driven[[:space:]]Image-Based[[:space:]]NeRF[[:space:]]Editing[[:space:]]With[[:space:]]Prior-Guided[[:space:]]Editing[[:space:]]Field/98bd5d41-9994-4c5a-92c8-200bf021b945_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SLACK_[[:space:]]Stable[[:space:]]Learning[[:space:]]of[[:space:]]Augmentations[[:space:]]With[[:space:]]Cold-Start[[:space:]]and[[:space:]]KL[[:space:]]Regularization/13983024-bdf8-451a-82ba-7c4e7f6f3424_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SLOPER4D_[[:space:]]A[[:space:]]Scene-Aware[[:space:]]Dataset[[:space:]]for[[:space:]]Global[[:space:]]4D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]in[[:space:]]Urban[[:space:]]Environments/6eb8c194-2d4e-4f05-826e-be6c8b0450e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SMAE_[[:space:]]Few-Shot[[:space:]]Learning[[:space:]]for[[:space:]]HDR[[:space:]]Deghosting[[:space:]]With[[:space:]]Saturation-Aware[[:space:]]Masked[[:space:]]Autoencoders/1480ea63-f1bc-4488-8d0b-6bc8008633c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SMOC-Net_[[:space:]]Leveraging[[:space:]]Camera[[:space:]]Pose[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Monocular[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/7ed77cdd-0ec8-4cc4-8097-ec9f36276521_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SMPConv_[[:space:]]Self-Moving[[:space:]]Point[[:space:]]Representations[[:space:]]for[[:space:]]Continuous[[:space:]]Convolution/a40d3fb1-72be-4c1d-a333-984a29892bca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SOOD_[[:space:]]Towards[[:space:]]Semi-Supervised[[:space:]]Oriented[[:space:]]Object[[:space:]]Detection/b6c29c1e-51a1-4935-bf5e-ec2e26810dec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SPARF_[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]From[[:space:]]Sparse[[:space:]]and[[:space:]]Noisy[[:space:]]Poses/9768904d-8df6-48f0-a9ea-13cb333d26ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SPIn-NeRF_[[:space:]]Multiview[[:space:]]Segmentation[[:space:]]and[[:space:]]Perceptual[[:space:]]Inpainting[[:space:]]With[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/a5f839e7-988d-4934-9aef-4c494df6e55a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SQUID_[[:space:]]Deep[[:space:]]Feature[[:space:]]In-Painting[[:space:]]for[[:space:]]Unsupervised[[:space:]]Anomaly[[:space:]]Detection/c1d458f7-f8ad-480d-9ab9-60ca7bc1a759_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/STAR[[:space:]]Loss_[[:space:]]Reducing[[:space:]]Semantic[[:space:]]Ambiguity[[:space:]]in[[:space:]]Facial[[:space:]]Landmark[[:space:]]Detection/fb30f024-6bbe-44bb-baaf-0659043475bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/STDLens_[[:space:]]Model[[:space:]]Hijacking-Resilient[[:space:]]Federated[[:space:]]Learning[[:space:]]for[[:space:]]Object[[:space:]]Detection/970c86e7-a18e-4fa1-bf10-060bfc2535dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/STMT_[[:space:]]A[[:space:]]Spatial-Temporal[[:space:]]Mesh[[:space:]]Transformer[[:space:]]for[[:space:]]MoCap-Based[[:space:]]Action[[:space:]]Recognition/ecec9e25-9648-4975-9559-d389674e900b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/STMixer_[[:space:]]A[[:space:]]One-Stage[[:space:]]Sparse[[:space:]]Action[[:space:]]Detector/7eef5975-ffbf-46c0-a31b-5b464eef7eef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SUDS_[[:space:]]Scalable[[:space:]]Urban[[:space:]]Dynamic[[:space:]]Scenes/e8fd5f67-d301-4dd3-8588-d4088862c6a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SVFormer_[[:space:]]Semi-Supervised[[:space:]]Video[[:space:]]Transformer[[:space:]]for[[:space:]]Action[[:space:]]Recognition/2bd91319-3d84-428a-8335-5833e2e27f2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SVGformer_[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Continuous[[:space:]]Vector[[:space:]]Graphics[[:space:]]Using[[:space:]]Transformers/3621baa2-936b-4fa2-8381-1968bf4d3112_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SViTT_[[:space:]]Temporal[[:space:]]Learning[[:space:]]of[[:space:]]Sparse[[:space:]]Video-Text[[:space:]]Transformers/a239fc53-51e4-48aa-8bdd-b397217aa71a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SadTalker_[[:space:]]Learning[[:space:]]Realistic[[:space:]]3D[[:space:]]Motion[[:space:]]Coefficients[[:space:]]for[[:space:]]Stylized[[:space:]]Audio-Driven[[:space:]]Single[[:space:]]Image[[:space:]]Talking[[:space:]]Face[[:space:]]Animation/aac86acd-650e-49a0-a046-054265e0fe1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Safe[[:space:]]Latent[[:space:]]Diffusion_[[:space:]]Mitigating[[:space:]]Inappropriate[[:space:]]Degeneration[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/27635077-d485-4fba-8dfd-7e03eb609fe0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sample-Level[[:space:]]Multi-View[[:space:]]Graph[[:space:]]Clustering/2dbc917c-e6ba-403e-80e3-a2fe15c249a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Samples[[:space:]]With[[:space:]]Low[[:space:]]Loss[[:space:]]Curvature[[:space:]]Improve[[:space:]]Data[[:space:]]Efficiency/0d4bc9ae-1b91-4517-8e70-147be86f88d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sampling[[:space:]]Is[[:space:]]Matter_[[:space:]]Point-Guided[[:space:]]3D[[:space:]]Human[[:space:]]Mesh[[:space:]]Reconstruction/ede83042-4d23-4ef9-a09b-ef8f9683611d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scalable,[[:space:]]Detailed[[:space:]]and[[:space:]]Mask-Free[[:space:]]Universal[[:space:]]Photometric[[:space:]]Stereo/b4164742-7be0-4231-ae8a-d7fbbefac394_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ScaleDet_[[:space:]]A[[:space:]]Scalable[[:space:]]Multi-Dataset[[:space:]]Object[[:space:]]Detector/483d2038-de2f-43cc-b2e7-cfb0359bee59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ScaleFL_[[:space:]]Resource-Adaptive[[:space:]]Federated[[:space:]]Learning[[:space:]]With[[:space:]]Heterogeneous[[:space:]]Clients/cb97df44-ce50-4c4d-9f39-8532b06bcaf1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ScaleKD_[[:space:]]Distilling[[:space:]]Scale-Aware[[:space:]]Knowledge[[:space:]]in[[:space:]]Small[[:space:]]Object[[:space:]]Detector/f4a95fb4-74c7-47ae-867c-7193aca0ac49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scaling[[:space:]]Language-Image[[:space:]]Pre-Training[[:space:]]via[[:space:]]Masking/72402de8-32e2-4c16-9bec-c8e1554c3fec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scaling[[:space:]]Up[[:space:]]GANs[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Synthesis/cc6d67d2-1178-4cd7-a0a4-6ace5123194c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ScanDMM_[[:space:]]A[[:space:]]Deep[[:space:]]Markov[[:space:]]Model[[:space:]]of[[:space:]]Scanpath[[:space:]]Prediction[[:space:]]for[[:space:]]360deg[[:space:]]Images/0e9438ec-529d-4504-8952-ad1887816ae6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ScarceNet_[[:space:]]Animal[[:space:]]Pose[[:space:]]Estimation[[:space:]]With[[:space:]]Scarce[[:space:]]Annotations/b512070b-870b-466d-ae4d-037690406953_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scene-Aware[[:space:]]Egocentric[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/3a95f507-702b-4e70-9e9b-d135557d5128_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SceneComposer_[[:space:]]Any-Level[[:space:]]Semantic[[:space:]]Image[[:space:]]Synthesis/a420eecc-24ac-471a-873a-4cbedabb0ab3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SceneTrilogy_[[:space:]]On[[:space:]]Human[[:space:]]Scene-Sketch[[:space:]]and[[:space:]]Its[[:space:]]Complementarity[[:space:]]With[[:space:]]Photo[[:space:]]and[[:space:]]Text/61c06fa2-0cf3-4cfc-a6de-25fbe7140dea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Score[[:space:]]Jacobian[[:space:]]Chaining_[[:space:]]Lifting[[:space:]]Pretrained[[:space:]]2D[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]3D[[:space:]]Generation/6ed9361a-e1bd-4c7d-893b-e63657bb23f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SeSDF_[[:space:]]Self-Evolved[[:space:]]Signed[[:space:]]Distance[[:space:]]Field[[:space:]]for[[:space:]]Implicit[[:space:]]3D[[:space:]]Clothed[[:space:]]Human[[:space:]]Reconstruction/568571cd-b563-47cc-be1f-30a698863adb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SeaThru-NeRF_[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]in[[:space:]]Scattering[[:space:]]Media/769a6e31-fc05-4ad4-bfda-b4fc158792ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Search-Map-Search_[[:space:]]A[[:space:]]Frame[[:space:]]Selection[[:space:]]Paradigm[[:space:]]for[[:space:]]Action[[:space:]]Recognition/72f3e04e-8ac3-45ed-bc77-477759a59e7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Seasoning[[:space:]]Model[[:space:]]Soups[[:space:]]for[[:space:]]Robustness[[:space:]]to[[:space:]]Adversarial[[:space:]]and[[:space:]]Natural[[:space:]]Distribution[[:space:]]Shifts/a6551453-1b77-4f6c-8df2-df18da134652_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Seeing[[:space:]]Beyond[[:space:]]the[[:space:]]Brain_[[:space:]]Conditional[[:space:]]Diffusion[[:space:]]Model[[:space:]]With[[:space:]]Sparse[[:space:]]Masked[[:space:]]Modeling[[:space:]]for[[:space:]]Vision[[:space:]]Decoding/fe08ae3e-3e6a-4967-96af-a2a0ffcad985_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Seeing[[:space:]]Through[[:space:]]the[[:space:]]Glass_[[:space:]]Neural[[:space:]]3D[[:space:]]Reconstruction[[:space:]]of[[:space:]]Object[[:space:]]Inside[[:space:]]a[[:space:]]Transparent[[:space:]]Container/26a8b2b4-a047-4aaa-a0e2-44b1f430ae84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Seeing[[:space:]]What[[:space:]]You[[:space:]]Miss_[[:space:]]Vision-Language[[:space:]]Pre-Training[[:space:]]With[[:space:]]Semantic[[:space:]]Completion[[:space:]]Learning/0dd82c17-0312-4de0-b06f-dd673d65e2b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Seeing[[:space:]]What[[:space:]]You[[:space:]]Said_[[:space:]]Talking[[:space:]]Face[[:space:]]Generation[[:space:]]Guided[[:space:]]by[[:space:]]a[[:space:]]Lip[[:space:]]Reading[[:space:]]Expert/f2651081-de91-4969-9dc4-565d10268fa2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Seeing[[:space:]]With[[:space:]]Sound_[[:space:]]Long-range[[:space:]]Acoustic[[:space:]]Beamforming[[:space:]]for[[:space:]]Multimodal[[:space:]]Scene[[:space:]]Understanding/fb091460-c1f3-4f4b-8f38-0da80d2617f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Seeing[[:space:]]a[[:space:]]Rose[[:space:]]in[[:space:]]Five[[:space:]]Thousand[[:space:]]Ways/d0859bbf-ed4b-4b6f-b924-83d3d600aaa3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Seeing'[[:space:]]Electric[[:space:]]Network[[:space:]]Frequency[[:space:]]From[[:space:]]Events/fe087304-aaeb-4b83-95f5-482f99b48d8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SegLoc_[[:space:]]Learning[[:space:]]Segmentation-Based[[:space:]]Representations[[:space:]]for[[:space:]]Privacy-Preserving[[:space:]]Visual[[:space:]]Localization/1cfa13a5-5a76-441e-9316-8ffeeadc7f19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Selective[[:space:]]Structured[[:space:]]State-Spaces[[:space:]]for[[:space:]]Long-Form[[:space:]]Video[[:space:]]Understanding/14c595cf-4963-4c6f-838c-aab6ba8e961d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Correctable[[:space:]]and[[:space:]]Adaptable[[:space:]]Inference[[:space:]]for[[:space:]]Generalizable[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/e99739e0-6424-49cc-bde8-f14fe85209d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Guided[[:space:]]Diffusion[[:space:]]Models/f7a1fbce-08b7-4a3e-bec6-f7f210a43bd4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Positioning[[:space:]]Point-Based[[:space:]]Transformer[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Understanding/ae9a2ced-fcbe-45f2-a028-30bec6cdce9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]3D[[:space:]]Scene[[:space:]]Flow[[:space:]]Estimation[[:space:]]Guided[[:space:]]by[[:space:]]Superpoints/59074c27-a020-40fd-a1b1-c4a4788b0f16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]AutoFlow/0468082f-9c1d-403d-ad48-bbd9972dc8a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Blind[[:space:]]Motion[[:space:]]Deblurring[[:space:]]With[[:space:]]Deep[[:space:]]Expectation[[:space:]]Maximization/01193d4a-7f2d-48c2-88a4-024ab66bb42e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Geometry-Aware[[:space:]]Encoder[[:space:]]for[[:space:]]Style-Based[[:space:]]3D[[:space:]]GAN[[:space:]]Inversion/93c57374-2532-4035-bc13-a000f9c980f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Image-to-Point[[:space:]]Distillation[[:space:]]via[[:space:]]Semantically[[:space:]]Tolerant[[:space:]]Contrastive[[:space:]]Loss/72e831f8-3c98-4889-bef1-a7e6e1731c8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Implicit[[:space:]]Glyph[[:space:]]Attention[[:space:]]for[[:space:]]Text[[:space:]]Recognition/d5fd183f-8dc2-4f32-b64c-07ecf54c5770_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Learning[[:space:]]From[[:space:]]Images[[:space:]]With[[:space:]]a[[:space:]]Joint-Embedding[[:space:]]Predictive[[:space:]]Architecture/2b14fc12-ce4b-4811-862f-99ba5bb50838_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Learning[[:space:]]for[[:space:]]Multimodal[[:space:]]Non-Rigid[[:space:]]3D[[:space:]]Shape[[:space:]]Matching/ccfe9ad5-7e9e-443b-9e12-06ccbcdc5e2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Non-Uniform[[:space:]]Kernel[[:space:]]Estimation[[:space:]]With[[:space:]]Flow-Based[[:space:]]Motion[[:space:]]Prior[[:space:]]for[[:space:]]Blind[[:space:]]Image[[:space:]]Deblurring/2c611962-37ed-4603-a449-0bd89aa64f24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Pre-Training[[:space:]]With[[:space:]]Masked[[:space:]]Shape[[:space:]]Prediction[[:space:]]for[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/8dee13eb-cc38-4b8c-80e6-9a5b48c6dadf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]CAD/4cc5a563-9690-458a-942b-a0d794493843_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Super-Plane[[:space:]]for[[:space:]]Neural[[:space:]]3D[[:space:]]Reconstruction/6e2d50ea-8ddc-42bb-8631-0ae6da7c60f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Video[[:space:]]Forensics[[:space:]]by[[:space:]]Audio-Visual[[:space:]]Anomaly[[:space:]]Detection/9a89e524-e7c2-4032-8f90-0dd8582e8fba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SelfME_[[:space:]]Self-Supervised[[:space:]]Motion[[:space:]]Learning[[:space:]]for[[:space:]]Micro-Expression[[:space:]]Recognition/b8214c0c-f617-479d-b314-a1181bda1897_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantic[[:space:]]Human[[:space:]]Parsing[[:space:]]via[[:space:]]Scalable[[:space:]]Semantic[[:space:]]Transfer[[:space:]]Over[[:space:]]Multiple[[:space:]]Label[[:space:]]Domains/ca102109-47d7-426b-b797-0ec2656b93f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantic[[:space:]]Ray_[[:space:]]Learning[[:space:]]a[[:space:]]Generalizable[[:space:]]Semantic[[:space:]]Field[[:space:]]With[[:space:]]Cross-Reprojection[[:space:]]Attention/f510e92b-c278-468a-98e5-fc928bb9d5b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantic[[:space:]]Scene[[:space:]]Completion[[:space:]]With[[:space:]]Cleaner[[:space:]]Self/b3db9b4f-80fc-4094-ae52-a4ba8502cd15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantic-Conditional[[:space:]]Diffusion[[:space:]]Networks[[:space:]]for[[:space:]]Image[[:space:]]Captioning/42eb23ca-5837-404f-a3e2-9c1d25d13215_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantic-Promoted[[:space:]]Debiasing[[:space:]]and[[:space:]]Background[[:space:]]Disambiguation[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Instance[[:space:]]Segmentation/3f3328dd-16a2-4514-926c-a52e422820ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-DETR_[[:space:]]Semi-Supervised[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]Detection[[:space:]]Transformers/4a665759-3896-41b6-90fe-effc63e754b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-Supervised[[:space:]]2D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]Driven[[:space:]]by[[:space:]]Position[[:space:]]Inconsistency[[:space:]]Pseudo[[:space:]]Label[[:space:]]Correction[[:space:]]Module/745b2939-dd3d-40e7-94da-6bb23dc90579_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-Supervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]With[[:space:]]Source[[:space:]]Label[[:space:]]Adaptation/475a5cdb-74cd-4405-8d4b-5cfa25ae25de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-Supervised[[:space:]]Hand[[:space:]]Appearance[[:space:]]Recovery[[:space:]]via[[:space:]]Structure[[:space:]]Disentanglement[[:space:]]and[[:space:]]Dual[[:space:]]Adversarial[[:space:]]Discrimination/fad467ff-879b-4cf1-8172-0343e6bcc3a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-Supervised[[:space:]]Learning[[:space:]]Made[[:space:]]Simple[[:space:]]With[[:space:]]Self-Supervised[[:space:]]Clustering/0e9efe28-710e-44d5-a83e-be7d49cfe986_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-Supervised[[:space:]]Parametric[[:space:]]Real-World[[:space:]]Image[[:space:]]Harmonization/f37ef072-b25a-4fa1-9f0a-d73638fa191a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-Supervised[[:space:]]Stereo-Based[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Cross-View[[:space:]]Consensus/561e552e-fa80-4b73-b5eb-ebc18269d4a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-Supervised[[:space:]]Video[[:space:]]Inpainting[[:space:]]With[[:space:]]Cycle[[:space:]]Consistency[[:space:]]Constraints/d1301105-29db-499a-b8ca-ca8df809c4b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-Weakly[[:space:]]Supervised[[:space:]]Object[[:space:]]Kinematic[[:space:]]Motion[[:space:]]Prediction/9e498cd8-36b3-488e-b44f-4b1d0cba5932_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SemiCVT_[[:space:]]Semi-Supervised[[:space:]]Convolutional[[:space:]]Vision[[:space:]]Transformer[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/937af7f5-c66b-43a5-8f2c-f0f00cbf4dcc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semidefinite[[:space:]]Relaxations[[:space:]]for[[:space:]]Robust[[:space:]]Multiview[[:space:]]Triangulation/48dd7206-cf0a-4e46-978e-d7e4dacfaafd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SeqTrack_[[:space:]]Sequence[[:space:]]to[[:space:]]Sequence[[:space:]]Learning[[:space:]]for[[:space:]]Visual[[:space:]]Object[[:space:]]Tracking/2a5c28d4-421e-4353-a18b-8806e4f4f180_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sequential[[:space:]]Training[[:space:]]of[[:space:]]GANs[[:space:]]Against[[:space:]]GAN-Classifiers[[:space:]]Reveals[[:space:]]Correlated[[:space:]]'Knowledge[[:space:]]Gaps'[[:space:]]Present[[:space:]]Among[[:space:]]Independently[[:space:]]Trained[[:space:]]GAN[[:space:]]Instances/f28dc581-33d7-45aa-acef-e350e328fe35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SfM-TTR_[[:space:]]Using[[:space:]]Structure[[:space:]]From[[:space:]]Motion[[:space:]]for[[:space:]]Test-Time[[:space:]]Refinement[[:space:]]of[[:space:]]Single-View[[:space:]]Depth[[:space:]]Networks/66822fab-0c0d-4ef7-a809-98c7e4012617_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ShadowDiffusion_[[:space:]]When[[:space:]]Degradation[[:space:]]Prior[[:space:]]Meets[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Shadow[[:space:]]Removal/3a1eccd7-262e-476f-ac9f-5fe889f56a8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ShadowNeuS_[[:space:]]Neural[[:space:]]SDF[[:space:]]Reconstruction[[:space:]]by[[:space:]]Shadow[[:space:]]Ray[[:space:]]Supervision/2faa6ae9-ac25-427f-a3c7-489f0e778ad9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shakes[[:space:]]on[[:space:]]a[[:space:]]Plane_[[:space:]]Unsupervised[[:space:]]Depth[[:space:]]Estimation[[:space:]]From[[:space:]]Unstabilized[[:space:]]Photography/1c92787e-15de-4a3b-8cf7-9963ea7ae2f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shape,[[:space:]]Pose,[[:space:]]and[[:space:]]Appearance[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]via[[:space:]]Bootstrapped[[:space:]]Radiance[[:space:]]Field[[:space:]]Inversion/f516b852-24c7-4f1d-a420-fbe9a6fae21d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shape-Aware[[:space:]]Text-Driven[[:space:]]Layered[[:space:]]Video[[:space:]]Editing/53846ee3-d438-4adf-be44-475d60acb994_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shape-Constraint[[:space:]]Recurrent[[:space:]]Flow[[:space:]]for[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/82f73773-b79f-4aa1-9f58-ccbefab952dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shape-Erased[[:space:]]Feature[[:space:]]Learning[[:space:]]for[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification/7ad164a0-bf09-4a6b-a7ac-872ea9e73c95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ShapeClipper_[[:space:]]Scalable[[:space:]]3D[[:space:]]Shape[[:space:]]Learning[[:space:]]From[[:space:]]Single-View[[:space:]]Images[[:space:]]via[[:space:]]Geometric[[:space:]]and[[:space:]]CLIP-Based[[:space:]]Consistency/0e2b39ff-b7cd-420b-8e94-cbccf774342d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ShapeTalk_[[:space:]]A[[:space:]]Language[[:space:]]Dataset[[:space:]]and[[:space:]]Framework[[:space:]]for[[:space:]]3D[[:space:]]Shape[[:space:]]Edits[[:space:]]and[[:space:]]Deformations/4468048a-b2b4-4115-a04f-bfc70f557c92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sharpness-Aware[[:space:]]Gradient[[:space:]]Matching[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/f6d9422f-9af8-4d97-b70b-a7dbbfcf9309_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shepherding[[:space:]]Slots[[:space:]]to[[:space:]]Objects_[[:space:]]Towards[[:space:]]Stable[[:space:]]and[[:space:]]Robust[[:space:]]Object-Centric[[:space:]]Learning/1a8f1387-bdcc-4715-9fc7-e01f069a1b9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shifted[[:space:]]Diffusion[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/3145c022-6d74-44cd-b233-9127ba843ffe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shortcomings[[:space:]]of[[:space:]]Top-Down[[:space:]]Randomization-Based[[:space:]]Sanity[[:space:]]Checks[[:space:]]for[[:space:]]Evaluations[[:space:]]of[[:space:]]Deep[[:space:]]Neural[[:space:]]Network[[:space:]]Explanations/2c0d762d-6883-47a8-9b20-4b93c21f1cfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Siamese[[:space:]]DETR/8c4c5d20-8553-4557-882d-88aa888befde_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Siamese[[:space:]]Image[[:space:]]Modeling[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Vision[[:space:]]Representation[[:space:]]Learning/9520e8f2-03c6-4904-9940-5bd97195c055_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sibling-Attack_[[:space:]]Rethinking[[:space:]]Transferable[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]Against[[:space:]]Face[[:space:]]Recognition/55038231-bffc-4bef-af3e-4c51741f8e70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Side[[:space:]]Adapter[[:space:]]Network[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/73852c8b-e982-467f-bffd-8f0bd9a07165_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Similarity[[:space:]]Maps[[:space:]]for[[:space:]]Self-Training[[:space:]]Weakly-Supervised[[:space:]]Phrase[[:space:]]Grounding/45ed3500-ed75-4c01-a48d-7e288ea844a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Similarity[[:space:]]Metric[[:space:]]Learning[[:space:]]for[[:space:]]RGB-Infrared[[:space:]]Group[[:space:]]Re-Identification/7f06c06a-3bc9-4c01-9629-fc40ba2da4b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SimpSON_[[:space:]]Simplifying[[:space:]]Photo[[:space:]]Cleanup[[:space:]]With[[:space:]]Single-Click[[:space:]]Distracting[[:space:]]Object[[:space:]]Segmentation[[:space:]]Network/bf805583-69d4-4923-ba15-d7df38269c5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Simple[[:space:]]Cues[[:space:]]Lead[[:space:]]to[[:space:]]a[[:space:]]Strong[[:space:]]Multi-Object[[:space:]]Tracker/999e696c-402d-4ec1-a2d3-bd7547b33bd9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SimpleNet_[[:space:]]A[[:space:]]Simple[[:space:]]Network[[:space:]]for[[:space:]]Image[[:space:]]Anomaly[[:space:]]Detection[[:space:]]and[[:space:]]Localization/6d5f841b-5c2d-4403-89d7-b4df42246b20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Simulated[[:space:]]Annealing[[:space:]]in[[:space:]]Early[[:space:]]Layers[[:space:]]Leads[[:space:]]to[[:space:]]Better[[:space:]]Generalization/848a9107-dad4-45f5-be6c-ac381e3d3ef2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Simultaneously[[:space:]]Short-[[:space:]]and[[:space:]]Long-Term[[:space:]]Temporal[[:space:]]Modeling[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Video[[:space:]]Semantic[[:space:]]Segmentation/c12ab650-b94e-4a61-9b08-cad2fe8bf36a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SinGRAF_[[:space:]]Learning[[:space:]]a[[:space:]]3D[[:space:]]Generative[[:space:]]Radiance[[:space:]]Field[[:space:]]for[[:space:]]a[[:space:]]Single[[:space:]]Scene/7b1cbab2-8fc0-4d4b-94a0-76a5dcdfe8bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Single[[:space:]]Domain[[:space:]]Generalization[[:space:]]for[[:space:]]LiDAR[[:space:]]Semantic[[:space:]]Segmentation/f7c54ffe-11fa-4c2e-9868-5b055a1deed4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Single[[:space:]]Image[[:space:]]Backdoor[[:space:]]Inversion[[:space:]]via[[:space:]]Robust[[:space:]]Smoothed[[:space:]]Classifiers/53ac1889-601f-4745-8836-7706cae4d524_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Single[[:space:]]Image[[:space:]]Depth[[:space:]]Prediction[[:space:]]Made[[:space:]]Better_[[:space:]]A[[:space:]]Multivariate[[:space:]]Gaussian[[:space:]]Take/6b4566b9-e782-421c-abeb-eb6a1c93b027_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Single[[:space:]]View[[:space:]]Scene[[:space:]]Scale[[:space:]]Estimation[[:space:]]Using[[:space:]]Scale[[:space:]]Field/3d921de9-7bde-42c3-9dc0-3e457e90a1d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sketch2Saliency_[[:space:]]Learning[[:space:]]To[[:space:]]Detect[[:space:]]Salient[[:space:]]Objects[[:space:]]From[[:space:]]Human[[:space:]]Drawings/0d42bab1-7cf6-4ecd-95ea-42c9e99b8061_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SketchXAI_[[:space:]]A[[:space:]]First[[:space:]]Look[[:space:]]at[[:space:]]Explainability[[:space:]]for[[:space:]]Human[[:space:]]Sketches/2c1a37e0-b57f-4b2e-ad69-44ecac6a37ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Skinned[[:space:]]Motion[[:space:]]Retargeting[[:space:]]With[[:space:]]Residual[[:space:]]Perception[[:space:]]of[[:space:]]Motion[[:space:]]Semantics[[:space:]]&[[:space:]]Geometry/05582d9a-66ac-4216-b8de-ff6dcd9441f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SkyEye_[[:space:]]Self-Supervised[[:space:]]Bird's-Eye-View[[:space:]]Semantic[[:space:]]Mapping[[:space:]]Using[[:space:]]Monocular[[:space:]]Frontal[[:space:]]View[[:space:]]Images/c28dfeac-ae1c-42b7-969c-8ab21879743b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SliceMatch_[[:space:]]Geometry-Guided[[:space:]]Aggregation[[:space:]]for[[:space:]]Cross-View[[:space:]]Pose[[:space:]]Estimation/45591055-c90c-49fb-8b35-751b6508a469_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sliced[[:space:]]Optimal[[:space:]]Partial[[:space:]]Transport/d2e8fcda-fa26-4ab0-b1ff-da1cea8e6046_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Slide-Transformer_[[:space:]]Hierarchical[[:space:]]Vision[[:space:]]Transformer[[:space:]]With[[:space:]]Local[[:space:]]Self-Attention/6be328bb-4525-4377-a906-80caf7311930_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Slimmable[[:space:]]Dataset[[:space:]]Condensation/289adc71-8843-4c5b-bae6-80f9e2fb50cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SlowLiDAR_[[:space:]]Increasing[[:space:]]the[[:space:]]Latency[[:space:]]of[[:space:]]LiDAR-Based[[:space:]]Detection[[:space:]]Using[[:space:]]Adversarial[[:space:]]Examples/716dd6c3-965c-4730-84e1-cb272c54effe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SmallCap_[[:space:]]Lightweight[[:space:]]Image[[:space:]]Captioning[[:space:]]Prompted[[:space:]]With[[:space:]]Retrieval[[:space:]]Augmentation/11351d4b-5d6e-4581-8f90-26e26058cbde_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SmartAssign_[[:space:]]Learning[[:space:]]a[[:space:]]Smart[[:space:]]Knowledge[[:space:]]Assignment[[:space:]]Strategy[[:space:]]for[[:space:]]Deraining[[:space:]]and[[:space:]]Desnowing/ff8d6ac9-0212-4539-8550-960075860e8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SmartBrush_[[:space:]]Text[[:space:]]and[[:space:]]Shape[[:space:]]Guided[[:space:]]Object[[:space:]]Inpainting[[:space:]]With[[:space:]]Diffusion[[:space:]]Model/478e2e10-e6c6-48ef-aa68-d04c6f6b27be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Soft[[:space:]]Augmentation[[:space:]]for[[:space:]]Image[[:space:]]Classification/805a3840-071e-461d-846c-53c8d3efecee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Soft-Landing[[:space:]]Strategy[[:space:]]for[[:space:]]Alleviating[[:space:]]the[[:space:]]Task[[:space:]]Discrepancy[[:space:]]Problem[[:space:]]in[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization[[:space:]]Tasks/7856315d-d9b4-459a-bfdc-45c65ae63b87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Solving[[:space:]]3D[[:space:]]Inverse[[:space:]]Problems[[:space:]]Using[[:space:]]Pre-Trained[[:space:]]2D[[:space:]]Diffusion[[:space:]]Models/935efc27-b56c-48fd-8842-4d9d46e27d5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Solving[[:space:]]Oscillation[[:space:]]Problem[[:space:]]in[[:space:]]Post-Training[[:space:]]Quantization[[:space:]]Through[[:space:]]a[[:space:]]Theoretical[[:space:]]Perspective/dcfbcb3d-004b-47ff-aa16-3b10992b8188_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Solving[[:space:]]Relaxations[[:space:]]of[[:space:]]MAP-MRF[[:space:]]Problems_[[:space:]]Combinatorial[[:space:]]In-Face[[:space:]]Frank-Wolfe[[:space:]]Directions/a3cb2ac3-8e10-4d42-a05a-29eb34905879_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sound[[:space:]]to[[:space:]]Visual[[:space:]]Scene[[:space:]]Generation[[:space:]]by[[:space:]]Audio-to-Visual[[:space:]]Latent[[:space:]]Alignment/1e4b43df-263c-4eed-8bff-07ead035debc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Source-Free[[:space:]]Adaptive[[:space:]]Gaze[[:space:]]Estimation[[:space:]]by[[:space:]]Uncertainty[[:space:]]Reduction/e7b90ce2-dd40-4a5b-8088-4e17214bda76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Source-Free[[:space:]]Video[[:space:]]Domain[[:space:]]Adaptation[[:space:]]With[[:space:]]Spatial-Temporal-Historical[[:space:]]Consistency[[:space:]]Learning/b19b9e2a-dd51-44cd-b440-df1eac75b7b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SpaText_[[:space:]]Spatio-Textual[[:space:]]Representation[[:space:]]for[[:space:]]Controllable[[:space:]]Image[[:space:]]Generation/81c6b9ca-4f91-4600-a996-a7109ffbd223_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sparse[[:space:]]Multi-Modal[[:space:]]Graph[[:space:]]Transformer[[:space:]]With[[:space:]]Shared-Context[[:space:]]Processing[[:space:]]for[[:space:]]Representation[[:space:]]Learning[[:space:]]of[[:space:]]Giga-Pixel[[:space:]]Images/c2fb7f4f-e155-4e96-8c8b-d8ca7e52cd6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SparseFusion_[[:space:]]Distilling[[:space:]]View-Conditioned[[:space:]]Diffusion[[:space:]]for[[:space:]]3D[[:space:]]Reconstruction/a61cca8b-ba5b-4180-9396-c56d50715190_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SparsePose_[[:space:]]Sparse-View[[:space:]]Camera[[:space:]]Pose[[:space:]]Regression[[:space:]]and[[:space:]]Refinement/75b2dc3a-534d-42d5-9cab-d3661394b1fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SparseViT_[[:space:]]Revisiting[[:space:]]Activation[[:space:]]Sparsity[[:space:]]for[[:space:]]Efficient[[:space:]]High-Resolution[[:space:]]Vision[[:space:]]Transformer/e681b42c-2066-472f-90c4-baa972562f69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sparsely[[:space:]]Annotated[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]With[[:space:]]Adaptive[[:space:]]Gaussian[[:space:]]Mixtures/a58c0df3-dca2-462c-9c3e-d56104def1f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sparsifiner_[[:space:]]Learning[[:space:]]Sparse[[:space:]]Instance-Dependent[[:space:]]Attention[[:space:]]for[[:space:]]Efficient[[:space:]]Vision[[:space:]]Transformers/c03cb7fb-a88d-4e8e-b2b6-fceb7e492c27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatial-Frequency[[:space:]]Mutual[[:space:]]Learning[[:space:]]for[[:space:]]Face[[:space:]]Super-Resolution/f534bdd2-2cf3-4328-82ac-6bfeeac24bf5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatial-Temporal[[:space:]]Concept[[:space:]]Based[[:space:]]Explanation[[:space:]]of[[:space:]]3D[[:space:]]ConvNets/0fcfdb15-3aac-4c8a-ab1e-8b3efc30afd3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatial-Then-Temporal[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]for[[:space:]]Video[[:space:]]Correspondence/e522e4b6-9ce5-42cb-8256-730717dc2cba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatially[[:space:]]Adaptive[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]for[[:space:]]Real-World[[:space:]]Image[[:space:]]Denoising/40e93627-de4f-483e-8de6-9b1000606de8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatio-Focal[[:space:]]Bidirectional[[:space:]]Disparity[[:space:]]Estimation[[:space:]]From[[:space:]]a[[:space:]]Dual-Pixel[[:space:]]Image/8857b05d-e159-4995-a78c-931b4fd2f3fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatio-Temporal[[:space:]]Pixel-Level[[:space:]]Contrastive[[:space:]]Learning-Based[[:space:]]Source-Free[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Video[[:space:]]Semantic[[:space:]]Segmentation/88747fa7-3afe-4a10-a659-26c876a06589_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatiotemporal[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]for[[:space:]]Point[[:space:]]Clouds[[:space:]]in[[:space:]]the[[:space:]]Wild/c35a94ac-f688-411c-8bc1-78950677b67e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Specialist[[:space:]]Diffusion_[[:space:]]Plug-and-Play[[:space:]]Sample-Efficient[[:space:]]Fine-Tuning[[:space:]]of[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]To[[:space:]]Learn[[:space:]]Any[[:space:]]Unseen[[:space:]]Style/b61a1d4d-6893-4a13-a3b5-b106207bc936_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spectral[[:space:]]Bayesian[[:space:]]Uncertainty[[:space:]]for[[:space:]]Image[[:space:]]Super-Resolution/4aa3b43d-2756-40c9-85e2-ab2407198e94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spectral[[:space:]]Enhanced[[:space:]]Rectangle[[:space:]]Transformer[[:space:]]for[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Denoising/94d3bea9-c068-46a6-9c1b-df7d0d960635_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sphere-Guided[[:space:]]Training[[:space:]]of[[:space:]]Neural[[:space:]]Implicit[[:space:]]Surfaces/af9152ab-bfa4-4e24-adbe-710da966f410_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spherical[[:space:]]Transformer[[:space:]]for[[:space:]]LiDAR-Based[[:space:]]3D[[:space:]]Recognition/f927e6e6-dd66-41aa-8c58-61de2b18ae2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spider[[:space:]]GAN_[[:space:]]Leveraging[[:space:]]Friendly[[:space:]]Neighbors[[:space:]]To[[:space:]]Accelerate[[:space:]]GAN[[:space:]]Training/ad16b436-1a5e-4d71-82c3-dce23fd0dde3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SplineCam_[[:space:]]Exact[[:space:]]Visualization[[:space:]]and[[:space:]]Characterization[[:space:]]of[[:space:]]Deep[[:space:]]Network[[:space:]]Geometry[[:space:]]and[[:space:]]Decision[[:space:]]Boundaries/96ddd7e1-d6dd-4c29-992c-828664089044_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spring_[[:space:]]A[[:space:]]High-Resolution[[:space:]]High-Detail[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Scene[[:space:]]Flow,[[:space:]]Optical[[:space:]]Flow[[:space:]]and[[:space:]]Stereo/20b8c8ab-8856-4417-a791-9e4af142e793_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Standing[[:space:]]Between[[:space:]]Past[[:space:]]and[[:space:]]Future_[[:space:]]Spatio-Temporal[[:space:]]Modeling[[:space:]]for[[:space:]]Multi-Camera[[:space:]]3D[[:space:]]Multi-Object[[:space:]]Tracking/de572537-e53f-4e81-b572-50e94f09bb7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StarCraftImage_[[:space:]]A[[:space:]]Dataset[[:space:]]for[[:space:]]Prototyping[[:space:]]Spatial[[:space:]]Reasoning[[:space:]]Methods[[:space:]]for[[:space:]]Multi-Agent[[:space:]]Environments/c8aa09bb-f40b-4426-97d0-ed6a237f2b90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Stare[[:space:]]at[[:space:]]What[[:space:]]You[[:space:]]See_[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling[[:space:]]Without[[:space:]]Reconstruction/a5a74b21-a9fa-48ef-9a55-0f9300e56af8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Starting[[:space:]]From[[:space:]]Non-Parametric[[:space:]]Networks[[:space:]]for[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Analysis/87b4219d-4109-4144-90e9-dbb06da19c5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SteerNeRF_[[:space:]]Accelerating[[:space:]]NeRF[[:space:]]Rendering[[:space:]]via[[:space:]]Smooth[[:space:]]Viewpoint[[:space:]]Trajectory/239818f9-6582-4b10-a603-f6a5e232a82b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StepFormer_[[:space:]]Self-Supervised[[:space:]]Step[[:space:]]Discovery[[:space:]]and[[:space:]]Localization[[:space:]]in[[:space:]]Instructional[[:space:]]Videos/2e0921be-d18d-40b5-bc54-1bf3f427a6d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Stimulus[[:space:]]Verification[[:space:]]Is[[:space:]]a[[:space:]]Universal[[:space:]]and[[:space:]]Effective[[:space:]]Sampler[[:space:]]in[[:space:]]Multi-Modal[[:space:]]Human[[:space:]]Trajectory[[:space:]]Prediction/98a1f951-091c-44b5-a3ee-fc0ff3a1a636_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Stitchable[[:space:]]Neural[[:space:]]Networks/58c7117b-0d9a-48d4-a3f2-2c82eddabe34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Streaming[[:space:]]Video[[:space:]]Model/5f628cc0-7b29-46f0-8435-346180e78402_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StructVPR_[[:space:]]Distill[[:space:]]Structural[[:space:]]Knowledge[[:space:]]With[[:space:]]Weighting[[:space:]]Samples[[:space:]]for[[:space:]]Visual[[:space:]]Place[[:space:]]Recognition/ad1e4666-6d01-42e6-b1fe-f213160d5ba1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Structural[[:space:]]Multiplane[[:space:]]Image_[[:space:]]Bridging[[:space:]]Neural[[:space:]]View[[:space:]]Synthesis[[:space:]]and[[:space:]]3D[[:space:]]Reconstruction/d01abaa5-0135-433b-83fa-882e39a28acd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Structure[[:space:]]Aggregation[[:space:]]for[[:space:]]Cross-Spectral[[:space:]]Stereo[[:space:]]Image[[:space:]]Guided[[:space:]]Denoising/f81a83a0-2faa-4877-8a8a-0a610504614c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Structured[[:space:]]3D[[:space:]]Features[[:space:]]for[[:space:]]Reconstructing[[:space:]]Controllable[[:space:]]Avatars/e451a1ab-dc78-421c-b81e-748a81f9e9ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Structured[[:space:]]Kernel[[:space:]]Estimation[[:space:]]for[[:space:]]Photon-Limited[[:space:]]Deconvolution/6df892c8-2e1a-4696-b021-641afb67c068_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Structured[[:space:]]Sparsity[[:space:]]Learning[[:space:]]for[[:space:]]Efficient[[:space:]]Video[[:space:]]Super-Resolution/ed077dac-8227-4e0e-9bc4-a7377d89e3b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyLess_[[:space:]]Boosting[[:space:]]the[[:space:]]Transferability[[:space:]]of[[:space:]]Adversarial[[:space:]]Examples/6f1fccc6-efbf-4f64-b140-75911087cba7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Style[[:space:]]Projected[[:space:]]Clustering[[:space:]]for[[:space:]]Domain[[:space:]]Generalized[[:space:]]Semantic[[:space:]]Segmentation/98057063-f492-4f6c-b6b6-c6b86b75d3df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyleAdv_[[:space:]]Meta[[:space:]]Style[[:space:]]Adversarial[[:space:]]Training[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Few-Shot[[:space:]]Learning/8636d260-66e4-4ec3-b4dc-94bb000bc5ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyleGAN[[:space:]]Salon_[[:space:]]Multi-View[[:space:]]Latent[[:space:]]Optimization[[:space:]]for[[:space:]]Pose-Invariant[[:space:]]Hairstyle[[:space:]]Transfer/ee5e8b03-df5c-443b-b56f-dfc54bf4046b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyleGene_[[:space:]]Crossover[[:space:]]and[[:space:]]Mutation[[:space:]]of[[:space:]]Region-Level[[:space:]]Facial[[:space:]]Genes[[:space:]]for[[:space:]]Kinship[[:space:]]Face[[:space:]]Synthesis/88326347-639e-4732-90cc-676114c49600_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyleIPSB_[[:space:]]Identity-Preserving[[:space:]]Semantic[[:space:]]Basis[[:space:]]of[[:space:]]StyleGAN[[:space:]]for[[:space:]]High[[:space:]]Fidelity[[:space:]]Face[[:space:]]Swapping/9436e560-374c-40b1-8fb0-26b0a7182b10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyleRF_[[:space:]]Zero-Shot[[:space:]]3D[[:space:]]Style[[:space:]]Transfer[[:space:]]of[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/5ea66819-1f3c-4dd7-a1a5-de9269668193_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyleRes_[[:space:]]Transforming[[:space:]]the[[:space:]]Residuals[[:space:]]for[[:space:]]Real[[:space:]]Image[[:space:]]Editing[[:space:]]With[[:space:]]StyleGAN/25e05c83-4dbb-4c83-a9f4-55eb76b40913_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyleSync_[[:space:]]High-Fidelity[[:space:]]Generalized[[:space:]]and[[:space:]]Personalized[[:space:]]Lip[[:space:]]Sync[[:space:]]in[[:space:]]Style-Based[[:space:]]Generator/16a4d87e-841b-4c93-8d78-c26bcd8dff23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SunStage_[[:space:]]Portrait[[:space:]]Reconstruction[[:space:]]and[[:space:]]Relighting[[:space:]]Using[[:space:]]the[[:space:]]Sun[[:space:]]as[[:space:]]a[[:space:]]Light[[:space:]]Stage/0c3e205e-2bfe-4ba0-8856-0ef79f6d0027_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Super-CLEVR_[[:space:]]A[[:space:]]Virtual[[:space:]]Benchmark[[:space:]]To[[:space:]]Diagnose[[:space:]]Domain[[:space:]]Robustness[[:space:]]in[[:space:]]Visual[[:space:]]Reasoning/e3182ce7-71fa-4a83-aeb4-c095eaebb6c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Super-Resolution[[:space:]]Neural[[:space:]]Operator/1174fe30-fb20-45e0-93e7-549281ad4be5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SuperDisco_[[:space:]]Super-Class[[:space:]]Discovery[[:space:]]Improves[[:space:]]Visual[[:space:]]Recognition[[:space:]]for[[:space:]]the[[:space:]]Long-Tail/c1ca0beb-07ce-4086-8c07-7b5c73f59b64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Superclass[[:space:]]Learning[[:space:]]With[[:space:]]Representation[[:space:]]Enhancement/2c1b722e-766f-4d50-b4f7-13afac67bce5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Supervised[[:space:]]Masked[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Few-Shot[[:space:]]Transformers/19caecc7-ab68-426b-8fbc-ffe20dc02942_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SurfelNeRF_[[:space:]]Neural[[:space:]]Surfel[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Online[[:space:]]Photorealistic[[:space:]]Reconstruction[[:space:]]of[[:space:]]Indoor[[:space:]]Scenes/ca098c26-5851-4e50-aa91-b22763ee66ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Swept-Angle[[:space:]]Synthetic[[:space:]]Wavelength[[:space:]]Interferometry/3690b80e-cefe-4aee-937f-0f4b7af2250e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Switchable[[:space:]]Representation[[:space:]]Learning[[:space:]]Framework[[:space:]]With[[:space:]]Self-Compatibility/879611fd-27b7-4de2-b6e0-22124b67ae58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Symmetric[[:space:]]Shape-Preserving[[:space:]]Autoencoder[[:space:]]for[[:space:]]Unsupervised[[:space:]]Real[[:space:]]Scene[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion/324b2114-0ac8-491b-ac58-edad9a307e8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SynthVSR_[[:space:]]Scaling[[:space:]]Up[[:space:]]Visual[[:space:]]Speech[[:space:]]Recognition[[:space:]]With[[:space:]]Synthetic[[:space:]]Supervision/6d7987b1-44f5-4f2e-8f41-c2dafd6e1f4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Synthesizing[[:space:]]Photorealistic[[:space:]]Virtual[[:space:]]Humans[[:space:]]Through[[:space:]]Cross-Modal[[:space:]]Disentanglement/f1ee982b-9321-4c5b-bbc8-a11edd225f8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/System-Status-Aware[[:space:]]Adaptive[[:space:]]Network[[:space:]]for[[:space:]]Online[[:space:]]Streaming[[:space:]]Video[[:space:]]Understanding/d274131a-49c0-4a6a-97bb-9cb59d470bd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/T-SEA_[[:space:]]Transfer-Based[[:space:]]Self-Ensemble[[:space:]]Attack[[:space:]]on[[:space:]]Object[[:space:]]Detection/8fc8a2c2-bc87-41bf-a7ad-e07bcd025a9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TAPS3D_[[:space:]]Text-Guided[[:space:]]3D[[:space:]]Textured[[:space:]]Shape[[:space:]]Generation[[:space:]]From[[:space:]]Pseudo[[:space:]]Supervision/cb0b5b28-71a3-4466-8edb-4420e1a78920_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TBP-Former_[[:space:]]Learning[[:space:]]Temporal[[:space:]]Bird's-Eye-View[[:space:]]Pyramid[[:space:]]for[[:space:]]Joint[[:space:]]Perception[[:space:]]and[[:space:]]Prediction[[:space:]]in[[:space:]]Vision-Centric[[:space:]]Autonomous[[:space:]]Driving/f7f592e0-a2b8-46fb-a0f7-4d6e06bddd8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TINC_[[:space:]]Tree-Structured[[:space:]]Implicit[[:space:]]Neural[[:space:]]Compression/8bbad193-1acf-4b1e-bca7-0998f55872d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TIPI_[[:space:]]Test[[:space:]]Time[[:space:]]Adaptation[[:space:]]With[[:space:]]Transformation[[:space:]]Invariance/5131537e-5515-4827-8458-d96e395e71bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TMO_[[:space:]]Textured[[:space:]]Mesh[[:space:]]Acquisition[[:space:]]of[[:space:]]Objects[[:space:]]With[[:space:]]a[[:space:]]Mobile[[:space:]]Device[[:space:]]by[[:space:]]Using[[:space:]]Differentiable[[:space:]]Rendering/a9ab43cd-15ce-44b4-8413-0c6fe896d349_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TOPLight_[[:space:]]Lightweight[[:space:]]Neural[[:space:]]Networks[[:space:]]With[[:space:]]Task-Oriented[[:space:]]Pretraining[[:space:]]for[[:space:]]Visible-Infrared[[:space:]]Recognition/fd571fdc-7e5b-4461-afda-4cb4debbc856_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TRACE_[[:space:]]5D[[:space:]]Temporal[[:space:]]Regression[[:space:]]of[[:space:]]Avatars[[:space:]]With[[:space:]]Dynamic[[:space:]]Cameras[[:space:]]in[[:space:]]3D[[:space:]]Environments/c8e8684e-51e8-46fa-8645-e310b1ba9073_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TTA-COPE_[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Category-Level[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/e0dd8ebd-75c4-4890-8983-be1eb1c01bc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TWINS_[[:space:]]A[[:space:]]Fine-Tuning[[:space:]]Framework[[:space:]]for[[:space:]]Improved[[:space:]]Transferability[[:space:]]of[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]and[[:space:]]Generalization/178a6a71-7aa4-4ba0-ae4e-9fff019b01f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Taming[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Audio-Driven[[:space:]]Co-Speech[[:space:]]Gesture[[:space:]]Generation/436e7035-a0f8-4cb2-9a25-a351d003ba3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tangentially[[:space:]]Elongated[[:space:]]Gaussian[[:space:]]Belief[[:space:]]Propagation[[:space:]]for[[:space:]]Event-Based[[:space:]]Incremental[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation/1e72f145-1dc6-4f05-89ce-8e276ebb5dbf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TarViS_[[:space:]]A[[:space:]]Unified[[:space:]]Approach[[:space:]]for[[:space:]]Target-Based[[:space:]]Video[[:space:]]Segmentation/7480a5f9-95b4-4d94-9341-e301cfb19950_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Target-Referenced[[:space:]]Reactive[[:space:]]Grasping[[:space:]]for[[:space:]]Dynamic[[:space:]]Objects/d5655194-9ce2-4b7e-be6f-e8df1e2c1030_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Task[[:space:]]Difficulty[[:space:]]Aware[[:space:]]Parameter[[:space:]]Allocation[[:space:]]&[[:space:]]Regularization[[:space:]]for[[:space:]]Lifelong[[:space:]]Learning/a54864ea-e9cc-490e-a81a-988ea47d70e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Task[[:space:]]Residual[[:space:]]for[[:space:]]Tuning[[:space:]]Vision-Language[[:space:]]Models/7e947285-2912-4e13-91da-b9b910f110fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Task-Specific[[:space:]]Fine-Tuning[[:space:]]via[[:space:]]Variational[[:space:]]Information[[:space:]]Bottleneck[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Pathology[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Classification/1b72dc26-76ea-4a15-b4ba-65955df2cecc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TeSLA_[[:space:]]Test-Time[[:space:]]Self-Learning[[:space:]]With[[:space:]]Automatic[[:space:]]Adversarial[[:space:]]Augmentation/ad821131-dbc0-46a1-909c-72d909b80828_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Teacher-Generated[[:space:]]Spatial-Attention[[:space:]]Labels[[:space:]]Boost[[:space:]]Robustness[[:space:]]and[[:space:]]Accuracy[[:space:]]of[[:space:]]Contrastive[[:space:]]Models/7ff3c0aa-d380-4920-a7cd-d62a1a9e78dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Teaching[[:space:]]Matters_[[:space:]]Investigating[[:space:]]the[[:space:]]Role[[:space:]]of[[:space:]]Supervision[[:space:]]in[[:space:]]Vision[[:space:]]Transformers/25874b86-4464-466c-bbd4-91f62653cf36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Teaching[[:space:]]Structured[[:space:]]Vision[[:space:]]&[[:space:]]Language[[:space:]]Concepts[[:space:]]to[[:space:]]Vision[[:space:]]&[[:space:]]Language[[:space:]]Models/9a8c6788-2d0c-421a-8a6a-0e2eb6239bd0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Teleidoscopic[[:space:]]Imaging[[:space:]]System[[:space:]]for[[:space:]]Microscale[[:space:]]3D[[:space:]]Shape[[:space:]]Reconstruction/d72f0f54-2daa-4009-a052-255877447821_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tell[[:space:]]Me[[:space:]]What[[:space:]]Happened_[[:space:]]Unifying[[:space:]]Text-Guided[[:space:]]Video[[:space:]]Completion[[:space:]]via[[:space:]]Multimodal[[:space:]]Masked[[:space:]]Video[[:space:]]Generation/0b8e3b80-d861-4dd7-999a-b659b22d0266_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TempSAL[[:space:]]-[[:space:]]Uncovering[[:space:]]Temporal[[:space:]]Information[[:space:]]for[[:space:]]Deep[[:space:]]Saliency[[:space:]]Prediction/a96ae848-89c1-42c8-883e-3ee6e955a530_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Temporal[[:space:]]Attention[[:space:]]Unit_[[:space:]]Towards[[:space:]]Efficient[[:space:]]Spatiotemporal[[:space:]]Predictive[[:space:]]Learning/d3109e90-c15d-418e-bbe6-8d0dca603ccc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Temporal[[:space:]]Consistent[[:space:]]3D[[:space:]]LiDAR[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Semantic[[:space:]]Perception[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/4a98930e-4a72-4866-a1b4-e6692bb0ab17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Temporal[[:space:]]Interpolation[[:space:]]Is[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]for[[:space:]]Dynamic[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/22a9031d-e022-4324-b56f-e7ab749525c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Temporally[[:space:]]Consistent[[:space:]]Online[[:space:]]Depth[[:space:]]Estimation[[:space:]]Using[[:space:]]Point-Based[[:space:]]Fusion/5039f27e-eabc-410c-b1b5-8a065a9cead0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TensoIR_[[:space:]]Tensorial[[:space:]]Inverse[[:space:]]Rendering/911491ec-c0e6-468f-9b5f-cd8676ecd8ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tensor4D_[[:space:]]Efficient[[:space:]]Neural[[:space:]]4D[[:space:]]Decomposition[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Dynamic[[:space:]]Reconstruction[[:space:]]and[[:space:]]Rendering/eeed7500-cf04-43cc-8f38-007884849085_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Test[[:space:]]Time[[:space:]]Adaptation[[:space:]]With[[:space:]]Regularized[[:space:]]Loss[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Salient[[:space:]]Object[[:space:]]Detection/2441aa83-753b-4443-ba38-3993bb6d5100_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Test[[:space:]]of[[:space:]]Time_[[:space:]]Instilling[[:space:]]Video-Language[[:space:]]Models[[:space:]]With[[:space:]]a[[:space:]]Sense[[:space:]]of[[:space:]]Time/f3a89bc6-be03-43ea-932f-9df8dc46f9ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TexPose_[[:space:]]Neural[[:space:]]Texture[[:space:]]Learning[[:space:]]for[[:space:]]Self-Supervised[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/3847ca6c-d45c-45c5-ba3b-fa7d3a58ce42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Text[[:space:]]With[[:space:]]Knowledge[[:space:]]Graph[[:space:]]Augmented[[:space:]]Transformer[[:space:]]for[[:space:]]Video[[:space:]]Captioning/4ae6afce-15c9-4a54-b4c2-e5743fb4d16e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Text-Guided[[:space:]]Unsupervised[[:space:]]Latent[[:space:]]Transformation[[:space:]]for[[:space:]]Multi-Attribute[[:space:]]Image[[:space:]]Manipulation/4306b93e-f301-4712-887d-3f71b781b5b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Text-Visual[[:space:]]Prompting[[:space:]]for[[:space:]]Efficient[[:space:]]2D[[:space:]]Temporal[[:space:]]Video[[:space:]]Grounding/86b06816-ef9f-43d9-b909-e5a92eff83a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Text2Scene_[[:space:]]Text-Driven[[:space:]]Indoor[[:space:]]Scene[[:space:]]Stylization[[:space:]]With[[:space:]]Part-Aware[[:space:]]Details/3c9b1104-12d9-49d4-abdb-99527f418bd3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Texts[[:space:]]as[[:space:]]Images[[:space:]]in[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Multi-Label[[:space:]]Image[[:space:]]Recognition/df1598eb-cae0-4ed0-bda2-3a1705488a58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Texture-Guided[[:space:]]Saliency[[:space:]]Distilling[[:space:]]for[[:space:]]Unsupervised[[:space:]]Salient[[:space:]]Object[[:space:]]Detection/099f5794-0ad0-4870-a20c-4c38e466b27c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Best[[:space:]]Defense[[:space:]]Is[[:space:]]a[[:space:]]Good[[:space:]]Offense_[[:space:]]Adversarial[[:space:]]Augmentation[[:space:]]Against[[:space:]]Adversarial[[:space:]]Attacks/2480b99e-0eac-4b8f-a85f-c4137baac3df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Dark[[:space:]]Side[[:space:]]of[[:space:]]Dynamic[[:space:]]Routing[[:space:]]Neural[[:space:]]Networks_[[:space:]]Towards[[:space:]]Efficiency[[:space:]]Backdoor[[:space:]]Injection/9563ac6b-8685-4660-8a83-51df260452bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Devil[[:space:]]Is[[:space:]]in[[:space:]]the[[:space:]]Points_[[:space:]]Weakly[[:space:]]Semi-Supervised[[:space:]]Instance[[:space:]]Segmentation[[:space:]]via[[:space:]]Point-Guided[[:space:]]Mask[[:space:]]Representation/56f30a3c-9c71-43a4-ad67-87e67797ecce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Dialog[[:space:]]Must[[:space:]]Go[[:space:]]On_[[:space:]]Improving[[:space:]]Visual[[:space:]]Dialog[[:space:]]via[[:space:]]Generative[[:space:]]Self-Training/d9898fec-8082-4fa6-bb0e-9fc0da60386d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Differentiable[[:space:]]Lens_[[:space:]]Compound[[:space:]]Lens[[:space:]]Search[[:space:]]Over[[:space:]]Glass[[:space:]]Surfaces[[:space:]]and[[:space:]]Materials[[:space:]]for[[:space:]]Object[[:space:]]Detection/36241643-32ab-4c21-8c20-aa46a37ff39f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Enemy[[:space:]]of[[:space:]]My[[:space:]]Enemy[[:space:]]Is[[:space:]]My[[:space:]]Friend_[[:space:]]Exploring[[:space:]]Inverse[[:space:]]Adversaries[[:space:]]for[[:space:]]Improving[[:space:]]Adversarial[[:space:]]Training/55dc6cd9-6a65-4c51-87d6-9b03a9bc9024_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]ObjectFolder[[:space:]]Benchmark_[[:space:]]Multisensory[[:space:]]Learning[[:space:]]With[[:space:]]Neural[[:space:]]and[[:space:]]Real[[:space:]]Objects/b4a6c67f-5481-42db-8b5d-95370684a90a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Resource[[:space:]]Problem[[:space:]]of[[:space:]]Using[[:space:]]Linear[[:space:]]Layer[[:space:]]Leakage[[:space:]]Attack[[:space:]]in[[:space:]]Federated[[:space:]]Learning/c319fec4-3d39-45ec-82b7-257ad738d84e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Treasure[[:space:]]Beneath[[:space:]]Multiple[[:space:]]Annotations_[[:space:]]An[[:space:]]Uncertainty-Aware[[:space:]]Edge[[:space:]]Detector/c007a047-d159-4c1a-8c41-39df50ca924d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Wisdom[[:space:]]of[[:space:]]Crowds_[[:space:]]Temporal[[:space:]]Progressive[[:space:]]Attention[[:space:]]for[[:space:]]Early[[:space:]]Action[[:space:]]Prediction/2dd8962b-43f9-49b3-b846-7788927b4a61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Therbligs[[:space:]]in[[:space:]]Action_[[:space:]]Video[[:space:]]Understanding[[:space:]]Through[[:space:]]Motion[[:space:]]Primitives/6c056490-f886-4578-aa36-41c3ac78d7c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Thermal[[:space:]]Spread[[:space:]]Functions[[:space:]](TSF)_[[:space:]]Physics-Guided[[:space:]]Material[[:space:]]Classification/ef9633bf-924d-4aeb-af05-e0b619ab1722_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Think[[:space:]]Twice[[:space:]]Before[[:space:]]Driving_[[:space:]]Towards[[:space:]]Scalable[[:space:]]Decoders[[:space:]]for[[:space:]]End-to-End[[:space:]]Autonomous[[:space:]]Driving/63a7606b-61c2-4e04-a444-05c617fcb360_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Three[[:space:]]Guidelines[[:space:]]You[[:space:]]Should[[:space:]]Know[[:space:]]for[[:space:]]Universally[[:space:]]Slimmable[[:space:]]Self-Supervised[[:space:]]Learning/4de122ae-59ff-44e7-98d8-08b1a285ff02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TimeBalance_[[:space:]]Temporally-Invariant[[:space:]]and[[:space:]]Temporally-Distinctive[[:space:]]Video[[:space:]]Representations[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Action[[:space:]]Recognition/5d5c4a69-a491-4e89-b854-5aa4e7a0100e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TinyMIM_[[:space:]]An[[:space:]]Empirical[[:space:]]Study[[:space:]]of[[:space:]]Distilling[[:space:]]MIM[[:space:]]Pre-Trained[[:space:]]Models/761d8c57-8108-457e-8fe1-6d5c1efc3b24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ToThePoint_[[:space:]]Efficient[[:space:]]Contrastive[[:space:]]Learning[[:space:]]of[[:space:]]3D[[:space:]]Point[[:space:]]Clouds[[:space:]]via[[:space:]]Recycling/9f6ffa5a-1011-40ef-bcc7-d4137dae7a78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Token[[:space:]]Boosting[[:space:]]for[[:space:]]Robust[[:space:]]Self-Supervised[[:space:]]Visual[[:space:]]Transformer[[:space:]]Pre-Training/bf670e6d-5ea9-4854-a8d9-74df579e9309_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Token[[:space:]]Contrast[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Semantic[[:space:]]Segmentation/80311486-b2cd-4d5f-b916-af23414538d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Token[[:space:]]Turing[[:space:]]Machines/cc96866d-6d7b-4175-986a-b6598360a4a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TokenHPE_[[:space:]]Learning[[:space:]]Orientation[[:space:]]Tokens[[:space:]]for[[:space:]]Efficient[[:space:]]Head[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Transformers/f1eb4d41-c084-4e28-bc61-080652f4ce5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Top-Down[[:space:]]Visual[[:space:]]Attention[[:space:]]From[[:space:]]Analysis[[:space:]]by[[:space:]]Synthesis/55ffb13b-50c1-421b-bf73-d9aafb69f622_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TopDiG_[[:space:]]Class-Agnostic[[:space:]]Topological[[:space:]]Directional[[:space:]]Graph[[:space:]]Extraction[[:space:]]From[[:space:]]Remote[[:space:]]Sensing[[:space:]]Images/f231b3a7-8f59-4f55-a995-be11d196b8f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TopNet_[[:space:]]Transformer-Based[[:space:]]Object[[:space:]]Placement[[:space:]]Network[[:space:]]for[[:space:]]Image[[:space:]]Compositing/ed51f82f-cfb5-4bb1-ad09-fe556e791d79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Topology-Guided[[:space:]]Multi-Class[[:space:]]Cell[[:space:]]Context[[:space:]]Generation[[:space:]]for[[:space:]]Digital[[:space:]]Pathology/681da271-7693-4748-9343-5a8b0296976e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Toward[[:space:]]Accurate[[:space:]]Post-Training[[:space:]]Quantization[[:space:]]for[[:space:]]Image[[:space:]]Super[[:space:]]Resolution/2d67c0f6-67b0-40a6-a959-5df1b1186b57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Toward[[:space:]]RAW[[:space:]]Object[[:space:]]Detection_[[:space:]]A[[:space:]]New[[:space:]]Benchmark[[:space:]]and[[:space:]]a[[:space:]]New[[:space:]]Model/ee690eaa-b7ed-4237-9fef-53fc4c1df50d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Toward[[:space:]]Stable,[[:space:]]Interpretable,[[:space:]]and[[:space:]]Lightweight[[:space:]]Hyperspectral[[:space:]]Super-Resolution/de262145-f062-4614-8067-1ac2eedec07f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Toward[[:space:]]Verifiable[[:space:]]and[[:space:]]Reproducible[[:space:]]Human[[:space:]]Evaluation[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/8ef44cdc-f064-4bc6-9db8-f2e1d56a6726_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Accurate[[:space:]]Image[[:space:]]Coding_[[:space:]]Improved[[:space:]]Autoregressive[[:space:]]Image[[:space:]]Generation[[:space:]]With[[:space:]]Dynamic[[:space:]]Vector[[:space:]]Quantization/fa07b829-bc44-4f66-a7ce-c5854d448113_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]All-in-One[[:space:]]Pre-Training[[:space:]]via[[:space:]]Maximizing[[:space:]]Multi-Modal[[:space:]]Mutual[[:space:]]Information/1e52309f-3b28-4db9-a4bc-1f6ac0495c82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Artistic[[:space:]]Image[[:space:]]Aesthetics[[:space:]]Assessment_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]and[[:space:]]a[[:space:]]New[[:space:]]Method/4ca55186-10d2-4df3-ba90-864b42e5c662_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Benchmarking[[:space:]]and[[:space:]]Assessing[[:space:]]Visual[[:space:]]Naturalness[[:space:]]of[[:space:]]Physical[[:space:]]World[[:space:]]Adversarial[[:space:]]Attacks/aa5ee8f8-503b-4378-bf50-f041f02a931c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Better[[:space:]]Decision[[:space:]]Forests_[[:space:]]Forest[[:space:]]Alternating[[:space:]]Optimization/91ec04b4-5b19-4f2a-9c04-5c73f3b2b2b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Better[[:space:]]Gradient[[:space:]]Consistency[[:space:]]for[[:space:]]Neural[[:space:]]Signed[[:space:]]Distance[[:space:]]Functions[[:space:]]via[[:space:]]Level[[:space:]]Set[[:space:]]Alignment/b3ba0145-35bb-4210-91d9-355b0231658e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Better[[:space:]]Stability[[:space:]]and[[:space:]]Adaptability_[[:space:]]Improve[[:space:]]Online[[:space:]]Self-Training[[:space:]]for[[:space:]]Model[[:space:]]Adaptation[[:space:]]in[[:space:]]Semantic[[:space:]]Segmentation/91dffe15-bcd3-4af5-8a0c-d8480447e038_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Bridging[[:space:]]the[[:space:]]Performance[[:space:]]Gaps[[:space:]]of[[:space:]]Joint[[:space:]]Energy-Based[[:space:]]Models/2f6bd5db-9d3b-4c87-855d-f2ee21b57e51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Building[[:space:]]Self-Aware[[:space:]]Object[[:space:]]Detectors[[:space:]]via[[:space:]]Reliable[[:space:]]Uncertainty[[:space:]]Quantification[[:space:]]and[[:space:]]Calibration/b02f6e4f-83fd-42b8-aed3-7097d527e1bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Compositional[[:space:]]Adversarial[[:space:]]Robustness_[[:space:]]Generalizing[[:space:]]Adversarial[[:space:]]Training[[:space:]]to[[:space:]]Composite[[:space:]]Semantic[[:space:]]Perturbations/689e1c83-f22c-442d-a6fc-73086816952f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Domain[[:space:]]Generalization[[:space:]]for[[:space:]]Multi-View[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]Bird-Eye-View/eca21448-b14b-4b39-ad47-374496474f54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Effective[[:space:]]Adversarial[[:space:]]Textured[[:space:]]3D[[:space:]]Meshes[[:space:]]on[[:space:]]Physical[[:space:]]Face[[:space:]]Recognition/264d253c-48b2-4041-8d90-f673300f89e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Effective[[:space:]]Visual[[:space:]]Representations[[:space:]]for[[:space:]]Partial-Label[[:space:]]Learning/b63097d7-26c3-4175-a94e-6f807cc1ee13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Efficient[[:space:]]Use[[:space:]]of[[:space:]]Multi-Scale[[:space:]]Features[[:space:]]in[[:space:]]Transformer-Based[[:space:]]Object[[:space:]]Detectors/d7de2354-15df-4e55-a898-27d68d3f6d5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]End-to-End[[:space:]]Generative[[:space:]]Modeling[[:space:]]of[[:space:]]Long[[:space:]]Videos[[:space:]]With[[:space:]]Memory-Efficient[[:space:]]Bidirectional[[:space:]]Transformers/42c21587-4445-4f1d-b897-3ef62ece5eaa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Fast[[:space:]]Adaptation[[:space:]]of[[:space:]]Pretrained[[:space:]]Contrastive[[:space:]]Models[[:space:]]for[[:space:]]Multi-Channel[[:space:]]Video-Language[[:space:]]Retrieval/2a0f744c-c057-4644-8b62-e4cc5a4a5bab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Flexible[[:space:]]Multi-Modal[[:space:]]Document[[:space:]]Models/4346ad36-4431-46ee-b275-bd3d199196bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Generalisable[[:space:]]Video[[:space:]]Moment[[:space:]]Retrieval_[[:space:]]Visual-Dynamic[[:space:]]Injection[[:space:]]to[[:space:]]Image-Text[[:space:]]Pre-Training/b4a0477a-2379-4183-8b9f-5416362ce296_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]High-Quality[[:space:]]and[[:space:]]Efficient[[:space:]]Video[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Spatial-Temporal[[:space:]]Data[[:space:]]Overfitting/73688438-a87d-4dae-9a11-4a6d08a10966_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Modality-Agnostic[[:space:]]Person[[:space:]]Re-Identification[[:space:]]With[[:space:]]Descriptive[[:space:]]Query/98341ca1-68e3-45d2-83a7-ad7e26e30b8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Open-World[[:space:]]Segmentation[[:space:]]of[[:space:]]Parts/487d5b04-95e2-4a7e-9488-e7dcda2c3bcf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Practical[[:space:]]Plug-and-Play[[:space:]]Diffusion[[:space:]]Models/5651d04e-1812-450e-8c5b-05bd13f8d653_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Professional[[:space:]]Level[[:space:]]Crowd[[:space:]]Annotation[[:space:]]of[[:space:]]Expert[[:space:]]Domain[[:space:]]Data/865bf5c6-8ec3-49ba-8b2b-48315fdf0500_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Realistic[[:space:]]Long-Tailed[[:space:]]Semi-Supervised[[:space:]]Learning_[[:space:]]Consistency[[:space:]]Is[[:space:]]All[[:space:]]You[[:space:]]Need/449a3136-5313-49cc-9868-b42b5b41e45a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Robust[[:space:]]Tampered[[:space:]]Text[[:space:]]Detection[[:space:]]in[[:space:]]Document[[:space:]]Image_[[:space:]]New[[:space:]]Dataset[[:space:]]and[[:space:]]New[[:space:]]Solution/8e7b7186-b75d-4184-96c9-b0d63fd2a9a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Scalable[[:space:]]Neural[[:space:]]Representation[[:space:]]for[[:space:]]Diverse[[:space:]]Videos/039af6ef-c478-4a5d-8d1d-f84ec7fa8b42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Stable[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Cross-View[[:space:]]Fusion[[:space:]]and[[:space:]]Foot[[:space:]]Stabilization/f1c255a2-c356-4c10-9b2d-42d8366ccd76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Transferable[[:space:]]Targeted[[:space:]]Adversarial[[:space:]]Examples/36351713-9f1e-40e5-984d-099d00ce677a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Trustable[[:space:]]Skin[[:space:]]Cancer[[:space:]]Diagnosis[[:space:]]via[[:space:]]Rewriting[[:space:]]Model's[[:space:]]Decision/6e3ac7f9-6fe7-432f-8830-1dff6d3ad53c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Unbiased[[:space:]]Volume[[:space:]]Rendering[[:space:]]of[[:space:]]Neural[[:space:]]Implicit[[:space:]]Surfaces[[:space:]]With[[:space:]]Geometry[[:space:]]Priors/1deb853e-23a6-4b6e-8979-e0948bb166cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Unified[[:space:]]Scene[[:space:]]Text[[:space:]]Spotting[[:space:]]Based[[:space:]]on[[:space:]]Sequence[[:space:]]Generation/e436d89a-2f25-409e-8608-dfdbf78596cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Universal[[:space:]]Fake[[:space:]]Image[[:space:]]Detectors[[:space:]]That[[:space:]]Generalize[[:space:]]Across[[:space:]]Generative[[:space:]]Models/d35cbe2c-5c1a-475d-9466-3d68e02ad94e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Unsupervised[[:space:]]Object[[:space:]]Detection[[:space:]]From[[:space:]]LiDAR[[:space:]]Point[[:space:]]Clouds/66ba3545-a5d8-406a-810f-5f7e58981079_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]a[[:space:]]Smaller[[:space:]]Student_[[:space:]]Capacity[[:space:]]Dynamic[[:space:]]Distillation[[:space:]]for[[:space:]]Efficient[[:space:]]Image[[:space:]]Retrieval/b406ec5e-aa9c-4529-bf2e-926000a14fc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Trace[[:space:]]and[[:space:]]Pace_[[:space:]]Controllable[[:space:]]Pedestrian[[:space:]]Animation[[:space:]]via[[:space:]]Guided[[:space:]]Trajectory[[:space:]]Diffusion/33b8027e-521c-4623-9621-614455ba3ade_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tracking[[:space:]]Multiple[[:space:]]Deformable[[:space:]]Objects[[:space:]]in[[:space:]]Egocentric[[:space:]]Videos/12ccbb34-801b-40c9-b5be-49f255aceca2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tracking[[:space:]]Through[[:space:]]Containers[[:space:]]and[[:space:]]Occluders[[:space:]]in[[:space:]]the[[:space:]]Wild/59dd3b8c-31e5-4646-918a-f97df2e1ee9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Trade-Off[[:space:]]Between[[:space:]]Robustness[[:space:]]and[[:space:]]Accuracy[[:space:]]of[[:space:]]Vision[[:space:]]Transformers/ec11fec8-ae22-4952-8103-aeeac532fee6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Train-Once-for-All[[:space:]]Personalization/70e4e3b8-e0fc-40b6-9c48-3d386aced8a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Train_Test-Time[[:space:]]Adaptation[[:space:]]With[[:space:]]Retrieval/41877221-a2fc-4b53-af7b-ed19c5ab60f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Trainable[[:space:]]Projected[[:space:]]Gradient[[:space:]]Method[[:space:]]for[[:space:]]Robust[[:space:]]Fine-Tuning/53b66bfc-ba36-45d6-95ce-0ed1a7aff555_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Training[[:space:]]Debiased[[:space:]]Subnetworks[[:space:]]With[[:space:]]Contrastive[[:space:]]Weight[[:space:]]Pruning/72c99801-93c3-4f50-8764-2126bcaa9b4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Trajectory-Aware[[:space:]]Body[[:space:]]Interaction[[:space:]]Transformer[[:space:]]for[[:space:]]Multi-Person[[:space:]]Pose[[:space:]]Forecasting/4e6bb4dd-41e5-43f5-924e-9b72e344aa66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TranSG_[[:space:]]Transformer-Based[[:space:]]Skeleton[[:space:]]Graph[[:space:]]Prototype[[:space:]]Contrastive[[:space:]]Learning[[:space:]]With[[:space:]]Structure-Trajectory[[:space:]]Prompted[[:space:]]Reconstruction[[:space:]]for[[:space:]]Person[[:space:]]Re-Identification/637d3a82-0d25-409a-a1b3-6f930b3189ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TransFlow_[[:space:]]Transformer[[:space:]]As[[:space:]]Flow[[:space:]]Learner/451e66cb-14d4-4aea-8632-e8b5ce82b34b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Transductive[[:space:]]Few-Shot[[:space:]]Learning[[:space:]]With[[:space:]]Prototype-Based[[:space:]]Label[[:space:]]Propagation[[:space:]]by[[:space:]]Iterative[[:space:]]Graph[[:space:]]Refinement/0b4668f3-4356-45b5-95ab-58f45e058426_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Transfer[[:space:]]Knowledge[[:space:]]From[[:space:]]Head[[:space:]]to[[:space:]]Tail_[[:space:]]Uncertainty[[:space:]]Calibration[[:space:]]Under[[:space:]]Long-Tailed[[:space:]]Distribution/0244fca2-e2bc-4a0a-be77-4822f4c9128b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Transfer4D_[[:space:]]A[[:space:]]Framework[[:space:]]for[[:space:]]Frugal[[:space:]]Motion[[:space:]]Capture[[:space:]]and[[:space:]]Deformation[[:space:]]Transfer/0944cb9f-9745-4881-b76e-8af12b43284c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Transferable[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]on[[:space:]]Vision[[:space:]]Transformers[[:space:]]With[[:space:]]Token[[:space:]]Gradient[[:space:]]Regularization/6c09193f-9b1b-470a-8d1d-80d53ae6697b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Transformer[[:space:]]Scale[[:space:]]Gate[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/6122053a-824a-436e-a44a-1b7667287787_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Transformer-Based[[:space:]]Learned[[:space:]]Optimization/d3991b61-af07-4811-a6d2-cd0820d481e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Transformer-Based[[:space:]]Unified[[:space:]]Recognition[[:space:]]of[[:space:]]Two[[:space:]]Hands[[:space:]]Manipulating[[:space:]]Objects/96c5e2f3-d7e4-4217-9e97-8243f58db85b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Transforming[[:space:]]Radiance[[:space:]]Field[[:space:]]With[[:space:]]Lipschitz[[:space:]]Network[[:space:]]for[[:space:]]Photorealistic[[:space:]]3D[[:space:]]Scene[[:space:]]Stylization/430e7276-5418-446b-bae4-18dd520b7e4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Trap[[:space:]]Attention_[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation[[:space:]]With[[:space:]]Manual[[:space:]]Traps/f08bc27f-6376-40dd-a90f-a5e65b203ca7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tree[[:space:]]Instance[[:space:]]Segmentation[[:space:]]With[[:space:]]Temporal[[:space:]]Contour[[:space:]]Graph/2788a9df-a4f4-4351-b819-3929301ac6a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tri-Perspective[[:space:]]View[[:space:]]for[[:space:]]Vision-Based[[:space:]]3D[[:space:]]Semantic[[:space:]]Occupancy[[:space:]]Prediction/646a06ad-b523-4bbc-a95f-33288db076a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TriDet_[[:space:]]Temporal[[:space:]]Action[[:space:]]Detection[[:space:]]With[[:space:]]Relative[[:space:]]Boundary[[:space:]]Modeling/9437ea6d-10cd-4c07-9205-4e29facb36f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TriVol_[[:space:]]Point[[:space:]]Cloud[[:space:]]Rendering[[:space:]]via[[:space:]]Triple[[:space:]]Volumes/8588dd47-f109-4c2c-9e39-5a57835eb5db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TrojDiff_[[:space:]]Trojan[[:space:]]Attacks[[:space:]]on[[:space:]]Diffusion[[:space:]]Models[[:space:]]With[[:space:]]Diverse[[:space:]]Targets/71878e3c-53d9-4a9a-946f-dfc6d854a0d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TrojViT_[[:space:]]Trojan[[:space:]]Insertion[[:space:]]in[[:space:]]Vision[[:space:]]Transformers/7a9b2a4c-ed7d-4d80-93ec-395c853b3fac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TruFor_[[:space:]]Leveraging[[:space:]]All-Round[[:space:]]Clues[[:space:]]for[[:space:]]Trustworthy[[:space:]]Image[[:space:]]Forgery[[:space:]]Detection[[:space:]]and[[:space:]]Localization/40350874-7a4f-490c-9c59-a2b3b8aa0ac5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TryOnDiffusion_[[:space:]]A[[:space:]]Tale[[:space:]]of[[:space:]]Two[[:space:]]UNets/17ba6fed-69a8-4a6b-adca-f7324f56e35c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tunable[[:space:]]Convolutions[[:space:]]With[[:space:]]Parametric[[:space:]]Multi-Loss[[:space:]]Optimization/5a93da61-dd4f-40f6-9e53-3301f4055d2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Turning[[:space:]]Strengths[[:space:]]Into[[:space:]]Weaknesses_[[:space:]]A[[:space:]]Certified[[:space:]]Robustness[[:space:]]Inspired[[:space:]]Attack[[:space:]]Framework[[:space:]]Against[[:space:]]Graph[[:space:]]Neural[[:space:]]Networks/d1b83acb-ee6b-4260-ae75-3a6cf88013e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Turning[[:space:]]a[[:space:]]CLIP[[:space:]]Model[[:space:]]Into[[:space:]]a[[:space:]]Scene[[:space:]]Text[[:space:]]Detector/45f8f50b-19fe-4aa2-9af6-d14b48f7cda2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Twin[[:space:]]Contrastive[[:space:]]Learning[[:space:]]With[[:space:]]Noisy[[:space:]]Labels/e832bc03-7222-4afb-85d3-0febdf1b1a15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Two-Shot[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/35a9cbed-73b1-44f4-a62c-e2bcd070dcb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Two-Stage[[:space:]]Co-Segmentation[[:space:]]Network[[:space:]]Based[[:space:]]on[[:space:]]Discriminative[[:space:]]Representation[[:space:]]for[[:space:]]Recovering[[:space:]]Human[[:space:]]Mesh[[:space:]]From[[:space:]]Videos/4ddaf86d-d6fb-4bb1-8a58-63f842026a49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Two-Stream[[:space:]]Networks[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization[[:space:]]With[[:space:]]Semantic-Aware[[:space:]]Mechanisms/82582932-0f6a-430b-a606-5b8d0ac6496d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Two-View[[:space:]]Geometry[[:space:]]Scoring[[:space:]]Without[[:space:]]Correspondences/3333749d-5ed9-45ab-a7c0-249dca26e47f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Two-Way[[:space:]]Multi-Label[[:space:]]Loss/8c3c7a29-0e3e-4fde-a515-8ea880d43ecd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UDE_[[:space:]]A[[:space:]]Unified[[:space:]]Driving[[:space:]]Engine[[:space:]]for[[:space:]]Human[[:space:]]Motion[[:space:]]Generation/51890b7b-5a0a-484f-af44-d196c0c3b295_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ULIP_[[:space:]]Learning[[:space:]]a[[:space:]]Unified[[:space:]]Representation[[:space:]]of[[:space:]]Language,[[:space:]]Images,[[:space:]]and[[:space:]]Point[[:space:]]Clouds[[:space:]]for[[:space:]]3D[[:space:]]Understanding/7604ec53-7759-4b99-8d3a-2c48883ae4bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UMat_[[:space:]]Uncertainty-Aware[[:space:]]Single[[:space:]]Image[[:space:]]High[[:space:]]Resolution[[:space:]]Material[[:space:]]Capture/db1e6c8b-88a7-48ff-a83b-45737998e762_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UTM_[[:space:]]A[[:space:]]Unified[[:space:]]Multiple[[:space:]]Object[[:space:]]Tracking[[:space:]]Model[[:space:]]With[[:space:]]Identity-Aware[[:space:]]Feature[[:space:]]Enhancement/624c289d-024f-46eb-a5f8-d63009bc6809_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UV[[:space:]]Volumes[[:space:]]for[[:space:]]Real-Time[[:space:]]Rendering[[:space:]]of[[:space:]]Editable[[:space:]]Free-View[[:space:]]Human[[:space:]]Performance/425d6a64-25ae-49e8-ba00-771369a74078_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ultra-High[[:space:]]Resolution[[:space:]]Segmentation[[:space:]]With[[:space:]]Ultra-Rich[[:space:]]Context_[[:space:]]A[[:space:]]Novel[[:space:]]Benchmark/52f4f6ad-31b9-4e8b-a706-539c487500c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ultrahigh[[:space:]]Resolution[[:space:]]Image_Video[[:space:]]Matting[[:space:]]With[[:space:]]Spatio-Temporal[[:space:]]Sparsity/c5bff5e9-401e-4ac1-94a9-0f41455f1eb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unbalanced[[:space:]]Optimal[[:space:]]Transport_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Object[[:space:]]Detection/3222c7a9-c45f-4020-9588-d3c6ac12cb39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unbiased[[:space:]]Multiple[[:space:]]Instance[[:space:]]Learning[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/4402254e-b7bf-4d42-9973-84bee73dff4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unbiased[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]in[[:space:]]Videos/00d911f9-8bc8-455f-81d2-fc5a3be0f2e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uncertainty-Aware[[:space:]]Optimal[[:space:]]Transport[[:space:]]for[[:space:]]Semantically[[:space:]]Coherent[[:space:]]Out-of-Distribution[[:space:]]Detection/41153b08-fedd-4bab-bf4f-c5853e3d6dc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uncertainty-Aware[[:space:]]Unsupervised[[:space:]]Image[[:space:]]Deblurring[[:space:]]With[[:space:]]Deep[[:space:]]Residual[[:space:]]Prior/2453cf22-2c2a-4301-8fb9-33d2a606580b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uncertainty-Aware[[:space:]]Vision-Based[[:space:]]Metric[[:space:]]Cross-View[[:space:]]Geolocalization/f660f7e8-6187-4903-b993-2e161929b068_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uncovering[[:space:]]the[[:space:]]Disentanglement[[:space:]]Capability[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/4e5bc085-b602-4820-8578-defea5fa97c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uncovering[[:space:]]the[[:space:]]Missing[[:space:]]Pattern_[[:space:]]Unified[[:space:]]Framework[[:space:]]Towards[[:space:]]Trajectory[[:space:]]Imputation[[:space:]]and[[:space:]]Prediction/5df10a40-1bce-41e8-a0cf-9fb5f4d8dc07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uncurated[[:space:]]Image-Text[[:space:]]Datasets_[[:space:]]Shedding[[:space:]]Light[[:space:]]on[[:space:]]Demographic[[:space:]]Bias/3bac0784-a27f-4de8-978e-35eb55263456_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Understanding[[:space:]]Deep[[:space:]]Generative[[:space:]]Models[[:space:]]With[[:space:]]Generalized[[:space:]]Empirical[[:space:]]Likelihoods/833954d7-0a8e-46cb-b0b1-41665896b55e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Understanding[[:space:]]Imbalanced[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]Through[[:space:]]Neural[[:space:]]Collapse/ff1d0c49-02df-45bb-857c-01a1a1b21662_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Understanding[[:space:]]Masked[[:space:]]Autoencoders[[:space:]]via[[:space:]]Hierarchical[[:space:]]Latent[[:space:]]Variable[[:space:]]Models/621d3a90-4555-44da-a2f6-238f7d63c0fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Understanding[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling[[:space:]]via[[:space:]]Learning[[:space:]]Occlusion[[:space:]]Invariant[[:space:]]Feature/4e696e5a-afe6-403e-88c1-aaae1fcac52b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Understanding[[:space:]]and[[:space:]]Constructing[[:space:]]Latent[[:space:]]Modality[[:space:]]Structures[[:space:]]in[[:space:]]Multi-Modal[[:space:]]Representation[[:space:]]Learning/c1e021ad-dca6-4888-9ebb-d8e755254788_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Understanding[[:space:]]and[[:space:]]Improving[[:space:]]Features[[:space:]]Learned[[:space:]]in[[:space:]]Deep[[:space:]]Functional[[:space:]]Maps/b4487695-d789-491a-9bb5-075f16d534f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Understanding[[:space:]]and[[:space:]]Improving[[:space:]]Visual[[:space:]]Prompting_[[:space:]]A[[:space:]]Label-Mapping[[:space:]]Perspective/30b17987-daad-422c-98ca-d563efa7190d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Understanding[[:space:]]the[[:space:]]Robustness[[:space:]]of[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]Bird's-Eye-View[[:space:]]Representations[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/6141f14b-886f-46ff-a330-b0574306f469_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uni-Perceiver[[:space:]]v2_[[:space:]]A[[:space:]]Generalist[[:space:]]Model[[:space:]]for[[:space:]]Large-Scale[[:space:]]Vision[[:space:]]and[[:space:]]Vision-Language[[:space:]]Tasks/59367da1-d8f2-4632-8c24-ad7b699a51cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uni3D_[[:space:]]A[[:space:]]Unified[[:space:]]Baseline[[:space:]]for[[:space:]]Multi-Dataset[[:space:]]3D[[:space:]]Object[[:space:]]Detection/1c841edf-4291-4e03-8807-6e74acddca11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniDAformer_[[:space:]]Unified[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]Transformer[[:space:]]via[[:space:]]Hierarchical[[:space:]]Mask[[:space:]]Calibration/29ed08fa-e548-469e-85b4-3007ad20d135_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniDexGrasp_[[:space:]]Universal[[:space:]]Robotic[[:space:]]Dexterous[[:space:]]Grasping[[:space:]]via[[:space:]]Learning[[:space:]]Diverse[[:space:]]Proposal[[:space:]]Generation[[:space:]]and[[:space:]]Goal-Conditioned[[:space:]]Policy/64edd7f0-cf37-4324-ad71-fbd0f8c51941_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniDistill_[[:space:]]A[[:space:]]Universal[[:space:]]Cross-Modality[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]Framework[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]Bird's-Eye[[:space:]]View/d1997566-9d78-4dc1-b9b2-8c96d7ef3c8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniHCP_[[:space:]]A[[:space:]]Unified[[:space:]]Model[[:space:]]for[[:space:]]Human-Centric[[:space:]]Perceptions/87261b9b-e480-4e77-b66b-baeebccbae7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniSim_[[:space:]]A[[:space:]]Neural[[:space:]]Closed-Loop[[:space:]]Sensor[[:space:]]Simulator/a64eb95a-aa1d-4a25-a16f-5464fd76cda9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unicode[[:space:]]Analogies_[[:space:]]An[[:space:]]Anti-Objectivist[[:space:]]Visual[[:space:]]Reasoning[[:space:]]Challenge/5137b37a-b2ae-416e-912d-f4b84e879283_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unified[[:space:]]Keypoint-Based[[:space:]]Action[[:space:]]Recognition[[:space:]]Framework[[:space:]]via[[:space:]]Structured[[:space:]]Keypoint[[:space:]]Pooling/882a8408-443b-403e-9041-44efb00b14e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unified[[:space:]]Mask[[:space:]]Embedding[[:space:]]and[[:space:]]Correspondence[[:space:]]Learning[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Video[[:space:]]Segmentation/734336d6-09d3-451f-8a54-f2692004671e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unified[[:space:]]Pose[[:space:]]Sequence[[:space:]]Modeling/24bb27b4-a1f6-4ef1-a0a6-bcfe910ca06c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unifying[[:space:]]Layout[[:space:]]Generation[[:space:]]With[[:space:]]a[[:space:]]Decoupled[[:space:]]Diffusion[[:space:]]Model/244d6f81-8171-45c0-ba85-c22d4324827b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unifying[[:space:]]Short[[:space:]]and[[:space:]]Long-Term[[:space:]]Tracking[[:space:]]With[[:space:]]Graph[[:space:]]Hierarchies/dfc814a2-7190-4878-b4d9-003bcffb32f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unifying[[:space:]]Vision,[[:space:]]Text,[[:space:]]and[[:space:]]Layout[[:space:]]for[[:space:]]Universal[[:space:]]Document[[:space:]]Processing/7d5f5f0d-fcab-4980-8cb1-61c2921f4f00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unite[[:space:]]and[[:space:]]Conquer_[[:space:]]Plug[[:space:]]&[[:space:]]Play[[:space:]]Multi-Modal[[:space:]]Synthesis[[:space:]]Using[[:space:]]Diffusion[[:space:]]Models/09532b4a-1d1c-4435-8004-0107b20fe7c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Universal[[:space:]]Instance[[:space:]]Perception[[:space:]]As[[:space:]]Object[[:space:]]Discovery[[:space:]]and[[:space:]]Retrieval/5731cf92-73a4-463b-b1f2-60f0b85afc69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unknown[[:space:]]Sniffer[[:space:]]for[[:space:]]Object[[:space:]]Detection_[[:space:]]Don't[[:space:]]Turn[[:space:]]a[[:space:]]Blind[[:space:]]Eye[[:space:]]to[[:space:]]Unknown[[:space:]]Objects/4ed53c3a-5d60-4c01-814e-690e3a2ed7cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unlearnable[[:space:]]Clusters_[[:space:]]Towards[[:space:]]Label-Agnostic[[:space:]]Unlearnable[[:space:]]Examples/2ab6e806-12b5-44a7-97b5-efea87cc6124_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unpaired[[:space:]]Image-to-Image[[:space:]]Translation[[:space:]]With[[:space:]]Shortest[[:space:]]Path[[:space:]]Regularization/846adc23-e0ff-4011-b440-ca531cb4ac38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Representation[[:space:]]Learning[[:space:]]by[[:space:]]Triangle[[:space:]]Constrained[[:space:]]Contrast[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/cd736d1a-a755-48f4-ab0f-1c09e524fefd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]3D[[:space:]]Shape[[:space:]]Reconstruction[[:space:]]by[[:space:]]Part[[:space:]]Retrieval[[:space:]]and[[:space:]]Assembly/8ace7411-301e-43ab-b02e-cfe983f923dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Continual[[:space:]]Semantic[[:space:]]Adaptation[[:space:]]Through[[:space:]]Neural[[:space:]]Rendering/5f5485a6-225a-4933-bdff-ff9de69e91da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Contour[[:space:]]Tracking[[:space:]]of[[:space:]]Live[[:space:]]Cells[[:space:]]by[[:space:]]Mechanical[[:space:]]and[[:space:]]Cycle[[:space:]]Consistency[[:space:]]Losses/ed077bf1-3bdf-4576-8f09-7346daa432b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Cumulative[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Foggy[[:space:]]Scene[[:space:]]Optical[[:space:]]Flow/0a9ca3cb-b73d-4ad7-be1a-2c9e6860c52e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Deep[[:space:]]Asymmetric[[:space:]]Stereo[[:space:]]Matching[[:space:]]With[[:space:]]Spatially-Adaptive[[:space:]]Self-Similarity/bbe00045-ba21-4b53-93e7-439b6b031a44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Deep[[:space:]]Probabilistic[[:space:]]Approach[[:space:]]for[[:space:]]Partial[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/ebde82dd-8afa-4f76-9a1a-7f9885329aeb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Domain[[:space:]]Adaption[[:space:]]With[[:space:]]Pixel-Level[[:space:]]Discriminator[[:space:]]for[[:space:]]Image-Aware[[:space:]]Layout[[:space:]]Generation/3881af3e-2c59-4828-afa1-b4895f1c59f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Inference[[:space:]]of[[:space:]]Signed[[:space:]]Distance[[:space:]]Functions[[:space:]]From[[:space:]]Single[[:space:]]Sparse[[:space:]]Point[[:space:]]Clouds[[:space:]]Without[[:space:]]Learning[[:space:]]Priors/e9a90f6a-c6f5-4eb5-b9b9-33d4504a11a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Intrinsic[[:space:]]Image[[:space:]]Decomposition[[:space:]]With[[:space:]]LiDAR[[:space:]]Intensity/35ea5174-effb-4c97-857c-8d3da07b5a4c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Object[[:space:]]Localization_[[:space:]]Observing[[:space:]]the[[:space:]]Background[[:space:]]To[[:space:]]Discover[[:space:]]Objects/1cb995f4-45e6-45a4-9b50-6eb7b974c912_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Sampling[[:space:]]Promoting[[:space:]]for[[:space:]]Stochastic[[:space:]]Human[[:space:]]Trajectory[[:space:]]Prediction/b5df89e4-4ecf-42c8-8718-07545bc0f6f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Space-Time[[:space:]]Network[[:space:]]for[[:space:]]Temporally-Consistent[[:space:]]Segmentation[[:space:]]of[[:space:]]Multiple[[:space:]]Motions/d2546b6c-fccb-415e-82b7-f83a8bc95a1b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification[[:space:]]via[[:space:]]Progressive[[:space:]]Graph[[:space:]]Matching[[:space:]]and[[:space:]]Alternate[[:space:]]Learning/e57e82e5-88ac-426a-a532-759e19819d22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Volumetric[[:space:]]Animation/a889d958-9019-4d50-a04e-3fe8005d0a99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Upcycling[[:space:]]Models[[:space:]]Under[[:space:]]Domain[[:space:]]and[[:space:]]Category[[:space:]]Shift/3c15dfe5-4f16-4f42-a8ec-c7637b296b92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Use[[:space:]]Your[[:space:]]Head_[[:space:]]Improving[[:space:]]Long-Tail[[:space:]]Video[[:space:]]Recognition/02c326fa-dacd-45cb-925e-17f5786781a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/V2V4Real_[[:space:]]A[[:space:]]Real-World[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]for[[:space:]]Vehicle-to-Vehicle[[:space:]]Cooperative[[:space:]]Perception/fdca9b81-9a0d-41af-8c88-1660cba5765d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/V2X-Seq_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Sequential[[:space:]]Dataset[[:space:]]for[[:space:]]Vehicle-Infrastructure[[:space:]]Cooperative[[:space:]]Perception[[:space:]]and[[:space:]]Forecasting/e83ea37e-f2ff-4f3b-88c4-f0bb731426e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VDN-NeRF_[[:space:]]Resolving[[:space:]]Shape-Radiance[[:space:]]Ambiguity[[:space:]]via[[:space:]]View-Dependence[[:space:]]Normalization/540bd833-4752-40a8-ab47-4b518dcfb3c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VGFlow_[[:space:]]Visibility[[:space:]]Guided[[:space:]]Flow[[:space:]]Network[[:space:]]for[[:space:]]Human[[:space:]]Reposing/fccfa24b-5aac-401b-8519-80e69c5f2cc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VILA_[[:space:]]Learning[[:space:]]Image[[:space:]]Aesthetics[[:space:]]From[[:space:]]User[[:space:]]Comments[[:space:]]With[[:space:]]Vision-Language[[:space:]]Pretraining/4ecfdbe3-e986-47dd-909d-fd3560531cf4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VIVE3D_[[:space:]]Viewpoint-Independent[[:space:]]Video[[:space:]]Editing[[:space:]]Using[[:space:]]3D-Aware[[:space:]]GANs/914339c1-7256-4062-a27c-0394e139b38e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VL-SAT_[[:space:]]Visual-Linguistic[[:space:]]Semantics[[:space:]]Assisted[[:space:]]Training[[:space:]]for[[:space:]]3D[[:space:]]Semantic[[:space:]]Scene[[:space:]]Graph[[:space:]]Prediction[[:space:]]in[[:space:]]Point[[:space:]]Cloud/9f03a635-774a-48bb-af48-0c4191918e67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VLPD_[[:space:]]Context-Aware[[:space:]]Pedestrian[[:space:]]Detection[[:space:]]via[[:space:]]Vision-Language[[:space:]]Semantic[[:space:]]Self-Supervision/b0007c8e-9aac-4b8b-8e66-70458abdf876_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VNE_[[:space:]]An[[:space:]]Effective[[:space:]]Method[[:space:]]for[[:space:]]Improving[[:space:]]Deep[[:space:]]Representation[[:space:]]by[[:space:]]Manipulating[[:space:]]Eigenvalue[[:space:]]Distribution/d4929f89-fed0-43c1-bcfa-3d6f742e201c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VQACL_[[:space:]]A[[:space:]]Novel[[:space:]]Visual[[:space:]]Question[[:space:]]Answering[[:space:]]Continual[[:space:]]Learning[[:space:]]Setting/30379e7e-4f44-47af-998c-d7ffc4dbb7f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Variational[[:space:]]Distribution[[:space:]]Learning[[:space:]]for[[:space:]]Unsupervised[[:space:]]Text-to-Image[[:space:]]Generation/5be8eccd-0a56-419f-a6eb-143c2eccfa07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VecFontSDF_[[:space:]]Learning[[:space:]]To[[:space:]]Reconstruct[[:space:]]and[[:space:]]Synthesize[[:space:]]High-Quality[[:space:]]Vector[[:space:]]Fonts[[:space:]]via[[:space:]]Signed[[:space:]]Distance[[:space:]]Functions/2cd194ae-3804-42af-b588-982f73d54ecc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Vector[[:space:]]Quantization[[:space:]]With[[:space:]]Self-Attention[[:space:]]for[[:space:]]Quality-Independent[[:space:]]Representation[[:space:]]Learning/6dae6e0e-85c2-462f-ad28-48a502e0ead4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VectorFloorSeg_[[:space:]]Two-Stream[[:space:]]Graph[[:space:]]Attention[[:space:]]Network[[:space:]]for[[:space:]]Vectorized[[:space:]]Roughcast[[:space:]]Floorplan[[:space:]]Segmentation/2d87f202-c410-4dc9-974c-32437e3f3c59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VectorFusion_[[:space:]]Text-to-SVG[[:space:]]by[[:space:]]Abstracting[[:space:]]Pixel-Based[[:space:]]Diffusion[[:space:]]Models/7acd215f-8fa7-43bb-a0fe-0b3af8d80b91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ViLEM_[[:space:]]Visual-Language[[:space:]]Error[[:space:]]Modeling[[:space:]]for[[:space:]]Image-Text[[:space:]]Retrieval/76d9e5e2-2c10-4447-b4f2-f0a59b348758_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ViP3D_[[:space:]]End-to-End[[:space:]]Visual[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]via[[:space:]]3D[[:space:]]Agent[[:space:]]Queries/27e174fb-a516-4d34-85c5-c27974ef67d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ViPLO_[[:space:]]Vision[[:space:]]Transformer[[:space:]]Based[[:space:]]Pose-Conditioned[[:space:]]Self-Loop[[:space:]]Graph[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection/d6bea0ce-740f-4f38-95a3-6ce4f1b1e783_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ViTs[[:space:]]for[[:space:]]SITS_[[:space:]]Vision[[:space:]]Transformers[[:space:]]for[[:space:]]Satellite[[:space:]]Image[[:space:]]Time[[:space:]]Series/c81be6d6-e82f-41f4-8e36-d3be189d9671_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Vid2Avatar_[[:space:]]3D[[:space:]]Avatar[[:space:]]Reconstruction[[:space:]]From[[:space:]]Videos[[:space:]]in[[:space:]]the[[:space:]]Wild[[:space:]]via[[:space:]]Self-Supervised[[:space:]]Scene[[:space:]]Decomposition/df1406f7-8062-49a0-8908-945594579c5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Vid2Seq_[[:space:]]Large-Scale[[:space:]]Pretraining[[:space:]]of[[:space:]]a[[:space:]]Visual[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Dense[[:space:]]Video[[:space:]]Captioning/b9c47081-ff83-4326-bb08-04c319cbcb40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]Compression[[:space:]]With[[:space:]]Entropy-Constrained[[:space:]]Neural[[:space:]]Representations/8518b033-f83c-4690-bd11-7f40c6d49fcf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]Dehazing[[:space:]]via[[:space:]]a[[:space:]]Multi-Range[[:space:]]Temporal[[:space:]]Alignment[[:space:]]Network[[:space:]]With[[:space:]]Physical[[:space:]]Prior/6413b92b-eb49-4efd-a354-6e883987ce8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]Event[[:space:]]Restoration[[:space:]]Based[[:space:]]on[[:space:]]Keyframes[[:space:]]for[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/c970c67e-3662-4ddf-90be-90ba10876112_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]Probabilistic[[:space:]]Diffusion[[:space:]]Models[[:space:]]in[[:space:]]Projected[[:space:]]Latent[[:space:]]Space/336566d0-dabc-4f14-9ca9-52a57277f742_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Action[[:space:]]Recognition/33f1c449-3f79-4508-bc03-b3bf3e85f530_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video-Text[[:space:]]As[[:space:]]Game[[:space:]]Players_[[:space:]]Hierarchical[[:space:]]Banzhaf[[:space:]]Interaction[[:space:]]for[[:space:]]Cross-Modal[[:space:]]Representation[[:space:]]Learning/ae4008cb-14e6-4b54-91f1-9a91d52ad592_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VideoMAE[[:space:]]V2_[[:space:]]Scaling[[:space:]]Video[[:space:]]Masked[[:space:]]Autoencoders[[:space:]]With[[:space:]]Dual[[:space:]]Masking/25e22258-562d-47aa-a018-73673a194e2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VideoTrack_[[:space:]]Learning[[:space:]]To[[:space:]]Track[[:space:]]Objects[[:space:]]via[[:space:]]Video[[:space:]]Transformer/adcfad4a-40e1-4dc5-b7a8-ba2dfa1ffa85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ViewNet_[[:space:]]A[[:space:]]Novel[[:space:]]Projection-Based[[:space:]]Backbone[[:space:]]With[[:space:]]View[[:space:]]Pooling[[:space:]]for[[:space:]]Few-Shot[[:space:]]Point[[:space:]]Cloud[[:space:]]Classification/fb5c7508-9ab8-48a8-8eaf-0644fc0c2430_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Viewpoint[[:space:]]Equivariance[[:space:]]for[[:space:]]Multi-View[[:space:]]3D[[:space:]]Object[[:space:]]Detection/e2a33c5b-9000-48df-9c5d-7c34cd772f80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VindLU_[[:space:]]A[[:space:]]Recipe[[:space:]]for[[:space:]]Effective[[:space:]]Video-and-Language[[:space:]]Pretraining/0d95737f-2410-41f9-9e5b-2f590234c1e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Virtual[[:space:]]Occlusions[[:space:]]Through[[:space:]]Implicit[[:space:]]Depth/01531298-cc68-451b-b2e2-ef0941b75d33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Virtual[[:space:]]Sparse[[:space:]]Convolution[[:space:]]for[[:space:]]Multimodal[[:space:]]3D[[:space:]]Object[[:space:]]Detection/3ffb3595-e9fe-496e-8907-7b93b54e3f79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VisFusion_[[:space:]]Visibility-Aware[[:space:]]Online[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]From[[:space:]]Videos/0ce3b4dc-c3ec-47ac-a5e0-ba59a1f938d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visibility[[:space:]]Aware[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Tracking[[:space:]]From[[:space:]]Single[[:space:]]RGB[[:space:]]Camera/f076d522-ec12-4b37-b209-a0e0a2a08fb9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visibility[[:space:]]Constrained[[:space:]]Wide-Band[[:space:]]Illumination[[:space:]]Spectrum[[:space:]]Design[[:space:]]for[[:space:]]Seeing-in-the-Dark/6fbe7647-e313-4c97-8f72-e27b9a926585_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Vision[[:space:]]Transformers[[:space:]]Are[[:space:]]Good[[:space:]]Mask[[:space:]]Auto-Labelers/f2e22d5e-4323-4dc7-bda5-f274d9f1cb09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Vision[[:space:]]Transformers[[:space:]]Are[[:space:]]Parameter-Efficient[[:space:]]Audio-Visual[[:space:]]Learners/1f0afbfe-c5e6-48f1-9df9-0417a3eab046_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Atoms_[[:space:]]Pre-Training[[:space:]]Vision[[:space:]]Transformers[[:space:]]With[[:space:]]Sinusoidal[[:space:]]Waves/ec60fe59-4411-4b1a-9d72-de7a8707a3c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]DNA_[[:space:]]Representing[[:space:]]and[[:space:]]Comparing[[:space:]]Images[[:space:]]Using[[:space:]]Distributions[[:space:]]of[[:space:]]Neuron[[:space:]]Activations/b78bb931-20bd-4385-a189-c871a0c2e531_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Dependency[[:space:]]Transformers_[[:space:]]Dependency[[:space:]]Tree[[:space:]]Emerges[[:space:]]From[[:space:]]Reversed[[:space:]]Attention/eb8a71e6-8687-4719-8599-3cc111277e00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Exemplar[[:space:]]Driven[[:space:]]Task-Prompting[[:space:]]for[[:space:]]Unified[[:space:]]Perception[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/ec72af2f-e872-4a27-9d78-76961a994e64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Language[[:space:]]Pretrained[[:space:]]Multiple[[:space:]]Instance[[:space:]]Zero-Shot[[:space:]]Transfer[[:space:]]for[[:space:]]Histopathology[[:space:]]Images/3eed8b63-db94-4386-9d6b-0afd57fe0568_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Localization[[:space:]]Using[[:space:]]Imperfect[[:space:]]3D[[:space:]]Models[[:space:]]From[[:space:]]the[[:space:]]Internet/021adfc1-9807-4bee-9de9-fa313a3d1fba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Programming_[[:space:]]Compositional[[:space:]]Visual[[:space:]]Reasoning[[:space:]]Without[[:space:]]Training/e98aa32c-22aa-4f46-97e9-5322514e7d89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Prompt[[:space:]]Multi-Modal[[:space:]]Tracking/be794b75-95bb-47fa-bb39-d63a19921375_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Generative[[:space:]]Transfer[[:space:]]Learning/eae3511a-ea5d-4f82-aed2-5e4864ef1ec8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Query[[:space:]]Tuning_[[:space:]]Towards[[:space:]]Effective[[:space:]]Usage[[:space:]]of[[:space:]]Intermediate[[:space:]]Representations[[:space:]]for[[:space:]]Parameter[[:space:]]and[[:space:]]Memory[[:space:]]Efficient[[:space:]]Transfer[[:space:]]Learning/731b5636-f122-4f27-b7f2-09d2dfcd7f43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Recognition[[:space:]]by[[:space:]]Request/ddc339ab-02e0-4aeb-99a8-ba74dd2b3566_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Recognition-Driven[[:space:]]Image[[:space:]]Restoration[[:space:]]for[[:space:]]Multiple[[:space:]]Degradation[[:space:]]With[[:space:]]Intrinsic[[:space:]]Semantics[[:space:]]Recovery/a137bf3b-d59d-44c8-98cc-f5005ef62816_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual-Language[[:space:]]Prompt[[:space:]]Tuning[[:space:]]With[[:space:]]Knowledge-Guided[[:space:]]Context[[:space:]]Optimization/999e3264-7e41-4db2-b70b-20730b6ee64d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual-Tactile[[:space:]]Sensing[[:space:]]for[[:space:]]In-Hand[[:space:]]Object[[:space:]]Reconstruction/0a04f1df-b372-48ff-a6af-93907a73626e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Vita-CLIP_[[:space:]]Video[[:space:]]and[[:space:]]Text[[:space:]]Adaptive[[:space:]]CLIP[[:space:]]via[[:space:]]Multimodal[[:space:]]Prompting/56dec0ed-0c95-4102-94ff-4aff4ae7b944_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VoP_[[:space:]]Text-Video[[:space:]]Co-Operative[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Cross-Modal[[:space:]]Retrieval/459441ac-5990-46e0-b1ab-75c0494563dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VolRecon_[[:space:]]Volume[[:space:]]Rendering[[:space:]]of[[:space:]]Signed[[:space:]]Ray[[:space:]]Distance[[:space:]]Functions[[:space:]]for[[:space:]]Generalizable[[:space:]]Multi-View[[:space:]]Reconstruction/b4cc2de1-5e12-400a-9325-16358db40ebd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VoxFormer_[[:space:]]Sparse[[:space:]]Voxel[[:space:]]Transformer[[:space:]]for[[:space:]]Camera-Based[[:space:]]3D[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion/d96bee95-35f9-4ce0-b771-d50d19387763_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VoxelNeXt_[[:space:]]Fully[[:space:]]Sparse[[:space:]]VoxelNet[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]and[[:space:]]Tracking/432f59b0-a396-432f-8ee6-b96bfe338bc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/WINNER_[[:space:]]Weakly-Supervised[[:space:]]hIerarchical[[:space:]]decompositioN[[:space:]]and[[:space:]]aligNment[[:space:]]for[[:space:]]Spatio-tEmporal[[:space:]]Video[[:space:]]gRounding/2c16fdab-ddbe-42a8-b456-4dec7f1f9e29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/WIRE_[[:space:]]Wavelet[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations/3d35a1f9-2925-4072-adf5-c46c6377bc1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Watch[[:space:]]or[[:space:]]Listen_[[:space:]]Robust[[:space:]]Audio-Visual[[:space:]]Speech[[:space:]]Recognition[[:space:]]With[[:space:]]Visual[[:space:]]Corruption[[:space:]]Modeling[[:space:]]and[[:space:]]Reliability[[:space:]]Scoring/136e52bf-791c-4dc2-a572-578b02e8c70f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Wavelet[[:space:]]Diffusion[[:space:]]Models[[:space:]]Are[[:space:]]Fast[[:space:]]and[[:space:]]Scalable[[:space:]]Image[[:space:]]Generators/7982af18-9a43-4bdf-ad38-4a383c96904d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weak-Shot[[:space:]]Object[[:space:]]Detection[[:space:]]Through[[:space:]]Mutual[[:space:]]Knowledge[[:space:]]Transfer/8c6a6762-d71f-4faa-b256-7597418d043c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly[[:space:]]Supervised[[:space:]]Class-Agnostic[[:space:]]Motion[[:space:]]Prediction[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/70679f57-3528-49d4-8d72-ecb7210e6b1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly[[:space:]]Supervised[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]Using[[:space:]]Multi-View[[:space:]]Projection[[:space:]]and[[:space:]]Direction[[:space:]]Consistency/0d41b6f5-6504-41a7-9240-1d4ed286e59e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly[[:space:]]Supervised[[:space:]]Posture[[:space:]]Mining[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Classification/a7e113b2-41b3-4706-8ee6-09f857d3e226_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly[[:space:]]Supervised[[:space:]]Segmentation[[:space:]]With[[:space:]]Point[[:space:]]Annotations[[:space:]]for[[:space:]]Histopathology[[:space:]]Images[[:space:]]via[[:space:]]Contrast-Based[[:space:]]Variational[[:space:]]Model/90b63037-8972-4f4a-8c41-d187b0bf325a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]via[[:space:]]Adversarial[[:space:]]Learning[[:space:]]of[[:space:]]Classifier[[:space:]]and[[:space:]]Reconstructor/246b634d-3ffe-4b54-b54a-b4209c598baf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly[[:space:]]Supervised[[:space:]]Temporal[[:space:]]Sentence[[:space:]]Grounding[[:space:]]With[[:space:]]Uncertainty-Guided[[:space:]]Self-Training/b8cf3c8b-20ab-480e-92e3-385e257b0c1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly[[:space:]]Supervised[[:space:]]Video[[:space:]]Emotion[[:space:]]Detection[[:space:]]and[[:space:]]Prediction[[:space:]]via[[:space:]]Cross-Modal[[:space:]]Temporal[[:space:]]Erasing[[:space:]]Network/1472d92e-555b-4fe7-a256-9432ba18084d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly[[:space:]]Supervised[[:space:]]Video[[:space:]]Representation[[:space:]]Learning[[:space:]]With[[:space:]]Unaligned[[:space:]]Text[[:space:]]for[[:space:]]Sequential[[:space:]]Videos/311757a5-eda6-46ca-aa36-aff47d290918_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly-Supervised[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]With[[:space:]]Prototypical[[:space:]]Contrastive[[:space:]]Learning/1e224334-9ad3-4b58-bfb4-98cde0d15658_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly-Supervised[[:space:]]Single-View[[:space:]]Image[[:space:]]Relighting/87d0d7e4-535e-4b61-892c-3c7cad55409a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/WeatherStream_[[:space:]]Light[[:space:]]Transport[[:space:]]Automation[[:space:]]of[[:space:]]Single[[:space:]]Image[[:space:]]Deweathering/5be309f3-341d-4dc0-ae7f-47b19e78bbce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/What[[:space:]]Can[[:space:]]Human[[:space:]]Sketches[[:space:]]Do[[:space:]]for[[:space:]]Object[[:space:]]Detection_/c43b140c-99b6-4cac-9f06-49dcf54cf293_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/What[[:space:]]Happened[[:space:]]3[[:space:]]Seconds[[:space:]]Ago_[[:space:]]Inferring[[:space:]]the[[:space:]]Past[[:space:]]With[[:space:]]Thermal[[:space:]]Imaging/416d4ca0-2dcd-4015-b260-859cdaac6b85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/What[[:space:]]You[[:space:]]Can[[:space:]]Reconstruct[[:space:]]From[[:space:]]a[[:space:]]Shadow/c29e7041-65bd-4aeb-bfd1-ceac46ce18c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Where[[:space:]]Is[[:space:]]My[[:space:]]Spot_[[:space:]]Few-Shot[[:space:]]Image[[:space:]]Generation[[:space:]]via[[:space:]]Latent[[:space:]]Subspace[[:space:]]Optimization/1031bb92-db13-42ec-9221-61b0dfa595d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Where[[:space:]]Is[[:space:]]My[[:space:]]Wallet_[[:space:]]Modeling[[:space:]]Object[[:space:]]Proposal[[:space:]]Sets[[:space:]]for[[:space:]]Egocentric[[:space:]]Visual[[:space:]]Query[[:space:]]Localization/e2a7879d-cafa-4f96-8e84-bef9e933eb6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Where[[:space:]]We[[:space:]]Are[[:space:]]and[[:space:]]What[[:space:]]We're[[:space:]]Looking[[:space:]]At_[[:space:]]Query[[:space:]]Based[[:space:]]Worldwide[[:space:]]Image[[:space:]]Geo-Localization[[:space:]]Using[[:space:]]Hierarchies[[:space:]]and[[:space:]]Scenes/f61294e6-7fe3-4a34-b145-41fa5cf48668_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Why[[:space:]]Is[[:space:]]the[[:space:]]Winner[[:space:]]the[[:space:]]Best_/c2aee3fd-954b-4d7e-8962-c326d7deb558_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Wide-Angle[[:space:]]Rectification[[:space:]]via[[:space:]]Content-Aware[[:space:]]Conformal[[:space:]]Mapping/2bf49664-8b7d-44ba-b318-8dc9cbed81d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/WildLight_[[:space:]]In-the-Wild[[:space:]]Inverse[[:space:]]Rendering[[:space:]]With[[:space:]]a[[:space:]]Flashlight/ad44787d-45e0-4c06-832c-fb431c8115c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/WinCLIP_[[:space:]]Zero-_Few-Shot[[:space:]]Anomaly[[:space:]]Classification[[:space:]]and[[:space:]]Segmentation/c208a5f9-4a69-4298-9ef4-2fc6584b97e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/X-Avatar_[[:space:]]Expressive[[:space:]]Human[[:space:]]Avatars/4d4638b1-270b-4ccf-a319-81705a8c8679_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/X-Pruner_[[:space:]]eXplainable[[:space:]]Pruning[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/6f66f8ed-cbf0-409c-a3fd-c69f5a71e613_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/X3KD_[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]Across[[:space:]]Modalities,[[:space:]]Tasks[[:space:]]and[[:space:]]Stages[[:space:]]for[[:space:]]Multi-Camera[[:space:]]3D[[:space:]]Object[[:space:]]Detection/f696bb03-5dbb-45b1-b60f-76f377d08211_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/YOLOv7_[[:space:]]Trainable[[:space:]]Bag-of-Freebies[[:space:]]Sets[[:space:]]New[[:space:]]State-of-the-Art[[:space:]]for[[:space:]]Real-Time[[:space:]]Object[[:space:]]Detectors/bad5ec70-8097-405f-b102-82274b9102e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/You[[:space:]]Are[[:space:]]Catching[[:space:]]My[[:space:]]Attention_[[:space:]]Are[[:space:]]Vision[[:space:]]Transformers[[:space:]]Bad[[:space:]]Learners[[:space:]]Under[[:space:]]Backdoor[[:space:]]Attacks_/bb764830-bdfe-4379-b9f8-43791f9b7d93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/You[[:space:]]Can[[:space:]]Ground[[:space:]]Earlier[[:space:]]Than[[:space:]]See_[[:space:]]An[[:space:]]Effective[[:space:]]and[[:space:]]Efficient[[:space:]]Pipeline[[:space:]]for[[:space:]]Temporal[[:space:]]Sentence[[:space:]]Grounding[[:space:]]in[[:space:]]Compressed[[:space:]]Videos/cce2da0a-23d0-4b85-8352-24557330c8e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/You[[:space:]]Do[[:space:]]Not[[:space:]]Need[[:space:]]Additional[[:space:]]Priors[[:space:]]or[[:space:]]Regularizers[[:space:]]in[[:space:]]Retinex-Based[[:space:]]Low-Light[[:space:]]Image[[:space:]]Enhancement/0febdd3e-df0c-45bf-a6a9-30b263888e4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/You[[:space:]]Need[[:space:]]Multiple[[:space:]]Exiting_[[:space:]]Dynamic[[:space:]]Early[[:space:]]Exiting[[:space:]]for[[:space:]]Accelerating[[:space:]]Unified[[:space:]]Vision[[:space:]]Language[[:space:]]Model/53a5c1c3-ace3-430d-945a-8a03bdb30ee9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/You[[:space:]]Only[[:space:]]Segment[[:space:]]Once_[[:space:]]Towards[[:space:]]Real-Time[[:space:]]Panoptic[[:space:]]Segmentation/7d312572-45e1-47db-9f44-a0eb92b0a329_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ZBS_[[:space:]]Zero-Shot[[:space:]]Background[[:space:]]Subtraction[[:space:]]via[[:space:]]Instance-Level[[:space:]]Background[[:space:]]Modeling[[:space:]]and[[:space:]]Foreground[[:space:]]Selection/e0b034d0-f139-4bb8-820c-8c8353faaa16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ZegCLIP_[[:space:]]Towards[[:space:]]Adapting[[:space:]]CLIP[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Semantic[[:space:]]Segmentation/5aa3c465-5800-4cf0-842d-3ff527e740db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Dual-Lens[[:space:]]Super-Resolution/e5623582-a7e1-46ef-8ecf-c8f00837c751_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Everything[[:space:]]Sketch-Based[[:space:]]Image[[:space:]]Retrieval,[[:space:]]and[[:space:]]in[[:space:]]Explainable[[:space:]]Style/225245ce-4b02-4bed-a89c-519460e55a67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Generative[[:space:]]Model[[:space:]]Adaptation[[:space:]]via[[:space:]]Image-Specific[[:space:]]Prompt[[:space:]]Learning/b1c7981e-30bf-4cd6-b2ff-8dd7951c08ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Model[[:space:]]Diagnosis/549384d4-f244-4966-8076-15abf189c0a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Noise2Noise_[[:space:]]Efficient[[:space:]]Image[[:space:]]Denoising[[:space:]]Without[[:space:]]Any[[:space:]]Data/2d536e4a-ee00-4291-84f0-1f5cbbcd1b0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Object[[:space:]]Counting/a8ad52a0-ee27-4576-981a-5efd9d6920f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Pose[[:space:]]Transfer[[:space:]]for[[:space:]]Unrigged[[:space:]]Stylized[[:space:]]3D[[:space:]]Characters/5ada3ef3-b974-4dfe-9b9c-f168fb79ad07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation[[:space:]]With[[:space:]]Global-Local[[:space:]]Context[[:space:]]Features/d944ca48-5a24-4209-88df-55a9c1e47851_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Text-to-Parameter[[:space:]]Translation[[:space:]]for[[:space:]]Game[[:space:]]Character[[:space:]]Auto-Creation/120ba89b-8ae1-4464-85a9-dac434567f6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/expOSE_[[:space:]]Accurate[[:space:]]Initialization-Free[[:space:]]Projective[[:space:]]Factorization[[:space:]]Using[[:space:]]Exponential[[:space:]]Regularization/6f578c7a-ccab-49cb-bc75-04e328397fea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/gSDF_[[:space:]]Geometry-Driven[[:space:]]Signed[[:space:]]Distance[[:space:]]Functions[[:space:]]for[[:space:]]3D[[:space:]]Hand-Object[[:space:]]Reconstruction/ab15db8d-1011-4beb-9528-c9481ccf45d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/iCLIP_[[:space:]]Bridging[[:space:]]Image[[:space:]]Classification[[:space:]]and[[:space:]]Contrastive[[:space:]]Language-Image[[:space:]]Pre-Training[[:space:]]for[[:space:]]Visual[[:space:]]Recognition/3b34aaa2-686f-4539-b795-03e3193c1a2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/iDisc_[[:space:]]Internal[[:space:]]Discretization[[:space:]]for[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/f111aba4-6ed5-4778-a051-8ab6247508d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/iQuery_[[:space:]]Instruments[[:space:]]As[[:space:]]Queries[[:space:]]for[[:space:]]Audio-Visual[[:space:]]Sound[[:space:]]Separation/79c43f28-c9f8-4b22-8dca-1b2b47c85f07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/itKD_[[:space:]]Interchange[[:space:]]Transfer-Based[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/0c6f9efe-676e-4d3a-914a-2dc77bae5a75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/pCON_[[:space:]]Polarimetric[[:space:]]Coordinate[[:space:]]Networks[[:space:]]for[[:space:]]Neural[[:space:]]Scene[[:space:]]Representations/762a8e4d-373c-4bab-83f9-f1ad8a1ae928_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/sRGB[[:space:]]Real[[:space:]]Noise[[:space:]]Synthesizing[[:space:]]With[[:space:]]Neighboring[[:space:]]Correlation-Aware[[:space:]]Noise[[:space:]]Model/4d3b606f-ee7b-460f-b2bf-d38a08fa5304_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/vMAP_[[:space:]]Vectorised[[:space:]]Object[[:space:]]Mapping[[:space:]]for[[:space:]]Neural[[:space:]]Field[[:space:]]SLAM/ff44bc7e-edea-45f5-9989-5020c0b824b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]the[[:space:]]Benefits[[:space:]]of[[:space:]]3D[[:space:]]Pose[[:space:]]and[[:space:]]Tracking[[:space:]]for[[:space:]]Human[[:space:]]Action[[:space:]]Recognition/0c904778-ffe0-4595-8277-48b7638fc1b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]the[[:space:]]Effectiveness[[:space:]]of[[:space:]]Partial[[:space:]]Variance[[:space:]]Reduction[[:space:]]in[[:space:]]Federated[[:space:]]Learning[[:space:]]With[[:space:]]Heterogeneous[[:space:]]Data/fbd20275-52b6-475a-b61f-5e4c3c892317_origin.pdf filter=lfs diff=lfs merge=lfs -text