*.7z filter=lfs diff=lfs merge=lfs -text *.arrow filter=lfs diff=lfs merge=lfs -text *.bin filter=lfs diff=lfs merge=lfs -text *.bz2 filter=lfs diff=lfs merge=lfs -text *.ckpt filter=lfs diff=lfs merge=lfs -text *.ftz filter=lfs diff=lfs merge=lfs -text *.gz filter=lfs diff=lfs merge=lfs -text *.h5 filter=lfs diff=lfs merge=lfs -text *.joblib filter=lfs diff=lfs merge=lfs -text *.lfs.* filter=lfs diff=lfs merge=lfs -text *.lz4 filter=lfs diff=lfs merge=lfs -text *.mds filter=lfs diff=lfs merge=lfs -text *.mlmodel filter=lfs diff=lfs merge=lfs -text *.model filter=lfs diff=lfs merge=lfs -text *.msgpack filter=lfs diff=lfs merge=lfs -text *.npy filter=lfs diff=lfs merge=lfs -text *.npz filter=lfs diff=lfs merge=lfs -text *.onnx filter=lfs diff=lfs merge=lfs -text *.ot filter=lfs diff=lfs merge=lfs -text *.parquet filter=lfs diff=lfs merge=lfs -text *.pb filter=lfs diff=lfs merge=lfs -text *.pickle filter=lfs diff=lfs merge=lfs -text *.pkl filter=lfs diff=lfs merge=lfs -text *.pt filter=lfs diff=lfs merge=lfs -text *.pth filter=lfs diff=lfs merge=lfs -text *.rar filter=lfs diff=lfs merge=lfs -text *.safetensors filter=lfs diff=lfs merge=lfs -text saved_model/**/* filter=lfs diff=lfs merge=lfs -text *.tar.* filter=lfs diff=lfs merge=lfs -text *.tar filter=lfs diff=lfs merge=lfs -text *.tflite filter=lfs diff=lfs merge=lfs -text *.tgz filter=lfs diff=lfs merge=lfs -text *.wasm filter=lfs diff=lfs merge=lfs -text *.xz filter=lfs diff=lfs merge=lfs -text *.zip filter=lfs diff=lfs merge=lfs -text *.zst filter=lfs diff=lfs merge=lfs -text *tfevents* filter=lfs diff=lfs merge=lfs -text # Audio files - uncompressed *.pcm filter=lfs diff=lfs merge=lfs -text *.sam filter=lfs diff=lfs merge=lfs -text *.raw filter=lfs diff=lfs merge=lfs -text # Audio files - compressed *.aac filter=lfs diff=lfs merge=lfs -text *.flac filter=lfs diff=lfs merge=lfs -text *.mp3 filter=lfs diff=lfs merge=lfs -text *.ogg filter=lfs diff=lfs merge=lfs -text *.wav filter=lfs diff=lfs merge=lfs -text # Image files - uncompressed *.bmp filter=lfs diff=lfs merge=lfs -text *.gif filter=lfs diff=lfs merge=lfs -text *.png filter=lfs diff=lfs merge=lfs -text *.tiff filter=lfs diff=lfs merge=lfs -text # Image files - compressed *.jpg filter=lfs diff=lfs merge=lfs -text *.jpeg filter=lfs diff=lfs merge=lfs -text *.webp filter=lfs diff=lfs merge=lfs -text # Video files - compressed *.mp4 filter=lfs diff=lfs merge=lfs -text *.webm filter=lfs diff=lfs merge=lfs -text 2025/2.5[[:space:]]Years[[:space:]]in[[:space:]]Class_[[:space:]]A[[:space:]]Multimodal[[:space:]]Textbook[[:space:]]for[[:space:]]Vision-Language[[:space:]]Pretraining/c315c786-998f-4e4e-beb4-40960bff2440_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/2D[[:space:]]Gaussian[[:space:]]Splatting-based[[:space:]]Sparse-view[[:space:]]Transparent[[:space:]]Object[[:space:]]Depth[[:space:]]Reconstruction[[:space:]]via[[:space:]]Physics[[:space:]]Simulation[[:space:]]for[[:space:]]Scene[[:space:]]Update/615e402b-d8fd-490d-ae92-02aaade5a1c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/2HandedAfforder_[[:space:]]Learning[[:space:]]Precise[[:space:]]Actionable[[:space:]]Bimanual[[:space:]]Affordances[[:space:]]from[[:space:]]Human[[:space:]]Videos/8585d5d6-260e-40e3-8a3d-423faa26a89c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D[[:space:]]Gaussian[[:space:]]Map[[:space:]]with[[:space:]]Open-Set[[:space:]]Semantic[[:space:]]Grouping[[:space:]]for[[:space:]]Vision-Language[[:space:]]Navigation/821d75f6-2708-421f-809a-0f68f030db87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Driven[[:space:]]Multi-View[[:space:]]Robust[[:space:]]Physical[[:space:]]Adversarial[[:space:]]Camouflage[[:space:]]Generation/fcad30ae-053c-4c1a-b886-bf0040b8b6ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D[[:space:]]Mesh[[:space:]]Editing[[:space:]]using[[:space:]]Masked[[:space:]]LRMs/1867613e-7c29-4005-a37a-db4fb6360da8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D[[:space:]]Test-time[[:space:]]Adaptation[[:space:]]via[[:space:]]Graph[[:space:]]Spectral[[:space:]]Driven[[:space:]]Point[[:space:]]Shift/0a891609-57d4-472a-a0f3-6de2d73d5c70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3D-MOOD_[[:space:]]Lifting[[:space:]]2D[[:space:]]to[[:space:]]3D[[:space:]]for[[:space:]]Monocular[[:space:]]Open-Set[[:space:]]Object[[:space:]]Detection/f445e101-4de0-4b26-b3af-a770583f8f62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3DGS-LM_[[:space:]]Faster[[:space:]]Gaussian-Splatting[[:space:]]Optimization[[:space:]]with[[:space:]]Levenberg-Marquardt/5b6cde06-e17b-4a1b-a66d-29e02f55a93d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3DGraphLLM_[[:space:]]Combining[[:space:]]Semantic[[:space:]]Graphs[[:space:]]and[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/6773f14e-c691-42a5-9104-c93f85b09206_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3DRealCar_[[:space:]]An[[:space:]]In-the-wild[[:space:]]RGB-D[[:space:]]Car[[:space:]]Dataset[[:space:]]with[[:space:]]360-degree[[:space:]]Views/3761be1c-b405-4d7a-8efc-c95a3e26fd6b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/3DSRBench_[[:space:]]A[[:space:]]Comprehensive[[:space:]]3D[[:space:]]Spatial[[:space:]]Reasoning[[:space:]]Benchmark/43e0b276-0db3-46cb-b420-e0da89085656_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/4D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]SLAM/a23669ea-d27e-417e-9767-5552932980a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/4D[[:space:]]Visual[[:space:]]Pre-training[[:space:]]for[[:space:]]Robot[[:space:]]Learning/c5936d3d-47df-4bf6-90e2-99689c77263e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/4D-Bench_[[:space:]]Benchmarking[[:space:]]Multi-modal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]4D[[:space:]]Object[[:space:]]Understanding/fa6c6a12-5a69-4725-9299-7f5a3aa2b23b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/4DSegStreamer_[[:space:]]Streaming[[:space:]]4D[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]via[[:space:]]Dual[[:space:]]Threads/f7b26332-5664-42b9-a56f-ae7c8fdb5588_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/6DOPE-GS_[[:space:]]Online[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]using[[:space:]]Gaussian[[:space:]]Splatting/9c1d20c3-0054-4369-a715-97da6d54ed7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/7DGS_[[:space:]]Unified[[:space:]]Spatial-Temporal-Angular[[:space:]]Gaussian[[:space:]]Splatting/0e168abc-5aff-4ae1-8723-bac8abf0692e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Conditional[[:space:]]Probability[[:space:]]Framework[[:space:]]for[[:space:]]Compositional[[:space:]]Zero-shot[[:space:]]Learning/c2142832-6bd2-44f6-bf81-22f562597be8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Constrained[[:space:]]Optimization[[:space:]]Approach[[:space:]]for[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Coarsely-posed[[:space:]]Images[[:space:]]and[[:space:]]Noisy[[:space:]]Lidar[[:space:]]Point[[:space:]]Clouds/4a409d43-089a-478e-8619-1edc7c687033_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Differentiable[[:space:]]Wave[[:space:]]Optics[[:space:]]Model[[:space:]]for[[:space:]]End-to-End[[:space:]]Computational[[:space:]]Imaging[[:space:]]System[[:space:]]Optimization/bb54bf5c-7243-47bf-b76e-d4fd7ee99367_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Framework[[:space:]]for[[:space:]]Double-Blind[[:space:]]Federated[[:space:]]Adaptation[[:space:]]of[[:space:]]Foundation[[:space:]]Models/a138201f-22d8-4697-b908-23db12352b14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Good[[:space:]]Teacher[[:space:]]Adapts[[:space:]]Their[[:space:]]Knowledge[[:space:]]for[[:space:]]Distillation/ab3259fc-bd06-4ff6-b5f8-f25d0dcd3a1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Hidden[[:space:]]Stumbling[[:space:]]Block[[:space:]]in[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery_[[:space:]]Distracted[[:space:]]Attention/b17028d0-c273-4565-9f0e-a5ec953fb140_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Hyperdimensional[[:space:]]One[[:space:]]Place[[:space:]]Signature[[:space:]]to[[:space:]]Represent[[:space:]]Them[[:space:]]All_[[:space:]]Stackable[[:space:]]Descriptors[[:space:]]For[[:space:]]Visual[[:space:]]Place[[:space:]]Recognition/cb0fe20b-feaa-4d0c-b8cb-0d351ea3e227_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Lesson[[:space:]]in[[:space:]]Splats_[[:space:]]Teacher-Guided[[:space:]]Diffusion[[:space:]]for[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splats[[:space:]]Generation[[:space:]]with[[:space:]]2D[[:space:]]Supervision/22ba9dcd-0796-422d-abf9-2cfa5b44334a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Linear[[:space:]]N-Point[[:space:]]Solver[[:space:]]for[[:space:]]Structure[[:space:]]and[[:space:]]Motion[[:space:]]from[[:space:]]Asynchronous[[:space:]]Tracks/a1a673db-5b6d-44cc-abd4-1d08688868a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Plug-and-Play[[:space:]]Physical[[:space:]]Motion[[:space:]]Restoration[[:space:]]Approach[[:space:]]for[[:space:]]In-the-Wild[[:space:]]High-Difficulty[[:space:]]Motions/99a335e6-813d-4ff4-ab34-c6e28f412480_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Quality-Guided[[:space:]]Mixture[[:space:]]of[[:space:]]Score-Fusion[[:space:]]Experts[[:space:]]Framework[[:space:]]for[[:space:]]Human[[:space:]]Recognition/bb9038ac-2b62-419b-997e-6bdc71d2c32d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Real-world[[:space:]]Display[[:space:]]Inverse[[:space:]]Rendering[[:space:]]Dataset/2cd1efd3-75cc-4c12-827f-39a7ed7a5d6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Recipe[[:space:]]for[[:space:]]Generating[[:space:]]3D[[:space:]]Worlds[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/3d20ae28-c95d-4006-a1a3-18620afb8229_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Simple[[:space:]]yet[[:space:]]Mighty[[:space:]]Hartley[[:space:]]Diffusion[[:space:]]Versatilist[[:space:]]for[[:space:]]Generalizable[[:space:]]Dense[[:space:]]Vision[[:space:]]Tasks/0ec1e285-377e-476c-8800-45167b9791df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Structure-aware[[:space:]]and[[:space:]]Motion-adaptive[[:space:]]Framework[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]Mamba/4a504570-03f1-4dca-adee-006a933e5720_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Tiny[[:space:]]Change,[[:space:]]A[[:space:]]Giant[[:space:]]Leap_[[:space:]]Long-Tailed[[:space:]]Class-Incremental[[:space:]]Learning[[:space:]]via[[:space:]]Geometric[[:space:]]Prototype[[:space:]]Alignment/fd7ddeb2-3030-42fa-a15e-366b5dc76154_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Token-level[[:space:]]Text[[:space:]]Image[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Document[[:space:]]Understanding/33b49a65-6431-4905-9c8c-d6d54b94a1f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Industrial[[:space:]]Cel-Animation[[:space:]]Colorization[[:space:]]with[[:space:]]Temporal-Structural[[:space:]]Awareness/9c3022ef-23cf-4aaf-8cb7-6851fa6c0a15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Motion[[:space:]]Reasoning[[:space:]]and[[:space:]]Generation[[:space:]]in[[:space:]]Human[[:space:]]Interaction/4f03f794-2650-47a5-86ad-744e17532321_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Unified[[:space:]]Framework[[:space:]]to[[:space:]]BRIDGE[[:space:]]Complete[[:space:]]and[[:space:]]Incomplete[[:space:]]Deep[[:space:]]Multi-View[[:space:]]Clustering[[:space:]]under[[:space:]]Non-IID[[:space:]]Missing[[:space:]]Patterns/bfa641ce-a498-4633-8908-973beee476ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Unified[[:space:]]Interpretation[[:space:]]of[[:space:]]Training-Time[[:space:]]Out-of-Distribution[[:space:]]Detection/c157585c-1fff-4b7e-bd9a-e651e62d1499_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]View-consistent[[:space:]]Sampling[[:space:]]Method[[:space:]]for[[:space:]]Regularized[[:space:]]Training[[:space:]]of[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/90cf7e87-1ed0-49c7-86fb-fbb63c8a8800_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A[[:space:]]Visual[[:space:]]Leap[[:space:]]in[[:space:]]CLIP[[:space:]]Compositionality[[:space:]]Reasoning[[:space:]]through[[:space:]]Generation[[:space:]]of[[:space:]]Counterfactual[[:space:]]Sets/3a990898-c16a-431f-9443-2af964a93a43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A0_[[:space:]]An[[:space:]]Affordance-Aware[[:space:]]Hierarchical[[:space:]]Model[[:space:]]for[[:space:]]General[[:space:]]Robotic[[:space:]]Manipulation/e1c76e8a-0356-415f-b666-9acb24cc149f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/A3GS_[[:space:]]Arbitrary[[:space:]]Artistic[[:space:]]Style[[:space:]]into[[:space:]]Arbitrary[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/f9cbeb2e-e6b6-4797-a906-0fc1aaf25add_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AAA-Gaussians_[[:space:]]Anti-Aliased[[:space:]]and[[:space:]]Artifact-Free[[:space:]]3D[[:space:]]Gaussian[[:space:]]Rendering/016cb452-3cf0-4124-be94-3a843aa45bac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ACAM-KD_[[:space:]]Adaptive[[:space:]]and[[:space:]]Cooperative[[:space:]]Attention[[:space:]]Masking[[:space:]]for[[:space:]]Knowledge[[:space:]]Distillation/c3c1af7e-cfca-49eb-8b92-794d6e4595ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ACE-G_[[:space:]]Improving[[:space:]]Generalization[[:space:]]of[[:space:]]Scene[[:space:]]Coordinate[[:space:]]Regression[[:space:]]Through[[:space:]]Query[[:space:]]Pre-Training/9695e05b-2ff2-4698-b5e9-559f39a8c855_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AD-GS_[[:space:]]Object-Aware[[:space:]]B-Spline[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Autonomous[[:space:]]Driving/7e0f40a8-6ee9-4e9a-a10d-d7c4af577a2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ADCD-Net_[[:space:]]Robust[[:space:]]Document[[:space:]]Image[[:space:]]Forgery[[:space:]]Localization[[:space:]]via[[:space:]]Adaptive[[:space:]]DCT[[:space:]]Feature[[:space:]]and[[:space:]]Hierarchical[[:space:]]Content[[:space:]]Disentanglement/dd36be76-543f-4693-b07c-6da6081010fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ADIEE_[[:space:]]Automatic[[:space:]]Dataset[[:space:]]Creation[[:space:]]and[[:space:]]Scorer[[:space:]]for[[:space:]]Instruction-Guided[[:space:]]Image[[:space:]]Editing[[:space:]]Evaluation/1573082a-4066-4bd5-b692-2a2b22ba9248_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AFUNet_[[:space:]]Cross-Iterative[[:space:]]Alignment-Fusion[[:space:]]Synergy[[:space:]]for[[:space:]]HDR[[:space:]]Reconstruction[[:space:]]via[[:space:]]Deep[[:space:]]Unfolding[[:space:]]Paradigm/71233aa6-2661-4120-8ae3-22626c005534_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AG2aussian_[[:space:]]Anchor-Graph[[:space:]]Structured[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Instance-Level[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding[[:space:]]and[[:space:]]Editing/12f086a3-3ec1-4fe1-8474-f4b0923c3ef6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AGO_[[:space:]]Adaptive[[:space:]]Grounding[[:space:]]for[[:space:]]Open[[:space:]]World[[:space:]]3D[[:space:]]Occupancy[[:space:]]Prediction/6ed96df1-e7ea-46a6-a9e4-b57d4261baaa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AHCPTQ_[[:space:]]Accurate[[:space:]]and[[:space:]]Hardware-Compatible[[:space:]]Post-Training[[:space:]]Quantization[[:space:]]for[[:space:]]Segment[[:space:]]Anything[[:space:]]Model/b50054a8-0f79-4f66-9d40-4be7a77748aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AIComposer_[[:space:]]Any[[:space:]]Style[[:space:]]and[[:space:]]Content[[:space:]]Image[[:space:]]Composition[[:space:]]via[[:space:]]Feature[[:space:]]Integration/e39e7ae3-0797-46bf-849a-a6b296484c94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AID_[[:space:]]Adapting[[:space:]]Image2Video[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Instruction-guided[[:space:]]Video[[:space:]]Prediction/5ef07eca-272f-442e-b3fa-cd39cf6bb0f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AIGI-Holmes_[[:space:]]Towards[[:space:]]Explainable[[:space:]]and[[:space:]]Generalizable[[:space:]]AI-Generated[[:space:]]Image[[:space:]]Detection[[:space:]]via[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/44cc3f4a-9c7b-4177-8014-c840cc8f3d0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AIM_[[:space:]]Adaptive[[:space:]]Inference[[:space:]]of[[:space:]]Multi-Modal[[:space:]]LLMs[[:space:]]via[[:space:]]Token[[:space:]]Merging[[:space:]]and[[:space:]]Pruning/e828101d-e0e7-439b-9b05-6a67b9b6fcc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AIM_[[:space:]]Amending[[:space:]]Inherent[[:space:]]Interpretability[[:space:]]via[[:space:]]Self-Supervised[[:space:]]Masking/5e1524ba-07b2-417c-9903-a15da3abf218_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AIRA_[[:space:]]Activation-Informed[[:space:]]Low-Rank[[:space:]]Adaptation[[:space:]]for[[:space:]]Large[[:space:]]Models/a367da2f-64d5-4c8c-aaff-7f75e9a9d22a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AJAHR_[[:space:]]Amputated[[:space:]]Joint[[:space:]]Aware[[:space:]]3D[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/95228a37-1454-4697-aafe-20ce614e0c02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ALOcc_[[:space:]]Adaptive[[:space:]]Lifting-Based[[:space:]]3D[[:space:]]Semantic[[:space:]]Occupancy[[:space:]]and[[:space:]]Cost[[:space:]]Volume-Based[[:space:]]Flow[[:space:]]Predictions/bebffc25-3e2b-43ac-9fa0-c86d44524165_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AM-Adapter_[[:space:]]Appearance[[:space:]]Matching[[:space:]]Adapter[[:space:]]for[[:space:]]Exemplar-based[[:space:]]Semantic[[:space:]]Image[[:space:]]Synthesis[[:space:]]in-the-Wild/4aa1205a-5a6c-4469-9471-38913c39e60e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AMDANet_[[:space:]]Attention-Driven[[:space:]]Multi-Perspective[[:space:]]Discrepancy[[:space:]]Alignment[[:space:]]for[[:space:]]RGB-Infrared[[:space:]]Image[[:space:]]Fusion[[:space:]]and[[:space:]]Segmentation/6651bfbe-0c52-49e2-a538-e5103159a324_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AMD_[[:space:]]Adaptive[[:space:]]Momentum[[:space:]]and[[:space:]]Decoupled[[:space:]]Contrastive[[:space:]]Learning[[:space:]]Framework[[:space:]]for[[:space:]]Robust[[:space:]]Long-Tail[[:space:]]Trajectory[[:space:]]Prediction/92918983-b9ca-4147-82c3-e92a6bf19f08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AR-1-to-3_[[:space:]]Single[[:space:]]Image[[:space:]]to[[:space:]]Consistent[[:space:]]3D[[:space:]]Object[[:space:]]via[[:space:]]Next-View[[:space:]]Prediction/deda316f-94fa-49a9-8d1e-7f791fbfc527_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AR-VRM_[[:space:]]Imitating[[:space:]]Human[[:space:]]Motions[[:space:]]for[[:space:]]Visual[[:space:]]Robot[[:space:]]Manipulation[[:space:]]with[[:space:]]Analogical[[:space:]]Reasoning/b95a7d6e-f79d-4355-a064-89e9825258c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ARGUS_[[:space:]]Hallucination[[:space:]]and[[:space:]]Omission[[:space:]]Evaluation[[:space:]]in[[:space:]]Video-LLMs/519b651b-5330-482b-81b7-967fcc24c370_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ARIG_[[:space:]]Autoregressive[[:space:]]Interactive[[:space:]]Head[[:space:]]Generation[[:space:]]for[[:space:]]Real-time[[:space:]]Conversations/50366525-3ff2-4b6c-bde6-e3edf94a525f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ARMO_[[:space:]]Autoregressive[[:space:]]Rigging[[:space:]]for[[:space:]]Multi-Category[[:space:]]Objects/998a011e-ba2a-4697-9933-e2c180b4d9f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ART_[[:space:]]Adaptive[[:space:]]Relation[[:space:]]Tuning[[:space:]]for[[:space:]]Generalized[[:space:]]Relation[[:space:]]Prediction/5091242a-7301-428b-ae2c-0fd97b4ca6a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ASCENT_[[:space:]]Annotation-free[[:space:]]Self-supervised[[:space:]]Contrastive[[:space:]]Embeddings[[:space:]]for[[:space:]]3D[[:space:]]Neuron[[:space:]]Tracking[[:space:]]in[[:space:]]Fluorescence[[:space:]]Microscopy/bf562c0e-220d-4cba-8698-d4e45d1fc9c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ASGS_[[:space:]]Single-Domain[[:space:]]Generalizable[[:space:]]Open-Set[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Adaptive[[:space:]]Subgraph[[:space:]]Searching/b967aebf-195a-4207-83ca-0064b797c9e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ATAS_[[:space:]]Any-to-Any[[:space:]]Self-Distillation[[:space:]]for[[:space:]]Enhanced[[:space:]]Open-Vocabulary[[:space:]]Dense[[:space:]]Prediction/f4790c17-66e5-4809-a9d4-8c8eed9334ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ATCTrack_[[:space:]]Aligning[[:space:]]Target-Context[[:space:]]Cues[[:space:]]with[[:space:]]Dynamic[[:space:]]Target[[:space:]]States[[:space:]]for[[:space:]]Robust[[:space:]]Vision-Language[[:space:]]Tracking/f2b9a1ec-0cb3-4cf4-978d-d9dce7fa404a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ATLAS_[[:space:]]Decoupling[[:space:]]Skeletal[[:space:]]and[[:space:]]Shape[[:space:]]Parameters[[:space:]]for[[:space:]]Expressive[[:space:]]Parametric[[:space:]]Human[[:space:]]Modeling/984acba8-651c-42fc-99ef-2e5fd9b70a5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AU-Blendshape[[:space:]]for[[:space:]]Fine-grained[[:space:]]Stylized[[:space:]]3D[[:space:]]Facial[[:space:]]Expression[[:space:]]Manipulation/dee184c7-41eb-4dd8-93a8-47a78f385cf4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AURELIA_[[:space:]]Test-time[[:space:]]Reasoning[[:space:]]Distillation[[:space:]]in[[:space:]]Audio-Visual[[:space:]]LLMs/00854c6e-06a0-4e2d-9e6d-9158f3b08f27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AV-Flow_[[:space:]]Transforming[[:space:]]Text[[:space:]]to[[:space:]]Audio-Visual[[:space:]]Human-like[[:space:]]Interactions/e354a88d-b5ec-4dcc-ae12-c56241f947c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AV-Link_[[:space:]]Temporally-Aligned[[:space:]]Diffusion[[:space:]]Features[[:space:]]for[[:space:]]Cross-Modal[[:space:]]Audio-Video[[:space:]]Generation/ade9c43e-e336-422a-8843-f6f5eae85850_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AVAM_[[:space:]]a[[:space:]]Universal[[:space:]]Training-free[[:space:]]Adaptive[[:space:]]Visual[[:space:]]Anchoring[[:space:]]Embedded[[:space:]]into[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Multi-image[[:space:]]Question[[:space:]]Answering/f1700933-8013-484f-ac78-44bf21117de1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AVTrustBench_[[:space:]]Assessing[[:space:]]and[[:space:]]Enhancing[[:space:]]Reliability[[:space:]]and[[:space:]]Robustness[[:space:]]in[[:space:]]Audio-Visual[[:space:]]LLMs/0f3978c0-6fee-44f0-b52d-521da36a0549_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AcZeroTS_[[:space:]]Active[[:space:]]Learning[[:space:]]for[[:space:]]Zero-shot[[:space:]]Tissue[[:space:]]Segmentation[[:space:]]in[[:space:]]Pathology[[:space:]]Images/bd073bea-9eb0-4e15-a348-99eec0536189_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Accelerate[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]Models[[:space:]]via[[:space:]]Zero-Shot[[:space:]]Attention[[:space:]]Key[[:space:]]Pruning/67f919e7-dc78-4b9a-b4fb-1881ece27bfd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Accelerating[[:space:]]Diffusion[[:space:]]Sampling[[:space:]]via[[:space:]]Exploiting[[:space:]]Local[[:space:]]Transition[[:space:]]Coherence/c1d35ad9-765a-4178-9bdb-b493649e94ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Accelerating[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]via[[:space:]]Gradient-Optimized[[:space:]]Cache/4e1e55ed-976e-4434-abb2-edb00b11e3e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AccidentalGS_[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Accidental[[:space:]]Camera[[:space:]]Motion/0e75f2f3-602e-45d4-861e-f9a90141ff3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Achieving[[:space:]]More[[:space:]]with[[:space:]]Less_[[:space:]]Additive[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Rehearsal-Free[[:space:]]Class-Incremental[[:space:]]Learning/286cdd29-cf9e-42b1-892e-790a7e4d9fbf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Acknowledging[[:space:]]Focus[[:space:]]Ambiguity[[:space:]]in[[:space:]]Visual[[:space:]]Questions/1567d663-4534-4102-9a87-8789686709c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Activation[[:space:]]Subspaces[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection/6e5672fc-dfa3-4abd-84c5-c9acedff4a4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Active[[:space:]]Learning[[:space:]]Meets[[:space:]]Foundation[[:space:]]Models_[[:space:]]Fast[[:space:]]Remote[[:space:]]Sensing[[:space:]]Data[[:space:]]Annotation[[:space:]]for[[:space:]]Object[[:space:]]Detection/669f9bfd-3a48-4061-b917-2ef1784b1f39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Active[[:space:]]Membership[[:space:]]Inference[[:space:]]Test[[:space:]](aMINT)_[[:space:]]Enhancing[[:space:]]Model[[:space:]]Auditability[[:space:]]with[[:space:]]Multi-Task[[:space:]]Learning./a245b5f9-4257-4181-88a6-6257771cb0e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Active[[:space:]]Perception[[:space:]]Meets[[:space:]]Rule-Guided[[:space:]]RL_[[:space:]]A[[:space:]]Two-Phase[[:space:]]Approach[[:space:]]for[[:space:]]Precise[[:space:]]Object[[:space:]]Navigation[[:space:]]in[[:space:]]Complex[[:space:]]Environments/d7663c0e-a29b-4363-a1e2-ea1fe3d6a373_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AdaDCP_[[:space:]]Learning[[:space:]]an[[:space:]]Adapter[[:space:]]with[[:space:]]Discrete[[:space:]]Cosine[[:space:]]Prior[[:space:]]for[[:space:]]Clear-to-Adverse[[:space:]]Domain[[:space:]]Generalization/b84cc67d-b915-4838-acc6-d437194ed84d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AdaDrive_[[:space:]]Self-Adaptive[[:space:]]Slow-Fast[[:space:]]System[[:space:]]for[[:space:]]Language-Grounded[[:space:]]Autonomous[[:space:]]Driving/2215d37c-6a5f-4b8a-95b7-fa7c8692af8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AdaHuman_[[:space:]]Animatable[[:space:]]Detailed[[:space:]]3D[[:space:]]Human[[:space:]]Generation[[:space:]]with[[:space:]]Compositional[[:space:]]Multiview[[:space:]]Diffusion/250bf69c-eea3-4a3d-919e-6dd25710b802_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adapt[[:space:]]Foundational[[:space:]]Segmentation[[:space:]]Models[[:space:]]with[[:space:]]Heterogeneous[[:space:]]Searching[[:space:]]Space/228e0ef5-8e58-45c9-9bce-f1f19fe4f4ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adapting[[:space:]]In-Domain[[:space:]]Few-Shot[[:space:]]Segmentation[[:space:]]to[[:space:]]New[[:space:]]Domains[[:space:]]without[[:space:]]Source[[:space:]]Domain[[:space:]]Retraining/1f509b59-c5a6-43d8-96e2-83f27bb2df69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adapting[[:space:]]Vehicle[[:space:]]Detectors[[:space:]]for[[:space:]]Aerial[[:space:]]Imagery[[:space:]]to[[:space:]]Unseen[[:space:]]Domains[[:space:]]with[[:space:]]Weak[[:space:]]Supervision/8c7d9202-bbbb-43b4-95a8-03f29294f574_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Articulated[[:space:]]Object[[:space:]]Manipulation[[:space:]]On[[:space:]]The[[:space:]]Fly[[:space:]]with[[:space:]]Foundation[[:space:]]Model[[:space:]]Reasoning[[:space:]]and[[:space:]]Part[[:space:]]Grounding/c6de4336-1123-46b0-b3d9-db39393a4712_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Caching[[:space:]]for[[:space:]]Faster[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion[[:space:]]Transformers/6ff754f4-9314-4139-9910-b0b3eedd5a86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Dual[[:space:]]Uncertainty[[:space:]]Optimization_[[:space:]]Boosting[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]under[[:space:]]Test-Time[[:space:]]Shifts/c498df0d-2910-4380-bea1-d1d32a2718e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Hyper-Graph[[:space:]]Convolution[[:space:]]Network[[:space:]]for[[:space:]]Skeleton-based[[:space:]]Human[[:space:]]Action[[:space:]]Recognition[[:space:]]with[[:space:]]Virtual[[:space:]]Connections/eb5cfddb-33b1-4050-a9eb-d937a7558bf3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Learning[[:space:]]of[[:space:]]High-Value[[:space:]]Regions[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/6962b9f6-7f28-4f4a-b794-c5ccf66928cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Prompt[[:space:]]Learning[[:space:]]via[[:space:]]Gaussian[[:space:]]Outlier[[:space:]]Synthesis[[:space:]]for[[:space:]]Out-of-distribution[[:space:]]Detection/64d4c9d4-3703-4d8c-8396-0f791cf0fe59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adaptive[[:space:]]Routing[[:space:]]of[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]Requests[[:space:]]Between[[:space:]]Large[[:space:]]Cloud[[:space:]]Model[[:space:]]and[[:space:]]Light-Weight[[:space:]]Edge[[:space:]]Model/da3df561-f44d-4027-8335-cb684e94422c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AdaptiveAE_[[:space:]]An[[:space:]]Adaptive[[:space:]]Exposure[[:space:]]Strategy[[:space:]]for[[:space:]]HDR[[:space:]]Capturing[[:space:]]in[[:space:]]Dynamic[[:space:]]Scenes/facb16c1-8dd7-4d43-a13f-75a64353a2a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adding[[:space:]]Additional[[:space:]]Control[[:space:]]to[[:space:]]One-Step[[:space:]]Diffusion[[:space:]]with[[:space:]]Joint[[:space:]]Distribution[[:space:]]Matching/c2b3835f-f805-4917-82a0-466f101155c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Addressing[[:space:]]Representation[[:space:]]Collapse[[:space:]]in[[:space:]]Vector[[:space:]]Quantized[[:space:]]Models[[:space:]]with[[:space:]]One[[:space:]]Linear[[:space:]]Layer/c9ee5a18-482c-48be-ab03-3fe420197718_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Addressing[[:space:]]Text[[:space:]]Embedding[[:space:]]Leakage[[:space:]]in[[:space:]]Diffusion-based[[:space:]]Image[[:space:]]Editing/7ef26365-88b2-4b63-b82e-fc5e555360c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AdsQA_[[:space:]]Towards[[:space:]]Advertisement[[:space:]]Video[[:space:]]Understanding/6375f078-3f0b-4521-b432-64d23e64bff4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AdvDreamer[[:space:]]Unveils_[[:space:]]Are[[:space:]]Vision-Language[[:space:]]Models[[:space:]]Truly[[:space:]]Ready[[:space:]]for[[:space:]]Real-World[[:space:]]3D[[:space:]]Variations_/74239f2c-88ab-4f0c-965d-c8f65d948118_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Advancing[[:space:]]Text-to-3D[[:space:]]Generation[[:space:]]with[[:space:]]Linearized[[:space:]]Lookahead[[:space:]]Variational[[:space:]]Score[[:space:]]Distillation/93194ba5-cad9-47e5-bf50-5b562a6b61ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Advancing[[:space:]]Textual[[:space:]]Prompt[[:space:]]Learning[[:space:]]with[[:space:]]Anchored[[:space:]]Attributes/4e7100e1-8488-4372-a2bb-73e12989371e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Advancing[[:space:]]Visual[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Multi-granular[[:space:]]Versatile[[:space:]]Perception/2e0e9703-37e8-4376-9e70-3d183c7f0c36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adversarial[[:space:]]Attention[[:space:]]Perturbations[[:space:]]for[[:space:]]Large[[:space:]]Object[[:space:]]Detection[[:space:]]Transformers/88192fda-87cb-4027-ba07-95fbfb69b360_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adversarial[[:space:]]Data[[:space:]]Augmentation[[:space:]]for[[:space:]]Single[[:space:]]Domain[[:space:]]Generalization[[:space:]]via[[:space:]]Lyapunov[[:space:]]Exponent-Guided[[:space:]]Optimization/f0ac30f5-bb31-44e2-94e3-b0f8627b8df9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adversarial[[:space:]]Distribution[[:space:]]Matching[[:space:]]for[[:space:]]Diffusion[[:space:]]Distillation[[:space:]]Towards[[:space:]]Efficient[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Synthesis/b23b5268-40ae-4571-a9ee-075cda737453_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adversarial[[:space:]]Exploitation[[:space:]]of[[:space:]]Data[[:space:]]Diversity[[:space:]]Improves[[:space:]]Visual[[:space:]]Localization/b8dc72b7-af77-4d02-8bf0-d21936019f63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adversarial[[:space:]]Purification[[:space:]]via[[:space:]]Super-Resolution[[:space:]]and[[:space:]]Diffusion/b9eb49c2-a69c-48c0-bba6-c208edb9f012_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adversarial[[:space:]]Reconstruction[[:space:]]Feedback[[:space:]]for[[:space:]]Robust[[:space:]]Fine-grained[[:space:]]Generalization/2762e731-fb17-490b-a90d-7e63ff84cb06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adversarial[[:space:]]Robust[[:space:]]Memory-Based[[:space:]]Continual[[:space:]]Learner/8c1dfab1-b341-43f5-bde7-c2f8e9dc2261_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adversarial[[:space:]]Robustness[[:space:]]of[[:space:]]Discriminative[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]in[[:space:]]Vision/793fb0cc-d29e-4453-8e4e-2aca5e095268_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Adversarial[[:space:]]Training[[:space:]]for[[:space:]]Probabilistic[[:space:]]Robustness/47fc789f-39f4-483e-ba24-6d818ddae7f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AerialVG_[[:space:]]A[[:space:]]Challenging[[:space:]]Benchmark[[:space:]]for[[:space:]]Aerial[[:space:]]Visual[[:space:]]Grounding[[:space:]]by[[:space:]]Exploring[[:space:]]Positional[[:space:]]Relations/b41713a2-2be5-479f-9826-9a9a4762ace5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Aether_[[:space:]]Geometric-Aware[[:space:]]Unified[[:space:]]World[[:space:]]Modeling/d408fc86-f2f5-4340-bdc7-9915c8ed0e52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AffordDexGrasp_[[:space:]]Open-set[[:space:]]Language-guided[[:space:]]Dexterous[[:space:]]Grasp[[:space:]]with[[:space:]]Generalizable-Instructive[[:space:]]Affordance/428be4db-41fb-452f-ae6c-be43892c88ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/After[[:space:]]the[[:space:]]Party_[[:space:]]Navigating[[:space:]]the[[:space:]]Mapping[[:space:]]From[[:space:]]Color[[:space:]]to[[:space:]]Ambient[[:space:]]Lighting/63a161d7-3eb4-437d-8d2d-2312e022984e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Agreement[[:space:]]aware[[:space:]]and[[:space:]]dissimilarity[[:space:]]oriented[[:space:]]GLOM/bd040c1e-9041-429f-8136-d5caa2846f28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AgroBench_[[:space:]]Vision-Language[[:space:]]Model[[:space:]]Benchmark[[:space:]]in[[:space:]]Agriculture/b17f54a3-0198-426c-98d3-db524df55ee2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AirCache_[[:space:]]Activating[[:space:]]Inter-modal[[:space:]]Relevancy[[:space:]]KV[[:space:]]Cache[[:space:]]Compression[[:space:]]for[[:space:]]Efficient[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Model[[:space:]]Inference/7536831b-10b6-4fe8-b035-0c2896718fa3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Align[[:space:]]Your[[:space:]]Rhythm_[[:space:]]Generating[[:space:]]Highly[[:space:]]Aligned[[:space:]]Dance[[:space:]]Poses[[:space:]]with[[:space:]]Gating-Enhanced[[:space:]]Rhythm-Aware[[:space:]]Feature[[:space:]]Representation/bae0438d-2c16-45f8-9ffb-98e7eef3786c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AlignDiff_[[:space:]]Learning[[:space:]]Physically-Grounded[[:space:]]Camera[[:space:]]Alignment[[:space:]]via[[:space:]]Diffusion/6380161a-ecce-4164-bb4a-61e264b72668_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AlignGuard_[[:space:]]Scalable[[:space:]]Safety[[:space:]]Alignment[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/821c3c22-a165-4d20-a058-8dccd9e55058_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Aligning[[:space:]]Constraint[[:space:]]Generation[[:space:]]with[[:space:]]Design[[:space:]]Intent[[:space:]]in[[:space:]]Parametric[[:space:]]CAD/cc603c3c-bad9-4030-b99c-93bffb00f1b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Aligning[[:space:]]Effective[[:space:]]Tokens[[:space:]]with[[:space:]]Video[[:space:]]Anomaly[[:space:]]in[[:space:]]Large[[:space:]]Language[[:space:]]Models/8e282f6a-9c9d-4443-a523-c2f4bcdbbda6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Aligning[[:space:]]Global[[:space:]]Semantics[[:space:]]and[[:space:]]Local[[:space:]]Textures[[:space:]]in[[:space:]]Generative[[:space:]]Video[[:space:]]Enhancement/48366b30-d2f0-4312-be7c-cf9d18325fdd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Aligning[[:space:]]Information[[:space:]]Capacity[[:space:]]Between[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]via[[:space:]]Dense-to-Sparse[[:space:]]Feature[[:space:]]Distillation[[:space:]]for[[:space:]]Image-Text[[:space:]]Matching/eee67cc2-5832-464d-b273-17cfe308efbf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Aligning[[:space:]]Moments[[:space:]]in[[:space:]]Time[[:space:]]using[[:space:]]Video[[:space:]]Queries/f1f18e54-b74a-4e85-8e01-d0d39e528c08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Aligning[[:space:]]Vision[[:space:]]to[[:space:]]Language_[[:space:]]Annotation-Free[[:space:]]Multimodal[[:space:]]Knowledge[[:space:]]Graph[[:space:]]Construction[[:space:]]for[[:space:]]Enhanced[[:space:]]LLMs[[:space:]]Reasoning/7ddf6f08-8bf9-46ae-a11c-83b7c2afd49c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/All[[:space:]]Parts[[:space:]]Matter_[[:space:]]A[[:space:]]Unified[[:space:]]Mask-Free[[:space:]]Virtual[[:space:]]Try-On[[:space:]]Framework/4207f645-0fb0-4a1e-b60a-46c12825cd89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/All[[:space:]]in[[:space:]]One_[[:space:]]Visual-Description-Guided[[:space:]]Unified[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation/9f45fe2a-eeec-496b-b486-eee2ccb9cf69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AllGCD_[[:space:]]Leveraging[[:space:]]All[[:space:]]Unlabeled[[:space:]]Data[[:space:]]for[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/87a08d4a-611f-4092-84c0-36e8cdf6fd1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AllTracker_[[:space:]]Efficient[[:space:]]Dense[[:space:]]Point[[:space:]]Tracking[[:space:]]at[[:space:]]High[[:space:]]Resolution/da1d8f6b-2320-4654-bad6-ea377ecc5c87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Alleviating[[:space:]]Textual[[:space:]]Reliance[[:space:]]in[[:space:]]Medical[[:space:]]Language-guided[[:space:]]Segmentation[[:space:]]via[[:space:]]Prototype-driven[[:space:]]Semantic[[:space:]]Approximation/bb0e3137-699a-4301-8697-612d0c48d0ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Allowing[[:space:]]Oscillation[[:space:]]Quantization_[[:space:]]Overcoming[[:space:]]Solution[[:space:]]Space[[:space:]]Limitation[[:space:]]in[[:space:]]Low[[:space:]]Bit-Width[[:space:]]Quantization/5d6dba58-3126-413c-a7af-b25ffe7d92f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Always[[:space:]]Skip[[:space:]]Attention/569226ba-e747-4584-b317-a54b6e4497d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Amodal[[:space:]]Depth[[:space:]]Anything_[[:space:]]Amodal[[:space:]]Depth[[:space:]]Estimation[[:space:]]in[[:space:]]the[[:space:]]Wild/dc30b08f-8ee9-4bca-9c02-871cd4377982_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Amodal3R_[[:space:]]Amodal[[:space:]]3D[[:space:]]Reconstruction[[:space:]]from[[:space:]]Occluded[[:space:]]2D[[:space:]]Images/88b04b9b-d30d-4c9a-91cb-8a21cd28defd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/An[[:space:]]Efficient[[:space:]]Hybrid[[:space:]]Vision[[:space:]]Transformer[[:space:]]for[[:space:]]TinyML[[:space:]]Applications/7d996d12-432e-4c35-99e7-c13b06e6d6ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/An[[:space:]]Efficient[[:space:]]Post-hoc[[:space:]]Framework[[:space:]]for[[:space:]]Reducing[[:space:]]Task[[:space:]]Discrepancy[[:space:]]of[[:space:]]Text[[:space:]]Encoders[[:space:]]for[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/a5a58fbf-87d0-4c6b-8e78-509d0d6c7143_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/An[[:space:]]Empirical[[:space:]]Study[[:space:]]of[[:space:]]Autoregressive[[:space:]]Pre-training[[:space:]]from[[:space:]]Videos/0fb150f7-5e61-492a-a2d2-7d4a735f9e3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/An[[:space:]]Information-Theoretic[[:space:]]Regularizer[[:space:]]for[[:space:]]Lossy[[:space:]]Neural[[:space:]]Image[[:space:]]Compression/cb73e4dd-eb4f-4fff-b4a2-c49c8fa70b5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/An[[:space:]]Inversion-based[[:space:]]Measure[[:space:]]of[[:space:]]Memorization[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/4b89bc09-28c7-40c0-9f29-6da101712bc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/An[[:space:]]OpenMind[[:space:]]for[[:space:]]3D[[:space:]]Medical[[:space:]]Vision[[:space:]]Self-supervised[[:space:]]Learning/237d5190-142b-4bfa-94c8-a21dfa727bcc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Analyzing[[:space:]]Finetuning[[:space:]]Representation[[:space:]]Shift[[:space:]]for[[:space:]]Multimodal[[:space:]]LLMs[[:space:]]Steering/3a9fad37-0f80-4ba9-981c-b874e41fd8d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Anchor[[:space:]]Token[[:space:]]Matching_[[:space:]]Implicit[[:space:]]Structure[[:space:]]Locking[[:space:]]for[[:space:]]Training-free[[:space:]]AR[[:space:]]Image[[:space:]]Editing/02ecc675-dff6-4c74-84e6-883c770f01a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnimalClue_[[:space:]]Recognizing[[:space:]]Animals[[:space:]]by[[:space:]]their[[:space:]]Traces/8e6d9b6c-ad00-4a08-84f7-166aef83cedd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Animate[[:space:]]Anyone[[:space:]]2_[[:space:]]High-Fidelity[[:space:]]Character[[:space:]]Image[[:space:]]Animation[[:space:]]with[[:space:]]Environment[[:space:]]Affordance/f25b16a0-97a2-4bcc-93f4-3c50d28debfd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnimateAnyMesh_[[:space:]]A[[:space:]]Feed-Forward[[:space:]]4D[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Text-Driven[[:space:]]Universal[[:space:]]Mesh[[:space:]]Animation/fc436c83-46ad-42d4-ab61-369656d1b892_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnimeGamer_[[:space:]]Infinite[[:space:]]Anime[[:space:]]Life[[:space:]]Simulation[[:space:]]with[[:space:]]Next[[:space:]]Game[[:space:]]State[[:space:]]Prediction/842749ed-2ab1-47b2-b183-403e0f55e3a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnnofreeOD_[[:space:]]Detecting[[:space:]]All[[:space:]]Classes[[:space:]]at[[:space:]]Low[[:space:]]Frame[[:space:]]Rates[[:space:]]Without[[:space:]]Human[[:space:]]Annotations/c469fc9e-69e6-4a2c-a3c4-94d82008c066_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Anomaly[[:space:]]Detection[[:space:]]of[[:space:]]Integrated[[:space:]]Circuits[[:space:]]Package[[:space:]]Substrates[[:space:]]Using[[:space:]]the[[:space:]]Large[[:space:]]Vision[[:space:]]Model[[:space:]]SAIC_[[:space:]]Dataset[[:space:]]Construction,[[:space:]]Methodology,[[:space:]]and[[:space:]]Application/d944de6f-b792-4861-9144-b4eef3b9b54d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Anti-Tamper[[:space:]]Protection[[:space:]]for[[:space:]]Unauthorized[[:space:]]Individual[[:space:]]Image[[:space:]]Generation/398b8e09-051b-45a4-b77b-d72f2adef0da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Any-SSR_[[:space:]]How[[:space:]]Recursive[[:space:]]Least[[:space:]]Squares[[:space:]]Works[[:space:]]in[[:space:]]Continual[[:space:]]Learning[[:space:]]of[[:space:]]Large[[:space:]]Language[[:space:]]Model/1c7ec03f-7650-4fc9-a2d3-9fcf35f3ddd3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Any2AnyTryon_[[:space:]]Leveraging[[:space:]]Adaptive[[:space:]]Position[[:space:]]Embeddings[[:space:]]for[[:space:]]Versatile[[:space:]]Virtual[[:space:]]Clothing[[:space:]]Tasks/ecd48321-8d60-4ab2-99a1-236e6399193e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnyBimanual_[[:space:]]Transferring[[:space:]]Unimanual[[:space:]]Policy[[:space:]]for[[:space:]]General[[:space:]]Bimanual[[:space:]]Manipulation/3881a5d6-b94c-42ae-bae9-1c8277df4bb2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnyCalib_[[:space:]]On-Manifold[[:space:]]Learning[[:space:]]for[[:space:]]Model-Agnostic[[:space:]]Single-View[[:space:]]Camera[[:space:]]Calibration/87cd3e23-a7af-4e95-87fa-7b2a48188ead_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnyI2V_[[:space:]]Animating[[:space:]]Any[[:space:]]Conditional[[:space:]]Image[[:space:]]with[[:space:]]Motion[[:space:]]Control/0d49c2bf-7b9d-4b66-b777-b638e8d29805_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AnyPortal_[[:space:]]Zero-Shot[[:space:]]Consistent[[:space:]]Video[[:space:]]Background[[:space:]]Replacement/273c7a82-033b-43c4-84d5-1f1008a90ca8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ArchiSet_[[:space:]]Benchmarking[[:space:]]Editable[[:space:]]and[[:space:]]Consistent[[:space:]]Single-View[[:space:]]3D[[:space:]]Reconstruction[[:space:]]of[[:space:]]Buildings[[:space:]]with[[:space:]]Specific[[:space:]]Window-to-Wall[[:space:]]Ratios/af7906ec-74c6-445d-972a-4120b395bf0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Are[[:space:]]They[[:space:]]the[[:space:]]Same_[[:space:]]Exploring[[:space:]]Visual[[:space:]]Correspondence[[:space:]]Shortcomings[[:space:]]of[[:space:]]Multimodal[[:space:]]LLMs/8cd09e3b-d25a-46f0-bc94-1d7e8c9618d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Are[[:space:]]VLMs[[:space:]]Ready[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving_[[:space:]]An[[:space:]]Empirical[[:space:]]Study[[:space:]]from[[:space:]]the[[:space:]]Reliability,[[:space:]]Data[[:space:]]and[[:space:]]Metric[[:space:]]Perspectives/387390b4-8a11-4710-8610-0b21730d9be6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ArgMatch_[[:space:]]Adaptive[[:space:]]Refinement[[:space:]]Gathering[[:space:]]for[[:space:]]Efficient[[:space:]]Dense[[:space:]]Matching/c8ba2ca7-20a5-4ff3-af4a-fb99a851f959_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ArgoTweak_[[:space:]]Towards[[:space:]]Self-Updating[[:space:]]HD[[:space:]]Maps[[:space:]]through[[:space:]]Structured[[:space:]]Priors/f5b94b59-8323-4836-a58d-9de52641e82e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ArtEditor_[[:space:]]Learning[[:space:]]Customized[[:space:]]Instructional[[:space:]]Image[[:space:]]Editor[[:space:]]from[[:space:]]Few-Shot[[:space:]]Examples/bef8242d-602d-4afe-b60e-5aefbead8a4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Arti-PG_[[:space:]]A[[:space:]]Toolbox[[:space:]]for[[:space:]]Procedurally[[:space:]]Synthesizing[[:space:]]Large-Scale[[:space:]]and[[:space:]]Diverse[[:space:]]Articulated[[:space:]]Objects[[:space:]]with[[:space:]]Rich[[:space:]]Annotations/5d16a0ed-b15e-4c1a-87aa-6b43c511b7ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Articulate3D_[[:space:]]Holistic[[:space:]]Understanding[[:space:]]of[[:space:]]3D[[:space:]]Scenes[[:space:]]as[[:space:]]Universal[[:space:]]Scene[[:space:]]Description/55837be9-6812-41d5-8e25-2e1815aaadfb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ask[[:space:]]and[[:space:]]Remember_[[:space:]]A[[:space:]]Questions-Only[[:space:]]Replay[[:space:]]Strategy[[:space:]]for[[:space:]]Continual[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/ef2ef1f8-6e35-4925-9da6-6edf0d26a701_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AstroLoc_[[:space:]]Robust[[:space:]]Space[[:space:]]to[[:space:]]Ground[[:space:]]Image[[:space:]]Localizer/a47f9788-ff10-40d0-8bfd-beccbbd4cae6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Asynchronous[[:space:]]Event[[:space:]]Error-Minimizing[[:space:]]Noise[[:space:]]for[[:space:]]Safeguarding[[:space:]]Event[[:space:]]Dataset/8473b7f6-ba54-4a6d-95ec-ad238db262f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Att-Adapter_[[:space:]]A[[:space:]]Robust[[:space:]]and[[:space:]]Precise[[:space:]]Domain-Specific[[:space:]]Multi-Attributes[[:space:]]T2I[[:space:]]Diffusion[[:space:]]Adapter[[:space:]]via[[:space:]]Conditional[[:space:]]Variational[[:space:]]Autoencoder/d560efa6-cc5a-478d-bbbc-d5e0776d5256_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Attention[[:space:]]to[[:space:]]Neural[[:space:]]Plagiarism_[[:space:]]Diffusion[[:space:]]Models[[:space:]]Can[[:space:]]Plagiarize[[:space:]]Your[[:space:]]Copyrighted[[:space:]]Images!/22bc5a08-5ff1-4a44-918a-87b1cd9b408d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Attention[[:space:]]to[[:space:]]Trajectory_[[:space:]]Trajectory-Aware[[:space:]]Open-Vocabulary[[:space:]]Tracking/110d82e5-d9ca-43a3-9a49-9a8e6391cbb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Attention[[:space:]]to[[:space:]]the[[:space:]]Burstiness[[:space:]]in[[:space:]]Visual[[:space:]]Prompt[[:space:]]Tuning!/6a717456-18b8-45a7-b437-39276a4f7001_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Audio-visual[[:space:]]Controlled[[:space:]]Video[[:space:]]Diffusion[[:space:]]with[[:space:]]Masked[[:space:]]Selective[[:space:]]State[[:space:]]Spaces[[:space:]]Modeling[[:space:]]for[[:space:]]Natural[[:space:]]Talking[[:space:]]Head[[:space:]]Generation/54257961-accc-40f4-a709-c43dc50d56a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Augmented[[:space:]]Mass-Spring[[:space:]]Model[[:space:]]for[[:space:]]Real-Time[[:space:]]Dense[[:space:]]Hair[[:space:]]Simulation/5f1680f0-d13e-41b9-8b57-43302084c352_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Augmented[[:space:]]and[[:space:]]Softened[[:space:]]Matching[[:space:]]for[[:space:]]Unsupervised[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification/658d0176-cc81-4231-9e7d-b19a170aaf1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Augmenting[[:space:]]Moment[[:space:]]Retrieval_[[:space:]]Zero-Dependency[[:space:]]Two-Stage[[:space:]]Learning/7ad477b0-629a-46d9-9244-38df2c532148_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Authentic[[:space:]]4D[[:space:]]Driving[[:space:]]Simulation[[:space:]]with[[:space:]]a[[:space:]]Video[[:space:]]Generation[[:space:]]Model/c4b0648d-9e94-4bfb-a24d-d6410577c4ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Auto-Controlled[[:space:]]Image[[:space:]]Perception[[:space:]]in[[:space:]]MLLMs[[:space:]]via[[:space:]]Visual[[:space:]]Perception[[:space:]]Tokens/c8eebe37-d780-4794-aae7-a19bf015e7a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Auto-Regressive[[:space:]]Transformation[[:space:]]for[[:space:]]Image[[:space:]]Alignment/c52aa418-0717-4374-b7e5-dac8aa2703bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Auto-Regressively[[:space:]]Generating[[:space:]]Multi-View[[:space:]]Consistent[[:space:]]Images/2159a4a8-7763-447f-9a63-fcccb1bd5b03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Auto-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/0f3512f6-9ec8-4ccb-95ba-f15696929556_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AutoComPose_[[:space:]]Automatic[[:space:]]Generation[[:space:]]of[[:space:]]Pose[[:space:]]Transition[[:space:]]Descriptions[[:space:]]for[[:space:]]Composed[[:space:]]Pose[[:space:]]Retrieval[[:space:]]Using[[:space:]]Multimodal[[:space:]]LLMs/0423c071-5af1-41f4-b79e-ba1e9dc7e5b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AutoOcc_[[:space:]]Automatic[[:space:]]Open-Ended[[:space:]]Semantic[[:space:]]Occupancy[[:space:]]Annotation[[:space:]]via[[:space:]]Vision-Language[[:space:]]Guided[[:space:]]Gaussian[[:space:]]Splatting/a5678c37-76e8-4e84-8893-77f73b3dbec5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AutoPrompt_[[:space:]]Automated[[:space:]]Red-Teaming[[:space:]]of[[:space:]]Text-to-Image[[:space:]]Models[[:space:]]via[[:space:]]LLM-Driven[[:space:]]Adversarial[[:space:]]Prompts/539b6f85-1215-4b31-bfdc-343456e91945_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/AutoScape_[[:space:]]Geometry-Consistent[[:space:]]Long-Horizon[[:space:]]Scene[[:space:]]Generation/f7761857-21a9-472d-adb8-6655dee4aa7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Automated[[:space:]]Model[[:space:]]Evaluation[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Prediction[[:space:]]Consistency[[:space:]]and[[:space:]]Reliability/56d0dda7-54b5-4b82-a538-8d51fd97536f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Automated[[:space:]]Red[[:space:]]Teaming[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Models[[:space:]]through[[:space:]]Feedback-Guided[[:space:]]Prompt[[:space:]]Iteration[[:space:]]with[[:space:]]Vision-Language[[:space:]]Models/26f077de-0c98-42e3-ba61-422e21feab7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Autoregressive[[:space:]]Denoising[[:space:]]Score[[:space:]]Matching[[:space:]]is[[:space:]]a[[:space:]]Good[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detector/be8da9d9-9d6a-40f1-ae54-1c09b8a93625_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Auxiliary[[:space:]]Prompt[[:space:]]Tuning[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models[[:space:]]for[[:space:]]Few-Shot[[:space:]]Out-of-Distribution[[:space:]]Detection/b2ae0546-1062-4dd0-915d-a29cdf0062f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Avat3r_[[:space:]]Large[[:space:]]Animatable[[:space:]]Gaussian[[:space:]]Reconstruction[[:space:]]Model[[:space:]]for[[:space:]]High-fidelity[[:space:]]3D[[:space:]]Head[[:space:]]Avatars/33025928-7bdf-4dc5-9ef0-20680ad8d9af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Axis-level[[:space:]]Symmetry[[:space:]]Detection[[:space:]]with[[:space:]]Group-Equivariant[[:space:]]Representation/9b1a165b-628f-43b6-888e-a507ddbef58a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/B-VLLM_[[:space:]]A[[:space:]]Vision[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]with[[:space:]]Balanced[[:space:]]Spatio-Temporal[[:space:]]Tokens/9bae0e8c-0744-43b8-b361-85b9e28599ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BANet_[[:space:]]Bilateral[[:space:]]Aggregation[[:space:]]Network[[:space:]]for[[:space:]]Mobile[[:space:]]Stereo[[:space:]]Matching/371a6e48-ef76-4ef5-94cb-44b977fc3aad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BASIC_[[:space:]]Boosting[[:space:]]Visual[[:space:]]Alignment[[:space:]]with[[:space:]]Intrinsic[[:space:]]Refined[[:space:]]Embeddings[[:space:]]in[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/cfe6018b-d012-4da5-a5b1-e85bd129ac40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BATCLIP_[[:space:]]Bimodal[[:space:]]Online[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]CLIP/095a5cfa-8193-42e5-9266-b959db1003a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BUFFER-X_[[:space:]]Towards[[:space:]]Zero-Shot[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration[[:space:]]in[[:space:]]Diverse[[:space:]]Scenes/3b868079-3684-4631-8177-346dc4932267_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BVINet_[[:space:]]Unlocking[[:space:]]Blind[[:space:]]Video[[:space:]]Inpainting[[:space:]]with[[:space:]]Zero[[:space:]]Annotations/02f927a1-1e70-4d16-a028-583b267f7d3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BabyVLM_[[:space:]]Data-Efficient[[:space:]]Pretraining[[:space:]]of[[:space:]]VLMs[[:space:]]Inspired[[:space:]]by[[:space:]]Infant[[:space:]]Learning/5d5c6c9c-eb3f-45fe-a3c0-70d0cb466306_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Back[[:space:]]on[[:space:]]Track_[[:space:]]Bundle[[:space:]]Adjustment[[:space:]]for[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Reconstruction/28032b83-7308-4f3f-9092-4ee41d294543_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Backdoor[[:space:]]Attacks[[:space:]]on[[:space:]]Neural[[:space:]]Networks[[:space:]]via[[:space:]]One-Bit[[:space:]]Flip/75348b6f-f141-4f8d-9983-a5b7097fd854_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Backdoor[[:space:]]Defense[[:space:]]via[[:space:]]Enhanced[[:space:]]Splitting[[:space:]]and[[:space:]]Trap[[:space:]]Isolation/b37324f9-0336-4fed-8641-591eef1a9c71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Backdoor[[:space:]]Mitigation[[:space:]]by[[:space:]]Distance-Driven[[:space:]]Detoxification/9ad9a972-e325-472f-94e4-a4c50530cdbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Backdooring[[:space:]]Self-Supervised[[:space:]]Contrastive[[:space:]]Learning[[:space:]]by[[:space:]]Noisy[[:space:]]Alignment/84a6a964-9479-46d4-ad6f-18993f049a42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Background[[:space:]]Invariance[[:space:]]Testing[[:space:]]According[[:space:]]to[[:space:]]Semantic[[:space:]]Proximity/ef48d2ec-737e-431b-a8d4-503616aab70c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BadVideo_[[:space:]]Stealthy[[:space:]]Backdoor[[:space:]]Attack[[:space:]]against[[:space:]]Text-to-Video[[:space:]]Generation/e350efe5-a457-47e9-b94c-0ce2ccd338ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Baking[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]into[[:space:]]Diffusion[[:space:]]Denoiser[[:space:]]for[[:space:]]Fast[[:space:]]and[[:space:]]Scalable[[:space:]]Single-stage[[:space:]]Image-to-3D[[:space:]]Generation[[:space:]]and[[:space:]]Reconstruction/0cf7f384-ec2a-431b-90f9-85970097b184_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Balanced[[:space:]]Image[[:space:]]Stylization[[:space:]]with[[:space:]]Style[[:space:]]Matching[[:space:]]Score/fc064b02-d92c-41ce-b8b8-b5c07ed15353_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Balanced[[:space:]]Sharpness-Aware[[:space:]]Minimization[[:space:]]for[[:space:]]Imbalanced[[:space:]]Regression/923c7e5f-a401-4f82-8f86-c8538bb5ca66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Balancing[[:space:]]Conservatism[[:space:]]and[[:space:]]Aggressiveness_[[:space:]]Prototype-Affinity[[:space:]]Hybrid[[:space:]]Network[[:space:]]for[[:space:]]Few-Shot[[:space:]]Segmentation/e423aeab-8d92-47ed-9c83-ea263c64ad95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Balancing[[:space:]]Task-invariant[[:space:]]Interaction[[:space:]]and[[:space:]]Task-specific[[:space:]]Adaptation[[:space:]]for[[:space:]]Unified[[:space:]]Image[[:space:]]Fusion/2e2201ad-55a4-4d8e-a96c-22f6059c8537_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bayesian-Inspired[[:space:]]Space-Time[[:space:]]Superpixels/2645687e-d38c-4a42-a4e8-e92c5dff48a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Benchmarking[[:space:]]Burst[[:space:]]Super-Resolution[[:space:]]for[[:space:]]Polarization[[:space:]]Images_[[:space:]]Noise[[:space:]]Dataset[[:space:]]and[[:space:]]Analysis/a425eb2c-09d4-4bf4-ab49-5fa35609fd84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Benchmarking[[:space:]]Egocentric[[:space:]]Visual-Inertial[[:space:]]SLAM[[:space:]]at[[:space:]]City[[:space:]]Scale/b0455543-9d8a-460a-9940-7cebc542ba33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Benchmarking[[:space:]]Multimodal[[:space:]]CoT[[:space:]]Reward[[:space:]]Model[[:space:]]Stepwise[[:space:]]by[[:space:]]Visual[[:space:]]Program/a70864d5-908a-45dd-ac06-83ea3ed64f62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Benchmarking[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]Against[[:space:]]Image[[:space:]]Corruptions/da99319b-1616-4f18-8d80-4298c6023dfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Benchmarking[[:space:]]and[[:space:]]Learning[[:space:]]Multi-Dimensional[[:space:]]Quality[[:space:]]Evaluator[[:space:]]for[[:space:]]Text-to-3D[[:space:]]Generation/98cba355-b64d-49bb-b2ef-39478a58d384_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Benefit[[:space:]]From[[:space:]]Seen_[[:space:]]Enhancing[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detection[[:space:]]by[[:space:]]Bridging[[:space:]]Visual[[:space:]]and[[:space:]]Textual[[:space:]]Co-Occurrence[[:space:]]Knowledge/e92edf69-f68d-48ee-bcc7-ef2288a56ca4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Blur_[[:space:]]A[[:space:]]Fluid[[:space:]]Perspective[[:space:]]on[[:space:]]Generative[[:space:]]Diffusion[[:space:]]Models/e104c1cd-a4ae-42d2-9c80-f4cfe0e68003_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Brain[[:space:]]Decoding_[[:space:]]Visual-Semantic[[:space:]]Reconstructions[[:space:]]to[[:space:]]Mental[[:space:]]Creation[[:space:]]Extension[[:space:]]Based[[:space:]]on[[:space:]]fMRI/67a27b01-beb4-4ee2-819d-c835b7e2eeeb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Isolated[[:space:]]Words_[[:space:]]Diffusion[[:space:]]Brush[[:space:]]for[[:space:]]Handwritten[[:space:]]Text-Line[[:space:]]Generation/dc5cf565-ad0b-4e42-aa11-573b6a8a2d1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Label[[:space:]]Semantics_[[:space:]]Language-Guided[[:space:]]Action[[:space:]]Anatomy[[:space:]]for[[:space:]]Few-shot[[:space:]]Action[[:space:]]Recognition/d02debe0-bba1-465d-a9e2-a8fcb6280be4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Losses[[:space:]]Reweighting_[[:space:]]Empowering[[:space:]]Multi-Task[[:space:]]Learning[[:space:]]via[[:space:]]the[[:space:]]Generalization[[:space:]]Perspective/75eedb7a-2ee5-45f4-9650-19a93dc79275_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Low-Rank[[:space:]]Tuning_[[:space:]]Model[[:space:]]Prior-Guided[[:space:]]Rank[[:space:]]Allocation[[:space:]]for[[:space:]]Effective[[:space:]]Transfer[[:space:]]in[[:space:]]Low-Data[[:space:]]and[[:space:]]Large-Gap[[:space:]]Regimes./f1d8d5a2-38b7-4e49-be1c-8475a18aa57c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Next-Token_[[:space:]]Next-X[[:space:]]Prediction[[:space:]]for[[:space:]]Autoregressive[[:space:]]Visual[[:space:]]Generation/16c00a59-671e-41f2-9ca2-5d913d66791b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]One[[:space:]]Shot,[[:space:]]Beyond[[:space:]]One[[:space:]]Perspective_[[:space:]]Cross-View[[:space:]]and[[:space:]]Long-Horizon[[:space:]]Distillation[[:space:]]for[[:space:]]Better[[:space:]]LiDAR[[:space:]]Representations/411fc8bc-d246-4dc7-8b7d-e77fdc996018_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Perspective_[[:space:]]Neural[[:space:]]360-Degree[[:space:]]Video[[:space:]]Compression/4b2008c1-868c-4c58-92e2-c31cb64a8b86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Pixel[[:space:]]Uncertainty_[[:space:]]Bounding[[:space:]]the[[:space:]]OoD[[:space:]]Objects[[:space:]]in[[:space:]]Road[[:space:]]Scenes/13a2c695-3f96-4e76-baf7-9cb2e4366475_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]RGB_[[:space:]]Adaptive[[:space:]]Parallel[[:space:]]Processing[[:space:]]for[[:space:]]RAW[[:space:]]Object[[:space:]]Detection/4eb59e5b-1f57-497c-978c-7200eb85b369_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Simple[[:space:]]Edits_[[:space:]]Composed[[:space:]]Video[[:space:]]Retrieval[[:space:]]with[[:space:]]Dense[[:space:]]Modifications/7374eb22-19fd-40b3-bcb7-dcde41ada27e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Single[[:space:]]Images_[[:space:]]Retrieval[[:space:]]Self-Augmented[[:space:]]Unsupervised[[:space:]]Camouflaged[[:space:]]Object[[:space:]]Detection/97e273cc-5238-4a9b-9faf-ebe9fe62d930_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Spatial[[:space:]]Frequency_[[:space:]]Pixel-wise[[:space:]]Temporal[[:space:]]Frequency-based[[:space:]]Deepfake[[:space:]]Video[[:space:]]Detection/366e3279-90a7-4c97-b868-1fc7506e7b8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Text-Visual[[:space:]]Attention_[[:space:]]Exploiting[[:space:]]Visual[[:space:]]Cues[[:space:]]for[[:space:]]Effective[[:space:]]Token[[:space:]]Pruning[[:space:]]in[[:space:]]VLMs/eafcca81-f214-4605-a52a-d1eb2ee2c994_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Training_[[:space:]]Dynamic[[:space:]]Token[[:space:]]Merging[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Video[[:space:]]Understanding/7baed4af-115a-44c0-9a2e-77849f053790_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]Walking_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Image-Text[[:space:]]Benchmark[[:space:]]for[[:space:]]Text-based[[:space:]]Person[[:space:]]Anomaly[[:space:]]Search/e900b99f-7757-4270-b1cd-fd4459a9328a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]\[cls\]_[[:space:]]Exploring[[:space:]]the[[:space:]]True[[:space:]]Potential[[:space:]]of[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling[[:space:]]Representations/c5c8af49-ecf3-4752-adb0-1e86aaff0867_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]the[[:space:]]Destination_[[:space:]]A[[:space:]]Novel[[:space:]]Benchmark[[:space:]]for[[:space:]]Exploration-Aware[[:space:]]Embodied[[:space:]]Question[[:space:]]Answering/7258dfaa-c243-40e3-a76a-9a662f9881a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]the[[:space:]]Frame_[[:space:]]Generating[[:space:]]360deg[[:space:]]Panoramic[[:space:]]Videos[[:space:]]from[[:space:]]Perspective[[:space:]]Videos/f127c640-86a7-4f40-a9be-2e54965f8987_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Beyond[[:space:]]the[[:space:]]Limits_[[:space:]]Overcoming[[:space:]]Negative[[:space:]]Correlation[[:space:]]of[[:space:]]Activation-Based[[:space:]]Training-Free[[:space:]]NAS/06340bee-5932-4dee-89d5-b15a79a2be91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BezierGS_[[:space:]]Dynamic[[:space:]]Urban[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]with[[:space:]]Bezier[[:space:]]Curve[[:space:]]Gaussian[[:space:]]Splatting/4010a7ee-2991-47dc-a024-cf1972398b96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bi-Level[[:space:]]Optimization[[:space:]]for[[:space:]]Self-Supervised[[:space:]]AI-Generated[[:space:]]Face[[:space:]]Detection/266f2139-68f1-49d6-866e-da985e83a28d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bias[[:space:]]in[[:space:]]Gender[[:space:]]Bias[[:space:]]Benchmarks_[[:space:]]How[[:space:]]Spurious[[:space:]]Features[[:space:]]Distort[[:space:]]Evaluation/a0e54493-9ff0-4d68-9480-5858e5123050_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bias-Resilient[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]Using[[:space:]]Normalizing[[:space:]]Flows/eed44ee3-e108-4775-9e90-5953e5912d3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bidirectional[[:space:]]Likelihood[[:space:]]Estimation[[:space:]]with[[:space:]]Multi-Modal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Text-Video[[:space:]]Retrieval/ed6270e0-5d3a-4db5-97a2-1285e80118b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bilateral[[:space:]]Collaboration[[:space:]]with[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]for[[:space:]]Open[[:space:]]Vocabulary[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection/47e608ed-fbb5-443f-ac40-8dfffde9189c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BillBoard[[:space:]]Splatting[[:space:]](BBSplat)_[[:space:]]Learnable[[:space:]]Textured[[:space:]]Primitives[[:space:]]for[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/5ff16eea-ddde-4d67-8d58-75a27e3f3bbf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bitrate-Controlled[[:space:]]Diffusion[[:space:]]for[[:space:]]Disentangling[[:space:]]Motion[[:space:]]and[[:space:]]Content[[:space:]]in[[:space:]]Video/0ada41b2-ae5c-4da0-bb66-ab8884073391_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Blended[[:space:]]Point[[:space:]]Cloud[[:space:]]Diffusion[[:space:]]for[[:space:]]Localized[[:space:]]Text-guided[[:space:]]Shape[[:space:]]Editing/bac707da-18eb-4a97-838e-5942186ea898_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Blind[[:space:]]Noisy[[:space:]]Image[[:space:]]Deblurring[[:space:]]Using[[:space:]]Residual[[:space:]]Guidance[[:space:]]Strategy/b3ef2f04-156c-447a-a625-e14483f4cc21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Blind[[:space:]]Video[[:space:]]Super-Resolution[[:space:]]based[[:space:]]on[[:space:]]Implicit[[:space:]]Kernels/c94ca448-71ce-4b13-a801-3c55ceb8b70c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Blind2Sound_[[:space:]]Self-Supervised[[:space:]]Image[[:space:]]Denoising[[:space:]]without[[:space:]]Residual[[:space:]]Noise/be445ee1-42f9-4396-9e6a-b38868614bd1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BlinkTrack_[[:space:]]Feature[[:space:]]Tracking[[:space:]]over[[:space:]]80[[:space:]]FPS[[:space:]]via[[:space:]]Events[[:space:]]and[[:space:]]Images/93ad602a-4bc5-47c3-afaf-6542df0081a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BlueNeg_[[:space:]]A[[:space:]]35mm[[:space:]]Negative[[:space:]]Film[[:space:]]Dataset[[:space:]]for[[:space:]]Restoring[[:space:]]Channel-Heterogeneous[[:space:]]Deterioration/450008ad-87fc-4769-bf93-2946d9523f0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BokehDiff_[[:space:]]Neural[[:space:]]Lens[[:space:]]Blur[[:space:]]with[[:space:]]One-Step[[:space:]]Diffusion/12ebee52-3cbc-4c02-933e-67cc6c4c4228_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bokehlicious_[[:space:]]Photorealistic[[:space:]]Bokeh[[:space:]]Rendering[[:space:]]with[[:space:]]Controllable[[:space:]]Apertures/cd01bf84-d7e5-4852-9d4d-d56e98c70fb3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bolt3D_[[:space:]]Generating[[:space:]]3D[[:space:]]Scenes[[:space:]]in[[:space:]]Seconds/efa7c237-8d25-4070-8d58-b74c414ec541_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boost[[:space:]]3D[[:space:]]Reconstruction[[:space:]]using[[:space:]]Diffusion-based[[:space:]]Monocular[[:space:]]Camera[[:space:]]Calibration/290d3795-e6fb-4654-9c2d-fbf01f2ab3b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]Adversarial[[:space:]]Transferability[[:space:]]via[[:space:]]Negative[[:space:]]Hessian[[:space:]]Trace[[:space:]]Regularization/d0669751-d806-4753-b603-d3b569ccaa23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]Adversarial[[:space:]]Transferability[[:space:]]via[[:space:]]Residual[[:space:]]Perturbation[[:space:]]Attack/c54a8e07-6dd4-44e9-8410-cdd1b2ba3431_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]Class[[:space:]]Representation[[:space:]]via[[:space:]]Semantically[[:space:]]Related[[:space:]]Instances[[:space:]]for[[:space:]]Robust[[:space:]]Long-Tailed[[:space:]]Learning[[:space:]]with[[:space:]]Noisy[[:space:]]Labels/2bad529a-50a1-46b2-b080-a54d048f2a6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]Domain[[:space:]]Generalized[[:space:]]and[[:space:]]Adaptive[[:space:]]Detection[[:space:]]with[[:space:]]Diffusion[[:space:]]Models_[[:space:]]Fitness,[[:space:]]Generalization,[[:space:]]and[[:space:]]Transferability/51372488-40a7-4c89-bb2d-5a4af5a3d142_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]Generative[[:space:]]Adversarial[[:space:]]Transferability[[:space:]]with[[:space:]]Self-supervised[[:space:]]Vision[[:space:]]Transformer[[:space:]]Features/232b97a7-2134-4265-a97a-c8ce5ef6426a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]MLLM[[:space:]]Reasoning[[:space:]]with[[:space:]]Text-Debiased[[:space:]]Hint-GRPO/9b98a3d4-840b-4e2f-b3c0-16f818b2953e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]Multi-View[[:space:]]Indoor[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Adaptive[[:space:]]3D[[:space:]]Volume[[:space:]]Construction/bf18b635-acb1-433f-b09e-5ed70027e232_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]Multimodal[[:space:]]Learning[[:space:]]via[[:space:]]Disentangled[[:space:]]Gradient[[:space:]]Learning/bbaa52e2-be6b-4b6a-89e9-a5413e1a712a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boosting[[:space:]]Vision[[:space:]]Semantic[[:space:]]Density[[:space:]]with[[:space:]]Anatomy[[:space:]]Normality[[:space:]]Modeling[[:space:]]for[[:space:]]Medical[[:space:]]Vision-language[[:space:]]Pre-training/1c57029b-2076-41fd-baae-3c0649760fbf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bootstrap3D_[[:space:]]Improving[[:space:]]Multi-view[[:space:]]Diffusion[[:space:]]Model[[:space:]]with[[:space:]]Synthetic[[:space:]]Data/6ed9b77f-ac9b-47c3-a021-d6f8b55b61cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bootstrapping[[:space:]]Grounded[[:space:]]Chain-of-Thought[[:space:]]in[[:space:]]Multimodal[[:space:]]LLMs[[:space:]]for[[:space:]]Data-Efficient[[:space:]]Model[[:space:]]Adaptation/ae05591a-d5bb-4b9d-9cb9-c6daf651b095_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Borrowing[[:space:]]Eyes[[:space:]]for[[:space:]]the[[:space:]]Blind[[:space:]]Spot_[[:space:]]Overcoming[[:space:]]Data[[:space:]]Scarcity[[:space:]]in[[:space:]]Malicious[[:space:]]Video[[:space:]]Detection[[:space:]]via[[:space:]]Cross-Domain[[:space:]]Retrieval[[:space:]]Augmentation/9f8877ef-1843-44d5-99d1-d4fb5c82dcae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Boundary[[:space:]]Probing[[:space:]]for[[:space:]]Input[[:space:]]Privacy[[:space:]]Protection[[:space:]]When[[:space:]]Using[[:space:]]LMM[[:space:]]Services/ce6cac13-4f16-4302-976f-477c88718405_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BoxDreamer_[[:space:]]Dreaming[[:space:]]Box[[:space:]]Corners[[:space:]]for[[:space:]]Generalizable[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/9898a77f-785b-4d97-bf4d-741986f15538_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Breaking[[:space:]]Grid[[:space:]]Constraints_[[:space:]]Dynamic[[:space:]]Graph[[:space:]]Reconstruction[[:space:]]Network[[:space:]]for[[:space:]]Multi-organ[[:space:]]Segmentation/c1147e80-35b6-4474-845f-ea63184e709a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Breaking[[:space:]]Rectangular[[:space:]]Shackles_[[:space:]]Cross-View[[:space:]]Object[[:space:]]Segmentation[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Object[[:space:]]Geo-Localization/1c0e1ea3-9196-4aed-844d-df557e67d799_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Breaking[[:space:]]the[[:space:]]Encoder[[:space:]]Barrier[[:space:]]for[[:space:]]Seamless[[:space:]]Video-Language[[:space:]]Understanding/e37dd9f4-a91c-4cb6-b14e-e901d26f263b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/BridgeDepth_[[:space:]]Bridging[[:space:]]Monocular[[:space:]]and[[:space:]]Stereo[[:space:]]Reasoning[[:space:]]with[[:space:]]Latent[[:space:]]Alignment/b70427a1-3c66-43ad-a017-36d72b21b9bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]3D[[:space:]]Anomaly[[:space:]]Localization[[:space:]]and[[:space:]]Repair[[:space:]]via[[:space:]]High-Quality[[:space:]]Continuous[[:space:]]Geometric[[:space:]]Representation/d80de6ff-9b89-4246-8e1a-093939d4191a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]Class[[:space:]]Imbalance[[:space:]]and[[:space:]]Partial[[:space:]]Labeling[[:space:]]via[[:space:]]Spectral-Balanced[[:space:]]Energy[[:space:]]Propagation[[:space:]]for[[:space:]]Skeleton-based[[:space:]]Action[[:space:]]Recognition/df0caeaa-19ff-40e4-94ab-fcfb91486537_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]Continuous[[:space:]]and[[:space:]]Discrete[[:space:]]Tokens[[:space:]]for[[:space:]]Autoregressive[[:space:]]Visual[[:space:]]Generation/d6554979-393d-4d14-bb41-03cd412c943b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]Diffusion[[:space:]]Models[[:space:]]and[[:space:]]3D[[:space:]]Representations_[[:space:]]A[[:space:]]3D[[:space:]]Consistent[[:space:]]Super-Resolution[[:space:]]Framework/c7991697-8db5-4c7d-9a57-8701a16307e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]Domain[[:space:]]Generalization[[:space:]]to[[:space:]]Multimodal[[:space:]]Domain[[:space:]]Generalization[[:space:]]via[[:space:]]Unified[[:space:]]Representations/7d824f18-c9e3-4945-9cdd-5dce04f5e29f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]Local[[:space:]]Inductive[[:space:]]Bias[[:space:]]and[[:space:]]Long-Range[[:space:]]Dependencies[[:space:]]with[[:space:]]Pixel-Mamba[[:space:]]for[[:space:]]End-to-end[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Analysis/1d0e34b8-16b3-4b60-8cff-441923314ded_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]the[[:space:]]Gap[[:space:]]Between[[:space:]]Ideal[[:space:]]and[[:space:]]Real-world[[:space:]]Evaluation_[[:space:]]Benchmarking[[:space:]]AI-Generated[[:space:]]Image[[:space:]]Detection[[:space:]]in[[:space:]]Challenging[[:space:]]Scenarios/0dafd2da-fa01-4a85-8cfc-bd72f2054664_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]the[[:space:]]Gap[[:space:]]between[[:space:]]Brain[[:space:]]and[[:space:]]Machine[[:space:]]in[[:space:]]Interpreting[[:space:]]Visual[[:space:]]Semantics_[[:space:]]Towards[[:space:]]Self-adaptive[[:space:]]Brain-to-Text[[:space:]]Decoding/237098f9-30f0-47ee-9ef3-f8d3c45a7842_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]the[[:space:]]Skeleton-Text[[:space:]]Modality[[:space:]]Gap_[[:space:]]Diffusion-Powered[[:space:]]Modality[[:space:]]Alignment[[:space:]]for[[:space:]]Zero-shot[[:space:]]Skeleton-based[[:space:]]Action[[:space:]]Recognition/749973bd-99ee-4b36-a489-ad2b57e63e19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bridging[[:space:]]the[[:space:]]Sky[[:space:]]and[[:space:]]Ground_[[:space:]]Towards[[:space:]]View-Invariant[[:space:]]Feature[[:space:]]Learning[[:space:]]for[[:space:]]Aerial-Ground[[:space:]]Person[[:space:]]Re-Identification/862fd8f1-d83c-489d-83b6-25dcdc39d449_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bring[[:space:]]Your[[:space:]]Rear[[:space:]]Cameras[[:space:]]for[[:space:]]Egocentric[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/6139a780-030c-4f10-9002-a037fcded436_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Bringing[[:space:]]RNNs[[:space:]]Back[[:space:]]to[[:space:]]Efficient[[:space:]]Open-Ended[[:space:]]Video[[:space:]]Understanding/d9a72757-de22-4a16-b97b-188028ce8bcf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/C2MIL_[[:space:]]Synchronizing[[:space:]]Semantic[[:space:]]and[[:space:]]Topological[[:space:]]Causalities[[:space:]]in[[:space:]]Multiple[[:space:]]Instance[[:space:]]Learning[[:space:]]for[[:space:]]Robust[[:space:]]and[[:space:]]Interpretable[[:space:]]Survival[[:space:]]Analysis/6f2a155e-a0ab-4f57-b8c3-ac3d9e3cc302_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/C4D_[[:space:]]4D[[:space:]]Made[[:space:]]from[[:space:]]3D[[:space:]]through[[:space:]]Dual[[:space:]]Correspondences/8ee8243e-620b-4043-b179-99be3c032189_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CA-I2P_[[:space:]]Channel-Adaptive[[:space:]]Registration[[:space:]]Network[[:space:]]with[[:space:]]Global[[:space:]]Optimal[[:space:]]Selection/b5af9489-5dec-4e7b-b03e-450eb2aa0a1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CA2C_[[:space:]]A[[:space:]]Prior-Knowledge-Free[[:space:]]Approach[[:space:]]for[[:space:]]Robust[[:space:]]Label[[:space:]]Noise[[:space:]]Learning[[:space:]]via[[:space:]]Asymmetric[[:space:]]Co-learning[[:space:]]and[[:space:]]Co-training/efaab113-c048-41e2-b77e-8779b1fc9332_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CABLD_[[:space:]]Contrast-Agnostic[[:space:]]Brain[[:space:]]Landmark[[:space:]]Detection[[:space:]]with[[:space:]]Consistency-Based[[:space:]]Regularization/a2c3461c-5b40-4a17-b40c-939c65065286_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CAD-Assistant_[[:space:]]Tool-Augmented[[:space:]]VLLMs[[:space:]]as[[:space:]]Generic[[:space:]]CAD[[:space:]]Task[[:space:]]Solvers/7d183306-a9e0-413e-9379-9fe106b3b2fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CAD-Recode_[[:space:]]Reverse[[:space:]]Engineering[[:space:]]CAD[[:space:]]Code[[:space:]]from[[:space:]]Point[[:space:]]Clouds/f1b353f1-c402-4453-adf3-d21da57883ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CAFA_[[:space:]]a[[:space:]]Controllable[[:space:]]Automatic[[:space:]]Foley[[:space:]]Artist/e52718db-14fb-4392-893f-4c0b4af3412e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CAPTURE_[[:space:]]Evaluating[[:space:]]Spatial[[:space:]]Reasoning[[:space:]]in[[:space:]]Vision[[:space:]]Language[[:space:]]Models[[:space:]]via[[:space:]]Occluded[[:space:]]Object[[:space:]]Counting/8ef58c08-87dc-4ce0-9e7a-0df7fa78b0ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CAP_[[:space:]]Evaluation[[:space:]]of[[:space:]]Persuasive[[:space:]]and[[:space:]]Creative[[:space:]]Image[[:space:]]Generation/19126ed7-01f4-4df4-aeb8-c4c97125a63a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CARIM_[[:space:]]Caption-Based[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Scene[[:space:]]Retrieval[[:space:]]via[[:space:]]Inclusive[[:space:]]Text[[:space:]]Matching/53bad9b6-e350-4182-bc26-af7ae4dbf97d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CARL_[[:space:]]Causality-guided[[:space:]]Architecture[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]an[[:space:]]Interpretable[[:space:]]Performance[[:space:]]Predictor/9032821e-dbf6-4585-a125-dd521fa29d36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CARP_[[:space:]]Visuomotor[[:space:]]Policy[[:space:]]Learning[[:space:]]via[[:space:]]Coarse-to-Fine[[:space:]]Autoregressive[[:space:]]Prediction/59ec4a8d-aa01-4eaa-b732-2709d4afe3f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CATP-LLM_[[:space:]]Empowering[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]for[[:space:]]Cost-Aware[[:space:]]Tool[[:space:]]Planning/ff124593-afca-4369-93cd-d5bdfdca0728_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CATSplat_[[:space:]]Context-Aware[[:space:]]Transformer[[:space:]]with[[:space:]]Spatial[[:space:]]Guidance[[:space:]]for[[:space:]]Generalizable[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]A[[:space:]]Single-View[[:space:]]Image/67e59e65-c63d-4054-aa41-386a513662f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CAT_[[:space:]]A[[:space:]]Unified[[:space:]]Click-and-Track[[:space:]]Framework[[:space:]]for[[:space:]]Realistic[[:space:]]Tracking/8aaf2137-1fce-4a96-a316-4d998f3230cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CAVIS_[[:space:]]Context-Aware[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation/fcd72415-5b1c-4828-9df0-c1f5a6e3f7ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CC-OCR_[[:space:]]A[[:space:]]Comprehensive[[:space:]]and[[:space:]]Challenging[[:space:]]OCR[[:space:]]Benchmark[[:space:]]for[[:space:]]Evaluating[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models[[:space:]]in[[:space:]]Literacy/270758c3-5044-488c-837f-97415cff5c82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CCL-LGS_[[:space:]]Contrastive[[:space:]]Codebook[[:space:]]Learning[[:space:]]for[[:space:]]3D[[:space:]]Language[[:space:]]Gaussian[[:space:]]Splatting/4629f162-0eeb-41d0-bf08-bb82ae4b075e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CCMNet_[[:space:]]Leveraging[[:space:]]Calibrated[[:space:]]Color[[:space:]]Correction[[:space:]]Matrices[[:space:]]for[[:space:]]Cross-Camera[[:space:]]Color[[:space:]]Constancy/6a191a53-023e-4e97-80d3-37c5ec3160ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CE-FAM_[[:space:]]Concept-Based[[:space:]]Explanation[[:space:]]via[[:space:]]Fusion[[:space:]]of[[:space:]]Activation[[:space:]]Maps/cb003548-537a-4771-bb5d-a738f4924e65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CF3_[[:space:]]Compact[[:space:]]and[[:space:]]Fast[[:space:]]3D[[:space:]]Feature[[:space:]]Fields/e7376295-ecc8-4cf7-b983-f8e0e2889ca8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CHARM3R_[[:space:]]Towards[[:space:]]Unseen[[:space:]]Camera[[:space:]]Height[[:space:]]Robust[[:space:]]Monocular[[:space:]]3D[[:space:]]Detector/edbcc5c0-df46-499a-9472-a1574043fa44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CHORDS_[[:space:]]Diffusion[[:space:]]Sampling[[:space:]]Accelerator[[:space:]]with[[:space:]]Multi-core[[:space:]]Hierarchical[[:space:]]ODE[[:space:]]Solvers/ea0fd999-c3d3-4926-afab-213aabec9c6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CHROME_[[:space:]]Clothed[[:space:]]Human[[:space:]]Reconstruction[[:space:]]with[[:space:]]Occlusion-Resilience[[:space:]]and[[:space:]]Multiview-Consistency[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/c7596452-50e4-4796-ae21-8ca273afea9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CIARD_[[:space:]]Cyclic[[:space:]]Iterative[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]Distillation/f46d7ff1-1185-4722-b40e-54c1bd985f05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CL-Splats_[[:space:]]Continual[[:space:]]Learning[[:space:]]of[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]Local[[:space:]]Optimization/aa083f47-22e6-4214-b786-911cc68dbb0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CLIP-Adapted[[:space:]]Region-to-Text[[:space:]]Learning[[:space:]]for[[:space:]]Generative[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/d5d3a546-5838-4e4f-966a-cc80167e6cba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CLIP-GS_[[:space:]]Unifying[[:space:]]Vision-Language[[:space:]]Representation[[:space:]]with[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/5006a90e-08dc-4564-b926-cb61cf4fe302_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CLIPSym_[[:space:]]Delving[[:space:]]into[[:space:]]Symmetry[[:space:]]Detection[[:space:]]with[[:space:]]CLIP/a1d8e2c4-84b4-4c58-bbc3-16401eefed3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CLIPer_[[:space:]]Hierarchically[[:space:]]Improving[[:space:]]Spatial[[:space:]]Representation[[:space:]]of[[:space:]]CLIP[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/9082efb3-147c-4475-98dc-dc0bbf401dbd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CLOT_[[:space:]]Closed[[:space:]]Loop[[:space:]]Optimal[[:space:]]Transport[[:space:]]for[[:space:]]Unsupervised[[:space:]]Action[[:space:]]Segmentation/914be78d-f00d-4321-a94e-4d344c4615b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CMAD_[[:space:]]Correlation-Aware[[:space:]]and[[:space:]]Modalities-Aware[[:space:]]Distillation[[:space:]]for[[:space:]]Multimodal[[:space:]]Sentiment[[:space:]]Analysis[[:space:]]with[[:space:]]Missing[[:space:]]Modalities/0940f9ac-200c-4ec3-a944-6ffd4ac83e03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CMB-ML_[[:space:]]A[[:space:]]Cosmic[[:space:]]Microwave[[:space:]]Background[[:space:]]Dataset[[:space:]]for[[:space:]]the[[:space:]]Oldest[[:space:]]Possible[[:space:]]Computer[[:space:]]Vision[[:space:]]Task/f7053019-584b-45d7-bb93-994aa63cc85f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CMT_[[:space:]]A[[:space:]]Cascade[[:space:]]MAR[[:space:]]with[[:space:]]Topology[[:space:]]Predictor[[:space:]]for[[:space:]]Multimodal[[:space:]]Conditional[[:space:]]CAD[[:space:]]Generation/a15dfd10-7528-4731-bad9-135c1ba83e4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CNS-Bench_[[:space:]]Benchmarking[[:space:]]Image[[:space:]]Classifier[[:space:]]Robustness[[:space:]]Under[[:space:]]Continuous[[:space:]]Nuisance[[:space:]]Shifts/9099dd2c-1c06-4fd4-ac5d-7f3fc9537ae3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CO2-Net_[[:space:]]A[[:space:]]Physics-Informed[[:space:]]Spatio-Temporal[[:space:]]Model[[:space:]]for[[:space:]]Global[[:space:]]Surface[[:space:]]CO2[[:space:]]Reconstruction/a96ed665-8838-4445-8355-e068cd6c9188_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CODA_[[:space:]]Repurposing[[:space:]]Continuous[[:space:]]VAEs[[:space:]]for[[:space:]]Discrete[[:space:]]Tokenization/6eb4ceed-e757-4444-a6e9-5aac45a7c69b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CODE-CL_[[:space:]]Conceptor-Based[[:space:]]Gradient[[:space:]]Projection[[:space:]]for[[:space:]]Deep[[:space:]]Continual[[:space:]]Learning/68e137e2-01b0-415a-94eb-32d65a535122_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/COIN_[[:space:]]Confidence[[:space:]]Score-Guided[[:space:]]Distillation[[:space:]]for[[:space:]]Annotation-Free[[:space:]]Cell[[:space:]]Segmentation/42fed33a-5700-4b51-8e9b-90fd0b6d8da4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/COME_[[:space:]]Dual[[:space:]]Structure-Semantic[[:space:]]Learning[[:space:]]with[[:space:]]Collaborative[[:space:]]MoE[[:space:]]for[[:space:]]Universal[[:space:]]Lesion[[:space:]]Detection[[:space:]]Across[[:space:]]Heterogeneous[[:space:]]Ultrasound[[:space:]]Datasets/ca2466e2-b988-4cd9-9400-2c0629172dc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/COSMO_[[:space:]]Combination[[:space:]]of[[:space:]]Selective[[:space:]]Memorization[[:space:]]for[[:space:]]Low-cost[[:space:]]Vision-and-Language[[:space:]]Navigation/3193a0f3-a8e5-4155-9b94-0a7e8b5e16f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/COSTARR_[[:space:]]Consolidated[[:space:]]Open[[:space:]]Set[[:space:]]Technique[[:space:]]with[[:space:]]Attenuation[[:space:]]for[[:space:]]Robust[[:space:]]Recognition/fb3a50aa-e77a-4e38-baab-2bb2ed4affc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/COVTrack_[[:space:]]Continuous[[:space:]]Open-Vocabulary[[:space:]]Tracking[[:space:]]via[[:space:]]Adaptive[[:space:]]Multi-Cue[[:space:]]Fusion/8eaee3e6-0a1d-45b4-9576-ef9fcb83a781_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CObL_[[:space:]]Toward[[:space:]]Zero-Shot[[:space:]]Ordinal[[:space:]]Layering[[:space:]]without[[:space:]]User[[:space:]]Prompting/3b21d04f-3977-4749-8ee2-5d060ab967ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CRAM_[[:space:]]Large[[:space:]]Scale[[:space:]]Video[[:space:]]Continual[[:space:]]Learning[[:space:]]with[[:space:]]Bootstrapped[[:space:]]Compression/d09a38af-79ec-4155-bbb6-ae4d82bd0a37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CSD-VAR_[[:space:]]Content-Style[[:space:]]Decomposition[[:space:]]in[[:space:]]Visual[[:space:]]Autoregressive[[:space:]]Models/4bda2207-6153-46cb-9aa3-f32fadbb43d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CT-ScanGaze_[[:space:]]A[[:space:]]Dataset[[:space:]]and[[:space:]]Baselines[[:space:]]for[[:space:]]3D[[:space:]]Volumetric[[:space:]]Scanpath[[:space:]]Modeling/293ab9cd-e58f-404b-a1b3-58d5382f70ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CULTURE3D_[[:space:]]A[[:space:]]Large-Scale[[:space:]]and[[:space:]]Diverse[[:space:]]Dataset[[:space:]]of[[:space:]]Cultural[[:space:]]Landmarks[[:space:]]and[[:space:]]Terrains[[:space:]]for[[:space:]]Gaussian-Based[[:space:]]Scene[[:space:]]Rendering/ce091b63-a4eb-4fe0-a67d-23e429687c5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CVFusion_[[:space:]]Cross-View[[:space:]]Fusion[[:space:]]of[[:space:]]4D[[:space:]]Radar[[:space:]]and[[:space:]]Camera[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/43a90e91-df8d-4c79-8c46-636747ed6c0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CVPT_[[:space:]]Cross[[:space:]]Visual[[:space:]]Prompt[[:space:]]Tuning/89faacc1-952e-4cc7-9a82-e4a750f52b6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CWNet_[[:space:]]Causal[[:space:]]Wavelet[[:space:]]Network[[:space:]]for[[:space:]]Low-Light[[:space:]]Image[[:space:]]Enhancement/8e8b06cf-2310-483c-a8ef-439d722d529b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CaO2_[[:space:]]Rectifying[[:space:]]Inconsistencies[[:space:]]in[[:space:]]Diffusion-Based[[:space:]]Dataset[[:space:]]Distillation/a49ceb97-e9a3-4071-8918-eb8bca6e5203_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CaliMatch_[[:space:]]Adaptive[[:space:]]Calibration[[:space:]]for[[:space:]]Improving[[:space:]]Safe[[:space:]]Semi-supervised[[:space:]]Learning/00bc9cad-afb5-452e-90f8-b9e83b7c194c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Calibrating[[:space:]]MLLM-as-a-judge[[:space:]]via[[:space:]]Multimodal[[:space:]]Bayesian[[:space:]]Prompt[[:space:]]Ensembles/13f4b251-cb91-4f2a-b959-196c806ede43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CalliReader_[[:space:]]Contextualizing[[:space:]]Chinese[[:space:]]Calligraphy[[:space:]]via[[:space:]]an[[:space:]]Embedding-Aligned[[:space:]]Vision-Language[[:space:]]Model/ffe88dd1-7d2a-41e9-b27f-56ba12683cc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CameraCtrl[[:space:]]II_[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Exploration[[:space:]]via[[:space:]]Camera-controlled[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/8352aa04-6293-4267-a43d-49a46bc72182_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Can[[:space:]]Generative[[:space:]]Geospatial[[:space:]]Diffusion[[:space:]]Models[[:space:]]Excel[[:space:]]as[[:space:]]Discriminative[[:space:]]Geospatial[[:space:]]Foundation[[:space:]]Models_/e031da02-662c-4b01-8a3f-1b44703d2601_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Can[[:space:]]Knowledge[[:space:]]be[[:space:]]Transferred[[:space:]]from[[:space:]]Unimodal[[:space:]]to[[:space:]]Multimodal_[[:space:]]Investigating[[:space:]]the[[:space:]]Transitivity[[:space:]]of[[:space:]]Multimodal[[:space:]]Knowledge[[:space:]]Editing/1ba1dde3-a767-4147-9663-83cd71491e2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Can[[:space:]]We[[:space:]]Achieve[[:space:]]Efficient[[:space:]]Diffusion[[:space:]]Without[[:space:]]Self-Attention_[[:space:]]Distilling[[:space:]]Self-Attention[[:space:]]into[[:space:]]Convolutions/6fcabdcc-1cbf-456b-90f2-161743b7000d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Can3Tok_[[:space:]]Canonical[[:space:]]3D[[:space:]]Tokenization[[:space:]]and[[:space:]]Latent[[:space:]]Modeling[[:space:]]of[[:space:]]Scene-Level[[:space:]]3D[[:space:]]Gaussians/e7e4733f-b792-4ad0-9c57-2e933abe7690_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CanFields_[[:space:]]Consolidating[[:space:]]Diffeomorphic[[:space:]]Flows[[:space:]]for[[:space:]]Non-Rigid[[:space:]]4D[[:space:]]Interpolation[[:space:]]from[[:space:]]Arbitrary-Length[[:space:]]Sequences/dad1473b-32fa-4035-8977-f588202d7bca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CanonSwap_[[:space:]]High-Fidelity[[:space:]]and[[:space:]]Consistent[[:space:]]Video[[:space:]]Face[[:space:]]Swapping[[:space:]]via[[:space:]]Canonical[[:space:]]Space[[:space:]]Modulation/ae478fce-b46c-48aa-8955-cc6fec376de4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CapeLLM_[[:space:]]Support-Free[[:space:]]Category-Agnostic[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/6ba06199-2fb8-46b6-a250-f3800e2eca2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CaptionSmiths_[[:space:]]Flexibly[[:space:]]Controlling[[:space:]]Language[[:space:]]Pattern[[:space:]]in[[:space:]]Image[[:space:]]Captioning/63955314-4e1b-4765-ad6b-8a1f9651bab3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Capturing[[:space:]]head[[:space:]]avatar[[:space:]]with[[:space:]]hand[[:space:]]contacts[[:space:]]from[[:space:]]a[[:space:]]monocular[[:space:]]video/eb3e7a48-922e-44a0-8820-78078e151427_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CarGait_[[:space:]]Cross-Attention[[:space:]]based[[:space:]]Re-ranking[[:space:]]for[[:space:]]Gait[[:space:]]recognition/1c88b96c-6147-4ab7-b494-e2b2ce320923_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CasP_[[:space:]]Improving[[:space:]]Semi-Dense[[:space:]]Feature[[:space:]]Matching[[:space:]]Pipeline[[:space:]]Leveraging[[:space:]]Cascaded[[:space:]]Correspondence[[:space:]]Priors[[:space:]]for[[:space:]]Guidance/3dc8c87d-3eb5-4b83-b64e-bf78d08655cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cassic_[[:space:]]Towards[[:space:]]Content-Adaptive[[:space:]]State-Space[[:space:]]Models[[:space:]]for[[:space:]]Learned[[:space:]]Image[[:space:]]Compression/02649d75-d848-4675-890d-9e6b2bf39e4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Category-Specific[[:space:]]Selective[[:space:]]Feature[[:space:]]Enhancement[[:space:]]for[[:space:]]Long-Tailed[[:space:]]Multi-Label[[:space:]]Image[[:space:]]Classification/118f1094-4cc7-46a2-8a34-034ca1d8df50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Causal[[:space:]]Disentanglement[[:space:]]and[[:space:]]Cross-Modal[[:space:]]Alignment[[:space:]]for[[:space:]]Enhanced[[:space:]]Few-Shot[[:space:]]Learning/0b1e0c49-dbbd-4c39-bcbd-410b45dbf527_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Causal-Entity[[:space:]]Reflected[[:space:]]Egocentric[[:space:]]Traffic[[:space:]]Accident[[:space:]]Video[[:space:]]Synthesis/991a42be-58c8-4be2-a48e-41d4d7877dd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Causality-guided[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Vision-language[[:space:]]Models[[:space:]]via[[:space:]]Visual[[:space:]]Granulation/601e9247-60d3-4c92-9508-1b5e34e9b62a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Certifiably[[:space:]]Optimal[[:space:]]Anisotropic[[:space:]]Rotation[[:space:]]Averaging/7ff0fca7-fa28-4bb9-a8b4-1487438be00b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CharaConsist_[[:space:]]Fine-Grained[[:space:]]Consistent[[:space:]]Character[[:space:]]Generation/34f4c915-b2ac-4ecd-be05-c32202c44e83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ChartCap_[[:space:]]Mitigating[[:space:]]Hallucination[[:space:]]of[[:space:]]Dense[[:space:]]Chart[[:space:]]Captioning/6addc65a-d44e-42e5-aab0-74a3ac6a1899_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ChartPoint_[[:space:]]Guiding[[:space:]]MLLMs[[:space:]]with[[:space:]]Grounding[[:space:]]Reflection[[:space:]]for[[:space:]]Chart[[:space:]]Reasoning/b2e71782-5aa8-4364-a4fe-19abbbeb00cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ChatReID_[[:space:]]Open-ended[[:space:]]Interactive[[:space:]]Person[[:space:]]Retrieval[[:space:]]via[[:space:]]Hierarchical[[:space:]]Progressive[[:space:]]Tuning[[:space:]]for[[:space:]]Vision[[:space:]]Language[[:space:]]Models/3205898c-0741-4f4a-921e-4bf2cf5ffbb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Chimera_[[:space:]]Improving[[:space:]]Generalist[[:space:]]Model[[:space:]]with[[:space:]]Domain-Specific[[:space:]]Experts/7b9a8c15-476b-41b7-9ec9-1d345c4cc2a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CityGS-X_[[:space:]]A[[:space:]]Scalable[[:space:]]Architecture[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Geometrically[[:space:]]Accurate[[:space:]]Large-Scale[[:space:]]Scene[[:space:]]Reconstruction/f148ae3f-5ec3-4762-bfdc-490ff0e12d81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CityNav_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]for[[:space:]]Real-World[[:space:]]Aerial[[:space:]]Navigation/fc2e0607-0eb1-4570-baa3-4dd942b42913_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ClaraVid_[[:space:]]A[[:space:]]Holistic[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]Benchmark[[:space:]]From[[:space:]]Aerial[[:space:]]Perspective[[:space:]]With[[:space:]]Delentropy-Based[[:space:]]Complexity[[:space:]]Profiling/5cedd07e-5f33-4602-855e-eccb93f583aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Class[[:space:]]Token[[:space:]]as[[:space:]]Proxy_[[:space:]]Optimal[[:space:]]Transport-assisted[[:space:]]Proxy[[:space:]]Learning[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/56946d0e-d7f3-422e-8c20-3a9bf71c93da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Class-Wise[[:space:]]Federated[[:space:]]Averaging[[:space:]]for[[:space:]]Efficient[[:space:]]Personalization/cbf82d9e-7455-4a08-b85d-80254c14d0a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CleanPose_[[:space:]]Category-Level[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Causal[[:space:]]Learning[[:space:]]and[[:space:]]Knowledge[[:space:]]Distillation/c7007f95-91c3-4ad7-8411-51ed429c5034_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ClearSight_[[:space:]]Human[[:space:]]Vision-Inspired[[:space:]]Solutions[[:space:]]for[[:space:]]Event-Based[[:space:]]Motion[[:space:]]Deblurring/1c4609b3-40c1-4b96-9c52-6f3d59bf25bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Client2Vec_[[:space:]]Improving[[:space:]]Federated[[:space:]]Learning[[:space:]]by[[:space:]]Distribution[[:space:]]Shifts[[:space:]]Aware[[:space:]]Client[[:space:]]Indexing/45df9d68-ad39-4c36-bb3c-2efd90e244ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Clink![[:space:]]Chop![[:space:]]Thud![[:space:]]-[[:space:]]Learning[[:space:]]Object[[:space:]]Sounds[[:space:]]from[[:space:]]Real-World[[:space:]]Interactions/f3519308-73a3-4235-af1a-ee2770a2139c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Closed-Loop[[:space:]]Transfer[[:space:]]for[[:space:]]Weakly-supervised[[:space:]]Affordance[[:space:]]Grounding/43dd40b6-3221-4554-89d0-978a843b4bcd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Co-Painter_[[:space:]]Fine-Grained[[:space:]]Controllable[[:space:]]Image[[:space:]]Stylization[[:space:]]via[[:space:]]Implicit[[:space:]]Decoupling[[:space:]]and[[:space:]]Adaptive[[:space:]]Injection/f801ce8a-aa7d-4173-9c4a-8eedc8e8094c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoA-VLA_[[:space:]]Improving[[:space:]]Vision-Language-Action[[:space:]]Models[[:space:]]via[[:space:]]Visual-Text[[:space:]]Chain-of-Affordance/6905d3e0-c99a-47f4-8e14-7deb6c650597_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoDa-4DGS_[[:space:]]Dynamic[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]Context[[:space:]]and[[:space:]]Deformation[[:space:]]Awareness[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/004ef6fd-2203-4af1-be6f-3c6f87de22ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoHD_[[:space:]]A[[:space:]]Counting-Aware[[:space:]]Hierarchical[[:space:]]Decoding[[:space:]]Framework[[:space:]]for[[:space:]]Generalized[[:space:]]Referring[[:space:]]Expression[[:space:]]Segmentation/7322e451-c522-4f48-bc2a-46ae73860970_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoLMDriver_[[:space:]]LLM-based[[:space:]]Negotiation[[:space:]]Benefits[[:space:]]Cooperative[[:space:]]Autonomous[[:space:]]Driving/a2276a5c-4607-4505-a10f-268c61072f9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoMPaSS_[[:space:]]Enhancing[[:space:]]Spatial[[:space:]]Understanding[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/a4597ce0-af0d-4ac4-b761-b1103f455496_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoMatch_[[:space:]]Dynamic[[:space:]]Covisibility-Aware[[:space:]]Transformer[[:space:]]for[[:space:]]Bilateral[[:space:]]Subpixel-Level[[:space:]]Semi-Dense[[:space:]]Image[[:space:]]Matching/51a93536-d483-42b7-baf5-f944d78a5a60_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoMoGaussian_[[:space:]]Continuous[[:space:]]Motion-Aware[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Motion-Blurred[[:space:]]Images/b7917893-9110-4ab1-a36c-b1bdfa578d8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoSMIC_[[:space:]]Continual[[:space:]]Self-supervised[[:space:]]Learning[[:space:]]for[[:space:]]Multi-Domain[[:space:]]Medical[[:space:]]Imaging[[:space:]]via[[:space:]]Conditional[[:space:]]Mutual[[:space:]]Information[[:space:]]Maximization/8b79d7c6-0a4f-40ec-854f-0e45aa0f3096_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoST_[[:space:]]Efficient[[:space:]]Collaborative[[:space:]]Perception[[:space:]]From[[:space:]]Unified[[:space:]]Spatiotemporal[[:space:]]Perspective/55b4bb74-a65d-4a18-95a3-15797cfba5a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoStoDet-DDPM_[[:space:]]Collaborative[[:space:]]Training[[:space:]]of[[:space:]]Stochastic[[:space:]]and[[:space:]]Deterministic[[:space:]]Models[[:space:]]Improves[[:space:]]Surgical[[:space:]]Workflow[[:space:]]Anticipation[[:space:]]and[[:space:]]Recognition/e8c7c91d-a15a-4f2b-9761-7f24ff867915_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoTMR_[[:space:]]Chain-of-Thought[[:space:]]Multi-Scale[[:space:]]Reasoning[[:space:]]for[[:space:]]Training-Free[[:space:]]Zero-Shot[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/54d09786-6292-4fca-9df0-8b9eaa057c32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoTracker3_[[:space:]]Simpler[[:space:]]and[[:space:]]Better[[:space:]]Point[[:space:]]Tracking[[:space:]]by[[:space:]]Pseudo-Labelling[[:space:]]Real[[:space:]]Videos/45675701-ef8f-4e27-bf82-2450697e4cc3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CogCM_[[:space:]]Cognition-Inspired[[:space:]]Contextual[[:space:]]Modeling[[:space:]]for[[:space:]]Audio-Visual[[:space:]]Speech[[:space:]]Enhancement/09393d1d-5299-42fe-a5fc-4e06e42de31a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CogNav_[[:space:]]Cognitive[[:space:]]Process[[:space:]]Modeling[[:space:]]for[[:space:]]Object[[:space:]]Goal[[:space:]]Navigation[[:space:]]with[[:space:]]LLMs/c4fa9da6-8968-4a30-a6f3-721d9b657f3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Collaborative[[:space:]]Instance[[:space:]]Object[[:space:]]Navigation_[[:space:]]Leveraging[[:space:]]Uncertainty-Awareness[[:space:]]to[[:space:]]Minimize[[:space:]]Human-Agent[[:space:]]Dialogues/fa6ea408-1ce1-4d7e-bc24-21de6a52c918_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Color[[:space:]]Matching[[:space:]]Using[[:space:]]Hypernetwork-Based[[:space:]]Kolmogorov-Arnold[[:space:]]Networks/9b3cc803-4088-4c9b-beaf-574f651d3f80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Colors[[:space:]]See[[:space:]]Colors[[:space:]]Ignore_[[:space:]]Clothes[[:space:]]Changing[[:space:]]ReID[[:space:]]with[[:space:]]Color[[:space:]]Disentanglement/e9fe5895-39ea-45a1-addb-c9602312a99e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CombatVLA_[[:space:]]An[[:space:]]Efficient[[:space:]]Vision-Language-Action[[:space:]]Model[[:space:]]for[[:space:]]Combat[[:space:]]Tasks[[:space:]]in[[:space:]]3D[[:space:]]Action[[:space:]]Role-Playing[[:space:]]Games/dabddea7-c4fe-4eae-be4c-839988b872e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Combinative[[:space:]]Matching[[:space:]]for[[:space:]]Geometric[[:space:]]Shape[[:space:]]Assembly/28f750d6-d11c-48ec-8826-6f5786683494_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Communication-Efficient[[:space:]]Multi-Vehicle[[:space:]]Collaborative[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]via[[:space:]]Sparse[[:space:]]3D[[:space:]]Gaussian[[:space:]]Sharing/1bdcd8c8-01f5-4b11-af54-2b5905ac4120_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CompCap_[[:space:]]Improving[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]with[[:space:]]Composite[[:space:]]Captions/9144eafb-eab1-4ab4-9430-1ee22d78f0de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CompSlider_[[:space:]]Compositional[[:space:]]Slider[[:space:]]for[[:space:]]Disentangled[[:space:]]Multiple-Attribute[[:space:]]Image[[:space:]]Generation/6e927e9b-4dc8-411d-a600-f522f6c3b0b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Competitive[[:space:]]Distillation_[[:space:]]A[[:space:]]Simple[[:space:]]Learning[[:space:]]Strategy[[:space:]]for[[:space:]]Improving[[:space:]]Visual[[:space:]]Classification/7edd4b92-6bcb-44f6-aedd-50a7f775a8f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CompleteMe_[[:space:]]Reference-based[[:space:]]Human[[:space:]]Image[[:space:]]Completion/48f487bc-d6ba-4172-a816-ddc8f73c657e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Completing[[:space:]]3D[[:space:]]Partial[[:space:]]Assemblies[[:space:]]with[[:space:]]View-Consistent[[:space:]]2D-3D[[:space:]]Correspondence/f3e32238-94f1-4410-9b3e-2207645cb98e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Compression[[:space:]]of[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]Optimized[[:space:]]Feature[[:space:]]Planes[[:space:]]and[[:space:]]Standard[[:space:]]Video[[:space:]]Codecs/2d24bd81-d5d6-4917-9048-1e609445c46b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Compression-Aware[[:space:]]One-Step[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]JPEG[[:space:]]Artifact[[:space:]]Removal/29e1b228-c5dc-44a0-a231-d496a91f81c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ConceptSplit_[[:space:]]Decoupled[[:space:]]Multi-Concept[[:space:]]Personalization[[:space:]]of[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Token-wise[[:space:]]Adaptation[[:space:]]and[[:space:]]Attention[[:space:]]Disentanglement/db120673-dd6b-4879-bb03-f9754fc69950_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Conditional[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Instance[[:space:]]Segmentation/6be8cade-4696-4221-9da9-991d224be160_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Conditional[[:space:]]Visual[[:space:]]Autoregressive[[:space:]]Modeling[[:space:]]for[[:space:]]Pathological[[:space:]]Image[[:space:]]Restoration/6d516831-0b88-499e-8ab7-e81bb3fbc5d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ConformalSAM_[[:space:]]Unlocking[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Foundational[[:space:]]Segmentation[[:space:]]Models[[:space:]]in[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Conformal[[:space:]]Prediction/f4741a74-9fb1-477a-b3f1-e773ceaae20f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Confound[[:space:]]from[[:space:]]All[[:space:]]Sides,[[:space:]]Distill[[:space:]]with[[:space:]]Resilience_[[:space:]]Multi-Objective[[:space:]]Adversarial[[:space:]]Paths[[:space:]]to[[:space:]]Zero-Shot[[:space:]]Robustness/078da4a2-a297-4fc7-9fb9-055101f70a35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ConsNoTrainLoRA_[[:space:]]Data-driven[[:space:]]Weight[[:space:]]Initialization[[:space:]]of[[:space:]]Low-rank[[:space:]]Adapters[[:space:]]using[[:space:]]Constraints/e3986096-fe50-4286-94ac-ff20c4e48086_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Consensus-Driven[[:space:]]Active[[:space:]]Model[[:space:]]Selection/a42b80a6-c172-4162-88e6-15c929e4eff4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Consistency[[:space:]]Trajectory[[:space:]]Matching[[:space:]]for[[:space:]]One-Step[[:space:]]Generative[[:space:]]Super-Resolution/60877472-e403-41ef-bf95-62e6bb9810c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Consistent[[:space:]]Time-of-Flight[[:space:]]Depth[[:space:]]Denoising[[:space:]]via[[:space:]]Graph-Informed[[:space:]]Geometric[[:space:]]Attention/8cd454c8-613d-4d8e-9cd7-c6fe2282285b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ConsistentCity_[[:space:]]Semantic[[:space:]]Flow-guided[[:space:]]Occupancy[[:space:]]DiT[[:space:]]for[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Driving[[:space:]]Scene[[:space:]]Synthesis/02ce02d4-cdad-4b2a-87b8-59c471b95058_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ConstStyle_[[:space:]]Robust[[:space:]]Domain[[:space:]]Generalization[[:space:]]with[[:space:]]Unified[[:space:]]Style[[:space:]]Transformation/6380018f-e123-4d20-aea5-50f62a92b263_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Constraint-Aware[[:space:]]Feature[[:space:]]Learning[[:space:]]for[[:space:]]Parametric[[:space:]]Point[[:space:]]Cloud/db4a3021-4b5d-4c8c-98d9-d23872bf7fd1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Constructing[[:space:]]Ophthalmic[[:space:]]MLLM[[:space:]]for[[:space:]]Positioning-diagnosis[[:space:]]Collaboration[[:space:]]Through[[:space:]]Clinical[[:space:]]Cognitive[[:space:]]Chain[[:space:]]Reasoning/88f1fbad-45ce-459c-94f3-b9a49b23cfc0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Contact-Aware[[:space:]]Amodal[[:space:]]Completion[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]via[[:space:]]Multi-Regional[[:space:]]Inpainting/637c15ef-632c-4152-9de9-b4adfa6bce41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Contact-Aware[[:space:]]Refinement[[:space:]]of[[:space:]]Human[[:space:]]Pose[[:space:]]Pseudo-Ground[[:space:]]Truth[[:space:]]via[[:space:]]Bioimpedance[[:space:]]Sensing/3dc9842e-c9a9-460a-8c32-203c12701927_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Context[[:space:]]Guided[[:space:]]Transformer[[:space:]]Entropy[[:space:]]Modeling[[:space:]]for[[:space:]]Video[[:space:]]Compression/a59b85b4-d17d-4619-bc0f-a484d87974cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Context-Aware[[:space:]]Academic[[:space:]]Emotion[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark/2e82cdd6-fd10-46f0-817d-16d027554fbd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ContextFace_[[:space:]]Generating[[:space:]]Facial[[:space:]]Expressions[[:space:]]from[[:space:]]Emotional[[:space:]]Contexts/ffc3d83c-8e85-4a2c-a2ba-afb568ed281e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Continual[[:space:]]Adaptation_[[:space:]]Environment-Conditional[[:space:]]Parameter[[:space:]]Generation[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]Dynamic[[:space:]]Scenarios/9cb0607a-aee5-447c-acde-eb5e4471c48e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Continual[[:space:]]Multiple[[:space:]]Instance[[:space:]]Learning[[:space:]]with[[:space:]]Enhanced[[:space:]]Localization[[:space:]]for[[:space:]]Histopathological[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Analysis/ef1071e7-781c-4b0e-b934-d8fa68077e7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Continual[[:space:]]Personalization[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/d282cc59-75ed-4370-99b6-389edf0c4c22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Continuous-Time[[:space:]]Human[[:space:]]Motion[[:space:]]Field[[:space:]]from[[:space:]]Event[[:space:]]Cameras/7a91838a-a867-4bdc-b7c6-0f51a1680158_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ContraGS_[[:space:]]Codebook-Condensed[[:space:]]and[[:space:]]Trainable[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Fast,[[:space:]]Memory-Efficient[[:space:]]Reconstruction/e32d558a-2b7f-4f2d-9bef-fe3aa2a692cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Contrastive[[:space:]]Flow[[:space:]]Matching/fe76106c-6cb2-4f7c-9f22-aed1ed1f34c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Contrastive[[:space:]]Test-Time[[:space:]]Composition[[:space:]]of[[:space:]]Multiple[[:space:]]LoRA[[:space:]]Models[[:space:]]for[[:space:]]Image[[:space:]]Generation/7aa7234b-9e0b-4189-9a03-2afaf17db967_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Controllable[[:space:]]3D[[:space:]]Outdoor[[:space:]]Scene[[:space:]]Generation[[:space:]]via[[:space:]]Scene[[:space:]]Graphs/006afe9f-026d-4e7f-b9d5-ebf9c644d000_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Controllable[[:space:]]Feature[[:space:]]Whitening[[:space:]]for[[:space:]]Hyperparameter-Free[[:space:]]Bias[[:space:]]Mitigation/069a2e2e-f955-49bf-8c7c-67d2751667f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Controllable[[:space:]]Latent[[:space:]]Space[[:space:]]Augmentation[[:space:]]for[[:space:]]Digital[[:space:]]Pathology/ba75370a-da2c-4877-b99a-2f690d48c463_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Controllable[[:space:]]Weather[[:space:]]Synthesis[[:space:]]and[[:space:]]Removal[[:space:]]with[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/a3af6e67-6402-42d0-9ec6-93d0bdc8798f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Controllable[[:space:]]and[[:space:]]Expressive[[:space:]]One-Shot[[:space:]]Video[[:space:]]Head[[:space:]]Swapping/e1cdd15f-2a8e-41be-b516-ccc33052b45c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Controllable-LPMoE_[[:space:]]Adapting[[:space:]]to[[:space:]]Challenging[[:space:]]Object[[:space:]]Segmentation[[:space:]]via[[:space:]]Dynamic[[:space:]]Local[[:space:]]Priors[[:space:]]from[[:space:]]Mixture-of-Experts/9e5ca2d8-256f-432a-9860-b99ac4cb96d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Controlling[[:space:]]Multimodal[[:space:]]LLMs[[:space:]]via[[:space:]]Reward-guided[[:space:]]Decoding/52c21544-e937-446c-bddc-cacfff212d36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoopTrack_[[:space:]]Exploring[[:space:]]End-to-End[[:space:]]Learning[[:space:]]for[[:space:]]Efficient[[:space:]]Cooperative[[:space:]]Sequential[[:space:]]Perception/6ee2a685-9968-4606-b86c-d084639af6d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cooperative[[:space:]]Pseudo[[:space:]]Labeling[[:space:]]for[[:space:]]Unsupervised[[:space:]]Federated[[:space:]]Classification/4d38cfb6-5ece-4d9e-b32d-8d486de04e9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Coordinate-based[[:space:]]Speed[[:space:]]of[[:space:]]Sound[[:space:]]Recovery[[:space:]]for[[:space:]]Aberration-Corrected[[:space:]]Photoacoustic[[:space:]]Computed[[:space:]]Tomography/dedf6ed0-a505-4ae7-af6d-d32083550a65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CopyrightShield_[[:space:]]Enhancing[[:space:]]Diffusion[[:space:]]Model[[:space:]]Security[[:space:]]Against[[:space:]]Copyright[[:space:]]Infringement[[:space:]]Attacks/03df74aa-06ca-4e8a-9fc3-59487938e4e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CoralSRT_[[:space:]]Revisiting[[:space:]]Coral[[:space:]]Reef[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]by[[:space:]]Feature[[:space:]]Rectification[[:space:]]via[[:space:]]Self-supervised[[:space:]]Guidance/cf660690-3462-405a-8c0e-f28990c811c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CorrCLIP_[[:space:]]Reconstructing[[:space:]]Patch[[:space:]]Correlations[[:space:]]in[[:space:]]CLIP[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/ed70e9ef-666b-4552-8bad-d930b13c9326_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Correspondence[[:space:]]as[[:space:]]Video_[[:space:]]Test-Time[[:space:]]Adaption[[:space:]]on[[:space:]]SAM2[[:space:]]for[[:space:]]Reference[[:space:]]Segmentation[[:space:]]in[[:space:]]the[[:space:]]Wild/a60ca03f-5cc5-4dc5-b6b2-7a87536bc2f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Correspondence-Free[[:space:]]Fast[[:space:]]and[[:space:]]Robust[[:space:]]Spherical[[:space:]]Point[[:space:]]Pattern[[:space:]]Registration/8ff0eadf-9f55-46dc-afdb-87c241f9e402_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Corvid_[[:space:]]Improving[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]Towards[[:space:]]Chain-of-Thought[[:space:]]Reasoning/0686a13d-b3a4-4b40-a4e8-a0ae8c7b8e50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CountSE_[[:space:]]Soft[[:space:]]Exemplar[[:space:]]Open-set[[:space:]]Object[[:space:]]Counting/0a7f1ca2-b8d6-422b-a07c-f80f69eb0c3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CounterPC_[[:space:]]Counterfactual[[:space:]]Feature[[:space:]]Realignment[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]on[[:space:]]Point[[:space:]]Clouds/2a42de9d-3478-4bb6-aef0-92ef5b9afac1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Counting[[:space:]]Stacked[[:space:]]Objects/4c2ce1ea-e450-4fb3-b691-e764b0ce759b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Coupling[[:space:]]the[[:space:]]Generator[[:space:]]with[[:space:]]Teacher[[:space:]]for[[:space:]]Effective[[:space:]]Data-Free[[:space:]]Knowledge[[:space:]]Distillation/72e586f7-07f9-415b-a025-837fa4e3696c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cracking[[:space:]]Instance[[:space:]]Jigsaw[[:space:]]Puzzles_[[:space:]]An[[:space:]]Alternative[[:space:]]to[[:space:]]Multiple[[:space:]]Instance[[:space:]]Learning[[:space:]]for[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Analysis/325830e1-17b9-49e5-b8be-0acce3a5a0bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CreatiLayout_[[:space:]]Siamese[[:space:]]Multimodal[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]for[[:space:]]Creative[[:space:]]Layout-to-Image[[:space:]]Generation/5120d911-005d-44cf-b3f0-14f6b0bbe550_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Creation-MMBench_[[:space:]]Assessing[[:space:]]Context-Aware[[:space:]]Creative[[:space:]]Intelligence[[:space:]]in[[:space:]]MLLMs/0d642316-e00d-4024-9a89-575c5f5d3527_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-Architecture[[:space:]]Distillation[[:space:]]Made[[:space:]]Simple[[:space:]]with[[:space:]]Redundancy[[:space:]]Suppression/3474b7ea-a6cd-4afa-b8d2-c4973693160f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-Category[[:space:]]Subjectivity[[:space:]]Generalization[[:space:]]for[[:space:]]Style-Adaptive[[:space:]]Sketch[[:space:]]Re-ID/bfd99126-cd74-44e7-a5a3-9d52c6a192b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-Granularity[[:space:]]Online[[:space:]]Optimization[[:space:]]with[[:space:]]Masked[[:space:]]Compensated[[:space:]]Information[[:space:]]for[[:space:]]Learned[[:space:]]Image[[:space:]]Compression/14d5a967-e16d-4509-b6a4-c22c3589af09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-Subject[[:space:]]Mind[[:space:]]Decoding[[:space:]]from[[:space:]]Inaccurate[[:space:]]Representations/c353d062-35fd-4897-a827-c725aaca5f89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-View[[:space:]]Isolated[[:space:]]Sign[[:space:]]Language[[:space:]]Recognition[[:space:]]via[[:space:]]View[[:space:]]Synthesis[[:space:]]and[[:space:]]Feature[[:space:]]Disentanglement/dda8b634-b995-49f6-871e-05b20f559536_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cross-modal[[:space:]]Ship[[:space:]]Re-Identification[[:space:]]via[[:space:]]Optical[[:space:]]and[[:space:]]SAR[[:space:]]Imagery_[[:space:]]A[[:space:]]Novel[[:space:]]Dataset[[:space:]]and[[:space:]]Method/4c9beb8e-ce40-4747-b930-49bee6162e3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CryoFastAR_[[:space:]]Fast[[:space:]]Cryo-EM[[:space:]]Ab[[:space:]]initio[[:space:]]Reconstruction[[:space:]]Made[[:space:]]Easy/9c884879-23a7-4204-b0f8-31093d7735db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CuMPerLay_[[:space:]]Learning[[:space:]]Cubical[[:space:]]Multiparameter[[:space:]]Persistence[[:space:]]Vectorizations/35d41ae2-a76e-4da4-8d5e-0cfeaad361a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CuRe_[[:space:]]Cultural[[:space:]]Gaps[[:space:]]in[[:space:]]the[[:space:]]Long[[:space:]]Tail[[:space:]]of[[:space:]]Text-to-Image[[:space:]]Systems/c5641cba-dfa5-4f2e-85d8-227cd2d88b4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Curve-Aware[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]3D[[:space:]]Parametric[[:space:]]Curve[[:space:]]Reconstruction/4a7c44f5-4152-4b10-a4be-5929ea45ca4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Customizing[[:space:]]Domain[[:space:]]Adapters[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/b17205a4-d28a-4995-a916-c7405aa49532_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CutS3D_[[:space:]]Cutting[[:space:]]Semantics[[:space:]]in[[:space:]]3D[[:space:]]for[[:space:]]2D[[:space:]]Unsupervised[[:space:]]Instance[[:space:]]Segmentation/67adb151-9308-4470-9d93-3e5e631d1aa9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cycle[[:space:]]Consistency[[:space:]]as[[:space:]]Reward_[[:space:]]Learning[[:space:]]Image-Text[[:space:]]Alignment[[:space:]]without[[:space:]]Human[[:space:]]Preferences/761eb42f-812e-4ae1-91c6-5f05dbe6998b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Cycle-Consistent[[:space:]]Learning[[:space:]]for[[:space:]]Joint[[:space:]]Layout-to-Image[[:space:]]Generation[[:space:]]and[[:space:]]Object[[:space:]]Detection/2baf9144-26c7-48b6-a5aa-20187b3fdab1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/CycleVAR_[[:space:]]Repurposing[[:space:]]Autoregressive[[:space:]]Model[[:space:]]for[[:space:]]Unsupervised[[:space:]]One-Step[[:space:]]Image[[:space:]]Translation/ee9e09b9-697a-4ddd-a6cb-9652e026d797_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/D-Attn_[[:space:]]Decomposed[[:space:]]Attention[[:space:]]for[[:space:]]Large[[:space:]]Vision-and-Language[[:space:]]Model/0a764c73-a275-43de-945f-df274ebe7798_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/D2ST-Adapter_[[:space:]]Disentangled-and-Deformable[[:space:]]Spatio-Temporal[[:space:]]Adapter[[:space:]]for[[:space:]]Few-shot[[:space:]]Action[[:space:]]Recognition/7ac1f9a3-50e7-4861-922e-b4f4818cac99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/D3QE_[[:space:]]Learning[[:space:]]Discrete[[:space:]]Distribution[[:space:]]Discrepancy-aware[[:space:]]Quantization[[:space:]]Error[[:space:]]for[[:space:]]Autoregressive-Generated[[:space:]]Image[[:space:]]Detection/6ba5a93e-816f-4d1c-a7c8-c05eb9b94d43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/D3_[[:space:]]Training-Free[[:space:]]AI-Generated[[:space:]]Video[[:space:]]Detection[[:space:]]Using[[:space:]]Second-Order[[:space:]]Features/44869553-3034-4013-b0ac-e64335af5ffa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DAA__[[:space:]]Deep[[:space:]]Angular[[:space:]]A[[:space:]]Star[[:space:]]for[[:space:]]Image-based[[:space:]]Path[[:space:]]Planning/f7d3bc1b-569c-4e3e-a742-162e3131d2e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DACoN_[[:space:]]DINO[[:space:]]for[[:space:]]Anime[[:space:]]Paint[[:space:]]Bucket[[:space:]]Colorization[[:space:]]with[[:space:]]Any[[:space:]]Number[[:space:]]of[[:space:]]Reference[[:space:]]Images/ad36cda1-152f-41dd-b893-90d3bd037ae8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DADM_[[:space:]]Dual[[:space:]]Alignment[[:space:]]of[[:space:]]Domain[[:space:]]and[[:space:]]Modality[[:space:]]for[[:space:]]Face[[:space:]]Anti-spoofing/a152821d-8158-44f7-a1b0-e8344d0fc6a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DADet_[[:space:]]Safeguarding[[:space:]]Image[[:space:]]Conditional[[:space:]]Diffusion[[:space:]]Models[[:space:]]against[[:space:]]Adversarial[[:space:]]and[[:space:]]Backdoor[[:space:]]Attacks[[:space:]]via[[:space:]]Diffusion[[:space:]]Anomaly[[:space:]]Detection/accb3dbe-4d68-4518-aa7c-62916b4b1d24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DALIP_[[:space:]]Distribution[[:space:]]Alignment-based[[:space:]]Language-Image[[:space:]]Pre-Training[[:space:]]for[[:space:]]Domain-Specific[[:space:]]Data/957a9523-30dd-4796-9c4c-63b40d8b16c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DAMap_[[:space:]]Distance-aware[[:space:]]MapNet[[:space:]]for[[:space:]]High[[:space:]]Quality[[:space:]]HD[[:space:]]Map[[:space:]]Construction/a5998075-1ad1-4690-be74-280051230454_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DAP-MAE_[[:space:]]Domain-Adaptive[[:space:]]Point[[:space:]]Cloud[[:space:]]Masked[[:space:]]Autoencoder[[:space:]]for[[:space:]]Effective[[:space:]]Cross-Domain[[:space:]]Learning/e93e1d83-5fac-4ec5-b34e-b9ef9e7b5e49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DASH_[[:space:]]4D[[:space:]]Hash[[:space:]]Encoding[[:space:]]with[[:space:]]Self-Supervised[[:space:]]Decomposition[[:space:]]for[[:space:]]Real-Time[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Rendering/20f00e52-bd77-4ab4-9290-b1f942558f93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DASH_[[:space:]]Detection[[:space:]]and[[:space:]]Assessment[[:space:]]of[[:space:]]Systematic[[:space:]]Hallucinations[[:space:]]of[[:space:]]VLMs/06c09e12-7e45-4cf7-b6b2-47081cc85938_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DATA_[[:space:]]Domain-And-Time[[:space:]]Alignment[[:space:]]for[[:space:]]High-Quality[[:space:]]Feature[[:space:]]Fusion[[:space:]]in[[:space:]]Collaborative[[:space:]]Perception/d082a012-3cb6-456f-99de-7017ee6d741c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DAViD_[[:space:]]Data-efficient[[:space:]]and[[:space:]]Accurate[[:space:]]Vision[[:space:]]Models[[:space:]]from[[:space:]]Synthetic[[:space:]]Data/0768eaf0-9ada-46e3-bc68-14a6ab335cc0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DAViD_[[:space:]]Modeling[[:space:]]Dynamic[[:space:]]Affordance[[:space:]]of[[:space:]]3D[[:space:]]Objects[[:space:]]Using[[:space:]]Pre-trained[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/66fb0948-395f-4309-8141-d14d71c2e21f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DC-AE[[:space:]]1.5_[[:space:]]Accelerating[[:space:]]Diffusion[[:space:]]Model[[:space:]]Convergence[[:space:]]with[[:space:]]Structured[[:space:]]Latent[[:space:]]Space/0991fd05-9d31-45e3-bf07-da8796e53db1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DC-AR_[[:space:]]Efficient[[:space:]]Masked[[:space:]]Autoregressive[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Deep[[:space:]]Compression[[:space:]]Hybrid[[:space:]]Tokenizer/71210c13-d799-49a5-b5d5-5666f3d17606_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DC-ControlNet_[[:space:]]Decoupling[[:space:]]Inter-[[:space:]]and[[:space:]]Intra-Element[[:space:]]Conditions[[:space:]]in[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/2838c612-50db-4a08-98ef-ff4c38d5c3cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DC-TTA_[[:space:]]Divide-and-Conquer[[:space:]]Framework[[:space:]]for[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]of[[:space:]]Interactive[[:space:]]Segmentation/9fb1aeee-c479-4d77-ad48-b66821ef44b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DCHM_[[:space:]]Depth-Consistent[[:space:]]Human[[:space:]]Modeling[[:space:]]for[[:space:]]Multiview[[:space:]]Detection/d761d034-3352-43ff-884d-a1e0b9c1d421_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DCT-Shield_[[:space:]]A[[:space:]]Robust[[:space:]]Frequency[[:space:]]Domain[[:space:]]Defense[[:space:]]against[[:space:]]Malicious[[:space:]]Image[[:space:]]Editing/9e25e5d0-f57f-4eb5-afa9-93d415d14faf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DDB_[[:space:]]Diffusion[[:space:]]Driven[[:space:]]Balancing[[:space:]]to[[:space:]]Address[[:space:]]Spurious[[:space:]]Correlations/4d22ee71-1442-48ff-8043-06cb0e7ecb68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DEPTHOR_[[:space:]]Depth[[:space:]]Enhancement[[:space:]]from[[:space:]]a[[:space:]]Practical[[:space:]]Light-Weight[[:space:]]dToF[[:space:]]Sensor[[:space:]]and[[:space:]]RGB[[:space:]]Image/b16b685c-4093-41fd-895a-9b0f47649f76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DGTalker_[[:space:]]Disentangled[[:space:]]Generative[[:space:]]Latent[[:space:]]Space[[:space:]]Learning[[:space:]]for[[:space:]]Audio-Driven[[:space:]]Gaussian[[:space:]]Talking[[:space:]]Heads/e8451aa8-7b1e-45dd-b729-3c6f687b5e03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DH-FaceVid-1K_[[:space:]]A[[:space:]]Large-Scale[[:space:]]High-Quality[[:space:]]Dataset[[:space:]]for[[:space:]]Face[[:space:]]Video[[:space:]]Generation/0a1704de-785d-4f54-98c9-d000e64f4572_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DIA_[[:space:]]The[[:space:]]Adversarial[[:space:]]Exposure[[:space:]]of[[:space:]]Deterministic[[:space:]]Inversion[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/87fcf2e9-ecdf-4417-b926-f0defbeffa58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DICE_[[:space:]]Staleness-Centric[[:space:]]Optimizations[[:space:]]for[[:space:]]Parallel[[:space:]]Diffusion[[:space:]]MoE[[:space:]]Inference/384b4045-2728-4623-9480-43ce818a5fd7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DIH-CLIP_[[:space:]]Unleashing[[:space:]]the[[:space:]]Diversity[[:space:]]of[[:space:]]Multi-Head[[:space:]]Self-Attention[[:space:]]for[[:space:]]Training-Free[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/8ce9588a-57d3-427a-a627-ec2a5739fa70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DIMCIM_[[:space:]]A[[:space:]]Quantitative[[:space:]]Evaluation[[:space:]]Framework[[:space:]]for[[:space:]]Default-mode[[:space:]]Diversity[[:space:]]and[[:space:]]Generalization[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Generative[[:space:]]Models/ceddef4b-e066-4bf1-a6c0-e24a954105f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DIMO_[[:space:]]Diverse[[:space:]]3D[[:space:]]Motion[[:space:]]Generation[[:space:]]for[[:space:]]Arbitrary[[:space:]]Objects/96ba5a43-d293-4513-9ee3-1255210620f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DIP_[[:space:]]Unsupervised[[:space:]]Dense[[:space:]]In-Context[[:space:]]Post-training[[:space:]]of[[:space:]]Visual[[:space:]]Representations/9c23df4f-1db2-4cd9-b244-86d951097422_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DISTA-Net_[[:space:]]Dynamic[[:space:]]Closely-Spaced[[:space:]]Infrared[[:space:]]Small[[:space:]]Target[[:space:]]Unmixing/90a69e89-5319-4b34-b605-32205c6ee9cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DISTIL_[[:space:]]Data-Free[[:space:]]Inversion[[:space:]]of[[:space:]]Suspicious[[:space:]]Trojan[[:space:]]Inputs[[:space:]]via[[:space:]]Latent[[:space:]]Diffusion/7fdc7b3b-aacf-432b-956b-737ad73f8e6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DIVE_[[:space:]]Taming[[:space:]]DINO[[:space:]]for[[:space:]]Subject-Driven[[:space:]]Video[[:space:]]Editing/218ccbea-6160-46a8-9ae5-8f67af89212d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DLFR-Gen_[[:space:]]Diffusion-based[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Dynamic[[:space:]]Latent[[:space:]]Frame[[:space:]]Rate/86307acb-6600-4537-8f9d-6bb3cbd20c3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DLF_[[:space:]]Extreme[[:space:]]Image[[:space:]]Compression[[:space:]]with[[:space:]]Dual-generative[[:space:]]Latent[[:space:]]Fusion/ef70f953-459b-4060-856d-d7799d3c2c84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DM-EFS_[[:space:]]Dynamically[[:space:]]Multiplexed[[:space:]]Expanded[[:space:]]Features[[:space:]]Set[[:space:]]Form[[:space:]]for[[:space:]]Robust[[:space:]]and[[:space:]]Efficient[[:space:]]Small[[:space:]]Object[[:space:]]Detection/34ecd42f-8f06-48c5-beae-8dec1b262a40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DMQ_[[:space:]]Dissecting[[:space:]]Outliers[[:space:]]of[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Post-Training[[:space:]]Quantization/c7b6552b-ad1f-4d7f-ada2-5c3bb748b2a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DMesh++_[[:space:]]An[[:space:]]Efficient[[:space:]]Differentiable[[:space:]]Mesh[[:space:]]for[[:space:]]Complex[[:space:]]Shapes/79f0bb8d-2945-451a-912a-101f0556b5d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DNF-Intrinsic_[[:space:]]Deterministic[[:space:]]Noise-Free[[:space:]]Diffusion[[:space:]]for[[:space:]]Indoor[[:space:]]Inverse[[:space:]]Rendering/9e03e1e9-511d-450c-98fb-af4a0198c135_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DOGR_[[:space:]]Towards[[:space:]]Versatile[[:space:]]Visual[[:space:]]Document[[:space:]]Grounding[[:space:]]and[[:space:]]Referring/452224d7-b239-490c-aaca-689e047fc504_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DOLLAR_[[:space:]]Few-Step[[:space:]]Video[[:space:]]Generation[[:space:]]via[[:space:]]Distillation[[:space:]]and[[:space:]]Latent[[:space:]]Reward[[:space:]]Optimization/8c8c316b-707c-44fe-a2b5-57a231b48959_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DONUT_[[:space:]]A[[:space:]]Decoder-Only[[:space:]]Model[[:space:]]for[[:space:]]Trajectory[[:space:]]Prediction/ba39018c-a30a-49ed-b7ec-093f8d6feb81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DPoser-X_[[:space:]]Diffusion[[:space:]]Model[[:space:]]as[[:space:]]Robust[[:space:]]3D[[:space:]]Whole-body[[:space:]]Human[[:space:]]Pose[[:space:]]Prior/0773b01c-f286-4aa6-80a2-62d8ebb219a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DRaM-LHM_[[:space:]]A[[:space:]]Quaternion[[:space:]]Framework[[:space:]]for[[:space:]]Iterative[[:space:]]Camera[[:space:]]Pose[[:space:]]Estimation/25a48c08-15e9-460b-80c6-d12abf74c8c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DSO_[[:space:]]Aligning[[:space:]]3D[[:space:]]Generators[[:space:]]with[[:space:]]Simulation[[:space:]]Feedback[[:space:]]for[[:space:]]Physical[[:space:]]Soundness/ae0feb1f-29f1-43a4-9436-3fa7c1d0bbec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DWIM_[[:space:]]Towards[[:space:]]Tool-aware[[:space:]]Visual[[:space:]]Reasoning[[:space:]]via[[:space:]]Discrepancy-aware[[:space:]]Workflow[[:space:]]Generation[[:space:]]&[[:space:]]Instruct-Masking[[:space:]]Tuning/666968c6-d5c2-43ef-9a89-146705cb3bef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DanceEditor_[[:space:]]Towards[[:space:]]Iterative[[:space:]]Editable[[:space:]]Music-driven[[:space:]]Dance[[:space:]]Generation[[:space:]]with[[:space:]]Open-Vocabulary[[:space:]]Descriptions/32861b19-b480-4386-b034-7866c8891681_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dark-ISP_[[:space:]]Enhancing[[:space:]]RAW[[:space:]]Image[[:space:]]Processing[[:space:]]for[[:space:]]Low-Light[[:space:]]Object[[:space:]]Detection/53732aa9-4ce6-4724-9a1e-440f63e8f1c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dataset[[:space:]]Distillation[[:space:]]as[[:space:]]Data[[:space:]]Compression_[[:space:]]A[[:space:]]Rate-Utility[[:space:]]Perspective/a56d7f84-a4be-4b58-8c18-2467956d47b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dataset[[:space:]]Distillation[[:space:]]via[[:space:]]Vision-Language[[:space:]]Category[[:space:]]Prototype/0d61e2e9-b5e8-494a-9ab1-f32fd8d336f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dataset[[:space:]]Distillation[[:space:]]via[[:space:]]the[[:space:]]Wasserstein[[:space:]]Metric/1ea23447-aeea-4511-8d18-83e52d179e11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dataset[[:space:]]Ownership[[:space:]]Verification[[:space:]]for[[:space:]]Pre-trained[[:space:]]Masked[[:space:]]Models/2cd3aeb9-4f68-46cb-b4da-29836e99ef28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeFSS_[[:space:]]Image-to-Mask[[:space:]]Denoising[[:space:]]Learning[[:space:]]for[[:space:]]Few-shot[[:space:]]Segmentation/a24f8032-6732-430c-a125-c1597af275b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeGauss_[[:space:]]Dynamic-Static[[:space:]]Decomposition[[:space:]]with[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Distractor-free[[:space:]]3D[[:space:]]Reconstruction/cec23c2e-8aea-4df8-9556-2f50c6afd012_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeRIS_[[:space:]]Decoupling[[:space:]]Perception[[:space:]]and[[:space:]]Cognition[[:space:]]for[[:space:]]Enhanced[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation[[:space:]]through[[:space:]]Loopback[[:space:]]Synergy/dd68472e-6167-4a30-bff3-eac7ed803f26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeSPITE_[[:space:]]Exploring[[:space:]]Contrastive[[:space:]]Deep[[:space:]]Skeleton-Pointcloud-IMU-Text[[:space:]]Embeddings[[:space:]]for[[:space:]]Advanced[[:space:]]Point[[:space:]]Cloud[[:space:]]Human[[:space:]]Activity[[:space:]]Understanding/2e8cafb2-bf41-4753-87cf-a03b6cdbe540_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Debiased[[:space:]]Curriculum[[:space:]]Adaptation[[:space:]]for[[:space:]]Safe[[:space:]]Transfer[[:space:]]Learning[[:space:]]in[[:space:]]Chest[[:space:]]X-ray[[:space:]]Classification/cff09aa9-b77f-4825-99e3-56a7fc497f44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Debiased[[:space:]]Teacher[[:space:]]for[[:space:]]Day-to-Night[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detection/e402ec66-cbec-43bc-8466-743fc63c73f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Debiasing[[:space:]]Trace[[:space:]]Guidance_[[:space:]]Top-down[[:space:]]Trace[[:space:]]Distillation[[:space:]]and[[:space:]]Bottom-up[[:space:]]Velocity[[:space:]]Alignment[[:space:]]for[[:space:]]Unsupervised[[:space:]]Anomaly[[:space:]]Detection/8ff214ba-deb5-41c4-89cc-d4b1d92ed6fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DecAD_[[:space:]]Decoupling[[:space:]]Anomalies[[:space:]]in[[:space:]]Latent[[:space:]]Space[[:space:]]for[[:space:]]Multi-Class[[:space:]]Unsupervised[[:space:]]Anomaly[[:space:]]Detection/104e643a-4201-4913-9238-e5645798c926_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Deciphering[[:space:]]Cross-Modal[[:space:]]Alignment[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]via[[:space:]]Modality[[:space:]]Integration[[:space:]]Rate/eee0e46b-2d71-4415-93c1-bd5422e68ac8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decoding[[:space:]]Correlation-Induced[[:space:]]Misalignment[[:space:]]in[[:space:]]the[[:space:]]Stable[[:space:]]Diffusion[[:space:]]Workflow[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/4700f8c0-cc0d-4188-8211-0ec1c4182bda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decouple[[:space:]]and[[:space:]]Track_[[:space:]]Benchmarking[[:space:]]and[[:space:]]Improving[[:space:]]Video[[:space:]]Diffusion[[:space:]]Transformers[[:space:]]For[[:space:]]Motion[[:space:]]Transfer/2815078c-1f0b-4015-9572-383046734683_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decouple[[:space:]]to[[:space:]]Reconstruct_[[:space:]]High[[:space:]]Quality[[:space:]]UHD[[:space:]]Restoration[[:space:]]via[[:space:]]Active[[:space:]]Feature[[:space:]]Disentanglement[[:space:]]and[[:space:]]Reversible[[:space:]]Fusion/d6f81016-06ea-4a39-af2a-5ad37aa766a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decoupled[[:space:]]Diffusion[[:space:]]Sparks[[:space:]]Adaptive[[:space:]]Scene[[:space:]]Generation/a55b328e-507c-4401-b176-9f66c33ecb98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Decoupled[[:space:]]Multi-Predictor[[:space:]]Optimization[[:space:]]for[[:space:]]Inference-Efficient[[:space:]]Model[[:space:]]Tuning/63a1968d-250c-4829-a5bd-f2aca13ca597_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Deep[[:space:]]Adaptive[[:space:]]Unfolded[[:space:]]Network[[:space:]]via[[:space:]]Spatial[[:space:]]Morphology[[:space:]]Stripping[[:space:]]and[[:space:]]Spectral[[:space:]]Filtration[[:space:]]for[[:space:]]Pan-sharpening/5f36f898-dc13-4c10-89d8-d6638590ec51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Deep[[:space:]]Incomplete[[:space:]]Multi-view[[:space:]]Clustering[[:space:]]with[[:space:]]Distribution[[:space:]]Dual-Consistency[[:space:]]Recovery[[:space:]]Guidance/42a07816-8432-4289-a30f-fa72e0de4e0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Deep[[:space:]]Space[[:space:]]Weather[[:space:]]Model_[[:space:]]Long-Range[[:space:]]Solar[[:space:]]Flare[[:space:]]Prediction[[:space:]]from[[:space:]]Multi-Wavelength[[:space:]]Images/3872e707-cc80-420b-98a9-bbbe814577aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeepMesh_[[:space:]]Auto-Regressive[[:space:]]Artist-mesh[[:space:]]Creation[[:space:]]with[[:space:]]Reinforcement[[:space:]]Learning/03994ed9-b2c2-491b-8ec3-39e554598987_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DeepShield_[[:space:]]Fortifying[[:space:]]Deepfake[[:space:]]Video[[:space:]]Detection[[:space:]]with[[:space:]]Local[[:space:]]and[[:space:]]Global[[:space:]]Forgery[[:space:]]Analysis/db469209-3bde-4f3e-ab52-26c1d2d3c65d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Deeply[[:space:]]Supervised[[:space:]]Flow-Based[[:space:]]Generative[[:space:]]Models/b6ef0e58-c67e-4811-a5b9-ba8562f322e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Degradation-Modeled[[:space:]]Multipath[[:space:]]Diffusion[[:space:]]for[[:space:]]Tunable[[:space:]]Metalens[[:space:]]Photography/43b9f660-f38e-43af-a859-5b73132a5c91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Demeter_[[:space:]]A[[:space:]]Parametric[[:space:]]Model[[:space:]]of[[:space:]]Crop[[:space:]]Plant[[:space:]]Morphology[[:space:]]from[[:space:]]the[[:space:]]Real[[:space:]]World/9d783bdb-685b-4e85-a09a-0b1960845eca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Democratizing[[:space:]]High-Fidelity[[:space:]]Co-Speech[[:space:]]Gesture[[:space:]]Video[[:space:]]Generation/a09a180d-9e43-4dde-b2fe-0fddaab99e29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Democratizing[[:space:]]Text-to-Image[[:space:]]Masked[[:space:]]Generative[[:space:]]Models[[:space:]]with[[:space:]]Compact[[:space:]]Text-Aware[[:space:]]One-Dimensional[[:space:]]Tokens/cf274573-cb97-4f00-8692-23d9b3cb6c99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Denoising[[:space:]]Token[[:space:]]Prediction[[:space:]]in[[:space:]]Masked[[:space:]]Autoregressive[[:space:]]Models/d0e535b1-f484-4ed9-9e8d-a999c03b8173_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dense[[:space:]]Policy_[[:space:]]Bidirectional[[:space:]]Autoregressive[[:space:]]Learning[[:space:]]of[[:space:]]Actions/b9a1aa41-79e4-43aa-aaac-572fab1be80b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dense2MoE_[[:space:]]Restructuring[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]to[[:space:]]MoE[[:space:]]for[[:space:]]Efficient[[:space:]]Text-to-Image[[:space:]]Generation/bb139216-d038-4688-9871-abeab65afec5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DepR_[[:space:]]Depth[[:space:]]Guided[[:space:]]Single-view[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]with[[:space:]]Instance-level[[:space:]]Diffusion/d745c363-0e8a-4b1f-a301-3282a9e01a51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Depth[[:space:]]Any[[:space:]]Event[[:space:]]Stream_[[:space:]]Enhancing[[:space:]]Event-based[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation[[:space:]]via[[:space:]]Dense-to-Sparse[[:space:]]Distillation/610215de-66f1-4132-a99a-235cefc95936_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Depth[[:space:]]AnyEvent_[[:space:]]A[[:space:]]Cross-Modal[[:space:]]Distillation[[:space:]]Paradigm[[:space:]]for[[:space:]]Event-Based[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/2d4c0221-9085-4ca7-8830-5b6d895d651e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DepthSync_[[:space:]]Diffusion[[:space:]]Guidance-Based[[:space:]]Depth[[:space:]]Synchronization[[:space:]]for[[:space:]]Scale-[[:space:]]and[[:space:]]Geometry-Consistent[[:space:]]Video[[:space:]]Depth[[:space:]]Estimation/2039af04-bc2c-41ce-8d36-cb2a37980d9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Derm1M_[[:space:]]A[[:space:]]Million-scale[[:space:]]Vision-Language[[:space:]]Dataset[[:space:]]Aligned[[:space:]]with[[:space:]]Clinical[[:space:]]Ontology[[:space:]]Knowledge[[:space:]]for[[:space:]]Dermatology/78678cc5-e4b5-461a-a64e-187f14fcd82e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Describe[[:space:]]Anything_[[:space:]]Detailed[[:space:]]Localized[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Captioning/22bba40a-fe55-4504-a91b-c128c9aa04f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Describe,[[:space:]]Adapt[[:space:]]and[[:space:]]Combine_[[:space:]]Empowering[[:space:]]CLIP[[:space:]]Encoders[[:space:]]for[[:space:]]Open-set[[:space:]]3D[[:space:]]Object[[:space:]]Retrieval/65002a8e-8ea6-4fc9-91ae-9c9d5cbfa659_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Describe,[[:space:]]Don't[[:space:]]Dictate_[[:space:]]Semantic[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Natural[[:space:]]Language[[:space:]]Intent/dc73ea42-98ce-4d6e-addb-e496f9a3af2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Details[[:space:]]Matter[[:space:]]for[[:space:]]Indoor[[:space:]]Open-vocabulary[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation/67e092d0-e8c5-4832-a1ce-008636845608_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Detect[[:space:]]Anything[[:space:]]3D[[:space:]]in[[:space:]]the[[:space:]]Wild/4762f117-64e7-4388-a8fa-d09b2e391820_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Detection,[[:space:]]Pose[[:space:]]Estimation[[:space:]]and[[:space:]]Segmentation[[:space:]]for[[:space:]]Multiple[[:space:]]Bodies_[[:space:]]Closing[[:space:]]the[[:space:]]Virtuous[[:space:]]Circle/4e20bf48-4e6a-463f-8303-a96794389913_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Deterministic[[:space:]]Object[[:space:]]Pose[[:space:]]Confidence[[:space:]]Region[[:space:]]Estimation/a166218d-b136-4fa3-917b-d62e1ddd64ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Devil[[:space:]]is[[:space:]]in[[:space:]]the[[:space:]]Uniformity_[[:space:]]Exploring[[:space:]]Diverse[[:space:]]Learners[[:space:]]within[[:space:]]Transformer[[:space:]]for[[:space:]]Image[[:space:]]Restoration/d5b19a6b-d06c-4b74-b4b3-ff6b4fdaff87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DexH2R_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Dynamic[[:space:]]Dexterous[[:space:]]Grasping[[:space:]]in[[:space:]]Human-to-Robot[[:space:]]Handover/eaf43253-1e74-44bf-a9a7-ffaba468efb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DexVLG_[[:space:]]Dexterous[[:space:]]Vision-Language-Grasp[[:space:]]Model[[:space:]]at[[:space:]]Scale/11e44a18-e2f5-4f3d-b3e2-dfc8659c0a48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiGA3D_[[:space:]]Coarse-to-Fine[[:space:]]Diffusional[[:space:]]Propagation[[:space:]]of[[:space:]]Geometry[[:space:]]and[[:space:]]Appearance[[:space:]]for[[:space:]]Versatile[[:space:]]3D[[:space:]]Inpainting/498d4798-8b86-499c-a626-9d43d7a73b03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiMPLe[[:space:]]-[[:space:]]Disentangled[[:space:]]Multi-Modal[[:space:]]Prompt[[:space:]]Learning_[[:space:]]Enhancing[[:space:]]Out-Of-Distribution[[:space:]]Alignment[[:space:]]with[[:space:]]Invariant[[:space:]]and[[:space:]]Spurious[[:space:]]Feature[[:space:]]Separation/58bc21ac-b64b-45f0-96c1-230a3bc49ee0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiSCO-3D[[:space:]]_[[:space:]]Discovering[[:space:]]and[[:space:]]Segmenting[[:space:]]Sub-Concepts[[:space:]]from[[:space:]]Open-vocabulary[[:space:]]Queries[[:space:]]in[[:space:]]NeRF/3f492d47-496d-40e7-8aea-e91714e855d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiST-4D_[[:space:]]Disentangled[[:space:]]Spatiotemporal[[:space:]]Diffusion[[:space:]]with[[:space:]]Metric[[:space:]]Depth[[:space:]]for[[:space:]]4D[[:space:]]Driving[[:space:]]Scene[[:space:]]Generation/8d95d4c6-b7b6-483e-b8b3-c239e47e9450_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiT4SR_[[:space:]]Taming[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]for[[:space:]]Real-World[[:space:]]Image[[:space:]]Super-Resolution/b8efa601-bad9-438a-b107-f179a56716fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiTFastAttnV2_[[:space:]]Head-wise[[:space:]]Attention[[:space:]]Compression[[:space:]]for[[:space:]]Multi-Modality[[:space:]]Diffusion[[:space:]]Transformers/e07fce6f-af4f-4914-852e-b81ccf820cb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiTaiListener_[[:space:]]Controllable[[:space:]]High[[:space:]]Fidelity[[:space:]]Listener[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion/35986d1a-e199-4e8e-a54f-b7addb417e6b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Di\[M\]O_[[:space:]]Distilling[[:space:]]Masked[[:space:]]Diffusion[[:space:]]Models[[:space:]]into[[:space:]]One-step[[:space:]]Generator/ee441443-9ff0-4da0-9b7e-3c4104011296_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diagnosing[[:space:]]Pretrained[[:space:]]Models[[:space:]]for[[:space:]]Out-of-distribution[[:space:]]Detection/7eafcf92-7a17-4d2e-bf34-f17f5cc7166a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DialNav_[[:space:]]Multi-turn[[:space:]]Dialog[[:space:]]Navigation[[:space:]]with[[:space:]]a[[:space:]]Remote[[:space:]]Guide/2af070bb-8ab6-43b0-b054-a16b7e514e37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DictAS_[[:space:]]A[[:space:]]Framework[[:space:]]for[[:space:]]Class-Generalizable[[:space:]]Few-Shot[[:space:]]Anomaly[[:space:]]Segmentation[[:space:]]via[[:space:]]Dictionary[[:space:]]Lookup/0a675cd5-14cc-4fea-b516-0808ecff9020_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diff2I2P_[[:space:]]Differentiable[[:space:]]Image-to-Point[[:space:]]Cloud[[:space:]]Registration[[:space:]]with[[:space:]]Diffusion[[:space:]]Prior/4da3bc1d-8ea6-4690-8f52-52e1eca287f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffDoctor_[[:space:]]Diagnosing[[:space:]]Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]Before[[:space:]]Treating/1eda5e65-eff7-48fa-937a-3f8b1b2d17ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffIP_[[:space:]]Representation[[:space:]]Fingerprints[[:space:]]for[[:space:]]Robust[[:space:]]IP[[:space:]]Protection[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/9f4d2559-18a8-4771-81f5-7ca172e93dce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffPCI_[[:space:]]Large[[:space:]]Motion[[:space:]]Point[[:space:]]Cloud[[:space:]]frame[[:space:]]Interpolation[[:space:]]with[[:space:]]Diffusion[[:space:]]Model/24aa2a9e-f405-4999-b735-bb6f409c7685_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffRefine_[[:space:]]Diffusion-based[[:space:]]Proposal[[:space:]]Specific[[:space:]]Point[[:space:]]Cloud[[:space:]]Densification[[:space:]]for[[:space:]]Cross-Domain[[:space:]]Object[[:space:]]Detection/2bb10626-951e-411d-a7cd-c6a47471526d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffSim_[[:space:]]Taming[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Evaluating[[:space:]]Visual[[:space:]]Similarity/80f749ee-2036-4359-9086-5d9aada7e6cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffTell_[[:space:]]A[[:space:]]High-Quality[[:space:]]Dataset[[:space:]]for[[:space:]]Describing[[:space:]]Image[[:space:]]Manipulation[[:space:]]Changes/8d31a853-1f64-4f20-85b4-93981ccab4c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffVSR_[[:space:]]Revealing[[:space:]]an[[:space:]]Effective[[:space:]]Recipe[[:space:]]for[[:space:]]Taming[[:space:]]Robust[[:space:]]Video[[:space:]]Super-Resolution[[:space:]]Against[[:space:]]Complex[[:space:]]Degradations/3f948368-15b1-4f30-9f01-fa500afb68fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Differentiable[[:space:]]Room[[:space:]]Acoustic[[:space:]]Rendering[[:space:]]with[[:space:]]Multi-View[[:space:]]Vision[[:space:]]Priors/db0156cb-5a3e-4d41-9327-f1d790570a62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Differential-informed[[:space:]]Sample[[:space:]]Selection[[:space:]]Accelerates[[:space:]]Multimodal[[:space:]]Contrastive[[:space:]]Learning/3d012e8e-f5dc-4983-99d6-04b0d824a420_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Differentially[[:space:]]Private[[:space:]]Fine-Tuning[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/f598eaa6-f932-4fb3-b339-88814f617fb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DiffuMatch_[[:space:]]Category-Agnostic[[:space:]]Spectral[[:space:]]Diffusion[[:space:]]Priors[[:space:]]for[[:space:]]Robust[[:space:]]Non-rigid[[:space:]]Shape[[:space:]]Matching/209ad058-ffa0-40fb-8607-6f1307899070_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffuman4D_[[:space:]]4D[[:space:]]Consistent[[:space:]]Human[[:space:]]View[[:space:]]Synthesis[[:space:]]from[[:space:]]Sparse-View[[:space:]]Videos[[:space:]]with[[:space:]]Spatio-Temporal[[:space:]]Diffusion[[:space:]]Models/e4a5a744-704a-4f99-bd71-4f3bef1da50e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion[[:space:]]Curriculum_[[:space:]]Synthetic-to-Real[[:space:]]Data[[:space:]]Curriculum[[:space:]]via[[:space:]]Image-Guided[[:space:]]Diffusion/0c7041a1-3d70-4e9f-82b7-fbde4cbee548_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion[[:space:]]Epistemic[[:space:]]Uncertainty[[:space:]]with[[:space:]]Asymmetric[[:space:]]Learning[[:space:]]for[[:space:]]Diffusion-Generated[[:space:]]Image[[:space:]]Detection/8969e4d9-9379-4f6d-8489-245219be741e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion[[:space:]]Guided[[:space:]]Adaptive[[:space:]]Augmentation[[:space:]]for[[:space:]]Generalization[[:space:]]in[[:space:]]Visual[[:space:]]Reinforcement[[:space:]]Learning/195de0f3-61b4-437c-a2aa-8c67b4db85c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion[[:space:]]Image[[:space:]]Prior/08dfda20-f5f1-4fd6-9bbe-8b45f9787234_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion[[:space:]]Transformer[[:space:]]meets[[:space:]]Multi-level[[:space:]]Wavelet[[:space:]]Spectrum[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Super-Resolution/d50cd5da-af6f-4a36-956d-814eea47016b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion-Based[[:space:]]Extreme[[:space:]]High-speed[[:space:]]Scenes[[:space:]]Reconstruction[[:space:]]with[[:space:]]the[[:space:]]Complementary[[:space:]]Vision[[:space:]]Sensor/05131e98-e2e9-4484-b739-8004a279b30f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion-Based[[:space:]]Imaginative[[:space:]]Coordination[[:space:]]for[[:space:]]Bimanual[[:space:]]Manipulation/ac53e74d-878e-4493-8c1e-684969497b8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion-based[[:space:]]3D[[:space:]]Hand[[:space:]]Motion[[:space:]]Recovery[[:space:]]with[[:space:]]Intuitive[[:space:]]Physics/5ff19ab9-038d-4534-afd3-b8b238b61a23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diffusion-based[[:space:]]Source-biased[[:space:]]Model[[:space:]]for[[:space:]]Single[[:space:]]Domain[[:space:]]Generalized[[:space:]]Object[[:space:]]Detection/94a127ee-6ca8-4911-8876-68243d5e3cb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DimensionX_[[:space:]]Create[[:space:]]Any[[:space:]]3D[[:space:]]and[[:space:]]4D[[:space:]]Scenes[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]with[[:space:]]Decoupled[[:space:]]Video[[:space:]]Diffusion/3372545f-e09a-43ba-ada2-2f35dc382757_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diorama_[[:space:]]Unleashing[[:space:]]Zero-shot[[:space:]]Single-view[[:space:]]3D[[:space:]]Indoor[[:space:]]Scene[[:space:]]Modeling/2b4c0291-e318-41b0-ace0-9d21641f7eca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dirichlet-Constrained[[:space:]]Variational[[:space:]]Codebook[[:space:]]Learning[[:space:]]for[[:space:]]Temporally[[:space:]]Coherent[[:space:]]Video[[:space:]]Face[[:space:]]Restoration/c08041c7-5407-4dd9-8b22-b78032fa866c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DisCo_[[:space:]]Towards[[:space:]]Distinct[[:space:]]and[[:space:]]Coherent[[:space:]]Visual[[:space:]]Encapsulation[[:space:]]in[[:space:]]Video[[:space:]]MLLMs/02a241e1-35d6-4d8d-bb84-d218302c9671_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DisCoPatch_[[:space:]]Taming[[:space:]]Adversarially-driven[[:space:]]Batch[[:space:]]Statistics[[:space:]]for[[:space:]]Improved[[:space:]]Out-of-Distribution[[:space:]]Detection/cdeb6e5d-d3d5-417d-8ca7-cf9fa11273a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DisCoRD_[[:space:]]Discrete[[:space:]]Tokens[[:space:]]to[[:space:]]Continuous[[:space:]]Motion[[:space:]]via[[:space:]]Rectified[[:space:]]Flow[[:space:]]Decoding/b9d7cb15-aaca-442c-b459-fe3fcb9fea59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DisTime_[[:space:]]Distribution-based[[:space:]]Time[[:space:]]Representation[[:space:]]for[[:space:]]Video[[:space:]]Large[[:space:]]Language[[:space:]]Models/193dc861-f4d5-4f52-a799-2da0ecca8617_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Discontinuity-aware[[:space:]]Normal[[:space:]]Integration[[:space:]]for[[:space:]]Generic[[:space:]]Central[[:space:]]Camera[[:space:]]Models/0831d4e3-d255-4916-a885-a2c9af659558_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Discovering[[:space:]]Divergent[[:space:]]Representations[[:space:]]between[[:space:]]Text-to-Image[[:space:]]Models/928eb480-0fbb-468f-8123-e3f3d726521c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Discretized[[:space:]]Gaussian[[:space:]]Representation[[:space:]]for[[:space:]]Tomographic[[:space:]]Reconstruction/3a10b8c3-53b1-49b7-a562-c85871c4708f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DisenQ_[[:space:]]Disentangling[[:space:]]Q-Former[[:space:]]for[[:space:]]Activity-Biometrics/c620ef84-cb57-4160-b8e5-f686f6abefa3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Disentangled[[:space:]]Clothed[[:space:]]Avatar[[:space:]]Generation[[:space:]]with[[:space:]]Layered[[:space:]]Representation/10ace222-7d49-4e0c-958f-cb57387c8b7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Disentangled[[:space:]]World[[:space:]]Models_[[:space:]]Learning[[:space:]]to[[:space:]]Transfer[[:space:]]Semantic[[:space:]]Knowledge[[:space:]]from[[:space:]]Distracting[[:space:]]Videos[[:space:]]for[[:space:]]Reinforcement[[:space:]]Learning/1d200702-3dee-4311-a2ca-7fd583ee8c24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Disentangling[[:space:]]Instance[[:space:]]and[[:space:]]Scene[[:space:]]Contexts[[:space:]]for[[:space:]]3D[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion/4431f600-6f79-4716-9409-d4f9d303bc51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Disrupting[[:space:]]Model[[:space:]]Merging_[[:space:]]A[[:space:]]Parameter-Level[[:space:]]Defense[[:space:]]Without[[:space:]]Sacrificing[[:space:]]Accuracy/e8caf0bb-543e-45d2-b0c5-efb082ffac2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dissecting[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery_[[:space:]]Multiplex[[:space:]]Consensus[[:space:]]under[[:space:]]Self-Deconstruction/a78a0fb9-48c2-440d-83d7-a8ade09a4c59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DistillDrive_[[:space:]]End-to-End[[:space:]]Multi-Mode[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Distillation[[:space:]]by[[:space:]]Isomorphic[[:space:]]Hetero-Source[[:space:]]Planning[[:space:]]Model/15e29f0d-53d2-46e2-b254-16c95bf8fb97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Distilling[[:space:]]Diffusion[[:space:]]Models[[:space:]]to[[:space:]]Efficient[[:space:]]3D[[:space:]]LiDAR[[:space:]]Scene[[:space:]]Completion/59e24b7e-1009-47c3-830e-20a7ba11eb32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Distilling[[:space:]]Parallel[[:space:]]Gradients[[:space:]]for[[:space:]]Fast[[:space:]]ODE[[:space:]]Solvers[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/2809df49-b24a-4993-80a0-03db60f4b179_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dita_[[:space:]]Scaling[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]for[[:space:]]Generalist[[:space:]]Vision-Language-Action[[:space:]]Policy/a1f7a15e-f8ed-40dc-a0d0-6e6306255908_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diversity-Enhanced[[:space:]]Distribution[[:space:]]Alignment[[:space:]]for[[:space:]]Dataset[[:space:]]Distillation/2ef82098-a03a-4a74-b9d7-ee3a8a25c584_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Divide-and-Conquer[[:space:]]for[[:space:]]Enhancing[[:space:]]Unlabeled[[:space:]]Learning,[[:space:]]Stability,[[:space:]]and[[:space:]]Plasticity[[:space:]]in[[:space:]]Semi-supervised[[:space:]]Continual[[:space:]]Learning/55412f60-0e17-4f97-b739-592f007cbb2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Diving[[:space:]]into[[:space:]]the[[:space:]]Fusion[[:space:]]of[[:space:]]Monocular[[:space:]]Priors[[:space:]]for[[:space:]]Generalized[[:space:]]Stereo[[:space:]]Matching/5978b78c-4f77-45c0-9ccf-de294b27342d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Do[[:space:]]It[[:space:]]Yourself_[[:space:]]Learning[[:space:]]Semantic[[:space:]]Correspondence[[:space:]]from[[:space:]]Pseudo-Labels/3a5d4464-d3c3-4d67-96cb-b1e5fa2f8792_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DocThinker_[[:space:]]Explainable[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]with[[:space:]]Rule-based[[:space:]]Reinforcement[[:space:]]Learning[[:space:]]for[[:space:]]Document[[:space:]]Understanding/d85d2101-a2d0-4138-a4f1-50af798ba6ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Does[[:space:]]Your[[:space:]]Vision-Language[[:space:]]Model[[:space:]]Get[[:space:]]Lost[[:space:]]in[[:space:]]the[[:space:]]Long[[:space:]]Video[[:space:]]Sampling[[:space:]]Dilemma_/92ead825-3f91-4fa8-87e9-4cf2fc448de0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Domain[[:space:]]Generalizable[[:space:]]Portrait[[:space:]]Style[[:space:]]Transfer/3f99f8be-4cfe-45d7-af38-0a505a5890d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Domain-aware[[:space:]]Category-level[[:space:]]Geometry[[:space:]]Learning[[:space:]]Segmentation[[:space:]]for[[:space:]]3D[[:space:]]Point[[:space:]]Clouds/5e959033-3720-406c-8b24-c664976882c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Doodle[[:space:]]Your[[:space:]]Keypoints_[[:space:]]Sketch-Based[[:space:]]Few-Shot[[:space:]]Keypoint[[:space:]]Detection/dff5865f-c7a4-4f77-ab92-94c17be0e8fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DoppDrive_[[:space:]]Doppler-Driven[[:space:]]Temporal[[:space:]]Aggregation[[:space:]]for[[:space:]]Improved[[:space:]]Radar[[:space:]]Object[[:space:]]Detection/b7bcdd54-390c-4c4d-a538-becaac536ef3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Doppler-Aware[[:space:]]LiDAR-RADAR[[:space:]]Fusion[[:space:]]for[[:space:]]Weather-Robust[[:space:]]3D[[:space:]]Detection/f9455af6-6dba-449a-905d-4de8ed00043a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Draw[[:space:]]Your[[:space:]]Mind_[[:space:]]Personalized[[:space:]]Generation[[:space:]]via[[:space:]]Condition-Level[[:space:]]Modeling[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/741949bb-d0fe-4fcb-a4ac-c11cd35e9412_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Drawing[[:space:]]Developmental[[:space:]]Trajectory[[:space:]]from[[:space:]]Cortical[[:space:]]Surface[[:space:]]Reconstruction/07c3b248-445e-4415-885a-f3afaf73fe8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dream-to-Recon_[[:space:]]Monocular[[:space:]]3D[[:space:]]Reconstruction[[:space:]]with[[:space:]]Diffusion-Depth[[:space:]]Distillation[[:space:]]from[[:space:]]Single[[:space:]]Images/b9572e31-c157-45c6-b931-3d898669c981_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DreamActor-M1_[[:space:]]Holistic,[[:space:]]Expressive[[:space:]]and[[:space:]]Robust[[:space:]]Human[[:space:]]Image[[:space:]]Animation[[:space:]]with[[:space:]]Hybrid[[:space:]]Guidance/32a35608-c60a-4a9c-b143-427a2f9c640e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DreamCube_[[:space:]]RGB-D[[:space:]]Panorama[[:space:]]Generation[[:space:]]via[[:space:]]Multi-plane[[:space:]]Synchronization/348de1c0-1446-4b35-b5c4-49bf905d4f84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DreamDance_[[:space:]]Animating[[:space:]]Human[[:space:]]Images[[:space:]]by[[:space:]]Enriching[[:space:]]3D[[:space:]]Geometry[[:space:]]Cues[[:space:]]from[[:space:]]2D[[:space:]]Poses/01f949d6-2e76-485c-b2c6-83126d3fbe5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DreamFuse_[[:space:]]Adaptive[[:space:]]Image[[:space:]]Fusion[[:space:]]with[[:space:]]Diffusion[[:space:]]Transformer/b4e07aee-7167-40dc-9bcf-9b711ae6514b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DreamLayer_[[:space:]]Simultaneous[[:space:]]Multi-Layer[[:space:]]Generation[[:space:]]via[[:space:]]Diffusion[[:space:]]Model/806d28dd-a0bb-4250-90fa-09d8d13c6907_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DreamRelation_[[:space:]]Relation-Centric[[:space:]]Video[[:space:]]Customization/efbd9993-ed72-4a51-940f-93f77550d447_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DreamRenderer_[[:space:]]Taming[[:space:]]Multi-Instance[[:space:]]Attribute[[:space:]]Control[[:space:]]in[[:space:]]Large-Scale[[:space:]]Text-to-Image[[:space:]]Models/a67926b8-e0e0-43b7-ae59-28a1ff080b5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DriveArena_[[:space:]]A[[:space:]]Closed-loop[[:space:]]Generative[[:space:]]Simulation[[:space:]]Platform[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/d9bd6be5-37d7-42ae-ae42-745735f493d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DriveX_[[:space:]]Omni[[:space:]]Scene[[:space:]]Modeling[[:space:]]for[[:space:]]Learning[[:space:]]Generalizable[[:space:]]World[[:space:]]Knowledge[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/a0d765a7-e791-4b96-8aac-9294f46703aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Driving[[:space:]]View[[:space:]]Synthesis[[:space:]]on[[:space:]]Free-form[[:space:]]Trajectories[[:space:]]with[[:space:]]Generative[[:space:]]Prior/c2c738be-410b-4597-9fca-ce2c33bbc39c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DrivingGPT_[[:space:]]Unifying[[:space:]]Driving[[:space:]]World[[:space:]]Modeling[[:space:]]and[[:space:]]Planning[[:space:]]with[[:space:]]Multi-modal[[:space:]]Autoregressive[[:space:]]Transformers/ced33ae0-83c7-493a-a2ec-072cd236ee15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DropletVideo_[[:space:]]A[[:space:]]Dataset[[:space:]]and[[:space:]]Approach[[:space:]]to[[:space:]]Explore[[:space:]]Integral[[:space:]]Spatio-Temporal[[:space:]]Consistent[[:space:]]Video[[:space:]]Generation/7e65525f-691f-4d61-8b39-fa22aee1231a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DuCos_[[:space:]]Duality[[:space:]]Constrained[[:space:]]Depth[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Foundation[[:space:]]Model/9d32a136-3635-4673-ad33-60242ad85178_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DuET_[[:space:]]Dual[[:space:]]Incremental[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Exemplar-Free[[:space:]]Task[[:space:]]Arithmetic/7e7a44ac-17c7-4ea9-9bde-375575b71e85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual[[:space:]]Domain[[:space:]]Control[[:space:]]via[[:space:]]Active[[:space:]]Learning[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Domain[[:space:]]Incremental[[:space:]]Object[[:space:]]Detection/852b8f51-8746-46e6-a383-7eb8d5661cac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual[[:space:]]Reciprocal[[:space:]]Learning[[:space:]]of[[:space:]]Language-based[[:space:]]Human[[:space:]]Motion[[:space:]]Understanding[[:space:]]and[[:space:]]Generation/01bd732f-a1f8-4922-9c67-efbe5f51832f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual[[:space:]]Recursive[[:space:]]Feedback[[:space:]]on[[:space:]]Generation[[:space:]]and[[:space:]]Appearance[[:space:]]Latents[[:space:]]for[[:space:]]Pose-Robust[[:space:]]Text-to-Image[[:space:]]Diffusion/402e68cb-5e95-4ee8-a029-857c5c569199_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual-Expert[[:space:]]Consistency[[:space:]]Model[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]High-Quality[[:space:]]Video[[:space:]]Generation/a71ee417-7511-4000-bc48-be8d306158b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual-Process[[:space:]]Image[[:space:]]Generation/8261b18a-8ce0-4dfa-8df8-e24547b06865_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual-Rate[[:space:]]Dynamic[[:space:]]Teacher[[:space:]]for[[:space:]]Source-Free[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detection/5e43aed5-6527-4eaa-b762-9164834c3e9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual-S3D_[[:space:]]Hierarchical[[:space:]]Dual-Path[[:space:]]Selective[[:space:]]SSM-CNN[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Implicit[[:space:]]Reconstruction/57d708cb-c272-498e-bffe-14a9c707753a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual-Temporal[[:space:]]Exemplar[[:space:]]Representation[[:space:]]Network[[:space:]]for[[:space:]]Video[[:space:]]Semantic[[:space:]]Segmentation/1dfe8c69-9e61-414f-b1dc-21ddf67a1fe8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dual-level[[:space:]]Prototype[[:space:]]Learning[[:space:]]for[[:space:]]Composite[[:space:]]Degraded[[:space:]]Image[[:space:]]Restoration/5da86314-6f17-490c-826d-d6e029ca3829_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DualReal_[[:space:]]Adaptive[[:space:]]Joint[[:space:]]Training[[:space:]]for[[:space:]]Lossless[[:space:]]Identity-Motion[[:space:]]Fusion[[:space:]]in[[:space:]]Video[[:space:]]Customization/e68e7434-01a7-42a0-9b2f-d0b450424e92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DuoCLR_[[:space:]]Dual-Surrogate[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Skeleton-based[[:space:]]Human[[:space:]]Action[[:space:]]Segmentation/a4d1c34d-9dee-4941-9149-7ce89e089c24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DuoLoRA[[:space:]]_[[:space:]]Cycle-consistent[[:space:]]and[[:space:]]Rank-disentangled[[:space:]]Content-Style[[:space:]]Personalization/b50bc0e5-fcd9-415f-aa7b-0124d967da90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DyGS-SLAM_[[:space:]]Real-Time[[:space:]]Accurate[[:space:]]Localization[[:space:]]and[[:space:]]Gaussian[[:space:]]Reconstruction[[:space:]]for[[:space:]]Dynamic[[:space:]]Scenes/cf5c56de-cd07-4297-a1d1-dac566fd2cf2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DyWA_[[:space:]]Dynamics-adaptive[[:space:]]World[[:space:]]Action[[:space:]]Model[[:space:]]for[[:space:]]Generalizable[[:space:]]Non-prehensile[[:space:]]Manipulation/2cb49262-70eb-45be-a01a-c67e85090c93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DynFaceRestore_[[:space:]]Balancing[[:space:]]Fidelity[[:space:]]and[[:space:]]Quality[[:space:]]in[[:space:]]Diffusion-Guided[[:space:]]Blind[[:space:]]Face[[:space:]]Restoration[[:space:]]with[[:space:]]Dynamic[[:space:]]Blur-Level[[:space:]]Mapping[[:space:]]and[[:space:]]Guidance/184f7040-fd88-4615-89ae-776eb9c1d1c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DynImg_[[:space:]]Key[[:space:]]Frames[[:space:]]with[[:space:]]Visual[[:space:]]Prompts[[:space:]]are[[:space:]]Good[[:space:]]Representation[[:space:]]for[[:space:]]Multi-Modal[[:space:]]Video[[:space:]]Understanding/3504abf0-1617-495e-9687-9d9a943fc9a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Dictionary[[:space:]]Learning[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Image[[:space:]]Segmentation/e9bf7a64-3add-459e-9e4b-0ab16af23631_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Group[[:space:]]Detection[[:space:]]using[[:space:]]VLM-augmented[[:space:]]Temporal[[:space:]]Groupness[[:space:]]Graph/1fc53a7c-4fc1-4e9c-8e1f-4015bb2a6be2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Multi-Layer[[:space:]]Null[[:space:]]Space[[:space:]]Projection[[:space:]]for[[:space:]]Vision-Language[[:space:]]Continual[[:space:]]Learning/2af212c1-116e-4a06-a548-3a30c98ce749_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Multimodal[[:space:]]Prototype[[:space:]]Learning[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models/61b3e272-efd9-42d7-8f0f-ab8b09b03d08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Point[[:space:]]Maps_[[:space:]]A[[:space:]]Versatile[[:space:]]Representation[[:space:]]for[[:space:]]Dynamic[[:space:]]3D[[:space:]]Reconstruction/d0107880-dae1-4822-aed3-80d8751413b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Reconstruction[[:space:]]of[[:space:]]Hand-Object[[:space:]]Interaction[[:space:]]with[[:space:]]Distributed[[:space:]]Force-aware[[:space:]]Contact[[:space:]]Representation/ee5a8186-bd46-4b47-9d66-b80b52d95f2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic[[:space:]]Typography_[[:space:]]Bringing[[:space:]]Text[[:space:]]to[[:space:]]Life[[:space:]]via[[:space:]]Video[[:space:]]Diffusion[[:space:]]Prior/b6877c5a-07c9-41e5-9da2-5607b80d1a30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic-DINO_[[:space:]]Fine-Grained[[:space:]]Mixture[[:space:]]of[[:space:]]Experts[[:space:]]Tuning[[:space:]]for[[:space:]]Real-time[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detection/3393d04b-878b-41cf-bfdd-2af3dbfd4c7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Dynamic-VLM_[[:space:]]Simple[[:space:]]Dynamic[[:space:]]Visual[[:space:]]Token[[:space:]]Compression[[:space:]]for[[:space:]]VideoLLM/f99a7d98-09ac-4dab-bfd6-bb7a5c33ee79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DynamicFace_[[:space:]]High-Quality[[:space:]]and[[:space:]]Consistent[[:space:]]Face[[:space:]]Swapping[[:space:]]for[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]using[[:space:]]Composable[[:space:]]3D[[:space:]]Facial[[:space:]]Priors/c494f9de-100d-4b29-b08a-b6b82d5e5fd4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/DynamicID_[[:space:]]Zero-Shot[[:space:]]Multi-ID[[:space:]]Image[[:space:]]Personalization[[:space:]]with[[:space:]]Flexible[[:space:]]Facial[[:space:]]Editability/1fad7f97-da33-417f-a5ff-c6344a9ec7af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/E-NeMF_[[:space:]]Event-based[[:space:]]Neural[[:space:]]Motion[[:space:]]Field[[:space:]]for[[:space:]]Novel[[:space:]]Space-time[[:space:]]View[[:space:]]Synthesis[[:space:]]of[[:space:]]Dynamic[[:space:]]Scenes/b1d24e42-13c5-4b4c-be5a-999e58bbd1c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/E-SAM_[[:space:]]Training-Free[[:space:]]Segment[[:space:]]Every[[:space:]]Entity[[:space:]]Model/2bc1188c-d2fb-46b3-9bff-90ed01dd1d31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EA-KD_[[:space:]]Entropy-based[[:space:]]Adaptive[[:space:]]Knowledge[[:space:]]Distillation/5d7c2c34-bc9d-41ab-b58b-b36dcb85e031_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EA-Vit_[[:space:]]Efficient[[:space:]]Adaptation[[:space:]]for[[:space:]]Elastic[[:space:]]Vision[[:space:]]Transformer/d07ea315-057d-4370-bea6-7830f65c8f9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EAMamba_[[:space:]]Efficient[[:space:]]All-Around[[:space:]]Vision[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]for[[:space:]]Image[[:space:]]Restoration/37f45fae-1c1a-4902-8813-8343e07775c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EC-Flow_[[:space:]]Enabling[[:space:]]Versatile[[:space:]]Robotic[[:space:]]Manipulation[[:space:]]from[[:space:]]Action-Unlabeled[[:space:]]Videos[[:space:]]via[[:space:]]Embodiment-Centric[[:space:]]Flow/d131af68-500c-47ab-8313-152badcef25a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EDFFDNet_[[:space:]]Towards[[:space:]]Accurate[[:space:]]and[[:space:]]Efficient[[:space:]]Unsupervised[[:space:]]Multi-Grid[[:space:]]Image[[:space:]]Registration/01025ef2-ed01-4606-adcf-707412b16e7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EDM_[[:space:]]Efficient[[:space:]]Deep[[:space:]]Feature[[:space:]]Matching/3875a24a-70f1-45f3-b185-5f2b9a3e9efb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EDiT_[[:space:]]Efficient[[:space:]]Diffusion[[:space:]]Transformers[[:space:]]with[[:space:]]Linear[[:space:]]Compressed[[:space:]]Attention/a402e4b5-b72d-4821-a0b2-c23a7d0e773d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EEGMirror_[[:space:]]Leveraging[[:space:]]EEG[[:space:]]Data[[:space:]]in[[:space:]]the[[:space:]]Wild[[:space:]]via[[:space:]]Montage-Agnostic[[:space:]]Self-Supervision[[:space:]]for[[:space:]]EEG[[:space:]]to[[:space:]]Video[[:space:]]Decoding/0cc8b244-de85-4123-9e84-d464d960f4e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EEdit[[:space:]]_[[:space:]]Rethinking[[:space:]]the[[:space:]]Spatial[[:space:]]and[[:space:]]Temporal[[:space:]]Redundancy[[:space:]]for[[:space:]]Efficient[[:space:]]Image[[:space:]]Editing/2e980da3-1e30-4953-adba-e4ac51cbcb7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EFTViT_[[:space:]]Efficient[[:space:]]Federated[[:space:]]Training[[:space:]]of[[:space:]]Vision[[:space:]]Transformers[[:space:]]with[[:space:]]Masked[[:space:]]Images[[:space:]]on[[:space:]]Resource-Constrained[[:space:]]Clients/585b9ed0-14ec-4270-ac43-0a0eff3cc9eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EMD_[[:space:]]Explicit[[:space:]]Motion[[:space:]]Modeling[[:space:]]for[[:space:]]High-Quality[[:space:]]Street[[:space:]]Gaussian[[:space:]]Splatting/5c25339e-6383-4948-8dac-d520c8ce96d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EMatch_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Event-based[[:space:]]Optical[[:space:]]Flow[[:space:]]and[[:space:]]Stereo[[:space:]]Matching/e49a6498-9195-4247-a00f-32dbb7bbea8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EMoTive_[[:space:]]Event-guided[[:space:]]Trajectory[[:space:]]Modeling[[:space:]]for[[:space:]]3D[[:space:]]Motion[[:space:]]Estimation/162c3375-fc4c-47ca-91ca-a91e69913a9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ERNet_[[:space:]]Efficient[[:space:]]Non-Rigid[[:space:]]Registration[[:space:]]Network[[:space:]]for[[:space:]]Point[[:space:]]Sequences/b401904b-ba0b-4f88-bd76-e645342316c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ESCNet_Edge-Semantic[[:space:]]Collaborative[[:space:]]Network[[:space:]]for[[:space:]]Camouflaged[[:space:]]Object[[:space:]]Detection/38039c08-c8ea-4174-b346-666949d80d5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ESSENTIAL_[[:space:]]Episodic[[:space:]]and[[:space:]]Semantic[[:space:]]Memory[[:space:]]Integration[[:space:]]for[[:space:]]Video[[:space:]]Class-Incremental[[:space:]]Learning/dff31a63-d139-46fd-81b8-93da7ed5f1e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ETA_[[:space:]]Efficiency[[:space:]]through[[:space:]]Thinking[[:space:]]Ahead,[[:space:]]A[[:space:]]Dual[[:space:]]Approach[[:space:]]to[[:space:]]Self-Driving[[:space:]]with[[:space:]]Large[[:space:]]Models/c0e0ca4c-efe6-4309-aca0-9ca01d854a08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ETA_[[:space:]]Energy-based[[:space:]]Test-time[[:space:]]Adaptation[[:space:]]for[[:space:]]Depth[[:space:]]Completion/2d32724a-e80b-4fa5-be54-c85e8ae0e4a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ETCH_[[:space:]]Generalizing[[:space:]]Body[[:space:]]Fitting[[:space:]]to[[:space:]]Clothed[[:space:]]Humans[[:space:]]via[[:space:]]Equivariant[[:space:]]Tightness/fdf5f400-18b3-4edb-af86-9b83b9e1a955_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ETVA_[[:space:]]Evaluation[[:space:]]of[[:space:]]Text-to-Video[[:space:]]Alignment[[:space:]]via[[:space:]]Fine-grained[[:space:]]Question[[:space:]]Generation[[:space:]]and[[:space:]]Answering/9ebd13ed-dc33-483b-8fe4-600d9033d23a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EVDM_[[:space:]]Event-based[[:space:]]Real-world[[:space:]]Video[[:space:]]Deblurring[[:space:]]with[[:space:]]Mamba/ec95cad2-8e2d-476e-b367-366116a088e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EVER_[[:space:]]Exact[[:space:]]Volumetric[[:space:]]Ellipsoid[[:space:]]Rendering[[:space:]]for[[:space:]]Real-time[[:space:]]View[[:space:]]Synthesis/2391ebf1-799e-4b3a-a589-f57f93683c15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EVEv2_[[:space:]]Improved[[:space:]]Baselines[[:space:]]for[[:space:]]Encoder-Free[[:space:]]Vision-Language[[:space:]]Models/898c4894-afab-45c9-adf0-da2b94f9d150_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EVOLVE_[[:space:]]Event-Guided[[:space:]]Deformable[[:space:]]Feature[[:space:]]Transfer[[:space:]]and[[:space:]]Dual-Memory[[:space:]]Refinement[[:space:]]for[[:space:]]Low-Light[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/1c89e658-03dd-4c4f-a547-fc48bb275863_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EVT_[[:space:]]Efficient[[:space:]]View[[:space:]]Transformation[[:space:]]for[[:space:]]Multi-Modal[[:space:]]3D[[:space:]]Object[[:space:]]Detection/5a4293e8-092b-49fe-a4ed-1d838f34bf9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EYE3_Turn[[:space:]]Anything[[:space:]]into[[:space:]]Naked-eye[[:space:]]3D/d99470d8-13c3-4305-bf37-1ce4c08ad45a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Early[[:space:]]Timestep[[:space:]]Zero-Shot[[:space:]]Candidate[[:space:]]Selection[[:space:]]for[[:space:]]Instruction-Guided[[:space:]]Image[[:space:]]Editing/ca303967-e288-4035-9a2c-c5aad718889a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Easi3R_[[:space:]]Estimating[[:space:]]Disentangled[[:space:]]Motion[[:space:]]from[[:space:]]DUSt3R[[:space:]]Without[[:space:]]Training/9a07de2d-b3df-4aca-9dce-3067726fe58a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Easy3D_[[:space:]]A[[:space:]]Simple[[:space:]]Yet[[:space:]]Effective[[:space:]]Method[[:space:]]for[[:space:]]3D[[:space:]]Interactive[[:space:]]Segmentation/df7a1024-2f08-4358-acb6-6e2473750d9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EasyControl_[[:space:]]Adding[[:space:]]Efficient[[:space:]]and[[:space:]]Flexible[[:space:]]Control[[:space:]]for[[:space:]]Diffusion[[:space:]]Transformer/3b3498ca-87fd-42a5-b025-e717e78105b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Edicho_[[:space:]]Consistent[[:space:]]Image[[:space:]]Editing[[:space:]]in[[:space:]]the[[:space:]]Wild/624759a8-118c-4eac-8304-fd34712dc0d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Edit360_[[:space:]]2D[[:space:]]Image[[:space:]]Edits[[:space:]]to[[:space:]]3D[[:space:]]Assets[[:space:]]from[[:space:]]Any[[:space:]]Angle/f1a652be-a430-4af1-9474-615f03aa5d06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EditCLIP_[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Image[[:space:]]Editing/91122731-fe44-4f31-a5ed-59bde7bff136_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Effective[[:space:]]Training[[:space:]]Data[[:space:]]Synthesis[[:space:]]for[[:space:]]Improving[[:space:]]MLLM[[:space:]]Chart[[:space:]]Understanding/3184609e-34fc-438b-87f4-d35aa0443024_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Adaptation[[:space:]]of[[:space:]]Pre-trained[[:space:]]Vision[[:space:]]Transformer[[:space:]]underpinned[[:space:]]by[[:space:]]Approximately[[:space:]]Orthogonal[[:space:]]Fine-Tuning[[:space:]]Strategy/560fe71d-4ff1-47cd-9604-62f38c8f3892_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Autoregressive[[:space:]]Shape[[:space:]]Generation[[:space:]]via[[:space:]]Octree-Based[[:space:]]Adaptive[[:space:]]Tokenization/90669462-d182-4f35-9816-1b763bfbc360_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Concertormer[[:space:]]for[[:space:]]Image[[:space:]]Deblurring[[:space:]]and[[:space:]]Beyond/d08e6c25-a0c0-4b36-ac46-e1691e49c6f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Event[[:space:]]Camera[[:space:]]Data[[:space:]]Pretraining[[:space:]]with[[:space:]]Adaptive[[:space:]]Prompt[[:space:]]Fusion/46c59fb3-815b-456c-8f37-ac3dc6a72604_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Fine-Tuning[[:space:]]of[[:space:]]Large[[:space:]]Models[[:space:]]via[[:space:]]Nested[[:space:]]Low-Rank[[:space:]]Adaptation/9f05468d-f1d1-4b28-bd42-a59748ac0859_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Input-level[[:space:]]Backdoor[[:space:]]Defense[[:space:]]on[[:space:]]Text-to-Image[[:space:]]Synthesis[[:space:]]via[[:space:]]Neuron[[:space:]]Activation[[:space:]]Variation/95a6207c-a31a-4936-b4e6-e36743bafd94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Multi-Person[[:space:]]Motion[[:space:]]Prediction[[:space:]]by[[:space:]]Lightweight[[:space:]]Spatial[[:space:]]and[[:space:]]Temporal[[:space:]]Interactions/95785fc8-8dd2-4fb8-9e8e-566d6df286c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Spiking[[:space:]]Point[[:space:]]Mamba[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Analysis/7122bff8-00f1-446d-b587-0cfa12efeba5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Track[[:space:]]Anything/f86b2c37-a72f-4152-8b81-4a1205533df8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Unsupervised[[:space:]]Shortcut[[:space:]]Learning[[:space:]]Detection[[:space:]]and[[:space:]]Mitigation[[:space:]]in[[:space:]]Transformers/5eba4ce1-7361-4b0d-bc6e-16286c46ab4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Efficient[[:space:]]Visual[[:space:]]Place[[:space:]]Recognition[[:space:]]Through[[:space:]]Multimodal[[:space:]]Semantic[[:space:]]Knowledge[[:space:]]Integration/cac53b3a-81a9-4ef6-8127-c3f1b59367c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EfficientMT_[[:space:]]Efficient[[:space:]]Temporal[[:space:]]Adaptation[[:space:]]for[[:space:]]Motion[[:space:]]Transfer[[:space:]]in[[:space:]]Text-to-Video[[:space:]]Diffusion[[:space:]]Models/1c5619bb-f938-41af-85f7-33c42dadd9fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EgoAdapt_[[:space:]]Adaptive[[:space:]]Multisensory[[:space:]]Distillation[[:space:]]and[[:space:]]Policy[[:space:]]Learning[[:space:]]for[[:space:]]Efficient[[:space:]]Egocentric[[:space:]]Perception/6c51cd88-ac07-4b6c-b3d5-ed8db9ddbdc5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EgoAgent_[[:space:]]A[[:space:]]Joint[[:space:]]Predictive[[:space:]]Agent[[:space:]]Model[[:space:]]in[[:space:]]Egocentric[[:space:]]Worlds/86beefbd-3471-46c1-9cb3-78bd2e6a7ec5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EgoM2P_[[:space:]]Egocentric[[:space:]]Multimodal[[:space:]]Multitask[[:space:]]Pretraining/b0e3bff3-4853-462c-88dd-9c08233a9686_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EgoMusic-driven[[:space:]]Human[[:space:]]Dance[[:space:]]Motion[[:space:]]Estimation[[:space:]]with[[:space:]]Skeleton[[:space:]]Mamba/b09b4ab0-d1ad-4ba9-acfa-49f50df0b0d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Egocentric[[:space:]]Action-aware[[:space:]]Inertial[[:space:]]Localization[[:space:]]in[[:space:]]Point[[:space:]]Clouds[[:space:]]with[[:space:]]Vision-Language[[:space:]]Guidance/69556f3e-9046-483f-a60f-7b3723c31855_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Embodied[[:space:]]Image[[:space:]]Captioning_[[:space:]]Self-supervised[[:space:]]Learning[[:space:]]Agents[[:space:]]for[[:space:]]Spatially[[:space:]]Coherent[[:space:]]Image[[:space:]]Descriptions/60e4736a-4341-43b1-8795-7707854358ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Embodied[[:space:]]Navigation[[:space:]]with[[:space:]]Auxiliary[[:space:]]Task[[:space:]]of[[:space:]]Action[[:space:]]Description[[:space:]]Prediction/a8259b3c-f4b9-4360-9aef-015762e621d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Embodied[[:space:]]Representation[[:space:]]Alignment[[:space:]]with[[:space:]]Mirror[[:space:]]Neurons/1a9ab08d-9a2c-43fd-94f5-5038fe26d6e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Embodied[[:space:]]VideoAgent_[[:space:]]Persistent[[:space:]]Memory[[:space:]]from[[:space:]]Egocentric[[:space:]]Videos[[:space:]]and[[:space:]]Embodied[[:space:]]Sensors[[:space:]]Enables[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Understanding/a017345c-038c-4d4f-96f4-badf8dc4eef1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EmbodiedOcc_[[:space:]]Embodied[[:space:]]3D[[:space:]]Occupancy[[:space:]]Prediction[[:space:]]for[[:space:]]Vision-based[[:space:]]Online[[:space:]]Scene[[:space:]]Understanding/ecffb22e-f04b-4d58-841a-1b052e2a208b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EmbodiedSplat_[[:space:]]Personalized[[:space:]]Real-to-Sim-to-Real[[:space:]]Navigation[[:space:]]with[[:space:]]Gaussian[[:space:]]Splats[[:space:]]from[[:space:]]a[[:space:]]Mobile[[:space:]]Device/eb28cbcd-1806-4f8d-ad73-ef4f7f494673_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EmotiCrafter_[[:space:]]Text-to-Emotional-Image[[:space:]]Generation[[:space:]]based[[:space:]]on[[:space:]]Valence-Arousal[[:space:]]Model/64d56063-afb5-47eb-8d97-ccd27f1db0d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Emulating[[:space:]]Self-attention[[:space:]]with[[:space:]]Convolution[[:space:]]for[[:space:]]Efficient[[:space:]]Image[[:space:]]Super-Resolution/28032f25-1d68-496c-851a-6c9d0a64d207_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/End-to-End[[:space:]]Driving[[:space:]]with[[:space:]]Online[[:space:]]Trajectory[[:space:]]Evaluation[[:space:]]via[[:space:]]BEV[[:space:]]World[[:space:]]Model/24c54d04-9c44-4257-9691-2d6e9dd0c2c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/End-to-End[[:space:]]Entity-Predicate[[:space:]]Association[[:space:]]Reasoning[[:space:]]for[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/331916de-5a77-43d5-b8b1-867a812cb4ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/End-to-End[[:space:]]Multi-Modal[[:space:]]Diffusion[[:space:]]Mamba/229968bf-b390-4563-8550-16e090dc25cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Engage[[:space:]]for[[:space:]]All_[[:space:]]Making[[:space:]]Ordinary[[:space:]]Image[[:space:]]Descriptions[[:space:]]Appealing[[:space:]]Again!/ad64f5b0-b3c7-476c-8285-9a78b446995d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhanced[[:space:]]Event-based[[:space:]]Dense[[:space:]]Stereo[[:space:]]via[[:space:]]Cross-Sensor[[:space:]]Knowledge[[:space:]]Distillation/f6b9298e-6b4d-4df1-b0c3-014f084ff091_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhanced[[:space:]]Pansharpening[[:space:]]via[[:space:]]Quaternion[[:space:]]Spatial-Spectral[[:space:]]Interactions/ead32439-3d1c-41ba-b332-4dc5bcc158be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Adversarial[[:space:]]Transferability[[:space:]]by[[:space:]]Balancing[[:space:]]Exploration[[:space:]]and[[:space:]]Exploitation[[:space:]]with[[:space:]]Gradient-Guided[[:space:]]Sampling/6da66727-c69d-4b80-9f15-0e506723c7cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Few-Shot[[:space:]]Vision-Language[[:space:]]Classification[[:space:]]with[[:space:]]Large[[:space:]]Multimodal[[:space:]]Model[[:space:]]Features/958199ee-2123-4605-9c2c-3f90d0a78ead_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Image[[:space:]]Restoration[[:space:]]Transformer[[:space:]]via[[:space:]]Adaptive[[:space:]]Translation[[:space:]]Equivariance/315e18c5-4a71-4399-bd7d-013d4c4b0332_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Mamba[[:space:]]Decoder[[:space:]]with[[:space:]]Bidirectional[[:space:]]Interaction[[:space:]]in[[:space:]]Multi-Task[[:space:]]Dense[[:space:]]Prediction/714687cb-b372-4c5e-8dc1-605cf43cbec1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Numerical[[:space:]]Prediction[[:space:]]of[[:space:]]MLLMs[[:space:]]with[[:space:]]Soft[[:space:]]Labeling/fe57947e-0779-4b96-b592-99d87fccfb07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Partially[[:space:]]Relevant[[:space:]]Video[[:space:]]Retrieval[[:space:]]with[[:space:]]Hyperbolic[[:space:]]Learning/8267e0d4-8636-4d06-abe0-cf72edf7f96f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Prompt[[:space:]]Generation[[:space:]]with[[:space:]]Adaptive[[:space:]]Refinement[[:space:]]for[[:space:]]Camouflaged[[:space:]]Object[[:space:]]Detection/31897efc-f494-43de-9990-b3b0b058743c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Reward[[:space:]]Models[[:space:]]for[[:space:]]High-quality[[:space:]]Image[[:space:]]Generation_[[:space:]]Beyond[[:space:]]Text-Image[[:space:]]Alignment/4fbc7473-ddd0-4038-9d67-401857bfa6dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Spatial[[:space:]]Reasoning[[:space:]]in[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]through[[:space:]]Reasoning-based[[:space:]]Segmentation/dda84173-1ef6-4cdd-867f-e3fcb580b499_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Transferability[[:space:]]of[[:space:]]Targeted[[:space:]]Adversarial[[:space:]]Examples[[:space:]]via[[:space:]]Inverse[[:space:]]Target[[:space:]]Gradient[[:space:]]Competition[[:space:]]and[[:space:]]Spatial[[:space:]]Distance[[:space:]]Stretching/b533dcf6-58e9-4bab-bdf5-cd55686734f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Transformers[[:space:]]Through[[:space:]]Conditioned[[:space:]]Embedded[[:space:]]Tokens/e4867dd3-615a-4802-a597-48b7c4a2f52c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enhancing[[:space:]]Zero-shot[[:space:]]Object[[:space:]]Counting[[:space:]]via[[:space:]]Text-guided[[:space:]]Local[[:space:]]Ranking[[:space:]]and[[:space:]]Number-evoked[[:space:]]Global[[:space:]]Attention/42dbced4-3afc-4fc3-a1f8-6ab3c4e3bb96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enpowering[[:space:]]Your[[:space:]]Pansharpening[[:space:]]Models[[:space:]]with[[:space:]]Generalizability_[[:space:]]Unified[[:space:]]Distribution[[:space:]]is[[:space:]]All[[:space:]]You[[:space:]]Need/84a6c0dd-500e-40a7-b080-833571baa669_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Enrich[[:space:]]and[[:space:]]Detect_[[:space:]]Video[[:space:]]Temporal[[:space:]]Grounding[[:space:]]with[[:space:]]Multimodal[[:space:]]LLMs/6520b7cd-57d9-4f71-bc18-4af7c56a0a7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ensemble[[:space:]]Foreground[[:space:]]Management[[:space:]]for[[:space:]]Unsupervised[[:space:]]Object[[:space:]]Discovery/8437cd25-5fba-48bf-86af-e65bf6a1e6c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Entropy-Adaptive[[:space:]]Diffusion[[:space:]]Policy[[:space:]]Optimization[[:space:]]with[[:space:]]Dynamic[[:space:]]Step[[:space:]]Alignment/c4a55a51-0565-4ccd-a910-9c67e543b847_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Environment-Agnostic[[:space:]]Pose_[[:space:]]Generating[[:space:]]Environment-independent[[:space:]]Object[[:space:]]Representations[[:space:]]for[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation/5521ae38-4b00-475c-a94f-31fec5839328_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Epipolar[[:space:]]Consistent[[:space:]]Attention[[:space:]]Aggregation[[:space:]]Network[[:space:]]for[[:space:]]Unsupervised[[:space:]]Light[[:space:]]Field[[:space:]]Disparity[[:space:]]Estimation/74749a68-b688-4175-99d0-00c425f7c7e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Epona_[[:space:]]Autoregressive[[:space:]]Diffusion[[:space:]]World[[:space:]]Model[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/96492e1e-bfce-4329-bd89-949aa9fef588_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EquiCaps_[[:space:]]Predictor-Free[[:space:]]Pose-Aware[[:space:]]Pre-Trained[[:space:]]Capsule[[:space:]]Networks/0be2dbf6-d8b3-41a1-9e9f-e1a2a2904ac5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Equipping[[:space:]]Vision[[:space:]]Foundation[[:space:]]Model[[:space:]]with[[:space:]]Mixture[[:space:]]of[[:space:]]Experts[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection/79e4b61c-cad3-4644-ae1d-912cf94be19d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Erasing[[:space:]]More[[:space:]]Than[[:space:]]Intended_[[:space:]]How[[:space:]]Concept[[:space:]]Erasure[[:space:]]Degrades[[:space:]]the[[:space:]]Generation[[:space:]]of[[:space:]]Non-Target[[:space:]]Concepts/872a71e9-733b-4e03-be37-36d9dbc5bdd0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Error[[:space:]]Recognition[[:space:]]in[[:space:]]Procedural[[:space:]]Videos[[:space:]]using[[:space:]]Generalized[[:space:]]Task[[:space:]]Graph/5a14de37-2719-437f-aefb-79f29e2ef354_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Estimating[[:space:]]2D[[:space:]]Camera[[:space:]]Motion[[:space:]]with[[:space:]]Hybrid[[:space:]]Motion[[:space:]]Basis/46352bd4-609c-49a6-a102-ed3ec1467e54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EvRT-DETR_[[:space:]]Latent[[:space:]]Space[[:space:]]Adaptation[[:space:]]of[[:space:]]Image[[:space:]]Detectors[[:space:]]for[[:space:]]Event-based[[:space:]]Vision/b434c8f2-0109-4d2a-8b19-39a0325126c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EvaGaussians_[[:space:]]Event[[:space:]]Stream[[:space:]]Assisted[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Blurry[[:space:]]Images/4ff03f6c-2177-47f2-8440-fcfd6f8d42f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Evading[[:space:]]Data[[:space:]]Provenance[[:space:]]in[[:space:]]Deep[[:space:]]Neural[[:space:]]Networks/22b6c5cf-c3b2-4e69-9394-11210dd9a3ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Event-Driven[[:space:]]Storytelling[[:space:]]with[[:space:]]Multiple[[:space:]]Lifelike[[:space:]]Humans[[:space:]]in[[:space:]]a[[:space:]]3D[[:space:]]Scene/d626a47c-0fd6-4b74-b622-4b22ef5d348b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Event-aided[[:space:]]Dense[[:space:]]and[[:space:]]Continuous[[:space:]]Point[[:space:]]Tracking_[[:space:]]Everywhere[[:space:]]and[[:space:]]Anytime/9ca8a16e-beb5-412b-96ff-28204d9a5d86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Event-based[[:space:]]Tiny[[:space:]]Object[[:space:]]Detection_[[:space:]]A[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]and[[:space:]]Baseline/4ab2a4db-cd75-44f4-866e-39574b448472_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Event-based[[:space:]]Visual[[:space:]]Vibrometry/b38fc5f7-5d86-436d-8460-9c3eb27d9fce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Event-boosted[[:space:]]Deformable[[:space:]]3D[[:space:]]Gaussians[[:space:]]for[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Reconstruction/0f6daa46-496f-4815-9cc1-8b104f1b964b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Event-guided[[:space:]]HDR[[:space:]]Reconstruction[[:space:]]with[[:space:]]Diffusion[[:space:]]Priors/1d6e20a9-59f5-4c36-8c14-b470c2a5ccd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Event-guided[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Low-light[[:space:]]Video[[:space:]]Enhancement,[[:space:]]Frame[[:space:]]Interpolation,[[:space:]]and[[:space:]]Deblurring/57e7702d-451a-45d5-82e6-7003cf61c7f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EventUPS_[[:space:]]Uncalibrated[[:space:]]Photometric[[:space:]]Stereo[[:space:]]Using[[:space:]]an[[:space:]]Event[[:space:]]Camera/c1fa3d5f-f9b9-4099-a85c-b7d3f0d0b412_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Everything[[:space:]]is[[:space:]]a[[:space:]]Video_[[:space:]]Unifying[[:space:]]Modalities[[:space:]]through[[:space:]]Next-Frame[[:space:]]Prediction/992f5003-7f3c-4061-aa0e-ae51513cab64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Evidential[[:space:]]Knowledge[[:space:]]Distillation/c547b9b0-9570-47d5-bf3e-877b877331fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/EvolvingGrasp_[[:space:]]Evolutionary[[:space:]]Grasp[[:space:]]Generation[[:space:]]via[[:space:]]Efficient[[:space:]]Preference[[:space:]]Alignment/a5b9e976-b527-49c5-a219-2287871511f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ExCap3D_[[:space:]]Expressive[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding[[:space:]]via[[:space:]]Object[[:space:]]Captioning[[:space:]]with[[:space:]]Varying[[:space:]]Detail/6db99a4e-cf3e-43fb-908c-7fbd63bfc3bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Explaining[[:space:]]Human[[:space:]]Preferences[[:space:]]via[[:space:]]Metrics[[:space:]]for[[:space:]]Structured[[:space:]]3D[[:space:]]Reconstruction/daa34d93-214e-4447-9b72-957738436c31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploiting[[:space:]]Diffusion[[:space:]]Prior[[:space:]]for[[:space:]]Task-driven[[:space:]]Image[[:space:]]Restoration/1dbcf4df-da9a-471c-b4a5-1d3b5d46842e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploiting[[:space:]]Domain[[:space:]]Properties[[:space:]]in[[:space:]]Language-Driven[[:space:]]Domain[[:space:]]Generalization[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/0339a1b5-4ddb-43d4-a904-de29094899c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploiting[[:space:]]Frequency[[:space:]]Dynamics[[:space:]]for[[:space:]]Enhanced[[:space:]]Multimodal[[:space:]]Event-based[[:space:]]Action[[:space:]]Recognition/ff885410-5312-499d-bf42-f2819adc9e95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploiting[[:space:]]Vision[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Training-Free[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]OOD[[:space:]]Detection[[:space:]]via[[:space:]]Graph[[:space:]]Score[[:space:]]Propagation/eecf87ed-7051-4e02-88a7-ac0885e7ef9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ExploreGS_[[:space:]]Explorable[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]with[[:space:]]Virtual[[:space:]]Camera[[:space:]]Samplings[[:space:]]and[[:space:]]Diffusion[[:space:]]Priors/3e64bd2c-ab83-411a-816b-90e85430a7c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Multimodal[[:space:]]Diffusion[[:space:]]Transformers[[:space:]]for[[:space:]]Enhanced[[:space:]]Prompt-based[[:space:]]Image[[:space:]]Editing/db432361-cc2c-467e-9c54-baa03e561a43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Probabilistic[[:space:]]Modeling[[:space:]]Beyond[[:space:]]Domain[[:space:]]Generalization[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/293d6594-cb1e-4cf4-8140-8fc705f162e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]The[[:space:]]Visual[[:space:]]Feature[[:space:]]Space[[:space:]]for[[:space:]]Multimodal[[:space:]]Neural[[:space:]]Decoding/cf7c857e-63f8-4a94-a425-38a73f920d8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]View[[:space:]]Consistency[[:space:]]for[[:space:]]Scene-Adaptive[[:space:]]Low-Light[[:space:]]Light[[:space:]]Field[[:space:]]Image[[:space:]]Enhancement/2b9eae5a-6bd0-47a6-941b-39ead0bda71c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]Weather-aware[[:space:]]Aggregation[[:space:]]and[[:space:]]Adaptation[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]under[[:space:]]Adverse[[:space:]]Conditions/bcf1a36a-351d-4939-805f-0e0dec332561_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Exploring[[:space:]]the[[:space:]]Adversarial[[:space:]]Vulnerabilities[[:space:]]of[[:space:]]Vision-Language-Action[[:space:]]Models[[:space:]]in[[:space:]]Robotics/654da609-1a74-4dd3-8326-891848884c54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Expressive[[:space:]]Talking[[:space:]]Human[[:space:]]from[[:space:]]Single-Image[[:space:]]with[[:space:]]Imperfect[[:space:]]Priors/d3058941-5a3c-47b1-a217-ebdb5b94b775_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Extending[[:space:]]Foundational[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimators[[:space:]]to[[:space:]]Fisheye[[:space:]]Cameras[[:space:]]with[[:space:]]Calibration[[:space:]]Tokens/ff8cd9c3-7a4b-4914-bf82-9a51ee69f9ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/External[[:space:]]Knowledge[[:space:]]Injection[[:space:]]for[[:space:]]CLIP-Based[[:space:]]Class-Incremental[[:space:]]Learning/59db92d7-ff41-4079-a08f-c630a2445462_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Extrapolated[[:space:]]Urban[[:space:]]View[[:space:]]Synthesis[[:space:]]Benchmark/e65fa038-3ae6-4620-9f47-ad1a63db0abf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/F-Bench_[[:space:]]Rethinking[[:space:]]Human[[:space:]]Preference[[:space:]]Evaluation[[:space:]]Metrics[[:space:]]for[[:space:]]Benchmarking[[:space:]]Face[[:space:]]Generation,[[:space:]]Customization,[[:space:]]and[[:space:]]Restoration/86d5740f-dff5-47ab-aa67-761b97593a44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FALCON_[[:space:]]Resolving[[:space:]]Visual[[:space:]]Redundancy[[:space:]]and[[:space:]]Fragmentation[[:space:]]in[[:space:]]High-resolution[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]via[[:space:]]Visual[[:space:]]Registers/3c258128-6a38-46a9-bf2a-aabe422e5840_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FA_[[:space:]]Forced[[:space:]]Prompt[[:space:]]Learning[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection/8ae5a438-b943-4e71-b751-475d9f24c3b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FB-Diff_[[:space:]]Fourier[[:space:]]Basis-guided[[:space:]]Diffusion[[:space:]]for[[:space:]]Temporal[[:space:]]Interpolation[[:space:]]of[[:space:]]4D[[:space:]]Medical[[:space:]]Imaging/c8ad7765-483e-400a-90bd-72090b0597d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FDPT_[[:space:]]Federated[[:space:]]Discrete[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Black-Box[[:space:]]Visual-Language[[:space:]]Models/5d5a1b2f-9957-4fb4-b461-7bf3e1ebccfb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FE-CLIP_[[:space:]]Frequency[[:space:]]Enhanced[[:space:]]CLIP[[:space:]]Model[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Anomaly[[:space:]]Detection[[:space:]]and[[:space:]]Segmentation/e20fdad0-0c02-434f-a4f5-0e729b38aab0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FED-PsyAU_[[:space:]]Privacy-Preserving[[:space:]]Micro-Expression[[:space:]]Recognition[[:space:]]via[[:space:]]Psychological[[:space:]]AU[[:space:]]Coordination[[:space:]]and[[:space:]]Dynamic[[:space:]]Facial[[:space:]]Motion[[:space:]]Modeling/e27ad4a5-3106-4216-bcea-a95709cd5f65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FEVER-OOD_[[:space:]]Free[[:space:]]Energy[[:space:]]Vulnerability[[:space:]]Elimination[[:space:]]for[[:space:]]Robust[[:space:]]Out-of-Distribution[[:space:]]Detection/0c16965c-d6fa-4547-a6ba-ab34512ce472_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FG-OrIU_[[:space:]]Towards[[:space:]]Better[[:space:]]Forgetting[[:space:]]via[[:space:]]Feature-Gradient[[:space:]]Orthogonality[[:space:]]for[[:space:]]Incremental[[:space:]]Unlearning/4a750062-438a-4c8c-b4db-9a6576c3de7a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FICGen_[[:space:]]Frequency-Inspired[[:space:]]Contextual[[:space:]]Disentanglement[[:space:]]for[[:space:]]Layout-driven[[:space:]]Degraded[[:space:]]Image[[:space:]]Generation/24509018-dca5-4420-ad1a-dd11a5974d9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FIND_[[:space:]]Few-Shot[[:space:]]Anomaly[[:space:]]Inspection[[:space:]]with[[:space:]]Normal-Only[[:space:]]Multi-Modal[[:space:]]Data/94dd2e05-5a4d-4859-bd44-9496d30c1756_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FLOAT_[[:space:]]Generative[[:space:]]Motion[[:space:]]Latent[[:space:]]Flow[[:space:]]Matching[[:space:]]for[[:space:]]Audio-driven[[:space:]]Talking[[:space:]]Portrait/5eb040a4-3c63-4680-a42f-da4ce86274a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FLOSS_[[:space:]]Free[[:space:]]Lunch[[:space:]]in[[:space:]]Open-vocabulary[[:space:]]Semantic[[:space:]]Segmentation/759adae1-3d57-4b38-8780-a8c8f1ca245e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FLSeg_[[:space:]]Enhancing[[:space:]]Privacy[[:space:]]and[[:space:]]Robustness[[:space:]]in[[:space:]]Federated[[:space:]]Learning[[:space:]]under[[:space:]]Heterogeneous[[:space:]]Data[[:space:]]via[[:space:]]Model[[:space:]]Segmentation/a4f90945-afa0-4f01-8edd-1b86f5d9d098_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FOLDER_[[:space:]]Accelerating[[:space:]]Multi-Modal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]with[[:space:]]Enhanced[[:space:]]Performance/7a17b26d-c0b2-4285-9105-7de16ee42381_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FPEM_[[:space:]]Face[[:space:]]Prior[[:space:]]Enhanced[[:space:]]Facial[[:space:]]Attractiveness[[:space:]]Prediction[[:space:]]for[[:space:]]Live[[:space:]]Videos[[:space:]]with[[:space:]]Face[[:space:]]Retouching/f47f8d6d-8a6c-491b-8f19-08123eb17e5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FREE-Merging_[[:space:]]Fourier[[:space:]]Transform[[:space:]]for[[:space:]]Efficient[[:space:]]Model[[:space:]]Merging/f3aa67c5-53d3-463b-8730-22ad4f95fe7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FRET_[[:space:]]Feature[[:space:]]Redundancy[[:space:]]Elimination[[:space:]]for[[:space:]]Test[[:space:]]Time[[:space:]]Adaptation/013e8d33-0504-4dd6-a3e9-326514c12e30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FROSS_[[:space:]]Faster-Than-Real-Time[[:space:]]Online[[:space:]]3D[[:space:]]Semantic[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]from[[:space:]]RGB-D[[:space:]]Images/f4b25ed8-2175-4e29-a597-e457f4795510_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FVGen_[[:space:]]Accelerating[[:space:]]Novel-View[[:space:]]Synthesis[[:space:]]with[[:space:]]Adversarial[[:space:]]Video[[:space:]]Diffusion[[:space:]]Distillation/381cfeb4-88c2-44b1-a694-860cc7f57161_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FW-Merging_[[:space:]]Scaling[[:space:]]Model[[:space:]]Merging[[:space:]]with[[:space:]]Frank-Wolfe[[:space:]]Optimization/56e48ade-9cb7-4688-a0be-edb71b4d459b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Face[[:space:]]Retouching[[:space:]]with[[:space:]]Diffusion[[:space:]]Data[[:space:]]Generation[[:space:]]and[[:space:]]Spectral[[:space:]]Restorement/2a12f446-c130-4e36-bc58-dba590a7c6d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FaceCraft4D_[[:space:]]Animated[[:space:]]3D[[:space:]]Facial[[:space:]]Avatar[[:space:]]Generation[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/c512f2aa-6fbd-4940-89be-2245ba51ffe2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FaceLift_[[:space:]]Learning[[:space:]]Generalizable[[:space:]]Single[[:space:]]Image[[:space:]]3D[[:space:]]Face[[:space:]]Reconstruction[[:space:]]from[[:space:]]Synthetic[[:space:]]Heads/1763108f-f851-4f26-ba5a-510a2c051a36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FaceShield_[[:space:]]Defending[[:space:]]Facial[[:space:]]Image[[:space:]]against[[:space:]]Deepfake[[:space:]]Threats/4436b2e6-eb11-4750-8df1-5fa6df4f36fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FaceXFormer_[[:space:]]A[[:space:]]Unified[[:space:]]Transformer[[:space:]]for[[:space:]]Facial[[:space:]]Analysis/43f275cd-265b-414a-a63d-4d3a6e26c5b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Factorized[[:space:]]Learning[[:space:]]for[[:space:]]Temporally[[:space:]]Grounded[[:space:]]Video-Language[[:space:]]Models/0391813b-b98c-4912-884e-a194864058af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Failure[[:space:]]Cases[[:space:]]Are[[:space:]]Better[[:space:]]Learned[[:space:]]But[[:space:]]Boundary[[:space:]]Says[[:space:]]Sorry_[[:space:]]Facilitating[[:space:]]Smooth[[:space:]]Perception[[:space:]]Change[[:space:]]for[[:space:]]Accuracy-Robustness[[:space:]]Trade-Off[[:space:]]in[[:space:]]Adversarial[[:space:]]Training/9886c220-8af0-438e-8e12-4de3189d4e86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fair[[:space:]]Generation[[:space:]]without[[:space:]]Unfair[[:space:]]Distortions_[[:space:]]Debiasing[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]with[[:space:]]Entanglement-Free[[:space:]]Attention/6ef80803-cc14-446f-a75b-f2593f51cc3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FairGen_[[:space:]]Enhancing[[:space:]]Fairness[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Self-Discovering[[:space:]]Latent[[:space:]]Directions/207183e6-db8a-436f-8e4f-faec1fb9d3a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FairHuman_[[:space:]]Boosting[[:space:]]Hand[[:space:]]and[[:space:]]Face[[:space:]]Quality[[:space:]]in[[:space:]]Human[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Minimum[[:space:]]Potential[[:space:]]Delay[[:space:]]Fairness[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/c1cadd0b-ab93-4bff-a36b-6e24c8c2c52f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FakeRadar_[[:space:]]Probing[[:space:]]Forgery[[:space:]]Outliers[[:space:]]to[[:space:]]Detect[[:space:]]Unknown[[:space:]]Deepfake[[:space:]]Videos/28a1e3bc-e2b9-40da-b7fe-fcc0d6369340_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fast[[:space:]]Globally[[:space:]]Optimal[[:space:]]and[[:space:]]Geometrically[[:space:]]Consistent[[:space:]]3D[[:space:]]Shape[[:space:]]Matching/cbbb1f39-340b-4752-9244-371644b15edf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fast[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Consistency[[:space:]]Rectified[[:space:]]Flow/5e42c446-52f8-4769-b738-a2b096f3a7a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FastJSMA_[[:space:]]Accelerating[[:space:]]Jacobian-based[[:space:]]Saliency[[:space:]]Map[[:space:]]Attacks[[:space:]]through[[:space:]]Gradient[[:space:]]Decoupling/caddbdfc-9052-49a6-8b20-7e4e162ec5a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FastPoint_[[:space:]]Accelerating[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Model[[:space:]]Inference[[:space:]]via[[:space:]]Sample[[:space:]]Point[[:space:]]Distance[[:space:]]Prediction/670cb46f-786b-455d-998f-328c26d7e47d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FastVAR_[[:space:]]Linear[[:space:]]Visual[[:space:]]Autoregressive[[:space:]]Modeling[[:space:]]via[[:space:]]Cached[[:space:]]Token[[:space:]]Pruning/70725db1-fcd5-4496-a95d-07038d174933_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Faster[[:space:]]and[[:space:]]Better[[:space:]]3D[[:space:]]Splatting[[:space:]]via[[:space:]]Group[[:space:]]Training/21033709-b705-4347-a1bb-c3e7a58591e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Feather[[:space:]]the[[:space:]]Throttle_[[:space:]]Revisiting[[:space:]]Visual[[:space:]]Token[[:space:]]Pruning[[:space:]]for[[:space:]]Vision-Language[[:space:]]Model[[:space:]]Acceleration/eb2a11ee-50f9-4c43-8cc0-429a524795eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Feature[[:space:]]Coding[[:space:]]in[[:space:]]the[[:space:]]Era[[:space:]]of[[:space:]]Large[[:space:]]Models_[[:space:]]Dataset,[[:space:]]Test[[:space:]]Conditions,[[:space:]]and[[:space:]]Benchmark/83895aed-d59c-465f-afa3-c8ed37ae4a4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Feature[[:space:]]Decomposition-Recomposition[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Model[[:space:]]for[[:space:]]Few-Shot[[:space:]]Class-Incremental[[:space:]]Learning/df51b4e9-12ac-40bf-8801-dbd6097351a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Feature[[:space:]]Extraction[[:space:]]and[[:space:]]Representation[[:space:]]of[[:space:]]Pre-training[[:space:]]Point[[:space:]]Cloud[[:space:]]Based[[:space:]]on[[:space:]]Diffusion[[:space:]]Models/0c060285-4bbf-430f-b421-79f016d6dd03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Feature[[:space:]]Purification[[:space:]]Matters_[[:space:]]Suppressing[[:space:]]Outlier[[:space:]]Propagation[[:space:]]for[[:space:]]Training-Free[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/4f979197-bed4-425a-a1d2-c2b5ca2160b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedAGC_[[:space:]]Federated[[:space:]]Continual[[:space:]]Learning[[:space:]]with[[:space:]]Asymmetric[[:space:]]Gradient[[:space:]]Correction/69b6bc67-ec8c-4e1a-b957-09f50cc21aab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedDifRC_[[:space:]]Unlocking[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]in[[:space:]]Heterogeneous[[:space:]]Federated[[:space:]]Learning/d48a8fed-132a-4154-9395-1fe6c487ac0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedMVP_[[:space:]]Federated[[:space:]]Multimodal[[:space:]]Visual[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/9a8979ca-f59a-40f4-a042-065bd341c994_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedMeNF_[[:space:]]Privacy-Preserving[[:space:]]Federated[[:space:]]Meta-Learning[[:space:]]for[[:space:]]Neural[[:space:]]Fields/97cb26d5-5659-4687-b57e-cc32fd64dcc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedPall_[[:space:]]Prototype-based[[:space:]]Adversarial[[:space:]]and[[:space:]]Collaborative[[:space:]]Learning[[:space:]]for[[:space:]]Federated[[:space:]]Learning[[:space:]]with[[:space:]]Feature[[:space:]]Drift/f6c6ce84-c8dc-4de0-bad0-3f6c4ed86bd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedVLA_[[:space:]]Federated[[:space:]]Vision-Language-Action[[:space:]]Learning[[:space:]]with[[:space:]]Dual[[:space:]]Gating[[:space:]]Mixture-of-Experts[[:space:]]for[[:space:]]Robotic[[:space:]]Manipulation/9830aa12-506c-4606-9232-07331ac373a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedWSQ_[[:space:]]Efficient[[:space:]]Federated[[:space:]]Learning[[:space:]]with[[:space:]]Weight[[:space:]]Standardization[[:space:]]and[[:space:]]Distribution-Aware[[:space:]]Non-Uniform[[:space:]]Quantization/1f166dee-20e5-452e-8e86-86019ce8d0bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FedXDS_[[:space:]]Leveraging[[:space:]]Model[[:space:]]Attribution[[:space:]]Methods[[:space:]]to[[:space:]]counteract[[:space:]]Data[[:space:]]Heterogeneity[[:space:]]in[[:space:]]Federated[[:space:]]Learning/3d4823b4-ec38-431a-8ed1-cb351135854c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Federated[[:space:]]Continual[[:space:]]Instruction[[:space:]]Tuning/c95b9507-af86-4883-a6e8-057630ebe555_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Federated[[:space:]]Continuous[[:space:]]Category[[:space:]]Discovery[[:space:]]and[[:space:]]Learning/7332a50f-74a4-4833-9616-ad03014bd631_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Federated[[:space:]]Domain[[:space:]]Generalization[[:space:]]with[[:space:]]Domain-specific[[:space:]]Soft[[:space:]]Prompts[[:space:]]Generation/b70289dd-c48a-4e53-8732-8d816d04b0c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Federated[[:space:]]Prompt-Tuning[[:space:]]with[[:space:]]Heterogeneous[[:space:]]and[[:space:]]Incomplete[[:space:]]Multimodal[[:space:]]Client[[:space:]]Data/c4864325-3bcd-4609-87ea-0a89ff648c25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Federated[[:space:]]Representation[[:space:]]Angle[[:space:]]Learning/f94f188a-83f1-436a-a6d8-36878318e078_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Feed-Forward[[:space:]]SceneDINO[[:space:]]for[[:space:]]Unsupervised[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion/0cb8dcd6-d20d-4d18-a810-629c106f3cdd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Few-Shot[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment[[:space:]]via[[:space:]]Adaptation[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models/89cdaf16-5828-4989-b825-83012badf535_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Few-Shot[[:space:]]Pattern[[:space:]]Detection[[:space:]]via[[:space:]]Template[[:space:]]Matching[[:space:]]and[[:space:]]Regression/9a1fc29b-9538-4897-943a-30b2211079d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fewer[[:space:]]Denoising[[:space:]]Steps[[:space:]]or[[:space:]]Cheaper[[:space:]]Per-Step[[:space:]]Inference_[[:space:]]Towards[[:space:]]Compute-Optimal[[:space:]]Diffusion[[:space:]]Model[[:space:]]Deployment/98cafbf9-4954-46a4-b5af-59f0bca621f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FiVE-Bench_[[:space:]]A[[:space:]]Fine-grained[[:space:]]Video[[:space:]]Editing[[:space:]]Benchmark[[:space:]]for[[:space:]]Evaluating[[:space:]]Emerging[[:space:]]Diffusion[[:space:]]and[[:space:]]Rectified[[:space:]]Flow[[:space:]]Models/61b80164-0e5a-42b8-b65d-d1598a219bbd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FiffDepth_[[:space:]]Feed-forward[[:space:]]Transformation[[:space:]]of[[:space:]]Diffusion-Based[[:space:]]Generators[[:space:]]for[[:space:]]Detailed[[:space:]]Depth[[:space:]]Estimation/3c8dc7fb-98b2-4d77-857e-35d78d6eaaf6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FinMMR_[[:space:]]Make[[:space:]]Financial[[:space:]]Numerical[[:space:]]Reasoning[[:space:]]More[[:space:]]Multimodal,[[:space:]]Comprehensive,[[:space:]]and[[:space:]]Challenging/9bdd874d-4c07-4812-a904-cd4ae1c8c840_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Find[[:space:]]Any[[:space:]]Part[[:space:]]in[[:space:]]3D/088883ee-ac1f-4d70-9944-8cc27e53fb77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Find[[:space:]]a[[:space:]]Scapegoat_[[:space:]]Poisoning[[:space:]]Membership[[:space:]]Inference[[:space:]]Attack[[:space:]]and[[:space:]]Defense[[:space:]]to[[:space:]]Federated[[:space:]]Learning/529b2c82-4128-4ab4-85c4-8de66f3e2ff1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fine-Grained[[:space:]]3D[[:space:]]Gaussian[[:space:]]Head[[:space:]]Avatars[[:space:]]Modeling[[:space:]]from[[:space:]]Static[[:space:]]Captures[[:space:]]via[[:space:]]Joint[[:space:]]Reconstruction[[:space:]]and[[:space:]]Registration/e721192b-de83-45d8-a5bf-c03ac12cb270_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fine-Grained[[:space:]]Evaluation[[:space:]]of[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/1969849d-58d7-41eb-87cc-aa7a483837e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fine-Tuning[[:space:]]Visual[[:space:]]Autogressive[[:space:]]Models[[:space:]]for[[:space:]]Subject-Driven[[:space:]]Generation/4c8031ed-11c5-42f3-b203-ffc0a08469b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fine-grained[[:space:]]Abnormality[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Zero-shot[[:space:]]Anomaly[[:space:]]Detection/c663af67-c6a8-4485-993d-8664f70e6bb9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fine-grained[[:space:]]Spatiotemporal[[:space:]]Grounding[[:space:]]on[[:space:]]Egocentric[[:space:]]Videos/0b2cb819-3068-430b-af48-9815e881ee12_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fine-structure[[:space:]]Preserved[[:space:]]Real-world[[:space:]]Image[[:space:]]Super-resolution[[:space:]]via[[:space:]]Transfer[[:space:]]VAE[[:space:]]Training/110e15b4-f11c-4a8d-9953-ae73df5ba306_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FineMotion_[[:space:]]A[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]with[[:space:]]both[[:space:]]Spatial[[:space:]]and[[:space:]]Temporal[[:space:]]Annotation[[:space:]]for[[:space:]]Fine-grained[[:space:]]Motion[[:space:]]Generation[[:space:]]and[[:space:]]Editing/80db9b9b-c681-4aec-a6dd-4f83bcbb0ab5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fish2Mesh[[:space:]]Transformer_[[:space:]]3D[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery[[:space:]]from[[:space:]]Egocentric[[:space:]]Vision/7ecd7e4c-8975-47e0-9687-e43c6d65f935_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fix-CLIP_[[:space:]]Dual-Branch[[:space:]]Hierarchical[[:space:]]Contrastive[[:space:]]Learning[[:space:]]via[[:space:]]Synthetic[[:space:]]Captions[[:space:]]for[[:space:]]Better[[:space:]]Understanding[[:space:]]of[[:space:]]Long[[:space:]]Text/428df74e-fc9d-4700-83c2-40f9e1a40319_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FixTalk_[[:space:]]Taming[[:space:]]Identity[[:space:]]Leakage[[:space:]]for[[:space:]]High-Quality[[:space:]]Talking[[:space:]]Head[[:space:]]Generation[[:space:]]in[[:space:]]Extreme[[:space:]]Cases/9391ce94-be2f-448f-b470-1aa69c368bd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Flash-VStream_[[:space:]]Efficient[[:space:]]Real-Time[[:space:]]Understanding[[:space:]]for[[:space:]]Long[[:space:]]Video[[:space:]]Streams/9ad578d2-2a13-4906-ab86-4e34da57d59b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlashDepth_[[:space:]]Real-time[[:space:]]Streaming[[:space:]]Video[[:space:]]Depth[[:space:]]Estimation[[:space:]]at[[:space:]]2K[[:space:]]Resolution/fd24c514-b148-4ed8-be2a-11f1bfe644f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlexGen_[[:space:]]Flexible[[:space:]]Multi-View[[:space:]]Generation[[:space:]]from[[:space:]]Text[[:space:]]and[[:space:]]Image[[:space:]]Inputs/8f7f66f5-df16-4bd1-a6e7-91c1d235f1ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Flexi-FSCIL_[[:space:]]Adaptive[[:space:]]Knowledge[[:space:]]Retention[[:space:]]for[[:space:]]Breaking[[:space:]]the[[:space:]]Stability-Plasticity[[:space:]]Dilemma[[:space:]]in[[:space:]]Few-Shot[[:space:]]Class-Incremental[[:space:]]Learning/2a182704-fa79-4364-b9ae-369b55239af6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Flow[[:space:]]Stochastic[[:space:]]Segmentation[[:space:]]Networks/98f25e18-74ac-4df9-8ae4-e60407cdb72e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Flow[[:space:]]to[[:space:]]the[[:space:]]Mode_[[:space:]]Mode-Seeking[[:space:]]Diffusion[[:space:]]Autoencoders[[:space:]]for[[:space:]]State-of-the-Art[[:space:]]Image[[:space:]]Tokenization/9aa9d640-ba2a-4d49-a011-61b42254c482_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Flow-MIL_[[:space:]]Constructing[[:space:]]Highly-expressive[[:space:]]Latent[[:space:]]Feature[[:space:]]Space[[:space:]]For[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Classification[[:space:]]Using[[:space:]]Normalizing[[:space:]]Flow/e00429a2-e9dd-4786-8652-0defd26eca6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Flow4Agent_[[:space:]]Long-form[[:space:]]Video[[:space:]]Understanding[[:space:]]via[[:space:]]Motion[[:space:]]Prior[[:space:]]from[[:space:]]Optical[[:space:]]Flow/a8de8822-13b0-4c9d-a15d-074c512b363e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlowChef_[[:space:]]Steering[[:space:]]of[[:space:]]Rectified[[:space:]]Flow[[:space:]]Models[[:space:]]for[[:space:]]Controlled[[:space:]]Generations/5b8d5c2c-0f10-4505-9f9e-ad746c713864_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlowDPS[[:space:]]_[[:space:]]Flow-Driven[[:space:]]Posterior[[:space:]]Sampling[[:space:]]for[[:space:]]Inverse[[:space:]]Problems/7679c8c5-c036-4af9-bdfc-f8b12469916a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlowEdit_[[:space:]]Inversion-Free[[:space:]]Text-Based[[:space:]]Editing[[:space:]]Using[[:space:]]Pre-Trained[[:space:]]Flow[[:space:]]Models/1c8b1f48-a6c8-4cb1-992b-2c4462ee41b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlowR_[[:space:]]Flowing[[:space:]]from[[:space:]]Sparse[[:space:]]to[[:space:]]Dense[[:space:]]3D[[:space:]]Reconstructions/ac5e994f-2583-496c-9765-563e409855b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlowSeek_[[:space:]]Optical[[:space:]]Flow[[:space:]]Made[[:space:]]Easier[[:space:]]with[[:space:]]Depth[[:space:]]Foundation[[:space:]]Models[[:space:]]and[[:space:]]Motion[[:space:]]Bases/90effc05-772b-48e4-a532-709f75e577c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlowStyler_[[:space:]]Artistic[[:space:]]Video[[:space:]]Stylization[[:space:]]via[[:space:]]Transformation[[:space:]]Fields[[:space:]]Transports/e9014c01-5080-4fde-acec-58a7aa6f1e44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FlowTok_[[:space:]]Flowing[[:space:]]Seamlessly[[:space:]]Across[[:space:]]Text[[:space:]]and[[:space:]]Image[[:space:]]Tokens/bad6735a-ab7b-4fd7-b3ec-c2f16b242765_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Focal[[:space:]]Plane[[:space:]]Visual[[:space:]]Feature[[:space:]]Generation[[:space:]]and[[:space:]]Matching[[:space:]]on[[:space:]]a[[:space:]]Pixel[[:space:]]Processor[[:space:]]Array/5d8ea9dc-5577-4013-be82-0616509dab4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FonTS_[[:space:]]Text[[:space:]]Rendering[[:space:]]With[[:space:]]Typography[[:space:]]and[[:space:]]Style[[:space:]]Controls/c426027e-1b67-47d6-a1fb-b0a38adf370a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FontAnimate_[[:space:]]High[[:space:]]Quality[[:space:]]Few-shot[[:space:]]Font[[:space:]]Generation[[:space:]]via[[:space:]]Animating[[:space:]]Font[[:space:]]Transfer[[:space:]]Process/f97d64b9-f43f-4137-bc3b-3b03f59bee8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ForCenNet_[[:space:]]Foreground-Centric[[:space:]]Network[[:space:]]for[[:space:]]Document[[:space:]]Image[[:space:]]Rectification/e18bc319-5145-4449-af12-e7e32480fbf2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ForeSight_[[:space:]]Multi-View[[:space:]]Streaming[[:space:]]Joint[[:space:]]Object[[:space:]]Detection[[:space:]]and[[:space:]]Trajectory[[:space:]]Forecasting/6edfed27-4444-47d5-81cc-bac42b8ea54d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Forecasting[[:space:]]Continuous[[:space:]]Non-Conservative[[:space:]]Dynamical[[:space:]]Systems[[:space:]]in[[:space:]]SO(3)/6e532ef7-f69a-4619-bc42-d33cf0eff731_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Forensic-MoE_[[:space:]]Exploring[[:space:]]Comprehensive[[:space:]]Synthetic[[:space:]]Image[[:space:]]Detection[[:space:]]Traces[[:space:]]with[[:space:]]Mixture[[:space:]]of[[:space:]]Experts/73cf7ecc-8bce-413c-ac62-4b0166782071_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Foresight[[:space:]]in[[:space:]]Motion_[[:space:]]Reinforcing[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]with[[:space:]]Reward[[:space:]]Heuristics/924f0bb3-0632-4ff5-bfe9-1be679b5873e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ForestFormer3D_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]End-to-End[[:space:]]Segmentation[[:space:]]of[[:space:]]Forest[[:space:]]LiDAR[[:space:]]3D[[:space:]]Point[[:space:]]Clouds/16ed1bf5-1877-4771-854d-69549d568323_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ForgeLens_[[:space:]]Data-Efficient[[:space:]]Forgery[[:space:]]Focus[[:space:]]for[[:space:]]Generalizable[[:space:]]Forgery[[:space:]]Image[[:space:]]Detection/a8bc193f-bb1d-4e2f-9837-d5fc5db302bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Forgetting[[:space:]]Through[[:space:]]Transforming_[[:space:]]Enabling[[:space:]]Federated[[:space:]]Unlearning[[:space:]]via[[:space:]]Class-Aware[[:space:]]Representation[[:space:]]Transformation/9fe83397-aa4f-4b52-a522-ae3f3a68fe61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FoundIR_[[:space:]]Unleashing[[:space:]]Million-scale[[:space:]]Training[[:space:]]Data[[:space:]]to[[:space:]]Advance[[:space:]]Foundation[[:space:]]Models[[:space:]]for[[:space:]]Image[[:space:]]Restoration/e5d26f60-fb4f-45cf-ac64-c2f306864c44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FrameFusion_[[:space:]]Combining[[:space:]]Similarity[[:space:]]and[[:space:]]Importance[[:space:]]for[[:space:]]Video[[:space:]]Token[[:space:]]Reduction[[:space:]]on[[:space:]]Large[[:space:]]Vision[[:space:]]Language[[:space:]]Models/9ef8a426-5732-4b21-b0be-ad7214f778c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FramePainter_[[:space:]]Endowing[[:space:]]Interactive[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Video[[:space:]]Diffusion[[:space:]]Priors/bd8826c0-cdff-4b90-87f0-caf2864a0668_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Free-Form[[:space:]]Motion[[:space:]]Control_[[:space:]]Controlling[[:space:]]the[[:space:]]6D[[:space:]]Poses[[:space:]]of[[:space:]]Camera[[:space:]]and[[:space:]]Objects[[:space:]]in[[:space:]]Video[[:space:]]Generation/eb7360e3-1ee8-4e4e-bc4c-bef05be637d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Free-MoRef_[[:space:]]Instantly[[:space:]]Multiplexing[[:space:]]Context[[:space:]]Perception[[:space:]]Capabilities[[:space:]]of[[:space:]]Video-MLLMs[[:space:]]within[[:space:]]Single[[:space:]]Inference/9095ddab-4ee4-4e80-94b2-8fc53f91b6ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Free-running[[:space:]]vs[[:space:]]Synchronous_[[:space:]]Single-Photon[[:space:]]Lidar[[:space:]]for[[:space:]]High-flux[[:space:]]3D[[:space:]]Imaging/fb37677e-353c-4aed-9a08-8c1e5111719e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Free2Guide_[[:space:]]Training-Free[[:space:]]Text-to-Video[[:space:]]Alignment[[:space:]]using[[:space:]]Image[[:space:]]LVLM/0e576880-5a47-4745-8bcf-fe2e833fa9d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Free4D_[[:space:]]Tuning-free[[:space:]]4D[[:space:]]Scene[[:space:]]Generation[[:space:]]with[[:space:]]Spatial-Temporal[[:space:]]Consistency/cdc3492c-4593-49ba-b611-a94468f26bc7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeCus_[[:space:]]Free[[:space:]]Lunch[[:space:]]Subject-driven[[:space:]]Customization[[:space:]]in[[:space:]]Diffusion[[:space:]]Transformers/4f8d8464-6452-45fc-933e-73e2bc913a36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeDNA_[[:space:]]Endowing[[:space:]]Domain[[:space:]]Adaptation[[:space:]]of[[:space:]]Diffusion-Based[[:space:]]Dense[[:space:]]Prediction[[:space:]]with[[:space:]]Training-Free[[:space:]]Domain[[:space:]]Noise[[:space:]]Alignment/f11f5cc6-1ff8-4655-a5d7-fbaf2507112b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeDance_[[:space:]]Towards[[:space:]]Harmonic[[:space:]]Free-Number[[:space:]]Group[[:space:]]Dance[[:space:]]Generation[[:space:]]via[[:space:]]a[[:space:]]Unified[[:space:]]Framework/53ab2a5f-cf9d-47e6-a6dd-d6374afaccd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeFlux_[[:space:]]Understanding[[:space:]]and[[:space:]]Exploiting[[:space:]]Layer-Specific[[:space:]]Roles[[:space:]]in[[:space:]]RoPE-Based[[:space:]]MMDiT[[:space:]]for[[:space:]]Versatile[[:space:]]Image[[:space:]]Editing/16c6f9f6-b989-4263-b033-0446362bf67a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeMorph_[[:space:]]Tuning-Free[[:space:]]Generalized[[:space:]]Image[[:space:]]Morphing[[:space:]]with[[:space:]]Diffusion[[:space:]]Model/7970cfd7-5bde-487d-8eb6-ce796d12d39f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeScale_[[:space:]]Unleashing[[:space:]]the[[:space:]]Resolution[[:space:]]of[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Tuning-Free[[:space:]]Scale[[:space:]]Fusion/49bf0fb1-1fce-4a09-b3da-7af66df60f07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreeSplatter_[[:space:]]Pose-free[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Sparse-view[[:space:]]3D[[:space:]]Reconstruction/eb0c0357-e174-472d-ae76-5c015ea2146c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FreqPDE_[[:space:]]Rethinking[[:space:]]Positional[[:space:]]Depth[[:space:]]Embedding[[:space:]]for[[:space:]]Multi-View[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]Transformers/afdb0ad4-8cac-4d1a-81fc-490dc4658926_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Frequency[[:space:]]Domain-Based[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Unpaired[[:space:]]Image[[:space:]]Dehazing/bbf8e2c6-5bf5-44eb-83ff-45eab337557f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Frequency-Aligned[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Lightweight[[:space:]]Spatiotemporal[[:space:]]Forecasting/2b08e445-be74-4d88-8397-c256a08162fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Frequency-Aware[[:space:]]Autoregressive[[:space:]]Modeling[[:space:]]for[[:space:]]Efficient[[:space:]]High-Resolution[[:space:]]Image[[:space:]]Synthesis/2d16156b-7563-42bc-87ce-0374e9392409_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Frequency-Dynamic[[:space:]]Attention[[:space:]]Modulation[[:space:]]For[[:space:]]Dense[[:space:]]Prediction/49885eb4-685a-47e8-8154-f478faea0bac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Frequency-Guided[[:space:]]Diffusion[[:space:]]for[[:space:]]Training-Free[[:space:]]Text-Driven[[:space:]]Image[[:space:]]Translation/4f83a419-e8b4-4f95-a5c7-9236f905e22b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Frequency-Guided[[:space:]]Posterior[[:space:]]Sampling[[:space:]]for[[:space:]]Diffusion-Based[[:space:]]Image[[:space:]]Restoration/3ffa9567-b938-404f-b8a2-375aa0913ea0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Frequency-Semantic[[:space:]]Enhanced[[:space:]]Variational[[:space:]]Autoencoder[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Skeleton-based[[:space:]]Action[[:space:]]Recognition/af432848-898b-432f-970c-3ce1b28396bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Abyssal[[:space:]]Darkness[[:space:]]to[[:space:]]Blinding[[:space:]]Glare_[[:space:]]A[[:space:]]Benchmark[[:space:]]on[[:space:]]Extreme[[:space:]]Exposure[[:space:]]Correction[[:space:]]in[[:space:]]Real[[:space:]]World/f710cad2-28b7-4ec5-9a70-eafcaaa2c099_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Easy[[:space:]]to[[:space:]]Hard_[[:space:]]Progressive[[:space:]]Active[[:space:]]Learning[[:space:]]Framework[[:space:]]for[[:space:]]Infrared[[:space:]]Small[[:space:]]Target[[:space:]]Detection[[:space:]]with[[:space:]]Single[[:space:]]Point[[:space:]]Supervision/05052ad4-e739-4bea-a923-f80426b52726_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Easy[[:space:]]to[[:space:]]Hard_[[:space:]]The[[:space:]]MIR[[:space:]]Benchmark[[:space:]]for[[:space:]]Progressive[[:space:]]Interleaved[[:space:]]Multi-Image[[:space:]]Reasoning/01d4b9f8-bd37-4104-aabf-c3db1b96b72b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Enhancement[[:space:]]to[[:space:]]Understanding_[[:space:]]Build[[:space:]]a[[:space:]]Generalized[[:space:]]Bridge[[:space:]]for[[:space:]]Low-light[[:space:]]Vision[[:space:]]via[[:space:]]Semantically[[:space:]]Consistent[[:space:]]Unsupervised[[:space:]]Fine-tuning/e55a76ae-0fb6-44ab-9b25-1103d787ff2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Gallery[[:space:]]to[[:space:]]Wrist_[[:space:]]Realistic[[:space:]]3D[[:space:]]Bracelet[[:space:]]Insertion[[:space:]]in[[:space:]]Videos/9d437b61-f674-4746-a483-3673b5a50f92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Gaze[[:space:]]to[[:space:]]Movement_[[:space:]]Predicting[[:space:]]Visual[[:space:]]Attention[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Human-Machine[[:space:]]Interaction[[:space:]]based[[:space:]]on[[:space:]]Programmatic[[:space:]]Imitation[[:space:]]Learning/701fd97a-bb3b-489c-a64b-61b055938b11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Holistic[[:space:]]to[[:space:]]Localized_[[:space:]]Local[[:space:]]Enhanced[[:space:]]Adapters[[:space:]]for[[:space:]]Efficient[[:space:]]Visual[[:space:]]Instruction[[:space:]]Fine-Tuning/5f7cfb96-d6ed-45b5-8e88-e19614984a0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Image[[:space:]]to[[:space:]]Video_[[:space:]]An[[:space:]]Empirical[[:space:]]Study[[:space:]]of[[:space:]]Diffusion[[:space:]]Representations/df5b2695-d7fa-4e6f-89e4-a8026db528fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Imitation[[:space:]]to[[:space:]]Innovation_[[:space:]]The[[:space:]]Emergence[[:space:]]of[[:space:]]AI's[[:space:]]Unique[[:space:]]Artistic[[:space:]]Styles[[:space:]]and[[:space:]]the[[:space:]]Challenge[[:space:]]of[[:space:]]Copyright[[:space:]]Protection/36cebc77-82bf-4089-92ca-13a4a418d841_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Linearity[[:space:]]to[[:space:]]Non-Linearity_[[:space:]]How[[:space:]]Masked[[:space:]]Autoencoders[[:space:]]Capture[[:space:]]Spatial[[:space:]]Correlations/9cd48e9f-64a3-413e-aa61-aad50c5ff7cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Objects[[:space:]]to[[:space:]]Events_[[:space:]]Unlocking[[:space:]]Complex[[:space:]]Visual[[:space:]]Understanding[[:space:]]in[[:space:]]Object[[:space:]]Detectors[[:space:]]via[[:space:]]LLM-guided[[:space:]]Symbolic[[:space:]]Reasoning/0c2f11d3-6ae3-4354-b8f2-929b7007c549_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]One[[:space:]]to[[:space:]]More_[[:space:]]Contextual[[:space:]]Part[[:space:]]Latents[[:space:]]for[[:space:]]3D[[:space:]]Generation/91d22616-cfab-4a3f-86cb-12e1f2b543d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Panels[[:space:]]to[[:space:]]Prose_[[:space:]]Generating[[:space:]]Literary[[:space:]]Narratives[[:space:]]from[[:space:]]Comics/e88a5226-331b-4c9e-bd6a-0caf37e3320c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Prompt[[:space:]]to[[:space:]]Progression_[[:space:]]Taming[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Seamless[[:space:]]Attribute[[:space:]]Transition/fa7224b2-b16c-4edf-a8c4-64f6f029d149_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Reflection[[:space:]]to[[:space:]]Perfection_[[:space:]]Scaling[[:space:]]Inference-Time[[:space:]]Optimization[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Reflection[[:space:]]Tuning/66756380-6b1f-4a6c-b620-8dfb6ee3c97a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Reusing[[:space:]]to[[:space:]]Forecasting_[[:space:]]Accelerating[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]TaylorSeers/915e0bb2-3846-48a8-89eb-a9a2038c3055_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Sharp[[:space:]]to[[:space:]]Blur_[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]2D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]Under[[:space:]]Extreme[[:space:]]Motion[[:space:]]Blur[[:space:]]Using[[:space:]]Event[[:space:]]Cameras/1d8160fb-056b-4809-bb5f-292067880dc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/From[[:space:]]Trial[[:space:]]to[[:space:]]Triumph_[[:space:]]Advancing[[:space:]]Long[[:space:]]Video[[:space:]]Understanding[[:space:]]via[[:space:]]Visual[[:space:]]Context[[:space:]]Sample[[:space:]]Scaling[[:space:]]and[[:space:]]Self-reward[[:space:]]Alignment/4a0f1359-9452-4469-b7d3-62005e604fbf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FuXi-RTM_[[:space:]]A[[:space:]]Physics-Guided[[:space:]]Prediction[[:space:]]Framework[[:space:]]with[[:space:]]Radiative[[:space:]]Transfer[[:space:]]Modeling/b1f77898-b0ac-400c-aec4-af24b6184408_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FullDiT_[[:space:]]Video[[:space:]]Generative[[:space:]]Foundation[[:space:]]Models[[:space:]]with[[:space:]]Multimodal[[:space:]]Control[[:space:]]via[[:space:]]Full[[:space:]]Attention/980a6e0d-b59a-4c02-a4b1-2f3876b164f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Function-centric[[:space:]]Bayesian[[:space:]]Network[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Object[[:space:]]Goal[[:space:]]Navigation/cdf344f6-7d57-4a47-8af8-907f5e688419_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fuse[[:space:]]Before[[:space:]]Transfer_[[:space:]]Knowledge[[:space:]]Fusion[[:space:]]for[[:space:]]Heterogeneous[[:space:]]Distillation/ab56fd7e-80c3-489f-9f84-c748afa3d076_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fusion[[:space:]]Meets[[:space:]]Diverse[[:space:]]Conditions_[[:space:]]A[[:space:]]High-diversity[[:space:]]Benchmark[[:space:]]and[[:space:]]Baseline[[:space:]]for[[:space:]]UAV-based[[:space:]]Multimodal[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Condition[[:space:]]Cues/a9a3c1c8-8b0d-4fbd-944a-fad6e862df4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/FusionPhys_[[:space:]]A[[:space:]]Flexible[[:space:]]Framework[[:space:]]for[[:space:]]Fusing[[:space:]]Complementary[[:space:]]Sensing[[:space:]]Modalities[[:space:]]in[[:space:]]Remote[[:space:]]Physiological[[:space:]]Measurement/2c71ed76-7389-4596-872e-527eb6e55fe6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Future-Aware[[:space:]]Interaction[[:space:]]Network[[:space:]]For[[:space:]]Motion[[:space:]]Forecasting/22797c33-a770-411f-9c99-d6408d0b7ad2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Fuzzy[[:space:]]Contrastive[[:space:]]Decoding[[:space:]]to[[:space:]]Alleviate[[:space:]]Object[[:space:]]Hallucination[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/c68673ea-6daf-49b3-9d90-2df2ca3e7d7a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/G-DexGrasp_[[:space:]]Generalizable[[:space:]]Dexterous[[:space:]]Grasping[[:space:]]Synthesis[[:space:]]Via[[:space:]]Part-Aware[[:space:]]Prior[[:space:]]Retrieval[[:space:]]and[[:space:]]Prior-Assisted[[:space:]]Generation/40e04304-e4ef-4586-b583-6bb560b28471_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/G2D_[[:space:]]Boosting[[:space:]]Multimodal[[:space:]]Learning[[:space:]]with[[:space:]]Gradient-Guided[[:space:]]Distillation/29255d2e-e3c1-4493-b0ff-b9a3b294ddeb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/G2PDiffusion_[[:space:]]Cross-Species[[:space:]]Genotype-to-Phenotype[[:space:]]Prediction[[:space:]]via[[:space:]]Evolutionary[[:space:]]Diffusion/9b73a7ba-899c-469c-bcd3-8d7b46ab3947_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/G2SF_[[:space:]]Geometry-Guided[[:space:]]Score[[:space:]]Fusion[[:space:]]for[[:space:]]Multimodal[[:space:]]Industrial[[:space:]]Anomaly[[:space:]]Detection/c2b7d61d-a16b-40fb-8ef4-64a5a585ee55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GAP_[[:space:]]Gaussianize[[:space:]]Any[[:space:]]Point[[:space:]]Clouds[[:space:]]with[[:space:]]Text[[:space:]]Guidance/c8c9f21b-c58c-4080-8e29-2f1a76f1c9df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GARF_[[:space:]]Learning[[:space:]]Generalizable[[:space:]]3D[[:space:]]Reassembly[[:space:]]for[[:space:]]Real-World[[:space:]]Fractures/18ab6f58-adda-492d-b614-a3890ecfe8ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GAS_[[:space:]]Generative[[:space:]]Avatar[[:space:]]Synthesis[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/245cb2df-c94a-4d7d-a42c-d2031558e23b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GCAV_[[:space:]]A[[:space:]]Global[[:space:]]Concept[[:space:]]Activation[[:space:]]Vector[[:space:]]Framework[[:space:]]for[[:space:]]Cross-Layer[[:space:]]Consistency[[:space:]]in[[:space:]]Interpretability/9ab9a64e-53a3-4411-919d-a8e6c4fa7450_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GCRayDiffusion_[[:space:]]Pose-Free[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]via[[:space:]]Geometric[[:space:]]Consistent[[:space:]]Ray[[:space:]]Diffusion/1d5edc52-84f3-4d87-919c-6651c5381ad0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GDKVM_[[:space:]]Echocardiography[[:space:]]Video[[:space:]]Segmentation[[:space:]]via[[:space:]]Spatiotemporal[[:space:]]Key-Value[[:space:]]Memory[[:space:]]with[[:space:]]Gated[[:space:]]Delta[[:space:]]Rule/0dca18d5-e42a-47f2-ba0e-8d768616f515_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GECKO_[[:space:]]Gigapixel[[:space:]]Vision-Concept[[:space:]]Contrastive[[:space:]]Pretraining[[:space:]]in[[:space:]]Histopathology/7bcc9000-8b7b-43b6-8cb7-a21120a8c446_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GECO_[[:space:]]Geometrically[[:space:]]Consistent[[:space:]]Embedding[[:space:]]with[[:space:]]Lightspeed[[:space:]]Inference/92c5652a-11cd-46b6-998f-de61681edae7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GEMeX_[[:space:]]A[[:space:]]Large-Scale,[[:space:]]Groundable,[[:space:]]and[[:space:]]Explainable[[:space:]]Medical[[:space:]]VQA[[:space:]]Benchmark[[:space:]]for[[:space:]]Chest[[:space:]]X-ray[[:space:]]Diagnosis/5f18f7f0-e42d-402e-aae0-6b6b53babbdb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GENMO_[[:space:]]A[[:space:]]GENeralist[[:space:]]Model[[:space:]]for[[:space:]]Human[[:space:]]MOtion/ee15da3f-de1a-4d40-b7ad-1da202734ed1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GEOBench-VLM_[[:space:]]Benchmarking[[:space:]]Vision-Language[[:space:]]Models[[:space:]]for[[:space:]]Geospatial[[:space:]]Tasks/f8d4fcc7-8784-45bc-897d-6071f0fc3b7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GEOPARD_[[:space:]]Geometric[[:space:]]Pretraining[[:space:]]for[[:space:]]Articulation[[:space:]]Prediction[[:space:]]in[[:space:]]3D[[:space:]]Shapes/229628f5-fc2c-43e7-95f6-cae7c8214633_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GFPack++_[[:space:]]Attention-Driven[[:space:]]Gradient[[:space:]]Fields[[:space:]]for[[:space:]]Optimizing[[:space:]]2D[[:space:]]Irregular[[:space:]]Packing/5f521088-c5c2-4ff1-a73a-698bebbedd3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GGTalker_[[:space:]]Talking[[:space:]]Head[[:space:]]Systhesis[[:space:]]with[[:space:]]Generalizable[[:space:]]Gaussian[[:space:]]Priors[[:space:]]and[[:space:]]Identity-Specific[[:space:]]Adaptation/d44c88ef-b726-45bd-ae2d-c1f1eb2ae22c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GIViC_[[:space:]]Generative[[:space:]]Implicit[[:space:]]Video[[:space:]]Compression/3c81a1ee-36ed-4243-a0f5-5be2b9a1a62e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GLEAM_[[:space:]]Enhanced[[:space:]]Transferable[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]for[[:space:]]Vision-Language[[:space:]]Pre-training[[:space:]]Models[[:space:]]via[[:space:]]Global-Local[[:space:]]Transformations/b9ddcb06-4f10-40f2-a642-17f6734f051a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GLEAM_[[:space:]]Learning[[:space:]]Generalizable[[:space:]]Exploration[[:space:]]Policy[[:space:]]for[[:space:]]Active[[:space:]]Mapping[[:space:]]in[[:space:]]Complex[[:space:]]3D[[:space:]]Indoor[[:space:]]Scene/2037116a-8269-4f6d-9b2d-526d722e31ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GM-MoE_[[:space:]]Low-Light[[:space:]]Enhancement[[:space:]]with[[:space:]]Gated-Mechanism[[:space:]]Mixture-of-Experts/06d78e4b-5b76-430e-bacf-04f5d134f3f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GMMamba_[[:space:]]Group[[:space:]]Masking[[:space:]]Mamba[[:space:]]for[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Classification/ca808fec-8708-4ebf-a7be-ed519a5ea370_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GRAB_[[:space:]]A[[:space:]]Challenging[[:space:]]GRaph[[:space:]]Analysis[[:space:]]Benchmark[[:space:]]for[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models/1549c76d-b5f6-41c7-a7d1-aaf40b01982f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GReg_[[:space:]]Geometry-Aware[[:space:]]Region[[:space:]]Refinement[[:space:]]for[[:space:]]Sign[[:space:]]Language[[:space:]]Video[[:space:]]Generation/9412e977-d0a4-4140-a84b-a21ef927131c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GS-ID_[[:space:]]Illumination[[:space:]]Decomposition[[:space:]]on[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]via[[:space:]]Adaptive[[:space:]]Light[[:space:]]Aggregation[[:space:]]and[[:space:]]Diffusion-Guided[[:space:]]Material[[:space:]]Priors/f686911f-9fe0-4e8e-a0be-8b5cda746f61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GS-LIVM_[[:space:]]Real-Time[[:space:]]Photo-Realistic[[:space:]]LiDAR-Inertial-Visual[[:space:]]Mapping[[:space:]]with[[:space:]]Gaussian[[:space:]]Splatting/43df5a0f-f5b5-48dd-9aaa-a0592f87e256_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GS-Occ3D_[[:space:]]Scaling[[:space:]]Vision-only[[:space:]]Occupancy[[:space:]]Reconstruction[[:space:]]with[[:space:]]Gaussian[[:space:]]Splatting/34a9640b-6b4b-40f5-994b-3dc79a5e9498_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GSOT3D_[[:space:]]Towards[[:space:]]Generic[[:space:]]3D[[:space:]]Single[[:space:]]Object[[:space:]]Tracking[[:space:]]in[[:space:]]the[[:space:]]Wild/07e978ac-556b-4f40-8717-0da03e621b9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GSRecon_[[:space:]]Efficient[[:space:]]Generalizable[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]from[[:space:]]Sparse[[:space:]]Views/a0980858-4397-429d-84e6-c6dceb41af02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GSV3D_[[:space:]]Gaussian[[:space:]]Splatting-based[[:space:]]Geometric[[:space:]]Distillation[[:space:]]with[[:space:]]Stable[[:space:]]Video[[:space:]]Diffusion[[:space:]]for[[:space:]]Single-Image[[:space:]]3D[[:space:]]Object[[:space:]]Generation/ec57455a-9353-4442-b69a-4e62cda29af8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GT-Loc_[[:space:]]Unifying[[:space:]]When[[:space:]]and[[:space:]]Where[[:space:]]in[[:space:]]Images[[:space:]]Through[[:space:]]a[[:space:]]Joint[[:space:]]Embedding[[:space:]]Space/e6fe20f1-6650-4f56-8a56-9289de09147e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GT-Mean[[:space:]]Loss_[[:space:]]A[[:space:]]Simple[[:space:]]Yet[[:space:]]Effective[[:space:]]Solution[[:space:]]for[[:space:]]Brightness[[:space:]]Mismatch[[:space:]]in[[:space:]]Low-Light[[:space:]]Image[[:space:]]Enhancement/a1d665cf-5643-46d9-9770-a927a416d5bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GTR_[[:space:]]Guided[[:space:]]Thought[[:space:]]Reinforcement[[:space:]]Prevents[[:space:]]Thought[[:space:]]Collapse[[:space:]]in[[:space:]]RL-based[[:space:]]VLM[[:space:]]Agent[[:space:]]Training/8d6a65fc-a5b8-4ac1-8162-da34115f8e7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GUAVA_[[:space:]]Generalizable[[:space:]]Upper[[:space:]]Body[[:space:]]3D[[:space:]]Gaussian[[:space:]]Avatar/c0c98a92-b53c-4e70-a42b-2a27bb507f42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GUIOdyssey_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Dataset[[:space:]]for[[:space:]]Cross-App[[:space:]]GUI[[:space:]]Navigation[[:space:]]on[[:space:]]Mobile[[:space:]]Devices/4ede5fa9-b934-4a09-844f-157df6c9d80b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GVDepth_[[:space:]]Zero-Shot[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation[[:space:]]for[[:space:]]Ground[[:space:]]Vehicles[[:space:]]based[[:space:]]on[[:space:]]Probabilistic[[:space:]]Cue[[:space:]]Fusion/3a4d0aff-64e6-4333-9616-0c5f08ae884a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GWM_[[:space:]]Towards[[:space:]]Scalable[[:space:]]Gaussian[[:space:]]World[[:space:]]Models[[:space:]]for[[:space:]]Robotic[[:space:]]Manipulation/c06a398f-0f37-4b2c-841a-f72762725efd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaRe_[[:space:]]Relightable[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Outdoor[[:space:]]Scenes[[:space:]]from[[:space:]]Unconstrained[[:space:]]Photo[[:space:]]Collections/fd69163d-c8b7-4b32-923c-0505677cc0f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaSLight_[[:space:]]Gaussian[[:space:]]Splats[[:space:]]for[[:space:]]Spatially-Varying[[:space:]]Lighting[[:space:]]in[[:space:]]HDR/bc703902-c35c-4461-a898-f4c3ead42c3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gain-MLP_[[:space:]]Improving[[:space:]]HDR[[:space:]]Gain[[:space:]]Map[[:space:]]Encoding[[:space:]]via[[:space:]]a[[:space:]]Lightweight[[:space:]]MLP/d6612fa3-5435-42aa-9471-d09cffa9ccc5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gait-X_[[:space:]]Exploring[[:space:]]X[[:space:]]modality[[:space:]]for[[:space:]]Generalized[[:space:]]Gait[[:space:]]Recognition/82087f31-0b39-4037-90a5-1e613163333e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GameFactory_[[:space:]]Creating[[:space:]]New[[:space:]]Games[[:space:]]with[[:space:]]Generative[[:space:]]Interactive[[:space:]]Videos/78e36c5f-5aee-48a3-8343-77df1d0998b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GauUpdate_[[:space:]]New[[:space:]]Object[[:space:]]Insertion[[:space:]]in[[:space:]]3D[[:space:]]Gaussian[[:space:]]Fields[[:space:]]with[[:space:]]Consistent[[:space:]]Global[[:space:]]Illumination/8678848a-7984-4e30-8d48-08c266102859_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GausSim_[[:space:]]Foreseeing[[:space:]]Reality[[:space:]]by[[:space:]]Gaussian[[:space:]]Simulator[[:space:]]for[[:space:]]Elastic[[:space:]]Objects/e987d6ae-2f53-4bda-81ba-5427f540c3b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussRender_[[:space:]]Learning[[:space:]]3D[[:space:]]Occupancy[[:space:]]with[[:space:]]Gaussian[[:space:]]Rendering/32137244-8441-41bb-894b-b2807565c2a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]Discretized[[:space:]]SDF[[:space:]]for[[:space:]]Relightable[[:space:]]Assets/99383a0f-f4aa-4e78-8dbf-a56905f800a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gaussian[[:space:]]Variation[[:space:]]Field[[:space:]]Diffusion[[:space:]]for[[:space:]]High-fidelity[[:space:]]Video-to-4D[[:space:]]Synthesis/a61f73cc-2180-4265-b989-c7ab82767a13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gaussian-based[[:space:]]World[[:space:]]Model_[[:space:]]Gaussian[[:space:]]Priors[[:space:]]for[[:space:]]Voxel-Based[[:space:]]Occupancy[[:space:]]Prediction[[:space:]]and[[:space:]]Future[[:space:]]Motion[[:space:]]Prediction/9bb0d531-6703-4c39-bfb4-c96b94cf8f4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussianFlowOcc_[[:space:]]Sparse[[:space:]]and[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Occupancy[[:space:]]Estimation[[:space:]]using[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]and[[:space:]]Temporal[[:space:]]Flow/31ca51a6-6827-41bb-b08b-6ac776277091_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussianOcc_[[:space:]]Fully[[:space:]]Self-supervised[[:space:]]and[[:space:]]Efficient[[:space:]]3D[[:space:]]Occupancy[[:space:]]Estimation[[:space:]]with[[:space:]]Gaussian[[:space:]]Splatting/3dad4538-f1d9-4a7a-9236-124b3f42cab8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussianProperty_[[:space:]]Integrating[[:space:]]Physical[[:space:]]Properties[[:space:]]to[[:space:]]3D[[:space:]]Gaussians[[:space:]]with[[:space:]]LMMs/f28f2398-cdaa-4130-b47e-22d976cb15a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussianReg_[[:space:]]Rapid[[:space:]]2D_3D[[:space:]]Registration[[:space:]]for[[:space:]]Emergency[[:space:]]Surgery[[:space:]]via[[:space:]]Explicit[[:space:]]3D[[:space:]]Modeling[[:space:]]with[[:space:]]Gaussian[[:space:]]Primitives/524c4b21-2587-478f-bcb3-441acdf9d447_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussianSpeech_[[:space:]]Audio-Driven[[:space:]]Personalized[[:space:]]3D[[:space:]]Gaussian[[:space:]]Avatars/78756e20-335e-49fd-bb91-edd5bd67ff25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussianUpdate_[[:space:]]Continual[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Update[[:space:]]for[[:space:]]Changing[[:space:]]Environments/d7ad1b8e-946c-47df-80d9-d1f3b99b7487_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GaussianVideo_[[:space:]]Efficient[[:space:]]Video[[:space:]]Representation[[:space:]]via[[:space:]]Hierarchical[[:space:]]Gaussian[[:space:]]Splatting/54e5a7d5-4b64-4d96-b39f-3fd606b13271_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gaze-Language[[:space:]]Alignment[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Prediction[[:space:]]of[[:space:]]Visual[[:space:]]Search[[:space:]]Targets[[:space:]]from[[:space:]]Human[[:space:]]Gaze[[:space:]]Scanpaths/d9fc8835-bc0e-44c2-b541-c5d6d9c10102_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GazeGaussian_[[:space:]]High-Fidelity[[:space:]]Gaze[[:space:]]Redirection[[:space:]]with[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/df420d8a-be10-4b21-b946-23d93b33d8eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Geminio_[[:space:]]Language-Guided[[:space:]]Gradient[[:space:]]Inversion[[:space:]]Attacks[[:space:]]in[[:space:]]Federated[[:space:]]Learning/bacba03d-bce1-464a-a6e5-9f2185250bb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GenDoP_[[:space:]]Auto-regressive[[:space:]]Camera[[:space:]]Trajectory[[:space:]]Generation[[:space:]]as[[:space:]]a[[:space:]]Director[[:space:]]of[[:space:]]Photography/574632cc-2118-4835-bb7b-4a6393098a4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GenFlow3D_[[:space:]]Generative[[:space:]]Scene[[:space:]]Flow[[:space:]]Estimation[[:space:]]and[[:space:]]Prediction[[:space:]]on[[:space:]]Point[[:space:]]Cloud[[:space:]]Sequences/9acd5a8d-fbfd-4e2d-ab56-0756f9fddd5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GenFlowRL_[[:space:]]Shaping[[:space:]]Rewards[[:space:]]with[[:space:]]Generative[[:space:]]Object-Centric[[:space:]]Flow[[:space:]]in[[:space:]]Visual[[:space:]]Reinforcement[[:space:]]Learning/0db99619-468e-4528-80a9-20912293d6bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GenHancer_[[:space:]]Imperfect[[:space:]]Generative[[:space:]]Models[[:space:]]are[[:space:]]Secretly[[:space:]]Strong[[:space:]]Vision-Centric[[:space:]]Enhancers/c13dc7c6-cf63-4057-b342-2b5e4cd32316_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GenHaze_[[:space:]]Pioneering[[:space:]]Controllable[[:space:]]One-Step[[:space:]]Realistic[[:space:]]Haze[[:space:]]Generation[[:space:]]for[[:space:]]Real-World[[:space:]]Dehazing/f5570052-a90e-447b-96f1-f3f76017f9e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GenM3_[[:space:]]Generative[[:space:]]Pretrained[[:space:]]Multi-path[[:space:]]Motion[[:space:]]Model[[:space:]]for[[:space:]]Text[[:space:]]Conditional[[:space:]]Human[[:space:]]Motion[[:space:]]Generation/0c5d5f6c-bbdb-4132-b23f-1e75da6907e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/General[[:space:]]Compression[[:space:]]Framework[[:space:]]for[[:space:]]Efficient[[:space:]]Transformer[[:space:]]Object[[:space:]]Tracking/9f20f082-1463-4db6-a2cc-2ec201439687_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalizable[[:space:]]Non-Line-of-Sight[[:space:]]Imaging[[:space:]]with[[:space:]]Learnable[[:space:]]Physical[[:space:]]Priors/393f7537-c0a8-4c70-b2bf-24472da80a1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalizable[[:space:]]Object[[:space:]]Re-Identification[[:space:]]via[[:space:]]Visual[[:space:]]In-Context[[:space:]]Prompting/a0dbf684-db1f-4d59-874e-478d9892dc8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalization-Preserved[[:space:]]Learning_[[:space:]]Closing[[:space:]]the[[:space:]]Backdoor[[:space:]]to[[:space:]]Catastrophic[[:space:]]Forgetting[[:space:]]in[[:space:]]Continual[[:space:]]Deepfake[[:space:]]Detection/a919de6d-4325-4862-bc22-0251c96b682f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalized[[:space:]]Deep[[:space:]]Multi-view[[:space:]]Clustering[[:space:]]via[[:space:]]Causal[[:space:]]Learning[[:space:]]with[[:space:]]Partially[[:space:]]Aligned[[:space:]]Cross-view[[:space:]]Correspondence/688daa7e-aaee-4a5d-8870-636bde6241ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalized[[:space:]]Few-Shot[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation[[:space:]]via[[:space:]]LLM-Assisted[[:space:]]Hyper-Relation[[:space:]]Matching/6311a54a-dc55-4427-a122-916031bf442c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalized[[:space:]]Tensor-based[[:space:]]Parameter-Efficient[[:space:]]Fine-Tuning[[:space:]]via[[:space:]]Lie[[:space:]]Group[[:space:]]Transformations/7a4db415-1c35-4df7-99b7-41a78e0e7a0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generalized[[:space:]]and[[:space:]]Efficient[[:space:]]2D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Arbitrary-scale[[:space:]]Super-Resolution/20678d4e-97aa-4911-a1a4-bd2bbb9b1ffc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generate,[[:space:]]Refine,[[:space:]]and[[:space:]]Encode_[[:space:]]Leveraging[[:space:]]Synthesized[[:space:]]Novel[[:space:]]Samples[[:space:]]for[[:space:]]On-the-Fly[[:space:]]Fine-Grained[[:space:]]Category[[:space:]]Discovery/b4eb0cd3-1b02-4ce1-b5db-f38de103f826_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generate,[[:space:]]Transduct,[[:space:]]Adapt_[[:space:]]Iterative[[:space:]]Transduction[[:space:]]with[[:space:]]VLMs/60c141fc-d254-4073-94a8-9bb602ffcc67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generating[[:space:]]Multi-Image[[:space:]]Synthetic[[:space:]]Data[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Customization/4fb3be0b-00e2-459d-9d24-b0020a054125_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generating[[:space:]]Physically[[:space:]]Stable[[:space:]]and[[:space:]]Buildable[[:space:]]Brick[[:space:]]Structures[[:space:]]from[[:space:]]Text/7e32684d-a2d8-4026-a133-e5a536509957_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generating,[[:space:]]Fast[[:space:]]and[[:space:]]Slow_[[:space:]]Scalable[[:space:]]Parallel[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Video[[:space:]]Interface[[:space:]]Networks/f026a02a-20a1-4c8b-a50a-71513f05811f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Active[[:space:]]Learning[[:space:]]for[[:space:]]Long-tail[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]via[[:space:]]Controllable[[:space:]]Diffusion[[:space:]]Model/3fe142d0-e9a6-4555-95bb-fcac157c8d3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Adversarial[[:space:]]Diffusion/cad38cac-a1f4-4a1f-9c60-e4c0d76cf634_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Gaussian[[:space:]]Splatting_[[:space:]]Generating[[:space:]]3D[[:space:]]Scenes[[:space:]]with[[:space:]]Video[[:space:]]Diffusion[[:space:]]Priors/de0f0bcc-fef3-4ac8-965a-3041130376f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Modeling[[:space:]]of[[:space:]]Shape-Dependent[[:space:]]Self-Contact[[:space:]]Human[[:space:]]Poses/a650d3d6-4ff5-4e51-be76-c03a1e2d8d9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Video[[:space:]]Bi-flow/81548ad3-fd4b-4d58-953a-4f3706a49405_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generative[[:space:]]Zoo/2fca6192-88ab-4e95-8dc2-8bf25588b2cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Generic[[:space:]]Event[[:space:]]Boundary[[:space:]]Detection[[:space:]]via[[:space:]]Denoising[[:space:]]Diffusion/d7b5bd04-277e-4439-8e6e-3469a69a4555_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GenieBlue_[[:space:]]Integrating[[:space:]]both[[:space:]]Linguistic[[:space:]]and[[:space:]]Multimodal[[:space:]]Capabilities[[:space:]]for[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]on[[:space:]]Mobile[[:space:]]Devices/948d09c3-27dd-4abc-a20c-478b5fe70f8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Geo4D_[[:space:]]Leveraging[[:space:]]Video[[:space:]]Generators[[:space:]]for[[:space:]]Geometric[[:space:]]4D[[:space:]]Scene[[:space:]]Reconstruction/39ae5c5c-1113-4bd5-9e5c-ab10bad7fcd3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GeoAvatar_[[:space:]]Adaptive[[:space:]]Geometrical[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]3D[[:space:]]Head[[:space:]]Avatar/4c4c67f2-21af-45b1-b82a-6bea93011845_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GeoDiffusion_[[:space:]]A[[:space:]]Training-Free[[:space:]]Framework[[:space:]]for[[:space:]]Accurate[[:space:]]3D[[:space:]]Geometric[[:space:]]Conditioning[[:space:]]in[[:space:]]Image[[:space:]]Generation/4309cc2f-464c-4b37-8957-e75010f05b3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GeoDistill_[[:space:]]Geometry-Guided[[:space:]]Self-Distillation[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Cross-View[[:space:]]Localization/065b583a-2364-49a2-ad73-734054d4f8f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GeoExplorer_[[:space:]]Active[[:space:]]Geo-localization[[:space:]]with[[:space:]]Curiosity-Driven[[:space:]]Exploration/6d6f11e7-0aeb-4fb0-8408-ffde7d4676a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GeoFormer_[[:space:]]Geometry[[:space:]]Point[[:space:]]Encoder[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Graph-based[[:space:]]Transformer/a49b0e01-d040-4ab3-b35b-4c8fe0a8d4a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GeoMan_[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Human[[:space:]]Geometry[[:space:]]Estimation[[:space:]]using[[:space:]]Image-to-Video[[:space:]]Diffusion/8186b55a-22c5-43f3-a671-5e9f12d26772_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GeoProg3D_[[:space:]]Compositional[[:space:]]Visual[[:space:]]Reasoning[[:space:]]for[[:space:]]City-Scale[[:space:]]3D[[:space:]]Language[[:space:]]Fields/7d892d07-2901-43c5-92cb-e267199aec2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GeoSplatting_[[:space:]]Towards[[:space:]]Geometry[[:space:]]Guided[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Physically-based[[:space:]]Inverse[[:space:]]Rendering/ceedb5ae-3e78-431e-8d40-34cefa595127_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Geometric[[:space:]]Alignment[[:space:]]and[[:space:]]Prior[[:space:]]Modulation[[:space:]]for[[:space:]]View-Guided[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion[[:space:]]on[[:space:]]Unseen[[:space:]]Categories/43686cd4-bd3d-4bbb-a881-0d99a37fbdc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Geometry[[:space:]]Distributions/24849854-056a-4414-afaf-e5bb200133cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GeometryCrafter_[[:space:]]Consistent[[:space:]]Geometry[[:space:]]Estimation[[:space:]]for[[:space:]]Open-world[[:space:]]Videos[[:space:]]with[[:space:]]Diffusion[[:space:]]Priors/c6a98414-1bb8-415c-8b89-0e910c18a3cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GestureHYDRA_[[:space:]]Semantic[[:space:]]Co-speech[[:space:]]Gesture[[:space:]]Synthesis[[:space:]]via[[:space:]]Hybrid[[:space:]]Modality[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]and[[:space:]]Cascaded-Synchronized[[:space:]]Retrieval-Augmented[[:space:]]Generation/e43b0fa0-b0bd-4784-99e8-1b9ffa866d9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GestureLSM_[[:space:]]Latent[[:space:]]Shortcut[[:space:]]based[[:space:]]Co-Speech[[:space:]]Gesture[[:space:]]Generation[[:space:]]with[[:space:]]Spatial-Temporal[[:space:]]Modeling/5180e772-2cd4-42fa-8379-fee981334086_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GigaTok_[[:space:]]Scaling[[:space:]]Visual[[:space:]]Tokenizers[[:space:]]to[[:space:]]3[[:space:]]Billion[[:space:]]Parameters[[:space:]]for[[:space:]]Autoregressive[[:space:]]Image[[:space:]]Generation/0192c2a3-0c4d-4114-910d-c3b2b9b233ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GlassWizard_[[:space:]]Harvesting[[:space:]]Diffusion[[:space:]]Priors[[:space:]]for[[:space:]]Glass[[:space:]]Surface[[:space:]]Detection/5eac0aec-e389-40e2-9bff-c9db100bc82a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GloPER_[[:space:]]Unsupervised[[:space:]]Animal[[:space:]]Pattern[[:space:]]Extraction[[:space:]]from[[:space:]]Local[[:space:]]Reconstruction/f61ffafa-b941-463e-8baf-d3e053062d0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Global[[:space:]]Motion[[:space:]]Corresponder[[:space:]]for[[:space:]]3D[[:space:]]Point-Based[[:space:]]Scene[[:space:]]Interpolation[[:space:]]under[[:space:]]Large[[:space:]]Motion/186bd700-a8c7-4f53-bb1b-b97021964c8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Global[[:space:]]Regulation[[:space:]]and[[:space:]]Excitation[[:space:]]via[[:space:]]Attention[[:space:]]Tuning[[:space:]]for[[:space:]]Stereo[[:space:]]Matching/3daef0d1-561c-48d3-81ca-53fecd0c0c69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Global[[:space:]]and[[:space:]]Local[[:space:]]Entailment[[:space:]]Learning[[:space:]]for[[:space:]]Natural[[:space:]]World[[:space:]]Imagery/d03a02a6-399b-4a05-81c1-531a4252bb61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Global-Aware[[:space:]]Monocular[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion[[:space:]]with[[:space:]]State[[:space:]]Space[[:space:]]Models/c6372e96-15c4-4340-bd71-3080d1ffe90c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Go[[:space:]]to[[:space:]]Zero_[[:space:]]Towards[[:space:]]Zero-shot[[:space:]]Motion[[:space:]]Generation[[:space:]]with[[:space:]]Million-scale[[:space:]]Data/2d511f4d-c8e4-4e25-93f8-ba18256636af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Golden[[:space:]]Noise[[:space:]]for[[:space:]]Diffusion[[:space:]]Models_[[:space:]]A[[:space:]]Learning[[:space:]]Framework/03a60141-a654-4975-92fe-ee00d48082fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gradient[[:space:]]Decomposition[[:space:]]and[[:space:]]Alignment[[:space:]]for[[:space:]]Incremental[[:space:]]Object[[:space:]]Detection/4cc216c5-6664-4802-9e00-81cad402a3b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gradient[[:space:]]Extrapolation[[:space:]]for[[:space:]]Debiased[[:space:]]Representation[[:space:]]Learning/fd83246e-837b-48d2-b5ed-1d2e10497426_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gradient[[:space:]]Short-Circuit_[[:space:]]Efficient[[:space:]]Out-of-Distribution[[:space:]]Detection[[:space:]]via[[:space:]]Feature[[:space:]]Intervention/43d61d92-7a85-4cd5-b0f9-dd289c0cd855_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Gradient-Reweighted[[:space:]]Adversarial[[:space:]]Camouflage[[:space:]]for[[:space:]]Physical[[:space:]]Object[[:space:]]Detection[[:space:]]Evasion/6de5e49a-62f6-4a07-ad80-6307b065aebc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Granular[[:space:]]Concept[[:space:]]Circuits_[[:space:]]Toward[[:space:]]a[[:space:]]Fine-Grained[[:space:]]Circuit[[:space:]]Discovery[[:space:]]for[[:space:]]Concept[[:space:]]Representations/f8140bd0-617a-4da2-9d4b-a0fe6266be44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Graph[[:space:]]Domain[[:space:]]Adaptation[[:space:]]with[[:space:]]Dual-branch[[:space:]]Encoder[[:space:]]and[[:space:]]Two-level[[:space:]]Alignment[[:space:]]for[[:space:]]Whole[[:space:]]Slide[[:space:]]Image-based[[:space:]]Survival[[:space:]]Prediction/5a1105c6-8a7b-4592-82c5-6ca5ada074c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GraspCoT_[[:space:]]Integrating[[:space:]]Physical[[:space:]]Property[[:space:]]Reasoning[[:space:]]for[[:space:]]6-DoF[[:space:]]Grasping[[:space:]]under[[:space:]]Flexible[[:space:]]Language[[:space:]]Instructions/d6e55b8a-a0ee-4f03-a9f6-a4708d7c2d16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Griffon[[:space:]]v2_[[:space:]]Advancing[[:space:]]Multimodal[[:space:]]Perception[[:space:]]with[[:space:]]High-Resolution[[:space:]]Scaling[[:space:]]and[[:space:]]Visual-Language[[:space:]]Co-Referring/7945e1df-7c18-4903-98b2-ee6ebfca5b6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GroundFlow_[[:space:]]A[[:space:]]Plug-in[[:space:]]Module[[:space:]]for[[:space:]]Temporal[[:space:]]Reasoning[[:space:]]on[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Sequential[[:space:]]Grounding/11a84876-6f7d-4377-990b-3afab748c14c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/GroundingSuite_[[:space:]]Measuring[[:space:]]Complex[[:space:]]Multi-Granular[[:space:]]Pixel[[:space:]]Grounding/82b13914-80b2-40a8-8fb1-fe8ab97b1377_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Group[[:space:]]Inertial[[:space:]]Poser_[[:space:]]Multi-Person[[:space:]]Pose[[:space:]]and[[:space:]]Global[[:space:]]Translation[[:space:]]from[[:space:]]Sparse[[:space:]]Inertial[[:space:]]Sensors[[:space:]]and[[:space:]]Ultra-Wideband[[:space:]]Ranging/1882ac30-44d0-4bc9-b055-e00352d6d65f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Group-wise[[:space:]]Scaling[[:space:]]and[[:space:]]Orthogonal[[:space:]]Decomposition[[:space:]]for[[:space:]]Domain-Invariant[[:space:]]Feature[[:space:]]Extraction[[:space:]]in[[:space:]]Face[[:space:]]Anti-Spoofing/19cc6eb6-3493-4da9-90c5-49a4139f434b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Grouped[[:space:]]Speculative[[:space:]]Decoding[[:space:]]for[[:space:]]Autoregressive[[:space:]]Image[[:space:]]Generation/7633fa04-45c5-4c08-926c-b4a1c5d7be43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Growing[[:space:]]a[[:space:]]Twig[[:space:]]to[[:space:]]Accelerate[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/b042f1e5-dfdf-456e-bc85-df078488f708_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Guiding[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Adaptive[[:space:]]Negative[[:space:]]Sampling[[:space:]]Without[[:space:]]External[[:space:]]Resources/e178a8b2-c48e-4308-8c21-e6cd6e2ef1cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Guiding[[:space:]]Diffusion-Based[[:space:]]Articulated[[:space:]]Object[[:space:]]Generation[[:space:]]by[[:space:]]Partial[[:space:]]Point[[:space:]]Cloud[[:space:]]Alignment[[:space:]]and[[:space:]]Physical[[:space:]]Plausibility[[:space:]]Constraints/6b4c8e1d-6e0e-4caa-9e4f-0862f7eaf083_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Guiding[[:space:]]Noisy[[:space:]]Label[[:space:]]Conditional[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Score-based[[:space:]]Discriminator[[:space:]]Correction/d672dbf6-da17-4589-9123-03b3615c55db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/H3R_[[:space:]]Hybrid[[:space:]]Multi-view[[:space:]]Correspondence[[:space:]]for[[:space:]]Generalizable[[:space:]]3D[[:space:]]Reconstruction/26659bb4-6510-4c32-998f-85b08b5291dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HADES_[[:space:]]Human[[:space:]]Avatar[[:space:]]with[[:space:]]Dynamic[[:space:]]Explicit[[:space:]]Hair[[:space:]]Strands/55982fcb-75e3-47e6-beeb-6b16f1c22ffe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HAMSt3R_[[:space:]]Human-Aware[[:space:]]Multi-view[[:space:]]Stereo[[:space:]]3D[[:space:]]Reconstruction/fce5f5cd-ed28-4130-a51e-ac808e658667_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HAMoBE_[[:space:]]Hierarchical[[:space:]]and[[:space:]]Adaptive[[:space:]]Mixture[[:space:]]of[[:space:]]Biometric[[:space:]]Experts[[:space:]]for[[:space:]]Video-based[[:space:]]Person[[:space:]]ReID/bf4ca273-3586-4ca3-a091-f5f54202ad5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HDR[[:space:]]Image[[:space:]]Generation[[:space:]]via[[:space:]]Gain[[:space:]]Map[[:space:]]Decomposed[[:space:]]Diffusion/5096adca-c269-48aa-b30a-74656711d6e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HERMES_[[:space:]]A[[:space:]]Unified[[:space:]]Self-Driving[[:space:]]World[[:space:]]Model[[:space:]]for[[:space:]]Simultaneous[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding[[:space:]]and[[:space:]]Generation/7839b82f-f843-4374-9e20-d0a28ad93d58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HERMES_[[:space:]]temporal-coHERent[[:space:]]long-forM[[:space:]]understanding[[:space:]]with[[:space:]]Episodes[[:space:]]and[[:space:]]Semantics/a622f6f1-6c09-4d19-a45e-dfbd63776357_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HERO_[[:space:]]Human[[:space:]]Reaction[[:space:]]Generation[[:space:]]from[[:space:]]Videos/8ec5852e-3b36-4263-8086-6935fe40c2d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HFD-Teacher_[[:space:]]High-Frequency[[:space:]]Depth[[:space:]]Distillation[[:space:]]from[[:space:]]Depth[[:space:]]Foundation[[:space:]]Models[[:space:]]for[[:space:]]Enhanced[[:space:]]Depth[[:space:]]Completion/6b52a789-55c4-406f-91cc-251df18a1dc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HIS-GPT_[[:space:]]Towards[[:space:]]3D[[:space:]]Human-In-Scene[[:space:]]Multimodal[[:space:]]Understanding/60847d65-c2df-4a39-81fb-e97deefe4b8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HOLa_[[:space:]]Zero-Shot[[:space:]]HOI[[:space:]]Detection[[:space:]]with[[:space:]]Low-Rank[[:space:]]Decomposed[[:space:]]VLM[[:space:]]Feature[[:space:]]Adaptation/5b531c6f-26b3-40fa-8810-a65f5ee63975_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HOMO-Feature_[[:space:]]Cross-Arbitrary-Modal[[:space:]]Image[[:space:]]Matching[[:space:]]with[[:space:]]Homomorphism[[:space:]]of[[:space:]]Organized[[:space:]]Major[[:space:]]Orientation/6aa6caba-5d51-471b-88e2-11eead961abd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HORT_[[:space:]]Monocular[[:space:]]Hand-held[[:space:]]Objects[[:space:]]Reconstruction[[:space:]]with[[:space:]]Transformers/5fa4e66f-9286-4da7-838d-281404146b23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HPSv3_[[:space:]]Towards[[:space:]]Wide-Spectrum[[:space:]]Human[[:space:]]Preference[[:space:]]Score/2ef1e1f0-8b09-4e9c-ba3d-79f94c85995d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HQ-CLIP_[[:space:]]Leveraging[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]to[[:space:]]Create[[:space:]]High-Quality[[:space:]]Image-Text[[:space:]]Datasets[[:space:]]and[[:space:]]CLIP[[:space:]]Models/00861017-3dd6-468a-9c35-be1ceb54fdaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HRScene_[[:space:]]How[[:space:]]Far[[:space:]]Are[[:space:]]VLMs[[:space:]]from[[:space:]]Effective[[:space:]]High-Resolution[[:space:]]Image[[:space:]]Understanding_/5525f367-aa5f-42ed-856e-0ab6ca80b8ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HUG_[[:space:]]Hierarchical[[:space:]]Urban[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]Block-Based[[:space:]]Reconstruction[[:space:]]for[[:space:]]Large-Scale[[:space:]]Aerial[[:space:]]Scenes/d765cc88-4a5a-4d43-aa8f-26ce764316c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HUMOTO_[[:space:]]A[[:space:]]4D[[:space:]]Dataset[[:space:]]of[[:space:]]Mocap[[:space:]]Human[[:space:]]Object[[:space:]]Interactions/97037554-d4d3-4868-8b23-7b33954091a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HUST_[[:space:]]High-Fidelity[[:space:]]Unbiased[[:space:]]Skin[[:space:]]Tone[[:space:]]Estimation[[:space:]]via[[:space:]]Texture[[:space:]]Quantization/04db57e4-586f-4a1f-ba81-70f8d503a339_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HVPUNet_[[:space:]]Hybrid-Voxel[[:space:]]Point-cloud[[:space:]]Upsampling[[:space:]]Network/902c59c0-0374-4b73-90d6-694966a16ac6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HairCUP_[[:space:]]Hair[[:space:]]Compositional[[:space:]]Universal[[:space:]]Prior[[:space:]]for[[:space:]]3D[[:space:]]Gaussian[[:space:]]Avatars/0f3dce30-67f5-4d1b-bbd0-49f74e7c0e5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hallucinatory[[:space:]]Image[[:space:]]Tokens_[[:space:]]A[[:space:]]Training-free[[:space:]]EAZY[[:space:]]Approach[[:space:]]to[[:space:]]Detecting[[:space:]]and[[:space:]]Mitigating[[:space:]]Object[[:space:]]Hallucinations[[:space:]]in[[:space:]]LVLMs/a9e996d0-73a8-40ee-9e61-1682f320695d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Harmonizing[[:space:]]Visual[[:space:]]Representations[[:space:]]for[[:space:]]Unified[[:space:]]Multimodal[[:space:]]Understanding[[:space:]]and[[:space:]]Generation/d04df740-73ac-4cd9-88aa-6b3a53a553fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HarmonySeg_[[:space:]]Tubular[[:space:]]Structure[[:space:]]Segmentation[[:space:]]with[[:space:]]Deep-Shallow[[:space:]]Feature[[:space:]]Fusion[[:space:]]and[[:space:]]Growth-Suppression[[:space:]]Balanced[[:space:]]Loss/6af921b0-cf81-4920-8c31-49ce1f979932_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Harnessing[[:space:]]Input-Adaptive[[:space:]]Inference[[:space:]]for[[:space:]]Efficient[[:space:]]VLN/db367b8e-43ab-4b27-8f06-b41b713b6aa6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Harnessing[[:space:]]Massive[[:space:]]Satellite[[:space:]]Imagery[[:space:]]with[[:space:]]Efficient[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling/b6ea627d-add3-4971-a537-ce79984bde6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Harnessing[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Self-Supervised[[:space:]]Learning/9e885722-8bcc-4e62-8f74-6cfb08ca039b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Harnessing[[:space:]]Uncertainty-aware[[:space:]]Bounding[[:space:]]Boxes[[:space:]]for[[:space:]]Unsupervised[[:space:]]3D[[:space:]]Object[[:space:]]Detection/622a1f38-f7cf-4a94-a632-0784d1467454_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Harnessing[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models[[:space:]]for[[:space:]]High-Performance,[[:space:]]Training-Free[[:space:]]Open[[:space:]]Vocabulary[[:space:]]Segmentation/79e9e37f-c9f0-450f-bafd-3419c84a420a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hate[[:space:]]in[[:space:]]Plain[[:space:]]Sight_[[:space:]]On[[:space:]]the[[:space:]]Risks[[:space:]]of[[:space:]]Moderating[[:space:]]AI-Generated[[:space:]]Hateful[[:space:]]Illusions/f7633723-45db-47c4-ac70-9e71ebebbe3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HazeFlow_[[:space:]]Revisit[[:space:]]Haze[[:space:]]Physical[[:space:]]Model[[:space:]]as[[:space:]]ODE[[:space:]]and[[:space:]]Non-Homogeneous[[:space:]]Haze[[:space:]]Generation[[:space:]]for[[:space:]]Real-World[[:space:]]Dehazing/774aadd6-c9c0-4f3d-95b5-9c4fb84c8cdf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HccePose(BF)_[[:space:]]Predicting[[:space:]]Front[[:space:]]&[[:space:]]Back[[:space:]]Surfaces[[:space:]]to[[:space:]]Construct[[:space:]]Ultra-Dense[[:space:]]2D-3D[[:space:]]Correspondences[[:space:]]for[[:space:]]Pose[[:space:]]Estimation/033dc6d6-e0fb-4358-9f94-89e9890bb0a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Head2Body_[[:space:]]Body[[:space:]]Pose[[:space:]]Generation[[:space:]]from[[:space:]]Multi-sensory[[:space:]]Head-mounted[[:space:]]Inputs/cea23e26-59ae-4155-98e7-23e4fccaf80e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Heatmap[[:space:]]Regression[[:space:]]without[[:space:]]Soft-Argmax[[:space:]]for[[:space:]]Facial[[:space:]]Landmark[[:space:]]Detection/5ea58eb1-3a2a-4eb8-a760-beed27e50a32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Heavy[[:space:]]Labels[[:space:]]Out![[:space:]]Dataset[[:space:]]Distillation[[:space:]]with[[:space:]]Label[[:space:]]Space[[:space:]]Lightening/4fe5f2e0-b3f4-45ca-951c-67ea60d45f03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Height-Fidelity[[:space:]]Dense[[:space:]]Global[[:space:]]Fusion[[:space:]]for[[:space:]]Multi-modal[[:space:]]3D[[:space:]]Object[[:space:]]Detection/e9dd5c78-f7bc-40a6-b71e-2dc99bfc101a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Heuristic-Induced[[:space:]]Multimodal[[:space:]]Risk[[:space:]]Distribution[[:space:]]Jailbreak[[:space:]]Attack[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/a662beb0-aaae-4c8d-81a1-99730614871f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hi-Gaussian_[[:space:]]Hierarchical[[:space:]]Gaussians[[:space:]]under[[:space:]]Normalized[[:space:]]Spherical[[:space:]]Projection[[:space:]]for[[:space:]]Single-View[[:space:]]3D[[:space:]]Reconstruction/dac504c1-dd89-4da6-bd66-6fb700efe9a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hi3DGen_[[:space:]]High-fidelity[[:space:]]3D[[:space:]]Geometry[[:space:]]Generation[[:space:]]from[[:space:]]Images[[:space:]]via[[:space:]]Normal[[:space:]]Bridging/60e9b436-d43a-4723-823e-670e19b03fc0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HiERO_[[:space:]]Understanding[[:space:]]the[[:space:]]Hierarchy[[:space:]]of[[:space:]]Human[[:space:]]Behavior[[:space:]]Enhances[[:space:]]Reasoning[[:space:]]on[[:space:]]Egocentric[[:space:]]Videos/4e0055ef-ae5b-49ce-b346-453e7160ce4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HiGarment_[[:space:]]Cross-modal[[:space:]]Harmony[[:space:]]Based[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Flat[[:space:]]Sketch[[:space:]]to[[:space:]]Realistic[[:space:]]Garment[[:space:]]Image/8ed195a0-2b02-45af-b0f3-1c132b5fcf07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HiMTok_[[:space:]]Learning[[:space:]]Hierarchical[[:space:]]Mask[[:space:]]Tokens[[:space:]]for[[:space:]]Image[[:space:]]Segmentation[[:space:]]with[[:space:]]Large[[:space:]]Multimodal[[:space:]]Model/cce1e939-48b9-4526-b2dc-49e9252ac1d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HiNeuS_[[:space:]]High-fidelity[[:space:]]Neural[[:space:]]Surface[[:space:]]Mitigating[[:space:]]Low-texture[[:space:]]and[[:space:]]Reflective[[:space:]]Ambiguity/d9db011f-3670-4218-a5b8-8853b3417e23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HiP-AD_[[:space:]]Hierarchical[[:space:]]and[[:space:]]Multi-Granularity[[:space:]]Planning[[:space:]]with[[:space:]]Deformable[[:space:]]Attention[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving[[:space:]]in[[:space:]]a[[:space:]]Single[[:space:]]Decoder/30239132-b2b8-486c-9ba6-f41d661abf9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]3D[[:space:]]Scene[[:space:]]Graphs[[:space:]]Construction[[:space:]]Outdoors/3c5e487d-ff62-43f1-ba40-c05db36f940b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]Cross-modal[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/b75d43af-eb63-48ed-942d-cc9873634efe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]Divide-and-Conquer[[:space:]]Grouping[[:space:]]for[[:space:]]Classification[[:space:]]Adaptation[[:space:]]of[[:space:]]Pre-Trained[[:space:]]Models/38b3a61a-8517-47b0-99c4-2cbd64aa4801_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]Event[[:space:]]Memory[[:space:]]for[[:space:]]Accurate[[:space:]]and[[:space:]]Low-latency[[:space:]]Online[[:space:]]Video[[:space:]]Temporal[[:space:]]Grounding/ae4226c7-4d1f-441f-a214-0f73133687f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]Material[[:space:]]Recognition[[:space:]]from[[:space:]]Local[[:space:]]Appearance/4f2ccd47-7059-4e7e-968d-a9a2b121a5cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]Variational[[:space:]]Test-Time[[:space:]]Prompt[[:space:]]Generation[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Generalization/30a69522-4559-440c-929e-b93e48ffd24e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical[[:space:]]Visual[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Continual[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation/e8e8d0bc-23bc-4b39-bbd7-7cde9087f698_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchical-aware[[:space:]]Orthogonal[[:space:]]Disentanglement[[:space:]]Framework[[:space:]]for[[:space:]]Fine-grained[[:space:]]Skeleton-based[[:space:]]Action[[:space:]]Recognition/f3e1a8e6-ef81-4943-ae34-0528df59d8b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchy[[:space:]]UGP_[[:space:]]Hierarchy[[:space:]]Unified[[:space:]]Gaussian[[:space:]]Primitive[[:space:]]for[[:space:]]Large-Scale[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Reconstruction/b34f4d44-8d3b-49dd-b709-fae10e60a485_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hierarchy-Aware[[:space:]]Pseudo[[:space:]]Word[[:space:]]Learning[[:space:]]with[[:space:]]Text[[:space:]]Adaptation[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/6b56f35f-d688-4693-b51a-4ce61a86e2c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/High-Precision[[:space:]]3D[[:space:]]Measurement[[:space:]]of[[:space:]]Complex[[:space:]]Textured[[:space:]]Surfaces[[:space:]]Using[[:space:]]Multiple[[:space:]]Filtering[[:space:]]Approach/ac034765-ab73-43a8-8507-f2c54a26d40b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/High-Resolution[[:space:]]Spatiotemporal[[:space:]]Modeling[[:space:]]with[[:space:]]Global-Local[[:space:]]State[[:space:]]Space[[:space:]]Models[[:space:]]for[[:space:]]Video-Based[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/1ed1cf4f-6ccb-47dd-970a-5d71673f28be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Highlight[[:space:]]What[[:space:]]You[[:space:]]Want_[[:space:]]Weakly-Supervised[[:space:]]Instance-Level[[:space:]]Controllable[[:space:]]Infrared-Visible[[:space:]]Image[[:space:]]Fusion/1de21735-785c-4cbb-a4a3-42e7d71bc7c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hints[[:space:]]of[[:space:]]Prompt_[[:space:]]Enhancing[[:space:]]Visual[[:space:]]Representation[[:space:]]for[[:space:]]Multimodal[[:space:]]LLMs[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/190720b4-f2af-42a4-a67b-3cf61fe738ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hipandas_[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Joint[[:space:]]Denoising[[:space:]]and[[:space:]]Super-Resolution[[:space:]]by[[:space:]]Image[[:space:]]Fusion[[:space:]]with[[:space:]]the[[:space:]]Panchromatic[[:space:]]Image/ef24e3f2-e971-4ffe-bfed-2166b4e47bed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HoliTracer_[[:space:]]Holistic[[:space:]]Vectorization[[:space:]]of[[:space:]]Geographic[[:space:]]Objects[[:space:]]from[[:space:]]Large-Size[[:space:]]Remote[[:space:]]Sensing[[:space:]]Imagery/c94f3b52-8898-462e-96f4-8c8af9ba33b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Holistic[[:space:]]Tokenizer[[:space:]]for[[:space:]]Autoregressive[[:space:]]Image[[:space:]]Generation/5caa4478-1280-457d-bedd-918fb7ddf9c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Holistic[[:space:]]Unlearning[[:space:]]Benchmark_[[:space:]]A[[:space:]]Multi-Faceted[[:space:]]Evaluation[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Model[[:space:]]Unlearning/c36a1c94-16f1-413f-9c57-0d3b286c1836_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HouseCrafter_[[:space:]]Lifting[[:space:]]Floorplans[[:space:]]to[[:space:]]3D[[:space:]]Scenes[[:space:]]with[[:space:]]2D[[:space:]]Diffusion[[:space:]]Models/f886b3d4-64d4-4d21-a0bf-1332044736c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HouseTour_[[:space:]]A[[:space:]]Virtual[[:space:]]Real[[:space:]]Estate[[:space:]]A(I)gent/cca057e1-98cf-4d48-b8af-dd722c88ff7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/How[[:space:]]Can[[:space:]]Objects[[:space:]]Help[[:space:]]Video-Language[[:space:]]Understanding_/84dbee9e-3165-434c-a753-48eec294085c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/How[[:space:]]Do[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]Handle[[:space:]]Complex[[:space:]]Multimodal[[:space:]]Reasoning_[[:space:]]Placing[[:space:]]Them[[:space:]]in[[:space:]]An[[:space:]]Extensible[[:space:]]Escape[[:space:]]Game/b2a6c399-1371-4f15-99b9-1112ebf8096f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/How[[:space:]]Do[[:space:]]Optical[[:space:]]Flow[[:space:]]and[[:space:]]Textual[[:space:]]Prompts[[:space:]]Collaborate[[:space:]]to[[:space:]]Assist[[:space:]]in[[:space:]]Audio-Visual[[:space:]]Semantic[[:space:]]Segmentation_/f6db52cd-d870-4761-9a24-e2526caa8e9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/How[[:space:]]Far[[:space:]]are[[:space:]]AI-generated[[:space:]]Videos[[:space:]]from[[:space:]]Simulating[[:space:]]the[[:space:]]3D[[:space:]]Visual[[:space:]]World_[[:space:]]A[[:space:]]Learned[[:space:]]3D[[:space:]]Evaluation[[:space:]]Approach/594deaab-7e32-4a0f-808e-5ef20b252f97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/How[[:space:]]To[[:space:]]Make[[:space:]]Your[[:space:]]Cell[[:space:]]Tracker[[:space:]]Say[[:space:]]'I[[:space:]]dunno!/c6165eaf-f023-4e5d-81cd-a42cc658d168_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/How[[:space:]]Would[[:space:]]It[[:space:]]Sound_[[:space:]]Material-Controlled[[:space:]]Multimodal[[:space:]]Acoustic[[:space:]]Profile[[:space:]]Generation[[:space:]]for[[:space:]]Indoor[[:space:]]Scenes/eed9dfdb-32c8-453d-a8e3-117738ed05ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Human-Object[[:space:]]Interaction[[:space:]]from[[:space:]]Human-Level[[:space:]]Instructions/965ba74b-7533-4018-932f-8c68c61d5a57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Human-in-the-Loop[[:space:]]Local[[:space:]]Corrections[[:space:]]of[[:space:]]3D[[:space:]]Scene[[:space:]]Layouts[[:space:]]via[[:space:]]Infilling/555142a6-4b12-40e5-9116-c1155f13d9ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HumanOLAT_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]for[[:space:]]Full-Body[[:space:]]Human[[:space:]]Relighting[[:space:]]and[[:space:]]Novel-View[[:space:]]Synthesis/f2bdbe62-bf64-417e-a6a4-e92cfa2ad25d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HumanSAM_[[:space:]]Classifying[[:space:]]Human-centric[[:space:]]Forgery[[:space:]]Videos[[:space:]]in[[:space:]]Human[[:space:]]Spatial,[[:space:]]Appearance,[[:space:]]and[[:space:]]Motion[[:space:]]Anomaly/f45216ad-63ee-47ac-bfd1-43a7928c15dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Humans[[:space:]]as[[:space:]]Checkerboards_[[:space:]]Calibrating[[:space:]]Camera[[:space:]]Motion[[:space:]]Scale[[:space:]]for[[:space:]]World-Coordinate[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/c476812d-2440-4c82-85bd-375a2e5bdfda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Humans[[:space:]]as[[:space:]]a[[:space:]]Calibration[[:space:]]Pattern_[[:space:]]Dynamic[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]from[[:space:]]Unsynchronized[[:space:]]and[[:space:]]Uncalibrated[[:space:]]Videos/66df3486-8a0c-4370-ba18-0f27a968b25b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HumorDB_[[:space:]]Can[[:space:]]AI[[:space:]]understand[[:space:]]graphical[[:space:]]humor_/56b8b480-a6b5-4580-9e2a-01639a43ed80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HyPiDecoder_[[:space:]]Hybrid[[:space:]]Pixel[[:space:]]Decoder[[:space:]]for[[:space:]]Efficient[[:space:]]Segmentation[[:space:]]and[[:space:]]Detection/6a0bf5d2-4524-4a2a-ab1a-376df424bc0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HyTIP_[[:space:]]Hybrid[[:space:]]Temporal[[:space:]]Information[[:space:]]Propagation[[:space:]]for[[:space:]]Masked[[:space:]]Conditional[[:space:]]Residual[[:space:]]Video[[:space:]]Coding/d29b7377-27e6-4904-8ab3-03657c8e20dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hybrid[[:space:]]Layout[[:space:]]Control[[:space:]]for[[:space:]]Diffusion[[:space:]]Transformer_[[:space:]]Fewer[[:space:]]Annotations,[[:space:]]Superior[[:space:]]Aesthetics/ae1e59b2-197a-4841-8853-922c8a836d50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hybrid-TTA_[[:space:]]Continual[[:space:]]Test-time[[:space:]]Adaptation[[:space:]]via[[:space:]]Dynamic[[:space:]]Domain[[:space:]]Shift[[:space:]]Detection/bcb7cb16-4727-40c7-bfd4-a30493141b83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hybrid-Tower_[[:space:]]Fine-grained[[:space:]]Pseudo-query[[:space:]]Interaction[[:space:]]and[[:space:]]Generation[[:space:]]for[[:space:]]Text-to-Video[[:space:]]Retrieval/7cedaf0e-f228-4b51-92e4-bb7f96e3ea22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hybrid-grained[[:space:]]Feature[[:space:]]Aggregation[[:space:]]with[[:space:]]Coarse-to-fine[[:space:]]Language[[:space:]]Guidance[[:space:]]for[[:space:]]Self-supervised[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/d8f58388-9087-4140-8fcd-3fb3175e5e61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hydra-NeXt_[[:space:]]Robust[[:space:]]Closed-Loop[[:space:]]Driving[[:space:]]with[[:space:]]Open-Loop[[:space:]]Training/9ad39af3-300c-4bd7-bb44-5a19870476d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HypDAE_[[:space:]]Hyperbolic[[:space:]]Diffusion[[:space:]]Autoencoders[[:space:]]for[[:space:]]Hierarchical[[:space:]]Few-shot[[:space:]]Image[[:space:]]Generation/f9f8dd56-6920-4113-ab91-70aa6d5d7b30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hyper-Depth_[[:space:]]Hypergraph-based[[:space:]]Multi-Scale[[:space:]]Representation[[:space:]]Fusion[[:space:]]for[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/9065bd02-894a-4492-84a7-2f41ed6a6b52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/HyperGCT_[[:space:]]A[[:space:]]Dynamic[[:space:]]Hyper-GNN-Learned[[:space:]]Geometric[[:space:]]Constraint[[:space:]]for[[:space:]]3D[[:space:]]Registration/47ca0850-f66e-4267-a9c8-3815b72effdc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Hypergraph[[:space:]]Clustering[[:space:]]Network[[:space:]]with[[:space:]]Partial[[:space:]]Attribute[[:space:]]Imputation/02fe5e20-640a-4f04-9577-3b19a285253b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/I[[:space:]]Am[[:space:]]Big,[[:space:]]You[[:space:]]Are[[:space:]]Little;[[:space:]]I[[:space:]]Am[[:space:]]Right,[[:space:]]You[[:space:]]Are[[:space:]]Wrong/93dc2a84-ee2a-4bbf-ac4a-8f024364dc0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/I2-World_[[:space:]]Intra-Inter[[:space:]]Tokenization[[:space:]]for[[:space:]]Efficient[[:space:]]Dynamic[[:space:]]4D[[:space:]]Scene[[:space:]]Forecasting/00ebed8e-83b6-42d4-9040-2931e585d1bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/I2V3D_[[:space:]]Controllable[[:space:]]Image-to-video[[:space:]]Generation[[:space:]]with[[:space:]]3D[[:space:]]Guidance/34908e4c-a1d2-4167-8e16-e13f8f0f79f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/I2VControl_[[:space:]]Disentangled[[:space:]]and[[:space:]]Unified[[:space:]]Video[[:space:]]Motion[[:space:]]Synthesis[[:space:]]Control/32e60292-bb6e-4339-824d-2e7360d03f42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IAP_[[:space:]]Invisible[[:space:]]Adversarial[[:space:]]Patch[[:space:]]Attack[[:space:]]through[[:space:]]Perceptibility-Aware[[:space:]]Localization[[:space:]]and[[:space:]]Perturbation[[:space:]]Optimization/f44a67b5-686d-4c28-8d64-841c9af4ca60_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ICE-Bench_[[:space:]]A[[:space:]]Unified[[:space:]]and[[:space:]]Comprehensive[[:space:]]Benchmark[[:space:]]for[[:space:]]Image[[:space:]]Creating[[:space:]]and[[:space:]]Editing/98156c06-fb1b-4306-b16c-31dd6dc2a05b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IDEATOR_[[:space:]]Jailbreaking[[:space:]]and[[:space:]]Benchmarking[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]Using[[:space:]]Themselves/45d8d302-b4d4-4da0-b545-ced98ef47fb7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IDF_[[:space:]]Iterative[[:space:]]Dynamic[[:space:]]Filtering[[:space:]]Networks[[:space:]]for[[:space:]]Generalizable[[:space:]]Image[[:space:]]Denoising/c2022b07-066b-45cb-941a-d87a4d6faead_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IDFace_[[:space:]]Face[[:space:]]Template[[:space:]]Protection[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Secure[[:space:]]Identification/747f110d-11a4-4062-8d8d-39ebd54b89f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IFAdapter_[[:space:]]Instance[[:space:]]Feature[[:space:]]Control[[:space:]]for[[:space:]]Grounded[[:space:]]Text-to-Image[[:space:]]Generation/73ad11f7-fce1-42a5-8996-86044a8d5d92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IGD_[[:space:]]Instructional[[:space:]]Graphic[[:space:]]Design[[:space:]]with[[:space:]]Multimodal[[:space:]]Layer[[:space:]]Generation/3b5edf5f-e4c2-48c6-a721-b7dffeee359b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IGL-Nav_[[:space:]]Incremental[[:space:]]3D[[:space:]]Gaussian[[:space:]]Localization[[:space:]]for[[:space:]]Image-goal[[:space:]]Navigation/2480563e-af1b-454d-9823-446ab2c18889_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ILLUME_[[:space:]]Illuminating[[:space:]]Your[[:space:]]LLMs[[:space:]]to[[:space:]]See,[[:space:]]Draw,[[:space:]]and[[:space:]]Self-Enhance/ffb49b41-a10c-49cd-bdfa-da203ee060d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IM-LUT_[[:space:]]Interpolation[[:space:]]Mixing[[:space:]]Look-Up[[:space:]]Tables[[:space:]]for[[:space:]]Image[[:space:]]Super-Resolution/8384010a-2025-4447-84e8-7f17cd121a80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IM360_[[:space:]]Large-scale[[:space:]]Indoor[[:space:]]Mapping[[:space:]]with[[:space:]]360[[:space:]]Cameras/ec132d3f-d3d3-4783-9e61-9ca3adea3f31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IMG_[[:space:]]Calibrating[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Implicit[[:space:]]Multimodal[[:space:]]Guidance/3b530dc1-965b-4273-9e99-80a18c3e4dd3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IMoRe_[[:space:]]Implicit[[:space:]]Program-Guided[[:space:]]Reasoning[[:space:]]for[[:space:]]Human[[:space:]]Motion[[:space:]]Q&A/89f73a18-4876-4ad1-bea2-88a5e937fb7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/INS-MMBench_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Benchmark[[:space:]]for[[:space:]]Evaluating[[:space:]]LVLMs'[[:space:]]Performance[[:space:]]in[[:space:]]Insurance/964a63fe-3963-439a-a4c5-e03caa1805e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/INSTINCT_[[:space:]]Instance-Level[[:space:]]Interaction[[:space:]]Architecture[[:space:]]for[[:space:]]Query-Based[[:space:]]Collaborative[[:space:]]Perception/71699507-04e2-4efc-87f9-0d91a8fe4524_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/INTER_[[:space:]]Mitigating[[:space:]]Hallucination[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]by[[:space:]]Interaction[[:space:]]Guidance[[:space:]]Sampling/8b4ebcd5-37f9-4ebe-ba49-6a35fc1237bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IQA-Adapter_[[:space:]]Exploring[[:space:]]Knowledge[[:space:]]Transfer[[:space:]]from[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment[[:space:]]to[[:space:]]Diffusion-based[[:space:]]Generative[[:space:]]Models/b8c0b25b-73bd-4ffe-9309-3fe52e52f97b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IRASim_[[:space:]]A[[:space:]]Fine-Grained[[:space:]]World[[:space:]]Model[[:space:]]for[[:space:]]Robot[[:space:]]Manipulation/d82a8bc7-4df5-4c4b-970b-dfc86f715eba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IRGPT_[[:space:]]Understanding[[:space:]]Real-world[[:space:]]Infrared[[:space:]]Image[[:space:]]with[[:space:]]Bi-cross-modal[[:space:]]Curriculum[[:space:]]on[[:space:]]Large-scale[[:space:]]Benchmark/cf73aea8-2657-4332-a281-4e009567983b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ISP2HRNet_[[:space:]]Learning[[:space:]]to[[:space:]]Reconstruct[[:space:]]High[[:space:]]Resolution[[:space:]]Image[[:space:]]from[[:space:]]Irregularly[[:space:]]Sampled[[:space:]]Pixels[[:space:]]via[[:space:]]Hierarchical[[:space:]]Gradient[[:space:]]Learning/3f18d859-4360-4f53-9b3e-5a36edc44bbf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Identity[[:space:]]Preserving[[:space:]]3D[[:space:]]Head[[:space:]]Stylization[[:space:]]with[[:space:]]Multiview[[:space:]]Score[[:space:]]Distillation/f2355bca-35f0-4470-88ed-e02e04703dbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Identity-aware[[:space:]]Language[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Open-vocabulary[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation/8fab492f-bba4-4fda-b200-52f75ddecdb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Im2Haircut_[[:space:]]Single-view[[:space:]]Strand-based[[:space:]]Hair[[:space:]]Reconstruction[[:space:]]for[[:space:]]Human[[:space:]]Avatars/fcc16100-9a30-4dc1-af13-9658c0c127ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ImHead_[[:space:]]A[[:space:]]Large-scale[[:space:]]Implicit[[:space:]]Morphable[[:space:]]Model[[:space:]]for[[:space:]]Localized[[:space:]]Head[[:space:]]Modeling/be3131b4-270a-4e13-868a-00bb9bc07fe4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Image[[:space:]]Intrinsic[[:space:]]Scale[[:space:]]Assessment_[[:space:]]Bridging[[:space:]]the[[:space:]]Gap[[:space:]]Between[[:space:]]Quality[[:space:]]and[[:space:]]Resolution/e31d0931-07ad-4aaa-b867-cf4b74c0fbdf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Image[[:space:]]as[[:space:]]an[[:space:]]IMU_[[:space:]]Estimating[[:space:]]Camera[[:space:]]Motion[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Motion-Blurred[[:space:]]Image/ee41ffb7-9dbd-40a7-bd0e-612926138961_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Image-Guided[[:space:]]Shape-from-Template[[:space:]]Using[[:space:]]Mesh[[:space:]]Inextensibility[[:space:]]Constraints/2c89590c-2db1-49b0-a4a9-81889811eb1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ImageGem_[[:space:]]In-the-wild[[:space:]]Generative[[:space:]]Image[[:space:]]Interaction[[:space:]]Dataset[[:space:]]for[[:space:]]Generative[[:space:]]Model[[:space:]]Personalization/07f986b8-430c-4c20-a4b6-b1a9d274c6e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ImageGen-CoT_[[:space:]]Enhancing[[:space:]]Text-to-Image[[:space:]]In-context[[:space:]]Learning[[:space:]]with[[:space:]]Chain-of-Thought[[:space:]]Reasoning/ac26b8d6-f4ca-4b6d-bd9c-6b59df193496_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Images[[:space:]]as[[:space:]]Noisy[[:space:]]Labels_[[:space:]]Unleashing[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]the[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/daed4996-19a5-4b34-9b9a-c91c1aabe556_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Imbalance[[:space:]]in[[:space:]]Balance_[[:space:]]Online[[:space:]]Concept[[:space:]]Balancing[[:space:]]in[[:space:]]Generation[[:space:]]Models/ef44cdaf-e268-4114-bad7-942ce19f280c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Implicit[[:space:]]Counterfactual[[:space:]]Learning[[:space:]]for[[:space:]]Audio-Visual[[:space:]]Segmentation/45bf89a7-fcec-4403-8d6d-9dfdac1522cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Importance-Based[[:space:]]Token[[:space:]]Merging[[:space:]]for[[:space:]]Efficient[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Generation/a0b24f99-b179-4777-95d9-12e24ffc73a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improved[[:space:]]Noise[[:space:]]Schedule[[:space:]]for[[:space:]]Diffusion[[:space:]]Training/4e080121-9724-4263-a31a-a10e306d71f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Large[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Models[[:space:]]by[[:space:]]Learning[[:space:]]from[[:space:]]a[[:space:]]Panel[[:space:]]of[[:space:]]Peers/3e7f63a4-b790-4e90-835f-268233c74fb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Multimodal[[:space:]]Learning[[:space:]]via[[:space:]]Imbalanced[[:space:]]Learning/47d2a2df-f87c-4dd8-96b5-5e3e456de276_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Noise[[:space:]]Efficiency[[:space:]]in[[:space:]]Privacy-preserving[[:space:]]Dataset[[:space:]]Distillation/89fd1f10-3ed4-4f82-a705-c28b281ea02a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]Rectified[[:space:]]Flow[[:space:]]with[[:space:]]Boundary[[:space:]]Conditions/03b146ca-6f32-410d-af30-65ea8a8df14a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Improving[[:space:]]SAM[[:space:]]for[[:space:]]Camouflaged[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Dual[[:space:]]Stream[[:space:]]Adapters/0399a812-eef2-401f-9e97-2c2eace272ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Incremental[[:space:]]Few-Shot[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]via[[:space:]]Multi-Level[[:space:]]Switchable[[:space:]]Visual[[:space:]]Prompts/50283e09-a8d8-4c79-91b1-15423091545c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InfGen_[[:space:]]A[[:space:]]Resolution-Agnostic[[:space:]]Paradigm[[:space:]]for[[:space:]]Scalable[[:space:]]Image[[:space:]]Synthesis/a0b8f9b9-6c61-4354-bb8c-43cd64b8f980_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Inference-Time[[:space:]]Diffusion[[:space:]]Model[[:space:]]Distillation/27c49697-1601-4e38-8fae-2a6e5d0bedd0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InfiniCube_[[:space:]]Unbounded[[:space:]]and[[:space:]]Controllable[[:space:]]Dynamic[[:space:]]3D[[:space:]]Driving[[:space:]]Scene[[:space:]]Generation[[:space:]]with[[:space:]]World-Guided[[:space:]]Video[[:space:]]Models/32192950-76dc-49a1-ba22-792c1a754407_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InfiniDreamer_[[:space:]]Arbitrarily[[:space:]]Long[[:space:]]Human[[:space:]]Motion[[:space:]]Generation[[:space:]]via[[:space:]]Segment[[:space:]]Score[[:space:]]Distillation/ca60fd0f-0089-4e40-aa44-40b0171d565d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InfiniteYou_[[:space:]]Flexible[[:space:]]Photo[[:space:]]Recrafting[[:space:]]While[[:space:]]Preserving[[:space:]]Your[[:space:]]Identity/e9969243-0d30-4beb-9f25-4be61a2fade2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InfoBridge_[[:space:]]Balanced[[:space:]]Multimodal[[:space:]]Integration[[:space:]]through[[:space:]]Conditional[[:space:]]Dependency[[:space:]]Modeling/484cd62a-a822-4476-8c1e-f550697ce1cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Information[[:space:]]Density[[:space:]]Principle[[:space:]]for[[:space:]]MLLM[[:space:]]Benchmarks/02b316e1-9a6f-4594-a2f1-a8ad2fddca19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Information-Bottleneck[[:space:]]Driven[[:space:]]Binary[[:space:]]Neural[[:space:]]Network[[:space:]]for[[:space:]]Change[[:space:]]Detection/03f651a5-7f55-4f7c-a098-ce9849315716_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Inpaint4Drag_[[:space:]]Repurposing[[:space:]]Inpainting[[:space:]]Models[[:space:]]for[[:space:]]Drag-Based[[:space:]]Image[[:space:]]Editing[[:space:]]via[[:space:]]Bidirectional[[:space:]]Warping/ce06495e-9535-4c82-bc48-6b99d4d3a6a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InsViE-1M_[[:space:]]Effective[[:space:]]Instruction-based[[:space:]]Video[[:space:]]Editing[[:space:]]with[[:space:]]Elaborate[[:space:]]Dataset[[:space:]]Construction/834d13ee-81f1-4e68-933f-3d5e48dae9ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InsideOut_[[:space:]]Integrated[[:space:]]RGB-Radiative[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Comprehensive[[:space:]]3D[[:space:]]Object[[:space:]]Representation/c7b12df3-a499-4deb-a3bd-20757028e1f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InstaDrive_[[:space:]]Instance-Aware[[:space:]]Driving[[:space:]]World[[:space:]]Models[[:space:]]for[[:space:]]Realistic[[:space:]]and[[:space:]]Consistent[[:space:]]Video[[:space:]]Generation/657a9961-ffcb-40c1-ae5e-dd5c77728ad7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InstaScene_[[:space:]]Towards[[:space:]]Complete[[:space:]]3D[[:space:]]Instance[[:space:]]Decomposition[[:space:]]and[[:space:]]Reconstruction[[:space:]]from[[:space:]]Cluttered[[:space:]]Scenes/a94c3c83-abeb-478e-b7f6-6c847d70363e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Instance-Level[[:space:]]Video[[:space:]]Depth[[:space:]]in[[:space:]]Groups[[:space:]]Beyond[[:space:]]Occlusions/9b3b552c-8ca9-4cc8-9e53-515673c11d56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Instant[[:space:]]GaussianImage_[[:space:]]A[[:space:]]Generalizable[[:space:]]and[[:space:]]Self-Adaptive[[:space:]]Image[[:space:]]Representation[[:space:]]via[[:space:]]2D[[:space:]]Gaussian[[:space:]]Splatting/7f8119c3-6d38-48ab-88fe-7f01f61c86ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InstantEdit_[[:space:]]Text-Guided[[:space:]]Few-Step[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Piecewise[[:space:]]Rectified[[:space:]]Flow/59ace991-a4c0-4de8-b2da-824b32195a11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InstructSeg_[[:space:]]Unifying[[:space:]]Instructed[[:space:]]Visual[[:space:]]Segmentation[[:space:]]with[[:space:]]Multi-modal[[:space:]]Large[[:space:]]Language[[:space:]]Models/31cd4d06-a652-4ebf-860e-edebd70056a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Instruction-Grounded[[:space:]]Visual[[:space:]]Projectors[[:space:]]for[[:space:]]Continual[[:space:]]Learning[[:space:]]of[[:space:]]Generative[[:space:]]Vision-Language[[:space:]]Models/cc1a573c-d883-4aae-87a6-207f42efaeca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Instruction-Oriented[[:space:]]Preference[[:space:]]Alignment[[:space:]]for[[:space:]]Enhancing[[:space:]]Multi-Modal[[:space:]]Comprehension[[:space:]]Capability[[:space:]]of[[:space:]]MLLMs/2dedb3ff-ee46-42f6-b8ea-73831b487c2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Instruction-based[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Planning,[[:space:]]Reasoning,[[:space:]]and[[:space:]]Generation/f131fae0-5c50-4f2e-8a5f-9e0b6e47107b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Integrating[[:space:]]Biological[[:space:]]Knowledge[[:space:]]for[[:space:]]Robust[[:space:]]Microscopy[[:space:]]Image[[:space:]]Profiling[[:space:]]on[[:space:]]De[[:space:]]Novo[[:space:]]Cell[[:space:]]Lines/ce63f5cf-7bdb-420b-9963-1d354dcf26e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Integrating[[:space:]]Task-Specific[[:space:]]and[[:space:]]Universal[[:space:]]Adapters[[:space:]]for[[:space:]]Pre-Trained[[:space:]]Model-based[[:space:]]Class-Incremental[[:space:]]Learning/58f8aeda-464c-4aa4-8259-d173990557ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Integrating[[:space:]]Visual[[:space:]]Interpretation[[:space:]]and[[:space:]]Linguistic[[:space:]]Reasoning[[:space:]]for[[:space:]]Geometric[[:space:]]Problem[[:space:]]Solving/39c2703f-ed62-4226-9464-8e49764d6091_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Inter2Former_[[:space:]]Dynamic[[:space:]]Hybrid[[:space:]]Attention[[:space:]]for[[:space:]]Efficient[[:space:]]High-Precision[[:space:]]Interactive[[:space:]]Segmentation/9089ef17-4cad-4e59-af3c-33a2c40b7600_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InterGSEdit_[[:space:]]Interactive[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Editing[[:space:]]with[[:space:]]3D[[:space:]]Geometry-Consistent[[:space:]]Attention[[:space:]]Prior/daadb754-7b6f-404e-9919-9d81d4f12d81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InterSyn_[[:space:]]Interleaved[[:space:]]Learning[[:space:]]for[[:space:]]Dynamic[[:space:]]Motion[[:space:]]Synthesis[[:space:]]in[[:space:]]the[[:space:]]Wild/cdcfdcf9-2088-45d4-8dcc-8a99b25db8f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InteractAvatar_[[:space:]]Modeling[[:space:]]Hand-Face[[:space:]]Interaction[[:space:]]in[[:space:]]Photorealistic[[:space:]]Avatars[[:space:]]with[[:space:]]Deformable[[:space:]]Gaussians/bdd96679-fb47-4ec5-b2d4-3739300fd41f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Interaction-Merged[[:space:]]Motion[[:space:]]Planning_[[:space:]]Effectively[[:space:]]Leveraging[[:space:]]Diverse[[:space:]]Motion[[:space:]]Datasets[[:space:]]for[[:space:]]Robust[[:space:]]Planning/85d40c7e-53c3-41af-a88f-db2ef8accfa1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Intermediate[[:space:]]Connectors[[:space:]]and[[:space:]]Geometric[[:space:]]Priors[[:space:]]for[[:space:]]Language-Guided[[:space:]]Affordance[[:space:]]Segmentation[[:space:]]on[[:space:]]Unseen[[:space:]]Object[[:space:]]Categories/63d9ae4a-0df0-4059-9e18-65ed61dea285_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Interpretable[[:space:]]Zero-Shot[[:space:]]Learning[[:space:]]with[[:space:]]Locally-Aligned[[:space:]]Vision-Language[[:space:]]Model/8d474245-4210-48e4-9cb5-dff75897bd85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Interpretable[[:space:]]point[[:space:]]cloud[[:space:]]classification[[:space:]]using[[:space:]]multiple[[:space:]]instance[[:space:]]learning/a81988e5-abba-4de9-9142-f9a36f14125e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Intervening[[:space:]]in[[:space:]]Black[[:space:]]Box_[[:space:]]Concept[[:space:]]Bottleneck[[:space:]]Model[[:space:]]for[[:space:]]Enhancing[[:space:]]Human[[:space:]]Neural[[:space:]]Network[[:space:]]Mutual[[:space:]]Understanding/137eb634-ef2a-45e3-bbfc-df7ea1b8115b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Intra-modal[[:space:]]and[[:space:]]Cross-modal[[:space:]]Synchronization[[:space:]]for[[:space:]]Audio-visual[[:space:]]Deepfake[[:space:]]Detection[[:space:]]and[[:space:]]Temporal[[:space:]]Localization/c6a839cf-ecb6-45b7-aa62-a7c0ec9aa2b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Intra-view[[:space:]]and[[:space:]]Inter-view[[:space:]]Correlation[[:space:]]Guided[[:space:]]Multi-view[[:space:]]Novel[[:space:]]Class[[:space:]]Discovery/e0f931a7-933e-49a6-9011-046b84d9772c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IntrinsicControlNet_[[:space:]]Cross-distribution[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Real[[:space:]]and[[:space:]]Unreal/a54a100a-0023-4974-b401-d5d7e5b64353_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/IntroStyle_[[:space:]]Training-Free[[:space:]]Introspective[[:space:]]Style[[:space:]]Attribution[[:space:]]using[[:space:]]Diffusion[[:space:]]Features/05410c10-c62e-48d3-aef5-6ed0e69f5149_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/InvRGB+L_[[:space:]]Inverse[[:space:]]Rendering[[:space:]]of[[:space:]]Complex[[:space:]]Scenes[[:space:]]with[[:space:]]Unified[[:space:]]Color[[:space:]]and[[:space:]]LiDAR[[:space:]]Reflectance[[:space:]]Modeling/f431b5e2-7627-45a1-bacd-e69436aa8b2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Inverse[[:space:]]3D[[:space:]]Microscopy[[:space:]]Rendering[[:space:]]for[[:space:]]Cell[[:space:]]Shape[[:space:]]Inference[[:space:]]with[[:space:]]Active[[:space:]]Mesh/2cc71dbf-705b-4a29-94ee-35adc63e8ca2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Inverse[[:space:]]Image-Based[[:space:]]Rendering[[:space:]]for[[:space:]]Light[[:space:]]Field[[:space:]]Generation[[:space:]]from[[:space:]]Single[[:space:]]Images/39f2b7af-c4d0-4d49-84ac-f88772dc429f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Invisible[[:space:]]Watermarks,[[:space:]]Visible[[:space:]]Gains_[[:space:]]Steering[[:space:]]Machine[[:space:]]Unlearning[[:space:]]with[[:space:]]Bi-Level[[:space:]]Watermarking[[:space:]]Design/ad9460aa-9fe3-48a0-90cc-30b4a4d5fcd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Iris_[[:space:]]Breaking[[:space:]]GUI[[:space:]]Complexity[[:space:]]with[[:space:]]Adaptive[[:space:]]Focus[[:space:]]and[[:space:]]Self-Refining/e0350618-d9cc-4c45-bd3b-355f1b9d8fba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Is[[:space:]]CLIP[[:space:]]ideal_[[:space:]]No.[[:space:]]Can[[:space:]]we[[:space:]]fix[[:space:]]it_[[:space:]]Yes!/f0b545cb-bfbf-419d-a3b4-68c5703379ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Is[[:space:]]Less[[:space:]]More_[[:space:]]Exploring[[:space:]]Token[[:space:]]Condensation[[:space:]]as[[:space:]]Training-free[[:space:]]Test-time[[:space:]]Adaptation/7c8fef83-7d78-44eb-93c8-9c249efba4f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Is[[:space:]]Meta-Learning[[:space:]]Out_[[:space:]]Rethinking[[:space:]]Unsupervised[[:space:]]Few-Shot[[:space:]]Classification[[:space:]]with[[:space:]]Limited[[:space:]]Entropy/a1b761e1-34a4-4312-a8ff-c4d7cf53898d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Is[[:space:]]Tracking[[:space:]]Really[[:space:]]More[[:space:]]Challenging[[:space:]]in[[:space:]]First[[:space:]]Person[[:space:]]Egocentric[[:space:]]Vision_/55a8a1d5-30e2-4fa4-80bf-ead871570d4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Is[[:space:]]Visual[[:space:]]in-Context[[:space:]]Learning[[:space:]]for[[:space:]]Compositional[[:space:]]Medical[[:space:]]Tasks[[:space:]]within[[:space:]]Reach_/2707997f-3774-4830-8339-796e987ce4df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/JPEG[[:space:]]Processing[[:space:]]Neural[[:space:]]Operator[[:space:]]for[[:space:]]Backward-Compatible[[:space:]]Coding/926ead2a-ebcc-4188-9da5-e5d0d151d2d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/JailbreakDiffBench_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Benchmark[[:space:]]for[[:space:]]Jailbreaking[[:space:]]Diffusion[[:space:]]Models/a850df95-90ac-477e-936c-6fc1a07da8c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Jailbreaking[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]via[[:space:]]Shuffle[[:space:]]Inconsistency/8840dd28-a2d6-49fc-b41e-8627a64b46f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Jigsaw++_[[:space:]]Imagining[[:space:]]Complete[[:space:]]Shape[[:space:]]Priors[[:space:]]for[[:space:]]Object[[:space:]]Reassembly/57087dd8-7f57-4ca4-88fb-903f25902ca5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Joint[[:space:]]Asymmetric[[:space:]]Loss[[:space:]]for[[:space:]]Learning[[:space:]]with[[:space:]]Noisy[[:space:]]Labels/e53fcf7d-7899-412a-83d4-413df7c89418_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Joint[[:space:]]Diffusion[[:space:]]Models[[:space:]]in[[:space:]]Continual[[:space:]]Learning/731611de-bb50-4ddc-9b19-68e9d0d9a1c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Joint[[:space:]]Learning[[:space:]]of[[:space:]]Pose[[:space:]]Regression[[:space:]]and[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]with[[:space:]]Score[[:space:]]Scaling[[:space:]]Sampling[[:space:]]for[[:space:]]Category-level[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation/2b66944b-30a5-4c2f-8d8a-4677509d5445_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Joint[[:space:]]Self-Supervised[[:space:]]Video[[:space:]]Alignment[[:space:]]and[[:space:]]Action[[:space:]]Segmentation/066fcacf-d0e4-4d90-8551-4c0b5f75a147_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Joint[[:space:]]Semantic[[:space:]]and[[:space:]]Rendering[[:space:]]Enhancements[[:space:]]in[[:space:]]3D[[:space:]]Gaussian[[:space:]]Modeling[[:space:]]with[[:space:]]Anisotropic[[:space:]]Local[[:space:]]Encoding/b3d4838d-1526-4cda-b686-8970ef9f95b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/JointDiT_[[:space:]]Enhancing[[:space:]]RGB-Depth[[:space:]]Joint[[:space:]]Modeling[[:space:]]with[[:space:]]Diffusion[[:space:]]Transformers/c99b7598-f64f-4da4-bde6-049bcd1d483f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/KDA_[[:space:]]Knowledge[[:space:]]Diffusion[[:space:]]Alignment[[:space:]]with[[:space:]]Enhanced[[:space:]]Context[[:space:]]for[[:space:]]Video[[:space:]]Temporal[[:space:]]Grounding/dc88eeb6-2102-4fba-9aef-7989318343b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/KOEnsAttack_[[:space:]]Towards[[:space:]]Efficient[[:space:]]Data-Free[[:space:]]Black-Box[[:space:]]Adversarial[[:space:]]Attacks[[:space:]]via[[:space:]]Knowledge-Orthogonalized[[:space:]]Substitute[[:space:]]Ensembles/63339767-c5e5-4156-8b1c-7e9e120dd664_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/KV-Edit_[[:space:]]Training-Free[[:space:]]Image[[:space:]]Editing[[:space:]]for[[:space:]]Precise[[:space:]]Background[[:space:]]Preservation/db1efb5a-cc80-4d92-9995-64a6f56f66bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Kaleidoscopic[[:space:]]Background[[:space:]]Attack_[[:space:]]Disrupting[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]Multi-Fold[[:space:]]Radial[[:space:]]Symmetry[[:space:]]Textures/d070b9ae-76c4-4e7e-87e1-7e1fd589d9a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Kaputt_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]for[[:space:]]Visual[[:space:]]Defect[[:space:]]Detection/3193448a-b76d-474e-a37f-e30835aa2c0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Keep[[:space:]]Your[[:space:]]Friends[[:space:]]Close,[[:space:]]and[[:space:]]Your[[:space:]]Enemies[[:space:]]Farther_[[:space:]]Distance-aware[[:space:]]Voxel-wise[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Multi-organ[[:space:]]Segmentation/1f40cb17-9cc4-4a3f-a2b0-e68776a3297c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Kestrel_[[:space:]]3D[[:space:]]Multimodal[[:space:]]LLM[[:space:]]for[[:space:]]Part-Aware[[:space:]]Grounded[[:space:]]Description/020760b7-4ef6-44d2-82be-57aea4fa8adb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Keyframe-oriented[[:space:]]Vision[[:space:]]Token[[:space:]]Pruning_[[:space:]]Enhancing[[:space:]]Efficiency[[:space:]]of[[:space:]]Large[[:space:]]Vision[[:space:]]Language[[:space:]]Models[[:space:]]on[[:space:]]Long-Form[[:space:]]Video[[:space:]]Processing/bde78e0b-ebd9-4465-b0bf-ab1297bdbdd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/KinMo_[[:space:]]Kinematic-aware[[:space:]]Human[[:space:]]Motion[[:space:]]Understanding[[:space:]]and[[:space:]]Generation/92a1e526-f2d3-44df-81ab-98dd75209dbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Know[[:space:]]'No'[[:space:]]Better_[[:space:]]A[[:space:]]Data-Driven[[:space:]]Approach[[:space:]]for[[:space:]]Enhancing[[:space:]]Negation[[:space:]]Awareness[[:space:]]in[[:space:]]CLIP/551daf80-bec0-402e-887d-c6411c967ceb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Know[[:space:]]Your[[:space:]]Attention[[:space:]]Maps_[[:space:]]Class-specific[[:space:]]Token[[:space:]]Masking[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/7be2e4f8-9ddb-405a-a0e6-ae804bfa4a31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Learned[[:space:]]Image[[:space:]]Compression/9696e170-d663-4dd2-a7ac-796eda2e41d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Knowledge[[:space:]]Distillation[[:space:]]with[[:space:]]Refined[[:space:]]Logits/2cedf8d5-4cdd-4e1e-9c45-617ceedccf62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Knowledge[[:space:]]Transfer[[:space:]]from[[:space:]]Interaction[[:space:]]Learning/6771f344-0dd8-4e1e-96a6-137c9e460bc0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Knowledge-Guided[[:space:]]Part[[:space:]]Segmentation/2c281643-fc70-4c9a-8025-59357eaf30eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LA-MOTR_[[:space:]]End-to-End[[:space:]]Multi-Object[[:space:]]Tracking[[:space:]]by[[:space:]]Learnable[[:space:]]Association/9f8417e0-e034-4e36-96e4-714630cd9589_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LACONIC_[[:space:]]A[[:space:]]3D[[:space:]]Layout[[:space:]]Adapter[[:space:]]for[[:space:]]Controllable[[:space:]]Image[[:space:]]Creation/baebfd64-f553-41fc-b854-03518c0954ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LANGTRAJ_[[:space:]]Diffusion[[:space:]]Model[[:space:]]and[[:space:]]Dataset[[:space:]]for[[:space:]]Language-Conditioned[[:space:]]Trajectory[[:space:]]Simulation/44795725-2a75-48f7-a39f-575aa7b51010_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LATINO-PRO_[[:space:]]LAtent[[:space:]]consisTency[[:space:]]INverse[[:space:]]sOlver[[:space:]]with[[:space:]]PRompt[[:space:]]Optimization/c09d3b93-b303-4fd5-9bcb-437aaad1dd3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LBM_[[:space:]]Latent[[:space:]]Bridge[[:space:]]Matching[[:space:]]for[[:space:]]Fast[[:space:]]Image-to-Image[[:space:]]Translation/b052ca5b-8f18-475c-8cf5-b82ee320f6e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LD-RPS_[[:space:]]Zero-Shot[[:space:]]Unified[[:space:]]Image[[:space:]]Restoration[[:space:]]via[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Recurrent[[:space:]]Posterior[[:space:]]Sampling/b5ef24c7-6ef0-48c2-83c6-496dde6ebfb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LDIP_[[:space:]]Long[[:space:]]Distance[[:space:]]Information[[:space:]]Propagation[[:space:]]for[[:space:]]Video[[:space:]]Super-Resolution/95a5b38b-f043-43e4-b9c0-35546726f6c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LDPose_[[:space:]]Towards[[:space:]]Inclusive[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]for[[:space:]]Limb-Deficient[[:space:]]Individuals[[:space:]]in[[:space:]]the[[:space:]]Wild/10a10d73-4c04-4f58-a48c-567b39302831_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LEGION_[[:space:]]Learning[[:space:]]to[[:space:]]Ground[[:space:]]and[[:space:]]Explain[[:space:]]for[[:space:]]Synthetic[[:space:]]Image[[:space:]]Detection/a0bf31b0-ae27-42e8-86e9-6b0299e41265_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LEGO-Maker_[[:space:]]A[[:space:]]Semantic-Driven[[:space:]]Algorithm[[:space:]]for[[:space:]]Text-to-3D[[:space:]]Generation/8857bd18-1e16-4a8d-9368-5724d44c8f42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LGA-Net_[[:space:]]Learning[[:space:]]Local[[:space:]]and[[:space:]]Global[[:space:]]Affinities[[:space:]]for[[:space:]]Sparse[[:space:]]Scribble[[:space:]]based[[:space:]]Image[[:space:]]Colorization/f9dde5a2-ae8a-4234-8d5b-fd9f949e7e5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LHM_[[:space:]]Large[[:space:]]Animatable[[:space:]]Human[[:space:]]Reconstruction[[:space:]]Model[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]to[[:space:]]3D[[:space:]]in[[:space:]]Seconds/4f23b92b-0c84-4b05-bc2c-eef2cde0c91c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LIFT_[[:space:]]Latent[[:space:]]Implicit[[:space:]]Functions[[:space:]]for[[:space:]]Task-[[:space:]]and[[:space:]]Data-Agnostic[[:space:]]Encoding/a4ac6768-a76e-4dd5-80cc-6ff40d372910_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LINR-PCGC_[[:space:]]Lossless[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Geometry[[:space:]]Compression/21d17f7e-a58b-403d-8284-8af3527e8cc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LIRA_[[:space:]]Inferring[[:space:]]Segmentation[[:space:]]in[[:space:]]Large[[:space:]]Multi-modal[[:space:]]Models[[:space:]]with[[:space:]]Local[[:space:]]Interleaved[[:space:]]Region[[:space:]]Assistance/80bc7142-58bd-411b-b390-3d54c5062f10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LIRA_[[:space:]]Reasoning[[:space:]]Reconstruction[[:space:]]via[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/ace4109c-425d-4425-b491-cccdb9481c92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLM[[:space:]]Thought[[:space:]]Divergence[[:space:]]and[[:space:]]Convergence[[:space:]]for[[:space:]]Dialogue-Based[[:space:]]Image[[:space:]]Generation[[:space:]]Control/4d208fe5-788e-4486-9c84-6566510d9d67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLM-Assisted[[:space:]]Semantic[[:space:]]Guidance[[:space:]]for[[:space:]]Sparsely[[:space:]]Annotated[[:space:]]Remote[[:space:]]Sensing[[:space:]]Object[[:space:]]Detection/58c323d1-580b-45ad-8f83-a442088e9949_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLM-assisted[[:space:]]Entropy-based[[:space:]]Adaptive[[:space:]]Distillation[[:space:]]for[[:space:]]Unsupervised[[:space:]]Fine-grained[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/42516c70-0f78-4fd4-87ad-b1b43d165bcb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLM-enhanced[[:space:]]Action-aware[[:space:]]Multi-modal[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Image-Text[[:space:]]Matching/c607cce6-6fa4-488f-a685-51c8ed519f9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLaFEA_[[:space:]]Frame-Event[[:space:]]Complementary[[:space:]]Fusion[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Spatiotemporal[[:space:]]Understanding[[:space:]]in[[:space:]]LMMs/d86966ae-3ab1-4902-9a62-ab205f6a73c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLaVA-3D_[[:space:]]A[[:space:]]Simple[[:space:]]yet[[:space:]]Effective[[:space:]]Pathway[[:space:]]to[[:space:]]Empowering[[:space:]]LMMs[[:space:]]with[[:space:]]3D[[:space:]]Capabilities/15470b0e-9711-496d-b72d-6a5ad7f54558_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLaVA-CoT_[[:space:]]Let[[:space:]]Vision[[:space:]]Language[[:space:]]Models[[:space:]]Reason[[:space:]]Step-by-Step/4db946f4-abd4-4fb8-9346-e0094fcaa93a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLaVA-KD_[[:space:]]A[[:space:]]Framework[[:space:]]of[[:space:]]Distilling[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/30473951-88d7-4992-9ec9-de3e72562a96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLaVA-PruMerge_[[:space:]]Adaptive[[:space:]]Token[[:space:]]Reduction[[:space:]]for[[:space:]]Efficient[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models/40e898a4-e40f-482b-91a3-81ffaf546cfe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LLaVA-SP_[[:space:]]Enhancing[[:space:]]Visual[[:space:]]Representation[[:space:]]with[[:space:]]Visual[[:space:]]Spatial[[:space:]]Tokens[[:space:]]for[[:space:]]MLLMs/359a7ff1-0a4b-4d5d-9d28-1b6e18c517cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LMM-Det_[[:space:]]Make[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models[[:space:]]Excel[[:space:]]in[[:space:]]Object[[:space:]]Detection/2f70d50b-4ef8-414a-84c6-a06aa7ec766d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LMM4LMM_[[:space:]]Benchmarking[[:space:]]and[[:space:]]Evaluating[[:space:]]Large-multimodal[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]LMMs/31d65c8b-3e7b-4424-a392-09b88df02a08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LOCATEdit_[[:space:]]Graph[[:space:]]Laplacian[[:space:]]Optimized[[:space:]]Cross[[:space:]]Attention[[:space:]]for[[:space:]]Localized[[:space:]]Text-Guided[[:space:]]Image[[:space:]]Editing/22d2894c-7434-4c09-992e-c68447654cff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LOMM_[[:space:]]Latest[[:space:]]Object[[:space:]]Memory[[:space:]]Management[[:space:]]for[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation/b3859b76-b90d-44c3-8d23-a63407f430b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LONG3R_[[:space:]]Long[[:space:]]Sequence[[:space:]]Streaming[[:space:]]3D[[:space:]]Reconstruction/794d0ef1-ec2a-4266-bc66-5e93c14737d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LOTA_[[:space:]]Bit-Planes[[:space:]]Guided[[:space:]]AI-Generated[[:space:]]Image[[:space:]]Detection/96d7ec2a-a235-401a-96fb-2a11d2bd32e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LOTS[[:space:]]of[[:space:]]Fashion![[:space:]]Multi-Conditioning[[:space:]]for[[:space:]]Image[[:space:]]Generation[[:space:]]via[[:space:]]Sketch-Text[[:space:]]Pairing/d4dc519f-d5d9-42ec-b3fe-6b29b8c743af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LUDVIG_[[:space:]]Learning-Free[[:space:]]Uplifting[[:space:]]of[[:space:]]2D[[:space:]]Visual[[:space:]]Features[[:space:]]to[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Scenes/4f7ece11-175b-4f16-abea-f13f4151de25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LUSD_[[:space:]]Localized[[:space:]]Update[[:space:]]Score[[:space:]]Distillation[[:space:]]for[[:space:]]Text-Guided[[:space:]]Image[[:space:]]Editing/34bc844a-2162-45e8-87dc-630540d31f09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LUT-Fuse_[[:space:]]Towards[[:space:]]Extremely[[:space:]]Fast[[:space:]]Infrared[[:space:]]and[[:space:]]Visible[[:space:]]Image[[:space:]]Fusion[[:space:]]via[[:space:]]Distillation[[:space:]]to[[:space:]]Learnable[[:space:]]Look-Up[[:space:]]Tables/ee9a5305-289d-472e-a618-99f8e9c3509a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LV-MAE_[[:space:]]Learning[[:space:]]Long[[:space:]]Video[[:space:]]Representations[[:space:]]through[[:space:]]Masked-Embedding[[:space:]]Autoencoders/d84dbc88-7dd7-4aa7-9665-f279572abc6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LVAgent_[[:space:]]Long[[:space:]]Video[[:space:]]Understanding[[:space:]]by[[:space:]]Multi-Round[[:space:]]Dynamical[[:space:]]Collaboration[[:space:]]of[[:space:]]MLLM[[:space:]]Agents/e9be5f77-d834-4871-8631-673cdd52d5ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LVBench_[[:space:]]An[[:space:]]Extreme[[:space:]]Long[[:space:]]Video[[:space:]]Understanding[[:space:]]Benchmark/49c64b43-b390-43f1-8825-d5aea85be994_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LVFace_[[:space:]]Progressive[[:space:]]Cluster[[:space:]]Optimization[[:space:]]for[[:space:]]Large[[:space:]]Vision[[:space:]]Models[[:space:]]in[[:space:]]Face[[:space:]]Recognition/6ab61968-3092-41c8-af46-e0cfaeb33bc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LaCoOT_[[:space:]]Layer[[:space:]]Collapse[[:space:]]through[[:space:]]Optimal[[:space:]]Transport/253581ce-3734-493a-a7b5-c5f624acb789_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LaRender_[[:space:]]Training-Free[[:space:]]Occlusion[[:space:]]Control[[:space:]]in[[:space:]]Image[[:space:]]Generation[[:space:]]via[[:space:]]Latent[[:space:]]Rendering/4f712f99-3635-443a-b0f1-2e74b32bdeeb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Laboring[[:space:]]on[[:space:]]less[[:space:]]labors_[[:space:]]RPCA[[:space:]]Paradigm[[:space:]]for[[:space:]]Pan-sharpening/da28cbd3-faca-4e11-bc04-f0d5b6674f4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LaneDiffusion_[[:space:]]Improving[[:space:]]Centerline[[:space:]]Graph[[:space:]]Learning[[:space:]]via[[:space:]]Prior[[:space:]]Injected[[:space:]]BEV[[:space:]]Feature[[:space:]]Generation/7cbb2e28-570e-41a9-b438-e0dcd5f7f762_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LangBridge_[[:space:]]Interpreting[[:space:]]Image[[:space:]]as[[:space:]]a[[:space:]]Combination[[:space:]]of[[:space:]]Language[[:space:]]Embeddings/d7f5c143-c799-4d04-9546-947ef8d14227_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LangScene-X_[[:space:]]Reconstruct[[:space:]]Generalizable[[:space:]]3D[[:space:]]Language-Embedded[[:space:]]Scenes[[:space:]]with[[:space:]]TriMap[[:space:]]Video[[:space:]]Diffusion/74352855-2b37-428c-8a80-632ea8b7f05b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Language[[:space:]]Decoupling[[:space:]]with[[:space:]]Fine-grained[[:space:]]Knowledge[[:space:]]Guidance[[:space:]]for[[:space:]]Referring[[:space:]]Multi-object[[:space:]]Tracking/7e0d8eb6-1ba8-4361-9879-4f24152100c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Language[[:space:]]Driven[[:space:]]Occupancy[[:space:]]Prediction/36dcfd9b-72eb-42fc-9432-7712302834e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Language-Driven[[:space:]]Multi-Label[[:space:]]Zero-Shot[[:space:]]Learning[[:space:]]with[[:space:]]Semantic[[:space:]]Granularity/f86f9f11-ea00-4b9e-bc00-d12fd2c052e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Large[[:space:]]Learning[[:space:]]Rates[[:space:]]Simultaneously[[:space:]]Achieve[[:space:]]Robustness[[:space:]]to[[:space:]]Spurious[[:space:]]Correlations[[:space:]]and[[:space:]]Compressibility/1d10e878-707f-4d7e-b1e5-2b4f55285661_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Large[[:space:]]Multi-modal[[:space:]]Models[[:space:]]Can[[:space:]]Interpret[[:space:]]Features[[:space:]]in[[:space:]]Large[[:space:]]Multi-modal[[:space:]]Models/ffc91543-0668-421b-ae87-8aca4593f6bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Large[[:space:]]Scene[[:space:]]Generation[[:space:]]with[[:space:]]Cube-Absorb[[:space:]]Discrete[[:space:]]Diffusion/42f51383-c577-42f2-b1de-7fb7b50e0621_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Large-scale[[:space:]]Pre-training[[:space:]]for[[:space:]]Grounded[[:space:]]Video[[:space:]]Caption[[:space:]]Generation/a22fa744-daa1-4bc5-9675-77c47060f431_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lark_[[:space:]]Low-Rank[[:space:]]Updates[[:space:]]After[[:space:]]Knowledge[[:space:]]Localization[[:space:]]for[[:space:]]Few-shot[[:space:]]Class-Incremental[[:space:]]Learning/1dd3c934-78f0-45d3-8d4a-2c24d6773a2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Latent[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Masked[[:space:]]AutoEncoders/38008586-d85b-4f64-8c43-74dd0860eaf3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Latent[[:space:]]Expression[[:space:]]Generation[[:space:]]for[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation[[:space:]]and[[:space:]]Grounding/0a400103-1872-44f2-9fd8-a65d83effa28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Latent[[:space:]]Swap[[:space:]]Joint[[:space:]]Diffusion[[:space:]]for[[:space:]]2D[[:space:]]Long-Form[[:space:]]Latent[[:space:]]Generation/6dadc9ec-9541-483a-a750-72e294d00c17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Latent-Reframe_[[:space:]]Enabling[[:space:]]Camera[[:space:]]Control[[:space:]]for[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models[[:space:]]without[[:space:]]Training/71cc9d4f-a575-4b55-8be4-061e70a71468_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Latte_[[:space:]]Collaborative[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models[[:space:]]in[[:space:]]Federated[[:space:]]Learning/ec8ef2d0-fe1e-418b-8965-4382a68a82a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LawDIS_[[:space:]]Language-Window-based[[:space:]]Controllable[[:space:]]Dichotomous[[:space:]]Image[[:space:]]Segmentation/bd582fff-c54d-4c1c-a050-f48866d185ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lay-Your-Scene_[[:space:]]Natural[[:space:]]Scene[[:space:]]Layout[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion[[:space:]]Transformers/afbb4a23-37d5-4e58-bdc0-be2ea046da4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lay2Story_[[:space:]]Extending[[:space:]]Diffusion[[:space:]]Transformers[[:space:]]for[[:space:]]Layout-Togglable[[:space:]]Story[[:space:]]Generation/009e958c-2378-4140-8a5f-b9f6ea9b47fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Layer-wise[[:space:]]Vision[[:space:]]Injection[[:space:]]with[[:space:]]Disentangled[[:space:]]Attention[[:space:]]for[[:space:]]Efficient[[:space:]]LVLMs/085214fc-2b87-4290-ba74-0542405b3c7a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LayerAnimate_[[:space:]]Layer-level[[:space:]]Control[[:space:]]for[[:space:]]Animation/31c501d6-67a0-4304-855f-36ed6dff4f4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LayerD_[[:space:]]Decomposing[[:space:]]Raster[[:space:]]Graphic[[:space:]]Designs[[:space:]]into[[:space:]]Layers/a0c7d108-ed3b-4802-bfd3-55c02a210890_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LayerLock_[[:space:]]Non-collapsing[[:space:]]Representation[[:space:]]Learning[[:space:]]with[[:space:]]Progressive[[:space:]]Freezing/fb95281f-f579-4953-a4b9-fc48ef2e2f82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LayerTracer_[[:space:]]Cognitive-Aligned[[:space:]]Layered[[:space:]]SVG[[:space:]]Synthesis[[:space:]]via[[:space:]]Diffusion[[:space:]]Transformer/06f37d5c-a93a-46bb-a897-bbd7428a6720_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LazyMAR_[[:space:]]Accelerating[[:space:]]Masked[[:space:]]Autoregressive[[:space:]]Models[[:space:]]via[[:space:]]Feature[[:space:]]Caching/18d2ca65-6b12-45c5-90d2-9fff3185bb14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LeGrad_[[:space:]]An[[:space:]]Explainability[[:space:]]Method[[:space:]]for[[:space:]]Vision[[:space:]]Transformers[[:space:]]via[[:space:]]Feature[[:space:]]Formation[[:space:]]Sensitivity/103dfa04-5bee-4785-aac6-d3991d19592f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LeanVAE_[[:space:]]An[[:space:]]Ultra-Efficient[[:space:]]Reconstruction[[:space:]]VAE[[:space:]]for[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/a61c3705-ae2a-4917-8df0-edfa13f143db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leaps[[:space:]]and[[:space:]]Bounds_[[:space:]]An[[:space:]]Improved[[:space:]]Point[[:space:]]Cloud[[:space:]]Winding[[:space:]]Number[[:space:]]Formulation[[:space:]]for[[:space:]]Fast[[:space:]]Normal[[:space:]]Estimation[[:space:]]and[[:space:]]Surface[[:space:]]Reconstruction/6927e9fd-3dba-4916-be90-4e1be50f8dae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learn2Synth_[[:space:]]Learning[[:space:]]Optimal[[:space:]]Data[[:space:]]Synthesis[[:space:]]Using[[:space:]]Hypergradients[[:space:]]for[[:space:]]Brain[[:space:]]Image[[:space:]]Segmentation/89c56d8c-15f6-4d6e-b11f-d8cde150b0a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learnable[[:space:]]Feature[[:space:]]Patches[[:space:]]and[[:space:]]Vectors[[:space:]]for[[:space:]]Boosting[[:space:]]Low-light[[:space:]]Image[[:space:]]Enhancement[[:space:]]without[[:space:]]External[[:space:]]Knowledge/db7ae19d-0f8b-48cb-953b-28ade69f5f44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learnable[[:space:]]Fractional[[:space:]]Reaction-Diffusion[[:space:]]Dynamics[[:space:]]for[[:space:]]Under-Display[[:space:]]ToF[[:space:]]Imaging[[:space:]]and[[:space:]]Beyond/8a9915c4-dba1-4fba-9d76-7b63f4d3f2f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learnable[[:space:]]Logit[[:space:]]Adjustment[[:space:]]for[[:space:]]Imbalanced[[:space:]]Semi-Supervised[[:space:]]Learning[[:space:]]under[[:space:]]Class[[:space:]]Distribution[[:space:]]Mismatch/adbbcd30-0e7e-47ee-97f1-452349329375_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learnable[[:space:]]Retrieval[[:space:]]Enhanced[[:space:]]Visual-Text[[:space:]]Alignment[[:space:]]and[[:space:]]Fusion[[:space:]]for[[:space:]]Radiology[[:space:]]Report[[:space:]]Generation/cfde0eef-6efd-4a1d-bd3c-5060785de305_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learned[[:space:]]Image[[:space:]]Compression[[:space:]]with[[:space:]]Hierarchical[[:space:]]Progressive[[:space:]]Context[[:space:]]Modeling/8f1da71f-dcee-4b84-9ac8-df57e3d7bbc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]3D[[:space:]]Object[[:space:]]Spatial[[:space:]]Relationships[[:space:]]from[[:space:]]Pre-trained[[:space:]]2D[[:space:]]Diffusion[[:space:]]Models/8588469b-025d-452a-abe4-fcf206f35f5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]3D[[:space:]]Scene[[:space:]]Analogies[[:space:]]with[[:space:]]Neural[[:space:]]Contextual[[:space:]]Scene[[:space:]]Maps/04ac297c-fdda-4bec-9503-90ecae6d2211_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]4D[[:space:]]Embodied[[:space:]]World[[:space:]]Models/6dc664ca-5ddf-47d2-a0b5-c59242b648e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]A[[:space:]]Unified[[:space:]]Template[[:space:]]for[[:space:]]Gait[[:space:]]Recognition/2f09a682-19a1-48b4-9402-d474aea84978_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Beyond[[:space:]]Still[[:space:]]Frames_[[:space:]]Scaling[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]Video/b9ab9888-ef9d-4d45-891c-36167c749cbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Counterfactually[[:space:]]Decoupled[[:space:]]Attention[[:space:]]for[[:space:]]Open-World[[:space:]]Model[[:space:]]Attribution/94ba5f04-a652-4fae-b282-d4cfa97db2a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Deblurring[[:space:]]Texture[[:space:]]Prior[[:space:]]from[[:space:]]Unpaired[[:space:]]Data[[:space:]]with[[:space:]]Diffusion[[:space:]]Model/754ad6a0-1617-44ba-b321-a25998e65ff6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Dense[[:space:]]Feature[[:space:]]Matching[[:space:]]via[[:space:]]Lifting[[:space:]]Single[[:space:]]2D[[:space:]]Image[[:space:]]to[[:space:]]3D[[:space:]]Space/9d82a358-1f2e-472c-8785-a3140cea69d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Efficient[[:space:]]and[[:space:]]Generalizable[[:space:]]Human[[:space:]]Representation[[:space:]]with[[:space:]]Human[[:space:]]Gaussian[[:space:]]Model/65b489b7-b57f-46e0-a4e2-da0ebcbf0882_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Few-Step[[:space:]]Diffusion[[:space:]]Models[[:space:]]by[[:space:]]Trajectory[[:space:]]Distribution[[:space:]]Matching/fb908000-38d0-4edc-9cb1-faab8ac746c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Hierarchical[[:space:]]Line[[:space:]]Buffer[[:space:]]for[[:space:]]Image[[:space:]]Processing/eba40168-823d-4461-ad01-07cc21f011cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Implicit[[:space:]]Features[[:space:]]with[[:space:]]Flow-Infused[[:space:]]Transformations[[:space:]]for[[:space:]]Realistic[[:space:]]Virtual[[:space:]]Try-On/8404fe48-4da8-4e2f-ab24-9cdcac851677_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Interpretable[[:space:]]Queries[[:space:]]for[[:space:]]Explainable[[:space:]]Image[[:space:]]Classification[[:space:]]with[[:space:]]Information[[:space:]]Pursuit/58259bd2-a30e-4eef-b22a-920a72fd00be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Large[[:space:]]Motion[[:space:]]Estimation[[:space:]]from[[:space:]]Intermediate[[:space:]]Representations[[:space:]]with[[:space:]]a[[:space:]]High-Resolution[[:space:]]Optical[[:space:]]Flow[[:space:]]Dataset[[:space:]]Featuring[[:space:]]Long-Range[[:space:]]Dynamic[[:space:]]Motion/a4edbd8f-c4e8-471f-88c9-4151c231b655_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Neural[[:space:]]Scene[[:space:]]Representation[[:space:]]from[[:space:]]iToF[[:space:]]Imaging/2efffc91-c7b7-45e7-87a5-d423a1df83a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Normal[[:space:]]Flow[[:space:]]Directly[[:space:]]From[[:space:]]Events/962e4da1-7a22-48dc-af48-6cda9a505fd9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Normals[[:space:]]of[[:space:]]Noisy[[:space:]]Points[[:space:]]by[[:space:]]Local[[:space:]]Gradient-Aware[[:space:]]Surface[[:space:]]Filtering/78a7da30-ae1b-49d9-903f-9ac1b52f4142_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Null[[:space:]]Geodesics[[:space:]]for[[:space:]]Gravitational[[:space:]]Lensing[[:space:]]Rendering[[:space:]]in[[:space:]]General[[:space:]]Relativity/2f4a1f8e-f812-4ed3-9e72-8aa8ca250b58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Pixel-adaptive[[:space:]]Multi-layer[[:space:]]Perceptrons[[:space:]]for[[:space:]]Real-time[[:space:]]Image[[:space:]]Enhancement/485d19b3-aa45-4103-b03d-84d253542d29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Precise[[:space:]]Affordances[[:space:]]from[[:space:]]Egocentric[[:space:]]Videos[[:space:]]for[[:space:]]Robotic[[:space:]]Manipulation/9b9b5d2f-6780-4b45-8e65-18cec66464eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Robust[[:space:]]Image[[:space:]]Watermarking[[:space:]]with[[:space:]]Lossless[[:space:]]Cover[[:space:]]Recovery/4127f17c-d1f7-4d0e-b3ca-6907ac310d1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Robust[[:space:]]Stereo[[:space:]]Matching[[:space:]]in[[:space:]]the[[:space:]]Wild[[:space:]]with[[:space:]]Selective[[:space:]]Mixture-of-Experts/11f1f69f-0ea0-429d-8d22-e829df7f7ff6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Separable[[:space:]]Fine-Grained[[:space:]]Representation[[:space:]]via[[:space:]]Dendrogram[[:space:]]Construction[[:space:]]from[[:space:]]Coarse[[:space:]]Labels[[:space:]]for[[:space:]]Fine-grained[[:space:]]Visual[[:space:]]Recognition/273af488-29d4-4aa8-94ac-7bb16d705a85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Streaming[[:space:]]Video[[:space:]]Representation[[:space:]]via[[:space:]]Multitask[[:space:]]Training/743906b9-cef9-488b-b116-b941c871742d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Visual[[:space:]]Hierarchies[[:space:]]in[[:space:]]Hyperbolic[[:space:]]Space[[:space:]]for[[:space:]]Image[[:space:]]Retrieval/d06973b2-7609-4f40-8e86-b7f97303be9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Visual[[:space:]]Proxy[[:space:]]for[[:space:]]Compositional[[:space:]]Zero-Shot[[:space:]]Learning/3eb1e1a8-85b9-47bd-8bfd-8af8555f6d47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]Yourself_[[:space:]]Class-Incremental[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Language-Inspired[[:space:]]Bootstrapped[[:space:]]Disentanglement/f6aa1420-dda0-43cb-99fe-038b53235e0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]an[[:space:]]Implicit[[:space:]]Physics[[:space:]]Model[[:space:]]for[[:space:]]Image-based[[:space:]]Fluid[[:space:]]Simulation/4da0074e-4f29-4322-8486-a4ccf078909e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]on[[:space:]]the[[:space:]]Go_[[:space:]]A[[:space:]]Meta-learning[[:space:]]Object[[:space:]]Navigation[[:space:]]Model/66a27178-b6ca-45e6-a633-c3dcf0b152c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]to[[:space:]]Generalize[[:space:]]without[[:space:]]Bias[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Action[[:space:]]Recognition/57792847-619b-4204-90ec-b6a2b9c4ba48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]to[[:space:]]Inference[[:space:]]Adaptively[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/48080f0e-cb48-44e9-89e1-99fa79a34aee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]to[[:space:]]See[[:space:]]Inside[[:space:]]Opaque[[:space:]]Liquid[[:space:]]Containers[[:space:]]using[[:space:]]Speckle[[:space:]]Vibrometry/f8855503-5c83-477f-86e9-1597fdb1ab54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]to[[:space:]]See[[:space:]]in[[:space:]]the[[:space:]]Extremely[[:space:]]Dark/e10f370e-a462-40b6-a7fd-a7f044f8d4d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Learning[[:space:]]to[[:space:]]Unlearn[[:space:]]while[[:space:]]Retaining_[[:space:]]Combating[[:space:]]Gradient[[:space:]]Conflicts[[:space:]]in[[:space:]]Machine[[:space:]]Unlearning/fcbf234f-543b-4dcd-be14-c0215f5ea471_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Less[[:space:]]Static,[[:space:]]More[[:space:]]Private_[[:space:]]Towards[[:space:]]Transferable[[:space:]]Privacy-Preserving[[:space:]]Action[[:space:]]Recognition[[:space:]]by[[:space:]]Generative[[:space:]]Decoupled[[:space:]]Learning/7ce0b40f-bcb9-4354-823e-baf380670b70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Less[[:space:]]is[[:space:]]More_[[:space:]]Empowering[[:space:]]GUI[[:space:]]Agent[[:space:]]with[[:space:]]Context-Aware[[:space:]]Simplification/139da5f2-e14c-42ff-b8ae-44d7d05ed4aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Less[[:space:]]is[[:space:]]More_[[:space:]]Improving[[:space:]]Motion[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Sparse[[:space:]]Keyframes/f154c857-669a-4682-8d7f-18abfd82749c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Less-to-More[[:space:]]Generalization_[[:space:]]Unlocking[[:space:]]More[[:space:]]Controllability[[:space:]]by[[:space:]]In-Context[[:space:]]Generation/c5b34fa8-6a00-40b8-87cc-f0731ca4cee6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]2D[[:space:]]Priors[[:space:]]and[[:space:]]SDF[[:space:]]Guidance[[:space:]]for[[:space:]]Urban[[:space:]]Scene[[:space:]]Rendering/d0288586-1d31-47d7-b530-3af5022aacc7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]BEV[[:space:]]Paradigm[[:space:]]for[[:space:]]Ground-to-Aerial[[:space:]]Image[[:space:]]Synthesis/5604a59b-5e4a-467f-b954-66e139cf8997_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]Debiased[[:space:]]Cross-modal[[:space:]]Attention[[:space:]]Maps[[:space:]]and[[:space:]]Code-based[[:space:]]Reasoning[[:space:]]for[[:space:]]Zero-shot[[:space:]]Referring[[:space:]]Expression[[:space:]]Comprehension/427e8f9b-0947-40ef-9ab1-8454ebe141a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]Local[[:space:]]Patch[[:space:]]Alignment[[:space:]]to[[:space:]]Seam-cutting[[:space:]]for[[:space:]]Large[[:space:]]Parallax[[:space:]]Image[[:space:]]Stitching/d2872463-c567-4120-99df-43165bcf42c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]Panoptic[[:space:]]Scene[[:space:]]Graph[[:space:]]for[[:space:]]Evaluating[[:space:]]Fine-Grained[[:space:]]Text-to-Image[[:space:]]Generation/d7658a10-858d-4cba-83b6-bc2dcc948d51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]Prior[[:space:]]Knowledge[[:space:]]of[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Person[[:space:]]Search/92d076b1-9a31-42fa-8620-ed34945db318_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]Spatial[[:space:]]Invariance[[:space:]]to[[:space:]]Boost[[:space:]]Adversarial[[:space:]]Transferability/0a74a1d9-d13b-4b59-b441-4f08adc6f84b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Leveraging[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]MLLMs[[:space:]]for[[:space:]]Gloss-Free[[:space:]]Sign[[:space:]]Language[[:space:]]Translation/6d1e2043-c879-4ef6-ab52-8f5bf311df63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LiON-LoRA_[[:space:]]Rethinking[[:space:]]LoRA[[:space:]]Fusion[[:space:]]to[[:space:]]Unify[[:space:]]Controllable[[:space:]]Spatial[[:space:]]and[[:space:]]Temporal[[:space:]]Generation[[:space:]]for[[:space:]]Video[[:space:]]Diffusion/ab578a7e-78c3-4dc1-bf05-733ed52ed0b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LiT_[[:space:]]Delving[[:space:]]into[[:space:]]a[[:space:]]Simple[[:space:]]Linear[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]for[[:space:]]Image[[:space:]]Generation/35ac3e04-1a30-41f8-9223-f922d0b30d7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Liberated-GS_[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Independent[[:space:]]from[[:space:]]SfM[[:space:]]Point[[:space:]]Clouds/e3c6c33b-85e8-45d5-a89a-fb0111f40e32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lidar[[:space:]]Waveforms[[:space:]]are[[:space:]]Worth[[:space:]]40x128x33[[:space:]]Words/0c71f570-d577-4cba-a328-9ffd839c8e21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lifting[[:space:]]the[[:space:]]Structural[[:space:]]Morphing[[:space:]]for[[:space:]]Wide-Angle[[:space:]]Images[[:space:]]Rectification_[[:space:]]Unified[[:space:]]Content[[:space:]]and[[:space:]]Boundary[[:space:]]Modeling/8c5d0845-b7b0-4f67-9bfe-86a24cd7b891_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Light-A-Video_[[:space:]]Training-free[[:space:]]Video[[:space:]]Relighting[[:space:]]via[[:space:]]Progressive[[:space:]]Light[[:space:]]Fusion/aead1fdc-c7e6-4a39-ae77-35fc1064514a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LightBSR_[[:space:]]Towards[[:space:]]Lightweight[[:space:]]Blind[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Discriminative[[:space:]]Implicit[[:space:]]Degradation[[:space:]]Representation[[:space:]]Learning/c9f27c48-91c2-4e79-a699-e3d96607cd8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LightCity_[[:space:]]An[[:space:]]Urban[[:space:]]Dataset[[:space:]]for[[:space:]]Outdoor[[:space:]]Inverse[[:space:]]Rendering[[:space:]]and[[:space:]]Reconstruction[[:space:]]under[[:space:]]Multi-illumination[[:space:]]Conditions/31dfc49d-3dd6-4558-9633-d7629233a757_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LightSwitch_[[:space:]]Multi-view[[:space:]]Relighting[[:space:]]with[[:space:]]Material-guided[[:space:]]Diffusion/0082079a-dc2a-4503-9438-da6b864f0db7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LightsOut_[[:space:]]Diffusion-based[[:space:]]Outpainting[[:space:]]for[[:space:]]Enhanced[[:space:]]Lens[[:space:]]Flare[[:space:]]Removal/40912b01-3e5f-4c4f-8aa6-de0707583805_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lightweight[[:space:]]Gradient-Aware[[:space:]]Upscaling[[:space:]]of[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Images/3ff12653-d198-421f-8aa4-4c763f4ed94b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lightweight[[:space:]]and[[:space:]]Fast[[:space:]]Real-time[[:space:]]Image[[:space:]]Enhancement[[:space:]]via[[:space:]]Decomposition[[:space:]]of[[:space:]]the[[:space:]]Spatial-aware[[:space:]]Lookup[[:space:]]Tables/0179e0a4-2b48-4269-ba7b-5ef6a58ec892_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LoD-Loc[[:space:]]v2_[[:space:]]Aerial[[:space:]]Visual[[:space:]]Localization[[:space:]]over[[:space:]]Low[[:space:]]Level-of-Detail[[:space:]]City[[:space:]]Models[[:space:]]using[[:space:]]Explicit[[:space:]]Silhouette[[:space:]]Alignment/d11360f6-eff6-46a3-980e-439167433ed2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LoRA-FAIR_[[:space:]]Federated[[:space:]]LoRA[[:space:]]Fine-Tuning[[:space:]]with[[:space:]]Aggregation[[:space:]]and[[:space:]]Initialization[[:space:]]Refinement/7a1f84cc-eddc-4946-a91c-881f764d86a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LoRA.rar_[[:space:]]Learning[[:space:]]to[[:space:]]Merge[[:space:]]LoRAs[[:space:]]via[[:space:]]Hypernetworks[[:space:]]for[[:space:]]Subject-Style[[:space:]]Conditioned[[:space:]]Image[[:space:]]Generation/fe40f4b1-82de-4562-97c1-4e841a4787b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LoRAverse_[[:space:]]A[[:space:]]Submodular[[:space:]]Framework[[:space:]]to[[:space:]]Retrieve[[:space:]]Diverse[[:space:]]Adapters[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/046c877e-4ccd-4488-9c08-6b10094cbe36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Local[[:space:]]Dense[[:space:]]Logit[[:space:]]Relations[[:space:]]for[[:space:]]Enhanced[[:space:]]Knowledge[[:space:]]Distillation/5ff3245f-a8b7-4bef-be54-a5c1dfccbc2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Local[[:space:]]Scale[[:space:]]Equivariance[[:space:]]with[[:space:]]Latent[[:space:]]Deep[[:space:]]Equilibrium[[:space:]]Canonicalizer/cd70f4c8-33f0-4d32-b4f0-f15d53afb28d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LocalDyGS_[[:space:]]Multi-view[[:space:]]Global[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Modeling[[:space:]]via[[:space:]]Adaptive[[:space:]]Local[[:space:]]Implicit[[:space:]]Feature[[:space:]]Decoupling/69820246-22fc-4b90-94e3-bfa5d0067118_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LoftUp_[[:space:]]Learning[[:space:]]a[[:space:]]Coordinate-Based[[:space:]]Feature[[:space:]]Upsampler[[:space:]]for[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models/2ce2806d-49e6-45f6-9c48-98f9cbb7dbb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Long[[:space:]]Context[[:space:]]Tuning[[:space:]]for[[:space:]]Video[[:space:]]Generation/f3ed7c2d-0f92-474b-932a-5ef6a21c31dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Long-Context[[:space:]]State-Space[[:space:]]Video[[:space:]]World[[:space:]]Models/f97137da-85ac-4dc7-be14-a5cfa43a84f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Long-LRM_[[:space:]]Long-sequence[[:space:]]Large[[:space:]]Reconstruction[[:space:]]Model[[:space:]]for[[:space:]]Wide-coverage[[:space:]]Gaussian[[:space:]]Splats/36b371e0-71e4-4305-99c0-9fc9e822321c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Long-Tailed[[:space:]]Classification[[:space:]]with[[:space:]]Multi-Granularity[[:space:]]Semantics/a9d2343e-b696-43bd-a2d2-71a267c8af01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Long-term[[:space:]]Traffic[[:space:]]Simulation[[:space:]]with[[:space:]]Interleaved[[:space:]]Autoregressive[[:space:]]Motion[[:space:]]and[[:space:]]Scenario[[:space:]]Generation/451c9a29-c451-4bbc-afff-c0cc70c6417c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LongAnimation_[[:space:]]Long[[:space:]]Animation[[:space:]]Generation[[:space:]]with[[:space:]]Dynamic[[:space:]]Global-Local[[:space:]]Memory/5cb0fd50-6a63-417d-b8d4-6e08388c9a15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LongSplat_[[:space:]]Robust[[:space:]]Unposed[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Casual[[:space:]]Long[[:space:]]Videos/5d687467-ee98-44f5-8437-ad7647dafd02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/LookOut_[[:space:]]Real-World[[:space:]]Humanoid[[:space:]]Egocentric[[:space:]]Navigation/f6c60a16-8f79-42ee-890a-60d18f7cba97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Looking[[:space:]]in[[:space:]]the[[:space:]]Mirror_[[:space:]]A[[:space:]]Faithful[[:space:]]Counterfactual[[:space:]]Explanation[[:space:]]Method[[:space:]]for[[:space:]]Interpreting[[:space:]]Deep[[:space:]]Image[[:space:]]Classification[[:space:]]Models/5655f403-23a8-4a2a-b5a2-e5c4caac19c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Loss[[:space:]]Functions[[:space:]]for[[:space:]]Predictor-based[[:space:]]Neural[[:space:]]Architecture[[:space:]]Search/00d00212-cd97-4803-9d33-3720fb5401ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Low-Light[[:space:]]Image[[:space:]]Enhancement[[:space:]]Using[[:space:]]Event-Based[[:space:]]Illumination[[:space:]]Estimation/95deb57f-4d99-4fec-a713-faa206cb7dd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lumina-Image[[:space:]]2.0_[[:space:]]A[[:space:]]Unified[[:space:]]and[[:space:]]Efficient[[:space:]]Image[[:space:]]Generative[[:space:]]Framework/220cd226-ffd4-42a3-9dff-3e322b2d5dad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Lyra_[[:space:]]An[[:space:]]Efficient[[:space:]]and[[:space:]]Speech-Centric[[:space:]]Framework[[:space:]]for[[:space:]]Omni-Cognition/5ecf84d3-d7c0-4a70-9d68-a44372d552ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/M-Net_[[:space:]]MRI[[:space:]]Brain[[:space:]]Tumor[[:space:]]Sequential[[:space:]]Segmentation[[:space:]]Network[[:space:]]via[[:space:]]Mesh-Cast/09de48b6-62a4-4473-a625-eaa3b4da3a9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/M-SpecGene_[[:space:]]Generalized[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]RGBT[[:space:]]Multispectral[[:space:]]Vision/0c7d4d5b-b486-4df9-9dfc-c04ed0d757d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/M2EIT_[[:space:]]Multi-Domain[[:space:]]Mixture[[:space:]]of[[:space:]]Experts[[:space:]]for[[:space:]]Robust[[:space:]]Neural[[:space:]]Inertial[[:space:]]Tracking/879bdd83-9db6-4ac1-be5f-33d0b4960073_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/M2SFormer_[[:space:]]Multi-Spectral[[:space:]]and[[:space:]]Multi-Scale[[:space:]]Attention[[:space:]]with[[:space:]]Edge-Aware[[:space:]]Difficulty[[:space:]]Guidance[[:space:]]for[[:space:]]Image[[:space:]]Forgery[[:space:]]Localization/ec2f5ef3-38a9-448d-a3ad-e3c8bd38a32b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MA-CIR_[[:space:]]A[[:space:]]Multimodal[[:space:]]Arithmetic[[:space:]]Benchmark[[:space:]]for[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/c7750d88-6e1f-4b96-9338-e19717fd2fbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MAESTRO_[[:space:]]Task-Relevant[[:space:]]Optimization[[:space:]]via[[:space:]]Adaptive[[:space:]]Feature[[:space:]]Enhancement[[:space:]]and[[:space:]]Suppression[[:space:]]for[[:space:]]Multi-task[[:space:]]3D[[:space:]]Perception/e423cd2c-6851-4524-a5eb-2082c03b13bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MATE_[[:space:]]Motion-Augmented[[:space:]]Temporal[[:space:]]Consistency[[:space:]]for[[:space:]]Event-based[[:space:]]Point[[:space:]]Tracking/a8e67856-0ba1-4e86-88ab-3528929593ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MAVFlow_[[:space:]]Preserving[[:space:]]Paralinguistic[[:space:]]Elements[[:space:]]with[[:space:]]Conditional[[:space:]]Flow[[:space:]]Matching[[:space:]]for[[:space:]]Zero-Shot[[:space:]]AV2AV[[:space:]]Multilingual[[:space:]]Translation/8cd0a0f4-c741-49a9-859c-efe1a39f1800_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MAVias_[[:space:]]Mitigate[[:space:]]any[[:space:]]Visual[[:space:]]Bias/6b15d040-86ca-4d2e-b390-5b207bc8dedd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MBTI_[[:space:]]Masked[[:space:]]Blending[[:space:]]Transformers[[:space:]]with[[:space:]]Implicit[[:space:]]Positional[[:space:]]Encoding[[:space:]]for[[:space:]]Frame-rate[[:space:]]Agnostic[[:space:]]Motion[[:space:]]Estimation/afdfb259-14dd-4848-8793-ea99a9704555_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MC-Bench_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Multi-Context[[:space:]]Visual[[:space:]]Grounding[[:space:]]in[[:space:]]the[[:space:]]Era[[:space:]]of[[:space:]]MLLMs/6f94713d-b538-4dae-a2cb-e7da91cfb8b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MCAM_[[:space:]]Multimodal[[:space:]]Causal[[:space:]]Analysis[[:space:]]Model[[:space:]]for[[:space:]]Ego-Vehicle-Level[[:space:]]Driving[[:space:]]Video[[:space:]]Understanding/a53040cc-9d44-4bef-a743-2617a3cacaa1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MCID_[[:space:]]Multi-aspect[[:space:]]Copyright[[:space:]]Infringement[[:space:]]Detection[[:space:]]for[[:space:]]Generated[[:space:]]Images/05d1bec8-906a-40d2-a93a-e4a58ff1240b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MCOP_[[:space:]]Multi-UAV[[:space:]]Collaborative[[:space:]]Occupancy[[:space:]]Prediction/2789da1d-1696-4902-82b3-76223f5c1139_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MDD_[[:space:]]A[[:space:]]Dataset[[:space:]]for[[:space:]]Text-and-Music[[:space:]]Conditioned[[:space:]]Duet[[:space:]]Dance[[:space:]]Generation/b143d93c-8c5e-4f89-b336-d2216cf62017_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MDP-Omni_[[:space:]]Parameter-free[[:space:]]Multimodal[[:space:]]Depth[[:space:]]Prior-based[[:space:]]Sampling[[:space:]]for[[:space:]]Omnidirectional[[:space:]]Stereo[[:space:]]Matching/74821c30-efce-435b-b247-a3ac482714be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MDP3_[[:space:]]A[[:space:]]Training-free[[:space:]]Approach[[:space:]]for[[:space:]]List-wise[[:space:]]Frame[[:space:]]Selection[[:space:]]in[[:space:]]Video-LLMs/b90fc24c-25f8-4660-a32a-39c74a3c43e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MEGA_[[:space:]]Memory-Efficient[[:space:]]4D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Dynamic[[:space:]]Scenes/c18cccbd-63be-4b13-8148-5f0ea29f4037_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MEH_[[:space:]]A[[:space:]]Multi-Style[[:space:]]Dataset[[:space:]]and[[:space:]]Toolkit[[:space:]]for[[:space:]]Advancing[[:space:]]Egyptian[[:space:]]Hieroglyph[[:space:]]Recognition/1fe05026-c49a-414f-8eb1-24d0526bb7d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MEMFOF_[[:space:]]High-Resolution[[:space:]]Training[[:space:]]for[[:space:]]Memory-Efficient[[:space:]]Multi-Frame[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation/5d520bd4-f777-4595-9cf4-6228c818c2d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/METEOR_[[:space:]]Multi-Encoder[[:space:]]Collaborative[[:space:]]Token[[:space:]]Pruning[[:space:]]for[[:space:]]Efficient[[:space:]]Vision[[:space:]]Language[[:space:]]Models/e218a5c2-17cc-4e00-80d6-8211c71df911_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MGSR_[[:space:]]2D_3D[[:space:]]Mutual-boosted[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]High-fidelity[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]under[[:space:]]Various[[:space:]]Light[[:space:]]Conditions/d12593eb-e14f-4c4d-aa57-40a16c4700fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MGSfM_[[:space:]]Multi-Camera[[:space:]]Geometry[[:space:]]Driven[[:space:]]Global[[:space:]]Structure-from-Motion/9ced54b7-eba2-49f2-bc7e-a0e65dde3e9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MH-LVC_[[:space:]]Multi-Hypothesis[[:space:]]Temporal[[:space:]]Prediction[[:space:]]for[[:space:]]Learned[[:space:]]Conditional[[:space:]]Residual[[:space:]]Video[[:space:]]Coding/f10df06d-d622-4ca6-9481-f5c89e167078_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MIEB_[[:space:]]Massive[[:space:]]Image[[:space:]]Embedding[[:space:]]Benchmark/2a6ac1e3-62a1-4936-ba29-dc517d816638_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MINERVA_[[:space:]]Evaluating[[:space:]]Complex[[:space:]]Video[[:space:]]Reasoning/56522b85-45a6-4d9e-84aa-c37ba2ddb7a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MIORe[[:space:]]&[[:space:]]VAR-MIORe_[[:space:]]Benchmarks[[:space:]]to[[:space:]]Push[[:space:]]the[[:space:]]Boundaries[[:space:]]of[[:space:]]Restoration/7d5e8ed0-864f-45e0-80a9-407a212e69ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MM-IFEngine_[[:space:]]Towards[[:space:]]Multimodal[[:space:]]Instruction[[:space:]]Following/d9f44f14-7429-4382-a042-d8f4db82266a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MM-Spatial_[[:space:]]Exploring[[:space:]]3D[[:space:]]Spatial[[:space:]]Understanding[[:space:]]in[[:space:]]Multimodal[[:space:]]LLMs/e2c27591-e872-476e-995d-31a672c4e0e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MMAD_[[:space:]]Multi-label[[:space:]]Micro-Action[[:space:]]Detection[[:space:]]in[[:space:]]Videos/f6159674-c99d-4a68-b5b5-fcc695377456_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MMAIF_[[:space:]]Multi-task[[:space:]]and[[:space:]]Multi-degradation[[:space:]]All-in-One[[:space:]]for[[:space:]]Image[[:space:]]Fusion[[:space:]]with[[:space:]]Language[[:space:]]Guidance/52b7b646-69c0-4db7-b16b-25fb1261603a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MMAT-1M_[[:space:]]A[[:space:]]Large[[:space:]]Reasoning[[:space:]]Dataset[[:space:]]for[[:space:]]Multimodal[[:space:]]Agent[[:space:]]Tuning/52445199-1ec2-44ae-8684-ebc7af90d2f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MMCR_[[:space:]]Benchmarking[[:space:]]Cross-Source[[:space:]]Reasoning[[:space:]]in[[:space:]]Scientific[[:space:]]Papers/df43d47e-586d-434b-b636-4429893cf5ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MMGeo_[[:space:]]Multimodal[[:space:]]Compositional[[:space:]]Geo-Localization[[:space:]]for[[:space:]]UAVs/795fbeb1-4d32-4cea-80af-b5026df6df65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MMOne_[[:space:]]Representing[[:space:]]Multiple[[:space:]]Modalities[[:space:]]in[[:space:]]One[[:space:]]Scene/bc330a1e-a3fe-4d16-9abc-f5487d57ea36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MMReason_[[:space:]]An[[:space:]]Open-Ended[[:space:]]Multi-Modal[[:space:]]Multi-Step[[:space:]]Reasoning[[:space:]]Benchmark[[:space:]]for[[:space:]]MLLMs[[:space:]]Toward[[:space:]]AGI/953781f5-7da8-4355-8369-f3c8132555c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MOBIUS_[[:space:]]Big-to-Mobile[[:space:]]Universal[[:space:]]Instance[[:space:]]Segmentation[[:space:]]via[[:space:]]Multi-modal[[:space:]]Bottleneck[[:space:]]Fusion[[:space:]]and[[:space:]]Calibrated[[:space:]]Decoder[[:space:]]Pruning/603c3d30-fab7-4cce-922c-473d64bcf902_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MOERL_[[:space:]]When[[:space:]]Mixture-of-Experts[[:space:]]Meet[[:space:]]Reinforcement[[:space:]]Learning[[:space:]]for[[:space:]]Adverse[[:space:]]Weather[[:space:]]Image[[:space:]]Restoration/2a24efbb-7f04-44ea-a7af-0f020c9be8ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MOSAIC_[[:space:]]Generating[[:space:]]Consistent,[[:space:]]Privacy-Preserving[[:space:]]Scenes[[:space:]]from[[:space:]]Multiple[[:space:]]Depth[[:space:]]Views[[:space:]]in[[:space:]]Multi-Room[[:space:]]Environments/eacb0ab0-b716-4372-bfaa-858b903a3e8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MOSCATO_[[:space:]]Predicting[[:space:]]Multiple[[:space:]]Object[[:space:]]State[[:space:]]Change[[:space:]]Through[[:space:]]Actions/0f77a7f0-5337-445c-bca6-3398b86851d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MOVE_[[:space:]]Motion-Guided[[:space:]]Few-Shot[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/627ae65b-36a5-450d-8a0b-88c83db89a8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MP-HSIR_[[:space:]]A[[:space:]]Multi-Prompt[[:space:]]Framework[[:space:]]for[[:space:]]Universal[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Restoration/285f17b1-2453-499c-b715-25c06c6e9c20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MPBR_[[:space:]]Multimodal[[:space:]]Progressive[[:space:]]Bidirectional[[:space:]]Reasoning[[:space:]]for[[:space:]]Open-Set[[:space:]]Fine-Grained[[:space:]]Recognition/2aa72232-1ca8-4842-aa95-aa7bf8d20d29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MPG-SAM[[:space:]]2_[[:space:]]Adapting[[:space:]]SAM[[:space:]]2[[:space:]]with[[:space:]]Mask[[:space:]]Priors[[:space:]]and[[:space:]]Global[[:space:]]Context[[:space:]]for[[:space:]]Referring[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/c995d143-4f3d-4c17-8b97-eac5a097368c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MR-FIQA_[[:space:]]Face[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment[[:space:]]with[[:space:]]Multi-Reference[[:space:]]Representations[[:space:]]from[[:space:]]Synthetic[[:space:]]Data[[:space:]]Generation/ac3bcbf3-eb35-40ea-bd4d-7633ee791e58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MRGen_[[:space:]]Segmentation[[:space:]]Data[[:space:]]Engine[[:space:]]For[[:space:]]Underrepresented[[:space:]]MRI[[:space:]]Modalities/d05decbb-0c0f-4f8a-b65e-4bbea63db075_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MS3D_[[:space:]]High-Quality[[:space:]]3D[[:space:]]Generation[[:space:]]via[[:space:]]Multi-Scale[[:space:]]Representation[[:space:]]Modeling/94a6f2e5-b487-4929-9014-d56838a9197d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MSA2_[[:space:]]Multi-task[[:space:]]Framework[[:space:]]with[[:space:]]Structure-aware[[:space:]]and[[:space:]]Style-adaptive[[:space:]]Character[[:space:]]Representation[[:space:]]for[[:space:]]Open-set[[:space:]]Chinese[[:space:]]Text[[:space:]]Recognition/5a2aa1fa-8b69-467b-9c2e-bdeae558f487_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MSQ_[[:space:]]Memory-Efficient[[:space:]]Bit[[:space:]]Sparsification[[:space:]]Quantization/ed536b3f-77bc-4172-a620-3e63eabe3fff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MUG_[[:space:]]Pseudo[[:space:]]Labeling[[:space:]]Augmented[[:space:]]Audio-Visual[[:space:]]Mamba[[:space:]]Network[[:space:]]for[[:space:]]Audio-Visual[[:space:]]Video[[:space:]]Parsing/9609dc25-5042-4bca-b11b-9dcf889eb008_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MUNBa_[[:space:]]Machine[[:space:]]Unlearning[[:space:]]via[[:space:]]Nash[[:space:]]Bargaining/b319cfd7-9fc1-4881-8094-36360759ec94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MUSE-VL_[[:space:]]Modeling[[:space:]]Unified[[:space:]]VLM[[:space:]]through[[:space:]]Semantic[[:space:]]Discrete[[:space:]]Encoding/329d8cbf-a14f-429b-91c0-168f31c7aef6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MUSE_[[:space:]]Multi-Subject[[:space:]]Unified[[:space:]]Synthesis[[:space:]]via[[:space:]]Explicit[[:space:]]Layout[[:space:]]Semantic[[:space:]]Expansion/358dbbeb-1ca4-4d1e-a0dd-c07c9a6d9d06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MV-Adapter_[[:space:]]Multi-View[[:space:]]Consistent[[:space:]]Image[[:space:]]Generation[[:space:]]Made[[:space:]]Easy/a174ac54-6a33-41fb-b4c5-d43775e4137b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MVGBench_[[:space:]]a[[:space:]]Comprehensive[[:space:]]Benchmark[[:space:]]for[[:space:]]Multi-view[[:space:]]Generation[[:space:]]Models/eb81092b-ce85-4636-9133-a32a981b655a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MVQA_[[:space:]]Mamba[[:space:]]with[[:space:]]Unified[[:space:]]Sampling[[:space:]]for[[:space:]]Efficient[[:space:]]Video[[:space:]]Quality[[:space:]]Assessment/3f7c9501-fb94-4cb9-8d8c-e18a060473b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MVTrajecter_[[:space:]]Multi-View[[:space:]]Pedestrian[[:space:]]Tracking[[:space:]]with[[:space:]]Trajectory[[:space:]]Motion[[:space:]]Cost[[:space:]]and[[:space:]]Trajectory[[:space:]]Appearance[[:space:]]Cost/103842ff-246f-43f8-963e-741d491eb618_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaGS_[[:space:]]Reconstructing[[:space:]]and[[:space:]]Simulating[[:space:]]Dynamic[[:space:]]3D[[:space:]]Objects[[:space:]]with[[:space:]]Mesh-adsorbed[[:space:]]Gaussian[[:space:]]Splatting/4d68afda-2e89-4c0b-b1ef-80f3a31d2d97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaTVLM_[[:space:]]Hybrid[[:space:]]Mamba-Transformer[[:space:]]for[[:space:]]Efficient[[:space:]]Vision-Language[[:space:]]Modeling/fd45a6ac-a98e-4153-bf9a-e49e62ac93b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaTe_[[:space:]]Images[[:space:]]Are[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]for[[:space:]]Material[[:space:]]Transfer[[:space:]]via[[:space:]]Diffusion[[:space:]]Transformer/3ae0ca4b-e0ef-4651-abb4-0f2961c4bcd4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MagShield_[[:space:]]Towards[[:space:]]Better[[:space:]]Robustness[[:space:]]in[[:space:]]Sparse[[:space:]]Inertial[[:space:]]Motion[[:space:]]Capture[[:space:]]Under[[:space:]]Magnetic[[:space:]]Disturbances/7d22c329-6932-455a-8391-f0415b7f7381_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Magic[[:space:]]Insert_[[:space:]]Style-Aware[[:space:]]Drag-and-Drop/3c3daa6a-97c3-4433-9ce9-096e9972a277_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MagicCity_[[:space:]]Geometry-Aware[[:space:]]3D[[:space:]]City[[:space:]]Generation[[:space:]]from[[:space:]]Satellite[[:space:]]Imagery[[:space:]]with[[:space:]]Multi-View[[:space:]]Consistency/08088e93-a4b2-4c1d-be64-ecc4185163f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MagicColor_[[:space:]]Multi-Instance[[:space:]]Sketch[[:space:]]Colorization/bf310549-bf55-4ecd-b783-f30b0e30bbdf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MagicDrive-V2_[[:space:]]High-Resolution[[:space:]]Long[[:space:]]Video[[:space:]]Generation[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving[[:space:]]with[[:space:]]Adaptive[[:space:]]Control/2f6a1591-462e-4815-bd1c-98fb6d608be8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MagicHOI_[[:space:]]Leveraging[[:space:]]3D[[:space:]]Priors[[:space:]]for[[:space:]]Accurate[[:space:]]Hand-object[[:space:]]Reconstruction[[:space:]]from[[:space:]]Short[[:space:]]Monocular[[:space:]]Video[[:space:]]Clips/baed5b3e-cc85-4805-b48a-6d1018b16ea9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MagicID_[[:space:]]Hybrid[[:space:]]Preference[[:space:]]Optimization[[:space:]]for[[:space:]]ID-Consistent[[:space:]]and[[:space:]]Dynamic-Preserved[[:space:]]Video[[:space:]]Customization/bc3669e7-6685-46de-9a3a-daed9384f625_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MagicMirror_[[:space:]]ID-Preserved[[:space:]]Video[[:space:]]Generation[[:space:]]in[[:space:]]Video[[:space:]]Diffusion[[:space:]]Transformers/a6614970-98a3-474d-bae2-83912be21239_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MagicMotion_[[:space:]]Controllable[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Dense-to-Sparse[[:space:]]Trajectory[[:space:]]Guidance/b9ee158f-3eff-45ff-b58f-dc0ba4eba8cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Make[[:space:]]Me[[:space:]]Happier_[[:space:]]Evoking[[:space:]]Emotions[[:space:]]Through[[:space:]]Image[[:space:]]Diffusion[[:space:]]Models/6d22bdc6-b259-4a53-a7c8-1a17742b58f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Make[[:space:]]Your[[:space:]]Training[[:space:]]Flexible_[[:space:]]Towards[[:space:]]Deployment-Efficient[[:space:]]Video[[:space:]]Models/5a324670-557f-4ffd-a5cb-4099fdfb1a08_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MamTiff-CAD_[[:space:]]Multi-Scale[[:space:]]Latent[[:space:]]Diffusion[[:space:]]with[[:space:]]Mamba+[[:space:]]for[[:space:]]Complex[[:space:]]Parametric[[:space:]]Sequence/6e354b74-7106-478e-9e40-9006fa0b707e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MamV2XCalib_[[:space:]]V2X-based[[:space:]]Target-less[[:space:]]Infrastructure[[:space:]]Camera[[:space:]]Calibration[[:space:]]with[[:space:]]State[[:space:]]Space[[:space:]]Model/30192a5f-9edd-4e11-ba0b-b472e678200d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mamba-3VL_[[:space:]]Taming[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]for[[:space:]]3D[[:space:]]Vision[[:space:]]Language[[:space:]]Learning/b38bcafa-521d-4aeb-b26f-864cb0ef6892_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MambaML_[[:space:]]Exploring[[:space:]]State[[:space:]]Space[[:space:]]Models[[:space:]]for[[:space:]]Multi-Label[[:space:]]Image[[:space:]]Classification/4fd80da5-8f9e-456c-9826-05a8dc397ab7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Manual-PA_[[:space:]]Learning[[:space:]]3D[[:space:]]Part[[:space:]]Assembly[[:space:]]from[[:space:]]Instruction[[:space:]]Diagrams/74b3b947-876e-452d-95aa-e01fda87a7f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Marigold-DC_[[:space:]]Zero-Shot[[:space:]]Monocular[[:space:]]Depth[[:space:]]Completion[[:space:]]with[[:space:]]Guided[[:space:]]Diffusion/cf367e26-2490-4b3d-8ee5-2b8d08d2289a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaskControl_[[:space:]]Spatio-Temporal[[:space:]]Control[[:space:]]for[[:space:]]Masked[[:space:]]Motion[[:space:]]Synthesis/d951b9cd-53fd-44a9-bd12-f0c45ed0cb0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaskHand_[[:space:]]Generative[[:space:]]Masked[[:space:]]Modeling[[:space:]]for[[:space:]]Robust[[:space:]]Hand[[:space:]]Mesh[[:space:]]Reconstruction[[:space:]]in[[:space:]]the[[:space:]]Wild/fb4a4782-380c-4485-ad2a-9b81184216c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaskSAM_[[:space:]]Auto-prompt[[:space:]]SAM[[:space:]]with[[:space:]]Mask[[:space:]]Classification[[:space:]]for[[:space:]]Volumetric[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/89bee5cf-5e07-4c8f-b0ff-80c591a6e9a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mastering[[:space:]]Collaborative[[:space:]]Multi-modal[[:space:]]Data[[:space:]]Selection_[[:space:]]A[[:space:]]Focus[[:space:]]on[[:space:]]Informativeness,[[:space:]]Uniqueness,[[:space:]]and[[:space:]]Representativeness/000cb9eb-f781-45b3-b965-f12cf174236a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MatchDiffusion_[[:space:]]Training-free[[:space:]]Generation[[:space:]]of[[:space:]]Match-Cuts/c98f1867-fbe6-478e-a279-d996d806d76c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MaterialMVP_[[:space:]]Illumination-Invariant[[:space:]]Material[[:space:]]Generation[[:space:]]via[[:space:]]Multi-view[[:space:]]PBR[[:space:]]Diffusion/8c9119ab-49f4-4e5a-93b3-2c93c46f6466_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MeasureXpert_[[:space:]]Automatic[[:space:]]Anthropometric[[:space:]]Measurement[[:space:]]Extraction[[:space:]]from[[:space:]]Two[[:space:]]Unregistered,[[:space:]]Partial,[[:space:]]Posed,[[:space:]]and[[:space:]]Dressed[[:space:]]Body[[:space:]]Scans/0c572caf-c1a9-4063-b590-901260e5bb71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Measuring[[:space:]]the[[:space:]]Impact[[:space:]]of[[:space:]]Rotation[[:space:]]Equivariance[[:space:]]on[[:space:]]Aerial[[:space:]]Object[[:space:]]Detection/a2510d9e-39f8-4a2f-98dd-d2a269dcff52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MedSegFactory_[[:space:]]Text-Guided[[:space:]]Generation[[:space:]]of[[:space:]]Medical[[:space:]]Image-Mask[[:space:]]Pairs/84a209f5-36a4-44f8-94a5-09ba63268967_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MedVSR_[[:space:]]Medical[[:space:]]Video[[:space:]]Super-Resolution[[:space:]]with[[:space:]]Cross[[:space:]]State-Space[[:space:]]Propagation/72a5300e-9ec3-40b5-9b3a-919bb4bdd0be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Medical[[:space:]]World[[:space:]]Model/14729f2c-bdbc-4837-a6ed-80275043fc16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MemDistill_[[:space:]]Distilling[[:space:]]LiDAR[[:space:]]Knowledge[[:space:]]into[[:space:]]Memory[[:space:]]for[[:space:]]Camera-Only[[:space:]]3D[[:space:]]Object[[:space:]]Detection/28201f57-4266-44c9-9caf-303dd1262fda_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Membership[[:space:]]Inference[[:space:]]Attacks[[:space:]]with[[:space:]]False[[:space:]]Discovery[[:space:]]Rate[[:space:]]Control/daf6e116-8bb5-4e74-acd1-2dfff950b4a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Memory-Efficient[[:space:]]4-bit[[:space:]]Preconditioned[[:space:]]Stochastic[[:space:]]Optimization/0cb885b3-78d4-4cbe-b7b7-72cade6a0361_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Memory-Efficient[[:space:]]Generative[[:space:]]Models[[:space:]]via[[:space:]]Product[[:space:]]Quantization/bb056e46-74c1-4dcc-876f-96af77b8bb75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MemoryTalker_[[:space:]]Personalized[[:space:]]Speech-Driven[[:space:]]3D[[:space:]]Facial[[:space:]]Animation[[:space:]]via[[:space:]]Audio-Guided[[:space:]]Stylization/d61de92e-9271-4c49-a938-e811c370e041_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MergeOcc_[[:space:]]Bridge[[:space:]]the[[:space:]]Domain[[:space:]]Gap[[:space:]]between[[:space:]]Different[[:space:]]LiDARs[[:space:]]for[[:space:]]Robust[[:space:]]Occupancy[[:space:]]Prediction/b3ddbea8-50d7-43b3-bab8-2602b27c10f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MeshAnything[[:space:]]V2_[[:space:]]Artist-Created[[:space:]]Mesh[[:space:]]Generation[[:space:]]with[[:space:]]Adjacent[[:space:]]Mesh[[:space:]]Tokenization/58bf335b-fe4d-4faf-90e3-2f45b68fde54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MeshLLM_[[:space:]]Empowering[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]to[[:space:]]Progressively[[:space:]]Understand[[:space:]]and[[:space:]]Generate[[:space:]]3D[[:space:]]Mesh/1fa31c30-dc31-423e-a785-0c002aae420d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MeshMamba_[[:space:]]State[[:space:]]Space[[:space:]]Models[[:space:]]for[[:space:]]Articulated[[:space:]]3D[[:space:]]Mesh[[:space:]]Generation[[:space:]]and[[:space:]]Reconstruction/b6d3343b-4841-4ebb-9905-d40a5409e5c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MeshPad_[[:space:]]Interactive[[:space:]]Sketch-Conditioned[[:space:]]Artist-Reminiscent[[:space:]]Mesh[[:space:]]Generation[[:space:]]and[[:space:]]Editing/54f3d4b8-867a-4b36-90cf-c96e58764346_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Met2Net_[[:space:]]A[[:space:]]Decoupled[[:space:]]Two-Stage[[:space:]]Spatio-Temporal[[:space:]]Forecasting[[:space:]]Model[[:space:]]for[[:space:]]Complex[[:space:]]Meteorological[[:space:]]Systems/1fbfabcc-332c-4414-b55b-c30738c4d30c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Meta-Learning[[:space:]]Dynamic[[:space:]]Center[[:space:]]Distance_[[:space:]]Hard[[:space:]]Sample[[:space:]]Mining[[:space:]]for[[:space:]]Learning[[:space:]]with[[:space:]]Noisy[[:space:]]Labels/e0651e8b-06bb-4867-81cd-98c347aea906_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Meta-Unlearning[[:space:]]on[[:space:]]Diffusion[[:space:]]Models_[[:space:]]Preventing[[:space:]]Relearning[[:space:]]Unlearned[[:space:]]Concepts/cef0b1ff-275e-4147-8652-b1c74c5cc0ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MetaMorph_[[:space:]]Multimodal[[:space:]]Understanding[[:space:]]and[[:space:]]Generation[[:space:]]via[[:space:]]Instruction[[:space:]]Tuning/c0f49922-7580-4ee0-b1fa-15611105fbe9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MetaScope_[[:space:]]Optics-Driven[[:space:]]Neural[[:space:]]Network[[:space:]]for[[:space:]]Ultra-Micro[[:space:]]Metalens[[:space:]]Endoscopy/288a8ede-12df-42de-95fa-4cca52a3027d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Metric[[:space:]]Convolutions_[[:space:]]A[[:space:]]Unifying[[:space:]]Theory[[:space:]]to[[:space:]]Adaptive[[:space:]]Image[[:space:]]Convolutions/42279ded-c6ed-49a1-8883-d98bc11b57bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MiDSummer_[[:space:]]Multi-Guidance[[:space:]]Diffusion[[:space:]]for[[:space:]]Controllable[[:space:]]Zero-Shot[[:space:]]Immersive[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Scene[[:space:]]Generation/c765d495-c9f1-4dad-a001-ec6d674ca975_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MikuDance_[[:space:]]Animating[[:space:]]Character[[:space:]]Art[[:space:]]with[[:space:]]Mixed[[:space:]]Motion[[:space:]]Dynamics/de08f484-e73c-423b-9a48-8ed91abffdc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MinCD-PnP_[[:space:]]Learning[[:space:]]2D-3D[[:space:]]Correspondences[[:space:]]with[[:space:]]Approximate[[:space:]]Blind[[:space:]]PnP/263d65a0-205e-4c5b-8b91-03badcdf7c38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mind[[:space:]]the[[:space:]]Cost[[:space:]]of[[:space:]]Scaffold![[:space:]]Benign[[:space:]]Clients[[:space:]]May[[:space:]]Even[[:space:]]Become[[:space:]]Accomplices[[:space:]]of[[:space:]]Backdoor[[:space:]]Attack/cf77c680-8736-4448-87f1-1fa719a0a0eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mind[[:space:]]the[[:space:]]Gap_[[:space:]]Aligning[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models[[:space:]]to[[:space:]]Image[[:space:]]Feature[[:space:]]Matching/4a79c4c6-2ae1-43c9-a364-e8258523ed0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mind[[:space:]]the[[:space:]]Gap_[[:space:]]Preserving[[:space:]]and[[:space:]]Compensating[[:space:]]for[[:space:]]the[[:space:]]Modality[[:space:]]Gap[[:space:]]in[[:space:]]CLIP-Based[[:space:]]Continual[[:space:]]Learning/b0da156e-1841-4f99-a384-d891bb262bb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MissRAG_[[:space:]]Addressing[[:space:]]the[[:space:]]Missing[[:space:]]Modality[[:space:]]Challenge[[:space:]]in[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/171da703-977d-4a7f-9bd2-937688db6d5e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MistSense_[[:space:]]Versatile[[:space:]]Online[[:space:]]Detection[[:space:]]of[[:space:]]Procedural[[:space:]]and[[:space:]]Execution[[:space:]]Mistakes/b81e6cfb-8577-4905-8c69-feb143b08209_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mitigating[[:space:]]Catastrophic[[:space:]]Overfitting[[:space:]]in[[:space:]]Fast[[:space:]]Adversarial[[:space:]]Training[[:space:]]via[[:space:]]Label[[:space:]]Information[[:space:]]Elimination/fbc88682-89a4-4a29-8c31-628bc5252167_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mitigating[[:space:]]Geometric[[:space:]]Degradation[[:space:]]in[[:space:]]Fast[[:space:]]DownSampling[[:space:]]via[[:space:]]FastAdapter[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation/9669b53d-bf78-4363-b001-523b0ef159d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mitigating[[:space:]]Object[[:space:]]Hallucinations[[:space:]]via[[:space:]]Sentence-Level[[:space:]]Early[[:space:]]Intervention/6c76d702-1672-4f68-ba9d-ada179e75f70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MixA-Q_[[:space:]]Revisiting[[:space:]]Activation[[:space:]]Sparsity[[:space:]]for[[:space:]]Vision[[:space:]]Transformers[[:space:]]from[[:space:]]a[[:space:]]Mixed-Precision[[:space:]]Quantization[[:space:]]Perspective/fc67ba09-f12f-40a8-bf6c-d17a5639c528_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MixANT_[[:space:]]Observation-dependent[[:space:]]Memory[[:space:]]Propagation[[:space:]]for[[:space:]]Stochastic[[:space:]]Dense[[:space:]]Action[[:space:]]Anticipation/a41f2350-3192-4fda-b614-10b00ebf5843_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MixA_[[:space:]]A[[:space:]]Mixed[[:space:]]Attention[[:space:]]approach[[:space:]]with[[:space:]]Stable[[:space:]]Lightweight[[:space:]]Linear[[:space:]]Attention[[:space:]]to[[:space:]]enhance[[:space:]]Efficiency[[:space:]]of[[:space:]]Vision[[:space:]]Transformers[[:space:]]at[[:space:]]the[[:space:]]Edge/da867698-7ede-4819-980e-cf45a6153784_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MixRI_[[:space:]]Mixing[[:space:]]Features[[:space:]]of[[:space:]]Reference[[:space:]]Images[[:space:]]for[[:space:]]Novel[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/c57d4296-f633-4634-ab0c-df7cc6046930_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mixed[[:space:]]Signals_[[:space:]]A[[:space:]]Diverse[[:space:]]Point[[:space:]]Cloud[[:space:]]Dataset[[:space:]]for[[:space:]]Heterogeneous[[:space:]]LiDAR[[:space:]]V2X[[:space:]]Collaboration/f57a9861-2b7b-4852-a144-996b4fabe70f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mixture[[:space:]]of[[:space:]]Experts[[:space:]]Guided[[:space:]]by[[:space:]]Gaussian[[:space:]]Splatters[[:space:]]Matters_[[:space:]]A[[:space:]]new[[:space:]]Approach[[:space:]]to[[:space:]]Weakly-Supervised[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/79b8f1f2-85bb-46e0-90db-508449adfafc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mixture-of-Scores_[[:space:]]Robust[[:space:]]Image-Text[[:space:]]Data[[:space:]]Valuation[[:space:]]via[[:space:]]Three[[:space:]]Lines[[:space:]]of[[:space:]]Code/4ed17ef1-9ddf-44ef-ac78-b8b842e9918e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoFRR_[[:space:]]Mixture[[:space:]]of[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Face[[:space:]]Retouching[[:space:]]Restoration/ccbed0f5-0cda-43ac-9f96-e65680e55d64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoGA_[[:space:]]3D[[:space:]]Generative[[:space:]]Avatar[[:space:]]Prior[[:space:]]for[[:space:]]Monocular[[:space:]]Gaussian[[:space:]]Avatar[[:space:]]Reconstruction/c6d041f6-cba8-439b-ac5d-543b196dfba3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoMa-Kitchen_[[:space:]]A[[:space:]]100K+[[:space:]]Benchmark[[:space:]]for[[:space:]]Affordance-Grounded[[:space:]]Last-Mile[[:space:]]Navigation[[:space:]]in[[:space:]]Mobile[[:space:]]Manipulation/04575eff-6149-4286-98db-5b872a8155bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoMaps_[[:space:]]Semantics-Aware[[:space:]]Scene[[:space:]]Motion[[:space:]]Generation[[:space:]]with[[:space:]]Motion[[:space:]]Maps/c6c7ca72-2f2e-4dd1-ba82-7f6d02fa53e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MoSiC_[[:space:]]Optimal-Transport[[:space:]]Motion[[:space:]]Trajectory[[:space:]]for[[:space:]]Dense[[:space:]]Self-Supervised[[:space:]]Learning/ac3ed535-2a82-4a34-8e92-8d1d840b8e34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Mobile[[:space:]]Video[[:space:]]Diffusion/a2a4808a-71e5-4b95-8480-18e27bd5c25d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MobileIE_[[:space:]]An[[:space:]]Extremely[[:space:]]Lightweight[[:space:]]and[[:space:]]Effective[[:space:]]ConvNet[[:space:]]for[[:space:]]Real-Time[[:space:]]Image[[:space:]]Enhancement[[:space:]]on[[:space:]]Mobile[[:space:]]Devices/8c06ff8a-63cf-448f-956e-3ba5dbc3d1f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MobileViCLIP_[[:space:]]An[[:space:]]Efficient[[:space:]]Video-Text[[:space:]]Model[[:space:]]for[[:space:]]Mobile[[:space:]]Devices/e75b4ef3-93c8-4f8f-837f-9dce69bcb661_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ModSkill_[[:space:]]Physical[[:space:]]Character[[:space:]]Skill[[:space:]]Modularization/89cbf093-3afb-47f3-8ce9-1d8953a9f46f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ModalTune_[[:space:]]Fine-Tuning[[:space:]]Slide-Level[[:space:]]Foundation[[:space:]]Models[[:space:]]with[[:space:]]Multi-Modal[[:space:]]Information[[:space:]]for[[:space:]]Multi-task[[:space:]]Learning[[:space:]]in[[:space:]]Digital[[:space:]]Pathology/6433a680-550d-4b98-9e8c-1945a64a8867_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Model[[:space:]]Reveals[[:space:]]What[[:space:]]to[[:space:]]Cache_[[:space:]]Profiling-Based[[:space:]]Feature[[:space:]]Reuse[[:space:]]for[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/2c23942a-5633-4c0f-83ef-a74a360d3e5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Modeling[[:space:]]Human[[:space:]]Gaze[[:space:]]Behavior[[:space:]]with[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Unified[[:space:]]Scanpath[[:space:]]Prediction/28aee7a3-79f1-4386-8f7a-b705b2b560a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Modeling[[:space:]]Saliency[[:space:]]Dataset[[:space:]]Bias/c0a859c9-2e9e-4850-8c72-5a80408acc25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Moderating[[:space:]]the[[:space:]]Generalization[[:space:]]of[[:space:]]Score-based[[:space:]]Generative[[:space:]]Model/a60cab8a-7677-416d-9adc-cf64532e3c45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MolParser_[[:space:]]End-to-end[[:space:]]Visual[[:space:]]Recognition[[:space:]]of[[:space:]]Molecule[[:space:]]Structures[[:space:]]in[[:space:]]the[[:space:]]Wild/9c40e628-1f14-4c08-8e06-3d3f2c4e760d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Moment[[:space:]]Quantization[[:space:]]for[[:space:]]Video[[:space:]]Temporal[[:space:]]Grounding/5c93f455-2739-43fb-97ab-df135c80dd01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Momentum-GS_[[:space:]]Momentum[[:space:]]Gaussian[[:space:]]Self-Distillation[[:space:]]for[[:space:]]High-Quality[[:space:]]Large[[:space:]]Scene[[:space:]]Reconstruction/87d70e7e-c5ff-457d-8425-2d368cf679de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MonSTeR_[[:space:]]a[[:space:]]Unified[[:space:]]Model[[:space:]]for[[:space:]]Motion,[[:space:]]Scene,[[:space:]]Text[[:space:]]Retrieval/67d906c5-0d3f-4134-bd5d-9bcb537d5534_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MonoFusion_[[:space:]]Sparse-View[[:space:]]4D[[:space:]]Reconstruction[[:space:]]via[[:space:]]Monocular[[:space:]]Fusion/849dde0e-b3c2-4411-8b3b-85e63f2caf1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MonoMVSNet_[[:space:]]Monocular[[:space:]]Priors[[:space:]]Guided[[:space:]]Multi-View[[:space:]]Stereo[[:space:]]Network/dada7867-8ef8-4647-a209-e9b4d30c390d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MonoMobility_[[:space:]]Zero-Shot[[:space:]]3D[[:space:]]Mobility[[:space:]]Analysis[[:space:]]from[[:space:]]Monocular[[:space:]]Videos/19262ecf-a81b-44e7-b3dd-1681d1aa520a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MonoSOWA_[[:space:]]Scalable[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detector[[:space:]]Without[[:space:]]Human[[:space:]]Annotations/957a8895-db49-431a-b94a-115def547ac9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Monocular[[:space:]]Facial[[:space:]]Appearance[[:space:]]Capture[[:space:]]in[[:space:]]the[[:space:]]Wild/48c58acd-fffa-4f67-ae06-ca330f10cdd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Monocular[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion[[:space:]]via[[:space:]]Masked[[:space:]]Recurrent[[:space:]]Networks/b048a874-593c-48ab-94c2-bb66539d61a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/More[[:space:]]Reliable[[:space:]]Pseudo-labels,[[:space:]]Better[[:space:]]Performance_[[:space:]]A[[:space:]]Generalized[[:space:]]Approach[[:space:]]to[[:space:]]Single[[:space:]]Positive[[:space:]]Multi-label[[:space:]]Learning/a76759ac-a643-448c-813c-435a290762fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Morph_[[:space:]]A[[:space:]]Motion-free[[:space:]]Physics[[:space:]]Optimization[[:space:]]Framework[[:space:]]for[[:space:]]Human[[:space:]]Motion[[:space:]]Generation/6f4b46c4-6fd9-4caf-a778-46addc441c5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MorphoGen_[[:space:]]Efficient[[:space:]]Unconditional[[:space:]]Generation[[:space:]]of[[:space:]]Long-Range[[:space:]]Projection[[:space:]]Neuronal[[:space:]]Morphology[[:space:]]via[[:space:]]a[[:space:]]Global-to-Local[[:space:]]Framework/59526794-7101-4a34-81d6-51d90d1bca1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MosaicDiff_[[:space:]]Training-free[[:space:]]Structural[[:space:]]Pruning[[:space:]]for[[:space:]]Diffusion[[:space:]]Model[[:space:]]Acceleration[[:space:]]Reflecting[[:space:]]Pretraining[[:space:]]Dynamics/fbff899d-a6c5-493b-ae72-d68995e13a99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Motal_[[:space:]]Unsupervised[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]by[[:space:]]Modality[[:space:]]and[[:space:]]Task-specific[[:space:]]Knowledge[[:space:]]Transfer/b9960c7d-1dc3-4d1a-b4e2-cdcc92805943_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Motion[[:space:]]Synthesis[[:space:]]with[[:space:]]Sparse[[:space:]]and[[:space:]]Flexible[[:space:]]Keyjoint[[:space:]]Control/2d8236a5-463e-456b-932d-ec2b7e294f3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Motion-2-to-3_[[:space:]]Leveraging[[:space:]]2D[[:space:]]Motion[[:space:]]Data[[:space:]]for[[:space:]]3D[[:space:]]Motion[[:space:]]Generations/f5a54a01-e9d9-404b-818c-67812ad7d710_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotionAgent_[[:space:]]Fine-grained[[:space:]]Controllable[[:space:]]Video[[:space:]]Generation[[:space:]]via[[:space:]]Motion[[:space:]]Field[[:space:]]Agent/ded27b9a-b5db-4b63-98ee-48196586201f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotionCtrl_[[:space:]]A[[:space:]]Real-time[[:space:]]Controllable[[:space:]]Vision-Language-Motion[[:space:]]Model/ec3d6711-ce42-4e2a-93aa-690051ee62bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotionDiff_[[:space:]]Training-free[[:space:]]Zero-shot[[:space:]]Interactive[[:space:]]Motion[[:space:]]Editing[[:space:]]via[[:space:]]Flow-assisted[[:space:]]Multi-view[[:space:]]Diffusion/69987dba-1c49-4dd9-a943-36ce80950e21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotionFollower_[[:space:]]Editing[[:space:]]Video[[:space:]]Motion[[:space:]]via[[:space:]]Score-Guided[[:space:]]Diffusion/d9bbdc1b-d5d3-4100-8384-80df2e169d20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotionLab_[[:space:]]Unified[[:space:]]Human[[:space:]]Motion[[:space:]]Generation[[:space:]]and[[:space:]]Editing[[:space:]]via[[:space:]]the[[:space:]]Motion-Condition-Motion[[:space:]]Paradigm/7aac3f8c-39c6-46e8-bdd7-233b53ebe19e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotionShot_[[:space:]]Adaptive[[:space:]]Motion[[:space:]]Transfer[[:space:]]across[[:space:]]Arbitrary[[:space:]]Objects[[:space:]]for[[:space:]]Text-to-Video[[:space:]]Generation/24ea4d82-751b-40ee-bd83-975a8c7443f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MotionStreamer_[[:space:]]Streaming[[:space:]]Motion[[:space:]]Generation[[:space:]]via[[:space:]]Diffusion-based[[:space:]]Autoregressive[[:space:]]Model[[:space:]]in[[:space:]]Causal[[:space:]]Latent[[:space:]]Space/70f20111-d99a-40ab-9636-9c306c59c412_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Moto_[[:space:]]Latent[[:space:]]Motion[[:space:]]Token[[:space:]]as[[:space:]]the[[:space:]]Bridging[[:space:]]Language[[:space:]]for[[:space:]]Learning[[:space:]]Robot[[:space:]]Manipulation[[:space:]]from[[:space:]]Videos/34ce4357-a504-4cd4-9edc-73e12f4ea820_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Move[[:space:]]to[[:space:]]Understand[[:space:]]a[[:space:]]3D[[:space:]]Scene_[[:space:]]Bridging[[:space:]]Visual[[:space:]]Grounding[[:space:]]and[[:space:]]Exploration[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Versatile[[:space:]]Embodied[[:space:]]Navigation/bd4e6e49-6f2e-4945-9184-3cfbceba25d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MuGS_[[:space:]]Multi-Baseline[[:space:]]Generalizable[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Reconstruction/14c018a9-7d2c-4e28-a996-e24b555e03c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Cache[[:space:]]Enhanced[[:space:]]Prototype[[:space:]]Learning[[:space:]]for[[:space:]]Test-Time[[:space:]]Generalization[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models/6cafc638-2b26-4336-ae6e-28bf81634544_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Granular[[:space:]]Spatio-Temporal[[:space:]]Token[[:space:]]Merging[[:space:]]for[[:space:]]Training-Free[[:space:]]Acceleration[[:space:]]of[[:space:]]Video[[:space:]]LLMs/e421d4db-9531-401d-bdbf-df8c32ab50f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Modal[[:space:]]Few-Shot[[:space:]]Temporal[[:space:]]Action[[:space:]]Segmentation/9a761976-11c0-4c70-b09a-6d7bb884d347_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Modal[[:space:]]Multi-Task[[:space:]]Unified[[:space:]]Embedding[[:space:]]Model[[:space:]](M3T-UEM)_[[:space:]]A[[:space:]]Task-Adaptive[[:space:]]Representation[[:space:]]Learning[[:space:]]Framework/e2f7a3a9-6b05-4254-bbfc-2817ecec023c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Object[[:space:]]Sketch[[:space:]]Animation[[:space:]]by[[:space:]]Scene[[:space:]]Decomposition[[:space:]]and[[:space:]]Motion[[:space:]]Planning/5709dc7c-9fec-439a-b1dc-a3ff1e1ea5f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-Schema[[:space:]]Proximity[[:space:]]Network[[:space:]]for[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval/276a58ac-1679-4902-bc1c-e9974e4727db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-View[[:space:]]3D[[:space:]]Point[[:space:]]Tracking/3a50f480-d0cc-42e1-b04b-7a7aed53cb6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-View[[:space:]]Slot[[:space:]]Attention[[:space:]]Using[[:space:]]Paraphrased[[:space:]]Texts[[:space:]]for[[:space:]]Face[[:space:]]Anti-Spoofing/3ae4be93-3dc1-49de-9fdf-b0df2083def3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-identity[[:space:]]Human[[:space:]]Image[[:space:]]Animation[[:space:]]with[[:space:]]Structural[[:space:]]Video[[:space:]]Diffusion/589837b5-df7c-43e3-8ac0-7f3886f830ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-modal[[:space:]]Identity[[:space:]]Extraction/a24df85c-1f05-4d92-977b-749551795bf2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-modal[[:space:]]Multi-platform[[:space:]]Person[[:space:]]Re-Identification_[[:space:]]Benchmark[[:space:]]and[[:space:]]Method/3826aa72-7b6e-401b-ba39-d78ba02805c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-modal[[:space:]]Segment[[:space:]]Anything[[:space:]]Model[[:space:]]for[[:space:]]Camouflaged[[:space:]]Scene[[:space:]]Segmentation/dbf76123-c26f-42dd-aa09-f3195c579240_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-scenario[[:space:]]Overlapping[[:space:]]Text[[:space:]]Segmentation[[:space:]]with[[:space:]]Depth[[:space:]]Awareness/6068a35e-2fef-46ca-93ba-910e01c4b9d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-turn[[:space:]]Consistent[[:space:]]Image[[:space:]]Editing/52c8740f-8a02-4db1-ac26-b2b9acf4b240_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multi-view[[:space:]]Gaze[[:space:]]Target[[:space:]]Estimation/50d738e2-c978-4d0c-af68-3721ee8bb370_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MultiADS_[[:space:]]Defect-aware[[:space:]]Supervision[[:space:]]for[[:space:]]Multi-type[[:space:]]Anomaly[[:space:]]Detection[[:space:]]and[[:space:]]Segmentation[[:space:]]in[[:space:]]Zero-Shot[[:space:]]Learning/71c2c8b1-2d10-495a-a612-c66e5e4c4789_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MultiModal[[:space:]]Action[[:space:]]Conditioned[[:space:]]Video[[:space:]]Simulation/fc0d1632-403d-4c1f-897e-a403007cb581_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MultiVerse_[[:space:]]A[[:space:]]Multi-Turn[[:space:]]Conversation[[:space:]]Benchmark[[:space:]]for[[:space:]]Evaluating[[:space:]]Large[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Models/9605f14b-6996-4043-94fe-40eae31bc26b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multidimensional[[:space:]]Byte[[:space:]]Pair[[:space:]]Encoding_[[:space:]]Shortened[[:space:]]Sequences[[:space:]]for[[:space:]]Improved[[:space:]]Visual[[:space:]]Data[[:space:]]Generation/0d58663f-8298-4306-aa37-2b0deac24245_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multimodal[[:space:]]LLM[[:space:]]Guided[[:space:]]Exploration[[:space:]]and[[:space:]]Active[[:space:]]Mapping[[:space:]]using[[:space:]]Fisher[[:space:]]Information/080348c2-9b8b-4644-bc55-54e35eed26de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multimodal[[:space:]]LLMs[[:space:]]as[[:space:]]Customized[[:space:]]Reward[[:space:]]Models[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/adb7e9c8-96f6-440f-89cc-3c03d9bea0d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model-Guided[[:space:]]ISP[[:space:]]Hyperparameter[[:space:]]Optimization[[:space:]]with[[:space:]]Dynamic[[:space:]]Preference[[:space:]]Learning/ae97d3a1-0e0f-4492-8fbe-e613b60836c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multimodal[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Complex[[:space:]]Sewing[[:space:]]Pattern[[:space:]]Generation/6dec07e3-ef2e-4ba0-aa9c-faf0bb38b642_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multimodal[[:space:]]Prompt[[:space:]]Alignment[[:space:]]for[[:space:]]Facial[[:space:]]Expression[[:space:]]Recognition/1db4c91e-2bd4-4f9d-8af4-ca336460c9da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Multispectral[[:space:]]Demosaicing[[:space:]]via[[:space:]]Dual[[:space:]]Cameras/4d055e3e-fdc1-4647-bb05-7242d4692f72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/MultiverSeg_[[:space:]]Scalable[[:space:]]Interactive[[:space:]]Segmentation[[:space:]]of[[:space:]]Biomedical[[:space:]]Imaging[[:space:]]Datasets[[:space:]]with[[:space:]]In-Context[[:space:]]Guidance/2c0ab0d2-f66c-45ff-8b4d-37232b8e7099_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Music[[:space:]]Grounding[[:space:]]by[[:space:]]Short[[:space:]]Video/10e3550f-6ae6-4ef2-9c3e-e51e71b230ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Music-Aligned[[:space:]]Holistic[[:space:]]3D[[:space:]]Dance[[:space:]]Generation[[:space:]]via[[:space:]]Hierarchical[[:space:]]Motion[[:space:]]Modeling/677bd9da-511f-4059-8f9f-e6eac9d62300_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NAPPure_[[:space:]]Adversarial[[:space:]]Purification[[:space:]]for[[:space:]]Robust[[:space:]]Image[[:space:]]Classification[[:space:]]under[[:space:]]Non-Additive[[:space:]]Perturbations/0e19a363-47a5-4749-8147-1304c8bdcf22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NATRA_[[:space:]]Noise-Agnostic[[:space:]]Framework[[:space:]]for[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]with[[:space:]]Noisy[[:space:]]Observations/95e29682-0349-4530-8ac4-a66a748de34f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NAVER_[[:space:]]A[[:space:]]Neuro-Symbolic[[:space:]]Compositional[[:space:]]Automaton[[:space:]]for[[:space:]]Visual[[:space:]]Grounding[[:space:]]with[[:space:]]Explicit[[:space:]]Logic[[:space:]]Reasoning/0de1b81d-113e-4231-8474-774b8c4fc3f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NETracer_[[:space:]]A[[:space:]]Topology-Aware[[:space:]]Iterative[[:space:]]Tracing[[:space:]]Approach[[:space:]]for[[:space:]]Tubular[[:space:]]Structure[[:space:]]Extraction/43c830b1-72a7-43a3-b1ea-bb86f8c497f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NGD_[[:space:]]Neural[[:space:]]Gradient[[:space:]]Based[[:space:]]Deformation[[:space:]]for[[:space:]]Monocular[[:space:]]Garment[[:space:]]Reconstruction/7a5a5fc5-79de-4745-a5f1-3964018485dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Nautilus_[[:space:]]Locality-aware[[:space:]]Autoencoder[[:space:]]for[[:space:]]Scalable[[:space:]]Mesh[[:space:]]Generation/115c12dd-51f0-4136-b512-4a739b4a5824_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NavMorph_[[:space:]]A[[:space:]]Self-Evolving[[:space:]]World[[:space:]]Model[[:space:]]for[[:space:]]Vision-and-Language[[:space:]]Navigation[[:space:]]in[[:space:]]Continuous[[:space:]]Environments/1bd990ce-9ece-4278-b0cd-15b41cc51e65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NavQ_[[:space:]]Learning[[:space:]]a[[:space:]]Q-Model[[:space:]]for[[:space:]]Foresighted[[:space:]]Vision-and-Language[[:space:]]Navigation/d97afcce-fcb7-42fe-9e21-f549314dfdfb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NeRF[[:space:]]Is[[:space:]]a[[:space:]]Valuable[[:space:]]Assistant[[:space:]]for[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/88028212-6453-4f3c-a1ac-d78cafbe36a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NegRefine_[[:space:]]Refining[[:space:]]Negative[[:space:]]Label-Based[[:space:]]Zero-Shot[[:space:]]OOD[[:space:]]Detection/e14aba48-5e41-459c-8325-1f1ded8d6e22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neighboring[[:space:]]Autoregressive[[:space:]]Modeling[[:space:]]for[[:space:]]Efficient[[:space:]]Visual[[:space:]]Generation/9dd22978-393b-495c-9be7-cdc80db48f2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NeuFrameQ_[[:space:]]Neural[[:space:]]Frame[[:space:]]Fields[[:space:]]for[[:space:]]Scalable[[:space:]]and[[:space:]]Generalizable[[:space:]]Anisotropic[[:space:]]Quadrangulation/052c2fc0-7ac0-4639-a583-8adf64d27f4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NeurOp-Diff_[[:space:]]Continuous[[:space:]]Remote[[:space:]]Sensing[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Neural[[:space:]]Operator[[:space:]]Diffusion/c0dd7f2b-37a0-4335-8f44-dfc2e42761e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NeuraLeaf_[[:space:]]Neural[[:space:]]Parametric[[:space:]]Leaf[[:space:]]Models[[:space:]]with[[:space:]]Shape[[:space:]]and[[:space:]]Deformation[[:space:]]Disentanglement/8671d383-605b-4e2c-b45f-23dec3c044c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neural[[:space:]]Architecture[[:space:]]Search[[:space:]]Driven[[:space:]]by[[:space:]]Locally[[:space:]]Guided[[:space:]]Diffusion[[:space:]]for[[:space:]]Personalized[[:space:]]Federated[[:space:]]Learning/7dbe6317-b829-49c5-b623-4e81dc8db9d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neural[[:space:]]Compression[[:space:]]for[[:space:]]3D[[:space:]]Geometry[[:space:]]Sets/ad1e00d1-6a15-4aaf-a42f-08d88505bea9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neural[[:space:]]Inverse[[:space:]]Rendering[[:space:]]for[[:space:]]High-Accuracy[[:space:]]3D[[:space:]]Measurement[[:space:]]of[[:space:]]Moving[[:space:]]Objects[[:space:]]with[[:space:]]Fewer[[:space:]]Phase-Shifting[[:space:]]Patterns/4ea12dc6-4c7d-4c5f-9eef-d929df96b273_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neural[[:space:]]Multi-View[[:space:]]Self-Calibrated[[:space:]]Photometric[[:space:]]Stereo[[:space:]]without[[:space:]]Photometric[[:space:]]Stereo[[:space:]]Cues/12410a0f-39ed-4fac-a931-41fd7a90af36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neural[[:space:]]Shell[[:space:]]Texture[[:space:]]Splatting_[[:space:]]More[[:space:]]Details[[:space:]]and[[:space:]]Fewer[[:space:]]Primitives/04fe172c-09a4-48d2-a637-e8f7d322688f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neural[[:space:]]Solver[[:space:]]of[[:space:]]Dichromatic[[:space:]]Reflection[[:space:]]Model[[:space:]]for[[:space:]]Specular[[:space:]]Highlight[[:space:]]Removal/7402af7a-ce66-49b6-9147-edae729050a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NeuralSVG_[[:space:]]An[[:space:]]Implicit[[:space:]]Representation[[:space:]]for[[:space:]]Text-to-Vector[[:space:]]Generation/a6866e3a-cec6-4b8a-808d-5a8db7f20712_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neuromanifold-Regularized[[:space:]]KANs[[:space:]]for[[:space:]]Shape-fair[[:space:]]Feature[[:space:]]Representations/d3b2d643-a15d-4913-a6ae-67b81c86e235_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neurons_[[:space:]]Emulating[[:space:]]the[[:space:]]Human[[:space:]]Visual[[:space:]]Cortex[[:space:]]Improves[[:space:]]Fidelity[[:space:]]and[[:space:]]Interpretability[[:space:]]in[[:space:]]fMRI-to-Video[[:space:]]Reconstruction/1a94d6fb-e1ca-4d74-b9b8-9bb3169c9fd7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Neuroverse3D_[[:space:]]Developing[[:space:]]In-Context[[:space:]]Learning[[:space:]]Universal[[:space:]]Model[[:space:]]for[[:space:]]Neuroimaging[[:space:]]in[[:space:]]3D/36ceca9c-ffdc-42a6-ad9c-5a18c4d47de2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/No[[:space:]]More[[:space:]]Sibling[[:space:]]Rivalry_[[:space:]]Debiasing[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection/4c30aca0-d936-45e5-957b-c91b73cd7113_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/No[[:space:]]Pose[[:space:]]at[[:space:]]All_[[:space:]]Self-Supervised[[:space:]]Pose-Free[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Sparse[[:space:]]Views/f8214a40-e553-423a-8e31-e4a3cdc10d95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Noise-Modeled[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Low-Light[[:space:]]Spike[[:space:]]Image[[:space:]]Restoration/ef9eaa1b-3967-44e2-a537-dc0dfa34fe1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Noise2Score3D_[[:space:]]Tweedie's[[:space:]]Approach[[:space:]]for[[:space:]]Unsupervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Denoising/3f7a3e35-5646-4be9-83b2-f202c6d0835c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NoiseController_[[:space:]]Towards[[:space:]]Consistent[[:space:]]Multi-view[[:space:]]Video[[:space:]]Generation[[:space:]]via[[:space:]]Noise[[:space:]]Decomposition[[:space:]]and[[:space:]]Collaboration/80da9020-5b51-4072-ba55-09fc53d54dc8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Normal[[:space:]]and[[:space:]]Abnormal[[:space:]]Pathology[[:space:]]Knowledge-Augmented[[:space:]]Vision-Language[[:space:]]Model[[:space:]]for[[:space:]]Anomaly[[:space:]]Detection[[:space:]]in[[:space:]]Pathology[[:space:]]Images/41398e12-3f00-4625-acd9-ff47ad825799_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NormalCrafter_[[:space:]]Learning[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Normals[[:space:]]from[[:space:]]Video[[:space:]]Diffusion[[:space:]]Priors/1b49986a-f0a1-43a8-beb6-cfa46741d4bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NormalLoc_[[:space:]]Visual[[:space:]]Localization[[:space:]]on[[:space:]]Textureless[[:space:]]3D[[:space:]]Models[[:space:]]using[[:space:]]Surface[[:space:]]Normals/b04cb193-d2f7-4b05-b2dd-28a6ff187b63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Not[[:space:]]All[[:space:]]Degradations[[:space:]]Are[[:space:]]Equal_[[:space:]]A[[:space:]]Targeted[[:space:]]Feature[[:space:]]Denoising[[:space:]]Framework[[:space:]]for[[:space:]]Generalizable[[:space:]]Image[[:space:]]Super-Resolution/0e8bcf79-1996-4058-8bbf-3ae85a41a565_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Not[[:space:]]All[[:space:]]Frame[[:space:]]Features[[:space:]]Are[[:space:]]Equal_[[:space:]]Video-to-4D[[:space:]]Generation[[:space:]]via[[:space:]]Decoupling[[:space:]]Dynamic-Static[[:space:]]Features/df23e57a-52ea-4b42-a993-627bbfb3a2f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Not[[:space:]]Only[[:space:]]Vision_[[:space:]]Evolve[[:space:]]Visual[[:space:]]Speech[[:space:]]Recognition[[:space:]]via[[:space:]]Peripheral[[:space:]]Information/3f569d32-c7e7-4f8c-922d-0bcc3f25f23c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Not[[:space:]]all[[:space:]]Views[[:space:]]are[[:space:]]Created[[:space:]]Equal_[[:space:]]Analyzing[[:space:]]Viewpoint[[:space:]]Instabilities[[:space:]]in[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models/eb7a47f8-d59c-408c-8c79-9c11ee8fdfcd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NuPlanQA_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Multi-View[[:space:]]Driving[[:space:]]Scene[[:space:]]Understanding[[:space:]]in[[:space:]]Multi-Modal[[:space:]]Large[[:space:]]Language[[:space:]]Models/b46218cb-bb98-4ac4-b263-b1e89fd042e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NuiScene_[[:space:]]Exploring[[:space:]]Efficient[[:space:]]Generation[[:space:]]of[[:space:]]Unbounded[[:space:]]Outdoor[[:space:]]Scenes/17f29c15-2cfe-40ce-8c8d-648da3cac6df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/NullSwap_[[:space:]]Proactive[[:space:]]Identity[[:space:]]Cloaking[[:space:]]Against[[:space:]]Deepfake[[:space:]]Face[[:space:]]Swapping/0b608084-3675-4607-8943-da754e90c335_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/O-MaMa_[[:space:]]Learning[[:space:]]Object[[:space:]]Mask[[:space:]]Matching[[:space:]]between[[:space:]]Egocentric[[:space:]]and[[:space:]]Exocentric[[:space:]]Views/d423303c-92a9-468d-8883-5cbdebc66327_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OCK_[[:space:]]Unsupervised[[:space:]]Dynamic[[:space:]]Video[[:space:]]Prediction[[:space:]]with[[:space:]]Object-Centric[[:space:]]Kinematics/7e6bc3e1-3dba-4597-9c32-5e03327b1c8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OCR[[:space:]]Hinders[[:space:]]RAG_[[:space:]]Evaluating[[:space:]]the[[:space:]]Cascading[[:space:]]Impact[[:space:]]of[[:space:]]OCR[[:space:]]on[[:space:]]Retrieval-Augmented[[:space:]]Generation/f6be1d70-029d-4404-9daa-e3d71e95f4a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OCSplats_[[:space:]]Observation[[:space:]]Completeness[[:space:]]Quantification[[:space:]]and[[:space:]]Label[[:space:]]Noise[[:space:]]Separation[[:space:]]in[[:space:]]3DGS/e3d26ef5-1992-48e7-8a50-bcbc19028d89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OD-RASE_[[:space:]]Ontology-Driven[[:space:]]Risk[[:space:]]Assessment[[:space:]]and[[:space:]]Safety[[:space:]]Enhancement[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/0a5eba6e-0d0a-448a-ab61-8885fd634288_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ODDR_[[:space:]]Outlier[[:space:]]Detection[[:space:]]&[[:space:]]Dimension[[:space:]]Reduction[[:space:]]Based[[:space:]]Defense[[:space:]]Against[[:space:]]Adversarial[[:space:]]Patches/fb2e8314-ca11-4103-b2bf-867621abb7a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ODP-Bench_[[:space:]]Benchmarking[[:space:]]Out-of-Distribution[[:space:]]Performance[[:space:]]Prediction/4905d954-700f-4d69-99f8-6934269bf151_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OMNI-DC_[[:space:]]Highly[[:space:]]Robust[[:space:]]Depth[[:space:]]Completion[[:space:]]with[[:space:]]Multiresolution[[:space:]]Depth[[:space:]]Integration/0cab73af-1573-4983-8bce-7d1e2a8356a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ONLY_[[:space:]]One-Layer[[:space:]]Intervention[[:space:]]Sufficiently[[:space:]]Mitigates[[:space:]]Hallucinations[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/3e3fa600-4de7-4dba-bfdc-03e9b57d4309_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ORION_[[:space:]]A[[:space:]]Holistic[[:space:]]End-to-End[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Framework[[:space:]]by[[:space:]]Vision-Language[[:space:]]Instructed[[:space:]]Action[[:space:]]Generation/bf0fa418-06a0-457e-9664-824ea58c4a88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OURO_[[:space:]]A[[:space:]]Self-Bootstrapped[[:space:]]Framework[[:space:]]for[[:space:]]Enhancing[[:space:]]Multimodal[[:space:]]Scene[[:space:]]Understanding/37f70ba8-ec63-47ed-8d32-90a38cb1d6ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OV-SCAN_[[:space:]]Semantically[[:space:]]Consistent[[:space:]]Alignment[[:space:]]for[[:space:]]Novel[[:space:]]Object[[:space:]]Discovery[[:space:]]in[[:space:]]Open-Vocabulary[[:space:]]3D[[:space:]]Object[[:space:]]Detection/f72deea6-33e3-4735-9be6-aeb4747089f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OV3D-CG_[[:space:]]Open-vocabulary[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation[[:space:]]with[[:space:]]Contextual[[:space:]]Guidance/0697bd6d-3ed0-4a2d-8b9b-016beb61b26f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OVA-Fields_[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Open-Vocabulary[[:space:]]Affordance[[:space:]]Fields[[:space:]]for[[:space:]]Robot[[:space:]]Operational[[:space:]]Part[[:space:]]Detection/4a5e3c5a-4482-4044-a131-1607df5665bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OVG-HQ_[[:space:]]Online[[:space:]]Video[[:space:]]Grounding[[:space:]]with[[:space:]]Hybrid-modal[[:space:]]Queries/9c359add-fe04-4d57-9f3c-f3b008a46803_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Oasis_[[:space:]]One[[:space:]]Image[[:space:]]is[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]for[[:space:]]Multimodal[[:space:]]Instruction[[:space:]]Data[[:space:]]Synthesis/11647c48-853c-481b-9533-5cdec21d143a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Object-centric[[:space:]]Video[[:space:]]Question[[:space:]]Answering[[:space:]]with[[:space:]]Visual[[:space:]]Grounding[[:space:]]and[[:space:]]Referring/9a56df3d-2a34-48c2-a76a-019ab23ad8de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Object-level[[:space:]]Correlation[[:space:]]for[[:space:]]Few-Shot[[:space:]]Segmentation/b5fb1f45-ec22-4a32-9539-092bfe988f6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ObjectGS_[[:space:]]Object-aware[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]and[[:space:]]Scene[[:space:]]Understanding[[:space:]]via[[:space:]]Gaussian[[:space:]]Splatting/839ec243-f5d8-48d6-8cdc-469a67e76247_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ObjectMate_[[:space:]]A[[:space:]]Recurrence[[:space:]]Prior[[:space:]]for[[:space:]]Object[[:space:]]Insertion[[:space:]]and[[:space:]]Subject-Driven[[:space:]]Generation/6482161e-5098-49ec-b0dc-f8959caa04b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ObjectRelator_[[:space:]]Enabling[[:space:]]Cross-View[[:space:]]Object[[:space:]]Relation[[:space:]]Understanding[[:space:]]Across[[:space:]]Ego-Centric[[:space:]]and[[:space:]]Exo-Centric[[:space:]]Perspectives/2bcea206-91ff-4a79-abd4-ec062b35cd75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OcRFDet_[[:space:]]Object-Centric[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Multi-View[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/4fa05516-f40e-463f-863b-11469db3bcc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OccluGaussian_[[:space:]]Occlusion-Aware[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Large[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]and[[:space:]]Rendering/bc1f5254-8a63-4266-b0da-d0431a911da1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Occlusion-robust[[:space:]]Stylization[[:space:]]for[[:space:]]Drawing-based[[:space:]]3D[[:space:]]Animation/01c58614-368d-4e46-80c7-02c2e1a3c55d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Occupancy[[:space:]]Learning[[:space:]]with[[:space:]]Spatiotemporal[[:space:]]Memory/a861c93e-eb97-42bd-a3b3-1b81d13e8990_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Omegance_[[:space:]]A[[:space:]]Single[[:space:]]Parameter[[:space:]]for[[:space:]]Various[[:space:]]Granularities[[:space:]]in[[:space:]]Diffusion-Based[[:space:]]Synthesis/c6240e85-7244-4180-ba3f-648e10eb25b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OminiControl_[[:space:]]Minimal[[:space:]]and[[:space:]]Universal[[:space:]]Control[[:space:]]for[[:space:]]Diffusion[[:space:]]Transformer/0efbd492-4dc6-4e12-9807-06d478a205b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Omni-scene[[:space:]]Perception-oriented[[:space:]]Point[[:space:]]Cloud[[:space:]]Geometry[[:space:]]Enhancement[[:space:]]for[[:space:]]Coordinate[[:space:]]Quantization/d644202b-7768-4493-8546-e72b5fa86e27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniCache_[[:space:]]A[[:space:]]Trajectory-Oriented[[:space:]]Global[[:space:]]Perspective[[:space:]]on[[:space:]]Training-Free[[:space:]]Cache[[:space:]]Reuse[[:space:]]for[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]Models/5214e236-2d17-4b8a-81fe-6e961feaad72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniDiff_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Benchmark[[:space:]]for[[:space:]]Fine-grained[[:space:]]Image[[:space:]]Difference[[:space:]]Captioning/559b28c0-1cf8-462a-8f59-78fcf4af2977_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniHuman-1_[[:space:]]Rethinking[[:space:]]the[[:space:]]Scaling-Up[[:space:]]of[[:space:]]One-Stage[[:space:]]Conditioned[[:space:]]Human[[:space:]]Animation[[:space:]]Models/4acdeec6-b69a-445a-a144-1b395822b96d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniPaint_[[:space:]]Mastering[[:space:]]Object-Oriented[[:space:]]Editing[[:space:]]via[[:space:]]Disentangled[[:space:]]Insertion-Removal[[:space:]]Inpainting/61627429-be97-4138-81ce-6a1ef9dd125b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniSAM_[[:space:]]Omnidirectional[[:space:]]Segment[[:space:]]Anything[[:space:]]Model[[:space:]]for[[:space:]]UDA[[:space:]]in[[:space:]]Panoramic[[:space:]]Semantic[[:space:]]Segmentation/a8f5a202-e50a-4caa-88e8-34f2e1a7dd85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OmniVTON_[[:space:]]Training-Free[[:space:]]Universal[[:space:]]Virtual[[:space:]]Try-On/62eda5ff-1ebb-400e-9713-865503937d21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On[[:space:]]Large[[:space:]]Multimodal[[:space:]]Models[[:space:]]as[[:space:]]Open-World[[:space:]]Image[[:space:]]Classifiers/c3eff69e-4f03-499f-9f34-be5015aa22d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On[[:space:]]the[[:space:]]Complexity-Faithfulness[[:space:]]Trade-off[[:space:]]of[[:space:]]Gradient-Based[[:space:]]Explanations/58ec82af-91cd-49cd-84da-41ec495566d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On[[:space:]]the[[:space:]]Generalization[[:space:]]of[[:space:]]Representation[[:space:]]Uncertainty[[:space:]]in[[:space:]]Earth[[:space:]]Observation/97c71b54-7ccc-40de-8091-3c206eff9a15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On[[:space:]]the[[:space:]]Provable[[:space:]]Importance[[:space:]]of[[:space:]]Gradients[[:space:]]for[[:space:]]Autonomous[[:space:]]Language-Assisted[[:space:]]Image[[:space:]]Clustering/9677d514-0eff-47ef-b2ac-73de1d613205_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On[[:space:]]the[[:space:]]Recovery[[:space:]]of[[:space:]]Cameras[[:space:]]from[[:space:]]Fundamental[[:space:]]Matrices/376c1e2f-61ce-43aa-8e96-1d8826067bef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On[[:space:]]the[[:space:]]Robustness[[:space:]]Tradeoff[[:space:]]in[[:space:]]Fine-Tuning/14e025a8-8324-4104-9781-f17ac621bcd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/On-Device[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]Policy[[:space:]]for[[:space:]]Efficient[[:space:]]Robot[[:space:]]Manipulation/0f24ec75-064b-436a-94f5-6a3d2094df9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One[[:space:]]Encoder[[:space:]]to[[:space:]]Rule[[:space:]]them[[:space:]]All_[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Model-free[[:space:]]Visual[[:space:]]Reinforcement[[:space:]]Learning[[:space:]]using[[:space:]]Fourier[[:space:]]Neural[[:space:]]Operators/b2f71d6a-f484-4bd1-94c9-1984c3d33756_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One[[:space:]]Last[[:space:]]Attention[[:space:]]for[[:space:]]Your[[:space:]]Vision-Language[[:space:]]Model/bdec3967-6429-468d-aa1e-73cf30365260_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One[[:space:]]Look[[:space:]]is[[:space:]]Enough_[[:space:]]Seamless[[:space:]]Patchwise[[:space:]]Refinement[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation[[:space:]]on[[:space:]]High-Resolution[[:space:]]Images/7ea58908-1a89-4fa6-b216-4e24940d158c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One[[:space:]]Object,[[:space:]]Multiple[[:space:]]Lies_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Cross-task[[:space:]]Adversarial[[:space:]]Attack[[:space:]]on[[:space:]]Unified[[:space:]]Vision-Language[[:space:]]Models/8ad51869-e58d-4a22-b107-ee707c02160b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One[[:space:]]Perturbation[[:space:]]is[[:space:]]Enough_[[:space:]]On[[:space:]]Generating[[:space:]]Universal[[:space:]]Adversarial[[:space:]]Perturbations[[:space:]]against[[:space:]]Vision-Language[[:space:]]Pre-training[[:space:]]Models/d9960338-4524-4ecd-a24f-8d20831fcf25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One[[:space:]]Polyp[[:space:]]Identifies[[:space:]]All_[[:space:]]One-Shot[[:space:]]Polyp[[:space:]]Segmentation[[:space:]]with[[:space:]]SAM[[:space:]]via[[:space:]]Cascaded[[:space:]]Priors[[:space:]]and[[:space:]]Iterative[[:space:]]Prompt[[:space:]]Evolution/a49579c5-8277-4e31-ac3c-6dd29bff12da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One[[:space:]]Trajectory,[[:space:]]One[[:space:]]Token_[[:space:]]Grounded[[:space:]]Video[[:space:]]Tokenization[[:space:]]via[[:space:]]Panoptic[[:space:]]Sub-object[[:space:]]Trajectory/4240b1f4-c37b-4e42-913c-297b7b077ef4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One-Shot[[:space:]]Knowledge[[:space:]]Transfer[[:space:]]for[[:space:]]Scalable[[:space:]]Person[[:space:]]Re-Identification/b18d6df3-a2dc-48db-a65f-af6713a41aa3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/One-Step[[:space:]]Specular[[:space:]]Highlight[[:space:]]Removal[[:space:]]with[[:space:]]Adapted[[:space:]]Diffusion[[:space:]]Models/00a9e9da-ef04-44ff-85c3-fae2ebdd4270_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OneGT_[[:space:]]One-Shot[[:space:]]Geometry-Texture[[:space:]]Neural[[:space:]]Rendering[[:space:]]for[[:space:]]Head[[:space:]]Avatars/8c1c0900-5b57-4438-8591-d00eaa0b8f3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Online[[:space:]]Dense[[:space:]]Point[[:space:]]Tracking[[:space:]]with[[:space:]]Streaming[[:space:]]Memory/b77372b1-220c-48a4-a1f7-03897c9e46bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Online[[:space:]]Generic[[:space:]]Event[[:space:]]Boundary[[:space:]]Detection/2c46d6f2-3fe8-4cbb-a760-a1384ac1d9e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Online[[:space:]]Language[[:space:]]Splatting/d3074254-d39f-4c83-a0b1-695ff3d065fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Online[[:space:]]Reasoning[[:space:]]Video[[:space:]]Segmentation[[:space:]]with[[:space:]]Just-in-Time[[:space:]]Digital[[:space:]]Twins/7de60053-07b8-4739-8339-ce7478544dac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Open-Unfairness[[:space:]]Adversarial[[:space:]]Mitigation[[:space:]]for[[:space:]]Generalized[[:space:]]Deepfake[[:space:]]Detection/b5957af4-a0ec-4002-9d0a-69065a237ec7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Open-Vocabulary[[:space:]]HOI[[:space:]]Detection[[:space:]]with[[:space:]]Interaction-aware[[:space:]]Prompt[[:space:]]and[[:space:]]Concept[[:space:]]Calibration/d364b960-6ae5-4574-ae75-c00d4acb5276_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Open-Vocabulary[[:space:]]Octree-Graph[[:space:]]for[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/01692f59-d598-49f8-a78f-da19a0dbd270_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Open-World[[:space:]]Skill[[:space:]]Discovery[[:space:]]from[[:space:]]Unsegmented[[:space:]]Demonstration[[:space:]]Videos/05f92ab7-d34a-4eab-b01d-01b6c42abb83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Open-ended[[:space:]]Hierarchical[[:space:]]Streaming[[:space:]]Video[[:space:]]Understanding[[:space:]]with[[:space:]]Vision[[:space:]]Language[[:space:]]Models/6b4ef31c-dc3f-43e7-b4c5-5736df81dff7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Open-set[[:space:]]Cross[[:space:]]Modal[[:space:]]Generalization[[:space:]]via[[:space:]]Multimodal[[:space:]]Unified[[:space:]]Representation/8fb89dae-a82d-44d6-b570-b41113ffa9f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OpenAnimals_[[:space:]]Revisiting[[:space:]]Person[[:space:]]Re-Identification[[:space:]]for[[:space:]]Animals[[:space:]]Towards[[:space:]]Better[[:space:]]Generalization/1e76ba30-0e24-4324-8bd1-dadfbc44faa2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OpenM3D_[[:space:]]Open[[:space:]]Vocabulary[[:space:]]Multi-view[[:space:]]Indoor[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]without[[:space:]]Human[[:space:]]Annotations/f809f56d-9654-4b55-9a43-c30e85820fd7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OpenRSD_[[:space:]]Towards[[:space:]]Open-prompts[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]Remote[[:space:]]Sensing[[:space:]]Images/838f0069-39e0-4609-a847-c733fffbf6a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OpenSubstance_[[:space:]]A[[:space:]]High-quality[[:space:]]Measured[[:space:]]Dataset[[:space:]]of[[:space:]]Multi-View[[:space:]]and[[:space:]]-Lighting[[:space:]]Images[[:space:]]and[[:space:]]Shapes/4e50d233-752b-4b0c-8e18-8a37b3b8b592_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OpenVision_[[:space:]]A[[:space:]]Fully-Open,[[:space:]]Cost-Effective[[:space:]]Family[[:space:]]of[[:space:]]Advanced[[:space:]]Vision[[:space:]]Encoders[[:space:]]for[[:space:]]Multimodal[[:space:]]Learning/5df37b49-916e-4142-b7d0-177f8d292ca2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OphCLIP_[[:space:]]Hierarchical[[:space:]]Retrieval-Augmented[[:space:]]Learning[[:space:]]for[[:space:]]Ophthalmic[[:space:]]Surgical[[:space:]]Video-Language[[:space:]]Pretraining/1ce8df48-7194-4ba8-bdea-72a545ff9a93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Optical[[:space:]]Model-Driven[[:space:]]Sharpness[[:space:]]Mapping[[:space:]]for[[:space:]]Autofocus[[:space:]]in[[:space:]]Small[[:space:]]Depth-of-Field[[:space:]]and[[:space:]]Severe[[:space:]]Defocus[[:space:]]Scenarios/18045a0e-9d47-4a84-b381-a7c5cc7a3de6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Optimal[[:space:]]Transport[[:space:]]for[[:space:]]Brain-Image[[:space:]]Alignment_[[:space:]]Unveiling[[:space:]]Redundancy[[:space:]]and[[:space:]]Synergy[[:space:]]in[[:space:]]Neural[[:space:]]Information[[:space:]]Processing/f7346415-79ce-4a30-8dc2-276363bff01b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OracleFusion_[[:space:]]Assisting[[:space:]]the[[:space:]]Decipherment[[:space:]]of[[:space:]]Oracle[[:space:]]Bone[[:space:]]Script[[:space:]]with[[:space:]]Structurally[[:space:]]Constrained[[:space:]]Semantic[[:space:]]Typography/dadfca99-2fe2-4fc4-afa7-2bb186e4860c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Orchid_[[:space:]]Image[[:space:]]Latent[[:space:]]Diffusion[[:space:]]for[[:space:]]Joint[[:space:]]Appearance[[:space:]]and[[:space:]]Geometry[[:space:]]Generation/5f89e0d9-0928-43ec-85ca-d7356a6c26ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OrderChain_[[:space:]]Towards[[:space:]]General[[:space:]]Instruct-Tuning[[:space:]]for[[:space:]]Stimulating[[:space:]]the[[:space:]]Ordinal[[:space:]]Understanding[[:space:]]Ability[[:space:]]of[[:space:]]MLLM/46d644d6-8708-4f76-8918-90a517db7194_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/OuroMamba_[[:space:]]A[[:space:]]Data-Free[[:space:]]Quantization[[:space:]]Framework[[:space:]]for[[:space:]]Vision[[:space:]]Mamba/2509e19a-7894-47d8-9bab-62cc4b17e170_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ouroboros_[[:space:]]Single-step[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Cycle-consistent[[:space:]]Forward[[:space:]]and[[:space:]]Inverse[[:space:]]Rendering/f8268f4c-0808-4117-802b-da9904583b3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Outdoor[[:space:]]Monocular[[:space:]]SLAM[[:space:]]with[[:space:]]Global[[:space:]]Scale-Consistent[[:space:]]3D[[:space:]]Gaussian[[:space:]]Pointmaps/f8984141-1b23-4311-90fb-d5826d818385_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Outlier-Aware[[:space:]]Post-Training[[:space:]]Quantization[[:space:]]for[[:space:]]Image[[:space:]]Super-Resolution/71c6002b-ff5e-472b-b341-4e45838643ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Overcoming[[:space:]]Dual[[:space:]]Drift[[:space:]]for[[:space:]]Continual[[:space:]]Long-Tailed[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/952fbd02-7d32-42ee-84cb-987d9ab127ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PAN-Crafter_[[:space:]]Learning[[:space:]]Modality-Consistent[[:space:]]Alignment[[:space:]]for[[:space:]]PAN-Sharpening/573a969d-172d-49d4-b34c-009cbd8c880c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PARTE_[[:space:]]Part-Guided[[:space:]]Texturing[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Reconstruction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/f03f2b8b-96e8-47b0-ad6f-ef2173142800_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PASD_[[:space:]]A[[:space:]]Pixel-Adaptive[[:space:]]Swarm[[:space:]]Dynamics[[:space:]]Approach[[:space:]]for[[:space:]]Unsupervised[[:space:]]Low-Light[[:space:]]Image[[:space:]]Enhancement/1b0a6d28-5568-430d-831e-25d18926ac4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PASG_[[:space:]]A[[:space:]]Closed-Loop[[:space:]]Framework[[:space:]]for[[:space:]]Automated[[:space:]]Geometric[[:space:]]Primitive[[:space:]]Extraction[[:space:]]and[[:space:]]Semantic[[:space:]]Anchoring[[:space:]]in[[:space:]]Robotic[[:space:]]Manipulation/5d43c587-cba7-4bca-a658-7f402c039bc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PASTA_[[:space:]]Part-Aware[[:space:]]Sketch-to-3D[[:space:]]Shape[[:space:]]Generation[[:space:]]with[[:space:]]Text-Aligned[[:space:]]Prior/969000c7-3137-4e2b-b18b-47a227821ec1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PBCAT_[[:space:]]Patch-Based[[:space:]]Composite[[:space:]]Adversarial[[:space:]]Training[[:space:]]against[[:space:]]Physically[[:space:]]Realizable[[:space:]]Attacks[[:space:]]on[[:space:]]Object[[:space:]]Detection/1a500e3f-634c-494e-a8ed-a7d2e6c6a740_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PBFG_[[:space:]]A[[:space:]]New[[:space:]]Physically-Based[[:space:]]Dataset[[:space:]]and[[:space:]]Removal[[:space:]]of[[:space:]]Lens[[:space:]]Flares[[:space:]]and[[:space:]]Glares/ee110c4e-c396-429a-a25a-68f69f21d1f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PCR-GS_[[:space:]]COLMAP-Free[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]via[[:space:]]Pose[[:space:]]Co-Regularizations/077e7048-601e-43c9-babc-489bc68a0027_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PEFTDiff_[[:space:]]Diffusion-Guided[[:space:]]Transferability[[:space:]]Estimation[[:space:]]for[[:space:]]Parameter-Efficient[[:space:]]Fine-Tuning/4c751e07-02a7-48a7-85e8-1dc4dc080625_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PERSONA_[[:space:]]Personalized[[:space:]]Whole-Body[[:space:]]3D[[:space:]]Avatar[[:space:]]with[[:space:]]Pose-Driven[[:space:]]Deformations[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/42344d95-9067-44a2-9af1-0be75cfa9327_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PHATNet_[[:space:]]A[[:space:]]Physics-guided[[:space:]]Haze[[:space:]]Transfer[[:space:]]Network[[:space:]]for[[:space:]]Domain-adaptive[[:space:]]Real-world[[:space:]]Image[[:space:]]Dehazing/965912d0-5b80-4b26-80fd-6a4448aff31e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PHD_[[:space:]]Personalized[[:space:]]3D[[:space:]]Human[[:space:]]Body[[:space:]]Fitting[[:space:]]with[[:space:]]Point[[:space:]]Diffusion/ce2a73d2-ac02-4530-8c48-0554935a4edf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PINO_[[:space:]]Person-Interaction[[:space:]]Noise[[:space:]]Optimization[[:space:]]for[[:space:]]Long-Duration[[:space:]]and[[:space:]]Customizable[[:space:]]Motion[[:space:]]Generation[[:space:]]of[[:space:]]Arbitrary-Sized[[:space:]]Groups/ae8230a6-84c3-4768-aa6c-ed4be6cd1dbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PLADIS_[[:space:]]Pushing[[:space:]]the[[:space:]]Limits[[:space:]]of[[:space:]]Attention[[:space:]]in[[:space:]]Diffusion[[:space:]]Models[[:space:]]at[[:space:]]Inference[[:space:]]Time[[:space:]]by[[:space:]]Leveraging[[:space:]]Sparsity/82c8e735-d2bf-4341-8332-5fad2c311d34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PLAN_[[:space:]]Proactive[[:space:]]Low-Rank[[:space:]]Allocation[[:space:]]for[[:space:]]Continual[[:space:]]Learning/01d232a7-2f59-4513-9309-faa11095f020_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PLA_[[:space:]]Prompt[[:space:]]Learning[[:space:]]Attack[[:space:]]against[[:space:]]Text-to-Image[[:space:]]Generative[[:space:]]Models/a7a206be-4736-4314-b417-4d689a64d159_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PLMP[[:space:]]-[[:space:]]Point-Line[[:space:]]Minimal[[:space:]]Problems[[:space:]]for[[:space:]]Projective[[:space:]]SfM/0d7fb231-624d-466b-808c-15e345b42b80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/POMATO_[[:space:]]Marrying[[:space:]]Pointmap[[:space:]]Matching[[:space:]]with[[:space:]]Temporal[[:space:]]Motions[[:space:]]for[[:space:]]Dynamic[[:space:]]3D[[:space:]]Reconstruction/1de0f91e-0150-4eab-a1fc-f5394801fd7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PRE-Mamba_[[:space:]]A[[:space:]]4D[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]for[[:space:]]Ultra-High-Frequent[[:space:]]Event[[:space:]]Camera[[:space:]]Deraining/521e29cf-6292-4b2a-8333-77ba2b903010_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PRIMAL_[[:space:]]Physically[[:space:]]Reactive[[:space:]]and[[:space:]]Interactive[[:space:]]Motor[[:space:]]Model[[:space:]]for[[:space:]]Avatar[[:space:]]Learning/fd4810f3-d689-4b89-b41b-90b3d1713c37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PRISM_[[:space:]]Reducing[[:space:]]Spurious[[:space:]]Implicit[[:space:]]Biases[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]LLM-Guided[[:space:]]Embedding[[:space:]]Projection/38f64838-a08c-4e4f-9654-04db04907cf0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PRM_[[:space:]]Photometric[[:space:]]Stereo[[:space:]]based[[:space:]]Large[[:space:]]Reconstruction[[:space:]]Model/376be5fb-cff2-44aa-87ae-beb1972ff68e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PRO-VPT_[[:space:]]Distribution-Adaptive[[:space:]]Visual[[:space:]]Prompt[[:space:]]Tuning[[:space:]]via[[:space:]]Prompt[[:space:]]Relocation/705f3e43-5f3f-4db6-89b2-0f1e38fc75e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PROGRESSOR_[[:space:]]A[[:space:]]Perceptually[[:space:]]Guided[[:space:]]Reward[[:space:]]Estimator[[:space:]]with[[:space:]]Self-Supervised[[:space:]]Online[[:space:]]Refinement/91086143-b5ba-4d82-87ab-d2e978da69f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PROL[[:space:]]_[[:space:]]Rehearsal[[:space:]]Free[[:space:]]Continual[[:space:]]Learning[[:space:]]in[[:space:]]Streaming[[:space:]]Data[[:space:]]via[[:space:]]Prompt[[:space:]]Online[[:space:]]Learning/d5f28d0d-4814-43b7-b4fc-d4fe29db4270_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PRVQL_[[:space:]]Progressive[[:space:]]Knowledge-guided[[:space:]]Refinement[[:space:]]for[[:space:]]Robust[[:space:]]Egocentric[[:space:]]Visual[[:space:]]Query[[:space:]]Localization/eb4759bd-cd46-42ae-a79f-87abd8db733e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PS-Mamba_[[:space:]]Spatial-Temporal[[:space:]]Graph[[:space:]]Mamba[[:space:]]for[[:space:]]Pose[[:space:]]Sequence[[:space:]]Refinement/2871c7f3-bcf2-48bb-a7f2-2d3979537c0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PS3_[[:space:]]A[[:space:]]Multimodal[[:space:]]Transformer[[:space:]]Integrating[[:space:]]Pathology[[:space:]]Reports[[:space:]]with[[:space:]]Histology[[:space:]]Images[[:space:]]and[[:space:]]Biological[[:space:]]Pathways[[:space:]]for[[:space:]]Cancer[[:space:]]Survival[[:space:]]Prediction/c9473abf-829d-47e9-a4dc-4c6013a76985_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PUMA_[[:space:]]Empowering[[:space:]]Unified[[:space:]]MLLM[[:space:]]with[[:space:]]Multi-granular[[:space:]]Visual[[:space:]]Generation/dcbe3e75-7b75-4b39-a12d-7aa36192d75e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PUMPS_[[:space:]]Skeleton-Agnostic[[:space:]]Point-based[[:space:]]Universal[[:space:]]Motion[[:space:]]Pre-Training[[:space:]]for[[:space:]]Synthesis[[:space:]]in[[:space:]]Human[[:space:]]Motion[[:space:]]Tasks/fa71b7dc-88d2-4d0a-b22f-6c63c5815e10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PVChat_[[:space:]]Personalized[[:space:]]Video[[:space:]]Chat[[:space:]]with[[:space:]]One-Shot[[:space:]]Learning/01ecb771-7a93-4ff6-b013-cfe8214300e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PVMamba_[[:space:]]Parallelizing[[:space:]]Vision[[:space:]]Mamba[[:space:]]via[[:space:]]Dynamic[[:space:]]State[[:space:]]Aggregation/b5acf676-d45e-478c-a205-37aafaf668da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PacGDC_[[:space:]]Label-Efficient[[:space:]]Generalizable[[:space:]]Depth[[:space:]]Completion[[:space:]]with[[:space:]]Projection[[:space:]]Ambiguity[[:space:]]and[[:space:]]Consistency/ae091d21-e39c-4bb1-bafa-5b8385ef73b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PanSt3R_[[:space:]]Multi-view[[:space:]]Consistent[[:space:]]Panoptic[[:space:]]Segmentation/a8dbe7c6-678d-44db-9657-79126771b75d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PanoLlama_[[:space:]]Generating[[:space:]]Endless[[:space:]]and[[:space:]]Coherent[[:space:]]Panoramas[[:space:]]with[[:space:]]Next-Token-Prediction[[:space:]]LLMs/7707fe69-4333-4323-9b1b-dc73ef886b5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PanoSplatt3R_[[:space:]]Leveraging[[:space:]]Perspective[[:space:]]Pretraining[[:space:]]for[[:space:]]Generalized[[:space:]]Unposed[[:space:]]Wide-Baseline[[:space:]]Panorama[[:space:]]Reconstruction/7d77afb7-3af8-4e70-97e8-dbc9cbf9bf81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Parameter-Efficient[[:space:]]Adaptation[[:space:]]of[[:space:]]Geospatial[[:space:]]Foundation[[:space:]]Models[[:space:]]through[[:space:]]Embedding[[:space:]]Deflection/4c675caa-d670-45a7-b5d5-2ad0188d6aeb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Parametric[[:space:]]Shadow[[:space:]]Control[[:space:]]for[[:space:]]Portrait[[:space:]]Generation[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/2f48f70e-ba7a-4b9b-a87a-aea20d846493_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PartField_[[:space:]]Learning[[:space:]]3D[[:space:]]Feature[[:space:]]Fields[[:space:]]for[[:space:]]Part[[:space:]]Segmentation[[:space:]]and[[:space:]]Beyond/319a69e8-211c-4099-ba4b-de064076be93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Partial[[:space:]]Forward[[:space:]]Blocking_[[:space:]]A[[:space:]]Novel[[:space:]]Data[[:space:]]Pruning[[:space:]]Paradigm[[:space:]]for[[:space:]]Lossless[[:space:]]Training[[:space:]]Acceleration/be2749cf-5871-470d-8c0a-db1272a3f609_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Partially[[:space:]]Matching[[:space:]]Submap[[:space:]]Helps_[[:space:]]Uncertainty[[:space:]]Modeling[[:space:]]and[[:space:]]Propagation[[:space:]]for[[:space:]]Text[[:space:]]to[[:space:]]Point[[:space:]]Cloud[[:space:]]Localization/146939b4-3bd1-45d3-b182-76e07e5e9e72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Passing[[:space:]]the[[:space:]]Driving[[:space:]]Knowledge[[:space:]]Test/38c53bea-e034-4bc8-a30f-2579326f191f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PatchScaler_[[:space:]]An[[:space:]]Efficient[[:space:]]Patch-Independent[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Image[[:space:]]Super-Resolution/93068f4c-f9db-464c-8bf3-024519f81d42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PathDiff_[[:space:]]Histopathology[[:space:]]Image[[:space:]]Synthesis[[:space:]]with[[:space:]]Unpaired[[:space:]]Text[[:space:]]and[[:space:]]Mask[[:space:]]Conditions/a41053a9-695c-45e8-b14d-349d91a2b039_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PathFinder_[[:space:]]A[[:space:]]Multi-Modal[[:space:]]Multi-Agent[[:space:]]System[[:space:]]for[[:space:]]Medical[[:space:]]Diagnostic[[:space:]]Decision-Making[[:space:]]Applied[[:space:]]to[[:space:]]Histopathology/73f052f3-693e-4b56-a9dc-b0211cf780a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Penalizing[[:space:]]Boundary[[:space:]]Activation[[:space:]]for[[:space:]]Object[[:space:]]Completeness[[:space:]]in[[:space:]]Diffusion[[:space:]]Models/655710c5-5ba6-48ea-a3de-a47fdb24473c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PerLDiff_[[:space:]]Controllable[[:space:]]Street[[:space:]]View[[:space:]]Synthesis[[:space:]]Using[[:space:]]Perspective-Layout[[:space:]]Diffusion[[:space:]]Model/4ffeba8b-41b6-4979-bdb4-21a2a8c0d351_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Perceive,[[:space:]]Understand[[:space:]]and[[:space:]]Restore_[[:space:]]Real-World[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]with[[:space:]]Autoregressive[[:space:]]Multimodal[[:space:]]Generative[[:space:]]Models/0f9ad6ad-e028-45bd-aae2-b58156738109_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Perceiving[[:space:]]and[[:space:]]Acting[[:space:]]in[[:space:]]First-Person_[[:space:]]A[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Egocentric[[:space:]]Human-Object-Human[[:space:]]Interactions/74e7b017-0cac-465c-886d-2c0991d82c47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Perception-as-Control_[[:space:]]Fine-grained[[:space:]]Controllable[[:space:]]Image[[:space:]]Animation[[:space:]]with[[:space:]]3D-aware[[:space:]]Motion[[:space:]]Representation/129c7bb5-45e8-4053-9383-9c38cbe63c5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Performing[[:space:]]Defocus[[:space:]]Deblurring[[:space:]]by[[:space:]]Modeling[[:space:]]its[[:space:]]Formation[[:space:]]Process/6a0f8ef8-50b8-4666-b55d-2d663d01f372_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PersPose_[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]Perspective[[:space:]]Encoding[[:space:]]and[[:space:]]Perspective[[:space:]]Rotation/5179649c-dd8e-4215-915e-721ff69c77ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PersonaCraft_[[:space:]]Personalized[[:space:]]and[[:space:]]Controllable[[:space:]]Full-Body[[:space:]]Multi-Human[[:space:]]Scene[[:space:]]Generation[[:space:]]Using[[:space:]]Occlusion-Aware[[:space:]]3D-Conditioned[[:space:]]Diffusion/b404a20a-223e-44ca-b6de-abe2b2a0740f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PersonalVideo_[[:space:]]High[[:space:]]ID-Fidelity[[:space:]]Video[[:space:]]Customization[[:space:]]without[[:space:]]Dynamic[[:space:]]and[[:space:]]Semantic[[:space:]]Degradation/6e540f26-560c-42d5-9428-011850311b1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Personalized[[:space:]]Federated[[:space:]]Learning[[:space:]]under[[:space:]]Local[[:space:]]Supervision/adb9235d-9a6c-4297-8c88-2b108f5dd2ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Perspective-Aware[[:space:]]Reasoning[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models[[:space:]]via[[:space:]]Mental[[:space:]]Imagery[[:space:]]Simulation/bc096b51-9096-4b8d-9ef4-1ecf094d0c77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Perspective-Aware[[:space:]]Teaching_[[:space:]]Adapting[[:space:]]Knowledge[[:space:]]for[[:space:]]Heterogeneous[[:space:]]Distillation/85c97e47-b15b-4355-9620-8be8513c9bf3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Perspective-Invariant[[:space:]]3D[[:space:]]Object[[:space:]]Detection/09da7f61-85a0-4476-b242-45d80ccfaa7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Perspective-aware[[:space:]]3D[[:space:]]Gaussian[[:space:]]Inpainting[[:space:]]with[[:space:]]Multi-view[[:space:]]Consistency/fe327956-1334-4511-bb1f-0307a9fecb0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ph-GAN_[[:space:]]Physics-Inspired[[:space:]]GAN[[:space:]]for[[:space:]]Generating[[:space:]]SAR[[:space:]]Images[[:space:]]Under[[:space:]]Limited[[:space:]]Data/bcd4cb4c-0b19-4378-8d0c-b95650e9eae7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Phantom_[[:space:]]Subject-Consistent[[:space:]]Video[[:space:]]Generation[[:space:]]via[[:space:]]Cross-Modal[[:space:]]Alignment/7aa26770-9621-40b5-b823-6de4eec087b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Photolithography[[:space:]]Overlay[[:space:]]Map[[:space:]]Generation[[:space:]]with[[:space:]]Implicit[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]Diffusion[[:space:]]Transformer/021073bc-0729-454b-b8f0-162b36b15d06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PhysRig_[[:space:]]Differentiable[[:space:]]Physics-Based[[:space:]]Skinning[[:space:]]and[[:space:]]Rigging[[:space:]]Framework[[:space:]]for[[:space:]]Realistic[[:space:]]Articulated[[:space:]]Object[[:space:]]Modeling/d5b4d5d7-6701-44b0-81f5-cd17d917cc0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PhysSplat_[[:space:]]Efficient[[:space:]]Physics[[:space:]]Simulation[[:space:]]for[[:space:]]3D[[:space:]]Scenes[[:space:]]via[[:space:]]MLLM-Guided[[:space:]]Gaussian[[:space:]]Splatting/c1f747b8-0c38-41cf-936f-13957227386a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PhysTwin_[[:space:]]Physics-Informed[[:space:]]Reconstruction[[:space:]]and[[:space:]]Simulation[[:space:]]of[[:space:]]Deformable[[:space:]]Objects[[:space:]]from[[:space:]]Videos/161f8a4d-daff-491e-a783-e7502e51931e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Physical[[:space:]]Degradation[[:space:]]Model-Guided[[:space:]]Interferometric[[:space:]]Hyperspectral[[:space:]]Reconstruction[[:space:]]with[[:space:]]Unfolding[[:space:]]Transformer/bf52801d-b78c-44e2-9e3a-f114d393d3e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Physics[[:space:]]Context[[:space:]]Builders_[[:space:]]A[[:space:]]Modular[[:space:]]Framework[[:space:]]for[[:space:]]Physical[[:space:]]Reasoning[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models/1fdfbc73-ca29-438d-b54a-d57e4710d398_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pi-GPS_[[:space:]]Enhancing[[:space:]]Geometry[[:space:]]Problem[[:space:]]Solving[[:space:]]by[[:space:]]Unleashing[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]Diagrammatic[[:space:]]Information/ca10b010-7560-4f0c-b8a3-2e2abdfd1503_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pinco_[[:space:]]Position-induced[[:space:]]Consistent[[:space:]]Adapter[[:space:]]for[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]in[[:space:]]Foreground-conditioned[[:space:]]Inpainting/6c38c1f9-d72a-440f-9d21-962b925a10d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PixTalk_[[:space:]]Controlling[[:space:]]Photorealistic[[:space:]]Image[[:space:]]Processing[[:space:]]and[[:space:]]Editing[[:space:]]with[[:space:]]Language/f8bc49a9-9e50-4575-8fd1-1c8c1ff1e2d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PixelStitch_[[:space:]]Structure-Preserving[[:space:]]Pixel-Wise[[:space:]]Bidirectional[[:space:]]Warps[[:space:]]for[[:space:]]Unsupervised[[:space:]]Image[[:space:]]Stitching/56471d07-cbd9-4fcc-8c67-b25f23885dee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PlaceIt3D_[[:space:]]Language-Guided[[:space:]]Object[[:space:]]Placement[[:space:]]in[[:space:]]Real[[:space:]]3D[[:space:]]Scenes/060d8314-3d2f-436d-b5f8-f32596d33859_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PlanGen_[[:space:]]Towards[[:space:]]Unified[[:space:]]Layout[[:space:]]Planning[[:space:]]and[[:space:]]Image[[:space:]]Generation[[:space:]]in[[:space:]]Auto-Regressive[[:space:]]Vision[[:space:]]Language[[:space:]]Models/f0ef0250-a754-449d-ac8e-de4e6ba4730c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Planar[[:space:]]Affine[[:space:]]Rectification[[:space:]]from[[:space:]]Local[[:space:]]Change[[:space:]]of[[:space:]]Scale[[:space:]]and[[:space:]]Orientation/ed6f73e1-72bc-43cc-bdd1-260737ecd60e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PlaneRAS_[[:space:]]Learning[[:space:]]Planar[[:space:]]Primitives[[:space:]]for[[:space:]]3D[[:space:]]Plane[[:space:]]Recovery/cef23686-644c-4d7b-863c-4d15976bf7e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Player-Centric[[:space:]]Multimodal[[:space:]]Prompt[[:space:]]Generation[[:space:]]for[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]Based[[:space:]]Identity-Aware[[:space:]]Basketball[[:space:]]Video[[:space:]]Captioning/01356fcc-483e-4949-9d0b-512986ef7af9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Plug-in[[:space:]]Feedback[[:space:]]Self-adaptive[[:space:]]Attention[[:space:]]in[[:space:]]CLIP[[:space:]]for[[:space:]]Training-free[[:space:]]Open-Vocabulary[[:space:]]Segmentation/44993567-0e9f-469b-8652-ce615ac68e01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PlugMark_[[:space:]]A[[:space:]]Plug-in[[:space:]]Zero-Watermarking[[:space:]]Framework[[:space:]]for[[:space:]]Diffusion[[:space:]]Models/0f214347-aff7-4bba-aae3-4ac1e9f7d47e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Point[[:space:]]Cloud[[:space:]]Self-supervised[[:space:]]Learning[[:space:]]via[[:space:]]3D[[:space:]]to[[:space:]]Multi-view[[:space:]]Masked[[:space:]]Learner/78eb5049-b4a6-48f9-90cd-05ec145869a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PointGAC_[[:space:]]Geometric-Aware[[:space:]]Codebook[[:space:]]for[[:space:]]Masked[[:space:]]Point[[:space:]]Modeling/256dab7d-197a-4a6e-a321-fe186b38ed9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PolGS_[[:space:]]Polarimetric[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Fast[[:space:]]Reflective[[:space:]]Surface[[:space:]]Reconstruction/960a376e-8681-4b7c-a7b6-adc05fed6ce7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PolarAnything_[[:space:]]Diffusion-based[[:space:]]Polarimetric[[:space:]]Image[[:space:]]Synthesis/7c17f0fb-413c-45b9-8b18-627fe88c21bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Polarimetric[[:space:]]Neural[[:space:]]Field[[:space:]]via[[:space:]]Unified[[:space:]]Complex-Valued[[:space:]]Wave[[:space:]]Representation/fca7a2d2-aaf9-40d0-b320-6575c46582cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ponimator_[[:space:]]Unfolding[[:space:]]Interactive[[:space:]]Pose[[:space:]]for[[:space:]]Versatile[[:space:]]Human-human[[:space:]]Interaction[[:space:]]Animation/02a78d17-ae86-44ba-825f-ad29044d63f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pose-Star_[[:space:]]Anatomy-Aware[[:space:]]Editing[[:space:]]for[[:space:]]Open-World[[:space:]]Fashion[[:space:]]Images/b89530bd-c937-48d3-8224-17332a894190_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PoseAnchor_[[:space:]]Robust[[:space:]]Root[[:space:]]Position[[:space:]]Estimation[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/f7679b11-9a46-4209-bdc8-8e542549b6fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PoseSyn_[[:space:]]Synthesizing[[:space:]]Diverse[[:space:]]3D[[:space:]]Pose[[:space:]]Data[[:space:]]from[[:space:]]In-the-Wild[[:space:]]2D[[:space:]]Data/b8ad1800-c09f-4bff-af4f-6e03c438e4a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PossLoss_[[:space:]]A[[:space:]]Reliable[[:space:]]and[[:space:]]Sensitive[[:space:]]Facial[[:space:]]Landmark[[:space:]]Detection[[:space:]]Loss[[:space:]]Function/dc34f475-fdab-4aee-8bd0-b120c7fe35a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Power[[:space:]]of[[:space:]]Cooperative[[:space:]]Supervision_[[:space:]]Multiple[[:space:]]Teachers[[:space:]]Framework[[:space:]]for[[:space:]]Advanced[[:space:]]3D[[:space:]]Semi-Supervised[[:space:]]Object[[:space:]]Detection/e7919580-4573-4fdb-aa8f-3ebf71e4a077_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Preacher_[[:space:]]Paper-to-Video[[:space:]]Agentic[[:space:]]System/3a51a60f-d8d8-49c7-b3c7-ae3a8b2f76ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Precise[[:space:]]Action-to-Video[[:space:]]Generation[[:space:]]Through[[:space:]]Visual[[:space:]]Action[[:space:]]Prompts/0bb33e0c-8791-4c9e-8bc0-5a04b2868bac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Predict-Optimize-Distill_[[:space:]]A[[:space:]]Self-Improving[[:space:]]Cycle[[:space:]]for[[:space:]]4D[[:space:]]Object[[:space:]]Understanding/f6353c48-e018-46a5-ae6c-8d9febab84ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Preserve[[:space:]]Anything_[[:space:]]Controllable[[:space:]]Image[[:space:]]Synthesis[[:space:]]with[[:space:]]Object[[:space:]]Preservation/9e888117-4aba-4a0e-81ef-1f0abcd103b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pretend[[:space:]]Benign_[[:space:]]A[[:space:]]Stealthy[[:space:]]Adversarial[[:space:]]Attack[[:space:]]by[[:space:]]Exploiting[[:space:]]Vulnerabilities[[:space:]]in[[:space:]]Cooperative[[:space:]]Perception/a361ffc8-7e04-41ae-8a31-f08e610ebbf5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pretrained[[:space:]]Reversible[[:space:]]Generation[[:space:]]as[[:space:]]Unsupervised[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/35f47a2e-af74-4186-ac3f-c3ac7ae7c2aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PriOr-Flow_[[:space:]]Enhancing[[:space:]]Primitive[[:space:]]Panoramic[[:space:]]Optical[[:space:]]Flow[[:space:]]with[[:space:]]Orthogonal[[:space:]]View/7175a2cc-d14e-421b-a526-96814cbe28a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PrimHOI_[[:space:]]Compositional[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]via[[:space:]]Reusable[[:space:]]Primitives/8a4c1133-1d96-48d4-9f83-e884b59471a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Princeton365_[[:space:]]A[[:space:]]Diverse[[:space:]]Dataset[[:space:]]with[[:space:]]Accurate[[:space:]]Camera[[:space:]]Pose/a729437b-aab0-43ba-9005-6ec634f481e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Principal[[:space:]]Components'[[:space:]]Enable[[:space:]]A[[:space:]]New[[:space:]]Language[[:space:]]of[[:space:]]Images/9970a29e-4738-417a-a9f9-e51287d10ca3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Principles[[:space:]]of[[:space:]]Visual[[:space:]]Tokens[[:space:]]for[[:space:]]Efficient[[:space:]]Video[[:space:]]Understanding/b9d020d2-93ab-4210-a8b8-81e64cfdd00c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prior-aware[[:space:]]Dynamic[[:space:]]Temporal[[:space:]]Modeling[[:space:]]Framework[[:space:]]for[[:space:]]Sequential[[:space:]]3D[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation/ec4c5f64-dd89-4b61-8d6c-891cd1a7e709_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prior2Former[[:space:]]-[[:space:]]Evidential[[:space:]]Modeling[[:space:]]of[[:space:]]Mask[[:space:]]Transformers[[:space:]]for[[:space:]]Assumption-Free[[:space:]]Open-World[[:space:]]Panoptic[[:space:]]Segmentation/b43bab6a-0b9a-4f3b-887c-ac197cd9ea49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PriorMotion_[[:space:]]Generative[[:space:]]Class-Agnostic[[:space:]]Motion[[:space:]]Prediction[[:space:]]with[[:space:]]Raster-Vector[[:space:]]Motion[[:space:]]Field[[:space:]]Priors/ed995fb3-b838-480f-be55-ccd8199a6800_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Privacy-centric[[:space:]]Deep[[:space:]]Motion[[:space:]]Retargeting[[:space:]]for[[:space:]]Anonymization[[:space:]]of[[:space:]]Skeleton-Based[[:space:]]Motion[[:space:]]Visualization/3773194d-a3bc-4782-b062-9628d619a829_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProGait_[[:space:]]A[[:space:]]Multi-Purpose[[:space:]]Video[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Transfemoral[[:space:]]Prosthesis[[:space:]]Users/a781c349-00ca-4d48-b665-158864febbe6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProJudge_[[:space:]]A[[:space:]]Multi-Modal[[:space:]]Multi-Discipline[[:space:]]Benchmark[[:space:]]and[[:space:]]Instruction-Tuning[[:space:]]Dataset[[:space:]]for[[:space:]]MLLM-based[[:space:]]Process[[:space:]]Judges/63702c5d-af55-47cc-988b-130ea642e17e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProSAM_[[:space:]]Enhancing[[:space:]]the[[:space:]]Robustness[[:space:]]of[[:space:]]SAM-based[[:space:]]Visual[[:space:]]Reference[[:space:]]Segmentation[[:space:]]with[[:space:]]Probabilistic[[:space:]]Prompts/4aca1ac0-75ff-4ad0-a5b6-2c22f7f80435_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Proactive[[:space:]]Scene[[:space:]]Decomposition[[:space:]]and[[:space:]]Reconstruction/4695e6b5-0aad-408f-b841-9ca84f161496_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProbMED_[[:space:]]A[[:space:]]Probabilistic[[:space:]]Framework[[:space:]]for[[:space:]]Medical[[:space:]]Multimodal[[:space:]]Binding/9fab1f6d-cca0-4c03-9394-87453bab99c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ProbRes_[[:space:]]Probabilistic[[:space:]]Jump[[:space:]]Diffusion[[:space:]]for[[:space:]]Open-World[[:space:]]Egocentric[[:space:]]Activity[[:space:]]Recognition/cfb3a0c2-f2d1-4101-9caa-5f91cca075cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Probabilistic[[:space:]]Inertial[[:space:]]Poser[[:space:]](ProbIP)_[[:space:]]Uncertainty-aware[[:space:]]Human[[:space:]]Motion[[:space:]]Modeling[[:space:]]from[[:space:]]Sparse[[:space:]]Inertial[[:space:]]Sensors/853c3321-a55a-4d9e-871a-4b9780db529a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Probabilistic[[:space:]]Prototype[[:space:]]Calibration[[:space:]]of[[:space:]]Vision-language[[:space:]]Models[[:space:]]for[[:space:]]Generalized[[:space:]]Few-shot[[:space:]]Semantic[[:space:]]Segmentation/8e105bc0-ca7b-4e67-a045-4398187ab03e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Processing[[:space:]]and[[:space:]]acquisition[[:space:]]traces[[:space:]]in[[:space:]]visual[[:space:]]encoders_[[:space:]]What[[:space:]]does[[:space:]]CLIP[[:space:]]know[[:space:]]about[[:space:]]your[[:space:]]camera_/2087b26c-22cb-49d8-94ee-fb95f0610c00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Progressive[[:space:]]Artwork[[:space:]]Outpainting[[:space:]]via[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models/076da141-4e72-4496-b4de-752cff3fb57a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Progressive[[:space:]]Distribution[[:space:]]Bridging_[[:space:]]Unsupervised[[:space:]]Adaptation[[:space:]]for[[:space:]]Large-scale[[:space:]]Pre-trained[[:space:]]Models[[:space:]]via[[:space:]]Adaptive[[:space:]]Auxiliary[[:space:]]Data/7eaf5dfe-a2ab-4f4e-a3d7-2b3e53d08314_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Progressive[[:space:]]Growing[[:space:]]of[[:space:]]Video[[:space:]]Tokenizers[[:space:]]for[[:space:]]Temporally[[:space:]]Compact[[:space:]]Latent[[:space:]]Spaces/335a379d-336b-48aa-a490-d148b61348d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Progressive[[:space:]]Homeostatic[[:space:]]and[[:space:]]Plastic[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Audio-Visual[[:space:]]Multi-Task[[:space:]]Incremental[[:space:]]Learning/162f2f5e-53be-483c-825d-8846a7c5fe1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Progressive[[:space:]]Test[[:space:]]Time[[:space:]]Energy[[:space:]]Adaptation[[:space:]]for[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/f1ee5fa8-d412-410f-94a0-7bfec3fcb8fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prompt[[:space:]]Guidance[[:space:]]and[[:space:]]Human[[:space:]]Proximal[[:space:]]Perception[[:space:]]for[[:space:]]HOT[[:space:]]Prediction[[:space:]]with[[:space:]]Regional[[:space:]]Joint[[:space:]]Loss/c7ad6979-6ed4-401b-80d5-ede69dddc5e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prompt-A-Video_[[:space:]]Prompt[[:space:]]Your[[:space:]]Video[[:space:]]Diffusion[[:space:]]Model[[:space:]]via[[:space:]]Preference-Aligned[[:space:]]LLM/927979c2-1c7d-4dcc-9ab7-15ff60978d0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prompt-driven[[:space:]]Transferable[[:space:]]Adversarial[[:space:]]Attack[[:space:]]on[[:space:]]Person[[:space:]]Re-Identification[[:space:]]with[[:space:]]Attribute-aware[[:space:]]Textual[[:space:]]Inversion/809738e1-2de1-4b27-bf04-7a2d46b9fe3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PromptDresser_[[:space:]]Improving[[:space:]]the[[:space:]]Quality[[:space:]]and[[:space:]]Controllability[[:space:]]of[[:space:]]Virtual[[:space:]]Try-On[[:space:]]via[[:space:]]Generative[[:space:]]Textual[[:space:]]Prompt[[:space:]]and[[:space:]]Prompt-aware[[:space:]]Mask/155dbdc6-42f8-4985-a140-c57049224a2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PropVG_[[:space:]]End-to-End[[:space:]]Proposal-Driven[[:space:]]Visual[[:space:]]Grounding[[:space:]]with[[:space:]]Multi-Granularity[[:space:]]Discrimination/98bab36e-ef91-4e4a-b112-bcf18b85e6ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prototype[[:space:]]Guided[[:space:]]Backdoor[[:space:]]Defense[[:space:]]via[[:space:]]Activation[[:space:]]Space[[:space:]]Manipulation/6dc2f893-7cbe-4330-a91c-fb6566f9020d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prototype-based[[:space:]]Contrastive[[:space:]]Learning[[:space:]]with[[:space:]]Stage-wise[[:space:]]Progressive[[:space:]]Augmentation[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Fine-Grained[[:space:]]Learning/5e7347d0-296f-4bc4-8225-debd60c96cec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Prototypes[[:space:]]are[[:space:]]Balanced[[:space:]]Units[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Effective[[:space:]]Partially[[:space:]]Relevant[[:space:]]Video[[:space:]]Retrieval/4f74cf9a-daf3-4a80-a712-7390913508c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Proxy-Bridged[[:space:]]Game[[:space:]]Transformer[[:space:]]for[[:space:]]Interactive[[:space:]]Extreme[[:space:]]Motion[[:space:]]Prediction/b953217b-d42c-4992-bd75-8c6210252ec0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pruning[[:space:]]All-Rounder_[[:space:]]Rethinking[[:space:]]and[[:space:]]Improving[[:space:]]Inference[[:space:]]Efficiency[[:space:]]for[[:space:]]Large[[:space:]]Vision[[:space:]]Language[[:space:]]Models/42a8d929-f53e-46a8-8186-c98c66dac97e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Pseudo-SD_[[:space:]]Pseudo[[:space:]]Controlled[[:space:]]Stable[[:space:]]Diffusion[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]and[[:space:]]Cross-Domain[[:space:]]Semantic[[:space:]]Segmentation/92c35442-b9fa-49e7-8027-49ad952cc7d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/PseudoMapTrainer_[[:space:]]Learning[[:space:]]Online[[:space:]]Mapping[[:space:]]without[[:space:]]HD[[:space:]]Maps/8460a162-f70b-4f48-b1d8-5673c525501a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Punching[[:space:]]Bag[[:space:]]vs.[[:space:]]Punching[[:space:]]Person_[[:space:]]Motion[[:space:]]Transferability[[:space:]]in[[:space:]]Videos/f78bf67d-6ba0-47d3-b924-650c57732265_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Puppet-Master_[[:space:]]Scaling[[:space:]]Interactive[[:space:]]Video[[:space:]]Generation[[:space:]]as[[:space:]]a[[:space:]]Motion[[:space:]]Prior[[:space:]]for[[:space:]]Part-Level[[:space:]]Dynamics/8a598ce5-ad3b-4420-8752-1941a27e71cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Purge-Gate_[[:space:]]Backpropagation-Free[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Point[[:space:]]Clouds[[:space:]]Classification[[:space:]]via[[:space:]]Token[[:space:]]purging/2429cca1-72c2-4c19-80bf-1716a5e84783_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Puzzle[[:space:]]Similarity_[[:space:]]A[[:space:]]Perceptually-guided[[:space:]]Cross-Reference[[:space:]]Metric[[:space:]]for[[:space:]]Artifact[[:space:]]Detection[[:space:]]in[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstructions/8da81524-0c7b-4cbb-81b4-318511362686_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Q-Frame_[[:space:]]Query-aware[[:space:]]Frame[[:space:]]Selection[[:space:]]and[[:space:]]Multi-Resolution[[:space:]]Adaptation[[:space:]]for[[:space:]]Video-LLMs/ab9a082a-1aa4-49c1-98fc-b024f141388a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Q-Norm_[[:space:]]Robust[[:space:]]Representation[[:space:]]Learning[[:space:]]via[[:space:]]Quality-Adaptive[[:space:]]Normalization/1576f49a-5601-427c-922f-b011709a52ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/QK-Edit_[[:space:]]Revisiting[[:space:]]Attention-based[[:space:]]Injection[[:space:]]in[[:space:]]MM-DiT[[:space:]]for[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Editing/91167f28-5944-489e-9c2c-7e5bb22f802d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/QR-LoRA_[[:space:]]Efficient[[:space:]]and[[:space:]]Disentangled[[:space:]]Fine-tuning[[:space:]]via[[:space:]]QR[[:space:]]Decomposition[[:space:]]for[[:space:]]Customized[[:space:]]Generation/ccada14c-8b66-401c-9ad7-1d8078de1b4c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/QuEST_[[:space:]]Low-bit[[:space:]]Diffusion[[:space:]]Model[[:space:]]Quantization[[:space:]]via[[:space:]]Efficient[[:space:]]Selective[[:space:]]Finetuning/38ba22b3-df29-4713-a428-1c3c71ce2b21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Quadratic[[:space:]]Gaussian[[:space:]]Splatting_[[:space:]]High[[:space:]]Quality[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]with[[:space:]]Second-order[[:space:]]Geometric[[:space:]]Primitives/2f602314-e15a-4fc7-939a-4b708ae2c21e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/QuantCache_[[:space:]]Adaptive[[:space:]]Importance-Guided[[:space:]]Quantization[[:space:]]with[[:space:]]Hierarchical[[:space:]]Latent[[:space:]]and[[:space:]]Layer[[:space:]]Caching[[:space:]]for[[:space:]]Video[[:space:]]Generation/6f7dad70-735e-4044-9864-b43ab499ba5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Quanta[[:space:]]Neural[[:space:]]Networks_[[:space:]]From[[:space:]]Photons[[:space:]]to[[:space:]]Perception/9eecf24f-6d2a-4417-88e0-f9ab971fbcce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Quantifying[[:space:]]and[[:space:]]Narrowing[[:space:]]the[[:space:]]Unknown_[[:space:]]Interactive[[:space:]]Text-to-Video[[:space:]]Retrieval[[:space:]]via[[:space:]]Uncertainty[[:space:]]Minimization/177cbc23-13b0-499e-8c2b-2d53003174fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/QuickSplat_[[:space:]]Fast[[:space:]]3D[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]via[[:space:]]Learned[[:space:]]Gaussian[[:space:]]Initialization/4f2624f6-40d3-46f2-9ae3-7887ba9e63ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/R-LiViT_[[:space:]]A[[:space:]]LiDAR-Visual-Thermal[[:space:]]Dataset[[:space:]]Enabling[[:space:]]Vulnerable[[:space:]]Road[[:space:]]User[[:space:]]Focused[[:space:]]Roadside[[:space:]]Perception/bf911c4f-1d0d-486a-80a9-1fff23731816_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/R1-Onevision_[[:space:]]Advancing[[:space:]]Generalized[[:space:]]Multimodal[[:space:]]Reasoning[[:space:]]through[[:space:]]Cross-Modal[[:space:]]Formalization/475ec1f3-e3db-4386-86c1-e4af56b1643d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/R1-VL_[[:space:]]Learning[[:space:]]to[[:space:]]Reason[[:space:]]with[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]via[[:space:]]Step-wise[[:space:]]Group[[:space:]]Relative[[:space:]]Policy[[:space:]]Optimization/e7a0ccdd-b0cc-4399-bb9b-7bd8e9beac35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RA-BUSSeg_[[:space:]]Relation-aware[[:space:]]Semi-supervised[[:space:]]Breast[[:space:]]Ultrasound[[:space:]]Image[[:space:]]Segmentation[[:space:]]via[[:space:]]Adjacent[[:space:]]Propagation[[:space:]]and[[:space:]]Cross-layer[[:space:]]Alignment/547a2e55-6b71-4048-91a3-d261e7b97d6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RAGD_[[:space:]]Regional-Aware[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/777252c2-8dce-4808-a8d6-efab81de4547_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RAGDiffusion_[[:space:]]Faithful[[:space:]]Cloth[[:space:]]Generation[[:space:]]via[[:space:]]External[[:space:]]Knowledge[[:space:]]Assimilation/e475c992-d95f-4c85-b7c9-c82b0c87a8ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RAGNet_[[:space:]]Large-scale[[:space:]]Reasoning-based[[:space:]]Affordance[[:space:]]Segmentation[[:space:]]Benchmark[[:space:]]towards[[:space:]]General[[:space:]]Grasping/f597b9c9-b0eb-4e05-8a43-19e44509d067_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RALoc_[[:space:]]Enhancing[[:space:]]Outdoor[[:space:]]LiDAR[[:space:]]Localization[[:space:]]via[[:space:]]Rotation[[:space:]]Awareness/a502ea9f-2ea1-4759-90cc-38d35aa41279_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RANKCLIP_[[:space:]]Ranking-Consistent[[:space:]]Language-Image[[:space:]]Pretraining/36b702cc-9b8c-4901-b825-63e6ad466de8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RARE_[[:space:]]Refine[[:space:]]Any[[:space:]]Registration[[:space:]]of[[:space:]]Pairwise[[:space:]]Point[[:space:]]Clouds[[:space:]]via[[:space:]]Zero-Shot[[:space:]]Learning/31084f33-ade9-4a22-b199-dbcd5f86db88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RCTDistill_[[:space:]]Cross-Modal[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]Framework[[:space:]]for[[:space:]]Radar-Camera[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Temporal[[:space:]]Fusion/c3344904-7a3d-450c-9b49-0bfc0eef5237_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/REDUCIO![[:space:]]Generating[[:space:]]1K[[:space:]]Video[[:space:]]within[[:space:]]16[[:space:]]Seconds[[:space:]]using[[:space:]]Extremely[[:space:]]Compressed[[:space:]]Motion[[:space:]]Latents/184a3ddf-9191-495c-ac71-0c98fabd0f16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/REGEN_[[:space:]]Learning[[:space:]]Compact[[:space:]]Video[[:space:]]Embedding[[:space:]]with[[:space:]](Re-)Generative[[:space:]]Decoder/fc895970-d957-4614-a966-4a644fd27e5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/REPA-E_[[:space:]]Unlocking[[:space:]]VAE[[:space:]]for[[:space:]]End-to-End[[:space:]]Tuning[[:space:]]of[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Transformers/04874538-2450-4e14-8844-b1d1f738bbdb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/REPARO_[[:space:]]Compositional[[:space:]]3D[[:space:]]Assets[[:space:]]Generation[[:space:]]with[[:space:]]Differentiable[[:space:]]3D[[:space:]]Layout[[:space:]]Alignment/e303d607-a2aa-4ba0-8faf-b6ee8aef2213_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RESCUE_[[:space:]]Crowd[[:space:]]Evacuation[[:space:]]Simulation[[:space:]]via[[:space:]]Controlling[[:space:]]SDM-United[[:space:]]Characters/38368614-6c60-4c3b-89b7-3b5161f0d296_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RGE-GS_[[:space:]]Reward-Guided[[:space:]]Expansive[[:space:]]Driving[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]via[[:space:]]Diffusion[[:space:]]Priors/f5c0c5e2-fe9c-4ffa-a82c-0d11d9a5d550_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RI3D_[[:space:]]Few-Shot[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]With[[:space:]]Repair[[:space:]]and[[:space:]]Inpainting[[:space:]]Diffusion[[:space:]]Priors/f87eaa72-b153-4f71-96eb-cd933a44db5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RIOcc_[[:space:]]Efficient[[:space:]]Cross-Modal[[:space:]]Fusion[[:space:]]Transformer[[:space:]]with[[:space:]]Collaborative[[:space:]]Feature[[:space:]]Refinement[[:space:]]for[[:space:]]3D[[:space:]]Semantic[[:space:]]Occupancy[[:space:]]Prediction/3e6e9afd-08c2-45b5-bccd-cd0e0a6e9794_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RIPE_[[:space:]]Reinforcement[[:space:]]Learning[[:space:]]on[[:space:]]Unlabeled[[:space:]]Image[[:space:]]Pairs[[:space:]]for[[:space:]]Robust[[:space:]]Keypoint[[:space:]]Extraction/bb17e26a-dee4-4a8f-a256-4affea9ba7e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RMultiplex200K_[[:space:]]Toward[[:space:]]Reliable[[:space:]]Multimodal[[:space:]]Process[[:space:]]Supervision[[:space:]]for[[:space:]]Visual[[:space:]]Language[[:space:]]Models[[:space:]]on[[:space:]]Telecommunications/21de1f49-8d33-4331-9b98-fe11d063adec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ROADWork_[[:space:]]A[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Learning[[:space:]]to[[:space:]]Recognize,[[:space:]]Observe,[[:space:]]Analyze[[:space:]]and[[:space:]]Drive[[:space:]]Through[[:space:]]Work[[:space:]]Zones/a282fced-716e-43b8-bc5e-05fba052ae84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ROAR_[[:space:]]Reducing[[:space:]]Inversion[[:space:]]Error[[:space:]]in[[:space:]]Generative[[:space:]]Image[[:space:]]Watermarking/967e3628-761e-4c34-b1f5-4f310b06c46e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ROVI_[[:space:]]A[[:space:]]VLM-LLM[[:space:]]Re-Captioned[[:space:]]Dataset[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Instance-Grounded[[:space:]]Text-to-Image[[:space:]]Generation/44a1ed72-9329-4443-a25d-f09953e202c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RS-vHeat_[[:space:]]Heat[[:space:]]Conduction[[:space:]]Guided[[:space:]]Efficient[[:space:]]Remote[[:space:]]Sensing[[:space:]]Foundation[[:space:]]Model/94cd09f3-cf1f-46f6-98bc-304743ee47c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RTMap_[[:space:]]Real-Time[[:space:]]Recursive[[:space:]]Mapping[[:space:]]with[[:space:]]Change[[:space:]]Detection[[:space:]]and[[:space:]]Localization/c637999d-0f0d-4030-97b8-2a89b2be133d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RadGPT_[[:space:]]Constructing[[:space:]]3D[[:space:]]Image-Text[[:space:]]Tumor[[:space:]]Datasets/8d30b116-8006-4dfe-9551-2c5c1f0a9398_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RadarSplat_[[:space:]]Radar[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Data[[:space:]]Synthesis[[:space:]]and[[:space:]]3D[[:space:]]Reconstruction[[:space:]]of[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Scenes/0e2a74b3-6b72-46b1-9a54-5ca1e9e4663e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Radiant[[:space:]]Foam_[[:space:]]Real-Time[[:space:]]Differentiable[[:space:]]Ray[[:space:]]Tracing/b4ba8099-5baa-4d40-b7b9-e018713360cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RainbowPrompt_[[:space:]]Diversity-Enhanced[[:space:]]Prompt-Evolving[[:space:]]for[[:space:]]Continual[[:space:]]Learning/a0c0b5bc-92d8-4fb7-af17-7e01c3190672_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Randomized[[:space:]]Autoregressive[[:space:]]Visual[[:space:]]Generation/7a271cf3-67f5-4d9e-8304-b922e236ce90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RapVerse_[[:space:]]Coherent[[:space:]]Vocals[[:space:]]and[[:space:]]Whole-Body[[:space:]]Motion[[:space:]]Generation[[:space:]]from[[:space:]]Text/83029d55-f117-48b0-b2a1-d5c53fa1a6f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RareCLIP_[[:space:]]Rarity-aware[[:space:]]Online[[:space:]]Zero-shot[[:space:]]Industrial[[:space:]]Anomaly[[:space:]]Detection/c89271c8-200b-4b69-9134-2b2e40fd2420_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RayGaussX_[[:space:]]Accelerating[[:space:]]Gaussian-Based[[:space:]]Ray[[:space:]]Marching[[:space:]]for[[:space:]]Real-Time[[:space:]]and[[:space:]]High-Quality[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/f1d600ef-cd99-42cb-9996-11692aa89243_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RayPose_[[:space:]]Ray[[:space:]]Bundling[[:space:]]Diffusion[[:space:]]for[[:space:]]Template[[:space:]]Views[[:space:]]in[[:space:]]Unseen[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/8e133c3d-e963-4ff1-a4d4-46823ed648d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RayZer_[[:space:]]A[[:space:]]Self-supervised[[:space:]]Large[[:space:]]View[[:space:]]Synthesis[[:space:]]Model/6c54b274-5971-48c3-9462-c7f7ddb43594_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RayletDF_[[:space:]]Raylet[[:space:]]Distance[[:space:]]Fields[[:space:]]for[[:space:]]Generalizable[[:space:]]3D[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]from[[:space:]]Point[[:space:]]Clouds[[:space:]]or[[:space:]]Gaussians/6772646b-a84b-4214-8cc8-8d4a062ae1b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReAL-AD_[[:space:]]Towards[[:space:]]Human-Like[[:space:]]Reasoning[[:space:]]in[[:space:]]End-to-End[[:space:]]Autonomous[[:space:]]Driving/0ac3ee88-2827-4a77-b75c-935c9749c9d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReCamMaster_[[:space:]]Camera-Controlled[[:space:]]Generative[[:space:]]Rendering[[:space:]]from[[:space:]]A[[:space:]]Single[[:space:]]Video/1a50cae9-992f-4437-a481-423135a77892_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReCoT_[[:space:]]Reflective[[:space:]]Self-Correction[[:space:]]Training[[:space:]]for[[:space:]]Mitigating[[:space:]]Confirmation[[:space:]]Bias[[:space:]]in[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models/a371b18d-1a91-428e-9232-7291cc721803_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReFlex_[[:space:]]Text-Guided[[:space:]]Editing[[:space:]]of[[:space:]]Real[[:space:]]Images[[:space:]]in[[:space:]]Rectified[[:space:]]Flow[[:space:]]via[[:space:]]Mid-Step[[:space:]]Feature[[:space:]]Extraction[[:space:]]and[[:space:]]Attention[[:space:]]Adaptation/2530398c-cd9d-4f17-964e-d2a853017942_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReME_[[:space:]]A[[:space:]]Data-Centric[[:space:]]Framework[[:space:]]for[[:space:]]Training-Free[[:space:]]Open-Vocabulary[[:space:]]Segmentation/3032e5d9-b631-4933-a4e0-7d455ca4d844_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReMP-AD_[[:space:]]Retrieval-enhanced[[:space:]]Multi-modal[[:space:]]Prompt[[:space:]]Fusion[[:space:]]for[[:space:]]Few-Shot[[:space:]]Industrial[[:space:]]Visual[[:space:]]Anomaly[[:space:]]Detection/bcba052f-fc64-4717-8010-aec1a7ee0c99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RePoseD_[[:space:]]Efficient[[:space:]]Relative[[:space:]]Pose[[:space:]]Estimation[[:space:]]With[[:space:]]Known[[:space:]]Depth[[:space:]]Information/40353887-4f73-4a0a-b981-5f923a53c68c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReTracker_[[:space:]]Exploring[[:space:]]Image[[:space:]]Matching[[:space:]]for[[:space:]]Robust[[:space:]]Online[[:space:]]Any[[:space:]]Point[[:space:]]Tracking/42118481-e852-4ebc-8501-2c156127aa2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Real3D_[[:space:]]Towards[[:space:]]Scaling[[:space:]]Large[[:space:]]Reconstruction[[:space:]]Models[[:space:]]with[[:space:]]Real[[:space:]]Images/f297df5f-7239-483e-af8f-937fc815c2f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RealCam-I2V_[[:space:]]Real-World[[:space:]]Image-to-Video[[:space:]]Generation[[:space:]]with[[:space:]]Interactive[[:space:]]Complex[[:space:]]Camera[[:space:]]Control/b63fbb92-0f5b-4445-8e06-8024ee574ac9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RealGeneral_[[:space:]]Unifying[[:space:]]Visual[[:space:]]Generation[[:space:]]via[[:space:]]Temporal[[:space:]]In-Context[[:space:]]Learning[[:space:]]with[[:space:]]Video[[:space:]]Models/c08ceab1-72c4-4e59-ba66-a43eddf28bb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reangle-A-Video_[[:space:]]4D[[:space:]]Video[[:space:]]Generation[[:space:]]as[[:space:]]Video-to-Video[[:space:]]Translation/642933ff-f262-4046-94fb-8c22ab2e3a58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReasonVQA_[[:space:]]A[[:space:]]Multi-hop[[:space:]]Reasoning[[:space:]]Benchmark[[:space:]]with[[:space:]]Structural[[:space:]]Knowledge[[:space:]]for[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/68918c4c-e885-4f4a-8dd7-71574a6a8b2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReassembleNet_[[:space:]]Learnable[[:space:]]Keypoints[[:space:]]and[[:space:]]Diffusion[[:space:]]for[[:space:]]2D[[:space:]]Fresco[[:space:]]Reconstruction/cb93cbc0-8a2b-4340-917d-d03dcb8e957f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Recognizing[[:space:]]Actions[[:space:]]from[[:space:]]Robotic[[:space:]]View[[:space:]]for[[:space:]]Natural[[:space:]]Human-Robot[[:space:]]Interaction/c5c90ef8-4e0a-470c-b38e-81e675e75ccc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReconDreamer++_[[:space:]]Harmonizing[[:space:]]Generative[[:space:]]and[[:space:]]Reconstructive[[:space:]]Models[[:space:]]for[[:space:]]Driving[[:space:]]Scene[[:space:]]Representation/062120a6-c966-496f-b5f3-707befbc6aaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Recover[[:space:]]Biological[[:space:]]Structure[[:space:]]from[[:space:]]Sparse-View[[:space:]]Diffraction[[:space:]]Images[[:space:]]with[[:space:]]Neural[[:space:]]Volumetric[[:space:]]Prior/b9ad909b-077c-4950-9bf7-53f3aac3aea6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Recovering[[:space:]]Parametric[[:space:]]Scenes[[:space:]]from[[:space:]]Very[[:space:]]Few[[:space:]]Time-of-Flight[[:space:]]Pixels/0d6707dd-8891-4089-87e1-808c4b20461d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rectifying[[:space:]]Magnitude[[:space:]]Neglect[[:space:]]in[[:space:]]Linear[[:space:]]Attention/e6100218-5fd4-4b92-a47e-baa795e53261_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reducing[[:space:]]Unimodal[[:space:]]Bias[[:space:]]in[[:space:]]Multi-Modal[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Multi-Scale[[:space:]]Functional[[:space:]]Entropy[[:space:]]Regularization/3f13f2ff-b34f-4073-be92-b7b0aa04657f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RefEdit_[[:space:]]A[[:space:]]Benchmark[[:space:]]and[[:space:]]Method[[:space:]]for[[:space:]]Improving[[:space:]]Instruction-based[[:space:]]Image[[:space:]]Editing[[:space:]]Model[[:space:]]on[[:space:]]Referring[[:space:]]Expressions/43bba521-e7da-412c-921a-566f4f8a1b53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Refer[[:space:]]to[[:space:]]Any[[:space:]]Segmentation[[:space:]]Mask[[:space:]]Group[[:space:]]With[[:space:]]Vision-Language[[:space:]]Prompts/980e7edd-bb53-44da-b915-887c539e6e01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReferDINO_[[:space:]]Referring[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation[[:space:]]with[[:space:]]Visual[[:space:]]Grounding[[:space:]]Foundations/20afbaf7-7893-4bfa-8b37-12b93f2016db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ReferEverything_[[:space:]]Towards[[:space:]]Segmenting[[:space:]]Everything[[:space:]]We[[:space:]]Can[[:space:]]Speak[[:space:]]of[[:space:]]in[[:space:]]Videos/f9eb3d3c-0357-4236-b785-119874e96abc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reference-based[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Image-based[[:space:]]Retrieval-Augmented[[:space:]]Generation[[:space:]]Diffusion/31188455-5d8d-4a90-8d1d-7770f1fdc725_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Referring[[:space:]]Expression[[:space:]]Comprehension[[:space:]]for[[:space:]]Small[[:space:]]Objects/9584ea89-42d5-450f-a1e2-eadc02ff019e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Referring[[:space:]]to[[:space:]]Any[[:space:]]Person/4946b02a-8894-40fb-97d2-3ddacd05e326_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reflect-DiT_[[:space:]]Inference-Time[[:space:]]Scaling[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Transformers[[:space:]]via[[:space:]]In-Context[[:space:]]Reflection/9bbce2c6-ee4f-46f6-a472-68b3931775ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RegGS_[[:space:]]Unposed[[:space:]]Sparse[[:space:]]Views[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]3DGS[[:space:]]Registration/24b277a5-3aec-4109-bd1f-58400c9ae49b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Region-Level[[:space:]]Data[[:space:]]Attribution[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generative[[:space:]]Models/05cd208e-da19-490c-bc06-67d7c367247a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Region-aware[[:space:]]Anchoring[[:space:]]Mechanism[[:space:]]for[[:space:]]Efficient[[:space:]]Referring[[:space:]]Visual[[:space:]]Grounding/41104eb8-8ebc-4a9f-a681-719616904b3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Region-based[[:space:]]Cluster[[:space:]]Discrimination[[:space:]]for[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/bd971583-854b-47cd-a1ce-a89140bf18b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Registration[[:space:]]beyond[[:space:]]Points_[[:space:]]General[[:space:]]Affine[[:space:]]Subspace[[:space:]]Alignment[[:space:]]via[[:space:]]Geodesic[[:space:]]Distance[[:space:]]on[[:space:]]Grassmann[[:space:]]Manifold/67b37a2a-6dc3-4edd-b50d-042d5da0f838_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reinforcement[[:space:]]Learning-Guided[[:space:]]Data[[:space:]]Selection[[:space:]]via[[:space:]]Redundancy[[:space:]]Assessment/5e32e63b-dbc8-4070-bbce-290817dadc09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Relative[[:space:]]Illumination[[:space:]]Fields_[[:space:]]Learning[[:space:]]Medium[[:space:]]and[[:space:]]Light[[:space:]]Independent[[:space:]]Underwater[[:space:]]Scenes/f2bff9df-c77d-4cb4-ae23-5f43b4bc3c16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reminiscence[[:space:]]Attack[[:space:]]on[[:space:]]Residuals_[[:space:]]Exploiting[[:space:]]Approximate[[:space:]]Machine[[:space:]]Unlearning[[:space:]]for[[:space:]]Privacy/ade4c134-2f62-448c-9914-d17757c580f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Removing[[:space:]]Cost[[:space:]]Volumes[[:space:]]from[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimators/5c1d73ab-8b89-40e5-a254-14c7d18d5363_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Removing[[:space:]]Out-of-Focus[[:space:]]Reflective[[:space:]]Flares[[:space:]]via[[:space:]]Color[[:space:]]Alignment/70315fc4-f942-4ea9-82b2-3f1b3b13eb67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rep-MTL_[[:space:]]Unleashing[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]Representation-level[[:space:]]Task[[:space:]]Saliency[[:space:]]for[[:space:]]Multi-Task[[:space:]]Learning/74e4c13f-f0fd-46af-9866-b91bbffced5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Representation[[:space:]]Shift_[[:space:]]Unifying[[:space:]]Token[[:space:]]Compression[[:space:]]with[[:space:]]FlashAttention/26aa3652-3129-495e-88f3-790b7421a62b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Representing[[:space:]]3D[[:space:]]Shapes[[:space:]]with[[:space:]]64[[:space:]]Latent[[:space:]]Vectors[[:space:]]for[[:space:]]3D[[:space:]]Diffusion[[:space:]]Models/9e1ce8a0-9ae0-47e7-ab6b-98d3658911a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Repurposing[[:space:]]2D[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Gaussian[[:space:]]Atlas[[:space:]]for[[:space:]]3D[[:space:]]Generation/7716ed45-364a-4f67-a590-a6006e1af52f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ResGS_[[:space:]]Residual[[:space:]]Densification[[:space:]]of[[:space:]]3D[[:space:]]Gaussian[[:space:]]for[[:space:]]Efficient[[:space:]]Detail[[:space:]]Recovery/ea8f33a4-cf0c-4f3b-a029-70a4f3eadc9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ResQ_[[:space:]]A[[:space:]]Novel[[:space:]]Framework[[:space:]]to[[:space:]]Implement[[:space:]]Residual[[:space:]]Neural[[:space:]]Networks[[:space:]]on[[:space:]]Analog[[:space:]]Rydberg[[:space:]]Atom[[:space:]]Quantum[[:space:]]Computers/18514cbe-ca9a-4cc2-9182-097b85868c6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ResidualViT[[:space:]]for[[:space:]]Efficient[[:space:]]Temporally[[:space:]]Dense[[:space:]]Video[[:space:]]Encoding/e7357eda-9c4d-4756-a51c-935ab0e33796_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Resolving[[:space:]]Token-Space[[:space:]]Gradient[[:space:]]Conflicts_[[:space:]]Token[[:space:]]Space[[:space:]]Manipulation[[:space:]]for[[:space:]]Transformer-Based[[:space:]]Multi-Task[[:space:]]Learning/d84395af-5a07-426f-a066-b7204dfbe77b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Resonance_[[:space:]]Learning[[:space:]]to[[:space:]]Predict[[:space:]]Social-Aware[[:space:]]Pedestrian[[:space:]]Trajectories[[:space:]]as[[:space:]]Co-Vibrations/88b7013c-c324-48da-b18c-bd6fa8c22193_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethink[[:space:]]Sparse[[:space:]]Signals[[:space:]]for[[:space:]]Pose-guided[[:space:]]Text-to-image[[:space:]]Generation/939bd1a6-6d51-4897-b577-5232b862ef5e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Bimanual[[:space:]]Robotic[[:space:]]Manipulation_[[:space:]]Learning[[:space:]]with[[:space:]]Decoupled[[:space:]]Interaction[[:space:]]Framework/37dab4a3-de4c-4d88-a54d-af55313399e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Cross-Modal[[:space:]]Interaction[[:space:]]in[[:space:]]Multimodal[[:space:]]Diffusion[[:space:]]Transformers/0f642901-6ad1-4e17-bcd5-c7b75d059fb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]DPO-style[[:space:]]Diffusion[[:space:]]Aligning[[:space:]]Frameworks/52856a50-434a-4a7f-8993-7a9fd5e8f691_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Detecting[[:space:]]Salient[[:space:]]and[[:space:]]Camouflaged[[:space:]]Objects[[:space:]]in[[:space:]]Unconstrained[[:space:]]Scenes/4b6d80af-f319-4b6f-a674-b96cf05b6cd4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Discrete[[:space:]]Tokens_[[:space:]]Treating[[:space:]]Them[[:space:]]as[[:space:]]Conditions[[:space:]]for[[:space:]]Continuous[[:space:]]Autoregressive[[:space:]]Image[[:space:]]Synthesis/d9603bf3-882e-455f-a588-46e61d648712_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Few[[:space:]]Shot[[:space:]]CLIP[[:space:]]Benchmarks_[[:space:]]A[[:space:]]Critical[[:space:]]Analysis[[:space:]]in[[:space:]]the[[:space:]]Inductive[[:space:]]Setting/6c86ab93-0fc4-4c20-be49-ae8f1e5ce8da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Key-frame-based[[:space:]]Micro-expression[[:space:]]Recognition_[[:space:]]A[[:space:]]Robust[[:space:]]and[[:space:]]Accurate[[:space:]]Framework[[:space:]]Against[[:space:]]Key-frame[[:space:]]Errors/303f401e-52fb-4695-beab-8441428174de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Layered[[:space:]]Graphic[[:space:]]Design[[:space:]]Generation[[:space:]]with[[:space:]]a[[:space:]]Top-Down[[:space:]]Approach/e44dfab2-499d-4513-b07e-10fccad954f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]Multi-modal[[:space:]]Object[[:space:]]Detection[[:space:]]from[[:space:]]the[[:space:]]Perspective[[:space:]]of[[:space:]]Mono-Modality[[:space:]]Feature[[:space:]]Learning/e2607e16-227a-44ff-ba5d-49ca94ab306a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]the[[:space:]]Embodied[[:space:]]Gap[[:space:]]in[[:space:]]Vision-and-Language[[:space:]]Navigation_[[:space:]]A[[:space:]]Holistic[[:space:]]Study[[:space:]]of[[:space:]]Physical[[:space:]]and[[:space:]]Visual[[:space:]]Disparities/2d54f040-e5ce-48d7-af3d-a61b07c7a6a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Rethinking[[:space:]]the[[:space:]]Upsampling[[:space:]]Process[[:space:]]in[[:space:]]Light[[:space:]]Field[[:space:]]Super-Resolution[[:space:]]with[[:space:]]Spatial-Epipolar[[:space:]]Implicit[[:space:]]Image[[:space:]]Function/96fd162c-a9ed-4a28-a688-bf2d29e708a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Retinex-MEF_[[:space:]]Retinex-based[[:space:]]Glare[[:space:]]Effects[[:space:]]Aware[[:space:]]Unsupervised[[:space:]]Multi-Exposure[[:space:]]Image[[:space:]]Fusion/352310b1-e32e-454f-9ece-777f202fc1ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RetinexMCNet_[[:space:]]A[[:space:]]Memory[[:space:]]Controller[[:space:]]Dominated[[:space:]]Network[[:space:]]for[[:space:]]Low-Light[[:space:]]Video[[:space:]]Enhancement[[:space:]]Based[[:space:]]on[[:space:]]Retinex/c758bfbb-509d-4d7f-b943-bed78c3b2ea7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reusing[[:space:]]Computation[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]for[[:space:]]Efficient[[:space:]]Generation[[:space:]]of[[:space:]]Image[[:space:]]Sets/fa0a87f7-7525-41f2-95f3-e6944c4ef380_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revelio_[[:space:]]Interpreting[[:space:]]and[[:space:]]leveraging[[:space:]]semantic[[:space:]]information[[:space:]]in[[:space:]]diffusion[[:space:]]models/7433fada-e812-408e-a299-b2985343c6a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Reverse[[:space:]]Convolution[[:space:]]and[[:space:]]Its[[:space:]]Applications[[:space:]]to[[:space:]]Image[[:space:]]Restoration/a4f330c7-a90b-47b8-ac12-43e41bf10069_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revisiting[[:space:]]Adversarial[[:space:]]Patch[[:space:]]Defenses[[:space:]]on[[:space:]]Object[[:space:]]Detectors_[[:space:]]Unified[[:space:]]Evaluation,[[:space:]]Large-Scale[[:space:]]Dataset,[[:space:]]and[[:space:]]New[[:space:]]Insights/a48f4320-f97b-4dc0-8cdb-096a6f193843_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revisiting[[:space:]]Efficient[[:space:]]Semantic[[:space:]]Segmentation_[[:space:]]Learning[[:space:]]Offsets[[:space:]]for[[:space:]]Better[[:space:]]Spatial[[:space:]]and[[:space:]]Class[[:space:]]Feature[[:space:]]Alignment/d11830e8-d201-4a41-b435-88acfc1d0006_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revisiting[[:space:]]Image[[:space:]]Fusion[[:space:]]for[[:space:]]Multi-Illuminant[[:space:]]White-Balance[[:space:]]Correction/9c440d82-0048-41ce-b7e7-3ee14f02370b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revisiting[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion_[[:space:]]Are[[:space:]]We[[:space:]]Ready[[:space:]]For[[:space:]]The[[:space:]]Real-World_/3fa00143-ea88-42b2-8a44-5c2b13e77edf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Revisiting[[:space:]]Pool-based[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Few-shot[[:space:]]Class-incremental[[:space:]]Learning/458b2414-5eb6-4011-a991-55dfe546f0c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RhythmGuassian_[[:space:]]Repurposing[[:space:]]Generalizable[[:space:]]Gaussian[[:space:]]Model[[:space:]]For[[:space:]]Remote[[:space:]]Physiological[[:space:]]Measurement/4917bcdf-a46c-4f95-84f9-159b7135b95b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Riemannian-Geometric[[:space:]]Fingerprints[[:space:]]of[[:space:]]Generative[[:space:]]Models/a6552543-e6fe-45d5-8e6e-b35c224e1bac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RnGCam_[[:space:]]High-speed[[:space:]]video[[:space:]]from[[:space:]]rolling[[:space:]]&[[:space:]]global[[:space:]]shutter[[:space:]]measurements/dbbc74b2-7b5f-417e-b3da-46239efd43ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoBridge_[[:space:]]A[[:space:]]Hierarchical[[:space:]]Architecture[[:space:]]Bridging[[:space:]]Cognition[[:space:]]and[[:space:]]Execution[[:space:]]for[[:space:]]General[[:space:]]Robotic[[:space:]]Manipulation/5ab8efd6-d27d-4321-9888-92a960391e6b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoCo-Sim_[[:space:]]Enhancing[[:space:]]Roadside[[:space:]]Collaborative[[:space:]]Perception[[:space:]]through[[:space:]]Foreground[[:space:]]Simulation/dd0cfc94-a4ef-4cc0-a2a4-698baf45c00c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoMo_[[:space:]]Robust[[:space:]]Motion[[:space:]]Segmentation[[:space:]]Improves[[:space:]]Structure[[:space:]]from[[:space:]]Motion/87f832e4-622b-499d-8c3b-6555eb345382_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RobAVA_[[:space:]]A[[:space:]]Large-scale[[:space:]]Dataset[[:space:]]and[[:space:]]Baseline[[:space:]]Towards[[:space:]]Video[[:space:]]based[[:space:]]Robotic[[:space:]]Arm[[:space:]]Action[[:space:]]Understanding/9168021d-e8fe-40b5-aee2-baa2b95ae99b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robin3D_[[:space:]]Improving[[:space:]]3D[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]via[[:space:]]Robust[[:space:]]Instruction[[:space:]]Tuning/6a9d3a84-40db-4d8f-a6d3-efeb653ca905_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboAnnotatorX_[[:space:]]A[[:space:]]Comprehensive[[:space:]]and[[:space:]]Universal[[:space:]]Annotation[[:space:]]Framework[[:space:]]for[[:space:]]Accurate[[:space:]]Understanding[[:space:]]of[[:space:]]Long-horizon[[:space:]]Robot[[:space:]]Demonstration/f66f5fe4-054b-4cff-8e20-4267c7dfeb6b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboFactory_[[:space:]]Exploring[[:space:]]Embodied[[:space:]]Agent[[:space:]]Collaboration[[:space:]]with[[:space:]]Compositional[[:space:]]Constraints/343afcf0-ff73-458e-8fba-e05e96b283f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboPearls_[[:space:]]Editable[[:space:]]Video[[:space:]]Simulation[[:space:]]for[[:space:]]Robot[[:space:]]Manipulation/b5d816a7-c248-4754-ab7f-df15a4fc862c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboTrom-Nav_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Embodied[[:space:]]Navigation[[:space:]]Integrating[[:space:]]Perception,[[:space:]]Planning,[[:space:]]and[[:space:]]Prediction/5cf61070-e2c5-495f-930d-bcb6f4546a80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboTron-Drive_[[:space:]]All-in-One[[:space:]]Large[[:space:]]Multimodal[[:space:]]Model[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/40d30cbc-6840-469c-b071-37cdaf04ff9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboTron-Mani_[[:space:]]All-in-One[[:space:]]Multimodal[[:space:]]Large[[:space:]]Model[[:space:]]for[[:space:]]Robotic[[:space:]]Manipulation/fa059703-e582-45fb-819c-27befc687c68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RoboTron-Sim_[[:space:]]Improving[[:space:]]Real-World[[:space:]]Driving[[:space:]]via[[:space:]]Simulated[[:space:]]Hard-Case/43e228e4-1ab7-481a-9789-346733c6c6d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RobuSTereo_[[:space:]]Robust[[:space:]]Zero-Shot[[:space:]]Stereo[[:space:]]Matching[[:space:]]under[[:space:]]Adverse[[:space:]]Weather/7fbb90b0-99b1-4c9f-8641-c8e30c417fdb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]using[[:space:]]Probabilistic[[:space:]]Point[[:space:]]Clouds[[:space:]]from[[:space:]]Single-Photon[[:space:]]LiDARs/9f8febf0-b817-400b-8412-92da98e2bc1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]3D-Masked[[:space:]]Part-level[[:space:]]Editing[[:space:]]in[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]with[[:space:]]Regularized[[:space:]]Score[[:space:]]Distillation[[:space:]]Sampling/8c68cabb-f5ac-4952-90e7-fd7f754c18df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]Adverse[[:space:]]Weather[[:space:]]Removal[[:space:]]via[[:space:]]Spectral-based[[:space:]]Spatial[[:space:]]Grouping/0e90ccb3-652c-475f-af9d-af974bf64a53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]Dataset[[:space:]]Condensation[[:space:]]using[[:space:]]Supervised[[:space:]]Contrastive[[:space:]]Learning/f8793aed-ba12-4172-a270-b29d0f50a099_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]Low-light[[:space:]]Scene[[:space:]]Restoration[[:space:]]via[[:space:]]Illumination[[:space:]]Transition/50bcb1da-1301-4e07-a545-114c913ec16d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]Machine[[:space:]]Unlearning[[:space:]]for[[:space:]]Quantized[[:space:]]Neural[[:space:]]Networks[[:space:]]via[[:space:]]Adaptive[[:space:]]Gradient[[:space:]]Reweighting[[:space:]]with[[:space:]]Similar[[:space:]]Labels/2601b190-051a-4ff9-b52d-3dd01f6b64bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]Multi-View[[:space:]]Learning[[:space:]]via[[:space:]]Representation[[:space:]]Fusion[[:space:]]of[[:space:]]Sample-Level[[:space:]]Attention[[:space:]]and[[:space:]]Alignment[[:space:]]of[[:space:]]Simulated[[:space:]]Perturbation/80914a28-0909-4f54-a37a-7922ac79c179_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Denoising[[:space:]]Using[[:space:]]Deep[[:space:]]Gaussian[[:space:]]Prior/7c456e93-c4ba-4026-b03e-658f93c3b052_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]Unfolding[[:space:]]Network[[:space:]]for[[:space:]]HDR[[:space:]]Imaging[[:space:]]with[[:space:]]Modulo[[:space:]]Cameras/e44946ea-bf4c-459a-a7b5-ff565aba7eb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robust[[:space:]]and[[:space:]]Efficient[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Urban[[:space:]]Scene[[:space:]]Reconstruction/e4ad7bc0-7409-4b1f-a1d2-2f3a333e783b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RobustSplat_[[:space:]]Decoupling[[:space:]]Densification[[:space:]]and[[:space:]]Dynamics[[:space:]]for[[:space:]]Transient-Free[[:space:]]3DGS/8475d6f9-fac1-42c3-ba00-a16e8a637015_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Robustifying[[:space:]]Zero-Shot[[:space:]]Vision[[:space:]]Language[[:space:]]Models[[:space:]]by[[:space:]]Subspaces[[:space:]]Alignment/f8c767c3-f7d0-4c52-b3fe-ae8863dddf16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RogSplat_[[:space:]]Robust[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]via[[:space:]]Generative[[:space:]]Priors/aee2af34-b308-4475-8e9e-caa47c120890_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/RomanTex_[[:space:]]Decoupling[[:space:]]3D-aware[[:space:]]Rotary[[:space:]]Positional[[:space:]]Embedded[[:space:]]Multi-Attention[[:space:]]Network[[:space:]]for[[:space:]]Texture[[:space:]]Synthesis/d3a8548e-cd70-4f26-9ef7-dc94ba850bd4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ross3D_[[:space:]]Reconstructive[[:space:]]Visual[[:space:]]Instruction[[:space:]]Tuning[[:space:]]with[[:space:]]3D-Awareness/60d216ff-b73f-40ba-a7b2-8f4d465527e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/S2M2_[[:space:]]Scalable[[:space:]]Stereo[[:space:]]Matching[[:space:]]Model[[:space:]]for[[:space:]]Reliable[[:space:]]Depth[[:space:]]Estimation/e7bb3d4a-b9f5-468e-8f9f-dd3f14d5fdc8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/S3E_[[:space:]]Self-Supervised[[:space:]]State[[:space:]]Estimation[[:space:]]for[[:space:]]Radar-Inertial[[:space:]]System/7fdc0a9d-6961-4746-b1cf-4d9fb66b4887_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/S3R-GS_[[:space:]]Streamlining[[:space:]]the[[:space:]]Pipeline[[:space:]]for[[:space:]]Large-Scale[[:space:]]Street[[:space:]]Scene[[:space:]]Reconstruction/c8af19d5-94dd-4b49-b94a-1fde74947f5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/S4M_[[:space:]]Boosting[[:space:]]Semi-Supervised[[:space:]]Instance[[:space:]]Segmentation[[:space:]]with[[:space:]]SAM/9e08d7fd-519a-425f-a6ba-6c77b0b71e10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SA-LUT_[[:space:]]Spatial[[:space:]]Adaptive[[:space:]]4D[[:space:]]Look-Up[[:space:]]Table[[:space:]]for[[:space:]]Photorealistic[[:space:]]Style[[:space:]]Transfer/6ce1c17c-a153-43de-b6b1-b11f09377561_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SA-Occ_[[:space:]]Satellite-Assisted[[:space:]]3D[[:space:]]Occupancy[[:space:]]Prediction[[:space:]]in[[:space:]]Real[[:space:]]World/956b8911-34b6-471e-ab9f-fc43a03677a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAC-GNC_[[:space:]]SAmple[[:space:]]Consensus[[:space:]]for[[:space:]]adaptive[[:space:]]Graduated[[:space:]]Non-Convexity/8cc77bba-a4a7-42b4-9408-9f23d811d9a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAFER_[[:space:]]Sharpness[[:space:]]Aware[[:space:]]layer-selective[[:space:]]Finetuning[[:space:]]for[[:space:]]Enhanced[[:space:]]Robustness[[:space:]]in[[:space:]]vision[[:space:]]transformers/0cd5fc8e-cb75-475b-9b78-38392775bdf7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAFT_[[:space:]]Shape[[:space:]]and[[:space:]]Appearance[[:space:]]of[[:space:]]Fabrics[[:space:]]from[[:space:]]Template[[:space:]]via[[:space:]]Differentiable[[:space:]]Physical[[:space:]]Simulations[[:space:]]from[[:space:]]Monocular[[:space:]]Video/c6759c13-fdf5-412f-9141-78adf63d2679_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAGI_[[:space:]]Semantically[[:space:]]Aligned[[:space:]]and[[:space:]]Uncertainty[[:space:]]Guided[[:space:]]AI[[:space:]]Image[[:space:]]Inpainting/3de45854-2db2-4424-b33f-2a82df7284fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SALAD[[:space:]]--[[:space:]]Semantics-Aware[[:space:]]Logical[[:space:]]Anomaly[[:space:]]Detection/3a0ca4fc-91bb-4c24-8c2e-18a8e9529957_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAM[[:space:]]Encoder[[:space:]]Breach[[:space:]]by[[:space:]]Adversarial[[:space:]]Simplicial[[:space:]]Complex[[:space:]]Triggers[[:space:]]Downstream[[:space:]]Model[[:space:]]Failures/580e549f-f87a-44b8-8bda-fd31161a24bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAM2Long_[[:space:]]Enhancing[[:space:]]SAM[[:space:]]2[[:space:]]for[[:space:]]Long[[:space:]]Video[[:space:]]Segmentation[[:space:]]with[[:space:]]a[[:space:]]Training-Free[[:space:]]Memory[[:space:]]Tree/55161fd3-63c3-4458-9773-51b075eb3c02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAM4D_[[:space:]]Segment[[:space:]]Anything[[:space:]]in[[:space:]]Camera[[:space:]]and[[:space:]]LiDAR[[:space:]]Streams/c01052e9-f8ce-4500-afa8-0066e9aa9f84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAME_[[:space:]]Learning[[:space:]]Generic[[:space:]]Language-Guided[[:space:]]Visual[[:space:]]Navigation[[:space:]]with[[:space:]]State-Adaptive[[:space:]]Mixture[[:space:]]of[[:space:]]Experts/8194ce86-eb6c-4c10-a9fa-d4cb58b8ea7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAMO_[[:space:]]A[[:space:]]Lightweight[[:space:]]Sharpness-Aware[[:space:]]Approach[[:space:]]for[[:space:]]Multi-Task[[:space:]]Optimization[[:space:]]with[[:space:]]Joint[[:space:]]Global-Local[[:space:]]Perturbation/c40b36bd-5368-4885-90e5-27d3f4371534_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAMPLE_[[:space:]]Semantic[[:space:]]Alignment[[:space:]]through[[:space:]]Temporal-Adaptive[[:space:]]Multimodal[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Event-Based[[:space:]]Open-Vocabulary[[:space:]]Action[[:space:]]Recognition/226c099a-5397-466f-a214-dc837ff1c164_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAMora_[[:space:]]Enhancing[[:space:]]SAM[[:space:]]through[[:space:]]Hierarchical[[:space:]]Self-Supervised[[:space:]]Pre-Training[[:space:]]for[[:space:]]Medical[[:space:]]Images/d73be635-24fb-41ee-8e17-538cff3be0ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SANA-Sprint_[[:space:]]One-Step[[:space:]]Diffusion[[:space:]]with[[:space:]]Continuous-Time[[:space:]]Consistency[[:space:]]Distillation/10606fea-7723-4a73-ba90-39649fbed303_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAS_[[:space:]]Segment[[:space:]]Any[[:space:]]3D[[:space:]]Scene[[:space:]]with[[:space:]]Integrated[[:space:]]2D[[:space:]]Priors/990d5716-df66-4bf1-8d78-8ce250deb1e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SAUCE_[[:space:]]Selective[[:space:]]Concept[[:space:]]Unlearning[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]Sparse[[:space:]]Autoencoders/f19f55b6-b58f-4e66-87eb-3d002422aab2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SC-Captioner_[[:space:]]Improving[[:space:]]Image[[:space:]]Captioning[[:space:]]with[[:space:]]Self-Correction[[:space:]]by[[:space:]]Reinforcement[[:space:]]Learning/07695081-246e-4f25-93dd-467806af5755_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SC-Lane_[[:space:]]Slope-aware[[:space:]]and[[:space:]]Consistent[[:space:]]Road[[:space:]]Height[[:space:]]Estimation[[:space:]]Framework[[:space:]]for[[:space:]]3D[[:space:]]Lane[[:space:]]Detection/165032b4-ae34-4f90-b84c-9e1d8860f0a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SCAN_[[:space:]]Bootstrapping[[:space:]]Contrastive[[:space:]]Pre-training[[:space:]]for[[:space:]]Data[[:space:]]Efficiency/1c1dbb5f-0f2a-411d-9b9d-75b8cf589dfe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SCFlow_[[:space:]]Implicitly[[:space:]]Learning[[:space:]]Style[[:space:]]and[[:space:]]Content[[:space:]]Disentanglement[[:space:]]with[[:space:]]Flow[[:space:]]Models/210472bd-a69c-42d2-83d3-aca94d045a19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SCORE_[[:space:]]Scene[[:space:]]Context[[:space:]]Matters[[:space:]]in[[:space:]]Open-Vocabulary[[:space:]]Remote[[:space:]]Sensing[[:space:]]Instance[[:space:]]Segmentation/cbd59ce2-2270-4f67-803e-71f3a8d3f10f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SD2Actor_[[:space:]]Continuous[[:space:]]State[[:space:]]Decomposition[[:space:]]via[[:space:]]Diffusion[[:space:]]Embeddings[[:space:]]for[[:space:]]Robotic[[:space:]]Manipulation/9b178221-568a-43b6-b265-e713912270e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SDFit_[[:space:]]3D[[:space:]]Object[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]by[[:space:]]Fitting[[:space:]]a[[:space:]]Morphable[[:space:]]SDF[[:space:]]to[[:space:]]a[[:space:]]Single[[:space:]]Image/8d65ef05-f711-4561-b8d8-23bde764d685_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SDFormer_[[:space:]]Vision-based[[:space:]]3D[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion[[:space:]]via[[:space:]]SAM-assisted[[:space:]]Dual-channel[[:space:]]Voxel[[:space:]]Transformer/8361fb40-bff6-4d7f-8cac-94b8d7233a46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SDMatte_[[:space:]]Grafting[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Interactive[[:space:]]Matting/499a5bf2-6bc0-4487-b57e-4e8ed3e2960d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SEAL_[[:space:]]Semantic[[:space:]]Aware[[:space:]]Image[[:space:]]Watermarking/24fa32d7-b4de-4e23-8a87-1b7645156009_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SEGA_[[:space:]]A[[:space:]]Stepwise[[:space:]]Evolution[[:space:]]Paradigm[[:space:]]for[[:space:]]Content-Aware[[:space:]]Layout[[:space:]]Generation[[:space:]]with[[:space:]]Design[[:space:]]Prior/80589886-7060-48e6-96c8-c53727d6e19c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SEGS-SLAM_[[:space:]]Structure-enhanced[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]SLAM[[:space:]]with[[:space:]]Appearance[[:space:]]Embedding/611eb9e2-977d-4789-be4d-ef9d556964e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SEHDR_[[:space:]]Single-Exposure[[:space:]]HDR[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]via[[:space:]]3D[[:space:]]Gaussian[[:space:]]Bracketing/501f645c-e4b8-49fb-898f-e6dd000f523c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SEREP_[[:space:]]Semantic[[:space:]]Facial[[:space:]]Expression[[:space:]]Representation[[:space:]]for[[:space:]]Robust[[:space:]]In-the-Wild[[:space:]]Capture[[:space:]]and[[:space:]]Retargeting/be7a4783-1d27-4901-9ebd-d981806e2993_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SFUOD_[[:space:]]Source-Free[[:space:]]Unknown[[:space:]]Object[[:space:]]Detection/afa8cfe8-ad89-41f5-ba01-ea4ad7d5ea44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SG-LDM_[[:space:]]Semantic-Guided[[:space:]]LiDAR[[:space:]]Generation[[:space:]]via[[:space:]]Latent-Aligned[[:space:]]Diffusion/3ea849b1-b5df-4d8b-b591-3a2d5a190919_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SGAD_[[:space:]]Semantic[[:space:]]and[[:space:]]Geometric-aware[[:space:]]Descriptor[[:space:]]for[[:space:]]Local[[:space:]]Feature[[:space:]]Matching/29ce3f6d-e830-4842-96d7-2b0b1bc91567_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SHIFT_[[:space:]]Smoothing[[:space:]]Hallucinations[[:space:]]by[[:space:]]Information[[:space:]]Flow[[:space:]]Tuning[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/24bee37f-188f-4328-83d1-339365bd9b90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SHeaP_[[:space:]]Self-Supervised[[:space:]]Head[[:space:]]Geometry[[:space:]]Predictor[[:space:]]Learned[[:space:]]via[[:space:]]2D[[:space:]]Gaussians/3ad848d7-b59c-489b-9f10-b91346521a68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SIC_[[:space:]]Similarity-Based[[:space:]]Interpretable[[:space:]]Image[[:space:]]Classification[[:space:]]with[[:space:]]Neural[[:space:]]Networks/77bb94c1-b520-408f-b4c5-8b6f4de61cbd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SIGMAN_[[:space:]]Scaling[[:space:]]3D[[:space:]]Human[[:space:]]Gaussian[[:space:]]Generation[[:space:]]with[[:space:]]Millions[[:space:]]of[[:space:]]Assets/a40a0b1f-0065-4573-8075-4c237198e8dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SILO_[[:space:]]Solving[[:space:]]Inverse[[:space:]]Problems[[:space:]]with[[:space:]]Latent[[:space:]]Operators/63ffabfd-3d67-4b0d-b6b6-6b10e4d16eef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SIMS_[[:space:]]Simulating[[:space:]]Stylized[[:space:]]Human-Scene[[:space:]]Interactions[[:space:]]with[[:space:]]Retrieval-Augmented[[:space:]]Script[[:space:]]Generation/dcd5f01f-10dd-4b91-872e-07e67d8d71b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SITE_[[:space:]]towards[[:space:]]Spatial[[:space:]]Intelligence[[:space:]]Thorough[[:space:]]Evaluation/f49033f0-0e45-44a2-abfb-519f5ae0298b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SKALD_[[:space:]]Learning-Based[[:space:]]Shot[[:space:]]Assembly[[:space:]]for[[:space:]]Coherent[[:space:]]Multi-Shot[[:space:]]Video[[:space:]]Creation/ec5fa6ec-cae6-4d6e-bff9-d8ede350b420_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SL2A-INR_[[:space:]]Single-Layer[[:space:]]Learnable[[:space:]]Activation[[:space:]]for[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representation/3ee63261-6528-4d4b-8b8f-4cac3c19ff16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SMARTIES_[[:space:]]Spectrum-Aware[[:space:]]Multi-Sensor[[:space:]]Auto-Encoder[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Images/e2bd0090-57a2-4e65-86b9-d6f7df1c65f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SMGDiff_[[:space:]]Soccer[[:space:]]Motion[[:space:]]Generation[[:space:]]using[[:space:]]Diffusion[[:space:]]Probabilistic[[:space:]]Models/30fbf042-cda9-4f0e-8871-c9b3c6d7d8da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SMP-Attack_[[:space:]]Boosting[[:space:]]the[[:space:]]Transferability[[:space:]]of[[:space:]]Feature[[:space:]]Importance-based[[:space:]]Adversarial[[:space:]]Attack[[:space:]]with[[:space:]]Semantics-aware[[:space:]]Multi-granularity[[:space:]]Patchout/ffa75b72-ebd6-4869-85ba-8020b166bdbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SMSTracker_[[:space:]]Tri-path[[:space:]]Score[[:space:]]Mask[[:space:]]Sigma[[:space:]]Fusion[[:space:]]for[[:space:]]Multi-Modal[[:space:]]Tracking/f3107544-21f3-48f3-aa75-4183b15aeee4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SMoLoRA_[[:space:]]Exploring[[:space:]]and[[:space:]]Defying[[:space:]]Dual[[:space:]]Catastrophic[[:space:]]Forgetting[[:space:]]in[[:space:]]Continual[[:space:]]Visual[[:space:]]Instruction[[:space:]]Tuning/b51b1c55-496e-490f-bc6a-6615976fb7bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SP2T_[[:space:]]Sparse[[:space:]]Proxy[[:space:]]Attention[[:space:]]for[[:space:]]Dual-stream[[:space:]]Point[[:space:]]Transformer/d7713683-21d1-4626-b4c3-bece5c4c8ff1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SPADE_[[:space:]]Spatial-Aware[[:space:]]Denoising[[:space:]]Network[[:space:]]for[[:space:]]Open-vocabulary[[:space:]]Panoptic[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]with[[:space:]]Long-[[:space:]]and[[:space:]]Local-range[[:space:]]Context[[:space:]]Reasoning/7002ff60-b6d4-4614-abf8-fe44b10d3313_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SPA_[[:space:]]Efficient[[:space:]]User-Preference[[:space:]]Alignment[[:space:]]against[[:space:]]Uncertainty[[:space:]]in[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/b69594dd-0243-4738-ae53-64224321ff55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SPD_[[:space:]]Shallow[[:space:]]Backdoor[[:space:]]Protecting[[:space:]]Deep[[:space:]]Backdoor[[:space:]]Against[[:space:]]Backdoor[[:space:]]Detection/14ac6a96-392d-457b-82be-dd31e1664d47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SRefiner_[[:space:]]Soft-Braid[[:space:]]Attention[[:space:]]for[[:space:]]Multi-Agent[[:space:]]Trajectory[[:space:]]Refinement/c8f45dae-b292-45ae-b91e-864559991250_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SSVQ_[[:space:]]Unleashing[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Vector[[:space:]]Quantization[[:space:]]with[[:space:]]Sign-Splitting/86761629-4cb3-4dd1-b2ba-5ed37977bbff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STAR_[[:space:]]Spatial-Temporal[[:space:]]Augmentation[[:space:]]with[[:space:]]Text-to-Video[[:space:]]Models[[:space:]]for[[:space:]]Real-World[[:space:]]Video[[:space:]]Super-Resolution/49006492-ae85-4278-aaeb-5a8a7210f180_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STD-GS_[[:space:]]Exploring[[:space:]]Frame-Event[[:space:]]Interaction[[:space:]]for[[:space:]]SpatioTemporal-Disentangled[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]to[[:space:]]Reconstruct[[:space:]]High-Dynamic[[:space:]]Scene/91dcae8a-d3b6-423a-9456-b34d3bb1786a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STDDNet_[[:space:]]Harnessing[[:space:]]Mamba[[:space:]]for[[:space:]]Video[[:space:]]Polyp[[:space:]]Segmentation[[:space:]]via[[:space:]]Spatial-aligned[[:space:]]Temporal[[:space:]]Modeling[[:space:]]and[[:space:]]Discriminative[[:space:]]Dynamic[[:space:]]Representation[[:space:]]Learning/ce8324fb-e0aa-48f4-8644-98f5248be52e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STEP-DETR_[[:space:]]Advancing[[:space:]]DETR-based[[:space:]]Semi-Supervised[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Super[[:space:]]Teacher[[:space:]]and[[:space:]]Pseudo-Label[[:space:]]Guided[[:space:]]Text[[:space:]]Queries/0bd8ea0e-9324-44c3-8f21-0e87718cbde6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STI-Bench_[[:space:]]Are[[:space:]]MLLMs[[:space:]]Ready[[:space:]]for[[:space:]]Precise[[:space:]]Spatial-Temporal[[:space:]]World[[:space:]]Understanding_/647f0831-8bb8-493c-9b91-ff6a53903710_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STIV_[[:space:]]Scalable[[:space:]]Text[[:space:]]and[[:space:]]Image[[:space:]]Conditioned[[:space:]]Video[[:space:]]Generation/e1b1ef4f-1214-4023-80c5-d298135470b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/STaR_[[:space:]]Seamless[[:space:]]Spatial-Temporal[[:space:]]Aware[[:space:]]Motion[[:space:]]Retargeting[[:space:]]with[[:space:]]Penetration[[:space:]]and[[:space:]]Consistency[[:space:]]Constraints/67301f20-0855-48c4-a509-a40f53f849da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SU-RGS_[[:space:]]Relightable[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Sparse[[:space:]]Views[[:space:]]under[[:space:]]Unconstrained[[:space:]]Illuminations/fb3eeb3c-b190-4a6e-9b7a-e226dbcd1415_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SUB_[[:space:]]Benchmarking[[:space:]]CBM[[:space:]]Generalization[[:space:]]via[[:space:]]Synthetic[[:space:]]Attribute[[:space:]]Substitutions/c78725e9-72f0-4e6d-80b3-f81b0faf53d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SUV_[[:space:]]Suppressing[[:space:]]Undesired[[:space:]]Video[[:space:]]Content[[:space:]]via[[:space:]]Semantic[[:space:]]Modulation[[:space:]]Based[[:space:]]on[[:space:]]Text[[:space:]]Embeddings/00e00b2a-2d33-4d06-a397-3e09efaebfad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SV4D[[:space:]]2.0_[[:space:]]Enhancing[[:space:]]Spatio-Temporal[[:space:]]Consistency[[:space:]]in[[:space:]]Multi-View[[:space:]]Video[[:space:]]Diffusion[[:space:]]for[[:space:]]High-Quality[[:space:]]4D[[:space:]]Generation/06fc2738-4b79-4d01-99d7-1efc0ae7b1a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SVG-Head_[[:space:]]Hybrid[[:space:]]Surface-Volumetric[[:space:]]Gaussians[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Head[[:space:]]Reconstruction[[:space:]]and[[:space:]]Real-Time[[:space:]]Editing/dc245795-ff5d-4e8f-9dac-3e46447c3803_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SVIP_[[:space:]]Semantically[[:space:]]Contextualized[[:space:]]Visual[[:space:]]Patches[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Learning/7395aa09-a2a6-4562-a76c-fd8d08c4fd7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SVTRv2_[[:space:]]CTC[[:space:]]Beats[[:space:]]Encoder-Decoder[[:space:]]Models[[:space:]]in[[:space:]]Scene[[:space:]]Text[[:space:]]Recognition/6c8633d2-9dca-4944-9b9b-c8363635f3a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SViM3D_[[:space:]]Stable[[:space:]]Video[[:space:]]Material[[:space:]]Diffusion[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]3D[[:space:]]Generation/3fabe82d-b5af-4114-83c5-842c88d4a7c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Safeguarding[[:space:]]Vision-Language[[:space:]]Models_[[:space:]]Mitigating[[:space:]]Vulnerabilities[[:space:]]to[[:space:]]Gaussian[[:space:]]Noise[[:space:]]in[[:space:]]Perturbation-based[[:space:]]Attacks/950e834f-b0f1-4e92-9dcc-faac704eccc0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Saliency-Aware[[:space:]]Quantized[[:space:]]Imitation[[:space:]]Learning[[:space:]]for[[:space:]]Efficient[[:space:]]Robotic[[:space:]]Control/2a537502-687b-49be-a0bd-ef85b0ab63fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Salvaging[[:space:]]the[[:space:]]Overlooked_[[:space:]]Leveraging[[:space:]]Class-Aware[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Multi-Class[[:space:]]Anomaly[[:space:]]Detection/e5659225-0fe1-4545-9ca2-9f5926d28cd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sat2City_[[:space:]]3D[[:space:]]City[[:space:]]Generation[[:space:]]from[[:space:]]A[[:space:]]Single[[:space:]]Satellite[[:space:]]Image[[:space:]]with[[:space:]]Cascaded[[:space:]]Latent[[:space:]]Diffusion/09f45ab0-7eef-49cf-9544-c0cd93d26c66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scalable[[:space:]]Dual[[:space:]]Fingerprinting[[:space:]]for[[:space:]]Hierarchical[[:space:]]Attribution[[:space:]]of[[:space:]]Text-to-Image[[:space:]]Models/6e5daf87-e14c-4ebc-b132-6f5262a81466_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scalable[[:space:]]Image[[:space:]]Tokenization[[:space:]]with[[:space:]]Index[[:space:]]Backpropagation[[:space:]]Quantization/7dabf22a-185d-49ce-8f9d-a61a9cf49852_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scalable[[:space:]]Ranked[[:space:]]Preference[[:space:]]Optimization[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/fbaca45f-41ea-42e5-8a84-4543c6a6a73b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scale[[:space:]]Your[[:space:]]Instructions_[[:space:]]Enhance[[:space:]]the[[:space:]]Instruction-Following[[:space:]]Fidelity[[:space:]]of[[:space:]]Unified[[:space:]]Image[[:space:]]Generation[[:space:]]Model[[:space:]]by[[:space:]]Self-Adaptive[[:space:]]Attention[[:space:]]Scaling/196008be-2aa6-4e9a-a127-8a7f835289ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]3D[[:space:]]Compositional[[:space:]]Models[[:space:]]for[[:space:]]Robust[[:space:]]Classification[[:space:]]and[[:space:]]Pose[[:space:]]Estimation/081d955d-8013-407f-a8f5-d3d15d6c303c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]Action[[:space:]]Detection_[[:space:]]AdaTAD++[[:space:]]with[[:space:]]Transformer-Enhanced[[:space:]]Temporal-Spatial[[:space:]]Adaptation/4279f865-68cf-4fb1-b906-d2667a7054bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]Inference-Time[[:space:]]Search[[:space:]]with[[:space:]]Vision[[:space:]]Value[[:space:]]Model[[:space:]]for[[:space:]]Improved[[:space:]]Visual[[:space:]]Comprehension/66dbd2a2-c0d3-493e-aebd-02c32762a66e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]Language-Free[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/130d8529-9d51-4e10-8b3b-06448318ad16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]Laws[[:space:]]for[[:space:]]Native[[:space:]]Multimodal[[:space:]]Models/03be59c7-3487-458a-af21-d31e774b920a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]Omni-modal[[:space:]]Pretraining[[:space:]]with[[:space:]]Multimodal[[:space:]]Context_[[:space:]]Advancing[[:space:]]Universal[[:space:]]Representation[[:space:]]Learning[[:space:]]Across[[:space:]]Modalities/ccc83d07-429f-46ed-8df2-f3d62cd574b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]Transformer-Based[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]Models[[:space:]]Token[[:space:]]Disentanglement[[:space:]]and[[:space:]]Synthetic[[:space:]]Data/1d070cf7-16fa-4adb-9c4f-eddeb7b49d59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]Tumor[[:space:]]Segmentation_[[:space:]]Best[[:space:]]Lessons[[:space:]]from[[:space:]]Real[[:space:]]and[[:space:]]Synthetic[[:space:]]Data/fc315cff-7a9c-4741-b2d0-9e9b8a893836_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scaling[[:space:]]and[[:space:]]Taming[[:space:]]Adversarial[[:space:]]Training[[:space:]]with[[:space:]]Synthetic[[:space:]]Data/2f606300-1a53-469b-8595-e831194d2041_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ScanEdit_[[:space:]]Hierarchically-Guided[[:space:]]Functional[[:space:]]3D[[:space:]]Scan[[:space:]]Editing/2654dbc5-1fe2-4c4f-817e-abbfd0152170_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scendi[[:space:]]Score_[[:space:]]Prompt-Aware[[:space:]]Diversity[[:space:]]Evaluation[[:space:]]via[[:space:]]Schur[[:space:]]Complement[[:space:]]of[[:space:]]CLIP[[:space:]]Embeddings/ca0ba2ac-8778-4c62-9c3e-a4dc899f161b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scene[[:space:]]Coordinate[[:space:]]Reconstruction[[:space:]]Priors/78c0cc72-6b59-44cf-b711-23ceb9b57115_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scene[[:space:]]Graph[[:space:]]Guided[[:space:]]Generation_[[:space:]]Enable[[:space:]]Accurate[[:space:]]Relations[[:space:]]Generation[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Models[[:space:]]via[[:space:]]Textural[[:space:]]Rectification/076da6dc-7d42-4320-bceb-8e07c0c15425_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SceneMI_[[:space:]]Motion[[:space:]]In-betweening[[:space:]]for[[:space:]]Modeling[[:space:]]Human-Scene[[:space:]]Interaction/3ef39760-7979-4139-8b04-b8f341fdcc6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ScenePainter_[[:space:]]Semantically[[:space:]]Consistent[[:space:]]Perpetual[[:space:]]3D[[:space:]]Scene[[:space:]]Generation[[:space:]]with[[:space:]]Concept[[:space:]]Relation[[:space:]]Alignment/11bf455e-9362-4428-a812-42a34955c98b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SceneSplat_[[:space:]]Gaussian[[:space:]]Splatting-based[[:space:]]Scene[[:space:]]Understanding[[:space:]]with[[:space:]]Vision-Language[[:space:]]Pretraining/a0699557-75fd-4112-84f6-c3aec5423ac9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scheduling[[:space:]]Weight[[:space:]]Transitions[[:space:]]for[[:space:]]Quantization-Aware[[:space:]]Training/b8d1d5c9-b168-470a-94da-d2f2ffb19806_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SciVid_[[:space:]]Cross-Domain[[:space:]]Evaluation[[:space:]]of[[:space:]]Video[[:space:]]Models[[:space:]]in[[:space:]]Scientific[[:space:]]Applications/897c133b-e223-460f-a3e4-19f4317f7ae4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ScoreHOI_[[:space:]]Physically[[:space:]]Plausible[[:space:]]Reconstruction[[:space:]]of[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]via[[:space:]]Score-Guided[[:space:]]Diffusion/4cfe16a9-80c4-4dd8-beea-def139abf561_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Scoring,[[:space:]]Remember,[[:space:]]and[[:space:]]Reference_[[:space:]]Catching[[:space:]]Camouflaged[[:space:]]Objects[[:space:]]in[[:space:]]Videos/5f947353-bfc5-4712-a12b-2af95dc8120b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sculpting[[:space:]]Memory_[[:space:]]Multi-Concept[[:space:]]Forgetting[[:space:]]in[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Dynamic[[:space:]]Mask[[:space:]]and[[:space:]]Concept-Aware[[:space:]]Optimization/84bc13bf-2fda-4c48-b24f-8b12c6b05cc5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SeaS_[[:space:]]Few-shot[[:space:]]Industrial[[:space:]]Anomaly[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Separation[[:space:]]and[[:space:]]Sharing[[:space:]]Fine-tuning/3385c760-e7d7-40ac-b131-efa4e4e01791_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seal[[:space:]]Your[[:space:]]Backdoor[[:space:]]with[[:space:]]Variational[[:space:]]Defense/9f954aa7-ed38-4758-9d11-a4ca43444671_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seam360GS_[[:space:]]Seamless[[:space:]]360deg[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]from[[:space:]]Real-World[[:space:]]Omnidirectional[[:space:]]Images/19b1313b-bb1d-4592-99cc-0991ca27d16d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Secure[[:space:]]On-Device[[:space:]]Video[[:space:]]OOD[[:space:]]Detection[[:space:]]Without[[:space:]]Backpropagation/c8a2067b-c271-45fb-baa5-ea7c6dda304f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeing[[:space:]]3D[[:space:]]Through[[:space:]]2D[[:space:]]Lenses_[[:space:]]3D[[:space:]]Few-Shot[[:space:]]Class-Incremental[[:space:]]Learning[[:space:]]via[[:space:]]Cross-Modal[[:space:]]Geometric[[:space:]]Rectification/2a0f9f4b-14d0-474c-a989-a551b14185cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeing[[:space:]]Through[[:space:]]Deepfakes_[[:space:]]A[[:space:]]Human-Inspired[[:space:]]Framework[[:space:]]for[[:space:]]Multi-Face[[:space:]]Detection/d2bd9e05-8709-4a85-9d4a-62e175354fe7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeing[[:space:]]and[[:space:]]Seeing[[:space:]]Through[[:space:]]the[[:space:]]Glass_[[:space:]]Real[[:space:]]and[[:space:]]Synthetic[[:space:]]Data[[:space:]]for[[:space:]]Multi-Layer[[:space:]]Depth[[:space:]]Estimation/f1dec45f-4722-4c54-86f2-5e85366a6d61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeing[[:space:]]the[[:space:]]Trees[[:space:]]for[[:space:]]the[[:space:]]Forest_[[:space:]]Rethinking[[:space:]]Weakly-Supervised[[:space:]]Medical[[:space:]]Visual[[:space:]]Grounding/f8e61514-0956-45d9-b914-c150beeff0f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Seeing[[:space:]]the[[:space:]]Unseen_[[:space:]]A[[:space:]]Semantic[[:space:]]Alignment[[:space:]]and[[:space:]]Context-Aware[[:space:]]Prompt[[:space:]]Framework[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Camouflaged[[:space:]]Object[[:space:]]Segmentation/0359656f-c2da-4225-9589-6a258678bfc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SegAnyPET_[[:space:]]Universal[[:space:]]Promptable[[:space:]]Segmentation[[:space:]]from[[:space:]]Positron[[:space:]]Emission[[:space:]]Tomography[[:space:]]Images/c4866e65-d14d-4c1a-bbd4-56bcc26dbf1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SegmentDreamer_[[:space:]]Towards[[:space:]]High-fidelity[[:space:]]Text-to-3D[[:space:]]Synthesis[[:space:]]with[[:space:]]Segmented[[:space:]]Consistency[[:space:]]Trajectory[[:space:]]Distillation/51fd7c46-9b39-4739-ab88-253ff2ae72e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Selective[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Affordance[[:space:]]Grounding/eb77a737-a1e3-4e9e-874f-4c95cc8ab8dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Calibrated[[:space:]]Variance-Stabilizing[[:space:]]Transformations[[:space:]]for[[:space:]]Real-World[[:space:]]Image[[:space:]]Denoising/cf97a8e6-bc10-44ce-8633-f6d96817a34a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Calibrating[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Large[[:space:]]Field-of-View[[:space:]]Reconstruction/3c36b91d-b749-4571-ae23-525eed266188_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Ensembling[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Few-Shot[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/93b3969c-504e-4190-b789-fb768c717c3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Reinforcing[[:space:]]Prototype[[:space:]]Evolution[[:space:]]with[[:space:]]Dual-Knowledge[[:space:]]Cooperation[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Lifelong[[:space:]]Person[[:space:]]Re-Identification/0d8ec9cc-91fd-4f6f-adc9-3bba1cad24c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Supervised[[:space:]]Monocular[[:space:]]4D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]for[[:space:]]Egocentric[[:space:]]Videos/0f34d99d-368d-4ba9-b31d-032659745be1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-Supervised[[:space:]]Sparse[[:space:]]Sensor[[:space:]]Fusion[[:space:]]for[[:space:]]Long[[:space:]]Range[[:space:]]Perception/6d4fd112-3fac-483b-b978-3c134ff1cdd1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Self-supervised[[:space:]]Learning[[:space:]]of[[:space:]]Hybrid[[:space:]]Part-aware[[:space:]]3D[[:space:]]Representations[[:space:]]of[[:space:]]2D[[:space:]]Gaussians[[:space:]]and[[:space:]]Superquadrics/b5799a7c-6527-47cd-81fd-ab5ff79ff559_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SemGes_[[:space:]]Semantics-aware[[:space:]]Co-Speech[[:space:]]Gesture[[:space:]]Generation[[:space:]]using[[:space:]]Semantic[[:space:]]Coherence[[:space:]]and[[:space:]]Relevance[[:space:]]Learning/d193eee8-ef57-4a70-9609-6d91a9953312_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SemTalk_[[:space:]]Holistic[[:space:]]Co-speech[[:space:]]Motion[[:space:]]Generation[[:space:]]with[[:space:]]Frame-level[[:space:]]Semantic[[:space:]]Emphasis/d0c52370-d24a-4c84-9ea3-ee5cd6ca6239_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semantic[[:space:]]Alignment[[:space:]]and[[:space:]]Reinforcement[[:space:]]for[[:space:]]Data-Free[[:space:]]Quantization[[:space:]]of[[:space:]]Vision[[:space:]]Transformers/33638821-3dce-4a55-a900-82748a75aeee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semantic[[:space:]]Causality-Aware[[:space:]]Vision-Based[[:space:]]3D[[:space:]]Occupancy[[:space:]]Prediction/218a3708-3415-4c53-9a50-ef5f14ed0f9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semantic[[:space:]]Discrepancy-aware[[:space:]]Detector[[:space:]]for[[:space:]]Image[[:space:]]Forgery[[:space:]]Identification/d257d4bc-770d-463b-8ab1-d4cb1749eea8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semantic[[:space:]]Equitable[[:space:]]Clustering_[[:space:]]A[[:space:]]Simple[[:space:]]and[[:space:]]Effective[[:space:]]Strategy[[:space:]]for[[:space:]]Clustering[[:space:]]Vision[[:space:]]Tokens/755b1314-9e24-437c-8f51-19b61bb62095_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semantic[[:space:]]Watermarking[[:space:]]Reinvented_[[:space:]]Enhancing[[:space:]]Robustness[[:space:]]and[[:space:]]Generation[[:space:]]Quality[[:space:]]with[[:space:]]Fourier[[:space:]]Integrity/4150a989-5307-4ff1-8e86-ca9b050e7e76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semantic[[:space:]]versus[[:space:]]Identity_[[:space:]]A[[:space:]]Divide-and-Conquer[[:space:]]Approach[[:space:]]towards[[:space:]]Adjustable[[:space:]]Medical[[:space:]]Image[[:space:]]De-Identification/2b39ad1e-d18f-4579-85b8-6d56df6a82db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semantic-guided[[:space:]]Camera[[:space:]]Ray[[:space:]]Regression[[:space:]]for[[:space:]]Visual[[:space:]]Localization/d007a0e2-f228-4563-89f9-6176e65aa729_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semi-ViM_[[:space:]]Bidirectional[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]for[[:space:]]Mitigating[[:space:]]Label[[:space:]]Imbalance[[:space:]]in[[:space:]]Semi-Supervised[[:space:]]Learning/694bc6b8-9d54-4920-9947-577e19b17917_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semi-supervised[[:space:]]Concept[[:space:]]Bottleneck[[:space:]]Models/790be7e9-6d82-433e-9aab-92f2c72b545a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Semi-supervised[[:space:]]Deep[[:space:]]Transfer[[:space:]]for[[:space:]]Regression[[:space:]]without[[:space:]]Domain[[:space:]]Alignment/60f16256-f509-4463-b38b-322283424ecd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SemiVisBooster_[[:space:]]Boosting[[:space:]]Semi-Supervised[[:space:]]Learning[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Classification[[:space:]]through[[:space:]]Pseudo-Label[[:space:]]Semantic[[:space:]]Guidance/2707eef4-2e69-485f-a5ee-896dce310667_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Separation[[:space:]]for[[:space:]]Better[[:space:]]Integration_[[:space:]]Disentangling[[:space:]]Edge[[:space:]]and[[:space:]]Motion[[:space:]]in[[:space:]]Event-based[[:space:]]Deblurring/9f65b86f-9c52-4842-ae57-85222dbb0f4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SeqGrowGraph_[[:space:]]Learning[[:space:]]Lane[[:space:]]Topology[[:space:]]as[[:space:]]a[[:space:]]Chain[[:space:]]of[[:space:]]Graph[[:space:]]Expansions/f66ab5d4-0d27-477f-96dd-6cf261bca1fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sequential[[:space:]]Gaussian[[:space:]]Avatars[[:space:]]with[[:space:]]Hierarchical[[:space:]]Motion[[:space:]]Context/ce271790-0752-41a2-ac20-e12f48492a27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sequential[[:space:]]keypoint[[:space:]]density[[:space:]]estimator_[[:space:]]an[[:space:]]overlooked[[:space:]]baseline[[:space:]]of[[:space:]]skeleton-based[[:space:]]video[[:space:]]anomaly[[:space:]]detection/a7b634d4-2fd6-469f-a18b-58cbc0beefcb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Serialization[[:space:]]based[[:space:]]Point[[:space:]]Cloud[[:space:]]Oversegmentation/02d08649-2762-474d-94e0-d2ae2cc700de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ShadowHack_[[:space:]]Hacking[[:space:]]Shadows[[:space:]]via[[:space:]]Luminance-Color[[:space:]]Divide[[:space:]]and[[:space:]]Conquer/d86f4be3-9e03-4b38-b628-22e777dbb589_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Shape[[:space:]]of[[:space:]]Motion_[[:space:]]4D[[:space:]]Reconstruction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Video/a825c689-ac4d-430b-9fb3-d2a95e2ffbee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ShortFT_[[:space:]]Diffusion[[:space:]]Model[[:space:]]Alignment[[:space:]]via[[:space:]]Shortcut-based[[:space:]]Fine-Tuning/447f852c-3aba-4c21-a78d-15886a337f7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ShortV_[[:space:]]Efficient[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models[[:space:]]by[[:space:]]Freezing[[:space:]]Visual[[:space:]]Tokens[[:space:]]in[[:space:]]Ineffective[[:space:]]Layers/213fbd78-7b79-438a-a9fe-b3e13e7f59de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Shot-by-Shot_[[:space:]]Film-Grammar-Aware[[:space:]]Training-Free[[:space:]]Audio[[:space:]]Description[[:space:]]Generation/b2b3829b-f130-4214-bc1a-c2339869b439_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SiM3D_[[:space:]]Single-instance[[:space:]]Multiview[[:space:]]Multimodal[[:space:]]and[[:space:]]Multisetup[[:space:]]3D[[:space:]]Anomaly[[:space:]]Detection[[:space:]]Benchmark/fc999caf-1202-46bd-84ea-94a9faab06d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sibai_[[:space:]]A[[:space:]]Few-Shot[[:space:]]Meta-Classifier[[:space:]]for[[:space:]]Poisoning[[:space:]]Detection[[:space:]]in[[:space:]]Federated[[:space:]]Learning/c815ddc4-d777-4075-b792-39640cd474e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SignRep_[[:space:]]Enhancing[[:space:]]Self-Supervised[[:space:]]Sign[[:space:]]Representations/b0873b30-3769-4f25-84b2-07088f813479_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Signs[[:space:]]as[[:space:]]Tokens_[[:space:]]A[[:space:]]Retrieval-Enhanced[[:space:]]Multilingual[[:space:]]Sign[[:space:]]Language[[:space:]]Generator/1cb6aa16-71d9-4a9a-a1f6-c59910d7b13b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sim-DETR_[[:space:]]Unlock[[:space:]]DETR[[:space:]]for[[:space:]]Temporal[[:space:]]Sentence[[:space:]]Grounding/461e9907-9af3-47f5-b610-f09041cded55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SimMLM_[[:space:]]A[[:space:]]Simple[[:space:]]Framework[[:space:]]for[[:space:]]Multi-modal[[:space:]]Learning[[:space:]]with[[:space:]]Missing[[:space:]]Modality/398740ea-2adc-472d-af59-846c4c676c5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Similarity[[:space:]]Memory[[:space:]]Prior[[:space:]]is[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]for[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/7124b243-3435-4191-a8a2-f64fe05a8d96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SimpleVQA_[[:space:]]Multimodal[[:space:]]Factuality[[:space:]]Evaluation[[:space:]]for[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/d1c51684-4f7b-42eb-97c5-436f06ea7c3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Simulating[[:space:]]Dual-Pixel[[:space:]]Images[[:space:]]From[[:space:]]Ray[[:space:]]Tracing[[:space:]]For[[:space:]]Depth[[:space:]]Estimation/daa8d996-9820-46cb-8540-26b2ea341407_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Simultaneous[[:space:]]Motion[[:space:]]And[[:space:]]Noise[[:space:]]Estimation[[:space:]]with[[:space:]]Event[[:space:]]Cameras/6f9dcd49-a085-404d-b287-141b9737e694_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Single-Scanline[[:space:]]Relative[[:space:]]Pose[[:space:]]Estimation[[:space:]]for[[:space:]]Rolling[[:space:]]Shutter[[:space:]]Cameras/1be14823-e24f-4aa3-8a5b-8d0994a9e44f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Skeleton[[:space:]]Motion[[:space:]]Words[[:space:]]for[[:space:]]Unsupervised[[:space:]]Skeleton-Based[[:space:]]Temporal[[:space:]]Action[[:space:]]Segmentation/d38bb523-583d-4f38-b51f-72d4fb904420_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SketchSplat_[[:space:]]3D[[:space:]]Edge[[:space:]]Reconstruction[[:space:]]via[[:space:]]Differentiable[[:space:]]Multi-view[[:space:]]Sketch[[:space:]]Splatting/03d6f604-4658-4429-92bd-44a5220bf2b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Skip-Vision_[[:space:]]Efficient[[:space:]]and[[:space:]]Scalable[[:space:]]Acceleration[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models[[:space:]]via[[:space:]]Adaptive[[:space:]]Token[[:space:]]Skipping/1d1ce44b-d4d0-40e2-b544-e98ad6e37776_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SkySense[[:space:]]V2_[[:space:]]A[[:space:]]Unified[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Multi-modal[[:space:]]Remote[[:space:]]Sensing/5fb33625-ede3-48c2-9ec0-3741fd1fba6c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sliced[[:space:]]Wasserstein[[:space:]]Bridge[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation/320731ee-724e-4243-b22e-a9cb1f48c443_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SliderSpace_[[:space:]]Decomposing[[:space:]]the[[:space:]]Visual[[:space:]]Capabilities[[:space:]]of[[:space:]]Diffusion[[:space:]]Models/50d20dec-7e48-4238-88fb-f013284e5706_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SmolDocling_[[:space:]]An[[:space:]]ultra-compact[[:space:]]vision-language[[:space:]]model[[:space:]]for[[:space:]]end-to-end[[:space:]]multi-modal[[:space:]]document[[:space:]]conversion/52fdd715-223a-408c-b30d-b6b134926ef6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Snakes[[:space:]]and[[:space:]]Ladders_[[:space:]]Two[[:space:]]Steps[[:space:]]Up[[:space:]]for[[:space:]]VideoMamba/b11c3116-c63e-4aac-8755-cf4eabafe782_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Social[[:space:]]Debiasing[[:space:]]for[[:space:]]Fair[[:space:]]Multi-modal[[:space:]]LLMs/c46b809c-5d37-4c83-a43f-cbf8d15658b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Soft[[:space:]]Local[[:space:]]Completeness_[[:space:]]Rethinking[[:space:]]Completeness[[:space:]]in[[:space:]]XAI/4e70f0b0-09f1-4a52-a8df-083516f56711_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Soft[[:space:]]Separation[[:space:]]and[[:space:]]Distillation_[[:space:]]Toward[[:space:]]Global[[:space:]]Uniformity[[:space:]]in[[:space:]]Federated[[:space:]]Unsupervised[[:space:]]Learning/a7c7f9f2-5e8e-4c1b-b520-1318204d80e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sparfels_[[:space:]]Fast[[:space:]]Reconstruction[[:space:]]from[[:space:]]Sparse[[:space:]]Unposed[[:space:]]Imagery/e0c26abe-dde6-464c-bb5e-932b397213e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sparse[[:space:]]Fine-Tuning[[:space:]]of[[:space:]]Transformers[[:space:]]for[[:space:]]Generative[[:space:]]Tasks/93b9d533-60f0-4a0a-a330-dae67ada8155_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sparse-Dense[[:space:]]Side-Tuner[[:space:]]for[[:space:]]efficient[[:space:]]Video[[:space:]]Temporal[[:space:]]Grounding/87762bca-719d-433c-abf0-05e54d888e6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SparseFlex_[[:space:]]High-Resolution[[:space:]]and[[:space:]]Arbitrary-Topology[[:space:]]3D[[:space:]]Shape[[:space:]]Modeling/fcf007d6-72eb-465a-b05f-460f5649b8d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SparseLaneSTP_[[:space:]]Leveraging[[:space:]]Spatio-Temporal[[:space:]]Priors[[:space:]]with[[:space:]]Sparse[[:space:]]Transformers[[:space:]]for[[:space:]]3D[[:space:]]Lane[[:space:]]Detection/1eda17b7-8ae2-4be3-89f2-07bc44a3e4eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SparseMM_[[:space:]]Head[[:space:]]Sparsity[[:space:]]Emerges[[:space:]]from[[:space:]]Visual[[:space:]]Concept[[:space:]]Responses[[:space:]]in[[:space:]]MLLMs/9ba0adfc-c24a-4716-85c4-644efaf54a4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SparseRecon_[[:space:]]Neural[[:space:]]Implicit[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]from[[:space:]]Sparse[[:space:]]Views[[:space:]]with[[:space:]]Feature[[:space:]]and[[:space:]]Depth[[:space:]]Consistencies/0728af14-5300-46a3-bf64-7d187d835efb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SparseVILA_[[:space:]]Decoupling[[:space:]]Visual[[:space:]]Sparsity[[:space:]]for[[:space:]]Efficient[[:space:]]VLM[[:space:]]Inference/9b15689c-1721-44a2-b85e-1b1667a45deb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Sparsity[[:space:]]Outperforms[[:space:]]Low-Rank[[:space:]]Projections[[:space:]]in[[:space:]]Few-Shot[[:space:]]Adaptation/6c4b6844-e0ee-4952-860e-16ecf63a3ae1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spatial[[:space:]]Alignment[[:space:]]and[[:space:]]Temporal[[:space:]]Matching[[:space:]]Adapter[[:space:]]for[[:space:]]Video-Radar[[:space:]]Remote[[:space:]]Physiological[[:space:]]Measurement/2a28716b-580e-47e9-adc9-9d410d63b302_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spatial[[:space:]]Preference[[:space:]]Rewarding[[:space:]]for[[:space:]]MLLMs[[:space:]]Spatial[[:space:]]Understanding/7f1d9c24-f4e8-482b-91ef-4b03709c5ebb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spatial-Temporal[[:space:]]Aware[[:space:]]Visuomotor[[:space:]]Diffusion[[:space:]]Policy[[:space:]]Learning/635ea516-8718-4cd2-bbf5-79fa32431478_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spatial-Temporal[[:space:]]Forgery[[:space:]]Trace[[:space:]]based[[:space:]]Forgery[[:space:]]Image[[:space:]]Identification/b7964ec0-870b-444a-822c-785edaff983a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpatialCrafter_[[:space:]]Unleashing[[:space:]]the[[:space:]]Imagination[[:space:]]of[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]from[[:space:]]Limited[[:space:]]Observations/359a0256-c24c-45c3-9f07-0205db934b25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpatialSplat_[[:space:]]Efficient[[:space:]]Semantic[[:space:]]3D[[:space:]]from[[:space:]]Sparse[[:space:]]Unposed[[:space:]]Images/b5246924-8fb7-4786-baa2-29a57614062d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpatialTrackerV2_[[:space:]]Advancing[[:space:]]3D[[:space:]]Point[[:space:]]Tracking[[:space:]]with[[:space:]]Explicit[[:space:]]Camera[[:space:]]Motion/925f4fd6-2e60-44c7-a108-57075a837941_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spatially-Varying[[:space:]]Autofocus/812fa83e-886a-4a5e-b27e-3f560550ef19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spatio-Spectral[[:space:]]Pattern[[:space:]]Illumination[[:space:]]for[[:space:]]Direct[[:space:]]and[[:space:]]Indirect[[:space:]]Separation[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Hyperspectral[[:space:]]Image/c416257e-42cc-4603-837a-a5cb6ea0c724_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpecGuard_[[:space:]]Spectral[[:space:]]Projection-based[[:space:]]Advanced[[:space:]]Invisible[[:space:]]Watermarking/1fcb8d1e-6633-4c96-8cd9-9ceeae669d64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spectral[[:space:]]Image[[:space:]]Tokenizer/2b3586c0-b936-4d11-a38a-2ff8e5859a65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spectral[[:space:]]Sensitivity[[:space:]]Estimation[[:space:]]with[[:space:]]an[[:space:]]Uncalibrated[[:space:]]Diffraction[[:space:]]Grating/79b5704f-97ff-445c-93ff-ff5edaa3c620_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpectralAR_[[:space:]]Spectral[[:space:]]Autoregressive[[:space:]]Visual[[:space:]]Generation/ddfacecc-ae50-4e90-8053-8ccf2e5c55b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Spherical[[:space:]]Epipolar[[:space:]]Rectification[[:space:]]for[[:space:]]Deep[[:space:]]Two-View[[:space:]]Absolute[[:space:]]Depth[[:space:]]Estimation/580cf2a2-db43-4692-a58d-cadb1e36cc3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpiLiFormer_[[:space:]]Enhancing[[:space:]]Spiking[[:space:]]Transformers[[:space:]]with[[:space:]]Lateral[[:space:]]Inhibition/75735d8f-465e-40da-8f9a-a276e0092626_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpikeDiff_[[:space:]]Zero-shot[[:space:]]High-Quality[[:space:]]Video[[:space:]]Reconstruction[[:space:]]from[[:space:]]Chromatic[[:space:]]Spike[[:space:]]Camera[[:space:]]and[[:space:]]Sub-millisecond[[:space:]]Spike[[:space:]]Streams/3f42c09f-75a5-4dd4-bd4a-8b0a20ea82c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpikePack_[[:space:]]Enhanced[[:space:]]Information[[:space:]]Flow[[:space:]]in[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks[[:space:]]with[[:space:]]High[[:space:]]Hardware[[:space:]]Compatibility/f9abc2f2-d2b5-43b8-9f56-aa36381da767_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SpinMeRound_[[:space:]]Consistent[[:space:]]Multi-View[[:space:]]Identity[[:space:]]Generation[[:space:]]Using[[:space:]]Diffusion[[:space:]]Models/d46ce965-2153-45b4-9566-3f0d4d7d97a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SplArt_[[:space:]]Articulation[[:space:]]Estimation[[:space:]]and[[:space:]]Part-Level[[:space:]]Reconstruction[[:space:]]with[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/99f05b1c-e917-486f-913a-4812f9c2d857_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Splat-LOAM_[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]LiDAR[[:space:]]Odometry[[:space:]]and[[:space:]]Mapping/21f438cd-a199-46e8-b668-60b759526f59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Splat-based[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]with[[:space:]]Extreme[[:space:]]Motion-blur/73a5b991-b304-4700-8590-8334fbf2bdee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SplatTalk_[[:space:]]3D[[:space:]]VQA[[:space:]]with[[:space:]]Gaussian[[:space:]]Splatting/71c4eb8c-9be8-49e8-bab3-c4048f8e7d2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Split-and-Combine_[[:space:]]Enhancing[[:space:]]Style[[:space:]]Augmentation[[:space:]]for[[:space:]]Single[[:space:]]Domain[[:space:]]Generalization/3714ebba-fcba-4212-aaf3-03100ed83e52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/St4RTrack_[[:space:]]Simultaneous[[:space:]]4D[[:space:]]Reconstruction[[:space:]]and[[:space:]]Tracking[[:space:]]in[[:space:]]the[[:space:]]World/8996ce18-d088-44bb-a34b-53f80b972348_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stable[[:space:]]Diffusion[[:space:]]Models[[:space:]]are[[:space:]]Secretly[[:space:]]Good[[:space:]]at[[:space:]]Visual[[:space:]]In-Context[[:space:]]Learning/94d38526-8561-492c-b6db-5bcfedbe9083_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stable[[:space:]]Score[[:space:]]Distillation/62930789-4f50-45e0-b2ce-8d1494eced86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stable[[:space:]]Virtual[[:space:]]Camera_[[:space:]]Generative[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/8df1f571-dcd1-4fc3-a74d-5f6f5bf1e485_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stable-Sim2Real_[[:space:]]Exploring[[:space:]]Simulation[[:space:]]of[[:space:]]Real-Captured[[:space:]]3D[[:space:]]Data[[:space:]]with[[:space:]]Two-Stage[[:space:]]Depth[[:space:]]Diffusion/ac3096b4-f6df-4b0e-b9d1-43e46507be43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StableCodec_[[:space:]]Taming[[:space:]]One-Step[[:space:]]Diffusion[[:space:]]for[[:space:]]Extreme[[:space:]]Image[[:space:]]Compression/5269d8ed-d6d6-4961-a311-a0e841365f77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StableDepth_[[:space:]]Scene-Consistent[[:space:]]and[[:space:]]Scale-Invariant[[:space:]]Monocular[[:space:]]Depth/73601462-88ee-4d47-a92f-ac7545019662_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Staining[[:space:]]and[[:space:]]Locking[[:space:]]Computer[[:space:]]Vision[[:space:]]Models[[:space:]]Without[[:space:]]Retraining/6d43a5d3-0678-494f-aae6-04c44e76e1c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Statistical[[:space:]]Confidence[[:space:]]Rescoring[[:space:]]for[[:space:]]Robust[[:space:]]3D[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]from[[:space:]]Multi-View[[:space:]]Images/00c53575-a21f-463b-8889-7a2370d2bff7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StealthAttack_[[:space:]]Robust[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Poisoning[[:space:]]via[[:space:]]Density-Guided[[:space:]]Illusions/b6c26c82-2907-466d-9aed-934611879ad9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stealthy[[:space:]]Backdoor[[:space:]]Attack[[:space:]]in[[:space:]]Federated[[:space:]]Learning[[:space:]]via[[:space:]]Adaptive[[:space:]]Layer-wise[[:space:]]Gradient[[:space:]]Alignment/f7dfb2e6-c547-4b0c-bb2e-4bfc0b02a71f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SteerX_[[:space:]]Creating[[:space:]]Any[[:space:]]Camera-Free[[:space:]]3D[[:space:]]and[[:space:]]4D[[:space:]]Scenes[[:space:]]with[[:space:]]Geometric[[:space:]]Steering/a06f04c2-df21-4c58-83ac-45d3ea205c44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Steering[[:space:]]Guidance[[:space:]]for[[:space:]]Personalized[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/34e2c957-5535-42df-a5b2-c4f7cd9d6450_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stepping[[:space:]]Out[[:space:]]of[[:space:]]Similar[[:space:]]Semantic[[:space:]]Space[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Segmentation/f495e889-2c8e-4e64-ba9d-86b1aec5716d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stereo[[:space:]]Any[[:space:]]Video_[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Stereo[[:space:]]Matching/dc20cc87-36ce-4c69-8c57-afd220728c97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stochastic[[:space:]]Gradient[[:space:]]Estimation[[:space:]]for[[:space:]]Higher-Order[[:space:]]Differentiable[[:space:]]Rendering/95d3584a-ba8b-4c05-9f1b-c00fa69cfc88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stochastic[[:space:]]Interpolants[[:space:]]for[[:space:]]Revealing[[:space:]]Stylistic[[:space:]]Flows[[:space:]]across[[:space:]]the[[:space:]]History[[:space:]]of[[:space:]]Art/16c23549-7e7f-4e3a-8944-314b6bb00228_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StochasticSplats_[[:space:]]Stochastic[[:space:]]Rasterization[[:space:]]for[[:space:]]Sorting-Free[[:space:]]3D[[:space:]]Gaussian[[:space:]]Splatting/d963f664-7a0a-43e1-95cb-e52f6169cd94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StolenLoRA_[[:space:]]Exploring[[:space:]]LoRA[[:space:]]Extraction[[:space:]]Attacks[[:space:]]via[[:space:]]Synthetic[[:space:]]Data/f4ddd460-9cf1-429e-89c2-3d861bf9c661_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Straighten[[:space:]]Viscous[[:space:]]Rectified[[:space:]]Flow[[:space:]]via[[:space:]]Noise[[:space:]]Optimization/7bcafc85-3856-49c2-a845-9dc4fd667b05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StrandHead_[[:space:]]Text[[:space:]]to[[:space:]]Hair-Disentangled[[:space:]]3D[[:space:]]Head[[:space:]]Avatars[[:space:]]Using[[:space:]]Human-Centric[[:space:]]Priors/5c673a06-82fc-4d7f-915c-54daefec29d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StreamDiffusion_[[:space:]]A[[:space:]]Pipeline-level[[:space:]]Solution[[:space:]]for[[:space:]]Real-Time[[:space:]]Interactive[[:space:]]Generation/522f37ba-5716-44cf-af98-f93d9030fe6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StreamGS_[[:space:]]Online[[:space:]]Generalizable[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Reconstruction[[:space:]]for[[:space:]]Unposed[[:space:]]Image[[:space:]]Streams/9d1940fe-237f-4495-9085-9441928f0c47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StreamMind_[[:space:]]Unlocking[[:space:]]Full[[:space:]]Frame[[:space:]]Rate[[:space:]]Streaming[[:space:]]Video[[:space:]]Dialogue[[:space:]]through[[:space:]]Event-Gated[[:space:]]Cognition/61e82f55-a3d1-4244-b3cc-22632c8cab6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Streaming[[:space:]]VideoLLMs[[:space:]]for[[:space:]]Real-Time[[:space:]]Procedural[[:space:]]Video[[:space:]]Understanding/e3ae8cdd-650d-48b8-9247-fe7363ebedec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Streamlining[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Layered[[:space:]]Diffusion[[:space:]]Brushes/6116442a-236c-4dff-8b3c-c87a9184cffb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Street[[:space:]]Gaussians[[:space:]]without[[:space:]]3D[[:space:]]Object[[:space:]]Tracker/ceea65cc-c6ed-4695-b497-f6fc77f3a4c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stroke2Sketch_[[:space:]]Harnessing[[:space:]]Stroke[[:space:]]Attributes[[:space:]]for[[:space:]]Training-Free[[:space:]]Sketch[[:space:]]Generation/88f15fb1-f5ab-4157-a23c-15d3cb3bc8f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stronger,[[:space:]]Steadier[[:space:]]&[[:space:]]Superior_[[:space:]]Geometric[[:space:]]Consistency[[:space:]]in[[:space:]]Depth[[:space:]]VFM[[:space:]]Forges[[:space:]]Domain[[:space:]]Generalized[[:space:]]Semantic[[:space:]]Segmentation/2e6ee106-ae7d-42e3-b798-bd6f97fafe0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StruMamba3D_[[:space:]]Exploring[[:space:]]Structural[[:space:]]Mamba[[:space:]]for[[:space:]]Self-supervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Representation[[:space:]]Learning/9ae9c283-762a-4256-b4de-303cb73e1e2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Structure[[:space:]]Matters_[[:space:]]Revisiting[[:space:]]Boundary[[:space:]]Refinement[[:space:]]in[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/bf659278-d12b-4982-824e-fbf802268f8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Structure-Guided[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Portrait[[:space:]]Shadow[[:space:]]Removal/73c9755b-db1d-47f9-8857-79c13b240555_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Structure-aware[[:space:]]Semantic[[:space:]]Discrepancy[[:space:]]and[[:space:]]Consistency[[:space:]]for[[:space:]]3D[[:space:]]Medical[[:space:]]Image[[:space:]]Self-supervised[[:space:]]Learning/d73adb04-404d-4f0b-9d9c-095f1f245a9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Structured[[:space:]]Policy[[:space:]]Optimization_[[:space:]]Enhance[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Model[[:space:]]via[[:space:]]Self-referenced[[:space:]]Dialogue/5161fa7a-2cd7-4a41-959a-e1ddb02c88dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StyleKeeper_[[:space:]]Prevent[[:space:]]Content[[:space:]]Leakage[[:space:]]using[[:space:]]Negative[[:space:]]Visual[[:space:]]Query[[:space:]]Guidance/45fcd356-f226-4ed7-980b-85432724bf8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StyleMotif_[[:space:]]Multi-Modal[[:space:]]Motion[[:space:]]Stylization[[:space:]]using[[:space:]]Style-Content[[:space:]]Cross[[:space:]]Fusion/ffae00f3-5394-4dd8-b9a4-bc3713aa96e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/StyleSRN_[[:space:]]Scene[[:space:]]Text[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]with[[:space:]]Text[[:space:]]Style[[:space:]]Embedding/34baee7e-bb1d-4173-bb75-506b85eb6a0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Stylized-Face_[[:space:]]A[[:space:]]Million-level[[:space:]]Stylized[[:space:]]Face[[:space:]]Dataset[[:space:]]for[[:space:]]Face[[:space:]]Recognition/fc351eb8-6a0f-465e-ac2f-6fa6da65b294_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SuMa_[[:space:]]A[[:space:]]Subspace[[:space:]]Mapping[[:space:]]Approach[[:space:]]for[[:space:]]Robust[[:space:]]and[[:space:]]Effective[[:space:]]Concept[[:space:]]Erasure[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/51080fa1-f759-4df5-8927-63f76b232e8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Subjective[[:space:]]Camera[[:space:]]1.0_[[:space:]]Bridging[[:space:]]Human[[:space:]]Cognition[[:space:]]and[[:space:]]Visual[[:space:]]Reconstruction[[:space:]]through[[:space:]]Sequence-Aware[[:space:]]Sketch-Guided[[:space:]]Diffusion/cae6ed8f-de57-4610-b479-1652f9260751_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SummDiff_[[:space:]]Generative[[:space:]]Modeling[[:space:]]of[[:space:]]Video[[:space:]]Summarization[[:space:]]with[[:space:]]Diffusion/314d20e1-b991-4676-b961-e90cddf3f9d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Super[[:space:]]Resolved[[:space:]]Imaging[[:space:]]with[[:space:]]Adaptive[[:space:]]Optics/330a7bf8-492d-4b73-b28f-8e96f1e5081a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SuperDec_[[:space:]]3D[[:space:]]Scene[[:space:]]Decomposition[[:space:]]with[[:space:]]Superquadrics[[:space:]]Primitives/ed2eb095-495d-4dd3-9f36-9c1449d1ce9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SuperEdit_[[:space:]]Rectifying[[:space:]]and[[:space:]]Facilitating[[:space:]]Supervision[[:space:]]for[[:space:]]Instruction-Based[[:space:]]Image[[:space:]]Editing/395fe9d1-b3f1-4c45-afff-e5eaba8e26ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SuperEvent_[[:space:]]Cross-Modal[[:space:]]Learning[[:space:]]of[[:space:]]Event-based[[:space:]]Keypoint[[:space:]]Detection[[:space:]]for[[:space:]]SLAM/2e684548-d342-4d0d-9d82-edd690c021a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SuperMat_[[:space:]]Physically[[:space:]]Consistent[[:space:]]PBR[[:space:]]Material[[:space:]]Estimation[[:space:]]at[[:space:]]Interactive[[:space:]]Rates/4daa8d06-6978-409c-9337-f3e4e925bd23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Supercharged[[:space:]]One-step[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Negative[[:space:]]Prompts/1564dd9e-386a-48aa-bdfb-0018fc932f68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Supercharging[[:space:]]Floorplan[[:space:]]Localization[[:space:]]with[[:space:]]Semantic[[:space:]]Rays/834ff1eb-78d1-464a-8d20-1e5cbed96504_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Superpowering[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detectors[[:space:]]for[[:space:]]X-ray[[:space:]]Vision/60987d09-4da2-484e-8cc1-bea5f240eaf8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Supervised[[:space:]]Exploratory[[:space:]]Learning[[:space:]]for[[:space:]]Long-Tailed[[:space:]]Visual[[:space:]]Recognition/3fb70b61-418e-4890-be11-1e54f33e6902_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SurfaceSplat_[[:space:]]Connecting[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]and[[:space:]]Gaussian[[:space:]]Splatting/386f37e7-3405-4869-8ba6-9588babfe21c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SweetTok_[[:space:]]Semantic-Aware[[:space:]]Spatial-Temporal[[:space:]]Tokenizer[[:space:]]for[[:space:]]Compact[[:space:]]Video[[:space:]]Discretization/394730b6-2240-4ddc-9556-9a00295dec81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Switch-a-View_[[:space:]]View[[:space:]]Selection[[:space:]]Learned[[:space:]]from[[:space:]]Unlabeled[[:space:]]In-the-wild[[:space:]]Videos/679bbc29-4d07-48b9-a239-1f0c15065380_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SynAD_[[:space:]]Enhancing[[:space:]]Real-World[[:space:]]End-to-End[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Models[[:space:]]through[[:space:]]Synthetic[[:space:]]Data[[:space:]]Integration/98f750ca-b29d-44cf-a005-9c501051463b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SynCity_[[:space:]]Training-Free[[:space:]]Generation[[:space:]]of[[:space:]]3D[[:space:]]Worlds/0d7bb992-f306-41cb-ba08-811e8364f528_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SynFER_[[:space:]]Towards[[:space:]]Boosting[[:space:]]Facial[[:space:]]Expression[[:space:]]Recognition[[:space:]]with[[:space:]]Synthetic[[:space:]]Data/1b1abe0d-673c-4757-97e7-7262600b5102_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SynTag_[[:space:]]Enhancing[[:space:]]the[[:space:]]Geometric[[:space:]]Robustness[[:space:]]of[[:space:]]Inversion-based[[:space:]]Generative[[:space:]]Image[[:space:]]Watermarking/94eed023-e594-46a7-a0d4-c4c09a3fe5b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/SyncDiff_[[:space:]]Synchronized[[:space:]]Motion[[:space:]]Diffusion[[:space:]]for[[:space:]]Multi-Body[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Synthesis/144abf90-8197-45c7-bc2f-75f04512f099_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Synchronization[[:space:]]of[[:space:]]Multiple[[:space:]]Videos/af4605e7-9aa0-4a23-926a-33856d420d35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Synchronizing[[:space:]]Task[[:space:]]Behavior_[[:space:]]Aligning[[:space:]]Multiple[[:space:]]Tasks[[:space:]]during[[:space:]]Test-Time[[:space:]]Training/60f0e363-9a39-4a78-803f-8a7a065d29c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Synergistic[[:space:]]Prompting[[:space:]]for[[:space:]]Robust[[:space:]]Visual[[:space:]]Recognition[[:space:]]with[[:space:]]Missing[[:space:]]Modalities/23443d08-f2d3-4457-b33f-3d22e53ffdbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Synthesizing[[:space:]]Near-Boundary[[:space:]]OOD[[:space:]]Samples[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection/c3c0a966-f219-49a1-a5df-dd3bae8aeed7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Synthetic[[:space:]]Video[[:space:]]Enhances[[:space:]]Physical[[:space:]]Fidelity[[:space:]]in[[:space:]]Video[[:space:]]Synthesis/59f9a5d1-fc8f-43b3-a954-394a680aed8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/T2Bs_[[:space:]]Text-to-Character[[:space:]]Blendshapes[[:space:]]via[[:space:]]Video[[:space:]]Generation/9a45a79c-5c5c-41c3-b749-b74beda00010_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/T2I-Copilot_[[:space:]]A[[:space:]]Training-Free[[:space:]]Multi-Agent[[:space:]]Text-to-Image[[:space:]]System[[:space:]]for[[:space:]]Enhanced[[:space:]]Prompt[[:space:]]Interpretation[[:space:]]and[[:space:]]Interactive[[:space:]]Generation/211d05eb-e5de-42fe-bf8d-1fbe51b9a0c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TAB_[[:space:]]Transformer[[:space:]]Attention[[:space:]]Bottlenecks[[:space:]]enable[[:space:]]User[[:space:]]Intervention[[:space:]]and[[:space:]]Debugging[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models/c47e8229-4bc7-4c48-877e-ff009c7d300e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TACO_[[:space:]]Taming[[:space:]]Diffusion[[:space:]]for[[:space:]]in-the-wild[[:space:]]Video[[:space:]]Amodal[[:space:]]Completion/9aef2cca-9deb-41e8-af6b-17686afaf49b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TAD-E2E_[[:space:]]A[[:space:]]Large-scale[[:space:]]End-to-end[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Dataset/e34dd6ef-6c84-4df1-9a84-f3853c7b546a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TAG-WM_[[:space:]]Tamper-Aware[[:space:]]Generative[[:space:]]Image[[:space:]]Watermarking[[:space:]]via[[:space:]]Diffusion[[:space:]]Inversion[[:space:]]Sensitivity/5c418f3b-3a84-4af8-86f8-844d737b1ba9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TAPNext_[[:space:]]Tracking[[:space:]]Any[[:space:]]Point[[:space:]](TAP)[[:space:]]as[[:space:]]Next[[:space:]]Token[[:space:]]Prediction/1570d2ea-8c3a-4e4f-8816-939792eaee46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TAR3D_[[:space:]]Creating[[:space:]]High-Quality[[:space:]]3D[[:space:]]Assets[[:space:]]via[[:space:]]Next-Part[[:space:]]Prediction/710fdc15-e281-4215-ba00-df2c8d33d122_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TARO_[[:space:]]Timestep-Adaptive[[:space:]]Representation[[:space:]]Alignment[[:space:]]with[[:space:]]Onset-Aware[[:space:]]Conditioning[[:space:]]for[[:space:]]Synchronized[[:space:]]Video-to-Audio[[:space:]]Synthesis/878e391d-0d53-4519-a75e-ad1361bd9aa4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TARS_[[:space:]]Traffic-Aware[[:space:]]Radar[[:space:]]Scene[[:space:]]Flow[[:space:]]Estimation/117d0b3b-bb15-46f7-a2e5-b92f9ed6f2e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TAViS_[[:space:]]Text-bridged[[:space:]]Audio-Visual[[:space:]]Segmentation[[:space:]]with[[:space:]]Foundation[[:space:]]Models/af82698e-0c3b-4eac-9384-3d90cd78e153_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TCFG_[[:space:]]Truncated[[:space:]]Classifier-Free[[:space:]]Guidance[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Scalable[[:space:]]Text-to-Image[[:space:]]Acceleration/1100162b-a47e-425a-9cfc-342eb2a6a1a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TESPEC_[[:space:]]Temporally-Enhanced[[:space:]]Self-Supervised[[:space:]]Pretraining[[:space:]]for[[:space:]]Event[[:space:]]Cameras/ac958e4a-297e-411f-8fa0-24835cab7ed2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TF-TI2I_[[:space:]]Training-Free[[:space:]]Text-and-Image-to-Image[[:space:]]Generation[[:space:]]via[[:space:]]Multi-Modal[[:space:]]Implicit-Context[[:space:]]Learning[[:space:]]In[[:space:]]Text-to-Image[[:space:]]Models/ed387077-7133-4437-b0c4-28989035bd93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TIP-I2V_[[:space:]]A[[:space:]]Million-Scale[[:space:]]Real[[:space:]]Text[[:space:]]and[[:space:]]Image[[:space:]]Prompt[[:space:]]Dataset[[:space:]]for[[:space:]]Image-to-Video[[:space:]]Generation/9814fe15-47cc-4a76-b295-d61762f914fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TITAN-Guide_[[:space:]]Taming[[:space:]]Inference-Time[[:space:]]Alignment[[:space:]]for[[:space:]]Guided[[:space:]]Text-to-Video[[:space:]]Diffusion[[:space:]]Models/5700097a-6a0d-4b54-99b1-75c9828be4bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TITAN_[[:space:]]Query-Token[[:space:]]based[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Adversarial[[:space:]]Learning/db6df430-7dc9-4f3a-93a3-c318df96384a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TLB-VFI_[[:space:]]Temporal-Aware[[:space:]]Latent[[:space:]]Brownian[[:space:]]Bridge[[:space:]]Diffusion[[:space:]]for[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation/eb4d704c-7bb0-44dd-80a8-10d5c4565d3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TOGA_[[:space:]]Temporally[[:space:]]Grounded[[:space:]]Open-Ended[[:space:]]Video[[:space:]]QA[[:space:]]with[[:space:]]Weak[[:space:]]Supervision/b6f43100-2638-4281-902e-bb9d7cbb9974_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TOTP_[[:space:]]Transferable[[:space:]]Online[[:space:]]Pedestrian[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]with[[:space:]]Temporal-Adaptive[[:space:]]Mamba[[:space:]]Latent[[:space:]]Diffusion/f1896b8b-1f67-4e76-b35a-1a9aca644b3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TPG-INR_[[:space:]]Target[[:space:]]Prior-Guided[[:space:]]Implicit[[:space:]]3D[[:space:]]CT[[:space:]]Reconstruction[[:space:]]for[[:space:]]Enhanced[[:space:]]Sparse-view[[:space:]]Imaging/69e9f8c1-06bd-4da4-a169-66ee558c851b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TR-PTS_[[:space:]]Task-Relevant[[:space:]]Parameter[[:space:]]and[[:space:]]Token[[:space:]]Selection[[:space:]]for[[:space:]]Efficient[[:space:]]Tuning/d171e406-f183-40a1-bfdb-c56ccea215c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TRACE_[[:space:]]Learning[[:space:]]3D[[:space:]]Gaussian[[:space:]]Physical[[:space:]]Dynamics[[:space:]]from[[:space:]]Multi-view[[:space:]]Videos/35c39f29-bbde-4dfe-a438-43609b544789_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TRCE_[[:space:]]Towards[[:space:]]Reliable[[:space:]]Malicious[[:space:]]Concept[[:space:]]Erasure[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/1fcd7e4c-5bc9-42a5-a64f-2b909c7996cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TREAD_[[:space:]]Token[[:space:]]Routing[[:space:]]for[[:space:]]Efficient[[:space:]]Architecture-agnostic[[:space:]]Diffusion[[:space:]]Training/f8e97f29-5ea0-4d4b-ad7e-c79cf279dcc5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TRKT_[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]with[[:space:]]Temporal-enhanced[[:space:]]Relation-aware[[:space:]]Knowledge[[:space:]]Transferring/7904d212-7f29-4708-9b23-841f48cb9b41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TRNAS_[[:space:]]A[[:space:]]Training-Free[[:space:]]Robust[[:space:]]Neural[[:space:]]Architecture[[:space:]]Search/c7b9475c-1445-489b-a85e-6a1917945609_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TWIST[[:space:]]&[[:space:]]SCOUT_[[:space:]]Grounding[[:space:]]Multimodal[[:space:]]LLM-Experts[[:space:]]by[[:space:]]Forget-Free[[:space:]]Tuning/134e44bc-92ea-480b-962b-d0feced2636a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Talking[[:space:]]to[[:space:]]DINO_[[:space:]]Bridging[[:space:]]Self-Supervised[[:space:]]Vision[[:space:]]Backbones[[:space:]]with[[:space:]]Language[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Segmentation/2d318053-4e42-4a22-ab3d-3d9554363088_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Taming[[:space:]]Flow[[:space:]]Matching[[:space:]]with[[:space:]]Unbalanced[[:space:]]Optimal[[:space:]]Transport[[:space:]]into[[:space:]]Fast[[:space:]]Pansharpening/3daa6182-60b7-4cc5-8a3e-4b4f5dd23c2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Taming[[:space:]]the[[:space:]]Untamed_[[:space:]]Graph-Based[[:space:]]Knowledge[[:space:]]Retrieval[[:space:]]and[[:space:]]Reasoning[[:space:]]for[[:space:]]MLLMs[[:space:]]to[[:space:]]Conquer[[:space:]]the[[:space:]]Unknown/072f85e2-8c40-4112-83d1-b7bcbaae60d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Target[[:space:]]Bias[[:space:]]Is[[:space:]]All[[:space:]]You[[:space:]]Need_[[:space:]]Zero-Shot[[:space:]]Debiasing[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]Bias[[:space:]]Corpus/ccbcadb7-2438-45c2-999a-7847b7b6dc5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Task[[:space:]]Vector[[:space:]]Quantization[[:space:]]for[[:space:]]Memory-Efficient[[:space:]]Model[[:space:]]Merging/12770e03-a2a2-413c-af25-77ecbf310064_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Task-Aware[[:space:]]Prompt[[:space:]]Gradient[[:space:]]Projection[[:space:]]for[[:space:]]Parameter-Efficient[[:space:]]Tuning[[:space:]]Federated[[:space:]]Class-Incremental[[:space:]]Learning/330ea2dc-46be-4eea-a815-ff2520dce576_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Task-Decoupled[[:space:]]Bezier[[:space:]]Surface[[:space:]]Constraint[[:space:]]for[[:space:]]Uneven[[:space:]]Low-Light[[:space:]]Image[[:space:]]Enhancement/cf609566-93da-426b-b3e8-e870273eba8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Task-Oriented[[:space:]]Human[[:space:]]Grasp[[:space:]]Synthesis[[:space:]]via[[:space:]]Context-[[:space:]]and[[:space:]]Task-Aware[[:space:]]Diffusers/78084cbf-2e36-4026-8fb6-9a359db5d4a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Task-Specific[[:space:]]Zero-shot[[:space:]]Quantization-Aware[[:space:]]Training[[:space:]]for[[:space:]]Object[[:space:]]Detection/f7694a0b-af92-4f31-9e1b-1cd427138e9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TaxaDiffusion_[[:space:]]Progressively[[:space:]]Trained[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Species[[:space:]]Generation/43f015a6-2b0b-4c2c-b0aa-5872afb41d47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TeEFusion_[[:space:]]Blending[[:space:]]Text[[:space:]]Embeddings[[:space:]]to[[:space:]]Distill[[:space:]]Classifier-Free[[:space:]]Guidance/7b4a5389-5346-4954-8bae-346372dd4918_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TeRA_[[:space:]]Rethinking[[:space:]]Text-guided[[:space:]]Realistic[[:space:]]3D[[:space:]]Avatar[[:space:]]Generation/208df374-ff90-4c93-976c-5d7baa48a4c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Teaching[[:space:]]AI[[:space:]]the[[:space:]]Anatomy[[:space:]]Behind[[:space:]]the[[:space:]]Scan_[[:space:]]Addressing[[:space:]]Anatomical[[:space:]]Flaws[[:space:]]in[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation[[:space:]]with[[:space:]]Learnable[[:space:]]Prior/1d92ab61-2930-488a-9688-8a5e3d69cd5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Teaching[[:space:]]VLMs[[:space:]]to[[:space:]]Localize[[:space:]]Specific[[:space:]]Objects[[:space:]]from[[:space:]]In-context[[:space:]]Examples/95bd80f4-1de1-433d-84f9-403f9f70c8ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Teeth[[:space:]]Reconstruction[[:space:]]and[[:space:]]Performance[[:space:]]Capture[[:space:]]Using[[:space:]]a[[:space:]]Phone[[:space:]]Camera/18c89b72-0459-47cb-9df0-d7f62a471bec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TeethGenerator_[[:space:]]A[[:space:]]two-stage[[:space:]]framework[[:space:]]for[[:space:]]paired[[:space:]]pre-[[:space:]]and[[:space:]]post-orthodontic[[:space:]]3D[[:space:]]dental[[:space:]]data[[:space:]]generation/b2d1d147-d6e7-49c7-973a-f45e15a0c12b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Teleportraits_[[:space:]]Training-Free[[:space:]]People[[:space:]]Insertion[[:space:]]into[[:space:]]Any[[:space:]]Scene/e1167a33-c43a-4294-8dbf-bb5f71014f4c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TemCoCo_[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Multi-modal[[:space:]]Video[[:space:]]Fusion[[:space:]]with[[:space:]]Visual-Semantic[[:space:]]Collaboration/68c19cd2-9acb-4a05-84fd-482318ad0922_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Temperature[[:space:]]in[[:space:]]Cosine-based[[:space:]]Softmax[[:space:]]Loss/88498a8a-8353-489f-9c49-393c2ca7e682_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Temporal[[:space:]]Overlapping[[:space:]]Prediction_[[:space:]]A[[:space:]]Self-supervised[[:space:]]Pre-training[[:space:]]Method[[:space:]]for[[:space:]]LiDAR[[:space:]]Moving[[:space:]]Object[[:space:]]Segmentation/a26d0920-1ecc-4a62-9e6f-658f417c86a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Temporal[[:space:]]Rate[[:space:]]Reduction[[:space:]]Clustering[[:space:]]for[[:space:]]Human[[:space:]]Motion[[:space:]]Segmentation/f14c6fc6-9149-4e8b-a6dd-a5de071b6ce5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Temporal[[:space:]]Unlearnable[[:space:]]Examples_[[:space:]]Preventing[[:space:]]Personal[[:space:]]Video[[:space:]]Data[[:space:]]from[[:space:]]Unauthorized[[:space:]]Exploitation[[:space:]]by[[:space:]]Object[[:space:]]Tracking/fe5a296a-27d1-4006-9167-632435e6a7a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Temporal-aware[[:space:]]Query[[:space:]]Routing[[:space:]]for[[:space:]]Real-time[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation/6956edf3-7339-4603-a3e5-6531a6c8894e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tensor-aggregated[[:space:]]LoRA[[:space:]]in[[:space:]]Federated[[:space:]]Fine-tuning/cfdeda6d-a34d-4c06-a268-6da1c14a07ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TerraMind_[[:space:]]Large-Scale[[:space:]]Generative[[:space:]]Multimodality[[:space:]]for[[:space:]]Earth[[:space:]]Observation/c9c92fd3-e6cf-49ea-bfb4-aa3c5b130a2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Test-Time[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Depth[[:space:]]Completion/0a2ddfb1-a84e-4b84-bcf1-12ab1957047a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Test-Time[[:space:]]Retrieval-Augmented[[:space:]]Adaptation[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/abcad282-0784-41dd-b230-4b9b3488f3c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Test-time[[:space:]]Adaptation[[:space:]]for[[:space:]]Foundation[[:space:]]Medical[[:space:]]Segmentation[[:space:]]Model[[:space:]]Without[[:space:]]Parametric[[:space:]]Updates/8a26d38c-cedb-499b-b41f-c629d5b94654_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Text[[:space:]]Embedding[[:space:]]Knows[[:space:]]How[[:space:]]to[[:space:]]Quantize[[:space:]]Text-Guided[[:space:]]Diffusion[[:space:]]Models/885bbebf-0678-4829-a8d8-7dbca73ecb4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Text-IRSTD_[[:space:]]Leveraging[[:space:]]Semantic[[:space:]]Text[[:space:]]to[[:space:]]Promote[[:space:]]Infrared[[:space:]]Small[[:space:]]Target[[:space:]]Detection[[:space:]]in[[:space:]]Complex[[:space:]]Scenes/f9aefd65-d52a-4731-ba81-d317b6edbcb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Text-guided[[:space:]]Visual[[:space:]]Prompt[[:space:]]DINO[[:space:]]for[[:space:]]Generic[[:space:]]Segmentation/0364b0ce-3511-4546-ac4b-3ddb77cbd170_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Text-to-Any-Skeleton[[:space:]]Motion[[:space:]]Generation[[:space:]]Without[[:space:]]Retargeting/a12eb351-f679-4d5e-8b40-f5c8f124976d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Text2Outfit_[[:space:]]Controllable[[:space:]]Outfit[[:space:]]Generation[[:space:]]with[[:space:]]Multimodal[[:space:]]Language[[:space:]]Models/bbd1b25f-9206-484c-b93d-8a139709e547_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Text2VDM_[[:space:]]Text[[:space:]]to[[:space:]]Vector[[:space:]]Displacement[[:space:]]Maps[[:space:]]for[[:space:]]Expressive[[:space:]]and[[:space:]]Interactive[[:space:]]3D[[:space:]]Sculpting/011b91bd-0c46-4896-9ef0-7e4bb90ad7cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TextMaster_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Realistic[[:space:]]Text[[:space:]]Editing[[:space:]]via[[:space:]]Glyph-Style[[:space:]]Dual-Control/948bb483-aee7-45ff-9c3e-421173343c17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TextSSR_[[:space:]]Diffusion-based[[:space:]]Data[[:space:]]Synthesis[[:space:]]for[[:space:]]Scene[[:space:]]Text[[:space:]]Recognition/6a7bfaa2-52e3-41ce-8b5a-5f025a33ee77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Textured[[:space:]]3D[[:space:]]Regenerative[[:space:]]Morphing[[:space:]]with[[:space:]]3D[[:space:]]Diffusion[[:space:]]Prior/0ee49b9a-7099-4408-8e6f-37bd5bc28baf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Best[[:space:]]of[[:space:]]Both[[:space:]]Worlds_[[:space:]]Integrating[[:space:]]Language[[:space:]]Models[[:space:]]and[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Video[[:space:]]Generation/b1aafd05-162c-4fda-a00d-cb706731d3d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Curse[[:space:]]of[[:space:]]Conditions_[[:space:]]Analyzing[[:space:]]and[[:space:]]Improving[[:space:]]Optimal[[:space:]]Transport[[:space:]]for[[:space:]]Conditional[[:space:]]Flow-Based[[:space:]]Generation/2ff56f19-5f85-4cca-a5d4-ab26816c514c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Devil[[:space:]]is[[:space:]]in[[:space:]]the[[:space:]]Spurious[[:space:]]Correlations_[[:space:]]Boosting[[:space:]]Moment[[:space:]]Retrieval[[:space:]]with[[:space:]]Dynamic[[:space:]]Learning/10bd97eb-949b-4ebe-ab75-25cc73578707_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Inter-Intra[[:space:]]Modal[[:space:]]Measure_[[:space:]]A[[:space:]]Predictive[[:space:]]Lens[[:space:]]on[[:space:]]Fine-Tuning[[:space:]]Outcomes[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models/f4964a00-88d3-4539-871c-39c0ed82d2c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Scalability[[:space:]]of[[:space:]]Simplicity_[[:space:]]Empirical[[:space:]]Analysis[[:space:]]of[[:space:]]Vision-Language[[:space:]]Learning[[:space:]]with[[:space:]]a[[:space:]]Single[[:space:]]Transformer/2b8a94f1-28e8-4218-82c7-da4dba2c45e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Silent[[:space:]]Assistant_[[:space:]]NoiseQuery[[:space:]]as[[:space:]]Implicit[[:space:]]Guidance[[:space:]]for[[:space:]]Goal-Driven[[:space:]]Image[[:space:]]Generation/76b9c942-da58-47e6-8ca6-ac0446eb367e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/The[[:space:]]Source[[:space:]]Image[[:space:]]is[[:space:]]the[[:space:]]Best[[:space:]]Attention[[:space:]]for[[:space:]]Infrared[[:space:]]and[[:space:]]Visible[[:space:]]Image[[:space:]]Fusion/cab6b62f-c3b3-45e2-8877-ae8b2379307d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Thermal[[:space:]]Polarimetric[[:space:]]Multi-view[[:space:]]Stereo/493fd467-1c08-4433-ac25-63db3b58338a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Think[[:space:]]Twice_[[:space:]]Test-Time[[:space:]]Reasoning[[:space:]]for[[:space:]]Robust[[:space:]]CLIP[[:space:]]Zero-Shot[[:space:]]Classification/87c172f1-4134-4138-9793-b4455c2d3c37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TikZero_[[:space:]]Zero-Shot[[:space:]]Text-Guided[[:space:]]Graphics[[:space:]]Program[[:space:]]Synthesis/5102f011-579e-4213-9591-ec3e5e763b92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tile-wise[[:space:]]vs.[[:space:]]Image-wise_[[:space:]]Random-Tile[[:space:]]Loss[[:space:]]and[[:space:]]Training[[:space:]]Paradigm[[:space:]]for[[:space:]]Gaussian[[:space:]]Splatting/e220fb2d-345d-4c10-96de-761db897dc9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tiling[[:space:]]artifacts[[:space:]]and[[:space:]]trade-offs[[:space:]]of[[:space:]]feature[[:space:]]normalization[[:space:]]in[[:space:]]the[[:space:]]segmentation[[:space:]]of[[:space:]]large[[:space:]]biological[[:space:]]images/eaeeea1f-eecf-4d2e-8480-0e89bb127b12_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Time-Aware[[:space:]]Auto[[:space:]]White[[:space:]]Balance[[:space:]]in[[:space:]]Mobile[[:space:]]Photography/59296f3d-511f-4955-b6aa-60113b360f05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TimeBooth_[[:space:]]Disentangled[[:space:]]Facial[[:space:]]Invariant[[:space:]]Representation[[:space:]]for[[:space:]]Diverse[[:space:]]and[[:space:]]Personalized[[:space:]]Face[[:space:]]Aging/b7999e96-8abf-47f1-9b7e-e5c9222126c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TimeExpert_[[:space:]]An[[:space:]]Expert-Guided[[:space:]]Video[[:space:]]LLM[[:space:]]for[[:space:]]Video[[:space:]]Temporal[[:space:]]Grounding/a4a233d0-73c0-4327-bc43-78ef9be737e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TimeFormer_[[:space:]]Capturing[[:space:]]Temporal[[:space:]]Relationships[[:space:]]of[[:space:]]Deformable[[:space:]]3D[[:space:]]Gaussians[[:space:]]for[[:space:]]Robust[[:space:]]Reconstruction/8cdc06bd-0ca5-4f6b-b012-50c8329288a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Timestep-Aware[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Extreme[[:space:]]Image[[:space:]]Rescaling/5b02e9a3-c733-4ab2-9aa6-549d5b3e9de2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TinyViM_[[:space:]]Frequency[[:space:]]Decoupling[[:space:]]for[[:space:]]Tiny[[:space:]]Hybrid[[:space:]]Vision[[:space:]]Mamba/f0ba958e-31c7-4013-97d5-2552433cb070_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/To[[:space:]]Label[[:space:]]or[[:space:]]Not[[:space:]]to[[:space:]]Label_[[:space:]]PALM[[:space:]]-[[:space:]]A[[:space:]]Predictive[[:space:]]Model[[:space:]]for[[:space:]]Evaluating[[:space:]]Sample[[:space:]]Efficiency[[:space:]]in[[:space:]]Active[[:space:]]Learning[[:space:]]Models/8bac4e62-edf8-4833-a2b8-a238e4c4c68b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ToF-Splatting_[[:space:]]Dense[[:space:]]SLAM[[:space:]]using[[:space:]]Sparse[[:space:]]Time-of-Flight[[:space:]]Depth[[:space:]]and[[:space:]]Multi-Frame[[:space:]]Integration/93b627e5-bb27-4dc6-aa8c-e29320f5f5eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Token[[:space:]]Activation[[:space:]]Map[[:space:]]to[[:space:]]Visually[[:space:]]Explain[[:space:]]Multimodal[[:space:]]LLMs/b852d151-66d5-47a7-8861-5fbe9cff3adc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Token-Efficient[[:space:]]VLM_[[:space:]]High-Resolution[[:space:]]Image[[:space:]]Understanding[[:space:]]via[[:space:]]Dynamic[[:space:]]Region[[:space:]]Proposal/b324faad-4fe9-44a7-a833-a82bac07835f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TokenUnify_[[:space:]]Scaling[[:space:]]Up[[:space:]]Autoregressive[[:space:]]Pretraining[[:space:]]for[[:space:]]Neuron[[:space:]]Segmentation/0bebb7e8-7094-4aab-a136-ac9635434593_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TokensGen_[[:space:]]Harnessing[[:space:]]Condensed[[:space:]]Tokens[[:space:]]for[[:space:]]Long[[:space:]]Video[[:space:]]Generation/439118b2-5f82-4c2e-9d89-cceba716e36e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ToolVQA_[[:space:]]A[[:space:]]Dataset[[:space:]]for[[:space:]]Multi-step[[:space:]]Reasoning[[:space:]]VQA[[:space:]]with[[:space:]]External[[:space:]]Tools/dd0c7eba-b554-4e21-9d01-9fb4080c4a97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Top2Pano_[[:space:]]Learning[[:space:]]to[[:space:]]Generate[[:space:]]Indoor[[:space:]]Panoramas[[:space:]]from[[:space:]]Top-Down[[:space:]]View/ddd3f9c4-d03d-4cd7-96eb-1ef2931311e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TopicGeo_[[:space:]]An[[:space:]]Efficient[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Geolocation/43f123eb-2c45-4de4-8a4a-4958a152775f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TopoTTA_[[:space:]]Topology-Enhanced[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Tubular[[:space:]]Structure[[:space:]]Segmentation/a88fe374-66af-4512-b87e-ddbd06fc2c78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TorchAdapt_[[:space:]]Towards[[:space:]]Light-Agnostic[[:space:]]Real-Time[[:space:]]Visual[[:space:]]Perception/60f819c6-a0ae-4fb0-be50-050c7068618b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Toward[[:space:]]Better[[:space:]]Out-painting_[[:space:]]Improving[[:space:]]the[[:space:]]Image[[:space:]]Composition[[:space:]]with[[:space:]]Initialization[[:space:]]Policy[[:space:]]Model/2c55e2cf-c2bf-4623-a4a2-6a1a33a77533_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Toward[[:space:]]Fair[[:space:]]and[[:space:]]Accurate[[:space:]]Cross-Domain[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation_[[:space:]]A[[:space:]]VLM-Driven[[:space:]]Active[[:space:]]Domain[[:space:]]Adaptation[[:space:]]Paradigm/0cab628c-59eb-4cba-b70b-2c418218511f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Toward[[:space:]]Long-Tailed[[:space:]]Online[[:space:]]Anomaly[[:space:]]Detection[[:space:]]through[[:space:]]Class-Agnostic[[:space:]]Concepts/331fec76-66a3-4c9b-96fc-eeb1ae13b1b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Toward[[:space:]]Material-Agnostic[[:space:]]System[[:space:]]Identification[[:space:]]from[[:space:]]Videos/58fcfc32-a696-4a6d-a6cd-3c218994c3ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Accurate[[:space:]]and[[:space:]]Efficient[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving_[[:space:]]A[[:space:]]Mixture[[:space:]]of[[:space:]]Experts[[:space:]]Computing[[:space:]]System[[:space:]]on[[:space:]]Edge/8be189e3-e5ec-449b-9eb4-67198582e1c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]via[[:space:]]Debiased[[:space:]]High-Confidence[[:space:]]Logit[[:space:]]Alignment/be7e63af-2995-4db4-b164-1814191af90b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Annotation-Free[[:space:]]Evaluation_[[:space:]]KPAScore[[:space:]]for[[:space:]]Human[[:space:]]Keypoint[[:space:]]Detection/c6c6dbbf-396b-41fc-b50d-460da67eb4ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Comprehensive[[:space:]]Lecture[[:space:]]Slides[[:space:]]Understanding_[[:space:]]Large-scale[[:space:]]Dataset[[:space:]]and[[:space:]]Effective[[:space:]]Method/58e52446-3fb4-48af-8a99-0aa5fc34cde3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Cross-modal[[:space:]]Backward-compatible[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/5c932632-b5a7-42ba-9baf-b3ca38b774d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Effective[[:space:]]Foundation[[:space:]]Model[[:space:]]Adaptation[[:space:]]for[[:space:]]Extreme[[:space:]]Cross-Domain[[:space:]]Few-Shot[[:space:]]Learning/2b5d04f4-f92c-4faf-b657-95623ea87551_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Efficient[[:space:]]General[[:space:]]Feature[[:space:]]Prediction[[:space:]]in[[:space:]]Masked[[:space:]]Skeleton[[:space:]]Modeling/23b6091d-1a61-4a72-bde7-0fb203b07465_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Explicit[[:space:]]Exoskeleton[[:space:]]for[[:space:]]the[[:space:]]Reconstruction[[:space:]]of[[:space:]]Complicated[[:space:]]3D[[:space:]]Human[[:space:]]Avatars/83fba3ab-8c7e-4045-994b-8f70e4fc8d14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Fine-grained[[:space:]]Interactive[[:space:]]Segmentation[[:space:]]in[[:space:]]Images[[:space:]]and[[:space:]]Videos/b24a62fe-1619-470e-918c-9d3158f1e3f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Foundational[[:space:]]Models[[:space:]]for[[:space:]]Single-Chip[[:space:]]Radar/6536d138-fd2a-4e4b-a53a-8d73d592cae1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Higher[[:space:]]Effective[[:space:]]Rank[[:space:]]in[[:space:]]Parameter-Efficient[[:space:]]Fine-tuning[[:space:]]using[[:space:]]Khatri-Rao[[:space:]]Product/5863579c-c0a2-49e2-85ef-c6b4d99e52c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Human-like[[:space:]]Virtual[[:space:]]Beings_[[:space:]]Simulating[[:space:]]Human[[:space:]]Behavior[[:space:]]in[[:space:]]3D[[:space:]]Scenes/099d59e1-3cf8-46ca-a7c2-5f8c85602d06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Immersive[[:space:]]Human-X[[:space:]]Interaction_[[:space:]]A[[:space:]]Real-Time[[:space:]]Framework[[:space:]]for[[:space:]]Physically[[:space:]]Plausible[[:space:]]Motion[[:space:]]Synthesis/4564e033-7cc9-41ba-b940-20704a40b38a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Long-Horizon[[:space:]]Vision-Language-Action[[:space:]]System_[[:space:]]Reasoning,[[:space:]]Acting[[:space:]]and[[:space:]]Memory/d546fd0e-5080-4957-803b-5e64e0ab9d88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]More[[:space:]]Diverse[[:space:]]and[[:space:]]Challenging[[:space:]]Pre-training[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Learning_[[:space:]]Self-Supervised[[:space:]]Cross[[:space:]]Reconstruction[[:space:]]with[[:space:]]Decoupled[[:space:]]Views/3caf6ed2-ef92-4e47-ab15-82d341d3a49f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Omnimodal[[:space:]]Expressions[[:space:]]and[[:space:]]Reasoning[[:space:]]in[[:space:]]Referring[[:space:]]Audio-Visual[[:space:]]Segmentation/7d9586d8-b2bc-4332-a23e-c79a83d7ffa9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Open-World[[:space:]]Generation[[:space:]]of[[:space:]]Stereo[[:space:]]Images[[:space:]]and[[:space:]]Unsupervised[[:space:]]Matching/2355f2bc-16e1-4dc5-88a7-53284ff42cd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Performance[[:space:]]Consistency[[:space:]]in[[:space:]]Multi-Level[[:space:]]Model[[:space:]]Collaboration/301f5fc3-3aa7-47fa-a32b-5000053a0dec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Privacy-preserved[[:space:]]Pre-training[[:space:]]of[[:space:]]Remote[[:space:]]Sensing[[:space:]]Foundation[[:space:]]Models[[:space:]]with[[:space:]]Federated[[:space:]]Mutual-guidance[[:space:]]Learning/4e8a08de-9e4f-42af-bb77-96df658a142a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Real[[:space:]]Unsupervised[[:space:]]Anomaly[[:space:]]Detection[[:space:]]Via[[:space:]]Confident[[:space:]]Meta-Learning/a98ede28-c496-4a14-ba48-1de754cd27a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Robust[[:space:]]Defense[[:space:]]against[[:space:]]Customization[[:space:]]via[[:space:]]Protective[[:space:]]Perturbation[[:space:]]Resistant[[:space:]]to[[:space:]]Diffusion-based[[:space:]]Purification/51e59d3c-d2c3-455f-a264-80568f0bd092_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Robustness[[:space:]]of[[:space:]]Person[[:space:]]Search[[:space:]]against[[:space:]]Corruptions/e2be793d-1b81-4ea7-b0e6-29ff47c041c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Safer[[:space:]]and[[:space:]]Understandable[[:space:]]Driver[[:space:]]Intention[[:space:]]Prediction/78fc6bd7-c976-4eea-b4c5-8b69ea6c8ae0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Scalable[[:space:]]Spatial[[:space:]]Intelligence[[:space:]]via[[:space:]]2D-to-3D[[:space:]]Data[[:space:]]Lifting/0b171c89-50c7-46cf-acd8-50048508d75c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Stabilized[[:space:]]and[[:space:]]Efficient[[:space:]]Diffusion[[:space:]]Transformers[[:space:]]through[[:space:]]Long-Skip-Connections[[:space:]]with[[:space:]]Spectral[[:space:]]Constraints/18aff051-28c8-4a64-84b7-b80f42fc7df0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Video[[:space:]]Thinking[[:space:]]Test_[[:space:]]A[[:space:]]Holistic[[:space:]]Benchmark[[:space:]]for[[:space:]]Advanced[[:space:]]Video[[:space:]]Reasoning[[:space:]]and[[:space:]]Understanding/5d6cf0c8-e101-4a9c-a41e-1a87a7f8247a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]Visual[[:space:]]Localization[[:space:]]Interoperability_[[:space:]]Cross-Feature[[:space:]]for[[:space:]]Collaborative[[:space:]]Visual[[:space:]]Localization[[:space:]]and[[:space:]]Mapping/6619fc98-74ec-4ad7-b2e5-d52df6cf4356_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]a[[:space:]]3D[[:space:]]Transfer-based[[:space:]]Black-box[[:space:]]Attack[[:space:]]via[[:space:]]Critical[[:space:]]Feature[[:space:]]Guidance/f48ba01b-3570-4f41-b2be-893add74aeb2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]a[[:space:]]Unified[[:space:]]Copernicus[[:space:]]Foundation[[:space:]]Model[[:space:]]for[[:space:]]Earth[[:space:]]Vision/66f2b950-975c-463c-8ab5-2e970fd9de61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]a[[:space:]]Universal[[:space:]]3D[[:space:]]Medical[[:space:]]Multi-modality[[:space:]]Generalization[[:space:]]via[[:space:]]Learning[[:space:]]Personalized[[:space:]]Invariant[[:space:]]Representation/bb68ec54-5128-4345-82f3-1c6d57dbd79b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Towards[[:space:]]a[[:space:]]Universal[[:space:]]Image[[:space:]]Degradation[[:space:]]Model[[:space:]]via[[:space:]]Content-Degradation[[:space:]]Disentanglement/132f7b63-3c80-4e29-b71c-12629d565072_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Trace3D_[[:space:]]Consistent[[:space:]]Segmentation[[:space:]]Lifting[[:space:]]via[[:space:]]Gaussian[[:space:]]Instance[[:space:]]Tracing/0cc1a493-961b-4f54-a730-2b271c982231_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tracing[[:space:]]Copied[[:space:]]Pixels[[:space:]]and[[:space:]]Regularizing[[:space:]]Patch[[:space:]]Affinity[[:space:]]in[[:space:]]Copy[[:space:]]Detection/9849795a-3297-49e7-b571-2752f89ff9b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TrackAny3D_[[:space:]]Transferring[[:space:]]Pretrained[[:space:]]3D[[:space:]]Models[[:space:]]for[[:space:]]Category-unified[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Tracking/b8270501-0dc2-4e54-a7bf-9436bd8455db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TrackVerse_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Object-Centric[[:space:]]Video[[:space:]]Dataset[[:space:]]for[[:space:]]Image-Level[[:space:]]Representation[[:space:]]Learning/40a64841-8da2-42b7-9621-1097b3664186_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tracking[[:space:]]Tiny[[:space:]]Drones[[:space:]]against[[:space:]]Clutter_[[:space:]]Large-Scale[[:space:]]Infrared[[:space:]]Benchmark[[:space:]]with[[:space:]]Motion-Centric[[:space:]]Adaptive[[:space:]]Algorithm/dbc967af-c8f8-41e6-aa8b-99ff344964d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Trade-offs[[:space:]]in[[:space:]]Image[[:space:]]Generation_[[:space:]]How[[:space:]]Do[[:space:]]Different[[:space:]]Dimensions[[:space:]]Interact_/10a5004e-ed73-49de-a1be-0a276f576fbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TrafficLoc_[[:space:]]Localizing[[:space:]]Traffic[[:space:]]Surveillance[[:space:]]Cameras[[:space:]]in[[:space:]]3D[[:space:]]Scenes/9b1112a2-23b7-4cfa-a27b-ef9713efdc41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Training-Free[[:space:]]Class[[:space:]]Purification[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/a596aa52-a563-40d9-a137-a01edbe2b11d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Training-Free[[:space:]]Industrial[[:space:]]Defect[[:space:]]Generation[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/9e5b7aa9-d7d8-487b-9879-649fcde9e5f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Training-Free[[:space:]]Personalization[[:space:]]via[[:space:]]Retrieval[[:space:]]and[[:space:]]Reasoning[[:space:]]on[[:space:]]Fingerprints/62ff830b-f652-458d-8077-3f6b1728e7f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Training-Free[[:space:]]Text-Guided[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Visual[[:space:]]Autoregressive[[:space:]]Model/aec6aa31-a674-4760-af58-3364247c6a92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Training-free[[:space:]]Generation[[:space:]]of[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Rewards[[:space:]]from[[:space:]]VLMs/7b1ff12e-cb6a-40b6-b044-96f95b4c4a63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Training-free[[:space:]]Geometric[[:space:]]Image[[:space:]]Editing[[:space:]]on[[:space:]]Diffusion[[:space:]]Models/6eadb843-68bf-4e8a-93b9-d9f59f84de0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Training-free[[:space:]]and[[:space:]]Adaptive[[:space:]]Sparse[[:space:]]Attention[[:space:]]for[[:space:]]Efficient[[:space:]]Long[[:space:]]Video[[:space:]]Generation/7c811e6b-af3e-4146-8962-68d025e67c25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TrajectoryCrafter_[[:space:]]Redirecting[[:space:]]Camera[[:space:]]Trajectory[[:space:]]for[[:space:]]Monocular[[:space:]]Videos[[:space:]]via[[:space:]]Diffusion[[:space:]]Models/a08ab0bf-5812-470e-bc64-4fb913fb6f09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Trans-Adapter_[[:space:]]A[[:space:]]Plug-and-Play[[:space:]]Framework[[:space:]]for[[:space:]]Transparent[[:space:]]Image[[:space:]]Inpainting/4a4b4d6a-69d7-4656-9d66-0743fc01f956_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Transformed[[:space:]]Low-rank[[:space:]]Adaptation[[:space:]]via[[:space:]]Tensor[[:space:]]Decomposition[[:space:]]and[[:space:]]Its[[:space:]]Applications[[:space:]]to[[:space:]]Text-to-image[[:space:]]Models/f493fb23-b16d-48fa-828f-3139ac0a189d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Transformer-based[[:space:]]Tooth[[:space:]]Alignment[[:space:]]Prediction[[:space:]]with[[:space:]]Occlusion[[:space:]]and[[:space:]]Collision[[:space:]]Constraints/91714285-fb54-4b24-bc5a-a0a1a6c06e7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TransiT_[[:space:]]Transient[[:space:]]Transformer[[:space:]]for[[:space:]]Non-line-of-sight[[:space:]]Videography/ce1e8b1e-20a2-4ef6-81a6-061d72d3a378_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Translation[[:space:]]of[[:space:]]Text[[:space:]]Embedding[[:space:]]via[[:space:]]Delta[[:space:]]Vector[[:space:]]to[[:space:]]Suppress[[:space:]]Strongly[[:space:]]Entangled[[:space:]]Content[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/4bf848a7-1a19-4877-93c6-82e3d1e8524e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Transparent[[:space:]]Vision_[[:space:]]A[[:space:]]Theory[[:space:]]of[[:space:]]Hierarchical[[:space:]]Invariant[[:space:]]Representations/5905193b-d457-4402-96b3-5830dd2c9411_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tree[[:space:]]Skeletonization[[:space:]]from[[:space:]]3D[[:space:]]Point[[:space:]]Clouds[[:space:]]by[[:space:]]Denoising[[:space:]]Diffusion/fd30c149-5fab-46d3-9884-62e9958d6b2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tree-NeRV_[[:space:]]Efficient[[:space:]]Non-Uniform[[:space:]]Sampling[[:space:]]for[[:space:]]Neural[[:space:]]Video[[:space:]]Representation[[:space:]]via[[:space:]]Tree-Structured[[:space:]]Feature[[:space:]]Grids/d97556f2-d762-4ece-87f0-4c6e09db2118_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TriDi_[[:space:]]Trilateral[[:space:]]Diffusion[[:space:]]of[[:space:]]3D[[:space:]]Humans,[[:space:]]Objects,[[:space:]]and[[:space:]]Interactions/29a2e2f9-097f-4177-b6ff-6ce33538487a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Triad_[[:space:]]Empowering[[:space:]]LMM-based[[:space:]]Anomaly[[:space:]]Detection[[:space:]]with[[:space:]]Expert-guided[[:space:]]Region-of-Interest[[:space:]]Tokenizer[[:space:]]and[[:space:]]Manufacturing[[:space:]]Process/6e8e4dd7-d19a-4c8c-86bc-72bd961ba497_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Trial-Oriented[[:space:]]Visual[[:space:]]Rearrangement/b0bd2e3f-6fae-4694-aff8-f088f8be239d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Trokens_[[:space:]]Semantic-Aware[[:space:]]Relational[[:space:]]Trajectory[[:space:]]Tokens[[:space:]]for[[:space:]]Few-Shot[[:space:]]Action[[:space:]]Recognition/5371ee09-662c-4b59-a273-a536d7a26afd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Trust[[:space:]]but[[:space:]]Verify_[[:space:]]Programmatic[[:space:]]VLM[[:space:]]Evaluation[[:space:]]in[[:space:]]the[[:space:]]Wild/92677331-1d56-4397-927a-e369a7c3a35f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TrustMark_[[:space:]]Robust[[:space:]]Watermarking[[:space:]]and[[:space:]]Watermark[[:space:]]Removal[[:space:]]for[[:space:]]Arbitrary[[:space:]]Resolution[[:space:]]Images/22c48031-2c56-4af2-9a2d-a8d9fa479123_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TruthPrInt_[[:space:]]Mitigating[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Models[[:space:]]Object[[:space:]]Hallucination[[:space:]]Via[[:space:]]Latent[[:space:]]Truthful-Guided[[:space:]]Pre-Intervention/b201ed13-f03a-484c-a46c-d68f884310d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TryOn-Refiner_[[:space:]]Conditional[[:space:]]Rectified-flow-based[[:space:]]TryOn[[:space:]]Refiner[[:space:]]for[[:space:]]More[[:space:]]Accurate[[:space:]]Detail[[:space:]]Reconstruction/45412c4a-69e4-46c1-8ca4-49949e18dc90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Tune-Your-Style_[[:space:]]Intensity-tunable[[:space:]]3D[[:space:]]Style[[:space:]]Transfer[[:space:]]with[[:space:]]Gaussian[[:space:]]Splatting/d3fb50e5-d2c3-43bd-b9b9-027cf9920f99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Turbo2K_[[:space:]]Towards[[:space:]]Ultra-Efficient[[:space:]]and[[:space:]]High-Quality[[:space:]]2K[[:space:]]Video[[:space:]]Synthesis/886fbc42-539a-474b-9403-39a131a3c0fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TurboReg_[[:space:]]TurboClique[[:space:]]for[[:space:]]Robust[[:space:]]and[[:space:]]Efficient[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/8f69f271-b4fa-440a-9251-a43872c96afa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TurboTrain_[[:space:]]Towards[[:space:]]Efficient[[:space:]]and[[:space:]]Balanced[[:space:]]Multi-Task[[:space:]]Learning[[:space:]]for[[:space:]]Multi-Agent[[:space:]]Perception[[:space:]]and[[:space:]]Prediction/1c97c7a4-0e96-445c-a586-d25226d759ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/TurboVSR_[[:space:]]Fantastic[[:space:]]Video[[:space:]]Upscalers[[:space:]]and[[:space:]]Where[[:space:]]to[[:space:]]Find[[:space:]]Them/36afc19e-5e2d-4874-a9f1-c08d0193802a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Two[[:space:]]Losses,[[:space:]]One[[:space:]]Goal_[[:space:]]Balancing[[:space:]]Conflict[[:space:]]Gradients[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Semantic[[:space:]]Segmentation/6fc6490b-2508-4234-ba59-b9cb28d2f1d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/U-ViLAR_[[:space:]]Uncertainty-Aware[[:space:]]Visual[[:space:]]Localization[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving[[:space:]]via[[:space:]]Differentiable[[:space:]]Association[[:space:]]and[[:space:]]Registration/ab6cb74e-e58d-4c83-a73f-709706fcfda7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UAVScenes_[[:space:]]A[[:space:]]Multi-Modal[[:space:]]Dataset[[:space:]]for[[:space:]]UAVs/0a5edaef-b38d-447a-9d4f-6cc303315a26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UDC-VIT_[[:space:]]A[[:space:]]Real-World[[:space:]]Video[[:space:]]Dataset[[:space:]]for[[:space:]]Under-Display[[:space:]]Cameras/41b8f13b-1842-4cfa-ab83-0c5b376cad9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UINavBench_[[:space:]]A[[:space:]]Framework[[:space:]]for[[:space:]]Comprehensive[[:space:]]Evaluation[[:space:]]of[[:space:]]Interactive[[:space:]]Digital[[:space:]]Agents/63c1e146-2caa-4b0c-9225-6b122aa9ef71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UIP2P_[[:space:]]Unsupervised[[:space:]]Instruction-based[[:space:]]Image[[:space:]]Editing[[:space:]]via[[:space:]]Edit[[:space:]]Reversibility[[:space:]]Constraint/4ff8e2fe-2674-4436-bfd3-5ac3e6bcebf6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UIPro_[[:space:]]Unleashing[[:space:]]Superior[[:space:]]Interaction[[:space:]]Capability[[:space:]]For[[:space:]]GUI[[:space:]]Agents/1ff43e17-165d-4b9d-b585-6f01cfeaf196_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UKBOB_[[:space:]]One[[:space:]]Billion[[:space:]]MRI[[:space:]]Labeled[[:space:]]Masks[[:space:]]for[[:space:]]Generalizable[[:space:]]3D[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/ce36597f-5431-48a5-912f-f53356d931c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ULTHO_[[:space:]]Ultra-Lightweight[[:space:]]yet[[:space:]]Efficient[[:space:]]Hyperparameter[[:space:]]Optimization[[:space:]]in[[:space:]]Deep[[:space:]]Reinforcement[[:space:]]Learning/033d4870-fc0c-4cf1-8631-72878b0c8766_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UMDATrack_[[:space:]]Unified[[:space:]]Multi-Domain[[:space:]]Adaptive[[:space:]]Tracking[[:space:]]Under[[:space:]]Adverse[[:space:]]Weather[[:space:]]Conditions/c458f856-58f5-41c1-a32e-36c3aa3f6c71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UNIS_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Achieving[[:space:]]Unbiased[[:space:]]Neural[[:space:]]Implicit[[:space:]]Surfaces[[:space:]]in[[:space:]]Volume[[:space:]]Rendering/2be466ea-ae25-48a4-a5b5-4eb564276457_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UPP_[[:space:]]Unified[[:space:]]Point-Level[[:space:]]Prompting[[:space:]]for[[:space:]]Robust[[:space:]]Point[[:space:]]Cloud[[:space:]]Analysis/dc90cce7-8792-4de3-9c0d-aa7f41a453fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UPRE_[[:space:]]Zero-Shot[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Unified[[:space:]]Prompt[[:space:]]and[[:space:]]Representation[[:space:]]Enhancement/ee84fafe-b2bc-4169-b41e-90567f2ae7c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/USP_[[:space:]]Unified[[:space:]]Self-Supervised[[:space:]]Pretraining[[:space:]]for[[:space:]]Image[[:space:]]Generation[[:space:]]and[[:space:]]Understanding/c136d3d3-2e0d-4d3b-9d4c-71ac3176d086_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UST-SSM_[[:space:]]Unified[[:space:]]Spatio-Temporal[[:space:]]State[[:space:]]Space[[:space:]]Models[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Video[[:space:]]Modeling/5c731df1-1167-4dce-95bc-19c244c61c9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ultra[[:space:]]High-Resolution[[:space:]]Image[[:space:]]Inpainting[[:space:]]with[[:space:]]Patch-Based[[:space:]]Content[[:space:]]Consistency[[:space:]]Adapter/f2baa4f7-f70f-4eba-b321-886a0972ab35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Ultra-Precision[[:space:]]6DoF[[:space:]]Pose[[:space:]]Estimation[[:space:]]Using[[:space:]]2-D[[:space:]]Interpolated[[:space:]]Discrete[[:space:]]Fourier[[:space:]]Transform/332233da-82be-4951-b6e9-b674b614759b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UnMix-NeRF_[[:space:]]Spectral[[:space:]]Unmixing[[:space:]]Meets[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/fd59934c-c695-47fa-b6ef-c7adad7dcf75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UnZipLoRA_[[:space:]]Separating[[:space:]]Content[[:space:]]and[[:space:]]Style[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/61454447-4d6f-4dc3-a032-88a30362e7e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unbiased[[:space:]]Missing-modality[[:space:]]Multimodal[[:space:]]Learning/5c9766e1-467c-4ff3-aa89-851f9ed98d48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unbiased[[:space:]]Region-Language[[:space:]]Alignment[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Dense[[:space:]]Prediction/c3d8ee55-8f3a-4c49-bfdc-17f4c2fee987_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Uncalibrated[[:space:]]Structure[[:space:]]from[[:space:]]Motion[[:space:]]on[[:space:]]a[[:space:]]Sphere/c152157a-c310-4ff9-8269-0155166a7163_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Uncertainty-Aware[[:space:]]Diffusion-Guided[[:space:]]Refinement[[:space:]]of[[:space:]]3D[[:space:]]Scenes/ec092315-f630-4c54-8c3f-0e910afd00a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Uncertainty-Aware[[:space:]]Gradient[[:space:]]Stabilization[[:space:]]for[[:space:]]Small[[:space:]]Object[[:space:]]Detection/28c6ee2f-ebb1-4e78-9084-2c2a3fa42057_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Uncertainty-Driven[[:space:]]Expert[[:space:]]Control_[[:space:]]Enhancing[[:space:]]the[[:space:]]Reliability[[:space:]]of[[:space:]]Medical[[:space:]]Vision-Language[[:space:]]Models/db1bd940-1a36-4181-896a-6b15d8f7326c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Uncover[[:space:]]Treasures[[:space:]]in[[:space:]]DCT_[[:space:]]Advancing[[:space:]]JPEG[[:space:]]Quality[[:space:]]Enhancement[[:space:]]by[[:space:]]Exploiting[[:space:]]Latent[[:space:]]Correlations/175ad537-877d-4cde-8a5d-957c10c04a04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Understanding[[:space:]]Co-speech[[:space:]]Gestures[[:space:]]in-the-wild/bf1e54ff-5e7b-4333-9ca3-ca0c877a6d75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Understanding[[:space:]]Flatness[[:space:]]in[[:space:]]Generative[[:space:]]Models_[[:space:]]Its[[:space:]]Role[[:space:]]and[[:space:]]Benefits/22f15534-014e-4320-bc67-09d3fbfe0de6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Understanding[[:space:]]Museum[[:space:]]Exhibits[[:space:]]using[[:space:]]Vision-Language[[:space:]]Reasoning/83ea511a-50fd-4886-ad4b-cd7c159b09f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Understanding[[:space:]]Personal[[:space:]]Concept[[:space:]]in[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation/38842030-bb13-4482-9b8a-3db2c49696fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Underwater[[:space:]]Visual[[:space:]]SLAM[[:space:]]with[[:space:]]Depth[[:space:]]Uncertainty[[:space:]]and[[:space:]]Medium[[:space:]]Modeling/75a7a7c7-150a-4b4e-b246-014114ab5a3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unfolding-Associative[[:space:]]Encoder-Decoder[[:space:]]Network[[:space:]]with[[:space:]]Progressive[[:space:]]Alignment[[:space:]]for[[:space:]]Pansharpening/dea6dc73-04a7-4048-8126-184a836b9cbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniCombine_[[:space:]]Unified[[:space:]]Multi-Conditional[[:space:]]Combination[[:space:]]with[[:space:]]Diffusion[[:space:]]Transformer/6cd6acf3-fe58-482b-a290-331d76a8f819_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniConvNet_[[:space:]]Expanding[[:space:]]Effective[[:space:]]Receptive[[:space:]]Field[[:space:]]while[[:space:]]Maintaining[[:space:]]Asymptotically[[:space:]]Gaussian[[:space:]]Distribution[[:space:]]for[[:space:]]ConvNets[[:space:]]of[[:space:]]Any[[:space:]]Scale/58ef4922-0dae-41e3-923a-b03102ac6ea9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniDxMD_[[:space:]]Towards[[:space:]]Unified[[:space:]]Representation[[:space:]]for[[:space:]]Cross-Modal[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]in[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation/837ba1a0-04ec-4b7d-84c2-46c1bd611503_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniEgoMotion_[[:space:]]A[[:space:]]Unified[[:space:]]Model[[:space:]]for[[:space:]]Egocentric[[:space:]]Motion[[:space:]]Reconstruction,[[:space:]]Forecasting,[[:space:]]and[[:space:]]Generation/8c434bc3-1ea7-48a1-90bd-41ece98501f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniFuse_[[:space:]]A[[:space:]]Unified[[:space:]]All-in-One[[:space:]]Framework[[:space:]]for[[:space:]]Multi-Modal[[:space:]]Medical[[:space:]]Image[[:space:]]Fusion[[:space:]]Under[[:space:]]Diverse[[:space:]]Degradations[[:space:]]and[[:space:]]Misalignments/35a52f87-645d-490c-96c2-6e533c9eada2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniGS_[[:space:]]Modeling[[:space:]]Unitary[[:space:]]3D[[:space:]]Gaussians[[:space:]]for[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]from[[:space:]]Sparse-view[[:space:]]Images/de178fcf-09c8-4523-b58f-685e6515e14b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniGlyph_[[:space:]]Unified[[:space:]]Segmentation-Conditioned[[:space:]]Diffusion[[:space:]]for[[:space:]]Precise[[:space:]]Visual[[:space:]]Text[[:space:]]Synthesis/b7f8af7d-cfa2-4609-9f7b-3068e41b027c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniMLVG_[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Multi-view[[:space:]]Long[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Comprehensive[[:space:]]Control[[:space:]]Capabilities[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/594fdb6c-496d-4fff-b6cb-f865aa077208_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniOcc_[[:space:]]A[[:space:]]Unified[[:space:]]Benchmark[[:space:]]for[[:space:]]Occupancy[[:space:]]Forecasting[[:space:]]and[[:space:]]Prediction[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/9430b682-b441-42d0-8475-c301d6fe3285_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniPhys_[[:space:]]Unified[[:space:]]Planner[[:space:]]and[[:space:]]Controller[[:space:]]with[[:space:]]Diffusion[[:space:]]for[[:space:]]Flexible[[:space:]]Physics-Based[[:space:]]Character[[:space:]]Control/53b5a63a-c2eb-4774-8e65-239c7f756b0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniPortrait_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Identity-Preserving[[:space:]]Single-[[:space:]]and[[:space:]]Multi-Human[[:space:]]Image[[:space:]]Personalization/ac1f4c8d-b706-476c-b724-873073e43fb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniRes_[[:space:]]Universal[[:space:]]Image[[:space:]]Restoration[[:space:]]for[[:space:]]Complex[[:space:]]Degradations/e485eedd-fd58-42d3-afd0-43a3b0a580cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniVG_[[:space:]]A[[:space:]]Generalist[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Unified[[:space:]]Image[[:space:]]Generation[[:space:]]and[[:space:]]Editing/f9139875-70ce-414f-997b-7b926147f5f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniVerse_[[:space:]]Unleashing[[:space:]]the[[:space:]]Scene[[:space:]]Prior[[:space:]]of[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Robust[[:space:]]Radiance[[:space:]]Field[[:space:]]Reconstruction/1113cb10-817b-436f-9eb6-a3b052766c61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unified[[:space:]]Adversarial[[:space:]]Augmentation[[:space:]]for[[:space:]]Improving[[:space:]]Palmprint[[:space:]]Recognition/1c95d82e-e9bf-409e-b6a7-8bd3468599ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unified[[:space:]]Category-Level[[:space:]]Object[[:space:]]Detection[[:space:]]and[[:space:]]Pose[[:space:]]Estimation[[:space:]]from[[:space:]]RGB[[:space:]]Images[[:space:]]using[[:space:]]3D[[:space:]]Prototypes/5a18f1d0-5854-4d52-9b3d-42b8f051f9c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unified[[:space:]]Multi-Agent[[:space:]]Trajectory[[:space:]]Modeling[[:space:]]with[[:space:]]Masked[[:space:]]Trajectory[[:space:]]Diffusion/08038fe6-7601-45f0-b544-974df2316e72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unified[[:space:]]Multimodal[[:space:]]Understanding[[:space:]]via[[:space:]]Byte-Pair[[:space:]]Visual[[:space:]]Encoding/2a101af7-25d6-43b6-aca5-d28a774d787a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unified[[:space:]]Open-World[[:space:]]Segmentation[[:space:]]with[[:space:]]Multi-Modal[[:space:]]Prompts/dfd9eff2-fa35-42fe-90a0-89d7c1f707fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unified[[:space:]]Video[[:space:]]Generation[[:space:]]via[[:space:]]Next-Set[[:space:]]Prediction[[:space:]]in[[:space:]]Continuous[[:space:]]Domain/d5450109-c459-4393-bfbc-3a6281b76964_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UniversalBooth_[[:space:]]Model-Agnostic[[:space:]]Personalized[[:space:]]Text-to-Image[[:space:]]Generation/5ed1061d-3c1f-4e57-b276-0976e7c57035_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unknown[[:space:]]Text[[:space:]]Learning[[:space:]]for[[:space:]]CLIP-based[[:space:]]Few-Shot[[:space:]]Open-set[[:space:]]Recognition/8df0b349-1f30-4db7-a5dd-d902f9f32a3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unlearning[[:space:]]the[[:space:]]Noisy[[:space:]]Correspondence[[:space:]]Makes[[:space:]]CLIP[[:space:]]More[[:space:]]Robust/12c7b5e4-01f6-4c86-937c-f87c5fd35d7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unleashing[[:space:]]High-Quality[[:space:]]Image[[:space:]]Generation[[:space:]]in[[:space:]]Diffusion[[:space:]]Sampling[[:space:]]Using[[:space:]]Second-Order[[:space:]]Levenberg-Marquardt-Langevin/13cbf79c-7d32-4412-9eee-95252f5be1a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unleashing[[:space:]]Vecset[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Fast[[:space:]]Shape[[:space:]]Generation/7b74d35d-3a86-45b7-bf2b-07600548528f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unleashing[[:space:]]the[[:space:]]Temporal[[:space:]]Potential[[:space:]]of[[:space:]]Stereo[[:space:]]Event[[:space:]]Cameras[[:space:]]for[[:space:]]Continuous-Time[[:space:]]3D[[:space:]]Object[[:space:]]Detection/32355f2d-a46d-49ce-b53e-d6bb8c2598ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unlocking[[:space:]]Constraints_[[:space:]]Source-Free[[:space:]]Occlusion-Aware[[:space:]]Seamless[[:space:]]Segmentation/918ec0ce-e376-4559-8192-30fc256f6bbd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unlocking[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Diffusion[[:space:]]Priors[[:space:]]in[[:space:]]Blind[[:space:]]Face[[:space:]]Restoration/29e022b8-6392-48f4-9ca2-0f6888716bc8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unraveling[[:space:]]the[[:space:]]Effects[[:space:]]of[[:space:]]Synthetic[[:space:]]Data[[:space:]]on[[:space:]]End-to-End[[:space:]]Autonomous[[:space:]]Driving/104f60be-097f-4b14-aff6-1eef2fd1a396_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unraveling[[:space:]]the[[:space:]]Smoothness[[:space:]]Properties[[:space:]]of[[:space:]]Diffusion[[:space:]]Models_[[:space:]]A[[:space:]]Gaussian[[:space:]]Mixture[[:space:]]Perspective/feb61fb0-3564-4496-9696-d1a9cebb7313_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UnrealZoo_[[:space:]]Enriching[[:space:]]Photo-realistic[[:space:]]Virtual[[:space:]]Worlds[[:space:]]for[[:space:]]Embodied[[:space:]]AI/6e3758e9-8534-4600-a9c9-4602b92c4dd7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unsupervised[[:space:]]Histopathological[[:space:]]Image[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Overlapping[[:space:]]Patches[[:space:]]Consistency[[:space:]]Constraint/c7ca9d94-b46a-45fa-8735-814dc1aec4fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unsupervised[[:space:]]Identification[[:space:]]of[[:space:]]Protein[[:space:]]Compositions[[:space:]]and[[:space:]]Conformations[[:space:]]via[[:space:]]Implicit[[:space:]]Content-Transformation[[:space:]]Disentanglement/b2d6ee64-c697-4606-86ea-b8852b053820_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unsupervised[[:space:]]Imaging[[:space:]]Inverse[[:space:]]Problems[[:space:]]with[[:space:]]Diffusion[[:space:]]Distribution[[:space:]]Matching/463f2cdf-f627-4efe-affd-1ebc9f1c243d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unsupervised[[:space:]]Joint[[:space:]]Learning[[:space:]]of[[:space:]]Optical[[:space:]]Flow[[:space:]]and[[:space:]]Intensity[[:space:]]with[[:space:]]Event[[:space:]]Cameras/92edc20b-e741-4fac-9960-c5dbfdfad535_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unsupervised[[:space:]]Part[[:space:]]Discovery[[:space:]]via[[:space:]]Descriptor-Based[[:space:]]Masked[[:space:]]Image[[:space:]]Restoration[[:space:]]with[[:space:]]Optimized[[:space:]]Constraints/ccf22bfc-f454-4ee8-984d-74df574ba826_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unsupervised[[:space:]]RGB-D[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration[[:space:]]for[[:space:]]Scenes[[:space:]]with[[:space:]]Low[[:space:]]Overlap[[:space:]]and[[:space:]]Photometric[[:space:]]Inconsistency/d34c24f1-e608-47c5-8100-5b5de8083fff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unsupervised[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-identification[[:space:]]under[[:space:]]Unpaired[[:space:]]Settings/8bad0627-426a-4c79-849d-e7d54bb417e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unsupervised[[:space:]]Visual[[:space:]]Chain-of-Thought[[:space:]]Reasoning[[:space:]]via[[:space:]]Preference[[:space:]]Optimization/2dfcc225-7530-4e71-9b57-14637900f483_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Unveiling[[:space:]]the[[:space:]]Invisible_[[:space:]]Reasoning[[:space:]]Complex[[:space:]]Occlusions[[:space:]]Amodally[[:space:]]with[[:space:]]AURA/5862fa00-d00b-4bf4-b924-a02792140ae7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/UrbanLLaVA_[[:space:]]A[[:space:]]Multi-modal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Urban[[:space:]]Intelligence/04c8cc16-b9e6-46a8-b406-679b33839262_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/V.I.P.[[:space:]]_[[:space:]]Iterative[[:space:]]Online[[:space:]]Preference[[:space:]]Distillation[[:space:]]for[[:space:]]Efficient[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/dd5c55db-7a1c-4c99-95eb-8d1eafcb989f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/V2M4_[[:space:]]4D[[:space:]]Mesh[[:space:]]Animation[[:space:]]Reconstruction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Monocular[[:space:]]Video/b113d5c5-a844-4d39-a907-37bd11792160_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/V2PE_[[:space:]]Improving[[:space:]]Multimodal[[:space:]]Long-Context[[:space:]]Capability[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models[[:space:]]with[[:space:]]Variable[[:space:]]Visual[[:space:]]Position[[:space:]]Encoding/75b04339-69b5-48e5-a71f-b17b99c6243a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/V2XPnP_[[:space:]]Vehicle-to-Everything[[:space:]]Spatio-Temporal[[:space:]]Fusion[[:space:]]for[[:space:]]Multi-Agent[[:space:]]Perception[[:space:]]and[[:space:]]Prediction/36b3719d-faf2-4104-a9f7-9253a3958218_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/V2XScenes_[[:space:]]A[[:space:]]Multiple[[:space:]]Challenging[[:space:]]Traffic[[:space:]]Conditions[[:space:]]Dataset[[:space:]]for[[:space:]]Large-Range[[:space:]]Vehicle-Infrastructure[[:space:]]Collaborative[[:space:]]Perception/e74c8f66-2f14-41e8-a64e-91a69d371436_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VA-MoE_[[:space:]]Variables-Adaptive[[:space:]]Mixture[[:space:]]of[[:space:]]Experts[[:space:]]for[[:space:]]Incremental[[:space:]]Weather[[:space:]]Forecasting/122157a9-e33d-4cdd-b0f9-d8b983f2a104_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VACE_[[:space:]]All-in-One[[:space:]]Video[[:space:]]Creation[[:space:]]and[[:space:]]Editing/bb0c9cf5-4f56-4fe0-bdd0-bf9454fc4819_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VAFlow_[[:space:]]Video-to-Audio[[:space:]]Generation[[:space:]]with[[:space:]]Cross-Modality[[:space:]]Flow[[:space:]]Matching/7812feed-641c-4918-a582-f934ca29c2a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VAGUE_[[:space:]]Visual[[:space:]]Contexts[[:space:]]Clarify[[:space:]]Ambiguous[[:space:]]Expressions/321d7adc-1768-447e-aad1-8d77c13702f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VALLR_[[:space:]]Visual[[:space:]]ASR[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Lip[[:space:]]Reading/88644d63-e953-49ce-b2d3-f0418686524b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VCA_[[:space:]]Video[[:space:]]Curious[[:space:]]Agent[[:space:]]for[[:space:]]Long[[:space:]]Video[[:space:]]Understanding/a73c253b-5afa-42a6-b79a-31559f118997_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VEGGIE_[[:space:]]Instructional[[:space:]]Editing[[:space:]]and[[:space:]]Reasoning[[:space:]]Video[[:space:]]Concepts[[:space:]]with[[:space:]]Grounded[[:space:]]Generation/d727aeb7-1f89-498a-ac85-9bcc05bddf8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VFlowOpt_[[:space:]]A[[:space:]]Token[[:space:]]Pruning[[:space:]]Framework[[:space:]]for[[:space:]]LMMs[[:space:]]with[[:space:]]Visual[[:space:]]Information[[:space:]]Flow-Guided[[:space:]]Optimization/b08d8e36-4b30-4260-aad4-6cb0fd4d67ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VGGSounder_[[:space:]]Audio-Visual[[:space:]]Evaluations[[:space:]]for[[:space:]]Foundation[[:space:]]Models/a6042e74-fb7f-4488-a331-f3b644cd7772_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VGMamba_[[:space:]]Attribute-to-Location[[:space:]]Clue[[:space:]]Reasoning[[:space:]]for[[:space:]]Quantity-Agnostic[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding/db8ed93c-5bcb-43cd-a62c-cc51ec338aab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VIGFace_[[:space:]]Virtual[[:space:]]Identity[[:space:]]Generation[[:space:]]for[[:space:]]Privacy-Free[[:space:]]Face[[:space:]]Recognition[[:space:]]Dataset/570e97bc-508f-4feb-946d-7b35db443845_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VIPerson_[[:space:]]Flexibly[[:space:]]Generating[[:space:]]Virtual[[:space:]]Identity[[:space:]]for[[:space:]]Person[[:space:]]Re-Identification/9c97a79e-c7c9-4628-ab6c-7d4d09ccbb66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VISION-XL_[[:space:]]High[[:space:]]Definition[[:space:]]Video[[:space:]]Inverse[[:space:]]Problem[[:space:]]Solver[[:space:]]using[[:space:]]Latent[[:space:]]Image[[:space:]]Diffusion[[:space:]]Models/339912e1-b9b8-4ac6-a699-6e214154b6e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VISO_[[:space:]]Accelerating[[:space:]]In-orbit[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Language-Guided[[:space:]]Mask[[:space:]]Learning[[:space:]]and[[:space:]]Sparse[[:space:]]Inference/bca87654-a43f-4442-9368-eee092911b00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VITAL_[[:space:]]More[[:space:]]Understandable[[:space:]]Feature[[:space:]]Visualization[[:space:]]through[[:space:]]Distribution[[:space:]]Alignment[[:space:]]and[[:space:]]Relevant[[:space:]]Information[[:space:]]Flow/65791404-56e7-4d3c-a033-7abfb8aab04e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VLABench_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Benchmark[[:space:]]for[[:space:]]Language-Conditioned[[:space:]]Robotics[[:space:]]Manipulation[[:space:]]with[[:space:]]Long-Horizon[[:space:]]Reasoning[[:space:]]Tasks/bbac0723-47a6-4c9f-bec1-1f18e03a36ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VLDrive_[[:space:]]Vision-Augmented[[:space:]]Lightweight[[:space:]]MLLMs[[:space:]]for[[:space:]]Efficient[[:space:]]Language-grounded[[:space:]]Autonomous[[:space:]]Driving/03e7ad20-412e-4720-b543-ad870d80aec2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VLIPP_[[:space:]]Towards[[:space:]]Physically[[:space:]]Plausible[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Informed[[:space:]]Physical[[:space:]]Prior/eba6c702-728f-4007-b7d6-27e6fa6404b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VLM4D_[[:space:]]Towards[[:space:]]Spatiotemporal[[:space:]]Awareness[[:space:]]in[[:space:]]Vision[[:space:]]Language[[:space:]]Models/7ad15710-76ed-462f-ba27-85159b1f2146_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VLR-Driver_[[:space:]]Large[[:space:]]Vision-Language-Reasoning[[:space:]]Models[[:space:]]for[[:space:]]Embodied[[:space:]]Autonomous[[:space:]]Driving/a365b909-1be9-4f17-b2ec-25f898d74a15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VLRMBench_[[:space:]]A[[:space:]]Comprehensive[[:space:]]and[[:space:]]Challenging[[:space:]]Benchmark[[:space:]]for[[:space:]]Vision-Language[[:space:]]Reward[[:space:]]Models/8e730ad2-3996-4a2e-9875-e28304c3e917_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VMBench_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Perception-Aligned[[:space:]]Video[[:space:]]Motion[[:space:]]Generation/1a1d91b4-eb97-44d0-a07d-41746c53ad96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VMem_[[:space:]]Consistent[[:space:]]Interactive[[:space:]]Video[[:space:]]Scene[[:space:]]Generation[[:space:]]with[[:space:]]Surfel-Indexed[[:space:]]View[[:space:]]Memory/1781a2ca-916a-47a8-9f71-bf548311fbcb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VOVTrack_[[:space:]]Exploring[[:space:]]the[[:space:]]Potentiality[[:space:]]in[[:space:]]Raw[[:space:]]Videos[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Multi-Object[[:space:]]Tracking/cb6ae00b-bc8e-44a2-b37e-6f565b52690f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VOccl3D_[[:space:]]A[[:space:]]Video[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Estimation[[:space:]]under[[:space:]]real[[:space:]]Occlusions/bee260a8-0889-4cd7-b8a5-2c0fb163c69d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VPO_[[:space:]]Aligning[[:space:]]Text-to-Video[[:space:]]Generation[[:space:]]Models[[:space:]]with[[:space:]]Prompt[[:space:]]Optimization/1912f989-0c41-45f7-ba83-c030d1f76cc0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VPR-Cloak_[[:space:]]A[[:space:]]First[[:space:]]Look[[:space:]]at[[:space:]]Privacy[[:space:]]Cloak[[:space:]]Against[[:space:]]Visual[[:space:]]Place[[:space:]]Recognition/ecc5979e-3031-4d35-8396-33fead75fa7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VQ-SGen_[[:space:]]A[[:space:]]Vector[[:space:]]Quantized[[:space:]]Stroke[[:space:]]Representation[[:space:]]for[[:space:]]Creative[[:space:]]Sketch[[:space:]]Generation/27ec5cff-30ea-4eda-ad45-f2004607bb53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VQ-VLA_[[:space:]]Improving[[:space:]]Vision-Language-Action[[:space:]]Models[[:space:]]via[[:space:]]Scaling[[:space:]]Vector-Quantized[[:space:]]Action[[:space:]]Tokenizers/a05d8d10-adb9-4812-8715-77d71aa82af2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VRBench_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Multi-Step[[:space:]]Reasoning[[:space:]]in[[:space:]]Long[[:space:]]Narrative[[:space:]]Videos/13199d03-d2f0-4f02-9446-3b78b6cbc2ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VRM_[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]via[[:space:]]Virtual[[:space:]]Relation[[:space:]]Matching/4752d372-1ed6-459b-bee6-6b202a028bae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VSC_[[:space:]]Visual[[:space:]]Search[[:space:]]Compositional[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Model/9f76eb31-5853-4235-b96e-6306079810f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VSP_[[:space:]]Diagnosing[[:space:]]the[[:space:]]Dual[[:space:]]Challenges[[:space:]]of[[:space:]]Perception[[:space:]]and[[:space:]]Reasoning[[:space:]]in[[:space:]]Spatial[[:space:]]Planning[[:space:]]Tasks[[:space:]]for[[:space:]]MLLMs/262f63b9-5c82-43f4-b645-d16941a910c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VSRM_[[:space:]]A[[:space:]]Robust[[:space:]]Mamba-Based[[:space:]]Framework[[:space:]]for[[:space:]]Video[[:space:]]Super-Resolution/6e341a1e-2a3f-4224-8262-4e1ca75e0c71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VSSD_[[:space:]]Vision[[:space:]]Mamba[[:space:]]with[[:space:]]Non-Causal[[:space:]]State[[:space:]]Space[[:space:]]Duality/4e862f57-d9f1-4a7e-9559-4423a19c05c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VTimeCoT_[[:space:]]Thinking[[:space:]]by[[:space:]]Drawing[[:space:]]for[[:space:]]Video[[:space:]]Temporal[[:space:]]Grounding[[:space:]]and[[:space:]]Reasoning/340dea8a-103d-45f5-8ee9-2f2409e86042_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vamba_[[:space:]]Understanding[[:space:]]Hour-Long[[:space:]]Videos[[:space:]]with[[:space:]]Hybrid[[:space:]]Mamba-Transformers/41032422-fba8-4ef6-8117-de06ce1069a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Variance-Based[[:space:]]Pruning[[:space:]]for[[:space:]]Accelerating[[:space:]]and[[:space:]]Compressing[[:space:]]Trained[[:space:]]Networks/5aed5135-0f01-4c2d-b7ee-d3f1987cb1ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vector[[:space:]]Contrastive[[:space:]]Learning[[:space:]]For[[:space:]]Pixel-Wise[[:space:]]Pretraining[[:space:]]In[[:space:]]Medical[[:space:]]Vision/b863182b-c21f-4dc0-8b6a-73de7593a982_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VehicleMAE_[[:space:]]View-asymmetry[[:space:]]Mutual[[:space:]]Learning[[:space:]]for[[:space:]]Vehicle[[:space:]]Re-identification[[:space:]]Pre-training[[:space:]]via[[:space:]]Masked[[:space:]]AutoEncoders/ec3290f4-caa9-4573-afb5-3440940874ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Verbalized[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Interpretable[[:space:]]Few-Shot[[:space:]]Generalization/e86880d5-4a52-49e7-83a5-bbf2d31e9dd1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Versatile[[:space:]]Transition[[:space:]]Generation[[:space:]]with[[:space:]]Image-to-Video[[:space:]]Diffusion/f3fc723c-8a6e-4204-9849-5843b598e850_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VertexRegen_[[:space:]]Mesh[[:space:]]Generation[[:space:]]with[[:space:]]Continuous[[:space:]]Level[[:space:]]of[[:space:]]Detail/342c3a2f-65c0-4b35-b717-21b0c8a1a0ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViCTr_[[:space:]]Vital[[:space:]]Consistency[[:space:]]Transfer[[:space:]]for[[:space:]]Pathology[[:space:]]Aware[[:space:]]Image[[:space:]]Synthesis/08391b0d-1c48-48a7-aee4-a1aed1fb2b8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViLLa_[[:space:]]Video[[:space:]]Reasoning[[:space:]]Segmentation[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Model/9374d1a1-c3e0-498b-83e7-5023233fa5ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViLU_[[:space:]]Learning[[:space:]]Vision-Language[[:space:]]Uncertainties[[:space:]]for[[:space:]]Failure[[:space:]]Prediction/9693ec7c-e63e-4549-b95a-8e863d6bee58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViM-VQ_[[:space:]]Efficient[[:space:]]Post-Training[[:space:]]Vector[[:space:]]Quantization[[:space:]]for[[:space:]]Visual[[:space:]]Mamba/bb5b37a9-8d33-488a-b6b3-ef3b094ba666_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViSpeak_[[:space:]]Visual[[:space:]]Instruction[[:space:]]Feedback[[:space:]]in[[:space:]]Streaming[[:space:]]Videos/4c2298b6-a1cd-4d0c-bdd9-742253ff8150_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViT-EnsembleAttack_[[:space:]]Augmenting[[:space:]]Ensemble[[:space:]]Models[[:space:]]for[[:space:]]Stronger[[:space:]]Adversarial[[:space:]]Transferability[[:space:]]in[[:space:]]Vision[[:space:]]Transformers/9e6fb5c3-a13e-4a97-9aed-2462169214a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViT-Linearizer_[[:space:]]Distilling[[:space:]]Quadratic[[:space:]]Knowledge[[:space:]]into[[:space:]]Linear-Time[[:space:]]Vision[[:space:]]Models/4d67ff7b-60ce-453d-9e8b-8fe983e07bca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViT-Split_[[:space:]]Unleashing[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]Vision[[:space:]]Foundation[[:space:]]Models[[:space:]]via[[:space:]]Efficient[[:space:]]Splitting[[:space:]]Heads/62f8e79c-d973-4405-a368-6eb7d50f7246_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vid-Group_[[:space:]]Temporal[[:space:]]Video[[:space:]]Grounding[[:space:]]Pretraining[[:space:]]from[[:space:]]Unlabeled[[:space:]]Videos[[:space:]]in[[:space:]]the[[:space:]]Wild/b7f7fdbe-e75b-4823-a9b7-19405b49fb65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video[[:space:]]Color[[:space:]]Grading[[:space:]]via[[:space:]]Look-Up[[:space:]]Table[[:space:]]Generation/b0b9c231-2082-4b4c-ad77-337ee0cc9045_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video[[:space:]]Individual[[:space:]]Counting[[:space:]]for[[:space:]]Moving[[:space:]]Drones/1cf4d87f-9c8a-44cd-a5ee-135b6f1daa18_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video[[:space:]]Motion[[:space:]]Graphs/78363dc2-8dfc-4b9b-80e6-1323ef2ac37e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video-T1_[[:space:]]Test-time[[:space:]]Scaling[[:space:]]for[[:space:]]Video[[:space:]]Generation/1019b1f8-19d0-4e70-9f39-8bbdb082c721_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Video2BEV_[[:space:]]Transforming[[:space:]]Drone[[:space:]]Videos[[:space:]]to[[:space:]]BEVs[[:space:]]for[[:space:]]Video-based[[:space:]]Geo-localization/1c533f62-65f2-4dff-a54f-3aa8e0de6fdc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoAds[[:space:]]for[[:space:]]Fast-Paced[[:space:]]Video[[:space:]]Understanding/8a639327-709f-4460-aabc-c77b069d813b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoAuteur_[[:space:]]Towards[[:space:]]Long[[:space:]]Narrative[[:space:]]Video[[:space:]]Generation/31b80179-4ba4-49ae-a3df-b51fb25c8cd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoLLaMB_[[:space:]]Long[[:space:]]Streaming[[:space:]]Video[[:space:]]Understanding[[:space:]]with[[:space:]]Recurrent[[:space:]]Memory[[:space:]]Bridges/ef7f527a-2c19-4223-8b21-ac615c5af575_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoMiner_[[:space:]]Iteratively[[:space:]]Grounding[[:space:]]Key[[:space:]]Frames[[:space:]]of[[:space:]]Hour-Long[[:space:]]Videos[[:space:]]via[[:space:]]Tree-based[[:space:]]Group[[:space:]]Relative[[:space:]]Policy[[:space:]]Optimization/d7b4639d-d9a0-4b6c-986f-11dd15c86ebf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoOrion_[[:space:]]Tokenizing[[:space:]]Object[[:space:]]Dynamics[[:space:]]in[[:space:]]Videos/872b582e-2d26-46e1-93e2-773aad7bd594_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoRFSplat_[[:space:]]Direct[[:space:]]Scene-Level[[:space:]]Text-to-3D[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]Generation[[:space:]]with[[:space:]]Flexible[[:space:]]Pose[[:space:]]and[[:space:]]Multi-View[[:space:]]Joint[[:space:]]Modeling/0f70facc-7c84-4b86-b0dc-d3d1bdbeaad1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoSetDiff_[[:space:]]Identifying[[:space:]]and[[:space:]]Reasoning[[:space:]]Similarities[[:space:]]and[[:space:]]Differences[[:space:]]in[[:space:]]Similar[[:space:]]Videos/3d626d6f-d212-46b3-bc48-4c5f49732e8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VideoVAE+_[[:space:]]Large[[:space:]]Motion[[:space:]]Video[[:space:]]Autoencoding[[:space:]]with[[:space:]]Cross-modal[[:space:]]Video[[:space:]]VAE/d40a1523-5eb7-42eb-ab5f-29118f57653b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ViewSRD_[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding[[:space:]]via[[:space:]]Structured[[:space:]]Multi-View[[:space:]]Decomposition/28fcbc42-0fe9-4106-81ab-90efedfc5cc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VisHall3D_[[:space:]]Monocular[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion[[:space:]]from[[:space:]]Reconstructing[[:space:]]the[[:space:]]Visible[[:space:]]Regions[[:space:]]to[[:space:]]Hallucinating[[:space:]]the[[:space:]]Invisible[[:space:]]Regions/329f4613-dbba-4972-811d-2a326b45a72a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VisNumBench_[[:space:]]Evaluating[[:space:]]Number[[:space:]]Sense[[:space:]]of[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/27f14a06-5371-423d-833b-707308fc0b78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VisRL_[[:space:]]Intention-Driven[[:space:]]Visual[[:space:]]Perception[[:space:]]via[[:space:]]Reinforced[[:space:]]Reasoning/f2419d2f-e576-4a4f-bc19-8d0454e83639_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vision-Language[[:space:]]Interactive[[:space:]]Relation[[:space:]]Mining[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/f8f524f5-9035-412a-9d49-1b195879895c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vision-Language[[:space:]]Models[[:space:]]Can't[[:space:]]See[[:space:]]the[[:space:]]Obvious/d8b7426a-cb65-4738-af2c-9b996a867e18_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vision-Language[[:space:]]Neural[[:space:]]Graph[[:space:]]Featurization[[:space:]]for[[:space:]]Extracting[[:space:]]Retinal[[:space:]]Lesions/7f2c87cd-d7a3-42c9-9013-568610396af7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VisionMath_[[:space:]]Vision-Form[[:space:]]Mathematical[[:space:]]Problem-Solving/e1365b02-1cb8-4330-95ba-78f25112df6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VistaDream_[[:space:]]Sampling[[:space:]]multiview[[:space:]]consistent[[:space:]]images[[:space:]]for[[:space:]]single-view[[:space:]]scene[[:space:]]reconstruction/7ae1ef6e-97d6-4f46-866a-e139e1269a78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Chronicles_[[:space:]]Using[[:space:]]Multimodal[[:space:]]LLMs[[:space:]]to[[:space:]]Analyze[[:space:]]Massive[[:space:]]Collections[[:space:]]of[[:space:]]Images/faf5d8e8-4101-4ccc-91eb-41873d923f20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Intention[[:space:]]Grounding[[:space:]]for[[:space:]]Egocentric[[:space:]]Assistants/eedc0963-0f97-46ed-a0a7-79c0f04d3f23_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Interestingness[[:space:]]Decoded_[[:space:]]How[[:space:]]GPT-4o[[:space:]]Mirrors[[:space:]]Human[[:space:]]Interests/38264d21-1016-40ad-892a-9a9170d58abd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Modality[[:space:]]Prompt[[:space:]]for[[:space:]]Adapting[[:space:]]Vision-Language[[:space:]]Object[[:space:]]Detectors/fcab79ff-b3e2-457c-a329-0dcc0ab0c321_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Relation[[:space:]]Diffusion[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection/dc9e94e3-6027-418f-9af5-8cec6e4a6997_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Surface[[:space:]]Wave[[:space:]]Elastography_[[:space:]]Revealing[[:space:]]Subsurface[[:space:]]Physical[[:space:]]Properties[[:space:]]via[[:space:]]Visible[[:space:]]Surface[[:space:]]Waves/f44e2841-6bf1-41ce-9953-fc74116e88aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Test-time[[:space:]]Scaling[[:space:]]for[[:space:]]GUI[[:space:]]Agent[[:space:]]Grounding/b0a9cda2-be4a-426f-a5b8-2a6b64ffb5fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual[[:space:]]Textualization[[:space:]]for[[:space:]]Image[[:space:]]Prompted[[:space:]]Object[[:space:]]Detection/5ef81f8a-58e2-436d-920c-1c45226041f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual-Oriented[[:space:]]Fine-Grained[[:space:]]Knowledge[[:space:]]Editing[[:space:]]for[[:space:]]MultiModal[[:space:]]Large[[:space:]]Language[[:space:]]Models/a7f3a1fd-fa58-4e7e-b24e-304e05b363a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Visual-RFT_[[:space:]]Visual[[:space:]]Reinforcement[[:space:]]Fine-Tuning/b1758928-1f95-4aa1-882d-f17b8d4c58c5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VisualCloze_[[:space:]]A[[:space:]]Universal[[:space:]]Image[[:space:]]Generation[[:space:]]Framework[[:space:]]via[[:space:]]Visual[[:space:]]In-Context[[:space:]]Learning/3aa630b3-d121-44bf-b130-2854f77fd1d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vivid4D_[[:space:]]Improving[[:space:]]4D[[:space:]]Reconstruction[[:space:]]from[[:space:]]Monocular[[:space:]]Video[[:space:]]by[[:space:]]Video[[:space:]]Inpainting/334ccb81-eef2-4518-a459-c4c4b31dd8c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VoiceCraft-Dub_[[:space:]]Automated[[:space:]]Video[[:space:]]Dubbing[[:space:]]with[[:space:]]Neural[[:space:]]Codec[[:space:]]Language[[:space:]]Models/68799307-ec96-4f0a-9ca1-088fb28e13e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VoluMe[[:space:]]-[[:space:]]Authentic[[:space:]]3D[[:space:]]Video[[:space:]]Calls[[:space:]]from[[:space:]]Live[[:space:]]Gaussian[[:space:]]Splat[[:space:]]Prediction/aca36b55-c429-4d0d-a0b8-cd49d838955e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VolumetricSMPL_[[:space:]]A[[:space:]]Neural[[:space:]]Volumetric[[:space:]]Body[[:space:]]Model[[:space:]]for[[:space:]]Efficient[[:space:]]Interactions,[[:space:]]Contacts,[[:space:]]and[[:space:]]Collisions/bb7a2147-d88a-4bc3-b1d6-b526e6f5b798_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VoteSplat_[[:space:]]Hough[[:space:]]Voting[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]3D[[:space:]]Scene[[:space:]]Understanding/5802dd71-01ac-4867-bdce-70fd12fae3bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/VoxelKP_[[:space:]]A[[:space:]]Voxel-based[[:space:]]Network[[:space:]]Architecture[[:space:]]for[[:space:]]Human[[:space:]]Keypoint[[:space:]]Estimation[[:space:]]in[[:space:]]LiDAR[[:space:]]Data/f5de6d36-8311-4c83-8578-66ba5526c326_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Voyaging[[:space:]]into[[:space:]]Perpetual[[:space:]]Dynamic[[:space:]]Scenes[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]View/c95a7c1e-c392-41cc-85d4-4436d0739bfa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Vulnerability-Aware[[:space:]]Spatio-Temporal[[:space:]]Learning[[:space:]]for[[:space:]]Generalizable[[:space:]]Deepfake[[:space:]]Video[[:space:]]Detection/66ddbb2b-1283-449c-8670-af8534967206_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WAVE_[[:space:]]Warp-Based[[:space:]]View[[:space:]]Guidance[[:space:]]for[[:space:]]Consistent[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]Using[[:space:]]a[[:space:]]Single[[:space:]]Image/d2ddfab3-354c-4385-b1b5-8f76650ff96d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WINS_[[:space:]]Winograd[[:space:]]Structured[[:space:]]Pruning[[:space:]]for[[:space:]]Fast[[:space:]]Winograd[[:space:]]Convolution/c1353ff7-9892-4a0b-8c5f-7b946b4ec7c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WIPES_[[:space:]]Wavelet-based[[:space:]]Visual[[:space:]]Primitives/5c2a3453-b8e9-4c45-a9c7-d765229caa07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WIR3D_[[:space:]]Visually-Informed[[:space:]]and[[:space:]]Geometry-Aware[[:space:]]3D[[:space:]]Shape[[:space:]]Abstraction/49efbaa1-4dc1-4f95-8026-29529291070b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WSI-LLaVA_[[:space:]]A[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Whole[[:space:]]Slide[[:space:]]Image/5c32ce28-a3b0-4e1d-be20-7568b907d88a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WalkVLM_[[:space:]]Aid[[:space:]]Visually[[:space:]]Impaired[[:space:]]People[[:space:]]Walking[[:space:]]by[[:space:]]Vision[[:space:]]Language[[:space:]]Model/f2f29bbd-c165-4e1f-b880-aba87bbd2c1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WarpHE4D_[[:space:]]Dense[[:space:]]4D[[:space:]]Head[[:space:]]Map[[:space:]]toward[[:space:]]Full[[:space:]]Head[[:space:]]Reconstruction/f316d5e0-ad4b-4e9b-ba4d-1c0b8a170fec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Wasserstein[[:space:]]Style[[:space:]]Distribution[[:space:]]Analysis[[:space:]]and[[:space:]]Transform[[:space:]]for[[:space:]]Stylized[[:space:]]Image[[:space:]]Generation/b8746cda-704c-4a0a-9467-542e4e841b49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Wave-MambaAD_[[:space:]]Wavelet-driven[[:space:]]State[[:space:]]Space[[:space:]]Model[[:space:]]for[[:space:]]Multi-class[[:space:]]Unsupervised[[:space:]]Anomaly[[:space:]]Detection/d167cc0d-f166-4fec-a4c9-a57721d19178_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WaveMamba_[[:space:]]Wavelet-Driven[[:space:]]Mamba[[:space:]]Fusion[[:space:]]for[[:space:]]RGB-Infrared[[:space:]]Object[[:space:]]Detection/c16d775a-5b9c-4030-b5af-7f8546eac6be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Wavelet[[:space:]]Policy_[[:space:]]Lifting[[:space:]]Scheme[[:space:]]for[[:space:]]Policy[[:space:]]Learning[[:space:]]in[[:space:]]Long-Horizon[[:space:]]Tasks/61340c8d-ce4e-42e4-9ab7-a22c1a9e1133_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Weakly[[:space:]]Supervised[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification[[:space:]]via[[:space:]]Heterogeneous[[:space:]]Expert[[:space:]]Collaborative[[:space:]]Consistency[[:space:]]Learning/1b97471f-d27a-4c34-a569-593af9bedc9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Weakly-Supervised[[:space:]]Learning[[:space:]]of[[:space:]]Dense[[:space:]]Functional[[:space:]]Correspondences/3dc2108f-7d0a-4c28-b604-10f9ccf91942_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WeaveSeg_[[:space:]]Iterative[[:space:]]Contrast-weaving[[:space:]]and[[:space:]]Spectral[[:space:]]Feature-refining[[:space:]]for[[:space:]]Nuclei[[:space:]]Instance[[:space:]]Segmentation/4176a988-4d3a-45b0-ab15-eb8f0b2fa50e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Web[[:space:]]Artifact[[:space:]]Attacks[[:space:]]Disrupt[[:space:]]Vision[[:space:]]Language[[:space:]]Models/793c8111-1366-4e5c-94bf-770223d04fae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/What[[:space:]]Changed[[:space:]]and[[:space:]]What[[:space:]]Could[[:space:]]Have[[:space:]]Changed_[[:space:]]State-Change[[:space:]]Counterfactuals[[:space:]]for[[:space:]]Procedure-Aware[[:space:]]Video[[:space:]]Representation[[:space:]]Learning/403d5bfc-2ad7-4b27-944d-e6bc63e5bcd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/What[[:space:]]Changed_[[:space:]]Detecting[[:space:]]and[[:space:]]Evaluating[[:space:]]Instruction-Guided[[:space:]]Image[[:space:]]Edits[[:space:]]with[[:space:]]Multimodal[[:space:]]Large[[:space:]]Language[[:space:]]Models/c0a4e339-69da-4cf2-9ffd-f439404dd16a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/What[[:space:]]If_[[:space:]]Understanding[[:space:]]Motion[[:space:]]Through[[:space:]]Sparse[[:space:]]Interactions/c33d9dcc-fde3-4776-81ca-8cda367e6981_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/What[[:space:]]Makes[[:space:]]for[[:space:]]Text[[:space:]]to[[:space:]]360-degree[[:space:]]Panorama[[:space:]]Generation[[:space:]]with[[:space:]]Stable[[:space:]]Diffusion_/cf8dae55-332f-44ad-840c-3f6b5a1ebfab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/What[[:space:]]You[[:space:]]Have[[:space:]]is[[:space:]]What[[:space:]]You[[:space:]]Track_[[:space:]]Adaptive[[:space:]]and[[:space:]]Robust[[:space:]]Multimodal[[:space:]]Tracking/c6b9c23e-b4e6-4935-b399-70cebfc453b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/What[[:space:]]to[[:space:]]Distill_[[:space:]]Fast[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]with[[:space:]]Adaptive[[:space:]]Sampling/5cc2db1b-d50c-4519-b11e-7cd7283b5e50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/What[[:space:]]we[[:space:]]need[[:space:]]is[[:space:]]explicit[[:space:]]controllability_[[:space:]]Training[[:space:]]3D[[:space:]]gaze[[:space:]]estimator[[:space:]]using[[:space:]]only[[:space:]]facial[[:space:]]images/130f2ff4-e0d9-47ca-b6cb-4080460048e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/What's[[:space:]]Making[[:space:]]That[[:space:]]Sound[[:space:]]Right[[:space:]]Now_[[:space:]]Video-centric[[:space:]]Audio-Visual[[:space:]]Localization/44803fdf-7cfa-4917-b402-d01e2447466c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/What's[[:space:]]in[[:space:]]a[[:space:]]Latent_[[:space:]]Leveraging[[:space:]]Diffusion[[:space:]]Latent[[:space:]]Space[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/81c74318-d891-44f5-8e24-de8ec3e2c413_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/When[[:space:]]Anchors[[:space:]]Meet[[:space:]]Cold[[:space:]]Diffusion_[[:space:]]A[[:space:]]Multi-Stage[[:space:]]Approach[[:space:]]to[[:space:]]Lane[[:space:]]Detection/03666e3c-da14-47a9-8be7-58e6b0821b95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/When[[:space:]]Confidence[[:space:]]Fails_[[:space:]]Revisiting[[:space:]]Pseudo-Label[[:space:]]Selection[[:space:]]in[[:space:]]Semi-supervised[[:space:]]Semantic[[:space:]]Segmentation/1e3de233-91fb-4c3e-99af-1b078b01bace_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/When[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Model[[:space:]]Meets[[:space:]]Large[[:space:]]Remote[[:space:]]Sensing[[:space:]]Imagery_[[:space:]]Coarse-to-Fine[[:space:]]Text-Guided[[:space:]]Token[[:space:]]Pruning/cc17937b-94d3-4352-8b5a-b4c323567743_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/When[[:space:]]Lighting[[:space:]]Deceives_[[:space:]]Exposing[[:space:]]Vision-Language[[:space:]]Models'[[:space:]]Illumination[[:space:]]Vulnerability[[:space:]]Through[[:space:]]Illumination[[:space:]]Transformation[[:space:]]Attack/7ce1661a-7863-4397-8464-08916ed25d3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/When[[:space:]]Pixel[[:space:]]Difference[[:space:]]Patterns[[:space:]]Meet[[:space:]]ViT_[[:space:]]PiDiViT[[:space:]]for[[:space:]]Few-Shot[[:space:]]Object[[:space:]]Detection/78d510bf-7798-4115-b8d7-929a950a8416_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/When[[:space:]]Schrodinger[[:space:]]Bridge[[:space:]]Meets[[:space:]]Real-World[[:space:]]Image[[:space:]]Dehazing[[:space:]]with[[:space:]]Unpaired[[:space:]]Training/c7189c5a-d018-4b96-8859-6c17a385c5d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/When[[:space:]]and[[:space:]]Where[[:space:]]do[[:space:]]Data[[:space:]]Poisons[[:space:]]Attack[[:space:]]Textual[[:space:]]Inversion_/3fa7061d-705e-43ad-9c51-b882bb9873d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Where[[:space:]]am[[:space:]]I_[[:space:]]Cross-View[[:space:]]Geo-localization[[:space:]]with[[:space:]]Natural[[:space:]]Language[[:space:]]Descriptions/f1023190-1ee9-442e-9a72-f9dec8a76f13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Where,[[:space:]]What,[[:space:]]Why_[[:space:]]Towards[[:space:]]Explainable[[:space:]]Driver[[:space:]]Attention[[:space:]]Prediction/8087ca0e-98c5-4d7a-a52f-09d0f857ac52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Who[[:space:]]Controls[[:space:]]the[[:space:]]Authorization_[[:space:]]Invertible[[:space:]]Networks[[:space:]]for[[:space:]]Copyright[[:space:]]Protection[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Synthesis/3f7258c8-a390-4cf7-9b1e-42cbf47a42a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Who[[:space:]]is[[:space:]]a[[:space:]]Better[[:space:]]Talker_[[:space:]]Subjective[[:space:]]and[[:space:]]Objective[[:space:]]Quality[[:space:]]Assessment[[:space:]]for[[:space:]]AI-Generated[[:space:]]Talking[[:space:]]Heads/31ef8567-3c00-4c44-b71f-af1f2b0b5c24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Why[[:space:]]LVLMs[[:space:]]Are[[:space:]]More[[:space:]]Prone[[:space:]]to[[:space:]]Hallucinations[[:space:]]in[[:space:]]Longer[[:space:]]Responses_[[:space:]]The[[:space:]]Role[[:space:]]of[[:space:]]Context/a7078303-ba33-4c42-a787-cdfd0a7d7b84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Wide2Long_[[:space:]]Learning[[:space:]]Lens[[:space:]]Compression[[:space:]]and[[:space:]]Perspective[[:space:]]Adjustment[[:space:]]for[[:space:]]Wide-Angle[[:space:]]to[[:space:]]Telephoto[[:space:]]Translation/baa830f2-e8bf-491c-a264-552b980530f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WikiAutoGen_[[:space:]]Towards[[:space:]]Multi-Modal[[:space:]]Wikipedia-Style[[:space:]]Article[[:space:]]Generation/707bfef2-e682-4352-8e81-2af4733b9fe7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WildSAT_[[:space:]]Learning[[:space:]]Satellite[[:space:]]Image[[:space:]]Representations[[:space:]]from[[:space:]]Wildlife[[:space:]]Observations/ffaa63f6-2b3f-41d0-a1eb-88d48e346985_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WildSeg3D_[[:space:]]Segment[[:space:]]Any[[:space:]]3D[[:space:]]Objects[[:space:]]in[[:space:]]the[[:space:]]Wild[[:space:]]from[[:space:]]2D[[:space:]]Images/c727212a-10ef-4c0d-a1c9-d78f5484a70b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WonderPlay_[[:space:]]Dynamic[[:space:]]3D[[:space:]]Scene[[:space:]]Generation[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]and[[:space:]]Actions/59b8ccd1-d5e7-4d4c-9298-ce3e755fe5ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WonderTurbo_[[:space:]]Generating[[:space:]]Interactive[[:space:]]3D[[:space:]]World[[:space:]]in[[:space:]]0.72[[:space:]]Seconds/004277e4-cc32-4a24-bc57-88909bd31122_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/World4Drive_[[:space:]]End-to-End[[:space:]]Autonomous[[:space:]]Driving[[:space:]]via[[:space:]]Intention-aware[[:space:]]Physical[[:space:]]Latent[[:space:]]World[[:space:]]Model/bd010426-5cc6-465c-8059-3ed79f4fec42_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/WorldScore_[[:space:]]A[[:space:]]Unified[[:space:]]Evaluation[[:space:]]Benchmark[[:space:]]for[[:space:]]World[[:space:]]Generation/2e63d834-5044-4966-887d-e7d4112f0b43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/X-Capture_[[:space:]]An[[:space:]]Open-Source[[:space:]]Portable[[:space:]]Device[[:space:]]for[[:space:]]Multi-Sensory[[:space:]]Learning/6b08175f-1aac-4b47-837d-556a41b9c21a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/X-Dancer_[[:space:]]Expressive[[:space:]]Music[[:space:]]to[[:space:]]Human[[:space:]]Dance[[:space:]]Video[[:space:]]Generation/893438c0-43d3-4ba2-a298-9c386258256a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/X-Fusion_[[:space:]]Introducing[[:space:]]New[[:space:]]Modality[[:space:]]to[[:space:]]Frozen[[:space:]]Large[[:space:]]Language[[:space:]]Models/425c73c4-35d6-4f59-b075-35ebadf4b8b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/X-Prompt_[[:space:]]Generalizable[[:space:]]Auto-Regressive[[:space:]]Visual[[:space:]]Learning[[:space:]]with[[:space:]]In-Context[[:space:]]Prompting/66c6a6ed-9847-4426-937e-5e7c6b19faf9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/X2-Gaussian_[[:space:]]4D[[:space:]]Radiative[[:space:]]Gaussian[[:space:]]Splatting[[:space:]]for[[:space:]]Continuous-time[[:space:]]Tomographic[[:space:]]Reconstruction/0bfd0212-1def-41aa-be95-53d5f33cfe97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/X2I_[[:space:]]Seamless[[:space:]]Integration[[:space:]]of[[:space:]]Multimodal[[:space:]]Understanding[[:space:]]into[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]via[[:space:]]Attention[[:space:]]Distillation/73b0573d-4884-4d06-9b66-e9a473055724_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/XTrack_[[:space:]]Multimodal[[:space:]]Training[[:space:]]Boosts[[:space:]]RGB-X[[:space:]]Video[[:space:]]Object[[:space:]]Trackers/c568565b-91f6-457d-aad4-2c25aae796ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/YOLO-Count_[[:space:]]Differentiable[[:space:]]Object[[:space:]]Counting[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation/cfefcd61-b5c9-4ed3-a2d8-a13711519bfe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/YOLOE_[[:space:]]Real-Time[[:space:]]Seeing[[:space:]]Anything/3dbff6c8-6a64-4ee0-abf9-c4b47ba4d2a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/You[[:space:]]Are[[:space:]]Your[[:space:]]Own[[:space:]]Best[[:space:]]Teacher_[[:space:]]Achieving[[:space:]]Centralized-level[[:space:]]Performance[[:space:]]in[[:space:]]Federated[[:space:]]Learning[[:space:]]under[[:space:]]Heterogeneous[[:space:]]and[[:space:]]Long-tailed[[:space:]]Data/24258e3b-6f86-4fb2-aca3-b713b31c24a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/You[[:space:]]Share[[:space:]]Beliefs,[[:space:]]I[[:space:]]Adapt_[[:space:]]Progressive[[:space:]]Heterogeneous[[:space:]]Collaborative[[:space:]]Perception/5467ef8e-c8b4-4c7d-b7ef-a1694767ed15_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/You[[:space:]]Think,[[:space:]]You[[:space:]]ACT_[[:space:]]The[[:space:]]New[[:space:]]Task[[:space:]]of[[:space:]]Arbitrary[[:space:]]Text[[:space:]]to[[:space:]]Motion[[:space:]]Generation/4bf51bb3-0b1c-40e9-80ac-090885fb49f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Your[[:space:]]Text[[:space:]]Encoder[[:space:]]Can[[:space:]]Be[[:space:]]An[[:space:]]Object-Level[[:space:]]Watermarking[[:space:]]Controller/a9813057-9d7e-495c-bd8d-0a15137ba2fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ZFusion_[[:space:]]Efficient[[:space:]]Deep[[:space:]]Compositional[[:space:]]Zero-shot[[:space:]]Learning[[:space:]]for[[:space:]]Blind[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]with[[:space:]]Generative[[:space:]]Diffusion[[:space:]]Prior/cb6554ce-9075-4801-b1eb-ec7a6cc16e64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ZIM_[[:space:]]Zero-Shot[[:space:]]Image[[:space:]]Matting[[:space:]]for[[:space:]]Anything/ec21a9a6-f4ee-473f-b698-87f36a029477_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ZIUM_[[:space:]]Zero-Shot[[:space:]]Intent-Aware[[:space:]]Adversarial[[:space:]]Attack[[:space:]]on[[:space:]]Unlearned[[:space:]]Models/2ecc5de9-caed-4861-8871-7280fe728f53_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-AVSR_[[:space:]]Zero-Shot[[:space:]]Audio-Visual[[:space:]]Speech[[:space:]]Recognition[[:space:]]with[[:space:]]LLMs[[:space:]]by[[:space:]]Learning[[:space:]]Language-Agnostic[[:space:]]Speech[[:space:]]Representations/c8281586-b056-41a9-b08b-fbc46de86111_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-Shot[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval[[:space:]]via[[:space:]]Dual-Stream[[:space:]]Instruction-Aware[[:space:]]Distillation/21e6de6e-abc8-43a4-bcfb-a4a84f3eeee1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-Shot[[:space:]]Compositional[[:space:]]Video[[:space:]]Learning[[:space:]]with[[:space:]]Coding[[:space:]]Rate[[:space:]]Reduction/28edbc35-c9cb-4584-bcf5-27480c4c0924_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-Shot[[:space:]]Depth[[:space:]]Aware[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/073d8c0b-4053-48f6-972c-264031ab2e86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-Shot[[:space:]]Vision[[:space:]]Encoder[[:space:]]Grafting[[:space:]]via[[:space:]]LLM[[:space:]]Surrogates/7a116134-5667-436c-b482-aee8bdfd9f65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zero-shot[[:space:]]Inexact[[:space:]]CAD[[:space:]]Model[[:space:]]Alignment[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/9abc6024-2f21-4050-a942-e3a3a7717cec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ZeroKey_[[:space:]]Point-Level[[:space:]]Reasoning[[:space:]]and[[:space:]]Zero-Shot[[:space:]]3D[[:space:]]Keypoint[[:space:]]Detection[[:space:]]from[[:space:]]Large[[:space:]]Language[[:space:]]Models/348d3cba-bca8-4d06-8d2b-4af61751a081_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ZeroStereo_[[:space:]]Zero-shot[[:space:]]Stereo[[:space:]]Matching[[:space:]]from[[:space:]]Single[[:space:]]Images/76bb9552-1902-4fcd-a11d-bcea4458ddbf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/Zeroth-Order[[:space:]]Fine-Tuning[[:space:]]of[[:space:]]LLMs[[:space:]]in[[:space:]]Random[[:space:]]Subspaces/30dae0e3-0c17-444c-9824-db8146e69eb8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/ZipVL_[[:space:]]Accelerating[[:space:]]Vision-Language[[:space:]]Models[[:space:]]through[[:space:]]Dynamic[[:space:]]Token[[:space:]]Sparsity/8cd7319e-aae8-4f84-8ca4-5b3ef6a42984_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/egoPPG_[[:space:]]Heart[[:space:]]Rate[[:space:]]Estimation[[:space:]]from[[:space:]]Eye-Tracking[[:space:]]Cameras[[:space:]]in[[:space:]]Egocentric[[:space:]]Systems[[:space:]]to[[:space:]]Benefit[[:space:]]Downstream[[:space:]]Vision[[:space:]]Tasks/f5964ab7-5331-420f-bb50-825216bb8b90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/iManip_[[:space:]]Skill-Incremental[[:space:]]Learning[[:space:]]for[[:space:]]Robotic[[:space:]]Manipulation/47162734-cb83-448a-b459-5b0fc2ff10ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/kh_[[:space:]]Symmetry[[:space:]]Understanding[[:space:]]of[[:space:]]3D[[:space:]]Shapes[[:space:]]via[[:space:]]Chirality[[:space:]]Disentanglement/83403940-00e7-485d-b6f5-2066a9a36d93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/mmCooper_[[:space:]]A[[:space:]]Multi-agent[[:space:]]Multi-stage[[:space:]]Communication-efficient[[:space:]]and[[:space:]]Collaboration-robust[[:space:]]Cooperative[[:space:]]Perception[[:space:]]Framework/3ae20890-4db9-45fb-acac-a9261ca916aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/monoVLN_[[:space:]]Bridging[[:space:]]the[[:space:]]Observation[[:space:]]Gap[[:space:]]between[[:space:]]Monocular[[:space:]]and[[:space:]]Panoramic[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Navigation/22fa6baf-2e02-4481-a20e-d89061e5dbe8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/p-AVAS_[[:space:]]Can[[:space:]]Physics-Integrated[[:space:]]Audio-Visual[[:space:]]Modeling[[:space:]]Boost[[:space:]]Neural[[:space:]]Acoustic[[:space:]]Synthesis_/87692093-3778-42e8-bdc6-19e39b1c5e1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2025/p-MoD_[[:space:]]Building[[:space:]]Mixture-of-Depths[[:space:]]MLLMs[[:space:]]via[[:space:]]Progressive[[:space:]]Ratio[[:space:]]Decay/4fc1e63c-b45e-4fff-aea4-5e3d73bc60ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/2D-3D[[:space:]]Interlaced[[:space:]]Transformer[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation[[:space:]]with[[:space:]]Scene-Level[[:space:]]Supervision/795103ed-ec59-49fd-adff-69d819d6a94f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/2D3D-MATR_[[:space:]]2D-3D[[:space:]]Matching[[:space:]]Transformer[[:space:]]for[[:space:]]Detection-Free[[:space:]]Registration[[:space:]]Between[[:space:]]Images[[:space:]]and[[:space:]]Point[[:space:]]Clouds/68525c70-fbeb-4652-9a57-0517f70184c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/360VOT_[[:space:]]A[[:space:]]New[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]for[[:space:]]Omnidirectional[[:space:]]Visual[[:space:]]Object[[:space:]]Tracking/89caa7e2-1f8d-4045-b07b-69d10a586ed8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Distillation_[[:space:]]Improving[[:space:]]Self-Supervised[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation[[:space:]]on[[:space:]]Reflective[[:space:]]Surfaces/940e4977-9767-4aba-96fa-fc3c8ae3c067_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery[[:space:]]with[[:space:]]Sequentially[[:space:]]Global[[:space:]]Rotation[[:space:]]Estimation/2b5b8d13-6922-455b-8df3-6f64fd86c392_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Implicit[[:space:]]Transporter[[:space:]]for[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Keypoint[[:space:]]Discovery/698aff1f-2865-4534-a833-ee939ed30f96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Instance[[:space:]]Segmentation[[:space:]]via[[:space:]]Enhanced[[:space:]]Spatial[[:space:]]and[[:space:]]Semantic[[:space:]]Supervision/9d33cce0-71c3-40c2-8ee4-379c4efd84e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Motion[[:space:]]Magnification_[[:space:]]Visualizing[[:space:]]Subtle[[:space:]]Motions[[:space:]]from[[:space:]]Time-Varying[[:space:]]Radiance[[:space:]]Fields/4df2fb33-78d6-4113-89b4-64d1cbf81974_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Neural[[:space:]]Embedding[[:space:]]Likelihood_[[:space:]]Probabilistic[[:space:]]Inverse[[:space:]]Graphics[[:space:]]for[[:space:]]Robust[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation/c48b080f-f62d-4927-852a-fc6ac1d32ef4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Segmentation[[:space:]]of[[:space:]]Humans[[:space:]]in[[:space:]]Point[[:space:]]Clouds[[:space:]]with[[:space:]]Synthetic[[:space:]]Data/c65358a3-b0da-49d0-a418-76dae25c71ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]Semantic[[:space:]]Subspace[[:space:]]Traverser_[[:space:]]Empowering[[:space:]]3D[[:space:]]Generative[[:space:]]Model[[:space:]]with[[:space:]]Shape[[:space:]]Editing[[:space:]]Capability/b4c79cab-55d2-458d-b2f7-b70ad976fff4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D[[:space:]]VR[[:space:]]Sketch[[:space:]]Guided[[:space:]]3D[[:space:]]Shape[[:space:]]Prototyping[[:space:]]and[[:space:]]Exploration/7651a5d3-6fec-45ce-bc0c-6bf897f82459_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D-Aware[[:space:]]Generative[[:space:]]Model[[:space:]]for[[:space:]]Improved[[:space:]]Side-View[[:space:]]Image[[:space:]]Synthesis/9c2d56de-d217-40ca-abcf-42949762d793_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D-Aware[[:space:]]Neural[[:space:]]Body[[:space:]]Fitting[[:space:]]for[[:space:]]Occlusion[[:space:]]Robust[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/2fa63626-9b1d-415b-b571-b36c444bdefd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D-VisTA_[[:space:]]Pre-trained[[:space:]]Transformer[[:space:]]for[[:space:]]3D[[:space:]]Vision[[:space:]]and[[:space:]]Text[[:space:]]Alignment/3be409d8-f8ae-4578-ad7f-e33a2ec9a2d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D-aware[[:space:]]Blending[[:space:]]with[[:space:]]Generative[[:space:]]NeRFs/c6b7200b-73ab-4d3a-939b-a50441252220_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3D-aware[[:space:]]Image[[:space:]]Generation[[:space:]]using[[:space:]]2D[[:space:]]Diffusion[[:space:]]Models/a1f92a77-46e3-4d3f-a4f1-2561244d5fed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3DHacker_[[:space:]]Spectrum-based[[:space:]]Decision[[:space:]]Boundary[[:space:]]Generation[[:space:]]for[[:space:]]Hard-label[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Attack/de922a90-2feb-49b4-b9ef-6ae72d2ba6c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3DHumanGAN_[[:space:]]3D-Aware[[:space:]]Human[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]3D[[:space:]]Pose[[:space:]]Mapping/be7a13b1-7ec6-43a9-8545-ac43941d6dd4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3DMOTFormer_[[:space:]]Graph[[:space:]]Transformer[[:space:]]for[[:space:]]Online[[:space:]]3D[[:space:]]Multi-Object[[:space:]]Tracking/b9c82067-861f-4512-94b6-942969831585_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3DMiner_[[:space:]]Discovering[[:space:]]Shapes[[:space:]]from[[:space:]]Large-Scale[[:space:]]Unannotated[[:space:]]Image[[:space:]]Datasets/bf290262-08ec-4505-a6e3-21359ef2f98a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/3DPPE_[[:space:]]3D[[:space:]]Point[[:space:]]Positional[[:space:]]Encoding[[:space:]]for[[:space:]]Transformer-based[[:space:]]Multi-Camera[[:space:]]3D[[:space:]]Object[[:space:]]Detection/f25e47ad-b689-4d43-88a2-6a53695395e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/4D[[:space:]]Myocardium[[:space:]]Reconstruction[[:space:]]with[[:space:]]Decoupled[[:space:]]Motion[[:space:]]and[[:space:]]Shape[[:space:]]Model/70020d06-1e1a-490b-8e43-9259afbc9e2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/4D[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]as[[:space:]]Invariant[[:space:]]and[[:space:]]Equivariant[[:space:]]Field[[:space:]]Prediction/0fc37ec4-09b6-4e35-bc1d-03f3b2bc2993_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]5-Point[[:space:]]Minimal[[:space:]]Solver[[:space:]]for[[:space:]]Event[[:space:]]Camera[[:space:]]Relative[[:space:]]Motion[[:space:]]Estimation/c422436e-a272-4ee2-b48b-107bc35d0c58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Benchmark[[:space:]]for[[:space:]]Chinese-English[[:space:]]Scene[[:space:]]Text[[:space:]]Image[[:space:]]Super-Resolution/5cd8e655-6903-4254-9ae3-be1671fa9f11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Complete[[:space:]]Recipe[[:space:]]for[[:space:]]Diffusion[[:space:]]Generative[[:space:]]Models/147ce3f9-90cc-4d35-8bc9-9a28a464c54e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Dynamic[[:space:]]Dual-Processing[[:space:]]Object[[:space:]]Detection[[:space:]]Framework[[:space:]]Inspired[[:space:]]by[[:space:]]the[[:space:]]Brain's[[:space:]]Recognition[[:space:]]Mechanism/2d53dd0a-d7f7-4ef8-b048-c3e0c581cedc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Fast[[:space:]]Unified[[:space:]]System[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]and[[:space:]]Tracking/3b81878d-9c49-42a2-b7a9-09f9ec85ba71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Game[[:space:]]of[[:space:]]Bundle[[:space:]]Adjustment[[:space:]]-[[:space:]]Learning[[:space:]]Efficient[[:space:]]Convergence/c4686e3e-c615-4ccb-8bca-a59e1222ac60_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Generalist[[:space:]]Framework[[:space:]]for[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]of[[:space:]]Images[[:space:]]and[[:space:]]Videos/b6b528c6-e7a4-47a6-a9ba-53bf9313429f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Good[[:space:]]Student[[:space:]]is[[:space:]]Cooperative[[:space:]]and[[:space:]]Reliable_[[:space:]]CNN-Transformer[[:space:]]Collaborative[[:space:]]Learning[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/d04739ce-3b4e-4bc6-801c-50d8fa2b1716_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Large-Scale[[:space:]]Outdoor[[:space:]]Multi-Modal[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]and[[:space:]]Implicit[[:space:]]Scene[[:space:]]Reconstruction/e94be951-0226-4d67-b430-8c286bbc39f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Large-scale[[:space:]]Study[[:space:]]of[[:space:]]Spatiotemporal[[:space:]]Representation[[:space:]]Learning[[:space:]]with[[:space:]]a[[:space:]]New[[:space:]]Benchmark[[:space:]]on[[:space:]]Action[[:space:]]Recognition/163fc3ef-1393-4b84-9c1e-ca21db22c3b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Latent[[:space:]]Space[[:space:]]of[[:space:]]Stochastic[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Image[[:space:]]Editing[[:space:]]and[[:space:]]Guidance/952640e2-2431-4706-99f1-0bc34493537b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Low-Shot[[:space:]]Object[[:space:]]Counting[[:space:]]Network[[:space:]]With[[:space:]]Iterative[[:space:]]Prototype[[:space:]]Adaptation/bb475c47-094c-482c-8b2a-adc1eecc6d8e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Multidimensional[[:space:]]Analysis[[:space:]]of[[:space:]]Social[[:space:]]Biases[[:space:]]in[[:space:]]Vision[[:space:]]Transformers/53956e48-c1c5-44d9-aefa-dd9658a0531a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Parse-Then-Place[[:space:]]Approach[[:space:]]for[[:space:]]Generating[[:space:]]Graphic[[:space:]]Layouts[[:space:]]from[[:space:]]Textual[[:space:]]Descriptions/e9b256d8-eff5-45fa-8342-68b6d6d33d99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Retrospect[[:space:]]to[[:space:]]Multi-prompt[[:space:]]Learning[[:space:]]across[[:space:]]Vision[[:space:]]and[[:space:]]Language/24b9bfd7-6f80-4967-b935-1ff376ff87c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Sentence[[:space:]]Speaks[[:space:]]a[[:space:]]Thousand[[:space:]]Images_[[:space:]]Domain[[:space:]]Generalization[[:space:]]through[[:space:]]Distilling[[:space:]]CLIP[[:space:]]with[[:space:]]Language[[:space:]]Guidance/f83d94e0-76cb-4d46-87ad-93993e89308b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Simple[[:space:]]Framework[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Segmentation[[:space:]]and[[:space:]]Detection/b130a1b4-b119-424c-9744-04602942fb1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Simple[[:space:]]Recipe[[:space:]]to[[:space:]]Meta-Learn[[:space:]]Forward[[:space:]]and[[:space:]]Backward[[:space:]]Transfer/81614f09-5767-4c8b-abc0-bdc69b9a56ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Simple[[:space:]]Vision[[:space:]]Transformer[[:space:]]for[[:space:]]Weakly[[:space:]]Semi-supervised[[:space:]]3D[[:space:]]Object[[:space:]]Detection/2d29d90d-28af-4b87-98eb-4f4c9905cb09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Skeletonization[[:space:]]Algorithm[[:space:]]for[[:space:]]Gradient-Based[[:space:]]Optimization/bac3c801-3295-4593-84cd-d26daa416a7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Soft[[:space:]]Nearest-Neighbor[[:space:]]Framework[[:space:]]for[[:space:]]Continual[[:space:]]Semi-Supervised[[:space:]]Learning/ad7acd04-2c9d-4f89-932e-49d655c127b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Theory[[:space:]]of[[:space:]]Topological[[:space:]]Derivatives[[:space:]]for[[:space:]]Inverse[[:space:]]Rendering[[:space:]]of[[:space:]]Geometry/d67c0fbd-4999-4410-a51c-4fdebf94074f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Unified[[:space:]]Continual[[:space:]]Learning[[:space:]]Framework[[:space:]]with[[:space:]]General[[:space:]]Parameter-Efficient[[:space:]]Tuning/aef84124-cdc4-48e5-8536-853c8c9134f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Robustness[[:space:]]on[[:space:]]Diverse[[:space:]]Sampling[[:space:]]Errors/15990be0-4b65-4af5-8bf7-368df0b35a5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A[[:space:]]step[[:space:]]towards[[:space:]]understanding[[:space:]]why[[:space:]]classification[[:space:]]helps[[:space:]]regression/c01757a5-7df6-4c1e-b50a-186b6f88d39d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A-STAR_[[:space:]]Test-time[[:space:]]Attention[[:space:]]Segregation[[:space:]]and[[:space:]]Retention[[:space:]]for[[:space:]]Text-to-image[[:space:]]Synthesis/c107af91-0063-46a9-b8fd-10dfe7105a81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/A2Q_[[:space:]]Accumulator-Aware[[:space:]]Quantization[[:space:]]with[[:space:]]Guaranteed[[:space:]]Overflow[[:space:]]Avoidance/cc333a04-1ae1-4de4-afe6-3aea8e62adc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ACLS_[[:space:]]Adaptive[[:space:]]and[[:space:]]Conditional[[:space:]]Label[[:space:]]Smoothing[[:space:]]for[[:space:]]Network[[:space:]]Calibration/473d4357-b44b-4777-abdf-1b1fbe63028a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fantasia3D_[[:space:]]Disentangling[[:space:]]Geometry[[:space:]]and[[:space:]]Appearance[[:space:]]for[[:space:]]High-quality[[:space:]]Text-to-3D[[:space:]]Content[[:space:]]Creation/5c5d4c5d-3820-44df-916a-bfeef9c352be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FashionNTM_[[:space:]]Multi-turn[[:space:]]Fashion[[:space:]]Image[[:space:]]Retrieval[[:space:]]via[[:space:]]Cascaded[[:space:]]Memory/96c19156-eda3-4f76-b877-c0799e40849d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fast[[:space:]]Adversarial[[:space:]]Training[[:space:]]with[[:space:]]Smooth[[:space:]]Convergence/23b31082-56d8-432b-80e3-85fc235c4ff5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fast[[:space:]]Full-frame[[:space:]]Video[[:space:]]Stabilization[[:space:]]with[[:space:]]Iterative[[:space:]]Optimization/ad2d1b40-bcdd-45b6-94bd-a7d9f53481b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fast[[:space:]]Globally[[:space:]]Optimal[[:space:]]Surface[[:space:]]Normal[[:space:]]Estimation[[:space:]]from[[:space:]]an[[:space:]]Affine[[:space:]]Correspondence/ecc07b62-9596-44a0-9745-a517e83aa84a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fast[[:space:]]Inference[[:space:]]and[[:space:]]Update[[:space:]]of[[:space:]]Probabilistic[[:space:]]Density[[:space:]]Estimation[[:space:]]on[[:space:]]Trajectory[[:space:]]Prediction/19d05057-5302-4e5f-a291-21f3188ae791_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fast[[:space:]]Neural[[:space:]]Scene[[:space:]]Flow/ffd5b6d9-5f0d-4f78-9688-f85e6ce5fa64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fast[[:space:]]and[[:space:]]Accurate[[:space:]]Transferability[[:space:]]Measurement[[:space:]]by[[:space:]]Evaluating[[:space:]]Intra-class[[:space:]]Feature[[:space:]]Variance/16d96c9b-c625-4fb0-aaab-b679d45a05de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FastRecon_[[:space:]]Few-shot[[:space:]]Industrial[[:space:]]Anomaly[[:space:]]Detection[[:space:]]via[[:space:]]Fast[[:space:]]Feature[[:space:]]Reconstruction/3fc31964-ced1-48a5-b5b7-b2709febc3e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FastViT_[[:space:]]A[[:space:]]Fast[[:space:]]Hybrid[[:space:]]Vision[[:space:]]Transformer[[:space:]]Using[[:space:]]Structural[[:space:]]Reparameterization/63912104-d421-4e14-8cba-166c1f78904b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FateZero_[[:space:]]Fusing[[:space:]]Attentions[[:space:]]for[[:space:]]Zero-shot[[:space:]]Text-based[[:space:]]Video[[:space:]]Editing/e5894958-e0d6-43ea-919c-f4a78223bdd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fcaformer_[[:space:]]Forward[[:space:]]Cross[[:space:]]Attention[[:space:]]in[[:space:]]Hybrid[[:space:]]Vision[[:space:]]Transformer/5726ddf1-d837-42aa-9a96-e54cdcd1f294_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FeatEnHancer_[[:space:]]Enhancing[[:space:]]Hierarchical[[:space:]]Features[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]and[[:space:]]Beyond[[:space:]]Under[[:space:]]Low-Light[[:space:]]Vision/d981278d-dbe2-41ac-9aab-c72923dbfb19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Feature[[:space:]]Modulation[[:space:]]Transformer_[[:space:]]Cross-Refinement[[:space:]]of[[:space:]]Global[[:space:]]Representation[[:space:]]via[[:space:]]High-Frequency[[:space:]]Prior[[:space:]]for[[:space:]]Image[[:space:]]Super-Resolution/358910c2-9bed-4301-a68b-c93d03912323_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Feature[[:space:]]Prediction[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/b7c9cb2d-d935-41b6-bf79-feb2138dade6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Feature[[:space:]]Proliferation[[:space:]]--[[:space:]]the[[:space:]]'Cancer'[[:space:]]in[[:space:]]StyleGAN[[:space:]]and[[:space:]]its[[:space:]]Treatments/585d65ee-f18b-4e36-8941-f08943b3d7c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FeatureNeRF_[[:space:]]Learning[[:space:]]Generalizable[[:space:]]NeRFs[[:space:]]by[[:space:]]Distilling[[:space:]]Foundation[[:space:]]Models/77c2128e-07a6-41a1-8c05-52f5da1cd81d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FedPD_[[:space:]]Federated[[:space:]]Open[[:space:]]Set[[:space:]]Recognition[[:space:]]with[[:space:]]Parameter[[:space:]]Disentanglement/edefee4b-642c-4a0c-ba36-c80cdd791aa1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FedPerfix_[[:space:]]Towards[[:space:]]Partial[[:space:]]Model[[:space:]]Personalization[[:space:]]of[[:space:]]Vision[[:space:]]Transformers[[:space:]]in[[:space:]]Federated[[:space:]]Learning/d543e771-be91-4881-a9ae-44abbc36ec5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Federated[[:space:]]Learning[[:space:]]Over[[:space:]]Images_[[:space:]]Vertical[[:space:]]Decompositions[[:space:]]and[[:space:]]Pre-Trained[[:space:]]Backbones[[:space:]]Are[[:space:]]Difficult[[:space:]]to[[:space:]]Beat/9da08db2-59f8-4fa7-be0b-195e30685450_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FemtoDet_[[:space:]]An[[:space:]]Object[[:space:]]Detection[[:space:]]Baseline[[:space:]]for[[:space:]]Energy[[:space:]]Versus[[:space:]]Performance[[:space:]]Tradeoffs/ae3858f0-bbf1-4f71-acdc-96b1b40f62b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FerKD_[[:space:]]Surgical[[:space:]]Label[[:space:]]Adaptation[[:space:]]for[[:space:]]Efficient[[:space:]]Distillation/37537680-ebe0-4e37-b920-5de9f0458feb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Few[[:space:]]Shot[[:space:]]Font[[:space:]]Generation[[:space:]]Via[[:space:]]Transferring[[:space:]]Similarity[[:space:]]Guided[[:space:]]Global[[:space:]]Style[[:space:]]and[[:space:]]Quantization[[:space:]]Local[[:space:]]Style/d538f1f0-c61f-4b9f-8a6c-8946cb533534_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Few-Shot[[:space:]]Common[[:space:]]Action[[:space:]]Localization[[:space:]]via[[:space:]]Cross-Attentional[[:space:]]Fusion[[:space:]]of[[:space:]]Context[[:space:]]and[[:space:]]Temporal[[:space:]]Dynamics/4021ee8f-1629-4cd1-8f02-6d3ae43c664d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Few-Shot[[:space:]]Dataset[[:space:]]Distillation[[:space:]]via[[:space:]]Translative[[:space:]]Pre-Training/72b4828a-ff69-4894-aad7-b57f445d822d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Few-Shot[[:space:]]Physically-Aware[[:space:]]Articulated[[:space:]]Mesh[[:space:]]Generation[[:space:]]via[[:space:]]Hierarchical[[:space:]]Deformation/557cad77-3efc-4a93-9f6c-1ad645fa6bd4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Few-Shot[[:space:]]Video[[:space:]]Classification[[:space:]]via[[:space:]]Representation[[:space:]]Fusion[[:space:]]and[[:space:]]Promotion[[:space:]]Learning/8b77d936-d037-44d1-977f-aa19f41ff071_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Few-shot[[:space:]]Continual[[:space:]]Infomax[[:space:]]Learning/4bda674e-273c-4a6b-8394-b466c9dde520_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fg-T2M_[[:space:]]Fine-Grained[[:space:]]Text-Driven[[:space:]]Human[[:space:]]Motion[[:space:]]Generation[[:space:]]via[[:space:]]Diffusion[[:space:]]Model/449cb3f1-1251-409f-b17a-0a1369fa81f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fine-grained[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Gait[[:space:]]Recognition/4500f2e6-687c-4c2c-b586-2f14ced6ff0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fine-grained[[:space:]]Visible[[:space:]]Watermark[[:space:]]Removal/3abb59a8-32e1-4765-b180-3a381113d9b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FineDance_[[:space:]]A[[:space:]]Fine-grained[[:space:]]Choreography[[:space:]]Dataset[[:space:]]for[[:space:]]3D[[:space:]]Full[[:space:]]Body[[:space:]]Dance[[:space:]]Generation/dfc4e0d1-1a1e-402c-8793-12d9907332cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FineRecon_[[:space:]]Depth-aware[[:space:]]Feed-forward[[:space:]]Network[[:space:]]for[[:space:]]Detailed[[:space:]]3D[[:space:]]Reconstruction/6c51d26e-268b-4ab2-8995-1d2780bb8d14_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fingerprinting[[:space:]]Deep[[:space:]]Image[[:space:]]Restoration[[:space:]]Models/2d0eb633-eeca-41e3-8998-eafb5489c443_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/First[[:space:]]Session[[:space:]]Adaptation_[[:space:]]A[[:space:]]Strong[[:space:]]Replay-Free[[:space:]]Baseline[[:space:]]for[[:space:]]Class-Incremental[[:space:]]Learning/34f3da95-3b97-4ef3-a0db-22b854de2065_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FishNet_[[:space:]]A[[:space:]]Large-scale[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark[[:space:]]for[[:space:]]Fish[[:space:]]Recognition,[[:space:]]Detection,[[:space:]]and[[:space:]]Functional[[:space:]]Trait[[:space:]]Prediction/76ca1838-244d-4142-bae4-84af44e0f40e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Flatness-Aware[[:space:]]Minimization[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/493c4d3d-8eb2-483f-be83-a0524b08186b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Flexible[[:space:]]Visual[[:space:]]Recognition[[:space:]]by[[:space:]]Evidential[[:space:]]Modeling[[:space:]]of[[:space:]]Confusion[[:space:]]and[[:space:]]Ignorance/c13c514a-2666-4045-89a1-55d246dbfa1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FlipNeRF_[[:space:]]Flipped[[:space:]]Reflection[[:space:]]Rays[[:space:]]for[[:space:]]Few-shot[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/b773a3a0-692f-4508-af4b-b8359e51e763_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Focal[[:space:]]Network[[:space:]]for[[:space:]]Image[[:space:]]Restoration/8fcb7aa2-a8ca-4dca-97f1-3bf63a15581e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FocalFormer3D_[[:space:]]Focusing[[:space:]]on[[:space:]]Hard[[:space:]]Instance[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/9285b2b7-cbfa-4b94-8361-0cc2abdf86b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Focus[[:space:]]on[[:space:]]Your[[:space:]]Target_[[:space:]]A[[:space:]]Dual[[:space:]]Teacher-Student[[:space:]]Framework[[:space:]]for[[:space:]]Domain-Adaptive[[:space:]]Semantic[[:space:]]Segmentation/93139a7c-5176-4d19-a1d3-a4deb1b068c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Focus[[:space:]]the[[:space:]]Discrepancy_[[:space:]]Intra-[[:space:]]and[[:space:]]Inter-Correlation[[:space:]]Learning[[:space:]]for[[:space:]]Image[[:space:]]Anomaly[[:space:]]Detection/9a627ddc-8c15-4383-a8ac-8266b31abdaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Forecast-MAE_[[:space:]]Self-supervised[[:space:]]Pre-training[[:space:]]for[[:space:]]Motion[[:space:]]Forecasting[[:space:]]with[[:space:]]Masked[[:space:]]Autoencoders/f4fc3c51-7cb9-491a-9580-072217774d3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Foreground[[:space:]]Object[[:space:]]Search[[:space:]]by[[:space:]]Distilling[[:space:]]Composite[[:space:]]Image[[:space:]]Feature/2f22dc81-3b09-4682-9d0e-7e47efe03306_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Foreground[[:space:]]and[[:space:]]Text-lines[[:space:]]Aware[[:space:]]Document[[:space:]]Image[[:space:]]Rectification/3650c92e-47cd-4562-b763-9e2e457518cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Foreground-Background[[:space:]]Distribution[[:space:]]Modeling[[:space:]]Transformer[[:space:]]for[[:space:]]Visual[[:space:]]Object[[:space:]]Tracking/82259de0-16cb-4749-9fd9-5e68b3f9f014_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Foreground-Background[[:space:]]Separation[[:space:]]through[[:space:]]Concept[[:space:]]Distillation[[:space:]]from[[:space:]]Generative[[:space:]]Image[[:space:]]Foundation[[:space:]]Models/5ee0597f-a7d5-4c09-a624-657832eb9d0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Forward[[:space:]]Flow[[:space:]]for[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]of[[:space:]]Dynamic[[:space:]]Scenes/0cdaae9f-cb85-4e33-8452-a1e60f8c4484_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FreeCOS_[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]from[[:space:]]Fractals[[:space:]]and[[:space:]]Unlabeled[[:space:]]Images[[:space:]]for[[:space:]]Curvilinear[[:space:]]Object[[:space:]]Segmentation/be0041d5-bbab-4e78-87f4-f20982dd3329_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FreeDoM_[[:space:]]Training-Free[[:space:]]Energy-Guided[[:space:]]Conditional[[:space:]]Diffusion[[:space:]]Model/6856ff82-c802-4528-9795-9924d03d4b41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Frequency[[:space:]]Guidance[[:space:]]Matters[[:space:]]in[[:space:]]Few-Shot[[:space:]]Learning/de0a02db-ed96-44b6-8f21-932746535501_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Frequency-aware[[:space:]]GAN[[:space:]]for[[:space:]]Adversarial[[:space:]]Manipulation[[:space:]]Generation/8bcc3433-dc75-4f7a-97ef-3808dc50d15f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/From[[:space:]]Chaos[[:space:]]Comes[[:space:]]Order_[[:space:]]Ordering[[:space:]]Event[[:space:]]Representations[[:space:]]for[[:space:]]Object[[:space:]]Recognition[[:space:]]and[[:space:]]Detection/d3f1e4dc-e3b8-48be-905a-0246e42863b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/From[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]to[[:space:]]Self-Knowledge[[:space:]]Distillation_[[:space:]]A[[:space:]]Unified[[:space:]]Approach[[:space:]]with[[:space:]]Normalized[[:space:]]Loss[[:space:]]and[[:space:]]Customized[[:space:]]Soft[[:space:]]Labels/7e406dac-df76-4483-b6be-35c6233f9e4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/From[[:space:]]Sky[[:space:]]to[[:space:]]the[[:space:]]Ground_[[:space:]]A[[:space:]]Large-scale[[:space:]]Benchmark[[:space:]]and[[:space:]]Simple[[:space:]]Baseline[[:space:]]Towards[[:space:]]Real[[:space:]]Rain[[:space:]]Removal/8d7e8632-869f-4c0d-a983-de51fca56097_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FrozenRecon_[[:space:]]Pose-free[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]with[[:space:]]Frozen[[:space:]]Depth[[:space:]]Models/db9b9018-f8d0-4575-958b-789f4b1d37cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Full-Body[[:space:]]Articulated[[:space:]]Human-Object[[:space:]]Interaction/4a2d2a53-946a-4894-a2ff-6a5f2112ea35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fully[[:space:]]Attentional[[:space:]]Networks[[:space:]]with[[:space:]]Self-emerging[[:space:]]Token[[:space:]]Labeling/caf0e66c-9593-4110-bbb4-fe33ef9d0d27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FunnyBirds_[[:space:]]A[[:space:]]Synthetic[[:space:]]Vision[[:space:]]Dataset[[:space:]]for[[:space:]]a[[:space:]]Part-Based[[:space:]]Analysis[[:space:]]of[[:space:]]Explainable[[:space:]]AI[[:space:]]Methods/07a936e0-4ced-42c0-9b0c-dba61ee8ef3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/G2L_[[:space:]]Semantically[[:space:]]Aligned[[:space:]]and[[:space:]]Uniform[[:space:]]Video[[:space:]]Grounding[[:space:]]via[[:space:]]Geodesic[[:space:]]and[[:space:]]Game[[:space:]]Theory/93524482-75b6-4526-b597-d7ac391c3846_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GACE_[[:space:]]Geometry[[:space:]]Aware[[:space:]]Confidence[[:space:]]Enhancement[[:space:]]for[[:space:]]Black-Box[[:space:]]3D[[:space:]]Object[[:space:]]Detectors[[:space:]]on[[:space:]]LiDAR-Data/4d4269c1-ea16-4959-9423-14c51ea10997_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GAFlow_[[:space:]]Incorporating[[:space:]]Gaussian[[:space:]]Attention[[:space:]]into[[:space:]]Optical[[:space:]]Flow/b25f2d39-f04e-4319-bbbf-98d2b295ade7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GAIT_[[:space:]]Generating[[:space:]]Aesthetic[[:space:]]Indoor[[:space:]]Tours[[:space:]]with[[:space:]]Deep[[:space:]]Reinforcement[[:space:]]Learning/04ba10a4-1e2a-4899-bb30-52b2658d3ca3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GECCO_[[:space:]]Geometrically-Conditioned[[:space:]]Point[[:space:]]Diffusion[[:space:]]Models/611629e6-6dc8-4865-b046-bb119108fba5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GEDepth_[[:space:]]Ground[[:space:]]Embedding[[:space:]]for[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/f5ecc1a6-422b-4c7e-892e-40a1a1d76504_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GETAvatar_[[:space:]]Generative[[:space:]]Textured[[:space:]]Meshes[[:space:]]for[[:space:]]Animatable[[:space:]]Human[[:space:]]Avatars/c8126114-d920-42c1-9bf4-5564d967b51e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GET_[[:space:]]Group[[:space:]]Event[[:space:]]Transformer[[:space:]]for[[:space:]]Event-Based[[:space:]]Vision/ff41b5a3-d7f5-415a-bed0-a19c923c4d13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GIFD_[[:space:]]A[[:space:]]Generative[[:space:]]Gradient[[:space:]]Inversion[[:space:]]Method[[:space:]]with[[:space:]]Feature[[:space:]]Domain[[:space:]]Optimization/17a535d7-b5f7-4771-a8b8-5dba5e56d0fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GLA-GCN_[[:space:]]Global-local[[:space:]]Adaptive[[:space:]]Graph[[:space:]]Convolutional[[:space:]]Network[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]from[[:space:]]Monocular[[:space:]]Video/b0eb3d3b-7d22-4407-8c2a-5fbe60f4e0f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GO-SLAM_[[:space:]]Global[[:space:]]Optimization[[:space:]]for[[:space:]]Consistent[[:space:]]3D[[:space:]]Instant[[:space:]]Reconstruction/364771f8-b04a-46fe-8ac3-e9da14debff5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GPA-3D_[[:space:]]Geometry-aware[[:space:]]Prototype[[:space:]]Alignment[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptive[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]from[[:space:]]Point[[:space:]]Clouds/face0a89-e82d-4526-a7a3-5e7b6bb9613e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GPFL_[[:space:]]Simultaneously[[:space:]]Learning[[:space:]]Global[[:space:]]and[[:space:]]Personalized[[:space:]]Feature[[:space:]]Information[[:space:]]for[[:space:]]Personalized[[:space:]]Federated[[:space:]]Learning/cea35509-746c-4db9-ab86-21390527316f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GPGait_[[:space:]]Generalized[[:space:]]Pose-based[[:space:]]Gait[[:space:]]Recognition/44bdc39f-9008-48dd-a141-cc76b96840e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GRAM-HD_[[:space:]]3D-Consistent[[:space:]]Image[[:space:]]Generation[[:space:]]at[[:space:]]High[[:space:]]Resolution[[:space:]]with[[:space:]]Generative[[:space:]]Radiance[[:space:]]Manifolds/76b956c0-f80b-4f3a-a5da-4d0648a8e06f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GaPro_[[:space:]]Box-Supervised[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Instance[[:space:]]Segmentation[[:space:]]Using[[:space:]]Gaussian[[:space:]]Processes[[:space:]]as[[:space:]]Pseudo[[:space:]]Labelers/5af19852-cfa8-4d73-9726-a85da972e1ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GameFormer_[[:space:]]Game-theoretic[[:space:]]Modeling[[:space:]]and[[:space:]]Learning[[:space:]]of[[:space:]]Transformer-based[[:space:]]Interactive[[:space:]]Prediction[[:space:]]and[[:space:]]Planning[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/06003916-848c-40d7-9139-3ed92acb8fb2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GasMono_[[:space:]]Geometry-Aided[[:space:]]Self-Supervised[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation[[:space:]]for[[:space:]]Indoor[[:space:]]Scenes/43dd8a5b-7260-470f-bbbd-2bb72e068faa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GePSAn_[[:space:]]Generative[[:space:]]Procedure[[:space:]]Step[[:space:]]Anticipation[[:space:]]in[[:space:]]Cooking[[:space:]]Videos/cff9c64b-0ec8-4c94-af71-5544ee4d8283_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GeT_[[:space:]]Generative[[:space:]]Target[[:space:]]Structure[[:space:]]Debiasing[[:space:]]for[[:space:]]Domain[[:space:]]Adaptation/b7a8b417-ca77-49f6-aad6-733e7ebca530_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gender[[:space:]]Artifacts[[:space:]]in[[:space:]]Visual[[:space:]]Datasets/e59e6d8f-5f2f-4f81-bb49-982635e41e5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/General[[:space:]]Image-to-Image[[:space:]]Translation[[:space:]]with[[:space:]]One-Shot[[:space:]]Image[[:space:]]Guidance/8746485e-c846-41f2-aee2-2337aea693f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/General[[:space:]]Planar[[:space:]]Motion[[:space:]]from[[:space:]]a[[:space:]]Pair[[:space:]]of[[:space:]]3D[[:space:]]Correspondences/f0bd6d2c-c5a0-4a9e-bf76-a76b38f1bd26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalizable[[:space:]]Decision[[:space:]]Boundaries_[[:space:]]Dualistic[[:space:]]Meta-Learning[[:space:]]for[[:space:]]Open[[:space:]]Set[[:space:]]Domain[[:space:]]Generalization/5eb1251b-94a4-48a6-9ba9-c2790464cd1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalizable[[:space:]]Neural[[:space:]]Fields[[:space:]]as[[:space:]]Partially[[:space:]]Observed[[:space:]]Neural[[:space:]]Processes/df74fa14-9d5e-4099-bb99-9f116dc41ff3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalized[[:space:]]Differentiable[[:space:]]RANSAC/bf8419d4-814a-484e-b729-cb7bc4e0e94e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalized[[:space:]]Few-Shot[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation[[:space:]]via[[:space:]]Geometric[[:space:]]Words/e2a12ac6-f8dc-44ba-86d9-79e807ec6ec7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalized[[:space:]]Lightness[[:space:]]Adaptation[[:space:]]with[[:space:]]Channel[[:space:]]Selective[[:space:]]Normalization/82cf6352-34bf-4246-bcdb-c8ed71254e69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalized[[:space:]]Sum[[:space:]]Pooling[[:space:]]for[[:space:]]Metric[[:space:]]Learning/6af11e08-e03d-45c8-bee4-df940ea0af2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalizing[[:space:]]Event-Based[[:space:]]Motion[[:space:]]Deblurring[[:space:]]in[[:space:]]Real-World[[:space:]]Scenarios/a3d9ba6c-db40-42f8-9d9d-103bccdabd2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generalizing[[:space:]]Neural[[:space:]]Human[[:space:]]Fitting[[:space:]]to[[:space:]]Unseen[[:space:]]Poses[[:space:]]With[[:space:]]Articulated[[:space:]]SE(3)[[:space:]]Equivariance/91278764-41e8-4233-9ea3-a37abd73e5d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generating[[:space:]]Dynamic[[:space:]]Kernels[[:space:]]via[[:space:]]Transformers[[:space:]]for[[:space:]]Lane[[:space:]]Detection/c8c321b3-0af5-4163-93f0-ca09862815a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generating[[:space:]]Instance-level[[:space:]]Prompts[[:space:]]for[[:space:]]Rehearsal-free[[:space:]]Continual[[:space:]]Learning/5b8dff06-947a-49a1-be3a-99252d12b332_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generating[[:space:]]Realistic[[:space:]]Images[[:space:]]from[[:space:]]In-the-wild[[:space:]]Sounds/5d4abf93-b22c-4037-a0d6-b1e30018b505_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generating[[:space:]]Visual[[:space:]]Scenes[[:space:]]from[[:space:]]Touch/da31f934-c707-4fe7-98b2-c1b00f383226_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generative[[:space:]]Action[[:space:]]Description[[:space:]]Prompts[[:space:]]for[[:space:]]Skeleton-based[[:space:]]Action[[:space:]]Recognition/17278522-bd9e-441a-985b-164a25c7e258_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generative[[:space:]]Gradient[[:space:]]Inversion[[:space:]]via[[:space:]]Over-Parameterized[[:space:]]Networks[[:space:]]in[[:space:]]Federated[[:space:]]Learning/8ac32b01-9f08-41e3-9c13-f4e826822f66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generative[[:space:]]Multiplane[[:space:]]Neural[[:space:]]Radiance[[:space:]]for[[:space:]]3D-Aware[[:space:]]Image[[:space:]]Generation/18b59b1b-b78a-4860-bd1b-40401f17988b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generative[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]3D-Aware[[:space:]]Diffusion[[:space:]]Models/7f1568de-8554-4f70-a788-975f161c4a3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Generative[[:space:]]Prompt[[:space:]]Model[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Object[[:space:]]Localization/f97bcd14-33fc-485b-ae4a-40bcbf47c58f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GeoMIM_[[:space:]]Towards[[:space:]]Better[[:space:]]3D[[:space:]]Knowledge[[:space:]]Transfer[[:space:]]via[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling[[:space:]]for[[:space:]]Multi-view[[:space:]]3D[[:space:]]Understanding/6799a7e2-aa26-47cc-b525-cc3cce31a086_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GeoUDF_[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]from[[:space:]]3D[[:space:]]Point[[:space:]]Clouds[[:space:]]via[[:space:]]Geometry-guided[[:space:]]Distance[[:space:]]Representation/96b8ebfa-5e31-4c56-afd3-c6f0347c8b9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Geometric[[:space:]]Viewpoint[[:space:]]Learning[[:space:]]with[[:space:]]Hyper-Rays[[:space:]]and[[:space:]]Harmonics[[:space:]]Encoding/c8f099b2-e69c-4454-939c-f9fa2a48baae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Geometrized[[:space:]]Transformer[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Homography[[:space:]]Estimation/0ff591f7-f322-49fb-bbec-dc7f3d1bf93e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Geometry-guided[[:space:]]Feature[[:space:]]Learning[[:space:]]and[[:space:]]Fusion[[:space:]]for[[:space:]]Indoor[[:space:]]Scene[[:space:]]Reconstruction/6c529b26-c681-429e-adda-d5d127c807f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Get[[:space:]]the[[:space:]]Best[[:space:]]of[[:space:]]Both[[:space:]]Worlds_[[:space:]]Improving[[:space:]]Accuracy[[:space:]]and[[:space:]]Transferability[[:space:]]by[[:space:]]Grassmann[[:space:]]Class[[:space:]]Representation/1b680402-a9a6-42ce-b862-c884eddebc43_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Get3DHuman_[[:space:]]Lifting[[:space:]]StyleGAN-Human[[:space:]]into[[:space:]]a[[:space:]]3D[[:space:]]Generative[[:space:]]Model[[:space:]]Using[[:space:]]Pixel-Aligned[[:space:]]Reconstruction[[:space:]]Priors/924997c1-f277-4fa5-921a-76f9d175dde7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Global[[:space:]]Adaptation[[:space:]]Meets[[:space:]]Local[[:space:]]Generalization_[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/ebda4117-98a5-4e55-a74b-e7c5dcf1dba5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Global[[:space:]]Balanced[[:space:]]Experts[[:space:]]for[[:space:]]Federated[[:space:]]Long-Tailed[[:space:]]Learning/e02ed9f1-14e4-4ee0-84d7-b2ea2ce44fe4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Global[[:space:]]Features[[:space:]]are[[:space:]]All[[:space:]]You[[:space:]]Need[[:space:]]for[[:space:]]Image[[:space:]]Retrieval[[:space:]]and[[:space:]]Reranking/5abfe831-5f9d-4032-9b7a-2e5ff4dfe904_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Global[[:space:]]Knowledge[[:space:]]Calibration[[:space:]]for[[:space:]]Fast[[:space:]]Open-Vocabulary[[:space:]]Segmentation/19a3cf55-549f-4105-b447-ec9384e528df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Global[[:space:]]Perception[[:space:]]Based[[:space:]]Autoregressive[[:space:]]Neural[[:space:]]Processes/1cc2c17d-7e2e-4353-ba09-6d3b854e8d21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GlobalMapper_[[:space:]]Arbitrary-Shaped[[:space:]]Urban[[:space:]]Layout[[:space:]]Generation/74f2331d-ff45-4e63-9a1f-5662b5ffdb31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gloss-Free[[:space:]]Sign[[:space:]]Language[[:space:]]Translation_[[:space:]]Improving[[:space:]]from[[:space:]]Visual-Language[[:space:]]Pretraining/cd6e9027-184a-44e1-a972-5d058001a3d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GlowGAN_[[:space:]]Unsupervised[[:space:]]Learning[[:space:]]of[[:space:]]HDR[[:space:]]Images[[:space:]]from[[:space:]]LDR[[:space:]]Images[[:space:]]in[[:space:]]the[[:space:]]Wild/842d2893-579c-455f-b796-382b3f5fb3e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GlueGen_[[:space:]]Plug[[:space:]]and[[:space:]]Play[[:space:]]Multi-modal[[:space:]]Encoders[[:space:]]for[[:space:]]X-to-image[[:space:]]Generation/6ac3a0c3-ca4b-46ae-a6f8-2ddaeed2ad8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GlueStick_[[:space:]]Robust[[:space:]]Image[[:space:]]Matching[[:space:]]by[[:space:]]Sticking[[:space:]]Points[[:space:]]and[[:space:]]Lines[[:space:]]Together/be32c2ea-a188-4693-933d-14fb72c51dfd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Going[[:space:]]Beyond[[:space:]]Nouns[[:space:]]With[[:space:]]Vision[[:space:]]&[[:space:]]Language[[:space:]]Models[[:space:]]Using[[:space:]]Synthetic[[:space:]]Data/8eeda7b0-f729-4c97-a9a8-4363ff8b63ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Going[[:space:]]Denser[[:space:]]with[[:space:]]Open-Vocabulary[[:space:]]Part[[:space:]]Segmentation/153b658d-6bfc-45b5-8ad9-c0001bb6eb05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gradient-Regulated[[:space:]]Meta-Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Generalizable[[:space:]]Vision-Language[[:space:]]Models/07b323a0-6097-4ee8-a487-4c3b59981f73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gradient-based[[:space:]]Sampling[[:space:]]for[[:space:]]Class[[:space:]]Imbalanced[[:space:]]Semi-supervised[[:space:]]Object[[:space:]]Detection/64f16eeb-dc06-4a74-9909-179e1c4d9a47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gram-based[[:space:]]Attentive[[:space:]]Neural[[:space:]]Ordinary[[:space:]]Differential[[:space:]]Equations[[:space:]]Network[[:space:]]for[[:space:]]Video[[:space:]]Nystagmography[[:space:]]Classification/f70bedf4-6c05-4ac1-8b05-2f58577d69e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Gramian[[:space:]]Attention[[:space:]]Heads[[:space:]]are[[:space:]]Strong[[:space:]]yet[[:space:]]Efficient[[:space:]]Vision[[:space:]]Learners/f2f327c5-7b7d-4958-93f4-0223a5bc14eb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Graph[[:space:]]Matching[[:space:]]with[[:space:]]Bi-level[[:space:]]Noisy[[:space:]]Correspondence/86964a07-63ce-497f-b058-80c473c3d4cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GraphAlign_[[:space:]]Enhancing[[:space:]]Accurate[[:space:]]Feature[[:space:]]Alignment[[:space:]]by[[:space:]]Graph[[:space:]]matching[[:space:]]for[[:space:]]Multi-Modal[[:space:]]3D[[:space:]]Object[[:space:]]Detection/490b4ef0-b516-4577-a0d6-c7455946bac0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GraphEcho_[[:space:]]Graph-Driven[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Echocardiogram[[:space:]]Video[[:space:]]Segmentation/594040d6-53b4-4c11-aa75-a08204fcd0de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Graphics2RAW_[[:space:]]Mapping[[:space:]]Computer[[:space:]]Graphics[[:space:]]Images[[:space:]]to[[:space:]]Sensor[[:space:]]RAW[[:space:]]Images/aa2b7962-2dff-4878-8a58-f91afd985cfb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GridMM_[[:space:]]Grid[[:space:]]Memory[[:space:]]Map[[:space:]]for[[:space:]]Vision-and-Language[[:space:]]Navigation/3a3a473d-aae4-427d-8997-1caf744d0ceb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GridPull_[[:space:]]Towards[[:space:]]Scalability[[:space:]]in[[:space:]]Learning[[:space:]]Implicit[[:space:]]Representations[[:space:]]from[[:space:]]3D[[:space:]]Point[[:space:]]Clouds/694f653d-a2d1-4ef0-9ac5-0fdfb6b8b37d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Grounded[[:space:]]Entity-Landmark[[:space:]]Adaptive[[:space:]]Pre-Training[[:space:]]for[[:space:]]Vision-and-Language[[:space:]]Navigation/66797f14-cf57-48b6-98a1-7ce74385a6fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Grounded[[:space:]]Image[[:space:]]Text[[:space:]]Matching[[:space:]]with[[:space:]]Mismatched[[:space:]]Relation[[:space:]]Reasoning/37f1017e-5814-46cc-ba5e-1fd9ffaa6a28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Grounding[[:space:]]3D[[:space:]]Object[[:space:]]Affordance[[:space:]]from[[:space:]]2D[[:space:]]Interactions[[:space:]]in[[:space:]]Images/7f72f78a-6126-465f-bd5b-53bbbe4b1d49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Group[[:space:]]DETR_[[:space:]]Fast[[:space:]]DETR[[:space:]]Training[[:space:]]with[[:space:]]Group-Wise[[:space:]]One-to-Many[[:space:]]Assignment/6e9ced6c-ddf3-4570-9bfd-71f343a00212_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Group[[:space:]]Pose_[[:space:]]A[[:space:]]Simple[[:space:]]Baseline[[:space:]]for[[:space:]]End-to-End[[:space:]]Multi-Person[[:space:]]Pose[[:space:]]Estimation/ebcfd119-a15f-45f3-824b-f01682d5c7b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/GrowCLIP_[[:space:]]Data-Aware[[:space:]]Automatic[[:space:]]Model[[:space:]]Growing[[:space:]]for[[:space:]]Large-scale[[:space:]]Contrastive[[:space:]]Language-Image[[:space:]]Pre-Training/84827675-3aac-4157-b576-07a125b493f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Growing[[:space:]]a[[:space:]]Brain[[:space:]]with[[:space:]]Sparsity-Inducing[[:space:]]Generation[[:space:]]for[[:space:]]Continual[[:space:]]Learning/d35af5fa-bfe4-478d-92d6-454b21b1d4e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Guided[[:space:]]Motion[[:space:]]Diffusion[[:space:]]for[[:space:]]Controllable[[:space:]]Human[[:space:]]Motion[[:space:]]Synthesis/e3abf2fb-5ad1-4ea8-9128-1e0bdc10e0b0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Guiding[[:space:]]Image[[:space:]]Captioning[[:space:]]Models[[:space:]]Toward[[:space:]]More[[:space:]]Specific[[:space:]]Captions/17ad3fd0-4183-40a9-ac84-18391fe9effe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Guiding[[:space:]]Local[[:space:]]Feature[[:space:]]Matching[[:space:]]with[[:space:]]Surface[[:space:]]Curvature/e3a43d92-ec36-432b-94e9-c4f32da40a2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/H3WB_[[:space:]]Human3.6M[[:space:]]3D[[:space:]]WholeBody[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark/a8f8b3db-4bb9-47fd-9ae5-943dc0af642d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HAL3D_[[:space:]]Hierarchical[[:space:]]Active[[:space:]]Learning[[:space:]]for[[:space:]]Fine-Grained[[:space:]]3D[[:space:]]Part[[:space:]]Labeling/4d9d11c1-a387-4840-b228-bd93e87c1665_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HDG-ODE_[[:space:]]A[[:space:]]Hierarchical[[:space:]]Continuous-Time[[:space:]]Model[[:space:]]for[[:space:]]Human[[:space:]]Pose[[:space:]]Forecasting/b32b33ed-c49a-4dab-8f95-5b4dd7695050_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HM-ViT_[[:space:]]Hetero-Modal[[:space:]]Vehicle-to-Vehicle[[:space:]]Cooperative[[:space:]]Perception[[:space:]]with[[:space:]]Vision[[:space:]]Transformer/9d343628-a95c-47d2-8ca4-db20691fefde_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HMD-NeMo_[[:space:]]Online[[:space:]]3D[[:space:]]Avatar[[:space:]]Motion[[:space:]]Generation[[:space:]]From[[:space:]]Sparse[[:space:]]Observations/c461bfb7-8a1f-4d23-a5a0-eb903d127c6c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HOSNeRF_[[:space:]]Dynamic[[:space:]]Human-Object-Scene[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Video/56bcebb1-a8ab-4faa-8362-f67dd5395540_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HRS-Bench_[[:space:]]Holistic,[[:space:]]Reliable[[:space:]]and[[:space:]]Scalable[[:space:]]Benchmark[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Models/c90d4d8a-8b52-4c5e-b753-56dc5d34d841_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HSE_[[:space:]]Hybrid[[:space:]]Species[[:space:]]Embedding[[:space:]]for[[:space:]]Deep[[:space:]]Metric[[:space:]]Learning/debbcae4-e366-4ac1-b461-19f60fcf7463_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HSR-Diff_[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Conditional[[:space:]]Diffusion[[:space:]]Models/aba8b2d5-ebce-4157-8dac-629ebe1b6925_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HTML_[[:space:]]Hybrid[[:space:]]Temporal-scale[[:space:]]Multimodal[[:space:]]Learning[[:space:]]Framework[[:space:]]for[[:space:]]Referring[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/17b3ea75-567e-4430-9688-bf87a6d34fd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HaMuCo_[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Multiview[[:space:]]Collaborative[[:space:]]Self-Supervised[[:space:]]Learning/20ed3f4b-5914-4843-b685-abff65a0a8b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HairCLIPv2_[[:space:]]Unifying[[:space:]]Hair[[:space:]]Editing[[:space:]]via[[:space:]]Proxy[[:space:]]Feature[[:space:]]Blending/ba9719c6-e989-4b1d-bc85-7328ef2439a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HairNeRF_[[:space:]]Geometry-Aware[[:space:]]Image[[:space:]]Synthesis[[:space:]]for[[:space:]]Hairstyle[[:space:]]Transfer/ad8409a2-1fc3-480b-99f2-83a8ffcfe8c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hallucination[[:space:]]Improves[[:space:]]the[[:space:]]Performance[[:space:]]of[[:space:]]Unsupervised[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/d1539e1e-e0b7-40bf-9663-58331a6a93fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HandR2N2_[[:space:]]Iterative[[:space:]]3D[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation[[:space:]]Using[[:space:]]a[[:space:]]Residual[[:space:]]Recurrent[[:space:]]Neural[[:space:]]Network/431e0555-f342-4159-8204-85714e05b7c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Handwritten[[:space:]]and[[:space:]]Printed[[:space:]]Text[[:space:]]Segmentation_[[:space:]]A[[:space:]]Signature[[:space:]]Case[[:space:]]Study/84e181a3-72a0-4a74-a144-02e187b2dd2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hard[[:space:]]No-Box[[:space:]]Adversarial[[:space:]]Attack[[:space:]]on[[:space:]]Skeleton-Based[[:space:]]Human[[:space:]]Action[[:space:]]Recognition[[:space:]]with[[:space:]]Skeleton-Motion-Informed[[:space:]]Gradient/58af11a8-198e-4cb8-99cf-9272139ca7d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Harnessing[[:space:]]the[[:space:]]Spatial-Temporal[[:space:]]Attention[[:space:]]of[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Text-to-Image[[:space:]]Synthesis/ae9152f5-e620-4815-87af-2959ad2b524b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Harvard[[:space:]]Glaucoma[[:space:]]Detection[[:space:]]and[[:space:]]Progression_[[:space:]]A[[:space:]]Multimodal[[:space:]]Multitask[[:space:]]Dataset[[:space:]]and[[:space:]]Generalization-Reinforced[[:space:]]Semi-Supervised[[:space:]]Learning/ed5ae2e0-94e5-433b-855f-28b687bebad2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hashing[[:space:]]Neural[[:space:]]Video[[:space:]]Decomposition[[:space:]]with[[:space:]]Multiplicative[[:space:]]Residuals[[:space:]]in[[:space:]]Space-Time/4db886e2-61ae-4d82-b252-4c366d1ec299_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Helping[[:space:]]Hands_[[:space:]]An[[:space:]]Object-Aware[[:space:]]Ego-Centric[[:space:]]Video[[:space:]]Recognition[[:space:]]Model/605acb09-5a50-4e6f-a563-f1799bbb19a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Heterogeneous[[:space:]]Diversity[[:space:]]Driven[[:space:]]Active[[:space:]]Learning[[:space:]]for[[:space:]]Multi-Object[[:space:]]Tracking/b3ed4007-51fd-4d18-ab48-0223562578b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Heterogeneous[[:space:]]Forgetting[[:space:]]Compensation[[:space:]]for[[:space:]]Class-Incremental[[:space:]]Learning/fd7dde4b-f839-40f9-bfea-101cb8081cd9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HiFace_[[:space:]]High-Fidelity[[:space:]]3D[[:space:]]Face[[:space:]]Reconstruction[[:space:]]by[[:space:]]Learning[[:space:]]Static[[:space:]]and[[:space:]]Dynamic[[:space:]]Details/a0404738-6577-4978-90b9-452e8141d3d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HiLo_[[:space:]]Exploiting[[:space:]]High[[:space:]]Low[[:space:]]Frequency[[:space:]]Relations[[:space:]]for[[:space:]]Unbiased[[:space:]]Panoptic[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/14111a75-4cc9-4c7b-8b2c-bff17e2c80b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HiTeA_[[:space:]]Hierarchical[[:space:]]Temporal-Aware[[:space:]]Video-Language[[:space:]]Pre-training/122f003d-e493-4776-85d5-ea92cef4fff9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HiVLP_[[:space:]]Hierarchical[[:space:]]Interactive[[:space:]]Video-Language[[:space:]]Pre-Training/e5850211-fab6-4860-b782-c2a5d6bad4e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hidden[[:space:]]Biases[[:space:]]of[[:space:]]End-to-End[[:space:]]Driving[[:space:]]Models/8f4d684a-1555-4674-8f25-206ee2d4af90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hiding[[:space:]]Visual[[:space:]]Information[[:space:]]via[[:space:]]Obfuscating[[:space:]]Adversarial[[:space:]]Perturbations/a690bb1f-1210-4744-bd66-f678e3096b46_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Pattern-Generalizable[[:space:]]Image[[:space:]]Corruption[[:space:]]Detection/2454a4c4-9bf9-4a9e-a255-d46d7a04e834_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Generation[[:space:]]of[[:space:]]Human-Object[[:space:]]Interactions[[:space:]]with[[:space:]]Diffusion[[:space:]]Probabilistic[[:space:]]Models/c8490beb-3559-4e1e-833f-b399932e108c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Point-based[[:space:]]Active[[:space:]]Learning[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Semantic[[:space:]]Segmentation/67399329-b319-4f94-bd3d-1cdd42128783_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Prior[[:space:]]Mining[[:space:]]for[[:space:]]Non-local[[:space:]]Multi-View[[:space:]]Stereo/32540140-b677-4721-9ff6-75d372d0916d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Spatio-Temporal[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Gait[[:space:]]Recognition/6240fae9-3133-48a9-86fe-481f88ab16fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Visual[[:space:]]Categories[[:space:]]Modeling_[[:space:]]A[[:space:]]Joint[[:space:]]Representation[[:space:]]Learning[[:space:]]and[[:space:]]Density[[:space:]]Estimation[[:space:]]Framework[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection/bd3288f5-edbd-4923-8ad4-4c190bb5dbf1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchical[[:space:]]Visual[[:space:]]Primitive[[:space:]]Experts[[:space:]]for[[:space:]]Compositional[[:space:]]Zero-Shot[[:space:]]Learning/d892901c-ab65-4154-a39e-e041473b0ce9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hierarchically[[:space:]]Decomposed[[:space:]]Graph[[:space:]]Convolutional[[:space:]]Networks[[:space:]]for[[:space:]]Skeleton-Based[[:space:]]Action[[:space:]]Recognition/4da439a0-ed06-4c50-a7bd-03aa64888b02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High[[:space:]]Quality[[:space:]]Entity[[:space:]]Segmentation/8c45756a-01dd-4842-b194-1be6a90fce5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/High-Resolution[[:space:]]Document[[:space:]]Shadow[[:space:]]Removal[[:space:]]via[[:space:]]A[[:space:]]Large-Scale[[:space:]]Real-World[[:space:]]Dataset[[:space:]]and[[:space:]]A[[:space:]]Frequency-Aware[[:space:]]Shadow[[:space:]]Erasing[[:space:]]Net/f8df0628-9a27-4d45-a2b6-2c8c71bf57b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Holistic[[:space:]]Geometric[[:space:]]Feature[[:space:]]Learning[[:space:]]for[[:space:]]Structured[[:space:]]Reconstruction/6569e137-9378-406b-9cde-ca409ae1924f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Holistic[[:space:]]Label[[:space:]]Correction[[:space:]]for[[:space:]]Noisy[[:space:]]Multi-Label[[:space:]]Classification/c24a07fb-5f1b-44b9-bb06-0d6a49b47146_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HollowNeRF_[[:space:]]Pruning[[:space:]]Hashgrid-Based[[:space:]]NeRFs[[:space:]]with[[:space:]]Trainable[[:space:]]Collision[[:space:]]Mitigation/1b4a3dc1-627b-4861-a9e0-292ebe468c19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HoloAssist_[[:space:]]an[[:space:]]Egocentric[[:space:]]Human[[:space:]]Interaction[[:space:]]Dataset[[:space:]]for[[:space:]]Interactive[[:space:]]AI[[:space:]]Assistants[[:space:]]in[[:space:]]the[[:space:]]Real[[:space:]]World/7b800728-4ea2-4a01-b47d-4a862ae2ec61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HoloFusion_[[:space:]]Towards[[:space:]]Photo-realistic[[:space:]]3D[[:space:]]Generative[[:space:]]Modeling/ded240ab-1a84-4b65-9fba-38d695e9e368_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Homeomorphism[[:space:]]Alignment[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/3818e9e0-d874-4841-b0b1-98c499dc8740_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Homography[[:space:]]Guided[[:space:]]Temporal[[:space:]]Fusion[[:space:]]for[[:space:]]Road[[:space:]]Line[[:space:]]and[[:space:]]Marking[[:space:]]Segmentation/44fc0793-39bb-4f98-abb1-f6daedfcefc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HopFIR_[[:space:]]Hop-wise[[:space:]]GraphFormer[[:space:]]with[[:space:]]Intragroup[[:space:]]Joint[[:space:]]Refinement[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/f663f584-1a4b-4181-b11b-1abe3d533ef4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Householder[[:space:]]Projector[[:space:]]for[[:space:]]Unsupervised[[:space:]]Latent[[:space:]]Semantics[[:space:]]Discovery/58ee27b2-47d7-4507-9418-162f31fa3470_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/How[[:space:]]Far[[:space:]]Pre-trained[[:space:]]Models[[:space:]]Are[[:space:]]from[[:space:]]Neural[[:space:]]Collapse[[:space:]]on[[:space:]]the[[:space:]]Target[[:space:]]Dataset[[:space:]]Informs[[:space:]]their[[:space:]]Transferability/336dd7d0-3ba4-4bc4-a56f-4e1851a25562_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/How[[:space:]]Much[[:space:]]Temporal[[:space:]]Long-Term[[:space:]]Context[[:space:]]is[[:space:]]Needed[[:space:]]for[[:space:]]Action[[:space:]]Segmentation_/94b447e5-f23e-4bb3-8d5c-681df55d2119_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/How[[:space:]]to[[:space:]]Boost[[:space:]]Face[[:space:]]Recognition[[:space:]]with[[:space:]]StyleGAN_/aa9ef81d-3574-478e-bb1c-7bc6363afaee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/How[[:space:]]to[[:space:]]Choose[[:space:]]your[[:space:]]Best[[:space:]]Allies[[:space:]]for[[:space:]]a[[:space:]]Transferable[[:space:]]Attack_/b91b40ef-6207-4d32-b5eb-5667ebb674ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Human[[:space:]]Part-wise[[:space:]]3D[[:space:]]Motion[[:space:]]Context[[:space:]]Learning[[:space:]]for[[:space:]]Sign[[:space:]]Language[[:space:]]Recognition/7a02bc72-2ece-4aac-a870-d0710b03a665_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Human[[:space:]]Preference[[:space:]]Score_[[:space:]]Better[[:space:]]Aligning[[:space:]]Text-to-Image[[:space:]]Models[[:space:]]with[[:space:]]Human[[:space:]]Preference/fe251d76-a611-4730-9cbd-1256284fc7ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Human[[:space:]]from[[:space:]]Blur_[[:space:]]Human[[:space:]]Pose[[:space:]]Tracking[[:space:]]from[[:space:]]Blurry[[:space:]]Images/4d60ff8e-1b94-47e5-a4eb-73a7f43d8a66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Human-Inspired[[:space:]]Facial[[:space:]]Sketch[[:space:]]Synthesis[[:space:]]with[[:space:]]Dynamic[[:space:]]Adaptation/b5a138b6-3e38-4d8f-9f0e-d8c3433e4be2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Human-centric[[:space:]]Scene[[:space:]]Understanding[[:space:]]for[[:space:]]3D[[:space:]]Large-scale[[:space:]]Scenarios/ecedad45-ccfc-4aaf-997e-73823120d034_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HumanMAC_[[:space:]]Masked[[:space:]]Motion[[:space:]]Completion[[:space:]]for[[:space:]]Human[[:space:]]Motion[[:space:]]Prediction/b293b992-1131-4888-b7ef-d32fc11b86d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HumanSD_[[:space:]]A[[:space:]]Native[[:space:]]Skeleton-Guided[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Human[[:space:]]Image[[:space:]]Generation/f976aa87-b6b5-4737-a3bf-c43f9251c15a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Humans[[:space:]]in[[:space:]]4D_[[:space:]]Reconstructing[[:space:]]and[[:space:]]Tracking[[:space:]]Humans[[:space:]]with[[:space:]]Transformers/7d94edb6-ca16-4a92-8e59-930c837b7cb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hybrid[[:space:]]Spectral[[:space:]]Denoising[[:space:]]Transformer[[:space:]]with[[:space:]]Guided[[:space:]]Attention/6adca05c-6238-47f4-9430-1a9c8d7f09ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HybridAugment++_[[:space:]]Unified[[:space:]]Frequency[[:space:]]Spectra[[:space:]]Perturbations[[:space:]]for[[:space:]]Model[[:space:]]Robustness/279315b6-b5e2-4d4b-981d-70b8aded9ab9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HyperDiffusion_[[:space:]]Generating[[:space:]]Implicit[[:space:]]Neural[[:space:]]Fields[[:space:]]with[[:space:]]Weight-Space[[:space:]]Diffusion/4b830781-c32c-4102-8b13-c84892303118_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/HyperReenact_[[:space:]]One-Shot[[:space:]]Reenactment[[:space:]]via[[:space:]]Jointly[[:space:]]Learning[[:space:]]to[[:space:]]Refine[[:space:]]and[[:space:]]Retarget[[:space:]]Faces/daf3d36b-9d42-438a-bbb4-f43af9baae57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hyperbolic[[:space:]]Audio-visual[[:space:]]Zero-shot[[:space:]]Learning/8817c007-d1dc-4c18-949c-bcbc9f88e37f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Hyperbolic[[:space:]]Chamfer[[:space:]]Distance[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion/af5f5b00-28d2-45cc-b62d-de69a73deaa4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/I[[:space:]]Can't[[:space:]]Believe[[:space:]]There's[[:space:]]No[[:space:]]Images![[:space:]]Learning[[:space:]]Visual[[:space:]]Tasks[[:space:]]Using[[:space:]]only[[:space:]]Language[[:space:]]Supervision/bc661605-97b3-4683-ba9c-a03d3d296ae5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/I-ViT_[[:space:]]Integer-only[[:space:]]Quantization[[:space:]]for[[:space:]]Efficient[[:space:]]Vision[[:space:]]Transformer[[:space:]]Inference/38eccc55-2982-4090-93b4-573f91a69836_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ICD-Face_[[:space:]]Intra-class[[:space:]]Compactness[[:space:]]Distillation[[:space:]]for[[:space:]]Face[[:space:]]Recognition/47a87cde-fe03-4b9c-b173-a39827ec9c87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ICE-NeRF_[[:space:]]Interactive[[:space:]]Color[[:space:]]Editing[[:space:]]of[[:space:]]NeRFs[[:space:]]via[[:space:]]Decomposition-Aware[[:space:]]Weight[[:space:]]Optimization/dbfb0232-39c8-401e-9705-517693e62b86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ICICLE_[[:space:]]Interpretable[[:space:]]Class[[:space:]]Incremental[[:space:]]Continual[[:space:]]Learning/8258f757-0562-423a-b152-f2c77159c6bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ICL-D3IE_[[:space:]]In-Context[[:space:]]Learning[[:space:]]with[[:space:]]Diverse[[:space:]]Demonstrations[[:space:]]Updating[[:space:]]for[[:space:]]Document[[:space:]]Information[[:space:]]Extraction/887c4dea-477d-4878-8791-5afad587305a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IDiff-Face_[[:space:]]Synthetic-based[[:space:]]Face[[:space:]]Recognition[[:space:]]through[[:space:]]Fizzy[[:space:]]Identity-Conditioned[[:space:]]Diffusion[[:space:]]Model/3e3e6596-116f-4618-8227-0ba783afce74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IHNet_[[:space:]]Iterative[[:space:]]Hierarchical[[:space:]]Network[[:space:]]Guided[[:space:]]by[[:space:]]High-Resolution[[:space:]]Estimated[[:space:]]Information[[:space:]]for[[:space:]]Scene[[:space:]]Flow[[:space:]]Estimation/ed2f5375-7229-4c40-ac54-c8533b741478_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IIEU_[[:space:]]Rethinking[[:space:]]Neural[[:space:]]Feature[[:space:]]Activation[[:space:]]from[[:space:]]Decision-Making/81971ace-ec45-436c-99e3-07f2786acdc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/INSTA-BNN_[[:space:]]Binary[[:space:]]Neural[[:space:]]Network[[:space:]]with[[:space:]]INSTAnce-aware[[:space:]]Threshold/f27d0e7a-9bfa-43cb-9f9b-993f5449f612_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/INT2_[[:space:]]Interactive[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]at[[:space:]]Intersections/8d2fd679-b885-4f4a-894c-7ac23d8fccfd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IOMatch_[[:space:]]Simplifying[[:space:]]Open-Set[[:space:]]Semi-Supervised[[:space:]]Learning[[:space:]]with[[:space:]]Joint[[:space:]]Inliers[[:space:]]and[[:space:]]Outliers[[:space:]]Utilization/622d052f-a2b8-4847-8454-6fc440f78280_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IST-Net_[[:space:]]Prior-Free[[:space:]]Category-Level[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]Implicit[[:space:]]Space[[:space:]]Transformation/0a6f237c-cfbd-45bd-bf15-b46f3537fde9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ITI-GEN_[[:space:]]Inclusive[[:space:]]Text-to-Image[[:space:]]Generation/467c7a9f-2662-47ce-bb67-cc2ae106548c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Identification[[:space:]]of[[:space:]]Systematic[[:space:]]Errors[[:space:]]of[[:space:]]Image[[:space:]]Classifiers[[:space:]]on[[:space:]]Rare[[:space:]]Subgroups/f5d8cffa-d787-441e-a432-072ebf914d7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Identity-Consistent[[:space:]]Aggregation[[:space:]]for[[:space:]]Video[[:space:]]Object[[:space:]]Detection/c05dbce7-e2c0-494d-b56c-6f670ec5ba85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Identity-Seeking[[:space:]]Self-Supervised[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Generalizable[[:space:]]Person[[:space:]]Re-Identification/36e60cf9-b0f6-4fcf-bdfe-3808ac35d312_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ImGeoNet_[[:space:]]Image-induced[[:space:]]Geometry-aware[[:space:]]Voxel[[:space:]]Representation[[:space:]]for[[:space:]]Multi-view[[:space:]]3D[[:space:]]Object[[:space:]]Detection/73ffc5f0-f100-4d81-b894-df9421797c13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Image-Free[[:space:]]Classifier[[:space:]]Injection[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Classification/ce407d99-1316-4952-ae2b-ab70c27c2176_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ImbSAM_[[:space:]]A[[:space:]]Closer[[:space:]]Look[[:space:]]at[[:space:]]Sharpness-Aware[[:space:]]Minimization[[:space:]]in[[:space:]]Class-Imbalanced[[:space:]]Recognition/3a1035e5-0c00-4bba-a556-fd6bacaccdb1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Imitator_[[:space:]]Personalized[[:space:]]Speech-driven[[:space:]]3D[[:space:]]Facial[[:space:]]Animation/f507ce68-acef-4a1f-be12-a4cc652f6471_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Implicit[[:space:]]Autoencoder[[:space:]]for[[:space:]]Point-Cloud[[:space:]]Self-Supervised[[:space:]]Representation[[:space:]]Learning/048dfad9-3954-498a-832e-89b1101dbed3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Implicit[[:space:]]Identity[[:space:]]Representation[[:space:]]Conditioned[[:space:]]Memory[[:space:]]Compensation[[:space:]]Network[[:space:]]for[[:space:]]Talking[[:space:]]Head[[:space:]]video[[:space:]]Generation/ca183109-6a47-49da-9201-002c79190ca5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Implicit[[:space:]]Neural[[:space:]]Representation[[:space:]]for[[:space:]]Cooperative[[:space:]]Low-light[[:space:]]Image[[:space:]]Enhancement/78b3f122-9032-4735-be20-52bac8bd5951_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Implicit[[:space:]]Temporal[[:space:]]Modeling[[:space:]]with[[:space:]]Learnable[[:space:]]Alignment[[:space:]]for[[:space:]]Video[[:space:]]Recognition/5bfa395f-480a-4134-9ded-c02d2b2a6c2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improved[[:space:]]Knowledge[[:space:]]Transfer[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]via[[:space:]]Trico[[:space:]]Training[[:space:]]Strategy/05be1a09-171d-4bef-b14f-c8ccbb0e48fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improved[[:space:]]Visual[[:space:]]Fine-tuning[[:space:]]with[[:space:]]Natural[[:space:]]Language[[:space:]]Supervision/6f043a80-e686-4404-9dcc-ab3e45db7394_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]3D[[:space:]]Imaging[[:space:]]with[[:space:]]Pre-Trained[[:space:]]Perpendicular[[:space:]]2D[[:space:]]Diffusion[[:space:]]Models/cd773ce1-78db-4f4d-a9ed-26142df5c837_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]of[[:space:]]Masked[[:space:]]Autoencoders[[:space:]]via[[:space:]]Test-time[[:space:]]Frequency-domain[[:space:]]Prompting/422cbabb-2e30-40c8-b907-14923feb5118_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]CLIP[[:space:]]Fine-tuning[[:space:]]Performance/b35f4dc0-6d3d-42b9-85bd-b60d7b0e949d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Continuous[[:space:]]Sign[[:space:]]Language[[:space:]]Recognition[[:space:]]with[[:space:]]Cross-Lingual[[:space:]]Signs/eb84f5ac-53d0-48e8-a763-5881e83dbf9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Diversity[[:space:]]in[[:space:]]Zero-Shot[[:space:]]GAN[[:space:]]Adaptation[[:space:]]with[[:space:]]Semantic[[:space:]]Variations/3f78f453-6e31-4a0a-b954-3f7577c76a72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Equivariance[[:space:]]in[[:space:]]State-of-the-Art[[:space:]]Supervised[[:space:]]Depth[[:space:]]and[[:space:]]Normal[[:space:]]Predictors/ab7b84dc-a82b-4dec-98b6-1dcd2b9444b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Generalization[[:space:]]in[[:space:]]Visual[[:space:]]Reinforcement[[:space:]]Learning[[:space:]]via[[:space:]]Conflict-aware[[:space:]]Gradient[[:space:]]Agreement[[:space:]]Augmentation/620b0ab7-3f50-4e9d-843c-1afe0f0c0d3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Generalization[[:space:]]of[[:space:]]Adversarial[[:space:]]Training[[:space:]]via[[:space:]]Robust[[:space:]]Critical[[:space:]]Fine-Tuning/c2359254-b39c-4110-ab86-adc2b8a6ad72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Lens[[:space:]]Flare[[:space:]]Removal[[:space:]]with[[:space:]]General-Purpose[[:space:]]Pipeline[[:space:]]and[[:space:]]Multiple[[:space:]]Light[[:space:]]Sources[[:space:]]Recovery/6e634441-3e78-4ae8-bc48-a37c3f1c8255_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Online[[:space:]]Lane[[:space:]]Graph[[:space:]]Extraction[[:space:]]by[[:space:]]Object-Lane[[:space:]]Clustering/7757d7f1-1c55-4e17-9abc-b098e1cfa64e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Pixel-based[[:space:]]MIM[[:space:]]by[[:space:]]Reducing[[:space:]]Wasted[[:space:]]Modeling[[:space:]]Capability/b457b56c-4727-4c14-9f6e-3d402575351b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Histopathologic[[:space:]]Images[[:space:]]with[[:space:]]Cluster[[:space:]]Constraints/ca8aab5c-9a54-4990-94c7-bb739f49f313_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Sample[[:space:]]Quality[[:space:]]of[[:space:]]Diffusion[[:space:]]Models[[:space:]]Using[[:space:]]Self-Attention[[:space:]]Guidance/f5c3b835-52ed-4882-a082-7569c21dcc5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Transformer-based[[:space:]]Image[[:space:]]Matching[[:space:]]by[[:space:]]Cascaded[[:space:]]Capturing[[:space:]]Spatially[[:space:]]Informative[[:space:]]Keypoints/ac869953-91a9-414e-af7d-387a46bb893a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Improving[[:space:]]Unsupervised[[:space:]]Visual[[:space:]]Program[[:space:]]Inference[[:space:]]with[[:space:]]Code[[:space:]]Rewriting[[:space:]]Families/ec912f89-6679-4196-b0c9-b5fc47ed3f4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/In-Style_[[:space:]]Bridging[[:space:]]Text[[:space:]]and[[:space:]]Uncurated[[:space:]]Videos[[:space:]]with[[:space:]]Style[[:space:]]Transfer[[:space:]]for[[:space:]]Text-Video[[:space:]]Retrieval/4779d7ee-cd2d-464f-881e-855883e95a65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Incremental[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/20c788af-3b9a-47f3-b0bd-613f4fd6f548_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Indoor[[:space:]]Depth[[:space:]]Recovery[[:space:]]Based[[:space:]]on[[:space:]]Deep[[:space:]]Unfolding[[:space:]]with[[:space:]]Non-Local[[:space:]]Prior/e5f5181b-406a-4253-ac28-b6d4219db732_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Inducing[[:space:]]Neural[[:space:]]Collapse[[:space:]]to[[:space:]]a[[:space:]]Fixed[[:space:]]Hierarchy-Aware[[:space:]]Frame[[:space:]]for[[:space:]]Reducing[[:space:]]Mistake[[:space:]]Severity/2e5d2b23-26dd-492c-bcdd-063d4b364cf6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/InfiniCity_[[:space:]]Infinite-Scale[[:space:]]City[[:space:]]Synthesis/6c3d591c-b1d8-4e81-8380-364812f493e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Informative[[:space:]]Data[[:space:]]Mining[[:space:]]for[[:space:]]One-Shot[[:space:]]Cross-Domain[[:space:]]Semantic[[:space:]]Segmentation/c3994c7c-dcd7-4d58-8fda-ae4443f93a6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Inherent[[:space:]]Redundancy[[:space:]]in[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks/c022e6b8-2a72-423f-82e2-127f9d38f67c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Innovating[[:space:]]Real[[:space:]]Fisheye[[:space:]]Image[[:space:]]Correction[[:space:]]with[[:space:]]Dual[[:space:]]Diffusion[[:space:]]Architecture/ec9c66a8-af0f-4cae-8ad5-3256273c5667_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Inspecting[[:space:]]the[[:space:]]Geographical[[:space:]]Representativeness[[:space:]]of[[:space:]]Images[[:space:]]from[[:space:]]Text-to-Image[[:space:]]Models/d9b27775-7113-477b-a914-ac530f639d68_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Instance[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field/98dbd449-8be5-4fff-b91e-ba4015672697_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Instance[[:space:]]and[[:space:]]Category[[:space:]]Supervision[[:space:]]are[[:space:]]Alternate[[:space:]]Learners[[:space:]]for[[:space:]]Continual[[:space:]]Learning/aafc881b-f2a1-4665-8eab-57c203bcadb2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Instance-aware[[:space:]]Dynamic[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Pre-trained[[:space:]]Point[[:space:]]Cloud[[:space:]]Models/afada580-f747-4ca2-9388-fa46a9a63bc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Instruct-NeRF2NeRF_[[:space:]]Editing[[:space:]]3D[[:space:]]Scenes[[:space:]]with[[:space:]]Instructions/33b7c9f7-bfe8-4305-b95e-982c533398a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Integrally[[:space:]]Migrating[[:space:]]Pre-trained[[:space:]]Transformer[[:space:]]Encoder-decoders[[:space:]]for[[:space:]]Visual[[:space:]]Object[[:space:]]Detection/891e76e1-cc41-450b-ab6f-c2b23f339599_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Integrating[[:space:]]Boxes[[:space:]]and[[:space:]]Masks_[[:space:]]A[[:space:]]Multi-Object[[:space:]]Framework[[:space:]]for[[:space:]]Unified[[:space:]]Visual[[:space:]]Tracking[[:space:]]and[[:space:]]Segmentation/1e21c85c-9373-43eb-841e-b122066e2125_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IntentQA_[[:space:]]Context-aware[[:space:]]Video[[:space:]]Intent[[:space:]]Reasoning/48e33eaa-87ef-46eb-866b-c7d0adb96b00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Inter-Realization[[:space:]]Channels_[[:space:]]Unsupervised[[:space:]]Anomaly[[:space:]]Detection[[:space:]]Beyond[[:space:]]One-Class[[:space:]]Classification/3a7d30b6-f128-4876-944b-c29e5df40040_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/InterDiff_[[:space:]]Generating[[:space:]]3D[[:space:]]Human-Object[[:space:]]Interactions[[:space:]]with[[:space:]]Physics-Informed[[:space:]]Diffusion/0df594ff-deaa-47b7-a9f9-c41aca07bfb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/InterFormer_[[:space:]]Real-time[[:space:]]Interactive[[:space:]]Image[[:space:]]Segmentation/20526d22-664e-4c59-9bb9-0e4046a73743_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Interaction-aware[[:space:]]Joint[[:space:]]Attention[[:space:]]Estimation[[:space:]]Using[[:space:]]People[[:space:]]Attributes/f04f1c8f-9c5d-40c6-98b3-ab3753de5077_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Interactive[[:space:]]Class-Agnostic[[:space:]]Object[[:space:]]Counting/917f66ef-a9c9-43a6-8b0f-cfa7ba1e7836_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/IntrinsicNeRF_[[:space:]]Learning[[:space:]]Intrinsic[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Editable[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/7268e5fe-c09f-4808-9549-aebf05859e17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Introducing[[:space:]]Language[[:space:]]Guidance[[:space:]]in[[:space:]]Prompt-based[[:space:]]Continual[[:space:]]Learning/2dc342e1-856c-4bcd-9615-20c11309439b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Invariant[[:space:]]Feature[[:space:]]Regularization[[:space:]]for[[:space:]]Fair[[:space:]]Face[[:space:]]Recognition/53296cd3-e51f-4c3a-affa-3ab8aa516bff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Invariant[[:space:]]Training[[:space:]]2D-3D[[:space:]]Joint[[:space:]]Hard[[:space:]]Samples[[:space:]]for[[:space:]]Few-Shot[[:space:]]Point[[:space:]]Cloud[[:space:]]Recognition/ccfbf079-9b74-4d80-9d0a-6f72a73f96d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Inverse[[:space:]]Compositional[[:space:]]Learning[[:space:]]for[[:space:]]Weakly-supervised[[:space:]]Relation[[:space:]]Grounding/7057271c-40b7-41d4-94ab-8f0afb5c3c0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Inverse[[:space:]]Problem[[:space:]]Regularization[[:space:]]with[[:space:]]Hierarchical[[:space:]]Variational[[:space:]]Autoencoders/ae05d22f-ddb5-4e7f-b033-093f58ef4087_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Is[[:space:]]Imitation[[:space:]]All[[:space:]]You[[:space:]]Need_[[:space:]]Generalized[[:space:]]Decision-Making[[:space:]]with[[:space:]]Dual-Phase[[:space:]]Training/2c796c81-f6d1-49ed-82e6-15393350ca9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Isomer_[[:space:]]Isomerous[[:space:]]Transformer[[:space:]]for[[:space:]]Zero-shot[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/be25be27-c581-48db-b93e-6518c57b72c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Iterative[[:space:]]Denoiser[[:space:]]and[[:space:]]Noise[[:space:]]Estimator[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Image[[:space:]]Denoising/10c9c3d9-e4cf-48e7-b01b-a5b1f402953a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Iterative[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Unsupervised[[:space:]]Backlit[[:space:]]Image[[:space:]]Enhancement/bdee9f40-0e51-4547-b1d6-138e6479f58d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Iterative[[:space:]]Soft[[:space:]]Shrinkage[[:space:]]Learning[[:space:]]for[[:space:]]Efficient[[:space:]]Image[[:space:]]Super-Resolution/babbb850-791e-4e54-8405-daa7b3af6e77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Iterative[[:space:]]Superquadric[[:space:]]Recomposition[[:space:]]of[[:space:]]3D[[:space:]]Objects[[:space:]]from[[:space:]]Multiple[[:space:]]Views/b5858f96-43ce-42f5-b073-7a408ba06abe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/JOTR_[[:space:]]3D[[:space:]]Joint[[:space:]]Contrastive[[:space:]]Learning[[:space:]]with[[:space:]]Transformers[[:space:]]for[[:space:]]Occluded[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/ceb9dd21-d0bf-4bd4-8923-e790e446aaf9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Joint[[:space:]]Demosaicing[[:space:]]and[[:space:]]Deghosting[[:space:]]of[[:space:]]Time-Varying[[:space:]]Exposures[[:space:]]for[[:space:]]Single-Shot[[:space:]]HDR[[:space:]]Imaging/780a07f6-c389-468f-948c-a47a744759a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Joint[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representation[[:space:]]for[[:space:]]High-fidelity[[:space:]]and[[:space:]]Compact[[:space:]]Vector[[:space:]]Fonts/9901c035-30b2-4140-af95-83253996a102_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Joint[[:space:]]Metrics[[:space:]]Matter_[[:space:]]A[[:space:]]Better[[:space:]]Standard[[:space:]]for[[:space:]]Trajectory[[:space:]]Forecasting/5224c072-abb5-4843-ab6f-b82ff4cf7c86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Joint-Relation[[:space:]]Transformer[[:space:]]for[[:space:]]Multi-Person[[:space:]]Motion[[:space:]]Prediction/060fc917-221c-4d9e-9ebd-453a15e5e51e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Jumping[[:space:]]through[[:space:]]Local[[:space:]]Minima_[[:space:]]Quantization[[:space:]]in[[:space:]]the[[:space:]]Loss[[:space:]]Landscape[[:space:]]of[[:space:]]Vision[[:space:]]Transformers/933b1fda-00af-4265-af5e-fcf9b90874ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/KECOR_[[:space:]]Kernel[[:space:]]Coding[[:space:]]Rate[[:space:]]Maximization[[:space:]]for[[:space:]]Active[[:space:]]3D[[:space:]]Object[[:space:]]Detection/806a3db2-d055-42bb-93a6-87978c2742ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Keep[[:space:]]It[[:space:]]SimPool_[[:space:]]Who[[:space:]]Said[[:space:]]Supervised[[:space:]]Transformers[[:space:]]Suffer[[:space:]]from[[:space:]]Attention[[:space:]]Deficit_/b6e1b4be-489e-444a-bfa0-3c68a252f3b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Kick[[:space:]]Back[[:space:]]&[[:space:]]Relax_[[:space:]]Learning[[:space:]]to[[:space:]]Reconstruct[[:space:]]the[[:space:]]World[[:space:]]by[[:space:]]Watching[[:space:]]SlowTV/1b31c4c6-3e4e-4143-ba2d-8891d10578c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Knowing[[:space:]]Where[[:space:]]to[[:space:]]Focus_[[:space:]]Event-aware[[:space:]]Transformer[[:space:]]for[[:space:]]Video[[:space:]]Grounding/516c97ae-15dc-46a8-90e8-c3e5b54d640c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Knowledge[[:space:]]Proxy[[:space:]]Intervention[[:space:]]for[[:space:]]Deconfounded[[:space:]]Video[[:space:]]Question[[:space:]]Answering/514a0700-d43c-4d83-9c86-3e2f76acb093_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Knowledge[[:space:]]Restore[[:space:]]and[[:space:]]Transfer[[:space:]]for[[:space:]]Multi-Label[[:space:]]Class-Incremental[[:space:]]Learning/6a89b426-4e44-4602-a6b8-b89d8022dd11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Knowledge-Aware[[:space:]]Federated[[:space:]]Active[[:space:]]Learning[[:space:]]with[[:space:]]Non-IID[[:space:]]Data/d33b9197-1e67-478c-85a0-027f1302e64f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Knowledge-Aware[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Generalizable[[:space:]]Vision-Language[[:space:]]Models/f80bdc7a-951f-458a-9041-c7ec710b627a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Knowledge-Spreader_[[:space:]]Learning[[:space:]]Semi-Supervised[[:space:]]Facial[[:space:]]Action[[:space:]]Dynamics[[:space:]]by[[:space:]]Consistifying[[:space:]]Knowledge[[:space:]]Granularity/034be0e4-fea1-4817-92cb-e4c77adf8b3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/L-DAWA_[[:space:]]Layer-wise[[:space:]]Divergence[[:space:]]Aware[[:space:]]Weight[[:space:]]Aggregation[[:space:]]in[[:space:]]Federated[[:space:]]Self-Supervised[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/9bdd1e99-85b5-41f4-8b6a-2802936bb131_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LA-Net_[[:space:]]Landmark-Aware[[:space:]]Learning[[:space:]]for[[:space:]]Reliable[[:space:]]Facial[[:space:]]Expression[[:space:]]Recognition[[:space:]]under[[:space:]]Label[[:space:]]Noise/3940474e-ef29-4460-bad1-b8ccaa50eae1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LAC[[:space:]]-[[:space:]]Latent[[:space:]]Action[[:space:]]Composition[[:space:]]for[[:space:]]Skeleton-based[[:space:]]Action[[:space:]]Segmentation/77f8111c-90e2-414f-869a-09da3b413c85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LAN-HDR_[[:space:]]Luminance-based[[:space:]]Alignment[[:space:]]Network[[:space:]]for[[:space:]]High[[:space:]]Dynamic[[:space:]]Range[[:space:]]Video[[:space:]]Reconstruction/4f661e1e-2b27-411a-962c-1b50da08b771_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LATR_[[:space:]]3D[[:space:]]Lane[[:space:]]Detection[[:space:]]from[[:space:]]Monocular[[:space:]]Images[[:space:]]with[[:space:]]Transformer/4eb6ee99-0a9b-486c-9c39-9964eb6313a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LAW-Diffusion_[[:space:]]Complex[[:space:]]Scene[[:space:]]Generation[[:space:]]by[[:space:]]Diffusion[[:space:]]with[[:space:]]Layouts/b4ac7e1d-15e1-4454-828d-3c1e6fd8fbd1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LD-ZNet_[[:space:]]A[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Approach[[:space:]]for[[:space:]]Text-Based[[:space:]]Image[[:space:]]Segmentation/0da00438-ff92-452d-99b0-81f30b81a30d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LDL_[[:space:]]Line[[:space:]]Distance[[:space:]]Functions[[:space:]]for[[:space:]]Panoramic[[:space:]]Localization/5b58b440-2b20-4389-a822-100317fa759f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LDP-Feat_[[:space:]]Image[[:space:]]Features[[:space:]]with[[:space:]]Local[[:space:]]Differential[[:space:]]Privacy/fa1ade99-98b7-4fc2-8070-579ba6f70b66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LEA2_[[:space:]]A[[:space:]]Lightweight[[:space:]]Ensemble[[:space:]]Adversarial[[:space:]]Attack[[:space:]]via[[:space:]]Non-overlapping[[:space:]]Vulnerable[[:space:]]Frequency[[:space:]]Regions/efd711f0-492e-4b76-ae59-c86ad2990d0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LERF_[[:space:]]Language[[:space:]]Embedded[[:space:]]Radiance[[:space:]]Fields/2617b38a-e247-4f02-88d2-1ce6ea0ef51c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LFS-GAN_[[:space:]]Lifelong[[:space:]]Few-Shot[[:space:]]Image[[:space:]]Generation/9715fb6d-dd86-45ab-87dc-e6db60ceaf24_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LIMITR_[[:space:]]Leveraging[[:space:]]Local[[:space:]]Information[[:space:]]for[[:space:]]Medical[[:space:]]Image-Text[[:space:]]Representation/c87c35b3-2e01-4e1b-8075-c5801c756064_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LISTER_[[:space:]]Neighbor[[:space:]]Decoding[[:space:]]for[[:space:]]Length-Insensitive[[:space:]]Scene[[:space:]]Text[[:space:]]Recognition/690259af-20d6-4662-96c7-081383ca1f72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LIST_[[:space:]]Learning[[:space:]]Implicitly[[:space:]]from[[:space:]]Spatial[[:space:]]Transformers[[:space:]]for[[:space:]]Single-View[[:space:]]3D[[:space:]]Reconstruction/2692589f-23ff-4b1f-b558-c4d997a79b19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LLM-Planner_[[:space:]]Few-Shot[[:space:]]Grounded[[:space:]]Planning[[:space:]]for[[:space:]]Embodied[[:space:]]Agents[[:space:]]with[[:space:]]Large[[:space:]]Language[[:space:]]Models/0793a62b-59a4-4b35-99f2-4c1d31a37c98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LMR_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Multi-Reference[[:space:]]Dataset[[:space:]]for[[:space:]]Reference-Based[[:space:]]Super-Resolution/4a5929cf-b059-4937-b700-f7e64ec67b29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LNPL-MIL_[[:space:]]Learning[[:space:]]from[[:space:]]Noisy[[:space:]]Pseudo[[:space:]]Labels[[:space:]]for[[:space:]]Promoting[[:space:]]Multiple[[:space:]]Instance[[:space:]]Learning[[:space:]]in[[:space:]]Whole[[:space:]]Slide[[:space:]]Image/117510f4-9d82-4166-9d6c-ddfdbc07eae1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LPFF_[[:space:]]A[[:space:]]Portrait[[:space:]]Dataset[[:space:]]for[[:space:]]Face[[:space:]]Generators[[:space:]]Across[[:space:]]Large[[:space:]]Poses/aa53b017-4614-413e-b0a8-3730db19687f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LRRU_[[:space:]]Long-short[[:space:]]Range[[:space:]]Recurrent[[:space:]]Updating[[:space:]]Networks[[:space:]]for[[:space:]]Depth[[:space:]]Completion/a5f82a91-11e6-456e-be48-177065abec92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LU-NeRF_[[:space:]]Scene[[:space:]]and[[:space:]]Pose[[:space:]]Estimation[[:space:]]by[[:space:]]Synchronizing[[:space:]]Local[[:space:]]Unposed[[:space:]]NeRFs/63d28ee9-f94d-44c9-a978-3bb759bfe469_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LVOS_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Long-term[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/71aa5bb7-373a-4425-9723-87dcfbcc268f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LaPE_[[:space:]]Layer-adaptive[[:space:]]Position[[:space:]]Embedding[[:space:]]for[[:space:]]Vision[[:space:]]Transformers[[:space:]]with[[:space:]]Independent[[:space:]]Layer[[:space:]]Normalization/881d69d4-8f69-41f3-95ce-4351c44e1c35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LaRS_[[:space:]]A[[:space:]]Diverse[[:space:]]Panoptic[[:space:]]Maritime[[:space:]]Obstacle[[:space:]]Detection[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmark/7b55dd4e-abd8-4f11-aca5-6c10f323df34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Label[[:space:]]Shift[[:space:]]Adapter[[:space:]]for[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]under[[:space:]]Covariate[[:space:]]and[[:space:]]Label[[:space:]]Shifts/8826026c-6da5-4749-8a23-22b76030d60b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Label-Efficient[[:space:]]Online[[:space:]]Continual[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]Streaming[[:space:]]Video/bb408240-93f4-4420-9e47-a5facd88851a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Label-Free[[:space:]]Event-based[[:space:]]Object[[:space:]]Recognition[[:space:]]via[[:space:]]Joint[[:space:]]Learning[[:space:]]with[[:space:]]Image[[:space:]]Reconstruction[[:space:]]from[[:space:]]Events/cebb8d22-3e6f-4881-895d-c913e7109230_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Label-Guided[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Continual[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]on[[:space:]]2D[[:space:]]Images[[:space:]]and[[:space:]]3D[[:space:]]Point[[:space:]]Clouds/3a801de3-50c8-469a-b26b-c00f633cc663_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Label-Noise[[:space:]]Learning[[:space:]]with[[:space:]]Intrinsically[[:space:]]Long-Tailed[[:space:]]Data/ca2911cd-1897-45f1-bc9c-ab89b815151c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Landscape[[:space:]]Learning[[:space:]]for[[:space:]]Neural[[:space:]]Network[[:space:]]Inversion/f1107abc-1d35-4283-b4db-d9fd38f81a61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Large[[:space:]]Selective[[:space:]]Kernel[[:space:]]Network[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Object[[:space:]]Detection/1c2f7154-2a88-4538-ac11-efbaf0ea97b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Large-Scale[[:space:]]Land[[:space:]]Cover[[:space:]]Mapping[[:space:]]with[[:space:]]Fine-Grained[[:space:]]Classes[[:space:]]via[[:space:]]Class-Aware[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation/85af6521-8e24-45e3-ac82-fb3352f9480e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Large-Scale[[:space:]]Person[[:space:]]Detection[[:space:]]and[[:space:]]Localization[[:space:]]Using[[:space:]]Overhead[[:space:]]Fisheye[[:space:]]Cameras/b691cb26-f3c1-4d93-ba53-9242b6258c4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Late[[:space:]]Stopping_[[:space:]]Avoiding[[:space:]]Confidently[[:space:]]Learning[[:space:]]from[[:space:]]Mislabeled[[:space:]]Examples/c0b4cc8e-d6dd-4adb-8592-11f986f8c0fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Latent-OFER_[[:space:]]Detect,[[:space:]]Mask,[[:space:]]and[[:space:]]Reconstruct[[:space:]]with[[:space:]]Latent[[:space:]]Vectors[[:space:]]for[[:space:]]Occluded[[:space:]]Facial[[:space:]]Expression[[:space:]]Recognition/b49e72cb-c8b1-4c66-830a-4cce4a977773_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LayoutDiffusion_[[:space:]]Improving[[:space:]]Graphic[[:space:]]Layout[[:space:]]Generation[[:space:]]by[[:space:]]Discrete[[:space:]]Diffusion[[:space:]]Probabilistic[[:space:]]Models/77145711-9475-40d3-bb1f-4deba9a37184_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LeaF_[[:space:]]Learning[[:space:]]Frames[[:space:]]for[[:space:]]4D[[:space:]]Point[[:space:]]Cloud[[:space:]]Sequence[[:space:]]Understanding/51c2f292-724b-40e6-8771-779ef4bbccc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Leaping[[:space:]]Into[[:space:]]Memories_[[:space:]]Space-Time[[:space:]]Deep[[:space:]]Feature[[:space:]]Synthesis/5ad3f2c3-b163-4339-b33a-6f2219ed23a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learn[[:space:]]TAROT[[:space:]]with[[:space:]]MENTOR_[[:space:]]A[[:space:]]Meta-Learned[[:space:]]Self-Supervised[[:space:]]Approach[[:space:]]for[[:space:]]Trajectory[[:space:]]Prediction/e84db066-834e-41d1-b5a0-8e0c6af0f9b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learned[[:space:]]Compressive[[:space:]]Representations[[:space:]]for[[:space:]]Single-Photon[[:space:]]3D[[:space:]]Imaging/78c901d2-1e87-453e-b9f2-75e42afe6005_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learned[[:space:]]Image[[:space:]]Reasoning[[:space:]]Prior[[:space:]]Penetrates[[:space:]]Deep[[:space:]]Unfolding[[:space:]]Network[[:space:]]for[[:space:]]Panchromatic[[:space:]]and[[:space:]]Multi-spectral[[:space:]]Image[[:space:]]Fusion/e7814c90-67a9-4d35-afdb-629b41a8c850_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Adaptive[[:space:]]Neighborhoods[[:space:]]for[[:space:]]Graph[[:space:]]Neural[[:space:]]Networks/002465ad-e9c1-4d74-bc7a-fd7e74285b28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Clothing[[:space:]]and[[:space:]]Pose[[:space:]]Invariant[[:space:]]3D[[:space:]]Shape[[:space:]]Representation[[:space:]]for[[:space:]]Long-Term[[:space:]]Person[[:space:]]Re-Identification/33efc350-232b-4a85-9272-b3026fb811bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Concise[[:space:]]and[[:space:]]Descriptive[[:space:]]Attributes[[:space:]]for[[:space:]]Visual[[:space:]]Recognition/eb8f500b-e381-4a8d-9629-e0fa391759b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Concordant[[:space:]]Attention[[:space:]]via[[:space:]]Target-aware[[:space:]]Alignment[[:space:]]for[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-identification/f96bbbd7-1bf4-40cd-9545-e535cf6aa841_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Continuous[[:space:]]Exposure[[:space:]]Value[[:space:]]Representations[[:space:]]for[[:space:]]Single-Image[[:space:]]HDR[[:space:]]Reconstruction/4c1481b1-79c6-4632-b011-bcc21d585466_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Correction[[:space:]]Filter[[:space:]]via[[:space:]]Degradation-Adaptive[[:space:]]Regression[[:space:]]for[[:space:]]Blind[[:space:]]Single[[:space:]]Image[[:space:]]Super-Resolution/eaad2ba2-88cf-4d2a-a334-fd43747c9ca0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Cross-Modal[[:space:]]Affinity[[:space:]]for[[:space:]]Referring[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation[[:space:]]Targeting[[:space:]]Limited[[:space:]]Samples/a0fe2dc9-eb10-4dc5-8bfa-8f5f03212fc3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Cross-Representation[[:space:]]Affinity[[:space:]]Consistency[[:space:]]for[[:space:]]Sparsely[[:space:]]Supervised[[:space:]]Biomedical[[:space:]]Instance[[:space:]]Segmentation/e6b0e920-39b7-4259-9938-916b268677a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Data-Driven[[:space:]]Vector-Quantized[[:space:]]Degradation[[:space:]]Model[[:space:]]for[[:space:]]Animation[[:space:]]Video[[:space:]]Super-Resolution/8790a28a-4db4-4d06-856c-466efed197ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Depth[[:space:]]Estimation[[:space:]]for[[:space:]]Transparent[[:space:]]and[[:space:]]Mirror[[:space:]]Surfaces/4e49b9b1-7456-432b-aaae-0da1220f3a37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Fine-Grained[[:space:]]Features[[:space:]]for[[:space:]]Pixel-Wise[[:space:]]Video[[:space:]]Correspondences/fd6b6adc-a24b-4ea2-b63f-ac1670224d69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Foresightful[[:space:]]Dense[[:space:]]Visual[[:space:]]Affordance[[:space:]]for[[:space:]]Deformable[[:space:]]Object[[:space:]]Manipulation/b7d619c9-c5f2-438e-896d-a2c31b01ebe0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Gabor[[:space:]]Texture[[:space:]]Features[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Recognition/6ca3f851-37f0-49f5-a456-09ea4b08c273_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Global-aware[[:space:]]Kernel[[:space:]]for[[:space:]]Image[[:space:]]Harmonization/edcde37a-d026-4b0d-9a7a-a8c99e117c7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Hierarchical[[:space:]]Features[[:space:]]with[[:space:]]Joint[[:space:]]Latent[[:space:]]Space[[:space:]]Energy-Based[[:space:]]Prior/240d7938-e848-4274-a9c2-b84ff0fe9ab2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Human[[:space:]]Dynamics[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving[[:space:]]Scenarios/cdd5a67c-1587-4c07-9231-ae20be13d8d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Human-Human[[:space:]]Interactions[[:space:]]in[[:space:]]Images[[:space:]]from[[:space:]]Weak[[:space:]]Textual[[:space:]]Supervision/6803bdfe-c2d2-4fd6-a5a5-6f813d705eb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Image[[:space:]]Harmonization[[:space:]]in[[:space:]]the[[:space:]]Linear[[:space:]]Color[[:space:]]Space/be202573-b803-48c5-8b87-29cda14139c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Image-Adaptive[[:space:]]Codebooks[[:space:]]for[[:space:]]Class-Agnostic[[:space:]]Image[[:space:]]Restoration/4f85f017-28ce-47a6-a95c-6487947424f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Long-Range[[:space:]]Information[[:space:]]with[[:space:]]Dual-Scale[[:space:]]Transformers[[:space:]]for[[:space:]]Indoor[[:space:]]Scene[[:space:]]Completion/4473583d-9b9f-42e0-9960-ac81cde7532b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Navigational[[:space:]]Visual[[:space:]]Representations[[:space:]]with[[:space:]]Semantic[[:space:]]Map[[:space:]]Supervision/94d1f5dd-be00-46b4-8d2e-2c6681cb4e80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Neural[[:space:]]Eigenfunctions[[:space:]]for[[:space:]]Unsupervised[[:space:]]Semantic[[:space:]]Segmentation/4dcc0481-9970-4a10-9577-a5671c84635b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Neural[[:space:]]Implicit[[:space:]]Surfaces[[:space:]]with[[:space:]]Object-Aware[[:space:]]Radiance[[:space:]]Fields/3c0bb7d1-9040-4c42-8743-ea0aac8c43c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Non-Local[[:space:]]Spatial-Angular[[:space:]]Correlation[[:space:]]for[[:space:]]Light[[:space:]]Field[[:space:]]Image[[:space:]]Super-Resolution/0facca26-8b6b-4a15-98c0-1ceb58a77d13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Optical[[:space:]]Flow[[:space:]]from[[:space:]]Event[[:space:]]Camera[[:space:]]with[[:space:]]Rendered[[:space:]]Dataset/9da70e78-d6a8-4bda-bae6-2f6d222d7409_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion[[:space:]]without[[:space:]]Complete[[:space:]]Point[[:space:]]Clouds_[[:space:]]A[[:space:]]Pose-Aware[[:space:]]Approach/c5a386b7-062e-40bf-ab26-f63c862742fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Pseudo-Relations[[:space:]]for[[:space:]]Cross-domain[[:space:]]Semantic[[:space:]]Segmentation/10b88ddb-f472-4de5-8cfb-db9e250c0176_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Rain[[:space:]]Location[[:space:]]Prior[[:space:]]for[[:space:]]Nighttime[[:space:]]Deraining/f13e3907-fb21-485a-916a-d1036549cfe5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Robust[[:space:]]Representations[[:space:]]with[[:space:]]Information[[:space:]]Bottleneck[[:space:]]and[[:space:]]Memory[[:space:]]Network[[:space:]]for[[:space:]]RGB-D-based[[:space:]]Gesture[[:space:]]Recognition/ffcfd550-d4a2-4cb4-a120-2a5976bb8508_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Semi-supervised[[:space:]]Gaussian[[:space:]]Mixture[[:space:]]Models[[:space:]]for[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/c328493e-9994-4352-8a93-3b569e354069_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Shape[[:space:]]Primitives[[:space:]]via[[:space:]]Implicit[[:space:]]Convexity[[:space:]]Regularization/4387a009-eca0-4928-920d-1473c95da518_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Spatial-context-aware[[:space:]]Global[[:space:]]Visual[[:space:]]Feature[[:space:]]Representation[[:space:]]for[[:space:]]Instance[[:space:]]Image[[:space:]]Retrieval/4a7da155-d324-4f0d-9452-e7f76abd7205_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Support[[:space:]]and[[:space:]]Trivial[[:space:]]Prototypes[[:space:]]for[[:space:]]Interpretable[[:space:]]Image[[:space:]]Classification/1bf5aedf-f1e8-4793-8d34-430db9b0cd03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Symmetry-Aware[[:space:]]Geometry[[:space:]]Correspondences[[:space:]]for[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/c3bb6c4d-c63d-4614-92bb-10b441ef4c78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Trajectory-Word[[:space:]]Alignments[[:space:]]for[[:space:]]Video-Language[[:space:]]Tasks/e49e204e-97bd-485d-8320-f15ff5bdec7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Unified[[:space:]]Decompositional[[:space:]]and[[:space:]]Compositional[[:space:]]NeRF[[:space:]]for[[:space:]]Editable[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/d796471e-04ce-409b-90cf-4f80b508e362_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Versatile[[:space:]]3D[[:space:]]Shape[[:space:]]Generation[[:space:]]with[[:space:]]Improved[[:space:]]Auto-regressive[[:space:]]Models/4e583ffa-62ea-4806-8a7a-46ee129f849e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]Vision-and-Language[[:space:]]Navigation[[:space:]]from[[:space:]]YouTube[[:space:]]Videos/05335e36-6140-429d-8434-dd018538220c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]a[[:space:]]More[[:space:]]Continuous[[:space:]]Zero[[:space:]]Level[[:space:]]Set[[:space:]]in[[:space:]]Unsigned[[:space:]]Distance[[:space:]]Fields[[:space:]]through[[:space:]]Level[[:space:]]Set[[:space:]]Projection/918f7816-6b1d-45a6-a6ec-23336f1dbd99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]a[[:space:]]Room[[:space:]]with[[:space:]]the[[:space:]]Occ-SDF[[:space:]]Hybrid_[[:space:]]Signed[[:space:]]Distance[[:space:]]Function[[:space:]]Mingled[[:space:]]with[[:space:]]Occupancy[[:space:]]Aids[[:space:]]Scene[[:space:]]Representation/c8aa23d3-c76a-4daf-a8f1-9ca330b3d300_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]by[[:space:]]Sorting_[[:space:]]Self-supervised[[:space:]]Learning[[:space:]]with[[:space:]]Group[[:space:]]Ordering[[:space:]]Constraints/f7288260-030f-4ffe-bd8d-4ee76970c742_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]from[[:space:]]Noisy[[:space:]]Data[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]3D[[:space:]]Object[[:space:]]Detection/a3c8a1dc-20e7-4f50-8e36-d4ddc2e6295d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]from[[:space:]]Noisy[[:space:]]Pseudo[[:space:]]Labels[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization/911172cc-132a-4651-a728-4f623936c716_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]from[[:space:]]Semantic[[:space:]]Alignment[[:space:]]between[[:space:]]Unpaired[[:space:]]Multiviews[[:space:]]for[[:space:]]Egocentric[[:space:]]Video[[:space:]]Recognition/3976a321-a1a8-401c-abba-91b9fbe78433_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]in[[:space:]]Imperfect[[:space:]]Environment_[[:space:]]Multi-Label[[:space:]]Classification[[:space:]]with[[:space:]]Long-Tailed[[:space:]]Distribution[[:space:]]and[[:space:]]Partial[[:space:]]Labels/0590a041-a804-4736-ab7e-fab59b5b776d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]to[[:space:]]Distill[[:space:]]Global[[:space:]]Representation[[:space:]]for[[:space:]]Sparse-View[[:space:]]CT/816ad801-b997-48a5-9faf-a09a225c6e9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]to[[:space:]]Generate[[:space:]]Semantic[[:space:]]Layouts[[:space:]]for[[:space:]]Higher[[:space:]]Text-Image[[:space:]]Correspondence[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Synthesis/580abc17-52d1-48f0-9975-3f39f0fd195f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]to[[:space:]]Ground[[:space:]]Instructional[[:space:]]Articles[[:space:]]in[[:space:]]Videos[[:space:]]through[[:space:]]Narrations/a5da633a-33ad-4510-84f7-55e37a75aa45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]to[[:space:]]Identify[[:space:]]Critical[[:space:]]States[[:space:]]for[[:space:]]Reinforcement[[:space:]]Learning[[:space:]]from[[:space:]]Videos/9d7591c9-99b4-4c7c-883c-f41dff434bc7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]to[[:space:]]Learn_[[:space:]]How[[:space:]]to[[:space:]]Continuously[[:space:]]Teach[[:space:]]Humans[[:space:]]and[[:space:]]Machines/794423e7-4b19-465b-9dc4-d9784b88b78e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]to[[:space:]]Transform[[:space:]]for[[:space:]]Generalizable[[:space:]]Instance-wise[[:space:]]Invariance/42cb6d78-d42d-4b01-aeff-7dfb1fcbef32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]to[[:space:]]Upsample[[:space:]]by[[:space:]]Learning[[:space:]]to[[:space:]]Sample/76a09629-5aaf-47bc-bbfc-722e93446189_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Learning[[:space:]]with[[:space:]]Diversity_[[:space:]]Self-Expanded[[:space:]]Equalization[[:space:]]for[[:space:]]Better[[:space:]]Generalized[[:space:]]Deep[[:space:]]Metric[[:space:]]Learning/3a1521d2-4ec7-4839-9c60-c7409d550c5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Lecture[[:space:]]Presentations[[:space:]]Multimodal[[:space:]]Dataset_[[:space:]]Towards[[:space:]]Understanding[[:space:]]Multimodality[[:space:]]in[[:space:]]Educational[[:space:]]Videos/c555fb73-35f1-474d-bacd-f2e336bcdb4f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Lens[[:space:]]Parameter[[:space:]]Estimation[[:space:]]for[[:space:]]Realistic[[:space:]]Depth[[:space:]]of[[:space:]]Field[[:space:]]Modeling/68931115-0a5c-4b32-83e2-387d3ca1ecf1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Less[[:space:]]is[[:space:]]More_[[:space:]]Focus[[:space:]]Attention[[:space:]]for[[:space:]]Efficient[[:space:]]DETR/2a4e88e8-851d-4c18-b213-e3ecc786fac1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Leveraging[[:space:]]Inpainting[[:space:]]for[[:space:]]Single-Image[[:space:]]Shadow[[:space:]]Removal/36723b14-e32c-4e97-9bce-c3c690534332_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Leveraging[[:space:]]Intrinsic[[:space:]]Properties[[:space:]]for[[:space:]]Non-Rigid[[:space:]]Garment[[:space:]]Alignment/76c61bef-ff1e-4c15-8f42-03a05ee216e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Leveraging[[:space:]]SE(3)[[:space:]]Equivariance[[:space:]]for[[:space:]]Learning[[:space:]]3D[[:space:]]Geometric[[:space:]]Shape[[:space:]]Assembly/860c6372-3554-4dfd-9464-7acd8a1a48da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Leveraging[[:space:]]Spatio-Temporal[[:space:]]Dependency[[:space:]]for[[:space:]]Skeleton-Based[[:space:]]Action[[:space:]]Recognition/ccc6fcda-0804-4c1f-9b33-bb3f31008209_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LexLIP_[[:space:]]Lexicon-Bottlenecked[[:space:]]Language-Image[[:space:]]Pre-Training[[:space:]]for[[:space:]]Large-Scale[[:space:]]Image-Text[[:space:]]Sparse[[:space:]]Retrieval/5bb38d4f-c86f-4dc7-9050-1b213a6c438d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LiDAR-Camera[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]via[[:space:]]Geometry-Consistent[[:space:]]and[[:space:]]Semantic-Aware[[:space:]]Alignment/5bfdc116-ec9e-4816-a47c-6c65df3a4f8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LiDAR-UDA_[[:space:]]Self-ensembling[[:space:]]Through[[:space:]]Time[[:space:]]for[[:space:]]Unsupervised[[:space:]]LiDAR[[:space:]]Domain[[:space:]]Adaptation/d4d3c041-43d8-4366-9140-ec16adc76dc5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LightDepth_[[:space:]]Single-View[[:space:]]Depth[[:space:]]Self-Supervision[[:space:]]from[[:space:]]Illumination[[:space:]]Decline/12c0e216-423d-4d20-b6d9-8142506010ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LightGlue_[[:space:]]Local[[:space:]]Feature[[:space:]]Matching[[:space:]]at[[:space:]]Light[[:space:]]Speed/0fc08733-6e3f-4ce8-8360-76ecf3bc7ab6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Lighting[[:space:]]Every[[:space:]]Darkness[[:space:]]in[[:space:]]Two[[:space:]]Pairs_[[:space:]]A[[:space:]]Calibration-Free[[:space:]]Pipeline[[:space:]]for[[:space:]]RAW[[:space:]]Denoising/4b55fdc1-47de-4dbb-86f4-b56402227b74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Lighting[[:space:]]up[[:space:]]NeRF[[:space:]]via[[:space:]]Unsupervised[[:space:]]Decomposition[[:space:]]and[[:space:]]Enhancement/aabd50cb-5914-4076-b6da-187d7e32d8e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Lightweight[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]with[[:space:]]Superpixel[[:space:]]Token[[:space:]]Interaction/79fde396-7940-42c9-8e78-b95022ba849b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Linear[[:space:]]Spaces[[:space:]]of[[:space:]]Meanings_[[:space:]]Compositional[[:space:]]Structures[[:space:]]in[[:space:]]Vision-Language[[:space:]]Models/db149a51-8d94-4de2-91c1-89b3c8518de7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Linear-Covariance[[:space:]]Loss[[:space:]]for[[:space:]]End-to-End[[:space:]]Learning[[:space:]]of[[:space:]]6D[[:space:]]Pose[[:space:]]Estimation/03f21eca-f84f-44f4-9cae-b923476022b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LinkGAN_[[:space:]]Linking[[:space:]]GAN[[:space:]]Latents[[:space:]]to[[:space:]]Pixels[[:space:]]for[[:space:]]Controllable[[:space:]]Image[[:space:]]Synthesis/b4385e01-04b7-4a01-902e-ccf701b4d33e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Lip[[:space:]]Reading[[:space:]]for[[:space:]]Low-resource[[:space:]]Languages[[:space:]]by[[:space:]]Learning[[:space:]]and[[:space:]]Combining[[:space:]]General[[:space:]]Speech[[:space:]]Knowledge[[:space:]]and[[:space:]]Language-specific[[:space:]]Knowledge/68b316c0-7588-474a-b180-0073d4ffb11f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Lip2Vec_[[:space:]]Efficient[[:space:]]and[[:space:]]Robust[[:space:]]Visual[[:space:]]Speech[[:space:]]Recognition[[:space:]]via[[:space:]]Latent-to-Latent[[:space:]]Visual[[:space:]]to[[:space:]]Audio[[:space:]]Representation[[:space:]]Mapping/87b706bc-939b-481d-8085-c9b557b268b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LiveHand_[[:space:]]Real-time[[:space:]]and[[:space:]]Photorealistic[[:space:]]Neural[[:space:]]Hand[[:space:]]Rendering/8f99ec7f-726d-4ff2-8806-030b764f37ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LivePose_[[:space:]]Online[[:space:]]3D[[:space:]]Reconstruction[[:space:]]from[[:space:]]Monocular[[:space:]]Video[[:space:]]with[[:space:]]Dynamic[[:space:]]Camera[[:space:]]Poses/9e9ec6b8-4b77-4480-8e91-6c3c1bdd74f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LivelySpeaker_[[:space:]]Towards[[:space:]]Semantic-Aware[[:space:]]Co-Speech[[:space:]]Gesture[[:space:]]Generation/cfcd51ff-f938-4877-865b-c93c4eb92efa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LoCUS_[[:space:]]Learning[[:space:]]Multiscale[[:space:]]3D-consistent[[:space:]]Features[[:space:]]from[[:space:]]Posed[[:space:]]Images/bdaeaa7c-3154-4b37-8417-4b69a9fde655_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LoGoPrompt_[[:space:]]Synthetic[[:space:]]Text[[:space:]]Images[[:space:]]Can[[:space:]]Be[[:space:]]Good[[:space:]]Visual[[:space:]]Prompts[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/cdb70ba6-3154-4ea5-8280-86e197015d54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LoLep_[[:space:]]Single-View[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]Locally-Learned[[:space:]]Planes[[:space:]]and[[:space:]]Self-Attention[[:space:]]Occlusion[[:space:]]Inference/421afef6-bb14-48d1-beda-63c945ec0084_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LoTE-Animal_[[:space:]]A[[:space:]]Long[[:space:]]Time-span[[:space:]]Dataset[[:space:]]for[[:space:]]Endangered[[:space:]]Animal[[:space:]]Behavior[[:space:]]Understanding/4dc27b24-fcd7-4a70-b637-438bfd277b20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Local[[:space:]]Context-Aware[[:space:]]Active[[:space:]]Domain[[:space:]]Adaptation/3e75f915-99b7-42dd-b5e3-b526d37b6e3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Local[[:space:]]and[[:space:]]Global[[:space:]]Logit[[:space:]]Adjustments[[:space:]]for[[:space:]]Long-Tailed[[:space:]]Learning/f228f66f-8e9b-4f60-9fb7-946b75e6b3e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Local[[:space:]]or[[:space:]]Global_[[:space:]]Selective[[:space:]]Knowledge[[:space:]]Assimilation[[:space:]]for[[:space:]]Federated[[:space:]]Learning[[:space:]]with[[:space:]]Limited[[:space:]]Labels/accff621-d681-458a-ba79-d752abe6d463_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Localizing[[:space:]]Moments[[:space:]]in[[:space:]]Long[[:space:]]Video[[:space:]]Via[[:space:]]Multimodal[[:space:]]Guidance/e6649f9e-dfb6-4bd3-9f05-e59d3919d706_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Localizing[[:space:]]Object-Level[[:space:]]Shape[[:space:]]Variations[[:space:]]with[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/2c468e24-56cf-4358-bc85-fde64ea6525e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Locally[[:space:]]Stylized[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/c57f70be-dfde-4f55-a20e-2bb86ddee8fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Locating[[:space:]]Noise[[:space:]]is[[:space:]]Halfway[[:space:]]Denoising[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Segmentation/6e47c59c-47ac-4877-b116-135b4077e772_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Locomotion-Action-Manipulation_[[:space:]]Synthesizing[[:space:]]Human-Scene[[:space:]]Interactions[[:space:]]in[[:space:]]Complex[[:space:]]3D[[:space:]]Environments/dd30424c-8f39-4f81-9610-40f8608f82c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Logic-induced[[:space:]]Diagnostic[[:space:]]Reasoning[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Semantic[[:space:]]Segmentation/18008f91-3ff4-45ed-b6ee-81b214f46c8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/LogicSeg_[[:space:]]Parsing[[:space:]]Visual[[:space:]]Semantics[[:space:]]with[[:space:]]Neural[[:space:]]Logic[[:space:]]Learning[[:space:]]and[[:space:]]Reasoning/b0bc60dd-fe96-4853-8e83-ec57e8130030_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Long-Range[[:space:]]Grouping[[:space:]]Transformer[[:space:]]for[[:space:]]Multi-View[[:space:]]3D[[:space:]]Reconstruction/4e9a3517-9427-4825-a666-7fb8fb3342ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Long-Term[[:space:]]Photometric[[:space:]]Consistent[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/8bbd38aa-6dc3-43b2-bd07-8bfc3d8084b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Long-range[[:space:]]Multimodal[[:space:]]Pretraining[[:space:]]for[[:space:]]Movie[[:space:]]Understanding/bec9d768-790d-4d3a-89f8-948b46f37c5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Look[[:space:]]at[[:space:]]the[[:space:]]Neighbor_[[:space:]]Distortion-aware[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Panoramic[[:space:]]Semantic[[:space:]]Segmentation/5dea4a43-f8f7-4a42-89d3-acf977588b12_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Lossy[[:space:]]and[[:space:]]Lossless[[:space:]](L2)[[:space:]]Post-training[[:space:]]Model[[:space:]]Size[[:space:]]Compression/9fe6614d-0e2c-4d57-9a3c-5f851c25c7f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Low-Light[[:space:]]Image[[:space:]]Enhancement[[:space:]]with[[:space:]]Illumination-Aware[[:space:]]Gamma[[:space:]]Correction[[:space:]]and[[:space:]]Complete[[:space:]]Image[[:space:]]Modelling[[:space:]]Network/c35d96eb-0cc7-4353-88f9-fddae6ab1f7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Low-Light[[:space:]]Image[[:space:]]Enhancement[[:space:]]with[[:space:]]Multi-Stage[[:space:]]Residue[[:space:]]Quantization[[:space:]]and[[:space:]]Brightness-Aware[[:space:]]Attention/007e11ea-7d58-4d62-993a-3faf53ebc19d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Luminance-aware[[:space:]]Color[[:space:]]Transform[[:space:]]for[[:space:]]Multiple[[:space:]]Exposure[[:space:]]Correction/151bf130-b428-401d-922d-e55f137b0cb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/M2T_[[:space:]]Masking[[:space:]]Transformers[[:space:]]Twice[[:space:]]for[[:space:]]Faster[[:space:]]Decoding/7fe45e5c-8294-43cb-9f0f-68176387d87b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAAL_[[:space:]]Multimodality-Aware[[:space:]]Autoencoder-Based[[:space:]]Affordance[[:space:]]Learning[[:space:]]for[[:space:]]3D[[:space:]]Articulated[[:space:]]Objects/7fbc4aa1-7a93-4d6e-ab50-9088c5ffe5ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAGI_[[:space:]]Multi-Annotated[[:space:]]Explanation-Guided[[:space:]]Learning/1dca00ec-5387-4988-a5b7-b9b58a03b227_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAMo_[[:space:]]Leveraging[[:space:]]Memory[[:space:]]and[[:space:]]Attention[[:space:]]for[[:space:]]Monocular[[:space:]]Video[[:space:]]Depth[[:space:]]Estimation/9b038a6e-8cec-41aa-9a60-6b5c149d4bcc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAPConNet_[[:space:]]Self-supervised[[:space:]]3D[[:space:]]Pose[[:space:]]Transfer[[:space:]]with[[:space:]]Mesh[[:space:]]and[[:space:]]Point[[:space:]]Contrastive[[:space:]]Learning/5ea6a702-87eb-41ed-9dc3-c7ff1bb4108e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAP_[[:space:]]Towards[[:space:]]Balanced[[:space:]]Generalization[[:space:]]of[[:space:]]IID[[:space:]]and[[:space:]]OOD[[:space:]]through[[:space:]]Model-Agnostic[[:space:]]Adapters/5221f0b5-da5c-4f1f-8373-51c14a7ee74e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MARS_[[:space:]]Model-agnostic[[:space:]]Biased[[:space:]]Object[[:space:]]Removal[[:space:]]without[[:space:]]Additional[[:space:]]Supervision[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Semantic[[:space:]]Segmentation/019cd255-48ce-4e7f-bc37-bdb0d0951d77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAS_[[:space:]]Towards[[:space:]]Resource-Efficient[[:space:]]Federated[[:space:]]Multiple-Task[[:space:]]Learning/949dea25-08fe-4a97-88d9-cba01d80d17f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MATE_[[:space:]]Masked[[:space:]]Autoencoders[[:space:]]are[[:space:]]Online[[:space:]]3D[[:space:]]Test-Time[[:space:]]Learners/33f725f9-63f7-44f1-bb31-b989f43bed82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MAtch,[[:space:]]eXpand[[:space:]]and[[:space:]]Improve_[[:space:]]Unsupervised[[:space:]]Finetuning[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Action[[:space:]]Recognition[[:space:]]with[[:space:]]Language[[:space:]]Knowledge/867cfa3f-7db4-4642-9d8e-f44db510f640_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MB-TaylorFormer_[[:space:]]Multi-Branch[[:space:]]Efficient[[:space:]]Transformer[[:space:]]Expanded[[:space:]]by[[:space:]]Taylor[[:space:]]Formula[[:space:]]for[[:space:]]Image[[:space:]]Dehazing/923931e1-44ae-46c8-a47e-873914dbf6e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MBPTrack_[[:space:]]Improving[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Tracking[[:space:]]with[[:space:]]Memory[[:space:]]Networks[[:space:]]and[[:space:]]Box[[:space:]]Priors/956ae52e-1f1a-4203-9786-92a3a80682aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MDCS_[[:space:]]More[[:space:]]Diverse[[:space:]]Experts[[:space:]]with[[:space:]]Consistency[[:space:]]Self-distillation[[:space:]]for[[:space:]]Long-tailed[[:space:]]Recognition/4759726a-56c7-4e00-9a4a-80fbdc668438_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MEFLUT_[[:space:]]Unsupervised[[:space:]]1D[[:space:]]Lookup[[:space:]]Tables[[:space:]]for[[:space:]]Multi-exposure[[:space:]]Image[[:space:]]Fusion/601a9827-54df-4685-968b-72e1c7050277_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MEGA_[[:space:]]Multimodal[[:space:]]Alignment[[:space:]]Aggregation[[:space:]]and[[:space:]]Distillation[[:space:]]For[[:space:]]Cinematic[[:space:]]Video[[:space:]]Segmentation/014aa68e-4ee6-4ed7-996a-a0fac673a258_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MGMAE_[[:space:]]Motion[[:space:]]Guided[[:space:]]Masking[[:space:]]for[[:space:]]Video[[:space:]]Masked[[:space:]]Autoencoding/913ed268-24e1-487c-b23b-995e3fca2c86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MHCN_[[:space:]]A[[:space:]]Hyperbolic[[:space:]]Neural[[:space:]]Network[[:space:]]Model[[:space:]]for[[:space:]]Multi-view[[:space:]]Hierarchical[[:space:]]Clustering/803b20b8-8ea9-4642-8b2e-1e4d0ac18614_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MHEntropy_[[:space:]]Entropy[[:space:]]Meets[[:space:]]Multiple[[:space:]]Hypotheses[[:space:]]for[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]Recovery/f9c54455-f324-48c7-9865-4e8d15caf336_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MI-GAN_[[:space:]]A[[:space:]]Simple[[:space:]]Baseline[[:space:]]for[[:space:]]Image[[:space:]]Inpainting[[:space:]]on[[:space:]]Mobile[[:space:]]Devices/af22c1bd-1590-4032-ac12-a1b776021514_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MIMO-NeRF_[[:space:]]Fast[[:space:]]Neural[[:space:]]Rendering[[:space:]]with[[:space:]]Multi-input[[:space:]]Multi-output[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/323ff8ab-5b3c-4603-a4d5-667546c604d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MMST-ViT_[[:space:]]Climate[[:space:]]Change-aware[[:space:]]Crop[[:space:]]Yield[[:space:]]Prediction[[:space:]]via[[:space:]]Multi-Modal[[:space:]]Spatial-Temporal[[:space:]]Vision[[:space:]]Transformer/8b54a15b-f3ee-46da-88b3-5c8950b9cb49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MMVP_[[:space:]]Motion-Matrix-Based[[:space:]]Video[[:space:]]Prediction/2d057465-769a-4de5-aaaa-89b4c330116e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MODA_[[:space:]]Mapping-Once[[:space:]]Audio-driven[[:space:]]Portrait[[:space:]]Animation[[:space:]]with[[:space:]]Dual[[:space:]]Attentions/cd6d82fe-3030-4ffc-8902-e0f3febe9ace_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MOSE_[[:space:]]A[[:space:]]New[[:space:]]Dataset[[:space:]]for[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation[[:space:]]in[[:space:]]Complex[[:space:]]Scenes/aa32c686-8817-467e-b52a-3ad29e222646_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MOST_[[:space:]]Multiple[[:space:]]Object[[:space:]]Localization[[:space:]]with[[:space:]]Self-Supervised[[:space:]]Transformers[[:space:]]for[[:space:]]Object[[:space:]]Discovery/b6bef65d-9eb6-4535-9de8-7c08bf88dd4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MPCViT_[[:space:]]Searching[[:space:]]for[[:space:]]Accurate[[:space:]]and[[:space:]]Efficient[[:space:]]MPC-Friendly[[:space:]]Vision[[:space:]]Transformer[[:space:]]with[[:space:]]Heterogeneous[[:space:]]Attention/844d9362-4725-44da-8709-8687e3c64eb2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MPI-Flow_[[:space:]]Learning[[:space:]]Realistic[[:space:]]Optical[[:space:]]Flow[[:space:]]with[[:space:]]Multiplane[[:space:]]Images/d01da3cc-92b1-43b1-b6d0-867a68823626_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MRM_[[:space:]]Masked[[:space:]]Relation[[:space:]]Modeling[[:space:]]for[[:space:]]Medical[[:space:]]Image[[:space:]]Pre-Training[[:space:]]with[[:space:]]Genetics/3ca9548a-636d-4952-80cd-106e032bd384_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MRN_[[:space:]]Multiplexed[[:space:]]Routing[[:space:]]Network[[:space:]]for[[:space:]]Incremental[[:space:]]Multilingual[[:space:]]Text[[:space:]]Recognition/6ed06787-7eb3-4534-9eb1-83af9b81cded_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MSI_[[:space:]]Maximize[[:space:]]Support-Set[[:space:]]Information[[:space:]]for[[:space:]]Few-Shot[[:space:]]Segmentation/18b3712a-f7d6-4a6b-8811-754fd084ad1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MST-compression_[[:space:]]Compressing[[:space:]]and[[:space:]]Accelerating[[:space:]]Binary[[:space:]]Neural[[:space:]]Networks[[:space:]]with[[:space:]]Minimum[[:space:]]Spanning[[:space:]]Tree/9bf308bf-7020-43ef-9313-33d2e81da894_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MULLER_[[:space:]]Multilayer[[:space:]]Laplacian[[:space:]]Resizer[[:space:]]for[[:space:]]Vision/540b48e0-0e9c-4eba-a984-866547201e9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MUVA_[[:space:]]A[[:space:]]New[[:space:]]Large-Scale[[:space:]]Benchmark[[:space:]]for[[:space:]]Multi-View[[:space:]]Amodal[[:space:]]Instance[[:space:]]Segmentation[[:space:]]in[[:space:]]the[[:space:]]Shopping[[:space:]]Scenario/4b78daa4-2084-48f2-bdbe-f8d1b9a4826f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MUter_[[:space:]]Machine[[:space:]]Unlearning[[:space:]]on[[:space:]]Adversarially[[:space:]]Trained[[:space:]]Models/794014c8-e2b9-40ef-b0f6-0010a00c9a4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MV-DeepSDF_[[:space:]]Implicit[[:space:]]Modeling[[:space:]]with[[:space:]]Multi-Sweep[[:space:]]Point[[:space:]]Clouds[[:space:]]for[[:space:]]3D[[:space:]]Vehicle[[:space:]]Reconstruction[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/0f77204f-207d-45fd-9794-f340dac7f292_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MV-Map_[[:space:]]Offboard[[:space:]]HD-Map[[:space:]]Generation[[:space:]]with[[:space:]]Multi-view[[:space:]]Consistency/59188c7c-19e6-4a92-b752-2ffcc7690379_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MVPSNet_[[:space:]]Fast[[:space:]]Generalizable[[:space:]]Multi-view[[:space:]]Photometric[[:space:]]Stereo/0b36b82c-cf6a-406b-8b06-65bc02693944_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MagicFusion_[[:space:]]Boosting[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]Performance[[:space:]]by[[:space:]]Fusing[[:space:]]Diffusion[[:space:]]Models/40801689-afc7-4c4a-b4e3-8cbca2d70a56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Make[[:space:]]Encoder[[:space:]]Great[[:space:]]Again[[:space:]]in[[:space:]]3D[[:space:]]GAN[[:space:]]Inversion[[:space:]]through[[:space:]]Geometry[[:space:]]and[[:space:]]Occlusion-Aware[[:space:]]Encoding/d0a41b80-a701-4955-834c-d42058b2d111_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Make-An-Animation_[[:space:]]Large-Scale[[:space:]]Text-conditional[[:space:]]3D[[:space:]]Human[[:space:]]Motion[[:space:]]Generation/0b6ce6f5-40a5-4972-8f2c-3c7fc7288f3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Make-It-3D_[[:space:]]High-fidelity[[:space:]]3D[[:space:]]Creation[[:space:]]from[[:space:]]A[[:space:]]Single[[:space:]]Image[[:space:]]with[[:space:]]Diffusion[[:space:]]Prior/3c32b216-e73e-471f-8cad-0ce51b8a9a7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Manipulate[[:space:]]by[[:space:]]Seeing_[[:space:]]Creating[[:space:]]Manipulation[[:space:]]Controllers[[:space:]]from[[:space:]]Pre-Trained[[:space:]]Representations/4fa47878-d18c-4c5d-89a6-b50ae3093aaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MapFormer_[[:space:]]Boosting[[:space:]]Change[[:space:]]Detection[[:space:]]by[[:space:]]Using[[:space:]]Pre-change[[:space:]]Information/8472e89b-fe96-412d-8c9c-414b2c42d2f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MapPrior_[[:space:]]Bird's-Eye[[:space:]]View[[:space:]]Map[[:space:]]Layout[[:space:]]Estimation[[:space:]]with[[:space:]]Generative[[:space:]]Models/80d4c617-535f-4e51-ba9c-d279342878ed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/March[[:space:]]in[[:space:]]Chat_[[:space:]]Interactive[[:space:]]Prompting[[:space:]]for[[:space:]]Remote[[:space:]]Embodied[[:space:]]Referring[[:space:]]Expression/dfdb2c34-61fe-4960-a099-a29f68122739_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Markov[[:space:]]Game[[:space:]]Video[[:space:]]Augmentation[[:space:]]for[[:space:]]Action[[:space:]]Segmentation/3ebb7b09-a6e7-4446-89ff-eb9741fd0b7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MasQCLIP[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Universal[[:space:]]Image[[:space:]]Segmentation/0636ab48-f4a3-4a28-9365-8ed3c11dca4b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MasaCtrl_[[:space:]]Tuning-Free[[:space:]]Mutual[[:space:]]Self-Attention[[:space:]]Control[[:space:]]for[[:space:]]Consistent[[:space:]]Image[[:space:]]Synthesis[[:space:]]and[[:space:]]Editing/6e476140-cab6-422d-9817-a489e1ead454_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mask-Attention-Free[[:space:]]Transformer[[:space:]]for[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation/2a13e96b-8070-4a45-a596-f26ce5eefa9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Autoencoders[[:space:]]Are[[:space:]]Stronger[[:space:]]Knowledge[[:space:]]Distillers/801459a1-6a61-45c0-b371-eb489abaf84f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Autoencoders[[:space:]]are[[:space:]]Efficient[[:space:]]Class[[:space:]]Incremental[[:space:]]Learners/3431d5c0-b1b6-47da-9e6e-bca6fba5a382_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Diffusion[[:space:]]Transformer[[:space:]]is[[:space:]]a[[:space:]]Strong[[:space:]]Image[[:space:]]Synthesizer/a412e570-59ba-43e9-aea5-2909edb799b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Motion[[:space:]]Predictors[[:space:]]are[[:space:]]Strong[[:space:]]3D[[:space:]]Action[[:space:]]Representation[[:space:]]Learners/06acdbbe-e15b-4021-a13d-3cb042fcaf0d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Retraining[[:space:]]Teacher-Student[[:space:]]Framework[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detection/5a41b22e-9755-4ada-9123-4e6d5ca57c32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Spatio-Temporal[[:space:]]Structure[[:space:]]Prediction[[:space:]]for[[:space:]]Self-supervised[[:space:]]Learning[[:space:]]on[[:space:]]Point[[:space:]]Cloud[[:space:]]Videos/e14b4021-babf-4da5-a112-3aa5b237ce50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Masked[[:space:]]Spiking[[:space:]]Transformer/7ef4a39e-394a-4406-bcd4-3ff4096115d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mastering[[:space:]]Spatial[[:space:]]Graph[[:space:]]Prediction[[:space:]]of[[:space:]]Road[[:space:]]Networks/e2e1da73-cf2c-48bb-9b35-8cf5cb0f98fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MatrixCity_[[:space:]]A[[:space:]]Large-scale[[:space:]]City[[:space:]]Dataset[[:space:]]for[[:space:]]City-scale[[:space:]]Neural[[:space:]]Rendering[[:space:]]and[[:space:]]Beyond/1fb06bbc-ee8a-4da1-8280-e317a61d1b94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MatrixVT_[[:space:]]Efficient[[:space:]]Multi-Camera[[:space:]]to[[:space:]]BEV[[:space:]]Transformation[[:space:]]for[[:space:]]3D[[:space:]]Perception/3d8f338d-7ad3-414d-a017-25b8e25151fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MeMOTR_[[:space:]]Long-Term[[:space:]]Memory-Augmented[[:space:]]Transformer[[:space:]]for[[:space:]]Multi-Object[[:space:]]Tracking/40deaafe-8b86-4716-a18c-f883acf32105_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MeViS_[[:space:]]A[[:space:]]Large-scale[[:space:]]Benchmark[[:space:]]for[[:space:]]Video[[:space:]]Segmentation[[:space:]]with[[:space:]]Motion[[:space:]]Expressions/422d1f87-df33-4f64-91c1-e11c03538001_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Measuring[[:space:]]Asymmetric[[:space:]]Gradient[[:space:]]Discrepancy[[:space:]]in[[:space:]]Parallel[[:space:]]Continual[[:space:]]Learning/e2e81f44-6f94-4699-81e2-292ae5d4c0a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MedKLIP_[[:space:]]Medical[[:space:]]Knowledge[[:space:]]Enhanced[[:space:]]Language-Image[[:space:]]Pre-Training[[:space:]]for[[:space:]]X-ray[[:space:]]Diagnosis/a294a7d4-55de-429f-8329-44c18b938170_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Membrane[[:space:]]Potential[[:space:]]Batch[[:space:]]Normalization[[:space:]]for[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks/cc22facf-9790-4a7c-8861-ea4adecda7ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Memory-and-Anticipation[[:space:]]Transformer[[:space:]]for[[:space:]]Online[[:space:]]Action[[:space:]]Understanding/a1b3a75e-11e3-4ea7-94a7-a5bef15574c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MemorySeg_[[:space:]]Online[[:space:]]LiDAR[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]a[[:space:]]Latent[[:space:]]Memory/9db8fa07-a91f-4043-9a71-b3d3e6b17388_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mesh2Tex_[[:space:]]Generating[[:space:]]Mesh[[:space:]]Textures[[:space:]]from[[:space:]]Image[[:space:]]Queries/743afd5f-14b2-4965-bbeb-bded0a673761_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Meta[[:space:]]OOD[[:space:]]Learning[[:space:]]For[[:space:]]Continuously[[:space:]]Adaptive[[:space:]]OOD[[:space:]]Detection/06d792b4-9c42-4f8e-baa6-449a229309f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Meta-ZSDETR_[[:space:]]Zero-shot[[:space:]]DETR[[:space:]]with[[:space:]]Meta-learning/ab81a251-c4da-4048-9092-c84f97030f0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MetaBEV_[[:space:]]Solving[[:space:]]Sensor[[:space:]]Failures[[:space:]]for[[:space:]]3D[[:space:]]Detection[[:space:]]and[[:space:]]Map[[:space:]]Segmentation/5204095b-1c9c-4150-9f9c-b1c27fdd82da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MetaF2N_[[:space:]]Blind[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]by[[:space:]]Learning[[:space:]]Efficient[[:space:]]Model[[:space:]]Adaptation[[:space:]]from[[:space:]]Faces/ed6327a3-66bd-4260-9c23-011c9182cd69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MetaGCD_[[:space:]]Learning[[:space:]]to[[:space:]]Continually[[:space:]]Learn[[:space:]]in[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/991c02f6-2cc1-4391-a19c-b8559690aace_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Metric3D_[[:space:]]Towards[[:space:]]Zero-shot[[:space:]]Metric[[:space:]]3D[[:space:]]Prediction[[:space:]]from[[:space:]]A[[:space:]]Single[[:space:]]Image/6cc2d292-7284-48d5-98ac-76d175845d01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mimic3D_[[:space:]]Thriving[[:space:]]3D-Aware[[:space:]]GANs[[:space:]]via[[:space:]]3D-to-2D[[:space:]]Imitation/d009dfab-0a08-423e-b77c-b88a4b5cf703_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MiniROAD_[[:space:]]Minimal[[:space:]]RNN[[:space:]]Framework[[:space:]]for[[:space:]]Online[[:space:]]Action[[:space:]]Detection/ea56ef7d-d0ee-4b4c-86bd-22698b72c0f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Minimal[[:space:]]Solutions[[:space:]]to[[:space:]]Generalized[[:space:]]Three-View[[:space:]]Relative[[:space:]]Pose[[:space:]]Problem/21f6bf16-a75f-4e22-8ec8-d9aabd4e98e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Minimal[[:space:]]Solutions[[:space:]]to[[:space:]]Uncalibrated[[:space:]]Two-view[[:space:]]Geometry[[:space:]]with[[:space:]]Known[[:space:]]Epipoles/1ff0de2c-2b50-437f-a3e0-afb03703f817_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Minimum[[:space:]]Latency[[:space:]]Deep[[:space:]]Online[[:space:]]Video[[:space:]]Stabilization/5c08d8de-7fcf-430c-a7d0-16aae326aed4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mining[[:space:]]bias-target[[:space:]]Alignment[[:space:]]from[[:space:]]Voronoi[[:space:]]Cells/45d7f31f-5be7-4011-90ca-1ebee3489169_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Misalign,[[:space:]]Contrast[[:space:]]then[[:space:]]Distill_[[:space:]]Rethinking[[:space:]]Misalignments[[:space:]]in[[:space:]]Language-Image[[:space:]]Pre-training/8db6862b-91a8-4f4b-b1af-4631b12bd46c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mitigating[[:space:]]Adversarial[[:space:]]Vulnerability[[:space:]]through[[:space:]]Causal[[:space:]]Parameter[[:space:]]Estimation[[:space:]]by[[:space:]]Adversarial[[:space:]]Double[[:space:]]Machine[[:space:]]Learning/720877af-e8c3-460a-ab3f-fbddf4379306_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mitigating[[:space:]]and[[:space:]]Evaluating[[:space:]]Static[[:space:]]Bias[[:space:]]of[[:space:]]Action[[:space:]]Representations[[:space:]]in[[:space:]]the[[:space:]]Background[[:space:]]and[[:space:]]the[[:space:]]Foreground/6ba02bea-6b1d-48df-8380-8f905fcf7514_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MixBag_[[:space:]]Bag-Level[[:space:]]Data[[:space:]]Augmentation[[:space:]]for[[:space:]]Learning[[:space:]]from[[:space:]]Label[[:space:]]Proportions/c288d48a-0e97-49d8-951a-99b2d027811b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MixCycle_[[:space:]]Mixup[[:space:]]Assisted[[:space:]]Semi-Supervised[[:space:]]3D[[:space:]]Single[[:space:]]Object[[:space:]]Tracking[[:space:]]with[[:space:]]Cycle[[:space:]]Consistency/1f3580cb-fdbe-4e7d-96d7-32968d08a569_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MixPath_[[:space:]]A[[:space:]]Unified[[:space:]]Approach[[:space:]]for[[:space:]]One-shot[[:space:]]Neural[[:space:]]Architecture[[:space:]]Search/7b5c5fe0-450e-42c0-9209-d2128ddb12f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MixReorg_[[:space:]]Cross-Modal[[:space:]]Mixed[[:space:]]Patch[[:space:]]Reorganization[[:space:]]is[[:space:]]a[[:space:]]Good[[:space:]]Mask[[:space:]]Learner[[:space:]]for[[:space:]]Open-World[[:space:]]Semantic[[:space:]]Segmentation/b98d5aa5-34b5-4693-a5bc-bc45e3fb8dcd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MixSpeech_[[:space:]]Cross-Modality[[:space:]]Self-Learning[[:space:]]with[[:space:]]Audio-Visual[[:space:]]Stream[[:space:]]Mixup[[:space:]]for[[:space:]]Visual[[:space:]]Speech[[:space:]]Translation[[:space:]]and[[:space:]]Recognition/8115edb8-8391-4f40-bf06-0cd9421711d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MixSynthFormer_[[:space:]]A[[:space:]]Transformer[[:space:]]Encoder-like[[:space:]]Structure[[:space:]]with[[:space:]]Mixed[[:space:]]Synthetic[[:space:]]Self-attention[[:space:]]for[[:space:]]Efficient[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/846dbe02-0b5b-4bf4-a391-96d12e988697_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Mixed[[:space:]]Neural[[:space:]]Voxels[[:space:]]for[[:space:]]Fast[[:space:]]Multi-view[[:space:]]Video[[:space:]]Synthesis/77672856-3b59-4e13-b4ae-857030e3ef1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MoTIF_[[:space:]]Learning[[:space:]]Motion[[:space:]]Trajectories[[:space:]]with[[:space:]]Local[[:space:]]Implicit[[:space:]]Neural[[:space:]]Functions[[:space:]]for[[:space:]]Continuous[[:space:]]Space-Time[[:space:]]Video[[:space:]]Super-Resolution/b7d53caf-5b95-474a-8203-b3df729d9009_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Modality[[:space:]]Unifying[[:space:]]Network[[:space:]]for[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification/c25cccee-cf9c-444a-bfaa-52844fa6cb10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Model[[:space:]]Calibration[[:space:]]in[[:space:]]Dense[[:space:]]Classification[[:space:]]with[[:space:]]Adaptive[[:space:]]Label[[:space:]]Perturbation/7e446bd7-c368-40d4-8d81-6aee3f5055e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ModelGiF_[[:space:]]Gradient[[:space:]]Fields[[:space:]]for[[:space:]]Model[[:space:]]Functional[[:space:]]Distance/5efa34c8-5ce4-47e6-bc15-889009199a89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Modeling[[:space:]]the[[:space:]]Relative[[:space:]]Visual[[:space:]]Tempo[[:space:]]for[[:space:]]Self-supervised[[:space:]]Skeleton-based[[:space:]]Action[[:space:]]Recognition/dbdfe6ed-ba3c-4daf-ac10-86c2b2f20cde_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MolGrapher_[[:space:]]Graph-based[[:space:]]Visual[[:space:]]Recognition[[:space:]]of[[:space:]]Chemical[[:space:]]Structures/0cbbbec6-ac8d-47e3-9896-a438f1a62616_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Moment[[:space:]]Detection[[:space:]]in[[:space:]]Long[[:space:]]Tutorial[[:space:]]Videos/d0d347cf-3a4a-4154-afe6-e8c1e6e24d00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MonoDETR_[[:space:]]Depth-guided[[:space:]]Transformer[[:space:]]for[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection/3836a1be-d313-4152-8663-e644f7bd2a45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MonoNeRD_[[:space:]]NeRF-like[[:space:]]Representations[[:space:]]for[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection/63ab896a-8106-40c7-82f9-2e2b03ab3262_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MonoNeRF_[[:space:]]Learning[[:space:]]a[[:space:]]Generalizable[[:space:]]Dynamic[[:space:]]Radiance[[:space:]]Field[[:space:]]from[[:space:]]Monocular[[:space:]]Videos/3b096572-4cf3-4c62-bc09-35a9369cf899_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Bounding[[:space:]]Box[[:space:]]Denoising[[:space:]]in[[:space:]]3D[[:space:]]by[[:space:]]Perceiver/3f374e4f-acb1-4e95-8005-f14f6138f707_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Monte[[:space:]]Carlo[[:space:]]Linear[[:space:]]Clustering[[:space:]]with[[:space:]]Single-Point[[:space:]]Supervision[[:space:]]is[[:space:]]Enough[[:space:]]for[[:space:]]Infrared[[:space:]]Small[[:space:]]Target[[:space:]]Detection/ff32047e-26fb-4725-b960-eee153eb3232_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MoreauGrad_[[:space:]]Sparse[[:space:]]and[[:space:]]Robust[[:space:]]Interpretation[[:space:]]of[[:space:]]Neural[[:space:]]Networks[[:space:]]via[[:space:]]Moreau[[:space:]]Envelope/99846963-8ced-4bc2-923b-b6a5c1cca433_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MosaiQ_[[:space:]]Quantum[[:space:]]Generative[[:space:]]Adversarial[[:space:]]Networks[[:space:]]for[[:space:]]Image[[:space:]]Generation[[:space:]]on[[:space:]]NISQ[[:space:]]Computers/f3fd0f01-346f-46fa-9180-3373ca2881cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Most[[:space:]]Important[[:space:]]Person-Guided[[:space:]]Dual-Branch[[:space:]]Cross-Patch[[:space:]]Attention[[:space:]]for[[:space:]]Group[[:space:]]Affect[[:space:]]Recognition/2995ddf2-76d9-45cb-9d66-1a50a95e4d2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Motion-Guided[[:space:]]Masking[[:space:]]for[[:space:]]Spatiotemporal[[:space:]]Representation[[:space:]]Learning/381ddc28-834d-4797-9d76-2bf9c96fea0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MotionBERT_[[:space:]]A[[:space:]]Unified[[:space:]]Perspective[[:space:]]on[[:space:]]Learning[[:space:]]Human[[:space:]]Motion[[:space:]]Representations/57fc0f5c-171f-41a1-94fe-f16ec1d8935d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MotionDeltaCNN_[[:space:]]Sparse[[:space:]]CNN[[:space:]]Inference[[:space:]]of[[:space:]]Frame[[:space:]]Differences[[:space:]]in[[:space:]]Moving[[:space:]]Camera[[:space:]]Videos[[:space:]]with[[:space:]]Spherical[[:space:]]Buffers[[:space:]]and[[:space:]]Padded[[:space:]]Convolutions/60cee84a-76a7-422c-a440-0ab4865a753d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/MotionLM_[[:space:]]Multi-Agent[[:space:]]Motion[[:space:]]Forecasting[[:space:]]as[[:space:]]Language[[:space:]]Modeling/675319b9-5800-4617-9b35-401666d76cde_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Movement[[:space:]]Enhancement[[:space:]]toward[[:space:]]Multi-Scale[[:space:]]Video[[:space:]]Feature[[:space:]]Representation[[:space:]]for[[:space:]]Temporal[[:space:]]Action[[:space:]]Detection/d8a8f490-e49c-49c9-b1d9-2afa438fb3bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Directional[[:space:]]Subspace[[:space:]]Editing[[:space:]]in[[:space:]]Style-Space/a84347b3-c2c0-42e5-a59b-c77ba1dcbff7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Event[[:space:]]Video-Text[[:space:]]Retrieval/4a8e429c-a424-42a1-b75b-812c6ff37eb5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Frequency[[:space:]]Representation[[:space:]]Enhancement[[:space:]]with[[:space:]]Privilege[[:space:]]Information[[:space:]]for[[:space:]]Video[[:space:]]Super-Resolution/80575c27-91e5-4c67-9202-9382a183799d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Label[[:space:]]Knowledge[[:space:]]Distillation/3f3ec8fe-1476-4cc1-910f-9ce888c57a74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Label[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]with[[:space:]]Scene[[:space:]]Images/d8308065-1607-45e1-b7cf-dd3679dcb186_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Metrics[[:space:]]Adaptively[[:space:]]Identifies[[:space:]]Backdoors[[:space:]]in[[:space:]]Federated[[:space:]]Learning/e9dae36d-9444-4c52-bedf-bba383c1ec78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Modal[[:space:]]Continual[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation/7a008f21-91bc-4a63-ae49-b60252dec431_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Modal[[:space:]]Gated[[:space:]]Mixture[[:space:]]of[[:space:]]Local-to-Global[[:space:]]Experts[[:space:]]for[[:space:]]Dynamic[[:space:]]Image[[:space:]]Fusion/6e316164-78c4-4054-8879-22d243fd30bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Modal[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field[[:space:]]for[[:space:]]Monocular[[:space:]]Dense[[:space:]]SLAM[[:space:]]with[[:space:]]a[[:space:]]Light-Weight[[:space:]]ToF[[:space:]]Sensor/d9ac6102-d7cb-4a42-a50b-5549d5aea3c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Object[[:space:]]Discovery[[:space:]]by[[:space:]]Low-Dimensional[[:space:]]Object[[:space:]]Motion/063ca8bc-3b85-47f8-b8af-3f433d446bab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Object[[:space:]]Navigation[[:space:]]with[[:space:]]Dynamically[[:space:]]Learned[[:space:]]Neural[[:space:]]Implicit[[:space:]]Representations/8c2cacd0-618b-43fb-afc2-cdc6e282531c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Scale[[:space:]]Bidirectional[[:space:]]Recurrent[[:space:]]Network[[:space:]]with[[:space:]]Hybrid[[:space:]]Correlation[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Based[[:space:]]Scene[[:space:]]Flow[[:space:]]Estimation/2097692e-e492-4ed9-98c6-50cba1cdd9a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Scale[[:space:]]Residual[[:space:]]Low-Pass[[:space:]]Filter[[:space:]]Network[[:space:]]for[[:space:]]Image[[:space:]]Deblurring/5c2653f2-ab7f-4799-9150-1c49cbf5bd6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-Task[[:space:]]Learning[[:space:]]with[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Dense[[:space:]]Prediction/333e6685-b40e-48b9-a0cd-db1868afbfa7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-View[[:space:]]Active[[:space:]]Fine-Grained[[:space:]]Visual[[:space:]]Recognition/ee209e69-47ab-4c40-9356-cb8b45ac9b6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-body[[:space:]]Depth[[:space:]]and[[:space:]]Camera[[:space:]]Pose[[:space:]]Estimation[[:space:]]from[[:space:]]Multiple[[:space:]]Views/92fd5224-0cb4-4518-837c-e6d30b98ac63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-grained[[:space:]]Temporal[[:space:]]Prototype[[:space:]]Learning[[:space:]]for[[:space:]]Few-shot[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/6d6e86b4-be03-4421-80b1-27dd9a45c2b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-granularity[[:space:]]Interaction[[:space:]]Simulation[[:space:]]for[[:space:]]Unsupervised[[:space:]]Interactive[[:space:]]Segmentation/860ca3f9-f21c-4c96-a4c1-8228c6546c93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-interactive[[:space:]]Feature[[:space:]]Learning[[:space:]]and[[:space:]]a[[:space:]]Full-time[[:space:]]Multi-modality[[:space:]]Benchmark[[:space:]]for[[:space:]]Image[[:space:]]Fusion[[:space:]]and[[:space:]]Segmentation/bd1a1af4-3a04-4546-9ec1-7f24749110ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-label[[:space:]]Affordance[[:space:]]Mapping[[:space:]]from[[:space:]]Egocentric[[:space:]]Vision/52183b21-9cd5-4c04-8bd7-3d9da9442890_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-task[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/7bdb9fe6-4a0e-4495-bf31-c15a718cc0e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-view[[:space:]]Self-supervised[[:space:]]Disentanglement[[:space:]]for[[:space:]]General[[:space:]]Image[[:space:]]Denoising/bd73600c-bf62-4573-910c-f059e97c644d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-view[[:space:]]Spectral[[:space:]]Polarization[[:space:]]Propagation[[:space:]]for[[:space:]]Video[[:space:]]Glass[[:space:]]Segmentation/da43f4ff-b9d6-45ae-8913-a53d60c1040d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi-weather[[:space:]]Image[[:space:]]Restoration[[:space:]]via[[:space:]]Domain[[:space:]]Translation/582b1bcf-2ccd-4515-9e72-bf5ada976e39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multi3DRefer_[[:space:]]Grounding[[:space:]]Text[[:space:]]Description[[:space:]]to[[:space:]]Multiple[[:space:]]3D[[:space:]]Objects/5fd3a41b-68c4-4dc5-a721-e608deb32134_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multimodal[[:space:]]Distillation[[:space:]]for[[:space:]]Egocentric[[:space:]]Action[[:space:]]Recognition/3b080565-3bdc-4aa7-989c-41508102f95e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multimodal[[:space:]]Garment[[:space:]]Designer_[[:space:]]Human-Centric[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Fashion[[:space:]]Image[[:space:]]Editing/aa9f2273-dd32-4864-b87b-f6ed07d1408f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multimodal[[:space:]]High-order[[:space:]]Relation[[:space:]]Transformer[[:space:]]for[[:space:]]Scene[[:space:]]Boundary[[:space:]]Detection/4648d770-3062-4c5f-b2d3-868636bc79b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multimodal[[:space:]]Motion[[:space:]]Conditioned[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Skeleton-based[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/c4ebdef1-c8e8-46eb-8870-f0578399391a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multimodal[[:space:]]Optimal[[:space:]]Transport-based[[:space:]]Co-Attention[[:space:]]Transformer[[:space:]]with[[:space:]]Global[[:space:]]Structure[[:space:]]Consistency[[:space:]]for[[:space:]]Survival[[:space:]]Prediction/5f20046b-2925-4246-b895-fa5914bd8de8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multimodal[[:space:]]Variational[[:space:]]Auto-encoder[[:space:]]based[[:space:]]Audio-Visual[[:space:]]Segmentation/b063a73b-af3c-49b4-be3e-5045777f08b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multiple[[:space:]]Instance[[:space:]]Learning[[:space:]]Framework[[:space:]]with[[:space:]]Masked[[:space:]]Hard[[:space:]]Instance[[:space:]]Mining[[:space:]]for[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Classification/4186ce96-ecad-4ae2-bdd7-3d5e8448e342_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multiple[[:space:]]Planar[[:space:]]Object[[:space:]]Tracking/16e4462f-086b-4e2c-b3e1-1c3bf9449c10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multiscale[[:space:]]Representation[[:space:]]for[[:space:]]Real-Time[[:space:]]Anti-Aliasing[[:space:]]Neural[[:space:]]Rendering/841627a5-4e79-4f55-b496-850f8cdca464_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Multiscale[[:space:]]Structure[[:space:]]Guided[[:space:]]Diffusion[[:space:]]for[[:space:]]Image[[:space:]]Deblurring/94fffe15-beec-48e1-85d7-b0694e8cbf05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Muscles[[:space:]]in[[:space:]]Action/c86143c5-a6a4-4687-9f7d-303faf2b4bed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NAPA-VQ_[[:space:]]Neighborhood-Aware[[:space:]]Prototype[[:space:]]Augmentation[[:space:]]with[[:space:]]Vector[[:space:]]Quantization[[:space:]]for[[:space:]]Continual[[:space:]]Learning/f23b053f-c500-4f74-9a98-3c849db1384c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NCHO_[[:space:]]Unsupervised[[:space:]]Learning[[:space:]]for[[:space:]]Neural[[:space:]]3D[[:space:]]Composition[[:space:]]of[[:space:]]Humans[[:space:]]and[[:space:]]Objects/8c18039b-91be-4047-a4da-aeaca86dd33b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NDC-Scene_[[:space:]]Boost[[:space:]]Monocular[[:space:]]3D[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion[[:space:]]in[[:space:]]Normalized[[:space:]]Device[[:space:]]Coordinates[[:space:]]Space/83ebd013-03f3-40f5-a192-e0becc9c98b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NDDepth_[[:space:]]Normal-Distance[[:space:]]Assisted[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/38398244-b647-41ab-881c-3f0e2f41102b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NEMTO_[[:space:]]Neural[[:space:]]Environment[[:space:]]Matting[[:space:]]for[[:space:]]Novel[[:space:]]View[[:space:]]and[[:space:]]Relighting[[:space:]]Synthesis[[:space:]]of[[:space:]]Transparent[[:space:]]Objects/eb787355-0cd4-4e39-95b5-7379a4a3c48e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NIR-assisted[[:space:]]Video[[:space:]]Enhancement[[:space:]]via[[:space:]]Unpaired[[:space:]]24-hour[[:space:]]Data/656d7a30-c6ff-4d92-ac0b-7ef289ef8ada_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NLOS-NeuS_[[:space:]]Non-line-of-sight[[:space:]]Neural[[:space:]]Implicit[[:space:]]Surface/49d8e7c9-c38b-4242-949c-d65e363866b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NPC_[[:space:]]Neural[[:space:]]Point[[:space:]]Characters[[:space:]]from[[:space:]]Video/9546f2d5-ee97-4006-ae1d-884b8a05b7ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NSF_[[:space:]]Neural[[:space:]]Surface[[:space:]]Fields[[:space:]]for[[:space:]]Human[[:space:]]Modeling[[:space:]]from[[:space:]]Monocular[[:space:]]Depth/fe28edc5-3dad-427d-a6ae-c5e1a54bf46f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Name[[:space:]]Your[[:space:]]Colour[[:space:]]For[[:space:]]the[[:space:]]Task_[[:space:]]Artificially[[:space:]]Discover[[:space:]]Colour[[:space:]]Naming[[:space:]]via[[:space:]]Colour[[:space:]]Quantisation[[:space:]]Transformer/6ba16a2a-a68b-450e-b345-9cf1c85165bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Narrator_[[:space:]]Towards[[:space:]]Natural[[:space:]]Control[[:space:]]of[[:space:]]Human-Scene[[:space:]]Interaction[[:space:]]Generation[[:space:]]via[[:space:]]Relationship[[:space:]]Reasoning/69436d15-f3df-44cd-955a-d687944123a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NaviNeRF_[[:space:]]NeRF-based[[:space:]]3D[[:space:]]Representation[[:space:]]Disentanglement[[:space:]]by[[:space:]]Latent[[:space:]]Semantic[[:space:]]Navigation/480e8bc4-5d43-421f-922a-baa2c4335eba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Navigating[[:space:]]to[[:space:]]Objects[[:space:]]Specified[[:space:]]by[[:space:]]Images/c3c478b7-05f2-4336-a1ce-35a2126ecb3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeILF++_[[:space:]]Inter-Reflectable[[:space:]]Light[[:space:]]Fields[[:space:]]for[[:space:]]Geometry[[:space:]]and[[:space:]]Material[[:space:]]Estimation/518abd7a-4c07-4e3d-92e1-78d0e45e2e44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeMF_[[:space:]]Inverse[[:space:]]Volume[[:space:]]Rendering[[:space:]]with[[:space:]]Neural[[:space:]]Microflake[[:space:]]Field/cca5e364-8f87-4845-aff6-9b1f10878055_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeO[[:space:]]360_[[:space:]]Neural[[:space:]]Fields[[:space:]]for[[:space:]]Sparse[[:space:]]View[[:space:]]Synthesis[[:space:]]of[[:space:]]Outdoor[[:space:]]Scenes/c18e454a-1d55-4386-8c69-6fa17a564816_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRF-Det_[[:space:]]Learning[[:space:]]Geometry-Aware[[:space:]]Volumetric[[:space:]]Representation[[:space:]]for[[:space:]]Multi-View[[:space:]]3D[[:space:]]Object[[:space:]]Detection/51c1f762-eba3-40ed-96a4-803e25d05a50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRF-LOAM_[[:space:]]Neural[[:space:]]Implicit[[:space:]]Representation[[:space:]]for[[:space:]]Large-Scale[[:space:]]Incremental[[:space:]]LiDAR[[:space:]]Odometry[[:space:]]and[[:space:]]Mapping/afc2f682-c3b9-40b2-a470-bc37ba34d191_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRF-MS_[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]with[[:space:]]Multi-Sequence/b7f8dde2-21b6-4f3b-8a0f-2928bb420c48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeRFrac_[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]through[[:space:]]Refractive[[:space:]]Surface/e0ec9f48-8e68-49f5-97a0-89b8c82d27b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeSS-ST_[[:space:]]Detecting[[:space:]]Good[[:space:]]and[[:space:]]Stable[[:space:]]Keypoints[[:space:]]with[[:space:]]a[[:space:]]Neural[[:space:]]Stability[[:space:]]Score[[:space:]]and[[:space:]]the[[:space:]]Shi-Tomasi[[:space:]]detector/23c2cb6c-a1e7-4a96-a394-6da1c492cd32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeTO_Neural[[:space:]]Reconstruction[[:space:]]of[[:space:]]Transparent[[:space:]]Objects[[:space:]]with[[:space:]]Self-Occlusion[[:space:]]Aware[[:space:]]Refraction-Tracing/5f754cdf-cfbe-4687-9cda-fdf9df976dc4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Nearest[[:space:]]Neighbor[[:space:]]Guidance[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection/92c25a69-e16b-40e1-91c8-97f41fc73445_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neglected[[:space:]]Free[[:space:]]Lunch[[:space:]]-[[:space:]]Learning[[:space:]]Image[[:space:]]Classifiers[[:space:]]Using[[:space:]]Annotation[[:space:]]Byproducts/662d0dba-c5aa-4a24-a88d-04f2a8b184c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NerfAcc_[[:space:]]Efficient[[:space:]]Sampling[[:space:]]Accelerates[[:space:]]NeRFs/85cb71f2-b32e-4d24-8a6d-79a9ec4f289f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Nerfbusters_[[:space:]]Removing[[:space:]]Ghostly[[:space:]]Artifacts[[:space:]]from[[:space:]]Casually[[:space:]]Captured[[:space:]]NeRFs/f4ea3c93-b06d-42b4-93d9-197fd1b7276d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeuRBF_[[:space:]]A[[:space:]]Neural[[:space:]]Fields[[:space:]]Representation[[:space:]]with[[:space:]]Adaptive[[:space:]]Radial[[:space:]]Basis[[:space:]]Functions/d96e11d2-19c5-4caf-bb6e-9670b5643682_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/NeuS2_[[:space:]]Fast[[:space:]]Learning[[:space:]]of[[:space:]]Neural[[:space:]]Implicit[[:space:]]Surfaces[[:space:]]for[[:space:]]Multi-view[[:space:]]Reconstruction/dea24b1e-2cd2-44dc-a0ee-869b942b504f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Characteristic[[:space:]]Function[[:space:]]Learning[[:space:]]for[[:space:]]Conditional[[:space:]]Image[[:space:]]Generation/74115898-bf74-43bb-ab26-eaa0248eb167_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Collage[[:space:]]Transfer_[[:space:]]Artistic[[:space:]]Reconstruction[[:space:]]via[[:space:]]Material[[:space:]]Manipulation/1aeba6e6-6d34-4f0e-8760-c49d417d1f8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Deformable[[:space:]]Models[[:space:]]for[[:space:]]3D[[:space:]]Bi-Ventricular[[:space:]]Heart[[:space:]]Shape[[:space:]]Reconstruction[[:space:]]and[[:space:]]Modeling[[:space:]]from[[:space:]]2D[[:space:]]Sparse[[:space:]]Cardiac[[:space:]]Magnetic[[:space:]]Resonance[[:space:]]Imaging/1292cfc8-0b91-4d7d-874c-2796f0188572_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Fields[[:space:]]for[[:space:]]Structured[[:space:]]Lighting/1749d8c1-e2a6-4f45-b5ed-31a3738e7d81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Haircut_[[:space:]]Prior-Guided[[:space:]]Strand-Based[[:space:]]Hair[[:space:]]Reconstruction/76b0f596-6cf5-4c66-ad43-bb0e9448ea86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Implicit[[:space:]]Surface[[:space:]]Evolution/7c4f31a8-a8bf-43b7-bfdb-ad3bb2700655_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Interactive[[:space:]]Keypoint[[:space:]]Detection/a7bd8b1b-6c0b-4cc0-93e0-b6b9efbc30a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]LiDAR[[:space:]]Fields[[:space:]]for[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/e68f42ac-9189-436c-b48c-749a9601565a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Microfacet[[:space:]]Fields[[:space:]]for[[:space:]]Inverse[[:space:]]Rendering/1762f104-412e-45db-81b7-76e05a6ae1c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Radiance[[:space:]]Field[[:space:]]with[[:space:]]LiDAR[[:space:]]maps/513631a6-b4ad-485d-9873-f4dd74c6a41b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Reconstruction[[:space:]]of[[:space:]]Relightable[[:space:]]Human[[:space:]]Model[[:space:]]from[[:space:]]Monocular[[:space:]]Video/f59cab9f-801c-4947-88bd-e8ce96e06d31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural[[:space:]]Video[[:space:]]Depth[[:space:]]Stabilizer/46ace2e3-533b-4af9-96b1-fd372dfe7d9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Neural-PBIR[[:space:]]Reconstruction[[:space:]]of[[:space:]]Shape,[[:space:]]Material,[[:space:]]and[[:space:]]Illumination/8184b355-5247-4a7d-b1e4-53d069607e88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/No[[:space:]]Fear[[:space:]]of[[:space:]]Classifier[[:space:]]Biases_[[:space:]]Neural[[:space:]]Collapse[[:space:]]Inspired[[:space:]]Federated[[:space:]]Learning[[:space:]]with[[:space:]]Synthetic[[:space:]]and[[:space:]]Fixed[[:space:]]Classifier/3ea28fb3-ffeb-4f76-a562-a40990180347_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Noise-Aware[[:space:]]Learning[[:space:]]from[[:space:]]Web-Crawled[[:space:]]Image-Text[[:space:]]Data[[:space:]]for[[:space:]]Image[[:space:]]Captioning/f9124a0d-b9fd-44d8-b4ee-be5f5ae464b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Noise2Info_[[:space:]]Noisy[[:space:]]Image[[:space:]]to[[:space:]]Information[[:space:]]of[[:space:]]Noise[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Image[[:space:]]Denoising/e57882b2-ff0a-4ce1-b179-2473f32f25c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Non-Coaxial[[:space:]]Event-Guided[[:space:]]Motion[[:space:]]Deblurring[[:space:]]with[[:space:]]Spatial[[:space:]]Alignment/9713b2e6-718f-46ab-b2cc-ada5e12e739c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Non-Semantics[[:space:]]Suppressed[[:space:]]Mask[[:space:]]Learning[[:space:]]for[[:space:]]Unsupervised[[:space:]]Video[[:space:]]Semantic[[:space:]]Compression/2e62575d-946c-441d-bc00-a1889eede50f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Nonrigid[[:space:]]Object[[:space:]]Contact[[:space:]]Estimation[[:space:]]With[[:space:]]Regional[[:space:]]Unwrapping[[:space:]]Transformer/1e33e6df-e4c8-431e-81c5-be2761008f3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Normalizing[[:space:]]Flows[[:space:]]for[[:space:]]Human[[:space:]]Pose[[:space:]]Anomaly[[:space:]]Detection/e81cf676-3dd7-4dc1-b887-260c659d44b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Not[[:space:]]All[[:space:]]Features[[:space:]]Matter_[[:space:]]Enhancing[[:space:]]Few-shot[[:space:]]CLIP[[:space:]]with[[:space:]]Adaptive[[:space:]]Prior[[:space:]]Refinement/3a4b3675-1cf6-474a-a26b-44c2b8efa528_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Not[[:space:]]All[[:space:]]Steps[[:space:]]are[[:space:]]Created[[:space:]]Equal_[[:space:]]Selective[[:space:]]Diffusion[[:space:]]Distillation[[:space:]]for[[:space:]]Image[[:space:]]Manipulation/c840435c-9a23-4d70-a3ff-f1fbb88b2e20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Not[[:space:]]Every[[:space:]]Side[[:space:]]Is[[:space:]]Equal_[[:space:]]Localization[[:space:]]Uncertainty[[:space:]]Estimation[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]3D[[:space:]]Object[[:space:]]Detection/13533891-e7ee-429a-9b79-234bdf016b85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Novel[[:space:]]Scenes[[:space:]]&[[:space:]]Classes_[[:space:]]Towards[[:space:]]Adaptive[[:space:]]Open-set[[:space:]]Object[[:space:]]Detection/d22cf2db-f1f1-4457-a16d-27579339fca6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Novel-View[[:space:]]Synthesis[[:space:]]and[[:space:]]Pose[[:space:]]Estimation[[:space:]]for[[:space:]]Hand-Object[[:space:]]Interaction[[:space:]]from[[:space:]]Sparse[[:space:]]Views/af37334b-df56-4359-805c-6e9273d598f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OCHID-Fi_[[:space:]]Occlusion-Robust[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation[[:space:]]in[[:space:]]3D[[:space:]]via[[:space:]]RF-Vision/57a51a94-eed7-412b-8305-093d8d78e4c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OFVL-MS_[[:space:]]Once[[:space:]]for[[:space:]]Visual[[:space:]]Localization[[:space:]]across[[:space:]]Multiple[[:space:]]Indoor[[:space:]]Scenes/59a9d8c1-fae3-438a-a6a9-e71f95058932_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OPERA_[[:space:]]Omni-Supervised[[:space:]]Representation[[:space:]]Learning[[:space:]]with[[:space:]]Hierarchical[[:space:]]Supervisions/14467c05-b629-4214-ba43-6a158c2e7e64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ORC_[[:space:]]Network[[:space:]]Group-based[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]using[[:space:]]Online[[:space:]]Role[[:space:]]Change/af3f44c8-5dc9-4ec4-80c3-560b605e9a74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Object[[:space:]]as[[:space:]]Query_[[:space:]]Lifting[[:space:]]Any[[:space:]]2D[[:space:]]Object[[:space:]]Detector[[:space:]]to[[:space:]]3D[[:space:]]Detection/55334971-0a86-4dc4-9a3c-1a9d5f74cd48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Object-Centric[[:space:]]Multiple[[:space:]]Object[[:space:]]Tracking/52df5362-9779-43c6-85e8-09c9401c623a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Object-aware[[:space:]]Gaze[[:space:]]Target[[:space:]]Detection/1b06fb37-2f19-4468-8298-3e4001ccb57b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ObjectFusion_[[:space:]]Multi-modal[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Object-Centric[[:space:]]Fusion/8de5d07f-c75a-4258-a9ff-ccb492571b06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ObjectSDF++_[[:space:]]Improved[[:space:]]Object-Compositional[[:space:]]Neural[[:space:]]Implicit[[:space:]]Surfaces/11e871fb-41ee-4aeb-b4c9-e91fe6564f19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Objects[[:space:]]Do[[:space:]]Not[[:space:]]Disappear_[[:space:]]Video[[:space:]]Object[[:space:]]Detection[[:space:]]by[[:space:]]Single-Frame[[:space:]]Object[[:space:]]Location[[:space:]]Anticipation/d1b31db7-4ab8-4b87-b637-ced95386b76f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OccFormer_[[:space:]]Dual-path[[:space:]]Transformer[[:space:]]for[[:space:]]Vision-based[[:space:]]3D[[:space:]]Semantic[[:space:]]Occupancy[[:space:]]Prediction/2544c5a9-a339-4037-a0c2-1dcce09d33f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Occ^2Net_[[:space:]]Robust[[:space:]]Image[[:space:]]Matching[[:space:]]Based[[:space:]]on[[:space:]]3D[[:space:]]Occupancy[[:space:]]Estimation[[:space:]]for[[:space:]]Occluded[[:space:]]Regions/66b3ab8b-115b-4a17-84ad-1e877343c1f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OmniLabel_[[:space:]]A[[:space:]]Challenging[[:space:]]Benchmark[[:space:]]for[[:space:]]Language-Based[[:space:]]Object[[:space:]]Detection/4b620b4b-80bd-4be7-9866-7593c35f96c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OmniZoomer_[[:space:]]Learning[[:space:]]to[[:space:]]Move[[:space:]]and[[:space:]]Zoom[[:space:]]in[[:space:]]on[[:space:]]Sphere[[:space:]]at[[:space:]]High-Resolution/395ae354-9af3-47ed-9602-d17a43166bd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Omnidirectional[[:space:]]Information[[:space:]]Gathering[[:space:]]for[[:space:]]Knowledge[[:space:]]Transfer-Based[[:space:]]Audio-Visual[[:space:]]Navigation/72dc1d1f-5280-4d9c-8818-e770a82a0024_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OmnimatteRF_[[:space:]]Robust[[:space:]]Omnimatte[[:space:]]with[[:space:]]3D[[:space:]]Background[[:space:]]Modeling/73afa19b-8029-4f53-ba4c-2298117fdb17_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]the[[:space:]]Audio-visual[[:space:]]Synchronization[[:space:]]for[[:space:]]Lip-to-Speech[[:space:]]Synthesis/55814ca0-2c20-4e6c-82a1-68a9be89cf81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]the[[:space:]]Effectiveness[[:space:]]of[[:space:]]Spectral[[:space:]]Discriminators[[:space:]]for[[:space:]]Perceptual[[:space:]]Quality[[:space:]]Improvement/4492cd99-75ed-4482-831c-b68b283af2af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]the[[:space:]]Robustness[[:space:]]of[[:space:]]Normalizing[[:space:]]Flows[[:space:]]for[[:space:]]Inverse[[:space:]]Problems[[:space:]]in[[:space:]]Imaging/98bdfe88-2941-4706-9bbc-dffb05bfc199_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/On[[:space:]]the[[:space:]]Robustness[[:space:]]of[[:space:]]Open-World[[:space:]]Test-Time[[:space:]]Training_[[:space:]]Self-Training[[:space:]]with[[:space:]]Dynamic[[:space:]]Prototype[[:space:]]Expansion/7d370e19-8314-43bf-9f3d-9ad2468edb5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Once[[:space:]]Detected,[[:space:]]Never[[:space:]]Lost_[[:space:]]Surpassing[[:space:]]Human[[:space:]]Performance[[:space:]]in[[:space:]]Offline[[:space:]]LiDAR[[:space:]]based[[:space:]]3D[[:space:]]Object[[:space:]]Detection/e02f67a5-87a9-4f76-b65a-2aadafce2c5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/One-Shot[[:space:]]Generative[[:space:]]Domain[[:space:]]Adaptation/a58b11fa-27a9-4584-b5c9-20f3e603ce99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/One-Shot[[:space:]]Recognition[[:space:]]of[[:space:]]Any[[:space:]]Material[[:space:]]Anywhere[[:space:]]Using[[:space:]]Contrastive[[:space:]]Learning[[:space:]]with[[:space:]]Physics-Based[[:space:]]Rendering/bfc242ee-5ea8-47b5-a188-de22e298e934_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/One-bit[[:space:]]Flip[[:space:]]is[[:space:]]All[[:space:]]You[[:space:]]Need_[[:space:]]When[[:space:]]Bit-flip[[:space:]]Attack[[:space:]]Meets[[:space:]]Model[[:space:]]Training/b98e0a61-e3d5-432c-97ad-a0eed8042df9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/One-shot[[:space:]]Implicit[[:space:]]Animatable[[:space:]]Avatars[[:space:]]with[[:space:]]Model-based[[:space:]]Priors/59cd4d9a-8f3c-4ceb-ad47-662e64ebe722_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Online[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning[[:space:]]on[[:space:]]Stochastic[[:space:]]Blurry[[:space:]]Task[[:space:]]Boundary[[:space:]]via[[:space:]]Mask[[:space:]]and[[:space:]]Visual[[:space:]]Prompt[[:space:]]Tuning/5b6381f8-7969-468c-a0be-114a72aa5c95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Online[[:space:]]Clustered[[:space:]]Codebook/97aa8f04-ba80-4c75-b17f-b9917eb745b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Online[[:space:]]Continual[[:space:]]Learning[[:space:]]on[[:space:]]Hierarchical[[:space:]]Label[[:space:]]Expansion/7b601a26-a1c7-418e-80e3-4d45f44b9d9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Online[[:space:]]Prototype[[:space:]]Learning[[:space:]]for[[:space:]]Online[[:space:]]Continual[[:space:]]Learning/9c82647c-1378-40ec-abdd-a71a155b5d29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OnlineRefer_[[:space:]]A[[:space:]]Simple[[:space:]]Online[[:space:]]Baseline[[:space:]]for[[:space:]]Referring[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/e4080b34-18b5-4538-a3e5-55070f5c89d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open[[:space:]]Set[[:space:]]Video[[:space:]]HOI[[:space:]]detection[[:space:]]from[[:space:]]Action-Centric[[:space:]]Chain-of-Look[[:space:]]Prompting/1b995da1-8e64-4e68-bda4-3001dda8a1d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-Vocabulary[[:space:]]Object[[:space:]]Detection[[:space:]]With[[:space:]]an[[:space:]]Open[[:space:]]Corpus/321d851b-9fec-47b6-91a3-7725374bf68d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Decoupled[[:space:]]One-Pass[[:space:]]Network/935a75a8-8df9-433e-97ad-cae6f8c6a8ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-domain[[:space:]]Visual[[:space:]]Entity[[:space:]]Recognition_[[:space:]]Towards[[:space:]]Recognizing[[:space:]]Millions[[:space:]]of[[:space:]]Wikipedia[[:space:]]Entities/339e005c-7008-4971-a1d5-4f81c39926be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-vocabulary[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]with[[:space:]]Embedding[[:space:]]Modulation/a7b3716b-5a91-4290-bfbd-d5e861677703_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Open-vocabulary[[:space:]]Video[[:space:]]Question[[:space:]]Answering_[[:space:]]A[[:space:]]New[[:space:]]Benchmark[[:space:]]for[[:space:]]Evaluating[[:space:]]the[[:space:]]Generalizability[[:space:]]of[[:space:]]Video[[:space:]]Question[[:space:]]Answering[[:space:]]Models/0c65c038-e204-4ace-9d2b-94e148c0b50c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OpenOccupancy_[[:space:]]A[[:space:]]Large[[:space:]]Scale[[:space:]]Benchmark[[:space:]]for[[:space:]]Surrounding[[:space:]]Semantic[[:space:]]Occupancy[[:space:]]Perception/9f3cbc21-4256-45a8-a124-5b104ccb60e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Optimizing[[:space:]]the[[:space:]]Placement[[:space:]]of[[:space:]]Roadside[[:space:]]LiDARs[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/e0d84c45-f07a-466c-b646-adaf94b12d33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ord2Seq_[[:space:]]Regarding[[:space:]]Ordinal[[:space:]]Regression[[:space:]]as[[:space:]]Label[[:space:]]Sequence[[:space:]]Prediction/d70da33a-7007-435b-81cf-8fcf72137e8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Order-Prompted[[:space:]]Tag[[:space:]]Sequence[[:space:]]Generation[[:space:]]for[[:space:]]Video[[:space:]]Tagging/2a21dd88-2d6f-4b4c-84c1-401543849924_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Order-preserving[[:space:]]Consistency[[:space:]]Regularization[[:space:]]for[[:space:]]Domain[[:space:]]Adaptation[[:space:]]and[[:space:]]Generalization/4fca9295-e5a4-43b6-a254-a392f0cb5721_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ordered[[:space:]]Atomic[[:space:]]Activity[[:space:]]for[[:space:]]Fine-grained[[:space:]]Interactive[[:space:]]Traffic[[:space:]]Scenario[[:space:]]Understanding/83d189b4-e0d5-4349-b324-433089abf30f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ordinal[[:space:]]Label[[:space:]]Distribution[[:space:]]Learning/c576d048-05c9-4065-bb52-15e57b5b2cbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OrthoPlanes_[[:space:]]A[[:space:]]Novel[[:space:]]Representation[[:space:]]for[[:space:]]Better[[:space:]]3D-Awareness[[:space:]]of[[:space:]]GANs/5beec214-7387-46a7-aa44-2c627b30130e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Out-of-Distribution[[:space:]]Detection[[:space:]]for[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/bb2807e5-08eb-4174-9a0d-76fdb821a0f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Out-of-Domain[[:space:]]GAN[[:space:]]Inversion[[:space:]]via[[:space:]]Invertibility[[:space:]]Decomposition[[:space:]]for[[:space:]]Photo-Realistic[[:space:]]Human[[:space:]]Face[[:space:]]Manipulation/182191b0-3bd4-4954-91f0-fc93d4d1bd45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Overcoming[[:space:]]Forgetting[[:space:]]Catastrophe[[:space:]]in[[:space:]]Quantization-Aware[[:space:]]Training/4d54eca7-8899-44e0-a4fd-f339a5ad51f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Overwriting[[:space:]]Pretrained[[:space:]]Bias[[:space:]]with[[:space:]]Finetuning[[:space:]]Data/d11febae-f1b9-445e-882d-a6d29abb4ac2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/OxfordTVG-HIC_[[:space:]]Can[[:space:]]Machine[[:space:]]Make[[:space:]]Humorous[[:space:]]Captions[[:space:]]from[[:space:]]Images_/186d0fc0-9204-4e7d-b394-8a550b5c7598_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/P1AC_[[:space:]]Revisiting[[:space:]]Absolute[[:space:]]Pose[[:space:]]From[[:space:]]a[[:space:]]Single[[:space:]]Affine[[:space:]]Correspondence/bde22a78-524e-4130-87fe-0098240a6601_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/P2C_[[:space:]]Self-Supervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion[[:space:]]from[[:space:]]Single[[:space:]]Partial[[:space:]]Clouds/7b935fff-6eda-4b0c-a92e-18fd51db41e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PADCLIP_[[:space:]]Pseudo-labeling[[:space:]]with[[:space:]]Adaptive[[:space:]]Debiasing[[:space:]]in[[:space:]]CLIP[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/57c5105d-6667-446a-930c-dfdf57f38638_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PADDLES_[[:space:]]Phase-Amplitude[[:space:]]Spectrum[[:space:]]Disentangled[[:space:]]Early[[:space:]]Stopping[[:space:]]for[[:space:]]Learning[[:space:]]with[[:space:]]Noisy[[:space:]]Labels/314a8ccb-681e-4a92-95f6-a8daae189809_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PARF_[[:space:]]Primitive-Aware[[:space:]]Radiance[[:space:]]Fusion[[:space:]]for[[:space:]]Indoor[[:space:]]Scene[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/60f7157e-d30f-4d45-be34-f4f1a5b8d769_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PARIS_[[:space:]]Part-level[[:space:]]Reconstruction[[:space:]]and[[:space:]]Motion[[:space:]]Analysis[[:space:]]for[[:space:]]Articulated[[:space:]]Objects/fb2ea5d6-5a22-410d-9a53-22e68018fe6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PARTNER_[[:space:]]Level[[:space:]]up[[:space:]]the[[:space:]]Polar[[:space:]]Representation[[:space:]]for[[:space:]]LiDAR[[:space:]]3D[[:space:]]Object[[:space:]]Detection/6ecf1ef6-90de-413d-9564-82dbf7f77f76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PASTA_[[:space:]]Proportional[[:space:]]Amplitude[[:space:]]Spectrum[[:space:]]Training[[:space:]]Augmentation[[:space:]]for[[:space:]]Syn-to-Real[[:space:]]Domain[[:space:]]Generalization/69f1a5f0-02f5-4b9a-ba6d-5cd4e1f70f40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PATMAT_[[:space:]]Person[[:space:]]Aware[[:space:]]Tuning[[:space:]]of[[:space:]]Mask-Aware[[:space:]]Transformer[[:space:]]for[[:space:]]Face[[:space:]]Inpainting/4646e369-c2e3-4cdb-91c4-15ee0f650618_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PC-Adapter_[[:space:]]Topology-Aware[[:space:]]Adapter[[:space:]]for[[:space:]]Efficient[[:space:]]Domain[[:space:]]Adaption[[:space:]]on[[:space:]]Point[[:space:]]Clouds[[:space:]]with[[:space:]]Rectified[[:space:]]Pseudo-label/9525778c-2637-4cbd-b71b-ea227b19c64d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PDiscoNet_[[:space:]]Semantically[[:space:]]consistent[[:space:]]part[[:space:]]discovery[[:space:]]for[[:space:]]fine-grained[[:space:]]recognition/1791f9bf-943d-4903-b8b4-d0a4bebdd6db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PEANUT_[[:space:]]Predicting[[:space:]]and[[:space:]]Navigating[[:space:]]to[[:space:]]Unseen[[:space:]]Targets/a1d7d4a8-cb4a-421b-89ad-31c0f8b7bc40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PETRv2_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]3D[[:space:]]Perception[[:space:]]from[[:space:]]Multi-Camera[[:space:]]Images/99614b3f-10db-4f4b-973c-58173b63b244_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PG-RCNN_[[:space:]]Semantic[[:space:]]Surface[[:space:]]Point[[:space:]]Generation[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/df5114db-c6da-4f05-bdd7-6a3181b11891_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PGFed_[[:space:]]Personalize[[:space:]]Each[[:space:]]Client's[[:space:]]Global[[:space:]]Objective[[:space:]]for[[:space:]]Federated[[:space:]]Learning/1e26cbc2-88c2-42bf-b6fc-90724a2fc315_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PHRIT_[[:space:]]Parametric[[:space:]]Hand[[:space:]]Representation[[:space:]]with[[:space:]]Implicit[[:space:]]Template/0d7b602d-7fb0-40b6-8352-5b5ba5dbfc73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PIDRo_[[:space:]]Parallel[[:space:]]Isomeric[[:space:]]Attention[[:space:]]with[[:space:]]Dynamic[[:space:]]Routing[[:space:]]for[[:space:]]Text-Video[[:space:]]Retrieval/d93eb355-3c9e-416a-887e-bdbdcc2c5831_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PIRNet_[[:space:]]Privacy-Preserving[[:space:]]Image[[:space:]]Restoration[[:space:]]Network[[:space:]]via[[:space:]]Wavelet[[:space:]]Lifting/687a6cf9-174d-46bc-9d95-cfda0b7f897d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PNI[[:space:]]_[[:space:]]Industrial[[:space:]]Anomaly[[:space:]]Detection[[:space:]]using[[:space:]]Position[[:space:]]and[[:space:]]Neighborhood[[:space:]]Information/3bbfce7a-52f6-4955-845a-a8271fd7f50d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PODA_[[:space:]]Prompt-driven[[:space:]]Zero-shot[[:space:]]Domain[[:space:]]Adaptation/73cf0969-dd90-4fc2-9829-bafbc5594f67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PODIA-3D_[[:space:]]Domain[[:space:]]Adaptation[[:space:]]of[[:space:]]3D[[:space:]]Generative[[:space:]]Model[[:space:]]Across[[:space:]]Large[[:space:]]Domain[[:space:]]Gap[[:space:]]Using[[:space:]]Pose-Preserved[[:space:]]Text-to-Image[[:space:]]Diffusion/01ed36c2-4d8d-4afd-a409-339f798b92e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PPR_[[:space:]]Physically[[:space:]]Plausible[[:space:]]Reconstruction[[:space:]]from[[:space:]]Monocular[[:space:]]Videos/00676a7a-51d9-45c2-8fbf-8117e7a5c7b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PRANC_[[:space:]]Pseudo[[:space:]]RAndom[[:space:]]Networks[[:space:]]for[[:space:]]Compacting[[:space:]]Deep[[:space:]]Models/4c05b56e-79d8-4cba-8901-7fd9fdc74bb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PRIOR_[[:space:]]Prototype[[:space:]]Representation[[:space:]]Joint[[:space:]]Learning[[:space:]]from[[:space:]]Medical[[:space:]]Images[[:space:]]and[[:space:]]Reports/ecda55b5-f6f1-4023-a61b-1ed74bcc857f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PVT++_[[:space:]]A[[:space:]]Simple[[:space:]]End-to-End[[:space:]]Latency-Aware[[:space:]]Visual[[:space:]]Tracking[[:space:]]Framework/97e033c1-6f76-4037-a6e0-fa18d0769030_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pairwise[[:space:]]Similarity[[:space:]]Learning[[:space:]]is[[:space:]]SimPLE/c214a7a7-c074-453b-8932-0d1f60cd1d51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PanFlowNet_[[:space:]]A[[:space:]]Flow-Based[[:space:]]Deep[[:space:]]Network[[:space:]]for[[:space:]]Pan-Sharpening/46953505-9c7c-4405-9302-1e4560cd5832_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Panoramas[[:space:]]from[[:space:]]Photons/f91257c3-0d84-4b94-a14d-7d6d84a283fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ParCNetV2_[[:space:]]Oversized[[:space:]]Kernel[[:space:]]with[[:space:]]Enhanced[[:space:]]Attention/967e9ea9-d8e7-457c-9a5d-e6c0f4dc68bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Parallax-Tolerant[[:space:]]Unsupervised[[:space:]]Deep[[:space:]]Image[[:space:]]Stitching/916b9fbd-8249-4c9c-93f4-e964d5b5a055_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Parallel[[:space:]]Attention[[:space:]]Interaction[[:space:]]Network[[:space:]]for[[:space:]]Few-Shot[[:space:]]Skeleton-Based[[:space:]]Action[[:space:]]Recognition/5fd8427a-d412-4f2d-8f55-fa3053f5935c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Parameterized[[:space:]]Cost[[:space:]]Volume[[:space:]]for[[:space:]]Stereo[[:space:]]Matching/f35bf9b3-fa1f-4819-96f6-a125ab3b774c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Parametric[[:space:]]Classification[[:space:]]for[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery_[[:space:]]A[[:space:]]Baseline[[:space:]]Study/9c0acd2f-f9ec-4657-a334-2d34ae518a34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Parametric[[:space:]]Depth[[:space:]]Based[[:space:]]Feature[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]and[[:space:]]Segmentation[[:space:]]in[[:space:]]Bird's-Eye[[:space:]]View/618ff43c-c49d-453d-a8e3-86ff140c77c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Parametric[[:space:]]Information[[:space:]]Maximization[[:space:]]for[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/0e1f6152-d15b-424d-9d6c-be9058aa2c45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Part-Aware[[:space:]]Transformer[[:space:]]for[[:space:]]Generalizable[[:space:]]Person[[:space:]]Re-identification/c6b01a9c-8789-4744-a2eb-2050f6c4d495_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Partition[[:space:]]Speeds[[:space:]]Up[[:space:]]Learning[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations[[:space:]]Based[[:space:]]on[[:space:]]Exponential-Increase[[:space:]]Hypothesis/aec7273e-a669-472a-8108-677f73809150_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Partition-And-Debias_[[:space:]]Agnostic[[:space:]]Biases[[:space:]]Mitigation[[:space:]]via[[:space:]]a[[:space:]]Mixture[[:space:]]of[[:space:]]Biases-Specific[[:space:]]Experts/8e505f00-4b39-4539-8b88-c36ec4ebea95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Passive[[:space:]]Ultra-Wideband[[:space:]]Single-Photon[[:space:]]Imaging/1b90d0de-2796-425d-97aa-73d3d01e8571_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PatchCT_[[:space:]]Aligning[[:space:]]Patch[[:space:]]Set[[:space:]]and[[:space:]]Label[[:space:]]Set[[:space:]]with[[:space:]]Conditional[[:space:]]Transport[[:space:]]for[[:space:]]Multi-Label[[:space:]]Image[[:space:]]Classification/492d51d9-64d8-4072-b631-4697ef5f2a28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Perceptual[[:space:]]Artifacts[[:space:]]Localization[[:space:]]for[[:space:]]Image[[:space:]]Synthesis[[:space:]]Tasks/8df34ed3-a38d-499c-94fc-6b5baef16728_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Perceptual[[:space:]]Grouping[[:space:]]in[[:space:]]Contrastive[[:space:]]Vision-Language[[:space:]]Models/0dc8422d-9ba0-41db-ad92-5f8691431d2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Periodically[[:space:]]Exchange[[:space:]]Teacher-Student[[:space:]]for[[:space:]]Source-Free[[:space:]]Object[[:space:]]Detection/ef588e47-8690-42ef-831a-63bb1e86bb36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Perpetual[[:space:]]Humanoid[[:space:]]Control[[:space:]]for[[:space:]]Real-time[[:space:]]Simulated[[:space:]]Avatars/af405af3-4457-47e7-9031-48415199c03b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Persistent-Transient[[:space:]]Duality_[[:space:]]A[[:space:]]Multi-Mechanism[[:space:]]Approach[[:space:]]for[[:space:]]Modeling[[:space:]]Human-Object[[:space:]]Interaction/bf8cb248-77f6-40bf-a652-5fed7366ac2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Person[[:space:]]Re-Identification[[:space:]]without[[:space:]]Identification[[:space:]]via[[:space:]]Event[[:space:]]anonymization/0d7a7e60-df95-4c26-9c1a-5edb5b3c1d80_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Personalized[[:space:]]Image[[:space:]]Generation[[:space:]]for[[:space:]]Color[[:space:]]Vision[[:space:]]Deficiency[[:space:]]Population/6d09ee2a-48ca-40d9-be63-ee6c24f51a3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Personalized[[:space:]]Semantics[[:space:]]Excitation[[:space:]]for[[:space:]]Federated[[:space:]]Image[[:space:]]Classification/23ce7f93-3be1-47e1-8169-833805630e8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PhaseMP_[[:space:]]Robust[[:space:]]3D[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Phase-conditioned[[:space:]]Human[[:space:]]Motion[[:space:]]Prior/1d831345-372a-4efe-ab08-b294d5674f04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Phasic[[:space:]]Content[[:space:]]Fusing[[:space:]]Diffusion[[:space:]]Model[[:space:]]with[[:space:]]Directional[[:space:]]Distribution[[:space:]]Consistency[[:space:]]for[[:space:]]Few-Shot[[:space:]]Model[[:space:]]Adaption/481fed3b-078d-49d5-9174-0732335f2e66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PhysDiff_[[:space:]]Physics-Guided[[:space:]]Human[[:space:]]Motion[[:space:]]Diffusion[[:space:]]Model/b5aef37d-dae7-4397-a21f-4174df3a3557_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Physically-Plausible[[:space:]]Illumination[[:space:]]Distribution[[:space:]]Estimation/7b6bb537-27c5-4e01-a46d-ef4b603694a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Physics-Augmented[[:space:]]Autoencoder[[:space:]]for[[:space:]]3D[[:space:]]Skeleton-Based[[:space:]]Gait[[:space:]]Recognition/d66b7750-0461-4bef-a60a-f8d3343ef2bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Physics-Driven[[:space:]]Turbulence[[:space:]]Image[[:space:]]Restoration[[:space:]]with[[:space:]]Stochastic[[:space:]]Refinement/b35636af-d631-47e0-b3b0-f6b99566fea1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PivotNet_[[:space:]]Vectorized[[:space:]]Pivot[[:space:]]Learning[[:space:]]for[[:space:]]End-to-end[[:space:]]HD[[:space:]]Map[[:space:]]Construction/e4482a7b-ca94-4dd0-8837-9b1913ad37aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pix2Video_[[:space:]]Video[[:space:]]Editing[[:space:]]using[[:space:]]Image[[:space:]]Diffusion/ec5dae6e-8fdf-4309-814e-a0799551bc3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pixel[[:space:]]Adaptive[[:space:]]Deep[[:space:]]Unfolding[[:space:]]Transformer[[:space:]]for[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Reconstruction/039e6737-5146-410d-bbd9-88cfa9d6d3f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pixel-Aligned[[:space:]]Recurrent[[:space:]]Queries[[:space:]]for[[:space:]]Multi-View[[:space:]]3D[[:space:]]Object[[:space:]]Detection/4d25f52d-4df8-46d4-be06-bc09eed27f6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pixel-Wise[[:space:]]Contrastive[[:space:]]Distillation/309b9bcd-3d01-4f8a-9e5c-d503b39ba168_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PlanarTrack_[[:space:]]A[[:space:]]Large-scale[[:space:]]Challenging[[:space:]]Benchmark[[:space:]]for[[:space:]]Planar[[:space:]]Object[[:space:]]Tracking/bd8dc73f-a6f0-4edc-ad39-375904f9911f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PlaneRecTR_[[:space:]]Unified[[:space:]]Query[[:space:]]Learning[[:space:]]for[[:space:]]3D[[:space:]]Plane[[:space:]]Recovery[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]View/c3bee3df-7aeb-4bad-99cd-cf9e60f858d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PlankAssembly_[[:space:]]Robust[[:space:]]3D[[:space:]]Reconstruction[[:space:]]from[[:space:]]Three[[:space:]]Orthographic[[:space:]]Views[[:space:]]with[[:space:]]Learnt[[:space:]]Shape[[:space:]]Programs/8e54e10f-db40-471e-b91a-5d9572827b02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Plausible[[:space:]]Uncertainties[[:space:]]for[[:space:]]Human[[:space:]]Pose[[:space:]]Regression/6054d10b-de23-4f60-9312-9597c2d957ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pluralistic[[:space:]]Aging[[:space:]]Diffusion[[:space:]]Autoencoder/3c7b90fb-6edd-47f8-8ec2-5341d502a4cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Poincare[[:space:]]ResNet/7624bd3c-72a1-420a-8498-df04edf9886d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Point[[:space:]]Contrastive[[:space:]]Prediction[[:space:]]with[[:space:]]Semantic[[:space:]]Clustering[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]on[[:space:]]Point[[:space:]]Cloud[[:space:]]Videos/82ffe7dc-9d7a-4edc-b935-623fd92d098f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Point-Query[[:space:]]Quadtree[[:space:]]for[[:space:]]Crowd[[:space:]]Counting,[[:space:]]Localization,[[:space:]]and[[:space:]]More/8a6b24c0-7bdd-472b-b42c-be46a4a22ad0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Point-SLAM_[[:space:]]Dense[[:space:]]Neural[[:space:]]Point[[:space:]]Cloud-based[[:space:]]SLAM/3b0d1f87-7626-44ea-bc6d-f8d9dc522f94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Point-TTA_[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration[[:space:]]Using[[:space:]]Multitask[[:space:]]Meta-Auxiliary[[:space:]]Learning/ece7d548-193c-4068-8a6d-ab1289010832_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Point2Mask_[[:space:]]Point-supervised[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]via[[:space:]]Optimal[[:space:]]Transport/8bc29df0-ee0a-42dd-8c19-adf69ec7682b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PointCLIP[[:space:]]V2_[[:space:]]Prompting[[:space:]]CLIP[[:space:]]and[[:space:]]GPT[[:space:]]for[[:space:]]Powerful[[:space:]]3D[[:space:]]Open-world[[:space:]]Learning/0caae07c-bb85-400e-a06d-6fd3678002d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PointDC_[[:space:]]Unsupervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]of[[:space:]]3D[[:space:]]Point[[:space:]]Clouds[[:space:]]via[[:space:]]Cross-Modal[[:space:]]Distillation[[:space:]]and[[:space:]]Super-Voxel[[:space:]]Clustering/b856c1fd-6d37-438e-94b6-fcbcce33eafd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PointMBF_[[:space:]]A[[:space:]]Multi-scale[[:space:]]Bidirectional[[:space:]]Fusion[[:space:]]Network[[:space:]]for[[:space:]]Unsupervised[[:space:]]RGB-D[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/02fcc136-0938-4745-96a3-86378f1a128b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PointOdyssey_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Synthetic[[:space:]]Dataset[[:space:]]for[[:space:]]Long-Term[[:space:]]Point[[:space:]]Tracking/b3ad7f4e-671c-4af0-9bfd-e784b8e14a49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PolicyCleanse_[[:space:]]Backdoor[[:space:]]Detection[[:space:]]and[[:space:]]Mitigation[[:space:]]for[[:space:]]Competitive[[:space:]]Reinforcement[[:space:]]Learning/ad34d0fe-6f70-4033-9eb0-d081c7f7a1ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ponder_[[:space:]]Point[[:space:]]Cloud[[:space:]]Pre-training[[:space:]]via[[:space:]]Neural[[:space:]]Rendering/430a9fbf-7de6-48ed-827f-b785403bd6c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pose-Free[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]via[[:space:]]Implicit[[:space:]]Pose[[:space:]]Regularization/58098ee9-b482-46cc-b4fa-83da0ad8343f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PoseDiffusion_[[:space:]]Solving[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Diffusion-aided[[:space:]]Bundle[[:space:]]Adjustment/a1f753b4-b03c-4ec9-b07f-9928b1daa1cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PoseFix_[[:space:]]Correcting[[:space:]]3D[[:space:]]Human[[:space:]]Poses[[:space:]]with[[:space:]]Natural[[:space:]]Language/65725baa-abf1-4f3c-98d9-7bc3bf984b89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PourIt!_[[:space:]]Weakly-Supervised[[:space:]]Liquid[[:space:]]Perception[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]for[[:space:]]Visual[[:space:]]Closed-Loop[[:space:]]Robotic[[:space:]]Pouring/728e9e0b-29e6-49f5-96a7-9886776507be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Practical[[:space:]]Membership[[:space:]]Inference[[:space:]]Attacks[[:space:]]Against[[:space:]]Large-Scale[[:space:]]Multi-Modal[[:space:]]Models_[[:space:]]A[[:space:]]Pilot[[:space:]]Study/2483ccb3-ea65-4217-92de-eeb8de678045_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pre-Training-Free[[:space:]]Image[[:space:]]Manipulation[[:space:]]Localization[[:space:]]through[[:space:]]Non-Mutually[[:space:]]Exclusive[[:space:]]Contrastive[[:space:]]Learning/f74e67d4-668e-477a-b970-ec07af63a23f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pre-training[[:space:]]Vision[[:space:]]Transformers[[:space:]]with[[:space:]]Very[[:space:]]Limited[[:space:]]Synthesized[[:space:]]Images/12a6d076-c92e-47fc-8b52-42b878f3f3b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PreSTU_[[:space:]]Pre-Training[[:space:]]for[[:space:]]Scene-Text[[:space:]]Understanding/c5c18218-53d8-4a71-9148-4d0d43178fa9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Predict[[:space:]]to[[:space:]]Detect_[[:space:]]Prediction-guided[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]using[[:space:]]Sequential[[:space:]]Images/fcbf30f9-f195-4f61-9083-e1dbafc732e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Preface_[[:space:]]A[[:space:]]Data-driven[[:space:]]Volumetric[[:space:]]Prior[[:space:]]for[[:space:]]Few-shot[[:space:]]Ultra[[:space:]]High-resolution[[:space:]]Face[[:space:]]Synthesis/904af390-9b66-44f1-b757-c369bcb42d27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Preparing[[:space:]]the[[:space:]]Future[[:space:]]for[[:space:]]Continual[[:space:]]Semantic[[:space:]]Segmentation/72d5919b-3823-475f-9939-cac5c29b48a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Preserve[[:space:]]Your[[:space:]]Own[[:space:]]Correlation_[[:space:]]A[[:space:]]Noise[[:space:]]Prior[[:space:]]for[[:space:]]Video[[:space:]]Diffusion[[:space:]]Models/12c6fb66-a455-4547-8b5a-4a378b5ee51f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Preserving[[:space:]]Modality[[:space:]]Structure[[:space:]]Improves[[:space:]]Multi-Modal[[:space:]]Learning/9786f414-b2fe-4822-86c7-2868d21a76fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Preserving[[:space:]]Tumor[[:space:]]Volumes[[:space:]]for[[:space:]]Unsupervised[[:space:]]Medical[[:space:]]Image[[:space:]]Registration/dbd41759-9a48-41b9-9206-f4ea354bc2a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pretrained[[:space:]]Language[[:space:]]Models[[:space:]]as[[:space:]]Visual[[:space:]]Planners[[:space:]]for[[:space:]]Human[[:space:]]Assistance/057af23b-bb0a-449d-acad-c434ed40a938_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Preventing[[:space:]]Zero-Shot[[:space:]]Transfer[[:space:]]Degradation[[:space:]]in[[:space:]]Continual[[:space:]]Learning[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models/438697ca-0d24-4663-a1e4-f11b5f684ada_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prior-guided[[:space:]]Source-free[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/67d12eb1-f1e4-4b90-aba2-4771e3bf7c39_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Priority-Centric[[:space:]]Human[[:space:]]Motion[[:space:]]Generation[[:space:]]in[[:space:]]Discrete[[:space:]]Latent[[:space:]]Space/58c7e478-f288-424e-a702-051324564633_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Privacy[[:space:]]Preserving[[:space:]]Localization[[:space:]]via[[:space:]]Coordinate[[:space:]]Permutations/f992eeb4-2e60-42f9-9e28-abc98f428b2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Privacy-Preserving[[:space:]]Face[[:space:]]Recognition[[:space:]]Using[[:space:]]Random[[:space:]]Frequency[[:space:]]Components/3da3fc63-806e-4e3e-99b2-e2f33c132bad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ProPainter_[[:space:]]Improving[[:space:]]Propagation[[:space:]]and[[:space:]]Transformer[[:space:]]for[[:space:]]Video[[:space:]]Inpainting/447df786-8aae-4bc4-b7d5-1e90caa96cc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ProbVLM_[[:space:]]Probabilistic[[:space:]]Adapter[[:space:]]for[[:space:]]Frozen[[:space:]]Vison-Language[[:space:]]Models/52935457-f9bc-4c99-866c-671f6012153b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Probabilistic[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery[[:space:]]in[[:space:]]3D[[:space:]]Scenes[[:space:]]from[[:space:]]Egocentric[[:space:]]Views/c490ef67-0a1e-4595-8f34-5035e971ee89_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Probabilistic[[:space:]]Modeling[[:space:]]of[[:space:]]Inter-[[:space:]]and[[:space:]]Intra-observer[[:space:]]Variability[[:space:]]in[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/bf4cff8f-5588-469f-97d0-102dd2317c9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Probabilistic[[:space:]]Precision[[:space:]]and[[:space:]]Recall[[:space:]]Towards[[:space:]]Reliable[[:space:]]Evaluation[[:space:]]of[[:space:]]Generative[[:space:]]Models/dd6377f8-01ac-488f-b3b2-5ae1ebd477f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Probabilistic[[:space:]]Triangulation[[:space:]]for[[:space:]]Uncalibrated[[:space:]]Multi-View[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/b6b4e04e-10fb-49dc-be18-fbc2e2b53da9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Progressive[[:space:]]Spatio-Temporal[[:space:]]Prototype[[:space:]]Matching[[:space:]]for[[:space:]]Text-Video[[:space:]]Retrieval/480eca60-b42e-4764-a2ee-88684bf2c389_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prompt[[:space:]]Switch_[[:space:]]Efficient[[:space:]]CLIP[[:space:]]Adaptation[[:space:]]for[[:space:]]Text-Video[[:space:]]Retrieval/408eb595-fd0b-42c6-8600-a2efb5916c5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prompt[[:space:]]Tuning[[:space:]]Inversion[[:space:]]for[[:space:]]Text-driven[[:space:]]Image[[:space:]]Editing[[:space:]]Using[[:space:]]Diffusion[[:space:]]Models/f6450b64-da9d-48f8-bcd6-4de4863bfc2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prompt-aligned[[:space:]]Gradient[[:space:]]for[[:space:]]Prompt[[:space:]]Tuning/7786570f-2ec2-4f85-a04c-e553c7832bc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PromptCap_[[:space:]]Prompt-Guided[[:space:]]Image[[:space:]]Captioning[[:space:]]for[[:space:]]VQA[[:space:]]with[[:space:]]GPT-3/4ffd89d6-8734-4748-beec-1543acfdd6d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/PromptStyler_[[:space:]]Prompt-driven[[:space:]]Style[[:space:]]Generation[[:space:]]for[[:space:]]Source-free[[:space:]]Domain[[:space:]]Generalization/f8d2a45d-4c84-4d88-913c-50e0d9fbe4a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ProtoFL_[[:space:]]Unsupervised[[:space:]]Federated[[:space:]]Learning[[:space:]]via[[:space:]]Prototypical[[:space:]]Distillation/c82e3c6e-4735-400a-b460-81642c25b787_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ProtoTransfer_[[:space:]]Cross-Modal[[:space:]]Prototype[[:space:]]Transfer[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation/4c052c14-482f-4c6c-bea8-5d4ef790aea8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prototype[[:space:]]Reminiscence[[:space:]]and[[:space:]]Augmented[[:space:]]Asymmetric[[:space:]]Knowledge[[:space:]]Aggregation[[:space:]]for[[:space:]]Non-Exemplar[[:space:]]Class-Incremental[[:space:]]Learning/f06e28ba-39c2-463d-9177-75781355ca62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prototype-based[[:space:]]Dataset[[:space:]]Comparison/02a62e7f-ba00-4044-92de-55f8e976a093_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prototypes-oriented[[:space:]]Transductive[[:space:]]Few-shot[[:space:]]Learning[[:space:]]with[[:space:]]Conditional[[:space:]]Transport/e3d7c832-dffc-42b7-b29a-faccf7386804_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prototypical[[:space:]]Kernel[[:space:]]Learning[[:space:]]and[[:space:]]Open-set[[:space:]]Foreground[[:space:]]Perception[[:space:]]for[[:space:]]Generalized[[:space:]]Few-shot[[:space:]]Semantic[[:space:]]Segmentation/ae5d3fad-adfb-4c23-912c-f33a2a374a40_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prototypical[[:space:]]Mixing[[:space:]]and[[:space:]]Retrieval-Based[[:space:]]Refinement[[:space:]]for[[:space:]]Label[[:space:]]Noise-Resistant[[:space:]]Image[[:space:]]Retrieval/ec71d51a-9735-4817-bbda-a4f4fe8a6bfa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Proxy[[:space:]]Anchor-based[[:space:]]Unsupervised[[:space:]]Learning[[:space:]]for[[:space:]]Continuous[[:space:]]Generalized[[:space:]]Category[[:space:]]Discovery/82f1fc34-be8f-46cf-8cc9-6e1e6fe0df48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Prune[[:space:]]Spatio-temporal[[:space:]]Tokens[[:space:]]by[[:space:]]Semantic-aware[[:space:]]Temporal[[:space:]]Accumulation/6db1efd1-b765-4bc8-8a5a-d2a425ae5e95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pseudo[[:space:]]Flow[[:space:]]Consistency[[:space:]]for[[:space:]]Self-Supervised[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/cf29abfd-328a-4679-9b7c-7340065b4fa2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pseudo-label[[:space:]]Alignment[[:space:]]for[[:space:]]Semi-supervised[[:space:]]Instance[[:space:]]Segmentation/42aab451-44d8-431c-bd3b-859749553f79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Pyramid[[:space:]]Dual[[:space:]]Domain[[:space:]]Injection[[:space:]]Network[[:space:]]for[[:space:]]Pan-sharpening/aa199a66-751b-4158-bd1f-1783173c94fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Q-Diffusion_[[:space:]]Quantizing[[:space:]]Diffusion[[:space:]]Models/64ad5dff-59b7-42a2-a4ac-ca7c9541d350_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/QD-BEV[[:space:]]_[[:space:]]Quantization-aware[[:space:]]View-guided[[:space:]]Distillation[[:space:]]for[[:space:]]Multi-view[[:space:]]3D[[:space:]]Object[[:space:]]Detection/ce2686ce-9d14-4ea8-b9f7-4395d3561b65_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Quality[[:space:]]Diversity[[:space:]]for[[:space:]]Visual[[:space:]]Pre-Training/3f69c090-f906-4d45-8423-292cfc679eb9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Quality-Agnostic[[:space:]]Deepfake[[:space:]]Detection[[:space:]]with[[:space:]]Intra-model[[:space:]]Collaborative[[:space:]]Learning/77142358-08da-4d47-9173-9c11400a028e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Query[[:space:]]Refinement[[:space:]]Transformer[[:space:]]for[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation/8435ea40-79e7-4103-b1c6-be175c97058d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Query6DoF_[[:space:]]Learning[[:space:]]Sparse[[:space:]]Queries[[:space:]]as[[:space:]]Implicit[[:space:]]Shape[[:space:]]Prior[[:space:]]for[[:space:]]Category-Level[[:space:]]6DoF[[:space:]]Pose[[:space:]]Estimation/4d1aafad-a31c-4773-aac8-62fd2260bfdc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/R-Pred_[[:space:]]Two-Stage[[:space:]]Motion[[:space:]]Prediction[[:space:]]Via[[:space:]]Tube-Query[[:space:]]Attention-Based[[:space:]]Trajectory[[:space:]]Refinement/cbd955af-2686-4f7c-843a-a3576e04bc6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/R3D3_[[:space:]]Dense[[:space:]]3D[[:space:]]Reconstruction[[:space:]]of[[:space:]]Dynamic[[:space:]]Scenes[[:space:]]from[[:space:]]Multiple[[:space:]]Cameras/df3f55bd-3906-48f7-856c-09ad31d9f51b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RANA_[[:space:]]Relightable[[:space:]]Articulated[[:space:]]Neural[[:space:]]Avatars/ae3b789d-a192-4696-b97f-daa6528a52ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RCA-NOC_[[:space:]]Relative[[:space:]]Contrastive[[:space:]]Alignment[[:space:]]for[[:space:]]Novel[[:space:]]Object[[:space:]]Captioning/cbdb5e43-66b1-4704-93d9-a919fc74c296_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/REAP_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Realistic[[:space:]]Adversarial[[:space:]]Patch[[:space:]]Benchmark/451492ac-a33d-4658-b61d-707c2258ba9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RED-PSM_[[:space:]]Regularization[[:space:]]by[[:space:]]Denoising[[:space:]]of[[:space:]]Partially[[:space:]]Separable[[:space:]]Models[[:space:]]for[[:space:]]Dynamic[[:space:]]Imaging/ffd9f980-64f4-43d0-a9f4-e03ede33c976_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RFD-ECNet_[[:space:]]Extreme[[:space:]]Underwater[[:space:]]Image[[:space:]]Compression[[:space:]]with[[:space:]]Reference[[:space:]]to[[:space:]]Feature[[:space:]]Dictionary/cdd3e31c-4266-47f7-a8c7-755cca493958_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RFLA_[[:space:]]A[[:space:]]Stealthy[[:space:]]Reflected[[:space:]]Light[[:space:]]Adversarial[[:space:]]Attack[[:space:]]in[[:space:]]the[[:space:]]Physical[[:space:]]World/bdd5612c-711f-444e-bf67-fd043897ebe9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RICO_[[:space:]]Regularizing[[:space:]]the[[:space:]]Unobservable[[:space:]]for[[:space:]]Indoor[[:space:]]Compositional[[:space:]]Reconstruction/9a248d4c-688f-46d5-8671-0975c0bef03f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RIGID_[[:space:]]Recurrent[[:space:]]GAN[[:space:]]Inversion[[:space:]]and[[:space:]]Editing[[:space:]]of[[:space:]]Real[[:space:]]Face[[:space:]]Videos/0ac9c66e-6779-4caf-9449-892ceb58d43e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RLIPv2_[[:space:]]Fast[[:space:]]Scaling[[:space:]]of[[:space:]]Relational[[:space:]]Language-Image[[:space:]]Pre-Training/dedbf300-120b-4912-a26e-f96e152258fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RLSAC_[[:space:]]Reinforcement[[:space:]]Learning[[:space:]]Enhanced[[:space:]]Sample[[:space:]]Consensus[[:space:]]for[[:space:]]End-to-End[[:space:]]Robust[[:space:]]Estimation/253f2fc2-5243-41a5-8404-4467c420f19c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RMP-Loss_[[:space:]]Regularizing[[:space:]]Membrane[[:space:]]Potential[[:space:]]Distribution[[:space:]]for[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks/8dd5399b-d34f-43a1-82f0-9e644d0103b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ROME_[[:space:]]Robustifying[[:space:]]Memory-Efficient[[:space:]]NAS[[:space:]]via[[:space:]]Topology[[:space:]]Disentanglement[[:space:]]and[[:space:]]Gradient[[:space:]]Accumulation/6ade4227-5686-42c9-830e-da85efbc5dca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RPEFlow_[[:space:]]Multimodal[[:space:]]Fusion[[:space:]]of[[:space:]]RGB-PointCloud-Event[[:space:]]for[[:space:]]Joint[[:space:]]Optical[[:space:]]Flow[[:space:]]and[[:space:]]Scene[[:space:]]Flow[[:space:]]Estimation/d027f499-898c-464c-acf6-a1eba174840a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RPG-Palm_[[:space:]]Realistic[[:space:]]Pseudo-data[[:space:]]Generation[[:space:]]for[[:space:]]Palmprint[[:space:]]Recognition/6f1b3792-7545-4d6e-8992-4404f77414a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RSFNet_[[:space:]]A[[:space:]]White-Box[[:space:]]Image[[:space:]]Retouching[[:space:]]Approach[[:space:]]using[[:space:]]Region-Specific[[:space:]]Color[[:space:]]Filters/a1a06f38-ffb3-49f3-88a3-88bb8d7af4de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Random[[:space:]]Boxes[[:space:]]Are[[:space:]]Open-world[[:space:]]Object[[:space:]]Detectors/99919d69-2dc0-454c-a4b9-63b911115db6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Random[[:space:]]Sub-Samples[[:space:]]Generation[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Real[[:space:]]Image[[:space:]]Denoising/0ac3ef7d-9ebc-4369-9f12-6e7b6e92e896_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Randomized[[:space:]]Quantization_[[:space:]]A[[:space:]]Generic[[:space:]]Augmentation[[:space:]]for[[:space:]]Data[[:space:]]Agnostic[[:space:]]Self-supervised[[:space:]]Learning/35e2ad13-e7ee-4623-a3bb-087851f2da1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RankMatch_[[:space:]]Fostering[[:space:]]Confidence[[:space:]]and[[:space:]]Consistency[[:space:]]in[[:space:]]Learning[[:space:]]with[[:space:]]Noisy[[:space:]]Labels/0d715541-80cd-458e-81f4-423a1c46d83e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RankMixup_[[:space:]]Ranking-Based[[:space:]]Mixup[[:space:]]Training[[:space:]]for[[:space:]]Network[[:space:]]Calibration/4c024ec3-bba1-4470-bd07-d0e7529be938_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rapid[[:space:]]Adaptation[[:space:]]in[[:space:]]Online[[:space:]]Continual[[:space:]]Learning_[[:space:]]Are[[:space:]]We[[:space:]]Evaluating[[:space:]]It[[:space:]]Right_/165500ab-b2c3-4db0-ba04-4570a2ad4282_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rapid[[:space:]]Network[[:space:]]Adaptation_[[:space:]]Learning[[:space:]]to[[:space:]]Adapt[[:space:]]Neural[[:space:]]Networks[[:space:]]Using[[:space:]]Test-Time[[:space:]]Feedback/32896e91-9d8e-4427-a165-f57c45c9afec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RawHDR_[[:space:]]High[[:space:]]Dynamic[[:space:]]Range[[:space:]]Image[[:space:]]Reconstruction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Raw[[:space:]]Image/fc353073-6b8a-4ffb-9c52-79bcc3452aaf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ray[[:space:]]Conditioning_[[:space:]]Trading[[:space:]]Photo-consistency[[:space:]]for[[:space:]]Photo-realism[[:space:]]in[[:space:]]Multi-view[[:space:]]Image[[:space:]]Generation/8b41c085-1da5-4856-9dcc-37e67b6f421c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RbA_[[:space:]]Segmenting[[:space:]]Unknown[[:space:]]Regions[[:space:]]Rejected[[:space:]]by[[:space:]]All/3679f972-2f48-4a98-b205-08350e1eb9c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Re-ReND_[[:space:]]Real-Time[[:space:]]Rendering[[:space:]]of[[:space:]]NeRFs[[:space:]]across[[:space:]]Devices/1bf3c862-c6bc-42e6-92ae-cf19aaa204db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Re-mine,[[:space:]]Learn[[:space:]]and[[:space:]]Reason_[[:space:]]Exploring[[:space:]]the[[:space:]]Cross-modal[[:space:]]Semantic[[:space:]]Correlations[[:space:]]for[[:space:]]Language-guided[[:space:]]HOI[[:space:]]detection/c598b516-59ad-4ec0-a52e-62dc384ae267_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ReFit_[[:space:]]Recurrent[[:space:]]Fitting[[:space:]]Network[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Recovery/b2c9efcc-2846-4559-9da6-6ddd113bec67_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ReGen_[[:space:]]A[[:space:]]good[[:space:]]Generative[[:space:]]Zero-Shot[[:space:]]Video[[:space:]]Classifier[[:space:]]Should[[:space:]]be[[:space:]]Rewarded/cc3e69d7-a9fb-4ba0-a557-ba6585418722_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ReLeaPS[[:space:]]_[[:space:]]Reinforcement[[:space:]]Learning-based[[:space:]]Illumination[[:space:]]Planning[[:space:]]for[[:space:]]Generalized[[:space:]]Photometric[[:space:]]Stereo/52829841-30ae-4d31-a0f8-9cf32450d650_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ReMoDiffuse_[[:space:]]Retrieval-Augmented[[:space:]]Motion[[:space:]]Diffusion[[:space:]]Model/c16cc50f-ca5f-4269-8a27-a7731683f961_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ReNeRF_[[:space:]]Relightable[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]with[[:space:]]Nearfield[[:space:]]Lighting/f4236930-92c8-4645-8688-589e6adc9f7d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ReST_[[:space:]]A[[:space:]]Reconfigurable[[:space:]]Spatial-Temporal[[:space:]]Graph[[:space:]]Model[[:space:]]for[[:space:]]Multi-Camera[[:space:]]Multi-Object[[:space:]]Tracking/f59bf130-110e-4b0f-b110-afa1ed8fc750_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Re_PolyWorld[[:space:]]-[[:space:]]A[[:space:]]Graph[[:space:]]Neural[[:space:]]Network[[:space:]]for[[:space:]]Polygonal[[:space:]]Scene[[:space:]]Parsing/68af7ba9-e2d9-461e-93f2-b88c2b579a79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ReactioNet_[[:space:]]Learning[[:space:]]High-Order[[:space:]]Facial[[:space:]]Behavior[[:space:]]from[[:space:]]Universal[[:space:]]Stimulus-Reaction[[:space:]]by[[:space:]]Dyadic[[:space:]]Relation[[:space:]]Reasoning/814713c2-4b57-4092-9ed5-36c25de28e98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Read-only[[:space:]]Prompt[[:space:]]Optimization[[:space:]]for[[:space:]]Vision-Language[[:space:]]Few-shot[[:space:]]Learning/870012ca-c33d-4e51-ad3f-d8a81141ba0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Real-Time[[:space:]]Neural[[:space:]]Rasterization[[:space:]]for[[:space:]]Large[[:space:]]Scenes/109ed94f-d329-47d1-b8b7-28d04bb3b4b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RealGraph_[[:space:]]A[[:space:]]Multiview[[:space:]]Dataset[[:space:]]for[[:space:]]4D[[:space:]]Real-world[[:space:]]Context[[:space:]]Graph[[:space:]]Generation/f36d1c24-e55e-43ab-8970-7494ad6fd015_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Realistic[[:space:]]Full-Body[[:space:]]Tracking[[:space:]]from[[:space:]]Sparse[[:space:]]Observations[[:space:]]via[[:space:]]Joint-Level[[:space:]]Modeling/9730a29f-54d0-4779-94cc-84ec645bc90c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RecRecNet_[[:space:]]Rectangling[[:space:]]Rectified[[:space:]]Wide-Angle[[:space:]]Images[[:space:]]by[[:space:]]Thin-Plate[[:space:]]Spline[[:space:]]Model[[:space:]]and[[:space:]]DoF-based[[:space:]]Curriculum[[:space:]]Learning/fe8c5fe7-9564-49ac-a29b-1aec3bc55fe4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reconciling[[:space:]]Object-Level[[:space:]]and[[:space:]]Global-Level[[:space:]]Objectives[[:space:]]for[[:space:]]Long-Tail[[:space:]]Detection/b9d5c464-a88a-4488-a5f6-473ffa48576a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reconstructed[[:space:]]Convolution[[:space:]]Module[[:space:]]Based[[:space:]]Look-Up[[:space:]]Tables[[:space:]]for[[:space:]]Efficient[[:space:]]Image[[:space:]]Super-Resolution/02a7a574-e6f7-4751-a4f4-5da27b5ac1e9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reconstructing[[:space:]]Groups[[:space:]]of[[:space:]]People[[:space:]]with[[:space:]]Hypergraph[[:space:]]Relational[[:space:]]Reasoning/27418ee6-2ea1-4dee-a93b-062c682c998d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reconstructing[[:space:]]Interacting[[:space:]]Hands[[:space:]]with[[:space:]]Interaction[[:space:]]Prior[[:space:]]from[[:space:]]Monocular[[:space:]]Images/a1a88ce5-e34d-4980-b092-49b03f98c1d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Recovering[[:space:]]a[[:space:]]Molecule's[[:space:]]3D[[:space:]]Dynamics[[:space:]]from[[:space:]]Liquid-phase[[:space:]]Electron[[:space:]]Microscopy[[:space:]]Movies/05fafcfe-ad73-4239-a781-0de1cd4fc5a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Recursive[[:space:]]Video[[:space:]]Lane[[:space:]]Detection/6f865887-4f69-4ddf-be08-fc6f56a6c479_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RecursiveDet_[[:space:]]End-to-End[[:space:]]Region-Based[[:space:]]Recursive[[:space:]]Object[[:space:]]Detection/6e06e604-8868-4928-8cdc-459b1fef71e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reducing[[:space:]]Training[[:space:]]Time[[:space:]]in[[:space:]]Cross-Silo[[:space:]]Federated[[:space:]]Learning[[:space:]]Using[[:space:]]Multigraph[[:space:]]Topology/4c7efd6a-b745-4cd2-a452-2c826bef363c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ref-NeuS_[[:space:]]Ambiguity-Reduced[[:space:]]Neural[[:space:]]Implicit[[:space:]]Surface[[:space:]]Learning[[:space:]]for[[:space:]]Multi-View[[:space:]]Reconstruction[[:space:]]with[[:space:]]Reflection/f3645140-6f08-4ea9-9a00-eeeec8105847_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RefEgo_[[:space:]]Referring[[:space:]]Expression[[:space:]]Comprehension[[:space:]]Dataset[[:space:]]from[[:space:]]First-Person[[:space:]]Perception[[:space:]]of[[:space:]]Ego4D/9aa5d5dc-884e-4266-8d26-ac03f92bc7f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reference-guided[[:space:]]Controllable[[:space:]]Inpainting[[:space:]]of[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/108a7538-6d0f-46fd-9434-8495a961d1f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Referring[[:space:]]Image[[:space:]]Segmentation[[:space:]]Using[[:space:]]Text[[:space:]]Supervision/02a327e9-853e-4010-afce-0986524c6ae1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RegFormer_[[:space:]]An[[:space:]]Efficient[[:space:]]Projection-Aware[[:space:]]Transformer[[:space:]]Network[[:space:]]for[[:space:]]Large-Scale[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/a338e428-4744-44b5-96b2-f572469e9303_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Regularized[[:space:]]Mask[[:space:]]Tuning_[[:space:]]Uncovering[[:space:]]Hidden[[:space:]]Knowledge[[:space:]]in[[:space:]]Pre-Trained[[:space:]]Vision-Language[[:space:]]Models/1e0a2f72-38fa-447b-b4fb-31f0c1375d4a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Regularized[[:space:]]Primitive[[:space:]]Graph[[:space:]]Learning[[:space:]]for[[:space:]]Unified[[:space:]]Vector[[:space:]]Mapping/41e38808-3f17-4ba0-add0-8c2563e2dc98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rehearsal-Free[[:space:]]Domain[[:space:]]Continual[[:space:]]Face[[:space:]]Anti-Spoofing_[[:space:]]Generalize[[:space:]]More[[:space:]]and[[:space:]]Forget[[:space:]]Less/7df59f20-4f77-4ab0-90ec-ec335102e71c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reinforce[[:space:]]Data,[[:space:]]Multiply[[:space:]]Impact_[[:space:]]Improved[[:space:]]Model[[:space:]]Accuracy[[:space:]]and[[:space:]]Robustness[[:space:]]with[[:space:]]Dataset[[:space:]]Reinforcement/8cbba0e0-5a68-42fe-a72b-eddf04c41c7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Reinforced[[:space:]]Disentanglement[[:space:]]for[[:space:]]Face[[:space:]]Swapping[[:space:]]without[[:space:]]Skip[[:space:]]Connection/364ec2cc-cdcd-4001-a450-eeeb9901cf5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Relightify_[[:space:]]Relightable[[:space:]]3D[[:space:]]Faces[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image[[:space:]]via[[:space:]]Diffusion[[:space:]]Models/a363b7d9-b21e-4562-9584-a7dd443a213e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Remembering[[:space:]]Normality_[[:space:]]Memory-guided[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Unsupervised[[:space:]]Anomaly[[:space:]]Detection/cf47b28a-3639-4252-9687-43df64f85447_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Removing[[:space:]]Anomalies[[:space:]]as[[:space:]]Noises[[:space:]]for[[:space:]]Industrial[[:space:]]Defect[[:space:]]Localization/5abbfd8c-dca1-4f3a-b838-eb4992249164_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RenderIH_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Synthetic[[:space:]]Dataset[[:space:]]for[[:space:]]3D[[:space:]]Interacting[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation/d5daffd4-f953-4866-96a1-a958da1c6771_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rendering[[:space:]]Humans[[:space:]]from[[:space:]]Object-Occluded[[:space:]]Monocular[[:space:]]Videos/7ec6056c-5f8a-43ac-b2e7-79805d07fdfa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/RepQ-ViT_[[:space:]]Scale[[:space:]]Reparameterization[[:space:]]for[[:space:]]Post-Training[[:space:]]Quantization[[:space:]]of[[:space:]]Vision[[:space:]]Transformers/beeaa5dc-3676-4c6e-8221-38fafd0d4f05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Replay_[[:space:]]Multi-modal[[:space:]]Multi-view[[:space:]]Acted[[:space:]]Videos[[:space:]]for[[:space:]]Casual[[:space:]]Holography/76ab5b02-1cf6-4751-9516-5e712a98fbc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Representation[[:space:]]Disparity-aware[[:space:]]Distillation[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/9b9d37fc-ae17-4ca5-afd0-eb8e2889c5cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Representation[[:space:]]Uncertainty[[:space:]]in[[:space:]]Self-Supervised[[:space:]]Learning[[:space:]]as[[:space:]]Variational[[:space:]]Inference/b8089fe3-8aec-4abb-a409-818092cc32a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ResQ_[[:space:]]Residual[[:space:]]Quantization[[:space:]]for[[:space:]]Video[[:space:]]Perception/a5baaaf2-41da-431a-8f3d-7c3524f496d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Residual[[:space:]]Pattern[[:space:]]Learning[[:space:]]for[[:space:]]Pixel-Wise[[:space:]]Out-of-Distribution[[:space:]]Detection[[:space:]]in[[:space:]]Semantic[[:space:]]Segmentation/ec924c06-cc92-46c4-826c-62079a01dca9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Amodal[[:space:]]Video[[:space:]]Segmentation[[:space:]]from[[:space:]]Learning[[:space:]]Supervised[[:space:]]Signals[[:space:]]with[[:space:]]Object-centric[[:space:]]Representation/77af8f5b-1889-41b9-9bf4-9a37028fd925_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Data[[:space:]]Distillation_[[:space:]]Do[[:space:]]Not[[:space:]]Overlook[[:space:]]Calibration/5a4503ef-8717-420c-9670-2b7977dd2da9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Fast[[:space:]]Fourier[[:space:]]Convolution[[:space:]]in[[:space:]]Image[[:space:]]Inpainting/bd48c9ca-d080-4461-9f6b-965884959b8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Mobile[[:space:]]Block[[:space:]]for[[:space:]]Efficient[[:space:]]Attention-based[[:space:]]Models/30d8f267-ffbf-4c7c-adc5-3b7b37b875e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Multi-Contrast[[:space:]]MRI[[:space:]]Super-Resolution_[[:space:]]Rectangle-Window[[:space:]]Cross-Attention[[:space:]]Transformer[[:space:]]and[[:space:]]Arbitrary-Scale[[:space:]]Upsampling/1c623158-e093-4054-8fb3-f1f3b6f067ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration[[:space:]]as[[:space:]]Masking[[:space:]]and[[:space:]]Reconstruction/4aee2148-7b07-491e-82b3-f27c8db418a5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Pose[[:space:]]Estimation[[:space:]]in[[:space:]]Crowds_[[:space:]]Overcoming[[:space:]]the[[:space:]]Detection[[:space:]]Information[[:space:]]Bottleneck[[:space:]]and[[:space:]]Ambiguity/a70bdb0c-96a6-44ac-b5d7-3a2fb1ecb054_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Range[[:space:]]View[[:space:]]Representation[[:space:]]for[[:space:]]LiDAR[[:space:]]Segmentation/1a20a55f-b2c2-4b94-a3b5-f9cecce9f1be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Safe[[:space:]]Semi-supervised[[:space:]]Learning_[[:space:]]Transferring[[:space:]]the[[:space:]]Open-set[[:space:]]Problem[[:space:]]to[[:space:]]A[[:space:]]Close-set[[:space:]]One/b7ea0291-c0e1-4a34-889b-a3c1cb6fd6af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation[[:space:]]from[[:space:]]Shutter[[:space:]]Mode[[:space:]]Induced[[:space:]]Degradation/c4a29997-ae0e-4c91-92da-e322eb41e5ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]Vision[[:space:]]Transformers[[:space:]]for[[:space:]]MobileNet[[:space:]]Size[[:space:]]and[[:space:]]Speed/5eeeadd3-2e29-4f93-bedf-31b242e23ee3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rethinking[[:space:]]the[[:space:]]Role[[:space:]]of[[:space:]]Pre-Trained[[:space:]]Networks[[:space:]]in[[:space:]]Source-Free[[:space:]]Domain[[:space:]]Adaptation/06d85838-283c-4479-927d-1ff6803ec90a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Retinexformer_[[:space:]]One-stage[[:space:]]Retinex-based[[:space:]]Transformer[[:space:]]for[[:space:]]Low-light[[:space:]]Image[[:space:]]Enhancement/7b55e77e-be73-4fd7-9bc1-825f54336c0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Retro-FPN_[[:space:]]Retrospective[[:space:]]Feature[[:space:]]Pyramid[[:space:]]Network[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Semantic[[:space:]]Segmentation/a2255514-2097-4f94-b325-e6b47eeec3d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisit[[:space:]]PCA-based[[:space:]]Technique[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection/de873e79-235d-4973-bcac-6d7b77de0a8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Domain-Adaptive[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]by[[:space:]]Reliable,[[:space:]]Diverse[[:space:]]and[[:space:]]Class-balanced[[:space:]]Pseudo-Labeling/34ed5af8-ff07-44ab-a9b5-6e896a2c304a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Foreground[[:space:]]and[[:space:]]Background[[:space:]]Separation[[:space:]]in[[:space:]]Weakly-supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization_[[:space:]]A[[:space:]]Clustering-based[[:space:]]Approach/cde310cc-8e00-403b-aa5d-0f1e8cd369fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Scene[[:space:]]Text[[:space:]]Recognition_[[:space:]]A[[:space:]]Data[[:space:]]Perspective/98fa85c0-b72a-45fe-92cb-cfea71ce6d51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]Vision[[:space:]]Transformer[[:space:]]from[[:space:]]the[[:space:]]View[[:space:]]of[[:space:]]Path[[:space:]]Ensemble/1da44968-3be8-4172-a1f0-73c4ebfbdd01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Revisiting[[:space:]]the[[:space:]]Parameter[[:space:]]Efficiency[[:space:]]of[[:space:]]Adapters[[:space:]]from[[:space:]]the[[:space:]]Perspective[[:space:]]of[[:space:]]Precision[[:space:]]Redundancy/eb8e064c-97d5-4820-bd7d-cb4ec3748138_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rickrolling[[:space:]]the[[:space:]]Artist_[[:space:]]Injecting[[:space:]]Backdoors[[:space:]]into[[:space:]]Text[[:space:]]Encoders[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Synthesis/7f374182-72fb-4899-806b-31beba883a60_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robo3D_[[:space:]]Towards[[:space:]]Robust[[:space:]]and[[:space:]]Reliable[[:space:]]3D[[:space:]]Perception[[:space:]]against[[:space:]]Corruptions/d9142cba-93fb-41a7-9617-144955c2ac93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Evaluation[[:space:]]of[[:space:]]Diffusion-Based[[:space:]]Adversarial[[:space:]]Purification/a893f09c-c4c3-4a48-8955-996e58abed2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Frame-to-Frame[[:space:]]Camera[[:space:]]Rotation[[:space:]]Estimation[[:space:]]in[[:space:]]Crowded[[:space:]]Scenes/76110b33-e493-40d6-8120-fdd5ec7aaba2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Geometry-Preserving[[:space:]]Depth[[:space:]]Estimation[[:space:]]Using[[:space:]]Differentiable[[:space:]]Rendering/28adc3ac-7c34-4fb3-be84-a71f67c92818_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Heterogeneous[[:space:]]Federated[[:space:]]Learning[[:space:]]under[[:space:]]Data[[:space:]]Corruption/3051a8f7-aef9-41e3-99c9-5d20f07a61ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Mixture-of-Expert[[:space:]]Training[[:space:]]for[[:space:]]Convolutional[[:space:]]Neural[[:space:]]Networks/262207c5-e6e4-4505-a5e6-b1bcfe89f378_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation[[:space:]]under[[:space:]]Challenging[[:space:]]Conditions/ad9a7732-257e-4c85-8ebb-4aa899115bea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Object[[:space:]]Modeling[[:space:]]for[[:space:]]Visual[[:space:]]Tracking/00314f98-ba13-4d7f-9471-c3995e660b9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]One-Shot[[:space:]]Face[[:space:]]Video[[:space:]]Re-enactment[[:space:]]using[[:space:]]Hybrid[[:space:]]Latent[[:space:]]Spaces[[:space:]]of[[:space:]]StyleGAN2/c048492c-ec95-4ab7-9ca6-9911d57b8897_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]Referring[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation[[:space:]]with[[:space:]]Cyclic[[:space:]]Structural[[:space:]]Consensus/84dfa689-ddee-430a-9dab-e5a1cd76f8b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robust[[:space:]]e-NeRF_[[:space:]]NeRF[[:space:]]from[[:space:]]Sparse[[:space:]]&[[:space:]]Noisy[[:space:]]Events[[:space:]]under[[:space:]]Non-Uniform[[:space:]]Motion/b941741f-c8b1-453a-a81c-5c8fe3fe75d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Robustifying[[:space:]]Token[[:space:]]Attention[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/65b16640-ac13-41a7-a616-4146806ebd0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Role-Aware[[:space:]]Interaction[[:space:]]Generation[[:space:]]from[[:space:]]Textual[[:space:]]Description/762b43c6-121b-4c4e-98c3-f302bdc252a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Root[[:space:]]Pose[[:space:]]Decomposition[[:space:]]Towards[[:space:]]Generic[[:space:]]Non-rigid[[:space:]]3D[[:space:]]Reconstruction[[:space:]]with[[:space:]]Monocular[[:space:]]Videos/8bbcfa14-6dfc-4abc-9874-b89dfb30e8c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Rosetta[[:space:]]Neurons_[[:space:]]Mining[[:space:]]the[[:space:]]Common[[:space:]]Units[[:space:]]in[[:space:]]a[[:space:]]Model[[:space:]]Zoo/d8b295e2-7272-4689-8836-09efe4ba4133_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/S-TREK_[[:space:]]Sequential[[:space:]]Translation[[:space:]]and[[:space:]]Rotation[[:space:]]Equivariant[[:space:]]Keypoints[[:space:]]for[[:space:]]Local[[:space:]]Feature[[:space:]]Extraction/99568c5a-035f-441c-b560-f0fa197b6253_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/S-VolSDF_[[:space:]]Sparse[[:space:]]Multi-View[[:space:]]Stereo[[:space:]]Regularization[[:space:]]of[[:space:]]Neural[[:space:]]Implicit[[:space:]]Surfaces/8bdd21fd-8585-4d3d-aec1-8cd8eba99bd5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/S3IM_[[:space:]]Stochastic[[:space:]]Structural[[:space:]]SIMilarity[[:space:]]and[[:space:]]Its[[:space:]]Unreasonable[[:space:]]Effectiveness[[:space:]]for[[:space:]]Neural[[:space:]]Fields/1a154fa8-60c5-4581-a2bc-523a2e4fafa6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SA-BEV_[[:space:]]Generating[[:space:]]Semantic-Aware[[:space:]]Bird's-Eye-View[[:space:]]Feature[[:space:]]for[[:space:]]Multi-view[[:space:]]3D[[:space:]]Object[[:space:]]Detection/8a4775de-65ba-4457-b860-c776e2f5bc59_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SAFARI_[[:space:]]Versatile[[:space:]]and[[:space:]]Efficient[[:space:]]Evaluations[[:space:]]for[[:space:]]Robustness[[:space:]]of[[:space:]]Interpretability/ad511e8b-9215-4559-b5e1-d99784b297c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SAFE_[[:space:]]Machine[[:space:]]Unlearning[[:space:]]With[[:space:]]Shard[[:space:]]Graphs/cf0c0568-1b2d-4072-ad46-e74a7a56584a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SAFE_[[:space:]]Sensitivity-Aware[[:space:]]Features[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Object[[:space:]]Detection/cd18c99e-1db4-4928-a365-356688ddaa7f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SAFL-Net_[[:space:]]Semantic-Agnostic[[:space:]]Feature[[:space:]]Learning[[:space:]]Network[[:space:]]with[[:space:]]Auxiliary[[:space:]]Plugins[[:space:]]for[[:space:]]Image[[:space:]]Manipulation[[:space:]]Detection/a2c6ae3a-82db-4b53-9097-67275112a298_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SAGA_[[:space:]]Spectral[[:space:]]Adversarial[[:space:]]Geometric[[:space:]]Attack[[:space:]]on[[:space:]]3D[[:space:]]Meshes/4103764f-1656-4824-a609-f6ef715d0f90_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SAL-ViT_[[:space:]]Towards[[:space:]]Latency[[:space:]]Efficient[[:space:]]Private[[:space:]]Inference[[:space:]]on[[:space:]]ViT[[:space:]]using[[:space:]]Selective[[:space:]]Attention[[:space:]]Search[[:space:]]with[[:space:]]a[[:space:]]Learnable[[:space:]]Softmax[[:space:]]Approximation/c127cea8-0dc6-49f2-8b8a-97ece8106b94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SALAD_[[:space:]]Part-Level[[:space:]]Latent[[:space:]]Diffusion[[:space:]]for[[:space:]]3D[[:space:]]Shape[[:space:]]Generation[[:space:]]and[[:space:]]Manipulation/66b22e52-c97e-4c39-8bea-45eef7c0bb60_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SAMPLING_[[:space:]]Scene-adaptive[[:space:]]Hierarchical[[:space:]]Multiplane[[:space:]]Images[[:space:]]Representation[[:space:]]for[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/4fb41d44-a542-4c4f-9a02-e45956b3e898_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SATR_[[:space:]]Zero-Shot[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]of[[:space:]]3D[[:space:]]Shapes/ff92bc68-718a-4fb8-b44f-ae3f2dec86fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SC3K_[[:space:]]Self-supervised[[:space:]]and[[:space:]]Coherent[[:space:]]3D[[:space:]]Keypoints[[:space:]]Estimation[[:space:]]from[[:space:]]Rotated,[[:space:]]Noisy,[[:space:]]and[[:space:]]Decimated[[:space:]]Point[[:space:]]Cloud[[:space:]]Data/e1a55925-ee8f-4109-936f-b6036efd1c35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SCANet_[[:space:]]Scene[[:space:]]Complexity[[:space:]]Aware[[:space:]]Network[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Video[[:space:]]Moment[[:space:]]Retrieval/0b0731d8-3653-4621-973c-2cea18eac709_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SCOB_[[:space:]]Universal[[:space:]]Text[[:space:]]Understanding[[:space:]]via[[:space:]]Character-wise[[:space:]]Supervised[[:space:]]Contrastive[[:space:]]Learning[[:space:]]with[[:space:]]Online[[:space:]]Text[[:space:]]Rendering[[:space:]]for[[:space:]]Bridging[[:space:]]Domain[[:space:]]Gap/78ac89c4-2c08-494b-8af8-758d8feb29c7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SEFD_[[:space:]]Learning[[:space:]]to[[:space:]]Distill[[:space:]]Complex[[:space:]]Pose[[:space:]]and[[:space:]]Occlusion/45abec4c-9744-4c3d-8454-9869b79b7d36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SEMPART_[[:space:]]Self-supervised[[:space:]]Multi-resolution[[:space:]]Partitioning[[:space:]]of[[:space:]]Image[[:space:]]Semantics/686e6157-a065-4d3c-91f4-925286d3ab47_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SFHarmony_[[:space:]]Source[[:space:]]Free[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Distributed[[:space:]]Neuroimaging[[:space:]]Analysis/356b08ce-81b4-4c11-85da-6e3abbafd4af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SG-Former_[[:space:]]Self-guided[[:space:]]Transformer[[:space:]]with[[:space:]]Evolving[[:space:]]Token[[:space:]]Reallocation/70b9d383-4c91-4bc9-8835-007c558b807f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SGAligner_[[:space:]]3D[[:space:]]Scene[[:space:]]Alignment[[:space:]]with[[:space:]]Scene[[:space:]]Graphs/4f8d52b1-c316-4db0-b56f-e697b2d1d861_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SHACIRA_[[:space:]]Scalable[[:space:]]HAsh-grid[[:space:]]Compression[[:space:]]for[[:space:]]Implicit[[:space:]]Neural[[:space:]]Representations/7795f23a-1209-4257-962c-1b66bc5543f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SHERF_[[:space:]]Generalizable[[:space:]]Human[[:space:]]NeRF[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/9049ecef-b26e-4633-904c-f2de020ae84d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SHIFT3D_[[:space:]]Synthesizing[[:space:]]Hard[[:space:]]Inputs[[:space:]]For[[:space:]]Tricking[[:space:]]3D[[:space:]]Detectors/364e348d-6a40-44c7-a6d3-0fe1fc006253_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SIDGAN_[[:space:]]High-Resolution[[:space:]]Dubbed[[:space:]]Video[[:space:]]Generation[[:space:]]via[[:space:]]Shift-Invariant[[:space:]]Learning/f4f83c37-390f-46c3-816f-a130cce0f3bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SIGMA_[[:space:]]Scale-Invariant[[:space:]]Global[[:space:]]Sparse[[:space:]]Shape[[:space:]]Matching/354345c9-184e-400e-85bb-b601e0ae186d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SILT_[[:space:]]Shadow-Aware[[:space:]]Iterative[[:space:]]Label[[:space:]]Tuning[[:space:]]for[[:space:]]Learning[[:space:]]to[[:space:]]Detect[[:space:]]Shadows[[:space:]]from[[:space:]]Noisy[[:space:]]Labels/6254a4e2-675f-4685-b538-d90e3b5e68f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SINC_[[:space:]]Self-Supervised[[:space:]]In-Context[[:space:]]Learning[[:space:]]for[[:space:]]Vision-Language[[:space:]]Tasks/d62b3033-2d93-47de-810b-4f5380e1b306_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SINC_[[:space:]]Spatial[[:space:]]Composition[[:space:]]of[[:space:]]3D[[:space:]]Human[[:space:]]Motions[[:space:]]for[[:space:]]Simultaneous[[:space:]]Action[[:space:]]Generation/2430bd53-9ab6-446a-bc9b-1abd0dcfeacd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SIRA-PCR_[[:space:]]Sim-to-Real[[:space:]]Adaptation[[:space:]]for[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/0581799c-81e3-4d8f-8a3e-f9e18b6dd939_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SKED_[[:space:]]Sketch-guided[[:space:]]Text-based[[:space:]]3D[[:space:]]Editing/9f9a2b38-ed9a-433a-951c-494c6eb700aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SKiT_[[:space:]]a[[:space:]]Fast[[:space:]]Key[[:space:]]Information[[:space:]]Video[[:space:]]Transformer[[:space:]]for[[:space:]]Online[[:space:]]Surgical[[:space:]]Phase[[:space:]]Recognition/a0ff069c-52c1-4f01-9242-27f0f8d9ff9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SLAN_[[:space:]]Self-Locator[[:space:]]Aided[[:space:]]Network[[:space:]]for[[:space:]]Vision-Language[[:space:]]Understanding/3a1e20d0-4005-4a4b-8cae-c4c6a2843100_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SLCA_[[:space:]]Slow[[:space:]]Learner[[:space:]]with[[:space:]]Classifier[[:space:]]Alignment[[:space:]]for[[:space:]]Continual[[:space:]]Learning[[:space:]]on[[:space:]]a[[:space:]]Pre-trained[[:space:]]Model/409eb288-aa08-4732-ad45-644c92c6c917_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SMAUG_[[:space:]]Sparse[[:space:]]Masked[[:space:]]Autoencoder[[:space:]]for[[:space:]]Efficient[[:space:]]Video-Language[[:space:]]Pre-Training/887d4d94-79bb-4eac-8f60-c6a6c1b28685_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SMMix_[[:space:]]Self-Motivated[[:space:]]Image[[:space:]]Mixing[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/451c0212-4651-4fbc-904f-b650149050bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SOAR_[[:space:]]Scene-debiasing[[:space:]]Open-set[[:space:]]Action[[:space:]]Recognition/6cac9206-337e-4081-be2d-05b245a850bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SOCS_[[:space:]]Semantically-Aware[[:space:]]Object[[:space:]]Coordinate[[:space:]]Space[[:space:]]for[[:space:]]Category-Level[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]under[[:space:]]Large[[:space:]]Shape[[:space:]]Variations/1f4be573-715b-48fe-80d6-a0303b2a0dc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SPACE_[[:space:]]Speech-driven[[:space:]]Portrait[[:space:]]Animation[[:space:]]with[[:space:]]Controllable[[:space:]]Expression/c5a105c8-22b3-40d8-b397-998426bf9a83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SPANet_[[:space:]]Frequency-balancing[[:space:]]Token[[:space:]]Mixer[[:space:]]using[[:space:]]Spectral[[:space:]]Pooling[[:space:]]Aggregation[[:space:]]Modulation/36acf2ab-2464-471c-824c-e1c84c31017e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SQAD_[[:space:]]Automatic[[:space:]]Smartphone[[:space:]]Camera[[:space:]]Quality[[:space:]]Assessment[[:space:]]and[[:space:]]Benchmarking/16c63254-a2b5-44c0-a6ce-055b2b99f569_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SRFormer_[[:space:]]Permuted[[:space:]]Self-Attention[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Super-Resolution/45b1226f-3e09-40bd-8998-19c0e2850f57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SSB_[[:space:]]Simple[[:space:]]but[[:space:]]Strong[[:space:]]Baseline[[:space:]]for[[:space:]]Boosting[[:space:]]Performance[[:space:]]of[[:space:]]Open-Set[[:space:]]Semi-Supervised[[:space:]]Learning/bba59860-99b4-4dca-86c2-bc3811ccd29e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SSDA_[[:space:]]Secure[[:space:]]Source-Free[[:space:]]Domain[[:space:]]Adaptation/601991a5-a75a-4070-bb07-88c905c3b6fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SSF_[[:space:]]Accelerating[[:space:]]Training[[:space:]]of[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks[[:space:]]with[[:space:]]Stabilized[[:space:]]Spiking[[:space:]]Flow/f0cb7cb3-6ab7-4be1-8594-7c050dba7aee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/STEERER_[[:space:]]Resolving[[:space:]]Scale[[:space:]]Variations[[:space:]]for[[:space:]]Counting[[:space:]]and[[:space:]]Localization[[:space:]]via[[:space:]]Selective[[:space:]]Inheritance[[:space:]]Learning/36b3fc2d-096c-4b5b-964f-c8f409224ed1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/STEPs_[[:space:]]Self-Supervised[[:space:]]Key[[:space:]]Step[[:space:]]Extraction[[:space:]]and[[:space:]]Localization[[:space:]]from[[:space:]]Unlabeled[[:space:]]Procedural[[:space:]]Videos/109515f0-41cc-45ad-85f0-5906855cc50d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/STPrivacy_[[:space:]]Spatio-Temporal[[:space:]]Privacy-Preserving[[:space:]]Action[[:space:]]Recognition/fcad4fac-73f3-472f-9077-c2d1bf1daec5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SUMMIT_[[:space:]]Source-Free[[:space:]]Adaptation[[:space:]]of[[:space:]]Uni-Modal[[:space:]]Models[[:space:]]to[[:space:]]Multi-Modal[[:space:]]Targets/7d70e245-7f7f-4dcf-a04c-423d6f1f5431_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SVDFormer_[[:space:]]Complementing[[:space:]]Point[[:space:]]Cloud[[:space:]]via[[:space:]]Self-view[[:space:]]Augmentation[[:space:]]and[[:space:]]Self-structure[[:space:]]Dual-generator/4eed359b-107d-45c5-b073-684a181aeba5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SVDiff_[[:space:]]Compact[[:space:]]Parameter[[:space:]]Space[[:space:]]for[[:space:]]Diffusion[[:space:]]Fine-Tuning/63aabc47-019a-46b1-bb3d-6797f238a337_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SVQNet_[[:space:]]Sparse[[:space:]]Voxel-Adjacent[[:space:]]Query[[:space:]]Network[[:space:]]for[[:space:]]4D[[:space:]]Spatio-Temporal[[:space:]]LiDAR[[:space:]]Semantic[[:space:]]Segmentation/ddb3d19b-e1a6-4421-94c3-1f7a0b67893f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SYENet_[[:space:]]A[[:space:]]Simple[[:space:]]Yet[[:space:]]Effective[[:space:]]Network[[:space:]]for[[:space:]]Multiple[[:space:]]Low-Level[[:space:]]Vision[[:space:]]Tasks[[:space:]]with[[:space:]]Real-Time[[:space:]]Performance[[:space:]]on[[:space:]]Mobile[[:space:]]Device/248a6325-372b-47a3-8d1f-c4e5560c41d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Saliency[[:space:]]Regularization[[:space:]]for[[:space:]]Self-Training[[:space:]]with[[:space:]]Partial[[:space:]]Annotations/a57d0621-983b-438f-8382-a45f1be4f38f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sample-adaptive[[:space:]]Augmentation[[:space:]]for[[:space:]]Point[[:space:]]Cloud[[:space:]]Recognition[[:space:]]Against[[:space:]]Real-world[[:space:]]Corruptions/1c882a35-b8e8-4735-9eed-d2aec0aad223_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sample-wise[[:space:]]Label[[:space:]]Confidence[[:space:]]Incorporation[[:space:]]for[[:space:]]Learning[[:space:]]with[[:space:]]Noisy[[:space:]]Labels/fe77abab-0be3-4339-b7a9-7b9c6e1d56da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sample4Geo_[[:space:]]Hard[[:space:]]Negative[[:space:]]Sampling[[:space:]]For[[:space:]]Cross-View[[:space:]]Geo-Localisation/600bc97d-511d-4452-909f-a8e04ad6aa3d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sat2Density_[[:space:]]Faithful[[:space:]]Density[[:space:]]Learning[[:space:]]from[[:space:]]Satellite-Ground[[:space:]]Image[[:space:]]Pairs/56378555-9bcd-4ee0-a1a5-b956e384547d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SatlasPretrain_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]for[[:space:]]Remote[[:space:]]Sensing[[:space:]]Image[[:space:]]Understanding/00bfc3a7-6a5e-4087-95c0-c0f4f107043d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scalable[[:space:]]Diffusion[[:space:]]Models[[:space:]]with[[:space:]]Transformers/88854950-0f86-40a1-9429-8e42fe8c73ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scalable[[:space:]]Multi-Temporal[[:space:]]Remote[[:space:]]Sensing[[:space:]]Change[[:space:]]Data[[:space:]]Generation[[:space:]]via[[:space:]]Simulating[[:space:]]Stochastic[[:space:]]Change[[:space:]]Process/62b67345-cc20-45fc-9137-a066d4908484_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scalable[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation[[:space:]]with[[:space:]]Simplified[[:space:]]Framework/a43981df-3cf4-4ea7-94dc-e5b8ee153b31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scale-Aware[[:space:]]Modulation[[:space:]]Meet[[:space:]]Transformer/d7c603f8-aab9-4224-9157-cd83aad177c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scale-MAE_[[:space:]]A[[:space:]]Scale-Aware[[:space:]]Masked[[:space:]]Autoencoder[[:space:]]for[[:space:]]Multiscale[[:space:]]Geospatial[[:space:]]Representation[[:space:]]Learning/c2c01f79-781d-4902-b5db-716835f78686_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scaling[[:space:]]Data[[:space:]]Generation[[:space:]]in[[:space:]]Vision-and-Language[[:space:]]Navigation/f9c91a6c-1f3e-402e-bee5-bc7e5a8e9aa8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ScanNet++_[[:space:]]A[[:space:]]High-Fidelity[[:space:]]Dataset[[:space:]]of[[:space:]]3D[[:space:]]Indoor[[:space:]]Scenes/745a73f8-04c7-403c-b4d4-cdb64cd7baad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scanning[[:space:]]Only[[:space:]]Once_[[:space:]]An[[:space:]]End-to-end[[:space:]]Framework[[:space:]]for[[:space:]]Fast[[:space:]]Temporal[[:space:]]Grounding[[:space:]]in[[:space:]]Long[[:space:]]Videos/369daa15-f8e0-450a-99a4-60c25942bec5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ScatterNeRF_[[:space:]]Seeing[[:space:]]Through[[:space:]]Fog[[:space:]]with[[:space:]]Physically-Based[[:space:]]Inverse[[:space:]]Neural[[:space:]]Rendering/6037bf6b-c6ca-4fb0-9181-f876ff712fcc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scene[[:space:]]Graph[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Embodied[[:space:]]Navigation/b694c6fe-e407-4788-acd7-e057a1f61a7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scene[[:space:]]Matters_[[:space:]]Model-based[[:space:]]Deep[[:space:]]Video[[:space:]]Compression/6f2fd7bb-e375-400c-ab9a-692c427a33d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scene[[:space:]]as[[:space:]]Occupancy/feddf2b2-1d9c-40c6-a0e4-9f157282a8dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scene-Aware[[:space:]]Feature[[:space:]]Matching/3280025b-b4e5-480a-9d5c-2306129d7ca3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scene-Aware[[:space:]]Label[[:space:]]Graph[[:space:]]Learning[[:space:]]for[[:space:]]Multi-Label[[:space:]]Image[[:space:]]Classification/c7bac3cf-5b3c-4840-8d20-bd4757318a72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SceneRF_[[:space:]]Self-Supervised[[:space:]]Monocular[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]with[[:space:]]Radiance[[:space:]]Fields/5e33dfeb-a888-4c19-bb66-b7283c1e3b2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scenimefy_[[:space:]]Learning[[:space:]]to[[:space:]]Craft[[:space:]]Anime[[:space:]]Scene[[:space:]]via[[:space:]]Semi-Supervised[[:space:]]Image-to-Image[[:space:]]Translation/7d21cfc4-58ea-44ed-a762-80cd2dc3a050_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Score[[:space:]]Priors[[:space:]]Guided[[:space:]]Deep[[:space:]]Variational[[:space:]]Inference[[:space:]]for[[:space:]]Unsupervised[[:space:]]Real-World[[:space:]]Single[[:space:]]Image[[:space:]]Denoising/bf59a31e-510c-458f-abc4-1cdcfc1e1a3c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Score-Based[[:space:]]Diffusion[[:space:]]Models[[:space:]]as[[:space:]]Principled[[:space:]]Priors[[:space:]]for[[:space:]]Inverse[[:space:]]Imaging/991ae81d-b645-44e7-8710-f474bf5c01ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scratch[[:space:]]Each[[:space:]]Other's[[:space:]]Back_[[:space:]]Incomplete[[:space:]]Multi-Modal[[:space:]]Brain[[:space:]]Tumor[[:space:]]Segmentation[[:space:]]via[[:space:]]Category[[:space:]]Aware[[:space:]]Group[[:space:]]Self-Support[[:space:]]Learning/09222efe-087b-4717-bc95-cff686f290c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Scratching[[:space:]]Visual[[:space:]]Transformer's[[:space:]]Back[[:space:]]with[[:space:]]Uniform[[:space:]]Attention/dba8bdba-edc7-4ecb-b304-fe8d9ecc24f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Seal-3D_[[:space:]]Interactive[[:space:]]Pixel-Level[[:space:]]Editing[[:space:]]for[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/1741e573-d4d9-401d-93db-9e3b2f4aeb72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Search[[:space:]]for[[:space:]]or[[:space:]]Navigate[[:space:]]to_[[:space:]]Dual[[:space:]]Adaptive[[:space:]]Thinking[[:space:]]for[[:space:]]Object[[:space:]]Navigation/c0d7466d-3c93-4785-bc77-847729f453de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/See[[:space:]]More[[:space:]]and[[:space:]]Know[[:space:]]More_[[:space:]]Zero-shot[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation[[:space:]]via[[:space:]]Multi-modal[[:space:]]Visual[[:space:]]Data/94e28caa-8e87-4cb0-b951-2d232741d35c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SeeABLE_[[:space:]]Soft[[:space:]]Discrepancies[[:space:]]and[[:space:]]Bounded[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Exposing[[:space:]]Deepfakes/c151129a-2489-45fb-b7ac-58b6efdc7464_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Seeing[[:space:]]Beyond[[:space:]]the[[:space:]]Patch_[[:space:]]Scale-Adaptive[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]of[[:space:]]High-resolution[[:space:]]Remote[[:space:]]Sensing[[:space:]]Imagery[[:space:]]based[[:space:]]on[[:space:]]Reinforcement[[:space:]]Learning/68f89365-7511-4189-bbf8-efed4158db57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SegGPT_[[:space:]]Towards[[:space:]]Segmenting[[:space:]]Everything[[:space:]]in[[:space:]]Context/944f3b6b-d48d-406d-be8d-7c6a789e22db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SegPrompt_[[:space:]]Boosting[[:space:]]Open-World[[:space:]]Segmentation[[:space:]]via[[:space:]]Category-Level[[:space:]]Prompt[[:space:]]Learning/74724532-8506-4515-bbca-59934f879476_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SegRCDB_[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]via[[:space:]]Formula-Driven[[:space:]]Supervised[[:space:]]Learning/2fe529d5-2e6f-4baa-ba37-46d22c0808e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Segment[[:space:]]Anything/eb57b157-6d50-47a0-b067-4d7f1974881f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Segment[[:space:]]Every[[:space:]]Reference[[:space:]]Object[[:space:]]in[[:space:]]Spatial[[:space:]]and[[:space:]]Temporal[[:space:]]Spaces/35c3a6f5-fc45-48a4-9e78-ffe1dfe4dfb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Segmentation[[:space:]]of[[:space:]]Tubular[[:space:]]Structures[[:space:]]Using[[:space:]]Iterative[[:space:]]Training[[:space:]]with[[:space:]]Tailored[[:space:]]Samples/2768a65b-50d0-48bb-ae69-564a7ae8fb84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Segmenting[[:space:]]Known[[:space:]]Objects[[:space:]]and[[:space:]]Unseen[[:space:]]Unknowns[[:space:]]without[[:space:]]Prior[[:space:]]Knowledge/6572dff6-13e7-41fd-8b60-890996677cc3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SeiT_[[:space:]]Storage-Efficient[[:space:]]Vision[[:space:]]Training[[:space:]]with[[:space:]]Tokens[[:space:]]Using[[:space:]]1%[[:space:]]of[[:space:]]Pixel[[:space:]]Storage/2820d200-3808-4ff1-b531-8f7526bf92de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Calibrated[[:space:]]Cross[[:space:]]Attention[[:space:]]Network[[:space:]]for[[:space:]]Few-Shot[[:space:]]Segmentation/ebbf1056-bd98-4376-9997-cf023fccd6de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Evolved[[:space:]]Dynamic[[:space:]]Expansion[[:space:]]Model[[:space:]]for[[:space:]]Task-Free[[:space:]]Continual[[:space:]]Learning/10fba97a-8259-4260-a6ee-de937bde9e97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Feedback[[:space:]]DETR[[:space:]]for[[:space:]]Temporal[[:space:]]Action[[:space:]]Detection/d4b1ba92-bc52-4804-9820-f30ac03251b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Ordering[[:space:]]Point[[:space:]]Clouds/9f805c6f-b12d-429a-aee2-763e3ebd55c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Organizing[[:space:]]Pathway[[:space:]]Expansion[[:space:]]for[[:space:]]Non-Exemplar[[:space:]]Class-Incremental[[:space:]]Learning/a9ed7969-be1d-4e9e-844b-eed1046d383f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Burst[[:space:]]Super-Resolution/95e19c5f-2b0f-43b6-814f-37123f95f9dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Character-to-Character[[:space:]]Distillation[[:space:]]for[[:space:]]Text[[:space:]]Recognition/bd7634f8-8c60-4951-b455-9c96c48981a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation[[:space:]]by[[:space:]]Direction-aware[[:space:]]Cumulative[[:space:]]Convolution[[:space:]]Network/44715a37-ff6c-4f7e-be40-d4e234cb91ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-Supervised[[:space:]]Object[[:space:]]Detection[[:space:]]from[[:space:]]Egocentric[[:space:]]Videos/fabad0e9-e9ff-4b22-bb49-561b673227e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-regulating[[:space:]]Prompts_[[:space:]]Foundational[[:space:]]Model[[:space:]]Adaptation[[:space:]]without[[:space:]]Forgetting/de94f2c6-8935-40d3-92b9-c3ec8bcfef62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-similarity[[:space:]]Driven[[:space:]]Scale-invariant[[:space:]]Learning[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Person[[:space:]]Search/33cd466f-3a57-456a-8d6e-1855da8fa091_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-supervised[[:space:]]Cross-view[[:space:]]Representation[[:space:]]Reconstruction[[:space:]]for[[:space:]]Change[[:space:]]Captioning/5f2ce86f-506e-442d-afdf-b4d723d95942_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-supervised[[:space:]]Image[[:space:]]Denoising[[:space:]]with[[:space:]]Downsampled[[:space:]]Invariance[[:space:]]Loss[[:space:]]and[[:space:]]Conditional[[:space:]]Blind-Spot[[:space:]]Network/eaa05795-c304-4369-b084-9c8d1819b943_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-supervised[[:space:]]Learning[[:space:]]of[[:space:]]Implicit[[:space:]]Shape[[:space:]]Representation[[:space:]]with[[:space:]]Dense[[:space:]]Correspondence[[:space:]]for[[:space:]]Deformable[[:space:]]Objects/46dc31d5-72ef-42cd-b1b5-e6a46a4978cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-supervised[[:space:]]Learning[[:space:]]to[[:space:]]Bring[[:space:]]Dual[[:space:]]Reversed[[:space:]]Rolling[[:space:]]Shutter[[:space:]]Images[[:space:]]Alive/0f7c9f43-96dc-4fd0-a054-e650a421a3a0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-supervised[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation_[[:space:]]Let's[[:space:]]Talk[[:space:]]About[[:space:]]The[[:space:]]Weather/7432c306-22bf-4ee4-b730-8425345482aa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-supervised[[:space:]]Monocular[[:space:]]Underwater[[:space:]]Depth[[:space:]]Recovery,[[:space:]]Image[[:space:]]Restoration,[[:space:]]and[[:space:]]a[[:space:]]Real-sea[[:space:]]Video[[:space:]]Dataset/e11d4111-32f8-407a-a327-9e6fbe49fac5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Self-supervised[[:space:]]Pre-training[[:space:]]for[[:space:]]Mirror[[:space:]]Detection/9faec277-19c4-4583-a3c8-84bf53084f2f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SemARFlow_[[:space:]]Injecting[[:space:]]Semantics[[:space:]]into[[:space:]]Unsupervised[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/fa7524e1-a76d-4844-9361-43c8b8046bf9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantic[[:space:]]Attention[[:space:]]Flow[[:space:]]Fields[[:space:]]for[[:space:]]Monocular[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Decomposition/68ea1ef8-9cb5-44da-894d-92e02632af66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantic[[:space:]]Information[[:space:]]in[[:space:]]Contrastive[[:space:]]Learning/c3144656-62d8-4975-97d6-ef4630f410df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantic-Aware[[:space:]]Dynamic[[:space:]]Parameter[[:space:]]for[[:space:]]Video[[:space:]]Inpainting[[:space:]]Transformer/706f36ef-c7c5-4ea7-8481-e3a3bb79a94c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantic-Aware[[:space:]]Implicit[[:space:]]Template[[:space:]]Learning[[:space:]]via[[:space:]]Part[[:space:]]Deformation[[:space:]]Consistency/840dd92d-8817-48f0-ac2c-586a3d27151d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantically[[:space:]]Structured[[:space:]]Image[[:space:]]Compression[[:space:]]via[[:space:]]Irregular[[:space:]]Group-Based[[:space:]]Decoupling/e3cccf23-d23a-4b30-90e1-8c04059b69dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantics[[:space:]]Meets[[:space:]]Temporal[[:space:]]Correspondence_[[:space:]]Self-supervised[[:space:]]Object-centric[[:space:]]Learning[[:space:]]in[[:space:]]Videos/0da0877f-f78d-4f97-9a49-55c50b2292de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantics-Consistent[[:space:]]Feature[[:space:]]Search[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Visual[[:space:]]Representation[[:space:]]Learning/fd85c784-c77b-49cb-91b7-f3a6f94e50c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semantify_[[:space:]]Simplifying[[:space:]]the[[:space:]]Control[[:space:]]of[[:space:]]3D[[:space:]]Morphable[[:space:]]Models[[:space:]]Using[[:space:]]CLIP/ff2d3341-5fbc-47bb-9ce8-0e16c61108bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-Supervised[[:space:]]Learning[[:space:]]via[[:space:]]Weight-Aware[[:space:]]Distillation[[:space:]]under[[:space:]]Class[[:space:]]Distribution[[:space:]]Mismatch/ab302b84-1623-4a04-b35b-84d3ca9ad536_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]under[[:space:]]Label[[:space:]]Noise[[:space:]]via[[:space:]]Diverse[[:space:]]Learning[[:space:]]Groups/8e5c15a5-f270-46c8-829c-81448a8fde6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-supervised[[:space:]]Semantics-guided[[:space:]]Adversarial[[:space:]]Training[[:space:]]for[[:space:]]Robust[[:space:]]Trajectory[[:space:]]Prediction/487dfb22-6094-4f31-976c-310213635421_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Semi-supervised[[:space:]]Speech-driven[[:space:]]3D[[:space:]]Facial[[:space:]]Animation[[:space:]]via[[:space:]]Cross-modal[[:space:]]Encoding/ebc67249-8be2-4ca5-975d-54a569184984_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sensitivity-Aware[[:space:]]Visual[[:space:]]Parameter-Efficient[[:space:]]Fine-Tuning/219bbdca-cfe8-433f-a6db-97c4773a1a10_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sentence[[:space:]]Attention[[:space:]]Blocks[[:space:]]for[[:space:]]Answer[[:space:]]Grounding/0622d151-144e-4e04-b267-1cccf0dba507_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sequential[[:space:]]Texts[[:space:]]Driven[[:space:]]Cohesive[[:space:]]Motions[[:space:]]Synthesis[[:space:]]with[[:space:]]Natural[[:space:]]Transitions/2c5b9c34-9105-42fb-b593-c16dc3db7c25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Set-level[[:space:]]Guidance[[:space:]]Attack_[[:space:]]Boosting[[:space:]]Adversarial[[:space:]]Transferability[[:space:]]of[[:space:]]Vision-Language[[:space:]]Pre-training[[:space:]]Models/4d2215bf-24a1-45b3-bcea-24623e459934_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shape[[:space:]]Analysis[[:space:]]of[[:space:]]Euclidean[[:space:]]Curves[[:space:]]under[[:space:]]Frenet-Serret[[:space:]]Framework/235fdcf4-691b-437b-8ee8-c7b76395f3ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shape[[:space:]]Anchor[[:space:]]Guided[[:space:]]Holistic[[:space:]]Indoor[[:space:]]Scene[[:space:]]Understanding/4f878220-d6cc-48e2-9d1e-82c1cf350095_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ShapeScaffolder_[[:space:]]Structure-Aware[[:space:]]3D[[:space:]]Shape[[:space:]]Generation[[:space:]]from[[:space:]]Text/3212f743-a520-438c-b4e2-ff41229e0a86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shatter[[:space:]]and[[:space:]]Gather_[[:space:]]Learning[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation[[:space:]]with[[:space:]]Text[[:space:]]Supervision/eba869e5-3a94-47e8-afcc-2f1d3dc46bfb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shift[[:space:]]from[[:space:]]Texture-bias[[:space:]]to[[:space:]]Shape-bias_[[:space:]]Edge[[:space:]]Deformation-based[[:space:]]Augmentation[[:space:]]for[[:space:]]Robust[[:space:]]Object[[:space:]]Recognition/e071fbaf-3036-4ac7-9bfd-fb6983ccf03f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ShiftNAS_[[:space:]]Improving[[:space:]]One-shot[[:space:]]NAS[[:space:]]via[[:space:]]Probability[[:space:]]Shift/9d0cd6fb-8343-4e85-b19e-ff450eb2e5e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shortcut-V2V_[[:space:]]Compression[[:space:]]Framework[[:space:]]for[[:space:]]Video-to-Video[[:space:]]Translation[[:space:]]Based[[:space:]]on[[:space:]]Temporal[[:space:]]Redundancy[[:space:]]Reduction/99da4ac0-6f5c-433f-bc7b-caa55aa672d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Shrinking[[:space:]]Class[[:space:]]Space[[:space:]]for[[:space:]]Enhanced[[:space:]]Certainty[[:space:]]in[[:space:]]Semi-Supervised[[:space:]]Learning/34917a9e-765e-47a5-bd3d-8044ab9de911_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SiLK_[[:space:]]Simple[[:space:]]Learned[[:space:]]Keypoints/545a50e4-a6f6-42d1-849e-c613959c21fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sigmoid[[:space:]]Loss[[:space:]]for[[:space:]]Language[[:space:]]Image[[:space:]]Pre-Training/85b123ac-a2cf-49ce-9282-93b24961a80b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sign[[:space:]]Language[[:space:]]Translation[[:space:]]with[[:space:]]Iterative[[:space:]]Prototype/f2c1bf6d-f55b-40d7-9372-42bff493c6d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SimFIR_[[:space:]]A[[:space:]]Simple[[:space:]]Framework[[:space:]]for[[:space:]]Fisheye[[:space:]]Image[[:space:]]Rectification[[:space:]]with[[:space:]]Self-supervised[[:space:]]Representation[[:space:]]Learning/e93fe221-1f98-47dd-bbe6-409ea24f23b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SimMatchV2_[[:space:]]Semi-Supervised[[:space:]]Learning[[:space:]]with[[:space:]]Graph[[:space:]]Consistency/4de24035-8115-4285-8e43-089c94dd354f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SimNP_[[:space:]]Learning[[:space:]]Self-Similarity[[:space:]]Priors[[:space:]]Between[[:space:]]Neural[[:space:]]Points/54ca6d2a-b8a1-4a4e-a6f5-57eb0fddb4dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Similarity[[:space:]]Min-Max_[[:space:]]Zero-Shot[[:space:]]Day-Night[[:space:]]Domain[[:space:]]Adaptation/d2d1b76d-0c03-43ab-96b8-c3ce7c667abf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Simoun_[[:space:]]Synergizing[[:space:]]Interactive[[:space:]]Motion-appearance[[:space:]]Understanding[[:space:]]for[[:space:]]Vision-based[[:space:]]Reinforcement[[:space:]]Learning/17dc049b-d1dc-4e17-921e-1317f662d966_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Simple[[:space:]]Baselines[[:space:]]for[[:space:]]Interactive[[:space:]]Video[[:space:]]Retrieval[[:space:]]with[[:space:]]Questions[[:space:]]and[[:space:]]Answers/149533a1-79f1-428b-885a-6c528f5bd0a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Simple[[:space:]]and[[:space:]]Effective[[:space:]]Out-of-Distribution[[:space:]]Detection[[:space:]]via[[:space:]]Cosine-based[[:space:]]Softmax[[:space:]]Loss/081c8308-e453-453c-8d9f-1e2b9b04b642_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SimpleClick_[[:space:]]Interactive[[:space:]]Image[[:space:]]Segmentation[[:space:]]with[[:space:]]Simple[[:space:]]Vision[[:space:]]Transformers/2b87aeaa-0b31-4e94-84d5-d9bfaea7e328_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Simulating[[:space:]]Fluids[[:space:]]in[[:space:]]Real-World[[:space:]]Still[[:space:]]Images/d51a02b7-7fc4-4f85-a456-fd173559877f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Single[[:space:]]Depth-image[[:space:]]3D[[:space:]]Reflection[[:space:]]Symmetry[[:space:]]and[[:space:]]Shape[[:space:]]Prediction/fcff96a5-45a1-4f7e-ab9f-0c309af710cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Single[[:space:]]Image[[:space:]]Deblurring[[:space:]]with[[:space:]]Row-dependent[[:space:]]Blur[[:space:]]Magnitude/05f13a7b-32f2-44f9-b585-5bd4c1beae6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Single[[:space:]]Image[[:space:]]Defocus[[:space:]]Deblurring[[:space:]]via[[:space:]]Implicit[[:space:]]Neural[[:space:]]Inverse[[:space:]]Kernels/bee3d863-7fc6-4e0b-a339-87a4afbe8bb4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Single[[:space:]]Image[[:space:]]Reflection[[:space:]]Separation[[:space:]]via[[:space:]]Component[[:space:]]Synergy/6169238f-a2a5-4863-b901-a9125cd1dec8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Single-Stage[[:space:]]Diffusion[[:space:]]NeRF_[[:space:]]A[[:space:]]Unified[[:space:]]Approach[[:space:]]to[[:space:]]3D[[:space:]]Generation[[:space:]]and[[:space:]]Reconstruction/06f68dbb-5c23-4704-ae55-c667675f6d9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Size[[:space:]]Does[[:space:]]Matter_[[:space:]]Size-aware[[:space:]]Virtual[[:space:]]Try-on[[:space:]]via[[:space:]]Clothing-oriented[[:space:]]Transformation[[:space:]]Try-on[[:space:]]Network/2e5e4941-f171-4e8c-be33-35affe248d0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SkeleTR_[[:space:]]Towards[[:space:]]Skeleton-based[[:space:]]Action[[:space:]]Recognition[[:space:]]in[[:space:]]the[[:space:]]Wild/27665a84-18aa-44bc-b25b-e6429d7734c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SkeletonMAE_[[:space:]]Graph-based[[:space:]]Masked[[:space:]]Autoencoder[[:space:]]for[[:space:]]Skeleton[[:space:]]Sequence[[:space:]]Pre-training/dcc19c04-02ae-4f8d-8f44-a86cdd91e8a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sketch[[:space:]]and[[:space:]]Text[[:space:]]Guided[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Colored[[:space:]]Point[[:space:]]Cloud[[:space:]]Generation/6a47e5bb-0a8f-4500-ae70-d8f17ed82ef9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Skill[[:space:]]Transformer_[[:space:]]A[[:space:]]Monolithic[[:space:]]Policy[[:space:]]for[[:space:]]Mobile[[:space:]]Manipulation/e273608a-9944-4287-91dc-881cf0220c1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Skip-Plan_[[:space:]]Procedure[[:space:]]Planning[[:space:]]in[[:space:]]Instructional[[:space:]]Videos[[:space:]]via[[:space:]]Condensed[[:space:]]Action[[:space:]]Space[[:space:]]Learning/c1a3a0ec-3885-4f2f-93dd-62793c50f314_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SlaBins_[[:space:]]Fisheye[[:space:]]Depth[[:space:]]Estimation[[:space:]]using[[:space:]]Slanted[[:space:]]Bins[[:space:]]on[[:space:]]Road[[:space:]]Environments/2f72858d-bf2e-45dd-b589-9c5d6578e810_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Small[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Coarse-to-fine[[:space:]]Proposal[[:space:]]Generation[[:space:]]and[[:space:]]Imitation[[:space:]]Learning/231f8e4e-8ec6-40cc-a455-514ae44e6fa1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Smoothness[[:space:]]Similarity[[:space:]]Regularization[[:space:]]for[[:space:]]Few-Shot[[:space:]]GAN[[:space:]]Adaptation/e521c479-803a-4c4d-bf48-e946e8872ca9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Snow[[:space:]]Removal[[:space:]]in[[:space:]]Video_[[:space:]]A[[:space:]]New[[:space:]]Dataset[[:space:]]and[[:space:]]A[[:space:]]Novel[[:space:]]Method/5ae50b63-d59f-4dc4-bf7f-4d6ac34210b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SoDaCam_[[:space:]]Software-defined[[:space:]]Cameras[[:space:]]via[[:space:]]Single-Photon[[:space:]]Imaging/ce1d43a9-f5c5-4d36-93a7-015029e926c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Social[[:space:]]Diffusion_[[:space:]]Long-term[[:space:]]Multiple[[:space:]]Human[[:space:]]Motion[[:space:]]Anticipation/e840bc61-69a6-4f9d-9e0d-f25143903533_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sound[[:space:]]Localization[[:space:]]from[[:space:]]Motion_[[:space:]]Jointly[[:space:]]Learning[[:space:]]Sound[[:space:]]Direction[[:space:]]and[[:space:]]Camera[[:space:]]Rotation/ebfa790a-8a91-4c3c-9e52-be53fe87e14f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sound[[:space:]]Source[[:space:]]Localization[[:space:]]is[[:space:]]All[[:space:]]about[[:space:]]Cross-Modal[[:space:]]Alignment/7b57a83c-7e70-4280-a6ee-f0f550384305_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Source-free[[:space:]]Depth[[:space:]]for[[:space:]]Object[[:space:]]Pop-out/1ac228ed-431b-40e3-b444-e5a574a99abb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Source-free[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/6db7e049-3b2d-4942-adce-fb8455dd18b2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Space[[:space:]]Engage_[[:space:]]Collaborative[[:space:]]Space[[:space:]]Supervision[[:space:]]for[[:space:]]Contrastive-Based[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation/5953e602-104c-4dc8-b226-b90796937563_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Space-time[[:space:]]Prompting[[:space:]]for[[:space:]]Video[[:space:]]Class-incremental[[:space:]]Learning/23f1a364-1b5a-487e-9a38-2d1ef0112aa7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SpaceEvo_[[:space:]]Hardware-Friendly[[:space:]]Search[[:space:]]Space[[:space:]]Design[[:space:]]for[[:space:]]Efficient[[:space:]]INT8[[:space:]]Inference/7d8d9b36-b64d-45d0-8aa5-d17e8abbee45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spacetime[[:space:]]Surface[[:space:]]Regularization[[:space:]]for[[:space:]]Neural[[:space:]]Dynamic[[:space:]]Scene[[:space:]]Reconstruction/11a08db4-fb69-460a-a705-16e05f3f48ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sparse[[:space:]]Instance[[:space:]]Conditioned[[:space:]]Multimodal[[:space:]]Trajectory[[:space:]]Prediction/5a87e2e4-dac0-41f3-9966-85985369c6c6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sparse[[:space:]]Point[[:space:]]Guided[[:space:]]3D[[:space:]]Lane[[:space:]]Detection/6f2348f9-ca1f-4bc7-8757-9bb55f6bb3b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Sparse[[:space:]]Sampling[[:space:]]Transformer[[:space:]]with[[:space:]]Uncertainty-Driven[[:space:]]Ranking[[:space:]]for[[:space:]]Unified[[:space:]]Removal[[:space:]]of[[:space:]]Raindrops[[:space:]]and[[:space:]]Rain[[:space:]]Streaks/32196ea3-316d-4633-8abd-b15ee0703f11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SparseBEV_[[:space:]]High-Performance[[:space:]]Sparse[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]from[[:space:]]Multi-Camera[[:space:]]Videos/96a859b1-e12e-4281-b4c4-b8ea44fdbf01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SparseDet_[[:space:]]Improving[[:space:]]Sparsely[[:space:]]Annotated[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Pseudo-positive[[:space:]]Mining/5aaa9a89-cdc6-4e0a-94f3-a3b82685ba6a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SparseFusion_[[:space:]]Fusing[[:space:]]Multi-Modal[[:space:]]Sparse[[:space:]]Representations[[:space:]]for[[:space:]]Multi-Sensor[[:space:]]3D[[:space:]]Object[[:space:]]Detection/280aa12d-c539-41fe-b78e-1bcf7c1aa152_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SparseMAE_[[:space:]]Sparse[[:space:]]Training[[:space:]]Meets[[:space:]]Masked[[:space:]]Autoencoders/a542f651-0510-4853-91de-1141d304dc5f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SparseNeRF_[[:space:]]Distilling[[:space:]]Depth[[:space:]]Ranking[[:space:]]for[[:space:]]Few-shot[[:space:]]Novel[[:space:]]View[[:space:]]Synthesis/0d56c181-4428-4dcb-b7bc-71667b8724d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatial[[:space:]]Self-Distillation[[:space:]]for[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Inaccurate[[:space:]]Bounding[[:space:]]Boxes/d223c2f5-fe18-4502-b91f-9060d747649d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatial-Aware[[:space:]]Token[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Object[[:space:]]Localization/f549b4c7-58f5-4874-8f3f-2978b440c9a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatially[[:space:]]and[[:space:]]Spectrally[[:space:]]Consistent[[:space:]]Deep[[:space:]]Functional[[:space:]]Maps/38616010-b1fd-470b-8376-180cbf13b5f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatially-Adaptive[[:space:]]Feature[[:space:]]Modulation[[:space:]]for[[:space:]]Efficient[[:space:]]Image[[:space:]]Super-Resolution/8203ecb7-a9b9-46de-8112-1277f4eeac63_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatio-Temporal[[:space:]]Crop[[:space:]]Aggregation[[:space:]]for[[:space:]]Video[[:space:]]Representation[[:space:]]Learning/6732d370-6205-414c-986c-288f16a786f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatio-Temporal[[:space:]]Domain[[:space:]]Awareness[[:space:]]for[[:space:]]Multi-Agent[[:space:]]Collaborative[[:space:]]Perception/fe9d96d8-4150-4b6d-a27c-f4b80a133514_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spatio-temporal[[:space:]]Prompting[[:space:]]Network[[:space:]]for[[:space:]]Robust[[:space:]]Video[[:space:]]Feature[[:space:]]Extraction/6b38d58c-2dfe-4bda-8611-478cf7d86d34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spectral[[:space:]]Graphormer_[[:space:]]Spectral[[:space:]]Graph-Based[[:space:]]Transformer[[:space:]]for[[:space:]]Egocentric[[:space:]]Two-Hand[[:space:]]Reconstruction[[:space:]]using[[:space:]]Multi-View[[:space:]]Color[[:space:]]Images/dbd672e5-b6e6-48ea-861d-97b4378c8ca3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spectrum-guided[[:space:]]Multi-granularity[[:space:]]Referring[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/dc8d764a-9214-4358-ad24-c2007b58a08c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Speech2Lip_[[:space:]]High-fidelity[[:space:]]Speech[[:space:]]to[[:space:]]Lip[[:space:]]Generation[[:space:]]by[[:space:]]Learning[[:space:]]from[[:space:]]a[[:space:]]Short[[:space:]]Video/b8d7dfde-aec2-4ac3-9ef3-b8789d6534b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Speech4Mesh_[[:space:]]Speech-Assisted[[:space:]]Monocular[[:space:]]3D[[:space:]]Facial[[:space:]]Reconstruction[[:space:]]for[[:space:]]Speech-Driven[[:space:]]3D[[:space:]]Facial[[:space:]]Animation/236896c7-9d5b-41e0-8f54-b3702ca337d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spherical[[:space:]]Space[[:space:]]Feature[[:space:]]Decomposition[[:space:]]for[[:space:]]Guided[[:space:]]Depth[[:space:]]Map[[:space:]]Super-Resolution/55278fbf-9750-4827-94bb-81181843441e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SpinCam_[[:space:]]High-Speed[[:space:]]Imaging[[:space:]]via[[:space:]]a[[:space:]]Rotating[[:space:]]Point-Spread[[:space:]]Function/280659a7-98c7-41f5-b75f-aea246ac862a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SportsMOT_[[:space:]]A[[:space:]]Large[[:space:]]Multi-Object[[:space:]]Tracking[[:space:]]Dataset[[:space:]]in[[:space:]]Multiple[[:space:]]Sports[[:space:]]Scenes/4b8d0490-50f4-45bc-9fad-a0ac515c1ca0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Spurious[[:space:]]Features[[:space:]]Everywhere[[:space:]]-[[:space:]]Large-Scale[[:space:]]Detection[[:space:]]of[[:space:]]Harmful[[:space:]]Spurious[[:space:]]Features[[:space:]]in[[:space:]]ImageNet/8d4ee68a-1062-4c0b-81d9-4f01bf36823f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Stabilizing[[:space:]]Visual[[:space:]]Reinforcement[[:space:]]Learning[[:space:]]via[[:space:]]Asymmetric[[:space:]]Interactive[[:space:]]Cooperation/3ce1c749-00ff-4ed7-ac11-5982848d2220_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Stable[[:space:]]Cluster[[:space:]]Discrimination[[:space:]]for[[:space:]]Deep[[:space:]]Clustering/9f881270-2ff5-4f87-a2db-8feef1f47071_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Stable[[:space:]]and[[:space:]]Causal[[:space:]]Inference[[:space:]]for[[:space:]]Discriminative[[:space:]]Self-supervised[[:space:]]Deep[[:space:]]Visual[[:space:]]Representations/fdb72a4a-94ed-4829-ab6f-8cf80d64f7f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StableVideo_[[:space:]]Text-driven[[:space:]]Consistency-aware[[:space:]]Diffusion[[:space:]]Video[[:space:]]Editing/bcc42e2b-6089-4c27-a5c2-cdec53620ca0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StageInteractor_[[:space:]]Query-based[[:space:]]Object[[:space:]]Detector[[:space:]]with[[:space:]]Cross-stage[[:space:]]Interaction/0e05e637-ac63-484f-9954-22b778c7c120_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Steered[[:space:]]Diffusion_[[:space:]]A[[:space:]]Generalized[[:space:]]Framework[[:space:]]for[[:space:]]Plug-and-Play[[:space:]]Conditional[[:space:]]Image[[:space:]]Synthesis/396c3711-d204-4df2-8e69-9bdf93d4fbff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StegaNeRF_[[:space:]]Embedding[[:space:]]Invisible[[:space:]]Information[[:space:]]within[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/3c8d0cf6-dc7d-44ba-a804-37f576c9cfcc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Stochastic[[:space:]]Segmentation[[:space:]]with[[:space:]]Conditional[[:space:]]Categorical[[:space:]]Diffusion[[:space:]]Models/36fc9ee6-a13a-40ed-bd76-54644b9d43ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Story[[:space:]]Visualization[[:space:]]by[[:space:]]Online[[:space:]]Text[[:space:]]Augmentation[[:space:]]with[[:space:]]Context[[:space:]]Memory/b1180507-958c-45ce-9ffe-139cc0187a6e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Strata-NeRF[[:space:]]_[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Stratified[[:space:]]Scenes/246a74f2-93a7-4d68-8843-6238c9108476_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Strip-MLP_[[:space:]]Efficient[[:space:]]Token[[:space:]]Interaction[[:space:]]for[[:space:]]Vision[[:space:]]MLP/5c4682dc-11cf-43c9-b82f-059171805e73_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Strivec_[[:space:]]Sparse[[:space:]]Tri-Vector[[:space:]]Radiance[[:space:]]Fields/35b59dcb-1ce5-40a4-83d2-07192fb62870_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Structural[[:space:]]Alignment[[:space:]]for[[:space:]]Network[[:space:]]Pruning[[:space:]]through[[:space:]]Partial[[:space:]]Regularization/bbe97c12-8d15-479d-81ec-fff8ed1209f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Structure[[:space:]]Invariant[[:space:]]Transformation[[:space:]]for[[:space:]]better[[:space:]]Adversarial[[:space:]]Transferability/303ed349-48a8-4393-8e2e-88000a7d20c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Structure[[:space:]]and[[:space:]]Content-Guided[[:space:]]Video[[:space:]]Synthesis[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/4fea04cc-819a-4990-99d9-0ccc1f6c300d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Structure-Aware[[:space:]]Surface[[:space:]]Reconstruction[[:space:]]via[[:space:]]Primitive[[:space:]]Assembly/4d750dca-9a8b-44f9-b884-62a2d6314ac8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Studying[[:space:]]How[[:space:]]to[[:space:]]Efficiently[[:space:]]and[[:space:]]Effectively[[:space:]]Guide[[:space:]]Models[[:space:]]with[[:space:]]Explanations/88a53c18-99d0-4b4c-b0b2-16a2556a29ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyleDiffusion_[[:space:]]Controllable[[:space:]]Disentangled[[:space:]]Style[[:space:]]Transfer[[:space:]]via[[:space:]]Diffusion[[:space:]]Models/fea588fd-ec6f-4543-ad58-5dd9ea412b01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyleDomain_[[:space:]]Efficient[[:space:]]and[[:space:]]Lightweight[[:space:]]Parameterizations[[:space:]]of[[:space:]]StyleGAN[[:space:]]for[[:space:]]One-shot[[:space:]]and[[:space:]]Few-shot[[:space:]]Domain[[:space:]]Adaptation/b30ac484-13c8-4c42-8711-cf47c43f4550_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyleGANEX_[[:space:]]StyleGAN-Based[[:space:]]Manipulation[[:space:]]Beyond[[:space:]]Cropped[[:space:]]Aligned[[:space:]]Faces/71e23834-fb4f-4c96-b627-797882a6ea6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyleInV_[[:space:]]A[[:space:]]Temporal[[:space:]]Style[[:space:]]Modulated[[:space:]]Inversion[[:space:]]Network[[:space:]]for[[:space:]]Unconditional[[:space:]]Video[[:space:]]Generation/ee18bf56-6e38-4711-9114-1ec06f896ae1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StyleLipSync_[[:space:]]Style-based[[:space:]]Personalized[[:space:]]Lip-sync[[:space:]]Video[[:space:]]Generation/35be9e19-3939-4344-a259-82c8df74d69e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/StylerDALLE_[[:space:]]Language-Guided[[:space:]]Style[[:space:]]Transfer[[:space:]]Using[[:space:]]a[[:space:]]Vector-Quantized[[:space:]]Tokenizer[[:space:]]of[[:space:]]a[[:space:]]Large-Scale[[:space:]]Generative[[:space:]]Model/39699168-4600-4fd3-a205-b0cf85e92a6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SuS-X_[[:space:]]Training-Free[[:space:]]Name-Only[[:space:]]Transfer[[:space:]]of[[:space:]]Vision-Language[[:space:]]Models/000612a7-f65d-40f9-b28d-4f843782dfff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Subclass-balancing[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Long-tailed[[:space:]]Recognition/bc05a493-9765-4ef8-8b9c-fc9c90911ac5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SupFusion_[[:space:]]Supervised[[:space:]]LiDAR-Camera[[:space:]]Fusion[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection/92e5510f-e140-45f8-8597-b810b193a66f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Supervised[[:space:]]Homography[[:space:]]Learning[[:space:]]with[[:space:]]Realistic[[:space:]]Dataset[[:space:]]Generation/56098eb5-516e-4294-9a4a-ac98eb40dc0f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Surface[[:space:]]Extraction[[:space:]]from[[:space:]]Neural[[:space:]]Unsigned[[:space:]]Distance[[:space:]]Fields/a0829e7a-732d-4761-bffa-ad35624b76ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Surface[[:space:]]Normal[[:space:]]Clustering[[:space:]]for[[:space:]]Implicit[[:space:]]Representation[[:space:]]of[[:space:]]Manhattan[[:space:]]Scenes/f357e4d1-f9a1-4f72-b3d1-9e59cc154adf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SurfsUP_[[:space:]]Learning[[:space:]]Fluid[[:space:]]Simulation[[:space:]]for[[:space:]]Novel[[:space:]]Surfaces/e47c0301-5f63-4201-a2ee-b44e21fee063_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SurroundOcc_[[:space:]]Multi-camera[[:space:]]3D[[:space:]]Occupancy[[:space:]]Prediction[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/0c733c34-bebb-47b3-8e25-0904ab5a04e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SwiftFormer_[[:space:]]Efficient[[:space:]]Additive[[:space:]]Attention[[:space:]]for[[:space:]]Transformer-based[[:space:]]Real-time[[:space:]]Mobile[[:space:]]Vision[[:space:]]Applications/3fb2491f-940c-4884-86b5-1f7be4430698_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SwinLSTM_[[:space:]]Improving[[:space:]]Spatiotemporal[[:space:]]Prediction[[:space:]]Accuracy[[:space:]]using[[:space:]]Swin[[:space:]]Transformer[[:space:]]and[[:space:]]LSTM/0bea32db-db74-4a9f-85a7-6fe6526e5e78_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/SynBody_[[:space:]]Synthetic[[:space:]]Dataset[[:space:]]with[[:space:]]Layered[[:space:]]Human[[:space:]]Models[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Perception[[:space:]]and[[:space:]]Modeling/f251ae55-2ca7-4578-ad92-a51119c290a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Synchronize[[:space:]]Feature[[:space:]]Extracting[[:space:]]and[[:space:]]Matching_[[:space:]]A[[:space:]]Single[[:space:]]Branch[[:space:]]Framework[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Tracking/d13aa22d-32ff-439b-9d2f-7deb1ed92309_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Synthesizing[[:space:]]Diverse[[:space:]]Human[[:space:]]Motions[[:space:]]in[[:space:]]3D[[:space:]]Indoor[[:space:]]Scenes/39947c2e-9c93-4c2e-8f0c-e7f92b1d1960_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TALL_[[:space:]]Thumbnail[[:space:]]Layout[[:space:]]for[[:space:]]Deepfake[[:space:]]Video[[:space:]]Detection/b759f55b-503f-4164-9793-b1ffe02056fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TAPIR_[[:space:]]Tracking[[:space:]]Any[[:space:]]Point[[:space:]]with[[:space:]]Per-Frame[[:space:]]Initialization[[:space:]]and[[:space:]]Temporal[[:space:]]Refinement/f0fe9ecb-f819-46e1-ba9a-de70f0257649_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TARGET_[[:space:]]Federated[[:space:]]Class-Continual[[:space:]]Learning[[:space:]]via[[:space:]]Exemplar-Free[[:space:]]Distillation/ff5641a5-17a3-470a-a519-ccac898db108_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TCOVIS_[[:space:]]Temporally[[:space:]]Consistent[[:space:]]Online[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation/806a2113-930c-496f-99e8-ccd02b9307f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TEMPO_[[:space:]]Efficient[[:space:]]Multi-View[[:space:]]Pose[[:space:]]Estimation,[[:space:]]Tracking,[[:space:]]and[[:space:]]Forecasting/5ca0ee4c-d0bf-43fa-950c-91f169b1d361_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TF-ICON_[[:space:]]Diffusion-Based[[:space:]]Training-Free[[:space:]]Cross-Domain[[:space:]]Image[[:space:]]Composition/082c8f3d-29c9-4e4b-9ab4-3aa55681552e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TIFA_[[:space:]]Accurate[[:space:]]and[[:space:]]Interpretable[[:space:]]Text-to-Image[[:space:]]Faithfulness[[:space:]]Evaluation[[:space:]]with[[:space:]]Question[[:space:]]Answering/ae7f1ffe-fe67-4917-84ae-5c196628c6e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TIJO_[[:space:]]Trigger[[:space:]]Inversion[[:space:]]with[[:space:]]Joint[[:space:]]Optimization[[:space:]]for[[:space:]]Defending[[:space:]]Multimodal[[:space:]]Backdoored[[:space:]]Models/7bfffc83-27ef-4c73-aa19-9a15a6ee13fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TM2D_[[:space:]]Bimodality[[:space:]]Driven[[:space:]]3D[[:space:]]Dance[[:space:]]Generation[[:space:]]via[[:space:]]Music-Text[[:space:]]Integration/1834d63b-e3e0-4b42-9900-788dbb6bee22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TMA_[[:space:]]Temporal[[:space:]]Motion[[:space:]]Aggregation[[:space:]]for[[:space:]]Event-based[[:space:]]Optical[[:space:]]Flow/9d269c18-5110-4a81-a1b9-317c1455548e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TMR_[[:space:]]Text-to-Motion[[:space:]]Retrieval[[:space:]]Using[[:space:]]Contrastive[[:space:]]3D[[:space:]]Human[[:space:]]Motion[[:space:]]Synthesis/d1f3cbdb-e07a-4952-8052-a9aa7bed81da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TORE_[[:space:]]Token[[:space:]]Reduction[[:space:]]for[[:space:]]Efficient[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery[[:space:]]with[[:space:]]Transformer/2d762d0d-bad4-403d-98f7-4210b7f7e39d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TRM-UAP_[[:space:]]Enhancing[[:space:]]the[[:space:]]Transferability[[:space:]]of[[:space:]]Data-Free[[:space:]]Universal[[:space:]]Adversarial[[:space:]]Perturbation[[:space:]]via[[:space:]]Truncated[[:space:]]Ratio[[:space:]]Maximization/6ac05fcd-de6a-4657-a3f2-a897a887dcbb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Take-A-Photo_[[:space:]]3D-to-2D[[:space:]]Generative[[:space:]]Pre-training[[:space:]]of[[:space:]]Point[[:space:]]Cloud[[:space:]]Models/e286768b-a5b2-4ea3-a9f1-4d06c6cc68f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Talking[[:space:]]Head[[:space:]]Generation[[:space:]]with[[:space:]]Probabilistic[[:space:]]Audio-to-Visual[[:space:]]Diffusion[[:space:]]Priors/7d213d41-b6ba-4fcf-a556-dbfb8dc18cb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Taming[[:space:]]Contrast[[:space:]]Maximization[[:space:]]for[[:space:]]Learning[[:space:]]Sequential,[[:space:]]Low-latency,[[:space:]]Event-based[[:space:]]Optical[[:space:]]Flow/ff7524e3-aa7d-4baa-a842-04712fa34f60_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tangent[[:space:]]Model[[:space:]]Composition[[:space:]]for[[:space:]]Ensembling[[:space:]]and[[:space:]]Continual[[:space:]]Fine-tuning/1ae449c1-93a5-4652-bb27-6a83c99dd85c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tangent[[:space:]]Sampson[[:space:]]Error_[[:space:]]Fast[[:space:]]Approximate[[:space:]]Two-view[[:space:]]Reprojection[[:space:]]Error[[:space:]]for[[:space:]]Central[[:space:]]Camera[[:space:]]Models/636b4c99-6986-4080-bf56-9a578e79f107_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Task[[:space:]]Agnostic[[:space:]]Restoration[[:space:]]of[[:space:]]Natural[[:space:]]Video[[:space:]]Dynamics/e82ebeb9-3e53-42eb-9d7f-02ba103a6afc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Task-Oriented[[:space:]]Multi-Modal[[:space:]]Mutual[[:space:]]Leaning[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/b6090cd5-35b5-49d1-9775-853efd91c055_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Task-aware[[:space:]]Adaptive[[:space:]]Learning[[:space:]]for[[:space:]]Cross-domain[[:space:]]Few-shot[[:space:]]Learning/d73ef7d6-b7de-4511-b8bf-4b4e20276122_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TaskExpert_[[:space:]]Dynamically[[:space:]]Assembling[[:space:]]Multi-Task[[:space:]]Representations[[:space:]]with[[:space:]]Memorial[[:space:]]Mixture-of-Experts/2e3b51ba-4260-4b59-a865-225826178569_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Taxonomy[[:space:]]Adaptive[[:space:]]Cross-Domain[[:space:]]Adaptation[[:space:]]in[[:space:]]Medical[[:space:]]Imaging[[:space:]]via[[:space:]]Optimization[[:space:]]Trajectory[[:space:]]Distillation/1f4cb2e9-7a0d-4b5a-99a2-329016955f96_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TeD-SPAD_[[:space:]]Temporal[[:space:]]Distinctiveness[[:space:]]for[[:space:]]Self-Supervised[[:space:]]Privacy-Preservation[[:space:]]for[[:space:]]Video[[:space:]]Anomaly[[:space:]]Detection/a12231ff-0413-4e22-8eb5-06598dd6fed6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Teaching[[:space:]]CLIP[[:space:]]to[[:space:]]Count[[:space:]]to[[:space:]]Ten/56d47f2c-86af-49f0-9e72-8decf5e81c8a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tem-Adapter_[[:space:]]Adapting[[:space:]]Image-Text[[:space:]]Pretraining[[:space:]]for[[:space:]]Video[[:space:]]Question[[:space:]]Answer/4a5fd117-e0c8-457c-9985-c705ba827466_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Template[[:space:]]Inversion[[:space:]]Attack[[:space:]]against[[:space:]]Face[[:space:]]Recognition[[:space:]]Systems[[:space:]]using[[:space:]]3D[[:space:]]Face[[:space:]]Reconstruction/203e5250-94d9-4833-ba8d-b9c548251baa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Template-guided[[:space:]]Hierarchical[[:space:]]Feature[[:space:]]Restoration[[:space:]]for[[:space:]]Anomaly[[:space:]]Detection/b5ce36c9-1639-4be6-bce9-ead12d7207f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Temporal[[:space:]]Collection[[:space:]]and[[:space:]]Distribution[[:space:]]for[[:space:]]Referring[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/3b08deeb-33eb-47b0-94a4-e0af89436d2d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Temporal[[:space:]]Enhanced[[:space:]]Training[[:space:]]of[[:space:]]Multi-view[[:space:]]3D[[:space:]]Object[[:space:]]Detector[[:space:]]via[[:space:]]Historical[[:space:]]Object[[:space:]]Prediction/553f0662-4b8d-4936-ab47-ad291ff03dd2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Temporal-Coded[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks[[:space:]]with[[:space:]]Dynamic[[:space:]]Firing[[:space:]]Threshold_[[:space:]]Learning[[:space:]]with[[:space:]]Event-Driven[[:space:]]Backpropagation/6f1f26ae-7b19-4337-978a-c5b559913901_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Test[[:space:]]Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Blind[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment/4f23b86f-f4db-4257-aee1-6b77c4f73657_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Test-time[[:space:]]Personalizable[[:space:]]Forecasting[[:space:]]of[[:space:]]3D[[:space:]]Human[[:space:]]Poses/205eafc7-d672-437b-9b70-81a2995fa72d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tetra-NeRF_[[:space:]]Representing[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]Using[[:space:]]Tetrahedra/8c410ac7-5eb0-4f17-a4bb-daadb2446edc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TexFusion_[[:space:]]Synthesizing[[:space:]]3D[[:space:]]Textures[[:space:]]with[[:space:]]Text-Guided[[:space:]]Image[[:space:]]Diffusion[[:space:]]Models/698eb5a8-8d74-4b19-b9a7-1fe068b4d1fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Text-Conditioned[[:space:]]Sampling[[:space:]]Framework[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]with[[:space:]]Masked[[:space:]]Generative[[:space:]]Models/c18683e6-ab91-4cf4-bf6b-203837440b00_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Text-Driven[[:space:]]Generative[[:space:]]Domain[[:space:]]Adaptation[[:space:]]with[[:space:]]Spectral[[:space:]]Consistency[[:space:]]Regularization/e13295e4-a033-4d38-a3a5-88f3b3385e81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Text2Performer_[[:space:]]Text-Driven[[:space:]]Human[[:space:]]Video[[:space:]]Generation/45c9b878-7941-4c19-aa7b-acd77ad00875_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Text2Room_[[:space:]]Extracting[[:space:]]Textured[[:space:]]3D[[:space:]]Meshes[[:space:]]from[[:space:]]2D[[:space:]]Text-to-Image[[:space:]]Models/ec345001-5311-431e-90bd-6b91f4b47f2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Text2Tex_[[:space:]]Text-driven[[:space:]]Texture[[:space:]]Synthesis[[:space:]]via[[:space:]]Diffusion[[:space:]]Models/8023656a-1169-49dd-b8cd-4db39941377d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Text2Video-Zero_[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]are[[:space:]]Zero-Shot[[:space:]]Video[[:space:]]Generators/56c5c84c-dd92-4857-8492-3e83266b25fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TextManiA_[[:space:]]Enriching[[:space:]]Visual[[:space:]]Feature[[:space:]]by[[:space:]]Text-driven[[:space:]]Manifold[[:space:]]Augmentation/21a8728d-cd5e-44be-95fc-d79147ae763b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TextPSG_[[:space:]]Panoptic[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]from[[:space:]]Textual[[:space:]]Descriptions/a412d306-b042-443c-958a-505db74fd5e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Texture[[:space:]]Generation[[:space:]]on[[:space:]]3D[[:space:]]Meshes[[:space:]]with[[:space:]]Point-UV[[:space:]]Diffusion/295801ea-f88d-446b-aeea-24167c36963d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Texture[[:space:]]Learning[[:space:]]Domain[[:space:]]Randomization[[:space:]]for[[:space:]]Domain[[:space:]]Generalized[[:space:]]Segmentation/5378f03d-031c-45ab-a522-6e62326e01af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Devil[[:space:]]is[[:space:]]in[[:space:]]the[[:space:]]Crack[[:space:]]Orientation_[[:space:]]A[[:space:]]New[[:space:]]Perspective[[:space:]]for[[:space:]]Crack[[:space:]]Detection/b4b142a9-63c6-4b31-9bbd-a4e1e31c2d72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Devil[[:space:]]is[[:space:]]in[[:space:]]the[[:space:]]Upsampling_[[:space:]]Architectural[[:space:]]Decisions[[:space:]]Made[[:space:]]Simpler[[:space:]]for[[:space:]]Denoising[[:space:]]with[[:space:]]Deep[[:space:]]Image[[:space:]]Prior/6b883d6c-8e38-4d25-996f-79941abad307_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Effectiveness[[:space:]]of[[:space:]]MAE[[:space:]]Pre-Pretraining[[:space:]]for[[:space:]]Billion-Scale[[:space:]]Pretraining/5b230860-aa61-4f33-9fe7-2407ba62f713_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Euclidean[[:space:]]Space[[:space:]]is[[:space:]]Evil_[[:space:]]Hyperbolic[[:space:]]Attribute[[:space:]]Editing[[:space:]]for[[:space:]]Few-shot[[:space:]]Image[[:space:]]Generation/e705b187-21ba-4812-b0e4-1feb752f1cdd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Making[[:space:]]and[[:space:]]Breaking[[:space:]]of[[:space:]]Camouflage/b706bafb-dae1-4647-ad2f-702e05fcaa82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Perils[[:space:]]of[[:space:]]Learning[[:space:]]From[[:space:]]Unlabeled[[:space:]]Data_[[:space:]]Backdoor[[:space:]]Attacks[[:space:]]on[[:space:]]Semi-supervised[[:space:]]Learning/b27d12fb-927b-406e-bce3-95c8be702c01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Power[[:space:]]of[[:space:]]Sound[[:space:]](TPoS)_[[:space:]]Audio[[:space:]]Reactive[[:space:]]Video[[:space:]]Generation[[:space:]]with[[:space:]]Stable[[:space:]]Diffusion/3b9298a9-5799-4904-96da-bef6d287733c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Stable[[:space:]]Signature_[[:space:]]Rooting[[:space:]]Watermarks[[:space:]]in[[:space:]]Latent[[:space:]]Diffusion[[:space:]]Models/c2da89cd-17c8-420c-b57b-7af434285ab7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Unreasonable[[:space:]]Effectiveness[[:space:]]of[[:space:]]Large[[:space:]]Language-Vision[[:space:]]Models[[:space:]]for[[:space:]]Source-Free[[:space:]]Video[[:space:]]Domain[[:space:]]Adaptation/c596fbdc-cd07-43b2-9a08-9301d722c491_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/The[[:space:]]Victim[[:space:]]and[[:space:]]The[[:space:]]Beneficiary_[[:space:]]Exploiting[[:space:]]a[[:space:]]Poisoned[[:space:]]Model[[:space:]]to[[:space:]]Train[[:space:]]a[[:space:]]Clean[[:space:]]Model[[:space:]]on[[:space:]]Poisoned[[:space:]]Data/fdfec7d9-d3ec-4468-9f42-c0ec2550345d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Theoretical[[:space:]]and[[:space:]]Numerical[[:space:]]Analysis[[:space:]]of[[:space:]]3D[[:space:]]Reconstruction[[:space:]]Using[[:space:]]Point[[:space:]]and[[:space:]]Line[[:space:]]Incidences/5fbfedc8-4ebc-4e53-8cf0-a4029a8a49ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Thinking[[:space:]]Image[[:space:]]Color[[:space:]]Aesthetics[[:space:]]Assessment_[[:space:]]Models,[[:space:]]Datasets[[:space:]]and[[:space:]]Benchmarks/42e3e8c9-44a4-4bef-b505-a2c5b22ff20f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TiDAL_[[:space:]]Learning[[:space:]]Training[[:space:]]Dynamics[[:space:]]for[[:space:]]Active[[:space:]]Learning/cbb37e36-1eb0-42b7-b0b1-3b2793b016b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TiDy-PSFs_[[:space:]]Computational[[:space:]]Imaging[[:space:]]with[[:space:]]Time-Averaged[[:space:]]Dynamic[[:space:]]Point-Spread-Functions/0fbb2c88-7b44-409c-9798-461b2922052e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tiled[[:space:]]Multiplane[[:space:]]Images[[:space:]]for[[:space:]]Practical[[:space:]]3D[[:space:]]Photography/332ad0e6-c530-4dba-82ae-058961745db1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Time[[:space:]]Does[[:space:]]Tell_[[:space:]]Self-Supervised[[:space:]]Time-Tuning[[:space:]]of[[:space:]]Dense[[:space:]]Image[[:space:]]Representations/1b69a328-51e9-4f52-9cd0-ceb8e6de4113_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Time-to-Contact[[:space:]]Map[[:space:]]by[[:space:]]Joint[[:space:]]Estimation[[:space:]]of[[:space:]]Up-to-Scale[[:space:]]Inverse[[:space:]]Depth[[:space:]]and[[:space:]]Global[[:space:]]Motion[[:space:]]using[[:space:]]a[[:space:]]Single[[:space:]]Event[[:space:]]Camera/0f2857ef-2537-43f4-b49d-a3319a025266_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tiny[[:space:]]Updater_[[:space:]]Towards[[:space:]]Efficient[[:space:]]Neural[[:space:]]Network-Driven[[:space:]]Software[[:space:]]Updating/23a6116f-5b95-4dcd-97e5-95528a6d1fff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TinyCLIP_[[:space:]]CLIP[[:space:]]Distillation[[:space:]]via[[:space:]]Affinity[[:space:]]Mimicking[[:space:]]and[[:space:]]Weight[[:space:]]Inheritance/7a0fee57-111a-443f-8a63-68f0c3d1c0b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/To[[:space:]]Adapt[[:space:]]or[[:space:]]Not[[:space:]]to[[:space:]]Adapt_[[:space:]]Real-Time[[:space:]]Adaptation[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/d5cc5a6e-449b-4085-98b6-42b0071a0627_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Token-Label[[:space:]]Alignment[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/ef46bf1f-34a3-4e70-bf1a-b3f5f27b821d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Too[[:space:]]Large;[[:space:]]Data[[:space:]]Reduction[[:space:]]for[[:space:]]Vision-Language[[:space:]]Pre-Training/bf4cd6c9-c260-42c2-88a5-f694382bd718_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ToonTalker_[[:space:]]Cross-Domain[[:space:]]Face[[:space:]]Reenactment/d3014894-36d6-4bc9-bdc4-403449ecf9bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TopoSeg_[[:space:]]Topology-Aware[[:space:]]Nuclear[[:space:]]Instance[[:space:]]Segmentation/a134e8ff-997a-47c5-810d-0149806d46fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Total-Recon_[[:space:]]Deformable[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]for[[:space:]]Embodied[[:space:]]View[[:space:]]Synthesis/49fdb57e-3f25-4f0f-85ac-577ce80d4f48_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Toward[[:space:]]Multi-Granularity[[:space:]]Decision-Making_[[:space:]]Explicit[[:space:]]Visual[[:space:]]Reasoning[[:space:]]with[[:space:]]Hierarchical[[:space:]]Knowledge/389ce227-ce67-4e7f-b3a7-67bbdf3737b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Toward[[:space:]]Unsupervised[[:space:]]Realistic[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/7aa5523f-be70-4113-8886-4fd1bb97fc1f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Attack-tolerant[[:space:]]Federated[[:space:]]Learning[[:space:]]via[[:space:]]Critical[[:space:]]Parameter[[:space:]]Analysis/02507d65-3ffa-43d1-a132-83fd9f04eaa0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Authentic[[:space:]]Face[[:space:]]Restoration[[:space:]]with[[:space:]]Iterative[[:space:]]Diffusion[[:space:]]Models[[:space:]]and[[:space:]]Beyond/15935391-63d5-4fdb-8bfc-918cd38734ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Better[[:space:]]Robustness[[:space:]]against[[:space:]]Common[[:space:]]Corruptions[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/59dcb24f-1686-40fa-9d89-bb27e72b782d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Building[[:space:]]More[[:space:]]Robust[[:space:]]Models[[:space:]]with[[:space:]]Frequency[[:space:]]Bias/bba1b8e8-55bc-4955-94e4-08c6ea66cfc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Content-based[[:space:]]Pixel[[:space:]]Retrieval[[:space:]]in[[:space:]]Revisited[[:space:]]Oxford[[:space:]]and[[:space:]]Paris/48cbf299-cb48-4284-a5e0-e11574316cd7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Deeply[[:space:]]Unified[[:space:]]Depth-aware[[:space:]]Panoptic[[:space:]]Segmentation[[:space:]]with[[:space:]]Bi-directional[[:space:]]Guidance[[:space:]]Learning/e62dbcc7-1d0a-4f47-a83b-273f11111b81_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Effective[[:space:]]Instance[[:space:]]Discrimination[[:space:]]Contrastive[[:space:]]Loss[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/d7d92e7a-0e90-4a85-8061-0c6b5875f070_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Fair[[:space:]]and[[:space:]]Comprehensive[[:space:]]Comparisons[[:space:]]for[[:space:]]Image-Based[[:space:]]3D[[:space:]]Object[[:space:]]Detection/919f3668-f220-4fa7-8982-072f75f2b2c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Fairness-aware[[:space:]]Adversarial[[:space:]]Network[[:space:]]Pruning/1eb44924-b623-45a1-b6e3-faeebcea0772_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]General[[:space:]]Low-Light[[:space:]]Raw[[:space:]]Noise[[:space:]]Synthesis[[:space:]]and[[:space:]]Modeling/d0080f03-1ea1-4222-9cf1-7905e1b14900_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Generic[[:space:]]Image[[:space:]]Manipulation[[:space:]]Detection[[:space:]]with[[:space:]]Weakly-Supervised[[:space:]]Self-Consistency[[:space:]]Learning/d0508e74-2a3e-4371-a769-b6a0b23be1f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Geospatial[[:space:]]Foundation[[:space:]]Models[[:space:]]via[[:space:]]Continual[[:space:]]Pretraining/f5d37ccf-f124-4c34-8689-3ac893ac6c26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Grand[[:space:]]Unified[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Unsupervised[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification/9e3a2586-6021-484c-9fc1-4dba6bb35ee0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]High-Fidelity[[:space:]]Text-Guided[[:space:]]3D[[:space:]]Face[[:space:]]Generation[[:space:]]and[[:space:]]Manipulation[[:space:]]Using[[:space:]]only[[:space:]]Images/ea19677c-655b-4353-a6ef-f64b901df36f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]High-Quality[[:space:]]Specular[[:space:]]Highlight[[:space:]]Removal[[:space:]]by[[:space:]]Leveraging[[:space:]]Large-Scale[[:space:]]Synthetic[[:space:]]Data/e4b28c26-5c3e-4a4a-abef-0ac6e5cb1f5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Improved[[:space:]]Input[[:space:]]Masking[[:space:]]for[[:space:]]Convolutional[[:space:]]Neural[[:space:]]Networks/2e54d932-3136-4f85-9d8a-e9d9c9b7fcdc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Inadequately[[:space:]]Pre-trained[[:space:]]Models[[:space:]]in[[:space:]]Transfer[[:space:]]Learning/37781ea1-8759-412e-90db-2b76b1876757_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Instance-adaptive[[:space:]]Inference[[:space:]]for[[:space:]]Federated[[:space:]]Learning/effcd160-8f35-4c75-822f-7f8beb31b21a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Memory-[[:space:]]and[[:space:]]Time-Efficient[[:space:]]Backpropagation[[:space:]]for[[:space:]]Training[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks/66075ff2-f851-4160-b628-a7149bc5465c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Models[[:space:]]that[[:space:]]Can[[:space:]]See[[:space:]]and[[:space:]]Read/e30fd25f-1e13-42d1-9f38-c668503e39b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Multi-Layered[[:space:]]3D[[:space:]]Garments[[:space:]]Animation/d0b1af1b-89f6-440f-9974-55a489065df8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Nonlinear-Motion-Aware[[:space:]]and[[:space:]]Occlusion-Robust[[:space:]]Rolling[[:space:]]Shutter[[:space:]]Correction/8f643363-6017-4a11-925e-658f3230a863_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Open-Set[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]Utilizing[[:space:]]the[[:space:]]Wisdom[[:space:]]of[[:space:]]Crowds[[:space:]]in[[:space:]]Entropy[[:space:]]Minimization/915dab93-b945-4ea6-9759-220d4cb65058_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Open-Vocabulary[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation/f0d9d521-0a5a-458f-bd3a-9f4a706333d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Real-World[[:space:]]Burst[[:space:]]Image[[:space:]]Super-Resolution_[[:space:]]Benchmark[[:space:]]and[[:space:]]Method/8f14d0e4-a8ef-4066-8bfd-5810cf451908_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Realistic[[:space:]]Evaluation[[:space:]]of[[:space:]]Industrial[[:space:]]Continual[[:space:]]Learning[[:space:]]Scenarios[[:space:]]with[[:space:]]an[[:space:]]Emphasis[[:space:]]on[[:space:]]Energy[[:space:]]Consumption[[:space:]]and[[:space:]]Computational[[:space:]]Footprint/015a0289-0201-44dc-9ecb-d24f89d8a220_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Robust[[:space:]]Model[[:space:]]Watermark[[:space:]]via[[:space:]]Reducing[[:space:]]Parametric[[:space:]]Vulnerability/255f44d2-23f9-4d14-8305-99d86c8e8c04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Robust[[:space:]]and[[:space:]]Smooth[[:space:]]3D[[:space:]]Multi-Person[[:space:]]Pose[[:space:]]Estimation[[:space:]]from[[:space:]]Monocular[[:space:]]Videos[[:space:]]in[[:space:]]the[[:space:]]Wild/a32752eb-c43b-45f4-8a1d-dd74e3e96498_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Saner[[:space:]]Deep[[:space:]]Image[[:space:]]Registration/969ee47c-2aba-47af-80c1-64648ec848d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Semi-supervised[[:space:]]Learning[[:space:]]with[[:space:]]Non-random[[:space:]]Missing[[:space:]]Labels/47882424-6fa5-429a-a713-c32d00e11996_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Understanding[[:space:]]the[[:space:]]Generalization[[:space:]]of[[:space:]]Deepfake[[:space:]]Detectors[[:space:]]from[[:space:]]a[[:space:]]Game-Theoretical[[:space:]]View/4026307c-9d06-438d-9eb5-fe702a9a25a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Unifying[[:space:]]Medical[[:space:]]Vision-and-Language[[:space:]]Pre-Training[[:space:]]via[[:space:]]Soft[[:space:]]Prompts/bd3eb3a3-07db-41ad-8836-1a7089aff4d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Universal[[:space:]]Image[[:space:]]Embeddings_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]and[[:space:]]Challenge[[:space:]]for[[:space:]]Generic[[:space:]]Image[[:space:]]Representations/8c2f1754-d8c5-4a78-9c4f-9c5e8430a234_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Universal[[:space:]]LiDAR-Based[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]by[[:space:]]Multi-Domain[[:space:]]Knowledge[[:space:]]Transfer/2aaf5794-b4eb-4cd3-8af5-6388e656faef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Generalization[[:space:]]for[[:space:]]Face[[:space:]]Anti-Spoofing/bd8ea241-2dde-4b9d-8b3f-c650fdd20668_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Viewpoint[[:space:]]Robustness[[:space:]]in[[:space:]]Bird's[[:space:]]Eye[[:space:]]View[[:space:]]Segmentation/16fee7e1-f620-4fac-9c0c-3162d1eef15b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Viewpoint-Invariant[[:space:]]Visual[[:space:]]Recognition[[:space:]]via[[:space:]]Adversarial[[:space:]]Training/70989692-bef1-4c0e-a587-c87710cbed88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Zero[[:space:]]Domain[[:space:]]Gap_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Study[[:space:]]of[[:space:]]Realistic[[:space:]]LiDAR[[:space:]]Simulation[[:space:]]for[[:space:]]Autonomy[[:space:]]Testing/79aa1948-5dc0-4d81-936f-a4119711ab31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Towards[[:space:]]Zero-Shot[[:space:]]Scale-Aware[[:space:]]Monocular[[:space:]]Depth[[:space:]]Estimation/d46e702e-9a81-465d-91c9-50d20773534a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tracing[[:space:]]the[[:space:]]Origin[[:space:]]of[[:space:]]Adversarial[[:space:]]Attack[[:space:]]for[[:space:]]Forensic[[:space:]]Investigation[[:space:]]and[[:space:]]Deterrence/920d8cc7-3efc-4e78-874f-ee26aa2a8840_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TrackFlow_[[:space:]]Multi-Object[[:space:]]tracking[[:space:]]with[[:space:]]Normalizing[[:space:]]Flows/bfdcf707-c7ec-4e18-99d1-67278282dd99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tracking[[:space:]]Anything[[:space:]]with[[:space:]]Decoupled[[:space:]]Video[[:space:]]Segmentation/40abd200-73bc-4936-bbf7-1961a1c819b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tracking[[:space:]]Everything[[:space:]]Everywhere[[:space:]]All[[:space:]]at[[:space:]]Once/439a90ad-d5cd-4c70-a5c8-4bd02f035b9d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tracking[[:space:]]by[[:space:]]3D[[:space:]]Model[[:space:]]Estimation[[:space:]]of[[:space:]]Unknown[[:space:]]Objects[[:space:]]in[[:space:]]Videos/24234f60-9d85-4c13-b53e-f957b9606dfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tracking[[:space:]]by[[:space:]]Natural[[:space:]]Language[[:space:]]Specification[[:space:]]with[[:space:]]Long[[:space:]]Short-term[[:space:]]Context[[:space:]]Decoupling/505e66ce-9cdf-440b-84fb-25fbc8bcdc04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tracking[[:space:]]without[[:space:]]Label_[[:space:]]Unsupervised[[:space:]]Multiple[[:space:]]Object[[:space:]]Tracking[[:space:]]via[[:space:]]Contrastive[[:space:]]Similarity[[:space:]]Learning/e8869ff0-ff3e-4e00-aabc-93034453e44a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Traj-MAE_[[:space:]]Masked[[:space:]]Autoencoders[[:space:]]for[[:space:]]Trajectory[[:space:]]Prediction/6c687680-4e70-4d32-b9d4-07b9bbd5c6b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TrajPAC_[[:space:]]Towards[[:space:]]Robustness[[:space:]]Verification[[:space:]]of[[:space:]]Pedestrian[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]Models/28661ab2-53f1-4220-8026-85117340c2a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Trajectory[[:space:]]Unified[[:space:]]Transformer[[:space:]]for[[:space:]]Pedestrian[[:space:]]Trajectory[[:space:]]Prediction/beb900fd-0e51-48d6-a50f-8abbafed55ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TrajectoryFormer_[[:space:]]3D[[:space:]]Object[[:space:]]Tracking[[:space:]]Transformer[[:space:]]with[[:space:]]Predictive[[:space:]]Trajectory[[:space:]]Hypotheses/6c339812-d974-4eee-801b-381997566611_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TransFace_[[:space:]]Calibrating[[:space:]]Transformer[[:space:]]Training[[:space:]]for[[:space:]]Face[[:space:]]Recognition[[:space:]]from[[:space:]]a[[:space:]]Data-Centric[[:space:]]Perspective/4dcce9f4-a956-41fd-bc36-ebad7c3548f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TransHuman_[[:space:]]A[[:space:]]Transformer-based[[:space:]]Human[[:space:]]Representation[[:space:]]for[[:space:]]Generalizable[[:space:]]Neural[[:space:]]Human[[:space:]]Rendering/315d76e3-6581-49f1-b8a1-28e488995c77_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TransIFF_[[:space:]]An[[:space:]]Instance-Level[[:space:]]Feature[[:space:]]Fusion[[:space:]]Framework[[:space:]]for[[:space:]]Vehicle-Infrastructure[[:space:]]Cooperative[[:space:]]3D[[:space:]]Detection[[:space:]]with[[:space:]]Transformers/86ca2520-8d5a-44c4-9b62-07437f6de57c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TransTIC_[[:space:]]Transferring[[:space:]]Transformer-based[[:space:]]Image[[:space:]]Compression[[:space:]]from[[:space:]]Human[[:space:]]Perception[[:space:]]to[[:space:]]Machine[[:space:]]Perception/480dc37a-13f4-4540-85ff-0c5c1e3e5ecb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Transferable[[:space:]]Adversarial[[:space:]]Attack[[:space:]]for[[:space:]]Both[[:space:]]Vision[[:space:]]Transformers[[:space:]]and[[:space:]]Convolutional[[:space:]]Networks[[:space:]]via[[:space:]]Momentum[[:space:]]Integrated[[:space:]]Gradients/fcba9473-26b2-40e2-873d-8a035fbc8440_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Transferable[[:space:]]Decoding[[:space:]]with[[:space:]]Visual[[:space:]]Entities[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Image[[:space:]]Captioning/32a08236-b67b-4a30-8e2c-c82aaeed9c50_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Translating[[:space:]]Images[[:space:]]to[[:space:]]Road[[:space:]]Network_[[:space:]]A[[:space:]]Non-Autoregressive[[:space:]]Sequence-to-Sequence[[:space:]]Approach/112a11a1-2f3e-4844-b498-8db6fc7ae12d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Transparent[[:space:]]Shape[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]View[[:space:]]Polarization[[:space:]]Image/6a7273bc-81f4-43ca-b469-d90d1069f397_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Treating[[:space:]]Pseudo-labels[[:space:]]Generation[[:space:]]as[[:space:]]Image[[:space:]]Matting[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/7a3d70aa-165b-435e-83be-0a3f1ad3ec99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tree-Structured[[:space:]]Shading[[:space:]]Decomposition/16a0e738-7f86-4f8f-bb7e-706fc0c6d9d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tri-MipRF_[[:space:]]Tri-Mip[[:space:]]Representation[[:space:]]for[[:space:]]Efficient[[:space:]]Anti-Aliasing[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/db21554a-3286-4436-99e3-6491e1afb027_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/TripLe_[[:space:]]Revisiting[[:space:]]Pretrained[[:space:]]Model[[:space:]]Reuse[[:space:]]and[[:space:]]Progressive[[:space:]]Learning[[:space:]]for[[:space:]]Efficient[[:space:]]Vision[[:space:]]Transformer[[:space:]]Scaling[[:space:]]and[[:space:]]Searching/d950cc72-af26-42e4-927f-79c6e09f86be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Troubleshooting[[:space:]]Ethnic[[:space:]]Quality[[:space:]]Bias[[:space:]]with[[:space:]]Curriculum[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Face[[:space:]]Image[[:space:]]Quality[[:space:]]Assessment/a5dd259c-b735-4b3f-9a5b-97a9d13d8e2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tube-Link_[[:space:]]A[[:space:]]Flexible[[:space:]]Cross[[:space:]]Tube[[:space:]]Framework[[:space:]]for[[:space:]]Universal[[:space:]]Video[[:space:]]Segmentation/d84eb9cf-b2db-4324-8da9-57d2a03ac277_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tubelet-Contrastive[[:space:]]Self-Supervision[[:space:]]for[[:space:]]Video-Efficient[[:space:]]Generalization/b167d6f6-888c-47d4-925f-e90f9d7439ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tune-A-Video_[[:space:]]One-Shot[[:space:]]Tuning[[:space:]]of[[:space:]]Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Text-to-Video[[:space:]]Generation/8e5a708b-3c22-4942-b1bf-5442af4badc3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Tuning[[:space:]]Pre-trained[[:space:]]Model[[:space:]]via[[:space:]]Moment[[:space:]]Probing/5d369f0c-36fc-4686-949e-9a5d015142de_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Two[[:space:]]Birds,[[:space:]]One[[:space:]]Stone_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Joint[[:space:]]Learning[[:space:]]of[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Style[[:space:]]Transfers/10af78bd-be4b-40b3-aa03-f7f7677aeac8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Two-in-One[[:space:]]Depth_[[:space:]]Bridging[[:space:]]the[[:space:]]Gap[[:space:]]Between[[:space:]]Monocular[[:space:]]and[[:space:]]Binocular[[:space:]]Self-Supervised[[:space:]]Depth[[:space:]]Estimation/c961c58c-866d-4edb-9151-c9cbfdb222ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/U-RED_[[:space:]]Unsupervised[[:space:]]3D[[:space:]]Shape[[:space:]]Retrieval[[:space:]]and[[:space:]]Deformation[[:space:]]for[[:space:]]Partial[[:space:]]Point[[:space:]]Clouds/9af690d9-d398-4b37-87dd-da93924a17bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UATVR_[[:space:]]Uncertainty-Adaptive[[:space:]]Text-Video[[:space:]]Retrieval/2128c674-369d-4d0e-a218-428eaae7e5fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UCF_[[:space:]]Uncovering[[:space:]]Common[[:space:]]Features[[:space:]]for[[:space:]]Generalizable[[:space:]]Deepfake[[:space:]]Detection/371cf0ee-61b6-4df3-9211-2f4b4b4126dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UGC_[[:space:]]Unified[[:space:]]GAN[[:space:]]Compression[[:space:]]for[[:space:]]Efficient[[:space:]]Image-to-Image[[:space:]]Translation/aed4a69f-15fe-4917-9cb7-a241157225fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UHDNeRF_[[:space:]]Ultra-High-Definition[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/c7c5058f-bb0d-4a9d-8582-614ecbdec97a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UMC_[[:space:]]A[[:space:]]Unified[[:space:]]Bandwidth-efficient[[:space:]]and[[:space:]]Multi-resolution[[:space:]]based[[:space:]]Collaborative[[:space:]]Perception[[:space:]]Framework/53f1d918-6c3a-441d-81ae-4ed5366adc29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UMFuse_[[:space:]]Unified[[:space:]]Multi[[:space:]]View[[:space:]]Fusion[[:space:]]for[[:space:]]Human[[:space:]]Editing[[:space:]]Applications/96d5328e-98b7-4a63-a75d-d8ee18e8e4ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UMIFormer_[[:space:]]Mining[[:space:]]the[[:space:]]Correlations[[:space:]]between[[:space:]]Similar[[:space:]]Tokens[[:space:]]for[[:space:]]Multi-View[[:space:]]3D[[:space:]]Reconstruction/81835dc5-dd52-49c5-a58f-a342d7efc075_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/USAGE_[[:space:]]A[[:space:]]Unified[[:space:]]Seed[[:space:]]Area[[:space:]]Generation[[:space:]]Paradigm[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/d7816efa-f5fc-4d56-80b2-8092acb2fb09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UnLoc_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Video[[:space:]]Localization[[:space:]]Tasks/fa2e30fc-efb6-44b2-abd4-c0bf4fa3d133_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unaligned[[:space:]]2D[[:space:]]to[[:space:]]3D[[:space:]]Translation[[:space:]]with[[:space:]]Conditional[[:space:]]Vector-Quantized[[:space:]]Code[[:space:]]Diffusion[[:space:]]using[[:space:]]Transformers/a11f235f-a032-47f5-ae18-efb9275aa8ec_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uncertainty[[:space:]]Guided[[:space:]]Adaptive[[:space:]]Warping[[:space:]]for[[:space:]]Robust[[:space:]]and[[:space:]]Efficient[[:space:]]Stereo[[:space:]]Matching/5ab4e148-48e0-410e-be16-9f73e023d9d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uncertainty-aware[[:space:]]State[[:space:]]Space[[:space:]]Transformer[[:space:]]for[[:space:]]Egocentric[[:space:]]3D[[:space:]]Hand[[:space:]]Trajectory[[:space:]]Forecasting/1b579985-afcd-4110-8ef2-574f8aa61a8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uncertainty-aware[[:space:]]Unsupervised[[:space:]]Multi-Object[[:space:]]Tracking/d44e5258-1aaa-4fc5-8cda-87432fc9299f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uncertainty-guided[[:space:]]Learning[[:space:]]for[[:space:]]Improving[[:space:]]Image[[:space:]]Manipulation[[:space:]]Detection/c3603b36-dcf8-471b-baa9-2d354b87dfed_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Under-Display[[:space:]]Camera[[:space:]]Image[[:space:]]Restoration[[:space:]]with[[:space:]]Scattering[[:space:]]Effect/ddf8842f-e02e-4a31-813a-3b106b0d79f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Understanding[[:space:]]3D[[:space:]]Object[[:space:]]Interaction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/a5e87c47-57be-42d6-aec3-fb3e08549db0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Understanding[[:space:]]Hessian[[:space:]]Alignment[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/abfc90b7-01c3-4328-93fb-0445f2d8a56e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Understanding[[:space:]]Self-attention[[:space:]]Mechanism[[:space:]]via[[:space:]]Dynamical[[:space:]]System[[:space:]]Perspective/29aaae24-247b-40d9-a6c0-93292eede8fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Understanding[[:space:]]the[[:space:]]Feature[[:space:]]Norm[[:space:]]for[[:space:]]Out-of-Distribution[[:space:]]Detection/98d3c64d-e712-4d45-86e6-ae3368962b91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unfolding[[:space:]]Framework[[:space:]]with[[:space:]]Prior[[:space:]]of[[:space:]]Convolution-Transformer[[:space:]]Mixture[[:space:]]and[[:space:]]Uncertainty[[:space:]]Estimation[[:space:]]for[[:space:]]Video[[:space:]]Snapshot[[:space:]]Compressive[[:space:]]Imaging/5b812264-48b2-4d3b-976e-3d5d330d10d4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Uni-3D_[[:space:]]A[[:space:]]Universal[[:space:]]Model[[:space:]]for[[:space:]]Panoptic[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction/3db97869-0931-437f-aaaf-4b56b51c40d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniDexGrasp++_[[:space:]]Improving[[:space:]]Dexterous[[:space:]]Grasping[[:space:]]Policy[[:space:]]Learning[[:space:]]via[[:space:]]Geometry-Aware[[:space:]]Curriculum[[:space:]]and[[:space:]]Iterative[[:space:]]Generalist-Specialist[[:space:]]Learning/99033397-91a9-4757-bba3-b4758532d85c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniFace_[[:space:]]Unified[[:space:]]Cross-Entropy[[:space:]]Loss[[:space:]]for[[:space:]]Deep[[:space:]]Face[[:space:]]Recognition/732d3e99-80f6-4f6e-bb6a-4c958df13b9c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniFormerV2_[[:space:]]Unlocking[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Image[[:space:]]ViTs[[:space:]]for[[:space:]]Video[[:space:]]Understanding/abda87df-f996-4ce1-9091-d7848e0625e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniFusion_[[:space:]]Unified[[:space:]]Multi-View[[:space:]]Fusion[[:space:]]Transformer[[:space:]]for[[:space:]]Spatial-Temporal[[:space:]]Representation[[:space:]]in[[:space:]]Bird's-Eye-View/54b4c6e6-eff7-4b28-8231-8a0c227b93f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniKD_[[:space:]]Universal[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Mimicking[[:space:]]Homogeneous[[:space:]]or[[:space:]]Heterogeneous[[:space:]]Object[[:space:]]Detectors/7961017b-ef3b-4578-9841-19260d58a3f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniSeg_[[:space:]]A[[:space:]]Unified[[:space:]]Multi-Modal[[:space:]]LiDAR[[:space:]]Segmentation[[:space:]]Network[[:space:]]and[[:space:]]the[[:space:]]OpenPCSeg[[:space:]]Codebase/d5a88638-a901-4da6-b9c8-38194d054706_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniT3D_[[:space:]]A[[:space:]]Unified[[:space:]]Transformer[[:space:]]for[[:space:]]3D[[:space:]]Dense[[:space:]]Captioning[[:space:]]and[[:space:]]Visual[[:space:]]Grounding/4c60a84b-c9f3-4157-bc7a-3b4354bafd76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniTR_[[:space:]]A[[:space:]]Unified[[:space:]]and[[:space:]]Efficient[[:space:]]Multi-Modal[[:space:]]Transformer[[:space:]]for[[:space:]]Bird's-Eye-View[[:space:]]Representation/003bc012-6987-41d4-9ab4-1c1de23603b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniVTG_[[:space:]]Towards[[:space:]]Unified[[:space:]]Video-Language[[:space:]]Temporal[[:space:]]Grounding/ebf58114-d72b-40a0-b808-2ef34101bdd8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unified[[:space:]]Adversarial[[:space:]]Patch[[:space:]]for[[:space:]]Cross-Modal[[:space:]]Attacks[[:space:]]in[[:space:]]the[[:space:]]Physical[[:space:]]World/12e0b8db-3730-4e1f-932b-57bc0b246c03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unified[[:space:]]Coarse-to-Fine[[:space:]]Alignment[[:space:]]for[[:space:]]Video-Text[[:space:]]Retrieval/f8e9ff41-8b7d-4ba6-ae76-44dc175bf721_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unified[[:space:]]Data-Free[[:space:]]Compression_[[:space:]]Pruning[[:space:]]and[[:space:]]Quantization[[:space:]]without[[:space:]]Fine-Tuning/6feeb25e-1869-49c7-9a5b-d828e418ac0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unified[[:space:]]Out-Of-Distribution[[:space:]]Detection_[[:space:]]A[[:space:]]Model-Specific[[:space:]]Perspective/d0dd8204-ce09-4a84-8bb8-f4324fe00f98_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unified[[:space:]]Pre-Training[[:space:]]with[[:space:]]Pseudo[[:space:]]Texts[[:space:]]for[[:space:]]Text-To-Image[[:space:]]Person[[:space:]]Re-Identification/3fe893ee-388e-4383-8320-db6905eb745e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unified[[:space:]]Visual[[:space:]]Relationship[[:space:]]Detection[[:space:]]with[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Models/1a462799-52dd-43b2-a43a-67001f1a64f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unify,[[:space:]]Align[[:space:]]and[[:space:]]Refine_[[:space:]]Multi-Level[[:space:]]Semantic[[:space:]]Alignment[[:space:]]for[[:space:]]Radiology[[:space:]]Report[[:space:]]Generation/a966e83b-7a17-4110-911d-fd2112ab98f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unilaterally[[:space:]]Aggregated[[:space:]]Contrastive[[:space:]]Learning[[:space:]]with[[:space:]]Hierarchical[[:space:]]Augmentation[[:space:]]for[[:space:]]Anomaly[[:space:]]Detection/a56874c1-f412-4f22-8058-de1ad673ed2b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UnitedHuman_[[:space:]]Harnessing[[:space:]]Multi-Source[[:space:]]Data[[:space:]]for[[:space:]]High-Resolution[[:space:]]Human[[:space:]]Generation/de409c66-cb18-4c9a-922d-c49366da24ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UniverSeg_[[:space:]]Universal[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/74efec43-18f2-4de7-b503-693f8927298f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Universal[[:space:]]Domain[[:space:]]Adaptation[[:space:]]via[[:space:]]Compressive[[:space:]]Attention[[:space:]]Matching/7a0e6510-66bb-4388-9fe5-0604ec64c3f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unleashing[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models[[:space:]]for[[:space:]]Visual[[:space:]]Perception/6458255d-1328-4166-8bbb-b043af1504a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unleashing[[:space:]]Vanilla[[:space:]]Vision[[:space:]]Transformer[[:space:]]with[[:space:]]Masked[[:space:]]Image[[:space:]]Modeling[[:space:]]for[[:space:]]Object[[:space:]]Detection/ff0026cf-4c7e-4dae-a39c-12d69d19042d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unleashing[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks[[:space:]]with[[:space:]]Dynamic[[:space:]]Confidence/135e7deb-97c9-4c04-901d-4d327fb1e37c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unleashing[[:space:]]the[[:space:]]Power[[:space:]]of[[:space:]]Gradient[[:space:]]Signal-to-Noise[[:space:]]Ratio[[:space:]]for[[:space:]]Zero-Shot[[:space:]]NAS/e953def6-e60e-4ece-b6b2-302b7444a210_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unmasked[[:space:]]Teacher_[[:space:]]Towards[[:space:]]Training-Efficient[[:space:]]Video[[:space:]]Foundation[[:space:]]Models/894e60a2-0689-4ace-8e2c-720148f835e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unmasking[[:space:]]Anomalies[[:space:]]in[[:space:]]Road-Scene[[:space:]]Segmentation/d5c28a07-dce2-4420-9a84-49fe86cc69a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unpaired[[:space:]]Multi-domain[[:space:]]Attribute[[:space:]]Translation[[:space:]]of[[:space:]]3D[[:space:]]Facial[[:space:]]Shapes[[:space:]]with[[:space:]]a[[:space:]]Square[[:space:]]and[[:space:]]Symmetric[[:space:]]Geometric[[:space:]]Map/1276a28e-d7bc-4d56-b6ca-81173d2a8000_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]3D[[:space:]]Perception[[:space:]]with[[:space:]]2D[[:space:]]Vision-Language[[:space:]]Distillation[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/dc84908d-0f3d-47a9-880b-ed1d2fde9946_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Accuracy[[:space:]]Estimation[[:space:]]of[[:space:]]Deep[[:space:]]Visual[[:space:]]Models[[:space:]]using[[:space:]]Domain-Adaptive[[:space:]]Adversarial[[:space:]]Perturbation[[:space:]]without[[:space:]]Source[[:space:]]Samples/192ae579-8b59-4d01-8f01-ba0ab4a1ab4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Compositional[[:space:]]Concepts[[:space:]]Discovery[[:space:]]with[[:space:]]Text-to-Image[[:space:]]Generative[[:space:]]Models/d0110522-4811-4ea4-bf19-35a0e738c71d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Training[[:space:]]Event-Based[[:space:]]Networks[[:space:]]Using[[:space:]]Contrastive[[:space:]]Learning[[:space:]]and[[:space:]]Uncorrelated[[:space:]]Conditioning/c65a0164-8ba5-4f3d-a512-012ac7f45e35_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Detection[[:space:]]with[[:space:]]Network[[:space:]]Stability[[:space:]]Analysis/811376ca-ba3b-4a8d-a49c-9c8f0b668dc6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Facial[[:space:]]Performance[[:space:]]Editing[[:space:]]via[[:space:]]Vector-Quantized[[:space:]]StyleGAN[[:space:]]Representations/9e463e07-6e3b-4cdf-86fe-f6b2f05307cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Feature[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Domain-generalized[[:space:]]Cross-domain[[:space:]]Image[[:space:]]Retrieval/8e840394-e1d5-489a-8b0e-990b50ede248_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Image[[:space:]]Denoising[[:space:]]in[[:space:]]Real-World[[:space:]]Scenarios[[:space:]]via[[:space:]]Self-Collaboration[[:space:]]Parallel[[:space:]]Generative[[:space:]]Adversarial[[:space:]]Branches/f3ca32a5-0a4e-4ec0-b714-0e0dd098ce19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Learning[[:space:]]of[[:space:]]Object-Centric[[:space:]]Embeddings[[:space:]]for[[:space:]]Cell[[:space:]]Instance[[:space:]]Segmentation[[:space:]]in[[:space:]]Microscopy[[:space:]]Images/a80542e5-0ebd-4d19-b88c-82d371e69149_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Manifold[[:space:]]Linearizing[[:space:]]and[[:space:]]Clustering/5d688e5d-f7e9-4c4f-b2b1-e91c71ddb42d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Object[[:space:]]Localization[[:space:]]with[[:space:]]Representer[[:space:]]Point[[:space:]]Selection/55f48842-e87d-4d03-bbad-4740a56ff028_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Localization[[:space:]]in[[:space:]]Videos/00389f49-7267-4132-a60c-3fe80d134c1e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Text-Driven[[:space:]]Object[[:space:]]Detection/d6effe98-ec12-44ca-8d88-1a3c8f7ac232_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Self-Driving[[:space:]]Attention[[:space:]]Prediction[[:space:]]via[[:space:]]Uncertainty[[:space:]]Mining[[:space:]]and[[:space:]]Knowledge[[:space:]]Embedding/48140d5b-dd2b-489a-98f8-b5ab258d0115_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Surface[[:space:]]Anomaly[[:space:]]Detection[[:space:]]with[[:space:]]Diffusion[[:space:]]Probabilistic[[:space:]]Model/480ebec9-cbcf-4e67-9c9f-2a10bb5d6c20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Video[[:space:]]Deraining[[:space:]]with[[:space:]]An[[:space:]]Event[[:space:]]Camera/74f999ec-bb7f-4790-aa59-58a7bd964c05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Unsupervised[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation[[:space:]]with[[:space:]]Online[[:space:]]Adversarial[[:space:]]Self-Tuning/569dfd78-9f76-450d-89c8-f1bce6cd4ffe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UpCycling_[[:space:]]Semi-supervised[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]without[[:space:]]Sharing[[:space:]]Raw-level[[:space:]]Unlabeled[[:space:]]Scenes/37323419-4923-4eb4-bb5b-fe59e37c4a1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Urban[[:space:]]Radiance[[:space:]]Field[[:space:]]Representation[[:space:]]with[[:space:]]Deformable[[:space:]]Neural[[:space:]]Mesh[[:space:]]Primitives/b95fc897-37fc-4b21-a7e0-2db896e228c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/UrbanGIRAFFE_[[:space:]]Representing[[:space:]]Urban[[:space:]]Scenes[[:space:]]as[[:space:]]Compositional[[:space:]]Generative[[:space:]]Neural[[:space:]]Feature[[:space:]]Fields/95f91179-18fd-4015-b10b-9eeff5d8adfc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Using[[:space:]]a[[:space:]]Waffle[[:space:]]Iron[[:space:]]for[[:space:]]Automotive[[:space:]]Point[[:space:]]Cloud[[:space:]]Semantic[[:space:]]Segmentation/f04d1aca-d6a4-4d82-bfb7-746cf22f81e1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/V-FUSE_[[:space:]]Volumetric[[:space:]]Depth[[:space:]]Map[[:space:]]Fusion[[:space:]]with[[:space:]]Long-Range[[:space:]]Constraints/02d75c3d-e5f5-4356-a602-ca3fc6929db5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/V3Det_[[:space:]]Vast[[:space:]]Vocabulary[[:space:]]Visual[[:space:]]Detection[[:space:]]Dataset/9d4274e7-df1c-45e4-880e-7119cdf14923_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VADER_[[:space:]]Video[[:space:]]Alignment[[:space:]]Differencing[[:space:]]and[[:space:]]Retrieval/14c60159-a9fa-456f-9b10-dceb191faecc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VAD_[[:space:]]Vectorized[[:space:]]Scene[[:space:]]Representation[[:space:]]for[[:space:]]Efficient[[:space:]]Autonomous[[:space:]]Driving/7df14791-6d0d-4356-816e-a71dd3d8537b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VAPCNet_[[:space:]]Viewpoint-Aware[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Completion/7854fc39-1e20-4883-8c49-87a5f3a1db9a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VI-Net_[[:space:]]Boosting[[:space:]]Category-level[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]via[[:space:]]Learning[[:space:]]Decoupled[[:space:]]Rotations[[:space:]]on[[:space:]]the[[:space:]]Spherical[[:space:]]Representations/ca5c8b5a-ce33-4e2e-8576-3a2a66776f49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VL-Match_[[:space:]]Enhancing[[:space:]]Vision-Language[[:space:]]Pretraining[[:space:]]with[[:space:]]Token-Level[[:space:]]and[[:space:]]Instance-Level[[:space:]]Matching/ea44ab83-0201-42b6-ba3b-7aee2967a113_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VL-PET_[[:space:]]Vision-and-Language[[:space:]]Parameter-Efficient[[:space:]]Tuning[[:space:]]via[[:space:]]Granularity[[:space:]]Control/18d614a1-5564-4e79-a525-66a43b960747_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VLN-PETL_[[:space:]]Parameter-Efficient[[:space:]]Transfer[[:space:]]Learning[[:space:]]for[[:space:]]Vision-and-Language[[:space:]]Navigation/956c4c9c-156f-47e6-9dc8-ecd7b46322ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VLSlice_[[:space:]]Interactive[[:space:]]Vision-and-Language[[:space:]]Slice[[:space:]]Discovery/1e3bc8c6-40ed-4190-8cc1-f0483eb19f05_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VQ3D_[[:space:]]Learning[[:space:]]a[[:space:]]3D-Aware[[:space:]]Generative[[:space:]]Model[[:space:]]on[[:space:]]ImageNet/6772d23e-7e0d-464c-bbbe-e2137459bbd3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VQA[[:space:]]Therapy_[[:space:]]Exploring[[:space:]]Answer[[:space:]]Differences[[:space:]]by[[:space:]]Visually[[:space:]]Grounding[[:space:]]Answers/1d68b644-ea5e-40bd-a981-7d091dfe73ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VQA-GNN_[[:space:]]Reasoning[[:space:]]with[[:space:]]Multimodal[[:space:]]Knowledge[[:space:]]via[[:space:]]Graph[[:space:]]Neural[[:space:]]Networks[[:space:]]for[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/fb5931f6-b8a5-499c-b95d-122044981a0c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Vanishing[[:space:]]Point[[:space:]]Estimation[[:space:]]in[[:space:]]Uncalibrated[[:space:]]Images[[:space:]]with[[:space:]]Prior[[:space:]]Gravity[[:space:]]Direction/1b66ee1e-ed71-4d13-a90e-f7a7349dcbe6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Variational[[:space:]]Causal[[:space:]]Inference[[:space:]]Network[[:space:]]for[[:space:]]Explanatory[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/8d8b2d72-f0aa-4110-b557-37ccce0ea9a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Variational[[:space:]]Degeneration[[:space:]]to[[:space:]]Structural[[:space:]]Refinement_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Superimposed[[:space:]]Image[[:space:]]Decomposition/0da1bd6f-7bc7-444b-8826-a01e2dfc8bdd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VeRi3D_[[:space:]]Generative[[:space:]]Vertex-based[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]3D[[:space:]]Controllable[[:space:]]Human[[:space:]]Image[[:space:]]Synthesis/bfc996cf-c0ee-4fe3-a02b-9796861d39d6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Verbs[[:space:]]in[[:space:]]Action_[[:space:]]Improving[[:space:]]Verb[[:space:]]Understanding[[:space:]]in[[:space:]]Video-Language[[:space:]]Models/eb8dc395-5b9d-427e-96f5-b07eb9261266_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Versatile[[:space:]]Diffusion_[[:space:]]Text,[[:space:]]Images[[:space:]]and[[:space:]]Variations[[:space:]]All[[:space:]]in[[:space:]]One[[:space:]]Diffusion[[:space:]]Model/f270447b-12b7-4de8-9fec-63d2d5be4585_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VertexSerum_[[:space:]]Poisoning[[:space:]]Graph[[:space:]]Neural[[:space:]]Networks[[:space:]]for[[:space:]]Link[[:space:]]Inference/9e74bc27-267b-4107-8eb1-95840677f227_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ViLLA_[[:space:]]Fine-Grained[[:space:]]Vision-Language[[:space:]]Representation[[:space:]]Learning[[:space:]]from[[:space:]]Real-World[[:space:]]Data/a73edae6-af72-4549-9a24-1caf5fb83436_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ViLTA_[[:space:]]Enhancing[[:space:]]Vision-Language[[:space:]]Pre-training[[:space:]]through[[:space:]]Textual[[:space:]]Augmentation/57804857-dfe9-481c-a6e4-24c0ff5b92a7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ViM_[[:space:]]Vision[[:space:]]Middleware[[:space:]]for[[:space:]]Unified[[:space:]]Downstream[[:space:]]Transferring/9bc8a264-8c6b-430d-bfad-aefae3498ee9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VidStyleODE_[[:space:]]Disentangled[[:space:]]Video[[:space:]]Editing[[:space:]]via[[:space:]]StyleGAN[[:space:]]and[[:space:]]NeuralODEs/f15e3f19-643c-446f-a7df-68ac5bfed9c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]Action[[:space:]]Recognition[[:space:]]with[[:space:]]Attentive[[:space:]]Semantic[[:space:]]Units/29fa5735-7b6d-42ec-aebc-cb618c8a3d04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]Action[[:space:]]Segmentation[[:space:]]via[[:space:]]Contextually[[:space:]]Refined[[:space:]]Temporal[[:space:]]Keypoints/8fa0757d-2401-44d1-b706-18b279cf4728_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]Adverse-Weather-Component[[:space:]]Suppression[[:space:]]Network[[:space:]]via[[:space:]]Weather[[:space:]]Messenger[[:space:]]and[[:space:]]Adversarial[[:space:]]Backpropagation/2665a1c9-e78f-4fc0-96be-4b3017b49a8f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]Anomaly[[:space:]]Detection[[:space:]]via[[:space:]]Sequentially[[:space:]]Learning[[:space:]]Multiple[[:space:]]Pretext[[:space:]]Tasks/6edd046a-5935-4461-965f-867acb539dca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]Background[[:space:]]Music[[:space:]]Generation_[[:space:]]Dataset,[[:space:]]Method[[:space:]]and[[:space:]]Evaluation/ba5466bd-08b5-4633-b6b2-01211e3e3a88_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]OWL-ViT_[[:space:]]Temporally-consistent[[:space:]]Open-world[[:space:]]Localization[[:space:]]in[[:space:]]Video/1f718f97-37a4-4316-8441-bec2d7cd438c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]Object[[:space:]]Segmentation-aware[[:space:]]Video[[:space:]]Frame[[:space:]]Interpolation/7b7eaabc-85a4-4593-bd7b-16e8851ef07a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]State-Changing[[:space:]]Object[[:space:]]Segmentation/f72d9388-12a7-444c-8736-1992c337b800_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video[[:space:]]Task[[:space:]]Decathlon_[[:space:]]Unifying[[:space:]]Image[[:space:]]and[[:space:]]Video[[:space:]]Tasks[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/ea94acb6-7137-48e4-9336-097fe4baf483_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Video-FocalNets_[[:space:]]Spatio-Temporal[[:space:]]Focal[[:space:]]Modulation[[:space:]]for[[:space:]]Video[[:space:]]Action[[:space:]]Recognition/6f7530ce-eb42-4afa-93b5-ecedb52aff06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VideoFlow_[[:space:]]Exploiting[[:space:]]Temporal[[:space:]]Cues[[:space:]]for[[:space:]]Multi-frame[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation/b36105bd-d1cc-497e-bc24-003f73501bbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/View[[:space:]]Consistent[[:space:]]Purification[[:space:]]for[[:space:]]Accurate[[:space:]]Cross-View[[:space:]]Localization/02a9acb5-5552-4804-8989-2f353cb8afb9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ViewRefer_[[:space:]]Grasp[[:space:]]the[[:space:]]Multi-view[[:space:]]Knowledge[[:space:]]for[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding/d4f32ad0-6cde-4730-9bd2-244e9454d42b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Viewing[[:space:]]Graph[[:space:]]Solvability[[:space:]]in[[:space:]]Practice/137a29e9-29f5-44f3-bc06-13c7c391cabe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Viewset[[:space:]]Diffusion_[[:space:]](0-)Image-Conditioned[[:space:]]3D[[:space:]]Generative[[:space:]]Models[[:space:]]from[[:space:]]2D[[:space:]]Data/ec728d7e-65e1-4d63-8d3d-5948311dae41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ViperGPT_[[:space:]]Visual[[:space:]]Inference[[:space:]]via[[:space:]]Python[[:space:]]Execution[[:space:]]for[[:space:]]Reasoning/f9402566-c621-4598-a0a6-af253e2d0235_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Virtual[[:space:]]Try-On[[:space:]]with[[:space:]]Pose-Garment[[:space:]]Keypoints[[:space:]]Guided[[:space:]]Inpainting/64da54d2-ca7d-497c-88f5-20bd29a61776_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification[[:space:]]via[[:space:]]Semantic[[:space:]]Alignment[[:space:]]and[[:space:]]Affinity[[:space:]]Inference/69713c73-312a-46fd-8c56-f72a91dbb924_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Vision[[:space:]]Grid[[:space:]]Transformer[[:space:]]for[[:space:]]Document[[:space:]]Layout[[:space:]]Analysis/3a116ca0-6788-4568-a9d6-9ac7d184d034_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Vision[[:space:]]HGNN_[[:space:]]An[[:space:]]Image[[:space:]]is[[:space:]]More[[:space:]]than[[:space:]]a[[:space:]]Graph[[:space:]]of[[:space:]]Nodes/e953f946-c784-4fc7-857d-3e979a98d5e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Vision[[:space:]]Relation[[:space:]]Transformer[[:space:]]for[[:space:]]Unbiased[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/bf9577e9-ac8c-4df7-9ee1-878130fd3950_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Vision[[:space:]]Transformer[[:space:]]Adapters[[:space:]]for[[:space:]]Generalizable[[:space:]]Multitask[[:space:]]Learning/c11c9b35-13b4-40fa-bdea-5c9b5ab0e575_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Explanations[[:space:]]via[[:space:]]Iterated[[:space:]]Integrated[[:space:]]Attributions/4585fab5-92a2-4f10-afda-c96ed68fe2a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visual[[:space:]]Traffic[[:space:]]Knowledge[[:space:]]Graph[[:space:]]Generation[[:space:]]from[[:space:]]Scene[[:space:]]Images/b17a2b91-1731-4b49-b785-da64b9d85cd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Visually-Prompted[[:space:]]Language[[:space:]]Model[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation[[:space:]]in[[:space:]]an[[:space:]]Open[[:space:]]World/4ce664fb-ee63-47d3-89fc-1aedbb978941_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/VoroMesh_[[:space:]]Learning[[:space:]]Watertight[[:space:]]Surface[[:space:]]Meshes[[:space:]]with[[:space:]]Voronoi[[:space:]]Diagrams/92e64de1-fb19-4803-b209-92439d4ca9f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Vox-E_[[:space:]]Text-Guided[[:space:]]Voxel[[:space:]]Editing[[:space:]]of[[:space:]]3D[[:space:]]Objects/cbefaa75-e9f9-4134-a56f-828f3cda546f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/WALDO_[[:space:]]Future[[:space:]]Video[[:space:]]Synthesis[[:space:]]Using[[:space:]]Object[[:space:]]Layer[[:space:]]Decomposition[[:space:]]and[[:space:]]Parametric[[:space:]]Flow[[:space:]]Prediction/5ec7da4a-46fc-4555-bafd-eb56e41b9709_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/WDiscOOD_[[:space:]]Out-of-Distribution[[:space:]]Detection[[:space:]]via[[:space:]]Whitened[[:space:]]Linear[[:space:]]Discriminant[[:space:]]Analysis/43abc2f7-5e29-4493-91e4-0b25545df112_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Waffling[[:space:]]Around[[:space:]]for[[:space:]]Performance_[[:space:]]Visual[[:space:]]Classification[[:space:]]with[[:space:]]Random[[:space:]]Words[[:space:]]and[[:space:]]Broad[[:space:]]Concepts/d34e70a8-8377-4593-a917-bf0e73418cea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Walking[[:space:]]Your[[:space:]]LiDOG_[[:space:]]A[[:space:]]Journey[[:space:]]Through[[:space:]]Multiple[[:space:]]Domains[[:space:]]for[[:space:]]LiDAR[[:space:]]Semantic[[:space:]]Segmentation/7d30bd5e-6e58-49a7-9745-c4f36aa6cf8d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Wasserstein[[:space:]]Expansible[[:space:]]Variational[[:space:]]Autoencoder[[:space:]]for[[:space:]]Discriminative[[:space:]]and[[:space:]]Generative[[:space:]]Continual[[:space:]]Learning/eca1f9cd-901a-441b-ad17-544e356ae569_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/WaterMask_[[:space:]]Instance[[:space:]]Segmentation[[:space:]]for[[:space:]]Underwater[[:space:]]Imagery/210abf8e-8364-4b66-84bc-a76cc31996a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/WaveIPT_[[:space:]]Joint[[:space:]]Attention[[:space:]]and[[:space:]]Flow[[:space:]]Alignment[[:space:]]in[[:space:]]the[[:space:]]Wavelet[[:space:]]domain[[:space:]]for[[:space:]]Pose[[:space:]]Transfer/3ec5f262-8d41-4e0f-8f40-2c2901b31700_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/WaveNeRF_[[:space:]]Wavelet-based[[:space:]]Generalizable[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/5b373e7c-96e4-4fc6-80d8-a59c573be144_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly[[:space:]]Supervised[[:space:]]Learning[[:space:]]of[[:space:]]Semantic[[:space:]]Correspondence[[:space:]]through[[:space:]]Cascaded[[:space:]]Online[[:space:]]Correspondence[[:space:]]Refinement/9ae6bb9b-0dc0-4018-8654-8e9c5aa06ddd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly[[:space:]]Supervised[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation[[:space:]]with[[:space:]]Intra-Chunk[[:space:]]and[[:space:]]Inter-Chunk[[:space:]]Consistency/69799bbc-060b-420e-8826-a85a16c6fbae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly-Supervised[[:space:]]Action[[:space:]]Localization[[:space:]]by[[:space:]]Hierarchically-Structured[[:space:]]Latent[[:space:]]Attention[[:space:]]Modeling/1235774c-1b7e-4cdb-8112-d97b5c970400_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly-Supervised[[:space:]]Action[[:space:]]Segmentation[[:space:]]and[[:space:]]Unseen[[:space:]]Error[[:space:]]Detection[[:space:]]in[[:space:]]Anomalous[[:space:]]Instructional[[:space:]]Videos/222f1add-78b4-4fa7-9827-90badedb6349_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly-Supervised[[:space:]]Text-Driven[[:space:]]Contrastive[[:space:]]Learning[[:space:]]for[[:space:]]Facial[[:space:]]Behavior[[:space:]]Understanding/84463813-320c-438c-957b-2196238ee6ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Weakly-supervised[[:space:]]3D[[:space:]]Pose[[:space:]]Transfer[[:space:]]with[[:space:]]Keypoints/9956cc3a-adfc-474c-a83c-48171fde5a86_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/What[[:space:]]Can[[:space:]]Simple[[:space:]]Arithmetic[[:space:]]Operations[[:space:]]Do[[:space:]]for[[:space:]]Temporal[[:space:]]Modeling_/aa0ecb1e-5e9c-40d5-86cd-6a395638be01_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/What[[:space:]]Can[[:space:]]a[[:space:]]Cook[[:space:]]in[[:space:]]Italy[[:space:]]Teach[[:space:]]a[[:space:]]Mechanic[[:space:]]in[[:space:]]India_[[:space:]]Action[[:space:]]Recognition[[:space:]]Generalisation[[:space:]]Over[[:space:]]Scenarios[[:space:]]and[[:space:]]Locations/6b464b32-1678-4949-9a56-cefc369ec1e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/What[[:space:]]Does[[:space:]]a[[:space:]]Platypus[[:space:]]Look[[:space:]]Like_[[:space:]]Generating[[:space:]]Customized[[:space:]]Prompts[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Image[[:space:]]Classification/b7c74bc7-2936-48e7-84e0-acbaae0c3bf5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/What[[:space:]]can[[:space:]]Discriminator[[:space:]]do_[[:space:]]Towards[[:space:]]Box-free[[:space:]]Ownership[[:space:]]Verification[[:space:]]of[[:space:]]Generative[[:space:]]Adversarial[[:space:]]Networks/3ffa3d93-7090-4355-be9b-717a5520bd54_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/What[[:space:]]do[[:space:]]neural[[:space:]]networks[[:space:]]learn[[:space:]]in[[:space:]]image[[:space:]]classification_[[:space:]]A[[:space:]]frequency[[:space:]]shortcut[[:space:]]perspective/ae160f4f-e78f-4b23-a073-c8547fbdd538_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/What[[:space:]]does[[:space:]]CLIP[[:space:]]know[[:space:]]about[[:space:]]a[[:space:]]red[[:space:]]circle_[[:space:]]Visual[[:space:]]prompt[[:space:]]engineering[[:space:]]for[[:space:]]VLMs/aff5c910-fbb2-49e1-8bfa-62a06056e2df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/When[[:space:]]Do[[:space:]]Curricula[[:space:]]Work[[:space:]]in[[:space:]]Federated[[:space:]]Learning_/c59a082f-3e79-4dc7-93aa-bc7af33e1cb0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/When[[:space:]]Epipolar[[:space:]]Constraint[[:space:]]Meets[[:space:]]Non-Local[[:space:]]Operators[[:space:]]in[[:space:]]Multi-View[[:space:]]Stereo/b104c0e5-28b4-43fb-89ce-ce1172b405e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/When[[:space:]]Noisy[[:space:]]Labels[[:space:]]Meet[[:space:]]Long[[:space:]]Tail[[:space:]]Dilemmas_[[:space:]]A[[:space:]]Representation[[:space:]]Calibration[[:space:]]Method/8e4730be-f5a9-498a-b495-3c25ba330a13_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/When[[:space:]]Prompt-based[[:space:]]Incremental[[:space:]]Learning[[:space:]]Does[[:space:]]Not[[:space:]]Meet[[:space:]]Strong[[:space:]]Pretraining/a7de6b99-ea1f-485c-a3bd-2453970d73fc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/When[[:space:]]to[[:space:]]Learn[[:space:]]What_[[:space:]]Model-Adaptive[[:space:]]Data[[:space:]]Augmentation[[:space:]]Curriculum/14115e65-4479-4ced-a042-67e0a20a72ac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Who[[:space:]]Are[[:space:]]You[[:space:]]Referring[[:space:]]To_[[:space:]]Coreference[[:space:]]Resolution[[:space:]]In[[:space:]]Image[[:space:]]Narrations/123a9be7-d202-4cb9-9413-cab55d11d2cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Why[[:space:]]Is[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models[[:space:]]Robust[[:space:]]to[[:space:]]Noisy[[:space:]]Labels_/c9eb5c74-620e-446b-a566-f806bad1df94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Why[[:space:]]do[[:space:]]networks[[:space:]]have[[:space:]]inhibitory_negative[[:space:]]connections_/8ce0a68a-68f1-4c0c-82e3-bb716d360442_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Will[[:space:]]Large-scale[[:space:]]Generative[[:space:]]Models[[:space:]]Corrupt[[:space:]]Future[[:space:]]Datasets_/65c63da0-b91c-4b40-970e-6c3c2a2cec91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Window-Based[[:space:]]Early-Exit[[:space:]]Cascades[[:space:]]for[[:space:]]Uncertainty[[:space:]]Estimation_[[:space:]]When[[:space:]]Deep[[:space:]]Ensembles[[:space:]]are[[:space:]]More[[:space:]]Efficient[[:space:]]than[[:space:]]Single[[:space:]]Models/600b1aed-8026-4dc6-87bf-4a603672e37c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/With[[:space:]]a[[:space:]]Little[[:space:]]Help[[:space:]]from[[:space:]]Your[[:space:]]Own[[:space:]]Past_[[:space:]]Prototypical[[:space:]]Memory[[:space:]]Networks[[:space:]]for[[:space:]]Image[[:space:]]Captioning/e40a79c9-9c1f-4a3b-9c88-2b989cae5c02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Workie-Talkie_[[:space:]]Accelerating[[:space:]]Federated[[:space:]]Learning[[:space:]]by[[:space:]]Overlapping[[:space:]]Computing[[:space:]]and[[:space:]]Communications[[:space:]]via[[:space:]]Contrastive[[:space:]]Regularization/74b5aebb-c147-4615-bd38-4ae95ce02071_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/X-Mesh_[[:space:]]Towards[[:space:]]Fast[[:space:]]and[[:space:]]Accurate[[:space:]]Text-driven[[:space:]]3D[[:space:]]Stylization[[:space:]]via[[:space:]]Dynamic[[:space:]]Textual[[:space:]]Guidance/92ffae4c-e0b0-4b55-8607-4bdc565e0f83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/X-VoE_[[:space:]]Measuring[[:space:]]eXplanatory[[:space:]]Violation[[:space:]]of[[:space:]]Expectation[[:space:]]in[[:space:]]Physical[[:space:]]Events/30bb9d0c-73da-4092-a920-8f6ab276da4c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/XMem++_[[:space:]]Production-level[[:space:]]Video[[:space:]]Segmentation[[:space:]]From[[:space:]]Few[[:space:]]Annotated[[:space:]]Frames/f3c75c12-8c8e-4f32-b61b-963ae7a9291d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/XNet_[[:space:]]Wavelet-Based[[:space:]]Low[[:space:]]and[[:space:]]High[[:space:]]Frequency[[:space:]]Fusion[[:space:]]Networks[[:space:]]for[[:space:]]Fully-[[:space:]]and[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]of[[:space:]]Biomedical[[:space:]]Images/aaeda918-00d7-414d-a844-057c3a91a4f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/XVO_[[:space:]]Generalized[[:space:]]Visual[[:space:]]Odometry[[:space:]]via[[:space:]]Cross-Modal[[:space:]]Self-Training/9ee8e316-e34b-4d76-ba0c-3d6de449fdb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/XiNet_[[:space:]]Efficient[[:space:]]Neural[[:space:]]Networks[[:space:]]for[[:space:]]tinyML/9814b73f-85e1-45ee-b61c-bb7329af94cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Yes,[[:space:]]we[[:space:]]CANN_[[:space:]]Constrained[[:space:]]Approximate[[:space:]]Nearest[[:space:]]Neighbors[[:space:]]for[[:space:]]Local[[:space:]]Feature-Based[[:space:]]Visual[[:space:]]Localization/8df9246f-790d-4813-8f7f-39b552c9db7c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/You[[:space:]]Never[[:space:]]Get[[:space:]]a[[:space:]]Second[[:space:]]Chance[[:space:]]To[[:space:]]Make[[:space:]]a[[:space:]]Good[[:space:]]First[[:space:]]Impression_[[:space:]]Seeding[[:space:]]Active[[:space:]]Learning[[:space:]]for[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation/1124eeca-6d9e-4358-8787-0beac157b5f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Your[[:space:]]Diffusion[[:space:]]Model[[:space:]]is[[:space:]]Secretly[[:space:]]a[[:space:]]Zero-Shot[[:space:]]Classifier/4fd32aec-71ef-4d9d-a1bc-a17570326a0a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zenseact[[:space:]]Open[[:space:]]Dataset_[[:space:]]A[[:space:]]Large-Scale[[:space:]]and[[:space:]]Diverse[[:space:]]Multimodal[[:space:]]Dataset[[:space:]]for[[:space:]]Autonomous[[:space:]]Driving/e03c73a4-c49f-4ab8-aab3-9df7248651e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-1-to-3_[[:space:]]Zero-shot[[:space:]]One[[:space:]]Image[[:space:]]to[[:space:]]3D[[:space:]]Object/6ea2be7f-e66b-4951-879d-11681606a777_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Composed[[:space:]]Image[[:space:]]Retrieval[[:space:]]with[[:space:]]Textual[[:space:]]Inversion/cd9f9855-ee97-4d0f-9820-86f8ba66dd45_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Contrastive[[:space:]]Loss[[:space:]]for[[:space:]]Text-Guided[[:space:]]Diffusion[[:space:]]Image[[:space:]]Style[[:space:]]Transfer/118dc630-0f18-4ba5-aa83-86f5a3896502_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Point[[:space:]]Cloud[[:space:]]Segmentation[[:space:]]by[[:space:]]Semantic-Visual[[:space:]]Aware[[:space:]]Synthesis/8c12dab2-2d72-49b0-ac49-dc993c2a8253_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-Shot[[:space:]]Spatial[[:space:]]Layout[[:space:]]Conditioning[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/2c8e4e39-3fad-47b1-8018-a0d078c8cf84_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zero-guidance[[:space:]]Segmentation[[:space:]]Using[[:space:]]Zero[[:space:]]Segment[[:space:]]Labels/de3eea11-fc15-47c3-aaca-2e33f37d07dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zip-NeRF_[[:space:]]Anti-Aliased[[:space:]]Grid-Based[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/e66cd95d-a2be-4b85-8a59-9beb18b18661_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Zolly_[[:space:]]Zoom[[:space:]]Focal[[:space:]]Length[[:space:]]Correctly[[:space:]]for[[:space:]]Perspective-Distorted[[:space:]]Human[[:space:]]Mesh[[:space:]]Reconstruction/68a81286-621e-4364-a79d-bd0bec8f9459_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/eP-ALM_[[:space:]]Efficient[[:space:]]Perceptual[[:space:]]Augmentation[[:space:]]of[[:space:]]Language[[:space:]]Models/7bc5ad5f-fd00-49dd-8bf4-2173a3217be4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/iDAG_[[:space:]]Invariant[[:space:]]DAG[[:space:]]Searching[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/a0bda018-acf5-4221-8260-06d382a24268_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/iVS-Net_[[:space:]]Learning[[:space:]]Human[[:space:]]View[[:space:]]Synthesis[[:space:]]from[[:space:]]Internet[[:space:]]Videos/3a4dd618-0665-4b90-b977-0565cab2230a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/s-Adaptive[[:space:]]Decoupled[[:space:]]Prototype[[:space:]]for[[:space:]]Few-Shot[[:space:]]Object[[:space:]]Detection/aa80d75e-714f-44e4-aed2-84d30f748ee4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/uSplit_[[:space:]]Image[[:space:]]Decomposition[[:space:]]for[[:space:]]Fluorescence[[:space:]]Microscopy/7a55fbcf-8bf2-4095-8dc6-20010dfc3d6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/zPROBE_[[:space:]]Zero[[:space:]]Peek[[:space:]]Robustness[[:space:]]Checks[[:space:]]for[[:space:]]Federated[[:space:]]Learning/e983c24e-2a73-41f0-a538-184e2f0c8e62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ACTIVE_[[:space:]]Towards[[:space:]]Highly[[:space:]]Transferable[[:space:]]3D[[:space:]]Physical[[:space:]]Camouflage[[:space:]]for[[:space:]]Universal[[:space:]]and[[:space:]]Robust[[:space:]]Vehicle[[:space:]]Evasion/50c3cd65-5eb2-41c9-9bca-82b78a66974f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ADAPT_[[:space:]]Efficient[[:space:]]Multi-Agent[[:space:]]Trajectory[[:space:]]Prediction[[:space:]]with[[:space:]]Adaptation/9a5596b0-23d0-4ec6-9e04-7e887623d8b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ADNet_[[:space:]]Lane[[:space:]]Shape[[:space:]]Prediction[[:space:]]via[[:space:]]Anchor[[:space:]]Decomposition/4a0e34ef-4bff-47a1-bcc8-d3ff7415c982_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AG3D_[[:space:]]Learning[[:space:]]to[[:space:]]Generate[[:space:]]3D[[:space:]]Avatars[[:space:]]from[[:space:]]2D[[:space:]]Image[[:space:]]Collections/ea4d6870-aa87-4abc-b3f4-f6a5755865f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AGG-Net_[[:space:]]Attention[[:space:]]Guided[[:space:]]Gated-Convolutional[[:space:]]Network[[:space:]]for[[:space:]]Depth[[:space:]]Image[[:space:]]Completion/21c733cf-4b85-45be-8b1d-c2b93fd1c65e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AIDE_[[:space:]]A[[:space:]]Vision-Driven[[:space:]]Multi-View,[[:space:]]Multi-Modal,[[:space:]]Multi-Tasking[[:space:]]Dataset[[:space:]]for[[:space:]]Assistive[[:space:]]Driving[[:space:]]Perception/94851846-f8ab-466c-82ca-2b426186631a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ALIP_[[:space:]]Adaptive[[:space:]]Language-Image[[:space:]]Pre-Training[[:space:]]with[[:space:]]Synthetic[[:space:]]Caption/a9eeb6f2-4bdf-4cee-a173-3e2969290245_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ALWOD_[[:space:]]Active[[:space:]]Learning[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Object[[:space:]]Detection/e06d7f60-ee2b-4486-a419-ef32c12c2779_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AREA_[[:space:]]Adaptive[[:space:]]Reweighting[[:space:]]via[[:space:]]Effective[[:space:]]Area[[:space:]]for[[:space:]]Long-Tailed[[:space:]]Classification/03e99916-f1f3-423a-a4fa-6b4af8ebe92c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ARNOLD_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Language-Grounded[[:space:]]Task[[:space:]]Learning[[:space:]]with[[:space:]]Continuous[[:space:]]States[[:space:]]in[[:space:]]Realistic[[:space:]]3D[[:space:]]Scenes/eed83e0d-4b1f-4139-9901-613e2051d04c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ASAG_[[:space:]]Building[[:space:]]Strong[[:space:]]One-Decoder-Layer[[:space:]]Sparse[[:space:]]Detectors[[:space:]]via[[:space:]]Adaptive[[:space:]]Sparse[[:space:]]Anchor[[:space:]]Generation/68734f14-4dca-4e01-90f0-f81ac30bdd3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ASIC_[[:space:]]Aligning[[:space:]]Sparse[[:space:]]in-the-wild[[:space:]]Image[[:space:]]Collections/1203f899-aed6-4c89-8fa4-b8e75c5d108b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ASM_[[:space:]]Adaptive[[:space:]]Skinning[[:space:]]Model[[:space:]]for[[:space:]]High-Quality[[:space:]]3D[[:space:]]Face[[:space:]]Modeling/7d5d17ae-4869-4f1a-9750-fa3c0945531a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ATT3D_[[:space:]]Amortized[[:space:]]Text-to-3D[[:space:]]Object[[:space:]]Synthesis/3373e3ba-c8b6-4957-807f-d8c3ac688d83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ablating[[:space:]]Concepts[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/9c021565-b1c8-4251-9401-e3558164b323_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AccFlow_[[:space:]]Backward[[:space:]]Accumulation[[:space:]]for[[:space:]]Long-Range[[:space:]]Optical[[:space:]]Flow/e5896194-f826-44fe-b36b-39eda9c29a37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Accurate[[:space:]]3D[[:space:]]Face[[:space:]]Reconstruction[[:space:]]with[[:space:]]Facial[[:space:]]Component[[:space:]]Tokens/640e260a-327e-4872-9b6d-5d5661b519dc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Accurate[[:space:]]and[[:space:]]Fast[[:space:]]Compressed[[:space:]]Video[[:space:]]Captioning/4f832d5d-d350-44d5-9f62-3e0bf6e26a99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Achievement-Based[[:space:]]Training[[:space:]]Progress[[:space:]]Balancing[[:space:]]for[[:space:]]Multi-Task[[:space:]]Learning/272271d0-f734-4249-a347-749bac4b8015_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ActFormer_[[:space:]]A[[:space:]]GAN-based[[:space:]]Transformer[[:space:]]towards[[:space:]]General[[:space:]]Action-Conditioned[[:space:]]3D[[:space:]]Human[[:space:]]Motion[[:space:]]Generation/b76da01a-6915-4f20-95e0-470a21a063d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Action[[:space:]]Sensitivity[[:space:]]Learning[[:space:]]for[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization/07053da8-e3f8-425e-9d4f-338c187ce557_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Activate[[:space:]]and[[:space:]]Reject_[[:space:]]Towards[[:space:]]Safe[[:space:]]Domain[[:space:]]Generalization[[:space:]]under[[:space:]]Category[[:space:]]Shift/1ff30265-9cf5-4ec1-a982-4f0c39c9392c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Active[[:space:]]Neural[[:space:]]Mapping/45e08854-ae2d-43e4-bc22-971696065335_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Active[[:space:]]Self-Supervised[[:space:]]Learning_[[:space:]]A[[:space:]]Few[[:space:]]Low-Cost[[:space:]]Relationships[[:space:]]Are[[:space:]]All[[:space:]]You[[:space:]]Need/bb9b412a-db56-444a-80cf-4e10a368832a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Active[[:space:]]Stereo[[:space:]]Without[[:space:]]Pattern[[:space:]]Projector/e4edf08c-4066-4c68-bc36-87ebb2f41a2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ActorsNeRF_[[:space:]]Animatable[[:space:]]Few-shot[[:space:]]Human[[:space:]]Rendering[[:space:]]with[[:space:]]Generalizable[[:space:]]NeRFs/819b87ee-4053-47b8-95f4-98ab375a06b4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AdVerb_[[:space:]]Visually[[:space:]]Guided[[:space:]]Audio[[:space:]]Dereverberation/0ef0b7d6-f655-4ab3-8486-3525169ecbf1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ada3D[[:space:]]_[[:space:]]Exploiting[[:space:]]the[[:space:]]Spatial[[:space:]]Redundancy[[:space:]]with[[:space:]]Adaptive[[:space:]]Inference[[:space:]]for[[:space:]]Efficient[[:space:]]3D[[:space:]]Object[[:space:]]Detection/62229bea-a76f-4616-a734-03a1a4a4104f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AdaMV-MoE_[[:space:]]Adaptive[[:space:]]Multi-Task[[:space:]]Vision[[:space:]]Mixture-of-Experts/a3b82d57-612a-4ba8-b11a-e5d8bcc51cc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AdaNIC_[[:space:]]Towards[[:space:]]Practical[[:space:]]Neural[[:space:]]Image[[:space:]]Compression[[:space:]]via[[:space:]]Dynamic[[:space:]]Transform[[:space:]]Routing/ce5a6ab4-4d64-4fac-9330-58921ef41afc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Calibrator[[:space:]]Ensemble_[[:space:]]Navigating[[:space:]]Test[[:space:]]Set[[:space:]]Difficulty[[:space:]]in[[:space:]]Out-of-Distribution[[:space:]]Scenarios/6840c81b-c2f8-423d-a4f4-3cc6357b227a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Frequency[[:space:]]Filters[[:space:]]As[[:space:]]Efficient[[:space:]]Global[[:space:]]Token[[:space:]]Mixers/34c8d01a-9aed-4e57-ba79-91027cdee634_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Illumination[[:space:]]Mapping[[:space:]]for[[:space:]]Shadow[[:space:]]Detection[[:space:]]in[[:space:]]Raw[[:space:]]Images/ec321973-a992-4412-b913-0b95ac8fa575_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Image[[:space:]]Anonymization[[:space:]]in[[:space:]]the[[:space:]]Context[[:space:]]of[[:space:]]Image[[:space:]]Classification[[:space:]]with[[:space:]]Neural[[:space:]]Networks/6b8f49ab-daf6-4127-8933-75035fd7eabf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Nonlinear[[:space:]]Latent[[:space:]]Transformation[[:space:]]for[[:space:]]Conditional[[:space:]]Face[[:space:]]Editing/5b66a900-de52-45bf-8e33-40c0cd8f8b27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Positional[[:space:]]Encoding[[:space:]]for[[:space:]]Bundle-Adjusting[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/8fe1cfda-cacd-4dbe-aabb-87a0e76c4309_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Reordering[[:space:]]Sampler[[:space:]]with[[:space:]]Neurally[[:space:]]Guided[[:space:]]MAGSAC/3e8cfc24-3a9e-4b85-a523-13f50248646c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Rotated[[:space:]]Convolution[[:space:]]for[[:space:]]Rotated[[:space:]]Object[[:space:]]Detection/3cbf325a-7f7f-4801-9da8-73feecf1c9fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Similarity[[:space:]]Bootstrapping[[:space:]]for[[:space:]]Self-Distillation[[:space:]]Based[[:space:]]Representation[[:space:]]Learning/78dccd14-aaee-4e8e-8c3b-2d9ba841e3df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Spiral[[:space:]]Layers[[:space:]]for[[:space:]]Efficient[[:space:]]3D[[:space:]]Representation[[:space:]]Learning[[:space:]]on[[:space:]]Meshes/10f4f8ab-1511-4ac7-bc37-c4104be042a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Superpixel[[:space:]]for[[:space:]]Active[[:space:]]Learning[[:space:]]in[[:space:]]Semantic[[:space:]]Segmentation/b1053ea5-a8dc-4be8-afac-ec4e09fcd994_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Template[[:space:]]Transformer[[:space:]]for[[:space:]]Mitochondria[[:space:]]Segmentation[[:space:]]in[[:space:]]Electron[[:space:]]Microscopy[[:space:]]Images/b066018d-9ac9-45a2-859a-89659fb52f25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]Testing[[:space:]]of[[:space:]]Computer[[:space:]]Vision[[:space:]]Models/65eacf8e-c31f-4e31-b168-625789d092af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adaptive[[:space:]]and[[:space:]]Background-Aware[[:space:]]Vision[[:space:]]Transformer[[:space:]]for[[:space:]]Real-Time[[:space:]]UAV[[:space:]]Tracking/b9e56fde-8156-4339-b3c6-bd1c645e78be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adding[[:space:]]Conditional[[:space:]]Control[[:space:]]to[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/18bb1d94-ef5e-4ef6-922b-af6830c153e7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AdvDiffuser_[[:space:]]Natural[[:space:]]Adversarial[[:space:]]Example[[:space:]]Synthesis[[:space:]]with[[:space:]]Diffusion[[:space:]]Models/14e90b42-6a27-4e59-bdae-11b80b095421_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Advancing[[:space:]]Example[[:space:]]Exploitation[[:space:]]Can[[:space:]]Alleviate[[:space:]]Critical[[:space:]]Challenges[[:space:]]in[[:space:]]Adversarial[[:space:]]Training/d7a5e4e9-c593-4f45-8d25-bbb3f8505e2a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Advancing[[:space:]]Referring[[:space:]]Expression[[:space:]]Segmentation[[:space:]]Beyond[[:space:]]Single[[:space:]]Image/de198203-e5bb-4a3e-b7ce-8d0e6ad3307f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adversarial[[:space:]]Bayesian[[:space:]]Augmentation[[:space:]]for[[:space:]]Single-Source[[:space:]]Domain[[:space:]]Generalization/90395b67-acec-451c-b373-771ab7954b61_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adversarial[[:space:]]Finetuning[[:space:]]with[[:space:]]Latent[[:space:]]Representation[[:space:]]Constraint[[:space:]]to[[:space:]]Mitigate[[:space:]]Accuracy-Robustness[[:space:]]Tradeoff/dfd2fc70-e586-4f60-943b-c4066fc85f83_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Adverse[[:space:]]Weather[[:space:]]Removal[[:space:]]with[[:space:]]Codebook[[:space:]]Priors/b9476420-c479-4089-a474-4d768096bf85_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AerialVLN_[[:space:]]Vision-and-Language[[:space:]]Navigation[[:space:]]for[[:space:]]UAVs/f4eb56f0-33bf-4710-8ab0-62a8620d75cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AesPA-Net_[[:space:]]Aesthetic[[:space:]]Pattern-Aware[[:space:]]Style[[:space:]]Transfer[[:space:]]Networks/e3d6fd42-76a1-46e8-93c0-9d532f013a3b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Affective[[:space:]]Image[[:space:]]Filter_[[:space:]]Reflecting[[:space:]]Emotions[[:space:]]from[[:space:]]Text[[:space:]]to[[:space:]]Images/3a1cd28b-9f9b-4430-8ea4-5f953b0e6ef4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Affine-Consistent[[:space:]]Transformer[[:space:]]for[[:space:]]Multi-Class[[:space:]]Cell[[:space:]]Nuclei[[:space:]]Detection/90dfbc17-b7a6-476f-9cb1-dd113144c7e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AffordPose_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]of[[:space:]]Hand-Object[[:space:]]Interactions[[:space:]]with[[:space:]]Affordance-Driven[[:space:]]Hand[[:space:]]Pose/b81558b9-e669-49af-b794-83290501b565_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Agglomerative[[:space:]]Transformer[[:space:]]for[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection/064d5556-b717-41a7-b94b-ef47add8ee22_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Aggregating[[:space:]]Feature[[:space:]]Point[[:space:]]Cloud[[:space:]]for[[:space:]]Depth[[:space:]]Completion/57327710-a5c3-4568-8fc7-8f5e90d22e76_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Agile[[:space:]]Modeling_[[:space:]]From[[:space:]]Concept[[:space:]]to[[:space:]]Classifier[[:space:]]in[[:space:]]Minutes/0996e453-5743-42f6-83f4-a1e89545c998_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Algebraically[[:space:]]Rigorous[[:space:]]Quaternion[[:space:]]Framework[[:space:]]for[[:space:]]the[[:space:]]Neural[[:space:]]Network[[:space:]]Pose[[:space:]]Estimation[[:space:]]Problem/c2789af8-ac66-4caf-97e7-22fd0dccc7e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AlignDet_[[:space:]]Aligning[[:space:]]Pre-training[[:space:]]and[[:space:]]Fine-tuning[[:space:]]in[[:space:]]Object[[:space:]]Detection/7951fef6-cf36-4371-abf5-9106ccdfaf44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Alignment[[:space:]]Before[[:space:]]Aggregation_[[:space:]]Trajectory[[:space:]]Memory[[:space:]]Retrieval[[:space:]]Network[[:space:]]for[[:space:]]Video[[:space:]]Object[[:space:]]Segmentation/3299758f-b79c-4070-8b73-a235dca68d0b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Alignment-free[[:space:]]HDR[[:space:]]Deghosting[[:space:]]with[[:space:]]Semantics[[:space:]]Consistent[[:space:]]Transformer/6397115b-d33f-402f-b8bc-aba5cf127af5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/All[[:space:]]in[[:space:]]Tokens_[[:space:]]Unifying[[:space:]]Output[[:space:]]Space[[:space:]]of[[:space:]]Visual[[:space:]]Tasks[[:space:]]via[[:space:]]Soft[[:space:]]Token/490d30a5-c9bc-4fb9-9197-aff0b0c766be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/All-to-Key[[:space:]]Attention[[:space:]]for[[:space:]]Arbitrary[[:space:]]Style[[:space:]]Transfer/7c265adc-8502-4af2-927e-48ebce6daac1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/All4One_[[:space:]]Symbiotic[[:space:]]Neighbour[[:space:]]Contrastive[[:space:]]Learning[[:space:]]via[[:space:]]Self-Attention[[:space:]]and[[:space:]]Redundancy[[:space:]]Reduction/eef37b9c-7d78-4872-b59a-eb29e99c62d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Alleviating[[:space:]]Catastrophic[[:space:]]Forgetting[[:space:]]of[[:space:]]Incremental[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Within-Class[[:space:]]and[[:space:]]Between-Class[[:space:]]Knowledge[[:space:]]Distillation/a698eef9-5b10-456d-be28-fa64b3c1034a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Among[[:space:]]Us_[[:space:]]Adversarially[[:space:]]Robust[[:space:]]Collaborative[[:space:]]Perception[[:space:]]by[[:space:]]Consensus/735f0555-6edf-4113-a56b-b6d688d630c9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/An[[:space:]]Adaptive[[:space:]]Model[[:space:]]Ensemble[[:space:]]Adversarial[[:space:]]Attack[[:space:]]for[[:space:]]Boosting[[:space:]]Adversarial[[:space:]]Transferability/16a4045e-5d9c-4f66-9792-e9962df21043_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/An[[:space:]]Embarrassingly[[:space:]]Simple[[:space:]]Backdoor[[:space:]]Attack[[:space:]]on[[:space:]]Self-supervised[[:space:]]Learning/c8954736-2e27-431e-8eb9-bb10b39c4f32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Anatomical[[:space:]]Invariance[[:space:]]Modeling[[:space:]]and[[:space:]]Semantic[[:space:]]Alignment[[:space:]]for[[:space:]]Self-supervised[[:space:]]Learning[[:space:]]in[[:space:]]3D[[:space:]]Medical[[:space:]]Image[[:space:]]Analysis/5d149076-2ca8-4ae3-94a0-78234aa410ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Anchor[[:space:]]Structure[[:space:]]Regularization[[:space:]]Induced[[:space:]]Multi-view[[:space:]]Subspace[[:space:]]Clustering[[:space:]]via[[:space:]]Enhanced[[:space:]]Tensor[[:space:]]Rank[[:space:]]Minimization/b4e8f409-7275-41bb-b806-f9df4d5b8be9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Anchor-Intermediate[[:space:]]Detector_[[:space:]]Decoupling[[:space:]]and[[:space:]]Coupling[[:space:]]Bounding[[:space:]]Boxes[[:space:]]for[[:space:]]Accurate[[:space:]]Object[[:space:]]Detection/2a217926-cbee-4fbd-9499-f3e276838473_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Animal3D_[[:space:]]A[[:space:]]Comprehensive[[:space:]]Dataset[[:space:]]of[[:space:]]3D[[:space:]]Animal[[:space:]]Pose[[:space:]]and[[:space:]]Shape/fb8a762f-f83e-436c-92f0-2531535a13e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Anomaly[[:space:]]Detection[[:space:]]Under[[:space:]]Distribution[[:space:]]Shift/614001b9-6d13-4a4d-a6f1-918463a8c4f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Anomaly[[:space:]]Detection[[:space:]]using[[:space:]]Score-based[[:space:]]Perturbation[[:space:]]Resilience/a762c86a-8277-45df-b03f-a1e661cb8396_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Anti-DreamBooth_[[:space:]]Protecting[[:space:]]Users[[:space:]]from[[:space:]]Personalized[[:space:]]Text-to-image[[:space:]]Synthesis/5200a5bc-ed03-417e-b1fb-928af2fb3ed8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Aperture[[:space:]]Diffraction[[:space:]]for[[:space:]]Compact[[:space:]]Snapshot[[:space:]]Spectral[[:space:]]Imaging/ec539aef-7f27-4fc7-9395-a5d336a964f0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Aria[[:space:]]Digital[[:space:]]Twin_[[:space:]]A[[:space:]]New[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]for[[:space:]]Egocentric[[:space:]]3D[[:space:]]Machine[[:space:]]Perception/2577e516-a079-45b6-8571-43491af89b32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AssetField_[[:space:]]Assets[[:space:]]Mining[[:space:]]and[[:space:]]Reconfiguration[[:space:]]in[[:space:]]Ground[[:space:]]Feature[[:space:]]Plane[[:space:]]Representation/4f308fb5-1497-4bbe-a6f5-230af735d7f3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Atmospheric[[:space:]]Transmission[[:space:]]and[[:space:]]Thermal[[:space:]]Inertia[[:space:]]Induced[[:space:]]Blind[[:space:]]Road[[:space:]]Segmentation[[:space:]]with[[:space:]]a[[:space:]]Large-Scale[[:space:]]Dataset[[:space:]]TBRSD/82f1298a-a557-4664-b3e7-8f3b34a002b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AttT2M_[[:space:]]Text-Driven[[:space:]]Human[[:space:]]Motion[[:space:]]Generation[[:space:]]with[[:space:]]Multi-Perspective[[:space:]]Attention[[:space:]]Mechanism/d8d016b4-1130-4ab5-880b-222bc2c5c97a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Attention[[:space:]]Discriminant[[:space:]]Sampling[[:space:]]for[[:space:]]Point[[:space:]]Clouds/c7989862-4b32-41c1-bafa-0397f2d6f268_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Attention[[:space:]]Where[[:space:]]It[[:space:]]Matters_[[:space:]]Rethinking[[:space:]]Visual[[:space:]]Document[[:space:]]Understanding[[:space:]]with[[:space:]]Selective[[:space:]]Region[[:space:]]Concentration/88cbdbad-7b4f-4b93-bc29-cd28808b1bc0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Attentive[[:space:]]Mask[[:space:]]CLIP/620a0c4c-7dfa-4a21-8cb0-f8b13fb60e93_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Audio-Enhanced[[:space:]]Text-to-Video[[:space:]]Retrieval[[:space:]]using[[:space:]]Text-Conditioned[[:space:]]Feature[[:space:]]Alignment/4641d8d4-90cc-4cc1-9f0d-98a434ae8655_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Audio-Visual[[:space:]]Class-Incremental[[:space:]]Learning/36127bb6-0e22-4beb-b983-8d9d10a2d518_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Audio-Visual[[:space:]]Deception[[:space:]]Detection_[[:space:]]DOLOS[[:space:]]Dataset[[:space:]]and[[:space:]]Parameter-Efficient[[:space:]]Crossmodal[[:space:]]Learning/660a5db4-4abf-4735-ac47-a0edf1cf44cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Audio-Visual[[:space:]]Glance[[:space:]]Network[[:space:]]for[[:space:]]Efficient[[:space:]]Video[[:space:]]Recognition/fdb4fe02-4f7d-4f8e-9527-900c7475bfcd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Audiovisual[[:space:]]Masked[[:space:]]Autoencoders/44d76770-e1a0-4e59-9206-6f0266ce5052_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Augmented[[:space:]]Box[[:space:]]Replay_[[:space:]]Overcoming[[:space:]]Foreground[[:space:]]Shift[[:space:]]for[[:space:]]Incremental[[:space:]]Object[[:space:]]Detection/5899cb6a-e303-4fac-a478-7145125ed278_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Augmenting[[:space:]]and[[:space:]]Aligning[[:space:]]Snippets[[:space:]]for[[:space:]]Few-Shot[[:space:]]Video[[:space:]]Domain[[:space:]]Adaptation/28037b88-d1b5-4db5-a929-1f91f907cf9f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AutoAD[[:space:]]II_[[:space:]]The[[:space:]]Sequel[[:space:]]-[[:space:]]Who,[[:space:]]When,[[:space:]]and[[:space:]]What[[:space:]]in[[:space:]]Movie[[:space:]]Audio[[:space:]]Description/8a4de17e-7514-48bf-86cf-33604815a974_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AutoDiffusion_[[:space:]]Training-Free[[:space:]]Optimization[[:space:]]of[[:space:]]Time[[:space:]]Steps[[:space:]]and[[:space:]]Architectures[[:space:]]for[[:space:]]Automated[[:space:]]Diffusion[[:space:]]Model[[:space:]]Acceleration/c2937e48-570b-4703-bb47-16acc1bcede8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AutoReP_[[:space:]]Automatic[[:space:]]ReLU[[:space:]]Replacement[[:space:]]for[[:space:]]Fast[[:space:]]Private[[:space:]]Network[[:space:]]Inference/bba066c6-078a-4719-bb4f-3847567b61fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AutoSynth_[[:space:]]Learning[[:space:]]to[[:space:]]Generate[[:space:]]3D[[:space:]]Training[[:space:]]Data[[:space:]]for[[:space:]]Object[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/2110ac07-8453-4a50-9e60-85ed462a8dc8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Automated[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]via[[:space:]]Monte[[:space:]]Carlo[[:space:]]Tree[[:space:]]Search/b9f40f64-d4d0-42f3-af61-c5f7fd7d8911_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Automatic[[:space:]]Animation[[:space:]]of[[:space:]]Hair[[:space:]]Blowing[[:space:]]in[[:space:]]Still[[:space:]]Portrait[[:space:]]Photos/dd0c2814-be7e-47d6-b16a-2739ee1306e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Automatic[[:space:]]Network[[:space:]]Pruning[[:space:]]via[[:space:]]Hilbert-Schmidt[[:space:]]Independence[[:space:]]Criterion[[:space:]]Lasso[[:space:]]under[[:space:]]Information[[:space:]]Bottleneck[[:space:]]Principle/f9247347-b031-4c9e-9fd2-9891379d5ff6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Auxiliary[[:space:]]Tasks[[:space:]]Benefit[[:space:]]3D[[:space:]]Skeleton-based[[:space:]]Human[[:space:]]Motion[[:space:]]Prediction/26b71198-5d38-4ad8-b3a0-c20e7c69884b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/AvatarCraft_[[:space:]]Transforming[[:space:]]Text[[:space:]]into[[:space:]]Neural[[:space:]]Human[[:space:]]Avatars[[:space:]]with[[:space:]]Parameterized[[:space:]]Shape[[:space:]]and[[:space:]]Pose[[:space:]]Control/548f9086-092d-4696-8de5-1c979355b556_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BANSAC_[[:space:]]A[[:space:]]Dynamic[[:space:]]BAyesian[[:space:]]Network[[:space:]]for[[:space:]]Adaptive[[:space:]]SAmple[[:space:]]Consensus/eb361771-fff9-4e8e-8ce5-fc722cb85052_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BEV-DG_[[:space:]]Cross-Modal[[:space:]]Learning[[:space:]]under[[:space:]]Bird's-Eye[[:space:]]View[[:space:]]for[[:space:]]Domain[[:space:]]Generalization[[:space:]]of[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation/1c05aa72-f6b0-4288-8ce7-70bef8c3c806_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BEVPlace_[[:space:]]Learning[[:space:]]LiDAR-based[[:space:]]Place[[:space:]]Recognition[[:space:]]using[[:space:]]Bird's[[:space:]]Eye[[:space:]]View[[:space:]]Images/145dd9f6-2ab0-4ace-9d25-36b5cff24f41_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BT^2_[[:space:]]Backward-compatible[[:space:]]Training[[:space:]]with[[:space:]]Basis[[:space:]]Transformation/e9fc0818-b369-4e3b-886a-5dae9fe85f75_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BUS_[[:space:]]Efficient[[:space:]]and[[:space:]]Effective[[:space:]]Vision-Language[[:space:]]Pre-Training[[:space:]]with[[:space:]]Bottom-Up[[:space:]]Patch[[:space:]]Summarization./964415d2-1e8c-4927-951d-fd3036bc7e71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BaRe-ESA_[[:space:]]A[[:space:]]Riemannian[[:space:]]Framework[[:space:]]for[[:space:]]Unregistered[[:space:]]Human[[:space:]]Body[[:space:]]Shapes/56ddf68f-a1dd-4043-8740-c0a43e880cc1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Backpropagation[[:space:]]Path[[:space:]]Search[[:space:]]On[[:space:]]Adversarial[[:space:]]Transferability/332f3014-8556-47ad-8d85-1800b5162bdb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BallGAN_[[:space:]]3D-aware[[:space:]]Image[[:space:]]Synthesis[[:space:]]with[[:space:]]a[[:space:]]Spherical[[:space:]]Background/0f4a1941-abf4-4ead-9ada-a4cc70e5832d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Batch-based[[:space:]]Model[[:space:]]Registration[[:space:]]for[[:space:]]Fast[[:space:]]3D[[:space:]]Sherd[[:space:]]Reconstruction/c63cf54c-fcd4-4aee-83fa-b17e4041378d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bayesian[[:space:]]Optimization[[:space:]]Meets[[:space:]]Self-Distillation/a5e81bb2-dc38-4324-80fb-5af603d2ec38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bayesian[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Image-Language[[:space:]]Model[[:space:]]Generalization/76432fc8-fad1-4e3f-a371-2311fce0e33e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Be[[:space:]]Everywhere[[:space:]]-[[:space:]]Hear[[:space:]]Everything[[:space:]](BEE)_[[:space:]]Audio[[:space:]]Scene[[:space:]]Reconstruction[[:space:]]by[[:space:]]Sparse[[:space:]]Audio-Visual[[:space:]]Samples/962c13d5-835e-4163-9939-82922c5c866a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BeLFusion_[[:space:]]Latent[[:space:]]Diffusion[[:space:]]for[[:space:]]Behavior-Driven[[:space:]]Human[[:space:]]Motion[[:space:]]Prediction/6d0f6649-de8c-4300-b553-157ee92e3972_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Beating[[:space:]]Backdoor[[:space:]]Attack[[:space:]]at[[:space:]]Its[[:space:]]Own[[:space:]]Game/a0d6c4ba-384b-43e6-889a-8fc72426ae2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Benchmarking[[:space:]]Algorithmic[[:space:]]Bias[[:space:]]in[[:space:]]Face[[:space:]]Recognition_[[:space:]]An[[:space:]]Experimental[[:space:]]Approach[[:space:]]Using[[:space:]]Synthetic[[:space:]]Faces[[:space:]]and[[:space:]]Human[[:space:]]Evaluation/25992a40-9f55-4552-bae7-dc4bf5b0168b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Benchmarking[[:space:]]Low-Shot[[:space:]]Robustness[[:space:]]to[[:space:]]Natural[[:space:]]Distribution[[:space:]]Shifts/f007f7a4-fa03-4f7d-8076-a960f89c1749_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Benchmarking[[:space:]]and[[:space:]]Analyzing[[:space:]]Robust[[:space:]]Point[[:space:]]Cloud[[:space:]]Recognition_[[:space:]]Bag[[:space:]]of[[:space:]]Tricks[[:space:]]for[[:space:]]Defending[[:space:]]Adversarial[[:space:]]Examples/b0eff3fe-514b-4dcd-bba1-e2779a9e415f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Betrayed[[:space:]]by[[:space:]]Captions_[[:space:]]Joint[[:space:]]Caption[[:space:]]Grounding[[:space:]]and[[:space:]]Generation[[:space:]]for[[:space:]]Open[[:space:]]Vocabulary[[:space:]]Instance[[:space:]]Segmentation/d4bc1f1e-c631-4c50-ac0e-60e80d3831ce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Better[[:space:]]May[[:space:]]Not[[:space:]]Be[[:space:]]Fairer_[[:space:]]A[[:space:]]Study[[:space:]]on[[:space:]]Subgroup[[:space:]]Discrepancy[[:space:]]in[[:space:]]Image[[:space:]]Classification/1fd68b2d-18df-4c00-8b64-52b47e610535_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Beyond[[:space:]]Image[[:space:]]Borders_[[:space:]]Learning[[:space:]]Feature[[:space:]]Extrapolation[[:space:]]for[[:space:]]Unbounded[[:space:]]Image[[:space:]]Composition/0105360a-81c5-4a52-a8d9-72e5c9e98239_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Beyond[[:space:]]Object[[:space:]]Recognition_[[:space:]]A[[:space:]]New[[:space:]]Benchmark[[:space:]]towards[[:space:]]Object[[:space:]]Concept[[:space:]]Learning/d2d0a888-0080-4fe0-b807-6ae33b15a1f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Beyond[[:space:]]One-to-One_[[:space:]]Rethinking[[:space:]]the[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation/82a0f339-1bf8-4918-8253-28efbae0b467_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Beyond[[:space:]]Single[[:space:]]Path[[:space:]]Integrated[[:space:]]Gradients[[:space:]]for[[:space:]]Reliable[[:space:]]Input[[:space:]]Attribution[[:space:]]via[[:space:]]Randomized[[:space:]]Path[[:space:]]Sampling/16846a68-2885-4bf6-adb6-495c55131b5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Beyond[[:space:]]Skin[[:space:]]Tone_[[:space:]]A[[:space:]]Multidimensional[[:space:]]Measure[[:space:]]of[[:space:]]Apparent[[:space:]]Skin[[:space:]]Color/94f0be82-2d17-47a3-a668-ea04929aac5c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Beyond[[:space:]]the[[:space:]]Limitation[[:space:]]of[[:space:]]Monocular[[:space:]]3D[[:space:]]Detector[[:space:]]via[[:space:]]Knowledge[[:space:]]Distillation/4f01d9ee-628c-4b8b-9605-683d846ae980_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Beyond[[:space:]]the[[:space:]]Pixel_[[:space:]]a[[:space:]]Photometrically[[:space:]]Calibrated[[:space:]]HDR[[:space:]]Dataset[[:space:]]for[[:space:]]Luminance[[:space:]]and[[:space:]]Color[[:space:]]Prediction/ad8f6ebe-8046-4ed5-af6a-0d780384ab74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BiFF_[[:space:]]Bi-level[[:space:]]Future[[:space:]]Fusion[[:space:]]with[[:space:]]Polyline-based[[:space:]]Coordinate[[:space:]]for[[:space:]]Interactive[[:space:]]Trajectory[[:space:]]Prediction/33c457fb-ce5d-40b7-8c50-b9c66b05b6a1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BiViT_[[:space:]]Extremely[[:space:]]Compressed[[:space:]]Binary[[:space:]]Vision[[:space:]]Transformers/6f702783-035a-409d-a78f-23b0f094efd0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bidirectional[[:space:]]Alignment[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Detection[[:space:]]with[[:space:]]Transformers/ac1366ff-cec7-4f5f-bd97-f79685ecef49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bidirectionally[[:space:]]Deformable[[:space:]]Motion[[:space:]]Modulation[[:space:]]For[[:space:]]Video-based[[:space:]]Human[[:space:]]Pose[[:space:]]Transfer/c3f86862-18ed-428e-a8a9-4aa17d3089f9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bird's-Eye-View[[:space:]]Scene[[:space:]]Graph[[:space:]]for[[:space:]]Vision-Language[[:space:]]Navigation/e576ab1f-e964-4702-a505-9686e1786a02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Black[[:space:]]Box[[:space:]]Few-Shot[[:space:]]Adaptation[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/fb6c9861-f3e8-4280-a4b4-690c978859ee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Black-Box[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation[[:space:]]with[[:space:]]Bi-Directional[[:space:]]Atkinson-Shiffrin[[:space:]]Memory/268f92d0-575d-408e-b0d2-6918ce4d3a44_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BlendFace_[[:space:]]Re-designing[[:space:]]Identity[[:space:]]Encoders[[:space:]]for[[:space:]]Face-Swapping/d289da29-f8bc-45f4-a6f9-faae129fc46a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Blending-NeRF_[[:space:]]Text-Driven[[:space:]]Localized[[:space:]]Editing[[:space:]]in[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/fc50f8f5-a5ad-48f1-9a4d-57fe6e91b778_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BlindHarmony_[[:space:]]'Blind'[[:space:]]Harmonization[[:space:]]for[[:space:]]MR[[:space:]]Images[[:space:]]via[[:space:]]Flow[[:space:]]Model/e0be287e-18d7-448d-9895-d7f9a400171f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BoMD_[[:space:]]Bag[[:space:]]of[[:space:]]Multi-label[[:space:]]Descriptors[[:space:]]for[[:space:]]Noisy[[:space:]]Chest[[:space:]]X-ray[[:space:]]Classification/e7ddbe22-5388-4e11-9ca7-5aa11e983bcc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Body[[:space:]]Knowledge[[:space:]]and[[:space:]]Uncertainty[[:space:]]Modeling[[:space:]]for[[:space:]]Monocular[[:space:]]3D[[:space:]]Human[[:space:]]Body[[:space:]]Reconstruction/a72e2d94-94c6-4d28-b536-4d62698e83cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bold[[:space:]]but[[:space:]]Cautious_[[:space:]]Unlocking[[:space:]]the[[:space:]]Potential[[:space:]]of[[:space:]]Personalized[[:space:]]Federated[[:space:]]Learning[[:space:]]through[[:space:]]Cautiously[[:space:]]Aggressive[[:space:]]Collaboration/a00f44f4-6f41-47ea-b292-61a0a1ca8222_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]3-DoF[[:space:]]Ground-to-Satellite[[:space:]]Camera[[:space:]]Localization[[:space:]]Accuracy[[:space:]]via[[:space:]]Geometry-Guided[[:space:]]Cross-View[[:space:]]Transformer/d1fab4de-11cc-4648-9320-ffd5858c4dff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Adversarial[[:space:]]Transferability[[:space:]]via[[:space:]]Gradient[[:space:]]Relevance[[:space:]]Attack/1d43d08a-ff52-4680-9a76-c2c11f354533_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Few-shot[[:space:]]Action[[:space:]]Recognition[[:space:]]with[[:space:]]Graph-guided[[:space:]]Hybrid[[:space:]]Matching/f4adc367-73d3-4cba-ab43-0614de5b0dac_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Long-tailed[[:space:]]Object[[:space:]]Detection[[:space:]]via[[:space:]]Step-wise[[:space:]]Learning[[:space:]]on[[:space:]]Smooth-tail[[:space:]]Data/a90cde8b-f45d-45dd-abe4-5a28c3b7c46b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Multi-modal[[:space:]]Model[[:space:]]Performance[[:space:]]with[[:space:]]Adaptive[[:space:]]Gradient[[:space:]]Modulation/850196e6-21b4-4984-8742-b9fc1cd1cf97_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Novel[[:space:]]Category[[:space:]]Discovery[[:space:]]Over[[:space:]]Domains[[:space:]]with[[:space:]]Soft[[:space:]]Contrastive[[:space:]]Learning[[:space:]]and[[:space:]]All[[:space:]]in[[:space:]]One[[:space:]]Classifier/ec551b03-9c67-4e98-a32c-d3cd2db75648_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Positive[[:space:]]Segments[[:space:]]for[[:space:]]Weakly-Supervised[[:space:]]Audio-Visual[[:space:]]Video[[:space:]]Parsing/65639b21-878a-426d-b0aa-0af6ca30ef7e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]from[[:space:]]the[[:space:]]Perspective[[:space:]]of[[:space:]]Explicit[[:space:]]Class[[:space:]]Embeddings/ada4b0cc-d28e-4ad9-89a9-0d62ce2d56dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Single[[:space:]]Image[[:space:]]Super-Resolution[[:space:]]via[[:space:]]Partial[[:space:]]Channel[[:space:]]Shifting/44007de9-c2c6-4ac1-b5f9-1942f53cbea6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boosting[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Classification[[:space:]]from[[:space:]]the[[:space:]]Perspectives[[:space:]]of[[:space:]]Distribution,[[:space:]]Correlation[[:space:]]and[[:space:]]Magnification/c3286ebc-a816-4e01-9bc7-9b64fb8c19ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bootstrap[[:space:]]Motion[[:space:]]Forecasting[[:space:]]With[[:space:]]Self-Consistent[[:space:]]Constraints/918f6124-b147-4f84-97c4-17522d0536df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Borrowing[[:space:]]Knowledge[[:space:]]From[[:space:]]Pre-trained[[:space:]]Language[[:space:]]Model_[[:space:]]A[[:space:]]New[[:space:]]Data-efficient[[:space:]]Visual[[:space:]]Learning[[:space:]]Paradigm/c4def6a4-c550-4777-9e9b-dce74ddaaf2e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Both[[:space:]]Diverse[[:space:]]and[[:space:]]Realism[[:space:]]Matter_[[:space:]]Physical[[:space:]]Attribute[[:space:]]and[[:space:]]Style[[:space:]]Alignment[[:space:]]for[[:space:]]Rainy[[:space:]]Image[[:space:]]Generation/57245fec-d2c0-4e16-b6c9-20af393b9de9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Boundary-Aware[[:space:]]Divide[[:space:]]and[[:space:]]Conquer_[[:space:]]A[[:space:]]Diffusion-Based[[:space:]]Solution[[:space:]]for[[:space:]]Unsupervised[[:space:]]Shadow[[:space:]]Removal/c312a37a-ce44-48b1-aa39-8046f1366449_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Box-based[[:space:]]Refinement[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]and[[:space:]]Unsupervised[[:space:]]Localization[[:space:]]Tasks/b0439873-3800-4f42-85d7-d84faadf5881_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BoxDiff_[[:space:]]Text-to-Image[[:space:]]Synthesis[[:space:]]with[[:space:]]Training-Free[[:space:]]Box-Constrained[[:space:]]Diffusion/a810d773-3d84-42ee-855d-74ff9090ca1d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/BoxSnake_[[:space:]]Polygonal[[:space:]]Instance[[:space:]]Segmentation[[:space:]]with[[:space:]]Box[[:space:]]Supervision/2e87fa44-b644-4bb9-875a-3241bace9044_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Breaking[[:space:]]Common[[:space:]]Sense_[[:space:]]WHOOPS![[:space:]]A[[:space:]]Vision-and-Language[[:space:]]Benchmark[[:space:]]of[[:space:]]Synthetic[[:space:]]and[[:space:]]Compositional[[:space:]]Images/bc416834-cbf6-4234-aed3-aa1e6726a42c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Breaking[[:space:]]Temporal[[:space:]]Consistency_[[:space:]]Generating[[:space:]]Video[[:space:]]Universal[[:space:]]Adversarial[[:space:]]Perturbations[[:space:]]Using[[:space:]]Image[[:space:]]Models/be812d96-d6d6-484e-93b6-7dee793fe02a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Breaking[[:space:]]The[[:space:]]Limits[[:space:]]of[[:space:]]Text-conditioned[[:space:]]3D[[:space:]]Motion[[:space:]]Synthesis[[:space:]]with[[:space:]]Elaborative[[:space:]]Descriptions/d8d12afc-1c77-4a98-b499-7ded386c9c55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bridging[[:space:]]Cross-task[[:space:]]Protocol[[:space:]]Inconsistency[[:space:]]for[[:space:]]Distillation[[:space:]]in[[:space:]]Dense[[:space:]]Object[[:space:]]Detection/007e3000-4134-4a11-8407-da4a1542cdfb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bridging[[:space:]]Vision[[:space:]]and[[:space:]]Language[[:space:]]Encoders_[[:space:]]Parameter-Efficient[[:space:]]Tuning[[:space:]]for[[:space:]]Referring[[:space:]]Image[[:space:]]Segmentation/7786eefd-5fbc-434b-a6cc-deb0a8fab5cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Bring[[:space:]]Clipart[[:space:]]to[[:space:]]Life/e8a23499-a451-4c9d-9f17-f1b8ed8c09da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Building[[:space:]]Bridge[[:space:]]Across[[:space:]]the[[:space:]]Time_[[:space:]]Disruption[[:space:]]and[[:space:]]Restoration[[:space:]]of[[:space:]]Murals[[:space:]]In[[:space:]]the[[:space:]]Wild/ed2a4336-8cdf-4df3-a6a3-d67a5915b6ad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Building[[:space:]]Vision[[:space:]]Transformers[[:space:]]with[[:space:]]Hierarchy[[:space:]]Aware[[:space:]]Feature[[:space:]]Aggregation/54e42720-aceb-4f75-893f-06805b825507_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Building[[:space:]]a[[:space:]]Winning[[:space:]]Team_[[:space:]]Selecting[[:space:]]Source[[:space:]]Model[[:space:]]Ensembles[[:space:]]using[[:space:]]a[[:space:]]Submodular[[:space:]]Transferability[[:space:]]Estimation[[:space:]]Approach/f84195db-fb7e-4c3d-b19c-4ebb3858fd92_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Building3D_[[:space:]]A[[:space:]]Urban-Scale[[:space:]]Dataset[[:space:]]and[[:space:]]Benchmarks[[:space:]]for[[:space:]]Learning[[:space:]]Roof[[:space:]]Structures[[:space:]]from[[:space:]]Point[[:space:]]Clouds/628b56f2-5e86-4617-b876-eac5510cb0a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/C2F2NeUS_[[:space:]]Cascade[[:space:]]Cost[[:space:]]Frustum[[:space:]]Fusion[[:space:]]for[[:space:]]High[[:space:]]Fidelity[[:space:]]and[[:space:]]Generalizable[[:space:]]Neural[[:space:]]Surface[[:space:]]Reconstruction/9a5cd15e-0ae5-4b7b-86d1-953eaec729d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/C2ST_[[:space:]]Cross-Modal[[:space:]]Contextualized[[:space:]]Sequence[[:space:]]Transduction[[:space:]]for[[:space:]]Continuous[[:space:]]Sign[[:space:]]Language[[:space:]]Recognition/cb8d47c8-feef-457c-9e7c-1134004c8fb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CAD-Estate_[[:space:]]Large-scale[[:space:]]CAD[[:space:]]Model[[:space:]]Annotation[[:space:]]in[[:space:]]RGB[[:space:]]Videos/e459e2aa-fd9b-49a0-bf2f-dbee0ddf85e2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CAFA_[[:space:]]Class-Aware[[:space:]]Feature[[:space:]]Alignment[[:space:]]for[[:space:]]Test-Time[[:space:]]Adaptation/13ff01bb-41c3-4962-9b0e-9fe34e86d66e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CAME_[[:space:]]Contrastive[[:space:]]Automated[[:space:]]Model[[:space:]]Evaluation/a9972025-ac0f-45b3-a963-20d9ccb39cdd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CASSPR_[[:space:]]Cross[[:space:]]Attention[[:space:]]Single[[:space:]]Scan[[:space:]]Place[[:space:]]Recognition/6f556eb3-49b7-4bf4-ad10-4222619fdae0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CBA_[[:space:]]Improving[[:space:]]Online[[:space:]]Continual[[:space:]]Learning[[:space:]]via[[:space:]]Continual[[:space:]]Bias[[:space:]]Adaptor/7be936d4-8bc8-4ea4-9731-13bab20c43f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CC3D_[[:space:]]Layout-Conditioned[[:space:]]Generation[[:space:]]of[[:space:]]Compositional[[:space:]]3D[[:space:]]Scenes/bcedc01e-866e-4543-8d3d-c580eee5d45b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CDAC_[[:space:]]Cross-domain[[:space:]]Attention[[:space:]]Consistency[[:space:]]in[[:space:]]Transformer[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Semantic[[:space:]]Segmentation/15020929-2b12-419a-be22-9e6e95d65e9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CDFSL-V_[[:space:]]Cross-Domain[[:space:]]Few-Shot[[:space:]]Learning[[:space:]]for[[:space:]]Videos/c6d2e0e1-1bfd-4dfe-a9f7-6831d31fad58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CDUL_[[:space:]]CLIP-Driven[[:space:]]Unsupervised[[:space:]]Learning[[:space:]]for[[:space:]]Multi-Label[[:space:]]Image[[:space:]]Classification/8a9bedd5-beba-4390-b380-198702a3c53b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CFCG_[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]via[[:space:]]Cross-Fusion[[:space:]]and[[:space:]]Contour[[:space:]]Guidance[[:space:]]Supervision/4ef5a4ee-3f5f-4066-bfb4-259b79db550a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CGBA_[[:space:]]Curvature-aware[[:space:]]Geometric[[:space:]]Black-box[[:space:]]Attack/bcc3971f-c940-4977-ae39-c3fd838c5124_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CHAMPAGNE_[[:space:]]Learning[[:space:]]Real-world[[:space:]]Conversation[[:space:]]from[[:space:]]Large-Scale[[:space:]]Web[[:space:]]Videos/354e527e-1fd4-4883-8b81-9e2d85228822_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CHORD_[[:space:]]Category-level[[:space:]]Hand-held[[:space:]]Object[[:space:]]Reconstruction[[:space:]]via[[:space:]]Shape[[:space:]]Deformation/5aa2b6c8-61a6-4d0f-a411-7bd968ab3e8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CIRI_[[:space:]]Curricular[[:space:]]Inactivation[[:space:]]for[[:space:]]Residue-aware[[:space:]]One-shot[[:space:]]Video[[:space:]]Inpainting/5d3624e9-a32e-4e35-b28a-bf70e6f8cceb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CL-MVSNet_[[:space:]]Unsupervised[[:space:]]Multi-View[[:space:]]Stereo[[:space:]]with[[:space:]]Dual-Level[[:space:]]Contrastive[[:space:]]Learning/6008f564-ab71-4bf4-a62e-7772d54b6132_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIP-Cluster_[[:space:]]CLIP-Guided[[:space:]]Attribute[[:space:]]Hallucination[[:space:]]for[[:space:]]Face[[:space:]]Clustering/2d55bed8-155f-40e9-af66-014228210f02_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIP-Driven[[:space:]]Universal[[:space:]]Model[[:space:]]for[[:space:]]Organ[[:space:]]Segmentation[[:space:]]and[[:space:]]Tumor[[:space:]]Detection/c723a803-3f13-4def-8096-fbfa0051c344_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIP2Point_[[:space:]]Transfer[[:space:]]CLIP[[:space:]]to[[:space:]]Point[[:space:]]Cloud[[:space:]]Classification[[:space:]]with[[:space:]]Image-Depth[[:space:]]Pre-Training/d237674f-5b78-4c19-8887-2dc61d0e0de4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIPN[[:space:]]for[[:space:]]Zero-Shot[[:space:]]OOD[[:space:]]Detection_[[:space:]]Teaching[[:space:]]CLIP[[:space:]]to[[:space:]]Say[[:space:]]No/e36cc0bd-2e25-449b-a125-09275b1a7cc9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIPTER_[[:space:]]Looking[[:space:]]at[[:space:]]the[[:space:]]Bigger[[:space:]]Picture[[:space:]]in[[:space:]]Scene[[:space:]]Text[[:space:]]Recognition/2e2f622b-ce47-482a-afa2-0b7b8d04153d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIPTrans_[[:space:]]Transferring[[:space:]]Visual[[:space:]]Knowledge[[:space:]]with[[:space:]]Pre-trained[[:space:]]Models[[:space:]]for[[:space:]]Multimodal[[:space:]]Machine[[:space:]]Translation/199df6c4-a234-4f90-a56c-5ad1a8e4e5b3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLIPascene_[[:space:]]Scene[[:space:]]Sketching[[:space:]]with[[:space:]]Different[[:space:]]Types[[:space:]]and[[:space:]]Levels[[:space:]]of[[:space:]]Abstraction/eabd525a-fc04-4438-b209-0bbb21ff39f4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLNeRF_[[:space:]]Continual[[:space:]]Learning[[:space:]]Meets[[:space:]]NeRF/5d251cdb-a00c-489e-aeb5-675366ebc581_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CLR_[[:space:]]Channel-wise[[:space:]]Lightweight[[:space:]]Reprogramming[[:space:]]for[[:space:]]Continual[[:space:]]Learning/8cb91868-875e-4aa4-83c0-c9b3d287744f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CMDA_[[:space:]]Cross-Modality[[:space:]]Domain[[:space:]]Adaptation[[:space:]]for[[:space:]]Nighttime[[:space:]]Semantic[[:space:]]Segmentation/6784169d-0815-4e47-9451-981dcbeb1c8b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CO-Net_[[:space:]]Learning[[:space:]]Multiple[[:space:]]Point[[:space:]]Cloud[[:space:]]Tasks[[:space:]]at[[:space:]]Once[[:space:]]with[[:space:]]A[[:space:]]Cohesive[[:space:]]Network/e12e368a-1e12-4394-a1b5-f1ed91c45726_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CO-PILOT_[[:space:]]Dynamic[[:space:]]Top-Down[[:space:]]Point[[:space:]]Cloud[[:space:]]with[[:space:]]Conditional[[:space:]]Neighborhood[[:space:]]Aggregation[[:space:]]for[[:space:]]Multi-Gigapixel[[:space:]]Histopathology[[:space:]]Image[[:space:]]Representation/809ec09c-1810-4886-bee7-8b044ab33a21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/COCO-O_[[:space:]]A[[:space:]]Benchmark[[:space:]]for[[:space:]]Object[[:space:]]Detectors[[:space:]]under[[:space:]]Natural[[:space:]]Distribution[[:space:]]Shifts/4bb376f3-3ab4-42a1-ab62-a89f02d9c986_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/COMPASS_[[:space:]]High-Efficiency[[:space:]]Deep[[:space:]]Image[[:space:]]Compression[[:space:]]with[[:space:]]Arbitrary-scale[[:space:]]Spatial[[:space:]]Scalability/71b890ab-daf2-45ff-9d26-ac19129c3af0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/COOL-CHIC_[[:space:]]Coordinate-based[[:space:]]Low[[:space:]]Complexity[[:space:]]Hierarchical[[:space:]]Image[[:space:]]Codec/52f65801-14ee-4c89-ab06-a461895e31d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/COOP_[[:space:]]Decoupling[[:space:]]and[[:space:]]Coupling[[:space:]]of[[:space:]]Whole-Body[[:space:]]Grasping[[:space:]]Pose[[:space:]]Generation/63687258-d79d-446f-8aa7-4fa43949ad04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/COPILOT_[[:space:]]Human-Environment[[:space:]]Collision[[:space:]]Prediction[[:space:]]and[[:space:]]Localization[[:space:]]from[[:space:]]Egocentric[[:space:]]Videos/0c8bf998-0418-4286-ad36-2baf369f9f3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CORE_[[:space:]]Co-planarity[[:space:]]Regularized[[:space:]]Monocular[[:space:]]Geometry[[:space:]]Estimation[[:space:]]with[[:space:]]Weak[[:space:]]Supervision/f2c51d8c-c4b2-4974-965e-09234155ab79_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CORE_[[:space:]]Cooperative[[:space:]]Reconstruction[[:space:]]for[[:space:]]Multi-Agent[[:space:]]Perception/b2d4f6a8-8a11-4088-a3bf-7458c1dc0239_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CPCM_[[:space:]]Contextual[[:space:]]Point[[:space:]]Cloud[[:space:]]Modeling[[:space:]]for[[:space:]]Weakly-supervised[[:space:]]Point[[:space:]]Cloud[[:space:]]Semantic[[:space:]]Segmentation/a437218a-1000-4404-a102-675411ef710d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CRN_[[:space:]]Camera[[:space:]]Radar[[:space:]]Net[[:space:]]for[[:space:]]Accurate,[[:space:]]Robust,[[:space:]]Efficient[[:space:]]3D[[:space:]]Perception/6892d69e-9b1e-41e8-8b94-4cda7b86f151_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CROSSFIRE_[[:space:]]Camera[[:space:]]Relocalization[[:space:]]On[[:space:]]Self-Supervised[[:space:]]Features[[:space:]]from[[:space:]]an[[:space:]]Implicit[[:space:]]Representation/34d18d94-3a25-473f-b994-3c894595a9fb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CSDA_[[:space:]]Learning[[:space:]]Category-Scale[[:space:]]Joint[[:space:]]Feature[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Object[[:space:]]Detection/d60668a8-565b-4964-8d5f-605128735b27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CTP_Towards[[:space:]]Vision-Language[[:space:]]Continual[[:space:]]Pretraining[[:space:]]via[[:space:]]Compatible[[:space:]]Momentum[[:space:]]Contrast[[:space:]]and[[:space:]]Topology[[:space:]]Preservation/3dfd70bc-0444-4a1e-a837-86154ea26b5a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CTVIS_[[:space:]]Consistent[[:space:]]Training[[:space:]]for[[:space:]]Online[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation/7f2234ea-9513-472a-895a-34530cd131e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CVRecon_[[:space:]]Rethinking[[:space:]]3D[[:space:]]Geometric[[:space:]]Feature[[:space:]]Learning[[:space:]]For[[:space:]]Neural[[:space:]]Reconstruction/ab34b620-6bc2-44e6-b1eb-a32a40d476af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CVSformer_[[:space:]]Cross-View[[:space:]]Synthesis[[:space:]]Transformer[[:space:]]for[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion/cc30b610-66d7-4d95-946f-7de3e12124d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CaPhy_[[:space:]]Capturing[[:space:]]Physical[[:space:]]Properties[[:space:]]for[[:space:]]Animatable[[:space:]]Human[[:space:]]Avatars/f7ab2d55-c226-4fc4-9d36-01ffa426d971_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Calibrating[[:space:]]Panoramic[[:space:]]Depth[[:space:]]Estimation[[:space:]]for[[:space:]]Practical[[:space:]]Localization[[:space:]]and[[:space:]]Mapping/8ee23286-5acc-4bf2-84d5-31c8b935436d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Calibrating[[:space:]]Uncertainty[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Crowd[[:space:]]Counting/e8b5ec09-3fec-482c-bc64-64a3df86a1df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Camera-Driven[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Person[[:space:]]Re-identification/4370f479-1a0e-4d7f-9b09-f95d7c7d643b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Can[[:space:]]Language[[:space:]]Models[[:space:]]Learn[[:space:]]to[[:space:]]Listen_/356854fb-a1ac-401f-ac4a-2667564e6381_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CancerUniT_[[:space:]]Towards[[:space:]]a[[:space:]]Single[[:space:]]Unified[[:space:]]Model[[:space:]]for[[:space:]]Effective[[:space:]]Detection,[[:space:]]Segmentation,[[:space:]]and[[:space:]]Diagnosis[[:space:]]of[[:space:]]Eight[[:space:]]Major[[:space:]]Cancers[[:space:]]Using[[:space:]]a[[:space:]]Large[[:space:]]Collection[[:space:]]of[[:space:]]CT[[:space:]]Scans/e5880b54-7701-4200-8655-2c0bd8e9c09d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Candidate-aware[[:space:]]Selective[[:space:]]Disambiguation[[:space:]]Based[[:space:]]On[[:space:]]Normalized[[:space:]]Entropy[[:space:]]for[[:space:]]Instance-dependent[[:space:]]Partial-label[[:space:]]Learning/f8eeec3e-6ffd-4ea2-86e4-2004ab3ed26c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Canonical[[:space:]]Factors[[:space:]]for[[:space:]]Hybrid[[:space:]]Neural[[:space:]]Fields/38785228-e9ec-45cb-85b7-ae96458a97da_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cascade-DETR_[[:space:]]Delving[[:space:]]into[[:space:]]High-Quality[[:space:]]Universal[[:space:]]Object[[:space:]]Detection/214af37a-b88c-4182-bb37-1ba9ede94aa7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Category-aware[[:space:]]Allocation[[:space:]]Transformer[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Object[[:space:]]Localization/d44dc0b1-f330-4e42-a8e8-77464be427a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CauSSL_[[:space:]]Causality-inspired[[:space:]]Semi-supervised[[:space:]]Learning[[:space:]]for[[:space:]]Medical[[:space:]]Image[[:space:]]Segmentation/76e7f5c0-0185-4535-97e3-6a8f385527cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Causal-DFQ_[[:space:]]Causality[[:space:]]Guided[[:space:]]Data-Free[[:space:]]Network[[:space:]]Quantization/bb17cacd-24fd-4c92-9846-a1acbc8e2134_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Center-Based[[:space:]]Decoupled[[:space:]]Point-cloud[[:space:]]Registration[[:space:]]for[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/cec9232d-701f-4e1e-8d85-55c51c64043f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Chaotic[[:space:]]World_[[:space:]]A[[:space:]]Large[[:space:]]and[[:space:]]Challenging[[:space:]]Benchmark[[:space:]]for[[:space:]]Human[[:space:]]Behavior[[:space:]]Understanding[[:space:]]in[[:space:]]Chaotic[[:space:]]Events/ad886da5-fee6-4e29-b145-b13d9861bb30_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ChartReader_[[:space:]]A[[:space:]]Unified[[:space:]]Framework[[:space:]]for[[:space:]]Chart[[:space:]]Derendering[[:space:]]and[[:space:]]Comprehension[[:space:]]without[[:space:]]Heuristic[[:space:]]Rules/e953209c-68eb-4462-9203-80cd3f44bba7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Chasing[[:space:]]Clouds_[[:space:]]Differentiable[[:space:]]Volumetric[[:space:]]Rasterisation[[:space:]]of[[:space:]]Point[[:space:]]Clouds[[:space:]]as[[:space:]]a[[:space:]]Highly[[:space:]]Efficient[[:space:]]and[[:space:]]Accurate[[:space:]]Loss[[:space:]]for[[:space:]]Large-Scale[[:space:]]Deformable[[:space:]]3D[[:space:]]Registration/c3d735f7-93f5-45d2-9047-ba0bf2c1a18c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CheckerPose_[[:space:]]Progressive[[:space:]]Dense[[:space:]]Keypoint[[:space:]]Localization[[:space:]]for[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]Graph[[:space:]]Neural[[:space:]]Network/bb072382-ea5d-4ea7-a764-2551a782d103_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ChildPlay_[[:space:]]A[[:space:]]New[[:space:]]Benchmark[[:space:]]for[[:space:]]Understanding[[:space:]]Children's[[:space:]]Gaze[[:space:]]Behaviour/5a323503-db88-4ce0-8a79-8509a9789c1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Chinese[[:space:]]Text[[:space:]]Recognition[[:space:]]with[[:space:]]A[[:space:]]Pre-Trained[[:space:]]CLIP-Like[[:space:]]Model[[:space:]]Through[[:space:]]Image-IDS[[:space:]]Aligning/648c4d14-cf6d-487d-9816-3638602336cb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Chop[[:space:]]&[[:space:]]Learn_[[:space:]]Recognizing[[:space:]]and[[:space:]]Generating[[:space:]]Object-State[[:space:]]Compositions/5239a163-2db2-47b2-9d01-04a2c49ef63a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Chordal[[:space:]]Averaging[[:space:]]on[[:space:]]Flag[[:space:]]Manifolds[[:space:]]and[[:space:]]Its[[:space:]]Applications/ff29854a-7e8b-49d1-ad00-e5f3e8b238bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Chupa_[[:space:]]Carving[[:space:]]3D[[:space:]]Clothed[[:space:]]Humans[[:space:]]from[[:space:]]Skinned[[:space:]]Shape[[:space:]]Priors[[:space:]]using[[:space:]]2D[[:space:]]Diffusion[[:space:]]Probabilistic[[:space:]]Models/faf27ff6-4cfe-49ce-b906-8cdf4bd3e203_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CiT_[[:space:]]Curation[[:space:]]in[[:space:]]Training[[:space:]]for[[:space:]]Effective[[:space:]]Vision-Language[[:space:]]Data/c6af1561-6737-4a38-8b4f-cd3d060a860a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CiteTracker_[[:space:]]Correlating[[:space:]]Image[[:space:]]and[[:space:]]Text[[:space:]]for[[:space:]]Visual[[:space:]]Tracking/c073587e-4f69-49db-804a-4983da8ea0d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class[[:space:]]Prior-Free[[:space:]]Positive-Unlabeled[[:space:]]Learning[[:space:]]with[[:space:]]Taylor[[:space:]]Variational[[:space:]]Loss[[:space:]]for[[:space:]]Hyperspectral[[:space:]]Remote[[:space:]]Sensing[[:space:]]Imagery/022667cb-8f02-4f8f-ad67-bacfb1a22ab6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class-Aware[[:space:]]Patch[[:space:]]Embedding[[:space:]]Adaptation[[:space:]]for[[:space:]]Few-Shot[[:space:]]Image[[:space:]]Classification/21179139-4508-4f4b-8ae4-46b728cb3343_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class-Incremental[[:space:]]Grouping[[:space:]]Network[[:space:]]for[[:space:]]Continual[[:space:]]Audio-Visual[[:space:]]Learning/6ccc4ada-c701-4dfe-9929-68c951495f07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class-incremental[[:space:]]Continual[[:space:]]Learning[[:space:]]for[[:space:]]Instance[[:space:]]Segmentation[[:space:]]with[[:space:]]Image-level[[:space:]]Weak[[:space:]]Supervision/2d912be0-30c0-43de-a08a-3d476e83c690_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Class-relation[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Novel[[:space:]]Class[[:space:]]Discovery/0b1ee8e9-a629-47b2-ad37-c4cadb65f0e3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CleanCLIP_[[:space:]]Mitigating[[:space:]]Data[[:space:]]Poisoning[[:space:]]Attacks[[:space:]]in[[:space:]]Multimodal[[:space:]]Contrastive[[:space:]]Learning/7a6d38cc-688b-48ba-9909-f69116a71ab9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ClimateNeRF_[[:space:]]Extreme[[:space:]]Weather[[:space:]]Synthesis[[:space:]]in[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field/aa5383cc-5ee1-41d0-a054-504eebbc5444_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cloth2Body_[[:space:]]Generating[[:space:]]3D[[:space:]]Human[[:space:]]Body[[:space:]]Mesh[[:space:]]from[[:space:]]2D[[:space:]]Clothing/0cee0b26-21d8-469b-8d57-0afd9f08cc07_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ClothPose_[[:space:]]A[[:space:]]Real-world[[:space:]]Benchmark[[:space:]]for[[:space:]]Visual[[:space:]]Analysis[[:space:]]of[[:space:]]Garment[[:space:]]Pose[[:space:]]via[[:space:]]An[[:space:]]Indirect[[:space:]]Recording[[:space:]]Solution/923e2f28-0d69-4cf7-8bcb-844c3645a8d9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ClothesNet_[[:space:]]An[[:space:]]Information-Rich[[:space:]]3D[[:space:]]Garment[[:space:]]Model[[:space:]]Repository[[:space:]]with[[:space:]]Simulated[[:space:]]Clothes[[:space:]]Environment/6e920fba-46cc-41e6-8277-2f780062343a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ClusT3_[[:space:]]Information[[:space:]]Invariant[[:space:]]Test-Time[[:space:]]Training/77bdb6dc-0ad6-40c6-959c-d04d2a3f6673_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Clusterformer_[[:space:]]Cluster-based[[:space:]]Transformer[[:space:]]for[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]in[[:space:]]Point[[:space:]]Clouds/3eb22086-9c9a-4d93-a6db-49d9415816b6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Clustering[[:space:]]based[[:space:]]Point[[:space:]]Cloud[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]3D[[:space:]]Analysis/f29e6307-e0bf-4c7b-9231-f52ca80a7dbe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Clutter[[:space:]]Detection[[:space:]]and[[:space:]]Removal[[:space:]]in[[:space:]]3D[[:space:]]Scenes[[:space:]]with[[:space:]]View-Consistent[[:space:]]Inpainting/a074d7fe-1b6f-4d7f-a7f8-01c00f833842_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Co-Evolution[[:space:]]of[[:space:]]Pose[[:space:]]and[[:space:]]Mesh[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Body[[:space:]]Estimation[[:space:]]from[[:space:]]Video/a5ef7844-b861-400b-8aa5-063489d8433e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CoIn_[[:space:]]Contrastive[[:space:]]Instance[[:space:]]Feature[[:space:]]Mining[[:space:]]for[[:space:]]Outdoor[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Very[[:space:]]Limited[[:space:]]Annotations/6dbd554f-b890-4527-91f6-f5c6788de158_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CoSign_[[:space:]]Exploring[[:space:]]Co-occurrence[[:space:]]Signals[[:space:]]in[[:space:]]Skeleton-based[[:space:]]Continuous[[:space:]]Sign[[:space:]]Language[[:space:]]Recognition/5b69e8a4-036e-4758-ba9d-aec108932ec8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CoTDet_[[:space:]]Affordance[[:space:]]Knowledge[[:space:]]Prompting[[:space:]]for[[:space:]]Task[[:space:]]Driven[[:space:]]Object[[:space:]]Detection/2e007673-bf51-4f3e-ae7b-517a8ad42500_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Coarse-to-Fine[[:space:]]Amodal[[:space:]]Segmentation[[:space:]]with[[:space:]]Shape[[:space:]]Prior/45375f31-b292-4467-9613-2e060bc6fd1c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Coarse-to-Fine_[[:space:]]Learning[[:space:]]Compact[[:space:]]Discriminative[[:space:]]Representation[[:space:]]for[[:space:]]Single-Stage[[:space:]]Image[[:space:]]Retrieval/14072a0e-8d6c-4878-a39b-9fc0c570ab74_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Coherent[[:space:]]Event[[:space:]]Guided[[:space:]]Low-Light[[:space:]]Video[[:space:]]Enhancement/20259ac6-d8f8-4616-a927-d32b0b13ae4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CoinSeg_[[:space:]]Contrast[[:space:]]Inter-[[:space:]]and[[:space:]]Intra-[[:space:]]Class[[:space:]]Representations[[:space:]]for[[:space:]]Incremental[[:space:]]Segmentation/5dac26d8-98d5-43ff-8044-1a7ca12076d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Collaborative[[:space:]]Propagation[[:space:]]on[[:space:]]Multiple[[:space:]]Instance[[:space:]]Graphs[[:space:]]for[[:space:]]3D[[:space:]]Instance[[:space:]]Segmentation[[:space:]]with[[:space:]]Single-point[[:space:]]Supervision/c0373650-91e6-4526-b454-e7be2fe6910a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Collaborative[[:space:]]Tracking[[:space:]]Learning[[:space:]]for[[:space:]]Frame-Rate-Insensitive[[:space:]]Multi-Object[[:space:]]Tracking/5c5cdc3f-b266-40ab-a800-7a8d0cb71166_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Collecting[[:space:]]The[[:space:]]Puzzle[[:space:]]Pieces_[[:space:]]Disentangled[[:space:]]Self-Driven[[:space:]]Human[[:space:]]Pose[[:space:]]Transfer[[:space:]]by[[:space:]]Permuting[[:space:]]Textures/7edb343d-b394-45f3-b920-6225cdc95a91_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Combating[[:space:]]Noisy[[:space:]]Labels[[:space:]]with[[:space:]]Sample[[:space:]]Selection[[:space:]]by[[:space:]]Mining[[:space:]]High-Discrepancy[[:space:]]Examples/99021fb7-ecf1-4b0f-820a-874c3dca1271_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Communication-Efficient[[:space:]]Vertical[[:space:]]Federated[[:space:]]Learning[[:space:]]with[[:space:]]Limited[[:space:]]Overlapping[[:space:]]Samples/990a4877-caab-4ef0-81e2-7ab0ce498ba6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Communication-efficient[[:space:]]Federated[[:space:]]Learning[[:space:]]with[[:space:]]Single-Step[[:space:]]Synthetic[[:space:]]Features[[:space:]]Compressor[[:space:]]for[[:space:]]Faster[[:space:]]Convergence/f0ca2f90-8d22-4ee4-a7b3-db7bd78d6ee4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Compatibility[[:space:]]of[[:space:]]Fundamental[[:space:]]Matrices[[:space:]]for[[:space:]]Complete[[:space:]]Viewing[[:space:]]Graphs/67975232-cda7-4384-a760-2b3dcd8567c1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Complementary[[:space:]]Domain[[:space:]]Adaptation[[:space:]]and[[:space:]]Generalization[[:space:]]for[[:space:]]Unsupervised[[:space:]]Continual[[:space:]]Domain[[:space:]]Shift[[:space:]]Learning/29b50774-602f-4b07-be37-0b4ce4e1138e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Compositional[[:space:]]Feature[[:space:]]Augmentation[[:space:]]for[[:space:]]Unbiased[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/9cb0a206-d3d7-4ab0-8299-854180a867a6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Computation[[:space:]]and[[:space:]]Data[[:space:]]Efficient[[:space:]]Backdoor[[:space:]]Attacks/794ff758-6ad2-4335-85f2-eb98aefa7fe2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Computational[[:space:]]3D[[:space:]]Imaging[[:space:]]with[[:space:]]Position[[:space:]]Sensors/58f362ca-0a42-4bbe-a2f1-c10aa8f94b27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Computationally-Efficient[[:space:]]Neural[[:space:]]Image[[:space:]]Compression[[:space:]]with[[:space:]]Shallow[[:space:]]Decoders/14826d0c-1667-4faa-8da5-f42d8faceffd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ConSlide_[[:space:]]Asynchronous[[:space:]]Hierarchical[[:space:]]Interaction[[:space:]]Transformer[[:space:]]with[[:space:]]Breakup-Reorganize[[:space:]]Rehearsal[[:space:]]for[[:space:]]Continual[[:space:]]Whole[[:space:]]Slide[[:space:]]Image[[:space:]]Analysis/7f7a4f25-adf3-4850-854d-a4623407bfb8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Concept-wise[[:space:]]Fine-tuning[[:space:]]Matters[[:space:]]in[[:space:]]Preventing[[:space:]]Negative[[:space:]]Transfer/b1c92b3a-c292-4387-af22-5a3afb0aeeee_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Conceptual[[:space:]]and[[:space:]]Hierarchical[[:space:]]Latent[[:space:]]Space[[:space:]]Decomposition[[:space:]]for[[:space:]]Face[[:space:]]Editing/d297c96e-99dd-4f0f-814d-57d536caf4e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Conditional[[:space:]]360-degree[[:space:]]Image[[:space:]]Synthesis[[:space:]]for[[:space:]]Immersive[[:space:]]Indoor[[:space:]]Scene[[:space:]]Decoration/d6a3c293-f12a-43aa-9b2d-fef12bb5b81b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Conditional[[:space:]]Cross[[:space:]]Attention[[:space:]]Network[[:space:]]for[[:space:]]Multi-Space[[:space:]]Embedding[[:space:]]without[[:space:]]Entanglement[[:space:]]in[[:space:]]Only[[:space:]]a[[:space:]]SINGLE[[:space:]]Network/05ab3d54-d369-4d9a-a07d-de60db03b349_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Confidence-aware[[:space:]]Pseudo-label[[:space:]]Learning[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Visual[[:space:]]Grounding/e13bd2fc-1b65-4c22-b34a-dc4c5e21b4a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Confidence-based[[:space:]]Visual[[:space:]]Dispersal[[:space:]]for[[:space:]]Few-shot[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/69a96626-c549-4487-bb43-31ba07caa563_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Consistent[[:space:]]Depth[[:space:]]Prediction[[:space:]]for[[:space:]]Transparent[[:space:]]Object[[:space:]]Reconstruction[[:space:]]from[[:space:]]RGB-D[[:space:]]Camera/b00582dd-7be3-47fd-aba4-42cc9a6242e5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Constraining[[:space:]]Depth[[:space:]]Map[[:space:]]Geometry[[:space:]]for[[:space:]]Multi-View[[:space:]]Stereo_[[:space:]]A[[:space:]]Dual-Depth[[:space:]]Approach[[:space:]]with[[:space:]]Saddle-shaped[[:space:]]Depth[[:space:]]Cells/76cfee17-8860-4ccc-ba7a-867c4c3bb8ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ContactGen_[[:space:]]Generative[[:space:]]Contact[[:space:]]Modeling[[:space:]]for[[:space:]]Grasp[[:space:]]Generation/b5347dd1-de8e-4cd3-83fe-4ffd25800e6f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Contactless[[:space:]]Pulse[[:space:]]Estimation[[:space:]]Leveraging[[:space:]]Pseudo[[:space:]]Labels[[:space:]]and[[:space:]]Self-Supervision/b347baa5-3e3b-4ff0-a8ca-1ff8868fc815_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Content-Aware[[:space:]]Local[[:space:]]GAN[[:space:]]for[[:space:]]Photo-Realistic[[:space:]]Super-Resolution/ec6bd9d2-2fee-457d-a3a3-ce401b4c9b55_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Context-Aware[[:space:]]Planning[[:space:]]and[[:space:]]Environment-Aware[[:space:]]Memory[[:space:]]for[[:space:]]Instruction[[:space:]]Following[[:space:]]Embodied[[:space:]]Agents/fec8491d-5511-4b8c-9131-2e62c03005ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Continual[[:space:]]Learning[[:space:]]for[[:space:]]Personalized[[:space:]]Co-speech[[:space:]]Gesture[[:space:]]Generation/7551a9f0-5533-4f1c-85d9-cff679d1aec9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Continual[[:space:]]Segment_[[:space:]]Towards[[:space:]]a[[:space:]]Single,[[:space:]]Unified[[:space:]]and[[:space:]]Non-forgetting[[:space:]]Continual[[:space:]]Segmentation[[:space:]]Model[[:space:]]of[[:space:]]143[[:space:]]Whole-body[[:space:]]Organs[[:space:]]in[[:space:]]CT[[:space:]]Scans/04a1c0ab-44a1-4c98-b6c4-3ca36e6978fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Continual[[:space:]]Zero-Shot[[:space:]]Learning[[:space:]]through[[:space:]]Semantically[[:space:]]Guided[[:space:]]Generative[[:space:]]Random[[:space:]]Walks/d0330bfc-f604-4dce-9b28-aab0e34aed38_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Continuously[[:space:]]Masked[[:space:]]Transformer[[:space:]]for[[:space:]]Image[[:space:]]Inpainting/c81b444c-83d0-4dbd-9576-8debef4b6a12_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Contrastive[[:space:]]Continuity[[:space:]]on[[:space:]]Augmentation[[:space:]]Stability[[:space:]]Rehearsal[[:space:]]for[[:space:]]Continual[[:space:]]Self-Supervised[[:space:]]Learning/94d34726-bc38-4f5d-825a-48168f3ea91a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Contrastive[[:space:]]Feature[[:space:]]Masking[[:space:]]Open-Vocabulary[[:space:]]Vision[[:space:]]Transformer/c6308eca-4bbf-4596-9fe1-23cc17facef2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Contrastive[[:space:]]Learning[[:space:]]Relies[[:space:]]More[[:space:]]on[[:space:]]Spatial[[:space:]]Inductive[[:space:]]Bias[[:space:]]Than[[:space:]]Supervised[[:space:]]Learning_[[:space:]]An[[:space:]]Empirical[[:space:]]Study/6a8f24be-46e6-4ba0-990b-97f4877ad3a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Contrastive[[:space:]]Model[[:space:]]Adaptation[[:space:]]for[[:space:]]Cross-Condition[[:space:]]Robustness[[:space:]]in[[:space:]]Semantic[[:space:]]Segmentation/5cb09725-cf90-4e29-927d-f6862c07cd5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Contrastive[[:space:]]Pseudo[[:space:]]Learning[[:space:]]for[[:space:]]Open-World[[:space:]]DeepFake[[:space:]]Attribution/0a6637f9-960a-45a2-9a06-58fec4c516b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Controllable[[:space:]]Guide-Space[[:space:]]for[[:space:]]Generalizable[[:space:]]Face[[:space:]]Forgery[[:space:]]Detection/921df931-3080-46a8-94c2-fd24997c9ffd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Controllable[[:space:]]Person[[:space:]]Image[[:space:]]Synthesis[[:space:]]with[[:space:]]Pose-Constrained[[:space:]]Latent[[:space:]]Diffusion/a92f7997-b241-47a0-974b-afa8881af7d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Controllable[[:space:]]Visual-Tactile[[:space:]]Synthesis/21420d03-b598-40e1-b713-1aa925ae2622_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Convex[[:space:]]Decomposition[[:space:]]of[[:space:]]Indoor[[:space:]]Scenes/0053bd3c-9d62-47e2-831c-d7cd54eb014d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Convolutional[[:space:]]Networks[[:space:]]with[[:space:]]Oriented[[:space:]]1D[[:space:]]Kernels/0c0ea477-d0a3-412f-a344-8bbc8f5bad82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Coordinate[[:space:]]Quantized[[:space:]]Neural[[:space:]]Implicit[[:space:]]Representations[[:space:]]for[[:space:]]Multi-view[[:space:]]Reconstruction/fd97227c-99f8-4706-ac31-91c22a9677e6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Coordinate[[:space:]]Transformer_[[:space:]]Achieving[[:space:]]Single-stage[[:space:]]Multi-person[[:space:]]Mesh[[:space:]]Recovery[[:space:]]from[[:space:]]Videos/09b6fb35-348f-46b5-a54f-da10d411f2c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CopyRNeRF_[[:space:]]Protecting[[:space:]]the[[:space:]]CopyRight[[:space:]]of[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/1b3d470e-8a07-4c7b-bd4a-53fae576b401_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Corrupting[[:space:]]Neuron[[:space:]]Explanations[[:space:]]of[[:space:]]Deep[[:space:]]Visual[[:space:]]Features/f817d1b9-5e8a-4ef1-b1a4-1f8994841143_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Counterfactual-based[[:space:]]Saliency[[:space:]]Map_[[:space:]]Towards[[:space:]]Visual[[:space:]]Contrastive[[:space:]]Explanations[[:space:]]for[[:space:]]Neural[[:space:]]Networks/ea1a7144-5002-419a-b976-34d3997c0893_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Counting[[:space:]]Crowds[[:space:]]in[[:space:]]Bad[[:space:]]Weather/81aaf155-f0ac-4eac-8b06-182baa32ca64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Creative[[:space:]]Birds_[[:space:]]Self-Supervised[[:space:]]Single-View[[:space:]]3D[[:space:]]Style[[:space:]]Transfer/3ebebeb3-ce0f-4e8f-803e-1ad4965b0ab3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CroCo[[:space:]]v2_[[:space:]]Improved[[:space:]]Cross-view[[:space:]]Completion[[:space:]]Pre-training[[:space:]]for[[:space:]]Stereo[[:space:]]Matching[[:space:]]and[[:space:]]Optical[[:space:]]Flow/411fa479-84b1-496e-8069-d97e46eea6b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross[[:space:]]Contrasting[[:space:]]Feature[[:space:]]Perturbation[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/1ade16e5-f3bd-4d39-8760-681206ffe1af_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross[[:space:]]Modal[[:space:]]Transformer_[[:space:]]Towards[[:space:]]Fast[[:space:]]and[[:space:]]Robust[[:space:]]3D[[:space:]]Object[[:space:]]Detection/a53ac45c-5a75-434b-9216-73f3717432f2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-Domain[[:space:]]Product[[:space:]]Representation[[:space:]]Learning[[:space:]]for[[:space:]]Rich-Content[[:space:]]E-Commerce/39ccf97c-d631-4b15-8fbb-6a53f093e54d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-Modal[[:space:]]Learning[[:space:]]with[[:space:]]3D[[:space:]]Deformable[[:space:]]Attention[[:space:]]for[[:space:]]Action[[:space:]]Recognition/42366ad6-da74-4315-b6f1-7a0f0da470a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-Modal[[:space:]]Orthogonal[[:space:]]High-Rank[[:space:]]Augmentation[[:space:]]for[[:space:]]RGB-Event[[:space:]]Transformer-Trackers/d5819b8b-12f6-4094-905a-6051e9687020_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-Modal[[:space:]]Translation[[:space:]]and[[:space:]]Alignment[[:space:]]for[[:space:]]Survival[[:space:]]Analysis/63d968a8-4600-4e01-898c-ba28ec01c3f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-Ray[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Novel-View[[:space:]]Synthesis[[:space:]]from[[:space:]]Unconstrained[[:space:]]Image[[:space:]]Collections/f215f158-43ee-4461-a24c-58f4b535f92f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-modal[[:space:]]Latent[[:space:]]Space[[:space:]]Alignment[[:space:]]for[[:space:]]Image[[:space:]]to[[:space:]]Avatar[[:space:]]Translation/8c43d47a-d520-4978-af55-7613e3ad852e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-modal[[:space:]]Scalable[[:space:]]Hyperbolic[[:space:]]Hierarchical[[:space:]]Clustering/aa8bc579-06d0-4870-babc-4460b8f45b66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-view[[:space:]]Semantic[[:space:]]Alignment[[:space:]]for[[:space:]]Livestreaming[[:space:]]Product[[:space:]]Recognition/6f6e1b77-59ab-4e51-bbf8-8fb612a34d03_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cross-view[[:space:]]Topology[[:space:]]Based[[:space:]]Consistent[[:space:]]and[[:space:]]Complementary[[:space:]]Information[[:space:]]for[[:space:]]Deep[[:space:]]Multi-view[[:space:]]Clustering/e404bf62-855d-4fb3-91db-60a7906875d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CrossLoc3D_[[:space:]]Aerial-Ground[[:space:]]Cross-Source[[:space:]]3D[[:space:]]Place[[:space:]]Recognition/53a6a222-1b17-458b-aa69-56e3d9a8e7bf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CrossMatch_[[:space:]]Source-Free[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]via[[:space:]]Cross-Modal[[:space:]]Consistency[[:space:]]Training/8998197b-b903-4195-a640-929a0e00c25b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/CuNeRF_[[:space:]]Cube-Based[[:space:]]Neural[[:space:]]Radiance[[:space:]]Field[[:space:]]for[[:space:]]Zero-Shot[[:space:]]Medical[[:space:]]Image[[:space:]]Arbitrary-Scale[[:space:]]Super[[:space:]]Resolution/d1ce0a63-9a82-4b38-8db8-eeccdbc79aca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cumulative[[:space:]]Spatial[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Vision[[:space:]]Transformers/84ec7df8-44a5-4724-b793-27c0bd087b57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Curvature-Aware[[:space:]]Training[[:space:]]for[[:space:]]Coordinate[[:space:]]Networks/48c1e39d-07e2-46bd-911c-89ea3d4fde26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cyclic[[:space:]]Test-Time[[:space:]]Adaptation[[:space:]]on[[:space:]]Monocular[[:space:]]Video[[:space:]]for[[:space:]]3D[[:space:]]Human[[:space:]]Mesh[[:space:]]Reconstruction/3222780b-fb7b-4fa3-adcc-d7dc12f9ac49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Cyclic-Bootstrap[[:space:]]Labeling[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Object[[:space:]]Detection/fd68102f-7c37-43a6-9999-e52c987e2934_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/D-IF_[[:space:]]Uncertainty-aware[[:space:]]Human[[:space:]]Digitization[[:space:]]via[[:space:]]Implicit[[:space:]]Distribution[[:space:]]Field/76a9ed73-ff53-4b89-bfb3-05adfd81b540_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/D3G_[[:space:]]Exploring[[:space:]]Gaussian[[:space:]]Prior[[:space:]]for[[:space:]]Temporal[[:space:]]Sentence[[:space:]]Grounding[[:space:]]with[[:space:]]Glance[[:space:]]Annotation/d2a40a0a-11a9-46f4-8114-4e69e0511656_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DALL-Eval_[[:space:]]Probing[[:space:]]the[[:space:]]Reasoning[[:space:]]Skills[[:space:]]and[[:space:]]Social[[:space:]]Biases[[:space:]]of[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]Models/1bb1cafc-db9a-4f8b-a414-ddf77c1bb800_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DARTH_[[:space:]]Holistic[[:space:]]Test-time[[:space:]]Adaptation[[:space:]]for[[:space:]]Multiple[[:space:]]Object[[:space:]]Tracking/2ae11728-b7c9-4c9f-8dcc-e75dbf8783c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DCPB_[[:space:]]Deformable[[:space:]]Convolution[[:space:]]Based[[:space:]]on[[:space:]]the[[:space:]]Poincare[[:space:]]Ball[[:space:]]for[[:space:]]Top-view[[:space:]]Fisheye[[:space:]]Cameras/d6216288-3c9e-4a4f-9a8b-4504a9fefa5e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DDColor_[[:space:]]Towards[[:space:]]Photo-Realistic[[:space:]]Image[[:space:]]Colorization[[:space:]]via[[:space:]]Dual[[:space:]]Decoders/12b53504-b584-4686-b0fc-89fbae3c933d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DDFM_[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Multi-Modality[[:space:]]Image[[:space:]]Fusion/8121062d-d344-464f-8ed9-3f39d512551e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DDG-Net_[[:space:]]Discriminability-Driven[[:space:]]Graph[[:space:]]Network[[:space:]]for[[:space:]]Weakly-supervised[[:space:]]Temporal[[:space:]]Action[[:space:]]Localization/10d518fc-211c-4652-b85d-bbbfe4a87bbd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DDIT_[[:space:]]Semantic[[:space:]]Scene[[:space:]]Completion[[:space:]]via[[:space:]]Deformable[[:space:]]Deep[[:space:]]Implicit[[:space:]]Templates/49506061-d949-4817-a21e-947bc7203755_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DDP_[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Dense[[:space:]]Visual[[:space:]]Prediction/7b0982cd-c7f6-48b6-b05e-1ccdfded4c70_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DDS2M_[[:space:]]Self-Supervised[[:space:]]Denoising[[:space:]]Diffusion[[:space:]]Spatio-Spectral[[:space:]]Model[[:space:]]for[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Restoration/a1d78241-550a-43fd-837a-a5c4ebb23964_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DECO_[[:space:]]Dense[[:space:]]Estimation[[:space:]]of[[:space:]]3D[[:space:]]Human-Scene[[:space:]]Contact[[:space:]]In[[:space:]]The[[:space:]]Wild/1835df8c-f954-405a-a5bb-a803801c8e31_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DEDRIFT_[[:space:]]Robust[[:space:]]Similarity[[:space:]]Search[[:space:]]under[[:space:]]Content[[:space:]]Drift/ad441f39-d483-4ffa-972a-8ad570791015_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DELFlow_[[:space:]]Dense[[:space:]]Efficient[[:space:]]Learning[[:space:]]of[[:space:]]Scene[[:space:]]Flow[[:space:]]for[[:space:]]Large-Scale[[:space:]]Point[[:space:]]Clouds/2dc41127-ad73-453a-8db0-46698b18a167_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DETA_[[:space:]]Denoised[[:space:]]Task[[:space:]]Adaptation[[:space:]]for[[:space:]]Few-Shot[[:space:]]Learning/a88fb5e5-f336-4545-9b6a-58907fa0b98b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DETR[[:space:]]Does[[:space:]]Not[[:space:]]Need[[:space:]]Multi-Scale[[:space:]]or[[:space:]]Locality[[:space:]]Design/8e3a5579-70a8-4ed6-a53d-22ff1cbe8e04_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DETRDistill_[[:space:]]A[[:space:]]Universal[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]Framework[[:space:]]for[[:space:]]DETR-families/d217f12d-e17d-4c97-9ec0-e1c1907c807a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DETRs[[:space:]]with[[:space:]]Collaborative[[:space:]]Hybrid[[:space:]]Assignments[[:space:]]Training/5dd8b567-95ed-4053-a422-0d2b02e106be_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DFA3D_[[:space:]]3D[[:space:]]Deformable[[:space:]]Attention[[:space:]]For[[:space:]]2D-to-3D[[:space:]]Feature[[:space:]]Lifting/6480b49a-0dc4-4859-bce2-b8812d494919_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DG-Recon_[[:space:]]Depth-Guided[[:space:]]Neural[[:space:]]3D[[:space:]]Scene[[:space:]]Reconstruction/df6da1a5-f433-4b01-bf08-6caea574736c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DG3D_[[:space:]]Generating[[:space:]]High[[:space:]]Quality[[:space:]]3D[[:space:]]Textured[[:space:]]Shapes[[:space:]]by[[:space:]]Learning[[:space:]]to[[:space:]]Discriminate[[:space:]]Multi-Modal[[:space:]]Diffusion-Renderings/521e639e-a1a6-4cba-8f9e-08456b5cf99d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DIFFGUARD_[[:space:]]Semantic[[:space:]]Mismatch-Guided[[:space:]]Out-of-Distribution[[:space:]]Detection[[:space:]]Using[[:space:]]Pre-Trained[[:space:]]Diffusion[[:space:]]Models/bf6a457f-f7cf-42b1-8ff1-48c9dfd1b7fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DIME-FM[[:space:]]_[[:space:]]DIstilling[[:space:]]Multimodal[[:space:]]and[[:space:]]Efficient[[:space:]]Foundation[[:space:]]Models/5342dd6d-e192-439f-859e-0071e750790c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DINAR_[[:space:]]Diffusion[[:space:]]Inpainting[[:space:]]of[[:space:]]Neural[[:space:]]Textures[[:space:]]for[[:space:]]One-Shot[[:space:]]Human[[:space:]]Avatars/557d2f15-98cf-4a13-80d4-2e6e62d9ca72_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DIRE[[:space:]]for[[:space:]]Diffusion-Generated[[:space:]]Image[[:space:]]Detection/1a2f1570-7ac4-4d96-8564-283fea75de52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DISeR_[[:space:]]Designing[[:space:]]Imaging[[:space:]]Systems[[:space:]]with[[:space:]]Reinforcement[[:space:]]Learning/e9ec8ef6-b929-498f-bd85-3023d7ae59e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DLGSANet_[[:space:]]Lightweight[[:space:]]Dynamic[[:space:]]Local[[:space:]]and[[:space:]]Global[[:space:]]Self-Attention[[:space:]]Networks[[:space:]]for[[:space:]]Image[[:space:]]Super-Resolution/d5872372-635f-430e-9e16-7cb33e6f8a37_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DLT_[[:space:]]Conditioned[[:space:]]layout[[:space:]]generation[[:space:]]with[[:space:]]Joint[[:space:]]Discrete-Continuous[[:space:]]Diffusion[[:space:]]Layout[[:space:]]Transformer/7cf07a5a-ddaa-4193-b8e3-9f2de5fa0017_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DMNet_[[:space:]]Delaunay[[:space:]]Meshing[[:space:]]Network[[:space:]]for[[:space:]]3D[[:space:]]Shape[[:space:]]Representation/32318fde-15f1-45d7-b425-1b1ab4745819_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DNA-Rendering_[[:space:]]A[[:space:]]Diverse[[:space:]]Neural[[:space:]]Actor[[:space:]]Repository[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Human-Centric[[:space:]]Rendering/6d2bbce4-ccc3-4c12-af5e-7b3b690bf021_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DOLCE_[[:space:]]A[[:space:]]Model-Based[[:space:]]Probabilistic[[:space:]]Diffusion[[:space:]]Framework[[:space:]]for[[:space:]]Limited-Angle[[:space:]]CT[[:space:]]Reconstruction/61eea2b6-a804-4229-9530-cc2f942778c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DOT_[[:space:]]A[[:space:]]Distillation-Oriented[[:space:]]Trainer/e7f4c96b-0222-4a9d-84d1-5e293ca2f570_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DPF-Net_[[:space:]]Combining[[:space:]]Explicit[[:space:]]Shape[[:space:]]Priors[[:space:]]in[[:space:]]Deformable[[:space:]]Primitive[[:space:]]Field[[:space:]]for[[:space:]]Unsupervised[[:space:]]Structural[[:space:]]Reconstruction[[:space:]]of[[:space:]]3D[[:space:]]Objects/b312178a-2442-4462-aafa-349ba20bf028_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DPM-OT_[[:space:]]A[[:space:]]New[[:space:]]Diffusion[[:space:]]Probabilistic[[:space:]]Model[[:space:]]Based[[:space:]]on[[:space:]]Optimal[[:space:]]Transport/784cecd7-e449-4a79-9ee7-8c3549dca67e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DPS-Net_[[:space:]]Deep[[:space:]]Polarimetric[[:space:]]Stereo[[:space:]]Depth[[:space:]]Estimation/2eae3868-fe7a-4398-887a-0055d0afa7c4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DQS3D_[[:space:]]Densely-matched[[:space:]]Quantization-aware[[:space:]]Semi-supervised[[:space:]]3D[[:space:]]Detection/56766d45-1e9e-4a4c-82f6-8a2e9a393b62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DR-Tune_[[:space:]]Improving[[:space:]]Fine-tuning[[:space:]]of[[:space:]]Pretrained[[:space:]]Visual[[:space:]]Models[[:space:]]by[[:space:]]Distribution[[:space:]]Regularization[[:space:]]with[[:space:]]Semantic[[:space:]]Calibration/2d7877bf-29dc-486a-a144-9e975e81bf27_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DRAW_[[:space:]]Defending[[:space:]]Camera-shooted[[:space:]]RAW[[:space:]]Against[[:space:]]Image[[:space:]]Manipulation/6b28d1a1-feed-44de-a984-ae3bff7e330c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DREAMWALKER_[[:space:]]Mental[[:space:]]Planning[[:space:]]for[[:space:]]Continuous[[:space:]]Vision-Language[[:space:]]Navigation/8bbb670f-0bce-43a4-85f8-ddca53805196_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DREAM_[[:space:]]Efficient[[:space:]]Dataset[[:space:]]Distillation[[:space:]]by[[:space:]]Representative[[:space:]]Matching/d9d791ed-76ee-47b3-b9b5-ec3c1b34210d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DReg-NeRF_[[:space:]]Deep[[:space:]]Registration[[:space:]]for[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/a2c31615-7066-442c-834e-8ab0938ae4b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DS-Fusion_[[:space:]]Artistic[[:space:]]Typography[[:space:]]via[[:space:]]Discriminated[[:space:]]and[[:space:]]Stylized[[:space:]]Diffusion/e42d312c-2577-41e1-ace7-f0029337b548_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DVGaze_[[:space:]]Dual-View[[:space:]]Gaze[[:space:]]Estimation/a414f000-f7e3-4326-bae9-08404594094e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DVIS_[[:space:]]Decoupled[[:space:]]Video[[:space:]]Instance[[:space:]]Segmentation[[:space:]]Framework/ee4c3f2b-f93f-4aac-9aa7-4d92848c125c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dancing[[:space:]]in[[:space:]]the[[:space:]]Dark_[[:space:]]A[[:space:]]Benchmark[[:space:]]towards[[:space:]]General[[:space:]]Low-light[[:space:]]Video[[:space:]]Enhancement/4c245c43-d3f3-4231-91b5-5fca4e3e0a32_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DandelionNet_[[:space:]]Domain[[:space:]]Composition[[:space:]]with[[:space:]]Instance[[:space:]]Adaptive[[:space:]]Classification[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/e3719c93-6b71-4e95-9537-10044397dd56_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DarSwin_[[:space:]]Distortion[[:space:]]Aware[[:space:]]Radial[[:space:]]Swin[[:space:]]Transformer/befdbfbc-7eed-4d6e-9c3e-57ea61f4ee28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dark[[:space:]]Side[[:space:]]Augmentation_[[:space:]]Generating[[:space:]]Diverse[[:space:]]Night[[:space:]]Examples[[:space:]]for[[:space:]]Metric[[:space:]]Learning/bb17f838-14f5-402d-bd34-e004258ee763_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Data[[:space:]]Augmented[[:space:]]Flatness-aware[[:space:]]Gradient[[:space:]]Projection[[:space:]]for[[:space:]]Continual[[:space:]]Learning/7149c2be-da80-45c8-9e57-bec9e59ad6f8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Data-Free[[:space:]]Class-Incremental[[:space:]]Hand[[:space:]]Gesture[[:space:]]Recognition/26137a77-66a1-4147-8a55-4ed0e87f06b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Data-free[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Fine-grained[[:space:]]Visual[[:space:]]Categorization/22b33a5c-9fd1-4497-9550-d039990d17ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DataDAM_[[:space:]]Efficient[[:space:]]Dataset[[:space:]]Distillation[[:space:]]with[[:space:]]Attention[[:space:]]Matching/0d82a377-a410-4540-b3bf-bc06d4e8a3df_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dataset[[:space:]]Quantization/f6b179cb-3687-4020-8462-2de788621c25_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeFormer_[[:space:]]Integrating[[:space:]]Transformers[[:space:]]with[[:space:]]Deformable[[:space:]]Models[[:space:]]for[[:space:]]3D[[:space:]]Shape[[:space:]]Abstraction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]Image/d3f6cd90-b4a7-44d9-b8a4-492fcd58dd5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeLiRa_[[:space:]]Self-Supervised[[:space:]]Depth,[[:space:]]Light,[[:space:]]and[[:space:]]Radiance[[:space:]]Fields/8021c922-a874-42ce-82a8-af8c4ebf0829_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dec-Adapter_[[:space:]]Exploring[[:space:]]Efficient[[:space:]]Decoder-Side[[:space:]]Adapter[[:space:]]for[[:space:]]Bridging[[:space:]]Screen[[:space:]]Content[[:space:]]and[[:space:]]Natural[[:space:]]Image[[:space:]]Compression/e17b2b08-29d4-4348-af46-c5b9dc30dc16_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decomposition-Based[[:space:]]Variational[[:space:]]Network[[:space:]]for[[:space:]]Multi-Contrast[[:space:]]MRI[[:space:]]Super-Resolution[[:space:]]and[[:space:]]Reconstruction/5ceb0081-a6ac-4ada-987e-7fb8090c2901_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decouple[[:space:]]Before[[:space:]]Interact_[[:space:]]Multi-Modal[[:space:]]Prompt[[:space:]]Learning[[:space:]]for[[:space:]]Continual[[:space:]]Visual[[:space:]]Question[[:space:]]Answering/0268b6e6-2000-49ff-a323-667d6632ab5b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decoupled[[:space:]]DETR_[[:space:]]Spatially[[:space:]]Disentangling[[:space:]]Localization[[:space:]]and[[:space:]]Classification[[:space:]]for[[:space:]]Improved[[:space:]]End-to-End[[:space:]]Object[[:space:]]Detection/15c43a21-bd59-464a-830a-0aaac42e5072_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Decoupled[[:space:]]Iterative[[:space:]]Refinement[[:space:]]Framework[[:space:]]for[[:space:]]Interacting[[:space:]]Hands[[:space:]]Reconstruction[[:space:]]from[[:space:]]a[[:space:]]Single[[:space:]]RGB[[:space:]]Image/be8a4262-c88d-4989-94c6-74325bf7d8a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeePoint_[[:space:]]Visual[[:space:]]Pointing[[:space:]]Recognition[[:space:]]and[[:space:]]Direction[[:space:]]Estimation/976ec122-95c1-4961-a261-c06cc576e960_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Active[[:space:]]Contours[[:space:]]for[[:space:]]Real-time[[:space:]]6-DoF[[:space:]]Object[[:space:]]Tracking/eca71987-cbbe-4656-a4f2-626e1ef283d0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Directly-Trained[[:space:]]Spiking[[:space:]]Neural[[:space:]]Networks[[:space:]]for[[:space:]]Object[[:space:]]Detection/01ec45e5-1ca7-4efd-9bdb-5525929d9893_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Equilibrium[[:space:]]Object[[:space:]]Detection/6a18cdf5-2b98-46fd-af6a-4b16d6189820_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Feature[[:space:]]Deblurring[[:space:]]Diffusion[[:space:]]for[[:space:]]Detecting[[:space:]]Out-of-Distribution[[:space:]]Objects/483a526e-dacc-4525-9092-f0b1849d773f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Fusion[[:space:]]Transformer[[:space:]]Network[[:space:]]with[[:space:]]Weighted[[:space:]]Vector-Wise[[:space:]]Keypoints[[:space:]]Voting[[:space:]]for[[:space:]]Robust[[:space:]]6D[[:space:]]Object[[:space:]]Pose[[:space:]]Estimation/5a7a9b92-08a1-455c-8a30-c155f14937d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Geometrized[[:space:]]Cartoon[[:space:]]Line[[:space:]]Inbetweening/6eba6358-ca07-4c5a-865d-d80873db4a28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Geometry-Aware[[:space:]]Camera[[:space:]]Self-Calibration[[:space:]]from[[:space:]]Video/9c4c4da8-db17-4bf5-abe7-bfb157f031b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Homography[[:space:]]Mixture[[:space:]]for[[:space:]]Single[[:space:]]Image[[:space:]]Rolling[[:space:]]Shutter[[:space:]]Correction/7236dd6e-d353-47d7-a792-384edd8a1cde_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Image[[:space:]]Harmonization[[:space:]]with[[:space:]]Globally[[:space:]]Guided[[:space:]]Feature[[:space:]]Transformation[[:space:]]and[[:space:]]Relation[[:space:]]Distillation/3575305f-76c9-41e0-b6f4-45837f2442fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Image[[:space:]]Harmonization[[:space:]]with[[:space:]]Learnable[[:space:]]Augmentation/dd6fe8ee-6307-4e3b-8a2b-a6897500b81b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Incubation_[[:space:]]Training[[:space:]]Large[[:space:]]Models[[:space:]]by[[:space:]]Divide-and-Conquering/7b60b154-fdfc-44f6-b7b0-5298b4aef187_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Multitask[[:space:]]Learning[[:space:]]with[[:space:]]Progressive[[:space:]]Parameter[[:space:]]Sharing/ab4fb7c7-9e23-4f19-971d-ef8b04c06065_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Multiview[[:space:]]Clustering[[:space:]]by[[:space:]]Contrasting[[:space:]]Cluster[[:space:]]Assignments/b1c8fb11-22b2-4dad-835e-ef60b2591108_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Optics[[:space:]]for[[:space:]]Video[[:space:]]Snapshot[[:space:]]Compressive[[:space:]]Imaging/142701f8-3ea8-437f-8bcb-31d675487c5e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deep[[:space:]]Video[[:space:]]Demoireing[[:space:]]via[[:space:]]Compact[[:space:]]Invertible[[:space:]]Dyadic[[:space:]]Decomposition/3807a777-554a-44fe-936d-1b47417b89d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeepChange_[[:space:]]A[[:space:]]Long-Term[[:space:]]Person[[:space:]]Re-Identification[[:space:]]Benchmark[[:space:]]with[[:space:]]Clothes[[:space:]]Change/30d3d505-dc54-4db0-9f77-58a2d447756e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DeformToon3D_[[:space:]]Deformable[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]3D[[:space:]]Toonification/2c77b009-01ce-4921-8e67-d791bf632cff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deformable[[:space:]]Model-Driven[[:space:]]Neural[[:space:]]Rendering[[:space:]]for[[:space:]]High-Fidelity[[:space:]]3D[[:space:]]Reconstruction[[:space:]]of[[:space:]]Human[[:space:]]Heads[[:space:]]Under[[:space:]]Low-View[[:space:]]Settings/ce2c3864-4009-4928-b5db-7ab2bb75a6e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deformable[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]using[[:space:]]RGB[[:space:]]and[[:space:]]Event[[:space:]]Cameras/2e368450-dc52-4021-a223-1741cc001f3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Deformer_[[:space:]]Dynamic[[:space:]]Fusion[[:space:]]Transformer[[:space:]]for[[:space:]]Robust[[:space:]]Hand[[:space:]]Pose[[:space:]]Estimation/08e57ea8-a97e-4118-852f-eecf34822068_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Degradation-Resistant[[:space:]]Unfolding[[:space:]]Network[[:space:]]for[[:space:]]Heterogeneous[[:space:]]Image[[:space:]]Fusion/c52f0049-569f-4ec2-9356-e0b5eb3c3b94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Delicate[[:space:]]Textured[[:space:]]Mesh[[:space:]]Recovery[[:space:]]from[[:space:]]NeRF[[:space:]]via[[:space:]]Adaptive[[:space:]]Surface[[:space:]]Refinement/940adf16-1d1b-4c43-95db-9838d3ac7178_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Delta[[:space:]]Denoising[[:space:]]Score/ebd0842e-8c98-4a4c-8c20-797889a0d850_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Delving[[:space:]]into[[:space:]]Motion-Aware[[:space:]]Matching[[:space:]]for[[:space:]]Monocular[[:space:]]3D[[:space:]]Object[[:space:]]Tracking/97273422-bff1-4334-95bc-8135e93cfccc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Democratising[[:space:]]2D[[:space:]]Sketch[[:space:]]to[[:space:]]3D[[:space:]]Shape[[:space:]]Retrieval[[:space:]]Through[[:space:]]Pivoting/719a7398-5b71-4343-9bd1-1c1d5a300d51_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Denoising[[:space:]]Diffusion[[:space:]]Autoencoders[[:space:]]are[[:space:]]Unified[[:space:]]Self-supervised[[:space:]]Learners/6041a9cd-2248-4e0b-b6ef-ccc2e844d503_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dense[[:space:]]2D-3D[[:space:]]Indoor[[:space:]]Prediction[[:space:]]with[[:space:]]Sound[[:space:]]via[[:space:]]Aligned[[:space:]]Cross-Modal[[:space:]]Distillation/665b654b-dd30-4b41-adab-99e965815165_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dense[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]with[[:space:]]Attention[[:space:]]Modulation/3efabcc3-6658-4d9b-89e5-48cf1db03e20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DenseShift_[[:space:]]Towards[[:space:]]Accurate[[:space:]]and[[:space:]]Efficient[[:space:]]Low-Bit[[:space:]]Power-of-Two[[:space:]]Quantization/6ea7ce49-2949-46b8-ac79-511fe5e724a3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Density-invariant[[:space:]]Features[[:space:]]for[[:space:]]Distant[[:space:]]Point[[:space:]]Cloud[[:space:]]Registration/d2de355a-0537-4948-b119-219cffc3cdb2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Designing[[:space:]]Phase[[:space:]]Masks[[:space:]]for[[:space:]]Under-Display[[:space:]]Cameras/8fd738c3-2282-4ec3-9c35-eb60f054b4fa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DetZero_[[:space:]]Rethinking[[:space:]]Offboard[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Long-term[[:space:]]Sequential[[:space:]]Point[[:space:]]Clouds/17fe57b8-7e19-43e4-a583-e64c2435dac6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Detecting[[:space:]]Objects[[:space:]]with[[:space:]]Context-Likelihood[[:space:]]Graphs[[:space:]]and[[:space:]]Graph[[:space:]]Refinement/0ba43fa8-4974-4499-a673-f2b6ec1490c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Detection[[:space:]]Transformer[[:space:]]with[[:space:]]Stable[[:space:]]Matching/f976bc2a-2bbf-44c0-9f3f-99aba9358d69_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DetermiNet_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Diagnostic[[:space:]]Dataset[[:space:]]for[[:space:]]Complex[[:space:]]Visually-Grounded[[:space:]]Referencing[[:space:]]using[[:space:]]Determiners/d2a41c07-2e50-4ffb-acc6-b38331f63ed7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiFaReli_[[:space:]]Diffusion[[:space:]]Face[[:space:]]Relighting/48011f93-9db6-4862-aacf-3b441f012c9b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiLiGenT-Pi_[[:space:]]Photometric[[:space:]]Stereo[[:space:]]for[[:space:]]Planar[[:space:]]Surfaces[[:space:]]with[[:space:]]Rich[[:space:]]Details[[:space:]]-[[:space:]]Benchmark[[:space:]]Dataset[[:space:]]and[[:space:]]Beyond/7100390e-d321-44da-aceb-52b29cbe0d12_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diff-Retinex_[[:space:]]Rethinking[[:space:]]Low-light[[:space:]]Image[[:space:]]Enhancement[[:space:]]with[[:space:]]A[[:space:]]Generative[[:space:]]Diffusion[[:space:]]Model/d2ed4208-2a47-4b34-ba95-4b492f0f89e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffCloth_[[:space:]]Diffusion[[:space:]]Based[[:space:]]Garment[[:space:]]Synthesis[[:space:]]and[[:space:]]Manipulation[[:space:]]via[[:space:]]Structural[[:space:]]Cross-modal[[:space:]]Semantic[[:space:]]Alignment/c487e753-950f-46b5-9f60-ac379b73bd33_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffDis_[[:space:]]Empowering[[:space:]]Generative[[:space:]]Diffusion[[:space:]]Model[[:space:]]with[[:space:]]Cross-Modal[[:space:]]Discrimination[[:space:]]Capability/51e38303-85c3-4188-81cf-88d73575a3bc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffDreamer_[[:space:]]Towards[[:space:]]Consistent[[:space:]]Unsupervised[[:space:]]Single-view[[:space:]]Scene[[:space:]]Extrapolation[[:space:]]with[[:space:]]Conditional[[:space:]]Diffusion[[:space:]]Models/585f95ef-7358-4daf-9486-32d2bfccdf28_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffFacto_[[:space:]]Controllable[[:space:]]Part-Based[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Generation[[:space:]]with[[:space:]]Cross[[:space:]]Diffusion/fed19c0c-6dcd-4d5c-b40b-dd3c3fccb109_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffFit_[[:space:]]Unlocking[[:space:]]Transferability[[:space:]]of[[:space:]]Large[[:space:]]Diffusion[[:space:]]Models[[:space:]]via[[:space:]]Simple[[:space:]]Parameter-efficient[[:space:]]Fine-Tuning/9ae0b6f3-5110-425e-aa9d-73ccaccec7f1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffIR_[[:space:]]Efficient[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Image[[:space:]]Restoration/bf43aa3d-4a15-46ce-9b3a-75e7b2488c87_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffPose_[[:space:]]Multi-hypothesis[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]using[[:space:]]Diffusion[[:space:]]Models/75925583-0181-4fcc-bee7-a9469a6dbed5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffPose_[[:space:]]SpatioTemporal[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Video-Based[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation/8d03d5d0-c622-4ca5-be97-31c025fba31e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffRate[[:space:]]_[[:space:]]Differentiable[[:space:]]Compression[[:space:]]Rate[[:space:]]for[[:space:]]Efficient[[:space:]]Vision[[:space:]]Transformers/450cb997-27a1-4109-acb9-faab40b963b7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffTAD_[[:space:]]Temporal[[:space:]]Action[[:space:]]Detection[[:space:]]with[[:space:]]Proposal[[:space:]]Denoising[[:space:]]Diffusion/fe2064f5-f955-4f16-9910-5c1f7e9695ae_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffV2S_[[:space:]]Diffusion-Based[[:space:]]Video-to-Speech[[:space:]]Synthesis[[:space:]]with[[:space:]]Vision-Guided[[:space:]]Speaker[[:space:]]Embedding/4f2df68a-7ec7-47e4-80d8-ccc65b7242f5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Differentiable[[:space:]]Transportation[[:space:]]Pruning/d11d74b9-6ef0-43ba-ba8a-9287afe133c8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffuMask_[[:space:]]Synthesizing[[:space:]]Images[[:space:]]with[[:space:]]Pixel-level[[:space:]]Annotations[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]Using[[:space:]]Diffusion[[:space:]]Models/d672ae46-ca48-4057-ba44-751f0f62c3a4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffuse3D_[[:space:]]Wide-Angle[[:space:]]3D[[:space:]]Photography[[:space:]]via[[:space:]]Bilateral[[:space:]]Diffusion/433137cb-be6c-4ce8-bcbe-f63ab5faf759_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion[[:space:]]Action[[:space:]]Segmentation/35c62501-965a-4d31-8309-2cf26e116eaa_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion[[:space:]]Model[[:space:]]as[[:space:]]Representation[[:space:]]Learner/38009ffd-c3b4-49ea-86b1-f317959ad8c0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion[[:space:]]Models[[:space:]]as[[:space:]]Masked[[:space:]]Autoencoders/c61d1821-1a7e-4122-8a41-ad3606a05b49_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion[[:space:]]in[[:space:]]Style/5e7eccda-caab-40ae-8966-857e1c59ae52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion-Based[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]Estimation[[:space:]]with[[:space:]]Multi-Hypothesis[[:space:]]Aggregation/647b91ac-431e-4849-9d95-ade60750d29a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion-Guided[[:space:]]Reconstruction[[:space:]]of[[:space:]]Everyday[[:space:]]Hand-Object[[:space:]]Interaction[[:space:]]Clips/6c19d36b-74bd-4770-b340-3fd3d2aa4d57_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion-SDF_[[:space:]]Conditional[[:space:]]Generative[[:space:]]Modeling[[:space:]]of[[:space:]]Signed[[:space:]]Distance[[:space:]]Functions/25283204-3afe-44e7-a616-9d93f00c5d3a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diffusion-based[[:space:]]Image[[:space:]]Translation[[:space:]]with[[:space:]]Label[[:space:]]Guidance[[:space:]]for[[:space:]]Domain[[:space:]]Adaptive[[:space:]]Semantic[[:space:]]Segmentation/ecacf5b5-bb10-4a27-9f51-19bdeec88118_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffusionDet_[[:space:]]Diffusion[[:space:]]Model[[:space:]]for[[:space:]]Object[[:space:]]Detection/c2cce1c3-20d4-4ddb-b5c9-29e3115ba5d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DiffusionRet_[[:space:]]Generative[[:space:]]Text-Video[[:space:]]Retrieval[[:space:]]with[[:space:]]Diffusion[[:space:]]Model/5bc640fe-6aab-4c61-bd50-46c7c8fe1016_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Discovering[[:space:]]Spatio-Temporal[[:space:]]Rationales[[:space:]]for[[:space:]]Video[[:space:]]Question[[:space:]]Answering/a8d44b81-cacd-42d7-a6a9-20c34b3b4afc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Discrepant[[:space:]]and[[:space:]]Multi-Instance[[:space:]]Proxies[[:space:]]for[[:space:]]Unsupervised[[:space:]]Person[[:space:]]Re-Identification/fe4e2da5-a08d-4e59-ac51-68d7d4cdf509_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Discriminative[[:space:]]Class[[:space:]]Tokens[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/f734d089-e80b-4290-9b7d-a57257e143ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Disentangle[[:space:]]then[[:space:]]Parse_[[:space:]]Night-time[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Illumination[[:space:]]Disentanglement/5c46b762-465a-495f-8b35-c45e3416038a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Disentangling[[:space:]]Spatial[[:space:]]and[[:space:]]Temporal[[:space:]]Learning[[:space:]]for[[:space:]]Efficient[[:space:]]Image-to-Video[[:space:]]Transfer[[:space:]]Learning/33e1b2b9-29dc-4722-8c13-ff8e7d985e0e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Disposable[[:space:]]Transfer[[:space:]]Learning[[:space:]]for[[:space:]]Selective[[:space:]]Source[[:space:]]Task[[:space:]]Unlearning/0846d1a5-75ee-4eff-ba45-3e065df9c576_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DistillBEV_[[:space:]]Boosting[[:space:]]Multi-Camera[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Cross-Modal[[:space:]]Knowledge[[:space:]]Distillation/a8973dcf-1513-4a6e-b18c-2bcbbe2d7a09_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distilled[[:space:]]Reverse[[:space:]]Attention[[:space:]]Network[[:space:]]for[[:space:]]Open-world[[:space:]]Compositional[[:space:]]Zero-Shot[[:space:]]Learning/45660a07-42f6-4099-bde3-32fed7a256ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distilling[[:space:]]Coarse-to-Fine[[:space:]]Semantic[[:space:]]Matching[[:space:]]Knowledge[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]3D[[:space:]]Visual[[:space:]]Grounding/1c211649-eb10-42b4-9147-40d84622b06e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distilling[[:space:]]DETR[[:space:]]with[[:space:]]Visual-Linguistic[[:space:]]Knowledge[[:space:]]for[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detection/18c17822-017e-4669-aa75-76d0b23b82c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distilling[[:space:]]Large[[:space:]]Vision-Language[[:space:]]Model[[:space:]]with[[:space:]]Out-of-Distribution[[:space:]]Generalizability/17151a04-37f6-4a1e-916e-9ddcea4439e0_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distilling[[:space:]]from[[:space:]]Similar[[:space:]]Tasks[[:space:]]for[[:space:]]Transfer[[:space:]]Learning[[:space:]]on[[:space:]]a[[:space:]]Budget/4803e5f1-e812-4435-8d4e-4c4fd362ed8c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distracting[[:space:]]Downpour_[[:space:]]Adversarial[[:space:]]Weather[[:space:]]Attacks[[:space:]]for[[:space:]]Motion[[:space:]]Estimation/1923605b-44a7-40e0-a466-8459d82ed92f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distributed[[:space:]]Bundle[[:space:]]Adjustment[[:space:]]with[[:space:]]Block-Based[[:space:]]Sparse[[:space:]]Matrix[[:space:]]Compression[[:space:]]for[[:space:]]Super[[:space:]]Large[[:space:]]Scale[[:space:]]Datasets/53795a7c-1110-4607-9d28-9efb85a08e29_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distribution[[:space:]]Shift[[:space:]]Matters[[:space:]]for[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]with[[:space:]]Webly[[:space:]]Collected[[:space:]]Images/96484d37-9a15-425e-8de7-57bec0d1751f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distribution-Aligned[[:space:]]Diffusion[[:space:]]for[[:space:]]Human[[:space:]]Mesh[[:space:]]Recovery/5a181a6f-194d-4165-960c-ba51a0598eb9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distribution-Aware[[:space:]]Prompt[[:space:]]Tuning[[:space:]]for[[:space:]]Vision-Language[[:space:]]Models/ae8dc403-b840-4f07-a507-933dc6568b82_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Distribution-Consistent[[:space:]]Modal[[:space:]]Recovering[[:space:]]for[[:space:]]Incomplete[[:space:]]Multimodal[[:space:]]Learning/667e9c7c-6836-479f-8993-9ebd4790ead9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diverse[[:space:]]Cotraining[[:space:]]Makes[[:space:]]Strong[[:space:]]Semi-Supervised[[:space:]]Segmentor/9071a1fa-151e-42c4-918d-adf21adcee66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diverse[[:space:]]Data[[:space:]]Augmentation[[:space:]]with[[:space:]]Diffusions[[:space:]]for[[:space:]]Effective[[:space:]]Test-time[[:space:]]Prompt[[:space:]]Tuning/182fdf57-a043-4a89-83a6-8b571702a70a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Diverse[[:space:]]Inpainting[[:space:]]and[[:space:]]Editing[[:space:]]with[[:space:]]GAN[[:space:]]Inversion/47036b5c-536d-4fce-b132-3d10dbb8269d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Divide[[:space:]]and[[:space:]]Conquer_[[:space:]]3D[[:space:]]Point[[:space:]]Cloud[[:space:]]Instance[[:space:]]Segmentation[[:space:]]With[[:space:]]Point-Wise[[:space:]]Binarization/8ccf557d-d5aa-4094-97fa-344dd214e48c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Divide[[:space:]]and[[:space:]]Conquer_[[:space:]]a[[:space:]]Two-Step[[:space:]]Method[[:space:]]for[[:space:]]High[[:space:]]Quality[[:space:]]Face[[:space:]]De-identification[[:space:]]with[[:space:]]Model[[:space:]]Explainability/569e3e0a-f60f-4ca7-b1f0-a8fdbd04f240_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Divide&Classify_[[:space:]]Fine-Grained[[:space:]]Classification[[:space:]]for[[:space:]]City-Wide[[:space:]]Visual[[:space:]]Geo-Localization/a5153619-6984-41bc-b294-4e967cf0b4ea_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Do[[:space:]]DALL-E[[:space:]]and[[:space:]]Flamingo[[:space:]]Understand[[:space:]]Each[[:space:]]Other_/f01e2256-671a-49eb-8582-625956d9e534_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DocTr_[[:space:]]Document[[:space:]]Transformer[[:space:]]for[[:space:]]Structured[[:space:]]Information[[:space:]]Extraction[[:space:]]in[[:space:]]Documents/8e47910e-d084-4191-b10b-4384fbebe1cc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Document[[:space:]]Understanding[[:space:]]Dataset[[:space:]]and[[:space:]]Evaluation[[:space:]](DUDE)/21ea7825-3529-4791-a3b5-88339563370d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Does[[:space:]]Physical[[:space:]]Adversarial[[:space:]]Example[[:space:]]Really[[:space:]]Matter[[:space:]]to[[:space:]]Autonomous[[:space:]]Driving_[[:space:]]Towards[[:space:]]System-Level[[:space:]]Effect[[:space:]]of[[:space:]]Adversarial[[:space:]]Object[[:space:]]Evasion[[:space:]]Attack/122f0a93-4edc-47f7-9e3f-d5cdc275647b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Domain[[:space:]]Adaptive[[:space:]]Few-Shot[[:space:]]Open-Set[[:space:]]Learning/5babb275-3e26-4251-83dd-a7be9b3324f7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Domain[[:space:]]Generalization[[:space:]]Guided[[:space:]]by[[:space:]]Gradient[[:space:]]Signal[[:space:]]to[[:space:]]Noise[[:space:]]Ratio[[:space:]]of[[:space:]]Parameters/dc968604-751a-45fb-819f-31c0bb63556b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Domain[[:space:]]Generalization[[:space:]]of[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]in[[:space:]]Autonomous[[:space:]]Driving/bb93f3ce-24dc-49d5-a338-c66a467c47a8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Domain[[:space:]]Generalization[[:space:]]via[[:space:]]Balancing[[:space:]]Training[[:space:]]Difficulty[[:space:]]and[[:space:]]Model[[:space:]]Capability/ccf97693-dee8-4a95-a718-235f2bc6a8a2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Domain[[:space:]]Generalization[[:space:]]via[[:space:]]Rationale[[:space:]]Invariance/ae49a822-8a67-4aa9-bf63-e455ad4004bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Domain[[:space:]]Specified[[:space:]]Optimization[[:space:]]for[[:space:]]Deployment[[:space:]]Authorization/3f30e18f-20ee-42e9-afb1-2b866b571946_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Domain-Specificity[[:space:]]Inducing[[:space:]]Transformers[[:space:]]for[[:space:]]Source-Free[[:space:]]Domain[[:space:]]Adaptation/0814fce5-e49b-4318-8b01-8e7f40c7d9d5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DomainAdaptor_[[:space:]]A[[:space:]]Novel[[:space:]]Approach[[:space:]]to[[:space:]]Test-time[[:space:]]Adaptation/fb56c215-5203-473e-8bff-3af20ecca890_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DomainDrop_[[:space:]]Suppressing[[:space:]]Domain-Sensitive[[:space:]]Channels[[:space:]]for[[:space:]]Domain[[:space:]]Generalization/21072815-7ee9-4df5-a941-e4d46477f9ba_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Doppelgangers_[[:space:]]Learning[[:space:]]to[[:space:]]Disambiguate[[:space:]]Images[[:space:]]of[[:space:]]Similar[[:space:]]Structures/fa41ffd6-8792-4712-966d-1ae4a3f88eca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Downscaled[[:space:]]Representation[[:space:]]Matters_[[:space:]]Improving[[:space:]]Image[[:space:]]Rescaling[[:space:]]with[[:space:]]Collaborative[[:space:]]Downscaled[[:space:]]Images/89813c80-5300-4c7d-bfff-383c563829b1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Downstream-agnostic[[:space:]]Adversarial[[:space:]]Examples/dd208220-99e8-4399-af60-6040dfdbb5db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DreamBooth3D_[[:space:]]Subject-Driven[[:space:]]Text-to-3D[[:space:]]Generation/ac713f40-5437-428c-b63e-4187e5d1becd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DreamPose_[[:space:]]Fashion[[:space:]]Video[[:space:]]Synthesis[[:space:]]with[[:space:]]Stable[[:space:]]Diffusion/15c5ba56-c8f2-4afe-8b0b-a1793d73d703_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DreamTeacher_[[:space:]]Pretraining[[:space:]]Image[[:space:]]Backbones[[:space:]]with[[:space:]]Deep[[:space:]]Generative[[:space:]]Models/3d4d5f82-c171-42bf-8e76-77ee7a8ead3f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DriveAdapter_[[:space:]]Breaking[[:space:]]the[[:space:]]Coupling[[:space:]]Barrier[[:space:]]of[[:space:]]Perception[[:space:]]and[[:space:]]Planning[[:space:]]in[[:space:]]End-to-End[[:space:]]Autonomous[[:space:]]Driving/fd3b0b94-16bf-4636-bef2-864ce97eb494_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dual[[:space:]]Aggregation[[:space:]]Transformer[[:space:]]for[[:space:]]Image[[:space:]]Super-Resolution/fcd844da-50a2-4a03-87eb-eecb5ecb7ef3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dual[[:space:]]Learning[[:space:]]with[[:space:]]Dynamic[[:space:]]Knowledge[[:space:]]Distillation[[:space:]]for[[:space:]]Partially[[:space:]]Relevant[[:space:]]Video[[:space:]]Retrieval/cc500bfb-9bd3-475a-8ba6-dc07e8e29f20_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dual[[:space:]]Meta-Learning[[:space:]]with[[:space:]]Longitudinally[[:space:]]Consistent[[:space:]]Regularization[[:space:]]for[[:space:]]One-Shot[[:space:]]Brain[[:space:]]Tissue[[:space:]]Segmentation[[:space:]]Across[[:space:]]the[[:space:]]Human[[:space:]]Lifespan/a2207fb4-e6aa-45da-97e2-2c6d02284ca7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dual[[:space:]]Pseudo-Labels[[:space:]]Interactive[[:space:]]Self-Training[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Visible-Infrared[[:space:]]Person[[:space:]]Re-Identification/762517c1-0d79-4e9e-b2a3-f4fbc72c550d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DyGait_[[:space:]]Exploiting[[:space:]]Dynamic[[:space:]]Representations[[:space:]]for[[:space:]]High-performance[[:space:]]Gait[[:space:]]Recognition/75558175-6ee4-4ec8-89cb-d61a40d2a2d2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DynaMITe_[[:space:]]Dynamic[[:space:]]Query[[:space:]]Bootstrapping[[:space:]]for[[:space:]]Multi-object[[:space:]]Interactive[[:space:]]Segmentation[[:space:]]Transformer/ef657f09-8043-4f9a-94a8-6c1b428d773c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Hyperbolic[[:space:]]Attention[[:space:]]Network[[:space:]]for[[:space:]]Fine[[:space:]]Hand-object[[:space:]]Reconstruction/6285bf43-6f94-4b08-b9b9-9dd232839221_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Mesh[[:space:]]Recovery[[:space:]]from[[:space:]]Partial[[:space:]]Point[[:space:]]Cloud[[:space:]]Sequence/b2694161-2730-41d5-8834-34ee3e8c1e66_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Mesh-Aware[[:space:]]Radiance[[:space:]]Fields/77a36c1c-e471-46bd-ba5a-c123b19ac20a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Perceiver[[:space:]]for[[:space:]]Efficient[[:space:]]Visual[[:space:]]Recognition/ce0526d3-88c9-4320-8eb0-a3470015f17b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]PlenOctree[[:space:]]for[[:space:]]Adaptive[[:space:]]Sampling[[:space:]]Refinement[[:space:]]in[[:space:]]Explicit[[:space:]]NeRF/6388bb58-6f1c-4e3d-89e6-0de8c2bd6d21_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Point[[:space:]]Fields/cf7ef324-9049-4a87-ae73-9f0b0101d650_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Residual[[:space:]]Classifier[[:space:]]for[[:space:]]Class[[:space:]]Incremental[[:space:]]Learning/62c82b6f-3751-4191-956d-b01ec127eb95_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Snake[[:space:]]Convolution[[:space:]]Based[[:space:]]on[[:space:]]Topological[[:space:]]Geometric[[:space:]]Constraints[[:space:]]for[[:space:]]Tubular[[:space:]]Structure[[:space:]]Segmentation/39d3dacb-060c-445f-a5d8-ccfa9642e941_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Dynamic[[:space:]]Token[[:space:]]Pruning[[:space:]]in[[:space:]]Plain[[:space:]]Vision[[:space:]]Transformers[[:space:]]for[[:space:]]Semantic[[:space:]]Segmentation/cc6d5c20-c5a9-420b-a93e-d1f86af4ebce_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/DynamicISP_[[:space:]]Dynamically[[:space:]]Controlled[[:space:]]Image[[:space:]]Signal[[:space:]]Processor[[:space:]]for[[:space:]]Image[[:space:]]Recognition/77a1f8e0-3ef6-4738-9cd5-2f4d4bdb0f5e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/E2E-LOAD_[[:space:]]End-to-End[[:space:]]Long-form[[:space:]]Online[[:space:]]Action[[:space:]]Detection/b1ac04a9-1cf1-47e9-a2d0-868cb2e9481e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/E2NeRF_[[:space:]]Event[[:space:]]Enhanced[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]from[[:space:]]Blurry[[:space:]]Images/f2381af6-ec4e-4fea-8edc-c217daab7a26_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/E3Sym_[[:space:]]Leveraging[[:space:]]E(3)[[:space:]]Invariance[[:space:]]for[[:space:]]Unsupervised[[:space:]]3D[[:space:]]Planar[[:space:]]Reflective[[:space:]]Symmetry[[:space:]]Detection/f5ba6edb-7c1e-4dbe-bba8-819d6eb99a94_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EDAPS_[[:space:]]Enhanced[[:space:]]Domain-Adaptive[[:space:]]Panoptic[[:space:]]Segmentation/50dd3c26-7c88-4b1c-abc1-a01935daa056_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EGC_[[:space:]]Image[[:space:]]Generation[[:space:]]and[[:space:]]Classification[[:space:]]via[[:space:]]a[[:space:]]Diffusion[[:space:]]Energy-Based[[:space:]]Model/eff10ce0-dc06-4ca0-8a7b-125b0c62378d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EGformer_[[:space:]]Equirectangular[[:space:]]Geometry-biased[[:space:]]Transformer[[:space:]]for[[:space:]]360[[:space:]]Depth[[:space:]]Estimation/5eb5fabf-36db-458c-8aec-541a92716091_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ELFNet_[[:space:]]Evidential[[:space:]]Local-global[[:space:]]Fusion[[:space:]]for[[:space:]]Stereo[[:space:]]Matching/7dbf1570-f70a-4e18-9c5e-8fdbe30035fd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ELITE_[[:space:]]Encoding[[:space:]]Visual[[:space:]]Concepts[[:space:]]into[[:space:]]Textual[[:space:]]Embeddings[[:space:]]for[[:space:]]Customized[[:space:]]Text-to-Image[[:space:]]Generation/6d4936ec-284d-4988-990c-e8a205fcb444_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EMDB_[[:space:]]The[[:space:]]Electromagnetic[[:space:]]Database[[:space:]]of[[:space:]]Global[[:space:]]3D[[:space:]]Human[[:space:]]Pose[[:space:]]and[[:space:]]Shape[[:space:]]in[[:space:]]the[[:space:]]Wild/1de4a602-ac24-48cc-b37d-49364bff4eef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EMMN_[[:space:]]Emotional[[:space:]]Motion[[:space:]]Memory[[:space:]]Network[[:space:]]for[[:space:]]Audio-driven[[:space:]]Emotional[[:space:]]Talking[[:space:]]Face[[:space:]]Generation/8fcf34d8-3f56-4b91-8654-29be49782197_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EMQ_[[:space:]]Evolving[[:space:]]Training-free[[:space:]]Proxies[[:space:]]for[[:space:]]Automated[[:space:]]Mixed[[:space:]]Precision[[:space:]]Quantization/6d8e651e-b434-4cde-b9e0-d85eeac9d152_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EMR-MSF_[[:space:]]Self-Supervised[[:space:]]Recurrent[[:space:]]Monocular[[:space:]]Scene[[:space:]]Flow[[:space:]]Exploiting[[:space:]]Ego-Motion[[:space:]]Rigidity/1fbe639c-4f1f-47f3-93d8-3cbecc9220d8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ENTL_[[:space:]]Embodied[[:space:]]Navigation[[:space:]]Trajectory[[:space:]]Learner/8f3edb8b-6baf-419a-ab56-721ab39ff56c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ENVIDR_[[:space:]]Implicit[[:space:]]Differentiable[[:space:]]Renderer[[:space:]]with[[:space:]]Neural[[:space:]]Environment[[:space:]]Lighting/44666fb0-422c-4250-a651-9295cd197941_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EP2P-Loc_[[:space:]]End-to-End[[:space:]]3D[[:space:]]Point[[:space:]]to[[:space:]]2D[[:space:]]Pixel[[:space:]]Localization[[:space:]]for[[:space:]]Large-Scale[[:space:]]Visual[[:space:]]Localization/f1194f48-fd00-462d-bbde-ffc5d94185ef_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EPiC_[[:space:]]Ensemble[[:space:]]of[[:space:]]Partial[[:space:]]Point[[:space:]]Clouds[[:space:]]for[[:space:]]Robust[[:space:]]Classification/06252bba-d1f8-46e1-8cf9-6a51590c8682_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EQ-Net_[[:space:]]Elastic[[:space:]]Quantization[[:space:]]Neural[[:space:]]Networks/9cd99c2e-eeb7-4d9a-b651-78e770185402_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ESSAformer_[[:space:]]Efficient[[:space:]]Transformer[[:space:]]for[[:space:]]Hyperspectral[[:space:]]Image[[:space:]]Super-resolution/d5e85607-179f-4363-9d03-56a1cfa245fe_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ESTextSpotter_[[:space:]]Towards[[:space:]]Better[[:space:]]Scene[[:space:]]Text[[:space:]]Spotting[[:space:]]with[[:space:]]Explicit[[:space:]]Synergy[[:space:]]in[[:space:]]Transformer/297320b5-19cf-411e-86ff-c415297f1027_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ETran_[[:space:]]Energy-Based[[:space:]]Transferability[[:space:]]Estimation/fcfaec75-e8dc-4b30-ba9d-96142d30f570_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/E^2VPT_[[:space:]]An[[:space:]]Effective[[:space:]]and[[:space:]]Efficient[[:space:]]Approach[[:space:]]for[[:space:]]Visual[[:space:]]Prompt[[:space:]]Tuning/4ba4da7f-33be-4ec4-8e69-143a3be16741_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EdaDet_[[:space:]]Open-Vocabulary[[:space:]]Object[[:space:]]Detection[[:space:]]Using[[:space:]]Early[[:space:]]Dense[[:space:]]Alignment/5215c942-b5e0-4203-8c3a-c499487355bd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Editable[[:space:]]Image[[:space:]]Geometric[[:space:]]Abstraction[[:space:]]via[[:space:]]Neural[[:space:]]Primitive[[:space:]]Assembly/f05838a4-cf25-4f3a-a089-696fd65a058a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Editing[[:space:]]Implicit[[:space:]]Assumptions[[:space:]]in[[:space:]]Text-to-Image[[:space:]]Diffusion[[:space:]]Models/15cdacdc-1f09-4143-98a9-32228ebb3ff3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Effective[[:space:]]Real[[:space:]]Image[[:space:]]Editing[[:space:]]with[[:space:]]Accelerated[[:space:]]Iterative[[:space:]]Diffusion[[:space:]]Inversion/b74ea1d7-f908-4dae-a180-1a6eed7acf3e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]3D[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]with[[:space:]]Superpoint[[:space:]]Transformer/206bf790-d1a0-4f0d-832c-59f71247b279_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Adaptive[[:space:]]Human-Object[[:space:]]Interaction[[:space:]]Detection[[:space:]]with[[:space:]]Concept-guided[[:space:]]Memory/6e9ae59f-4eee-40cd-bfa7-6de3ce71a2ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Computation[[:space:]]Sharing[[:space:]]for[[:space:]]Multi-Task[[:space:]]Visual[[:space:]]Scene[[:space:]]Understanding/e94bb98e-2aa8-4b05-b1e9-3f1e11b2e803_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Controllable[[:space:]]Multi-Task[[:space:]]Architectures/88104c66-7c15-4031-b124-e54abe7e7e4d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Converted[[:space:]]Spiking[[:space:]]Neural[[:space:]]Network[[:space:]]for[[:space:]]3D[[:space:]]and[[:space:]]2D[[:space:]]Classification/b31d6b1e-22ca-4cb3-b994-aeca38818030_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Decision-based[[:space:]]Black-box[[:space:]]Patch[[:space:]]Attacks[[:space:]]on[[:space:]]Video[[:space:]]Recognition/434f89a5-95e3-4370-b166-9d0412393d52_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Deep[[:space:]]Space[[:space:]]Filling[[:space:]]Curve/0765ec00-2640-448f-8fd2-4aab219af5b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Diffusion[[:space:]]Training[[:space:]]via[[:space:]]Min-SNR[[:space:]]Weighting[[:space:]]Strategy/f0086622-e6cb-4657-9e9c-1b7eb5b82299_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Discovery[[:space:]]and[[:space:]]Effective[[:space:]]Evaluation[[:space:]]of[[:space:]]Visual[[:space:]]Perceptual[[:space:]]Similarity_[[:space:]]A[[:space:]]Benchmark[[:space:]]and[[:space:]]Beyond/fc5bfd7a-2ab1-4479-b7a8-45c53f5a356c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Emotional[[:space:]]Adaptation[[:space:]]for[[:space:]]Audio-Driven[[:space:]]Talking-Head[[:space:]]Generation/9cfb8824-700a-426a-a5f5-c1bed0f5ff62_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Joint[[:space:]]Optimization[[:space:]]of[[:space:]]Layer-Adaptive[[:space:]]Weight[[:space:]]Pruning[[:space:]]in[[:space:]]Deep[[:space:]]Neural[[:space:]]Networks/a6e44cb4-0fcb-41a9-8253-10c5e7ea7f9e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]LiDAR[[:space:]]Point[[:space:]]Cloud[[:space:]]Oversegmentation[[:space:]]Network/a55e6c76-dbb6-41e8-bd41-cdd1a4eed22b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Model[[:space:]]Personalization[[:space:]]in[[:space:]]Federated[[:space:]]Learning[[:space:]]via[[:space:]]Client-Specific[[:space:]]Prompt[[:space:]]Generation/1c970683-6066-493a-90e6-53abffe7552e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Neural[[:space:]]Supersampling[[:space:]]on[[:space:]]a[[:space:]]Novel[[:space:]]Gaming[[:space:]]Dataset/55a519e8-1ca0-492b-b09f-678c470380d3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Region-Aware[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]High-Fidelity[[:space:]]Talking[[:space:]]Portrait[[:space:]]Synthesis/f9465311-edc9-4819-82b7-52057d42e358_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Transformer-based[[:space:]]3D[[:space:]]Object[[:space:]]Detection[[:space:]]with[[:space:]]Dynamic[[:space:]]Token[[:space:]]Halting/a85e0f70-3f90-4bda-b3c4-905473850984_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Unified[[:space:]]Demosaicing[[:space:]]for[[:space:]]Bayer[[:space:]]and[[:space:]]Non-Bayer[[:space:]]Patterned[[:space:]]Image[[:space:]]Sensors/d8b8bef8-569f-462c-a21c-239b2e40dc36_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Video[[:space:]]Action[[:space:]]Detection[[:space:]]with[[:space:]]Token[[:space:]]Dropout[[:space:]]and[[:space:]]Context[[:space:]]Refinement/7cdb6216-f2b9-4450-8e73-7e89950cc751_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]Video[[:space:]]Prediction[[:space:]]via[[:space:]]Sparsely[[:space:]]Conditioned[[:space:]]Flow[[:space:]]Matching/72637b83-0d8f-4088-ad30-1ca50b79a7cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient[[:space:]]View[[:space:]]Synthesis[[:space:]]with[[:space:]]Neural[[:space:]]Radiance[[:space:]]Distribution[[:space:]]Field/bd28ffdf-3517-4f9b-abac-2bff42627b4e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficient-VQGAN_[[:space:]]Towards[[:space:]]High-Resolution[[:space:]]Image[[:space:]]Generation[[:space:]]with[[:space:]]Efficient[[:space:]]Vision[[:space:]]Transformers/89255077-81d8-4c2e-9ab0-4dcaeeec1c6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EfficientTrain_[[:space:]]Exploring[[:space:]]Generalized[[:space:]]Curriculum[[:space:]]Learning[[:space:]]for[[:space:]]Training[[:space:]]Visual[[:space:]]Backbones/6f1e8bb5-b725-45e4-84d5-118c77a14661_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EfficientViT_[[:space:]]Lightweight[[:space:]]Multi-Scale[[:space:]]Attention[[:space:]]for[[:space:]]High-Resolution[[:space:]]Dense[[:space:]]Prediction/68d89dfe-de19-4936-9629-8321499e808a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Efficiently[[:space:]]Robustify[[:space:]]Pre-Trained[[:space:]]Models/9cc5eaad-90c6-4c57-bf85-0846ad30b0e8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ego-Humans_[[:space:]]An[[:space:]]Ego-Centric[[:space:]]3D[[:space:]]Multi-Human[[:space:]]Benchmark/1c11e310-e3fd-4aff-9650-9bca43d5b6ab_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Ego-Only_[[:space:]]Egocentric[[:space:]]Action[[:space:]]Detection[[:space:]]without[[:space:]]Exocentric[[:space:]]Transferring/f76ca444-21a7-4789-9919-832280d47d6d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EgoLoc_[[:space:]]Revisiting[[:space:]]3D[[:space:]]Object[[:space:]]Localization[[:space:]]from[[:space:]]Egocentric[[:space:]]Videos[[:space:]]with[[:space:]]Visual[[:space:]]Queries/cc7f248c-0f72-4a4f-9c8c-bb5875c7702c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EgoObjects_[[:space:]]A[[:space:]]Large-Scale[[:space:]]Egocentric[[:space:]]Dataset[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Object[[:space:]]Understanding/168ad744-f46b-47c9-a747-a8aee89d361d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EgoPCA_[[:space:]]A[[:space:]]New[[:space:]]Framework[[:space:]]for[[:space:]]Egocentric[[:space:]]Hand-Object[[:space:]]Interaction[[:space:]]Understanding/9ebf682e-84c8-4a61-b73b-72f2f13cb397_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EgoTV_[[:space:]]Egocentric[[:space:]]Task[[:space:]]Verification[[:space:]]from[[:space:]]Natural[[:space:]]Language[[:space:]]Task[[:space:]]Descriptions/861b6197-cdb8-4406-bcf7-c894924ae91f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EgoVLPv2_[[:space:]]Egocentric[[:space:]]Video-Language[[:space:]]Pre-training[[:space:]]with[[:space:]]Fusion[[:space:]]in[[:space:]]the[[:space:]]Backbone/4046dd2f-509b-4755-be31-c2a4774af5f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EigenPlaces_[[:space:]]Training[[:space:]]Viewpoint[[:space:]]Robust[[:space:]]Models[[:space:]]for[[:space:]]Visual[[:space:]]Place[[:space:]]Recognition/1d87f5dc-6dda-448c-be5d-9406e9b60346_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EigenTrajectory_[[:space:]]Low-Rank[[:space:]]Descriptors[[:space:]]for[[:space:]]Multi-Modal[[:space:]]Trajectory[[:space:]]Forecasting/4dc73c56-4c71-4b1c-8b72-c726c25b4581_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ElasticViT_[[:space:]]Conflict-aware[[:space:]]Supernet[[:space:]]Training[[:space:]]for[[:space:]]Deploying[[:space:]]Fast[[:space:]]Vision[[:space:]]Transformer[[:space:]]on[[:space:]]Diverse[[:space:]]Mobile[[:space:]]Devices/b3399b8b-6a5d-480e-851b-f2b849f5dfbc_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EmoSet_[[:space:]]A[[:space:]]Large-scale[[:space:]]Visual[[:space:]]Emotion[[:space:]]Dataset[[:space:]]with[[:space:]]Rich[[:space:]]Attributes/423ac1f3-e6af-4796-b860-2399494ecb2c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EmoTalk_[[:space:]]Speech-Driven[[:space:]]Emotional[[:space:]]Disentanglement[[:space:]]for[[:space:]]3D[[:space:]]Face[[:space:]]Animation/09d8e1ef-dcfc-49d5-8c5a-3e9251dc1aa5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Emotional[[:space:]]Listener[[:space:]]Portrait_[[:space:]]Neural[[:space:]]Listener[[:space:]]Head[[:space:]]Generation[[:space:]]with[[:space:]]Emotion/9d6c9ada-5aba-494f-ae06-7a1e7daecfd6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Empowering[[:space:]]Low-Light[[:space:]]Image[[:space:]]Enhancer[[:space:]]through[[:space:]]Customized[[:space:]]Learnable[[:space:]]Priors/706a09ea-04bb-41a7-8e71-5ae2fcb394db_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Encyclopedic[[:space:]]VQA_[[:space:]]Visual[[:space:]]Questions[[:space:]]About[[:space:]]Detailed[[:space:]]Properties[[:space:]]of[[:space:]]Fine-Grained[[:space:]]Categories/ba285a2f-2b2b-4a3f-a668-d5b5141f838f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/End-to-End[[:space:]]Diffusion[[:space:]]Latent[[:space:]]Optimization[[:space:]]Improves[[:space:]]Classifier[[:space:]]Guidance/e3c78165-4121-4a76-af64-2398b7625453_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/End-to-end[[:space:]]3D[[:space:]]Tracking[[:space:]]with[[:space:]]Decoupled[[:space:]]Queries/5a984014-38cc-4e6c-b86a-0e7b1fb63ba6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/End2End[[:space:]]Multi-View[[:space:]]Feature[[:space:]]Matching[[:space:]]with[[:space:]]Differentiable[[:space:]]Pose[[:space:]]Optimization/cada1448-a478-4cfc-b8e3-440ed13be109_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Energy-based[[:space:]]Self-Training[[:space:]]and[[:space:]]Normalization[[:space:]]for[[:space:]]Unsupervised[[:space:]]Domain[[:space:]]Adaptation/531ae82c-29ee-425c-bbd5-38c9aea53848_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhanced[[:space:]]Meta[[:space:]]Label[[:space:]]Correction[[:space:]]for[[:space:]]Coping[[:space:]]with[[:space:]]Label[[:space:]]Corruption/15847755-93d3-4db9-b5d4-9aac37f1dc99_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhanced[[:space:]]Soft[[:space:]]Label[[:space:]]for[[:space:]]Semi-Supervised[[:space:]]Semantic[[:space:]]Segmentation/8ad3292a-b1aa-4c2e-a3c3-facdacd29dc8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhancing[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]in[[:space:]]Low-Label[[:space:]]Regime[[:space:]]via[[:space:]]Adaptively[[:space:]]Weighted[[:space:]]Regularization[[:space:]]and[[:space:]]Knowledge[[:space:]]Distillation/5b20948c-7578-4f78-9ef5-295a08b56c64_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhancing[[:space:]]Fine-Tuning[[:space:]]Based[[:space:]]Backdoor[[:space:]]Defense[[:space:]]with[[:space:]]Sharpness-Aware[[:space:]]Minimization/fd991178-a503-478a-9bf1-7144851286b5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhancing[[:space:]]Generalization[[:space:]]of[[:space:]]Universal[[:space:]]Adversarial[[:space:]]Perturbation[[:space:]]through[[:space:]]Gradient[[:space:]]Aggregation/2053f0e4-6385-404d-bc0c-80d101158bad_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhancing[[:space:]]Modality-Agnostic[[:space:]]Representations[[:space:]]via[[:space:]]Meta-Learning[[:space:]]for[[:space:]]Brain[[:space:]]Tumor[[:space:]]Segmentation/e83a2e09-3cda-446b-8415-ef045a10bfe5_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhancing[[:space:]]NeRF[[:space:]]akin[[:space:]]to[[:space:]]Enhancing[[:space:]]LLMs_[[:space:]]Generalizable[[:space:]]NeRF[[:space:]]Transformer[[:space:]]with[[:space:]]Mixture-of-View-Experts/789b5b35-746d-41f6-b000-29d462744766_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhancing[[:space:]]Non-line-of-sight[[:space:]]Imaging[[:space:]]via[[:space:]]Learnable[[:space:]]Inverse[[:space:]]Kernel[[:space:]]and[[:space:]]Attention[[:space:]]Mechanisms/bc26baaa-3fdb-4899-b2b1-cd5549bb970c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhancing[[:space:]]Privacy[[:space:]]Preservation[[:space:]]in[[:space:]]Federated[[:space:]]Learning[[:space:]]via[[:space:]]Learning[[:space:]]Rate[[:space:]]Perturbation/8f3f28ea-02eb-4bfc-936c-166bf61a0a7b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Enhancing[[:space:]]Sample[[:space:]]Utilization[[:space:]]through[[:space:]]Sample[[:space:]]Adaptive[[:space:]]Augmentation[[:space:]]in[[:space:]]Semi-Supervised[[:space:]]Learning/b499a7bd-d5c6-4a3f-96e2-a386fb4f001a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Environment[[:space:]]Agnostic[[:space:]]Representation[[:space:]]for[[:space:]]Visual[[:space:]]Reinforcement[[:space:]]Learning/fcace76e-c3fc-45c3-bf6f-6cc92b443422_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Environment-Invariant[[:space:]]Curriculum[[:space:]]Relation[[:space:]]Learning[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Scene[[:space:]]Graph[[:space:]]Generation/e22d8745-758d-4a8d-8b9c-7a64886a7762_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Equivariant[[:space:]]Similarity[[:space:]]for[[:space:]]Vision-Language[[:space:]]Foundation[[:space:]]Models/a9194639-5422-418f-871a-5d12a835f108_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Erasing[[:space:]]Concepts[[:space:]]from[[:space:]]Diffusion[[:space:]]Models/f6da9568-2bde-4c71-81d3-615801a78c19_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Essential[[:space:]]Matrix[[:space:]]Estimation[[:space:]]using[[:space:]]Convex[[:space:]]Relaxations[[:space:]]in[[:space:]]Orthogonal[[:space:]]Space/0f679c85-6b31-44b5-82a4-c5a95ad9b9b9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Estimator[[:space:]]Meets[[:space:]]Equilibrium[[:space:]]Perspective_[[:space:]]A[[:space:]]Rectified[[:space:]]Straight[[:space:]]Through[[:space:]]Estimator[[:space:]]for[[:space:]]Binary[[:space:]]Neural[[:space:]]Networks[[:space:]]Training/038c9c58-b7c1-4689-98fe-c23fc3667613_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Eulerian[[:space:]]Single-Photon[[:space:]]Vision/f1e73f92-a2fb-4130-a805-d3edd9ab8f06_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Evaluating[[:space:]]Data[[:space:]]Attribution[[:space:]]for[[:space:]]Text-to-Image[[:space:]]Models/8a13797d-5742-41e1-9d0b-bca4200e14c3_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Evaluation[[:space:]]and[[:space:]]Improvement[[:space:]]of[[:space:]]Interpretability[[:space:]]for[[:space:]]Self-Explainable[[:space:]]Part-Prototype[[:space:]]Networks/5ed03f23-86ed-4eff-a798-495874543417_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Event[[:space:]]Camera[[:space:]]Data[[:space:]]Pre-training/7199ac1f-d0e9-4f78-a6b6-34444716d1ff_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Event-Guided[[:space:]]Procedure[[:space:]]Planning[[:space:]]from[[:space:]]Instructional[[:space:]]Videos[[:space:]]with[[:space:]]Text[[:space:]]Supervision/865b64dd-fd95-4b59-9a64-93db04db263b_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Event-based[[:space:]]Temporally[[:space:]]Dense[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation[[:space:]]with[[:space:]]Sequential[[:space:]]Learning/a16715ab-7bf5-4cbe-9c69-167ccd53abb6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Eventful[[:space:]]Transformers_[[:space:]]Leveraging[[:space:]]Temporal[[:space:]]Redundancy[[:space:]]in[[:space:]]Vision[[:space:]]Transformers/b0f6e3b8-2625-4fcd-a25c-6d4c2e2a14e4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/EverLight_[[:space:]]Indoor-Outdoor[[:space:]]Editable[[:space:]]HDR[[:space:]]Lighting[[:space:]]Estimation/3a730283-cedc-440c-ade0-c3f840da0650_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ExBluRF_[[:space:]]Efficient[[:space:]]Radiance[[:space:]]Fields[[:space:]]for[[:space:]]Extreme[[:space:]]Motion[[:space:]]Blurred[[:space:]]Images/0353d7dd-628a-4d68-a945-ef6e9a8afef7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Examining[[:space:]]Autoexposure[[:space:]]for[[:space:]]Challenging[[:space:]]Scenes/a5522d0a-a5c7-4ef9-988a-d1316244fdc2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exemplar-Free[[:space:]]Continual[[:space:]]Transformer[[:space:]]with[[:space:]]Convolutions/af50468d-c112-4d91-8c08-d06ac026e29d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Explaining[[:space:]]Adversarial[[:space:]]Robustness[[:space:]]of[[:space:]]Neural[[:space:]]Networks[[:space:]]from[[:space:]]Clustering[[:space:]]Effect[[:space:]]Perspective/ebc2dc5e-4f57-408f-aac2-3497cdfd01cd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Explicit[[:space:]]Motion[[:space:]]Disentangling[[:space:]]for[[:space:]]Efficient[[:space:]]Optical[[:space:]]Flow[[:space:]]Estimation/f61131d6-1693-4338-ad7e-9ef999e05ff2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploiting[[:space:]]Proximity-Aware[[:space:]]Tasks[[:space:]]for[[:space:]]Embodied[[:space:]]Social[[:space:]]Navigation/853a56da-8560-4c39-86c9-51141a00a1bb_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Explore[[:space:]]and[[:space:]]Tell_[[:space:]]Embodied[[:space:]]Visual[[:space:]]Captioning[[:space:]]in[[:space:]]3D[[:space:]]Environments/e39551fd-b074-4196-868b-a22dd2decd4c_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Group[[:space:]]Video[[:space:]]Captioning[[:space:]]with[[:space:]]Efficient[[:space:]]Relational[[:space:]]Approximation/af8df661-05b3-4a4b-b3f9-620449ef9b58_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Lightweight[[:space:]]Hierarchical[[:space:]]Vision[[:space:]]Transformers[[:space:]]for[[:space:]]Efficient[[:space:]]Visual[[:space:]]Tracking/2bfb7b34-7f8d-47b0-b324-0817e9dd459e_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Model[[:space:]]Transferability[[:space:]]through[[:space:]]the[[:space:]]Lens[[:space:]]of[[:space:]]Potential[[:space:]]Energy/1dd05f2f-c895-4d07-a2e4-9ab8a335e257_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Object-Centric[[:space:]]Temporal[[:space:]]Modeling[[:space:]]for[[:space:]]Efficient[[:space:]]Multi-View[[:space:]]3D[[:space:]]Object[[:space:]]Detection/2eaffbdc-1c46-4211-8ecc-dba2fbb36023_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Open-Vocabulary[[:space:]]Semantic[[:space:]]Segmentation[[:space:]]from[[:space:]]CLIP[[:space:]]Vision[[:space:]]Encoder[[:space:]]Distillation[[:space:]]Only/e071c408-c408-4047-a62d-c3714313f031_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Positional[[:space:]]Characteristics[[:space:]]of[[:space:]]Dual-Pixel[[:space:]]Data[[:space:]]for[[:space:]]Camera[[:space:]]Autofocus/30c15f86-3160-4142-99db-8db8fea33211_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Predicate[[:space:]]Visual[[:space:]]Context[[:space:]]in[[:space:]]Detecting[[:space:]]of[[:space:]]Human-Object[[:space:]]Interactions/9d7d5493-0002-40c7-8cfd-27ec12b12a5d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Temporal[[:space:]]Concurrency[[:space:]]for[[:space:]]Video-Language[[:space:]]Representation[[:space:]]Learning/de72ae1b-fa2f-4664-9042-11c4c1e36d34_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Temporal[[:space:]]Frequency[[:space:]]Spectrum[[:space:]]in[[:space:]]Deep[[:space:]]Video[[:space:]]Deblurring/21732e8c-3f62-415c-a009-700974ba8411_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Transformers[[:space:]]for[[:space:]]Open-world[[:space:]]Instance[[:space:]]Segmentation/4bd0a7e1-f652-4150-9a04-b8bf97221f1a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]Video[[:space:]]Quality[[:space:]]Assessment[[:space:]]on[[:space:]]User[[:space:]]Generated[[:space:]]Contents[[:space:]]from[[:space:]]Aesthetic[[:space:]]and[[:space:]]Technical[[:space:]]Perspectives/3179c812-deff-4e60-9abf-21602fc6f94d_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]the[[:space:]]Benefits[[:space:]]of[[:space:]]Visual[[:space:]]Prompting[[:space:]]in[[:space:]]Differential[[:space:]]Privacy/abbeecd3-3916-43f3-8845-3b8b51db8247_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Exploring[[:space:]]the[[:space:]]Sim2Real[[:space:]]Gap[[:space:]]Using[[:space:]]Digital[[:space:]]Twins/117e1060-86aa-4ca1-a194-49adb636ea11_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/ExposureDiffusion_[[:space:]]Learning[[:space:]]to[[:space:]]Expose[[:space:]]for[[:space:]]Low-light[[:space:]]Image[[:space:]]Enhancement/e100a96c-bc65-40cc-afe3-27ddfb28f48a_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Expressive[[:space:]]Text-to-Image[[:space:]]Generation[[:space:]]with[[:space:]]Rich[[:space:]]Text/0ca67ade-7baf-4bcc-b5eb-3e0c91f797cf_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Extensible[[:space:]]and[[:space:]]Efficient[[:space:]]Proxy[[:space:]]for[[:space:]]Neural[[:space:]]Architecture[[:space:]]Search/9e4fac96-8ab1-4ec7-a547-c645996269ca_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/F&F[[:space:]]Attack_[[:space:]]Adversarial[[:space:]]Attack[[:space:]]against[[:space:]]Multiple[[:space:]]Object[[:space:]]Trackers[[:space:]]by[[:space:]]Inducing[[:space:]]False[[:space:]]Negatives[[:space:]]and[[:space:]]False[[:space:]]Positives/fe54c5d9-0d2a-4043-8b59-54724bce8cb8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FACET_[[:space:]]Fairness[[:space:]]in[[:space:]]Computer[[:space:]]Vision[[:space:]]Evaluation[[:space:]]Benchmark/df38e774-83c7-46fc-bb94-1e1ad374973f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FACTS_[[:space:]]First[[:space:]]Amplify[[:space:]]Correlations[[:space:]]and[[:space:]]Then[[:space:]]Slice[[:space:]]to[[:space:]]Discover[[:space:]]Bias/71f78e0a-fb43-4f08-a623-feee8d9687a9_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FB-BEV_[[:space:]]BEV[[:space:]]Representation[[:space:]]from[[:space:]]Forward-Backward[[:space:]]View[[:space:]]Transformations/64d62057-0642-4c0a-af0f-a23225c099c2_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FBLNet_[[:space:]]FeedBack[[:space:]]Loop[[:space:]]Network[[:space:]]for[[:space:]]Driver[[:space:]]Attention[[:space:]]Prediction/8d64da33-7cea-41fb-a8ae-c020ef457d71_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FCCNs_[[:space:]]Fully[[:space:]]Complex-valued[[:space:]]Convolutional[[:space:]]Networks[[:space:]]using[[:space:]]Complex-valued[[:space:]]Color[[:space:]]Model[[:space:]]and[[:space:]]Loss[[:space:]]Function/b7773e41-b998-4160-87f3-a31b58ad3609_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FDViT_[[:space:]]Improve[[:space:]]the[[:space:]]Hierarchical[[:space:]]Architecture[[:space:]]of[[:space:]]Vision[[:space:]]Transformer/34d14881-370a-4f54-96ee-8e6950cd20d7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FLIP_[[:space:]]Cross-domain[[:space:]]Face[[:space:]]Anti-spoofing[[:space:]]with[[:space:]]Language[[:space:]]Guidance/3647baa8-d5f0-46c6-b925-fa4f8c8e764f_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FLatten[[:space:]]Transformer_[[:space:]]Vision[[:space:]]Transformer[[:space:]]using[[:space:]]Focused[[:space:]]Linear[[:space:]]Attention/0473946d-360e-416c-9e49-d7a86cd5bae7_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FPR_[[:space:]]False[[:space:]]Positive[[:space:]]Rectification[[:space:]]for[[:space:]]Weakly[[:space:]]Supervised[[:space:]]Semantic[[:space:]]Segmentation/23a6eebe-8c7f-4868-9a86-a93635590661_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FRAug_[[:space:]]Tackling[[:space:]]Federated[[:space:]]Learning[[:space:]]with[[:space:]]Non-IID[[:space:]]Features[[:space:]]via[[:space:]]Representation[[:space:]]Augmentation/ab485480-1175-4656-9a32-fd39ae969299_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FS-DETR_[[:space:]]Few-Shot[[:space:]]DEtection[[:space:]]TRansformer[[:space:]]with[[:space:]]Prompting[[:space:]]and[[:space:]]without[[:space:]]Re-Training/05f1e760-0f28-4518-aa51-71cf05e4e6f6_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FSAR_[[:space:]]Federated[[:space:]]Skeleton-based[[:space:]]Action[[:space:]]Recognition[[:space:]]with[[:space:]]Adaptive[[:space:]]Topology[[:space:]]Structure[[:space:]]and[[:space:]]Knowledge[[:space:]]Distillation/73880105-40da-4085-a9a9-6786a89af620_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FSI_[[:space:]]Frequency[[:space:]]and[[:space:]]Spatial[[:space:]]Interactive[[:space:]]Learning[[:space:]]for[[:space:]]Image[[:space:]]Restoration[[:space:]]in[[:space:]]Under-Display[[:space:]]Cameras/605d5da0-0822-4093-967e-22f22f0067d1_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FULLER_[[:space:]]Unified[[:space:]]Multi-modality[[:space:]]Multi-task[[:space:]]3D[[:space:]]Perception[[:space:]]via[[:space:]]Multi-level[[:space:]]Gradient[[:space:]]Calibration/1ddd7908-f78d-4029-9fee-0b0d027735b8_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Face[[:space:]]Clustering[[:space:]]via[[:space:]]Graph[[:space:]]Convolutional[[:space:]]Networks[[:space:]]with[[:space:]]Confidence[[:space:]]Edges/3f7b5ff5-2652-4ec0-9d28-4de6563d9be4_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/FaceCLIPNeRF_[[:space:]]Text-driven[[:space:]]3D[[:space:]]Face[[:space:]]Manipulation[[:space:]]using[[:space:]]Deformable[[:space:]]Neural[[:space:]]Radiance[[:space:]]Fields/e9b62c55-9dec-4fc0-ae34-206cef93b3dd_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Factorized[[:space:]]Inverse[[:space:]]Path[[:space:]]Tracing[[:space:]]for[[:space:]]Efficient[[:space:]]and[[:space:]]Accurate[[:space:]]Material-Lighting[[:space:]]Estimation/d681a9b9-626b-4659-8518-238571b79521_origin.pdf filter=lfs diff=lfs merge=lfs -text 2023/Fan-Beam[[:space:]]Binarization[[:space:]]Difference[[:space:]]Projection[[:space:]](FB-BDP)_[[:space:]]A[[:space:]]Novel[[:space:]]Local[[:space:]]Object[[:space:]]Descriptor[[:space:]]for[[:space:]]Fine-Grained[[:space:]]Leaf[[:space:]]Image[[:space:]]Retrieval/96bdcc1f-9ca5-4074-b1f9-4dabaad67964_origin.pdf filter=lfs diff=lfs merge=lfs -text