SPES Collection Pretrained models for paper "Pretraining A Large Language Model using Distributed GPUs: A Memory-Efficient Decentralized Paradigm" • 3 items • Updated 4 days ago
SPES Collection Pretrained models for paper "Pretraining A Large Language Model using Distributed GPUs: A Memory-Efficient Decentralized Paradigm" • 3 items • Updated 4 days ago
Caption Anything: Interactive Image Description with Diverse Multimodal Controls Paper • 2305.02677 • Published May 4, 2023 • 1
Transferable Decoding with Visual Entities for Zero-Shot Image Captioning Paper • 2307.16525 • Published Jul 31, 2023 • 1
LongVALE: Vision-Audio-Language-Event Benchmark Towards Time-Aware Omni-Modal Perception of Long Videos Paper • 2411.19772 • Published Nov 29, 2024 • 1
Pretraining A Large Language Model using Distributed GPUs: A Memory-Efficient Decentralized Paradigm Paper • 2602.11543 • Published 30 days ago • 6 • 4
Pretraining A Large Language Model using Distributed GPUs: A Memory-Efficient Decentralized Paradigm Paper • 2602.11543 • Published 30 days ago • 6
Pretraining A Large Language Model using Distributed GPUs: A Memory-Efficient Decentralized Paradigm Paper • 2602.11543 • Published 30 days ago • 6
GENIUS: Generative Fluid Intelligence Evaluation Suite Paper • 2602.11144 • Published about 1 month ago • 53
Diversity-Preserved Distribution Matching Distillation for Fast Visual Synthesis Paper • 2602.03139 • Published Feb 3 • 42
ARC-Chapter: Structuring Hour-Long Videos into Navigable Chapters and Hierarchical Summaries Paper • 2511.14349 • Published Nov 18, 2025 • 18