VoiceGRPO: Modern MoE Transformers with Group Relative Policy Optimization GRPO for AI Voice Health Care Applications on Voice Pathology Detection Paper • 2503.03797 • Published Mar 5, 2025 • 1
A Survey on Computational Pathology Foundation Models: Datasets, Adaptation Strategies, and Evaluation Tasks Paper • 2501.15724 • Published Jan 27, 2025 • 1
view article Article EMO: Pretraining mixture of experts for emergent modularity allenai • 7 days ago • 33
view article Article How to build scalable web apps with OpenAI's Privacy Filter +1 ysharma, freddyaboulton, abidlabs • 18 days ago • 8
view article Article DeepInfra on Hugging Face Inference Providers 🔥 +6 araikin, shang-pin-deepinfra, Pernekhan, yessenzhar, ovuruska, celinah, sbrandeis, Wauplin • 16 days ago • 9
view article Article Building Blocks for Foundation Model Training and Inference on AWS amazon • 3 days ago • 19
TMAS: Scaling Test-Time Compute via Multi-Agent Synergy Paper • 2605.10344 • Published 4 days ago • 47
CollabVR: Collaborative Video Reasoning with Vision-Language and Video Generation Models Paper • 2605.08735 • Published 6 days ago • 67
Mean Mode Screaming: Mean--Variance Split Residuals for 1000-Layer Diffusion Transformers Paper • 2605.06169 • Published 8 days ago • 182
Detection Transformer for Teeth Detection, Segmentation, and Numbering in Oral Rare Diseases: Focus on Data Augmentation and Inpainting Techniques Paper • 2402.04408 • Published Feb 6, 2024 • 1
Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference Paper • 2303.06182 • Published Mar 10, 2023 • 2
CAMERA: Multi-Matrix Joint Compression for MoE Models via Micro-Expert Redundancy Analysis Paper • 2508.02322 • Published Aug 4, 2025 • 1
Sub-MoE: Efficient Mixture-of-Expert LLMs Compression via Subspace Expert Merging Paper • 2506.23266 • Published Jun 29, 2025 • 1
Shortcut-connected Expert Parallelism for Accelerating Mixture-of-Experts Paper • 2404.05019 • Published Apr 7, 2024 • 2
Expert Race: A Flexible Routing Strategy for Scaling Diffusion Transformer with Mixture of Experts Paper • 2503.16057 • Published Mar 20, 2025 • 15