<...,"['RAG Support', 'Multi-turn']"
facebook/MobileLLM-Pro,MobileLLM-Pro,facebook,2025-09-10T18:40:06+00:00,4486,138,transformers,"['transformers', 'safetensors', 'llama4_text', 'text-generation', 'facebook', 'meta', 'pytorch', 'conversational', 'custom_code', 'en', 'base_model:facebook/MobileLLM-Pro-base', 'base_model:finetune:facebook/MobileLLM-Pro-base', 'license:fair-noncommercial-research-license', 'autotrain_compatible', 'region:us']",10,196,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:01:04.530889,No README available,[]
nineninesix/kani-tts-370m,kani-tts-370m,nineninesix,2025-09-30T07:31:22+00:00,8444,136,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'text-to-speech', 'en', 'de', 'ar', 'zh', 'es', 'ko', 'arxiv:2505.20506', 'base_model:nineninesix/kani-tts-450m-0.2-pt', 'base_model:finetune:nineninesix/kani-tts-450m-0.2-pt', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",10,197,Lfm2ForCausalLM,lfm2,283798528.0,128000.0,1024.0,16.0,16.0,80539.0,Language Model,"['Fine-tuned', 'Specialized']",['conversation'],"['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Legal']",{},75.0,Medium,0.75,2025-11-02T06:01:04.690562,"license: apache-2.0 language: - en - de - ar - zh - es - ko pipeline_tag: text-to-speech library_name: transformers base_model: - nineninesix/kani-tts-450m-0.2-pt
- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']"
LiquidAI/LFM2-8B-A1B,LFM2-8B-A1B,LiquidAI,2025-10-07T13:55:39+00:00,13348,231,transformers,"['transformers', 'safetensors', 'lfm2_moe', 'text-generation', 'liquid', 'lfm2', 'edge', 'moe', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,207,Lfm2MoeForCausalLM,lfm2_moe,1342177280.0,128000.0,2048.0,32.0,24.0,65536.0,Language Model,"['Quantized', 'LoRA']","['text-generation', 'question-answering', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing', 'Healthcare']",{},88.0,Easy,0.88,2025-11-02T06:01:12.046492,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge - moe ...,"['Function Calling', 'RAG Support', 'Fast Inference', 'Multi-turn', 'Safety Aligned']"
cerebras/Qwen3-Coder-REAP-25B-A3B,Qwen3-Coder-REAP-25B-A3B,cerebras,2025-10-20T15:40:03+00:00,971,32,transformers,"['transformers', 'qwen3_moe', 'text-generation', 'qwen-coder', 'MOE', 'pruning', 'compression', 'conversational', 'en', 'arxiv:2510.13999', 'base_model:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'base_model:finetune:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,208,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,262144.0,2048.0,32.0,48.0,151936.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'reasoning']","['English', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},85.0,Easy,0.85,2025-11-02T06:01:12.522267,language: - en library_name: transformers tags: - qwen-coder - MOE - pruning - compression license: apache-2.0 name: cerebras/Qwen3-Coder-REAP-25B-A3B description: > This model was obtained by uniform...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']"
meta-llama/Llama-3.2-1B-Instruct,Llama-3.2-1B-Instruct,meta-llama,2024-09-18T15:12:47+00:00,3843095,1141,transformers,"['transformers', 'safetensors', 'llama', 'text-generation', 'facebook', 'meta', 'pytorch', 'llama-3', 'conversational', 'en', 'de', 'fr', 'it', 'pt', 'hi', 'es', 'th', 'arxiv:2204.05149', 'arxiv:2405.16406', 'license:llama3.2', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,209,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:01:13.704023,No README available,[]
microsoft/Phi-4-mini-instruct,Phi-4-mini-instruct,microsoft,2025-02-19T01:00:58+00:00,249315,623,transformers,"['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'ar', 'zh', 'cs', 'da', 'nl', 'en', 'fi', 'fr', 'de', 'he', 'hu', 'it', 'ja', 'ko', 'no', 'pl', 'pt', 'ru', 'es', 'sv', 'th', 'tr', 'uk', 'arxiv:2503.01743', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,210,Phi3ForCausalLM,phi3,4238475264.0,131072.0,3072.0,24.0,32.0,200064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Business', 'Healthcare', 'Legal']",{},86.0,Easy,0.86,2025-11-02T06:01:13.889560,language: - multilingual - ar - zh - cs - da - nl - en - fi - fr - de - he - hu - it - ja - ko - 'no' - pl - pt - ru - es - sv - th - tr - uk library_name: transformers license: mit license_link: http...,"['Function Calling', 'RAG Support', 'Long Context', 'Multi-turn', 'Safety Aligned']"
Qwen/Qwen3-30B-A3B,Qwen3-30B-A3B,Qwen,2025-04-27T03:43:05+00:00,421099,808,transformers,"['transformers', 'safetensors', 'qwen3_moe', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-30B-A3B-Base', 'base_model:finetune:Qwen/Qwen3-30B-A3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",8,211,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,40960.0,2048.0,32.0,48.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:01:16.176834,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-30B-A3B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-30B-A3B-Base