model_id,model_name,author,created_at,downloads,likes,library,tags,trending_score,trending_rank,architecture,model_type,num_parameters,max_position_embeddings,hidden_size,num_attention_heads,num_hidden_layers,vocab_size,primary_category,secondary_categories,task_types,language_support,use_cases,performance_metrics,a2ap_compatibility_score,merge_difficulty,evolution_potential,analysis_timestamp,readme_summary,special_features MiniMaxAI/MiniMax-M2,MiniMax-M2,MiniMaxAI,2025-10-22T13:45:10+00:00,529835,912,transformers,"['transformers', 'safetensors', 'minimax', 'text-generation', 'conversational', 'arxiv:2504.07164', 'arxiv:2509.06501', 'arxiv:2509.13160', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'fp8', 'region:us']",912,1,MiniMaxM2ForCausalLM,minimax,7635861504.0,196608.0,3072.0,48.0,62.0,200064.0,Language Model,['LoRA'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production']",{},91.0,Easy,0.91,2025-11-02T05:59:02.143282,"pipeline_tag: text-generation license: mit library_name: transformers
- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ inference: false fine-tuning: false langu...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" PokeeAI/pokee_research_7b,pokee_research_7b,PokeeAI,2025-10-17T20:38:37+00:00,5574,95,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'agent', 'deepresearch', 'llm', 'rl', 'reinforcementlearning', 'conversational', 'en', 'dataset:miromind-ai/MiroRL-GenQA', 'arxiv:2510.15862', 'base_model:Qwen/Qwen2.5-7B-Instruct', 'base_model:finetune:Qwen/Qwen2.5-7B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",24,27,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged', 'Specialized']","['text-generation', 'question-answering', 'summarization', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education', 'Healthcare', 'Legal', 'Finance']",{},78.0,Medium,0.78,2025-11-02T05:59:14.893408,base_model: - Qwen/Qwen2.5-7B-Instruct datasets: - miromind-ai/MiroRL-GenQA language: - en license: apache-2.0 tags: - agent - deepresearch - llm - rl - reinforcementlearning pipeline_tag: text-genera...,"['RAG Support', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" ibm-granite/granite-4.0-350m-base,granite-4.0-350m-base,ibm-granite,2025-10-07T20:25:24+00:00,495,23,transformers,"['transformers', 'safetensors', 'granitemoehybrid', 'text-generation', 'language', 'granite-4.0', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",23,28,GraniteMoeHybridForCausalLM,granitemoehybrid,455081984.0,32768.0,1024.0,16.0,28.0,100352.0,Language Model,"['LoRA', 'Specialized']","['text-generation', 'question-answering', 'text-classification', 'summarization']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Education']",{},86.0,Easy,0.86,2025-11-02T05:59:15.069927,license: apache-2.0 library_name: transformers tags: - language - granite-4.0 **Model Summary:** Granite-4.0-350M-Base is a lightweight decoder-only language model designed for scenarios where efficie...,"['Long Context', 'Fast Inference', 'Safety Aligned']" internlm/JanusCoder-14B,JanusCoder-14B,internlm,2025-10-27T09:34:49+00:00,165,23,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",23,29,Qwen3ForCausalLM,qwen3,13360824320.0,32768.0,5120.0,40.0,40.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},58.0,Hard,0.58,2025-11-02T05:59:15.259607,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] ByteDance/Ouro-2.6B,Ouro-2.6B,ByteDance,2025-10-28T22:19:46+00:00,227,21,transformers,"['transformers', 'safetensors', 'ouro', 'text-generation', 'looped-language-model', 'reasoning', 'recurrent-depth', 'conversational', 'custom_code', 'arxiv:2510.25741', 'license:apache-2.0', 'autotrain_compatible', 'region:us']",21,30,OuroForCausalLM,ouro,2516582400.0,65536.0,2048.0,16.0,48.0,49152.0,Language Model,[],"['text-generation', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education']",{},78.0,Medium,0.78,2025-11-02T05:59:15.435642,library_name: transformers license: apache-2.0 pipeline_tag: text-generation tags: - looped-language-model - reasoning - recurrent-depth ![Ouro Logo](assets/logo.png) [📚 Paper (Hugging Face)](https://...,"['RAG Support', 'Long Context']" HuggingFaceTB/SmolLM3-3B,SmolLM3-3B,HuggingFaceTB,2025-07-08T10:11:45+00:00,56521,768,transformers,"['transformers', 'safetensors', 'smollm3', 'text-generation', 'conversational', 'en', 'fr', 'es', 'it', 'pt', 'zh', 'ar', 'ru', 'base_model:HuggingFaceTB/SmolLM3-3B-Base', 'base_model:finetune:HuggingFaceTB/SmolLM3-3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",20,31,SmolLM3ForCausalLM,smollm3,2074607616.0,65536.0,2048.0,16.0,36.0,128256.0,Language Model,['Quantized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},83.0,Medium,0.83,2025-11-02T05:59:17.626701,library_name: transformers license: apache-2.0 language: - en - fr - es - it - pt - zh - ar - ru base_model: - HuggingFaceTB/SmolLM3-3B-Base ![image/png](https://cdn-uploads.huggingface.co/production/...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient']" Qwen/Qwen3-0.6B,Qwen3-0.6B,Qwen,2025-04-27T03:40:08+00:00,7327000,752,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-0.6B-Base', 'base_model:finetune:Qwen/Qwen3-0.6B-Base', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",19,32,Qwen3ForCausalLM,qwen3,507904000.0,40960.0,1024.0,16.0,28.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},78.0,Medium,0.78,2025-11-02T05:59:17.864018,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-0.6B-Base This model was obtained by uniformly pruning 25...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" Qwen/Qwen2.5-7B-Instruct,Qwen2.5-7B-Instruct,Qwen,2024-09-16T11:55:40+00:00,8169288,849,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-7B', 'base_model:finetune:Qwen/Qwen2.5-7B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",17,36,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},86.0,Easy,0.86,2025-11-02T05:59:20.058011,license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE language: - en pipeline_tag: text-generation base_model: Qwen/Qwen2.5-7B tags: - chat library_name: ...,"['Long Context', 'Fast Inference', 'Safety Aligned']" deepseek-ai/DeepSeek-R1,DeepSeek-R1,deepseek-ai,2025-01-20T03:46:07+00:00,425526,12820,transformers,"['transformers', 'safetensors', 'deepseek_v3', 'text-generation', 'conversational', 'custom_code', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'fp8', 'region:us']",17,37,DeepseekV3ForCausalLM,deepseek_v3,38537003008.0,163840.0,7168.0,128.0,61.0,129280.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged']","['text-generation', 'question-answering', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business']",{},48.0,Hard,0.48,2025-11-02T05:59:20.248560,"license: mit library_name: transformers
- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" microsoft/VibeVoice-1.5B,VibeVoice-1.5B,microsoft,2025-08-25T13:46:48+00:00,223609,1943,transformers,"['transformers', 'safetensors', 'vibevoice', 'text-generation', 'Podcast', 'text-to-speech', 'en', 'zh', 'arxiv:2508.19205', 'arxiv:2412.08635', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,56,VibeVoiceForConditionalGeneration,vibevoice,,,,,,,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business', 'Legal']",{},53.0,Hard,0.53,2025-11-02T05:59:33.358457,"language: - en - zh license: mit pipeline_tag: text-to-speech tags: - Podcast library_name: transformers VibeVoice is a novel framework designed for generating expressive, long-form, multi-speaker con...","['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" LiquidAI/LFM2-350M-Extract,LFM2-350M-Extract,LiquidAI,2025-09-03T17:13:53+00:00,5852,61,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'liquid', 'edge', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'base_model:LiquidAI/LFM2-350M', 'base_model:finetune:LiquidAI/LFM2-350M', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,57,Lfm2ForCausalLM,lfm2,268435456.0,128000.0,1024.0,16.0,16.0,65536.0,Language Model,['LoRA'],"['text-generation', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},70.0,Medium,0.7,2025-11-02T05:59:33.566029,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge base_m...,['Multi-turn'] tencent/HunyuanImage-3.0,HunyuanImage-3.0,tencent,2025-09-25T06:28:28+00:00,101407,947,transformers,"['transformers', 'safetensors', 'hunyuan_image_3_moe', 'text-generation', 'text-to-image', 'custom_code', 'arxiv:2509.23951', 'license:other', 'autotrain_compatible', 'region:us']",12,58,HunyuanImage3ForCausalMM,hunyuan_image_3_moe,6987710464.0,12800.0,4096.0,32.0,32.0,133120.0,Language Model,['RLHF'],"['text-generation', 'translation', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Creative Writing']",{},55.0,Hard,0.55,2025-11-02T05:59:33.722563,"license: other license_name: tencent-hunyuan-community license_link: LICENSE pipeline_tag: text-to-image library_name: transformers
[!NOTE] > GGUF uploads with our fixes. More details and [Read o...,"['Function Calling', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-Next-80B-A3B-Instruct,Qwen3-Next-80B-A3B-Instruct,Qwen,2025-09-09T15:40:56+00:00,1663221,845,transformers,"['transformers', 'safetensors', 'qwen3_next', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2404.06654', 'arxiv:2505.09388', 'arxiv:2501.15383', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",11,65,Qwen3NextForCausalLM,qwen3_next,2727084032.0,262144.0,2048.0,16.0,48.0,151936.0,Language Model,['Merged'],"['text-generation', 'question-answering', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T05:59:38.987336,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Instruct/blob/main/LICENSE pipeline_tag: text-generation This model was obtained by uniformly pruning 4...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" internlm/JanusCoder-8B,JanusCoder-8B,internlm,2025-10-27T09:33:54+00:00,139,11,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",11,68,Qwen3ForCausalLM,qwen3,7870087168.0,32768.0,4096.0,32.0,36.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},68.0,Medium,0.68,2025-11-02T05:59:39.634162,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] unsloth/MiniMax-M2,MiniMax-M2,unsloth,2025-10-28T12:05:08+00:00,264,11,transformers,"['transformers', 'safetensors', 'minimax', 'text-generation', 'conversational', 'arxiv:2504.07164', 'arxiv:2509.06501', 'arxiv:2509.13160', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'fp8', 'region:us']",11,69,MiniMaxM2ForCausalLM,minimax,7635861504.0,196608.0,3072.0,48.0,62.0,200064.0,Language Model,['LoRA'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production']",{},91.0,Easy,0.91,2025-11-02T05:59:39.970608,"pipeline_tag: text-generation license: mit library_name: transformers
<...,"['RAG Support', 'Multi-turn']" facebook/MobileLLM-Pro,MobileLLM-Pro,facebook,2025-09-10T18:40:06+00:00,4486,138,transformers,"['transformers', 'safetensors', 'llama4_text', 'text-generation', 'facebook', 'meta', 'pytorch', 'conversational', 'custom_code', 'en', 'base_model:facebook/MobileLLM-Pro-base', 'base_model:finetune:facebook/MobileLLM-Pro-base', 'license:fair-noncommercial-research-license', 'autotrain_compatible', 'region:us']",10,76,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T05:59:44.420226,No README available,[] nineninesix/kani-tts-370m,kani-tts-370m,nineninesix,2025-09-30T07:31:22+00:00,8444,136,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'text-to-speech', 'en', 'de', 'ar', 'zh', 'es', 'ko', 'arxiv:2505.20506', 'base_model:nineninesix/kani-tts-450m-0.2-pt', 'base_model:finetune:nineninesix/kani-tts-450m-0.2-pt', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",10,77,Lfm2ForCausalLM,lfm2,283798528.0,128000.0,1024.0,16.0,16.0,80539.0,Language Model,"['Fine-tuned', 'Specialized']",['conversation'],"['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Legal']",{},75.0,Medium,0.75,2025-11-02T05:59:44.638776,"license: apache-2.0 language: - en - de - ar - zh - es - ko pipeline_tag: text-to-speech library_name: transformers base_model: - nineninesix/kani-tts-450m-0.2-pt

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" LiquidAI/LFM2-8B-A1B,LFM2-8B-A1B,LiquidAI,2025-10-07T13:55:39+00:00,13348,231,transformers,"['transformers', 'safetensors', 'lfm2_moe', 'text-generation', 'liquid', 'lfm2', 'edge', 'moe', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,87,Lfm2MoeForCausalLM,lfm2_moe,1342177280.0,128000.0,2048.0,32.0,24.0,65536.0,Language Model,"['Quantized', 'LoRA']","['text-generation', 'question-answering', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing', 'Healthcare']",{},88.0,Easy,0.88,2025-11-02T05:59:51.806756,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge - moe ...,"['Function Calling', 'RAG Support', 'Fast Inference', 'Multi-turn', 'Safety Aligned']" cerebras/Qwen3-Coder-REAP-25B-A3B,Qwen3-Coder-REAP-25B-A3B,cerebras,2025-10-20T15:40:03+00:00,971,32,transformers,"['transformers', 'qwen3_moe', 'text-generation', 'qwen-coder', 'MOE', 'pruning', 'compression', 'conversational', 'en', 'arxiv:2510.13999', 'base_model:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'base_model:finetune:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,88,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,262144.0,2048.0,32.0,48.0,151936.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'reasoning']","['English', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},85.0,Easy,0.85,2025-11-02T05:59:51.992032,language: - en library_name: transformers tags: - qwen-coder - MOE - pruning - compression license: apache-2.0 name: cerebras/Qwen3-Coder-REAP-25B-A3B description: > This model was obtained by uniform...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" meta-llama/Llama-3.2-1B-Instruct,Llama-3.2-1B-Instruct,meta-llama,2024-09-18T15:12:47+00:00,3843095,1141,transformers,"['transformers', 'safetensors', 'llama', 'text-generation', 'facebook', 'meta', 'pytorch', 'llama-3', 'conversational', 'en', 'de', 'fr', 'it', 'pt', 'hi', 'es', 'th', 'arxiv:2204.05149', 'arxiv:2405.16406', 'license:llama3.2', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,89,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T05:59:53.195984,No README available,[] microsoft/Phi-4-mini-instruct,Phi-4-mini-instruct,microsoft,2025-02-19T01:00:58+00:00,249315,623,transformers,"['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'ar', 'zh', 'cs', 'da', 'nl', 'en', 'fi', 'fr', 'de', 'he', 'hu', 'it', 'ja', 'ko', 'no', 'pl', 'pt', 'ru', 'es', 'sv', 'th', 'tr', 'uk', 'arxiv:2503.01743', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,90,Phi3ForCausalLM,phi3,4238475264.0,131072.0,3072.0,24.0,32.0,200064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Business', 'Healthcare', 'Legal']",{},86.0,Easy,0.86,2025-11-02T05:59:53.379300,language: - multilingual - ar - zh - cs - da - nl - en - fi - fr - de - he - hu - it - ja - ko - 'no' - pl - pt - ru - es - sv - th - tr - uk library_name: transformers license: mit license_link: http...,"['Function Calling', 'RAG Support', 'Long Context', 'Multi-turn', 'Safety Aligned']" Qwen/Qwen3-30B-A3B,Qwen3-30B-A3B,Qwen,2025-04-27T03:43:05+00:00,421099,808,transformers,"['transformers', 'safetensors', 'qwen3_moe', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-30B-A3B-Base', 'base_model:finetune:Qwen/Qwen3-30B-A3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",8,91,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,40960.0,2048.0,32.0,48.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T05:59:55.596033,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-30B-A3B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-30B-A3B-Base - If you want to learn more about how we process your perso...,[] TheBloke/Mistral-7B-Instruct-v0.2-GGUF,Mistral-7B-Instruct-v0.2-GGUF,TheBloke,2023-12-11T22:18:46+00:00,60731,475,transformers,"['transformers', 'gguf', 'mistral', 'finetuned', 'text-generation', 'arxiv:2310.06825', 'base_model:mistralai/Mistral-7B-Instruct-v0.2', 'base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2', 'license:apache-2.0', 'region:us', 'conversational']",6,106,Unknown,mistral,,,,,,,Language Model,"['Fine-tuned', 'Quantized']","['text-generation', 'question-answering', 'summarization', 'conversation']","['English', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Production', 'Creative Writing']",{},63.0,Hard,0.63,2025-11-02T06:00:07.525595,base_model: mistralai/Mistral-7B-Instruct-v0.2 inference: false license: apache-2.0 model_creator: Mistral AI_ model_name: Mistral 7B Instruct v0.2 model_type: mistral pipeline_tag: text-generation pr...,"['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-32B,Qwen3-32B,Qwen,2025-04-27T03:52:59+00:00,1610353,561,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",6,107,Qwen3ForCausalLM,qwen3,20910571520.0,40960.0,5120.0,64.0,64.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},61.0,Hard,0.61,2025-11-02T06:00:07.692824,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-32B/blob/main/LICENSE pipeline_tag: text-generation

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ inference: false fine-tuning: false langu...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" PokeeAI/pokee_research_7b,pokee_research_7b,PokeeAI,2025-10-17T20:38:37+00:00,5574,95,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'agent', 'deepresearch', 'llm', 'rl', 'reinforcementlearning', 'conversational', 'en', 'dataset:miromind-ai/MiroRL-GenQA', 'arxiv:2510.15862', 'base_model:Qwen/Qwen2.5-7B-Instruct', 'base_model:finetune:Qwen/Qwen2.5-7B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",24,147,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged', 'Specialized']","['text-generation', 'question-answering', 'summarization', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education', 'Healthcare', 'Legal', 'Finance']",{},78.0,Medium,0.78,2025-11-02T06:00:30.941112,base_model: - Qwen/Qwen2.5-7B-Instruct datasets: - miromind-ai/MiroRL-GenQA language: - en license: apache-2.0 tags: - agent - deepresearch - llm - rl - reinforcementlearning pipeline_tag: text-genera...,"['RAG Support', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" ibm-granite/granite-4.0-350m-base,granite-4.0-350m-base,ibm-granite,2025-10-07T20:25:24+00:00,495,23,transformers,"['transformers', 'safetensors', 'granitemoehybrid', 'text-generation', 'language', 'granite-4.0', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",23,148,GraniteMoeHybridForCausalLM,granitemoehybrid,455081984.0,32768.0,1024.0,16.0,28.0,100352.0,Language Model,"['LoRA', 'Specialized']","['text-generation', 'question-answering', 'text-classification', 'summarization']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Education']",{},86.0,Easy,0.86,2025-11-02T06:00:31.109072,license: apache-2.0 library_name: transformers tags: - language - granite-4.0 **Model Summary:** Granite-4.0-350M-Base is a lightweight decoder-only language model designed for scenarios where efficie...,"['Long Context', 'Fast Inference', 'Safety Aligned']" internlm/JanusCoder-14B,JanusCoder-14B,internlm,2025-10-27T09:34:49+00:00,165,23,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",23,149,Qwen3ForCausalLM,qwen3,13360824320.0,32768.0,5120.0,40.0,40.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},58.0,Hard,0.58,2025-11-02T06:00:31.304356,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] ByteDance/Ouro-2.6B,Ouro-2.6B,ByteDance,2025-10-28T22:19:46+00:00,227,21,transformers,"['transformers', 'safetensors', 'ouro', 'text-generation', 'looped-language-model', 'reasoning', 'recurrent-depth', 'conversational', 'custom_code', 'arxiv:2510.25741', 'license:apache-2.0', 'autotrain_compatible', 'region:us']",21,150,OuroForCausalLM,ouro,2516582400.0,65536.0,2048.0,16.0,48.0,49152.0,Language Model,[],"['text-generation', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education']",{},78.0,Medium,0.78,2025-11-02T06:00:31.505272,library_name: transformers license: apache-2.0 pipeline_tag: text-generation tags: - looped-language-model - reasoning - recurrent-depth ![Ouro Logo](assets/logo.png) [📚 Paper (Hugging Face)](https://...,"['RAG Support', 'Long Context']" HuggingFaceTB/SmolLM3-3B,SmolLM3-3B,HuggingFaceTB,2025-07-08T10:11:45+00:00,56521,768,transformers,"['transformers', 'safetensors', 'smollm3', 'text-generation', 'conversational', 'en', 'fr', 'es', 'it', 'pt', 'zh', 'ar', 'ru', 'base_model:HuggingFaceTB/SmolLM3-3B-Base', 'base_model:finetune:HuggingFaceTB/SmolLM3-3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",20,151,SmolLM3ForCausalLM,smollm3,2074607616.0,65536.0,2048.0,16.0,36.0,128256.0,Language Model,['Quantized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},83.0,Medium,0.83,2025-11-02T06:00:33.760902,library_name: transformers license: apache-2.0 language: - en - fr - es - it - pt - zh - ar - ru base_model: - HuggingFaceTB/SmolLM3-3B-Base ![image/png](https://cdn-uploads.huggingface.co/production/...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient']" Qwen/Qwen3-0.6B,Qwen3-0.6B,Qwen,2025-04-27T03:40:08+00:00,7327000,752,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-0.6B-Base', 'base_model:finetune:Qwen/Qwen3-0.6B-Base', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",19,152,Qwen3ForCausalLM,qwen3,507904000.0,40960.0,1024.0,16.0,28.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},78.0,Medium,0.78,2025-11-02T06:00:33.951165,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-0.6B-Base This model was obtained by uniformly pruning 25...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" Qwen/Qwen2.5-7B-Instruct,Qwen2.5-7B-Instruct,Qwen,2024-09-16T11:55:40+00:00,8169288,849,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-7B', 'base_model:finetune:Qwen/Qwen2.5-7B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",17,156,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},86.0,Easy,0.86,2025-11-02T06:00:34.643187,license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE language: - en pipeline_tag: text-generation base_model: Qwen/Qwen2.5-7B tags: - chat library_name: ...,"['Long Context', 'Fast Inference', 'Safety Aligned']" deepseek-ai/DeepSeek-R1,DeepSeek-R1,deepseek-ai,2025-01-20T03:46:07+00:00,425526,12820,transformers,"['transformers', 'safetensors', 'deepseek_v3', 'text-generation', 'conversational', 'custom_code', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'fp8', 'region:us']",17,157,DeepseekV3ForCausalLM,deepseek_v3,38537003008.0,163840.0,7168.0,128.0,61.0,129280.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged']","['text-generation', 'question-answering', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business']",{},48.0,Hard,0.48,2025-11-02T06:00:34.828024,"license: mit library_name: transformers
- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" microsoft/VibeVoice-1.5B,VibeVoice-1.5B,microsoft,2025-08-25T13:46:48+00:00,223609,1943,transformers,"['transformers', 'safetensors', 'vibevoice', 'text-generation', 'Podcast', 'text-to-speech', 'en', 'zh', 'arxiv:2508.19205', 'arxiv:2412.08635', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,176,VibeVoiceForConditionalGeneration,vibevoice,,,,,,,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business', 'Legal']",{},53.0,Hard,0.53,2025-11-02T06:00:52.407117,"language: - en - zh license: mit pipeline_tag: text-to-speech tags: - Podcast library_name: transformers VibeVoice is a novel framework designed for generating expressive, long-form, multi-speaker con...","['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" LiquidAI/LFM2-350M-Extract,LFM2-350M-Extract,LiquidAI,2025-09-03T17:13:53+00:00,5852,61,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'liquid', 'edge', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'base_model:LiquidAI/LFM2-350M', 'base_model:finetune:LiquidAI/LFM2-350M', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,177,Lfm2ForCausalLM,lfm2,268435456.0,128000.0,1024.0,16.0,16.0,65536.0,Language Model,['LoRA'],"['text-generation', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},70.0,Medium,0.7,2025-11-02T06:00:52.573790,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge base_m...,['Multi-turn'] tencent/HunyuanImage-3.0,HunyuanImage-3.0,tencent,2025-09-25T06:28:28+00:00,101407,947,transformers,"['transformers', 'safetensors', 'hunyuan_image_3_moe', 'text-generation', 'text-to-image', 'custom_code', 'arxiv:2509.23951', 'license:other', 'autotrain_compatible', 'region:us']",12,178,HunyuanImage3ForCausalMM,hunyuan_image_3_moe,6987710464.0,12800.0,4096.0,32.0,32.0,133120.0,Language Model,['RLHF'],"['text-generation', 'translation', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Creative Writing']",{},55.0,Hard,0.55,2025-11-02T06:00:53.047824,"license: other license_name: tencent-hunyuan-community license_link: LICENSE pipeline_tag: text-to-image library_name: transformers
[!NOTE] > GGUF uploads with our fixes. More details and [Read o...,"['Function Calling', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-Next-80B-A3B-Instruct,Qwen3-Next-80B-A3B-Instruct,Qwen,2025-09-09T15:40:56+00:00,1663221,845,transformers,"['transformers', 'safetensors', 'qwen3_next', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2404.06654', 'arxiv:2505.09388', 'arxiv:2501.15383', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",11,185,Qwen3NextForCausalLM,qwen3_next,2727084032.0,262144.0,2048.0,16.0,48.0,151936.0,Language Model,['Merged'],"['text-generation', 'question-answering', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:00:59.666276,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Instruct/blob/main/LICENSE pipeline_tag: text-generation This model was obtained by uniformly pruning 4...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" internlm/JanusCoder-8B,JanusCoder-8B,internlm,2025-10-27T09:33:54+00:00,139,11,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",11,188,Qwen3ForCausalLM,qwen3,7870087168.0,32768.0,4096.0,32.0,36.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},68.0,Medium,0.68,2025-11-02T06:01:00.137897,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] unsloth/MiniMax-M2,MiniMax-M2,unsloth,2025-10-28T12:05:08+00:00,264,11,transformers,"['transformers', 'safetensors', 'minimax', 'text-generation', 'conversational', 'arxiv:2504.07164', 'arxiv:2509.06501', 'arxiv:2509.13160', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'fp8', 'region:us']",11,189,MiniMaxM2ForCausalLM,minimax,7635861504.0,196608.0,3072.0,48.0,62.0,200064.0,Language Model,['LoRA'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production']",{},91.0,Easy,0.91,2025-11-02T06:01:00.311076,"pipeline_tag: text-generation license: mit library_name: transformers
<...,"['RAG Support', 'Multi-turn']" facebook/MobileLLM-Pro,MobileLLM-Pro,facebook,2025-09-10T18:40:06+00:00,4486,138,transformers,"['transformers', 'safetensors', 'llama4_text', 'text-generation', 'facebook', 'meta', 'pytorch', 'conversational', 'custom_code', 'en', 'base_model:facebook/MobileLLM-Pro-base', 'base_model:finetune:facebook/MobileLLM-Pro-base', 'license:fair-noncommercial-research-license', 'autotrain_compatible', 'region:us']",10,196,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:01:04.530889,No README available,[] nineninesix/kani-tts-370m,kani-tts-370m,nineninesix,2025-09-30T07:31:22+00:00,8444,136,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'text-to-speech', 'en', 'de', 'ar', 'zh', 'es', 'ko', 'arxiv:2505.20506', 'base_model:nineninesix/kani-tts-450m-0.2-pt', 'base_model:finetune:nineninesix/kani-tts-450m-0.2-pt', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",10,197,Lfm2ForCausalLM,lfm2,283798528.0,128000.0,1024.0,16.0,16.0,80539.0,Language Model,"['Fine-tuned', 'Specialized']",['conversation'],"['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Legal']",{},75.0,Medium,0.75,2025-11-02T06:01:04.690562,"license: apache-2.0 language: - en - de - ar - zh - es - ko pipeline_tag: text-to-speech library_name: transformers base_model: - nineninesix/kani-tts-450m-0.2-pt

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" LiquidAI/LFM2-8B-A1B,LFM2-8B-A1B,LiquidAI,2025-10-07T13:55:39+00:00,13348,231,transformers,"['transformers', 'safetensors', 'lfm2_moe', 'text-generation', 'liquid', 'lfm2', 'edge', 'moe', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,207,Lfm2MoeForCausalLM,lfm2_moe,1342177280.0,128000.0,2048.0,32.0,24.0,65536.0,Language Model,"['Quantized', 'LoRA']","['text-generation', 'question-answering', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing', 'Healthcare']",{},88.0,Easy,0.88,2025-11-02T06:01:12.046492,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge - moe ...,"['Function Calling', 'RAG Support', 'Fast Inference', 'Multi-turn', 'Safety Aligned']" cerebras/Qwen3-Coder-REAP-25B-A3B,Qwen3-Coder-REAP-25B-A3B,cerebras,2025-10-20T15:40:03+00:00,971,32,transformers,"['transformers', 'qwen3_moe', 'text-generation', 'qwen-coder', 'MOE', 'pruning', 'compression', 'conversational', 'en', 'arxiv:2510.13999', 'base_model:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'base_model:finetune:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,208,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,262144.0,2048.0,32.0,48.0,151936.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'reasoning']","['English', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},85.0,Easy,0.85,2025-11-02T06:01:12.522267,language: - en library_name: transformers tags: - qwen-coder - MOE - pruning - compression license: apache-2.0 name: cerebras/Qwen3-Coder-REAP-25B-A3B description: > This model was obtained by uniform...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" meta-llama/Llama-3.2-1B-Instruct,Llama-3.2-1B-Instruct,meta-llama,2024-09-18T15:12:47+00:00,3843095,1141,transformers,"['transformers', 'safetensors', 'llama', 'text-generation', 'facebook', 'meta', 'pytorch', 'llama-3', 'conversational', 'en', 'de', 'fr', 'it', 'pt', 'hi', 'es', 'th', 'arxiv:2204.05149', 'arxiv:2405.16406', 'license:llama3.2', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,209,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:01:13.704023,No README available,[] microsoft/Phi-4-mini-instruct,Phi-4-mini-instruct,microsoft,2025-02-19T01:00:58+00:00,249315,623,transformers,"['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'ar', 'zh', 'cs', 'da', 'nl', 'en', 'fi', 'fr', 'de', 'he', 'hu', 'it', 'ja', 'ko', 'no', 'pl', 'pt', 'ru', 'es', 'sv', 'th', 'tr', 'uk', 'arxiv:2503.01743', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,210,Phi3ForCausalLM,phi3,4238475264.0,131072.0,3072.0,24.0,32.0,200064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Business', 'Healthcare', 'Legal']",{},86.0,Easy,0.86,2025-11-02T06:01:13.889560,language: - multilingual - ar - zh - cs - da - nl - en - fi - fr - de - he - hu - it - ja - ko - 'no' - pl - pt - ru - es - sv - th - tr - uk library_name: transformers license: mit license_link: http...,"['Function Calling', 'RAG Support', 'Long Context', 'Multi-turn', 'Safety Aligned']" Qwen/Qwen3-30B-A3B,Qwen3-30B-A3B,Qwen,2025-04-27T03:43:05+00:00,421099,808,transformers,"['transformers', 'safetensors', 'qwen3_moe', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-30B-A3B-Base', 'base_model:finetune:Qwen/Qwen3-30B-A3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",8,211,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,40960.0,2048.0,32.0,48.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:01:16.176834,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-30B-A3B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-30B-A3B-Base - If you want to learn more about how we process your perso...,[] TheBloke/Mistral-7B-Instruct-v0.2-GGUF,Mistral-7B-Instruct-v0.2-GGUF,TheBloke,2023-12-11T22:18:46+00:00,60731,475,transformers,"['transformers', 'gguf', 'mistral', 'finetuned', 'text-generation', 'arxiv:2310.06825', 'base_model:mistralai/Mistral-7B-Instruct-v0.2', 'base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2', 'license:apache-2.0', 'region:us', 'conversational']",6,226,Unknown,mistral,,,,,,,Language Model,"['Fine-tuned', 'Quantized']","['text-generation', 'question-answering', 'summarization', 'conversation']","['English', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Production', 'Creative Writing']",{},63.0,Hard,0.63,2025-11-02T06:01:27.594162,base_model: mistralai/Mistral-7B-Instruct-v0.2 inference: false license: apache-2.0 model_creator: Mistral AI_ model_name: Mistral 7B Instruct v0.2 model_type: mistral pipeline_tag: text-generation pr...,"['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-32B,Qwen3-32B,Qwen,2025-04-27T03:52:59+00:00,1610353,561,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",6,227,Qwen3ForCausalLM,qwen3,20910571520.0,40960.0,5120.0,64.0,64.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},61.0,Hard,0.61,2025-11-02T06:01:27.791859,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-32B/blob/main/LICENSE pipeline_tag: text-generation

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ inference: false fine-tuning: false langu...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" PokeeAI/pokee_research_7b,pokee_research_7b,PokeeAI,2025-10-17T20:38:37+00:00,5574,95,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'agent', 'deepresearch', 'llm', 'rl', 'reinforcementlearning', 'conversational', 'en', 'dataset:miromind-ai/MiroRL-GenQA', 'arxiv:2510.15862', 'base_model:Qwen/Qwen2.5-7B-Instruct', 'base_model:finetune:Qwen/Qwen2.5-7B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",24,267,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged', 'Specialized']","['text-generation', 'question-answering', 'summarization', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education', 'Healthcare', 'Legal', 'Finance']",{},78.0,Medium,0.78,2025-11-02T06:01:50.783799,base_model: - Qwen/Qwen2.5-7B-Instruct datasets: - miromind-ai/MiroRL-GenQA language: - en license: apache-2.0 tags: - agent - deepresearch - llm - rl - reinforcementlearning pipeline_tag: text-genera...,"['RAG Support', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" ibm-granite/granite-4.0-350m-base,granite-4.0-350m-base,ibm-granite,2025-10-07T20:25:24+00:00,495,23,transformers,"['transformers', 'safetensors', 'granitemoehybrid', 'text-generation', 'language', 'granite-4.0', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",23,268,GraniteMoeHybridForCausalLM,granitemoehybrid,455081984.0,32768.0,1024.0,16.0,28.0,100352.0,Language Model,"['LoRA', 'Specialized']","['text-generation', 'question-answering', 'text-classification', 'summarization']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Education']",{},86.0,Easy,0.86,2025-11-02T06:01:51.839468,license: apache-2.0 library_name: transformers tags: - language - granite-4.0 **Model Summary:** Granite-4.0-350M-Base is a lightweight decoder-only language model designed for scenarios where efficie...,"['Long Context', 'Fast Inference', 'Safety Aligned']" internlm/JanusCoder-14B,JanusCoder-14B,internlm,2025-10-27T09:34:49+00:00,165,23,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",23,269,Qwen3ForCausalLM,qwen3,13360824320.0,32768.0,5120.0,40.0,40.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},58.0,Hard,0.58,2025-11-02T06:01:52.083284,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] ByteDance/Ouro-2.6B,Ouro-2.6B,ByteDance,2025-10-28T22:19:46+00:00,227,21,transformers,"['transformers', 'safetensors', 'ouro', 'text-generation', 'looped-language-model', 'reasoning', 'recurrent-depth', 'conversational', 'custom_code', 'arxiv:2510.25741', 'license:apache-2.0', 'autotrain_compatible', 'region:us']",21,270,OuroForCausalLM,ouro,2516582400.0,65536.0,2048.0,16.0,48.0,49152.0,Language Model,[],"['text-generation', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education']",{},78.0,Medium,0.78,2025-11-02T06:01:52.242806,library_name: transformers license: apache-2.0 pipeline_tag: text-generation tags: - looped-language-model - reasoning - recurrent-depth ![Ouro Logo](assets/logo.png) [📚 Paper (Hugging Face)](https://...,"['RAG Support', 'Long Context']" HuggingFaceTB/SmolLM3-3B,SmolLM3-3B,HuggingFaceTB,2025-07-08T10:11:45+00:00,56521,768,transformers,"['transformers', 'safetensors', 'smollm3', 'text-generation', 'conversational', 'en', 'fr', 'es', 'it', 'pt', 'zh', 'ar', 'ru', 'base_model:HuggingFaceTB/SmolLM3-3B-Base', 'base_model:finetune:HuggingFaceTB/SmolLM3-3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",20,271,SmolLM3ForCausalLM,smollm3,2074607616.0,65536.0,2048.0,16.0,36.0,128256.0,Language Model,['Quantized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},83.0,Medium,0.83,2025-11-02T06:01:54.440328,library_name: transformers license: apache-2.0 language: - en - fr - es - it - pt - zh - ar - ru base_model: - HuggingFaceTB/SmolLM3-3B-Base ![image/png](https://cdn-uploads.huggingface.co/production/...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient']" Qwen/Qwen3-0.6B,Qwen3-0.6B,Qwen,2025-04-27T03:40:08+00:00,7327000,752,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-0.6B-Base', 'base_model:finetune:Qwen/Qwen3-0.6B-Base', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",19,272,Qwen3ForCausalLM,qwen3,507904000.0,40960.0,1024.0,16.0,28.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},78.0,Medium,0.78,2025-11-02T06:01:54.580882,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-0.6B-Base This model was obtained by uniformly pruning 25...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" Qwen/Qwen2.5-7B-Instruct,Qwen2.5-7B-Instruct,Qwen,2024-09-16T11:55:40+00:00,8169288,849,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-7B', 'base_model:finetune:Qwen/Qwen2.5-7B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",17,276,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},86.0,Easy,0.86,2025-11-02T06:01:55.320492,license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE language: - en pipeline_tag: text-generation base_model: Qwen/Qwen2.5-7B tags: - chat library_name: ...,"['Long Context', 'Fast Inference', 'Safety Aligned']" deepseek-ai/DeepSeek-R1,DeepSeek-R1,deepseek-ai,2025-01-20T03:46:07+00:00,425526,12820,transformers,"['transformers', 'safetensors', 'deepseek_v3', 'text-generation', 'conversational', 'custom_code', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'fp8', 'region:us']",17,277,DeepseekV3ForCausalLM,deepseek_v3,38537003008.0,163840.0,7168.0,128.0,61.0,129280.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged']","['text-generation', 'question-answering', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business']",{},48.0,Hard,0.48,2025-11-02T06:01:55.499319,"license: mit library_name: transformers
- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" microsoft/VibeVoice-1.5B,VibeVoice-1.5B,microsoft,2025-08-25T13:46:48+00:00,223609,1943,transformers,"['transformers', 'safetensors', 'vibevoice', 'text-generation', 'Podcast', 'text-to-speech', 'en', 'zh', 'arxiv:2508.19205', 'arxiv:2412.08635', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,296,VibeVoiceForConditionalGeneration,vibevoice,,,,,,,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business', 'Legal']",{},53.0,Hard,0.53,2025-11-02T06:02:09.310733,"language: - en - zh license: mit pipeline_tag: text-to-speech tags: - Podcast library_name: transformers VibeVoice is a novel framework designed for generating expressive, long-form, multi-speaker con...","['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" LiquidAI/LFM2-350M-Extract,LFM2-350M-Extract,LiquidAI,2025-09-03T17:13:53+00:00,5852,61,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'liquid', 'edge', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'base_model:LiquidAI/LFM2-350M', 'base_model:finetune:LiquidAI/LFM2-350M', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,297,Lfm2ForCausalLM,lfm2,268435456.0,128000.0,1024.0,16.0,16.0,65536.0,Language Model,['LoRA'],"['text-generation', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},70.0,Medium,0.7,2025-11-02T06:02:09.475790,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge base_m...,['Multi-turn'] tencent/HunyuanImage-3.0,HunyuanImage-3.0,tencent,2025-09-25T06:28:28+00:00,101407,947,transformers,"['transformers', 'safetensors', 'hunyuan_image_3_moe', 'text-generation', 'text-to-image', 'custom_code', 'arxiv:2509.23951', 'license:other', 'autotrain_compatible', 'region:us']",12,298,HunyuanImage3ForCausalMM,hunyuan_image_3_moe,6987710464.0,12800.0,4096.0,32.0,32.0,133120.0,Language Model,['RLHF'],"['text-generation', 'translation', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Creative Writing']",{},55.0,Hard,0.55,2025-11-02T06:02:09.657653,"license: other license_name: tencent-hunyuan-community license_link: LICENSE pipeline_tag: text-to-image library_name: transformers
[!NOTE] > GGUF uploads with our fixes. More details and [Read o...,"['Function Calling', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-Next-80B-A3B-Instruct,Qwen3-Next-80B-A3B-Instruct,Qwen,2025-09-09T15:40:56+00:00,1663221,845,transformers,"['transformers', 'safetensors', 'qwen3_next', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2404.06654', 'arxiv:2505.09388', 'arxiv:2501.15383', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",11,305,Qwen3NextForCausalLM,qwen3_next,2727084032.0,262144.0,2048.0,16.0,48.0,151936.0,Language Model,['Merged'],"['text-generation', 'question-answering', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:02:17.030640,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Instruct/blob/main/LICENSE pipeline_tag: text-generation This model was obtained by uniformly pruning 4...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" internlm/JanusCoder-8B,JanusCoder-8B,internlm,2025-10-27T09:33:54+00:00,139,11,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",11,308,Qwen3ForCausalLM,qwen3,7870087168.0,32768.0,4096.0,32.0,36.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},68.0,Medium,0.68,2025-11-02T06:02:17.742509,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] unsloth/MiniMax-M2,MiniMax-M2,unsloth,2025-10-28T12:05:08+00:00,264,11,transformers,"['transformers', 'safetensors', 'minimax', 'text-generation', 'conversational', 'arxiv:2504.07164', 'arxiv:2509.06501', 'arxiv:2509.13160', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'fp8', 'region:us']",11,309,MiniMaxM2ForCausalLM,minimax,7635861504.0,196608.0,3072.0,48.0,62.0,200064.0,Language Model,['LoRA'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production']",{},91.0,Easy,0.91,2025-11-02T06:02:17.948843,"pipeline_tag: text-generation license: mit library_name: transformers
<...,"['RAG Support', 'Multi-turn']" facebook/MobileLLM-Pro,MobileLLM-Pro,facebook,2025-09-10T18:40:06+00:00,4486,138,transformers,"['transformers', 'safetensors', 'llama4_text', 'text-generation', 'facebook', 'meta', 'pytorch', 'conversational', 'custom_code', 'en', 'base_model:facebook/MobileLLM-Pro-base', 'base_model:finetune:facebook/MobileLLM-Pro-base', 'license:fair-noncommercial-research-license', 'autotrain_compatible', 'region:us']",10,316,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:02:22.274794,No README available,[] nineninesix/kani-tts-370m,kani-tts-370m,nineninesix,2025-09-30T07:31:22+00:00,8444,136,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'text-to-speech', 'en', 'de', 'ar', 'zh', 'es', 'ko', 'arxiv:2505.20506', 'base_model:nineninesix/kani-tts-450m-0.2-pt', 'base_model:finetune:nineninesix/kani-tts-450m-0.2-pt', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",10,317,Lfm2ForCausalLM,lfm2,283798528.0,128000.0,1024.0,16.0,16.0,80539.0,Language Model,"['Fine-tuned', 'Specialized']",['conversation'],"['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Legal']",{},75.0,Medium,0.75,2025-11-02T06:02:22.404526,"license: apache-2.0 language: - en - de - ar - zh - es - ko pipeline_tag: text-to-speech library_name: transformers base_model: - nineninesix/kani-tts-450m-0.2-pt

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" LiquidAI/LFM2-8B-A1B,LFM2-8B-A1B,LiquidAI,2025-10-07T13:55:39+00:00,13348,231,transformers,"['transformers', 'safetensors', 'lfm2_moe', 'text-generation', 'liquid', 'lfm2', 'edge', 'moe', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,327,Lfm2MoeForCausalLM,lfm2_moe,1342177280.0,128000.0,2048.0,32.0,24.0,65536.0,Language Model,"['Quantized', 'LoRA']","['text-generation', 'question-answering', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing', 'Healthcare']",{},88.0,Easy,0.88,2025-11-02T06:02:31.730781,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge - moe ...,"['Function Calling', 'RAG Support', 'Fast Inference', 'Multi-turn', 'Safety Aligned']" cerebras/Qwen3-Coder-REAP-25B-A3B,Qwen3-Coder-REAP-25B-A3B,cerebras,2025-10-20T15:40:03+00:00,971,32,transformers,"['transformers', 'qwen3_moe', 'text-generation', 'qwen-coder', 'MOE', 'pruning', 'compression', 'conversational', 'en', 'arxiv:2510.13999', 'base_model:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'base_model:finetune:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,328,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,262144.0,2048.0,32.0,48.0,151936.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'reasoning']","['English', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},85.0,Easy,0.85,2025-11-02T06:02:32.037782,language: - en library_name: transformers tags: - qwen-coder - MOE - pruning - compression license: apache-2.0 name: cerebras/Qwen3-Coder-REAP-25B-A3B description: > This model was obtained by uniform...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" meta-llama/Llama-3.2-1B-Instruct,Llama-3.2-1B-Instruct,meta-llama,2024-09-18T15:12:47+00:00,3843095,1141,transformers,"['transformers', 'safetensors', 'llama', 'text-generation', 'facebook', 'meta', 'pytorch', 'llama-3', 'conversational', 'en', 'de', 'fr', 'it', 'pt', 'hi', 'es', 'th', 'arxiv:2204.05149', 'arxiv:2405.16406', 'license:llama3.2', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,329,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:02:33.286959,No README available,[] microsoft/Phi-4-mini-instruct,Phi-4-mini-instruct,microsoft,2025-02-19T01:00:58+00:00,249315,623,transformers,"['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'ar', 'zh', 'cs', 'da', 'nl', 'en', 'fi', 'fr', 'de', 'he', 'hu', 'it', 'ja', 'ko', 'no', 'pl', 'pt', 'ru', 'es', 'sv', 'th', 'tr', 'uk', 'arxiv:2503.01743', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,330,Phi3ForCausalLM,phi3,4238475264.0,131072.0,3072.0,24.0,32.0,200064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Business', 'Healthcare', 'Legal']",{},86.0,Easy,0.86,2025-11-02T06:02:33.514613,language: - multilingual - ar - zh - cs - da - nl - en - fi - fr - de - he - hu - it - ja - ko - 'no' - pl - pt - ru - es - sv - th - tr - uk library_name: transformers license: mit license_link: http...,"['Function Calling', 'RAG Support', 'Long Context', 'Multi-turn', 'Safety Aligned']" Qwen/Qwen3-30B-A3B,Qwen3-30B-A3B,Qwen,2025-04-27T03:43:05+00:00,421099,808,transformers,"['transformers', 'safetensors', 'qwen3_moe', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-30B-A3B-Base', 'base_model:finetune:Qwen/Qwen3-30B-A3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",8,331,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,40960.0,2048.0,32.0,48.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:02:35.665325,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-30B-A3B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-30B-A3B-Base - If you want to learn more about how we process your perso...,[] TheBloke/Mistral-7B-Instruct-v0.2-GGUF,Mistral-7B-Instruct-v0.2-GGUF,TheBloke,2023-12-11T22:18:46+00:00,60731,475,transformers,"['transformers', 'gguf', 'mistral', 'finetuned', 'text-generation', 'arxiv:2310.06825', 'base_model:mistralai/Mistral-7B-Instruct-v0.2', 'base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2', 'license:apache-2.0', 'region:us', 'conversational']",6,346,Unknown,mistral,,,,,,,Language Model,"['Fine-tuned', 'Quantized']","['text-generation', 'question-answering', 'summarization', 'conversation']","['English', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Production', 'Creative Writing']",{},63.0,Hard,0.63,2025-11-02T06:02:49.275008,base_model: mistralai/Mistral-7B-Instruct-v0.2 inference: false license: apache-2.0 model_creator: Mistral AI_ model_name: Mistral 7B Instruct v0.2 model_type: mistral pipeline_tag: text-generation pr...,"['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-32B,Qwen3-32B,Qwen,2025-04-27T03:52:59+00:00,1610353,561,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",6,347,Qwen3ForCausalLM,qwen3,20910571520.0,40960.0,5120.0,64.0,64.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},61.0,Hard,0.61,2025-11-02T06:02:49.430634,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-32B/blob/main/LICENSE pipeline_tag: text-generation

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ inference: false fine-tuning: false langu...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" PokeeAI/pokee_research_7b,pokee_research_7b,PokeeAI,2025-10-17T20:38:37+00:00,5574,95,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'agent', 'deepresearch', 'llm', 'rl', 'reinforcementlearning', 'conversational', 'en', 'dataset:miromind-ai/MiroRL-GenQA', 'arxiv:2510.15862', 'base_model:Qwen/Qwen2.5-7B-Instruct', 'base_model:finetune:Qwen/Qwen2.5-7B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",24,387,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged', 'Specialized']","['text-generation', 'question-answering', 'summarization', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education', 'Healthcare', 'Legal', 'Finance']",{},78.0,Medium,0.78,2025-11-02T06:03:11.532085,base_model: - Qwen/Qwen2.5-7B-Instruct datasets: - miromind-ai/MiroRL-GenQA language: - en license: apache-2.0 tags: - agent - deepresearch - llm - rl - reinforcementlearning pipeline_tag: text-genera...,"['RAG Support', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" ibm-granite/granite-4.0-350m-base,granite-4.0-350m-base,ibm-granite,2025-10-07T20:25:24+00:00,495,23,transformers,"['transformers', 'safetensors', 'granitemoehybrid', 'text-generation', 'language', 'granite-4.0', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",23,388,GraniteMoeHybridForCausalLM,granitemoehybrid,455081984.0,32768.0,1024.0,16.0,28.0,100352.0,Language Model,"['LoRA', 'Specialized']","['text-generation', 'question-answering', 'text-classification', 'summarization']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Education']",{},86.0,Easy,0.86,2025-11-02T06:03:11.713311,license: apache-2.0 library_name: transformers tags: - language - granite-4.0 **Model Summary:** Granite-4.0-350M-Base is a lightweight decoder-only language model designed for scenarios where efficie...,"['Long Context', 'Fast Inference', 'Safety Aligned']" internlm/JanusCoder-14B,JanusCoder-14B,internlm,2025-10-27T09:34:49+00:00,165,23,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",23,389,Qwen3ForCausalLM,qwen3,13360824320.0,32768.0,5120.0,40.0,40.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},58.0,Hard,0.58,2025-11-02T06:03:11.847754,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] ByteDance/Ouro-2.6B,Ouro-2.6B,ByteDance,2025-10-28T22:19:46+00:00,227,21,transformers,"['transformers', 'safetensors', 'ouro', 'text-generation', 'looped-language-model', 'reasoning', 'recurrent-depth', 'conversational', 'custom_code', 'arxiv:2510.25741', 'license:apache-2.0', 'autotrain_compatible', 'region:us']",21,390,OuroForCausalLM,ouro,2516582400.0,65536.0,2048.0,16.0,48.0,49152.0,Language Model,[],"['text-generation', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education']",{},78.0,Medium,0.78,2025-11-02T06:03:12.171111,library_name: transformers license: apache-2.0 pipeline_tag: text-generation tags: - looped-language-model - reasoning - recurrent-depth ![Ouro Logo](assets/logo.png) [📚 Paper (Hugging Face)](https://...,"['RAG Support', 'Long Context']" HuggingFaceTB/SmolLM3-3B,SmolLM3-3B,HuggingFaceTB,2025-07-08T10:11:45+00:00,56521,768,transformers,"['transformers', 'safetensors', 'smollm3', 'text-generation', 'conversational', 'en', 'fr', 'es', 'it', 'pt', 'zh', 'ar', 'ru', 'base_model:HuggingFaceTB/SmolLM3-3B-Base', 'base_model:finetune:HuggingFaceTB/SmolLM3-3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",20,391,SmolLM3ForCausalLM,smollm3,2074607616.0,65536.0,2048.0,16.0,36.0,128256.0,Language Model,['Quantized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},83.0,Medium,0.83,2025-11-02T06:03:14.683845,library_name: transformers license: apache-2.0 language: - en - fr - es - it - pt - zh - ar - ru base_model: - HuggingFaceTB/SmolLM3-3B-Base ![image/png](https://cdn-uploads.huggingface.co/production/...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient']" Qwen/Qwen3-0.6B,Qwen3-0.6B,Qwen,2025-04-27T03:40:08+00:00,7327000,752,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-0.6B-Base', 'base_model:finetune:Qwen/Qwen3-0.6B-Base', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",19,392,Qwen3ForCausalLM,qwen3,507904000.0,40960.0,1024.0,16.0,28.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},78.0,Medium,0.78,2025-11-02T06:03:14.825492,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-0.6B-Base This model was obtained by uniformly pruning 25...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" Qwen/Qwen2.5-7B-Instruct,Qwen2.5-7B-Instruct,Qwen,2024-09-16T11:55:40+00:00,8169288,849,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-7B', 'base_model:finetune:Qwen/Qwen2.5-7B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",17,396,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},86.0,Easy,0.86,2025-11-02T06:03:15.566097,license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE language: - en pipeline_tag: text-generation base_model: Qwen/Qwen2.5-7B tags: - chat library_name: ...,"['Long Context', 'Fast Inference', 'Safety Aligned']" deepseek-ai/DeepSeek-R1,DeepSeek-R1,deepseek-ai,2025-01-20T03:46:07+00:00,425526,12820,transformers,"['transformers', 'safetensors', 'deepseek_v3', 'text-generation', 'conversational', 'custom_code', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'fp8', 'region:us']",17,397,DeepseekV3ForCausalLM,deepseek_v3,38537003008.0,163840.0,7168.0,128.0,61.0,129280.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged']","['text-generation', 'question-answering', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business']",{},48.0,Hard,0.48,2025-11-02T06:03:15.873743,"license: mit library_name: transformers
- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" microsoft/VibeVoice-1.5B,VibeVoice-1.5B,microsoft,2025-08-25T13:46:48+00:00,223609,1943,transformers,"['transformers', 'safetensors', 'vibevoice', 'text-generation', 'Podcast', 'text-to-speech', 'en', 'zh', 'arxiv:2508.19205', 'arxiv:2412.08635', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,416,VibeVoiceForConditionalGeneration,vibevoice,,,,,,,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business', 'Legal']",{},53.0,Hard,0.53,2025-11-02T06:03:30.555314,"language: - en - zh license: mit pipeline_tag: text-to-speech tags: - Podcast library_name: transformers VibeVoice is a novel framework designed for generating expressive, long-form, multi-speaker con...","['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" LiquidAI/LFM2-350M-Extract,LFM2-350M-Extract,LiquidAI,2025-09-03T17:13:53+00:00,5852,61,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'liquid', 'edge', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'base_model:LiquidAI/LFM2-350M', 'base_model:finetune:LiquidAI/LFM2-350M', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,417,Lfm2ForCausalLM,lfm2,268435456.0,128000.0,1024.0,16.0,16.0,65536.0,Language Model,['LoRA'],"['text-generation', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},70.0,Medium,0.7,2025-11-02T06:03:30.714878,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge base_m...,['Multi-turn'] tencent/HunyuanImage-3.0,HunyuanImage-3.0,tencent,2025-09-25T06:28:28+00:00,101407,947,transformers,"['transformers', 'safetensors', 'hunyuan_image_3_moe', 'text-generation', 'text-to-image', 'custom_code', 'arxiv:2509.23951', 'license:other', 'autotrain_compatible', 'region:us']",12,418,HunyuanImage3ForCausalMM,hunyuan_image_3_moe,6987710464.0,12800.0,4096.0,32.0,32.0,133120.0,Language Model,['RLHF'],"['text-generation', 'translation', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Creative Writing']",{},55.0,Hard,0.55,2025-11-02T06:03:30.841044,"license: other license_name: tencent-hunyuan-community license_link: LICENSE pipeline_tag: text-to-image library_name: transformers
[!NOTE] > GGUF uploads with our fixes. More details and [Read o...,"['Function Calling', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-Next-80B-A3B-Instruct,Qwen3-Next-80B-A3B-Instruct,Qwen,2025-09-09T15:40:56+00:00,1663221,845,transformers,"['transformers', 'safetensors', 'qwen3_next', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2404.06654', 'arxiv:2505.09388', 'arxiv:2501.15383', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",11,425,Qwen3NextForCausalLM,qwen3_next,2727084032.0,262144.0,2048.0,16.0,48.0,151936.0,Language Model,['Merged'],"['text-generation', 'question-answering', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:03:36.135334,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Instruct/blob/main/LICENSE pipeline_tag: text-generation This model was obtained by uniformly pruning 4...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" internlm/JanusCoder-8B,JanusCoder-8B,internlm,2025-10-27T09:33:54+00:00,139,11,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",11,428,Qwen3ForCausalLM,qwen3,7870087168.0,32768.0,4096.0,32.0,36.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},68.0,Medium,0.68,2025-11-02T06:03:36.685705,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] unsloth/MiniMax-M2,MiniMax-M2,unsloth,2025-10-28T12:05:08+00:00,264,11,transformers,"['transformers', 'safetensors', 'minimax', 'text-generation', 'conversational', 'arxiv:2504.07164', 'arxiv:2509.06501', 'arxiv:2509.13160', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'fp8', 'region:us']",11,429,MiniMaxM2ForCausalLM,minimax,7635861504.0,196608.0,3072.0,48.0,62.0,200064.0,Language Model,['LoRA'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production']",{},91.0,Easy,0.91,2025-11-02T06:03:36.854738,"pipeline_tag: text-generation license: mit library_name: transformers
<...,"['RAG Support', 'Multi-turn']" facebook/MobileLLM-Pro,MobileLLM-Pro,facebook,2025-09-10T18:40:06+00:00,4486,138,transformers,"['transformers', 'safetensors', 'llama4_text', 'text-generation', 'facebook', 'meta', 'pytorch', 'conversational', 'custom_code', 'en', 'base_model:facebook/MobileLLM-Pro-base', 'base_model:finetune:facebook/MobileLLM-Pro-base', 'license:fair-noncommercial-research-license', 'autotrain_compatible', 'region:us']",10,436,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:03:41.632278,No README available,[] nineninesix/kani-tts-370m,kani-tts-370m,nineninesix,2025-09-30T07:31:22+00:00,8444,136,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'text-to-speech', 'en', 'de', 'ar', 'zh', 'es', 'ko', 'arxiv:2505.20506', 'base_model:nineninesix/kani-tts-450m-0.2-pt', 'base_model:finetune:nineninesix/kani-tts-450m-0.2-pt', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",10,437,Lfm2ForCausalLM,lfm2,283798528.0,128000.0,1024.0,16.0,16.0,80539.0,Language Model,"['Fine-tuned', 'Specialized']",['conversation'],"['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Legal']",{},75.0,Medium,0.75,2025-11-02T06:03:41.800663,"license: apache-2.0 language: - en - de - ar - zh - es - ko pipeline_tag: text-to-speech library_name: transformers base_model: - nineninesix/kani-tts-450m-0.2-pt

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" LiquidAI/LFM2-8B-A1B,LFM2-8B-A1B,LiquidAI,2025-10-07T13:55:39+00:00,13348,231,transformers,"['transformers', 'safetensors', 'lfm2_moe', 'text-generation', 'liquid', 'lfm2', 'edge', 'moe', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,447,Lfm2MoeForCausalLM,lfm2_moe,1342177280.0,128000.0,2048.0,32.0,24.0,65536.0,Language Model,"['Quantized', 'LoRA']","['text-generation', 'question-answering', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing', 'Healthcare']",{},88.0,Easy,0.88,2025-11-02T06:03:48.680263,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge - moe ...,"['Function Calling', 'RAG Support', 'Fast Inference', 'Multi-turn', 'Safety Aligned']" cerebras/Qwen3-Coder-REAP-25B-A3B,Qwen3-Coder-REAP-25B-A3B,cerebras,2025-10-20T15:40:03+00:00,971,32,transformers,"['transformers', 'qwen3_moe', 'text-generation', 'qwen-coder', 'MOE', 'pruning', 'compression', 'conversational', 'en', 'arxiv:2510.13999', 'base_model:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'base_model:finetune:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,448,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,262144.0,2048.0,32.0,48.0,151936.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'reasoning']","['English', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},85.0,Easy,0.85,2025-11-02T06:03:48.842699,language: - en library_name: transformers tags: - qwen-coder - MOE - pruning - compression license: apache-2.0 name: cerebras/Qwen3-Coder-REAP-25B-A3B description: > This model was obtained by uniform...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" meta-llama/Llama-3.2-1B-Instruct,Llama-3.2-1B-Instruct,meta-llama,2024-09-18T15:12:47+00:00,3843095,1141,transformers,"['transformers', 'safetensors', 'llama', 'text-generation', 'facebook', 'meta', 'pytorch', 'llama-3', 'conversational', 'en', 'de', 'fr', 'it', 'pt', 'hi', 'es', 'th', 'arxiv:2204.05149', 'arxiv:2405.16406', 'license:llama3.2', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,449,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:03:50.032006,No README available,[] microsoft/Phi-4-mini-instruct,Phi-4-mini-instruct,microsoft,2025-02-19T01:00:58+00:00,249315,623,transformers,"['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'ar', 'zh', 'cs', 'da', 'nl', 'en', 'fi', 'fr', 'de', 'he', 'hu', 'it', 'ja', 'ko', 'no', 'pl', 'pt', 'ru', 'es', 'sv', 'th', 'tr', 'uk', 'arxiv:2503.01743', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,450,Phi3ForCausalLM,phi3,4238475264.0,131072.0,3072.0,24.0,32.0,200064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Business', 'Healthcare', 'Legal']",{},86.0,Easy,0.86,2025-11-02T06:03:50.417744,language: - multilingual - ar - zh - cs - da - nl - en - fi - fr - de - he - hu - it - ja - ko - 'no' - pl - pt - ru - es - sv - th - tr - uk library_name: transformers license: mit license_link: http...,"['Function Calling', 'RAG Support', 'Long Context', 'Multi-turn', 'Safety Aligned']" Qwen/Qwen3-30B-A3B,Qwen3-30B-A3B,Qwen,2025-04-27T03:43:05+00:00,421099,808,transformers,"['transformers', 'safetensors', 'qwen3_moe', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-30B-A3B-Base', 'base_model:finetune:Qwen/Qwen3-30B-A3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",8,451,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,40960.0,2048.0,32.0,48.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:03:52.610146,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-30B-A3B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-30B-A3B-Base - If you want to learn more about how we process your perso...,[] TheBloke/Mistral-7B-Instruct-v0.2-GGUF,Mistral-7B-Instruct-v0.2-GGUF,TheBloke,2023-12-11T22:18:46+00:00,60731,475,transformers,"['transformers', 'gguf', 'mistral', 'finetuned', 'text-generation', 'arxiv:2310.06825', 'base_model:mistralai/Mistral-7B-Instruct-v0.2', 'base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2', 'license:apache-2.0', 'region:us', 'conversational']",6,466,Unknown,mistral,,,,,,,Language Model,"['Fine-tuned', 'Quantized']","['text-generation', 'question-answering', 'summarization', 'conversation']","['English', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Production', 'Creative Writing']",{},63.0,Hard,0.63,2025-11-02T06:04:04.367912,base_model: mistralai/Mistral-7B-Instruct-v0.2 inference: false license: apache-2.0 model_creator: Mistral AI_ model_name: Mistral 7B Instruct v0.2 model_type: mistral pipeline_tag: text-generation pr...,"['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-32B,Qwen3-32B,Qwen,2025-04-27T03:52:59+00:00,1610353,561,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",6,467,Qwen3ForCausalLM,qwen3,20910571520.0,40960.0,5120.0,64.0,64.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},61.0,Hard,0.61,2025-11-02T06:04:04.610446,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-32B/blob/main/LICENSE pipeline_tag: text-generation

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ inference: false fine-tuning: false langu...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" PokeeAI/pokee_research_7b,pokee_research_7b,PokeeAI,2025-10-17T20:38:37+00:00,5574,95,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'agent', 'deepresearch', 'llm', 'rl', 'reinforcementlearning', 'conversational', 'en', 'dataset:miromind-ai/MiroRL-GenQA', 'arxiv:2510.15862', 'base_model:Qwen/Qwen2.5-7B-Instruct', 'base_model:finetune:Qwen/Qwen2.5-7B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",24,507,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged', 'Specialized']","['text-generation', 'question-answering', 'summarization', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education', 'Healthcare', 'Legal', 'Finance']",{},78.0,Medium,0.78,2025-11-02T06:04:24.383264,base_model: - Qwen/Qwen2.5-7B-Instruct datasets: - miromind-ai/MiroRL-GenQA language: - en license: apache-2.0 tags: - agent - deepresearch - llm - rl - reinforcementlearning pipeline_tag: text-genera...,"['RAG Support', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" ibm-granite/granite-4.0-350m-base,granite-4.0-350m-base,ibm-granite,2025-10-07T20:25:24+00:00,495,23,transformers,"['transformers', 'safetensors', 'granitemoehybrid', 'text-generation', 'language', 'granite-4.0', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",23,508,GraniteMoeHybridForCausalLM,granitemoehybrid,455081984.0,32768.0,1024.0,16.0,28.0,100352.0,Language Model,"['LoRA', 'Specialized']","['text-generation', 'question-answering', 'text-classification', 'summarization']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Education']",{},86.0,Easy,0.86,2025-11-02T06:04:24.541922,license: apache-2.0 library_name: transformers tags: - language - granite-4.0 **Model Summary:** Granite-4.0-350M-Base is a lightweight decoder-only language model designed for scenarios where efficie...,"['Long Context', 'Fast Inference', 'Safety Aligned']" internlm/JanusCoder-14B,JanusCoder-14B,internlm,2025-10-27T09:34:49+00:00,165,23,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",23,509,Qwen3ForCausalLM,qwen3,13360824320.0,32768.0,5120.0,40.0,40.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},58.0,Hard,0.58,2025-11-02T06:04:24.735748,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] ByteDance/Ouro-2.6B,Ouro-2.6B,ByteDance,2025-10-28T22:19:46+00:00,227,21,transformers,"['transformers', 'safetensors', 'ouro', 'text-generation', 'looped-language-model', 'reasoning', 'recurrent-depth', 'conversational', 'custom_code', 'arxiv:2510.25741', 'license:apache-2.0', 'autotrain_compatible', 'region:us']",21,510,OuroForCausalLM,ouro,2516582400.0,65536.0,2048.0,16.0,48.0,49152.0,Language Model,[],"['text-generation', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education']",{},78.0,Medium,0.78,2025-11-02T06:04:24.927183,library_name: transformers license: apache-2.0 pipeline_tag: text-generation tags: - looped-language-model - reasoning - recurrent-depth ![Ouro Logo](assets/logo.png) [📚 Paper (Hugging Face)](https://...,"['RAG Support', 'Long Context']" HuggingFaceTB/SmolLM3-3B,SmolLM3-3B,HuggingFaceTB,2025-07-08T10:11:45+00:00,56521,768,transformers,"['transformers', 'safetensors', 'smollm3', 'text-generation', 'conversational', 'en', 'fr', 'es', 'it', 'pt', 'zh', 'ar', 'ru', 'base_model:HuggingFaceTB/SmolLM3-3B-Base', 'base_model:finetune:HuggingFaceTB/SmolLM3-3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",20,511,SmolLM3ForCausalLM,smollm3,2074607616.0,65536.0,2048.0,16.0,36.0,128256.0,Language Model,['Quantized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},83.0,Medium,0.83,2025-11-02T06:04:27.385153,library_name: transformers license: apache-2.0 language: - en - fr - es - it - pt - zh - ar - ru base_model: - HuggingFaceTB/SmolLM3-3B-Base ![image/png](https://cdn-uploads.huggingface.co/production/...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient']" Qwen/Qwen3-0.6B,Qwen3-0.6B,Qwen,2025-04-27T03:40:08+00:00,7327000,752,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-0.6B-Base', 'base_model:finetune:Qwen/Qwen3-0.6B-Base', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",19,512,Qwen3ForCausalLM,qwen3,507904000.0,40960.0,1024.0,16.0,28.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},78.0,Medium,0.78,2025-11-02T06:04:27.646742,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-0.6B-Base This model was obtained by uniformly pruning 25...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" Qwen/Qwen2.5-7B-Instruct,Qwen2.5-7B-Instruct,Qwen,2024-09-16T11:55:40+00:00,8169288,849,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-7B', 'base_model:finetune:Qwen/Qwen2.5-7B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",17,516,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},86.0,Easy,0.86,2025-11-02T06:04:28.967496,license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE language: - en pipeline_tag: text-generation base_model: Qwen/Qwen2.5-7B tags: - chat library_name: ...,"['Long Context', 'Fast Inference', 'Safety Aligned']" deepseek-ai/DeepSeek-R1,DeepSeek-R1,deepseek-ai,2025-01-20T03:46:07+00:00,425526,12820,transformers,"['transformers', 'safetensors', 'deepseek_v3', 'text-generation', 'conversational', 'custom_code', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'fp8', 'region:us']",17,517,DeepseekV3ForCausalLM,deepseek_v3,38537003008.0,163840.0,7168.0,128.0,61.0,129280.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged']","['text-generation', 'question-answering', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business']",{},48.0,Hard,0.48,2025-11-02T06:04:29.139725,"license: mit library_name: transformers
- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" microsoft/VibeVoice-1.5B,VibeVoice-1.5B,microsoft,2025-08-25T13:46:48+00:00,223609,1943,transformers,"['transformers', 'safetensors', 'vibevoice', 'text-generation', 'Podcast', 'text-to-speech', 'en', 'zh', 'arxiv:2508.19205', 'arxiv:2412.08635', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,536,VibeVoiceForConditionalGeneration,vibevoice,,,,,,,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business', 'Legal']",{},53.0,Hard,0.53,2025-11-02T06:04:41.697053,"language: - en - zh license: mit pipeline_tag: text-to-speech tags: - Podcast library_name: transformers VibeVoice is a novel framework designed for generating expressive, long-form, multi-speaker con...","['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" LiquidAI/LFM2-350M-Extract,LFM2-350M-Extract,LiquidAI,2025-09-03T17:13:53+00:00,5852,61,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'liquid', 'edge', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'base_model:LiquidAI/LFM2-350M', 'base_model:finetune:LiquidAI/LFM2-350M', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,537,Lfm2ForCausalLM,lfm2,268435456.0,128000.0,1024.0,16.0,16.0,65536.0,Language Model,['LoRA'],"['text-generation', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},70.0,Medium,0.7,2025-11-02T06:04:41.856905,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge base_m...,['Multi-turn'] tencent/HunyuanImage-3.0,HunyuanImage-3.0,tencent,2025-09-25T06:28:28+00:00,101407,947,transformers,"['transformers', 'safetensors', 'hunyuan_image_3_moe', 'text-generation', 'text-to-image', 'custom_code', 'arxiv:2509.23951', 'license:other', 'autotrain_compatible', 'region:us']",12,538,HunyuanImage3ForCausalMM,hunyuan_image_3_moe,6987710464.0,12800.0,4096.0,32.0,32.0,133120.0,Language Model,['RLHF'],"['text-generation', 'translation', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Creative Writing']",{},55.0,Hard,0.55,2025-11-02T06:04:42.507409,"license: other license_name: tencent-hunyuan-community license_link: LICENSE pipeline_tag: text-to-image library_name: transformers
[!NOTE] > GGUF uploads with our fixes. More details and [Read o...,"['Function Calling', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-Next-80B-A3B-Instruct,Qwen3-Next-80B-A3B-Instruct,Qwen,2025-09-09T15:40:56+00:00,1663221,845,transformers,"['transformers', 'safetensors', 'qwen3_next', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2404.06654', 'arxiv:2505.09388', 'arxiv:2501.15383', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",11,545,Qwen3NextForCausalLM,qwen3_next,2727084032.0,262144.0,2048.0,16.0,48.0,151936.0,Language Model,['Merged'],"['text-generation', 'question-answering', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:04:48.062881,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Instruct/blob/main/LICENSE pipeline_tag: text-generation This model was obtained by uniformly pruning 4...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" internlm/JanusCoder-8B,JanusCoder-8B,internlm,2025-10-27T09:33:54+00:00,139,11,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",11,548,Qwen3ForCausalLM,qwen3,7870087168.0,32768.0,4096.0,32.0,36.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},68.0,Medium,0.68,2025-11-02T06:04:48.730231,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] unsloth/MiniMax-M2,MiniMax-M2,unsloth,2025-10-28T12:05:08+00:00,264,11,transformers,"['transformers', 'safetensors', 'minimax', 'text-generation', 'conversational', 'arxiv:2504.07164', 'arxiv:2509.06501', 'arxiv:2509.13160', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'fp8', 'region:us']",11,549,MiniMaxM2ForCausalLM,minimax,7635861504.0,196608.0,3072.0,48.0,62.0,200064.0,Language Model,['LoRA'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production']",{},91.0,Easy,0.91,2025-11-02T06:04:48.913609,"pipeline_tag: text-generation license: mit library_name: transformers
<...,"['RAG Support', 'Multi-turn']" facebook/MobileLLM-Pro,MobileLLM-Pro,facebook,2025-09-10T18:40:06+00:00,4486,138,transformers,"['transformers', 'safetensors', 'llama4_text', 'text-generation', 'facebook', 'meta', 'pytorch', 'conversational', 'custom_code', 'en', 'base_model:facebook/MobileLLM-Pro-base', 'base_model:finetune:facebook/MobileLLM-Pro-base', 'license:fair-noncommercial-research-license', 'autotrain_compatible', 'region:us']",10,556,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:04:53.726917,No README available,[] nineninesix/kani-tts-370m,kani-tts-370m,nineninesix,2025-09-30T07:31:22+00:00,8444,136,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'text-to-speech', 'en', 'de', 'ar', 'zh', 'es', 'ko', 'arxiv:2505.20506', 'base_model:nineninesix/kani-tts-450m-0.2-pt', 'base_model:finetune:nineninesix/kani-tts-450m-0.2-pt', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",10,557,Lfm2ForCausalLM,lfm2,283798528.0,128000.0,1024.0,16.0,16.0,80539.0,Language Model,"['Fine-tuned', 'Specialized']",['conversation'],"['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Legal']",{},75.0,Medium,0.75,2025-11-02T06:04:53.846237,"license: apache-2.0 language: - en - de - ar - zh - es - ko pipeline_tag: text-to-speech library_name: transformers base_model: - nineninesix/kani-tts-450m-0.2-pt

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" LiquidAI/LFM2-8B-A1B,LFM2-8B-A1B,LiquidAI,2025-10-07T13:55:39+00:00,13348,231,transformers,"['transformers', 'safetensors', 'lfm2_moe', 'text-generation', 'liquid', 'lfm2', 'edge', 'moe', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,567,Lfm2MoeForCausalLM,lfm2_moe,1342177280.0,128000.0,2048.0,32.0,24.0,65536.0,Language Model,"['Quantized', 'LoRA']","['text-generation', 'question-answering', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing', 'Healthcare']",{},88.0,Easy,0.88,2025-11-02T06:05:00.694981,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge - moe ...,"['Function Calling', 'RAG Support', 'Fast Inference', 'Multi-turn', 'Safety Aligned']" cerebras/Qwen3-Coder-REAP-25B-A3B,Qwen3-Coder-REAP-25B-A3B,cerebras,2025-10-20T15:40:03+00:00,971,32,transformers,"['transformers', 'qwen3_moe', 'text-generation', 'qwen-coder', 'MOE', 'pruning', 'compression', 'conversational', 'en', 'arxiv:2510.13999', 'base_model:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'base_model:finetune:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,568,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,262144.0,2048.0,32.0,48.0,151936.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'reasoning']","['English', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},85.0,Easy,0.85,2025-11-02T06:05:00.907388,language: - en library_name: transformers tags: - qwen-coder - MOE - pruning - compression license: apache-2.0 name: cerebras/Qwen3-Coder-REAP-25B-A3B description: > This model was obtained by uniform...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" meta-llama/Llama-3.2-1B-Instruct,Llama-3.2-1B-Instruct,meta-llama,2024-09-18T15:12:47+00:00,3843095,1141,transformers,"['transformers', 'safetensors', 'llama', 'text-generation', 'facebook', 'meta', 'pytorch', 'llama-3', 'conversational', 'en', 'de', 'fr', 'it', 'pt', 'hi', 'es', 'th', 'arxiv:2204.05149', 'arxiv:2405.16406', 'license:llama3.2', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,569,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:05:02.115343,No README available,[] microsoft/Phi-4-mini-instruct,Phi-4-mini-instruct,microsoft,2025-02-19T01:00:58+00:00,249315,623,transformers,"['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'ar', 'zh', 'cs', 'da', 'nl', 'en', 'fi', 'fr', 'de', 'he', 'hu', 'it', 'ja', 'ko', 'no', 'pl', 'pt', 'ru', 'es', 'sv', 'th', 'tr', 'uk', 'arxiv:2503.01743', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,570,Phi3ForCausalLM,phi3,4238475264.0,131072.0,3072.0,24.0,32.0,200064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Business', 'Healthcare', 'Legal']",{},86.0,Easy,0.86,2025-11-02T06:05:02.271039,language: - multilingual - ar - zh - cs - da - nl - en - fi - fr - de - he - hu - it - ja - ko - 'no' - pl - pt - ru - es - sv - th - tr - uk library_name: transformers license: mit license_link: http...,"['Function Calling', 'RAG Support', 'Long Context', 'Multi-turn', 'Safety Aligned']" Qwen/Qwen3-30B-A3B,Qwen3-30B-A3B,Qwen,2025-04-27T03:43:05+00:00,421099,808,transformers,"['transformers', 'safetensors', 'qwen3_moe', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-30B-A3B-Base', 'base_model:finetune:Qwen/Qwen3-30B-A3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",8,571,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,40960.0,2048.0,32.0,48.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:05:04.542970,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-30B-A3B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-30B-A3B-Base - If you want to learn more about how we process your perso...,[] TheBloke/Mistral-7B-Instruct-v0.2-GGUF,Mistral-7B-Instruct-v0.2-GGUF,TheBloke,2023-12-11T22:18:46+00:00,60731,475,transformers,"['transformers', 'gguf', 'mistral', 'finetuned', 'text-generation', 'arxiv:2310.06825', 'base_model:mistralai/Mistral-7B-Instruct-v0.2', 'base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2', 'license:apache-2.0', 'region:us', 'conversational']",6,586,Unknown,mistral,,,,,,,Language Model,"['Fine-tuned', 'Quantized']","['text-generation', 'question-answering', 'summarization', 'conversation']","['English', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Production', 'Creative Writing']",{},63.0,Hard,0.63,2025-11-02T06:05:16.419112,base_model: mistralai/Mistral-7B-Instruct-v0.2 inference: false license: apache-2.0 model_creator: Mistral AI_ model_name: Mistral 7B Instruct v0.2 model_type: mistral pipeline_tag: text-generation pr...,"['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-32B,Qwen3-32B,Qwen,2025-04-27T03:52:59+00:00,1610353,561,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",6,587,Qwen3ForCausalLM,qwen3,20910571520.0,40960.0,5120.0,64.0,64.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},61.0,Hard,0.61,2025-11-02T06:05:16.666075,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-32B/blob/main/LICENSE pipeline_tag: text-generation

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ inference: false fine-tuning: false langu...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" PokeeAI/pokee_research_7b,pokee_research_7b,PokeeAI,2025-10-17T20:38:37+00:00,5574,95,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'agent', 'deepresearch', 'llm', 'rl', 'reinforcementlearning', 'conversational', 'en', 'dataset:miromind-ai/MiroRL-GenQA', 'arxiv:2510.15862', 'base_model:Qwen/Qwen2.5-7B-Instruct', 'base_model:finetune:Qwen/Qwen2.5-7B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",24,627,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged', 'Specialized']","['text-generation', 'question-answering', 'summarization', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education', 'Healthcare', 'Legal', 'Finance']",{},78.0,Medium,0.78,2025-11-02T06:05:40.411269,base_model: - Qwen/Qwen2.5-7B-Instruct datasets: - miromind-ai/MiroRL-GenQA language: - en license: apache-2.0 tags: - agent - deepresearch - llm - rl - reinforcementlearning pipeline_tag: text-genera...,"['RAG Support', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" ibm-granite/granite-4.0-350m-base,granite-4.0-350m-base,ibm-granite,2025-10-07T20:25:24+00:00,495,23,transformers,"['transformers', 'safetensors', 'granitemoehybrid', 'text-generation', 'language', 'granite-4.0', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",23,628,GraniteMoeHybridForCausalLM,granitemoehybrid,455081984.0,32768.0,1024.0,16.0,28.0,100352.0,Language Model,"['LoRA', 'Specialized']","['text-generation', 'question-answering', 'text-classification', 'summarization']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Education']",{},86.0,Easy,0.86,2025-11-02T06:05:40.562937,license: apache-2.0 library_name: transformers tags: - language - granite-4.0 **Model Summary:** Granite-4.0-350M-Base is a lightweight decoder-only language model designed for scenarios where efficie...,"['Long Context', 'Fast Inference', 'Safety Aligned']" internlm/JanusCoder-14B,JanusCoder-14B,internlm,2025-10-27T09:34:49+00:00,165,23,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",23,629,Qwen3ForCausalLM,qwen3,13360824320.0,32768.0,5120.0,40.0,40.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},58.0,Hard,0.58,2025-11-02T06:05:40.742049,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] ByteDance/Ouro-2.6B,Ouro-2.6B,ByteDance,2025-10-28T22:19:46+00:00,227,21,transformers,"['transformers', 'safetensors', 'ouro', 'text-generation', 'looped-language-model', 'reasoning', 'recurrent-depth', 'conversational', 'custom_code', 'arxiv:2510.25741', 'license:apache-2.0', 'autotrain_compatible', 'region:us']",21,630,OuroForCausalLM,ouro,2516582400.0,65536.0,2048.0,16.0,48.0,49152.0,Language Model,[],"['text-generation', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education']",{},78.0,Medium,0.78,2025-11-02T06:05:40.859257,library_name: transformers license: apache-2.0 pipeline_tag: text-generation tags: - looped-language-model - reasoning - recurrent-depth ![Ouro Logo](assets/logo.png) [📚 Paper (Hugging Face)](https://...,"['RAG Support', 'Long Context']" HuggingFaceTB/SmolLM3-3B,SmolLM3-3B,HuggingFaceTB,2025-07-08T10:11:45+00:00,56521,768,transformers,"['transformers', 'safetensors', 'smollm3', 'text-generation', 'conversational', 'en', 'fr', 'es', 'it', 'pt', 'zh', 'ar', 'ru', 'base_model:HuggingFaceTB/SmolLM3-3B-Base', 'base_model:finetune:HuggingFaceTB/SmolLM3-3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",20,631,SmolLM3ForCausalLM,smollm3,2074607616.0,65536.0,2048.0,16.0,36.0,128256.0,Language Model,['Quantized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},83.0,Medium,0.83,2025-11-02T06:05:43.126512,library_name: transformers license: apache-2.0 language: - en - fr - es - it - pt - zh - ar - ru base_model: - HuggingFaceTB/SmolLM3-3B-Base ![image/png](https://cdn-uploads.huggingface.co/production/...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient']" Qwen/Qwen3-0.6B,Qwen3-0.6B,Qwen,2025-04-27T03:40:08+00:00,7327000,752,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-0.6B-Base', 'base_model:finetune:Qwen/Qwen3-0.6B-Base', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",19,632,Qwen3ForCausalLM,qwen3,507904000.0,40960.0,1024.0,16.0,28.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},78.0,Medium,0.78,2025-11-02T06:05:44.087197,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-0.6B-Base This model was obtained by uniformly pruning 25...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" Qwen/Qwen2.5-7B-Instruct,Qwen2.5-7B-Instruct,Qwen,2024-09-16T11:55:40+00:00,8169288,849,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-7B', 'base_model:finetune:Qwen/Qwen2.5-7B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",17,636,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},86.0,Easy,0.86,2025-11-02T06:05:45.137638,license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE language: - en pipeline_tag: text-generation base_model: Qwen/Qwen2.5-7B tags: - chat library_name: ...,"['Long Context', 'Fast Inference', 'Safety Aligned']" deepseek-ai/DeepSeek-R1,DeepSeek-R1,deepseek-ai,2025-01-20T03:46:07+00:00,425526,12820,transformers,"['transformers', 'safetensors', 'deepseek_v3', 'text-generation', 'conversational', 'custom_code', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'fp8', 'region:us']",17,637,DeepseekV3ForCausalLM,deepseek_v3,38537003008.0,163840.0,7168.0,128.0,61.0,129280.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged']","['text-generation', 'question-answering', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business']",{},48.0,Hard,0.48,2025-11-02T06:05:45.306025,"license: mit library_name: transformers
- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" microsoft/VibeVoice-1.5B,VibeVoice-1.5B,microsoft,2025-08-25T13:46:48+00:00,223609,1943,transformers,"['transformers', 'safetensors', 'vibevoice', 'text-generation', 'Podcast', 'text-to-speech', 'en', 'zh', 'arxiv:2508.19205', 'arxiv:2412.08635', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,656,VibeVoiceForConditionalGeneration,vibevoice,,,,,,,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business', 'Legal']",{},53.0,Hard,0.53,2025-11-02T06:06:01.739478,"language: - en - zh license: mit pipeline_tag: text-to-speech tags: - Podcast library_name: transformers VibeVoice is a novel framework designed for generating expressive, long-form, multi-speaker con...","['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" LiquidAI/LFM2-350M-Extract,LFM2-350M-Extract,LiquidAI,2025-09-03T17:13:53+00:00,5852,61,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'liquid', 'edge', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'base_model:LiquidAI/LFM2-350M', 'base_model:finetune:LiquidAI/LFM2-350M', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,657,Lfm2ForCausalLM,lfm2,268435456.0,128000.0,1024.0,16.0,16.0,65536.0,Language Model,['LoRA'],"['text-generation', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},70.0,Medium,0.7,2025-11-02T06:06:01.909892,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge base_m...,['Multi-turn'] tencent/HunyuanImage-3.0,HunyuanImage-3.0,tencent,2025-09-25T06:28:28+00:00,101407,947,transformers,"['transformers', 'safetensors', 'hunyuan_image_3_moe', 'text-generation', 'text-to-image', 'custom_code', 'arxiv:2509.23951', 'license:other', 'autotrain_compatible', 'region:us']",12,658,HunyuanImage3ForCausalMM,hunyuan_image_3_moe,6987710464.0,12800.0,4096.0,32.0,32.0,133120.0,Language Model,['RLHF'],"['text-generation', 'translation', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Creative Writing']",{},55.0,Hard,0.55,2025-11-02T06:06:02.099876,"license: other license_name: tencent-hunyuan-community license_link: LICENSE pipeline_tag: text-to-image library_name: transformers
[!NOTE] > GGUF uploads with our fixes. More details and [Read o...,"['Function Calling', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-Next-80B-A3B-Instruct,Qwen3-Next-80B-A3B-Instruct,Qwen,2025-09-09T15:40:56+00:00,1663221,845,transformers,"['transformers', 'safetensors', 'qwen3_next', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2404.06654', 'arxiv:2505.09388', 'arxiv:2501.15383', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",11,665,Qwen3NextForCausalLM,qwen3_next,2727084032.0,262144.0,2048.0,16.0,48.0,151936.0,Language Model,['Merged'],"['text-generation', 'question-answering', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:06:07.545109,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Instruct/blob/main/LICENSE pipeline_tag: text-generation This model was obtained by uniformly pruning 4...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" internlm/JanusCoder-8B,JanusCoder-8B,internlm,2025-10-27T09:33:54+00:00,139,11,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",11,668,Qwen3ForCausalLM,qwen3,7870087168.0,32768.0,4096.0,32.0,36.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},68.0,Medium,0.68,2025-11-02T06:06:08.076246,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] unsloth/MiniMax-M2,MiniMax-M2,unsloth,2025-10-28T12:05:08+00:00,264,11,transformers,"['transformers', 'safetensors', 'minimax', 'text-generation', 'conversational', 'arxiv:2504.07164', 'arxiv:2509.06501', 'arxiv:2509.13160', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'fp8', 'region:us']",11,669,MiniMaxM2ForCausalLM,minimax,7635861504.0,196608.0,3072.0,48.0,62.0,200064.0,Language Model,['LoRA'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production']",{},91.0,Easy,0.91,2025-11-02T06:06:08.231981,"pipeline_tag: text-generation license: mit library_name: transformers
<...,"['RAG Support', 'Multi-turn']" facebook/MobileLLM-Pro,MobileLLM-Pro,facebook,2025-09-10T18:40:06+00:00,4486,138,transformers,"['transformers', 'safetensors', 'llama4_text', 'text-generation', 'facebook', 'meta', 'pytorch', 'conversational', 'custom_code', 'en', 'base_model:facebook/MobileLLM-Pro-base', 'base_model:finetune:facebook/MobileLLM-Pro-base', 'license:fair-noncommercial-research-license', 'autotrain_compatible', 'region:us']",10,676,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:06:12.455641,No README available,[] nineninesix/kani-tts-370m,kani-tts-370m,nineninesix,2025-09-30T07:31:22+00:00,8444,136,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'text-to-speech', 'en', 'de', 'ar', 'zh', 'es', 'ko', 'arxiv:2505.20506', 'base_model:nineninesix/kani-tts-450m-0.2-pt', 'base_model:finetune:nineninesix/kani-tts-450m-0.2-pt', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",10,677,Lfm2ForCausalLM,lfm2,283798528.0,128000.0,1024.0,16.0,16.0,80539.0,Language Model,"['Fine-tuned', 'Specialized']",['conversation'],"['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Legal']",{},75.0,Medium,0.75,2025-11-02T06:06:12.819920,"license: apache-2.0 language: - en - de - ar - zh - es - ko pipeline_tag: text-to-speech library_name: transformers base_model: - nineninesix/kani-tts-450m-0.2-pt

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" LiquidAI/LFM2-8B-A1B,LFM2-8B-A1B,LiquidAI,2025-10-07T13:55:39+00:00,13348,231,transformers,"['transformers', 'safetensors', 'lfm2_moe', 'text-generation', 'liquid', 'lfm2', 'edge', 'moe', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,687,Lfm2MoeForCausalLM,lfm2_moe,1342177280.0,128000.0,2048.0,32.0,24.0,65536.0,Language Model,"['Quantized', 'LoRA']","['text-generation', 'question-answering', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing', 'Healthcare']",{},88.0,Easy,0.88,2025-11-02T06:06:20.119550,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge - moe ...,"['Function Calling', 'RAG Support', 'Fast Inference', 'Multi-turn', 'Safety Aligned']" cerebras/Qwen3-Coder-REAP-25B-A3B,Qwen3-Coder-REAP-25B-A3B,cerebras,2025-10-20T15:40:03+00:00,971,32,transformers,"['transformers', 'qwen3_moe', 'text-generation', 'qwen-coder', 'MOE', 'pruning', 'compression', 'conversational', 'en', 'arxiv:2510.13999', 'base_model:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'base_model:finetune:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,688,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,262144.0,2048.0,32.0,48.0,151936.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'reasoning']","['English', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},85.0,Easy,0.85,2025-11-02T06:06:20.501870,language: - en library_name: transformers tags: - qwen-coder - MOE - pruning - compression license: apache-2.0 name: cerebras/Qwen3-Coder-REAP-25B-A3B description: > This model was obtained by uniform...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" meta-llama/Llama-3.2-1B-Instruct,Llama-3.2-1B-Instruct,meta-llama,2024-09-18T15:12:47+00:00,3843095,1141,transformers,"['transformers', 'safetensors', 'llama', 'text-generation', 'facebook', 'meta', 'pytorch', 'llama-3', 'conversational', 'en', 'de', 'fr', 'it', 'pt', 'hi', 'es', 'th', 'arxiv:2204.05149', 'arxiv:2405.16406', 'license:llama3.2', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,689,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:06:21.686870,No README available,[] microsoft/Phi-4-mini-instruct,Phi-4-mini-instruct,microsoft,2025-02-19T01:00:58+00:00,249315,623,transformers,"['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'ar', 'zh', 'cs', 'da', 'nl', 'en', 'fi', 'fr', 'de', 'he', 'hu', 'it', 'ja', 'ko', 'no', 'pl', 'pt', 'ru', 'es', 'sv', 'th', 'tr', 'uk', 'arxiv:2503.01743', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,690,Phi3ForCausalLM,phi3,4238475264.0,131072.0,3072.0,24.0,32.0,200064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Business', 'Healthcare', 'Legal']",{},86.0,Easy,0.86,2025-11-02T06:06:21.861792,language: - multilingual - ar - zh - cs - da - nl - en - fi - fr - de - he - hu - it - ja - ko - 'no' - pl - pt - ru - es - sv - th - tr - uk library_name: transformers license: mit license_link: http...,"['Function Calling', 'RAG Support', 'Long Context', 'Multi-turn', 'Safety Aligned']" Qwen/Qwen3-30B-A3B,Qwen3-30B-A3B,Qwen,2025-04-27T03:43:05+00:00,421099,808,transformers,"['transformers', 'safetensors', 'qwen3_moe', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-30B-A3B-Base', 'base_model:finetune:Qwen/Qwen3-30B-A3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",8,691,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,40960.0,2048.0,32.0,48.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:06:24.105877,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-30B-A3B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-30B-A3B-Base - If you want to learn more about how we process your perso...,[] TheBloke/Mistral-7B-Instruct-v0.2-GGUF,Mistral-7B-Instruct-v0.2-GGUF,TheBloke,2023-12-11T22:18:46+00:00,60731,475,transformers,"['transformers', 'gguf', 'mistral', 'finetuned', 'text-generation', 'arxiv:2310.06825', 'base_model:mistralai/Mistral-7B-Instruct-v0.2', 'base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2', 'license:apache-2.0', 'region:us', 'conversational']",6,706,Unknown,mistral,,,,,,,Language Model,"['Fine-tuned', 'Quantized']","['text-generation', 'question-answering', 'summarization', 'conversation']","['English', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Production', 'Creative Writing']",{},63.0,Hard,0.63,2025-11-02T06:06:35.190803,base_model: mistralai/Mistral-7B-Instruct-v0.2 inference: false license: apache-2.0 model_creator: Mistral AI_ model_name: Mistral 7B Instruct v0.2 model_type: mistral pipeline_tag: text-generation pr...,"['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-32B,Qwen3-32B,Qwen,2025-04-27T03:52:59+00:00,1610353,561,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",6,707,Qwen3ForCausalLM,qwen3,20910571520.0,40960.0,5120.0,64.0,64.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},61.0,Hard,0.61,2025-11-02T06:06:35.316618,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-32B/blob/main/LICENSE pipeline_tag: text-generation

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ inference: false fine-tuning: false langu...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" PokeeAI/pokee_research_7b,pokee_research_7b,PokeeAI,2025-10-17T20:38:37+00:00,5574,95,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'agent', 'deepresearch', 'llm', 'rl', 'reinforcementlearning', 'conversational', 'en', 'dataset:miromind-ai/MiroRL-GenQA', 'arxiv:2510.15862', 'base_model:Qwen/Qwen2.5-7B-Instruct', 'base_model:finetune:Qwen/Qwen2.5-7B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",24,747,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged', 'Specialized']","['text-generation', 'question-answering', 'summarization', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education', 'Healthcare', 'Legal', 'Finance']",{},78.0,Medium,0.78,2025-11-02T06:06:55.153433,base_model: - Qwen/Qwen2.5-7B-Instruct datasets: - miromind-ai/MiroRL-GenQA language: - en license: apache-2.0 tags: - agent - deepresearch - llm - rl - reinforcementlearning pipeline_tag: text-genera...,"['RAG Support', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" ibm-granite/granite-4.0-350m-base,granite-4.0-350m-base,ibm-granite,2025-10-07T20:25:24+00:00,495,23,transformers,"['transformers', 'safetensors', 'granitemoehybrid', 'text-generation', 'language', 'granite-4.0', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",23,748,GraniteMoeHybridForCausalLM,granitemoehybrid,455081984.0,32768.0,1024.0,16.0,28.0,100352.0,Language Model,"['LoRA', 'Specialized']","['text-generation', 'question-answering', 'text-classification', 'summarization']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Education']",{},86.0,Easy,0.86,2025-11-02T06:06:55.302105,license: apache-2.0 library_name: transformers tags: - language - granite-4.0 **Model Summary:** Granite-4.0-350M-Base is a lightweight decoder-only language model designed for scenarios where efficie...,"['Long Context', 'Fast Inference', 'Safety Aligned']" internlm/JanusCoder-14B,JanusCoder-14B,internlm,2025-10-27T09:34:49+00:00,165,23,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",23,749,Qwen3ForCausalLM,qwen3,13360824320.0,32768.0,5120.0,40.0,40.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},58.0,Hard,0.58,2025-11-02T06:06:55.433706,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] ByteDance/Ouro-2.6B,Ouro-2.6B,ByteDance,2025-10-28T22:19:46+00:00,227,21,transformers,"['transformers', 'safetensors', 'ouro', 'text-generation', 'looped-language-model', 'reasoning', 'recurrent-depth', 'conversational', 'custom_code', 'arxiv:2510.25741', 'license:apache-2.0', 'autotrain_compatible', 'region:us']",21,750,OuroForCausalLM,ouro,2516582400.0,65536.0,2048.0,16.0,48.0,49152.0,Language Model,[],"['text-generation', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education']",{},78.0,Medium,0.78,2025-11-02T06:06:55.563280,library_name: transformers license: apache-2.0 pipeline_tag: text-generation tags: - looped-language-model - reasoning - recurrent-depth ![Ouro Logo](assets/logo.png) [📚 Paper (Hugging Face)](https://...,"['RAG Support', 'Long Context']" HuggingFaceTB/SmolLM3-3B,SmolLM3-3B,HuggingFaceTB,2025-07-08T10:11:45+00:00,56521,768,transformers,"['transformers', 'safetensors', 'smollm3', 'text-generation', 'conversational', 'en', 'fr', 'es', 'it', 'pt', 'zh', 'ar', 'ru', 'base_model:HuggingFaceTB/SmolLM3-3B-Base', 'base_model:finetune:HuggingFaceTB/SmolLM3-3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",20,751,SmolLM3ForCausalLM,smollm3,2074607616.0,65536.0,2048.0,16.0,36.0,128256.0,Language Model,['Quantized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},83.0,Medium,0.83,2025-11-02T06:06:57.776558,library_name: transformers license: apache-2.0 language: - en - fr - es - it - pt - zh - ar - ru base_model: - HuggingFaceTB/SmolLM3-3B-Base ![image/png](https://cdn-uploads.huggingface.co/production/...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient']" Qwen/Qwen3-0.6B,Qwen3-0.6B,Qwen,2025-04-27T03:40:08+00:00,7327000,752,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-0.6B-Base', 'base_model:finetune:Qwen/Qwen3-0.6B-Base', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",19,752,Qwen3ForCausalLM,qwen3,507904000.0,40960.0,1024.0,16.0,28.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},78.0,Medium,0.78,2025-11-02T06:06:57.908805,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-0.6B-Base This model was obtained by uniformly pruning 25...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" Qwen/Qwen2.5-7B-Instruct,Qwen2.5-7B-Instruct,Qwen,2024-09-16T11:55:40+00:00,8169288,849,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-7B', 'base_model:finetune:Qwen/Qwen2.5-7B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",17,756,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},86.0,Easy,0.86,2025-11-02T06:06:58.578005,license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE language: - en pipeline_tag: text-generation base_model: Qwen/Qwen2.5-7B tags: - chat library_name: ...,"['Long Context', 'Fast Inference', 'Safety Aligned']" deepseek-ai/DeepSeek-R1,DeepSeek-R1,deepseek-ai,2025-01-20T03:46:07+00:00,425526,12820,transformers,"['transformers', 'safetensors', 'deepseek_v3', 'text-generation', 'conversational', 'custom_code', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'fp8', 'region:us']",17,757,DeepseekV3ForCausalLM,deepseek_v3,38537003008.0,163840.0,7168.0,128.0,61.0,129280.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged']","['text-generation', 'question-answering', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business']",{},48.0,Hard,0.48,2025-11-02T06:06:58.706613,"license: mit library_name: transformers
- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" microsoft/VibeVoice-1.5B,VibeVoice-1.5B,microsoft,2025-08-25T13:46:48+00:00,223609,1943,transformers,"['transformers', 'safetensors', 'vibevoice', 'text-generation', 'Podcast', 'text-to-speech', 'en', 'zh', 'arxiv:2508.19205', 'arxiv:2412.08635', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,776,VibeVoiceForConditionalGeneration,vibevoice,,,,,,,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business', 'Legal']",{},53.0,Hard,0.53,2025-11-02T06:07:12.044701,"language: - en - zh license: mit pipeline_tag: text-to-speech tags: - Podcast library_name: transformers VibeVoice is a novel framework designed for generating expressive, long-form, multi-speaker con...","['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" LiquidAI/LFM2-350M-Extract,LFM2-350M-Extract,LiquidAI,2025-09-03T17:13:53+00:00,5852,61,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'liquid', 'edge', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'base_model:LiquidAI/LFM2-350M', 'base_model:finetune:LiquidAI/LFM2-350M', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,777,Lfm2ForCausalLM,lfm2,268435456.0,128000.0,1024.0,16.0,16.0,65536.0,Language Model,['LoRA'],"['text-generation', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},70.0,Medium,0.7,2025-11-02T06:07:12.213770,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge base_m...,['Multi-turn'] tencent/HunyuanImage-3.0,HunyuanImage-3.0,tencent,2025-09-25T06:28:28+00:00,101407,947,transformers,"['transformers', 'safetensors', 'hunyuan_image_3_moe', 'text-generation', 'text-to-image', 'custom_code', 'arxiv:2509.23951', 'license:other', 'autotrain_compatible', 'region:us']",12,778,HunyuanImage3ForCausalMM,hunyuan_image_3_moe,6987710464.0,12800.0,4096.0,32.0,32.0,133120.0,Language Model,['RLHF'],"['text-generation', 'translation', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Creative Writing']",{},55.0,Hard,0.55,2025-11-02T06:07:12.456853,"license: other license_name: tencent-hunyuan-community license_link: LICENSE pipeline_tag: text-to-image library_name: transformers
[!NOTE] > GGUF uploads with our fixes. More details and [Read o...,"['Function Calling', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-Next-80B-A3B-Instruct,Qwen3-Next-80B-A3B-Instruct,Qwen,2025-09-09T15:40:56+00:00,1663221,845,transformers,"['transformers', 'safetensors', 'qwen3_next', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2404.06654', 'arxiv:2505.09388', 'arxiv:2501.15383', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",11,785,Qwen3NextForCausalLM,qwen3_next,2727084032.0,262144.0,2048.0,16.0,48.0,151936.0,Language Model,['Merged'],"['text-generation', 'question-answering', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:07:17.873192,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Instruct/blob/main/LICENSE pipeline_tag: text-generation This model was obtained by uniformly pruning 4...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" internlm/JanusCoder-8B,JanusCoder-8B,internlm,2025-10-27T09:33:54+00:00,139,11,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",11,788,Qwen3ForCausalLM,qwen3,7870087168.0,32768.0,4096.0,32.0,36.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},68.0,Medium,0.68,2025-11-02T06:07:18.631013,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] unsloth/MiniMax-M2,MiniMax-M2,unsloth,2025-10-28T12:05:08+00:00,264,11,transformers,"['transformers', 'safetensors', 'minimax', 'text-generation', 'conversational', 'arxiv:2504.07164', 'arxiv:2509.06501', 'arxiv:2509.13160', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'fp8', 'region:us']",11,789,MiniMaxM2ForCausalLM,minimax,7635861504.0,196608.0,3072.0,48.0,62.0,200064.0,Language Model,['LoRA'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production']",{},91.0,Easy,0.91,2025-11-02T06:07:18.837348,"pipeline_tag: text-generation license: mit library_name: transformers
<...,"['RAG Support', 'Multi-turn']" facebook/MobileLLM-Pro,MobileLLM-Pro,facebook,2025-09-10T18:40:06+00:00,4486,138,transformers,"['transformers', 'safetensors', 'llama4_text', 'text-generation', 'facebook', 'meta', 'pytorch', 'conversational', 'custom_code', 'en', 'base_model:facebook/MobileLLM-Pro-base', 'base_model:finetune:facebook/MobileLLM-Pro-base', 'license:fair-noncommercial-research-license', 'autotrain_compatible', 'region:us']",10,796,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:07:23.117546,No README available,[] nineninesix/kani-tts-370m,kani-tts-370m,nineninesix,2025-09-30T07:31:22+00:00,8444,136,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'text-to-speech', 'en', 'de', 'ar', 'zh', 'es', 'ko', 'arxiv:2505.20506', 'base_model:nineninesix/kani-tts-450m-0.2-pt', 'base_model:finetune:nineninesix/kani-tts-450m-0.2-pt', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",10,797,Lfm2ForCausalLM,lfm2,283798528.0,128000.0,1024.0,16.0,16.0,80539.0,Language Model,"['Fine-tuned', 'Specialized']",['conversation'],"['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Legal']",{},75.0,Medium,0.75,2025-11-02T06:07:23.243490,"license: apache-2.0 language: - en - de - ar - zh - es - ko pipeline_tag: text-to-speech library_name: transformers base_model: - nineninesix/kani-tts-450m-0.2-pt

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" LiquidAI/LFM2-8B-A1B,LFM2-8B-A1B,LiquidAI,2025-10-07T13:55:39+00:00,13348,231,transformers,"['transformers', 'safetensors', 'lfm2_moe', 'text-generation', 'liquid', 'lfm2', 'edge', 'moe', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,807,Lfm2MoeForCausalLM,lfm2_moe,1342177280.0,128000.0,2048.0,32.0,24.0,65536.0,Language Model,"['Quantized', 'LoRA']","['text-generation', 'question-answering', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing', 'Healthcare']",{},88.0,Easy,0.88,2025-11-02T06:07:30.121317,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge - moe ...,"['Function Calling', 'RAG Support', 'Fast Inference', 'Multi-turn', 'Safety Aligned']" cerebras/Qwen3-Coder-REAP-25B-A3B,Qwen3-Coder-REAP-25B-A3B,cerebras,2025-10-20T15:40:03+00:00,971,32,transformers,"['transformers', 'qwen3_moe', 'text-generation', 'qwen-coder', 'MOE', 'pruning', 'compression', 'conversational', 'en', 'arxiv:2510.13999', 'base_model:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'base_model:finetune:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,808,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,262144.0,2048.0,32.0,48.0,151936.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'reasoning']","['English', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},85.0,Easy,0.85,2025-11-02T06:07:30.253734,language: - en library_name: transformers tags: - qwen-coder - MOE - pruning - compression license: apache-2.0 name: cerebras/Qwen3-Coder-REAP-25B-A3B description: > This model was obtained by uniform...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" meta-llama/Llama-3.2-1B-Instruct,Llama-3.2-1B-Instruct,meta-llama,2024-09-18T15:12:47+00:00,3843095,1141,transformers,"['transformers', 'safetensors', 'llama', 'text-generation', 'facebook', 'meta', 'pytorch', 'llama-3', 'conversational', 'en', 'de', 'fr', 'it', 'pt', 'hi', 'es', 'th', 'arxiv:2204.05149', 'arxiv:2405.16406', 'license:llama3.2', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,809,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:07:32.503941,No README available,[] microsoft/Phi-4-mini-instruct,Phi-4-mini-instruct,microsoft,2025-02-19T01:00:58+00:00,249315,623,transformers,"['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'ar', 'zh', 'cs', 'da', 'nl', 'en', 'fi', 'fr', 'de', 'he', 'hu', 'it', 'ja', 'ko', 'no', 'pl', 'pt', 'ru', 'es', 'sv', 'th', 'tr', 'uk', 'arxiv:2503.01743', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,810,Phi3ForCausalLM,phi3,4238475264.0,131072.0,3072.0,24.0,32.0,200064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Business', 'Healthcare', 'Legal']",{},86.0,Easy,0.86,2025-11-02T06:07:32.808225,language: - multilingual - ar - zh - cs - da - nl - en - fi - fr - de - he - hu - it - ja - ko - 'no' - pl - pt - ru - es - sv - th - tr - uk library_name: transformers license: mit license_link: http...,"['Function Calling', 'RAG Support', 'Long Context', 'Multi-turn', 'Safety Aligned']" Qwen/Qwen3-30B-A3B,Qwen3-30B-A3B,Qwen,2025-04-27T03:43:05+00:00,421099,808,transformers,"['transformers', 'safetensors', 'qwen3_moe', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-30B-A3B-Base', 'base_model:finetune:Qwen/Qwen3-30B-A3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",8,811,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,40960.0,2048.0,32.0,48.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:07:35.033318,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-30B-A3B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-30B-A3B-Base - If you want to learn more about how we process your perso...,[] TheBloke/Mistral-7B-Instruct-v0.2-GGUF,Mistral-7B-Instruct-v0.2-GGUF,TheBloke,2023-12-11T22:18:46+00:00,60731,475,transformers,"['transformers', 'gguf', 'mistral', 'finetuned', 'text-generation', 'arxiv:2310.06825', 'base_model:mistralai/Mistral-7B-Instruct-v0.2', 'base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2', 'license:apache-2.0', 'region:us', 'conversational']",6,826,Unknown,mistral,,,,,,,Language Model,"['Fine-tuned', 'Quantized']","['text-generation', 'question-answering', 'summarization', 'conversation']","['English', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Production', 'Creative Writing']",{},63.0,Hard,0.63,2025-11-02T06:07:46.779221,base_model: mistralai/Mistral-7B-Instruct-v0.2 inference: false license: apache-2.0 model_creator: Mistral AI_ model_name: Mistral 7B Instruct v0.2 model_type: mistral pipeline_tag: text-generation pr...,"['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-32B,Qwen3-32B,Qwen,2025-04-27T03:52:59+00:00,1610353,561,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",6,827,Qwen3ForCausalLM,qwen3,20910571520.0,40960.0,5120.0,64.0,64.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},61.0,Hard,0.61,2025-11-02T06:07:46.928794,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-32B/blob/main/LICENSE pipeline_tag: text-generation

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ inference: false fine-tuning: false langu...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" PokeeAI/pokee_research_7b,pokee_research_7b,PokeeAI,2025-10-17T20:38:37+00:00,5574,95,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'agent', 'deepresearch', 'llm', 'rl', 'reinforcementlearning', 'conversational', 'en', 'dataset:miromind-ai/MiroRL-GenQA', 'arxiv:2510.15862', 'base_model:Qwen/Qwen2.5-7B-Instruct', 'base_model:finetune:Qwen/Qwen2.5-7B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",24,867,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged', 'Specialized']","['text-generation', 'question-answering', 'summarization', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education', 'Healthcare', 'Legal', 'Finance']",{},78.0,Medium,0.78,2025-11-02T06:08:06.270542,base_model: - Qwen/Qwen2.5-7B-Instruct datasets: - miromind-ai/MiroRL-GenQA language: - en license: apache-2.0 tags: - agent - deepresearch - llm - rl - reinforcementlearning pipeline_tag: text-genera...,"['RAG Support', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" ibm-granite/granite-4.0-350m-base,granite-4.0-350m-base,ibm-granite,2025-10-07T20:25:24+00:00,495,23,transformers,"['transformers', 'safetensors', 'granitemoehybrid', 'text-generation', 'language', 'granite-4.0', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",23,868,GraniteMoeHybridForCausalLM,granitemoehybrid,455081984.0,32768.0,1024.0,16.0,28.0,100352.0,Language Model,"['LoRA', 'Specialized']","['text-generation', 'question-answering', 'text-classification', 'summarization']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Education']",{},86.0,Easy,0.86,2025-11-02T06:08:06.532795,license: apache-2.0 library_name: transformers tags: - language - granite-4.0 **Model Summary:** Granite-4.0-350M-Base is a lightweight decoder-only language model designed for scenarios where efficie...,"['Long Context', 'Fast Inference', 'Safety Aligned']" internlm/JanusCoder-14B,JanusCoder-14B,internlm,2025-10-27T09:34:49+00:00,165,23,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",23,869,Qwen3ForCausalLM,qwen3,13360824320.0,32768.0,5120.0,40.0,40.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},58.0,Hard,0.58,2025-11-02T06:08:06.715875,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] ByteDance/Ouro-2.6B,Ouro-2.6B,ByteDance,2025-10-28T22:19:46+00:00,227,21,transformers,"['transformers', 'safetensors', 'ouro', 'text-generation', 'looped-language-model', 'reasoning', 'recurrent-depth', 'conversational', 'custom_code', 'arxiv:2510.25741', 'license:apache-2.0', 'autotrain_compatible', 'region:us']",21,870,OuroForCausalLM,ouro,2516582400.0,65536.0,2048.0,16.0,48.0,49152.0,Language Model,[],"['text-generation', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education']",{},78.0,Medium,0.78,2025-11-02T06:08:06.912174,library_name: transformers license: apache-2.0 pipeline_tag: text-generation tags: - looped-language-model - reasoning - recurrent-depth ![Ouro Logo](assets/logo.png) [📚 Paper (Hugging Face)](https://...,"['RAG Support', 'Long Context']" HuggingFaceTB/SmolLM3-3B,SmolLM3-3B,HuggingFaceTB,2025-07-08T10:11:45+00:00,56521,768,transformers,"['transformers', 'safetensors', 'smollm3', 'text-generation', 'conversational', 'en', 'fr', 'es', 'it', 'pt', 'zh', 'ar', 'ru', 'base_model:HuggingFaceTB/SmolLM3-3B-Base', 'base_model:finetune:HuggingFaceTB/SmolLM3-3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",20,871,SmolLM3ForCausalLM,smollm3,2074607616.0,65536.0,2048.0,16.0,36.0,128256.0,Language Model,['Quantized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},83.0,Medium,0.83,2025-11-02T06:08:09.140947,library_name: transformers license: apache-2.0 language: - en - fr - es - it - pt - zh - ar - ru base_model: - HuggingFaceTB/SmolLM3-3B-Base ![image/png](https://cdn-uploads.huggingface.co/production/...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient']" Qwen/Qwen3-0.6B,Qwen3-0.6B,Qwen,2025-04-27T03:40:08+00:00,7327000,752,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-0.6B-Base', 'base_model:finetune:Qwen/Qwen3-0.6B-Base', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",19,872,Qwen3ForCausalLM,qwen3,507904000.0,40960.0,1024.0,16.0,28.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},78.0,Medium,0.78,2025-11-02T06:08:09.288942,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-0.6B-Base This model was obtained by uniformly pruning 25...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" Qwen/Qwen2.5-7B-Instruct,Qwen2.5-7B-Instruct,Qwen,2024-09-16T11:55:40+00:00,8169288,849,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-7B', 'base_model:finetune:Qwen/Qwen2.5-7B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",17,876,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},86.0,Easy,0.86,2025-11-02T06:08:10.177229,license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE language: - en pipeline_tag: text-generation base_model: Qwen/Qwen2.5-7B tags: - chat library_name: ...,"['Long Context', 'Fast Inference', 'Safety Aligned']" deepseek-ai/DeepSeek-R1,DeepSeek-R1,deepseek-ai,2025-01-20T03:46:07+00:00,425526,12820,transformers,"['transformers', 'safetensors', 'deepseek_v3', 'text-generation', 'conversational', 'custom_code', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'fp8', 'region:us']",17,877,DeepseekV3ForCausalLM,deepseek_v3,38537003008.0,163840.0,7168.0,128.0,61.0,129280.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged']","['text-generation', 'question-answering', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business']",{},48.0,Hard,0.48,2025-11-02T06:08:10.332094,"license: mit library_name: transformers
- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" microsoft/VibeVoice-1.5B,VibeVoice-1.5B,microsoft,2025-08-25T13:46:48+00:00,223609,1943,transformers,"['transformers', 'safetensors', 'vibevoice', 'text-generation', 'Podcast', 'text-to-speech', 'en', 'zh', 'arxiv:2508.19205', 'arxiv:2412.08635', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,896,VibeVoiceForConditionalGeneration,vibevoice,,,,,,,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business', 'Legal']",{},53.0,Hard,0.53,2025-11-02T06:08:28.944760,"language: - en - zh license: mit pipeline_tag: text-to-speech tags: - Podcast library_name: transformers VibeVoice is a novel framework designed for generating expressive, long-form, multi-speaker con...","['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" LiquidAI/LFM2-350M-Extract,LFM2-350M-Extract,LiquidAI,2025-09-03T17:13:53+00:00,5852,61,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'liquid', 'edge', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'base_model:LiquidAI/LFM2-350M', 'base_model:finetune:LiquidAI/LFM2-350M', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",12,897,Lfm2ForCausalLM,lfm2,268435456.0,128000.0,1024.0,16.0,16.0,65536.0,Language Model,['LoRA'],"['text-generation', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},70.0,Medium,0.7,2025-11-02T06:08:29.107744,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge base_m...,['Multi-turn'] tencent/HunyuanImage-3.0,HunyuanImage-3.0,tencent,2025-09-25T06:28:28+00:00,101407,947,transformers,"['transformers', 'safetensors', 'hunyuan_image_3_moe', 'text-generation', 'text-to-image', 'custom_code', 'arxiv:2509.23951', 'license:other', 'autotrain_compatible', 'region:us']",12,898,HunyuanImage3ForCausalMM,hunyuan_image_3_moe,6987710464.0,12800.0,4096.0,32.0,32.0,133120.0,Language Model,['RLHF'],"['text-generation', 'translation', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Creative Writing']",{},55.0,Hard,0.55,2025-11-02T06:08:29.245911,"license: other license_name: tencent-hunyuan-community license_link: LICENSE pipeline_tag: text-to-image library_name: transformers
[!NOTE] > GGUF uploads with our fixes. More details and [Read o...,"['Function Calling', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-Next-80B-A3B-Instruct,Qwen3-Next-80B-A3B-Instruct,Qwen,2025-09-09T15:40:56+00:00,1663221,845,transformers,"['transformers', 'safetensors', 'qwen3_next', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2404.06654', 'arxiv:2505.09388', 'arxiv:2501.15383', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",11,905,Qwen3NextForCausalLM,qwen3_next,2727084032.0,262144.0,2048.0,16.0,48.0,151936.0,Language Model,['Merged'],"['text-generation', 'question-answering', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:08:37.447860,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-Next-80B-A3B-Instruct/blob/main/LICENSE pipeline_tag: text-generation This model was obtained by uniformly pruning 4...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" internlm/JanusCoder-8B,JanusCoder-8B,internlm,2025-10-27T09:33:54+00:00,139,11,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",11,908,Qwen3ForCausalLM,qwen3,7870087168.0,32768.0,4096.0,32.0,36.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},68.0,Medium,0.68,2025-11-02T06:08:37.876663,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] unsloth/MiniMax-M2,MiniMax-M2,unsloth,2025-10-28T12:05:08+00:00,264,11,transformers,"['transformers', 'safetensors', 'minimax', 'text-generation', 'conversational', 'arxiv:2504.07164', 'arxiv:2509.06501', 'arxiv:2509.13160', 'license:mit', 'autotrain_compatible', 'endpoints_compatible', 'fp8', 'region:us']",11,909,MiniMaxM2ForCausalLM,minimax,7635861504.0,196608.0,3072.0,48.0,62.0,200064.0,Language Model,['LoRA'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production']",{},91.0,Easy,0.91,2025-11-02T06:08:38.087306,"pipeline_tag: text-generation license: mit library_name: transformers
<...,"['RAG Support', 'Multi-turn']" facebook/MobileLLM-Pro,MobileLLM-Pro,facebook,2025-09-10T18:40:06+00:00,4486,138,transformers,"['transformers', 'safetensors', 'llama4_text', 'text-generation', 'facebook', 'meta', 'pytorch', 'conversational', 'custom_code', 'en', 'base_model:facebook/MobileLLM-Pro-base', 'base_model:finetune:facebook/MobileLLM-Pro-base', 'license:fair-noncommercial-research-license', 'autotrain_compatible', 'region:us']",10,916,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:08:42.499813,No README available,[] nineninesix/kani-tts-370m,kani-tts-370m,nineninesix,2025-09-30T07:31:22+00:00,8444,136,transformers,"['transformers', 'safetensors', 'lfm2', 'text-generation', 'text-to-speech', 'en', 'de', 'ar', 'zh', 'es', 'ko', 'arxiv:2505.20506', 'base_model:nineninesix/kani-tts-450m-0.2-pt', 'base_model:finetune:nineninesix/kani-tts-450m-0.2-pt', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",10,917,Lfm2ForCausalLM,lfm2,283798528.0,128000.0,1024.0,16.0,16.0,80539.0,Language Model,"['Fine-tuned', 'Specialized']",['conversation'],"['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Legal']",{},75.0,Medium,0.75,2025-11-02T06:08:42.677657,"license: apache-2.0 language: - en - de - ar - zh - es - ko pipeline_tag: text-to-speech library_name: transformers base_model: - nineninesix/kani-tts-450m-0.2-pt

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ pipeline_tag: text-generation datasets: -...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" LiquidAI/LFM2-8B-A1B,LFM2-8B-A1B,LiquidAI,2025-10-07T13:55:39+00:00,13348,231,transformers,"['transformers', 'safetensors', 'lfm2_moe', 'text-generation', 'liquid', 'lfm2', 'edge', 'moe', 'conversational', 'en', 'ar', 'zh', 'fr', 'de', 'ja', 'ko', 'es', 'license:other', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,927,Lfm2MoeForCausalLM,lfm2_moe,1342177280.0,128000.0,2048.0,32.0,24.0,65536.0,Language Model,"['Quantized', 'LoRA']","['text-generation', 'question-answering', 'translation', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Creative Writing', 'Healthcare']",{},88.0,Easy,0.88,2025-11-02T06:08:49.721474,library_name: transformers license: other license_name: lfm1.0 license_link: LICENSE language: - en - ar - zh - fr - de - ja - ko - es pipeline_tag: text-generation tags: - liquid - lfm2 - edge - moe ...,"['Function Calling', 'RAG Support', 'Fast Inference', 'Multi-turn', 'Safety Aligned']" cerebras/Qwen3-Coder-REAP-25B-A3B,Qwen3-Coder-REAP-25B-A3B,cerebras,2025-10-20T15:40:03+00:00,971,32,transformers,"['transformers', 'qwen3_moe', 'text-generation', 'qwen-coder', 'MOE', 'pruning', 'compression', 'conversational', 'en', 'arxiv:2510.13999', 'base_model:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'base_model:finetune:Qwen/Qwen3-Coder-30B-A3B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",9,928,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,262144.0,2048.0,32.0,48.0,151936.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'reasoning']","['English', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Creative Writing']",{},85.0,Easy,0.85,2025-11-02T06:08:49.894040,language: - en library_name: transformers tags: - qwen-coder - MOE - pruning - compression license: apache-2.0 name: cerebras/Qwen3-Coder-REAP-25B-A3B description: > This model was obtained by uniform...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" meta-llama/Llama-3.2-1B-Instruct,Llama-3.2-1B-Instruct,meta-llama,2024-09-18T15:12:47+00:00,3843095,1141,transformers,"['transformers', 'safetensors', 'llama', 'text-generation', 'facebook', 'meta', 'pytorch', 'llama-3', 'conversational', 'en', 'de', 'fr', 'it', 'pt', 'hi', 'es', 'th', 'arxiv:2204.05149', 'arxiv:2405.16406', 'license:llama3.2', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,929,Unknown,unknown,,,,,,,General Language Model,[],['text-generation'],['English'],['General Purpose'],{},40.0,Critical,0.4,2025-11-02T06:08:51.191167,No README available,[] microsoft/Phi-4-mini-instruct,Phi-4-mini-instruct,microsoft,2025-02-19T01:00:58+00:00,249315,623,transformers,"['transformers', 'safetensors', 'phi3', 'text-generation', 'nlp', 'code', 'conversational', 'custom_code', 'multilingual', 'ar', 'zh', 'cs', 'da', 'nl', 'en', 'fi', 'fr', 'de', 'he', 'hu', 'it', 'ja', 'ko', 'no', 'pl', 'pt', 'ru', 'es', 'sv', 'th', 'tr', 'uk', 'arxiv:2503.01743', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",8,930,Phi3ForCausalLM,phi3,4238475264.0,131072.0,3072.0,24.0,32.0,200064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Production', 'Education', 'Business', 'Healthcare', 'Legal']",{},86.0,Easy,0.86,2025-11-02T06:08:51.372004,language: - multilingual - ar - zh - cs - da - nl - en - fi - fr - de - he - hu - it - ja - ko - 'no' - pl - pt - ru - es - sv - th - tr - uk library_name: transformers license: mit license_link: http...,"['Function Calling', 'RAG Support', 'Long Context', 'Multi-turn', 'Safety Aligned']" Qwen/Qwen3-30B-A3B,Qwen3-30B-A3B,Qwen,2025-04-27T03:43:05+00:00,421099,808,transformers,"['transformers', 'safetensors', 'qwen3_moe', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-30B-A3B-Base', 'base_model:finetune:Qwen/Qwen3-30B-A3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",8,931,Qwen3MoeForCausalLM,qwen3_moe,2727084032.0,40960.0,2048.0,32.0,48.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},86.0,Easy,0.86,2025-11-02T06:08:53.629210,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-30B-A3B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-30B-A3B-Base - If you want to learn more about how we process your perso...,[] TheBloke/Mistral-7B-Instruct-v0.2-GGUF,Mistral-7B-Instruct-v0.2-GGUF,TheBloke,2023-12-11T22:18:46+00:00,60731,475,transformers,"['transformers', 'gguf', 'mistral', 'finetuned', 'text-generation', 'arxiv:2310.06825', 'base_model:mistralai/Mistral-7B-Instruct-v0.2', 'base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2', 'license:apache-2.0', 'region:us', 'conversational']",6,946,Unknown,mistral,,,,,,,Language Model,"['Fine-tuned', 'Quantized']","['text-generation', 'question-answering', 'summarization', 'conversation']","['English', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Production', 'Creative Writing']",{},63.0,Hard,0.63,2025-11-02T06:09:05.269373,base_model: mistralai/Mistral-7B-Instruct-v0.2 inference: false license: apache-2.0 model_creator: Mistral AI_ model_name: Mistral 7B Instruct v0.2 model_type: mistral pipeline_tag: text-generation pr...,"['RAG Support', 'Long Context', 'Fast Inference', 'Multi-turn']" Qwen/Qwen3-32B,Qwen3-32B,Qwen,2025-04-27T03:52:59+00:00,1610353,561,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2309.00071', 'arxiv:2505.09388', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",6,947,Qwen3ForCausalLM,qwen3,20910571520.0,40960.0,5120.0,64.0,64.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},61.0,Hard,0.61,2025-11-02T06:09:05.400342,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-32B/blob/main/LICENSE pipeline_tag: text-generation

- https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/ inference: false fine-tuning: false langu...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" PokeeAI/pokee_research_7b,pokee_research_7b,PokeeAI,2025-10-17T20:38:37+00:00,5574,95,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'agent', 'deepresearch', 'llm', 'rl', 'reinforcementlearning', 'conversational', 'en', 'dataset:miromind-ai/MiroRL-GenQA', 'arxiv:2510.15862', 'base_model:Qwen/Qwen2.5-7B-Instruct', 'base_model:finetune:Qwen/Qwen2.5-7B-Instruct', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",24,987,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged', 'Specialized']","['text-generation', 'question-answering', 'summarization', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education', 'Healthcare', 'Legal', 'Finance']",{},78.0,Medium,0.78,2025-11-02T06:09:24.045792,base_model: - Qwen/Qwen2.5-7B-Instruct datasets: - miromind-ai/MiroRL-GenQA language: - en license: apache-2.0 tags: - agent - deepresearch - llm - rl - reinforcementlearning pipeline_tag: text-genera...,"['RAG Support', 'Fast Inference', 'Memory Efficient', 'Multi-turn', 'Safety Aligned']" ibm-granite/granite-4.0-350m-base,granite-4.0-350m-base,ibm-granite,2025-10-07T20:25:24+00:00,495,23,transformers,"['transformers', 'safetensors', 'granitemoehybrid', 'text-generation', 'language', 'granite-4.0', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",23,988,GraniteMoeHybridForCausalLM,granitemoehybrid,455081984.0,32768.0,1024.0,16.0,28.0,100352.0,Language Model,"['LoRA', 'Specialized']","['text-generation', 'question-answering', 'text-classification', 'summarization']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Research', 'Education']",{},86.0,Easy,0.86,2025-11-02T06:09:24.182033,license: apache-2.0 library_name: transformers tags: - language - granite-4.0 **Model Summary:** Granite-4.0-350M-Base is a lightweight decoder-only language model designed for scenarios where efficie...,"['Long Context', 'Fast Inference', 'Safety Aligned']" internlm/JanusCoder-14B,JanusCoder-14B,internlm,2025-10-27T09:34:49+00:00,165,23,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'image-text-to-text', 'conversational', 'arxiv:2510.23538', 'arxiv:2403.14734', 'arxiv:2510.09724', 'arxiv:2507.22080', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",23,989,Qwen3ForCausalLM,qwen3,13360824320.0,32768.0,5120.0,40.0,40.0,151936.0,Language Model,['Specialized'],"['text-generation', 'code-generation', 'conversation']","['English', 'Chinese', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']",['Research'],{},58.0,Hard,0.58,2025-11-02T06:09:24.594361,license: apache-2.0 pipeline_tag: image-text-to-text library_name: transformers [💻Github Repo](https://github.com/InternLM/JanusCoder) • [🤗Model Collections](https://huggingface.co/collections/internl...,['Safety Aligned'] ByteDance/Ouro-2.6B,Ouro-2.6B,ByteDance,2025-10-28T22:19:46+00:00,227,21,transformers,"['transformers', 'safetensors', 'ouro', 'text-generation', 'looped-language-model', 'reasoning', 'recurrent-depth', 'conversational', 'custom_code', 'arxiv:2510.25741', 'license:apache-2.0', 'autotrain_compatible', 'region:us']",21,990,OuroForCausalLM,ouro,2516582400.0,65536.0,2048.0,16.0,48.0,49152.0,Language Model,[],"['text-generation', 'reasoning']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Production', 'Education']",{},78.0,Medium,0.78,2025-11-02T06:09:25.149327,library_name: transformers license: apache-2.0 pipeline_tag: text-generation tags: - looped-language-model - reasoning - recurrent-depth ![Ouro Logo](assets/logo.png) [📚 Paper (Hugging Face)](https://...,"['RAG Support', 'Long Context']" HuggingFaceTB/SmolLM3-3B,SmolLM3-3B,HuggingFaceTB,2025-07-08T10:11:45+00:00,56521,768,transformers,"['transformers', 'safetensors', 'smollm3', 'text-generation', 'conversational', 'en', 'fr', 'es', 'it', 'pt', 'zh', 'ar', 'ru', 'base_model:HuggingFaceTB/SmolLM3-3B-Base', 'base_model:finetune:HuggingFaceTB/SmolLM3-3B-Base', 'license:apache-2.0', 'autotrain_compatible', 'endpoints_compatible', 'region:us']",20,991,SmolLM3ForCausalLM,smollm3,2074607616.0,65536.0,2048.0,16.0,36.0,128256.0,Language Model,['Quantized'],"['text-generation', 'question-answering', 'summarization', 'code-generation', 'reasoning', 'conversation']","['English', 'Chinese', 'Korean', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},83.0,Medium,0.83,2025-11-02T06:09:27.355254,library_name: transformers license: apache-2.0 language: - en - fr - es - it - pt - zh - ar - ru base_model: - HuggingFaceTB/SmolLM3-3B-Base ![image/png](https://cdn-uploads.huggingface.co/production/...,"['RAG Support', 'Long Context', 'Fast Inference', 'Memory Efficient']" Qwen/Qwen3-0.6B,Qwen3-0.6B,Qwen,2025-04-27T03:40:08+00:00,7327000,752,transformers,"['transformers', 'safetensors', 'qwen3', 'text-generation', 'conversational', 'arxiv:2505.09388', 'base_model:Qwen/Qwen3-0.6B-Base', 'base_model:finetune:Qwen/Qwen3-0.6B-Base', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",19,992,Qwen3ForCausalLM,qwen3,507904000.0,40960.0,1024.0,16.0,28.0,151936.0,Language Model,[],"['text-generation', 'question-answering', 'translation', 'code-generation', 'reasoning', 'conversation']","['English', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']","['Production', 'Creative Writing']",{},78.0,Medium,0.78,2025-11-02T06:09:32.776449,"library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-0.6B/blob/main/LICENSE pipeline_tag: text-generation base_model: - Qwen/Qwen3-0.6B-Base This model was obtained by uniformly pruning 25...,"['Function Calling', 'Fast Inference', 'Memory Efficient', 'Multi-turn']" Qwen/Qwen2.5-7B-Instruct,Qwen2.5-7B-Instruct,Qwen,2024-09-16T11:55:40+00:00,8169288,849,transformers,"['transformers', 'safetensors', 'qwen2', 'text-generation', 'chat', 'conversational', 'en', 'arxiv:2309.00071', 'arxiv:2407.10671', 'base_model:Qwen/Qwen2.5-7B', 'base_model:finetune:Qwen/Qwen2.5-7B', 'license:apache-2.0', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'region:us']",17,996,Qwen2ForCausalLM,qwen2,4860936192.0,32768.0,3584.0,28.0,28.0,152064.0,Language Model,['Specialized'],"['text-generation', 'question-answering', 'code-generation', 'conversation']","['English', 'Chinese', 'Korean', 'Japanese', 'Spanish', 'French', 'German', 'Russian', 'Arabic', 'Multilingual']",['Production'],{},86.0,Easy,0.86,2025-11-02T06:09:33.809068,license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen2.5-7B-Instruct/blob/main/LICENSE language: - en pipeline_tag: text-generation base_model: Qwen/Qwen2.5-7B tags: - chat library_name: ...,"['Long Context', 'Fast Inference', 'Safety Aligned']" deepseek-ai/DeepSeek-R1,DeepSeek-R1,deepseek-ai,2025-01-20T03:46:07+00:00,425526,12820,transformers,"['transformers', 'safetensors', 'deepseek_v3', 'text-generation', 'conversational', 'custom_code', 'arxiv:2501.12948', 'license:mit', 'autotrain_compatible', 'text-generation-inference', 'endpoints_compatible', 'fp8', 'region:us']",17,997,DeepseekV3ForCausalLM,deepseek_v3,38537003008.0,163840.0,7168.0,128.0,61.0,129280.0,Language Model,"['Fine-tuned', 'RLHF', 'Merged']","['text-generation', 'question-answering', 'summarization', 'reasoning', 'conversation']","['English', 'Chinese', 'Spanish', 'French', 'German', 'Russian', 'Arabic']","['Research', 'Education', 'Business']",{},48.0,Hard,0.48,2025-11-02T06:09:34.069626,"license: mit library_name: transformers