_id string | id string | author string | baseModels dict | downloads int64 | downloads_all_time int64 | gated string | created_at timestamp[us, tz=UTC] | last_modified timestamp[us, tz=UTC] | library_name string | likes int64 | trending_score float64 | model_index string | pipeline_tag string | safetensors string | siblings list | sizes list | total_size int64 | sha string | tags list | gguf string | card string | spaces list | licenses list | datasets list | languages list | safetensors_params float64 | gguf_params float64 | tasks list | metrics list | architectures list | modalities list | input_modalities list | output_modalities list | org_model string | org_type string | org_country list | a_gated string | a_baseModels string | a_input_modalities list | a_output_modalities list | a_architectures list | a_languages list | a_training_methods list | a_ddpa string | annotator int64 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
68ac69484a1f0871ddf555e4 | microsoft/VibeVoice-1.5B | microsoft | null | 87,188 | 87,188 | False | 2025-08-25T13:46:48Z | 2025-08-28T04:57:59Z | null | 1,117 | 1,117 | null | text-to-speech | {"parameters": {"BF16": 2704021985}, "total": 2704021985} | [
".gitattributes",
"README.md",
"config.json",
"figures/Fig1.png",
"model-00001-of-00003.safetensors",
"model-00002-of-00003.safetensors",
"model-00003-of-00003.safetensors",
"model.safetensors.index.json",
"preprocessor_config.json"
] | [
1603,
7273,
2762,
153971,
1975317828,
1983051688,
1449832938,
122616,
351
] | 5,408,491,030 | cf42b8ff262f8a286bcbe580835cfaad62d277ca | [
"safetensors",
"vibevoice",
"Podcast",
"text-to-speech",
"en",
"zh",
"arxiv:2508.19205",
"arxiv:2412.08635",
"license:mit",
"region:us"
] | null | ## VibeVoice: A Frontier Open-Source Text-to-Speech Model
VibeVoice is a novel framework designed for generating expressive, long-form, multi-speaker conversational audio, such as podcasts, from text. It addresses significant challenges in traditional Text-to-Speech (TTS) systems, particularly in scalability, speaker ... | [
"broadfield-dev/VibeVoice-demo",
"yasserrmd/VibeVoice",
"broadfield-dev/VibeVoice-demo-dev",
"akhaliq/VibeVoice-1.5B",
"mrfakename/VibeVoice-1.5B",
"NeuralFalcon/VibeVoice-Colab",
"thelip/VibeVoice",
"ReallyFloppyPenguin/VibeVoice-demo",
"Xenobd/VibeVoice-demo",
"Dorjzodovsuren/VibeVoice",
"umin... | [
"mit"
] | null | [
"en",
"zh"
] | 2,704,021,985 | null | [
"text-to-speech"
] | null | [
"VibeVoiceForConditionalGeneration",
"vibevoice"
] | [
"audio"
] | [
"text"
] | [
"audio"
] | free | company | [
"United States of America",
"International",
"India",
"Belgium"
] | null | null | null | null | null | null | null | null | null |
68aaebfbfe684542cfc51e66 | openbmb/MiniCPM-V-4_5 | openbmb | null | 9,706 | 9,706 | False | 2025-08-24T10:39:55Z | 2025-08-31T14:57:14Z | transformers | 747 | 747 | null | image-text-to-text | {"parameters": {"BF16": 8695895280}, "total": 8695895280} | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"configuration_minicpm.py",
"generation_config.json",
"image_processing_minicpmv.py",
"merges.txt",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-000... | [
1570,
24775,
2862,
1461,
3288,
268,
20757,
1671853,
5286612176,
5301855088,
4546851120,
2256571800,
72172,
17754,
41835,
714,
11026,
11732,
12103,
1647,
11437868,
25786,
2776833
] | 17,408,026,488 | 17353d11601386fac6cca5a541e84b85928bd4ae | [
"transformers",
"safetensors",
"minicpmv",
"feature-extraction",
"minicpm-v",
"vision",
"ocr",
"multi-image",
"video",
"custom_code",
"image-text-to-text",
"conversational",
"multilingual",
"dataset:openbmb/RLAIF-V-Dataset",
"arxiv:2403.11703",
"region:us"
] | null | <h1>A GPT-4o Level MLLM for Single Image, Multi Image and High-FPS Video Understanding on Your Phone</h1>
[GitHub](https://github.com/OpenBMB/MiniCPM-o) | [Demo](http://101.126.42.235:30910/)</a>
## MiniCPM-V 4.5
**MiniCPM-V 4.5** is the latest and most capable model in the MiniCPM-V series. The model is built on... | [
"akhaliq/MiniCPM-V-4_5",
"orrzxz/MiniCPM-V-4_5",
"WYC-2025/MiniCPM-V-4_5",
"CGQN/MiniCPM-V-4_5",
"CGQN/MiniCPM-V-4_5-from_gpt5",
"CGQN/MiniCPM-V-4_5-CPU-0"
] | null | [
"openbmb/RLAIF-V-Dataset"
] | [
"multilingual"
] | 8,695,895,280 | null | [
"feature-extraction",
"image-text-to-text"
] | null | [
"modeling_minicpmv.MiniCPMV",
"MiniCPMV",
"AutoModel",
"minicpmv"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"embeddings",
"text"
] | free | community | [
"China"
] | null | null | null | null | null | null | null | null | null |
68a8de283195d5730fd2c5b8 | xai-org/grok-2 | xai-org | null | 4,047 | 4,047 | False | 2025-08-22T21:16:24Z | 2025-08-24T00:59:56Z | null | 879 | 485 | null | null | null | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"pytorch_model-00000-TP-common.safetensors",
"pytorch_model-00001-TP-common.safetensors",
"pytorch_model-00002-TP-common.safetensors",
"pytorch_model-00003-TP-common.safetensors",
"pytorch_model-00004-TP-common.safetensors",
"pytorch_model-0... | [
1519,
5362,
1583,
947,
2147483760,
2147483744,
16472,
34359745872,
34359745872,
34359745744,
17179936544,
17179936544,
17179936544,
17179936544,
17179936544,
17179936544,
17179936544,
17179936544,
17179936544,
17179936544,
17179936544,
17179936544,
17179936544,
17179936... | 539,040,431,560 | d60cbe267db8bb43be676bc80e200c64268ea8ec | [
"git",
"region:us"
] | null | # Grok 2
This repository contains the weights of Grok 2, a model trained and used at xAI in 2024.
## Usage: Serving with SGLang
- Download the weights. You can replace `/local/grok-2` with any other folder name you prefer.
```
hf download xai-org/grok-2 --local-dir /local/grok-2
```
You might encounter som... | [
"umint/o4-mini",
"AnilNiraula/FinChat",
"umint/gpt-4.1-nano",
"umint/o3"
] | null | null | null | null | null | null | null | [
"Grok1ForCausalLM",
"git"
] | null | null | null | team | company | [
"United States of America"
] | null | null | null | null | null | null | null | null | null |
68a19381db43c983deb63fa5 | Qwen/Qwen-Image-Edit | Qwen | null | 75,516 | 75,516 | False | 2025-08-17T08:32:01Z | 2025-08-25T04:41:11Z | diffusers | 1,545 | 359 | null | image-to-image | null | [
".gitattributes",
"README.md",
"model_index.json",
"processor/added_tokens.json",
"processor/chat_template.jinja",
"processor/merges.txt",
"processor/preprocessor_config.json",
"processor/special_tokens_map.json",
"processor/tokenizer.json",
"processor/tokenizer_config.json",
"processor/video_pr... | [
1580,
11747,
512,
605,
1017,
1671853,
788,
613,
11421896,
4727,
904,
2776833,
485,
3217,
244,
4968243304,
4991495816,
4932751040,
1691924384,
57655,
605,
2427,
1671853,
613,
4686,
3383407,
339,
4989364312,
4984214160,
4946470000,
4984213736,
4946471896,
... | 57,720,467,613 | ac7f9318f633fc4b5778c59367c8128225f1e3de | [
"diffusers",
"safetensors",
"image-to-image",
"en",
"zh",
"arxiv:2508.02324",
"license:apache-2.0",
"diffusers:QwenImageEditPipeline",
"region:us"
] | null | <p align="center">
<img src="https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen-Image/qwen_image_edit_logo.png" width="400"/>
<p>
<p align="center">
💜 <a href="https://chat.qwen.ai/"><b>Qwen Chat</b></a>   |   🤗 <a href="https://huggingface.co/Qwen/Qwen-Image-Edit">Hugging Face</a>&nbs... | [
"multimodalart/Qwen-Image-Edit-Fast",
"Qwen/Qwen-Image-Edit",
"zerogpu-aoti/Qwen-Image-Edit-Relight",
"zerogpu-aoti/Qwen-Image-Edit-Outpaint",
"llamameta/nano-banana-experimental",
"zerogpu-aoti/Qwen-Image-Edit-Multi-Image",
"bep40/Nano-Banana",
"LPX55/Qwen-Image-Edit_Fast-Presets",
"VirtualKimi/Nan... | [
"apache-2.0"
] | null | [
"en",
"zh"
] | null | null | [
"image-to-image"
] | null | null | [
"vision"
] | [
"image"
] | [
"image"
] | team | company | [
"China"
] | null | null | null | null | null | null | null | null | null |
68abccbf1935e46075b39df2 | Wan-AI/Wan2.2-S2V-14B | Wan-AI | null | 9,959 | 9,959 | False | 2025-08-25T02:38:55Z | 2025-08-28T02:36:24Z | diffusers | 197 | 197 | null | null | null | [
".gitattributes",
"README.md",
"Wan2.1_VAE.pth",
"assets/471504690-b63bfa58-d5d7-4de6-a1a2-98970b06d9a7.mp4",
"assets/comp_effic.png",
"assets/logo.png",
"assets/moe_2.png",
"assets/moe_arch.png",
"assets/performance.png",
"assets/vae.png",
"config.json",
"configuration.json",
"diffusion_pyt... | [
1300,
18697,
507609880,
9193286,
202156,
56322,
527914,
74900,
306535,
165486,
890,
43,
9968229352,
9891539248,
9956985634,
2774887624,
113150,
6623,
4548313,
16837417,
61728,
11361920418,
2328,
36,
5327,
200,
1531,
86,
6198,
1261905572,
1372,
78,
8629... | 49,148,819,983 | eff0178482d4d6e1fed7763f6c3b3f480be908c0 | [
"diffusers",
"safetensors",
"s2v",
"arxiv:2503.20314",
"arxiv:2508.18621",
"license:apache-2.0",
"region:us"
] | null | # Wan2.2
<p align="center">
<img src="assets/logo.png" width="400"/>
<p>
<p align="center">
💜 <a href="https://wan.video"><b>Wan</b></a>    |    🖥️ <a href="https://github.com/Wan-Video/Wan2.2">GitHub</a>    |   🤗 <a href="https://huggingface.co/Wan-AI/">Hugging Face</a>&nb... | [
"Wan-AI/Wan2.2-S2V",
"mjinabq/Wan2.2-S2V",
"opparco/Wan2.2-S2V",
"ItsMpilo/Wan2.2-S2V"
] | [
"apache-2.0"
] | null | null | null | null | null | null | [
"s2v"
] | null | null | null | free | company | [
"China"
] | null | null | null | null | null | null | null | null | null |
688a4ad0a0c7bbd72715e857 | Phr00t/WAN2.2-14B-Rapid-AllInOne | Phr00t | {
"models": [
{
"_id": "6881e60ffcffaee6d84fe9e4",
"id": "Wan-AI/Wan2.2-I2V-A14B"
}
],
"relation": "finetune"
} | 0 | 0 | False | 2025-07-30T16:39:44Z | 2025-08-23T23:51:11Z | wan2.2 | 494 | 166 | null | image-to-video | null | [
".gitattributes",
"README.md",
"v2/wan2.2-i2v-aio-v2.safetensors",
"v2/wan2.2-t2v-aio-v2.safetensors",
"v3/wan2.2-i2v-rapid-aio-540p-v3.safetensors",
"v3/wan2.2-i2v-rapid-aio-720p-v3.safetensors",
"v3/wan2.2-t2v-rapid-aio-v3.safetensors",
"v4/wan2.2-i2v-rapid-aio-v4.safetensors",
"v4/wan2.2-t2v-rapi... | null | null | 6c7be992d665858c886ad1c7791b7a83db2478c1 | [
"wan2.2",
"wan",
"accelerator",
"image-to-video",
"base_model:Wan-AI/Wan2.2-I2V-A14B",
"base_model:finetune:Wan-AI/Wan2.2-I2V-A14B",
"region:us"
] | null | These are mixtures of WAN 2.2 and other WAN-like models and accelerators (with CLIP and VAE also included) to provide a fast, "all in one" solution for making videos as easily and quickly as possible. FP8 precision. Generally the latest version available for each type of model (image to video or text to video) is recom... | null | null | null | null | null | null | [
"image-to-video"
] | null | null | [
"vision"
] | [
"text",
"image"
] | [
"video"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68a686808e8db90f8998697a | deepseek-ai/DeepSeek-V3.1 | deepseek-ai | null | 76,644 | 76,644 | False | 2025-08-21T02:37:52Z | 2025-08-26T08:14:11Z | transformers | 668 | 163 | null | text-generation | {"parameters": {"BF16": 3918786560, "F8_E4M3": 680571043840, "F32": 41555600}, "total": 684531386000} | [
".gitattributes",
"LICENSE",
"README.md",
"assets/chat_template.jinja",
"assets/code_agent_trajectory.html",
"assets/search_python_tool_trajectory.html",
"assets/search_tool_trajectory.html",
"config.json",
"configuration_deepseek.py",
"generation_config.json",
"model-00001-of-000163.safetensors... | [
1519,
1084,
11296,
3330,
22659,
19652,
10272,
1686,
9897,
171,
5234139343,
4302383966,
4302384375,
4302349996,
4302384154,
4372073602,
4306080097,
4302384356,
4302350190,
4302383960,
4302384375,
1321583941,
4302317244,
4302384328,
4302350218,
4302383932,
430238437... | 688,603,634,706 | 9e6c48c3fa6bb3e1cf684675dc02e813ca45d20f | [
"transformers",
"safetensors",
"deepseek_v3",
"text-generation",
"conversational",
"custom_code",
"arxiv:2412.19437",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"fp8",
"region:us"
] | null | # DeepSeek-V3.1
<!-- markdownlint-disable first-line-h1 -->
<!-- markdownlint-disable html -->
<!-- markdownlint-disable no-duplicate-header -->
<div align="center">
<img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" />
</div>
<hr>
<div align="cen... | [
"enzostvs/deepsite",
"umint/ai",
"ReallyFloppyPenguin/DeepSeek-V3.1-Superintell",
"nazdridoy/inferoxy-hub",
"Humbl3m33/deepseek-ai-DeepSeek-V3.1",
"umint/o4-mini",
"Xavernox/Orionixlabs-ai-DeepSeek-V3.1",
"KhushParikh/deepseek-ai-DeepSeek-V3.1",
"birde2003/for4-ai-Seek-V3.1",
"HgThazh/chat",
"yz... | [
"mit"
] | null | null | 684,531,386,000 | null | [
"text-generation"
] | null | [
"DeepseekV3ForCausalLM",
"deepseek_v3",
"AutoModelForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | free | company | [
"China"
] | null | null | null | null | null | null | null | null | null |
68913539bd3d0a833438591d | openai/gpt-oss-20b | openai | null | 8,811,370 | 8,811,370 | False | 2025-08-04T22:33:29Z | 2025-08-26T17:25:47Z | transformers | 3,342 | 126 | null | text-generation | {"parameters": {"BF16": 1804459584, "U8": 19707494400}, "total": 21511953984} | [
".gitattributes",
"LICENSE",
"README.md",
"USAGE_POLICY",
"chat_template.jinja",
"config.json",
"generation_config.json",
"metal/model.bin",
"model-00000-of-00002.safetensors",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"original/c... | [
1570,
11357,
7095,
200,
16738,
1806,
177,
13750886400,
4792272488,
4798702184,
4170342232,
36355,
376,
13082,
13761300984,
98,
27868174,
4200
] | 41,301,465,516 | 6cee5e81ee83917806bbde320786a8fb61efebee | [
"transformers",
"safetensors",
"gpt_oss",
"text-generation",
"vllm",
"conversational",
"arxiv:2508.10925",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"8-bit",
"mxfp4",
"region:us"
] | null | <p align="center">
<img alt="gpt-oss-20b" src="https://raw.githubusercontent.com/openai/gpt-oss/main/docs/gpt-oss-20b.svg">
</p>
<p align="center">
<a href="https://gpt-oss.com"><strong>Try gpt-oss</strong></a> ·
<a href="https://cookbook.openai.com/topic/gpt-oss"><strong>Guides</strong></a> ·
<a href="https:/... | [
"umint/ai",
"ArthT/openai-gpt-oss-20b",
"MGZON/mgzon-app",
"SustainabilityLabIITGN/VayuChat",
"merterbak/gpt-oss-20b-demo",
"fastrtc/talk-to-oai-gpt-oss-20b",
"fdaudens/gpt-oss-news-agent",
"mAI-models/m-4.0",
"Kunal444/KunalGPT",
"Paulwalker4884/Nursa",
"DESTINY21/mychabot",
"Ansjsn/litert-co... | [
"apache-2.0"
] | null | null | 21,511,953,984 | null | [
"text-generation"
] | null | [
"GptOssForCausalLM",
"AutoModelForCausalLM",
"gpt_oss"
] | [
"text"
] | [
"text"
] | [
"text"
] | enterprise | company | [
"United States of America"
] | null | null | null | null | null | null | null | null | null |
68b159467c2485b297655f40 | meituan-longcat/LongCat-Flash-Chat | meituan-longcat | null | 9 | 9 | False | 2025-08-29T07:39:50Z | 2025-08-31T09:12:12Z | LongCat-Flash-Chat | 126 | 126 | null | text-generation | {"parameters": {"BF16": 561730738176, "F32": 132142080}, "total": 561862880256} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"configuration_longcat_flash.py",
"generation_config.json",
"model.safetensors.index.json",
"model_00001-of-00075.safetensors",
"model_00002-of-00075.safetensors",
"model_00003-of-00075.safetensors",
"model_00004-of-00075.safetensors",
"... | null | null | 6d2d483a1112bce151bcba600d84329c40eb72dd | [
"LongCat-Flash-Chat",
"safetensors",
"text-generation",
"transformers",
"conversational",
"custom_code",
"license:mit",
"region:us"
] | null | # LongCat-Flash-Chat
<div align="center">
<img src="https://raw.githubusercontent.com/meituan-longcat/LongCat-Flash-Chat/main/figures/longcat_logo.svg"
width="300"
alt="LongCat Logo"/>
</div>
<hr>
<div align="center" style="line-height: 1;">
<a href="https://longcat.ai/" target="_blank" style="m... | null | [
"mit"
] | null | null | 561,862,880,256 | null | [
"text-generation"
] | null | [
"LongcatFlashForCausalLM",
"AutoModelForCausalLM",
"modeling_longcat_flash.LongcatFlashForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68913522f16f3c8aaffccf1f | openai/gpt-oss-120b | openai | null | 2,333,920 | 2,333,920 | False | 2025-08-04T22:33:06Z | 2025-08-26T17:25:03Z | transformers | 3,669 | 113 | null | text-generation | {"parameters": {"BF16": 2167371072, "U8": 118244966400}, "total": 120412337472} | [
".gitattributes",
"LICENSE",
"README.md",
"USAGE_POLICY",
"chat_template.jinja",
"config.json",
"generation_config.json",
"metal/model.bin",
"model-00000-of-00014.safetensors",
"model-00001-of-00014.safetensors",
"model-00002-of-00014.safetensors",
"model-00003-of-00014.safetensors",
"model-... | [
1570,
11357,
7111,
201,
16738,
2089,
177,
65238253568,
4625017896,
4115586736,
4625017888,
4115586752,
4625017896,
4115586696,
4625017856,
4060267176,
4625017896,
4170906304,
4625017896,
4115586752,
4064660808,
4625017896,
4115586736,
54511,
377,
19658,
1054404068... | 195,764,040,609 | b5c939de8f754692c1647ca79fbf85e8c1e70f8a | [
"transformers",
"safetensors",
"gpt_oss",
"text-generation",
"vllm",
"conversational",
"arxiv:2508.10925",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"8-bit",
"mxfp4",
"region:us"
] | null | <p align="center">
<img alt="gpt-oss-120b" src="https://raw.githubusercontent.com/openai/gpt-oss/main/docs/gpt-oss-120b.svg">
</p>
<p align="center">
<a href="https://gpt-oss.com"><strong>Try gpt-oss</strong></a> ·
<a href="https://cookbook.openai.com/topic/gpt-oss"><strong>Guides</strong></a> ·
<a href="https... | [
"amd/gpt-oss-120b-chatbot",
"umint/ai",
"MGZON/mgzon-app",
"SustainabilityLabIITGN/VayuChat",
"fdaudens/gpt-oss-news-agent",
"Arphd4/ARK.AI",
"Wenxi123/openai-gpt-oss-120b",
"nazdridoy/inferoxy-hub",
"Humbl3m33/openai-gpt-oss-120b",
"umint/o4-mini",
"openfree/OpenAI-gpt-oss",
"ginipick/FLUXlla... | [
"apache-2.0"
] | null | null | 120,412,337,472 | null | [
"text-generation"
] | null | [
"GptOssForCausalLM",
"AutoModelForCausalLM",
"gpt_oss"
] | [
"text"
] | [
"text"
] | [
"text"
] | enterprise | company | [
"United States of America"
] | null | null | null | null | null | null | null | null | null |
68b0017230f0a85ce3af1186 | stepfun-ai/Step-Audio-2-mini | stepfun-ai | null | 567 | 567 | False | 2025-08-28T07:12:50Z | 2025-08-29T10:53:00Z | null | 111 | 111 | null | null | {"parameters": {"BF16": 8315179264}, "total": 8315179264} | [
".gitattributes",
"README.md",
"added_tokens.json",
"assets/architecture5.png",
"assets/arxiv.svg",
"assets/logo.png",
"assets/qrcode.jpg",
"assets/radar.png",
"assets/wechat_group.jpg",
"config.json",
"configuration_step_audio_2.py",
"merges.txt",
"model-00001-of-00004.safetensors",
"mode... | null | null | 6da8a9a58eb45f0448dffa63411be5cb410cfb70 | [
"onnx",
"safetensors",
"step_audio_2",
"custom_code",
"arxiv:2507.16632",
"license:apache-2.0",
"region:us"
] | null | <div align="center">
<img src="assets/logo.png" height=100>
</div>
<div align="center" style="line-height: 1;">
<a href="https://github.com/stepfun-ai/Step-Audio2" target="_blank"><img alt="GitHub" src="https://img.shields.io/badge/GitHub-StepFun-white?logo=github&logoColor=white"/></a>  
<a href="https://... | [
"Steveeeeeeen/Step-Audio-2-mini",
"reach-vb/Step-Audio-2-mini"
] | [
"apache-2.0"
] | null | null | 8,315,179,264 | null | null | null | [
"step_audio_2",
"StepAudio2ForCausalLM"
] | null | null | null | team | company | [
"China"
] | null | null | null | null | null | null | null | null | null |
68ac80cfba9b1e4d828e0fe5 | OpenGVLab/InternVL3_5-241B-A28B | OpenGVLab | {
"models": [
{
"_id": "68ac918a70eb335713fa922c",
"id": "OpenGVLab/InternVL3_5-241B-A28B-MPO"
}
],
"relation": "finetune"
} | 2,458 | 2,458 | False | 2025-08-25T15:27:11Z | 2025-08-29T17:57:02Z | transformers | 108 | 108 | null | image-text-to-text | {"parameters": {"BF16": 240699370368}, "total": 240699370368} | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.jinja",
"config.json",
"configuration_intern_vit.py",
"configuration_internvl_chat.py",
"conversation.py",
"examples/image1.jpg",
"examples/image2.jpg",
"examples/red-panda.mp4",
"generation_config.json",
"images/.DS_Store",
... | [
2801,
53717,
892,
475,
2682,
5546,
4700,
15309,
78073,
125656,
1867237,
69,
6148,
294354,
80900,
345911,
71966,
156486,
581610,
828601,
265469,
634719,
226506,
314378,
650028,
958889,
479663,
289495,
336405,
496867,
342672,
529283,
73983,
1671853,
... | 481,433,952,804 | 8e5cfda1dabab9bdc76a3931bf77ecae150a33c8 | [
"transformers",
"safetensors",
"internvl_chat",
"feature-extraction",
"internvl",
"custom_code",
"image-text-to-text",
"conversational",
"multilingual",
"dataset:OpenGVLab/MMPR-v1.2",
"dataset:OpenGVLab/MMPR-Tiny",
"arxiv:2312.14238",
"arxiv:2404.16821",
"arxiv:2412.05271",
"arxiv:2411.1... | null | # InternVL3_5-241B-A28B
[\[📂 GitHub\]](https://github.com/OpenGVLab/InternVL) [\[📜 InternVL 1.0\]](https://huggingface.co/papers/2312.14238) [\[📜 InternVL 1.5\]](https://huggingface.co/papers/2404.16821) [\[📜 InternVL 2.5\]](https://huggingface.co/papers/2412.05271) [\[📜 InternVL2.5-MPO\]](https://huggingface... | null | [
"apache-2.0"
] | [
"OpenGVLab/MMPR-v1.2",
"OpenGVLab/MMPR-Tiny"
] | [
"multilingual"
] | 240,699,370,368 | null | [
"feature-extraction",
"image-text-to-text"
] | null | [
"modeling_internvl_chat.InternVLChatModel",
"AutoModel",
"InternVLChatModel",
"internvl_chat"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"embeddings",
"text"
] | free | community | [
"China"
] | null | null | null | null | null | null | null | null | null |
68a34925f057ffe5051b5b6b | NousResearch/Hermes-4-70B | NousResearch | {
"models": [
{
"_id": "66944fd095c7fa6e68c314ae",
"id": "meta-llama/Llama-3.1-70B"
}
],
"relation": "finetune"
} | 2,387 | 2,387 | False | 2025-08-18T15:39:17Z | 2025-08-26T18:44:49Z | transformers | 96 | 96 | [{"name": "Hermes-4-Llama-3.1-70B", "results": []}] | text-generation | {"parameters": {"BF16": 70553706496}, "total": 70553706496} | [
".gitattributes",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00030.safetensors",
"model-00002-of-00030.safetensors",
"model-00003-of-00030.safetensors",
"model-00004-of-00030.safetensors",
"model-00005-of-00030.safetensors",
"model-00006-of-00030... | [
1570,
9258,
4121,
840,
180,
4584408808,
4664167376,
4999711704,
4966157032,
4664134408,
4664167408,
4664167408,
4999711728,
4966157056,
4664134408,
4664167408,
4664167408,
4999711728,
4966157056,
4664134408,
4664167408,
4664167408,
4999711728,
4966157056,
4664134408,
... | 141,124,834,214 | 8635c3f88cc33e405a8ad297f05fb4f33042a533 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"Llama-3.1",
"instruct",
"finetune",
"reasoning",
"hybrid-mode",
"chatml",
"function calling",
"tool use",
"json mode",
"structured outputs",
"atropos",
"dataforge",
"long context",
"roleplaying",
"chat",
"conversationa... | null | # Hermes 4 — Llama-3.1 70B

## Model Description
Hermes 4 70B is a frontier, hybrid-mode **reasoning** model based on Llama-3.1-70B by Nous Research that is aligned to **you**.
Read the Hermes 4 t... | [
"ReallyFloppyPenguin/NousResearch-Hermes-4-70B"
] | [
"llama3"
] | null | [
"en"
] | 70,553,706,496 | null | [
"text-generation"
] | null | [
"llama",
"AutoModelForCausalLM",
"LlamaForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | team | company | [
"Online"
] | null | null | null | null | null | null | null | null | null |
688d9adf9f62ee5c9a3804eb | Qwen/Qwen-Image | Qwen | null | 182,088 | 182,088 | False | 2025-08-02T04:58:07Z | 2025-08-18T02:42:19Z | diffusers | 1,933 | 92 | null | text-to-image | null | [
".gitattributes",
"LICENSE",
"README.md",
"model_index.json",
"scheduler/scheduler_config.json",
"text_encoder/config.json",
"text_encoder/generation_config.json",
"text_encoder/model-00001-of-00004.safetensors",
"text_encoder/model-00002-of-00004.safetensors",
"text_encoder/model-00003-of-00004.s... | [
1519,
11343,
6778,
443,
485,
3217,
244,
4968243304,
4991495816,
4932751040,
1691924384,
57655,
605,
2427,
1671853,
613,
4686,
3383407,
371,
4989364312,
4984214160,
4946470000,
4984213736,
4946471896,
4946451560,
4908690520,
4984232856,
1170918840,
198887,
73... | 57,704,594,653 | 75e0b4be04f60ec59a75f475837eced720f823b6 | [
"diffusers",
"safetensors",
"text-to-image",
"en",
"zh",
"arxiv:2508.02324",
"license:apache-2.0",
"diffusers:QwenImagePipeline",
"region:us"
] | null | <p align="center">
<img src="https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen-Image/qwen_image_logo.png" width="400"/>
<p>
<p align="center">
💜 <a href="https://chat.qwen.ai/"><b>Qwen Chat</b></a>   |   🤗 <a href="https://huggingface.co/Qwen/Qwen-Image">Hugging Face</a>   | &... | [
"Qwen/Qwen-Image",
"multimodalart/Qwen-Image-Fast",
"InstantX/Qwen-Image-ControlNet",
"multimodalart/Qwen-Image-LoRA-Explorer",
"Heartsync/Qwen-Image-LORA",
"instaagent/Qwen-Image-Fast-8steps",
"prithivMLmods/Qwen-Image-LoRA-DLC",
"prithivMLmods/Qwen-Image-Diffusion",
"Arphd4/ARK.AI",
"nazdridoy/i... | [
"apache-2.0"
] | null | [
"en",
"zh"
] | null | null | [
"text-to-image"
] | null | null | [
"vision"
] | [
"text"
] | [
"image"
] | team | company | [
"China"
] | null | null | null | null | null | null | null | null | null |
68a5e3be37edead1bf9642f1 | ByteDance-Seed/Seed-OSS-36B-Instruct | ByteDance-Seed | null | 15,080 | 15,080 | False | 2025-08-20T15:03:26Z | 2025-08-26T02:33:00Z | transformers | 372 | 89 | null | text-generation | {"parameters": {"BF16": 36151104512}, "total": 36151104512} | [
".gitattributes",
"LICENSE.txt",
"MODEL_CARD.md",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00015.safetensors",
"model-00002-of-00015.safetensors",
"model-00003-of-00015.safetensors",
"model-00004-of-00015.safetensors",
"model-00005-of-00015.s... | [
1626,
11356,
4829,
19944,
7705,
770,
172,
4954686296,
4991407840,
4834167328,
4886550176,
4834167360,
4886550176,
4834167360,
4886550176,
4834167360,
4886550176,
4834167360,
4886550176,
4834167360,
4886550176,
4031898896,
63285,
432,
190266,
11883696,
23849
] | 72,314,506,146 | 497f1dca95ebdec98e41d517b9f060ee753c902f | [
"transformers",
"safetensors",
"seed_oss",
"text-generation",
"vllm",
"conversational",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | <div align="center">
👋 Hi, everyone!
<br>
We are <b>ByteDance Seed Team.</b>
</div>
<p align="center">
You can get to know us better through the following channels👇
<br>
<a href="https://seed.bytedance.com/">
<img src="https://img.shields.io/badge/Website-%231e37ff?style=for-the-badge&logo=bytedan... | [
"umint/o4-mini",
"umint/gpt-4.1-nano",
"umint/o3"
] | [
"apache-2.0"
] | null | null | 36,151,104,512 | null | [
"text-generation"
] | null | [
"AutoModelForCausalLM",
"seed_oss",
"SeedOssForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68a6c4929160f033a5f10c57 | tencent/HunyuanVideo-Foley | tencent | null | 577 | 577 | False | 2025-08-21T07:02:42Z | 2025-08-27T13:04:08Z | hunyuanvideo-foley | 89 | 89 | null | text-to-audio | null | [
".gitattributes",
"LICENSE",
"NOTICE",
"README.md",
"assets/data_pipeline.png",
"assets/logo.png",
"assets/model_arch.png",
"assets/pan_chart.png",
"config.yaml",
"hunyuanvideo_foley.pth",
"synchformer_state_dict.pth",
"vae_128d_48k.pth"
] | null | null | f3fda473b027296b5680981b10657d599763d5c9 | [
"hunyuanvideo-foley",
"text-to-audio",
"video-to-audio",
"text-video-to-audio",
"en",
"zh",
"arxiv:2508.16930",
"license:other",
"region:us"
] | null | null | [
"tencent/HunyuanVideo-Foley",
"Bils/ShortiFoley",
"svjack/HunyuanVideo-Foley"
] | [
"other",
"tencent-hunyuan-community",
"https://huggingface.co/tencent/HunyuanVideo-Foley/blob/main/LICENSE"
] | null | [
"en",
"zh"
] | null | null | [
"text-to-audio"
] | null | null | [
"text"
] | [
"text"
] | [
"audio"
] | free | null | [
"China"
] | null | null | null | null | null | null | null | null | null |
68acd0e86c89708a5657c8ca | WestZhang/VibeVoice-Large-pt | WestZhang | null | 16,019 | 16,019 | False | 2025-08-25T21:08:56Z | 2025-08-25T22:00:21Z | null | 80 | 80 | null | null | {"parameters": {"BF16": 9343355361}, "total": 9343355361} | [
".gitattributes",
"README.md",
"config.json",
"model-00001-of-00010.safetensors",
"model-00002-of-00010.safetensors",
"model-00003-of-00010.safetensors",
"model-00004-of-00010.safetensors",
"model-00005-of-00010.safetensors",
"model-00006-of-00010.safetensors",
"model-00007-of-00010.safetensors",
... | [
1519,
156,
2786,
1886424044,
1864468520,
1864468520,
1864468544,
1864468568,
1864468568,
1864468568,
1972552744,
1959739938,
1681341960,
122675,
349
] | 18,686,997,459 | 0b68ee6da8ca6bca98484758d06cbe9c33f49e7b | [
"safetensors",
"vibevoice",
"region:us"
] | null | null | null | null | null | null | 9,343,355,361 | null | null | null | [
"VibeVoiceForConditionalGeneration",
"vibevoice"
] | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68aec35be2c318d75ca7cd3c | bytedance-research/USO | bytedance-research | {
"models": [
{
"_id": "66aaa908fc35e079a941470d",
"id": "black-forest-labs/FLUX.1-dev"
}
],
"relation": "finetune"
} | 128 | 128 | False | 2025-08-27T08:35:39Z | 2025-08-31T09:01:17Z | transformers | 79 | 79 | null | text-to-image | null | [
".gitattributes",
"README.md",
"assets/teaser.webp",
"assets/uso.webp",
"config.json",
"uso_flux_v1.0/dit_lora.safetensors",
"uso_flux_v1.0/projector.safetensors"
] | null | null | b745e66613531e71fd84a4e66120c16b88e670d2 | [
"transformers",
"image-generation",
"subject-personalization",
"style-transfer",
"Diffusion-Transformer",
"text-to-image",
"en",
"arxiv:2508.18966",
"base_model:black-forest-labs/FLUX.1-dev",
"base_model:finetune:black-forest-labs/FLUX.1-dev",
"license:apache-2.0",
"endpoints_compatible",
"r... | null | <p align="center">
<img src="assets/uso.webp" width="100"/>
<p>
<h3 align="center">
Unified Style and Subject-Driven Generation via Disentangled and Reward Learning
</h3>
Paper: [USO: Unified Style and Subject-Driven Generation via Disentangled and Reward Learning](https://huggingface.co/papers/2508.18966)
<p... | [
"bytedance-research/USO",
"bep40/USO",
"svjack/USO"
] | [
"apache-2.0"
] | null | [
"en"
] | null | null | [
null,
"text-to-image"
] | null | [
"AutoModel"
] | [
"vision"
] | [
"text"
] | [
"image"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
689252773b8900ddb9116aed | google/gemma-3-270m | google | null | 112,750 | 112,768 | manual | 2025-08-05T18:50:31Z | 2025-08-14T07:35:01Z | transformers | 702 | 78 | null | text-generation | {"parameters": {"BF16": 268098176}, "total": 268098176} | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"generation_config.json",
"model.safetensors",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer.model",
"tokenizer_config.json"
] | [
1570,
28276,
35,
1352,
133,
536223056,
662,
33384570,
4689074,
1155375
] | 575,484,103 | 9b0cfec892e2bc2afd938c98eabe4e4a7b1e0ca1 | [
"transformers",
"safetensors",
"gemma3_text",
"text-generation",
"gemma3",
"gemma",
"google",
"arxiv:2503.19786",
"arxiv:1905.07830",
"arxiv:1905.10044",
"arxiv:1911.11641",
"arxiv:1705.03551",
"arxiv:1911.01547",
"arxiv:1907.10641",
"arxiv:2311.07911",
"arxiv:2311.12022",
"arxiv:241... | null | null | [
"Fraser/web-chat",
"hari7261/Super-text-generation",
"umint/o4-mini",
"Fraser/piclets",
"simata/webui",
"ReallyFloppyPenguin/NanoAISuperHub",
"Pranav9605/AI_Travel_Rihla",
"hingep/apartmint-llm",
"AIPretender/AIPDF",
"ShahzebKhoso/Gamm3_270M_Chat",
"ByteMeHarder-404/gemma_chatbot",
"JJflying/g... | [
"gemma"
] | null | null | 268,098,176 | null | [
"text-generation"
] | null | [
"gemma3_text",
"Gemma3ForCausalLM",
"AutoModelForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | enterprise | company | [
"United States of America"
] | null | null | null | null | null | null | null | null | null |
68ac97aacd5cacefdd04852f | apple/FastVLM-0.5B | apple | null | 2,039 | 2,039 | False | 2025-08-25T17:04:42Z | 2025-08-29T17:22:08Z | ml-fastvlm | 77 | 77 | null | text-generation | {"parameters": {"BF16": 758833760}, "total": 758833760} | [
".gitattributes",
"LICENSE",
"README.md",
"acc_vs_latency_qwen-2.png",
"added_tokens.json",
"config.json",
"generation_config.json",
"llava_qwen.py",
"merges.txt",
"model.safetensors",
"special_tokens_map.json",
"tokenizer_config.json",
"trainer_state.json",
"training_args.bin",
"vocab.j... | [
1581,
5820,
4527,
228925,
80,
1404,
100,
82304,
1670344,
1517793184,
367,
1325,
6817275,
6584,
3383407
] | 1,529,997,227 | 139ff37252cf3d85ffbf7bc02436bbe0c6e5cc72 | [
"ml-fastvlm",
"safetensors",
"llava_qwen2",
"text-generation",
"transformers",
"conversational",
"custom_code",
"arxiv:2412.13303",
"license:apple-amlr",
"region:us"
] | null | # FastVLM: Efficient Vision Encoding for Vision Language Models
FastVLM was introduced in
**[FastVLM: Efficient Vision Encoding for Vision Language Models](https://www.arxiv.org/abs/2412.13303). (CVPR 2025)**
[//]: # ()
<p align="center">
<img src="acc_vs_lat... | [
"akhaliq/FastVLM-0.5B-gradio",
"akhaliq/FastVLM-0.5B-stream-gradio"
] | [
"apple-amlr",
"apple-ascl",
"https://github.com/apple/ml-fastvlm/blob/main/LICENSE_MODEL"
] | null | null | 758,833,760 | null | [
"text-generation"
] | null | [
"LlavaQwen2ForCausalLM",
"llava_qwen.LlavaQwen2ForCausalLM",
"llava_qwen2",
"AutoModelForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | free | null | [
"United States of America"
] | null | null | null | null | null | null | null | null | null |
68a44b5c81d6dff37afe329f | deepseek-ai/DeepSeek-V3.1-Base | deepseek-ai | null | 22,672 | 22,672 | False | 2025-08-19T10:01:00Z | 2025-08-26T08:15:20Z | transformers | 955 | 71 | null | text-generation | {"parameters": {"BF16": 3918786560, "F8_E4M3": 680571043840, "F32": 41555600}, "total": 684531386000} | [
".gitattributes",
"LICENSE",
"README.md",
"assets/chat_template.jinja",
"assets/code_agent_trajectory.html",
"assets/search_python_tool_trajectory.html",
"assets/search_tool_trajectory.html",
"config.json",
"configuration_deepseek.py",
"generation_config.json",
"model-00001-of-000163.safetensors... | [
1519,
1084,
11296,
3330,
22659,
19652,
10272,
1686,
9897,
171,
5234139343,
4302383966,
4302384375,
4302349996,
4302384154,
4372073602,
4306080097,
4302384356,
4302350190,
4302383960,
4302384375,
1321583941,
4302317244,
4302384328,
4302350218,
4302383932,
430238437... | 688,603,634,706 | d3d4eafdc470de44bbf6f0a74f852eb522357be8 | [
"transformers",
"safetensors",
"deepseek_v3",
"text-generation",
"conversational",
"custom_code",
"arxiv:2412.19437",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"fp8",
"region:us"
] | null | # DeepSeek-V3.1
<!-- markdownlint-disable first-line-h1 -->
<!-- markdownlint-disable html -->
<!-- markdownlint-disable no-duplicate-header -->
<div align="center">
<img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" />
</div>
<hr>
<div align="cen... | [
"umint/ai",
"Arphd4/ARK.AI",
"umint/o4-mini",
"qualybittech/bharath",
"ajay5364747/Ajay",
"juliusNice/deepseekv3.1",
"Xavernox/deepseek-ai-DeepSeek-V3.1-Base",
"fokan/train-modle",
"umint/gpt-4.1-nano",
"umint/o3"
] | [
"mit"
] | null | null | 684,531,386,000 | null | [
"text-generation"
] | null | [
"DeepseekV3ForCausalLM",
"deepseek_v3",
"AutoModelForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | free | company | [
"China"
] | null | null | null | null | null | null | null | null | null |
68ac9838d3bca8f4ccc00251 | apple/FastVLM-7B | apple | null | 427 | 427 | False | 2025-08-25T17:07:04Z | 2025-08-29T17:22:02Z | ml-fastvlm | 68 | 68 | null | text-generation | {"parameters": {"BF16": 7764588000}, "total": 7764588000} | [
".gitattributes",
"LICENSE",
"README.md",
"acc_vs_latency_qwen-2.png",
"added_tokens.json",
"config.json",
"generation_config.json",
"llava_qwen.py",
"merges.txt",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-0000... | null | null | 15f26f4273f3adfbefc39db5ef2cb993103a2282 | [
"ml-fastvlm",
"safetensors",
"llava_qwen2",
"text-generation",
"transformers",
"conversational",
"custom_code",
"arxiv:2412.13303",
"license:apple-amlr",
"region:us"
] | null | # FastVLM: Efficient Vision Encoding for Vision Language Models
FastVLM was introduced in
**[FastVLM: Efficient Vision Encoding for Vision Language Models](https://www.arxiv.org/abs/2412.13303). (CVPR 2025)**
[//]: # ()
<p align="center">
<img src="acc_vs_lat... | null | [
"apple-amlr",
"apple-ascl",
"https://github.com/apple/ml-fastvlm/blob/main/LICENSE_MODEL"
] | null | null | 7,764,588,000 | null | [
"text-generation"
] | null | [
"LlavaQwen2ForCausalLM",
"llava_qwen.LlavaQwen2ForCausalLM",
"llava_qwen2",
"AutoModelForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | free | null | [
"United States of America"
] | null | null | null | null | null | null | null | null | null |
66aaa908fc35e079a941470d | black-forest-labs/FLUX.1-dev | black-forest-labs | null | 1,371,856 | 21,186,921 | auto | 2024-07-31T21:13:44Z | 2025-06-27T16:22:19Z | diffusers | 11,311 | 65 | null | text-to-image | null | [
".gitattributes",
"LICENSE.md",
"README.md",
"ae.safetensors",
"dev_grid.jpg",
"flux1-dev.safetensors",
"model_index.json",
"scheduler/scheduler_config.json",
"text_encoder/config.json",
"text_encoder/model.safetensors",
"text_encoder_2/config.json",
"text_encoder_2/model-00001-of-00002.safete... | [
1655,
18621,
4394,
335304388,
1301528,
23802932552,
536,
273,
613,
246144352,
782,
4994582224,
4530066360,
19885,
524619,
588,
705,
1059962,
2543,
791656,
2424235,
20817,
378,
9983040304,
9949328904,
3870584832,
121262,
820,
167666902
] | 57,885,946,690 | 3de623fc3c33e44ffbe2bad470d0f45bccf2eb21 | [
"diffusers",
"safetensors",
"text-to-image",
"image-generation",
"flux",
"en",
"license:other",
"endpoints_compatible",
"diffusers:FluxPipeline",
"region:us"
] | null | null | [
"bytedance-research/USO",
"black-forest-labs/FLUX.1-dev",
"jasperai/Flux.1-dev-Controlnet-Upscaler",
"black-forest-labs/FLUX.1-Krea-dev",
"multimodalart/flux-lora-the-explorer",
"prithivMLmods/FLUX-REALISM",
"ameerazam08/FLUX.1-dev-Inpainting-Model-Beta-GPU",
"Nymbo/Serverless-ImgGen-Hub",
"Yuanshi/... | [
"other",
"flux-1-dev-non-commercial-license",
"LICENSE.md"
] | null | [
"en"
] | null | null | [
"text-to-image"
] | null | null | [
"vision"
] | [
"text"
] | [
"image"
] | team | company | [
"Germany"
] | Other agreement/info requirements | null | [
"Text"
] | [
"Image Generation"
] | [
"Diffusion-based Network"
] | [
"en"
] | [
"Knowledge distillation"
] | Not disclosed | 6 |
689bc3944b86fdeb37e9a08d | nvidia/NVIDIA-Nemotron-Nano-9B-v2 | nvidia | {
"models": [
{
"_id": "68a677062fc955d300dc18f9",
"id": "nvidia/NVIDIA-Nemotron-Nano-12B-v2"
}
],
"relation": "finetune"
} | 59,361 | 59,361 | False | 2025-08-12T22:43:32Z | 2025-08-30T01:41:18Z | transformers | 280 | 64 | null | text-generation | {"parameters": {"BF16": 8888227328}, "total": 8888227328} | [
".gitattributes",
"README.md",
"acc-vs-budget.png",
"accuracy_chart.png",
"bias.md",
"config.json",
"configuration_nemotron_h.py",
"explainability.md",
"generation_config.json",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model... | [
1679,
47188,
491034,
169781,
2277,
1557,
12176,
2635,
158,
4924823528,
4937507160,
4871563216,
3042598608,
26843,
78798,
3723,
2297,
2300,
422,
17078330,
181326
] | 17,794,595,036 | dc376c20a64208fc2cb4667e00af485eeced8ae4 | [
"transformers",
"safetensors",
"nvidia",
"pytorch",
"text-generation",
"conversational",
"en",
"es",
"fr",
"de",
"it",
"ja",
"dataset:nvidia/Nemotron-Post-Training-Dataset-v1",
"dataset:nvidia/Nemotron-Post-Training-Dataset-v2",
"dataset:nvidia/Nemotron-Pretraining-Dataset-sample",
"da... | null | # NVIDIA-Nemotron-Nano-9B-v2

**Model Developer:** NVIDIA Corporation
**Model Dates:**
June 2025 \- August 2025
**Data Freshness:**
September 2024
The pretraining data has a cutoff date of September 2024.
## Model Overview
NVIDIA-Nemotron-Nano-9B-v2 is a large language model (LLM) tra... | [
"akhaliq/NVIDIA-Nemotron-Nano-9B-v2"
] | [
"other",
"nvidia-open-model-license",
"https://www.nvidia.com/en-us/agreements/enterprise-software/nvidia-open-model-license/"
] | [
"nvidia/Nemotron-Post-Training-Dataset-v1",
"nvidia/Nemotron-Post-Training-Dataset-v2",
"nvidia/Nemotron-Pretraining-Dataset-sample",
"nvidia/Nemotron-CC-v2",
"nvidia/Nemotron-CC-Math-v1",
"nvidia/Nemotron-Pretraining-SFT-v1"
] | [
"en",
"es",
"fr",
"de",
"it",
"ja"
] | 8,888,227,328 | null | [
null,
"text-generation"
] | null | [
"AutoModel"
] | [
"text"
] | [
"text"
] | [
"text"
] | enterprise_plus | company | [
"United States of America"
] | null | null | null | null | null | null | null | null | null |
689761ce345c5cafa2ebc6a6 | lightx2v/Qwen-Image-Lightning | lightx2v | {
"models": [
{
"_id": "688d9adf9f62ee5c9a3804eb",
"id": "Qwen/Qwen-Image"
}
],
"relation": "finetune"
} | 498,394 | 498,394 | False | 2025-08-09T14:57:18Z | 2025-08-24T12:17:13Z | diffusers | 306 | 59 | null | text-to-image | null | [
".gitattributes",
"Qwen-Image-Edit-Lightning-4steps-V1.0-bf16.safetensors",
"Qwen-Image-Edit-Lightning-4steps-V1.0.safetensors",
"Qwen-Image-Edit-Lightning-8steps-V1.0-bf16.safetensors",
"Qwen-Image-Edit-Lightning-8steps-V1.0.safetensors",
"Qwen-Image-Lightning-4steps-V1.0-bf16.safetensors",
"Qwen-Image... | [
1519,
849608296,
1698951104,
849608296,
1698951104,
849608296,
1698951104,
1698951104,
849608296,
1698951104,
1919
] | 11,893,192,142 | 430a8879074ce23ac1e2784f778401c97ac2fee7 | [
"diffusers",
"Qwen-Image;",
"distillation;",
"LoRA",
"text-to-image",
"en",
"zh",
"base_model:Qwen/Qwen-Image",
"base_model:finetune:Qwen/Qwen-Image",
"license:apache-2.0",
"region:us"
] | null | Please refer to [Qwen-Image-Lightning github](https://github.com/ModelTC/Qwen-Image-Lightning/) to learn how to use the models.
use with diffusers 🧨:
make sure to install diffusers from `main` (`pip install git+https://github.com/huggingface/diffusers.git`)
```
from diffusers import DiffusionPipeline, FlowMatchEuler... | [
"multimodalart/Qwen-Image-Edit-Fast",
"zerogpu-aoti/Qwen-Image-Edit-Relight",
"zerogpu-aoti/Qwen-Image-Edit-Outpaint",
"multimodalart/Qwen-Image-Fast",
"zerogpu-aoti/Qwen-Image-Edit-Multi-Image",
"bep40/Nano-Banana",
"LPX55/Qwen-Image-Edit_Fast-Presets",
"multimodalart/Qwen-Image-LoRA-Explorer",
"Vi... | [
"apache-2.0"
] | null | [
"en",
"zh"
] | null | null | [
"text-to-image"
] | null | null | [
"vision"
] | [
"text"
] | [
"image"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
6895d02b321732d7f2a35ad9 | lodestones/Chroma1-HD | lodestones | null | 38,700 | 38,700 | False | 2025-08-08T10:23:39Z | 2025-08-23T12:33:55Z | diffusers | 155 | 56 | null | text-to-image | null | [
".gitattributes",
"Chroma1-HD.safetensors",
"README.md",
"images/FictionalChromaBanner_1.png",
"model_index.json",
"scheduler/scheduler_config.json",
"text_encoder/config.json",
"text_encoder/model-00001-of-00002.safetensors",
"text_encoder/model-00002-of-00002.safetensors",
"text_encoder/model.sa... | [
1660,
17800038288,
7983,
1008192,
494,
482,
741,
4994582224,
4530066360,
19921,
2593,
2543,
791656,
20847,
490,
9946193392,
7853894360,
106695,
819,
167666902
] | 45,294,406,642 | ed274863c4b9e1bba19e191d4e941778d2139232 | [
"diffusers",
"safetensors",
"text-to-image",
"license:apache-2.0",
"diffusers:ChromaPipeline",
"region:us"
] | null | # Chroma1-HD
Chroma1-HD is an **8.9B** parameter text-to-image foundational model based on **FLUX.1-schnell**. It is fully **Apache 2.0 licensed**, ensuring that anyone can use, modify, and build upon it.
As a **base model**, Chroma1 is intentionally designed to be an excellent starting point for **finetuning**. It p... | [
"multimodalart/Chroma1-HD",
"gokaygokay/Chroma"
] | [
"apache-2.0"
] | null | null | null | null | [
"text-to-image"
] | null | null | [
"vision"
] | [
"text"
] | [
"image"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
67be03c930eecba21c83a91e | Kijai/WanVideo_comfy | Kijai | {
"models": [
{
"_id": "6822f6e25843aa0767e04828",
"id": "Wan-AI/Wan2.1-VACE-1.3B"
}
],
"relation": "finetune"
} | 4,247,109 | 5,829,604 | False | 2025-02-25T17:54:17Z | 2025-08-22T11:44:43Z | diffusion-single-file | 1,312 | 55 | null | null | null | [
".gitattributes",
"CineScale/README.md",
"CineScale/Wan2.1_I2V_14B_CineScale_ntk20_lora_rank16_fp16.safetensors",
"CineScale/Wan2.1_T2V_1.3B_CineScale_ntk20_lora_rank16_fp16.safetensors",
"CineScale/Wan2.1_T2V_14B_CineScale_ntk20_lora_rank16_fp16.safetensors",
"EchoShot/Wan2_1-T2V-1-3B-EchoShot_fp16.safet... | [
1672,
151,
153445960,
43813344,
153445960,
2838276200,
358479336,
2046114748,
1253192432,
163327688,
630697104,
9999659744,
660874456,
3129105448,
16595124448,
16595124448,
16594632928,
16594632928,
17648319713,
16595462552,
5125258232,
5124439112,
758,
2555119088,
14... | 1,089,376,430,378 | 471420e5eb7df39671327e69b68f6fb1e3920dcb | [
"diffusion-single-file",
"comfyui",
"base_model:Wan-AI/Wan2.1-VACE-1.3B",
"base_model:finetune:Wan-AI/Wan2.1-VACE-1.3B",
"region:us"
] | null | Combined and quantized models for WanVideo, originating from here:
https://huggingface.co/Wan-AI/
Can be used with: https://github.com/kijai/ComfyUI-WanVideoWrapper and ComfyUI native WanVideo nodes.
I've also started to do fp8_scaled versions over here: https://huggingface.co/Kijai/WanVideo_comfy_fp8_scaled
Other ... | [
"zerogpu-aoti/wan2-2-fp8da-aoti-faster",
"zerogpu-aoti/wan2-2-fp8da-aoti",
"multimodalart/wan2-1-fast",
"ginigen/Nano-Banana-Video",
"alexnasa/OmniAvatar",
"ginigen/VEO3-Free",
"jbilcke-hf/InstaVideo",
"ginigen/VEO3-Free-mirror",
"Heartsync/wan2-1-fast-security",
"rahul7star/wan-fusionx-lora",
"... | null | null | null | null | null | null | null | null | null | null | null | user | user | [
"user"
] | null | Wan-AI/Wan2.1-VACE-1.3B | [
"Image"
] | [
"Video Generation"
] | [
"Transformer: Image Encoder-Decoder"
] | null | [
"Finetuning: Supervised",
" Quantization"
] | Not disclosed | 7 |
68a2e834fdfab51069736146 | MeiGen-AI/InfiniteTalk | MeiGen-AI | null | 0 | 0 | False | 2025-08-18T08:45:40Z | 2025-08-19T02:03:00Z | null | 102 | 55 | null | null | null | [
".gitattributes",
"README.md",
"assets/logo2.jpg",
"comfyui/infinitetalk_multi.safetensors",
"comfyui/infinitetalk_single.safetensors",
"multi/infinitetalk.safetensors",
"quant_models/infinitetalk_multi_fp8.json",
"quant_models/infinitetalk_multi_fp8.safetensors",
"quant_models/infinitetalk_multi_fp... | null | null | 527d6cadd8d066d70b3625f00a15f1dea5b8a6fc | [
"en",
"zh",
"license:apache-2.0",
"region:us"
] | null | <p align="center">
<img src="assets/logo2.jpg" alt="InfiniteTalk" width="500"/>
</p>
# InfiniteTalk: Audio-driven Video Generation for Sparse-Frame Video Dubbing
<p align="left">
<a href="">
<img
src="https://img.shields.io/badge/InfiniteTalk-Website-0A66C2?logo=safari&logoColor=white" style="display: i... | null | [
"apache-2.0"
] | null | [
"en",
"zh"
] | null | null | null | null | null | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68378cef5cbef05290b4d045 | black-forest-labs/FLUX.1-Kontext-dev | black-forest-labs | null | 420,319 | 887,866 | auto | 2025-05-28T22:23:43Z | 2025-06-27T21:55:46Z | diffusers | 2,208 | 54 | null | image-to-image | null | [
".gitattributes",
"LICENSE.md",
"README.md",
"ae.safetensors",
"flux1-kontext-dev.safetensors",
"model_index.json",
"scheduler/scheduler_config.json",
"teaser.png",
"text_encoder/config.json",
"text_encoder/model.safetensors",
"text_encoder_2/config.json",
"text_encoder_2/model-00001-of-00002.... | [
1656,
18621,
9640,
335304388,
23802947360,
688,
486,
6171062,
561,
246144352,
741,
4994582224,
4530066360,
19885,
524619,
588,
735,
1059962,
2543,
791656,
2424235,
20847,
452,
9983040304,
9949328904,
3870584832,
121262,
914,
167666902
] | 57,890,836,779 | af58063aa431f4d2bbc11ae46f57451d4416a170 | [
"diffusers",
"safetensors",
"image-generation",
"flux",
"diffusion-single-file",
"image-to-image",
"en",
"arxiv:2506.15742",
"license:other",
"diffusers:FluxKontextPipeline",
"region:us"
] | null | null | [
"black-forest-labs/FLUX.1-Kontext-Dev",
"umint/ai",
"gparmar/Group-Inference-FLUX.1-Kontext",
"kontext-community/kontext-relight",
"kontext-community/FLUX.1-Kontext-multi-image",
"Nymbo/FLUX.1-Kontext-Dev",
"AlekseyCalvin/fast-Kontext-Flux-LoRAs-bySilverAgePoets",
"fffiloni/reachy-mini-doll",
"Sahil... | [
"other",
"flux-1-dev-non-commercial-license",
"LICENSE.md"
] | null | [
"en"
] | null | null | [
"image-to-image"
] | null | null | [
"vision"
] | [
"image"
] | [
"image"
] | team | company | [
"Germany"
] | Accept to share username & email | enhanceaiteam/FLUX.1-Pro | [
"Image",
" Text"
] | [
"Image Generation"
] | [
"Diffusion-based Network"
] | [
"en"
] | [
"Knowledge distillation",
" Instruction finetuning"
] | Not disclosed | 3 |
676ca1388118866906abbd7c | hexgrad/Kokoro-82M | hexgrad | {
"models": [
{
"_id": "655c07aeceb07624c6e64c54",
"id": "yl4579/StyleTTS2-LJSpeech"
}
],
"relation": "finetune"
} | 2,352,148 | 13,084,911 | False | 2024-12-26T00:20:08Z | 2025-04-10T18:12:48Z | null | 4,960 | 52 | null | text-to-speech | null | [
".gitattributes",
"DONATE.md",
"EVAL.md",
"README.md",
"SAMPLES.md",
"VOICES.md",
"config.json",
"eval/ArtificialAnalysis-2025-02-26.jpeg",
"eval/TTS_Arena-2025-02-26.jpeg",
"eval/TTS_Spaces_Arena-2025-02-26.jpeg",
"kokoro-v1_0.pth",
"samples/HEARME.wav",
"samples/af_heart_0.wav",
"samples... | [
1913,
2562,
534,
6348,
5956,
7625,
2351,
939504,
560066,
515255,
327212226,
996044,
237644,
517244,
496844,
1407644,
1116044,
1033244,
523425,
523425,
523425,
523425,
523435,
523420,
523430,
523420,
523425,
523425,
523351,
523420,
523420,
523420,
52343... | 363,323,757 | f3ff3571791e39611d31c381e3a41a3af07b4987 | [
"text-to-speech",
"en",
"arxiv:2306.07691",
"arxiv:2203.02395",
"base_model:yl4579/StyleTTS2-LJSpeech",
"base_model:finetune:yl4579/StyleTTS2-LJSpeech",
"doi:10.57967/hf/4329",
"license:apache-2.0",
"region:us"
] | null | **Kokoro** is an open-weight TTS model with 82 million parameters. Despite its lightweight architecture, it delivers comparable quality to larger models while being significantly faster and more cost-efficient. With Apache-licensed weights, Kokoro can be deployed anywhere from production environments to personal projec... | [
"hexgrad/Kokoro-TTS",
"TTS-AGI/TTS-Arena-V2",
"m-ric/open-notebooklm",
"Inferless/Open-Source-TTS-Gallary",
"Pendrokar/TTS-Spaces-Arena",
"Steveeeeeeen/SpeechLLM-Playbook",
"TTS-AGI/TTS-Arena",
"NeuralFalcon/Kokoro-TTS-Subtitle",
"Nymbo/Tools",
"nazdridoy/inferoxy-hub",
"aiqtech/Open-Source-TTS-... | [
"apache-2.0"
] | null | [
"en"
] | null | null | [
"text-to-speech"
] | null | null | [
"audio"
] | [
"text"
] | [
"audio"
] | user | user | [
"user"
] | null | yl4579/StyleTTS2-LJSpeech | [
"Text"
] | [
"Speech Generation"
] | [
"Transformer: Text Decoder-only",
" Transformer: Speech Decoder-only"
] | [
"en",
" ja",
" zh",
" es",
" fr",
" hi",
" it",
" pt"
] | [
"Finetuning: Supervised"
] | Partially disclosed: unavailable | 6 |
68ac918db9bc800b65f56cf6 | OpenGVLab/InternVL3_5-GPT-OSS-20B-A4B-Preview | OpenGVLab | {
"models": [
{
"_id": "674022eb1b99fe8e53fde088",
"id": "OpenGVLab/InternViT-300M-448px-V2_5"
},
{
"_id": "68913539bd3d0a833438591d",
"id": "openai/gpt-oss-20b"
}
],
"relation": "merge"
} | 5,376 | 5,376 | False | 2025-08-25T16:38:37Z | 2025-08-29T17:59:02Z | transformers | 51 | 51 | null | image-text-to-text | {"parameters": {"BF16": 21232768704}, "total": 392282304} | [
".gitattributes",
"README.md",
"chat_template.jinja",
"config.json",
"configuration_intern_vit.py",
"configuration_internvl_chat.py",
"conversation.py",
"examples/image1.jpg",
"examples/image2.jpg",
"examples/red-panda.mp4",
"generation_config.json",
"model-00001-of-00009.safetensors",
"mode... | [
1685,
81438,
15934,
3361,
5546,
4869,
16350,
78073,
125656,
1867237,
68,
4521940256,
4939128416,
4939128416,
4939128448,
4939128464,
4939128464,
4939128464,
4939128464,
3369791600,
69381,
18151,
17106,
666,
72,
440,
27869826,
5762,
1345
] | 42,495,813,958 | aaabe6aa487a7b3db734b104e72b7e85afcd9093 | [
"transformers",
"safetensors",
"internvl_chat",
"feature-extraction",
"internvl",
"custom_code",
"image-text-to-text",
"conversational",
"multilingual",
"dataset:OpenGVLab/MMPR-v1.2",
"dataset:OpenGVLab/MMPR-Tiny",
"arxiv:2312.14238",
"arxiv:2404.16821",
"arxiv:2412.05271",
"arxiv:2411.1... | null | # InternVL3_5-GPT-OSS-20B-A4B-Preview
[\[📂 GitHub\]](https://github.com/OpenGVLab/InternVL) [\[📜 InternVL 1.0\]](https://huggingface.co/papers/2312.14238) [\[📜 InternVL 1.5\]](https://huggingface.co/papers/2404.16821) [\[📜 InternVL 2.5\]](https://huggingface.co/papers/2412.05271) [\[📜 InternVL2.5-MPO\]](https... | null | [
"apache-2.0"
] | [
"OpenGVLab/MMPR-v1.2",
"OpenGVLab/MMPR-Tiny"
] | [
"multilingual"
] | 392,282,304 | null | [
"feature-extraction",
"image-text-to-text"
] | null | [
"modeling_internvl_chat.InternVLChatModel",
"AutoModel",
"InternVLChatModel",
"internvl_chat"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"embeddings",
"text"
] | free | community | [
"China"
] | null | null | null | null | null | null | null | null | null |
6889ec2029577a514ed82d21 | rednote-hilab/dots.ocr | rednote-hilab | null | 173,926 | 174,431 | False | 2025-07-30T09:55:44Z | 2025-08-18T04:26:38Z | dots_ocr | 874 | 45 | null | image-text-to-text | {"parameters": {"BF16": 3039179264}, "total": 3039179264} | [
".gitattributes",
"NOTICE",
"README.md",
"chat_template.json",
"config.json",
"configuration_dots.py",
"dots.ocr LICENSE AGREEMENT",
"generation_config.json",
"merges.txt",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"modeling_dots_... | [
1519,
117509,
31079,
1109,
1474,
2927,
15532,
74,
1671839,
4292758192,
1785673544,
52250,
4981,
17510,
19435,
347,
494,
7036028,
9310,
2776833
] | 6,090,191,986 | 325ed02afb60352c2976f755e595653aae6908f6 | [
"dots_ocr",
"safetensors",
"image-to-text",
"ocr",
"document-parse",
"layout",
"table",
"formula",
"image-text-to-text",
"conversational",
"custom_code",
"en",
"zh",
"multilingual",
"license:mit",
"region:us"
] | null | <div align="center">
<p align="center">
<img src="https://raw.githubusercontent.com/rednote-hilab/dots.ocr/master/assets/logo.png" width="300"/>
<p>
<h1 align="center">
dots.ocr: Multilingual Document Layout Parsing in a Single Vision-Language Model
</h1>
[
Type | Name | Location | Download
| ------------ | -------------------------------------------------- | ------------------------... | null | [
"apache-2.0"
] | null | [
"en",
"zh"
] | null | 20,430,401,088 | [
"image-to-image"
] | null | [
"qwen_image"
] | [
"vision"
] | [
"image"
] | [
"image"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
6892c3259e05f65e4e6f20a7 | NousResearch/Hermes-4-405B | NousResearch | {
"models": [
{
"_id": "6695cd12321386ed51d7bc22",
"id": "meta-llama/Llama-3.1-405B"
}
],
"relation": "finetune"
} | 237 | 237 | False | 2025-08-06T02:51:17Z | 2025-08-26T18:45:42Z | transformers | 44 | 44 | [{"name": "Hermes-4-Llama-3.1-405B", "results": []}] | text-generation | {"parameters": {"BF16": 405853388800}, "total": 405853388800} | [
".gitattributes",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00191.safetensors",
"model-00002-of-00191.safetensors",
"model-00003-of-00191.safetensors",
"model-00004-of-00191.safetensors",
"model-00005-of-00191.safetensors",
"model-00006-of-00191... | null | null | ded87685955a04e0d2f0b4e80aef173b3e2ec24f | [
"transformers",
"safetensors",
"llama",
"text-generation",
"Llama-3.1",
"instruct",
"finetune",
"reasoning",
"hybrid-mode",
"chatml",
"function calling",
"tool use",
"json mode",
"structured outputs",
"atropos",
"dataforge",
"long context",
"roleplaying",
"chat",
"conversationa... | null | # Hermes 4 — Llama-3.1 405B

## Model Description
Hermes 4 405B is a frontier, hybrid-mode **reasoning** model based on Llama-3.1-405B by Nous Research that is aligned to **you**.
Read the Hermes ... | [
"ReallyFloppyPenguin/NousResearch-Hermes-4-405B",
"Humbl3m33/NousResearch-Hermes-4-405B"
] | [
"llama3"
] | null | [
"en"
] | 405,853,388,800 | null | [
"text-generation"
] | null | [
"llama",
"AutoModelForCausalLM",
"LlamaForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | team | company | [
"Online"
] | null | null | null | null | null | null | null | null | null |
68a5958fa06c5a7e6e4e7fc7 | InstantX/Qwen-Image-ControlNet-Union | InstantX | {
"models": [
{
"_id": "688d9adf9f62ee5c9a3804eb",
"id": "Qwen/Qwen-Image"
}
],
"relation": "finetune"
} | 3,540 | 3,540 | False | 2025-08-20T09:29:51Z | 2025-08-26T18:09:04Z | diffusers | 45 | 43 | null | image-to-image | null | [
".gitattributes",
"README.md",
"conds/canny.png",
"conds/canny1.png",
"conds/depth.png",
"conds/pose.png",
"conds/soft_edge.png",
"config.json",
"controlnet_qwenimage.py",
"diffusion_pytorch_model.safetensors",
"infer_qwenimage_cn_union.py",
"outputs/canny.png",
"outputs/canny1.png",
"outp... | [
2057,
5786,
448365,
278157,
105657,
114547,
930559,
491,
15790,
3536027816,
3214,
1482438,
1475396,
1152512,
1471965,
1147515,
40760,
27421
] | 3,544,730,446 | b13036f066d6dee7c20513e263d3d673055e9de8 | [
"diffusers",
"safetensors",
"Image-to-Image",
"ControlNet",
"Diffusers",
"QwenImageControlNetPipeline",
"Qwen-Image",
"image-to-image",
"en",
"base_model:Qwen/Qwen-Image",
"base_model:finetune:Qwen/Qwen-Image",
"license:apache-2.0",
"region:us"
] | null | # Qwen-Image-ControlNet-Union
This repository provides a unified ControlNet that supports 4 common control types (canny, soft edge, depth, pose) for [Qwen-Image](https://github.com/QwenLM/Qwen-Image).
# Model Cards
- This ControlNet consists of 5 double blocks copied from the pretrained transformer layers.
- We train... | [
"InstantX/Qwen-Image-ControlNet"
] | [
"apache-2.0"
] | null | [
"en"
] | null | null | [
"image-to-image"
] | null | null | [
"vision"
] | [
"image"
] | [
"image"
] | free | community | [
"Online",
"China"
] | null | null | null | null | null | null | null | null | null |
6698d8a0653e4babe21e1e7d | meta-llama/Llama-3.1-8B-Instruct | meta-llama | {
"models": [
{
"_id": "66944f1fe0c5c2e493a804f5",
"id": "meta-llama/Llama-3.1-8B"
}
],
"relation": "finetune"
} | 12,432,304 | 79,425,304 | manual | 2024-07-18T08:56:00Z | 2024-09-25T17:00:57Z | transformers | 4,549 | 40 | null | text-generation | {"parameters": {"BF16": 8030261248}, "total": 8030261248} | [
".gitattributes",
"LICENSE",
"README.md",
"USE_POLICY.md",
"config.json",
"generation_config.json",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"original/consolid... | [
1519,
7627,
44044,
4691,
855,
184,
4976698672,
4999802720,
4915916176,
1168138808,
23950,
16060617592,
199,
2183982,
296,
9085657,
55351
] | 32,132,582,323 | 0e9e39f249a16976918f6564b8830bc894c89659 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pytorch",
"llama-3",
"conversational",
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th",
"arxiv:2204.05149",
"base_model:meta-llama/Llama-3.1-8B",
"base_model:finetune:meta-llama/Llama-3.1-8B",
"lic... | null | null | [
"umint/ai",
"eduagarcia/open_pt_llm_leaderboard",
"Nymbo/Serverless-TextGen-Hub",
"cvachet/pdf-chatbot",
"allenai/reward-bench",
"KingNish/OpenGPT-4o",
"flowers-team/StickToYourRoleLeaderboard",
"GIZ/audit_assistant",
"baconnier/prompt-plus-plus",
"data-agents/jupyter-agent",
"nvidia/kvpress",
... | [
"llama3.1"
] | null | [
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th"
] | 8,030,261,248 | null | [
"text-generation"
] | null | [
"llama",
"AutoModelForCausalLM",
"LlamaForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | user | user | [
"user"
] | Accept to share username & email | meta-llama/Llama-3.1-8B | [
"text"
] | [
"Text Generation"
] | [
"Transformer: Text Decoder-only"
] | [
"en",
" fr",
" de",
" hi",
" it",
" pt",
" es",
" th"
] | [
"Pretraining: Causal Language Modeling (CLM)",
" Finetuning: Supervised",
" Reinforcement learning from feedback"
] | Partially disclosed: unavailable | 13 |
689fc0902706443d5b9e1a78 | NexaAI/OmniNeural-4B | NexaAI | null | 332 | 332 | False | 2025-08-15T23:19:44Z | 2025-08-28T20:42:07Z | null | 143 | 40 | null | null | null | [
".gitattributes",
"LICENSE",
"README.md",
"assets/MOBILE_50MB.mp4",
"assets/PC_Demo_Agent.mov",
"assets/PC_Demo_Audio.mov",
"assets/PC_demo_2_image.mov",
"audio/attachements-3-3.nexa",
"config.json",
"files-1-1.nexa",
"llm/attachements-1-3.nexa",
"vit/attachement-2-3.nexa",
"weights-1-8.nexa... | null | null | f8db61dffd2b634a10e630048fd4975daef5e6ab | [
"multimodal",
"NPU",
"On-device",
"Snapdragon PC",
"Android",
"license:other",
"region:us"
] | null | <p align="center">
<img alt="omnineural" src="https://cdn-uploads.huggingface.co/production/uploads/6618e0424dbef6bd3c72f89a/zRUnoWmw43fl9hrXHg0pE.png">
</p>
# **OmniNeural** — World’s First NPU-aware Multimodal Model
## **Overview**
**OmniNeural** is the first fully multimodal model designed specifically for Ne... | null | [
"other",
"nexa-research",
"LICENSE"
] | null | null | null | null | null | null | null | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68af04f9b4b5e5f5380bbe90 | CohereLabs/command-a-translate-08-2025 | CohereLabs | {
"models": [
{
"_id": "67cffded2c8bb280124570e9",
"id": "CohereLabs/c4ai-command-a-03-2025"
}
],
"relation": "finetune"
} | 98 | 98 | auto | 2025-08-27T13:15:37Z | 2025-08-28T14:51:33Z | transformers | 39 | 40 | null | text-generation | {"parameters": {"BF16": 111057580032}, "total": 111057580032} | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"model-00001-of-00049.safetensors",
"model-00002-of-00049.safetensors",
"model-00003-of-00049.safetensors",
"model-00004-of-00049.safetensors",
"model-00005-of-00049.safetensors",
"model-00006-of-00049.safetensors",
"model-0... | null | null | 368e28e1039dce40fff82b9470aa83240b348ee8 | [
"transformers",
"safetensors",
"cohere2",
"text-generation",
"conversational",
"en",
"fr",
"de",
"es",
"it",
"pt",
"ja",
"ko",
"zh",
"ar",
"el",
"fa",
"pl",
"id",
"cs",
"he",
"hi",
"nl",
"ro",
"ru",
"tr",
"uk",
"vi",
"base_model:CohereLabs/c4ai-command-a-03-20... | null | null | [
"CohereLabs/command-a-translate"
] | [
"cc-by-nc-4.0"
] | null | [
"en",
"fr",
"de",
"es",
"it",
"pt",
"ja",
"ko",
"zh",
"ar",
"el",
"fa",
"pl",
"id",
"cs",
"he",
"hi",
"nl",
"ro",
"ru",
"tr",
"uk",
"vi"
] | 111,057,580,032 | null | [
"text-generation"
] | null | [
"Cohere2ForCausalLM",
"AutoModelForCausalLM",
"cohere2"
] | [
"text"
] | [
"text"
] | [
"text"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
6890b6d148203522b7cece41 | nvidia/parakeet-tdt-0.6b-v3 | nvidia | null | 34,932 | 34,932 | False | 2025-08-04T13:34:09Z | 2025-08-20T11:21:08Z | nemo | 220 | 38 | [{"name": "parakeet-tdt-0.6b-v3", "results": [{"task": {"name": "Automatic Speech Recognition", "type": "automatic-speech-recognition"}, "dataset": {"name": "AMI (Meetings test)", "type": "edinburghcstr/ami", "config": "ihm", "split": "test", "args": {"language": "en"}}, "metrics": [{"name": "Test WER", "type": "wer", ... | automatic-speech-recognition | null | [
".gitattributes",
"README.md",
"parakeet-tdt-0.6b-v3.nemo",
"plots/asr.png"
] | [
1679,
37439,
2509332480,
114075
] | 2,509,485,673 | bc3e42c344d9127e85c2d2f92be914f57d741b59 | [
"nemo",
"automatic-speech-recognition",
"speech",
"audio",
"Transducer",
"TDT",
"FastConformer",
"Conformer",
"pytorch",
"NeMo",
"hf-asr-leaderboard",
"en",
"es",
"fr",
"de",
"bg",
"hr",
"cs",
"da",
"nl",
"et",
"fi",
"el",
"hu",
"it",
"lv",
"lt",
"mt",
"pl",
... | null | # **<span style="color:#76b900;">🦜 parakeet-tdt-0.6b-v3: Multilingual Speech-to-Text Model</span>**
<style>
img {
display: inline;
}
</style>
[](#model-architecture)
| [ [\[📜 InternVL 1.0\]](https://huggingface.co/papers/2312.14238) [\[📜 InternVL 1.5\]](https://huggingface.co/papers/2404.16821) [\[📜 InternVL 2.5\]](https://huggingface.co/papers/2412.05271) [\[📜 InternVL2.5-MPO\]](https://huggingface.co/pap... | null | [
"apache-2.0"
] | [
"OpenGVLab/MMPR-v1.2",
"OpenGVLab/MMPR-Tiny"
] | [
"multilingual"
] | 8,528,318,464 | null | [
"feature-extraction",
"image-text-to-text"
] | null | [
"modeling_internvl_chat.InternVLChatModel",
"AutoModel",
"InternVLChatModel",
"internvl_chat"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"embeddings",
"text"
] | free | community | [
"China"
] | null | null | null | null | null | null | null | null | null |
68aeba9d1c034ae61a80d946 | QuantStack/Wan2.2-S2V-14B-GGUF | QuantStack | {
"models": [
{
"_id": "68abccbf1935e46075b39df2",
"id": "Wan-AI/Wan2.2-S2V-14B"
}
],
"relation": "quantized"
} | 17,515 | 17,515 | False | 2025-08-27T07:58:21Z | 2025-08-29T08:07:49Z | gguf | 38 | 38 | null | text-to-video | null | [
".gitattributes",
"README.md",
"Wan2.2-S2V-14B-Q2_K.gguf",
"Wan2.2-S2V-14B-Q3_K_M.gguf",
"Wan2.2-S2V-14B-Q3_K_S.gguf",
"Wan2.2-S2V-14B-Q4_0.gguf",
"Wan2.2-S2V-14B-Q4_1.gguf",
"Wan2.2-S2V-14B-Q4_K_M.gguf",
"Wan2.2-S2V-14B-Q4_K_S.gguf",
"Wan2.2-S2V-14B-Q5_0.gguf",
"Wan2.2-S2V-14B-Q5_1.gguf",
"Wa... | [
2324,
1611,
9510925920,
11386074720,
10724980320,
12768065120,
13469300320,
13861697120,
12958119520,
14524429920,
15225665120,
15002023520,
14347482720,
16213620320,
19616577120,
23037
] | 179,608,988,732 | f3e3bb574e18741431bc6ba84892f06286c29328 | [
"gguf",
"s2v",
"text-to-video",
"base_model:Wan-AI/Wan2.2-S2V-14B",
"base_model:quantized:Wan-AI/Wan2.2-S2V-14B",
"license:apache-2.0",
"region:us"
] | {"total": 16295755609, "architecture": "wan"} | This GGUF file is a direct conversion of [Wan-AI/Wan2.2-S2V-14B](https://huggingface.co/Wan-AI/Wan2.2-S2V-14B)
Type | Name | Location | Download
| ------------- | -------------------------------------------------- | ----------------... | null | [
"apache-2.0"
] | null | null | null | 16,295,755,609 | [
"text-to-video"
] | null | [
"wan"
] | [
"vision"
] | [
"text"
] | [
"video"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
6885cd8c6963bab90aab7f6f | Comfy-Org/Wan_2.2_ComfyUI_Repackaged | Comfy-Org | null | 5,249,135 | 6,095,000 | False | 2025-07-27T06:56:12Z | 2025-08-29T09:09:10Z | diffusion-single-file | 269 | 36 | null | null | null | [
".gitattributes",
"README.md",
"split_files/audio_encoders/wav2vec2_large_english_fp16.safetensors",
"split_files/diffusion_models/wan2.2_fun_camera_high_noise_14B_bf16.safetensors",
"split_files/diffusion_models/wan2.2_fun_camera_high_noise_14B_fp8_scaled.safetensors",
"split_files/diffusion_models/wan2.... | [
1519,
115,
630997322,
29584580312,
15301408048,
29584580312,
15301408048,
10003303280,
28579237064,
14296064656,
28579237064,
14296064656,
10000937656,
28577915912,
14294743520,
28577915912,
14294743520,
28577914792,
14294742832,
28577914792,
14294742832,
32591643778,
163... | 537,050,059,968 | e58ce6c7762f9662699e3479ba30b5004204f93c | [
"diffusion-single-file",
"comfyui",
"region:us"
] | null | Examples: https://comfyanonymous.github.io/ComfyUI_examples/wan22/ | null | null | null | null | null | null | null | null | null | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68ab60cf3ee871412992e835 | Marvis-AI/marvis-tts-250m-v0.1 | Marvis-AI | {
"models": [
{
"_id": "68a73f3ac81d3434c34d6360",
"id": "Marvis-AI/marvis-tts-250m-v0.1-base-pt"
}
],
"relation": "finetune"
} | 1,189 | 1,189 | False | 2025-08-24T18:58:23Z | 2025-08-26T19:10:01Z | transformers | 35 | 35 | null | text-to-audio | null | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model.safetensors",
"prompts/conversational_a.txt",
"prompts/conversational_a.wav",
"prompts/conversational_b.txt",
"prompts/conversational_b.wav",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer_c... | null | null | 0fcb1428603acf0f6f4cdaad9edc6b351f7a4650 | [
"transformers",
"safetensors",
"csm",
"text-to-audio",
"mlx",
"mlx-audio",
"en",
"dataset:amphion/Emilia-Dataset",
"base_model:Marvis-AI/marvis-tts-250m-v0.1-base-pt",
"base_model:finetune:Marvis-AI/marvis-tts-250m-v0.1-base-pt",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | # Introduction
[[code](https://github.com/Marvis-Labs/marvis-tts)]
Marvis is a cutting-edge conversational speech model designed to enable real-time streaming text-to-speech synthesis. Built with efficiency and accessibility in mind, Marvis addresses the growing need for high-quality, real-time voice synthesis that ca... | null | [
"apache-2.0"
] | [
"amphion/Emilia-Dataset"
] | [
"en"
] | null | null | [
"text-to-audio"
] | null | [
"CsmForConditionalGeneration",
"csm",
"AutoModelForTextToWaveform"
] | [
"text"
] | [
"text"
] | [
"audio"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
654a84cadff2f49007ce6c37 | openai/whisper-large-v3 | openai | null | 4,517,007 | 77,013,476 | False | 2023-11-07T18:41:14Z | 2024-08-12T10:20:10Z | transformers | 4,835 | 33 | null | automatic-speech-recognition | {"parameters": {"F16": 1543490560}, "total": 1543490560} | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"flax_model.msgpack",
"generation_config.json",
"merges.txt",
"model.fp32-00001-of-00002.safetensors",
"model.fp32-00002-of-00002.safetensors",
"model.safetensors",
"model.safetensors.index.fp32.json",
"normalizer.json",
"pre... | [
1519,
21829,
34648,
1272,
6174007324,
3903,
493869,
4993448880,
1180663192,
3087130976,
117893,
52666,
340,
3087394553,
117957,
4993677094,
1180725482,
2072,
2480617,
282843,
1036558
] | 24,701,695,487 | 06f233fe06e710322aca913c1bc4249a0d71fce1 | [
"transformers",
"pytorch",
"jax",
"safetensors",
"whisper",
"automatic-speech-recognition",
"audio",
"hf-asr-leaderboard",
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"... | null | # Whisper
Whisper is a state-of-the-art model for automatic speech recognition (ASR) and speech translation, proposed in the paper
[Robust Speech Recognition via Large-Scale Weak Supervision](https://huggingface.co/papers/2212.04356) by Alec Radford
et al. from OpenAI. Trained on >5M hours of labeled data, Whisper d... | [
"openai/whisper",
"hf-audio/whisper-large-v3",
"wcy1122/MGM-Omni",
"gobeldan/insanely-fast-whisper-webui",
"allenai/OLMoASR",
"MGZON/mgzon-app",
"devilent2/whisper-v3-zero",
"reedmayhew/insanely-fast-whisper-webui-zero",
"artificialguybr/Video-Transcription-Smart-Summary",
"MERaLiON/AudioBench-Lea... | [
"apache-2.0"
] | null | [
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"el",
"ms",
"cs",
"ro",
"da",
"hu",
"ta",
"no",
"th",
"ur",
"hr",
"bg",
"lt",
"la",
"mi",
"ml",
"cy",
"sk"... | 1,543,490,560 | null | [
"automatic-speech-recognition"
] | null | [
"WhisperForConditionalGeneration",
"AutoModelForSpeechSeq2Seq",
"whisper"
] | [
"multimodal"
] | [
"audio"
] | [
"text"
] | enterprise | company | [
"United States of America"
] | null | null | [
"Speech"
] | [
"Text Generation"
] | [
"Transformer: Speech Encoder-only",
" Transformer: Text Decoder-only"
] | [
"multilingual"
] | [
"Pretraining: Causal Language Modeling (CLM)",
" Pretraining: Supervised"
] | Not disclosed | 5 |
687f6a6e50a26ff8c82d9a00 | Kijai/WanVideo_comfy_fp8_scaled | Kijai | {
"models": [
{
"_id": "6822f6e25843aa0767e04828",
"id": "Wan-AI/Wan2.1-VACE-1.3B"
}
],
"relation": "finetune"
} | 274,597 | 348,580 | False | 2025-07-22T10:39:42Z | 2025-08-29T12:12:17Z | diffusion-single-file | 186 | 33 | null | null | null | [
".gitattributes",
"Fun/Wan2_2-Fun-Control-A14B-HIGH_fp8_e4m3fn_scaled_KJ_fixed.safetensors",
"Fun/Wan2_2-Fun-Control-A14B-HIGH_fp8_e5m2_scaled_KJ_fixed.safetensors",
"Fun/Wan2_2-Fun-Control-A14B-LOW_fp8_e4m3fn_scaled_KJ_fixed.safetensors",
"Fun/Wan2_2-Fun-Control-A14B-LOW_fp8_e5m2_scaled_KJ_fixed.safetensor... | [
1519,
14530495778,
14530492906,
14530495778,
14530492906,
15535183490,
15535183490,
14528518994,
14528518986,
17135485186,
17135485186,
66,
16643349018,
16643349018,
16643349018,
16643349018,
17135485186,
17135485186,
15002999858,
15002999858,
15002999858,
15002999858,
27... | 516,996,521,648 | 5765da72e9fee3c08a1fe79d5c44e98eab2d5339 | [
"diffusion-single-file",
"comfyui",
"base_model:Wan-AI/Wan2.1-VACE-1.3B",
"base_model:finetune:Wan-AI/Wan2.1-VACE-1.3B",
"license:apache-2.0",
"region:us"
] | null | Better fp8 scaled models (when measured against fp16) based on quantization code from https://github.com/Tencent-Hunyuan/HunyuanVideo/blob/main/hyvideo/modules/fp8_optimization.py
Can be used with: https://github.com/kijai/ComfyUI-WanVideoWrapper (latest version) and ComfyUI native WanVideo nodes.
14B-T2V comparison ... | null | [
"apache-2.0"
] | null | null | null | null | null | null | null | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
6890b6f154a8b9ff771a29c9 | nvidia/canary-1b-v2 | nvidia | null | 8,435 | 8,435 | False | 2025-08-04T13:34:41Z | 2025-08-20T11:20:40Z | nemo | 220 | 32 | [{"name": "canary-1b-v2", "results": [{"task": {"type": "Automatic Speech Recognition", "name": "automatic-speech-recognition"}, "dataset": {"name": "FLEURS", "type": "google/fleurs", "config": "bg_bg", "split": "test", "args": {"language": "bg"}}, "metrics": [{"name": "Test WER (Bg)", "type": "wer", "value": 9.25, "ve... | automatic-speech-recognition | null | [
".gitattributes",
"README.md",
"canary-1b-v2.nemo",
"plots/asr.png",
"plots/en_x.png",
"plots/x_en.png"
] | [
1615,
67988,
6358958080,
146866,
134001,
125547
] | 6,359,434,097 | 21c939b94d98894647bdc73fa86e19c458f7bfc9 | [
"nemo",
"automatic-speech-recognition",
"automatic-speech-translation",
"speech",
"audio",
"Transformer",
"FastConformer",
"Conformer",
"pytorch",
"NeMo",
"hf-asr-leaderboard",
"bg",
"hr",
"cs",
"da",
"nl",
"en",
"et",
"fi",
"fr",
"de",
"el",
"hu",
"it",
"lv",
"lt",... | null | ## <span style="color:#ffb300;">🐤 Canary 1B v2: Multitask Speech Transcription and Translation Model </span>
**``Canary-1b-v2``** is a powerful 1-billion parameter model built for high-quality speech transcription and translation across 25 European languages.
It excels at both automatic speech recognition (ASR) and... | [
"nvidia/canary-1b-v2",
"nvidia/canary-1b-flash"
] | [
"cc-by-4.0"
] | [
"nvidia/Granary",
"nvidia/nemo-asr-set-3.0"
] | [
"bg",
"hr",
"cs",
"da",
"nl",
"en",
"et",
"fi",
"fr",
"de",
"el",
"hu",
"it",
"lv",
"lt",
"mt",
"pl",
"pt",
"ro",
"sk",
"sl",
"es",
"sv",
"ru",
"uk"
] | null | null | [
"automatic-speech-recognition"
] | [
"bleu",
"wer",
"comet"
] | null | [
"multimodal"
] | [
"audio"
] | [
"text"
] | enterprise_plus | company | [
"United States of America"
] | null | null | null | null | null | null | null | null | null |
687c61b066d1fc7f73788fc2 | zai-org/GLM-4.5 | zai-org | null | 86,977 | 95,519 | False | 2025-07-20T03:25:36Z | 2025-08-11T13:27:03Z | transformers | 1,277 | 30 | null | text-generation | {"parameters": {"BF16": 358337776896, "F32": 14400}, "total": 358337791296} | [
".gitattributes",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00093.safetensors",
"model-00002-of-00093.safetensors",
"model-00003-of-00093.safetensors",
"model-00004-of-00093.safetensors",
"model-00005-of-00093.safetensors",
"model-00006-of-00093... | [
1570,
9670,
3242,
1007,
155,
3753953568,
650168352,
650168352,
7871313120,
7871313120,
7871313120,
7871313120,
7871313120,
7871313120,
7871313120,
7871313616,
7871313616,
7871313616,
7871313616,
7871313616,
7871313616,
7871313616,
7871313616,
7871313616,
7871313616,
... | 716,705,229,651 | cbb2c7cfb52fa128a9660cb1a7a78e017899e115 | [
"transformers",
"safetensors",
"glm4_moe",
"text-generation",
"conversational",
"en",
"zh",
"arxiv:2508.06471",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | # GLM-4.5
<div align="center">
<img src=https://raw.githubusercontent.com/zai-org/GLM-4.5/refs/heads/main/resources/logo.svg width="15%"/>
</div>
<p align="center">
👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.
<br>
📖 Check out the GLM-4.5 <a href="https://z.ai... | [
"umint/ai",
"zai-org/GLM-4.5-Space",
"Arphd4/ARK.AI",
"nazdridoy/inferoxy-hub",
"umint/o4-mini",
"Tigasturned/GLM-4.5-WebDev",
"taha092/zai-org-GLM-4.5",
"rajibsalui/zai-org-GLM-4.5",
"ReallyFloppyPenguin/zai-org-GLM-4.5",
"Sam3838/zai-org-GLM-4.5",
"llamameta/glm4.5-free-unlimited-chatbot",
"... | [
"mit"
] | null | [
"en",
"zh"
] | 358,337,791,296 | null | [
"text-generation"
] | null | [
"Glm4MoeForCausalLM",
"glm4_moe",
"AutoModelForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
688a5f23e7cb4413ca5b8678 | google/gemma-3-270m-it | google | {
"models": [
{
"_id": "689252773b8900ddb9116aed",
"id": "google/gemma-3-270m"
}
],
"relation": "finetune"
} | 149,636 | 149,654 | manual | 2025-07-30T18:06:27Z | 2025-08-14T07:35:07Z | transformers | 367 | 30 | null | text-generation | {"parameters": {"BF16": 268098176}, "total": 268098176} | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model.safetensors",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer.model",
"tokenizer_config.json"
] | [
1570,
28308,
35,
1532,
1352,
173,
536223056,
662,
33384570,
4689074,
1155375
] | 575,485,707 | ac82b4e820549b854eebf28ce6dedaf9fdfa17b3 | [
"transformers",
"safetensors",
"gemma3_text",
"text-generation",
"gemma3",
"gemma",
"google",
"conversational",
"arxiv:2503.19786",
"arxiv:1905.07830",
"arxiv:1905.10044",
"arxiv:1911.11641",
"arxiv:1705.03551",
"arxiv:1911.01547",
"arxiv:1907.10641",
"arxiv:2311.07911",
"arxiv:2311.... | null | null | [
"anakin87/gemma-3-270m-it",
"Monster/gemma-3-270m-it",
"broadfield-dev/gemma-3-270m-it-demo",
"akhaliq/gemma-3-270m-gradio-coder",
"umint/o4-mini",
"Luigi/ZeroGPU-LLM-Inference",
"salmankhanpm/Telugu_Vocab_Evaluation",
"tigeryfan/Anything2Cal",
"daniel-dona/gemma-3-270m",
"TakiTakiTa/gemma-3-270m"... | [
"gemma"
] | null | null | 268,098,176 | null | [
"text-generation"
] | null | [
"gemma3_text",
"Gemma3ForCausalLM",
"AutoModelForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | enterprise | company | [
"United States of America"
] | null | null | null | null | null | null | null | null | null |
6895269304b1e0b7de0d0a47 | DavidAU/OpenAi-GPT-oss-20b-abliterated-uncensored-NEO-Imatrix-gguf | DavidAU | {
"models": [
{
"_id": "689374f41dc3bcaaf05e4963",
"id": "huihui-ai/Huihui-gpt-oss-20b-BF16-abliterated"
}
],
"relation": "quantized"
} | 226,350 | 226,350 | False | 2025-08-07T22:20:03Z | 2025-08-30T05:25:26Z | null | 127 | 30 | null | text-generation | null | [
".gitattributes",
"OpenAI-20B-NEO-CODE-DI-Uncensored-Q5_1.gguf",
"OpenAI-20B-NEO-CODE-DI-Uncensored-Q8_0.gguf",
"OpenAI-20B-NEO-CODE2-Plus-Uncensored-IQ4_NL.gguf",
"OpenAI-20B-NEO-CODEPlus-Uncensored-IQ4_NL.gguf",
"OpenAI-20B-NEO-CODEPlus-Uncensored-Q5_1.gguf",
"OpenAI-20B-NEO-CODEPlus16-Uncensored-IQ4_... | [
3032,
15728919168,
22080931968,
11815759488,
11815759456,
15728919136,
12648263776,
22080931968,
11815759488,
15728919168,
21972344448,
15728919168,
22080931968,
11815759488,
12648263776,
16452835936,
11815759456,
15728919136,
22080931936,
39635,
147332
] | 289,769,018,927 | 6e9bdcc3a8f9da44f0cdcbf4ec822b4d08decf9b | [
"gguf",
"gpt_oss",
"gpt-oss",
"openai",
"mxfp4",
"programming",
"code generation",
"code",
"coding",
"coder",
"chat",
"reasoning",
"thinking",
"r1",
"cot",
"deepseek",
"128k context",
"general usage",
"problem solving",
"brainstorming",
"solve riddles",
"uncensored",
"abl... | {"total": 20914757184, "architecture": "gpt-oss", "context_length": 131072, "quantize_imatrix_file": "E:/_imx/OpenAi-GPT-oss-20b-abliterated-uncensored-NEO-CODE-OT.gguf", "chat_template": "{# Copyright 2025-present Unsloth. Apache 2.0 License. Unsloth chat template fixes. Edited from ggml-org & OpenAI #}\n{#-\n In add... | <small><font color="red">Specialized uncensored/abliterated quants for new OpenAI 20B MOE - Mixture of Experts Model at 80+ T/S. See settings and special instructions for using abliterated models below.</font></small>
<h2>OpenAi-GPT-oss-20b-abliterated-uncensored-NEO-Imatrix-gguf</h2>
<img src="power-the-matrix.gif" ... | [
"Exmanq/qqqq"
] | [
"apache-2.0"
] | null | [
"en"
] | null | 20,914,757,184 | [
"text-generation"
] | null | [
"gpt-oss"
] | [
"text"
] | [
"text"
] | [
"text"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
664dc170474f2283fa5c8659 | mistralai/Mistral-7B-Instruct-v0.3 | mistralai | {
"models": [
{
"_id": "664dc156dba1a2aeb958dc90",
"id": "mistralai/Mistral-7B-v0.3"
}
],
"relation": "finetune"
} | 761,130 | 15,719,806 | auto | 2024-05-22T09:57:04Z | 2025-07-24T16:43:00Z | vllm | 2,054 | 29 | null | null | {"parameters": {"BF16": 7248023552}, "total": 7248023552} | [
".gitattributes",
"README.md",
"config.json",
"consolidated.safetensors",
"generation_config.json",
"model-00001-of-00003.safetensors",
"model-00002-of-00003.safetensors",
"model-00003-of-00003.safetensors",
"model.safetensors.index.json",
"params.json",
"special_tokens_map.json",
"tokenizer.j... | [
1519,
7876,
601,
14496078512,
116,
4949453792,
4999819336,
4546807800,
23950,
202,
414,
1961548,
587404,
587404,
140874
] | 28,995,471,348 | 0d4b76e1efeb5eb6f6b5e757c79870472e04bd3a | [
"vllm",
"safetensors",
"mistral",
"mistral-common",
"base_model:mistralai/Mistral-7B-v0.3",
"base_model:finetune:mistralai/Mistral-7B-v0.3",
"license:apache-2.0",
"region:us"
] | null | null | [
"gokaygokay/FLUX-Prompt-Generator",
"eduagarcia/open_pt_llm_leaderboard",
"Nymbo/Serverless-TextGen-Hub",
"cvachet/pdf-chatbot",
"flowers-team/StickToYourRoleLeaderboard",
"bhaskartripathi/LLM_Quantization",
"niltoncota/equiframe_llm",
"Shankarm08/LLMsintro",
"Skier8402/mistral-super-fast",
"hysts... | [
"apache-2.0"
] | null | null | 7,248,023,552 | null | null | null | [
"MistralForCausalLM",
"mistral"
] | null | null | null | team | company | [
"France"
] | Accept to share username & email | mistralai/Mistral-7B-v0.3 | [
"Text"
] | [
"Text Generation"
] | [
"Transformer: Text Encoder-Decoder"
] | [
"EN"
] | [
"Instruction finetuning"
] | Disclosed: unavailable | 7 |
683f05dc98de733bf4c59f3d | Qwen/Qwen3-Embedding-0.6B | Qwen | {
"models": [
{
"_id": "680f0da2f99a20ddfc9a9f26",
"id": "Qwen/Qwen3-0.6B-Base"
}
],
"relation": "finetune"
} | 3,827,194 | 8,792,135 | False | 2025-06-03T14:25:32Z | 2025-06-20T09:31:05Z | sentence-transformers | 537 | 29 | null | feature-extraction | {"parameters": {"BF16": 595776512}, "total": 595776512} | [
".gitattributes",
"1_Pooling/config.json",
"README.md",
"config.json",
"config_sentence_transformers.json",
"generation_config.json",
"merges.txt",
"model.safetensors",
"modules.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"
] | [
1570,
313,
17237,
727,
215,
117,
1671853,
1191586416,
349,
11423705,
9706,
2776833
] | 1,207,489,041 | c54f2e6e80b2d7b7de06f51cec4959f6b3e03418 | [
"sentence-transformers",
"safetensors",
"qwen3",
"text-generation",
"transformers",
"sentence-similarity",
"feature-extraction",
"text-embeddings-inference",
"arxiv:2506.05176",
"base_model:Qwen/Qwen3-0.6B-Base",
"base_model:finetune:Qwen/Qwen3-0.6B-Base",
"license:apache-2.0",
"autotrain_co... | null | # Qwen3-Embedding-0.6B
<p align="center">
<img src="https://qianwen-res.oss-accelerate-overseas.aliyuncs.com/logo_qwen3.png" width="400"/>
<p>
## Highlights
The Qwen3 Embedding model series is the latest proprietary model of the Qwen family, specifically designed for text embedding and ranking tasks. Building up... | [
"davanstrien/huggingface-datasets-search-v2",
"dzenzzz/sq",
"wublewobble/genre-classifier",
"Omartificial-Intelligence-Space/qwen-arabic-semantic-suite",
"rwitz/Qwen3-Embedding-0.6B-Demo",
"Shamik/ml_research_assistant_and_tutor",
"Agents-MCP-Hackathon/ml_research_agent",
"mouryachinta/gwen-cohere",
... | [
"apache-2.0"
] | null | null | 595,776,512 | null | [
"sentence-similarity",
"text-generation",
"feature-extraction"
] | null | [
"AutoModelForCausalLM",
"Qwen3ForCausalLM",
"qwen3"
] | [
"multimodal",
"text"
] | [
"text"
] | [
"embeddings",
"logits",
"text"
] | team | company | [
"China"
] | null | Qwen/Qwen3-0.6B-Base | [
"Text"
] | [
"Text Embedding"
] | [
"Transformer: Text Decoder-only"
] | [
"multilingual"
] | [
"Finetuning: Supervised",
" Multi-task finetuning",
" Model Merging"
] | Partially disclosed: unavailable | 3 |
6890b5cd9db9549fd73fca6f | lightx2v/Wan2.2-Lightning | lightx2v | {
"models": [
{
"_id": "6881e60ffcffaee6d84fe9e4",
"id": "Wan-AI/Wan2.2-I2V-A14B"
}
],
"relation": "finetune"
} | 0 | 0 | False | 2025-08-04T13:29:49Z | 2025-08-13T10:12:42Z | null | 323 | 29 | null | text-to-video | null | [
".gitattributes",
"README.md",
"Wan2.2-I2V-A14B-4steps-lora-rank64-Seko-V1/Wan2.2-I2V-A14B-4steps-lora-rank64-Seko-V1-NativeComfy.json",
"Wan2.2-I2V-A14B-4steps-lora-rank64-Seko-V1/Wan2.2-I2V-A14B-4steps-lora-rank64-Seko-V1-forKJ.json",
"Wan2.2-I2V-A14B-4steps-lora-rank64-Seko-V1/Wan2.2-I2V-A14B-4steps-lora... | null | null | 5a786cebbd8af0e5b0cc8162312e054ce6b97d8a | [
"safetensors",
"text-to-video;",
"image-to-video;",
"comfyUI;",
"video-generation;",
"text-to-video",
"en",
"arxiv:2309.14509",
"base_model:Wan-AI/Wan2.2-I2V-A14B",
"base_model:finetune:Wan-AI/Wan2.2-I2V-A14B",
"license:apache-2.0",
"region:us"
] | null | You're welcome to visit our [GitHub repository](https://github.com/ModelTC/Wan2.2-Lightning) for the latest model releases or to reproduce our results.
# Wan2.2-Lightning
<!-- [**Wan2.2-Lightning: Distill Wan2.2 Family into 4 Steps**] <be> -->
We are excited to release the distilled version of <a href="https://wan.... | [
"rahul7star/wan2-2-T2V-EXP"
] | [
"apache-2.0"
] | null | [
"en"
] | null | null | [
"text-to-video"
] | null | null | [
"vision"
] | [
"text"
] | [
"video"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
621ffdc136468d709f180294 | sentence-transformers/all-MiniLM-L6-v2 | sentence-transformers | null | 91,725,092 | 1,318,484,604 | False | 2022-03-02T23:29:05Z | 2025-03-06T13:37:44Z | sentence-transformers | 3,821 | 28 | null | sentence-similarity | {"parameters": {"I64": 512, "F32": 22713216}, "total": 22713728} | [
".gitattributes",
"1_Pooling/config.json",
"README.md",
"config.json",
"config_sentence_transformers.json",
"data_config.json",
"model.safetensors",
"modules.json",
"onnx/model.onnx",
"onnx/model_O1.onnx",
"onnx/model_O2.onnx",
"onnx/model_O3.onnx",
"onnx/model_O4.onnx",
"onnx/model_qint8_... | [
1229,
190,
10454,
612,
116,
39265,
90868376,
349,
90405214,
90360328,
90326566,
90326497,
45212349,
23026053,
23026053,
23026053,
23046789,
90265744,
211315,
22933664,
368006,
90888945,
90887379,
53,
112,
91005696,
466247,
350,
13156,
231508
] | 976,948,668 | c9745ed1d9f207416be6d2e6f8de32d1f16199bf | [
"sentence-transformers",
"pytorch",
"tf",
"rust",
"onnx",
"safetensors",
"openvino",
"bert",
"feature-extraction",
"sentence-similarity",
"transformers",
"en",
"dataset:s2orc",
"dataset:flax-sentence-embeddings/stackexchange_xml",
"dataset:ms_marco",
"dataset:gooaq",
"dataset:yahoo_a... | null | # all-MiniLM-L6-v2
This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and can be used for tasks like clustering or semantic search.
## Usage (Sentence-Transformers)
Using this model becomes easy when you have [sentence-transformers](ht... | [
"mteb/leaderboard",
"Jawahars/flan-t5-base-kyndryl2023-24",
"seanpedrickcase/Light-PDF-Web-QA-Chatbot",
"cvachet/pdf-chatbot",
"JournalistsonHF/ai-scraper",
"HusseinEid/RAG-enabled-Chatbot",
"katanaml/sparrow-ml",
"opensearch-project/OpenSearch-AI",
"K00B404/InstructIR",
"Jeice/AgenteHelpN8n",
"... | [
"apache-2.0"
] | [
"s2orc",
"flax-sentence-embeddings/stackexchange_xml",
"ms_marco",
"gooaq",
"yahoo_answers_topics",
"code_search_net",
"search_qa",
"eli5",
"snli",
"multi_nli",
"wikihow",
"natural_questions",
"trivia_qa",
"embedding-data/sentence-compression",
"embedding-data/flickr30k-captions",
"emb... | [
"en"
] | 22,713,728 | null | [
"sentence-similarity",
"feature-extraction"
] | null | [
"AutoModel",
"BertModel",
"bert"
] | [
"multimodal",
"text"
] | [
"text"
] | [
"embeddings",
"logits"
] | free | university | [
"Germany"
] | null | nreimers/MiniLM-L6-H384-uncased | [
"Text"
] | [
"Text Embedding"
] | [
"Transformer: Text Encoder-only"
] | [
"en"
] | [
"Finetuning: Supervised"
] | Disclosed: available | 0 |
6540d2d50cb8e9d8e63a1e1f | coqui/XTTS-v2 | coqui | null | 3,131,770 | 29,008,587 | False | 2023-10-31T10:11:33Z | 2023-12-11T17:50:00Z | coqui | 3,010 | 28 | null | text-to-speech | null | [
".gitattributes",
"LICENSE.txt",
"README.md",
"config.json",
"dvae.pth",
"hash.md5",
"mel_stats.pth",
"model.pth",
"samples/de_sample.wav",
"samples/en_sample.wav",
"samples/es_sample.wav",
"samples/fr_sample.wav",
"samples/ja-sample.wav",
"samples/pt_sample.wav",
"samples/tr_sample.wav"... | [
1519,
4014,
4263,
4368,
210514388,
32,
1067,
1867929118,
299066,
299066,
329786,
432186,
284730,
445498,
299066,
383034,
7754818,
361219
] | 2,089,347,238 | 6c2b0d75eae4b7047358e3b6bd9325f857d43f77 | [
"coqui",
"text-to-speech",
"license:other",
"region:us"
] | null | # ⓍTTS
ⓍTTS is a Voice generation model that lets you clone voices into different languages by using just a quick 6-second audio clip. There is no need for an excessive amount of training data that spans countless hours.
This is the same or similar model to what powers [Coqui Studio](https://coqui.ai/) and [Coqui API]... | [
"tonyassi/voice-clone",
"Inferless/Open-Source-TTS-Gallary",
"Pendrokar/TTS-Spaces-Arena",
"coqui/xtts",
"TTS-AGI/TTS-Arena",
"yuAIDaren/BG-XTTS-v2",
"aiqtech/Open-Source-TTS-Gallary",
"coqui/voice-chat-with-mistral",
"kerncraze/XTTS_V1_CPU",
"awacke1/voice-chat-with-mistral",
"MXNXVMadman/space... | [
"other",
"coqui-public-model-license",
"https://coqui.ai/cpml"
] | null | null | null | null | [
"text-to-speech"
] | null | null | [
"audio"
] | [
"text"
] | [
"audio"
] | free | company | [
"Germany"
] | null | null | [
"Text"
] | [
"Speech Generation"
] | [
"Transformer: Text Encoder-only",
" Transformer: Speech Encoder-only",
" Transformer: Speech Decoder-only",
" Variational Autoencoder",
" Diffusion-based Network"
] | [
"EN",
" ES",
"FR",
" DE",
" IT",
" PT",
"PL",
" TR",
" RU",
" NL",
" CS",
" AR",
" ZH",
" JA",
" HU",
" KO",
" HI"
] | [
"Pretraining: Multimodal joint-embeddings",
" Finetuning: Supervised"
] | Partially disclosed: unavailable | 7 |
684375c01253c8d254bb3c9e | Motif-Technologies/Motif-2.6B | Motif-Technologies | null | 817 | 6,001 | False | 2025-06-06T23:12:00Z | 2025-08-28T08:48:38Z | null | 74 | 28 | null | text-generation | {"parameters": {"F32": 2597218432}, "total": 2597218432} | [
".gitattributes",
"LICENSE",
"README.md",
"added_tokens.json",
"config.json",
"configuration_motif.py",
"generation_config.json",
"merges.txt",
"model-00001-of-00003.safetensors",
"model-00002-of-00003.safetensors",
"model-00003-of-00003.safetensors",
"model.safetensors.index.json",
"modelin... | [
1570,
7580,
12083,
3600,
944,
9805,
164,
2741072,
4952662512,
4966459400,
469808712,
41422,
66965,
589,
17264873,
23699,
4377405
] | 10,413,482,395 | 5b901f07cdecaa3a3f55ad1a1a89ef78bf2beb4d | [
"safetensors",
"Motif",
"text-generation-inference",
"conversational",
"motif",
"text-generation",
"custom_code",
"en",
"ko",
"arxiv:2508.09148",
"arxiv:2310.06825",
"arxiv:2408.00118",
"arxiv:2503.19786",
"arxiv:2407.21783",
"arxiv:2404.14219",
"arxiv:2412.15115",
"license:apache-2.... | null | *Last update: 14th august 2025*
# New
Now you can try out Motif 2.6B on Model Hub: https://model-hub.motiftech.io/
**Select 'Motif 2.6B' from the dropdown next to the Send button.**
# Introduction
We announce **Motif 2.6B**, a 2.6 billion parameter language model trained from scratch on AMD Instinct™ MI250 GPUs. Mo... | null | [
"apache-2.0"
] | null | [
"en",
"ko"
] | 2,597,218,432 | null | [
"text-generation"
] | null | [
"Motif",
"MotifForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
689d9bf926fe49e8ad685a63 | nunchaku-tech/nunchaku-qwen-image | nunchaku-tech | {
"models": [
{
"_id": "688d9adf9f62ee5c9a3804eb",
"id": "Qwen/Qwen-Image"
}
],
"relation": "quantized"
} | 48,386 | 48,386 | False | 2025-08-14T08:19:05Z | 2025-08-27T14:26:33Z | diffusers | 144 | 28 | null | text-to-image | null | [
".gitattributes",
"README.md",
"svdq-fp4_r128-qwen-image-lightningv1.0-4steps.safetensors",
"svdq-fp4_r128-qwen-image-lightningv1.1-8steps.safetensors",
"svdq-fp4_r128-qwen-image.safetensors",
"svdq-fp4_r32-qwen-image-lightningv1.0-4steps.safetensors",
"svdq-fp4_r32-qwen-image-lightningv1.1-8steps.safet... | [
1519,
6634,
13081386856,
13081386856,
13081386856,
11948923656,
11948923656,
11948923656,
12654443144,
12654443144,
12654443144,
11521979944,
11521979944,
11521979944
] | 147,620,208,953 | fd7d045cae217cb9366fa28c35598457e2a6c93f | [
"diffusers",
"text-to-image",
"SVDQuant",
"Qwen-Image",
"Diffusion",
"Quantization",
"ICLR2025",
"en",
"dataset:mit-han-lab/svdquant-datasets",
"arxiv:2411.05007",
"base_model:Qwen/Qwen-Image",
"base_model:quantized:Qwen/Qwen-Image",
"license:apache-2.0",
"region:us"
] | null | <p align="center" style="border-radius: 10px">
<img src="https://huggingface.co/datasets/nunchaku-tech/cdn/resolve/main/nunchaku/assets/nunchaku.svg" width="30%" alt="Nunchaku Logo"/>
</p>
# Model Card for nunchaku-qwen-image
 models proposed in the [OLMoASR: Open Models and Data for Training Robust Speech Recognition Models](https://github.com/allenai/OLMoASR.git)
paper by Huong Ngo et al. from Ai2. Trained on 440K hours of weakly-supervised audio-text pairs collec... | [
"allenai/OLMoASR"
] | [
"apache-2.0"
] | null | null | null | null | [
"audio-text-to-text"
] | null | null | null | null | null | enterprise | non-profit | [
"United Kingdom"
] | null | null | null | null | null | null | null | null | null |
68aac0cdbf149075fd46a89f | thedeoxen/refcontrol-flux-kontext-reference-pose-lora | thedeoxen | {
"models": [
{
"_id": "68378cef5cbef05290b4d045",
"id": "black-forest-labs/FLUX.1-Kontext-dev"
}
],
"relation": "adapter"
} | 0 | 0 | False | 2025-08-24T07:35:41Z | 2025-08-27T05:42:30Z | null | 26 | 26 | null | image-to-image | null | [
".gitattributes",
"README.md",
"examples/10a.png",
"examples/10b.png",
"examples/11a.png",
"examples/11b.png",
"examples/12a.png",
"examples/12b.png",
"examples/13a.png",
"examples/13b.png",
"examples/14a.png",
"examples/14b.png",
"examples/1a.png",
"examples/1b.png",
"examples/2a.png",
... | null | null | 34dbf37679158c536eb89ad822b2780bb418c1bb | [
"flux",
"kontext",
"img2img",
"controlnet",
"flux-kontext",
"image",
"editing",
"lora",
"image-to-image",
"base_model:black-forest-labs/FLUX.1-Kontext-dev",
"base_model:adapter:black-forest-labs/FLUX.1-Kontext-dev",
"license:apache-2.0",
"region:us"
] | null | license: apache-2.0
base_model:
- black-forest-labs/FLUX.1-Kontext-dev
pipeline_tag: image-to-image
---
# RefControl Flux Kontext – Reference Pose LoRA
## 📝 Short description
A LoRA for **Flux Kontext Dev** that fuses a **reference image** (left) with a **pose control map** (right).
It preserves **identity and sty... | null | [
"apache-2.0"
] | null | null | null | null | [
"image-to-image"
] | null | null | [
"vision"
] | [
"image"
] | [
"image"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68ac918d66dab09772f3021c | OpenGVLab/InternVL3_5-30B-A3B | OpenGVLab | {
"models": [
{
"_id": "68ac918d65fc0297eeace1e9",
"id": "OpenGVLab/InternVL3_5-30B-A3B-MPO"
}
],
"relation": "finetune"
} | 10,604 | 10,604 | False | 2025-08-25T16:38:37Z | 2025-08-29T17:57:02Z | transformers | 26 | 26 | null | image-text-to-text | {"parameters": {"BF16": 30848730112}, "total": 30848730112} | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.jinja",
"config.json",
"configuration_intern_vit.py",
"configuration_internvl_chat.py",
"conversation.py",
"generation_config.json",
"merges.txt",
"model-00001-of-00013.safetensors",
"model-00002-of-00013.safetensors",
"model... | [
1570,
53713,
892,
475,
2594,
5546,
4700,
15309,
69,
1671853,
4997840368,
4998902680,
4997987088,
4997833552,
4999945200,
4999878000,
4998894032,
4999884976,
4999410600,
4999952016,
4997258696,
4998338400,
1713990752,
2012346,
18151,
16521,
666,
72,
744,
1142... | 61,718,131,223 | ba1762d830e2aeddb3272b66ba9e0abf7e69f626 | [
"transformers",
"safetensors",
"internvl_chat",
"feature-extraction",
"internvl",
"custom_code",
"image-text-to-text",
"conversational",
"multilingual",
"dataset:OpenGVLab/MMPR-v1.2",
"dataset:OpenGVLab/MMPR-Tiny",
"arxiv:2312.14238",
"arxiv:2404.16821",
"arxiv:2412.05271",
"arxiv:2411.1... | null | # InternVL3_5-30B-A3B
[\[📂 GitHub\]](https://github.com/OpenGVLab/InternVL) [\[📜 InternVL 1.0\]](https://huggingface.co/papers/2312.14238) [\[📜 InternVL 1.5\]](https://huggingface.co/papers/2404.16821) [\[📜 InternVL 2.5\]](https://huggingface.co/papers/2412.05271) [\[📜 InternVL2.5-MPO\]](https://huggingface.c... | null | [
"apache-2.0"
] | [
"OpenGVLab/MMPR-v1.2",
"OpenGVLab/MMPR-Tiny"
] | [
"multilingual"
] | 30,848,730,112 | null | [
"feature-extraction",
"image-text-to-text"
] | null | [
"modeling_internvl_chat.InternVLChatModel",
"AutoModel",
"InternVLChatModel",
"internvl_chat"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"embeddings",
"text"
] | free | community | [
"China"
] | null | null | null | null | null | null | null | null | null |
66aa974d1f83b210ae7f74ae | black-forest-labs/FLUX.1-schnell | black-forest-labs | null | 620,901 | 14,330,004 | auto | 2024-07-31T19:58:05Z | 2024-08-16T14:37:56Z | diffusers | 4,220 | 24 | null | text-to-image | null | [
".gitattributes",
"README.md",
"ae.safetensors",
"flux1-schnell.safetensors",
"model_index.json",
"scheduler/scheduler_config.json",
"schnell_grid.jpeg",
"text_encoder/config.json",
"text_encoder/model.safetensors",
"text_encoder_2/config.json",
"text_encoder_2/model-00001-of-00002.safetensors",... | [
1668,
3681,
335304388,
23782506688,
536,
274,
1671111,
613,
246144352,
782,
4994582224,
4530066360,
19885,
524619,
588,
705,
1059962,
2543,
791656,
2424235,
20817,
321,
9962580296,
9949328904,
3870584832,
120822,
774,
167666902
] | 57,845,410,538 | 741f7c3ce8b383c54771c7003378a50191e9efe9 | [
"diffusers",
"safetensors",
"text-to-image",
"image-generation",
"flux",
"en",
"license:apache-2.0",
"endpoints_compatible",
"diffusers:FluxPipeline",
"region:us"
] | null | ![FLUX.1 [schnell] Grid](./schnell_grid.jpeg)
`FLUX.1 [schnell]` is a 12 billion parameter rectified flow transformer capable of generating images from text descriptions.
For more information, please read our [blog post](https://blackforestlabs.ai/announcing-black-forest-labs/).
# Key Features
1. Cutting-edge output ... | [
"bytedance-research/USO",
"black-forest-labs/FLUX.1-schnell",
"Kunbyte/OmniTry",
"gparmar/Group-Inference-FLUX.1-Schnell",
"ginigen/Fashion-Fit360",
"Nymbo/Serverless-ImgGen-Hub",
"Yuanshi/OminiControl_Art",
"Yntec/ToyWorld",
"prithivMLmods/FLUX-LoRA-DLC",
"yanze/PuLID-FLUX",
"bobber/DiffuseCraf... | [
"apache-2.0"
] | null | [
"en"
] | null | null | [
"text-to-image"
] | null | null | [
"vision"
] | [
"text"
] | [
"image"
] | team | company | [
"Germany"
] | Other agreement/info requirements | null | [
"Text"
] | [
"Image Generation"
] | [
"Diffusion-based Network"
] | [
"en"
] | [
"Knowledge distillation"
] | Not disclosed | 6 |
672e53b40b534c9deab3be09 | polyglots/SinLlama_v01 | polyglots | {
"models": [
{
"_id": "661f97d48e7f3438386f755d",
"id": "meta-llama/Meta-Llama-3-8B"
}
],
"relation": "adapter"
} | 630 | 1,111 | False | 2024-11-08T18:08:52Z | 2025-08-30T02:50:18Z | peft | 24 | 24 | null | null | null | [
".gitattributes",
"README.md",
"adapter_config.json",
"adapter_model.safetensors",
"asserts/SinLlama.png",
"optimizer.pt",
"rng_state_0.pth",
"rng_state_1.pth",
"rng_state_2.pth",
"rng_state_3.pth",
"scheduler.pt",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer_config.json",
"tr... | null | null | 4f464b14d56935e20d31a07e9a1ac3e5f31d8dfe | [
"peft",
"safetensors",
"si",
"dataset:polyglots/MADLAD_CulturaX_cleaned",
"arxiv:2508.09115",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:adapter:meta-llama/Meta-Llama-3-8B",
"region:us"
] | null | base_model: meta-llama/Meta-Llama-3-8B
library_name: peft
---
# Model Card for SinLlama
SinLlama is the first large language model specifically extended for Sinhala. It is based on Meta-Llama-3-8B and adapted through tokenizer vocabulary extension and continual pretraining on a 10M sentence Sinhala corpus. SinLlama s... | [
"Ayesh84/Sinhala-bot"
] | null | [
"polyglots/MADLAD_CulturaX_cleaned"
] | [
"si"
] | null | null | null | [
"precision",
"recall",
"f1"
] | null | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
689cd02c6582d51ced24852d | zju-community/matchanything_eloftr | zju-community | null | 1,355 | 1,355 | False | 2025-08-13T17:49:32Z | 2025-08-21T04:27:53Z | transformers | 67 | 24 | null | null | {"parameters": {"F32": 16050816}, "total": 16050816} | [
".gitattributes",
"README.md",
"config.json",
"model.safetensors",
"preprocessor_config.json"
] | null | null | 7bd52a4d5e2ca0f7c4edfaa518a25fb1cd6eea47 | [
"transformers",
"safetensors",
"efficientloftr",
"keypoint-matching",
"arxiv:2501.07556",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | # MatchAnything-ELOFTR
The MatchAnything-ELOFTR model was proposed in **"MatchAnything: Universal Cross-Modality Image Matching with Large-Scale Pre-Training"** by Xingyi He, Hao Yu, Sida Peng, Dongli Tan, Zehong Shen, Hujun Bao, and Xiaowei Zhou from Zhejiang University and Shandong University.
This model is a versi... | [
"zju-community/efficientloftr"
] | [
"apache-2.0"
] | null | null | 16,050,816 | null | [
null
] | null | [
"efficientloftr",
"EfficientLoFTRForKeypointMatching"
] | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68a3deeddc7c9dccc5113f56 | Comfy-Org/Qwen-Image-Edit_ComfyUI | Comfy-Org | null | 219,174 | 219,174 | False | 2025-08-19T02:18:21Z | 2025-08-19T02:41:23Z | diffusion-single-file | 116 | 24 | null | null | null | [
".gitattributes",
"README.md",
"split_files/diffusion_models/qwen_image_edit_bf16.safetensors",
"split_files/diffusion_models/qwen_image_edit_fp8_e4m3fn.safetensors"
] | [
1519,
72,
40861031488,
20430635136
] | 61,291,668,215 | abbda39d0283bbafdb48ae1ffa0f4c0d60bd8717 | [
"diffusion-single-file",
"comfyui",
"license:apache-2.0",
"region:us"
] | null | null | [
"apache-2.0"
] | null | null | null | null | null | null | null | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null | |
68adca40759ab009f3c16b37 | TheDrummer/GLM-Steam-106B-A12B-v1 | TheDrummer | {
"models": [
{
"_id": "687c61c324649ecb26a748f0",
"id": "zai-org/GLM-4.5-Air"
}
],
"relation": "finetune"
} | 181 | 181 | False | 2025-08-26T14:52:48Z | 2025-08-29T09:17:10Z | null | 24 | 24 | null | null | {"parameters": {"F32": 128, "BF16": 110468824704}, "total": 106852245504} | [
".gitattributes",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00043.safetensors",
"model-00002-of-00043.safetensors",
"model-00003-of-00043.safetensors",
"model-00004-of-00043.safetensors",
"model-00005-of-00043.safetensors",
"model-00006-of-00043... | null | null | 851c5aa95110966e2dd7f1e5f5fe7cba94cad4c4 | [
"safetensors",
"glm4_moe",
"base_model:zai-org/GLM-4.5-Air",
"base_model:finetune:zai-org/GLM-4.5-Air",
"region:us"
] | null | # Join our Discord! https://discord.gg/BeaverAI
## Nearly 7000 members strong 💪 A hub for users and makers alike!
---
### Thank you to everyone who subscribed through [Patreon](https://www.patreon.com/TheDrummer). Your suppprt helps me chug along in this brave new world.
---
[Drummer](https://huggingface.co/TheDrumme... | null | null | null | null | 106,852,245,504 | null | null | null | [
"Glm4MoeForCausalLM",
"glm4_moe"
] | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68ae7e692e4f6737fab0edf9 | starsfriday/Qwen-Image-Edit-Remove-Clothes | starsfriday | {
"models": [
{
"_id": "68a19381db43c983deb63fa5",
"id": "Qwen/Qwen-Image-Edit"
}
],
"relation": "adapter"
} | 1,971 | 1,971 | False | 2025-08-27T03:41:29Z | 2025-08-28T05:40:02Z | diffusers | 24 | 24 | null | image-to-image | null | [
".gitattributes",
"Qwen-Edit-LORA.json",
"README.md",
"qwen-edit-remove-clothes.safetensors",
"result/result1.png",
"result/result2.png",
"result/result3.png",
"result/test.jpg"
] | [
1684,
18792,
3011,
472047184,
4402875,
2024841,
5457153,
17115
] | 483,972,655 | d3f9caf27e7bf81e2799fd0f57b07640ee524409 | [
"diffusers",
"image-generation",
"lora",
"Qwen-Image",
"image-to-image",
"en",
"base_model:Qwen/Qwen-Image-Edit",
"base_model:adapter:Qwen/Qwen-Image-Edit",
"license:apache-2.0",
"region:us"
] | null | # starsfriday Qwen-Image-Edit LoRA
<Gallery />
## Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
This is a model for object removal, trained on ```Qwen/Qwen-Image-Edit```, and it is mainly used to remove clothes from characters.For use in ```ComfyUI```.
The greatest advantage o... | null | [
"apache-2.0"
] | null | [
"en"
] | null | null | [
"image-to-image"
] | null | null | [
"vision"
] | [
"image"
] | [
"image"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
6555d075325783cc791fb794 | pyannote/speaker-diarization-3.1 | pyannote | null | 18,594,861 | 191,900,740 | auto | 2023-11-16T08:19:01Z | 2024-05-10T19:43:23Z | pyannote-audio | 1,089 | 23 | null | automatic-speech-recognition | null | [
".gitattributes",
".github/workflows/sync_to_hub.yaml",
"README.md",
"config.yaml",
"handler.py",
"reproducible_research/AISHELL.SpeakerDiarization.Benchmark.test.eval",
"reproducible_research/AISHELL.SpeakerDiarization.Benchmark.test.rttm",
"reproducible_research/AMI-SDM.SpeakerDiarization.Benchmark.... | [
1519,
467,
10985,
469,
2167,
3359,
624408,
2699,
572959,
2699,
572817,
8235,
365238,
3479,
940136,
37082,
3600418,
65701,
648416,
10724,
1441591,
31859,
1979557,
21
] | 10,927,005 | 84fd25912480287da0247647c3d2b4853cb3ee5d | [
"pyannote-audio",
"pyannote",
"pyannote-audio-pipeline",
"audio",
"voice",
"speech",
"speaker",
"speaker-diarization",
"speaker-change-detection",
"voice-activity-detection",
"overlapped-speech-detection",
"automatic-speech-recognition",
"arxiv:2111.14448",
"arxiv:2012.01477",
"license:m... | null | null | [
"r3gm/SoniTranslate_translate_audio_of_a_video_content",
"JonnyTran/SoniTranslate",
"jhj0517/Whisper-WebUI",
"RO-Rtechs/Translate_Video_language",
"waloneai/VideoTranslate_translate_audio_of_a_video_content",
"ganga4364/stt-tibetan",
"pyannote/pretrained-pipelines",
"avans06/whisper-webui-translate",
... | [
"mit"
] | null | null | null | null | [
"automatic-speech-recognition"
] | null | null | [
"multimodal"
] | [
"audio"
] | [
"text"
] | free | non-profit | [
"France"
] | null | null | [
"Speech"
] | [
"Text Classification"
] | [
"Transformer: Speech Encoder-only"
] | [
"en"
] | [
"Finetuning: Supervised"
] | Disclosed: available | 7 |
689b5d4ecb854cb523d91d09 | CohereLabs/command-a-reasoning-08-2025 | CohereLabs | {
"models": [
{
"_id": "67cffded2c8bb280124570e9",
"id": "CohereLabs/c4ai-command-a-03-2025"
}
],
"relation": "finetune"
} | 8,917 | 8,917 | auto | 2025-08-12T15:27:10Z | 2025-08-21T14:41:43Z | transformers | 105 | 23 | null | text-generation | {"parameters": {"BF16": 111057580032}, "total": 111057580032} | [
".gitattributes",
"README.md",
"additional_chat_templates/rag.jinja",
"additional_chat_templates/tool_use.jinja",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00049.safetensors",
"model-00002-of-00049.safetensors",
"model-00003-of-00049.safetensors",
"model-000... | [
1570,
10607,
13966,
13945,
23757,
2562,
139,
6291456144,
4932527624,
4278215728,
4932552312,
4278215728,
4278215728,
4932552312,
4278215728,
4278215744,
4932552328,
4278215736,
4278215736,
4932552328,
4278215736,
4278215736,
4932552328,
4278215736,
4278215736,
49325... | 222,137,990,910 | 68e8ae22cb0ef42e87d3e0347b94bfc338e004c0 | [
"transformers",
"safetensors",
"cohere2",
"text-generation",
"conversational",
"en",
"fr",
"de",
"es",
"it",
"pt",
"ja",
"ko",
"zh",
"ar",
"el",
"fa",
"pl",
"id",
"cs",
"he",
"hi",
"nl",
"ro",
"ru",
"tr",
"uk",
"vi",
"base_model:CohereLabs/c4ai-command-a-03-20... | null | null | [
"nazdridoy/inferoxy-hub",
"ReallyFloppyPenguin/CohereLabs-command-a-reasoning-08-2025",
"dnzzh/CohereLabs-command-a-reasoning-08-2025",
"CodeHubb/CohereLabs-command-a-reasoning-08-2025",
"AstralVisions/CohereLabs-command-a-reasoning-08-2025"
] | [
"cc-by-nc-4.0"
] | null | [
"en",
"fr",
"de",
"es",
"it",
"pt",
"ja",
"ko",
"zh",
"ar",
"el",
"fa",
"pl",
"id",
"cs",
"he",
"hi",
"nl",
"ro",
"ru",
"tr",
"uk",
"vi"
] | 111,057,580,032 | null | [
"text-generation"
] | null | [
"Cohere2ForCausalLM",
"AutoModelForCausalLM",
"cohere2"
] | [
"text"
] | [
"text"
] | [
"text"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
689ecb9afc79f2b3ec739847 | AIDC-AI/Ovis2.5-9B | AIDC-AI | null | 10,903 | 10,903 | False | 2025-08-15T05:54:34Z | 2025-08-23T04:52:56Z | transformers | 273 | 23 | null | image-text-to-text | {"parameters": {"BF16": 9174807784}, "total": 9174807784} | [
".gitattributes",
"LICENSE",
"NOTICE",
"README.md",
"added_tokens.json",
"chat_template.json",
"config.json",
"configuration_ovis2_5.py",
"generation_config.json",
"merges.txt",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"mod... | [
1570,
548,
416,
14606,
707,
1616,
2071,
3936,
277,
1671853,
4905356464,
4915960936,
4974672744,
3553737368,
84690,
44342,
394,
613,
11422654,
7018,
2776833
] | 18,365,761,656 | ab82b02a8ded7852b2062ad8ededb9f2b3ddb599 | [
"transformers",
"safetensors",
"ovis2_5",
"text-generation",
"MLLM",
"image-text-to-text",
"conversational",
"custom_code",
"en",
"zh",
"dataset:AIDC-AI/Ovis-dataset",
"arxiv:2508.11737",
"arxiv:2405.20797",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | null | # Ovis2.5-9B
<div align="center">
<img src=https://cdn-uploads.huggingface.co/production/uploads/637aebed7ce76c3b834cea37/3IK823BZ8w-mz_QfeYkDn.png width="30%"/>
</div>
<p align="center">
<a href="https://arxiv.org/abs/2508.11737"><img src="https://img.shields.io/badge/📖_Technical_Report-Ovis2.5-b31b1b.svg" alt="... | [
"AIDC-AI/Ovis2.5-9B",
"davanstrien/ocr-time-machine",
"AIDC-AI/Ovis2.5-2B",
"Agung1453/Ovis2.5-9B",
"storytracer/ocr-time-machine"
] | [
"apache-2.0"
] | [
"AIDC-AI/Ovis-dataset"
] | [
"en",
"zh"
] | 9,174,807,784 | null | [
"text-generation",
"image-text-to-text"
] | null | [
"AutoModelForCausalLM",
"modeling_ovis2_5.Ovis2_5",
"Ovis2_5",
"ovis2_5"
] | [
"multimodal",
"text"
] | [
"text",
"image"
] | [
"text"
] | team | company | [
"China"
] | null | null | null | null | null | null | null | null | null |
689f39ffe0f517eca6293d1d | Alibaba-DAMO-Academy/RynnEC-7B | Alibaba-DAMO-Academy | null | 171 | 171 | False | 2025-08-15T13:45:35Z | 2025-08-26T08:19:00Z | null | 28 | 23 | null | null | {"parameters": {"BF16": 8297112866}, "total": 8297112866} | [
".gitattributes",
"README.md",
"added_tokens.json",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00004.safetensors",
"model-00002-of-00004.safetensors",
"model-00003-of-00004.safetensors",
"model-00004-of-00004.safetensors",
"model.safetensors.index.json",
"special_to... | null | null | c4014a63a198e553b85e7bf6bead605dc0bb8035 | [
"safetensors",
"rynnec_qwen2",
"arxiv:2508.14160",
"license:apache-2.0",
"region:us"
] | null | <p align="center">
<img src="https://github.com/alibaba-damo-academy/RynnEC/blob/main/assets/logo.jpg?raw=true" width="150" style="margin-bottom: 0.2;"/>
<p>
<h3 align="center"><a href="" style="color:#9C276A">
RynnEC: Bringing MLLMs into Embodied World</a></h3>
<h5 align="center"> If our project helps you, please... | null | [
"apache-2.0"
] | null | null | 8,297,112,866 | null | null | null | [
"RynnecQwen2ForCausalLM",
"rynnec_qwen2"
] | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68ac918d4fcf2623b747ff00 | OpenGVLab/InternVL3_5-38B | OpenGVLab | {
"models": [
{
"_id": "68ac918d0c2b29fb0cd80406",
"id": "OpenGVLab/InternVL3_5-38B-MPO"
}
],
"relation": "finetune"
} | 1,885 | 1,885 | False | 2025-08-25T16:38:37Z | 2025-08-29T17:57:02Z | transformers | 23 | 23 | null | image-text-to-text | {"parameters": {"BF16": 38390405504}, "total": 38390405504} | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.jinja",
"config.json",
"configuration_intern_vit.py",
"configuration_internvl_chat.py",
"conversation.py",
"generation_config.json",
"merges.txt",
"model-00001-of-00016.safetensors",
"model-00002-of-00016.safetensors",
"model... | null | null | de99855be3642cd44fe97c9b72d70e5ce2c07f69 | [
"transformers",
"safetensors",
"internvl_chat",
"feature-extraction",
"internvl",
"custom_code",
"image-text-to-text",
"conversational",
"multilingual",
"dataset:OpenGVLab/MMPR-v1.2",
"dataset:OpenGVLab/MMPR-Tiny",
"arxiv:2312.14238",
"arxiv:2404.16821",
"arxiv:2412.05271",
"arxiv:2411.1... | null | # InternVL3_5-38B
[\[📂 GitHub\]](https://github.com/OpenGVLab/InternVL) [\[📜 InternVL 1.0\]](https://huggingface.co/papers/2312.14238) [\[📜 InternVL 1.5\]](https://huggingface.co/papers/2404.16821) [\[📜 InternVL 2.5\]](https://huggingface.co/papers/2412.05271) [\[📜 InternVL2.5-MPO\]](https://huggingface.co/pa... | null | [
"apache-2.0"
] | [
"OpenGVLab/MMPR-v1.2",
"OpenGVLab/MMPR-Tiny"
] | [
"multilingual"
] | 38,390,405,504 | null | [
"feature-extraction",
"image-text-to-text"
] | null | [
"modeling_internvl_chat.InternVLChatModel",
"AutoModel",
"InternVLChatModel",
"internvl_chat"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"embeddings",
"text"
] | free | community | [
"China"
] | null | null | null | null | null | null | null | null | null |
6641997de7d4af2dcc8c77ce | microsoft/kosmos-2.5 | microsoft | null | 2,134 | 46,269 | False | 2024-05-13T04:39:25Z | 2025-08-28T14:15:13Z | transformers | 220 | 22 | null | image-text-to-text | {"parameters": {"F32": 1374646272}, "total": 1374646272} | [
".gitattributes",
"README.md",
"ckpt.pt",
"config.json",
"generation_config.json",
"md.py",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"ocr.py",
"output.png",
"preprocessor_config.json",
"receipt_00008.png",
"special_tokens_map.j... | [
1620,
6198,
6165757107,
4058,
178,
1097,
4995252144,
503408384,
56481,
2487,
1663635,
107,
1853544,
629,
8859691,
1434622
] | 11,678,301,982 | ec3c8051b697166514a31d646cfa36d6ef4c93d7 | [
"transformers",
"safetensors",
"kosmos-2.5",
"image-to-text",
"image-text-to-text",
"en",
"arxiv:2309.11419",
"license:mit",
"endpoints_compatible",
"region:us"
] | null | # Kosmos-2.5
[Microsoft Document AI](https://www.microsoft.com/en-us/research/project/document-ai/) | [GitHub](https://github.com/microsoft/unilm/tree/master/kosmos-2.5)
## Model description
Kosmos-2.5 is a multimodal literate model for machine reading of text-intensive images. Pre-trained on large-scale text-intens... | [
"nielsr/kosmos-2.5-demo",
"dan-durbin/Kosmos-2.5",
"eric-23fe2/microsoft-kosmos-2.5",
"IndianChessMans/microsoft-kosmos-2.5",
"jazzisfuture/microsoft-kosmos-2.5",
"srangaiah/microsoft-kosmos-2.5",
"iovex/microsoft-kosmos-2.5"
] | [
"mit"
] | null | [
"en"
] | 1,374,646,272 | null | [
"image-to-text",
"image-text-to-text"
] | null | [
"kosmos-2.5",
"Kosmos2_5ForConditionalGeneration",
"AutoModelForVision2Seq"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] | free | company | [
"United States of America",
"International",
"India",
"Belgium"
] | null | null | null | null | null | null | null | null | null |
6795ffcd88cd7c0294702a72 | Qwen/Qwen2.5-VL-7B-Instruct | Qwen | null | 3,660,403 | 24,552,651 | False | 2025-01-26T09:26:37Z | 2025-04-06T16:23:01Z | transformers | 1,185 | 22 | null | image-text-to-text | {"parameters": {"BF16": 8292166656}, "total": 8292166656} | [
".gitattributes",
"README.md",
"chat_template.json",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00005.safetensors",
"model-00002-of-00005.safetensors",
"model-00003-of-00005.safetensors",
"model-00004-of-00005.safetensors",
"model-00005-of-00005.safetensors",
"model... | [
1519,
18574,
1050,
1374,
216,
1671839,
3900233256,
3864726320,
3864726424,
3864733680,
1089994880,
57619,
350,
7031645,
5702,
2776833
] | 16,595,981,281 | cc594898137f460bfe9f0759e9844b3ce807cfb5 | [
"transformers",
"safetensors",
"qwen2_5_vl",
"image-to-text",
"multimodal",
"image-text-to-text",
"conversational",
"en",
"arxiv:2309.00071",
"arxiv:2409.12191",
"arxiv:2308.12966",
"license:apache-2.0",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | # Qwen2.5-VL-7B-Instruct
<a href="https://chat.qwenlm.ai/" target="_blank" style="margin: 2px;">
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
</a>
## Introduction
In the past five months since Qwen2-VL’s r... | [
"multimodalart/Qwen-Image-Edit-Fast",
"Qwen/Qwen-Image-Edit",
"zerogpu-aoti/Qwen-Image-Edit-Relight",
"zerogpu-aoti/Qwen-Image-Edit-Outpaint",
"fffiloni/Meigen-MultiTalk",
"Wan-AI/Wan-2.2-5B",
"zerogpu-aoti/Qwen-Image-Edit-Multi-Image",
"wcy1122/MGM-Omni",
"innoai/self-forcing",
"Heartsync/VEO3-Re... | [
"apache-2.0"
] | null | [
"en"
] | 8,292,166,656 | null | [
"image-to-text",
"image-text-to-text"
] | null | [
"AutoModelForVision2Seq",
"Qwen2_5_VLForConditionalGeneration",
"qwen2_5_vl"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] | team | company | [
"China"
] | null | Qwen/Qwen2-VL-7B | [
"Image",
" Video",
" Text"
] | [
"Text Generation",
" Image Bounding Boxes",
" Text Classification",
" Image Classification",
" Other"
] | [
"Transformer: Text Decoder-only",
" Transformer: Image Encoder-only"
] | [
"EN"
] | [
"Pretraining: Multimodal joint-embeddings",
" Pretraining: Supervised",
" Finetuning: Supervised",
" Instruction finetuning"
] | Not disclosed | 9 |
686ba3cf000440bf9721d5a6 | black-forest-labs/FLUX.1-Krea-dev | black-forest-labs | {
"models": [
{
"_id": "66aaa908fc35e079a941470d",
"id": "black-forest-labs/FLUX.1-dev"
}
],
"relation": "finetune"
} | 113,422 | 134,516 | auto | 2025-07-07T10:39:11Z | 2025-07-31T14:33:39Z | diffusers | 721 | 22 | null | text-to-image | null | [
".gitattributes",
"LICENSE.md",
"README.md",
"ae.safetensors",
"flux1-krea-dev.safetensors",
"model_index.json",
"scheduler/scheduler_config.json",
"teaser.png",
"text_encoder/config.json",
"text_encoder/model.safetensors",
"text_encoder_2/config.json",
"text_encoder_2/model-00001-of-00002.saf... | [
1655,
18640,
8164,
335304388,
23802958224,
536,
273,
12641428,
613,
246144352,
782,
4994582224,
4530066360,
19885,
524619,
588,
705,
1059962,
2543,
791656,
2424235,
20817,
394,
9983040304,
9949328904,
3870584832,
121262,
820,
167666902
] | 57,897,316,067 | 8162a9c7b05a641be098422bf2fcf335615c2f28 | [
"diffusers",
"safetensors",
"text-to-image",
"image-generation",
"flux",
"en",
"base_model:black-forest-labs/FLUX.1-dev",
"base_model:finetune:black-forest-labs/FLUX.1-dev",
"license:other",
"endpoints_compatible",
"diffusers:FluxPipeline",
"region:us"
] | null | null | [
"bytedance-research/USO",
"black-forest-labs/FLUX.1-Krea-dev",
"prithivMLmods/FLUX-REALISM",
"yanze/PuLID-FLUX",
"Nymbo/FLUX.1-Krea-dev",
"Nymbo/Tools",
"Nymbo/FLUX.1-Krea-Serverless",
"nazdridoy/inferoxy-hub",
"bep40/USO",
"AlekseyCalvin/soonfactory",
"ovi054/FLUX.Dev-LORA-Serverless",
"jiufa... | [
"other",
"flux-1-dev-non-commercial-license",
"LICENSE.md"
] | null | [
"en"
] | null | null | [
"text-to-image"
] | null | null | [
"vision"
] | [
"text"
] | [
"image"
] | team | company | [
"Germany"
] | Accept to share username & email | black-forest-labs/FLUX.1-dev | [
"Text"
] | [
"Image Generation"
] | [
"Diffusion-based Network"
] | [
"en"
] | [
"Knowledge distillation"
] | Not disclosed | 3 |
687c61c324649ecb26a748f0 | zai-org/GLM-4.5-Air | zai-org | null | 195,192 | 200,657 | False | 2025-07-20T03:25:55Z | 2025-08-11T13:25:37Z | transformers | 402 | 22 | null | text-generation | {"parameters": {"BF16": 110468818944, "F32": 5888}, "total": 110468824832} | [
".gitattributes",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00047.safetensors",
"model-00002-of-00047.safetensors",
"model-00003-of-00047.safetensors",
"model-00004-of-00047.safetensors",
"model-00005-of-00047.safetensors",
"model-00006-of-00047... | [
1570,
3001,
3242,
1008,
155,
2970138176,
4683035216,
4683035216,
4683035216,
4683035216,
4683035216,
4683035216,
4683035216,
4683035216,
4683035216,
4683035616,
4683035616,
4683035616,
4683035616,
4683035616,
4683035616,
4683035616,
4683035616,
4683035616,
4683035616,... | 220,961,581,797 | a24ceef6ce4f3536971efe9b778bdaa1bab18daa | [
"transformers",
"safetensors",
"glm4_moe",
"text-generation",
"conversational",
"en",
"zh",
"arxiv:2508.06471",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | # GLM-4.5-Air
<div align="center">
<img src=https://raw.githubusercontent.com/zai-org/GLM-4.5/refs/heads/main/resources/logo.svg width="15%"/>
</div>
<p align="center">
👋 Join our <a href="https://discord.gg/QR7SARHRxK" target="_blank">Discord</a> community.
<br>
📖 Check out the GLM-4.5 <a href="https://... | [
"umint/ai",
"umint/o4-mini",
"William9875/zai-org-GLM-4.5-Air",
"wuhuizgptamd/ai",
"AXJD/zai-org-GLM-4.5-Air",
"umint/gpt-4.1-nano",
"umint/o3",
"stackway-ai/openwebui",
"umint/openwebui"
] | [
"mit"
] | null | [
"en",
"zh"
] | 110,468,824,832 | null | [
"text-generation"
] | null | [
"Glm4MoeForCausalLM",
"glm4_moe",
"AutoModelForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
6881e60ffcffaee6d84fe9e4 | Wan-AI/Wan2.2-I2V-A14B | Wan-AI | null | 10,637 | 13,737 | False | 2025-07-24T07:51:43Z | 2025-08-07T09:42:48Z | wan2.2 | 270 | 22 | null | image-to-video | null | [
".gitattributes",
"README.md",
"Wan2.1_VAE.pth",
"assets/comp_effic.png",
"assets/logo.png",
"assets/moe_2.png",
"assets/moe_arch.png",
"assets/performance.png",
"assets/vae.png",
"configuration.json",
"examples/i2v_input.JPG",
"google/umt5-xxl/special_tokens_map.json",
"google/umt5-xxl/spie... | [
1866,
16192,
507609880,
202156,
56322,
527914,
74900,
306535,
165486,
47,
250628,
6623,
4548313,
16837417,
61728,
250,
9994119944,
9943937936,
9943979184,
9839059744,
9839059744,
7595559224,
96805,
250,
9994119944,
9943937936,
9943979184,
9839059744,
983905974... | 126,205,359,703 | 206a9ee1b7bfaaf8f7e4d81335650533490646a3 | [
"wan2.2",
"diffusers",
"safetensors",
"image-to-video",
"en",
"zh",
"arxiv:2503.20314",
"license:apache-2.0",
"region:us"
] | null | # Wan2.2
<p align="center">
<img src="assets/logo.png" width="400"/>
<p>
<p align="center">
💜 <a href="https://wan.video"><b>Wan</b></a>    |    🖥️ <a href="https://github.com/Wan-Video/Wan2.2">GitHub</a>    |   🤗 <a href="https://huggingface.co/Wan-AI/">Hugging Face</a>&nb... | [
"wavespeed/wan2.2",
"Fgasa/Wan-AI-Wan2.2-I2V-A14B",
"dapersonperson/Image2VIdeo",
"ufogr/Wan-AI-Wan2.2-I2V-A14B",
"Beatwrecka/Wan-AI-Wan2.2-I2V-A14B",
"ziwaixian009/wan",
"Jatin264/Wan-AI-Wan2.2-I2V-A14B",
"Gee6ix/Wan-AI-Wan2.2-I2V-A14B",
"hugger666666/Wan-AI-Wan2.2-I2V-A14B",
"Surendara1991/Wan-A... | [
"apache-2.0"
] | null | [
"en",
"zh"
] | null | null | [
"image-to-video"
] | null | null | [
"vision"
] | [
"text",
"image"
] | [
"video"
] | free | company | [
"China"
] | null | null | null | null | null | null | null | null | null |
688b1597e5e83e19d1b3238a | Qwen/Qwen3-Coder-30B-A3B-Instruct | Qwen | null | 367,357 | 373,293 | False | 2025-07-31T07:04:55Z | 2025-08-21T10:18:32Z | transformers | 528 | 22 | null | text-generation | {"parameters": {"BF16": 30532122624}, "total": 30532122624} | [
".gitattributes",
"LICENSE",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00016.safetensors",
"model-00002-of-00016.safetensors",
"model-00003-of-00016.safetensors",
"model-00004-of-00016.safetensors",
"model-00005-of-00016.safetens... | [
1519,
11343,
5425,
6211,
992,
180,
1671839,
3998893112,
3999974192,
3997360832,
3999975056,
3999975400,
3999975400,
3999975472,
3997362064,
3999975408,
3999975400,
3999975408,
3987924896,
3999975088,
3999975400,
3999975400,
1085307128,
1699758,
31613,
7032399,
130... | 61,079,826,823 | 573fa3901e5799703b1e60825b0ec024a4c0f1d3 | [
"transformers",
"safetensors",
"qwen3_moe",
"text-generation",
"conversational",
"arxiv:2505.09388",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | # Qwen3-Coder-30B-A3B-Instruct
<a href="https://chat.qwen.ai/" target="_blank" style="margin: 2px;">
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
</a>
## Highlights
**Qwen3-Coder** is available in multiple... | [
"nazdridoy/inferoxy-hub",
"RaulGuo1/ttt1",
"AheedTahir/First_agent_template",
"sadsawq/Flower",
"Semnykcz/Qwen3",
"jameshazra220/Qwen-Qwen3-Coder-30B-A3B-Instruct",
"soupstick/advanced-fraud-analyst",
"jrmagallanes/First_agent_template",
"Kucjt/Qwen-Qwen3-Coder-30B-A3B-Instruct",
"johnflash2007/Qw... | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen3-Coder-30B-A3B-Instruct/blob/main/LICENSE"
] | null | null | 30,532,122,624 | null | [
"text-generation"
] | null | [
"Qwen3MoeForCausalLM",
"qwen3_moe",
"AutoModelForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | team | company | [
"China"
] | null | null | null | null | null | null | null | null | null |
6891e3bb084ce75acffb033d | Qwen/Qwen3-4B-Instruct-2507 | Qwen | null | 629,200 | 629,200 | False | 2025-08-05T10:58:03Z | 2025-08-06T11:08:47Z | transformers | 254 | 22 | null | text-generation | {"parameters": {"BF16": 4022468096}, "total": 4022468096} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00003.safetensors",
"model-00002-of-00003.safetensors",
"model-00003-of-00003.safetensors",
"model.safetensors.index.json",
"tokenizer.json",
"tokenizer_config.json",
"vocab.json"... | [
1570,
11343,
8168,
727,
238,
1671839,
3957900840,
3987450520,
99630640,
32819,
11422654,
10824,
2776833
] | 8,060,919,015 | eb25fbe4f35f7147763bc24445679d1c00588d89 | [
"transformers",
"safetensors",
"qwen3",
"text-generation",
"conversational",
"arxiv:2505.09388",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | # Qwen3-4B-Instruct-2507
<a href="https://chat.qwen.ai" target="_blank" style="margin: 2px;">
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
</a>
## Highlights
We introduce the updated version of the **Qwen3... | [
"yellowcandle/chinese-essay-streamlit",
"Nishath2025/qwen-finetuned",
"akhaliq/Qwen3-4B-Instruct-2507",
"BolaNash/New_Brain",
"hubvale/Qwen-Qwen3-4B-Instruct-2507",
"Pavan7424/Task_Priority",
"Myoussef11/Voice_Analysis_Toolkit",
"MightyOctopus/mockup-data-generator",
"biaogd/Qwen-Qwen3-4B-Instruct-2... | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507/blob/main/LICENSE"
] | null | null | 4,022,468,096 | null | [
"text-generation"
] | null | [
"AutoModelForCausalLM",
"Qwen3ForCausalLM",
"qwen3"
] | [
"text"
] | [
"text"
] | [
"text"
] | team | company | [
"China"
] | null | null | null | null | null | null | null | null | null |
689cad98683d345ff0e9fff9 | nasa-ibm-ai4science/Surya-1.0 | nasa-ibm-ai4science | null | 212 | 212 | False | 2025-08-13T15:22:00Z | 2025-08-20T00:57:57Z | null | 80 | 22 | null | null | null | [
".gitattributes",
"README.md",
"SuryaFM.pdf",
"config.yaml",
"scalers.yaml",
"surya.366m.v1.pt",
"surya_arch.png"
] | null | null | 5cc4b5386d5f78fda3896b1389589d4e173bf212 | [
"Pytorch",
"Heliophysics",
"Space Weather",
"Time Series",
"Foundation Model",
"NASA",
"IBM",
"SDO",
"license:apache-2.0",
"region:us"
] | null | # Surya 1.0
NASA, IBM, and partners present **Surya**, the first open-source AI **foundation model for heliophysics**.
Surya is a 366M-parameter transformer model pretrained on **9 years (≈218 TB)** of multi-instrument data from NASA’s [Solar Dynamics Observatory (SDO)](https://sdo.gsfc.nasa.gov/), including 8 Atmos... | [
"AndersonConforto/Test",
"broadfield-dev/surya-demo",
"johannesschmude/surya_visual_forecasting_demo"
] | [
"apache-2.0"
] | null | null | null | null | null | null | null | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68a1ddddd451385a609815bd | FreeSEED-AI/gpt-oss-120b-mandarin-thinking | FreeSEED-AI | {
"models": [
{
"_id": "68913522f16f3c8aaffccf1f",
"id": "openai/gpt-oss-120b"
}
],
"relation": "finetune"
} | 460 | 460 | False | 2025-08-17T13:49:17Z | 2025-08-17T14:47:17Z | null | 27 | 22 | null | null | {"parameters": {"BF16": 116829156672}, "total": 116829156672} | [
".gitattributes",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00073.safetensors",
"model-00002-of-00073.safetensors",
"model-00003-of-00073.safetensors",
"model-00004-of-00073.safetensors",
"model-00005-of-00073.safetensors",
"model-00006-of-00073... | null | null | b1a21e693fc05ed76cbc4505e7bd2936a759157c | [
"safetensors",
"gpt_oss",
"zh",
"base_model:openai/gpt-oss-120b",
"base_model:finetune:openai/gpt-oss-120b",
"license:apache-2.0",
"region:us"
] | null | # GPT-OSS-ZhTW-Thinking
[](https://huggingface.co/FreeSEED-AI/gpt-oss-zhtw-thinking)
[](LICENSE)
A specialized language model optimized for thinking in Traditional Chine... | null | [
"apache-2.0"
] | null | [
"zh"
] | 116,829,156,672 | null | null | null | [
"GptOssForCausalLM",
"gpt_oss"
] | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
678098bbf05671ace4c99e1c | blurgy/CoMPaSS-FLUX.1 | blurgy | {
"models": [
{
"_id": "66aaa908fc35e079a941470d",
"id": "black-forest-labs/FLUX.1-dev"
}
],
"relation": "adapter"
} | 197 | 197 | False | 2025-01-10T03:49:15Z | 2025-08-26T11:30:47Z | diffusers | 21 | 21 | null | text-to-image | null | [
".gitattributes",
"LICENSE",
"README.md",
"images/bird-below-skateboard.jpg",
"images/horse-left-bottle.jpg",
"images/laptop-above-dog.jpg",
"lora.safetensors"
] | null | null | b834f5fdeda29572d6b4e6e00f59730d42f54f64 | [
"diffusers",
"text-to-image",
"lora",
"template:diffusion-lora",
"arxiv:2412.13195",
"base_model:black-forest-labs/FLUX.1-dev",
"base_model:adapter:black-forest-labs/FLUX.1-dev",
"license:other",
"region:us"
] | null | # CoMPaSS-FLUX.1
\[[Project Page]\]
\[[code]\]
\[[arXiv]\]
<Gallery />
## Model description
# CoMPaSS-FLUX.1
A LoRA adapter that enhances spatial understanding capabilities of the FLUX.1 text-to-image
diffusion model. This model demonstrates significant improvements in generating images with specific
spatial rela... | null | [
"other",
"compass-lora-weights-nc-license",
"LICENSE"
] | null | null | null | null | [
"text-to-image"
] | null | null | [
"vision"
] | [
"text"
] | [
"image"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
678dc6fff905d106be796d8a | deepseek-ai/DeepSeek-R1 | deepseek-ai | null | 493,704 | 10,810,897 | False | 2025-01-20T03:46:07Z | 2025-03-27T04:01:59Z | transformers | 12,662 | 21 | null | text-generation | {"parameters": {"BF16": 3918786560, "F8_E4M3": 680571043840, "F32": 41555600}, "total": 684531386000} | [
".gitattributes",
"LICENSE",
"README.md",
"config.json",
"configuration_deepseek.py",
"figures/benchmark.jpg",
"generation_config.json",
"model-00001-of-000163.safetensors",
"model-00002-of-000163.safetensors",
"model-00003-of-000163.safetensors",
"model-00004-of-000163.safetensors",
"model-00... | [
1519,
1064,
15994,
1660,
9897,
777314,
171,
5234139343,
4302383966,
4302384375,
4302349996,
4302384154,
4372073602,
4306080097,
4302384356,
4302350190,
4302383960,
4302384375,
1321583941,
4302317244,
4302384328,
4302350218,
4302383932,
4302384377,
4302350026,
430238... | 688,604,360,633 | 56d4cbbb4d29f4355bab4b9a39ccb717a14ad5ad | [
"transformers",
"safetensors",
"deepseek_v3",
"text-generation",
"conversational",
"custom_code",
"arxiv:2501.12948",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"fp8",
"region:us"
] | null | # DeepSeek-R1
<!-- markdownlint-disable first-line-h1 -->
<!-- markdownlint-disable html -->
<!-- markdownlint-disable no-duplicate-header -->
<div align="center">
<img src="https://github.com/deepseek-ai/DeepSeek-V2/blob/main/figures/logo.svg?raw=true" width="60%" alt="DeepSeek-V3" />
</div>
<hr>
<div align="center... | [
"umint/ai",
"bhaskartripathi/LLM_Quantization",
"nazdridoy/inferoxy-hub",
"umint/o4-mini",
"sbudni/sk",
"migueldeguzmandev/migueldeguzmandev-papercliptodd_v2",
"Dorjzodovsuren/DorjGPT_assistant",
"alx-d/philosophy_aristotle",
"Lakshan2003/llama-chat",
"ruslanmv/convert_to_gguf",
"dlflannery/Grad... | [
"mit"
] | null | null | 684,531,386,000 | null | [
"text-generation"
] | null | [
"DeepseekV3ForCausalLM",
"deepseek_v3",
"AutoModelForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | free | company | [
"China"
] | null | deepseek-ai/DeepSeek-V3-Base | [
"Text"
] | [
"Text Generation"
] | [
"Transformer: Text Decoder-only"
] | [
"zh",
" en"
] | [
"Finetuning: Supervised",
" Reinforcement learning from feedback"
] | Partially disclosed: unavailable | 6 |
682b9a1c4e0d74489a402e3c | google/medgemma-4b-it | google | {
"models": [
{
"_id": "682b9a06abb94133d05acbfb",
"id": "google/medgemma-4b-pt"
}
],
"relation": "finetune"
} | 108,842 | 362,684 | auto | 2025-05-19T20:52:44Z | 2025-07-09T18:14:57Z | transformers | 641 | 20 | null | image-text-to-text | {"parameters": {"BF16": 4300079472}, "total": 4971331952} | [
".gitattributes",
"README.md",
"added_tokens.json",
"chat_template.jinja",
"config.json",
"generation_config.json",
"model-00001-of-00002.safetensors",
"model-00002-of-00002.safetensors",
"model.safetensors.index.json",
"preprocessor_config.json",
"processor_config.json",
"special_tokens_map.j... | [
4626,
35060,
35,
1532,
2469,
156,
4961251752,
3639026128,
90594,
570,
70,
662,
33384570,
4689074,
1157001
] | 8,639,644,299 | efe6cc02361759b6bd501c654ddb7c9d25ec509d | [
"transformers",
"safetensors",
"gemma3",
"image-text-to-text",
"medical",
"radiology",
"clinical-reasoning",
"dermatology",
"pathology",
"ophthalmology",
"chest-x-ray",
"conversational",
"arxiv:2303.15343",
"arxiv:2507.05201",
"arxiv:2405.03162",
"arxiv:2106.14463",
"arxiv:2412.03555... | null | null | [
"google/rad_explain",
"jbondy007/MedgemmaRadiology",
"mol-wise/MOLx-Powered_by_MedRAX",
"rishiraj/radiology",
"warshanks/medgemma-4b-it",
"bala00712200502/A6B5B",
"khalednabawi11/Medical-Scan-Gemma",
"CAGILDENIZCANDURGUN/llm",
"raheelnext/Gemma",
"yigitsoy/medgemma-4b-it",
"goog-sean/radiology",... | [
"other",
"health-ai-developer-foundations",
"https://developers.google.com/health-ai-developer-foundations/terms"
] | null | null | 4,971,331,952 | null | [
"image-text-to-text"
] | null | [
"AutoModelForImageTextToText",
"Gemma3ForConditionalGeneration",
"gemma3"
] | [
"multimodal"
] | [
"text",
"image"
] | [
"text"
] | enterprise | company | [
"United States of America"
] | null | null | null | null | null | null | null | null | null |
687fa5b678e85b9b93b7bac2 | Qwen/Qwen3-Coder-480B-A35B-Instruct | Qwen | null | 138,706 | 164,142 | False | 2025-07-22T14:52:38Z | 2025-08-21T10:18:07Z | transformers | 1,139 | 20 | null | text-generation | {"parameters": {"BF16": 480154875392}, "total": 480154875392} | [
".gitattributes",
"LICENSE",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"merges.txt",
"model-00001-of-00241.safetensors",
"model-00002-of-00241.safetensors",
"model-00003-of-00241.safetensors",
"model-00004-of-00241.safetensors",
"model-00005-of-00241.safetens... | [
1519,
11343,
5500,
6211,
994,
180,
1671839,
3987218640,
3995090312,
3995090328,
3995090448,
3978206784,
3995090320,
3995090344,
3995090448,
3978206760,
3995090320,
3995090360,
3995090440,
3978206744,
3995090312,
3995090376,
3890125904,
3988799064,
3995090320,
399509... | 960,327,832,085 | 9d90cf8fca1bf7b7acca42d3fc9ae694a2194069 | [
"transformers",
"safetensors",
"qwen3_moe",
"text-generation",
"conversational",
"arxiv:2505.09388",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | # Qwen3-Coder-480B-A35B-Instruct
<a href="https://chat.qwen.ai/" target="_blank" style="margin: 2px;">
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
</a>
## Highlights
Today, we're announcing **Qwen3-Coder*... | [
"enzostvs/deepsite",
"lvwerra/jupyter-agent-2",
"umint/ai",
"Svngoku/jupyter2agent",
"zenafey/Qwen-Qwen3-Coder-480B-A35B-Instruct",
"umint/o4-mini",
"HPAI-BSC/TuRTLe-Leaderboard",
"Elias-CIC/Final_Assignment_Template",
"mgbam/builder",
"WilliamRabuel/GAIA_Agent",
"androaichain/Qwen-Qwen3-Coder-4... | [
"apache-2.0",
"https://huggingface.co/Qwen/Qwen3-Coder-480B-A35B-Instruct/blob/main/LICENSE"
] | null | null | 480,154,875,392 | null | [
"text-generation"
] | null | [
"Qwen3MoeForCausalLM",
"qwen3_moe",
"AutoModelForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | team | company | [
"China"
] | null | null | null | null | null | null | null | null | null |
688833e80f0a1154e901f0d5 | lodestones/Chroma1-Base | lodestones | null | 3,590 | 3,825 | False | 2025-07-29T02:37:28Z | 2025-08-23T10:18:47Z | diffusers | 66 | 20 | null | text-to-image | null | [
".gitattributes",
"Chroma1-Base.safetensors",
"README.md",
"images/FictionalChromaBanner_1.png",
"model_index.json",
"scheduler/scheduler_config.json",
"text_encoder/config.json",
"text_encoder/model-00001-of-00002.safetensors",
"text_encoder/model-00002-of-00002.safetensors",
"text_encoder/model.... | [
1590,
17800038288,
5605,
1008192,
494,
147,
741,
4994582224,
4530066360,
19921,
2593,
2543,
791656,
20847,
490,
9946193392,
7853894360,
106695,
819,
167666902
] | 45,294,403,859 | 50ebe125b67794f4ed8e4985d26a4414ca62aed3 | [
"diffusers",
"safetensors",
"text-to-image",
"license:apache-2.0",
"diffusers:ChromaPipeline",
"region:us"
] | null | # Chroma1-Base
Chroma1-Base is an **8.9B** parameter text-to-image foundational model based on **FLUX.1-schnell**. It is fully **Apache 2.0 licensed**, ensuring that anyone can use, modify, and build upon it.
As a **base model**, Chroma1 is intentionally designed to be an excellent starting point for **finetuning**. ... | null | [
"apache-2.0"
] | null | null | null | null | [
"text-to-image"
] | null | null | [
"vision"
] | [
"text"
] | [
"image"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68923b71467354ab9147aa88 | unsloth/gpt-oss-20b-GGUF | unsloth | {
"models": [
{
"_id": "68913539bd3d0a833438591d",
"id": "openai/gpt-oss-20b"
}
],
"relation": "quantized"
} | 711,207 | 711,207 | False | 2025-08-05T17:12:17Z | 2025-08-21T12:36:55Z | transformers | 350 | 20 | null | text-generation | null | [
".gitattributes",
"README.md",
"config.json",
"gpt-oss-20b-F16.gguf",
"gpt-oss-20b-Q2_K.gguf",
"gpt-oss-20b-Q2_K_L.gguf",
"gpt-oss-20b-Q3_K_M.gguf",
"gpt-oss-20b-Q3_K_S.gguf",
"gpt-oss-20b-Q4_0.gguf",
"gpt-oss-20b-Q4_1.gguf",
"gpt-oss-20b-Q4_K_M.gguf",
"gpt-oss-20b-Q4_K_S.gguf",
"gpt-oss-20b... | [
2760,
8849,
1643,
13792639168,
11468317888,
11757884608,
11506103488,
11463894208,
11501495488,
11577504448,
11624759488,
11618492608,
11717357248,
11711827648,
12041000128,
12109567168,
11872347328,
12041000128,
13195442368,
149,
7355
] | 190,999,654,164 | c3303d94926e0e2262aacdd0fac4b18e1a29468e | [
"transformers",
"gguf",
"gpt_oss",
"text-generation",
"openai",
"unsloth",
"base_model:openai/gpt-oss-20b",
"base_model:quantized:openai/gpt-oss-20b",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us",
"conversational"
] | {"total": 20914757184, "architecture": "gpt-oss", "context_length": 131072, "chat_template": "{# Chat template fixes by Unsloth #}\n{#-\n In addition to the normal inputs of `messages` and `tools`, this template also accepts the\n following kwargs:\n - \"builtin_tools\": A list, can contain \"browser\" and/or \"pyth... | > [!NOTE]
> GGUF uploads with our fixes. More details and [Read our guide here.](https://docs.unsloth.ai/basics/gpt-oss)
>
<div>
<p style="margin-bottom: 0; margin-top: 0;">
<strong>See <a href="https://huggingface.co/collections/unsloth/gpt-oss-6892433695ce0dee42f31681">our collection</a> for all versions of gpt... | [
"dbmoradi60/gpt-oss-20b-cpu",
"Monster/gpt-oss-20b",
"Ahkjtgcfdhzjzxk/New-space"
] | [
"apache-2.0"
] | null | null | null | 20,914,757,184 | [
"text-generation"
] | null | [
"gpt-oss",
"GptOssForCausalLM",
"AutoModelForCausalLM",
"gpt_oss"
] | [
"text"
] | [
"text"
] | [
"text"
] | team | company | [
"United States of America"
] | null | null | null | null | null | null | null | null | null |
6899a242df5caf44d2304d68 | BasedBase/Qwen3-Coder-30B-A3B-Instruct-480B-Distill-V2 | BasedBase | {
"models": [
{
"_id": "688b1597e5e83e19d1b3238a",
"id": "Qwen/Qwen3-Coder-30B-A3B-Instruct"
}
],
"relation": "quantized"
} | 11,239 | 11,239 | False | 2025-08-11T07:56:50Z | 2025-08-18T00:48:51Z | transformers | 51 | 20 | null | null | null | [
".gitattributes",
"Qwen3-30B-A3B-Instruct-Coder-480B-Distill-v2-Q8_0.gguf",
"Qwen3-Coder-30B-A3B-Instruct-480B-Distill-V2-Q2_K.gguf",
"Qwen3-Coder-30B-A3B-Instruct-480B-Distill-V2-Q3_K_M.gguf",
"Qwen3-Coder-30B-A3B-Instruct-480B-Distill-V2-Q4_0.gguf",
"Qwen3-Coder-30B-A3B-Instruct-480B-Distill-V2-Q4_K_M.g... | [
2521,
32483934208,
11258611712,
14711848960,
17304492032,
18556688384,
17456011264,
21080512512,
21725583360,
21080512512,
25092534272,
4252
] | 200,750,735,989 | 493912de63169cf6d7dd84c445fd563bfdc10bc4 | [
"transformers",
"gguf",
"causal-lm",
"moe",
"mixture-of-experts",
"qwen",
"distillation",
"svd",
"lora-merged",
"code-generation",
"en",
"code",
"base_model:Qwen/Qwen3-Coder-30B-A3B-Instruct",
"base_model:quantized:Qwen/Qwen3-Coder-30B-A3B-Instruct",
"license:apache-2.0",
"endpoints_co... | {"total": 30532122624, "architecture": "qwen3moe", "context_length": 262144, "chat_template": "{% macro render_item_list(item_list, tag_name='required') %}\n {%- if item_list is defined and item_list is iterable and item_list | length > 0 %}\n {%- if tag_name %}{{- '\\n<' ~ tag_name ~ '>' -}}{% endif %}\n ... | # A SVD based Distillation of Qwen3-Coder-480B for better code generation
## Model Description
This model is a distilled version of **`Qwen/Qwen3-Coder-30B-A3B-Instruct`** designed to achieve coding and reasoning capabilities approaching those of a much larger teacher model.
It is the result of applying a LoRA made ... | null | [
"apache-2.0"
] | null | [
"en",
"code"
] | null | 30,532,122,624 | [
null
] | null | [
"qwen3moe",
"AutoModel"
] | null | null | null | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
68b1d8c5629f7849e2a37694 | peteromallet/Qwen-Image-Edit-InStyle | peteromallet | {
"models": [
{
"_id": "68a19381db43c983deb63fa5",
"id": "Qwen/Qwen-Image-Edit"
}
],
"relation": "adapter"
} | 0 | 0 | False | 2025-08-29T16:43:49Z | 2025-08-29T19:14:56Z | null | 20 | 20 | null | image-to-image | null | [
".gitattributes",
"InStyle-0.5.safetensors",
"README.md",
"samples.png"
] | null | null | 1ed1b237b5624aaa6e878779f87ea6d1ca089734 | [
"image",
"editing",
"lora",
"style-transfer",
"qwen",
"image-to-image",
"dataset:peteromallet/high-quality-midjouney-srefs",
"base_model:Qwen/Qwen-Image-Edit",
"base_model:adapter:Qwen/Qwen-Image-Edit",
"license:apache-2.0",
"region:us"
] | null | null | null | [
"apache-2.0"
] | [
"peteromallet/high-quality-midjouney-srefs"
] | null | null | null | [
"image-to-image"
] | null | null | [
"vision"
] | [
"image"
] | [
"image"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
65f683e9df29f02c6da01279 | xai-org/grok-1 | xai-org | null | 383 | 59,363 | False | 2024-03-17T05:47:21Z | 2024-03-28T16:25:32Z | grok | 2,361 | 19 | null | text-generation | null | [
".gitattributes",
"README.md",
"RELEASE",
"ckpt-0/tensor00000_000",
"ckpt-0/tensor00001_000",
"ckpt-0/tensor00002_000",
"ckpt-0/tensor00003_000",
"ckpt-0/tensor00004_000",
"ckpt-0/tensor00005_000",
"ckpt-0/tensor00006_000",
"ckpt-0/tensor00007_000",
"ckpt-0/tensor00008_000",
"ckpt-0/tensor00... | [
1570,
975,
1145,
3221225637,
24727,
1611137347,
1611399491,
1611137347,
6293814,
37847359,
37761334,
6293814,
24727,
24727,
24727,
24727,
196770,
1611137347,
1611399491,
1611137347,
6293814,
37847359,
37761334,
6293814,
24727,
24727,
24727,
24727,
196770,
16... | 318,239,889,830 | 5de83eb225f49624b424f1c8aa74f96983b5885c | [
"grok",
"grok-1",
"text-generation",
"license:apache-2.0",
"region:us"
] | null | # Grok-1
This repository contains the weights of the Grok-1 open-weights model. You can find the code in the [GitHub Repository](https://github.com/xai-org/grok-1/tree/main).
# Download instruction
Clone the repo & download the `int8` checkpoint to the `checkpoints` directory by executing this command in the repo roo... | [
"Xenova/the-tokenizer-playground",
"yhavinga/dutch-tokenizer-arena",
"Omnibus/grok-1-test",
"Itsmade/Grok",
"shams1992/the-tokenizer-playground",
"DearGreen/MySpace_04",
"prompts-dot-com/Prompts.com-grok-1",
"doctorsafe/the-tokenizer-playground",
"marlonbarrios/the-tokenizer-playground",
"CLSDNZ/t... | [
"apache-2.0"
] | null | null | null | null | [
"text-generation"
] | null | null | [
"text"
] | [
"text"
] | [
"text"
] | team | company | [
"United States of America"
] | null | null | [
"Text"
] | [
"Text Generation"
] | [
"Transformer: Text Decoder-only"
] | [
"en"
] | [
"Pretraining: Causal Language Modeling (CLM)"
] | Not disclosed | 5 |
682527d3e3eb09c41abcf704 | onnx-community/FastVLM-0.5B-ONNX | onnx-community | null | 4,680 | 5,609 | False | 2025-05-14T23:31:31Z | 2025-07-01T21:21:39Z | transformers.js | 32 | 19 | null | image-text-to-text | null | [
".gitattributes",
"LICENSE",
"README.md",
"added_tokens.json",
"config.json",
"generation_config.json",
"merges.txt",
"onnx/decoder_model_merged.onnx",
"onnx/decoder_model_merged_bnb4.onnx",
"onnx/decoder_model_merged_fp16.onnx",
"onnx/decoder_model_merged_int8.onnx",
"onnx/decoder_model_merge... | [
1570,
5814,
3937,
80,
1328,
121,
1670344,
1983654988,
286590783,
991914846,
502744473,
317445767,
282252137,
502744556,
502744556,
543621359,
543621378,
271810890,
135905761,
543621378,
271810909,
135905761,
135905761,
505205879,
505205898,
252699038,
222787236,
... | 10,657,548,707 | 0e77df5563c7789544499b2e5b34bb2182b38301 | [
"transformers.js",
"onnx",
"llava_qwen2",
"text-generation",
"fastvlm",
"image-text-to-text",
"conversational",
"license:apple-amlr",
"region:us"
] | null | null | null | [
"apple-amlr"
] | null | null | null | null | [
"text-generation",
"image-text-to-text"
] | null | [
"LlavaQwen2ForCausalLM",
"llava_qwen2",
"AutoModelForCausalLM"
] | [
"multimodal",
"text"
] | [
"text",
"image"
] | [
"text"
] | free | community | [
"Online"
] | null | null | null | null | null | null | null | null | null |
687060f05721fba56ca177a8 | moonshotai/Kimi-K2-Instruct | moonshotai | null | 400,648 | 765,924 | False | 2025-07-11T00:55:12Z | 2025-08-11T13:45:09Z | transformers | 2,117 | 19 | null | text-generation | null | [
".gitattributes",
"LICENSE",
"README.md",
"THIRD_PARTY_NOTICES.md",
"chat_template.jinja",
"config.json",
"configuration_deepseek.py",
"docs/deploy_guidance.md",
"docs/tool_call_guidance.md",
"figures/Base-Evaluation.png",
"figures/banner.png",
"figures/kimi-logo.png",
"generation_config.jso... | [
1695,
1463,
25472,
1664,
1882,
1725,
10652,
8903,
10280,
245449,
291736,
87988,
52,
87988,
2846451040,
17066593104,
17066595432,
17066595432,
17066595432,
17066595432,
17066595432,
17066595432,
17066595432,
17066595432,
17066595432,
17066593104,
17066595432,
170... | 1,029,207,174,310 | c52f808f632c07eb8361388616b1d04749373a94 | [
"transformers",
"safetensors",
"kimi_k2",
"text-generation",
"conversational",
"custom_code",
"doi:10.57967/hf/5976",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"fp8",
"region:us"
] | null | <div align="center">
<picture>
<img src="figures/kimi-logo.png" width="30%" alt="Kimi K2: Open Agentic Intellignece">
</picture>
</div>
<hr>
<div align="center" style="line-height:1">
<a href="https://www.kimi.com" target="_blank"><img alt="Chat" src="https://img.shields.io/badge/🤖%20Chat-Kimi%20K2-ff6b6... | [
"enzostvs/deepsite",
"umint/ai",
"ISEEKYAN/megatron_memory_estimator",
"nazdridoy/inferoxy-hub",
"umint/o4-mini",
"KrishnaVelama/Roberta-nemotron4-MentalHealth-Analyzer",
"akiko19191/Better_tool_calling",
"Quliyev/NeuroX",
"joseassuno1/penseai",
"raoufjat/tunisian-comedy-generator",
"blakeurmos/... | [
"other",
"modified-mit"
] | null | null | null | null | [
"text-generation"
] | null | [
"DeepseekV3ForCausalLM",
"kimi_k2",
"AutoModelForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | user | user | [
"user"
] | null | moonshotai/Kimi-K2-Base | [
"Text"
] | [
"Text Generation"
] | [
"Mixture-of-Experts",
" Transformer: Text Decoder-only"
] | [
"zh",
" en"
] | [
"Instruction finetuning"
] | Not disclosed | 3 |
689752deb3ed7f312e97b38c | DatarusAI/Datarus-R1-14B-preview | DatarusAI | {
"models": [
{
"_id": "66e6d08a5c06b7719cebd8ec",
"id": "Qwen/Qwen2.5-14B"
}
],
"relation": "finetune"
} | 3,866 | 3,866 | False | 2025-08-09T13:53:34Z | 2025-08-20T13:52:18Z | transformers | 124 | 19 | null | text-generation | {"parameters": {"BF16": 14770033664}, "total": 14770033664} | [
".gitattributes",
"README.md",
"config.json",
"generation_config.json",
"model-00001-of-00006.safetensors",
"model-00002-of-00006.safetensors",
"model-00003-of-00006.safetensors",
"model-00004-of-00006.safetensors",
"model-00005-of-00006.safetensors",
"model-00006-of-00006.safetensors",
"model.s... | [
1570,
5609,
713,
168,
4986211280,
4954847344,
4954847392,
4954847392,
4954847392,
4734533160,
47472,
485,
11425050,
8920,
8760
] | 29,551,632,707 | 60b4cb859cdcdc323702d4898f8916e207df6191 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"en",
"arxiv:2508.13382",
"base_model:Qwen/Qwen2.5-14B",
"base_model:finetune:Qwen/Qwen2.5-14B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | null | # Datarus-R1-14B-preview
<div align="center">
<img src="https://i.postimg.cc/7hsStNgm/logo-icon-2-1.png" alt="Datarus Logo" width="150"/>
[](https://huggingface.co/DatarusAI/Datarus-R1-14B-preview)
[
## Model Description
Hermes 4 70B is a frontier, hybrid-mode **reasoning** model based on Llama-3.1-70B by Nous Research that is aligned to **you**.
Read the Herm... | null | [
"llama3"
] | null | [
"en"
] | 70,560,423,936 | null | [
"text-generation"
] | null | [
"llama",
"AutoModelForCausalLM",
"LlamaForCausalLM"
] | [
"text"
] | [
"text"
] | [
"text"
] | team | company | [
"Online"
] | null | null | null | null | null | null | null | null | null |
68a82b499a90732a36195279 | kurakurai/Luth-LFM2-1.2B | kurakurai | {
"models": [
{
"_id": "686fabae2fb74b6dbfe2dc8b",
"id": "LiquidAI/LFM2-1.2B"
}
],
"relation": "finetune"
} | 301 | 301 | False | 2025-08-22T08:33:13Z | 2025-08-25T17:36:02Z | transformers | 20 | 19 | null | text-generation | {"parameters": {"BF16": 1170340608}, "total": 1170340608} | [
".gitattributes",
"LICENSE",
"README.md",
"chat_template.jinja",
"config.json",
"generation_config.json",
"lfm2-luth.png",
"media/lfm2-luth.png",
"media/logo_collab.png",
"model.safetensors",
"special_tokens_map.json",
"tokenizer.json",
"tokenizer_config.json"
] | null | null | 5b7edc4bc70afbcb8a37ccf984c32e4259cf4339 | [
"transformers",
"safetensors",
"lfm2",
"text-generation",
"liquid",
"luth",
"conversational",
"fr",
"en",
"dataset:kurakurai/luth-sft",
"base_model:LiquidAI/LFM2-1.2B",
"base_model:finetune:LiquidAI/LFM2-1.2B",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"region:us... | null | null | [
"MisterAI/Try_Small_Models02"
] | [
"other",
"lfm1.0",
"LICENSE"
] | [
"kurakurai/luth-sft"
] | [
"fr",
"en"
] | 1,170,340,608 | null | [
"text-generation"
] | null | [
"AutoModelForCausalLM",
"Lfm2ForCausalLM",
"lfm2"
] | [
"text"
] | [
"text"
] | [
"text"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
64bfcd5ff462a99a04fd1ec8 | stabilityai/stable-diffusion-xl-base-1.0 | stabilityai | null | 2,156,015 | 96,112,479 | False | 2023-07-25T13:25:51Z | 2023-10-30T16:03:47Z | diffusers | 6,882 | 18 | null | text-to-image | null | [
".gitattributes",
"01.png",
"LICENSE.md",
"README.md",
"comparison.png",
"model_index.json",
"pipeline.png",
"scheduler/scheduler_config.json",
"sd_xl_base_1.0.safetensors",
"sd_xl_base_1.0_0.9vae.safetensors",
"sd_xl_offset_example-lora_1.0.safetensors",
"text_encoder/config.json",
"text_en... | [
1562,
4608613,
14109,
8668,
130252,
609,
80188,
479,
6938078334,
6938078334,
49553604,
565,
492248682,
246144152,
492587457,
492265168,
492242672,
1057789,
575,
2778657095,
1389382176,
1041992,
2778639360,
2778702264,
2778640120,
2790191,
524619,
472,
737,
1... | 76,912,765,291 | 462165984030d82259a11f4367a4eed129e94a7b | [
"diffusers",
"onnx",
"safetensors",
"text-to-image",
"stable-diffusion",
"arxiv:2307.01952",
"arxiv:2211.01324",
"arxiv:2108.01073",
"arxiv:2112.10752",
"license:openrail++",
"autotrain_compatible",
"endpoints_compatible",
"diffusers:StableDiffusionXLPipeline",
"region:us"
] | null | # SD-XL 1.0-base Model Card

## Model

[SDXL](https://arxiv.org/abs/2307.01952) consists of an [ensemble of experts](https://arxiv.org/abs/2211.01324) pipeline for latent diffusion:
In a first step, the base model is used to generate (noisy) latents,
which are then further ... | [
"jallenjia/Change-Clothes-AI",
"google/sdxl",
"yisol/IDM-VTON",
"optimum/neuron-export",
"fffiloni/InstantIR",
"Nymbo/Serverless-ImgGen-Hub",
"frogleo/Image-to-3D",
"jbilcke-hf/OmniAvatar",
"fffiloni/StyleAligned_Transfer",
"bghira/Glyph-SDXL-v2",
"yanze/PuLID-FLUX",
"bobber/DiffuseCraft",
"... | [
"openrail++"
] | null | null | null | null | [
"text-to-image"
] | null | null | [
"vision"
] | [
"text"
] | [
"image"
] | user | user | [
"user"
] | null | null | [
"Text",
" Image"
] | [
"Image Generation"
] | [
"Diffusion-based Network"
] | [
"en"
] | [
"Pretraining: Denoising Autoencoder",
" Pretraining: Contrastive Learning",
" Pretraining: Multimodal joint-embeddings"
] | Not disclosed | 6 |
65b53851e602b6c2c96e78da | BAAI/bge-m3 | BAAI | null | 5,717,245 | 47,041,490 | False | 2024-01-27T17:07:29Z | 2024-07-03T14:50:10Z | sentence-transformers | 2,311 | 18 | null | sentence-similarity | null | [
".gitattributes",
"1_Pooling/config.json",
"README.md",
"colbert_linear.pt",
"config.json",
"config_sentence_transformers.json",
"imgs/.DS_Store",
"imgs/bm25.jpg",
"imgs/long.jpg",
"imgs/miracl.jpg",
"imgs/mkqa.jpg",
"imgs/nqa.jpg",
"imgs/others.webp",
"long.jpg",
"modules.json",
"onnx... | [
1627,
191,
15822,
2100674,
687,
123,
6148,
131849,
485432,
576482,
608027,
158358,
20984,
126894,
349,
65552,
698,
724923,
2266820608,
5069051,
964,
17082821,
1173,
2271145830,
54,
5069051,
3516,
964,
17098108,
444
] | 4,587,317,404 | 5617a9f61b028005a4858fdac845db406aefb181 | [
"sentence-transformers",
"pytorch",
"onnx",
"xlm-roberta",
"feature-extraction",
"sentence-similarity",
"arxiv:2402.03216",
"arxiv:2004.04906",
"arxiv:2106.14807",
"arxiv:2107.05720",
"arxiv:2004.12832",
"license:mit",
"autotrain_compatible",
"text-embeddings-inference",
"endpoints_compa... | null | For more details please refer to our github repo: https://github.com/FlagOpen/FlagEmbedding
# BGE-M3 ([paper](https://arxiv.org/pdf/2402.03216.pdf), [code](https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/BGE_M3))
In this project, we introduce BGE-M3, which is distinguished for its versatility in M... | [
"mteb/leaderboard",
"GIZ/audit_assistant",
"ChatVLD/CHATVLD",
"istat-ai/auto-ateco",
"saqib7/BAAI-bge-m3",
"Clarymind/BAAI-bge-m3",
"lucas-wa/rag-chat",
"SujonPro24/sentence_similairty",
"mikeee/baai-m3",
"ShivanshMathur007/MoP",
"TheDrakosfire/RuleLawyer",
"ahmedkasem/quran-nlp",
"danieldux... | [
"mit"
] | null | null | null | null | [
"sentence-similarity",
"feature-extraction"
] | null | [
"xlm-roberta",
"XLMRobertaModel"
] | [
"multimodal",
"text"
] | [
"text"
] | [
"embeddings",
"logits"
] | team | non-profit | [
"China"
] | null | BAAI/bge-m3-unsupervised | [
"Text"
] | [
"Text Embedding"
] | [
"Transformer: Text Encoder-only"
] | [
"Multilingual"
] | [
"Pretraining: Contrastive Learning",
" Instruction finetuning",
" Knowledge distillation"
] | Disclosed: available | 10 |
6881cd19c508ec44951b7620 | internlm/Intern-S1 | internlm | null | 64,081 | 71,506 | False | 2025-07-24T06:05:13Z | 2025-08-29T02:56:15Z | transformers | 235 | 18 | null | image-text-to-text | {"parameters": {"BF16": 240709856128}, "total": 240709856128} | [
".gitattributes",
"LICENSE.txt",
"README.md",
"chat_template.jinja",
"config.json",
"configuration_interns1.py",
"generation_config.json",
"merges.txt",
"model-00001-of-00097.safetensors",
"model-00002-of-00097.safetensors",
"model-00003-of-00097.safetensors",
"model-00004-of-00097.safetensors... | [
1519,
11357,
20741,
4561,
2621,
10210,
126,
1671853,
4988565184,
4937275136,
4999804368,
4988147640,
4988147640,
4988147640,
4988147640,
4988147640,
4988147640,
4988147640,
4988147640,
4988147640,
4988147848,
4988148032,
4988148032,
4988148032,
4988148032,
498814803... | 481,434,170,855 | 0ac91986960cf9631098a00d92b47e2319f1eed9 | [
"transformers",
"safetensors",
"interns1",
"text-generation",
"image-text-to-text",
"conversational",
"custom_code",
"arxiv:2508.15763",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | null | ## Intern-S1
<div align="center">
<img src="https://cdn-uploads.huggingface.co/production/uploads/642695e5274e7ad464c8a5ba/E43cgEXBRWjVJlU_-hdh6.png" />
<div> </div>
[💻Github Repo](https://github.com/InternLM/Intern-S1) • [🤗Model Collections](https://huggingface.co/collections/internlm/intern-s1-6882e325e8a... | null | [
"apache-2.0"
] | null | null | 240,709,856,128 | null | [
"text-generation",
"image-text-to-text"
] | null | [
"InternS1ForConditionalGeneration",
"modeling_interns1.InternS1ForConditionalGeneration",
"AutoModelForCausalLM",
"interns1"
] | [
"multimodal",
"text"
] | [
"text",
"image"
] | [
"text"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
6885ada1ede5d03681df23f6 | TheDrummer/RimTalk-Mini-v1-GGUF | TheDrummer | {
"models": [
{
"_id": "66eaef786865fea1324edb5d",
"id": "meta-llama/Llama-3.2-3B-Instruct"
}
],
"relation": "quantized"
} | 1,536 | 1,550 | False | 2025-07-27T04:40:01Z | 2025-08-29T06:11:41Z | null | 18 | 18 | null | null | null | [
".gitattributes",
"README.md",
"RimDialogue-3B-v1a-Q2_K.gguf",
"RimDialogue-3B-v1a-Q3_K_M.gguf",
"RimDialogue-3B-v1a-Q4_K_M.gguf",
"RimDialogue-3B-v1a-Q5_K_M.gguf",
"RimDialogue-3B-v1a-Q6_K.gguf",
"RimDialogue-3B-v1a-Q8_0.gguf"
] | null | null | fbaecf4be445e7717d9960d37637deaac72b449f | [
"gguf",
"base_model:meta-llama/Llama-3.2-3B-Instruct",
"base_model:quantized:meta-llama/Llama-3.2-3B-Instruct",
"endpoints_compatible",
"region:us",
"conversational"
] | {"total": 3212749888, "architecture": "llama", "context_length": 131072, "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_st... | # RimTalk Mini v1

Mod Link: https://steamcommunity.com/sharedfiles/filedetails/?id=3365889763
---
 • [🤗Model Collections](https://huggingface.co/collections/internlm/intern-s1-6882e3... | null | [
"apache-2.0"
] | null | null | 8,538,804,224 | null | [
"text-generation",
"image-text-to-text"
] | null | [
"InternS1ForConditionalGeneration",
"modeling_interns1.InternS1ForConditionalGeneration",
"AutoModelForCausalLM",
"interns1"
] | [
"multimodal",
"text"
] | [
"text",
"image"
] | [
"text"
] | user | user | [
"user"
] | null | null | null | null | null | null | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.