| { | |
| "EleutherAI/pythia-160m-deduped": { | |
| "model_name": "EleutherAI/pythia-160m-deduped", | |
| "total_params": 123689472, | |
| "embedding_params": 38633472, | |
| "non_embedding_params": 85056000, | |
| "model_size": 0.16, | |
| "model_label": "Pythia", | |
| "vocab_size": 50304 | |
| }, | |
| "EleutherAI/pythia-410m-deduped": { | |
| "model_name": "EleutherAI/pythia-410m-deduped", | |
| "total_params": 353822720, | |
| "embedding_params": 51511296, | |
| "non_embedding_params": 302311424, | |
| "model_size": 0.41, | |
| "model_label": "Pythia", | |
| "vocab_size": 50304 | |
| }, | |
| "facebook/opt-125m": { | |
| "model_name": "facebook/opt-125m", | |
| "total_params": 125239296, | |
| "embedding_params": 40183296, | |
| "non_embedding_params": 85056000, | |
| "model_size": 0.125, | |
| "model_label": "OPT", | |
| "vocab_size": 50272 | |
| }, | |
| "facebook/opt-350m": { | |
| "model_name": "facebook/opt-350m", | |
| "total_params": 331196416, | |
| "embedding_params": 27838464, | |
| "non_embedding_params": 303357952, | |
| "model_size": 0.35, | |
| "model_label": "OPT", | |
| "vocab_size": 50272 | |
| }, | |
| "facebook/opt-1.3b": { | |
| "model_name": "facebook/opt-1.3b", | |
| "total_params": 1315758080, | |
| "embedding_params": 107155456, | |
| "non_embedding_params": 1208602624, | |
| "model_size": 1.3, | |
| "model_label": "OPT", | |
| "vocab_size": 50272 | |
| }, | |
| "facebook/opt-2.7b": { | |
| "model_name": "facebook/opt-2.7b", | |
| "total_params": 2651596800, | |
| "embedding_params": 133944320, | |
| "non_embedding_params": 2517652480, | |
| "model_size": 2.7, | |
| "model_label": "OPT", | |
| "vocab_size": 50272 | |
| }, | |
| "facebook/opt-6.7b": { | |
| "model_name": "facebook/opt-6.7b", | |
| "total_params": 6658473984, | |
| "embedding_params": 214310912, | |
| "non_embedding_params": 6444163072, | |
| "model_size": 6.7, | |
| "model_label": "OPT", | |
| "vocab_size": 50272 | |
| }, | |
| "EleutherAI/gpt-neo-125m": { | |
| "model_name": "EleutherAI/gpt-neo-125m", | |
| "total_params": 125198592, | |
| "embedding_params": 40170240, | |
| "non_embedding_params": 85028352, | |
| "model_size": 0.125, | |
| "model_label": "GPT-Neo", | |
| "vocab_size": 50257 | |
| }, | |
| "distilbert/distilgpt2": { | |
| "model_name": "distilbert/distilgpt2", | |
| "total_params": 81912576, | |
| "embedding_params": 39383808, | |
| "non_embedding_params": 42528768, | |
| "model_size": 0.082, | |
| "model_label": "DistilGPT2", | |
| "vocab_size": 50257 | |
| }, | |
| "Qwen/Qwen2.5-0.5B": { | |
| "model_name": "Qwen/Qwen2.5-0.5B", | |
| "total_params": 494032768, | |
| "embedding_params": 136134656, | |
| "non_embedding_params": 357898112, | |
| "model_size": 0.5, | |
| "model_label": "QWen2.5", | |
| "vocab_size": 151936 | |
| }, | |
| "Qwen/Qwen2.5-1.5B": { | |
| "model_name": "Qwen/Qwen2.5-1.5B", | |
| "total_params": 1543714304, | |
| "embedding_params": 233373696, | |
| "non_embedding_params": 1310340608, | |
| "model_size": 1.5, | |
| "model_label": "QWen2.5", | |
| "vocab_size": 151936 | |
| }, | |
| "Qwen/Qwen2.5-3B": { | |
| "model_name": "Qwen/Qwen2.5-3B", | |
| "total_params": 3085938688, | |
| "embedding_params": 311164928, | |
| "non_embedding_params": 2774773760, | |
| "model_size": 3, | |
| "model_label": "QWen2.5", | |
| "vocab_size": 151936 | |
| }, | |
| "Qwen/Qwen2.5-7B": { | |
| "model_name": "Qwen/Qwen2.5-7B", | |
| "total_params": 7070619136, | |
| "embedding_params": 544997376, | |
| "non_embedding_params": 6525621760, | |
| "model_size": 7, | |
| "model_label": "QWen2.5", | |
| "vocab_size": 152064 | |
| }, | |
| "Qwen/Qwen2.5-14B": { | |
| "model_name": "Qwen/Qwen2.5-14B", | |
| "total_params": 13991465984, | |
| "embedding_params": 778567680, | |
| "non_embedding_params": 13212898304, | |
| "model_size": 14, | |
| "model_label": "QWen2.5", | |
| "vocab_size": 152064 | |
| }, | |
| "Qwen/Qwen2.5-32B": { | |
| "model_name": "Qwen/Qwen2.5-32B", | |
| "total_params": 31985308672, | |
| "embedding_params": 778567680, | |
| "non_embedding_params": 31206740992, | |
| "model_size": 32, | |
| "model_label": "QWen2.5", | |
| "vocab_size": 152064 | |
| }, | |
| "Qwen/Qwen2.5-72B": { | |
| "model_name": "Qwen/Qwen2.5-72B", | |
| "total_params": 71460495360, | |
| "embedding_params": 1245708288, | |
| "non_embedding_params": 70214787072, | |
| "model_size": 72, | |
| "model_label": "QWen2.5", | |
| "vocab_size": 152064 | |
| }, | |
| "Qwen/Qwen3-0.6B":{ | |
| "model_name": "Qwen/Qwen3-0.6B", | |
| "total_params": 596049920, | |
| "embedding_params": 155582464, | |
| "non_embedding_params": 440467456, | |
| "model_size": 0.6, | |
| "model_label": "QWen3", | |
| "vocab_size": 151936 | |
| }, | |
| "Qwen/Qwen3-1.7B":{ | |
| "model_name": "Qwen/Qwen3-1.7B", | |
| "total_params": 1720574976, | |
| "embedding_params": 311164928, | |
| "non_embedding_params": 1409410048, | |
| "model_size": 1.7, | |
| "model_label": "QWen3", | |
| "vocab_size": 151936 | |
| }, | |
| "Qwen/Qwen3-4B":{ | |
| "model_name": "Qwen/Qwen3-4B", | |
| "total_params": 4022468096, | |
| "embedding_params": 388956160, | |
| "non_embedding_params": 3633511936, | |
| "model_size": 4, | |
| "model_label": "QWen3", | |
| "vocab_size": 151936 | |
| }, | |
| "Qwen/Qwen3-8B":{ | |
| "model_name": "Qwen/Qwen3-8B", | |
| "total_params": 7568405504, | |
| "embedding_params": 622329856, | |
| "non_embedding_params": 6946075648, | |
| "model_size": 8, | |
| "model_label": "QWen3", | |
| "vocab_size": 151936 | |
| }, | |
| "Qwen/Qwen3-14B":{ | |
| "model_name": "Qwen/Qwen3-14B", | |
| "total_params": 13990394880, | |
| "embedding_params": 777912320, | |
| "non_embedding_params": 13212482560, | |
| "model_size": 14, | |
| "model_label": "QWen3", | |
| "vocab_size": 151936 | |
| }, | |
| "Qwen/Qwen3-32B": { | |
| "model_name": "Qwen/Qwen3-32B", | |
| "total_params": 31984210944, | |
| "embedding_params": 777912320, | |
| "non_embedding_params": 31206298624, | |
| "model_size": 32, | |
| "model_label": "Qwen3", | |
| "vocab_size": 151 | |
| }, | |
| "Qwen/Qwen3-0.6B-Base": { | |
| "model_name": "Qwen/Qwen3-0.6B-Base", | |
| "total_params": 596049920, | |
| "embedding_params": 155582464, | |
| "non_embedding_params": 440467456, | |
| "model_size": 32, | |
| "model_label": "Qwen3", | |
| "vocab_size": 151936 | |
| }, | |
| "Qwen/Qwen3-1.7B-Base": { | |
| "model_name": "Qwen/Qwen3-1.7B-Base", | |
| "total_params": 1720574976, | |
| "embedding_params": 311164928, | |
| "non_embedding_params": 1409410048, | |
| "model_size": 32, | |
| "model_label": "Qwen3", | |
| "vocab_size": 151936 | |
| }, | |
| "Qwen/Qwen3-4B-Base": { | |
| "model_name": "Qwen/Qwen3-4B-Base", | |
| "total_params": 4022468096, | |
| "embedding_params": 388956160, | |
| "non_embedding_params": 3633511936, | |
| "model_size": 32, | |
| "model_label": "Qwen3", | |
| "vocab_size": 151936 | |
| }, | |
| "Qwen/Qwen3-8B-Base": { | |
| "model_name": "Qwen/Qwen3-8B-Base", | |
| "total_params": 7568405504, | |
| "embedding_params": 622329856, | |
| "non_embedding_params": 6946075648, | |
| "model_size": 32, | |
| "model_label": "Qwen3", | |
| "vocab_size": 151936 | |
| }, | |
| "Qwen/Qwen3-14B-Base": { | |
| "model_name": "Qwen/Qwen3-14B-Base", | |
| "total_params": 13990394880, | |
| "embedding_params": 777912320, | |
| "non_embedding_params": 13212482560, | |
| "model_size": 32, | |
| "model_label": "Qwen3", | |
| "vocab_size": 151936 | |
| }, | |
| "meta-llama/Meta-Llama-3-8B": { | |
| "model_name": "meta-llama/Meta-Llama-3-8B", | |
| "total_params": 7504924672, | |
| "embedding_params": 525336576, | |
| "non_embedding_params": 6979588096, | |
| "model_size": 8, | |
| "model_label": "Llama3", | |
| "vocab_size": 128256 | |
| }, | |
| "meta-llama/Meta-Llama-3-70B": { | |
| "model_name": "meta-llama/Meta-Llama-3-70B", | |
| "total_params": 69503033344, | |
| "embedding_params": 1050673152, | |
| "non_embedding_params": 68452360192, | |
| "model_size": 70, | |
| "model_label": "Llama3", | |
| "vocab_size": 128256 | |
| }, | |
| "meta-llama/Llama-2-13b-hf": { | |
| "model_name": "meta-llama/Llama-2-13b-hf", | |
| "total_params": 12852024320, | |
| "embedding_params": 163840000, | |
| "non_embedding_params": 12688184320, | |
| "model_size": 13, | |
| "model_label": "Llama2", | |
| "vocab_size": 32000 | |
| }, | |
| "meta-llama/Llama-2-7b-hf": { | |
| "model_name": "meta-llama/Llama-2-7b-hf", | |
| "total_params": 6607343616, | |
| "embedding_params": 131072000, | |
| "non_embedding_params": 6476271616, | |
| "model_size": 7, | |
| "model_label": "Llama2", | |
| "vocab_size": 32000 | |
| }, | |
| "meta-llama/Llama-2-70b-hf": { | |
| "model_name": "meta-llama/Llama-2-70b-hf", | |
| "total_params": 68714504192, | |
| "embedding_params": 262144000, | |
| "non_embedding_params": 68452360192, | |
| "model_size": 70, | |
| "model_label": "Llama2", | |
| "vocab_size": 32000 | |
| }, | |
| "openai-community/gpt2": { | |
| "model_name": "openai-community/gpt2", | |
| "total_params": 124439808, | |
| "embedding_params": 39383808, | |
| "non_embedding_params": 85056000, | |
| "model_size": 0.124, | |
| "model_label": "GPT2", | |
| "vocab_size": 50257 | |
| }, | |
| "openai-community/gpt2-medium": { | |
| "model_name": "openai-community/gpt2-medium", | |
| "total_params": 354823168, | |
| "embedding_params": 52511744, | |
| "non_embedding_params": 302311424, | |
| "model_size": 0.355, | |
| "model_label": "GPT2", | |
| "vocab_size": 50257 | |
| }, | |
| "openai-community/gpt2-large": { | |
| "model_name": "openai-community/gpt2-large", | |
| "total_params": 774030080, | |
| "embedding_params": 65639680, | |
| "non_embedding_params": 708390400, | |
| "model_size": 0.774, | |
| "model_label": "GPT2", | |
| "vocab_size": 50257 | |
| }, | |
| "openai-community/gpt2-xl": { | |
| "model_name": "openai-community/gpt2-xl", | |
| "total_params": 1557611200, | |
| "embedding_params": 82049600, | |
| "non_embedding_params": 1475561600, | |
| "model_size": 1.5, | |
| "model_label": "GPT2", | |
| "vocab_size": 50257 | |
| }, | |
| "huggyllama/llama-7b": { | |
| "model_name": "huggyllama/llama-7b", | |
| "total_params": 6607343616, | |
| "embedding_params": 131072000, | |
| "non_embedding_params": 6476271616, | |
| "model_size": 7, | |
| "model_label": "Llama1", | |
| "vocab_size": 32000 | |
| }, | |
| "huggyllama/llama-65b": { | |
| "model_name": "huggyllama/llama-65b", | |
| "total_params": 65023516672, | |
| "embedding_params": 262144000, | |
| "non_embedding_params": 64761372672, | |
| "model_size": 65, | |
| "model_label": "Llama1", | |
| "vocab_size": 32000 | |
| }, | |
| "microsoft/phi-2": { | |
| "model_name": "microsoft/phi-2", | |
| "total_params": 2648560640, | |
| "embedding_params": 131072000, | |
| "non_embedding_params": 2517488640, | |
| "model_size": 2, | |
| "model_label": "Phi-2", | |
| "vocab_size": 51200 | |
| }, | |
| "meta-llama/Llama-3.2-1B": { | |
| "model_name": "meta-llama/Llama-3.2-1B", | |
| "total_params": 1235814400, | |
| "embedding_params": 262668288, | |
| "non_embedding_params": 973146112, | |
| "model_size": 1, | |
| "model_label": "Llama3.2", | |
| "vocab_size": 128256 | |
| }, | |
| "meta-llama/Llama-3.2-3B": { | |
| "model_name": "meta-llama/Llama-3.2-3B", | |
| "total_params": 3212749824, | |
| "embedding_params": 394002432, | |
| "non_embedding_params": 2818747392, | |
| "model_size": 3, | |
| "model_label": "Llama3.2", | |
| "vocab_size": 128256 | |
| }, | |
| "mistralai/Mistral-7B-v0.1": { | |
| "model_name": "mistralai/Mistral-7B-v0.1", | |
| "total_params": 7110660096, | |
| "embedding_params": 131072000, | |
| "non_embedding_params": 6979588096, | |
| "model_size": 7, | |
| "model_label": "Mistral", | |
| "vocab_size": 32000 | |
| }, | |
| "baichuan-inc/Baichuan-M1-14B-Base": { | |
| "model_name": "baichuan-inc/Baichuan-M1-14B-Base", | |
| "total_params": 13789189920, | |
| "embedding_params": 681574400, | |
| "non_embedding_params": 13107615520, | |
| "model_size": 14, | |
| "model_label": "BaichuanM1", | |
| "vocab_size": 133120 | |
| }, | |
| "THUDM/glm-4-9b": { | |
| "model_name": "THUDM/glm-4-9b", | |
| "total_params": 9399951360, | |
| "embedding_params": 620756992, | |
| "non_embedding_params": 8779194368, | |
| "model_size": 9, | |
| "model_label": "GLM4", | |
| "vocab_size": 151552 | |
| }, | |
| "deepseek-ai/DeepSeek-V2-Lite": { | |
| "model_name": "deepseek-ai/DeepSeek-V2-Lite", | |
| "total_params": 15496769024, | |
| "embedding_params": 209715200, | |
| "non_embedding_params": 15287053824, | |
| "model_size": 16, | |
| "model_label": "DeepSeek-V2", | |
| "vocab_size": 102400 | |
| }, | |
| "roneneldan/TinyStories-1M": { | |
| "model_name": "roneneldan/TinyStories-1M", | |
| "total_params": 3745984, | |
| "embedding_params": 3347520, | |
| "non_embedding_params": 398464, | |
| "model_size": 0.001, | |
| "model_label": "TinyStories", | |
| "vocab_size": 50257 | |
| }, | |
| "roneneldan/TinyStories-3M": { | |
| "model_name": "roneneldan/TinyStories-3M", | |
| "total_params": 8278400, | |
| "embedding_params": 6695040, | |
| "non_embedding_params": 1583360, | |
| "model_size": 0.003, | |
| "model_label": "TinyStories", | |
| "vocab_size": 50257 | |
| }, | |
| "roneneldan/TinyStories-8M": { | |
| "model_name": "roneneldan/TinyStories-8M", | |
| "total_params": 19702528, | |
| "embedding_params": 13390080, | |
| "non_embedding_params": 6312448, | |
| "model_size": 0.008, | |
| "model_label": "TinyStories", | |
| "vocab_size": 50257 | |
| }, | |
| "EleutherAI/pythia-14m": { | |
| "model_name": "EleutherAI/pythia-14m", | |
| "total_params": 7628800, | |
| "embedding_params": 6438912, | |
| "non_embedding_params": 1189888, | |
| "model_size": 0.014, | |
| "model_label": "Pythia", | |
| "vocab_size": 50304 | |
| }, | |
| "EleutherAI/pythia-70m": { | |
| "model_name": "EleutherAI/pythia-70m", | |
| "total_params": 44670976, | |
| "embedding_params": 25755648, | |
| "non_embedding_params": 18915328, | |
| "model_size": 0.07, | |
| "model_label": "Pythia", | |
| "vocab_size": 50304 | |
| }, | |
| "EleutherAI/pythia-160m": { | |
| "model_name": "EleutherAI/pythia-160m", | |
| "total_params": 123689472, | |
| "embedding_params": 38633472, | |
| "non_embedding_params": 85056000, | |
| "model_size": 0.16, | |
| "model_label": "Pythia", | |
| "vocab_size": 50304 | |
| }, | |
| "EleutherAI/pythia-410m": { | |
| "model_name": "EleutherAI/pythia-410m", | |
| "total_params": 353822720, | |
| "embedding_params": 51511296, | |
| "non_embedding_params": 302311424, | |
| "model_size": 0.41, | |
| "model_label": "Pythia", | |
| "vocab_size": 50304 | |
| }, | |
| "EleutherAI/pythia-31m": { | |
| "model_name": "EleutherAI/pythia-31m", | |
| "total_params": 17616896, | |
| "embedding_params": 12877824, | |
| "non_embedding_params": 4739072, | |
| "model_size": "31m", | |
| "model_label": "Pythia", | |
| "vocab_size": 50304 | |
| }, | |
| "EleutherAI/pythia-1b": { | |
| "model_name": "EleutherAI/pythia-1b", | |
| "total_params": 908759040, | |
| "embedding_params": 103022592, | |
| "non_embedding_params": 805736448, | |
| "model_size": "1b", | |
| "model_label": "Pythia", | |
| "vocab_size": 50304 | |
| }, | |
| "EleutherAI/pythia-1.4b": { | |
| "model_name": "EleutherAI/pythia-1.4b", | |
| "total_params": 1311625216, | |
| "embedding_params": 103022592, | |
| "non_embedding_params": 1208602624, | |
| "model_size": "1.4b", | |
| "model_label": "Pythia", | |
| "vocab_size": 50304 | |
| }, | |
| "EleutherAI/pythia-2.8b": { | |
| "model_name": "EleutherAI/pythia-2.8b", | |
| "total_params": 2646430720, | |
| "embedding_params": 128778240, | |
| "non_embedding_params": 2517652480, | |
| "model_size": "2.8b", | |
| "model_label": "Pythia", | |
| "vocab_size": 50304 | |
| }, | |
| "EleutherAI/pythia-6.9b": { | |
| "model_name": "EleutherAI/pythia-6.9b", | |
| "total_params": 6650732544, | |
| "embedding_params": 206569472, | |
| "non_embedding_params": 6444163072, | |
| "model_size": "6.9b", | |
| "model_label": "Pythia", | |
| "vocab_size": 50432 | |
| }, | |
| "EleutherAI/pythia-12b": { | |
| "model_name": "EleutherAI/pythia-12b", | |
| "total_params": 11586549760, | |
| "embedding_params": 259522560, | |
| "non_embedding_params": 11327027200, | |
| "model_size": "12b", | |
| "model_label": "Pythia", | |
| "vocab_size": 50688 | |
| }, | |
| "openai-community/openai-gpt": { | |
| "model_name": "openai-community/openai-gpt", | |
| "total_params": 116534784, | |
| "embedding_params": 31480320, | |
| "non_embedding_params": 85054464, | |
| "model_size": 0.12, | |
| "model_label": "GPT1", | |
| "vocab_size": 40478 | |
| }, | |
| "DeepSeek-V3-Base": { | |
| "model_name": "DeepSeek-V3-Base", | |
| "total_params": 671000000000, | |
| "embedding_params": 900000000, | |
| "non_embedding_params": 669200000000, | |
| "model_size": 671, | |
| "model_label": "DeepSeek-V3-Base", | |
| "vocab_size": 129280 | |
| }, | |
| "deepseek-ai/DeepSeek-V3-Base": { | |
| "model_name": "DeepSeek-V3-Base", | |
| "total_params": 671000000000, | |
| "embedding_params": 900000000, | |
| "non_embedding_params": 669200000000, | |
| "model_size": 671, | |
| "model_label": "DeepSeek-V3-Base", | |
| "vocab_size": 129280 | |
| }, | |
| "meta-llama/Llama-3.1-8B": { | |
| "total_params": 7504924672, | |
| "embedding_params": 525336576, | |
| "non_embedding_params": 6979588096, | |
| "vocab_size": 128256, | |
| "model_size": 8, | |
| "model_label": "Llama-3.1" | |
| }, | |
| "meta-llama/Llama-3.1-70B": { | |
| "model_name": "meta-llama/Llama-3.1-70B", | |
| "total_params": 69503033344, | |
| "embedding_params": 1050673152, | |
| "non_embedding_params": 68452360192, | |
| "model_size": 70, | |
| "model_label": "Llama3.1", | |
| "vocab_size": 128256 | |
| }, | |
| "meta-llama/Llama-3.1-405B": { | |
| "total_params": 403752042496, | |
| "embedding_params": 2101346304, | |
| "non_embedding_params": 401650696192, | |
| "vocab_size": 128256, | |
| "model_size": 405, | |
| "model_label": "Llama-3.1" | |
| }, | |
| "meta-llama/Llama-4-Maverick-17B-128E": { | |
| "total_params": 399677363200, | |
| "embedding_params": 1034485760, | |
| "non_embedding_params": 398642877440, | |
| "vocab_size": 202048 | |
| }, | |
| "meta-llama/Llama-4-Scout-17B-16E": { | |
| "total_params": 106735375360, | |
| "embedding_params": 1034485760, | |
| "non_embedding_params": 105700889600, | |
| "vocab_size": 202048 | |
| } | |
| } |