index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,800 | AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-50-percent-med-high-bt-rouge-1 | mistral | ["MistralForCausalLM"] | 32064 | bfloat16 | 4.41.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
1,801 | AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-75-percent-high-bt-rouge-1 | mistral | ["MistralForCausalLM"] | 32064 | bfloat16 | 4.41.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
1,802 | AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-75-percent-low-bt-rouge-1 | mistral | ["MistralForCausalLM"] | 32064 | bfloat16 | 4.41.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
1,803 | AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-75-percent-low-high-bt-rouge-1 | mistral | ["MistralForCausalLM"] | 32064 | bfloat16 | 4.41.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
1,804 | AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-75-percent-low-med-bt-rouge-1 | mistral | ["MistralForCausalLM"] | 32064 | bfloat16 | 4.41.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
1,805 | AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-75-percent-low-med-high-bt-rouge-1 | mistral | ["MistralForCausalLM"] | 32064 | bfloat16 | 4.41.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
1,806 | AdamKasumovic/phi3-mini-4k-instruct-bactrian-x-xh-75-percent-med-bt-rouge-1 | mistral | ["MistralForCausalLM"] | 32064 | bfloat16 | 4.41.2 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
1,807 | AdamKasumovic/phi3-mini-4k-instruct-mmlu-college-medicine-af-mmlu-mmlu-low | mistral | ["MistralForCausalLM"] | 32064 | bfloat16 | 4.42.3 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | true | false | false | true | 3.623879 | medium | medium |
1,808 | AdamLucek/Orpo-Llama-3.2-1B-40k | llama | ["LlamaForCausalLM"] | 128258 | float16 | 4.45.2 | 2,048 | 8,192 | 16 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128256 | 128257 | true | 64 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 0.805306 | small | very_long |
1,809 | AdamLucek/Phi-3-mini-EmoMarketing-DELLA | phi3 | ["Phi3ForCausalLM"] | 32011 | float16 | 4.44.0 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 131,072 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | very_long |
1,810 | AdamLuwiko/gpt2-chatbot-voucheran | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.48.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,811 | AdamLuwiko/my-chatbot-model | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.48.3 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,812 | AdamRTomkins/phi-kal | phi | ["PhiForCausalLM"] | 51200 | float16 | 4.39.0.dev0 | 2,048 | 8,192 | 24 | 32 | 32 | gelu_new | 0 | False | 2,048 | 10,000 | null | 0.02 | null | null | false | 64 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 1.20796 | medium | medium |
1,813 | Adammz/cs_roberta_base-1 | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,814 | Adammz/ruBert-base-1-4 | bert | ["BertForSequenceClassification"] | 120138 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,815 | Adammz/ruBert-base-1-third | bert | ["BertForSequenceClassification"] | 120138 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,816 | Adammz/rubert-base-cased-1-third | bert | ["BertForSequenceClassification"] | 119547 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,817 | Adammz/rubert-tiny2-1-4 | bert | ["BertForSequenceClassification"] | 83828 | float32 | 4.35.2 | 312 | 600 | 3 | 12 | null | gelu | null | True | 2,048 | null | null | 0.02 | null | null | null | 26 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.003504 | small | medium |
1,818 | AdanLee/distilhubert-finetuned-gtzan | hubert | ["HubertForSequenceClassification"] | 32 | float32 | 4.32.0 | 768 | 3,072 | 2 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.014156 | small | null |
1,819 | AdanLee/whisper-tiny-minds14-en | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.32.0 | null | null | 4 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
1,820 | Adansonia/internal_audit_new | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.46.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128040 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
1,821 | AdaptLLM/biomed-InternVL3-1B | internvl_chat | ["InternVLChatModel"] | null | bfloat16 | null | 896 | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
1,822 | AdaptLLM/biomed-Qwen2.5-VL-3B-Instruct | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 151936 | bfloat16 | 4.51.3 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | False | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
1,823 | AdaptLLM/biomed-gemma-3-4b-it | gemma3 | ["Gemma3ForConditionalGeneration"] | null | bfloat16 | 4.51.3 | 2,560 | null | null | null | null | null | null | False | null | null | null | 0.02 | null | [1, 106] | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
1,824 | AdaptLLM/finance-LLM | llama | ["LLaMAForCausalLM"] | 32001 | float16 | 4.27.0.dev0 | 4,096 | 11,008 | 32 | 32 | null | silu | null | True | null | null | 0.000001 | 0.02 | 0 | 1 | null | 128 | null | false | null | null | true | false | false | false | false | false | false | false | false | false | 6.442451 | medium | null |
1,825 | AdaptLLM/finance-chat | llama | ["LlamaForCausalLM"] | 32001 | float16 | 4.31.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,826 | AdaptLLM/food-Qwen2.5-VL-3B-Instruct | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 151936 | bfloat16 | 4.51.3 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | False | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
1,827 | AdaptLLM/medicine-chat | llama | ["LlamaForCausalLM"] | 32001 | float16 | 4.31.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,828 | AdaptLLM/remote-sensing-LLaVA-NeXT-Llama3-8B | llava_next | ["LlavaNextForConditionalGeneration"] | 128257 | float16 | 4.39.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
1,829 | AdaptLLM/remote-sensing-Qwen2-VL-2B-Instruct | qwen2_vl | ["Qwen2VLForConditionalGeneration"] | 151936 | bfloat16 | 4.45.0.dev0 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | False | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.792723 | small | very_long |
1,830 | AdaptLLM/remote-sensing-Qwen2.5-VL-3B-Instruct | qwen2_5_vl | ["Qwen2_5_VLForConditionalGeneration"] | 151936 | bfloat16 | 4.51.3 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | False | 128,000 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
1,831 | AdaptLLM/visual-instruction-synthesizer | llava_next | ["LlavaNextForConditionalGeneration"] | 128257 | float16 | 4.39.0.dev0 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
1,832 | AdapterOcean/expert_53c4f9908e134f65867c5b72c836d718 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,833 | AdapterOcean/expert_8758633174e848f1a94fa8449fe4c5c2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,834 | AdapterOcean/expert_a6ae28522de54fbf9bdf7ae624dde68c | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,835 | AdapterOcean/expert_e6e205c66b39455bafc9c960f5d2da5e | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,836 | Adapting/bert-base-chinese-finetuned-NER-biomedical | bert | ["BertForTokenClassification"] | 21128 | float32 | 4.22.1 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,837 | Adarsh-aot/ai_class | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.38.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,838 | Adarsh-aot/ai_poet | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.38.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,839 | Adarsh203/new_mT5_Sum | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.37.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,840 | Adbhut/gpt2_ft-ai4privacy-open-pii-masking-500k-ai4privacy | gpt2 | ["GPT2LMHeadModel"] | 50261 | float32 | 4.52.4 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,841 | Adbhut/speecht5-finetuned-voxpopuli_sl | speecht5 | ["SpeechT5ForTextToSpeech"] | 81 | float32 | 4.33.2 | 768 | null | null | null | null | gelu | 0.1 | False | null | null | null | 0.02 | 0 | 2 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,842 | Addaci/bert-base-multilingual-cased-finetuned-yiddish-experiment-2 | bert | ["BertForMaskedLM"] | 119547 | float32 | 4.47.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,843 | Addaci/bert-base-multilingual-cased-finetuned-yiddish-experiment-3 | bert | ["BertForMaskedLM"] | 119547 | float32 | 4.47.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,844 | AddieFoote0/arithmetic-.3b-reference | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float32 | 4.49.0 | 768 | 3,072 | 14 | 8 | 4 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 96 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.09909 | small | long |
1,845 | AddieFoote0/arithmetic-300M-MaxEnt | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.49.0 | 768 | 3,072 | 14 | 8 | 4 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 96 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.09909 | small | long |
1,846 | AddieFoote0/arithmetic-300M-reference | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float32 | 4.49.0 | 768 | 3,072 | 14 | 8 | 4 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 96 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.09909 | small | long |
1,847 | AddieFoote0/language-100M-MaxEnt | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.49.0 | 320 | 1,280 | 14 | 8 | 4 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 40 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.017203 | small | long |
1,848 | AddieFoote0/language-100M-MaxEnt-distilled | gemma2 | ["Gemma2ForCausalLM"] | 256000 | float32 | 4.49.0 | 320 | 1,280 | 14 | 8 | 4 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 40 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.017203 | small | long |
1,849 | AddieFoote0/language-100M-MaxEnt-distilled-relearned | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.49.0 | 320 | 1,280 | 14 | 8 | 4 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 40 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.017203 | small | long |
1,850 | AddieFoote0/language-100M-MaxEnt-relearned | gemma2 | ["Gemma2ForCausalLM"] | 256000 | bfloat16 | 4.49.0 | 320 | 1,280 | 14 | 8 | 4 | gelu_pytorch_tanh | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 40 | 2 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.017203 | small | long |
1,851 | Addy2903/gita-text-generation-gpt2 | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.46.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,852 | Ade-Pyaar/eng_fr_model | marian | ["MarianMTModel"] | 59514 | float32 | 4.47.1 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,853 | Adebolajo/llama3-instruction-AWQ | null | [] | 128256 | null | null | 4,096 | 14,336 | 32 | 32 | 8 | silu | null | null | 8,192 | null | null | null | null | null | null | 128 | 4 | false | null | null | false | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
1,854 | Adekiitan11/bert-email-classifier | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.49.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,855 | Adelina111/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 | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.42.4 | 384 | 1,536 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 32 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.021234 | small | short |
1,856 | Adeptschneider/dyu-fr-opus-v1.0 | marian | ["MarianMTModel"] | 59514 | null | 4.38.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,857 | Adeptschneider/dyu-fr-opus-v2.0 | marian | ["MarianMTModel"] | 59514 | null | 4.38.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,858 | Adeptschneider/dyu_to_fr_v2.0 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.43.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,859 | Adeptschneider/dyu_to_fr_v4.0 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.43.0.dev0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,860 | Adeptschneider/opus-mt-dyu-fr | marian | ["MarianMTModel"] | 59514 | null | 4.38.2 | null | null | 6 | null | null | null | 0 | True | 512 | null | null | null | 0 | 0 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,861 | AdhamEhab/ImageCaptioning | blip | ["BlipForConditionalGeneration"] | null | float32 | 4.40.2 | null | null | null | null | null | null | null | null | null | null | null | 0.02 | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,862 | AdhamEhab/StoryGen | gpt2 | ["GPT2LMHeadModel"] | 50266 | float32 | 4.40.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,863 | Adhithyamadesh2001/DeepSeek-R1-ApmChatbot_tanglish | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.51.1 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
1,864 | Adi-25/phi-3-mini-instruct-openvino-int4-model | phi3 | ["Phi3ForCausalLM"] | 32064 | bfloat16 | 4.49.0 | 3,072 | 8,192 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 96 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 3.623879 | medium | medium |
1,865 | Adignite/LawRL_NewDF_Flant5_PPO | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,866 | Adignite/MailSense_Classifier-chat-llama7b | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.33.1 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,867 | Adiii02/medicine-gpt_llama2_7B_qlora | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
1,868 | Adiii143/bert-finetuned-ner | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,869 | AdilHayat173/token_classification | bert | ["BertForTokenClassification"] | 28996 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,870 | Adilmar/test_trainer | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.30.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,871 | Adipta/audio_classification | wav2vec2 | ["Wav2Vec2ForSequenceClassification"] | 32 | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | 0.1 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
1,872 | Adipta/image_classification | vit | ["ViTForImageClassification"] | null | float32 | 4.42.4 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
1,873 | AdirK/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,874 | Adish14/llama3-8b-instruct-raft-merged-fullmodel-6056-v2 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.51.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 131,072 | 500,000 | 0.00001 | 0.02 | 128000 | 128009 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | very_long |
1,875 | AditKaushik/Bert_scratch1 | bert | ["BertForMaskedLM"] | 30522 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,876 | AdithyaSK/LLama3-Gaja-Hindi-8B-Instruct-alpha | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
1,877 | AdithyaSK/LLama3Tokenizer | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.0.dev0 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
1,878 | Aditya0619/Podcaster_phi4 | llama | ["LlamaForCausalLM"] | 100352 | float16 | 4.47.1 | 5,120 | 17,920 | 40 | 40 | 10 | silu | 0 | True | 16,384 | 250,000 | 0.00001 | 0.02 | 100257 | 100265 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 12.582912 | large | long |
1,879 | Aditya1000101/trained_summarization_model | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.48.3 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,880 | Aditya757864/whisper-tiny-hi | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.38.0.dev0 | null | null | 4 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
1,881 | AdityaAKUMAR/Memma_More_Humorous | gemma2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,882 | AdityaBajracharya/ppo-lunarlander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,883 | AdityaBavadekar/gpt2-medium-finetuned-qamodel | gpt2 | ["GPT2LMHeadModel"] | 50261 | float32 | 4.31.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
1,884 | AdityaBorse11/poca-SoccerTwos | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,885 | AdityaK10/xlm-roberta-base-finetuned-panx-de | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.16.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,886 | AdityaK10/xlm-roberta-base-finetuned-panx-de-fr | xlm-roberta | ["XLMRobertaForTokenClassification"] | 250002 | float32 | 4.16.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,887 | AdityaKothari/WellnessAI-7B-5-bit | mistral | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | true | false | false | false | null | null | null |
1,888 | AdityaKumar2408/rare-puppers | vit | ["ViTForImageClassification"] | null | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
1,889 | AdityaLavaniya/TinyLlama-Fitness-Instructor | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.35.2 | 2,048 | 5,632 | 22 | 32 | 4 | silu | null | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
1,890 | AdityaNamuduri/t5_recommendation_sports_equipment_english | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.26.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
1,891 | AdityaNamuduri/testing | bart | ["BartForConditionalGeneration"] | 50265 | float32 | 4.39.3 | null | null | 6 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
1,892 | AdityaXPV/Llama-3-8b-Instruct-law-sage-v0.1 | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
1,893 | Adityaganesh/Telugu_Sentiment_Analysis | bert | ["BertForSequenceClassification"] | 119547 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,894 | AdityasArsenal/finetuned-for-YogaPosesv4 | mobilenet_v2 | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
1,895 | Adnan-AI-Labs/DistilBERT-ProductClassifier | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.46.1 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
1,896 | AdnanRiaz107/CBertbase-APPS10k | roberta | ["RobertaForCausalLM"] | 50265 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | False | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,897 | AdnanRiaz107/CBertbase-mlm-APPS10k | roberta | ["RobertaForCausalLM"] | 50265 | float32 | 4.38.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | False | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
1,898 | AdnanRiaz107/CodeGemma2b-300APPS | gemma | ["GemmaForCausalLM"] | 256000 | bfloat16 | 4.40.2 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
1,899 | AdnanRiaz107/CodeLLAMA3-8BI-300APPS | llama | ["LlamaForCausalLM"] | 128256 | bfloat16 | 4.40.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.