index int64 0 125k | modelId stringlengths 6 115 | config_model_type stringlengths 2 46 ⌀ | config_architectures stringlengths 2 91 ⌀ | config_vocab_size stringlengths 1 8 ⌀ | config_torch_dtype stringclasses 7 values | config_transformers_version stringclasses 228 values | config_hidden_size float64 0 18.4k ⌀ | config_intermediate_size float64 0 25.2M ⌀ | config_num_hidden_layers float64 -1 260 ⌀ | config_num_attention_heads stringclasses 47 values | config_num_key_value_heads float64 0 4.1k ⌀ | config_hidden_act stringclasses 19 values | config_attention_dropout float64 0 0.5 ⌀ | config_use_cache stringclasses 3 values | config_max_position_embeddings float64 -1 10.5M ⌀ | config_rope_theta float64 256 100B ⌀ | config_rms_norm_eps float64 0 0 ⌀ | config_initializer_range float64 0 2 ⌀ | config_bos_token_id stringclasses 158 values | config_eos_token_id stringclasses 339 values | config_tie_word_embeddings bool 2 classes | config_head_dimension float64 0.5 3.07k ⌀ | config_gqa_ratio float64 0.5 64 ⌀ | config_moe_enabled bool 1 class | config_n_routed_experts float64 1 384 ⌀ | config_num_experts_per_tok float64 1 64 ⌀ | is_llama_family bool 2 classes | is_bert_family bool 2 classes | is_gpt_family bool 2 classes | is_t5_family bool 2 classes | is_whisper_family bool 2 classes | is_deepseek_family bool 2 classes | is_mistral_family bool 2 classes | uses_moe bool 2 classes | uses_gqa bool 2 classes | uses_rope bool 2 classes | config_approx_params_billions float64 -0.2 606 ⌀ | size_category stringclasses 4 values | context_category stringclasses 4 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
3,300 | AmirlyPhd/V11-bert-text-classification-model | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,301 | AmirlyPhd/V4-bert-text-classification-model | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,302 | AmirlyPhd/cybercrime_LSTM_GLOVE | lstm | [] | null | null | null | 128 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,303 | AmirlyPhd/final_V0-bert-text-classification-model | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,304 | AmirlyPhd/final_V1-roberta-text-classification-model | roberta | ["RobertaForSequenceClassification"] | 50265 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,305 | AmirlyPhd/final_V1_xai-bert-text-classification-model | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,306 | AmirlyPhd/final_V4_resized_balanced_Bert_balanced_dataset-after-adding-new-words-text-classification-model | bert | ["BertForSequenceClassification"] | 30641 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,307 | AmirlyPhd/merged_model_V4_merged_all_IEEEdatasets_threat | llama | ["LlamaForCausalLM"] | 32002 | float16 | 4.36.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
3,308 | AmirlyPhd/merged_model_V7_merged_all_IEEEdatasets_threat | llama | ["LlamaForCausalLM"] | 32002 | float16 | 4.36.2 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | True | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
3,309 | AmirlyPhd/v10-bert-combined-dataset-text-classification-model | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.39.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,310 | Amirmerfan/bert-base-uncased-persian-ner-10k-mini | bert | ["BertForTokenClassification"] | 105879 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,311 | Amirmerfan/bert-base-uncased-persian-ner-50k-base | bert | ["BertForTokenClassification"] | 105879 | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,312 | Amirmerfan/distilbert-finetuned-ner-restaurant | distilbert | ["DistilBertForTokenClassification"] | 30522 | float32 | 4.48.3 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,313 | Amirmerfan/t5-dialogsum-summarizer | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.48.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,314 | Amirmerfan/vit-base-indian-food-classifier | vit | ["ViTForImageClassification"] | null | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
3,315 | Amirmerfan/vit-persian-food-classifier-mini | vit | ["ViTForImageClassification"] | null | float32 | 4.48.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.084935 | small | null |
3,316 | Amirmerfan/whisper-small-finetuned-large-clean | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.51.1 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,317 | Amirmerfan/whisper-small-finetuned-medium-noisy | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.51.1 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,318 | Amirmerfan/whisper-small-finetuned-small-noisy | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.51.1 | null | null | 12 | null | null | null | 0 | False | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,319 | Amirmnsh/ppo-LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,320 | AmirrezaV1/PsychologistChatbot_emotional_model | xlm-roberta | ["XLMRobertaForSequenceClassification"] | 250002 | float32 | 4.38.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,321 | AmitTewari/test_trainer | bert | ["BertForSequenceClassification"] | 28996 | float32 | 4.40.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,322 | Amitesh007/tw-sentiment-finetuned | distilbert | ["DistilBertForSequenceClassification"] | 28996 | null | 4.26.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,323 | AmithAdiraju1694/Cerebras-GPT-590M_menuitemexp | gpt2 | ["GPT2LMHeadModel"] | 51552 | float32 | 4.47.1 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
3,324 | Amitz244/output_dir_controlnet | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,325 | AmiyendraOP/llama3-legal-finetuned | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.51.3 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
3,326 | Amjad123/Llama-2-7b-QA-Telecom | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
3,327 | Amjad123/Llama-2-7b-chat-finetune | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
3,328 | AmjadKha/Boppy | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.41.2 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
3,329 | Ammad1Ali/Llongorca_16k | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
3,330 | Ammad1Ali/bert2bert | encoder-decoder | ["EncoderDecoderModel"] | 30522 | float32 | 4.38.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | 102 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,331 | Ammad1Ali/llama-v2-7B-alt | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.31.0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
3,332 | Ammar-1/llama2-7b-parenting-finetuned | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.32.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | null | True | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
3,333 | Ammok/LunarLander-v2 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,334 | Ammok/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,335 | Ammok/predict-llm | bert | ["BertForSequenceClassification"] | 30522 | float32 | 4.33.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,336 | Ammonsh/wav2vec2-common_voice-tr-demo | wav2vec2 | ["Wav2Vec2ForCTC"] | 41 | float32 | 4.21.0.dev0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
3,337 | Amna100/DebertaMLMnc2c | deberta | ["DebertaForMaskedLM"] | 50265 | float32 | 4.35.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | null | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,338 | Amo/gpt-neo-125m-mlp-micro | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float16 | 4.17.0 | 768 | null | null | null | null | null | 0 | False | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
3,339 | Amod/falcon7b-fine-tuned-therapy-merged | RefinedWebModel | ["RWForCausalLM"] | 65024 | float16 | 4.30.2 | 4,544 | null | null | null | null | null | 0 | True | null | null | null | 0.02 | 11 | 11 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,340 | Amogh01/whisper-small-hi | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.46.2 | null | null | 12 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,341 | Amogh06/DeTr-Table-Detection-Model-public3 | detr | ["DetrForObjectDetection"] | null | float32 | null | null | null | 6 | null | null | null | 0 | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,342 | AmoooEBI/Bert-fa-qa-finetuned-on-PersianQA | bert | ["BertForQuestionAnswering"] | 100000 | float32 | 4.51.3 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,343 | AmoooEBI/xlm-roberta-fa-qa-finetuned-on-PersianQA | xlm-roberta | ["XLMRobertaForQuestionAnswering"] | 250002 | float32 | 4.51.3 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,344 | Amoros/Amoros_Beaugosse_test-large-2025_05_20_53326-bs32_freeze | dinov2 | ["Dinov2ForImageClassification"] | null | float32 | 4.48.0 | 1,024 | null | 24 | 16 | null | gelu | null | null | null | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
3,345 | Amr-h/AmrdisBert | distilbert | ["DistilBertForQuestionAnswering"] | 28996 | float32 | 4.40.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,346 | Amr-h/DisBertAmrHamedAbdelmoniem | distilbert | ["DistilBertForQuestionAnswering"] | 28996 | float32 | 4.40.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,347 | AmrMKayid/synchformer-hf | synchformer | ["SynchformerModel"] | null | null | 4.45.2 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,348 | AmrSheta/Meme | bert | ["BertModel"] | 28996 | float32 | 4.17.0 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,349 | Amrinkar/BoxModel | llava | ["LISAForCausalLM"] | 32003 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
3,350 | Amrinkar/CartModel2 | llava | ["LISAForCausalLM"] | 32003 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
3,351 | Amrinkar/CoffeeMachineModel2 | llava | ["LISAForCausalLM"] | 32003 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
3,352 | Amrinkar/LisaSelected5 | llava | ["LISAForCausalLM"] | 32003 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
3,353 | Amrinkar/PARIS3D | llava | ["LISAForCausalLM"] | 32003 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
3,354 | Amrinkar/PartNameModel | llava | ["LISAForCausalLM"] | 32003 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
3,355 | Amrinkar/lisa2prompt | llava | ["LISAForCausalLM"] | 32003 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
3,356 | Amrinkar/lisanoprompt | llava | ["LISAForCausalLM"] | 32003 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
3,357 | Amrinkar/lisatest80 | llava | ["LISAForCausalLM"] | 32003 | bfloat16 | 4.31.0 | 5,120 | 13,824 | 40 | 40 | 40 | silu | null | False | 4,096 | null | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | false | false | false | false | false | false | false | false | false | true | 12.582912 | large | medium |
3,358 | AmritaBha/sd15_fill50k | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,359 | Amritha1903/flan-t5-houseprice | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.52.4 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,360 | AmrutaMuthal/mero_sd15_controlnet_inpaint_masked_loss_wt_parts_ellipse | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,361 | Amshaker/Mobile-VideoGPT-0.5B | MobileVideoGPT_qwen | ["MobileVideoGPTQwenForCausalLM"] | 151666 | float16 | 4.42.4 | 896 | 4,864 | 24 | 14 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 64 | 7 | false | null | null | false | false | true | false | false | false | false | false | true | true | 0.231211 | small | very_long |
3,362 | Amshaker/Mobile-VideoGPT-1.5B | MobileVideoGPT_qwen | ["MobileVideoGPTQwenForCausalLM"] | 151666 | float16 | 4.42.4 | 1,536 | 8,960 | 28 | 12 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 6 | false | null | null | false | false | true | false | false | false | false | false | true | true | 0.792723 | small | very_long |
3,363 | Amu/t1-3B | qwen2 | ["Qwen2ForCausalLM"] | 151936 | float16 | 4.49.0 | 2,048 | 11,008 | 36 | 16 | 2 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | true | 128 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 1.811939 | medium | very_long |
3,364 | Amu/zen | mistral | ["MistralForCausalLM"] | 32002 | bfloat16 | 4.36.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 10,000 | 0.00001 | 0.02 | 1 | 32000 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | long |
3,365 | Amusam/pegasus-multipak | pegasus | ["PegasusForConditionalGeneration"] | 96103 | float32 | 4.44.2 | null | null | 16 | null | null | null | 0.1 | True | 1,024 | null | null | null | 0 | 1 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
3,366 | An0rak/alpaca_unsloth_gguf | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
3,367 | AnAmbitiousMonk/ppo-LunarLander-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,368 | AnAmbitiousMonk/ppo-LunarLander-v4 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,369 | AnHoangbg/t5-small-finetuned-manimml-1.2 | t5 | ["T5ForConditionalGeneration"] | 32128 | float32 | 4.42.3 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | null | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,370 | AnHoangbg/tinyllama-manim-v1 | llama | ["LlamaForCausalLM"] | 32000 | float16 | 4.42.4 | 2,048 | 5,632 | 22 | 32 | 4 | silu | 0 | True | 2,048 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 64 | 8 | false | null | null | true | false | false | false | false | false | false | false | true | true | 1.107296 | medium | medium |
3,371 | AnOng/whisper-large-v2-vi | whisper | ["WhisperForConditionalGeneration"] | 51865 | float32 | 4.27.0.dev0 | null | null | 32 | null | null | null | 0 | True | null | null | null | null | 50257 | 50257 | null | null | null | false | null | null | false | false | false | false | true | false | false | false | false | false | null | null | null |
3,372 | AnReu/math_pretrained_bert | bert | ["BertForPreTraining"] | 29497 | float32 | 4.9.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,373 | AnReu/math_pretrained_roberta | roberta | ["RobertaForMaskedLM"] | 50769 | float32 | 4.9.2 | 768 | 3,072 | 12 | 12 | null | gelu | null | True | 514 | null | null | 0.02 | 0 | 2 | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.084935 | small | short |
3,374 | AnabSohail/ahmed-detr-finetuned-tumor | detr | ["DetrForObjectDetection"] | null | float32 | null | null | null | 6 | null | null | null | 0 | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,375 | Anaelle27/distilbert-base-uncased-finetuned-sentiment | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.49.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,376 | AnakinCodebreaker/spk_qna_model | t5 | ["T5ForQuestionAnswering"] | 32128 | float32 | 4.34.1 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,377 | AnaluRRamos/mt5-biomedical-translation-mode1 | mt5 | ["MT5ForConditionalGeneration"] | 250112 | float32 | 4.49.0 | null | null | null | null | null | null | null | True | null | null | null | null | null | 1 | false | null | null | false | null | null | false | false | false | true | false | false | false | false | false | false | null | null | null |
3,378 | Anamta98/Qwen2-VL-7B-Instruct-LoRA-FT | qwen2_vl | ["Qwen2VLForConditionalGeneration"] | 152064 | bfloat16 | 4.46.1 | 3,584 | 18,944 | 28 | 28 | 4 | silu | 0 | True | 32,768 | 1,000,000 | 0.000001 | 0.02 | 151643 | 151645 | false | 128 | 7 | false | null | null | false | false | false | false | false | false | false | false | true | true | 4.315939 | medium | very_long |
3,379 | Anand-Keshavan/model | gpt_neo | ["GPTNeoForCausalLM"] | 50257 | float32 | 4.26.0 | 2,048 | null | null | null | null | null | 0 | True | 2,048 | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | medium |
3,380 | AnandBhat/my_awesome_wnut_model | distilbert | ["DistilBertForTokenClassification"] | 30522 | float32 | 4.30.2 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,381 | Anandappu/bert_comedy | bart | ["BartForConditionalGeneration"] | 50264 | float32 | 4.38.2 | null | null | 12 | null | null | null | 0 | True | 1,024 | null | null | null | 0 | 2 | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | short |
3,382 | Anandhd/my_awesome_model | distilbert | ["DistilBertForSequenceClassification"] | 30522 | float32 | 4.44.0 | null | null | null | null | null | null | 0.1 | null | 512 | null | null | 0.02 | null | null | null | null | null | false | null | null | false | true | false | false | false | false | false | false | false | false | null | null | short |
3,383 | Ananjas/AwooAI | gpt2 | ["GPT2LMHeadModel"] | 50257 | float32 | 4.24.0 | null | null | null | null | null | null | null | True | null | null | null | 0.02 | 50256 | 50256 | null | null | null | false | null | null | false | false | true | false | false | false | false | false | false | false | null | null | null |
3,384 | Anant58/a2c-PandaReachDense-v3 | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | true | null | null | null |
3,385 | Anant58/ppo-Pyramid | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,386 | Anant58/ppo-SnowballTarget | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,387 | Anant58/vizdoom_health_gathering_supreme | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,388 | Ananthu357/Ananthus-BAAI-for-contracts2.0 | bert | ["BertModel"] | 30522 | float32 | 4.41.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,389 | Ananthu357/BAAI-first-train | bert | ["BertModel"] | 30522 | float32 | 4.41.2 | 1,024 | 4,096 | 24 | 16 | null | gelu | null | True | 512 | null | null | 0.02 | null | null | null | 64 | null | false | null | null | false | true | false | false | false | false | false | false | false | false | 0.30199 | small | short |
3,390 | Anarchist/myLora | llama | ["LlamaForCausalLM"] | 32000 | float32 | 4.37.0.dev0 | 4,096 | 11,008 | 32 | 32 | 32 | silu | 0 | False | 4,096 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 1 | false | null | null | true | false | false | false | false | false | false | false | false | true | 6.442451 | medium | medium |
3,391 | Anaryl/ppo-Huggy | null | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | false | false | false | false | false | false | false | false | false | false | null | null | null |
3,392 | Anas00/abcd | wav2vec2 | ["Wav2Vec2ForCTC"] | 112 | float32 | 4.18.0 | 1,024 | 4,096 | 24 | 16 | null | gelu | 0 | null | null | null | null | 0.02 | 1 | 2 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | null |
3,393 | Anas099/GumBit_AI | llama | [] | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | false | null | null | true | false | false | false | false | false | false | false | false | false | null | null | null |
3,394 | Anas989898/gemma_2b_it_ds | gemma | ["GemmaForCausalLM"] | 256000 | float16 | 4.38.2 | 2,048 | 16,384 | 18 | 8 | 1 | gelu | 0 | True | 8,192 | 10,000 | 0.000001 | 0.02 | 2 | 1 | null | 256 | 8 | false | null | null | false | false | false | false | false | false | false | false | true | true | 0.90597 | small | long |
3,395 | AnasAber/seamless-darija-eng | seamless_m4t_v2 | ["SeamlessM4Tv2ForTextToText"] | 256102 | float32 | 4.44.0 | 1,024 | null | 24 | 16 | null | null | 0.1 | True | 4,096 | null | null | 0.02 | 2 | 3 | null | 64 | null | false | null | null | false | false | false | false | false | false | false | false | false | false | 0.30199 | small | medium |
3,396 | Anasrafiq/llama3-unsloth-16bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
3,397 | AnassMzn/ECE-AFA | mistral | ["MistralForCausalLM"] | 32000 | bfloat16 | 4.46.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | False | 32,768 | 10,000 | 0.00001 | 0.02 | 1 | 2 | false | 128 | 4 | false | null | null | false | false | false | false | false | false | true | false | true | true | 6.442451 | medium | very_long |
3,398 | AnatolyBelov/CB_16bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
3,399 | AnatolyBelov/CB_4bit | llama | ["LlamaForCausalLM"] | 128256 | float16 | 4.44.2 | 4,096 | 14,336 | 32 | 32 | 8 | silu | 0 | True | 8,192 | 500,000 | 0.00001 | 0.02 | 128000 | 128001 | false | 128 | 4 | false | null | null | true | false | false | false | false | false | false | false | true | true | 6.442451 | medium | long |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.