roshniramesh 's Collections int8 llm
updated
meta-llama/Llama-Guard-3-8B-INT8
Text Generation
• 8B • Updated • 13.5k
• 38
google/gemma-7b-quant-pytorch
Text Generation
• Updated • 58
• 2
Intel/gpt-j-6B-int8-dynamic-inc
Text Generation
• Updated • 20
• 16
Intel/t5-small-xsum-int8-dynamic-inc
Updated • 60
• 1
INC4AI/bert-base-uncased-mrpc-int8-static-inc
Text Classification
• Updated • 33
Intel/bert-large-uncased-cola-int8-inc
Text Classification
• Updated • 30
INC4AI/vit-base-patch16-224-int8-static-inc
Image Classification
• Updated • 34
• 1
INC4AI/albert-base-v2-sst2-int8-static-inc
Text Classification
• Updated • 39
Intel/roberta-base-mrpc-int8-dynamic-inc
Text Classification
• Updated • 13
INC4AI/roberta-base-mrpc-int8-static-inc
Text Classification
• Updated • 29
Intel/xlm-roberta-base-mrpc-int8-dynamic-inc
Text Classification
• Updated • 15
Intel/dynamic-minilmv2-L6-H384-squad1.1-int8-static
Question Answering
• 30.1M • Updated • 35
Intel/MiniLM-L12-H384-uncased-mrpc-int8-dynamic-inc
Text Classification
• Updated • 12
Intel/bart-large-cnn-int8-dynamic-inc
Updated • 34
• 1
Intel/distilbart-cnn-12-6-int8-dynamic-inc
Updated • 27
• 2
Intel/gpt-j-6B-int8-static-inc
Text Generation
• Updated • 28
• 9
Intel/gpt-j-6B-pytorch-int8-static-inc
Text Generation
• Updated • 21
Intel/bert-base-cased-finetuned-sst2-int8-inc
Text Classification
• Updated • 35
Intel/bert-base-uncased-CoLA-int8-inc
Text Classification
• Updated • 33
Intel/bert-base-uncased-QNLI-int8-inc
Text Classification
• Updated • 34
Intel/bert-base-uncased-STS-B-int8-inc
Text Classification
• Updated • 32
INC4AI/bert-base-uncased-mrpc-int8-qat-inc
Text Classification
• Updated • 40
• 1
Intel/bert-large-uncased-rte-int8-dynamic-inc
Text Classification
• Updated • 27
Intel/bert-large-uncased-rte-int8-static-inc
Text Classification
• Updated • 35
Intel/distilbert-base-uncased-distilled-squad-int8-static-inc
Question Answering
• Updated • 78
• 5
Intel/distilbert-base-uncased-MRPC-int8-dynamic-inc
Text Classification
• Updated • 21
• 1
Intel/distilbert-base-uncased-MRPC-int8-static-inc
Text Classification
• Updated • 20
Intel/albert-base-v2-sst2-int8-dynamic-inc
Text Classification
• Updated • 32
Intel/albert-base-v2-MRPC-int8-inc
Text Classification
• Updated • 31
Intel/bge-small-en-v1.5-rag-int8-static
Feature Extraction
• Updated • 64
• 2
Intel/bge-base-en-v1.5-rag-int8-static
Feature Extraction
• Updated • 39
INC4AI/falcon-7b-sq-int8-inc
Text Generation
• Updated • 33
amd/Llama-3.1-8B-Instruct-w-int8-a-int8-sym-test
8B • Updated • 15.7k
RedHatAI/Llama-3.2-1B-Instruct-quantized.w8a8
Text Generation
• 1B • Updated • 27.6k
• 8
FriendliAI/Meta-Llama-3-8B-int8
Text Generation
• 8B • Updated • 11
• 1
google/gemma-7b-it-quant-pytorch
Text Generation
• Updated • 55
• 11
OpenVINO/mistral-7b-instruct-v0.1-int8-ov
Text Generation
• Updated • 38
• 1
FriendliAI/Meta-Llama-3.1-8B-Instruct-int8
Text Generation
• 8B • Updated • 12.3k
• 1
Text Generation
• 14B • Updated • 91
• 7
Text Generation
• 8B • Updated • 111
• 9
Text Generation
• 2B • Updated • 107
• 5
Qwen/Qwen1.5-1.8B-Chat-GPTQ-Int8
Text Generation
• 2B • Updated • 47
• 2
Qwen/Qwen1.5-14B-Chat-GPTQ-Int8
Text Generation
• 15B • Updated • 51
• 11
Qwen/Qwen1.5-4B-Chat-GPTQ-Int8
Text Generation
• 4B • Updated • 43
• 6
Qwen/Qwen1.5-72B-Chat-GPTQ-Int8
Text Generation
• 72B • Updated • 55
• 7
Qwen/Qwen1.5-4B-Chat-GGUF
Text Generation
• 4B • Updated • 832
• 16
Qwen/Qwen1.5-0.5B-Chat-GGUF
Text Generation
• 0.6B • Updated • 7.38k
• 35
Qwen/Qwen1.5-7B-Chat-GGUF
Text Generation
• 8B • Updated • 945
• 70
Qwen/CodeQwen1.5-7B-Chat-GGUF
Text Generation
• 7B • Updated • 973
• 110
Qwen/Qwen2.5-1.5B-Instruct-GPTQ-Int8
Text Generation
• 2B • Updated • 150
• 6
Qwen/Qwen2.5-0.5B-Instruct-GPTQ-Int8
Text Generation
• 0.5B • Updated • 728
• 10
Qwen/Qwen2.5-0.5B-Instruct-GGUF
Text Generation
• 0.6B • Updated • 90.2k
• 93
Qwen/Qwen2-1.5B-Instruct-GGUF
Text Generation
• 2B • Updated • 25.8k
• 29
Qwen/Qwen2-0.5B-Instruct-GGUF
Text Generation
• 0.5B • Updated • 46.8k
• 72
Qwen/Qwen2-7B-Instruct-GGUF
Text Generation
• 8B • Updated • 9.73k
• 179
Qwen/Qwen2-0.5B-Instruct-GPTQ-Int8
Text Generation
• 0.6B • Updated • 84
• 4
Qwen/Qwen2-1.5B-Instruct-GPTQ-Int8
Text Generation
• 2B • Updated • 63
• 4
Qwen/Qwen2-7B-Instruct-GPTQ-Int8
Text Generation
• 8B • Updated • 104
• 17
Qwen/Qwen2-72B-Instruct-GPTQ-Int8
Text Generation
• 73B • Updated • 51
• 15