Kwen-1.5B

Model Description

Kwen-1.5B is a fine-tuned version of the Qwen2.5-1.5B base model. This model has been specifically trained to embody the identity of Kwen, an AI assistant developed by The Kwen Foundation.

Through fine-tuning, the model has integrated its new identity while maintaining the high reasoning and linguistic capabilities of the original Qwen architecture.

Key Features

  • Identity-Aware: Recognizes itself as Kwen and references the Kwen Foundation.
  • Efficient Size: Optimized for consumer GPUs (like the RTX 4060).
  • Architecture: Based on Qwen2.5, utilizing Causal Language Modeling with advanced attention mechanisms.

Model Lineage

  • Base Model: Qwen/Qwen2.5-1.5B
  • Developer: The Kwen Foundation
  • Fine-tuning Method: LoRA/QLoRA (via Unsloth)

Quants (GGUF)

The weights provided in this repository include quantized versions created using llama.cpp.

File Name Method Size Description
kwen_q4_k_m.gguf Q4_K_M ~935 MB High compression, balanced quality (Recommended)
kwen_f16.gguf F16 ~3.0 GB Uncompressed weights for maximum precision

How to use with Ollama

  1. Run ollama run hf.co/TheKwenFoundation/Kwen-1.5B:Q4_K_M
Downloads last month
5
GGUF
Model size
2B params
Architecture
qwen2
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ 1 Ask for provider support

Model tree for TheKwenFoundation/Kwen-1.5B

Quantized
(65)
this model