Kirim 2 (ๅ้ 2)
Collection
5 items
โข
Updated
Kirim-V2 is an advanced general-purpose language model with 26 billion parameters, featuring an innovative sparse activation architecture where 18 billion parameters are actively engaged during inference. This design delivers high performance while maintaining computational efficiency.
Core Competencies
Advanced Features
Kirim-V2 represents a significant advancement over Kirim-V1, featuring:
Architecture: Sparse Transformer
Training Data: Diverse web corpus, code, and specialized datasets
Tokenizer: Custom trained tokenizer optimized for multilingual performance
Optimization: Mixed precision training with gradient checkpointing
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "Kirim-ai/Kirim-V2"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
device_map="auto",
torch_dtype="auto"
)
prompt = "Explain quantum entanglement in simple terms."
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
This model should be used responsibly. Users should verify critical information independently and be aware of potential biases in generated content. The model is not intended for making decisions in high-stakes scenarios without human oversight.
This model is released under the Apache 2.0 License.
@model{kirimv2_2025,
title={Kirim-V2: A 26B Parameter Sparse Activation Language Model},
author={Qiling Tech},
year={2026},
publisher={Hugging Face},
url={https://huggingface.co/Kirim-ai/Kirim-V2}
}
Release Date: 2026 Model Type: Causal Language Model