HybriKo - Korean Hybrid LLM
Griffin-inspired hybrid architecture combining RNN and Attention mechanisms.
Model Details
- Architecture: Hybrid RNN + Attention (2:1 ratio)
- Parameters: 117.8M
- Training: Continued pretraining on exp4_plus dataset
- Base Model: exp4 (Wikipedia pretrained)
Training Data
- korean_textbooks_tiny (50K samples)
- korean_textbooks_edu (50K samples)
- korean_public_corpus (50K samples)
Usage
from hybridko.model import HybriKoModel, HybriKoConfig
config = HybriKoConfig.from_yaml("config.yaml")
model = HybriKoModel(config)
# Load checkpoint
checkpoint = torch.load("checkpoints/checkpoint_step_XXX.pt")
model.load_state_dict(checkpoint["model_state_dict"])
Citation
@misc{hybridko2024,
title={HybriKo: Korean Hybrid LLM},
year={2024},
}
- Downloads last month
- 14
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support