Collections

Discover the best community collections!

Collections trending this week
SozKZ MoE: Mixture of Experts
Mixture-of-Experts models for Kazakh — upcycled and domain-pretrained MoE architectures
SozKZ Core: Kazakh Language Models
Base, instruct, and balanced Kazakh language models trained from scratch — Llama (50M–600M), GPT2, Pythia architectures
SozKZ Corpora: Kazakh Training Datasets
Training corpora for Kazakh LLMs — raw, cleaned, deduplicated, tokenized, synthetic, and parallel datasets
SozKZ GEC: Kazakh Grammar Error Correction
Grammar error correction models and datasets for Kazakh — Llama GEC (300M, 600M), mT5 GEC, morphology models
SozKZ Corpora: Kazakh Training Datasets
Training corpora for Kazakh LLMs — raw, cleaned, deduplicated, tokenized, synthetic, and parallel datasets
SozKZ MoE: Mixture of Experts
Mixture-of-Experts models for Kazakh — upcycled and domain-pretrained MoE architectures
SozKZ GEC: Kazakh Grammar Error Correction
Grammar error correction models and datasets for Kazakh — Llama GEC (300M, 600M), mT5 GEC, morphology models
SozKZ Core: Kazakh Language Models
Base, instruct, and balanced Kazakh language models trained from scratch — Llama (50M–600M), GPT2, Pythia architectures