SozKZ Core: Kazakh Language Models
Collection
Base, instruct, and balanced Kazakh language models trained from scratch — Llama (50M–600M), GPT2, Pythia architectures • 22 items • Updated
Domain-adaptive pretraining (DAPT) of EleutherAI/pythia-14m on Kazakh text. The first Kazakh language model experiment in the Soz project — a proof of concept.
| Property | Value |
|---|---|
| Base model | EleutherAI/pythia-14m |
| Training steps | 13,000 |
| Method | Domain-adaptive pretraining |
| Language | Kazakh |
| License | Apache 2.0 |
from transformers import AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("EleutherAI/pythia-14m")
model = AutoModelForCausalLM.from_pretrained("stukenov/slm-kk-pythia14m-dapt-13k")
input_ids = tokenizer("Қазақстан — ", return_tensors="pt").input_ids
output = model.generate(input_ids, max_new_tokens=50, do_sample=True, temperature=0.8)
print(tokenizer.decode(output[0], skip_special_tokens=True))
This was the first experiment in the Soz project — a proof of concept for Kazakh language modeling. The Pythia-14m base model uses an English-centric tokenizer, making it suboptimal for Kazakh. Later models in the project use Kazakh-native tokenizers.
Part of the Soz — Kazakh Language Models project, a research effort to build open-source language models for Kazakh.
@misc{tukenov2026soz,
title={Soz: Small Language Models for Kazakh},
author={Tukenov, Saken},
year={2026},
url={https://huggingface.co/stukenov/slm-kk-pythia14m-dapt-13k}
}
Apache 2.0
Base model
EleutherAI/pythia-14m