Turkish-LLM-32B-Instruct

The largest open-source Turkish-enhanced language model. Fine-tuned from Qwen2.5-32B-Instruct with QLoRA on a carefully curated 173K Turkish instruction dataset.

Part of the Turkish LLM Family - a complete suite of Turkish language models from 7B to 32B.

Highlights

  • 32B parameters - largest openly available Turkish fine-tuned model
  • Outperforms base model on MMLU-TR (+2.71) and XCOPA-TR (+1.00)
  • 67.89% MMLU-TR - significant improvement through iterative dataset engineering
  • GGUF available - Q4/Q5/Q8 quantizations for local inference

Benchmark Results

Benchmark Base (Qwen2.5-32B) v1 (Mar 21) v2 (Current) Delta vs Base
MMLU-TR (57 categories) 0.6518 0.6564 0.6789 +2.71
XCOPA-TR (Causal Reasoning) 0.6800 0.6740 0.6900 +1.00
XNLI-TR (NLI) 0.4578 0.4610 0.4514 -0.64

Iterative Improvement

This model is the result of systematic dataset engineering across multiple iterations:

  • v1 (Mar 21): Initial fine-tune with 242K examples. Improved MMLU-TR and XNLI-TR but regressed on XCOPA-TR.
  • v2 (Mar 29): Rebalanced dataset (173K examples) with XCOPA augmentation and evaluation-aligned NLI formatting. Achieved improvements on both MMLU-TR and XCOPA-TR.

Key insight: reducing dataset size from 242K to 173K while improving data quality led to better results — quality over quantity.

MMLU-TR: Strongest Category Improvements (v2)

Category Base Ours Delta
College Computer Science 0.545 0.616 +7.1
Logical Fallacies 0.640 0.696 +5.6
College Mathematics 0.530 0.580 +5.0
Formal Logic 0.508 0.556 +4.8
High School Mathematics 0.507 0.548 +4.1

Quick Start

With Transformers

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained(
    "ogulcanaydogan/Turkish-LLM-32B-Instruct",
    torch_dtype="auto",
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("ogulcanaydogan/Turkish-LLM-32B-Instruct")

messages = [
    {"role": "system", "content": "Sen yardimci bir Turkce asistansin."},
    {"role": "user", "content": "Yapay zekanin saglik sektorundeki uygulamalarini acikla."}
]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer([text], return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

With Ollama (GGUF)

ollama run hf.co/ogulcanaydogan/Turkish-LLM-32B-Instruct-GGUF:Q4_K_M

With vLLM

vllm serve ogulcanaydogan/Turkish-LLM-32B-Instruct --dtype auto --max-model-len 4096

Training Details

Parameter Value
Base Model Qwen/Qwen2.5-32B-Instruct
Method QLoRA (4-bit NF4 + double quantization)
LoRA rank / alpha 32 / 64
Learning rate 1e-5 (cosine schedule)
Epochs 1
Effective batch size 16
Max sequence length 2048
Training time ~55 hours on NVIDIA A100 80GB
Dataset 173K Turkish instruction examples (v7.1)

Dataset Composition (v7.1)

Source Examples Percentage
Turkish Math 100,000 57.9%
Turkish Exam Instructions 41,297 23.9%
XNLI Augmented (MC format) 10,000 5.8%
GSM8K Turkish 8,760 5.1%
Alignment Data 7,245 4.2%
XCOPA Augmented 5,000 2.9%
GPQA Turkish 545 0.3%

Turkish LLM Family

Model Size MMLU-TR Download
Turkish-LLM-7B-Instruct 7B - GGUF
Turkish-LLM-14B-Instruct 14B 0.5977 GGUF
Turkish-LLM-32B-Instruct 32B 0.6789 GGUF

Limitations

  • Slight regression on XNLI-TR natural language inference (-0.64 points)
  • Inherits base model limitations for very long contexts
  • Best suited for Turkish STEM, reasoning, and general knowledge tasks

Citation

@misc{aydogan2026turkishllm,
  title={Turkish LLM Family: Open-Source Turkish Language Models},
  author={Ogulcan Aydogan},
  year={2026},
  url={https://huggingface.co/collections/ogulcanaydogan/turkish-llm-family-69b303b4ef1c36caffca4e94}
}
Downloads last month
332
Safetensors
Model size
33B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ogulcanaydogan/Turkish-LLM-32B-Instruct

Base model

Qwen/Qwen2.5-32B
Finetuned
(1215)
this model
Quantizations
3 models

Dataset used to train ogulcanaydogan/Turkish-LLM-32B-Instruct

Collection including ogulcanaydogan/Turkish-LLM-32B-Instruct

Evaluation results