Qwen3-0.6B - Kto
Model Description
This model is a LoRA Adapter fine-tuned from Qwen/Qwen3-0.6B using the Kto training method.
Kahneman-Tversky Optimization - Binary preference optimization based on Prospect Theory
This model was developed as part of thesis research on LLM Alignment using Preference Optimization Methods.
Model Details
| Property | Value |
|---|---|
| Base Model | Qwen/Qwen3-0.6B |
| Training Method | Kto |
| Model Type | LoRA Adapter |
| Training Date | December 2025 |
| Framework | PyTorch + Transformers + PEFT |
Benchmark Results
| Benchmark | Score |
|---|---|
| HellaSwag (10-shot) | 0.264 |
| TruthfulQA (0-shot MC2) | 0.486 |
| MMLU-Mini (5-shot) | 0.269 |
Comparative Analysis
The following chart compares this method against other training approaches on the same base model:
Training Configuration
| Parameter | Value |
|---|---|
| Epochs | 1 |
| Batch Size | 2 |
| Gradient Accumulation | 8 |
| Effective Batch Size | 16 |
| Learning Rate | 2e-4 |
| Max Sequence Length | 512 |
| LoRA Rank | 16 |
| LoRA Alpha | 32 |
| Dataset | UltraFeedback Binarized |
Usage
Loading as LoRA Adapter
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
# Load base model
base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen3-0.6B")
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen3-0.6B")
# Load adapter
model = PeftModel.from_pretrained(base_model, "Nishef/Qwen3-0.6B-Full_KTO_20251225_102050")
# Generate text
inputs = tokenizer("Hello, how are you?", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0]))
Training Methodology
Kto
Kahneman-Tversky Optimization - Binary preference optimization based on Prospect Theory
Key Features:
- Binary feedback signals (thumbs up/down)
- No need for paired preference data
- Reference model for KL divergence regularization
- Prospect Theory-inspired loss function
Citation
If you use this model in your research, please cite:
@misc{qwen3_0.6b_kto_2025,
title = {Qwen3-0.6B Fine-tuned with Kto},
author = {Thesis Research},
year = {2025},
publisher = {HuggingFace},
url = {https://huggingface.co/Nishef/Qwen3-0.6B-Full_KTO_20251225_102050}
}
Repository Structure
.
βββ adapter_config.json # LoRA configuration
βββ adapter_model.safetensors # Model weights
βββ tokenizer files # Tokenizer configuration
βββ eval_summary.csv # Evaluation results
βββ thesis_plots/ # Visualization assets
β βββ benchmark_results.png
β βββ training_loss.png
βββ README.md # This file
Acknowledgments
- Base Model: Qwen/Qwen3-0.6B
- Training Framework: Hugging Face Transformers
- Fine-tuning Library: PEFT
License
This model is released under the Apache 2.0 license.
This model was created as part of thesis research on LLM alignment using preference optimization methods.
