Qwen3-0.6B - Kto

Benchmark Results

Model Description

This model is a LoRA Adapter fine-tuned from Qwen/Qwen3-0.6B using the Kto training method.

Kahneman-Tversky Optimization - Binary preference optimization based on Prospect Theory

This model was developed as part of thesis research on LLM Alignment using Preference Optimization Methods.

Model Details

Property Value
Base Model Qwen/Qwen3-0.6B
Training Method Kto
Model Type LoRA Adapter
Training Date December 2025
Framework PyTorch + Transformers + PEFT

Benchmark Results

Benchmark Score
HellaSwag (10-shot) 0.264
TruthfulQA (0-shot MC2) 0.486
MMLU-Mini (5-shot) 0.269

Comparative Analysis

The following chart compares this method against other training approaches on the same base model:

Training Loss Curves

Training Configuration

Parameter Value
Epochs 1
Batch Size 2
Gradient Accumulation 8
Effective Batch Size 16
Learning Rate 2e-4
Max Sequence Length 512
LoRA Rank 16
LoRA Alpha 32
Dataset UltraFeedback Binarized

Usage

Loading as LoRA Adapter

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

# Load base model
base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen3-0.6B")
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen3-0.6B")

# Load adapter
model = PeftModel.from_pretrained(base_model, "Nishef/Qwen3-0.6B-Full_KTO_20251225_102050")

# Generate text
inputs = tokenizer("Hello, how are you?", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0]))

Training Methodology

Kto

Kahneman-Tversky Optimization - Binary preference optimization based on Prospect Theory

Key Features:

  • Binary feedback signals (thumbs up/down)
  • No need for paired preference data
  • Reference model for KL divergence regularization
  • Prospect Theory-inspired loss function

Citation

If you use this model in your research, please cite:

@misc{qwen3_0.6b_kto_2025,
  title = {Qwen3-0.6B Fine-tuned with Kto},
  author = {Thesis Research},
  year = {2025},
  publisher = {HuggingFace},
  url = {https://huggingface.co/Nishef/Qwen3-0.6B-Full_KTO_20251225_102050}
}

Repository Structure

.
β”œβ”€β”€ adapter_config.json      # LoRA configuration
β”œβ”€β”€ adapter_model.safetensors # Model weights
β”œβ”€β”€ tokenizer files          # Tokenizer configuration
β”œβ”€β”€ eval_summary.csv         # Evaluation results
β”œβ”€β”€ thesis_plots/            # Visualization assets
β”‚   β”œβ”€β”€ benchmark_results.png
β”‚   └── training_loss.png
└── README.md               # This file

Acknowledgments

License

This model is released under the Apache 2.0 license.


This model was created as part of thesis research on LLM alignment using preference optimization methods.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Nishef/Qwen3-0.6B-Full_KTO_20251225_102050

Finetuned
Qwen/Qwen3-0.6B
Finetuned
(544)
this model

Dataset used to train Nishef/Qwen3-0.6B-Full_KTO_20251225_102050