SelectiveDPO
Collection
Released models trained by Selective DPO. • 5 items • Updated
This model is fine-tuned from the HuggingFaceH4/mistral-7b-sft-beta model using SelectiveDPO on the Ultrafeedback_binarized dataset.
For the recipe to reproduce this model, please visit our GitHub page.
Base model
mistralai/Mistral-7B-v0.1