Home

Awesome

ORPO

Updates (24.03.25)

 

This is the official repository for <a class="link" href="https://arxiv.org/abs/2403.07691">ORPO: Monolithic Preference Optimization without Reference Model</a>. The detailed results in the paper can be found in:

Model Checkpoints

Our models trained with ORPO can be found in:

And the corresponding logs for the average log probabilities of chosen/rejected responses during training are reported in:

 

AlpacaEval

<figure> <img class="png" src="/assets/img/alpaca_blog.png" alt="Description of the image"> <figcaption><b>Figure 1.</b> AlpacaEval 2.0 score for the models trained with different alignment methods.</figcaption> </figure>

 

MT-Bench

<figure> <img class="png" src="/assets/img/mtbench_hf.png" alt="Description of the image"> <figcaption><b>Figure 2.</b> MT-Bench result by category.</figcaption> </figure>

 

IFEval

IFEval scores are measured with <a class="link" href="https://github.com/EleutherAI/lm-evaluation-harness">EleutherAI/lm-evaluation-harness</a> by applying the chat template. The scores for Llama-2-Chat (70B), Zephyr-β (7B), and Mixtral-8X7B-Instruct-v0.1 are originally reported in <a class="link" href="https://twitter.com/wiskojo/status/1739767758462877823">this tweet</a>.

Model TypePrompt-StrictPrompt-LooseInst-StrictInst-Loose
Llama-2-Chat (70B)0.44360.53420.54680.6319
Zephyr-β (7B)0.42330.45470.54920.5767
Mixtral-8X7B-Instruct-v0.10.52130.57120.63430.6823
Mistral-ORPO-⍺ (7B)0.50090.50830.59950.6163
Mistral-ORPO-β (7B)0.52870.55640.63550.6619