angelahzyuan
commited on
Commit
•
5e37a0b
1
Parent(s):
9b2b8e0
Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,7 @@ Self-Play Preference Optimization for Language Model Alignment (https://arxiv.or
|
|
10 |
|
11 |
# Mistral7B-PairRM-SPPO-Iter1
|
12 |
|
13 |
-
This model was developed using [Self-Play Preference Optimization](https://arxiv.org/abs/2405.00675) at iteration
|
14 |
|
15 |
**This is the model reported in the paper** , with K=5 (generate 5 responses per iteration). We attached the Arena-Hard eval results in this model page.
|
16 |
|
|
|
10 |
|
11 |
# Mistral7B-PairRM-SPPO-Iter1
|
12 |
|
13 |
+
This model was developed using [Self-Play Preference Optimization](https://arxiv.org/abs/2405.00675) at iteration 2, based on the [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) architecture as starting point. We utilized the prompt sets from the [openbmb/UltraFeedback](https://huggingface.co/datasets/openbmb/UltraFeedback) dataset, splited to 3 parts for 3 iterations by [snorkelai/Snorkel-Mistral-PairRM-DPO-Dataset](https://huggingface.co/datasets/snorkelai/Snorkel-Mistral-PairRM-DPO-Dataset). All responses used are synthetic.
|
14 |
|
15 |
**This is the model reported in the paper** , with K=5 (generate 5 responses per iteration). We attached the Arena-Hard eval results in this model page.
|
16 |
|