legolasyiu commited on
Commit
5dc0892
1 Parent(s): a6da892

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -9,6 +9,8 @@ tags:
9
  - unsloth
10
  - llama
11
  - trl
 
 
12
  ---
13
  <img src="https://huggingface.co/EpistemeAI/Fireball-Llama-3.1-8B-v1dpo/resolve/main/fireball-llama.JPG" width="200"/>
14
 
@@ -104,7 +106,7 @@ Where to send questions or comments about the model Instructions on how to provi
104
 
105
  ## Training
106
  **KTO Fine tuning**:
107
- Experimental: KTO fine tuning
108
 
109
  KTO - Kahneman-Tversky Optimization (KTO) that makes it easier and cheaper than ever before to align LLMs on your data without compromising performance
110
 
@@ -225,4 +227,4 @@ But Llama 3.1 is a new technology, and like any new technology, there are risks
225
 
226
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
227
 
228
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
9
  - unsloth
10
  - llama
11
  - trl
12
+ datasets:
13
+ - argilla/distilabel-intel-orca-kto
14
  ---
15
  <img src="https://huggingface.co/EpistemeAI/Fireball-Llama-3.1-8B-v1dpo/resolve/main/fireball-llama.JPG" width="200"/>
16
 
 
106
 
107
  ## Training
108
  **KTO Fine tuning**:
109
+ Experimental: KTO fine tuning with dataset- argilla/distilabel-intel-orca-kto
110
 
111
  KTO - Kahneman-Tversky Optimization (KTO) that makes it easier and cheaper than ever before to align LLMs on your data without compromising performance
112
 
 
227
 
228
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
229
 
230
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)