Update README.md
Browse files
README.md
CHANGED
@@ -17,7 +17,7 @@ We present Meltemi 7B Instruct v1.5 Large Language Model (LLM), a new and improv
|
|
17 |
|
18 |
# Model Information
|
19 |
|
20 |
-
- Vocabulary extension of the Mistral 7b tokenizer with Greek tokens
|
21 |
- 8192 context length
|
22 |
- Fine-tuning has been done with the [Odds Ratio Preference Optimization (ORPO)](https://arxiv.org/abs/2403.07691) algorithm using 97k preference data:
|
23 |
* 89,730 Greek preference data which are mostly translated versions of high-quality datasets on Hugging Face
|
|
|
17 |
|
18 |
# Model Information
|
19 |
|
20 |
+
- Vocabulary extension of the Mistral 7b tokenizer with Greek tokens for lower costs and faster inference (**1.52** vs. 6.80 tokens/word for Greek)
|
21 |
- 8192 context length
|
22 |
- Fine-tuning has been done with the [Odds Ratio Preference Optimization (ORPO)](https://arxiv.org/abs/2403.07691) algorithm using 97k preference data:
|
23 |
* 89,730 Greek preference data which are mostly translated versions of high-quality datasets on Hugging Face
|