FremyCompany
commited on
Commit
•
6a08784
1
Parent(s):
224f136
Update README.md
Browse files
README.md
CHANGED
@@ -41,4 +41,19 @@ Training utilized Nvidia H100 and A100 GPUs. Inference is accessible on lower-en
|
|
41 |
### Model Weights
|
42 |
|
43 |
- This model was trained in bfloat16.
|
44 |
-
- [GGUF weights](https://huggingface.co/BramVanroy/tweety-7b-dutch-v24a-GGUF) are released by Bram Vanroy.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
41 |
### Model Weights
|
42 |
|
43 |
- This model was trained in bfloat16.
|
44 |
+
- [GGUF weights](https://huggingface.co/BramVanroy/tweety-7b-dutch-v24a-GGUF) are released by Bram Vanroy.
|
45 |
+
|
46 |
+
|
47 |
+
## Citation
|
48 |
+
|
49 |
+
If you use this model, please cite our work as:
|
50 |
+
|
51 |
+
```
|
52 |
+
@article{tweeties2024,
|
53 |
+
title = {Trans-Tokenization and Cross-lingual Vocabulary Transfers: Language Adaptation of LLMs for Low-Resource NLP},
|
54 |
+
author = {François Remy and Pieter Delobelle and Hayastan Avetisyan and Alfiya Khabibullina and Miryam de Lhoneux and Thomas Demeester},
|
55 |
+
url = {https://arxiv.org/abs/2408.04303},
|
56 |
+
year = {2024},
|
57 |
+
note = {Accepted at COLM 2024}
|
58 |
+
}
|
59 |
+
```
|