File size: 443 Bytes
d4f603a
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
---
license: mit
---

# This Model 
This is a partially continued pretrained Llama 3.1 8B LLM (using unsloth/Meta-Llama-3.1-8B). Training was done on [200k articles from Arabic Wikipedia 2023](akhooli/arwiki_128).  
This is just a proof of concept demo and should never be used for production. 
The model was then used to for instruction fine-tuning for classical Arabic poetry generation (toy model: https://huggingface.co/akhooli/llama31ft).