--- license: apache-2.0 datasets: - timdettmers/openassistant-guanaco pipeline_tag: text-generation library_name: transformers --- This is a Llama-2-7b-chat-hf model fine-tuned using QLoRA (4-bit precision) on the mlabonne/guanaco-llama2-1k dataset(a subset of OpenAssistant/oasst1). This model was trained on a single A100. ``` {'train_runtime': 2368.6484, 'train_samples_per_second': 4.157, 'train_steps_per_second': 1.039, 'train_loss': 1.3207073250405679, 'epoch': 1.0} ``` This model was finetuned primarily for educational purposes.