Intel LLM Leaderboard Submitted LoRA Model
Collection
LoRA Models
•
3 items
•
Updated
•
1
This model was fine-tuned from meta-llama/Meta-Llama-3-8B
The gopalakrishnan-d/meta-llama3-8b-alpaca-v1 model is a fine-tuned variant of the Llama3 architecture with 8 billion parameters. This version has been specifically enhanced for better performance on diverse language tasks, utilizing the Gaudi 2 Accelerator to optimize the training process.
- learning_rate: 5e-06 (Low Rate)
- train_batch_size: 8
- seed: 100
- gradient_accumulation_steps: 1
- optimizer: Adam
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.03
- lora_rank=16
- lora_alpha=32
Will be update..!