|
--- |
|
license: llama3 |
|
--- |
|
|
|
# Weights from the Llama-3-8B Self-Align Experiments |
|
|
|
[WEIGHTS TO BE UPLOADED ONCE DONE] |
|
|
|
## Training Config |
|
|
|
The `config.yaml` should be used during `accelerate launch`, and `run.sh` was used to launch the training using the [StarCoder2 Self-Align training script](https://github.com/bigcode-project/starcoder2-self-align?tab=readme-ov-file#training-details). |
|
Some tweaks were performed to get this working on 48GB vRAM: |
|
- FSDP was used |
|
- `per_device_batch_size` is `2` |
|
- A learning rate of 3e-6 was used |
|
|
|
|
|
## Environment: |
|
|
|
- Trained with 2x4090 GPUs |
|
- 128GB RAM |