--- license: apache-2.0 base_model: alignment-handbook/zephyr-7b-sft-full tags: - generated_from_trainer model-index: - name: spin-v-high-loss results: [] --- # spin-v-high-loss This model is a fine-tuned version of [alignment-handbook/zephyr-7b-sft-full](https://huggingface.co/alignment-handbook/zephyr-7b-sft-full) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0069 - Rewards/real: -10.1415 - Rewards/generated: -55.1541 - Rewards/accuracies: 1.0 - Rewards/margins: 45.0126 - Logps/generated: -5640.6729 - Logps/real: -1151.2217 - Logits/generated: 3.0744 - Logits/real: 1.9177 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-07 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 4 - total_train_batch_size: 32 - total_eval_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rewards/real | Rewards/generated | Rewards/accuracies | Rewards/margins | Logps/generated | Logps/real | Logits/generated | Logits/real | |:-------------:|:-----:|:----:|:---------------:|:------------:|:-----------------:|:------------------:|:---------------:|:---------------:|:----------:|:----------------:|:-----------:| | 0.0717 | 0.13 | 50 | 0.0490 | -3.1258 | -37.3431 | 0.9907 | 34.2173 | -3859.5708 | -449.6532 | 3.4831 | 1.9303 | | 0.0323 | 0.27 | 100 | 0.0300 | -3.9959 | -38.8380 | 0.9973 | 34.8421 | -4009.0552 | -536.6592 | -0.0155 | -0.1626 | | 0.026 | 0.4 | 150 | 0.0158 | -8.2107 | -50.0493 | 0.9947 | 41.8386 | -5130.1880 | -958.1443 | 1.0207 | 1.0071 | | 0.0106 | 0.53 | 200 | 0.0087 | -9.2505 | -61.7325 | 0.9960 | 52.4820 | -6298.5093 | -1062.1265 | 2.2349 | 1.2992 | | 0.0071 | 0.67 | 250 | 0.0106 | -11.4051 | -49.3118 | 0.9987 | 37.9067 | -5056.4409 | -1277.5874 | 2.8798 | 3.2925 | | 0.0121 | 0.8 | 300 | 0.0074 | -9.0224 | -49.1152 | 1.0 | 40.0928 | -5036.7827 | -1039.3110 | 2.8713 | 2.6792 | | 0.0013 | 0.93 | 350 | 0.0069 | -10.1415 | -55.1541 | 1.0 | 45.0126 | -5640.6729 | -1151.2217 | 3.0744 | 1.9177 | ### Framework versions - Transformers 4.37.0 - Pytorch 2.1.2+cu121 - Datasets 2.14.6 - Tokenizers 0.15.2