Update README.md
Browse files
README.md
CHANGED
@@ -88,5 +88,5 @@ There is also some issue with handling long system messages for RP, I was planni
|
|
88 |
|
89 |
I will probably be working on de-contaminating base Yi-34B model now. \
|
90 |
My second run of AEZAKMI v2 fine-tune was just 0.15 epochs and I really like how natural this model is and how rich is it's vocabulary. I will try to train less to hit the sweetspot. \
|
91 |
-
I will be uploading LoRA adapter for that second run that was just 0.15 epochs.
|
92 |
I believe that I might have gotten what I want if I would have stopped training sooner. I don't have checkpoints older than 1500 steps back so I would need to re-run training to get it back.
|
|
|
88 |
|
89 |
I will probably be working on de-contaminating base Yi-34B model now. \
|
90 |
My second run of AEZAKMI v2 fine-tune was just 0.15 epochs and I really like how natural this model is and how rich is it's vocabulary. I will try to train less to hit the sweetspot. \
|
91 |
+
I will be uploading LoRA adapter for that second run that was just 0.15 epochs. \
|
92 |
I believe that I might have gotten what I want if I would have stopped training sooner. I don't have checkpoints older than 1500 steps back so I would need to re-run training to get it back.
|