timpal0l commited on
Commit
5b07b5c
1 Parent(s): c028d91

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -0
README.md CHANGED
@@ -24,6 +24,10 @@ widget:
24
  Llama-3-8B-flashback-v1 is a continuation of the pretraining process for the base meta-llama/Meta-Llama-3-8B model, utilizing 2 251 233 forum threads from the Swedish website https://www.flashback.org/. Which is rougly 40GB of text.
25
  It is a full finetune for one epoch.
26
 
 
 
 
 
27
  ## How to use:
28
  ```python
29
  import torch
 
24
  Llama-3-8B-flashback-v1 is a continuation of the pretraining process for the base meta-llama/Meta-Llama-3-8B model, utilizing 2 251 233 forum threads from the Swedish website https://www.flashback.org/. Which is rougly 40GB of text.
25
  It is a full finetune for one epoch.
26
 
27
+ ```html
28
+ * tokens: 8_271_170_155
29
+ * checkpoint 4/5
30
+ ```
31
  ## How to use:
32
  ```python
33
  import torch