mtasic85 commited on
Commit
5a757e9
1 Parent(s): 3890e98

readme context length

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -25,7 +25,7 @@ tags:
25
 
26
  A pretrained language model based on the Llama model with about **33M** parameters. This model has been trained on **4.2B** (`4,252,334,823`) tokens from more than **6.2M** (`6,271,145`) dataset rows.
27
 
28
- This model **isn't** designed for immediate use but rather for Continued Pretraining and Finetuning on a downstream task. While it can handle a context length of up to **32K** (`32,768`) tokens, it was pretrained with sequences of **32K** (`32768`) tokens.
29
 
30
  The objective is to streamline the cognitive or reasoning core, eliminating any redundant knowledge from the model.
31
 
 
25
 
26
  A pretrained language model based on the Llama model with about **33M** parameters. This model has been trained on **4.2B** (`4,252,334,823`) tokens from more than **6.2M** (`6,271,145`) dataset rows.
27
 
28
+ This model **isn't** designed for immediate use but rather for Continued Pretraining and Finetuning on a downstream task. It can handle a context length of up to **32K** (`32,768`) tokens, and it was pretrained with sequences of **32K** (`32,768`) tokens.
29
 
30
  The objective is to streamline the cognitive or reasoning core, eliminating any redundant knowledge from the model.
31