Update config.json to accurately reflect the 32k context window.
#73
by
Kearm
- opened
Replace the config.json to reflect the lack of a sliding context window, have the context window accurately reflect the 32k context window, and update to the latest transformers version.
The original config is correct.
The original config is correct.
Then why can't I use all 32k context? I have SFT'ed the Base 7b 0.2 and it will summarize 29k Mistral tokens of text without issue. This one wil not.