How to increase or decrease the context length?
#9
by
CouchCommander
- opened
I am trying to run my model with dynamic context length but it is not specified how to configure it.
Hi
@CouchCommander
, You can adjust the context length of the inputs using max_length
or max_new_tokens
in outputs = model.generate(**input_ids, max_length=100)
or while instantiating the model.