Any one who use the script in the Model Card for inference purpose?
#64
by
disper84
- opened
the tokenizer you're using is not fixed correctly, try checking for PRs in the community tab here to see if someone has provided it
Hi
@disper84
, Could you please try again and let us know if the issue still persists? Please try changing the max_length
by setting some increased number, outputs = model.generate(**input_ids, max_length=200)
to control the output generation length as the above warning states.
Also, Pretrained (PT) versions of the model are not trained on any specific tasks or instructions beyond the Gemma core data training set whcih might cause inconsistency in the output. You should not deploy these models in applications or use for inference without performing some tuning.