Why is the special token for padding `A` in `special_tokens_map.json`?
#3
by
ardhe-qb
- opened
Hi,
Why is the special token for padding A
in special_tokens_map.json
? It seems like the padding token is <pad>
in the vocabulary, in the default instantiation of EsmTokenizer
, etc.
Is there a reason for this? It seems strange to add random Alanine's to pad a protein so that the whole batch is of the same size.
Thanks in advance for your reply!