roberta_zinc_480m / README.md
entropy's picture
Update README.md
40a96d5
|
raw
history blame
No virus
369 Bytes

Roberta Zinc 480m

This is a Roberta style masked language model trained on ~480m SMILES strings from the ZINC database. The model has ~102m parameters and was trained for 150000 iterations with a batch size of 4096 to a validation loss of ~0.122. This model is useful for generating embeddings from SMILES strings.


license: mit