roberta_zinc_480m / README.md
entropy's picture
Update README.md
40a96d5
|
raw
history blame
369 Bytes
# Roberta Zinc 480m
This is a Roberta style masked language model trained on ~480m SMILES strings from the [ZINC database](https://zinc.docking.org/).
The model has ~102m parameters and was trained for 150000 iterations with a batch size of 4096 to a validation loss of ~0.122.
This model is useful for generating embeddings from SMILES strings.
---
license: mit
---