File size: 369 Bytes
40a96d5
 
 
 
 
 
03c6d05
 
 
1
2
3
4
5
6
7
8
9
10
# Roberta Zinc 480m

This is a Roberta style masked language model trained on ~480m SMILES strings from the [ZINC database](https://zinc.docking.org/).
The model has ~102m parameters and was trained for 150000 iterations with a batch size of 4096 to a validation loss of ~0.122.
This model is useful for generating embeddings from SMILES strings.

---
license: mit
---