Mol ID
A transformer encoder model pretrained on 50M ZINC SMILES string using flash attention 2
Hardware:
- gpu that support flash attention 2 and bf16
Software:
- flash attention 2
- lightning for mixed precision (bf16-mixed)
- wandb for logging
- huggingface
- tokenizers
- datasets
github repo: link
- Downloads last month
- 11
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support