Update README.md
Browse files
README.md
CHANGED
|
@@ -26,6 +26,8 @@ Paper: [Arxiv Link](https://github.com/IBM/materials/blob/main/smi-ted/paper/smi
|
|
| 26 |
|
| 27 |
For more information contact: [email protected] or [email protected].
|
| 28 |
|
|
|
|
|
|
|
| 29 |
## Introduction
|
| 30 |
|
| 31 |
We present a large encoder-decoder chemical foundation model, SMILES-based Transformer Encoder-Decoder (SMI-TED), pre-trained on a curated dataset of 91 million SMILES samples sourced from PubChem, equivalent to 4 billion molecular tokens. SMI-TED supports various complex tasks, including quantum property prediction, with two main variants ($289M$ and $8 \times 289M$). Our experiments across multiple benchmark datasets demonstrate state-of-the-art performance for various tasks. For more information contact: [email protected] or [email protected].
|
|
|
|
| 26 |
|
| 27 |
For more information contact: [email protected] or [email protected].
|
| 28 |
|
| 29 |
+

|
| 30 |
+
|
| 31 |
## Introduction
|
| 32 |
|
| 33 |
We present a large encoder-decoder chemical foundation model, SMILES-based Transformer Encoder-Decoder (SMI-TED), pre-trained on a curated dataset of 91 million SMILES samples sourced from PubChem, equivalent to 4 billion molecular tokens. SMI-TED supports various complex tasks, including quantum property prediction, with two main variants ($289M$ and $8 \times 289M$). Our experiments across multiple benchmark datasets demonstrate state-of-the-art performance for various tasks. For more information contact: [email protected] or [email protected].
|