eduardosoares99
commited on
Commit
•
4c21438
1
Parent(s):
d20b340
Update README.md
Browse files
README.md
CHANGED
@@ -22,6 +22,8 @@ GitHub: [GitHub Link](https://github.com/IBM/materials/tree/main)
|
|
22 |
|
23 |
# SMILES-based Transformer Encoder-Decoder (SMI-TED)
|
24 |
|
|
|
|
|
25 |
This repository provides PyTorch source code associated with our publication, "A Large Encoder-Decoder Family of Foundation Models for Chemical Language".
|
26 |
|
27 |
Paper: [Arxiv Link](https://github.com/IBM/materials/blob/main/smi-ted/paper/smi_ted_preprint.pdf)
|
@@ -33,8 +35,6 @@ We provide the model weights in two formats:
|
|
33 |
|
34 |
For more information contact: [email protected] or [email protected].
|
35 |
|
36 |
-
![ted-smi](smi-ted.png)
|
37 |
-
|
38 |
## Introduction
|
39 |
|
40 |
We present a large encoder-decoder chemical foundation model, SMILES-based Transformer Encoder-Decoder (SMI-TED), pre-trained on a curated dataset of 91 million SMILES samples sourced from PubChem, equivalent to 4 billion molecular tokens. SMI-TED supports various complex tasks, including quantum property prediction, with two main variants (289M and 8X289M). Our experiments across multiple benchmark datasets demonstrate state-of-the-art performance for various tasks. For more information contact: [email protected] or [email protected].
|
|
|
22 |
|
23 |
# SMILES-based Transformer Encoder-Decoder (SMI-TED)
|
24 |
|
25 |
+
![ted-smi](smi-ted.png)
|
26 |
+
|
27 |
This repository provides PyTorch source code associated with our publication, "A Large Encoder-Decoder Family of Foundation Models for Chemical Language".
|
28 |
|
29 |
Paper: [Arxiv Link](https://github.com/IBM/materials/blob/main/smi-ted/paper/smi_ted_preprint.pdf)
|
|
|
35 |
|
36 |
For more information contact: [email protected] or [email protected].
|
37 |
|
|
|
|
|
38 |
## Introduction
|
39 |
|
40 |
We present a large encoder-decoder chemical foundation model, SMILES-based Transformer Encoder-Decoder (SMI-TED), pre-trained on a curated dataset of 91 million SMILES samples sourced from PubChem, equivalent to 4 billion molecular tokens. SMI-TED supports various complex tasks, including quantum property prediction, with two main variants (289M and 8X289M). Our experiments across multiple benchmark datasets demonstrate state-of-the-art performance for various tasks. For more information contact: [email protected] or [email protected].
|