eduardosoares99 commited on
Commit
4c21438
1 Parent(s): d20b340

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -22,6 +22,8 @@ GitHub: [GitHub Link](https://github.com/IBM/materials/tree/main)
22
 
23
  # SMILES-based Transformer Encoder-Decoder (SMI-TED)
24
 
 
 
25
  This repository provides PyTorch source code associated with our publication, "A Large Encoder-Decoder Family of Foundation Models for Chemical Language".
26
 
27
  Paper: [Arxiv Link](https://github.com/IBM/materials/blob/main/smi-ted/paper/smi_ted_preprint.pdf)
@@ -33,8 +35,6 @@ We provide the model weights in two formats:
33
 
34
  For more information contact: [email protected] or [email protected].
35
 
36
- ![ted-smi](smi-ted.png)
37
-
38
  ## Introduction
39
 
40
  We present a large encoder-decoder chemical foundation model, SMILES-based Transformer Encoder-Decoder (SMI-TED), pre-trained on a curated dataset of 91 million SMILES samples sourced from PubChem, equivalent to 4 billion molecular tokens. SMI-TED supports various complex tasks, including quantum property prediction, with two main variants (289M and 8X289M). Our experiments across multiple benchmark datasets demonstrate state-of-the-art performance for various tasks. For more information contact: [email protected] or [email protected].
 
22
 
23
  # SMILES-based Transformer Encoder-Decoder (SMI-TED)
24
 
25
+ ![ted-smi](smi-ted.png)
26
+
27
  This repository provides PyTorch source code associated with our publication, "A Large Encoder-Decoder Family of Foundation Models for Chemical Language".
28
 
29
  Paper: [Arxiv Link](https://github.com/IBM/materials/blob/main/smi-ted/paper/smi_ted_preprint.pdf)
 
35
 
36
  For more information contact: [email protected] or [email protected].
37
 
 
 
38
  ## Introduction
39
 
40
  We present a large encoder-decoder chemical foundation model, SMILES-based Transformer Encoder-Decoder (SMI-TED), pre-trained on a curated dataset of 91 million SMILES samples sourced from PubChem, equivalent to 4 billion molecular tokens. SMI-TED supports various complex tasks, including quantum property prediction, with two main variants (289M and 8X289M). Our experiments across multiple benchmark datasets demonstrate state-of-the-art performance for various tasks. For more information contact: [email protected] or [email protected].