ZJU-Fangyin
commited on
Commit
•
e0d4260
1
Parent(s):
f30da1a
Update README.md
Browse files
README.md
CHANGED
@@ -15,7 +15,7 @@ With a training corpus of over 100 million molecules in SELFIES representation,
|
|
15 |
Specifically, MolGen-large employs a bidirectional Transformer as its encoder and an autoregressive Transformer as its decoder.
|
16 |
Through its carefully designed multi-task molecular prefix tuning (MPT), MolGen-large-opt can generate molecules with desired properties, making it a valuable tool for molecular optimization.
|
17 |
|
18 |
-
![image.png](./
|
19 |
|
20 |
## Intended uses
|
21 |
You can use the fine-tuned model for molecule optimization for downstream tasks. See the [repository](https://github.com/zjunlp/MolGen) to look for fine-tune details on a task that interests you.
|
|
|
15 |
Specifically, MolGen-large employs a bidirectional Transformer as its encoder and an autoregressive Transformer as its decoder.
|
16 |
Through its carefully designed multi-task molecular prefix tuning (MPT), MolGen-large-opt can generate molecules with desired properties, making it a valuable tool for molecular optimization.
|
17 |
|
18 |
+
![image.png](./molgen.png)
|
19 |
|
20 |
## Intended uses
|
21 |
You can use the fine-tuned model for molecule optimization for downstream tasks. See the [repository](https://github.com/zjunlp/MolGen) to look for fine-tune details on a task that interests you.
|