TravelLeraLone
commited on
Commit
•
44e0761
1
Parent(s):
811dc4c
release ChemDFM-v1.5-8B
Browse files
README.md
CHANGED
@@ -15,9 +15,10 @@ ChemDFM is the pioneering open-sourced dialogue foundation model for Chemistry a
|
|
15 |
|
16 |
## News
|
17 |
|
|
|
18 |
* **2024-06-13**: The results on the comprehensive science benchmark [SciKnowEval](https://huggingface.co/datasets/hicai-zju/SciKnowEval) show that "ChemDFM emerged as one of the top open-source models by continuing pre-training and fine-tuning on a vast corpus of scientific literature".
|
19 |
* **2024-04-17**: The evaluation data (including instructions) we used in our paper is released on [GitHub](https://github.com/OpenDFM/ChemDFM)
|
20 |
-
* **2024-03-12**: The parameter of ChemDFM-13B is open-sourced!
|
21 |
* **2024-01-26**: The paper of ChemDFM-13B is released on arXiv: [ChemDFM: Dialogue Foundation Model for Chemistry](https://arxiv.org/abs/2401.14818)
|
22 |
|
23 |
## Usage Details
|
@@ -32,7 +33,7 @@ To load and run ChemDFM locally, here is an example:
|
|
32 |
import torch
|
33 |
from transformers import LlamaTokenizer, LlamaForCausalLM, GenerationConfig
|
34 |
|
35 |
-
model_name_or_id = "
|
36 |
tokenizer = LlamaTokenizer.from_pretrained(model_name_or_id)
|
37 |
model = LlamaForCausalLM.from_pretrained(model_name_or_id, torch_dtype=torch.float16, device_map="auto")
|
38 |
|
|
|
15 |
|
16 |
## News
|
17 |
|
18 |
+
* 2024-11-09: [ChemDFM-v1.5-8B](https://huggingface.co/OpenDFM/ChemDFM-v1.5-8B) is released! We implemented our domain pre-training and instruction tuning precedure on a stronger base model LLaMA-3-8B.
|
19 |
* **2024-06-13**: The results on the comprehensive science benchmark [SciKnowEval](https://huggingface.co/datasets/hicai-zju/SciKnowEval) show that "ChemDFM emerged as one of the top open-source models by continuing pre-training and fine-tuning on a vast corpus of scientific literature".
|
20 |
* **2024-04-17**: The evaluation data (including instructions) we used in our paper is released on [GitHub](https://github.com/OpenDFM/ChemDFM)
|
21 |
+
* **2024-03-12**: The parameter of [ChemDFM-13B](https://huggingface.co/OpenDFM/ChemDFM-13B-v1.0) is open-sourced!
|
22 |
* **2024-01-26**: The paper of ChemDFM-13B is released on arXiv: [ChemDFM: Dialogue Foundation Model for Chemistry](https://arxiv.org/abs/2401.14818)
|
23 |
|
24 |
## Usage Details
|
|
|
33 |
import torch
|
34 |
from transformers import LlamaTokenizer, LlamaForCausalLM, GenerationConfig
|
35 |
|
36 |
+
model_name_or_id = "OpenDFM/ChemDFM-13B-v1.0"
|
37 |
tokenizer = LlamaTokenizer.from_pretrained(model_name_or_id)
|
38 |
model = LlamaForCausalLM.from_pretrained(model_name_or_id, torch_dtype=torch.float16, device_map="auto")
|
39 |
|