guoday commited on
Commit
505e2a1
1 Parent(s): c050327

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -21,7 +21,7 @@ Deepseek Coder comprises a series of code language models trained on both 87% co
21
 
22
 
23
  ### 2. Model Summary
24
- deepseek-coder-5.7b-instruct is a 5.7B parameter model initialized from deepseek-coder-5.7b-base and fine-tuned on 2B tokens of instruction data.
25
  - **Home Page:** [DeepSeek](https://deepseek.com/)
26
  - **Repository:** [deepseek-ai/deepseek-coder](https://github.com/deepseek-ai/deepseek-coder)
27
  - **Chat With DeepSeek Coder:** [DeepSeek-Coder](https://coder.deepseek.com/)
@@ -32,8 +32,8 @@ Here give some examples of how to use our model.
32
  #### Chat Model Inference
33
  ```python
34
  from transformers import AutoTokenizer, AutoModelForCausalLM
35
- tokenizer = AutoTokenizer.from_pretrained("deepseek-coder-5.7b-instruct", trust_remote_code=True)
36
- model = AutoModelForCausalLM.from_pretrained("deepseek-coder-5.7b-instruct", trust_remote_code=True).cuda()
37
  system_prompt = "You are an AI programming assistant, utilizing the Deepseek Coder model, developed by Deepseek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer.\n"
38
  messages=[
39
  { 'role': 'user', 'content': "write a quick sort algorithm in python."}
 
21
 
22
 
23
  ### 2. Model Summary
24
+ deepseek-coder-6.7b-instruct is a 6.7B parameter model initialized from deepseek-coder-6.7b-base and fine-tuned on 2B tokens of instruction data.
25
  - **Home Page:** [DeepSeek](https://deepseek.com/)
26
  - **Repository:** [deepseek-ai/deepseek-coder](https://github.com/deepseek-ai/deepseek-coder)
27
  - **Chat With DeepSeek Coder:** [DeepSeek-Coder](https://coder.deepseek.com/)
 
32
  #### Chat Model Inference
33
  ```python
34
  from transformers import AutoTokenizer, AutoModelForCausalLM
35
+ tokenizer = AutoTokenizer.from_pretrained("deepseek-coder-6.7b-instruct", trust_remote_code=True)
36
+ model = AutoModelForCausalLM.from_pretrained("deepseek-coder-6.7b-instruct", trust_remote_code=True).cuda()
37
  system_prompt = "You are an AI programming assistant, utilizing the Deepseek Coder model, developed by Deepseek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer.\n"
38
  messages=[
39
  { 'role': 'user', 'content': "write a quick sort algorithm in python."}