feihu.hf commited on
Commit
83c0546
•
1 Parent(s): e3b2348

update README & LICENSE

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -1,5 +1,6 @@
1
  ---
2
  license: apache-2.0
 
3
  language:
4
  - en
5
  base_model:
@@ -25,7 +26,7 @@ Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (
25
  - A more comprehensive foundation for real-world applications such as **Code Agents**. Not only enhancing coding capabilities but also maintaining its strengths in mathematics and general competencies.
26
  - **Long-context Support** up to 128K tokens and can generate up to 8K tokens.
27
 
28
- **This repo contains the 1.5B Qwen2.5-Coder model**, which has the following features:
29
  - Type: Causal Language Models
30
  - Training Stage: Pretraining & Post-training
31
  - Architecture: transformers with RoPE, SwiGLU, RMSNorm, Attention QKV bias and tied word embeddings
@@ -33,7 +34,7 @@ Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (
33
  - Number of Paramaters (Non-Embedding): 1.31B
34
  - Number of Layers: 28
35
  - Number of Attention Heads (GQA): 12 for Q and 2 for KV
36
- - Context Length: Full 32,768 tokens and generation 8192 tokens
37
 
38
  For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2.5-coder/), [GitHub](https://github.com/QwenLM/Qwen2.5-Coder), and [Documentation](https://qwen.readthedocs.io/en/latest/).
39
 
@@ -85,6 +86,7 @@ generated_ids = [
85
  response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
86
  ```
87
 
 
88
  ## Evaluation & Performance
89
 
90
  Detailed evaluation results are reported in this [📑 blog](https://qwenlm.github.io/blog/qwen2.5-coder/).
 
1
  ---
2
  license: apache-2.0
3
+ license_link: https://huggingface.co/Qwen/Qwen2.5-Coder-1.5B-Instruct/blob/main/LICENSE
4
  language:
5
  - en
6
  base_model:
 
26
  - A more comprehensive foundation for real-world applications such as **Code Agents**. Not only enhancing coding capabilities but also maintaining its strengths in mathematics and general competencies.
27
  - **Long-context Support** up to 128K tokens and can generate up to 8K tokens.
28
 
29
+ **This repo contains the instruction-tuned 1.5B Qwen2.5-Coder model**, which has the following features:
30
  - Type: Causal Language Models
31
  - Training Stage: Pretraining & Post-training
32
  - Architecture: transformers with RoPE, SwiGLU, RMSNorm, Attention QKV bias and tied word embeddings
 
34
  - Number of Paramaters (Non-Embedding): 1.31B
35
  - Number of Layers: 28
36
  - Number of Attention Heads (GQA): 12 for Q and 2 for KV
37
+ - Context Length: Full 32,768 tokens
38
 
39
  For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2.5-coder/), [GitHub](https://github.com/QwenLM/Qwen2.5-Coder), and [Documentation](https://qwen.readthedocs.io/en/latest/).
40
 
 
86
  response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
87
  ```
88
 
89
+
90
  ## Evaluation & Performance
91
 
92
  Detailed evaluation results are reported in this [📑 blog](https://qwenlm.github.io/blog/qwen2.5-coder/).