feihu.hf commited on
Commit
19a986b
•
1 Parent(s): d85b5e3

update README & LICENSE

Browse files
Files changed (1) hide show
  1. README.md +6 -5
README.md CHANGED
@@ -1,5 +1,6 @@
1
  ---
2
  license: apache-2.0
 
3
  language:
4
  - en
5
  base_model:
@@ -9,8 +10,8 @@ library_name: transformers
9
  tags:
10
  - code
11
  - qwen
12
- - codeqwen
13
  - qwen-coder
 
14
  ---
15
 
16
  # Qwen2.5-Coder-1.5B
@@ -23,6 +24,7 @@ Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (
23
  - A more comprehensive foundation for real-world applications such as **Code Agents**. Not only enhancing coding capabilities but also maintaining its strengths in mathematics and general competencies.
24
  - **Long-context Support** up to 128K tokens and can generate up to 8K tokens.
25
 
 
26
  **This repo contains the 1.5B Qwen2.5-Coder model**, which has the following features:
27
  - Type: Causal Language Models
28
  - Training Stage: Pretraining & Post-training
@@ -31,7 +33,7 @@ Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (
31
  - Number of Paramaters (Non-Embedding): 1.31B
32
  - Number of Layers: 28
33
  - Number of Attention Heads (GQA): 12 for Q and 2 for KV
34
- - Context Length: Full 32,768 tokens and generation 8192 tokens
35
 
36
  **We do not recommend using base language models for conversations.** Instead, you can apply post-training, e.g., SFT, RLHF, continued pretraining, etc., or fill in the middle tasks on this model.
37
 
@@ -46,6 +48,7 @@ With `transformers<4.37.0`, you will encounter the following error:
46
  KeyError: 'qwen2'
47
  ```
48
 
 
49
  ## Evaluation & Performance
50
 
51
  Detailed evaluation results are reported in this [📑 blog](https://qwenlm.github.io/blog/qwen2.5-coder/).
@@ -70,6 +73,4 @@ If you find our work helpful, feel free to give us a cite.
70
  journal={arXiv preprint arXiv:2407.10671},
71
  year={2024}
72
  }
73
- ```
74
-
75
-
 
1
  ---
2
  license: apache-2.0
3
+ license_link: https://huggingface.co/Qwen/Qwen2.5-Coder-1.5B/blob/main/LICENSE
4
  language:
5
  - en
6
  base_model:
 
10
  tags:
11
  - code
12
  - qwen
 
13
  - qwen-coder
14
+ - codeqwen
15
  ---
16
 
17
  # Qwen2.5-Coder-1.5B
 
24
  - A more comprehensive foundation for real-world applications such as **Code Agents**. Not only enhancing coding capabilities but also maintaining its strengths in mathematics and general competencies.
25
  - **Long-context Support** up to 128K tokens and can generate up to 8K tokens.
26
 
27
+
28
  **This repo contains the 1.5B Qwen2.5-Coder model**, which has the following features:
29
  - Type: Causal Language Models
30
  - Training Stage: Pretraining & Post-training
 
33
  - Number of Paramaters (Non-Embedding): 1.31B
34
  - Number of Layers: 28
35
  - Number of Attention Heads (GQA): 12 for Q and 2 for KV
36
+ - Context Length: Full 32,768 tokens
37
 
38
  **We do not recommend using base language models for conversations.** Instead, you can apply post-training, e.g., SFT, RLHF, continued pretraining, etc., or fill in the middle tasks on this model.
39
 
 
48
  KeyError: 'qwen2'
49
  ```
50
 
51
+
52
  ## Evaluation & Performance
53
 
54
  Detailed evaluation results are reported in this [📑 blog](https://qwenlm.github.io/blog/qwen2.5-coder/).
 
73
  journal={arXiv preprint arXiv:2407.10671},
74
  year={2024}
75
  }
76
+ ```