feihu.hf
		
	commited on
		
		
					Commit 
							
							·
						
						4bbe56d
	
1
								Parent(s):
							
							1a90645
								
update README.md
Browse files
    	
        README.md
    CHANGED
    
    | @@ -20,12 +20,11 @@ tags: | |
| 20 |  | 
| 21 | 
             
            ## Introduction
         | 
| 22 |  | 
| 23 | 
            -
            Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (formerly known as CodeQwen). As of now, Qwen2.5-Coder has covered six mainstream model sizes, 0.5, 1.5, 3, 7, 14, 32 billion parameters, to meet the needs of different developers | 
| 24 |  | 
| 25 | 
             
            - Significantly improvements in **code generation**, **code reasoning** and **code fixing**. Base on the strong Qwen2.5, we scale up the training tokens into 5.5 trillion including source code, text-code grounding, Synthetic data, etc. Qwen2.5-Coder-32B has become the current state-of-the-art open-source codeLLM, with its coding abilities matching those of GPT-4o.
         | 
| 26 | 
             
            - A more comprehensive foundation for real-world applications such as **Code Agents**. Not only enhancing coding capabilities but also maintaining its strengths in mathematics and general competencies.
         | 
| 27 | 
             
            - **Long-context Support** up to 128K tokens.
         | 
| 28 | 
            -
            - 
         | 
| 29 |  | 
| 30 | 
             
            **This repo contains the instruction-tuned 7B Qwen2.5-Coder model**, which has the following features:
         | 
| 31 | 
             
            - Type: Causal Language Models
         | 
| @@ -122,10 +121,10 @@ If you find our work helpful, feel free to give us a cite. | |
| 122 |  | 
| 123 | 
             
            ```
         | 
| 124 | 
             
            @article{hui2024qwen2,
         | 
| 125 | 
            -
             | 
| 126 | 
            -
             | 
| 127 | 
            -
             | 
| 128 | 
            -
             | 
| 129 | 
             
            }
         | 
| 130 | 
             
            @article{qwen2,
         | 
| 131 | 
             
                  title={Qwen2 Technical Report}, 
         | 
|  | |
| 20 |  | 
| 21 | 
             
            ## Introduction
         | 
| 22 |  | 
| 23 | 
            +
            Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (formerly known as CodeQwen). As of now, Qwen2.5-Coder has covered six mainstream model sizes, 0.5, 1.5, 3, 7, 14, 32 billion parameters, to meet the needs of different developers. Qwen2.5-Coder brings the following improvements upon CodeQwen1.5:
         | 
| 24 |  | 
| 25 | 
             
            - Significantly improvements in **code generation**, **code reasoning** and **code fixing**. Base on the strong Qwen2.5, we scale up the training tokens into 5.5 trillion including source code, text-code grounding, Synthetic data, etc. Qwen2.5-Coder-32B has become the current state-of-the-art open-source codeLLM, with its coding abilities matching those of GPT-4o.
         | 
| 26 | 
             
            - A more comprehensive foundation for real-world applications such as **Code Agents**. Not only enhancing coding capabilities but also maintaining its strengths in mathematics and general competencies.
         | 
| 27 | 
             
            - **Long-context Support** up to 128K tokens.
         | 
|  | |
| 28 |  | 
| 29 | 
             
            **This repo contains the instruction-tuned 7B Qwen2.5-Coder model**, which has the following features:
         | 
| 30 | 
             
            - Type: Causal Language Models
         | 
|  | |
| 121 |  | 
| 122 | 
             
            ```
         | 
| 123 | 
             
            @article{hui2024qwen2,
         | 
| 124 | 
            +
                  title={Qwen2. 5-Coder Technical Report},
         | 
| 125 | 
            +
                  author={Hui, Binyuan and Yang, Jian and Cui, Zeyu and Yang, Jiaxi and Liu, Dayiheng and Zhang, Lei and Liu, Tianyu and Zhang, Jiajun and Yu, Bowen and Dang, Kai and others},
         | 
| 126 | 
            +
                  journal={arXiv preprint arXiv:2409.12186},
         | 
| 127 | 
            +
                  year={2024}
         | 
| 128 | 
             
            }
         | 
| 129 | 
             
            @article{qwen2,
         | 
| 130 | 
             
                  title={Qwen2 Technical Report}, 
         | 
