Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,47 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
tags:
|
| 4 |
+
- code generation
|
| 5 |
+
base_model: internlm/AlchemistCoder-L-7B
|
| 6 |
+
pipeline_tag: text-generation
|
| 7 |
+
---
|
| 8 |
+
# QuantFactory/AlchemistCoder-L-7B-GGUF
|
| 9 |
+
This is quantized version of [internlm/AlchemistCoder-L-7B](https://huggingface.co/internlm/AlchemistCoder-L-7B) created using llama.cpp
|
| 10 |
+
|
| 11 |
+
## Model Description: AlchemistCoder
|
| 12 |
+
|
| 13 |
+
[[π Paper](https://arxiv.org/abs/2405.19265)]
|
| 14 |
+
[[π Project Page](https://internlm.github.io/AlchemistCoder/)]
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
## β¨ Highlights
|
| 18 |
+
> **Abstract:** *Open-source Large Language Models (LLMs) and their specialized variants, particularly Code LLMs, have recently delivered impressive performance. However, previous Code LLMs are typically fine-tuned on single-source data with limited quality and diversity, which may insufficiently elicit the potential of pre-trained Code LLMs. In this paper, we present AlchemistCoder, a series of Code LLMs with enhanced code generation and generalization capabilities fine-tuned on multi-source data. To achieve this, we pioneer to unveil inherent conflicts among the various styles and qualities in multi-source code corpora and introduce data-specific prompts with hindsight relabeling, termed AlchemistPrompts, to harmonize different data sources and instruction-response pairs. Additionally, we propose incorporating the data construction process into the fine-tuning data as code comprehension tasks, including instruction evolution, data filtering, and code review. Extensive experiments demonstrate that AlchemistCoder holds a clear lead among all models of the same size (6.7B/7B) and rivals or even surpasses larger models (15B/33B/70B), showcasing the efficacy of our method in refining instruction-following capabilities and advancing the boundaries of code intelligence.*
|
| 19 |
+
|
| 20 |
+
- **AlchemistPrompts**: Designed as data-specific prompts for harmonizing inherent conflicts in multi-source data and mitigating the instruction/response misalignment at a fined-grained level.
|
| 21 |
+
- **Code Comprehenstion Tasks**: Sourced from the process of data construction, consisting of instruction evolution, data filtering, and code review.
|
| 22 |
+
- **Harmonized Multi-source Data**: Instruction tuned on 200M tokens, including 6 types of high-quality data.
|
| 23 |
+
- **Superior Model Performance**: Surpassing all the open-source models of the same size (6.7/7B), and rivaling or even beating larger models (15B/33B/70B/ChatGPT) on 6 code benchmarks.
|
| 24 |
+
- **Advanced generic capabilities**: Demonstrated by the significant improvements on MMLU, BBH, and GSM8K.
|
| 25 |
+
|
| 26 |
+
|
| 27 |
+
## π Quick Start
|
| 28 |
+
```python
|
| 29 |
+
import torch
|
| 30 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 31 |
+
|
| 32 |
+
tokenizer = AutoTokenizer.from_pretrained("internlm/AlchemistCoder-L-7B", trust_remote_code=True)
|
| 33 |
+
model = AutoModelForCausalLM.from_pretrained("internlm/AlchemistCoder-L-7B", trust_remote_code=True, torch_dtype=torch.bfloat16).cuda()
|
| 34 |
+
model = model.eval()
|
| 35 |
+
|
| 36 |
+
input_text = "Implement the Dijkstra algorithm in Python"
|
| 37 |
+
inputs = tokenizer(input_text, return_tensors="pt").to(model.device)
|
| 38 |
+
outputs = model.generate(**inputs, max_length=128)
|
| 39 |
+
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
| 40 |
+
```
|
| 41 |
+
|
| 42 |
+
|
| 43 |
+
## π§ͺ Evaluation and Fine-tune
|
| 44 |
+
Please refer to [**AlchemistCoder**](https://github.com/InternLM/AlchemistCoder) and [**InternLM**](https://github.com/InternLM/InternLM/tree/main).
|
| 45 |
+
|
| 46 |
+
## π Acknowledgments
|
| 47 |
+
*AlchemistCoder* is built with [**InternLM**](https://github.com/InternLM) and [**OpenCompass**](https://github.com/open-compass). Thanks for their awesome work!
|