# aiXcoder-7B Code Large Language Model
🏠 Official website|🛠 VS Code Plugin|🛠 Jetbrains Plugin|Github Project
Welcome to the official repository of aiXcoder-7B Code Large Language Model. This model is designed to understand and generate code across multiple programming languages, offering state-of-the-art performance in code completion, comprehension, generation, and more tasks about programming languages. Table of Contents 1. [Model Introduction](#model-introduction) 2. [Quickstart](#quickstart) - [Environment Requirements](#environment-requirements) - [Model Weights](#model-weights) - [Inference Example](#inference-example) 3. [License](#license) 4. [Acknowledgments](#acknowledgments) ## Model Introduction As the capabilities of large code models are gradually being unearthed, aiXcoder has consistently pondered on how to make these models more beneficial in real development scenarios. To this end, we have open-sourced aiXcoder 7B Base, which has undergone extensive training on 1.2T Unique Tokens, and the model's pre-training tasks as well as the contextual information have been uniquely designed for real-world code generation contexts. aiXcoder 7B Base stands out as the most effective model in code completion scenarios among all models of similar parameter sizes, and it also surpasses mainstream models like codellama 34B and StarCoder2 15B in the average performance on the multilingual nl2code benchmark. In our ongoing exploration to apply large code models, the release of aiXcoder 7B Base represents a significant milestone. The current version of aiXcoder 7B Base is a foundational model that focuses on improving the efficiency and accuracy of code completion and code generation tasks, aiming to provide robust support for developers in these scenarios. It is important to note that this version has not undergone specific instruct-tuning, which means it might not yet offer optimal performance for specialized higher-level tasks such as test case generation and code debugging. However, we have plans for further development of the aiXcoder model series already in motion. In the near future, we aim to release new versions of the model that have been meticulously instruct-tuned for a wider range of programming tasks, including but not limited to test case generation and code debugging. Through these instruct-tuned models, we anticipate offering developers more comprehensive and deeper programming support, helping them to maximize efficiency at every stage of software development. ## Quickstart ### Environment Requirements #### Option 1: Build Env To run the model inference code, you'll need the following environment setup: - Python 3.8 or higher - PyTorch 2.1.0 or higher - sentencepiece 0.2.0 or higher - transformers 4.34.1 or higher (if run inference by transformers library) Please ensure all dependencies are installed using the following command: ```bash conda create -n aixcoder-7b python=3.11 conda activate aixcoder-7b git clone git@github.com:aixcoder-plugin/aiXcoder-7b.git cd aiXcoder-7b pip install -r requirements.txt ``` `requirements.txt` listed all necessary libraries and their versions. To achieve faster inference speeds, especially for large models, we recommend installing `flash attention`. `Flash attention` is an optimized attention mechanism that significantly reduces computation time for transformer-based models without sacrificing accuracy. Before proceeding, ensure your environment meets the CUDA requirements as `flash attention` leverages GPU acceleration. Follow these steps to install `flash attention`: ```bash git clone git@github.com:Dao-AILab/flash-attention.git cd flash-attention MAX_JOBS=8 python setup.py install ``` #### Option 2: Docker For a consistent and isolated environment, we recommend running the model inference code using Docker. Here's how to set up and use Docker for our model: 1. Install Docker: If you haven't already, install Docker on your machine. 2. Pull the Docker Image: Pull the Docker image from Docker Hub. ```bash docker pull pytorch/pytorch:2.1.0-cuda11.8-cudnn8-devel ``` 3. Run the Container: Once the image is pulled, you can run the model inside a Docker container. ```bash docker run --gpus all -it -v /dev/shm:/dev/shm --name aix_instance pytorch/pytorch:2.1.0-cuda11.8-cudnn8-devel /bin/bash pip install sentencepiece git clone git@github.com:aixcoder-plugin/aiXcoder-7b.git cd aiXcoder-7b ``` This command starts a container named aix_instance from the pytorch image. You can interact with the model inside this container. To achieve faster inference speeds, especially for large models, we recommend installing `flash attention`. ```bash git clone git@github.com:Dao-AILab/flash-attention.git cd flash-attention MAX_JOBS=8 python setup.py install ``` 4. Model Inference: Within the Docker container, you can run the model inference code as described in the Inference Example section. Using Docker provides a clean, controlled environment that minimizes issues related to software versions and dependencies. ### Model Weights You can download the model weights from the following link: - [aiXcoder Base Download](https://huggingface.co/aiXcoder/aixcoder-7b-base) - aiXcoder Instruct Download (Comming soon...) ### Inference Example #### Command Line Execution For a quick start, you can run the model inference directly from the command line: ```bash torchrun --nproc_per_node 1 sess_megatron.py --model_dir "path/to/model_weights_dir" ``` Replace "path/to/model_weights_dir" with the actual path to your downloaded model weights. or run inference with huggingface's transformers: ```bash python sess_huggingface.py ``` #### Python Script Execution Alternatively, you can invoke the model programmatically within your Python scripts. This method provides more flexibility for integrating the model into your applications or workflows. Here's a simple example on how to do it: ```python from sess_megatron import TestInference infer = TestInference() res = infer.run_infer( # for FIM style input, code_string stands for prefix context code_string="""# 快速排序算法""", # for FIM style input, later_code stands for suffix context later_code="\n", # file_path should be a path from project to file file_path="test.py", # max num for generated tokens max_new_tokens=256, ) print(res) """output: def quick_sort(arr): if len(arr) <= 1: return arr pivot = arr[0] less = [i for i in arr[1:] if i <= pivot] greater = [i for i in arr[1:] if i > pivot] return quick_sort(less) + [pivot] + quick_sort(greater) # 测试 arr = [3, 2, 1, 4, 5] print(quick_sort(arr)) # [1, 2, 3, 4, 5] """ ``` ```python import torch import sys from hf_mini.utils import input_wrapper from transformers import AutoModelForCausalLM, AutoTokenizer device = "cuda" # the device to load the model onto tokenizer = AutoTokenizer.from_pretrained("aiXcoder/aixcoder-7b-base") model = AutoModelForCausalLM.from_pretrained("aiXcoder/aixcoder-7b-base", torch_dtype=torch.bfloat16) text = input_wrapper( # for FIM style input, code_string stands for prefix context code_string="# 快速排序算法", # for FIM style input, later_code stands for suffix context later_code="\n# 测试\narr = [3, 2, 1, 4, 5]\nprint(quick_sort(arr)) # [1, 2, 3, 4, 5]", # file_path should be a path from project to file path="test.py" ) if len(text) == 0: sys.exit() inputs = tokenizer(text, return_tensors="pt", return_token_type_ids=False) inputs = inputs.to(device) model.to(device) outputs = model.generate(**inputs, max_new_tokens=256) print(tokenizer.decode(outputs[0], skip_special_tokens=False)) """output: def quick_sort(arr): # 如果数组长度小于等于1,直接返回 if len(arr) <= 1: return arr # 选择数组的第一个元素作为基准 pivot = arr[0] # 初始化左右指针 left, right = 1, len(arr) - 1 # 循环直到左指针小于右指针 while left < right: # 从右到左找到第一个小于基准的元素,与左指针元素交换 if arr[right] < pivot: arr[left], arr[right] = arr[right], arr[left] left += 1 # 从左到右找到第一个大于等于基准的元素,与右指针元素交换 if arr[left] >= pivot: right -= 1 # 将基准元素与左指针元素交换 arr[left], arr[0] = arr[0], arr[left] # 对左半部分进行递归排序 quick_sort(arr[:left]) # 对右半部分进行递归排序 quick_sort(arr[left + 1:]) return arr """ ``` ## License The model weights are licensed under the [Model License](./MODEL_LICENSE) for academic research use; for commercial use, please apply by sending an email to support@aiXcoder.com. ## Acknowledgments We would like to thank all contributors to the open-source projects and datasets that made this work possible. Thank you for your interest in our Code Large Language Model. We look forward to your contributions and feedback!