--- license: apache-2.0 datasets: - mlabonne/CodeLlama-2-20k language: - zh --- # Chinese-CodeLlama-7B-SFT We implemented SFT based on our [Chinese-CodeLlama-7B-PT](https://huggingface.co/frankminors123/Chinese-CodeLlama-7B-PT). The dataset comes from [CodeLlama-2-20k](https://huggingface.co/datasets/mlabonne/CodeLlama-2-20k), we used Google Translate to translate it into Chinese. In addition, we designed appropriate Chinese prompt template for coding tasks, and during the fine-tuning stage, `memory efficient attention` was applied which save us a lot of GPU memory space.