We introduce Kimi-Dev-72B, our new open-source coding LLM for software engineering tasks. Kimi-Dev-72B achieves a new state-of-the-art on SWE-bench Verified among open-source models.
-
Kimi-Dev-72B achieves 60.4% performance on SWE-bench Verified. It surpasses the runner-up, setting a new state-of-the-art result among open-source models.
-
Kimi-Dev-72B is optimized via large-scale reinforcement learning. It autonomously patches real repositories in Docker and gains rewards only when the entire test suite passes. This ensures correct and robust solutions, aligning with real-world development standards.
-
Kimi-Dev-72B is available for download and deployment on Hugging Face and GitHub. We welcome developers and researchers to explore its capabilities and contribute to development.
# clone repo
git clone https://github.com/MoonshotAI/Kimi-Dev.git
# create env
conda create -n kimidev python=3.12
# local install
pip install -e .
Prepare repo structure [From Agentless]
Since for each issue in the benchmark (both SWE-Bench Lite and SWE-Bench Verified) we need to checkout the repository and process the files, you might want to save some time by downloading the preprocessed data here: swebench_repo_structure.zip. After downloading, please unzip and export the location as such
export PROJECT_FILE_LOC={folder which you saved}
# Install vLLM with CUDA 12.8.
# If you are using pip.
pip install vllm --extra-index-url https://download.pytorch.org/whl/cu128
# If you are using uv.
uv pip install vllm --torch-backend=auto
vllm serve Kimi-Dev-72B --served-model-name kimi-dev --host 0.0.0.0 --port 8000 --gpu-memory-utilization 0.95 --max-seq-len-to-capture 131072 --tensor-parallel-size 8
Kimi-Dev adopts a simplified two-stage framework for handling code repair and test writing tasks:
- File Localization: Intelligently identify key files that need modification based on problem descriptions and repository structure
- Code Editing: Perform precise code modifications on the located files, including bug fixes or unit test insertions
Compared to multi-step localization methods, we perform localization at the file level and then pass the complete file to the repair step for more detailed reasoning.
Run rollout script:
conda activate kimidev
# Bugfixer
python kimidev/examples/rollout_messages_bugfixer.py --model_name {vllm_serve_model}
# Testwriter
python kimidev/examples/rollout_messages_testwriter.py --model_name {vllm_serve_model}
We provide some example result files as well as the files required for test-time scaling here.
You can also download these files from Google Drive.
Welcome to submit Pull Requests or create Issues to help improve the project.
If you have any questions, please feel free to submit a GitHub issue or contact zhuhan@moonshot.cn.
If you find our code and models useful, please kindly cite the following information.
@misc{kimi_dev_72b_2025,
title = {Introducing Kimi-Dev-72B: A Strong and Open Coding LLM for Issue Resolution},
author = {{Kimi-Dev Team}},
year = {2025},
month = {June},
url = {\url{https://www.moonshot.cn/Kimi-Dev}}
}