Ant Group adds AI ecosystem, open-sources code model Ling-Coder-Lite

Written by
Silas Grey
Updated on:July-09th-2025
Recommendation

Ant Group open-sources Ling-Coder-Lite, a new breakthrough in AI code big model.

Core content:
1. Code big model redefines programming efficiency and possibilities
2. MoE architecture improves code big model performance and efficiency
3. Ling-Coder-Lite is open source and supports multi-language and multi-task scenarios

Yang Fangxian
Founder of 53AI/Most Valuable Expert of Tencent Cloud (TVP)



From LLM Challenge to MoE Breakthrough

With the rapid development of artificial intelligence technology, the Code Large Language Model (Code LLM) is becoming an important part of the developer tool chain. From code completion to error repair, from multi-language support to the automation of complex tasks, the Code Large Language Model is redefining the efficiency and possibilities of programming.

However, despite the continuous improvement of the capabilities of large code models, developers still face two major challenges in practical applications: balancing performance and efficiency, and supporting multiple languages ​​and multiple tasks. To solve these problems, the Mixture of Experts (MoE) model came into being. The MoE architecture significantly reduces the computational cost by dynamically activating some parameters, while improving the flexibility and efficiency of the model .
Ant Group officially open-sources Ling-Coder-Lite, a large code language model based on MoE architecture, which improves reasoning efficiency by 1.5-2 times and provides efficient solutions for scenarios such as AI-IDE code completion. It hopes to provide developers with efficient, versatile and easy-to-integrate code generation and understanding tools.
Key highlights

The key summary of this release is as follows:

  • Model and dataset open source : Two lightweight code models Ling-Coder-Lite and Ling-Coder-Lite-Base have been open sourced on Hugging Face and ModelScope. At the same time, Ant Group open sourced SyntheticQA for annealing training, SFT (Supervised Fine-tuning) and DPO (Direct Preference Optimization) for post-training, totaling about 30 million data (Figure 1(a)), to support further research and development in the community.

  • Disclosure of technical details : A technical report was released simultaneously with this open source release, disclosing more details about the method of building a high-quality training code dataset and the phased mixing and matching strategy of data during training, to help the industry jointly advance the research on large code models.

  • Balanced upgrade of efficiency and effect : Based on the Ling-MoE architecture, the total number of parameters of Ling-Coder-Lite is 16.8B, and the activation parameters during inference are only 2.75B, while taking into account higher efficiency and better effect.

  • Multi-language and multi-task support : Ling-Coder-Lite supports dozens of commonly used programming languages ​​such as Python, Java, C++, JavaScript, and performs well in multi-language benchmarks such as MultiPL-E and MBXP. In addition to simple and multi-language code generation, it also supports multiple task scenarios such as competition and application advanced code generation, code understanding and input-output reasoning, data science and SQL data analysis, and code repair.

Model Effect

In 12 code benchmarks, Ling-Coder's performance is comparable to the best model of similar size (Qwen2.5-Coder-7B) (7 out of 12), ahead of OpenCoder-8B and DeepSeek-Coder-V2-lite, see Figure 1(b) for details; the reasoning efficiency is  1.5X~2X faster than Qwen2.5-Coder-7B (Figure 1(c)), which is particularly suitable for scenarios that require low-latency response, such as code completion in AI-IDE. In actual internal use, Ling-Coder-Lite saves half the deployment resources compared to previous models of similar size based on dense architecture under the same latency setting .


Explore the wonderful world of Markdown

Figure 1: Ling-Coder-Lite open source data, model code capabilities, and theoretical reasoning efficiency

  1. Some high-quality data (about 30M samples) used by Ling-Coder-Lite during annealing and post-training have been open sourced; 
  2. Performance of LLM codes of similar parameter size on 12 benchmarks; 
  3. Comparison of various models’ performance (mean evaluation score) versus theoretical computational effort (TFLOPs required for a single inference with a context length of 4096).
Future plans

In the future, we plan to continue to optimize Ling-Coder-Lite in multiple dimensions, including:

  • Continue to push the boundaries of performance and efficiency for large code models.
  • By introducing reinforcement learning and execution feedback, the model's reasoning ability in handling actual software engineering tasks is improved .
  • Continue to improve code data quality, especially the quality of synthetic data .