Kwai AI´s SRPO Framework Drastically Boosts LLM Reinforcement Learning Efficiency

Kwai AI´s new SRPO method streamlines large language model training, achieving DeepSeek-level math and code results with only a fraction of the effort—reshaping Artificial Intelligence reinforcement learning.

Recent research from Kwai AI´s Kwaipilot team has demonstrated a significant leap in large language model (LLM) reinforcement learning, introducing the Two-Staged history-Resampling Policy Optimization (SRPO) framework. This approach addresses well-known inefficiencies and bottlenecks in standard General Reinforcement Learning from Preference Optimization (GRPO) by reducing post-training steps by 90% while matching the high-performance benchmarks set by models like DeepSeek-R1-Zero in both mathematical and coding domains. By building on the same Qwen2.5-32B base as DeepSeek and open-sourcing their SRPO-Qwen-32B model, the team showcases a combination of technical transparency and empirical success.

The SRPO framework utilizes a two-stage reinforcement learning process tailored for cross-domain reasoning. The first stage exclusively trains on challenging mathematical data to enhance deep, reflective reasoning abilities, incentivizing behaviors such as step-by-step problem decomposition and self-correction. In the subsequent phase, code data is integrated, reinforcing not only programming skills but also advancing procedural thinking and tool-use capabilities. This staged approach resolves cross-domain optimization conflicts inherent in mixed training and ensures the model´s reasoning skills are robust and transferable. Notably, SRPO achieves state-of-the-art results on the AIME24 (50) and LiveCodeBench (41.6) benchmarks with only one-tenth the RL steps required by comparative methods.

Key innovations in SRPO include the history resampling technique, which filters out redundant or overly simple samples that provide little learning signal, and instead prioritizes examples that generate informative gradient updates. This substantially improves training efficiency and maintains effective learning curves, even in later training stages where traditional methods plateau. Rigorous data preprocessing further enhances model reliability by verifying solution correctness and categorizing problem difficulty. Experiments reveal that SRPO-trained models not only produce longer, more detailed responses but also spontaneously display advanced reflective behaviors like code-based self-verification, mirroring human-like problem-solving adaptability. This research marks a robust advancement in efficient, cross-domain RL for large language models and has set a new standard for open and reproducible Artificial Intelligence model development.

77

Impact Score

Hybrid Web3 strategies for the artificial intelligence era

Enterprises are starting to blend Web2 infrastructure with decentralized Web3 technologies to cut costs, improve resilience, and support artificial intelligence workloads, while navigating persistent interoperability, regulatory, and user experience challenges.

Artificial Intelligence, chips, and robots set the tone at CES 2026

CES 2026 in Las Vegas put Artificial Intelligence at the center of nearly every major announcement, with chipmakers and robotics firms using the show to preview their next wave of platforms and humanoid systems. Nvidia, AMD, Intel, Qualcomm, Google, Samsung, Hyundai, and Boston Dynamics all leaned on Artificial Intelligence to anchor their product strategies.

Inside the UK’s artificial intelligence security institute

The UK’s artificial intelligence security institute has found that popular frontier models can be jailbroken at scale, exposing reliability gaps and security risks for governments and regulated industries that rely on trusted vendors.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.