Verl: Reinforcement Learning Framework for Post-Training Large Language Models

Verl streamlines reinforcement learning post-training for large language models and integrates seamlessly with major machine learning tools, optimizing throughput and scalability for Artificial Intelligence applications.

Verl is an open source, flexible reinforcement learning (RL) training framework specifically designed for post-training large language models (LLMs). Built as an implementation of the HybridFlow architecture, verl offers a user-friendly platform for constructing and executing sophisticated RL dataflows, enabling researchers and practitioners to efficiently extend and experiment with diverse RL algorithms. Its modular APIs decouple computation and data dependencies, allowing seamless integration with existing LLM infrastructures such as PyTorch FSDP, Megatron-LM, and vLLM, as well as with HuggingFace models.

The framework is engineered for performance and scalability, supporting flexible device mapping and parallelism to optimize GPU utilization across various cluster sizes. Verls´s adoption of the 3D-HybridEngine technology ensures efficient actor model resharding, which eliminates memory redundancies and reduces communication overhead during training and generation transitions. The architecture leverages the strengths of both single-controller and multi-controller paradigms, streamlining the execution of complex post-training workflows using a concise and extensible codebase.

Comprehensive documentation and a suite of practical guides—covering installation, backend choices, multi-node training, programming with the HybridFlow model, data preparation, configuration management, and performance tuning—enables rapid onboarding and experimentation. Verls supports community collaboration under Apache License 2.0, with contributions encouraged via GitHub, Slack, or WeChat. The framework adopts modern code quality practices, employing ´ruff´ for linting/formatting and ´pre-commit´ for code management. Continuous integration guidance and open project roadmaps further bolster community engagement, positioning verl as a robust resource for advancing RL-based post-training in cutting-edge Artificial Intelligence systems.

71

Impact Score

Siemens debuts digital twin composer for industrial metaverse deployments

Siemens has introduced digital twin composer, a software tool that builds industrial metaverse environments at scale by merging comprehensive digital twins with real-time physical data, enabling faster virtual decision making. Early deployments with PepsiCo report higher throughput, shorter design cycles and reduced capital expenditure through physics-accurate simulations and artificial intelligence driven optimization.

Cadence builds chiplet partner ecosystem for physical artificial intelligence and data center designs

Cadence has introduced a Chiplet Spec-to-Packaged Parts ecosystem aimed at simplifying chiplet design for physical artificial intelligence, data center and high performance computing workloads, backed by a roster of intellectual property and foundry partners. The program centers on a physical artificial intelligence chiplet platform and framework that integrates prevalidated components to cut risk and speed commercial deployment.

Patch notes detail split compute and IO tiles in Intel Diamond Rapids Xeon 7

Linux kernel patch notes reveal that Intel’s upcoming Diamond Rapids Xeon 7 server processors separate compute and IO tiles and adopt new performance monitoring and PCIe 6.0 support. The changes point to a more modular architecture and a streamlined product stack focused on 16-channel memory configurations.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.