LeRobot v0.5.0: Scaling Every Dimension

-



With over 200 merged PRs and over 50 latest contributors since v0.4.0, LeRobot v0.5.0 is our biggest release yet — expanding in every direction directly. More robots (including our first humanoid), more policies (including the comeback of autoregressive VLAs), faster datasets, simulation environments you’ll be able to load straight from the Hub, and a modernized codebase running on Python 3.12 and Transformers v5. Whether you are training policies in simulation or deploying them on real hardware, v0.5.0 has something for you.



TL;DR

LeRobot v0.5.0 adds full Unitree G1 humanoid support (whole-body control models), latest policies –including Pi0-FAST autoregressive VLAs and Real-Time Chunking for responsive inference–, and streaming video encoding that eliminates wait times between recording episodes. The discharge also introduces EnvHub for loading simulation environments from the Hugging Face Hub, NVIDIA IsaacLab-Arena integration, and a significant codebase modernization with Python 3.12+, Transformers v5, and third-party policy plugins.



Table of Contents



Hardware: More Robots Than Ever

LeRobot v0.5.0 dramatically expands the roster of supported hardware — from arms and mobile robots to a full humanoid.



Unitree G1 Humanoid

The most important hardware addition on this release: full Unitree G1 humanoid support. That is LeRobot’s first humanoid integration, and it’s comprehensive:

  • Locomotion: Walk, navigate, and move through environments.
  • Manipulation: Perform dexterous object manipulation tasks.
  • Teleoperation: Control the G1 remotely with an intuitive teleoperation interface.
  • Whole-Body Control (WBC): Coordinate locomotion and manipulation concurrently for complex, real-world tasks.

The G1 integration represents a significant step toward general-purpose robotics inside LeRobot — moving beyond tabletop arms into full-body embodied AI. Try it out yourself by following the documentation.

unitree-boss



OpenArm & OpenArm Mini

We have added support for the OpenArm robot and its companion OpenArm Mini teleoperator. OpenArm is a capable robot arm with full LeRobot integration, and the Mini serves as its natural teleoperation device. Each support bi-manual configurations, enabling dual-arm setups for more complex manipulation tasks. Test it out within the documentation.



More Robots

The hardware ecosystem keeps growing:

  • Earth Rover: Our first mobile robot integration, bringing LeRobot to outdoor navigation and ground-level robotics.
  • OMX Robot: A brand new robot arm with configurable gripper settings and calibration support.
  • SO-100/SO-101 Consolidation: We have unified the SO-100 and SO-101 implementations right into a single, cleaner codebase — including bi-manual setups. Less code duplication, easier maintenance, same great robots.



CAN Bus Motors

Recent motor controller support via CAN (Controller Area Network) bus opens the door to higher-performance actuators:

  • RobStride: A CAN-based motor controller for high-torque applications.
  • Damiao: One other CAN bus motor controller, expanding the range of compatible hardware.

These additions mean LeRobot can now drive a greater diversity of professional-grade actuators beyond the present Dynamixel and Feetech ecosystem.



Policies: A Growing Model Zoo

This release brings six latest policies and techniques into LeRobot, pushing the boundaries of what is possible with open-source robot learning.



Pi0-FAST: Autoregressive VLAs

Pi0-FAST brings autoregressive Vision-Language-Motion models to LeRobot with FAST (Frequency-space Motion Sequence Tokenization). Unlike the flow-matching approach of Pi0, Pi0-FAST uses an autoregressive motion expert (based on Gemma 300M) that generates discretized motion tokens, enabling:

  • FAST tokenization: Actions are tokenized for autoregressive decoding, with a dedicated FAST motion tokenizer.
  • Flexible decoding: Configurable temperature and max decoding steps for balancing speed and quality.
  • RTC-compatible: Works with Real-Time Chunking (see next section) for responsive inference.
lerobot-train 
  --policy.type=pi0_fast 
  --dataset.repo_id=lerobot/aloha_sim_insertion_human 
  --policy.device=cuda



Real-Time Chunking (RTC)

Real-Time Chunking is an inference-time technique from Physical Intelligence that makes flow-matching policies dramatically more responsive. As a substitute of waiting for a full motion chunk to complete before replanning, RTC repeatedly blends latest predictions with in-progress actions, producing smoother and more reactive behavior.

RTC will not be a standalone policy — it’s an enhancement that plugs into existing flow-matching policies (Pi0 family, SmolVLA & Diffusion). Configure it via --policy.rtc_config.enabled=true.

It is a game-changer for real-world deployment where latency matters. Read the original paper for the technical details and our documentation.



Wall-X

Wall-X is a brand new VLA policy built on Qwen2.5-VL with flow-matching motion prediction. It combines the strong vision-language understanding of Qwen2.5-VL with a flow-matching head for cross-embodiment robotic control.

pip install lerobot[wall_x]
lerobot-train 
  --policy.type=wall_x 
  --dataset.repo_id=lerobot/aloha_sim_insertion_human



X-VLA

X-VLA brings a Florence2-based VLA to LeRobot. Built on Microsoft’s Florence-2 vision-language model, X-VLA offers an alternate backbone for VLA policies, expanding the range of foundation models available for robot learning. Take a look at the training guide for setup instructions and the base model.

pip install lerobot[xvla]
lerobot-train 
  --policy.type=xvla 
  --dataset.repo_id=lerobot/bimanual-so100-handover-cube



SARM

SARM (Stage-Aware Reward Modeling) tackles certainly one of the toughest problems in robot learning: long-horizon tasks. As a substitute of using a single global linear progress signal over the entire episode, it models progress in a stage-aware manner by predicting each the duty stage and the progress inside that stage. This makes it much easier to coach policies for complex, multi-step manipulation tasks. Start experimenting with it by following the documentation.

sarm-community



PEFT Support

You’ll be able to now fine-tune large VLAs using LoRA (and other PEFT methods) without modifying the core training pipeline. PEFT configuration lives on the policy level, making it straightforward to adapt massive foundation models to your specific robot and task with a fraction of the compute. Learn more reading the documentation.

lerobot-train 
  --policy.type=pi0 
  --policy.peft_config.use_peft=true 
  --dataset.repo_id=lerobot/aloha_sim_insertion_human



Datasets: Faster Recording, Faster Training

The dataset pipeline gets major performance improvements on this release, making each data collection and training significantly faster.



Streaming Video Encoding

Previously, recording a dataset meant waiting after each episode for video encoding to complete. No more. With streaming video encoding, frames are encoded in real-time as they’re captured — meaning zero wait time between episodes. Just finish one episode and immediately start the subsequent.

Streaming encoding also supports hardware encoder auto-detection, so in case your system has a GPU-accelerated video encoder, LeRobot will use it mechanically:

dataset = LeRobotDataset.create(
    repo_id="my/dataset",
    fps=30,
    video_backend="auto",       
    streaming_encoding=True,    
)



10x Faster Image Training, 3x Faster Encoding

Under the hood, we have fixed key data access bottlenecks and overhauled image processing:

  • 10x faster image training: Improved image transform support and glued data access bottlenecks that were silently slowing down training.
  • 3x faster encoding: Parallel encoding is now the default across all platforms, with dynamic compression levels that adapt to your dataset type (video vs. image), when not using streaming.
  • Higher CPU utilization: More efficient resource usage during recording and dataset creation.



Recent Dataset Tools

The dataset editing toolkit continues to grow:

  • Subtask support: Annotate and query subtasks inside episodes for hierarchical task learning.
  • Image-to-video conversion: Convert existing image-based datasets to video format for higher storage efficiency, with support for multiple episodes per video file.
  • More editing operations: Recent info operation for inspecting datasets, task modification tools, and diverse fixes to existing operations (splitting, merging, feature editing).
  • Expose more options: Configurable video codecs, tolerance settings, and metadata buffer sizes for fine-grained control over dataset creation.



EnvHub: Environments from the Hub

EnvHub is a brand new technique to use simulation environments in LeRobot: load them directly from the Hugging Face Hub. As a substitute of putting in environment packages locally and wiring up registration, you’ll be able to now point LeRobot at a Hub repository and it handles every little thing — downloading the environment code, registering it with Gymnasium, and making it available for training and evaluation.

Hub environments use HubEnvConfig, which downloads and executes distant make_env functions:

lerobot-train 
  --env.type=hub 
  --env.hub_path="username/my-custom-env" 
  --policy.type=act

This lowers the barrier for sharing custom simulation environments with the community. Package your environment, push it to the Hub, and anyone can train on it. Take a look at the documentation to learn more. Here’s an example to start: LeIsaac x LeRobot EnvHub tutorial.



NVIDIA IsaacLab-Arena

We have integrated NVIDIA IsaacLab-Arena, bringing GPU-accelerated simulation to LeRobot. IsaacLab-Arena provides a set of manipulation tasks running on NVIDIA’s Isaac Sim, offering massively parallel environment instances for fast reinforcement learning. The combination includes dedicated pre/post-processing steps and full compatibility with LeRobot’s training pipeline. Take a look at the documentation.



Codebase: A Modern Foundation

This release modernizes the codebase:

  • Python 3.12+: LeRobot now requires Python 3.12 because the minimum version, enabling modern syntax and higher performance.
  • Transformers v5: We have migrated to Hugging Face Transformers v5, staying current with the most recent model ecosystem.
  • Third-party policy plugins: Identical to v0.4.0’s hardware plugin system, you’ll be able to now register custom policies as installable packages — pip install lerobot_policy_mypolicy and use it with --policy.type=mypolicy. No core library changes needed. Learn easy methods to do it by following the documentation.
  • Distant Rerun visualization: Visualize your robot’s telemetry remotely using Rerun, with compressed image support for bandwidth-efficient streaming.
  • Installation improvements: Added uv installation instructions, clarified setup steps, and improved dependency management. Sequential install steps at the moment are clearly documented.
  • Documentation versioning: Docs at the moment are versioned, so you’ll be able to at all times find documentation matching your installed release.
  • PyTorch version bump: Updated PyTorch version bounds to support NVIDIA Blackwell GPUs.


Community & Ecosystem

  • Modernized Discord: Updated probably the most vibrant community hub with a greater channel organization.
  • GitHub README, templates & automated labeling: A refreshed README, latest issue and PR templates, contributing guidelines, and automatic labeling of tickets — making it easier for everybody to contribute.
  • ICLR 2026 paper acceptance: The LeRobot paper has been accepted to ICLR 2026!
  • LeRobot Visualizer refresh: The visualization tool got a refresh with latest dataset visualization badges and improved functionality. Test it out !
  • LeRobot Annotation Studio: A HuggingFace Space designed to simply annotate every moment of your dataset with natural language subtasks. Test it out !

visualizer



Final Thoughts

Beyond these headline features, v0.5.0 includes tons of of bug fixes, documentation improvements, CI/CD enhancements, and quality-of-life improvements across all the codebase. From higher type checking to more robust test infrastructure, we’re investing within the foundations that make LeRobot reliable and maintainable because it scales.

We would like to increase an enormous thanks to everyone locally — contributors, users, and collaborators alike — for helping LeRobot grow into what it’s today. Every bug report, PR, and discussion makes this project higher.

Stay tuned for more to return 🤗 Start here!
– The LeRobot team ❤️

There’s an enormous surprise coming good across the corner, stay tuned! 👕



Source link

ASK ANA

What are your thoughts on this topic?
Let us know in the comments below.

0 0 votes
Article Rating
guest
0 Comments
Oldest
Newest Most Voted
Inline Feedbacks
View all comments

Share this article

Recent posts

0
Would love your thoughts, please comment.x
()
x