Training from Demonstrations (LeRobot)
rfx datasets are LeRobot-compatible by default. Use LeRobot’s training pipeline with your collected demos:LeRobot supports multiple policy architectures: ACT (Action Chunking Transformer), Diffusion Policy, VQ-BeT, and more. See LeRobot docs for architecture details.
Training in Simulation
For reinforcement learning or sim-to-real transfer, use rfx’s GPU-accelerated simulation:Parallel Simulation Training
Simulation Backends
rfx supports multiple physics backends:| Backend | Description | Best For |
|---|---|---|
mock | Pure PyTorch, no physics | Fast prototyping, testing |
genesis | Genesis physics engine | GPU-accelerated parallel sim |
mjx | MuJoCo XLA | High-fidelity physics |
Saving Trained Models
rfx policies are self-describing and portable:Training Configuration
Hyperparameters
Robot Configuration
Specify robot parameters for simulation:Domain Randomization
For sim-to-real transfer, apply domain randomization:Monitoring Training
WandB Integration (LeRobot)
Manual Logging
Curriculum Learning
Progressively increase task difficulty:Best Practices
Start Small
Start Small
Begin with 50-100 demos and a simple policy architecture. Scale up once you verify the pipeline works.
Data Quality > Quantity
Data Quality > Quantity
10 high-quality demos are better than 100 noisy ones. Ensure consistent, smooth demonstrations.
Validation Split
Validation Split
Hold out 10-20% of episodes for validation to detect overfitting.
Checkpoint Regularly
Checkpoint Regularly
Save checkpoints every 10-50k steps. Training can be unstable, especially early on.
Next Steps
Deploy Policy
Deploy your trained policy to real hardware
Hub Integration
Share your trained models on HuggingFace Hub
