Two Types of Policies
Type Definition Use Case Learned Neural network trained on demonstrations Tasks requiring visual adaptation Replay Recorded action sequence Repeatable, fixed motionsLearned Skills (ACT Policy)
Learned skills use Action Chunking with Transformers (ACT)—a neural network architecture that takes camera images and joint positions as input, and outputs action sequences.Metadata Structure
Execution Flow
- Load Policy: Neural network checkpoint loaded into GPU memory
- Move to Start Pose: Arm moves to consistent initial configuration
-
Inference Loop (25Hz): Every 40ms:
- Capture images from both cameras
- Read current joint positions
- Run policy forward pass
- Output: 6 joint commands + 2 base velocity commands
- Progress Monitoring: Early termination when progress > 95%
- End Pose: Optionally return to safe configuration
Key Characteristics
- Reactive: Continuously adjusts based on visual feedback
- Adaptive: Handles variation in object position, lighting, orientation
- Coordinated: Can move arm and base simultaneously
Replay Skills
Replay skills play back pre-recorded action sequences. Simpler than learned skills, but deterministic and reliable.Metadata Structure
Execution Flow
- Load Recording: H5 file containing timestamped actions
- Move to Start Pose: Arm moves to recorded initial position
- Playback (50Hz): Execute recorded actions in sequence
- End Pose: Return to specified configuration
Execution Pipeline
Both skill types are executed by the BehaviorServer:Creating New Skills
Learned Skill Workflow
- Collect Demonstrations: Teleoperate robot through task 10-100 times
- Train Policy: Run ACT training pipeline to produce checkpoint
-
Create Metadata: Add
metadata.jsonwith execution parameters -
Deploy: Place in
skills/<name>/directory
Replay Skill Workflow
- Record Episode: Teleoperate through motion once, save to H5
-
Create Metadata: Add
metadata.jsonwith replay parameters -
Deploy: Place in
skills/<name>/directory

