Clone with submodules to include LIBERO benchmark and LeRobot dependencies.
Set up a conda environment with Python 3.12 and install PyTorch with CUDA support.
Install LeRobot (for Pi0.5 support) and the project package.
Download the Pi0.5 fine-tuned checkpoint from HuggingFace.
Run a quick check to make sure everything is configured correctly.
Run the full mechanistic interpretability pipeline on Pi0.5 (3B, dual-pathway, flow matching).
Train TopK SAEs (k=64, 8x expansion) on action expert residual stream activations.
TORCH_COMPILE_DISABLE=1 for experiment scripts.
Identify task-specific features using Cohen's d effect size scoring.
Zero-ablate or steer specific SAE features during live rollouts to establish causality.
Validate causality with probes and test activation transfer across conditions.
Additional setup for OpenVLA-OFT (7B, LLaMA-2 backbone, L1 regression action head).
Install the prismatic-vlms and openvla packages alongside the base environment.
Train SAEs on all 32 LLaMA-2 layers. Note: apply SAEs to action tokens only (last 7 of 605 total tokens).
Run ablation experiments across all 4 LIBERO suites.
Action Atlas is deployed at action-atlas.com, but you can also run it locally with full data access.
The Flask backend serves API endpoints and data from local storage.
The Next.js frontend provides the interactive visualization interface.
Visit http://localhost:3002 to access Action Atlas locally.
Run both backend and frontend in a single container.
Browse UMAP scatter plots of SAE features, search by semantic query, and inspect individual feature activations across layers and suites.
Filter through 49,000+ rollout videos by model (Pi0.5 / OFT), experiment type, suite, and success/failure outcome.
Compare baseline vs. ablated behavior side-by-side, view success rate deltas per concept, and explore layer-phase ablation matrices.
Test model robustness to vision perturbations (flip, blur, noise, crop) with real experiment data from 5,691 episodes across Pi0.5 and OFT.