Installation¶
Set up the full AlphaBrain runtime environment.
System Requirements¶
- Python 3.10+
- CUDA 11.8+ / CUDA 12.x
- PyTorch 2.1+
- GPU ≥ 1 (A100 / H100 recommended)
1. Clone the Repository¶
2. Install Core Dependencies¶
conda create -n alphabrain python=3.10 -y
conda activate alphabrain
pip install -r requirements.txt
pip install -e .
Install Flash Attention:
Flash Attention build time
The first flash-attn install compiles from source and takes 10–30 minutes.
3. Install Evaluation Dependencies¶
Evaluation uses a separate conda environment to avoid conflicts with the training environment. LIBERO example (see the LIBERO official docs):
conda create -n libero python=3.10 -y
conda activate libero
git clone https://github.com/Lifelong-Robot-Learning/LIBERO.git
cd LIBERO
pip install -r requirements.txt
pip install -e .
pip install tyro matplotlib mediapy websockets msgpack rich "numpy==1.24.4"
4. Download Pretrained Models and Data¶
Place model weights under PRETRAINED_MODELS_DIR (default data/pretrained_models/), or use symlinks:
data/pretrained_models/
├── Qwen2.5-VL-3B-Instruct/
├── Qwen2.5-VL-7B-Instruct/
├── Qwen3-VL-4B-Instruct/
└── Qwen3-VL-9B-Instruct/
Download from HuggingFace:
huggingface-cli download Qwen/Qwen2.5-VL-3B-Instruct \
--local-dir data/pretrained_models/Qwen2.5-VL-3B-Instruct
Model selection
| Model | Params | Recommended Use |
|---|---|---|
| Qwen2.5-VL-3B-Instruct | 3B | Quick experiments, limited VRAM |
| Qwen2.5-VL-7B-Instruct | 7B | High-accuracy tasks |
| Qwen3-VL-4B-Instruct | 4B | Balanced performance and efficiency |
| Qwen3-VL-9B-Instruct | 9B | Highest accuracy |
One-click LIBERO data download:
The script downloads and symlinks the following datasets:
| Dataset | Description |
|---|---|
libero_spatial | Spatial understanding |
libero_object | Object manipulation |
libero_goal | Goal-directed |
libero_10 | Long-horizon |
Directory layout (LeRobot format, used for training):
${LEROBOT_LIBERO_DATA_DIR}/
├── libero_spatial_no_noops_1.0.0_lerobot/
├── libero_object_no_noops_1.0.0_lerobot/
├── libero_goal_no_noops_1.0.0_lerobot/
└── libero_10_no_noops_1.0.0_lerobot/
5. Configure Environment Variables¶
Copy .env.example and fill in the local paths:
# Pretrained model root (Qwen2.5-VL-3B-Instruct/, paligemma-3b-pt-224/, ...)
PRETRAINED_MODELS_DIR=/your/path/to/pretrained_models
# LeRobot-format LIBERO data root (used for training)
LEROBOT_LIBERO_DATA_DIR=/your/path/to/LEROBOT_LIBERO_DATA
# RLDS-format LIBERO data root (used for evaluation)
LIBERO_DATA_ROOT=/your/path/to/IPEC-COMMUNITY
# LIBERO project root (simulation env)
LIBERO_HOME=/your/path/to/LIBERO
# Python interpreter of the separate LIBERO evaluation conda env
LIBERO_PYTHON=/your/path/to/miniconda3/envs/libero/bin/python
.env is not version-controlled
.env is already in .gitignore, so you can safely store local paths and sensitive info.
6. Verify¶
CUDA errors
Make sure your PyTorch build matches your CUDA version:
Next¶
Head to Baseline VLA for the default finetune + eval walk-through — a short run there also serves as an end-to-end check that your install works.