vla_foundry
Collection
VLA Foundry: pretrained LLM, VLM, and VLA checkpoints. • 8 items • Updated • 3
A 1.7B parameter vision-language-action model for bimanual robotic manipulation, part of the VLA Foundry collection. Trained on real-world manipulation data only.
Success rates on 16 seen tasks and 3 unseen tasks (200 rollouts per task):
| Simulator | Seen (16 tasks) | Unseen (3 tasks) |
|---|---|---|
| CS | 1.0% | 0.3% |
| OSS | 2.2% | 4.5% |
Note: This model was trained on real-world data only and does not transfer well to simulation.
git clone https://github.com/TRI-ML/vla_foundry.git
cd vla_foundry
pip install -e .
from vla_foundry.models.base_model import BaseModel
model = BaseModel.from_pretrained("TRI-ML/Foundry-VLA-1.7B-real")