PhAIL runs four models (OpenPI/pi0.5, GR00T, ACT, SmolVLA) on bin-to-bin order picking – one of the most common warehouse operations. Same robot (Franka FR3), same objects, hundreds of blind runs. The operator doesn't know which model is running.
Best model: 64 UPH. Human teleoperating the same robot: 330. Human by hand: 1,300+.
Everything is public – every run with synced video and telemetry, the fine-tuning dataset, training scripts. The leaderboard is open for submissions.
Happy to answer questions about methodology, the models, or what we observed.
[1] Vision-Language-Action: https://en.wikipedia.org/wiki/Vision-language-action_model
anna_pozniak•4h ago
vertix•4h ago