All activity
Tektonianleft a comment
Happy to share this. In our experience, running robot benchmarks often ends up being more about setup and debugging than evaluation itself. This was built to remove that overhead. Curious how others are dealing with this today: - Are you running everything locally? - Any workflows for parallel execution or scaling experiments? Would be great to hear what’s actually working in practice.
Simulac Accelerate physical AI (VLA) evaluation
