Robot Manipulation On Simplerenv Widow X
Metriken
Average
Put Carrot on Plate
Put Eggplant in Yellow Basket
Put Spoon on Towel
Stack Green Block on Yellow Block
Ergebnisse
Leistungsergebnisse verschiedener Modelle zu diesem Benchmark
| Paper Title | ||||||
|---|---|---|---|---|---|---|
| SoFar | 0.583 | 0.667 | 0.375 | 0.583 | 0.708 | SoFar: Language-Grounded Orientation Bridges Spatial Reasoning and Object Manipulation |
| SpatialVLA | 0.344 | 0.208 | - | 0.208 | 0.250 | SpatialVLA: Exploring Spatial Representations for Visual-Language-Action Model |
| Octo-Small | 0.300 | 0.097 | - | 0.472 | 0.042 | Octo: An Open-Source Generalist Robot Policy |
| Octo-Base | 0.160 | 0.083 | - | 0.125 | 0.000 | Octo: An Open-Source Generalist Robot Policy |
| RoboVLM | 0.135 | 0.250 | - | 0.208 | 0.083 | Towards Generalist Robot Policies: What Matters in Building Vision-Language-Action Models |
| RT-1-X | 0.011 | 0.042 | - | 0.000 | 0.000 | RT-1: Robotics Transformer for Real-World Control at Scale |
| OpenVLA | 0.010 | 0.000 | - | 0.000 | 0.000 | OpenVLA: An Open-Source Vision-Language-Action Model |
0 of 7 row(s) selected.