HyperAIHyperAI

Command Palette

Search for a command to run...

pi0.7: 発現的な能力を備えたステアラブルな汎用ロボット基盤モデル(Steerable Generalist Robotic Foundation Model)

概要

我々は、広範なシナリオにおいて強力なout-of-the-box(即時的な)性能を実現する、新しいロボティクス基盤モデル「π0.7\pi_{0.7}π0.7」を提案する。π0.7\pi_{0.7}π0.7は、未知の環境において多様な言語指示に従うことが可能であり、これには様々なキッチン家電を用いたマルチステージ(多段階)タスクが含まれる。また、ゼロショットでのcross-embodiment(異種エンボディメント間)の汎化性能も備えており、例えば、洗濯物を畳むタスクを事前に学習することなく実行できる。さらに、エスプレッソマシンの操作といった困難なタスクにおいても、より特化型のRL-finetuned(強化学習による微調整済み)モデルに匹敵するレベルの性能を、学習なし(out of the box)で発揮する。π0.7\pi_{0.7}π0.7の根幹となるアイデアは、学習プロセスにおいて多様なcontext conditioning(コンテキスト条件付け)を用いることにある。promptに含まれるこの条件付け情報により、モデルを精密に制御し、異なる戦略を用いた多くのタスクを実行することが可能となる。この条件付けは、単に行うべき内容を記述する言語コマンドだけでなく、タスクの遂行方法や戦略を記述する追加的なマルチモーダル情報(タスクのパフォーマンスに関するメタデータやsubgoal(部分目標)の画像など)に基づいている。これにより、π0.7\pi_{0.7}π0.7は極めて多様なデータを利用することが可能となっている。

One-sentence Summary

pi0.7 is a steerable generalist robotic foundation model utilizing diverse context conditioning with multimodal prompt information to precisely steer task strategies, delivering strong out-of-the-box performance in unseen environments and zero-shot cross-embodiment generalization for tasks like laundry folding while matching specialized RL-finetuned models on challenging tasks such as operating an espresso machine.

Key Contributions

  • The paper introduces π0.7\pi_{0.7}π0.7, a robotic foundation model designed to deliver strong out-of-the-box performance across a wide range of scenarios without task-specific post-training.
  • The method utilizes diverse context conditioning during training by augmenting language commands with strategy metadata and subgoal images to resolve ambiguity in diverse datasets.
  • Evaluation results demonstrate zero-shot cross-embodiment generalization and the ability to perform challenging tasks at a level matching specialized RL-finetuned models.

Introduction

Physical intelligence seeks to establish generalist capabilities in robotics similar to large language models, but prior vision-language-action models lack compositional generalization and often require task-specific fine-tuning. Training on diverse datasets often leads models to average out different strategies, resulting in suboptimal performance. The authors introduce pi_0.7, a steerable generalist robot foundation model that leverages diverse context conditioning to resolve ambiguity in mixed-quality data. By enriching prompts with detailed language, subgoal images, and strategy metadata, the model learns to compose skills effectively without fine-tuning, enabling zero-shot cross-embodiment transfer and robust performance on complex dexterous tasks.

Dataset

  • Composition and Sources: The authors aggregate demonstration data from diverse robot platforms (static, mobile, single, and bimanual) operating in lab, home, and wild environments. The mixture also includes autonomous data from policy evaluations, human interventions, open-source robot datasets, egocentric human videos, and auxiliary web data for visual question answering and object prediction.
  • Suboptimal Data Strategy: Departing from classic pipelines, the dataset intentionally includes lower quality demonstrations, failure episodes, and trajectories from prior model versions. This approach enables the model to distill capabilities from RL-trained specialists and improves robustness across varied states.
  • Metadata Processing: Episode metadata is constructed to label task execution attributes. Speed is discretized into 500-step intervals, quality receives a score from 1 to 5, and human annotators identify mistake segments within action sequences.
  • Training and Usage: Context modalities including instructions, images, and metadata undergo dropout during training to ensure flexible prompting. At inference, the model uses ground-truth metadata to condition performance on desired speed, quality, and accuracy.

Method

The π0.7\pi_{0.7}π0.7 model is a Vision-Language-Action (VLA) foundation model designed for generalist robot manipulation. It builds upon the π0.6\pi_{0.6}π0.6 architecture and the MEM memory system, extending them with multi-modal context conditioning. The model consists of a 4B-parameter VLM backbone initialized from Gemma 3, which includes a 400M-parameter SigLIP vision encoder, and a separate 860M-parameter action expert. The total parameter count is approximately 5B.

The authors leverage a flow matching objective for the action expert to predict continuous action chunks. The VLM backbone processes visual observations and language inputs, while the action expert attends to these activations to generate robot commands. This separation allows for fast inference at runtime while maintaining stable training for the backbone via discrete cross-entropy loss on FAST tokens, a technique known as Knowledge Insulation.

Refer to the architecture diagram below for a detailed view of the model components and data flow.

A key innovation in π0.7\pi_{0.7}π0.7 is the expansion of the context prompt CtC_tCt beyond simple language instructions. The model accepts a rich set of inputs including multi-view observation memory, task instructions, subtask instructions, episode metadata, and subgoal images. This multi-modal prompting enables the model to learn from diverse and heterogeneous datasets, including suboptimal behaviors and failures.

The system integrates a high-level policy and a world model to generate these contextual elements at runtime. The high-level policy produces semantic subtask instructions, while the world model generates subgoal images that depict the desired near-future state of the scene. These subgoal images provide spatial grounding that language alone may lack.

Refer to the system overview below illustrating how robot and non-robot data feed into the training pipeline and how the model operates during inference.

During training, the model is exposed to a combination of real future images and generated subgoal images. To handle the variability in image quality and delay, the authors employ a specific sampling scheme where real images are sampled from future timesteps or generated by the world model. The training objective maximizes the log-likelihood of the action chunk given the observations and context:

maxθED[logπθ(at:t+HotT:t,Ct)]\operatorname* { m a x } _ { \theta } \mathbb { E } _ { \mathcal { D } } \left[ \log \pi _ { \theta } ( \mathbf { a } _ { t : t + H } \mid \mathbf { o } _ { t - T : t } , \mathcal { C } _ { t } ) \right]maxθED[logπθ(at:t+HotT:t,Ct)]

The model utilizes a block-causal masking scheme where observation and subgoal tokens use bidirectional attention, while text tokens use causal attention. This structure is visualized in the attention mask diagram below.

At inference time, the model supports Classifier-Free Guidance (CFG) on the episode metadata to elicit specific behaviors such as higher speed or quality. The subtask instructions and subgoal images are refreshed whenever the semantic intent changes or after a fixed time interval. The following sequence demonstrates the model executing a complex task involving an air fryer using step-by-step verbal coaching and subtask instructions.

Experiment

The evaluation assesses the π0.7\pi_{0.7}π0.7 model across diverse robot platforms and tasks, specifically testing out-of-the-box dexterity, instruction following, cross-embodiment transfer, and compositional generalization. Results demonstrate that π0.7\pi_{0.7}π0.7 matches specialized fine-tuned models on complex manipulation tasks without post-training and successfully transfers skills to unseen robot morphologies by adapting manipulation strategies. Furthermore, the model exhibits superior language following capabilities that allow it to overcome dataset biases and perform new long-horizon tasks through verbal coaching, while effectively leveraging large, mixed-quality datasets for improved generalization.


AIでAIを構築

アイデアからローンチまで — 無料のAIコーディング支援、すぐに使える環境、最高のGPU価格でAI開発を加速。

AI コーディング補助
すぐに使える GPU
最適な料金体系

HyperAI Newsletters

最新情報を購読する
北京時間 毎週月曜日の午前9時 に、その週の最新情報をメールでお届けします
メール配信サービスは MailChimp によって提供されています