HyperAIHyperAI

Command Palette

Search for a command to run...

هل يمكن لنماذج الرؤية واللغة حل لعبة القشرة؟

Tiedong Liu Wee Sun Lee

الملخص

تعد تتبع الكيانات البصرية قدرة معرفية فطرية لدى البشر، إلا أنها لا تزال تمثل عنق زجاجة حاسم في نماذج الرؤية واللغة (VLMs). وغالبًا ما يُخفي هذا القصور في معايير التقييم الحالية الخاصة بالفيديو بفضل اختصارات بصرية. نقدم في هذا البحث منصة تشخيصية اصطناعية تُعرف بـ VET-Bench، تتميز بوجود كائنات متطابقة بصريًا تتطلب التتبع حصريًا عبر الاستمرارية الزمكانية. وكشفت تجاربنا أن أحدث نماذج الرؤية واللغة المتطورة (SOTA) تؤدي على منصة VET-Bench بمستوى يقترب من الصدفة أو يساويها، مما يكشف عن قيد جوهري: الاعتماد المفرط على السمات الثابتة على مستوى الإطار الواحد، وفشل في الحفاظ على تمثيلات الكيانات عبر الزمن. وقد قدمنا تحليلًا نظريًا يربط هذه الظاهرة بمسألة تتبع الحالة، مبرهنين على أن نماذج الرؤية واللغة القائمة على محولات (Transformers) ذات عمق ثابت تعاني من قيود جوهرية في تتبع الكائنات غير القابلة للتمييز دون إشراف وسيط، وذلك نتيجة لقيود القدرة التعبيرية. وللتغلب على هذا التحدي، نقترح منهجية "سلسلة التفكير المرساة زمنيًا ومكانيًا" (Spatiotemporal Grounded Chain-of-Thought - SGCoT)، التي تولد مسارات الكائنات كحالات وسيطة صريحة. واستنادًا إلى قدرات تتبع الكائنات في نموذج Molmo2، حفزنا الاستدلال عبر منهجية SGCoT من خلال الضبط الدقيق على بيانات نصية اصطناعية لضمان المحاذاة. وقد حققت طريقتنا دقة متطورة تتجاوز 90% على منصة VET-Bench، مما يثبت أن نماذج الرؤية واللغة قادرة على حل مهمة "لعبة القشرة" في الفيديو بشكل كامل من البداية إلى النهاية دون الحاجة إلى أدوات خارجية. وتتوفر شيفرتنا البرمجية وبياناتنا على الرابط: https://vetbench.github.io.

One-sentence Summary

Researchers from the National University of Singapore introduce VET-Bench to expose tracking failures in Vision-Language Models and propose Spatiotemporal Grounded Chain-of-Thought, a technique that generates explicit object trajectories to overcome expressivity limits and achieve over 90% accuracy on indistinguishable object tracking tasks.

Key Contributions

  • Existing video benchmarks obscure the critical bottleneck of visual entity tracking by allowing models to rely on static appearance cues, prompting the introduction of VET-Bench, a synthetic testbed featuring visually identical objects that necessitate tracking through spatiotemporal continuity alone.
  • Theoretical analysis proves that visual entity tracking is NC1-complete, demonstrating that fixed-depth transformer-based VLMs are fundamentally limited in solving this task without intermediate supervision due to expressivity constraints.
  • The proposed Spatiotemporal Grounded Chain-of-Thought method elicits explicit object trajectory generation as intermediate reasoning states, enabling the Molmo2 model to achieve state-of-the-art accuracy exceeding 90% on VET-Bench without external tools.

Introduction

Visual entity tracking is a foundational capability for embodied AI and game-playing agents, yet current Vision-Language Models (VLMs) struggle with this task due to an over-reliance on static appearance cues rather than genuine spatiotemporal continuity. Existing benchmarks often mask this deficit by including visual shortcuts, such as distinctive object features, which allow models to achieve high scores without performing actual tracking across frames. The authors address these limitations by introducing VET-Bench, a synthetic testbed using visually identical objects to force models to rely solely on motion continuity, while also proving that fixed-depth transformers are theoretically limited in solving such tasks without intermediate computation. To overcome these expressivity constraints, they propose Spatiotemporal Grounded Chain-of-Thought (SGCoT), a method that elicits explicit object trajectory generation as intermediate reasoning steps, enabling models to achieve over 90% accuracy on the benchmark without external tools.

Dataset

  • Dataset Composition and Sources: The authors introduce VET-Bench, a fully synthetic dataset generated via a three.js pipeline to evaluate visual entity tracking. Unlike real-world benchmarks, this approach offers fine-grained control over environmental parameters such as color, material, texture, lighting, and camera viewpoint to prevent data leakage and overfitting.

  • Key Subset Details: The benchmark focuses on two canonical tasks:

    • Cups Game: Modeled after the Shell Game, this task requires tracking a ball hidden under identical opaque containers that undergo positional swaps.
    • Cards Game: Inspired by Three-Card Monte, this task involves tracking a card after it is flipped face-down and shuffled.
    • The pipeline allows precise adjustment of object counts and swap counts to create unlimited episodes and enable diagnostic evaluation of specific factors.
  • Usage and Training Strategy: The dataset is designed for diagnostic evaluation rather than standard training, forcing models to rely exclusively on fine-grained spatiotemporal perception. The authors use it to demonstrate that models achieving high scores on other benchmarks often fail here because they cannot solve the task without explicit frame-level cues.

  • Processing and Filtering Rules: To ensure realism and prevent shortcuts, the generation process enforces strict constraints:

    • No single frame reveals the target identity or the swap operation.
    • All containers are visually identical and opaque to block appearance-based re-identification.
    • The dataset removes static cues and symbolic annotations (such as arrows) found in other benchmarks like VideoReasonBench.
    • This design ensures that correct answers depend on exploiting spatiotemporal continuity across frames rather than static in-frame information.

Method

The authors formulate the visual entity tracking task as determining the terminal index π(i)\pi(i)π(i) of a target object iii within a video sequence V={F0,,FT}\mathcal{V} = \{F_0, \ldots, F_T\}V={F0,,FT} containing NNN visually indistinguishable objects. To ensure the problem is well-posed, a continuity constraint is enforced where the maximum displacement ddd between consecutive frames satisfies 2d<Δ2d < \Delta2d<Δ, preventing identity aliasing during object crossovers. As illustrated in the benchmark overview below, the Visual Entity Tracking Benchmark (VET-Bench) evaluates this capability through scenarios like the "Cups Game" and "Cards Game," where objects undergo shuffling permutations. The figure highlights that while humans can intuitively track these entities, standard Vision-Language Models (VLMs) often fail to provide correct answers even when attempting step-by-step reasoning.

Theoretical analysis reveals that for k5k \geq 5k5 objects, the tracking problem is NC1^11-complete, placing it beyond the theoretical capacity of constant-depth transformers to solve via direct end-to-end training. Empirical results confirm that training with direct-answer supervision leads to stagnant loss at random chance levels. To overcome these limitations, the authors propose Spatiotemporal Grounded Chain-of-Thought (SGCoT). This method leverages Molmo2, a model pre-trained on video object tracking, to generate explicit spatiotemporal trajectories as intermediate reasoning steps.

Instead of providing a direct answer, the model is prompted to output a structured trajectory in the format <tracks coords="timestamp object_idx x y;...">, where timestamps are spaced at 0.5-second intervals and coordinates are normalized. This trajectory serves as the Chain-of-Thought, explicitly aligning when events occur and where entities are located. The training process is designed to be highly efficient and text-only. The authors synthesize trajectories using a Python script and align the model by masking the loss on the generated trajectory tokens while supervising only the final answer. This approach encourages the model to retain its grounding capabilities while learning to derive the final answer from the explicit state representation provided by the SGCoT. By discretizing time at consistent intervals and ensuring precise spatial states for each timestamp, SGCoT avoids the temporal misalignment and underspecification issues found in generic descriptive CoTs.

Experiment

  • Evaluation of diverse proprietary and open-source video-language models on the VET-Bench shell game reveals that all systems perform near random chance, indicating a universal failure in fine-grained spatiotemporal entity tracking.
  • Qualitative analysis identifies three primary failure modes: direct guessing without reasoning, coarse semantic descriptions that miss specific swap events, and hallucinated swap sequences where models generate logically coherent but visually incorrect tracking steps.
  • Experiments varying swap counts show that while models succeed on zero-swap tasks by relying on static visual cues, performance collapses to random guessing with just a single swap, proving an inability to maintain object continuity.
  • Tests with varying object counts demonstrate that accuracy scales inversely with the number of objects, confirming that models do not perform genuine entity tracking but instead resort to statistical guessing.
  • A filtered subset of videos with identical opaque cups and no visual shortcuts confirms that current models excel only when exploiting dataset artifacts and fail completely when robust visual perception is strictly required.
  • Fine-tuning a model with Spatiotemporal Grounded Chain-of-Thought improves reasoning structure but still fails when the model cannot distinguish between visually identical objects, leading to tracking jumps and incorrect final predictions.

بناء الذكاء الاصطناعي بالذكاء الاصطناعي

من الفكرة إلى الإطلاق — سرّع تطوير الذكاء الاصطناعي الخاص بك مع المساعدة البرمجية المجانية بالذكاء الاصطناعي، وبيئة جاهزة للاستخدام، وأفضل أسعار لوحدات معالجة الرسومات.

البرمجة التعاونية باستخدام الذكاء الاصطناعي
وحدات GPU جاهزة للعمل
أفضل الأسعار

HyperAI Newsletters

اشترك في آخر تحديثاتنا
سنرسل لك أحدث التحديثات الأسبوعية إلى بريدك الإلكتروني في الساعة التاسعة من صباح كل يوم اثنين
مدعوم بواسطة MailChimp