HyperAIHyperAI

Command Palette

Search for a command to run...

Videonalyse und -generierung mittels einer semantischen Progress-Funktion

Gal Metzer Sagi Polaczek Ali Mahdavi-Amiri Raja Giryes Daniel Cohen-Or

Zusammenfassung

Die durch Bild- und Videogenerierungsmodelle erzeugten Transformationen verlaufen oft hochgradig nichtlinear: Lange Phasen, in denen sich der Inhalt kaum verändert, werden von plötzlichen, abrupten semantischen Sprüngen gefolgt. Um dieses Verhalten zu analysieren und zu korrigieren, führen wir eine Semantic Progress Function ein – eine eindimensionale Repräsentation, welche die zeitliche Entwicklung der Bedeutung einer gegebenen Sequenz erfasst. Für jedes Frame berechnen wir die Distanzen zwischen den semantischen Embeddings und passen eine glatte Kurve an, die die kumulative semantische Verschiebung über die gesamte Sequenz hinweg widerspiegelt. Abweichungen dieser Kurve von einer Geraden offenbaren ein ungleichmäßiges semantisches Tempo (semantic pacing).Basierend auf dieser Erkenntnis schlagen wir ein Verfahren zur semantischen Linearisierung vor, welches die Sequenz reparametrisiert (oder umzeitet), sodass die semantische Veränderung mit einer konstanten Rate erfolgt, was zu glatteren und kohärenteren Übergängen führt. Über die Linearisierung hinaus bietet unser Framework eine modellagnostische Grundlage, um zeitliche Unregelmäßigkeiten zu identifizieren, das semantische Tempo zwischen verschiedenen Generatoren zu vergleichen und sowohl generierte als auch reale Videosequenzen auf ein beliebiges Zieltempo zu steuern.

One-sentence Summary

The authors introduce a Semantic Progress Function that captures semantic evolution by fitting a smooth curve to semantic embedding distances to reveal uneven pacing and propose a semantic linearization procedure that retimes sequences for constant semantic change to yield smoother transitions, providing a model-agnostic foundation for identifying temporal irregularities and steering both generated and real-world video sequences toward arbitrary target pacing.

Key Contributions

  • A Semantic Progress Function is introduced to represent semantic evolution as a one-dimensional curve by computing distances between semantic embeddings across a sequence. This metric objectively quantifies temporal linearity and reveals uneven semantic pacing.
  • A semantic linearization procedure is proposed to reparameterize video sequences so that semantic change unfolds at a constant rate. This method warps temporal positions based on measured semantic content via the Semantic Progress Function to yield smoother and more coherent transitions.
  • The framework provides a model-agnostic foundation for identifying temporal irregularities and steering both generated and real-world video sequences toward arbitrary target pacing. This approach enables the transformation of in-the-wild videos into a constant pace without requiring the manual user annotation necessitated by existing guidance-based methods.

Introduction

Generative video models frequently produce transitions where semantic meaning evolves unevenly, causing abrupt jumps after long static periods that undermine perceptual coherence. While prior work addresses temporal smoothness or latent interpolation, these methods fail to quantify the rate of semantic change or compare pacing across different generators. To resolve this, the authors introduce the Semantic Progress Function, a one-dimensional representation that measures cumulative semantic shift to identify irregularities. Building on this metric, they propose a model-agnostic linearization procedure that reparameterizes video sequences to ensure transformations unfold at a constant rate without requiring fine-tuning.

Method

The authors introduce the Semantic Progress Function (SPF) as a model-agnostic formulation to capture semantic evolution over time. Formally, given a video consisting of TTT frames {x1,x2,,xT}\{x_1, x_2, \ldots, x_T\}{x1,x2,,xT}, the SPF is defined as a scalar-valued function SiRS_i \in \mathbb{R}SiR mapped from the frame index iii. This representation distills complex visual transformations into a one-dimensional trajectory. The construction proceeds in two stages: first computing pairwise semantic distances between frames, and then integrating these differences over time.

To measure semantic differences, the method utilizes pretrained semantic image embedders such as SigLIP. Each video frame xix_ixi is mapped to a semantic embedding ziRdz_i \in \mathbb{R}^dziRd. The semantic distance between frames iii and jjj is computed using an angular metric in the embedding space: dij=arccos(zizj)d_{ij} = \operatorname{arccos}(z_i^\top z_j)dij=arccos(zizj) The SPF vector SRTS \in \mathbb{R}^TSRT is estimated such that its pairwise temporal differences approximate these semantic distances. This is formulated as a regularized, weighted least-squares objective: minSRT(ASb)W(ASb)+λSS\min_{S \in \mathbb{R}^T} (AS - b)^\top W (AS - b) + \lambda S^\top SminSRT(ASb)W(ASb)+λSS where AAA encodes the linear constraints for frame pairs, bbb collects the distances, and WWW is a diagonal weighting matrix that favors temporally local constraints.

As shown in the figure below, the SPF effectively visualizes semantic pacing. The top graph depicts the SPF of a raw input video where a cat abruptly transforms into a lion. The slope of the function increases sharply at the transition point, reflecting the semantic discontinuity. The bottom graph illustrates the SPF after retiming, where the semantic progression appears significantly steadier.

Building on this analysis, the authors propose semantic linearization via ReTime to reparameterize time so that semantic change progresses at a constant rate. For generated videos, this is achieved by warping the model's temporal positional encodings. The SPF SSS is normalized to [0,1][0, 1][0,1], and warped temporal positions τk\tau_kτk are computed via inversion: τk=S1(kT1)\tau_k = S^{-1}\left( \frac{k}{T-1} \right)τk=S1(T1k) This stretches time in regions of rapid semantic change and compresses stable regions. Since modern video diffusion transformers employ Rotary Position Embeddings (RoPE), the authors introduce frequency-aware warping. Low-frequency bands, which control long-range structure, are warped more strongly, while high-frequency bands remain closer to linear time to preserve local motion smoothness: pt(b)=(1αb)t+αbτtp_t^{(b)} = (1 - \alpha_b) t + \alpha_b \tau_tpt(b)=(1αb)t+αbτt The warping strength αb\alpha_bαb decays exponentially across frequency bands. Additionally, a timestep-dependent modulation applies stronger warping early in the denoising process to concentrate semantic correction during structure formation.

For existing videos where generation is not controllable, the method segments the SPF into piecewise linear components using segmented least squares. This isolates regions of near-constant semantic velocity. Refer to the framework diagram for an example of this process on a cinematic clip, where the segmented SPF (dotted line) guides the redistribution of temporal capacity to smooth abrupt transitions.

Intermediate clips are then regenerated for each segment to ensure uniform pacing. The first and last frames of each segment serve as semantic keyframes for regeneration. This approach allows the use of various open or closed-source models, as long as they can be conditioned on keyframes or first-last frames, ensuring that the duration of each segment is proportional to the magnitude of the semantic change between its boundary frames.

Experiment

The evaluation suite validates the framework by comparing retiming strategies against baselines, applying the method to real cinematic footage, and verifying accuracy through controlled synthetic experiments. Qualitative findings show that operating directly on model features prevents ghosting artifacts and external quality bottlenecks, enabling smooth semantic transitions that baseline methods fail to resolve. Synthetic benchmarks confirm the Semantic Progress Function accurately tracks pacing profiles independent of pixel motion, while ablation studies identify SigLIP as the optimal embedder for capturing semantic shifts. Finally, quantitative metrics and user studies demonstrate that the approach maintains visual fidelity while significantly improving semantic pacing.

The authors evaluate the visual fidelity of their retiming framework by comparing original and retimed outputs from Wan2.2 and LTX-2 models using VBench metrics. The results demonstrate that the retimed videos maintain quality levels nearly identical to the original generations across aesthetic, motion, and temporal dimensions. This confirms that the proposed temporal manipulation preserves the intrinsic visual capabilities of the base models. Aesthetic quality scores for retimed videos remain comparable to the original model outputs. Motion smoothness and temporal fidelity metrics show negligible deviation between original and retimed sequences. The evaluation confirms that the retiming process preserves the visual fidelity of the underlying generative models.

The authors evaluate the visual fidelity of their retiming framework by comparing original and retimed outputs from Wan2.2 and LTX-2 models using VBench metrics. Results demonstrate that retimed videos maintain quality levels nearly identical to the original generations across aesthetic, motion, and temporal dimensions. This confirms that the proposed temporal manipulation preserves the intrinsic visual capabilities of the base models.


KI mit KI entwickeln

Von der Idee bis zum Launch – beschleunigen Sie Ihre KI-Entwicklung mit kostenlosem KI-Co-Coding, sofort einsatzbereiter Umgebung und bestem GPU-Preis.

KI-gestütztes kollaboratives Programmieren
Sofort einsatzbereite GPUs
Die besten Preise

HyperAI Newsletters

Abonnieren Sie unsere neuesten Updates
Wir werden die neuesten Updates der Woche in Ihren Posteingang liefern um neun Uhr jeden Montagmorgen
Unterstützt von MailChimp