Streaming Video Generation · USTC · FrameX.AI

Stream-R1 Reliability-Perplexity Aware Reward Distillation for streaming video generation

Bin Wu1 Mengqi Huang1†‡ Shaojin Wu3 Weinan Jia1 Yuxin Wang2 Zhendong Mao1 Yongdong Zhang1
1 University of Science and Technology of China 2 FrameX.AI 3 Independent Researcher
Corresponding author Project lead

Hover any clip to pause · click to play fullscreen

Abstract

Stream-R1 reweights distribution-matching distillation along two complementary axes — Inter-Reliability across rollouts and Intra-Perplexity across spatiotemporal regions — using a single shared video reward model.

Existing DMD methods for streaming video diffusion treat every rollout, frame, and pixel as equally reliable supervision. We argue this caps the upper bound of distilled quality, because it overlooks two complementary axes of variance in the DMD signal. Stream-R1 rescales each rollout’s loss by an exponential of a pretrained reward score — rollouts whose gradient genuinely points toward the teacher’s high-quality mode dominate — and back-propagates the same reward model to obtain per-pixel saliency that concentrates optimization where the local reward landscape has not yet flattened. An adaptive balancing mechanism keeps visual quality, motion quality, and text alignment improving in lockstep. The resulting student surpasses its multi-step Wan2.1 teacher on VBench Total and Semantic at 23.1 FPS, with no architectural change and zero inference overhead.

01

Inter-Reliability Weighting

The DMD gradient g = ffake − freal varies in reliability across rollouts. Stream-R1 rescales each rollout’s loss by exp(β · rfinal), so rollouts whose gradient genuinely points toward the teacher’s high-quality mode dominate the supervision.

02

Intra-Perplexity Weighting

Back-propagating the reward model yields a per-pixel saliency volume S ∈ RF×H×W; factorized into spatial and temporal weights, it concentrates optimization where further refinement yields the largest expected gain — regions whose local reward landscape has not yet flattened.

03

Adaptive Reward Balancing

A sliding window tracks per-axis (Visual / Motion / Text-align) improvement and softly down-weights axes that are already improving, keeping all three quality dimensions advancing together.

Motivation

Two complementary axes of variance

The DMD supervision signal varies across rollouts — the gradient on rollouts far from the teacher’s high-quality mode encodes within-low-quality refinement rather than a path toward it — and across spatiotemporal regions, since some areas have already saturated under the reward while others still lie in the steepest part of the reward landscape. Treating them uniformly dilutes the supervision and wastes optimization budget on already-saturated regions while leaving high-perplexity content under-supervised.

Motivation of Stream-R1
Figure 1. The DMD supervision signal exhibits Inter-Reliability across rollouts and Intra-Perplexity across spatiotemporal regions. Existing DMD applies a uniform weight; Stream-R1 upweights rollouts on which the supervision is reliable and concentrates optimization on regions where further refinement yields the largest expected gain, all driven by a single reward model.

Method

A single reward model, two complementary weights

Stream-R1 retains the tractability of the DMD objective while replacing its uniform weighting with reliability- and perplexity-aware guidance. The same reward model produces both a rollout-level scalar and a per-pixel gradient saliency volume, with adaptive fusion across visual, motion, and text-alignment axes.

Stream-R1 method overview
Figure 2. Inter-Reliability score extraction yields Winter; adaptive saliency aggregation factorizes the back-propagated reward into spatial and temporal weights forming Wintra. The composed loss is LStream-R1 = Winter · (Wintra ⊙ LDMD).

Long Video Generation

Four durations, no quality drift

Each row plays clips that were independently rolled out at a different duration — 30 seconds, 60 seconds, 2 minutes, 3 minutes. Hover a row to pause its scroll and zoom the clip under your cursor; click any clip to play it fullscreen.

30seconds
60seconds
2minutes
3minutes

Method Comparison

Reward Forcing vs. Stream-R1

Same prompt, same seed. Reward Forcing applies a single global scalar reward; Stream-R1 additionally redistributes the optimization signal across reliable rollouts and high-perplexity regions. Each row plays the two clips in sync the moment it scrolls into view.

Reward Forcing Stream-R1 (ours)
Baseline
Ours
Baseline
Ours
Baseline
Ours
Baseline
Ours
Baseline
Ours
Baseline
Ours

Spatiotemporal Saliency · Controlled Stress Test

Where the reward gradient looks

To probe whether the spatiotemporal weights actually respond to local quality deficiency, we run a controlled experiment: Gaussian blur is injected only into the lower half of each frame, so every frame contains a clean (top) versus degraded (bottom) contrast. Across the four frames, the corrupted area progressively expands from left to right. We then back-propagate the reward score through the vision encoder to recover a per-pixel gradient saliency. Drag the slider to step through.

Degraded frame
Frame 1: small blur region in lower half Frame 2: growing blur region Frame 3: larger blur region Frame 4: largest blur region
Reward saliency overlay
Frame 1 saliency overlay Frame 2 saliency overlay Frame 3 saliency overlay Frame 4 saliency overlay
Per-frame temporal weight wt 0.587

Within each frame, saliency is biased toward the lower (blurred) half rather than the visually intact upper half. As the blurred region grows, the spatial saliency tightens onto its interior. Per-frame temporal weights climb monotonically from 0.587 to 2.117 — not hand-engineered, this behavior emerges purely from the reward-model gradient, automatically allocating more learning signal to frames that need it most.

Quantitative Results

Beating the multi-step teacher at 30× speed

Despite being distilled into a 4-step model, Stream-R1 surpasses its own diffusion teacher Wan2.1 on VBench Total and Semantic, while improving on Reward Forcing across every axis at the same throughput.

VBench — 5-second video, 832×480

Best in highlighted mark. Stream-R1 achieves the highest Total of any compared method, surpassing both its diffusion teacher (84.26) and the previous best distilled model Reward Forcing (84.13).

Model Params FPS↑ Total↑ Quality↑ Semantic↑
Diffusion models
LTX-Video1.9B8.9880.0082.3070.79
Wan2.1 (teacher)1.3B0.7884.2685.3080.09
Autoregressive / streaming models
SkyReels-V21.3B0.4982.6784.7074.53
MAGI-14.5B0.1979.1882.0467.74
NOVA0.6B0.8880.1280.3979.05
Pyramid Flow2B6.781.7284.7469.62
CausVid1.3B17.082.8883.9378.69
Self Forcing1.3B17.083.8084.5980.64
LongLive1.3B20.783.2283.6881.37
Rolling Forcing1.3B17.581.2284.0869.78
Reward-guided distillation
Reward Forcing1.3B23.184.1384.8481.32
Stream-R1 (ours) 1.3B 23.1 84.40 85.14 81.44

Qwen3-VL judgement — 60-second video

Each video scored 1–5 on three axes by a strong VLM judge. Stream-R1 attains the best Visual Quality and Text Alignment; we trade a small Dynamic margin for gains on the other two axes, consistent with the balanced multi-dimensional design.

Model Visual↑ Dynamic↑ Text↑
SkyReels-V23.303.052.70
CausVid4.663.163.32
Self Forcing3.893.443.11
LongLive4.793.813.98
Reward Forcing4.824.184.04
Stream-R1 (ours) 4.92 4.04 4.11
Per-metric quality versus video length
Figure 3. Per-metric quality vs. video length. Stream-R1 widens its lead over Reward Forcing as duration grows — especially at 120s and 180s, confirming that spatiotemporal reward-guided weighting mitigates the quality drift accumulated during long autoregressive rollouts.

Ablation

Each component contributes

We progressively add each component on top of the standard DMD baseline, evaluating short videos on VBench (5s) and long videos on a 60-second protocol. Drift measures autoregressive quality decay (lower is better).

Variant VBench (5 s) Long Video (60 s)
Total↑Quality↑Semantic↑ Total↑Drift↓
DMD baseline 83.4484.1680.5579.452.479
+ Spatial reward Ws  min=0.15) 83.6784.4680.5180.712.653
+ Balanced multi-dim reward 83.6884.4480.6280.732.697
+ Temporal reward Wt (full) 84.40 85.14 81.44 80.86 2.417

BibTeX

Cite Stream-R1

Citation forthcoming — we’ll publish a BibTeX entry once the preprint is available on arXiv.