Advances in Interactive Video Generation and Animation

The field of interactive video generation and animation is increasingly focused on creating immersive and realistic experiences. Recent developments have centered around improving the controllability and coherence of generated videos, particularly in areas such as game video generation, human-scene interaction, and 3D animation. Notable advancements include the use of hybrid history-conditioned training strategies, joint video-pose diffusion models, and causal-aware reinforcement learning. These innovations have enabled significant improvements in visual fidelity, realism, and action controllability. Some particularly noteworthy papers in this regard include Hunyuan-GameCraft, which introduces a novel framework for high-dynamic interactive video generation in game environments, and GenHSI, which proposes a training-free method for controllable generation of long human-scene interaction videos. AnimaX is also noteworthy for its feed-forward 3D animation framework that bridges the motion priors of video diffusion models with the controllable structure of skeleton-based animation.

Sources

Hunyuan-GameCraft: High-dynamic Interactive Game Video Generation with Hybrid History Condition

GenHSI: Controllable Generation of Human-Scene Interaction Videos

AnimaX: Animating the Inanimate in 3D with Joint Video-Pose Diffusion Models

Causal-Aware Intelligent QoE Optimization for VR Interaction with Adaptive Keyframe Extraction

Rethink Sparse Signals for Pose-guided Text-to-image Generation

PoseMaster: Generating 3D Characters in Arbitrary Poses from a Single Image

FairyGen: Storied Cartoon Video from a Single Child-Drawn Character

Built with on top of