Advances in 3D Vision and Human Motion Modeling

The field of 3D vision and human motion modeling is rapidly advancing, with a focus on developing more accurate and efficient methods for tasks such as 3D shape completion, human pose estimation, and motion generation. Recent research has explored the use of generative models, diffusion-based methods, and transformation-based approaches to improve the accuracy and robustness of these tasks. Notably, the use of latent generative paradigms and autoregressive diffusion models has shown promising results in 3D shape completion and motion generation. Additionally, the development of new datasets and evaluation metrics has enabled more comprehensive assessment of model performance. Some noteworthy papers in this area include: Evaluating Latent Generative Paradigms for High-Fidelity 3D Shape Completion, which compares the performance of different generative models for 3D shape completion; Free3D, which proposes a framework for synthesizing realistic 3D motions without 3D motion annotations; and TriDiff-4D, which introduces a novel 4D generative pipeline for producing high-quality, temporally coherent 4D avatars.

Sources

Evaluating Latent Generative Paradigms for High-Fidelity 3D Shape Completion from a Single Depth Image

YCB-Ev SD: Synthetic event-vision dataset for 6DoF object pose estimation

Free3D: 3D Human Motion Emerges from Single-View 2D Supervision

Rethinking Multimodal Point Cloud Completion: A Completion-by-Correction Perspective

Towards Metric-Aware Multi-Person Mesh Recovery by Jointly Optimizing Human Crowd in Camera Space

RSPose: Ranking Based Losses for Human Pose Estimation

B2F: End-to-End Body-to-Face Motion Generation with Style Reference

StreamingTalker: Audio-driven 3D Facial Animation with Autoregressive Diffusion Model

BEDLAM2.0: Synthetic Humans and Cameras in Motion

Simba: Towards High-Fidelity and Geometrically-Consistent Point Cloud Completion via Transformation Diffusion

TriDiff-4D: Fast 4D Generation through Diffusion-based Triplane Re-posing

NoPo-Avatar: Generalizable and Animatable Avatars from Sparse Inputs without Human Poses

Built with on top of