This is the official repository of V2M4. We're actively organizing the code to enhance the user experience—stay tuned for updates! 🚀
If you encounter any question about the paper, please feel free to contact us. You can create an issue or just send email to me [email protected]. Also welcome for any idea exchange and discussion.
[03/18/2025] Repository init.
- Code release
We present V2M4, a novel 4D reconstruction method that directly generates a usable 4D mesh animation asset from a single monocular video. Unlike existing approaches that rely on priors from multi-view image and video generation models, our method is based on native 3D mesh generation models. Naively applying 3D mesh generation models to generate a mesh for each frame in a 4D task can lead to issues such as incorrect mesh poses, misalignment of mesh appearance, and inconsistencies in mesh geometry and texture maps. To address these problems, we propose a structured workflow that includes camera search and mesh reposing, condition embedding optimization for mesh appearance refinement, pairwise mesh registration for topology consistency, and global texture map optimization for texture consistency. Our method outputs high-quality 4D animated assets that are compatible with mainstream graphics and game software. Experimental results across a variety of animation types and motion amplitudes demonstrate the generalization and effectiveness of our method.
If you find this paper useful in your research, please consider citing:
@article{chen2025v2m4,
title={V2M4: 4D Mesh Animation Reconstruction from a Single Monocular Video},
author={Chen, Jianqi and Zhang, Biao and Tang, Xiangjun and Wonka, Peter},
journal={arXiv preprint arXiv:2503.09631},
year={2025}
}
This project is licensed under the MIT license. See LICENSE for details.