This project page is a collection of papers (and possibly code) that has something to do with human motion and novel view synthesis in 2D and 3D space. There are two major classes of methods when doing it in 2D space:
- Spatial Transformation (using flow information, etc)
- Conditioned Generation (pose label map, etc)
- Texture-space and Geometry generation (UV correspondence, texture generation, using 3D, etc)
- Feature-space manipulation (Spatial attention, kernels, latent embeddings, etc)
Papers are also split into several sub-categories (ones with the project page):
- Novel View Synthesis - This category is about rendering/generating an image of a human within another pose, may be a single picture or a video
- Motion Retargeting, Representation, Continuation - This category is about the modeling of human motion as well as skeletal prediction. May also include temporally-coherent pose estimation
- 3D Human Digitization - From a(n ideally) single image, this category deals with 3D human generation along with texture and/or shape inference from unseen angles
Papers that do not have first-party implementations are put into another category with two sub-categories:
- Architectural - Contribution is mainly about architecture
- Representation - Contribution is related to information representation, spatial representation, or feature extraction methods
This project page is only intended to supplement my studies and research where everything is (hopefully nicely) organized into one place.
- [ICCV 2019] Everybody Dance Now
- [ICCV 2019] Liquid Warping GAN: A Unified Framework for Human Motion Imitation, Appearance Transfer and Novel View Synthesis
- [CVPR 2019] Progressive Pose Attention for Person Image Generation
- [NeurIPS 2019] First Order Motion Model for Image Animation
- [CVPR 2020] Deep Image Spatial Transformation for Person Image Generation
- [arXiv 2020] Liquid Warping GAN with Attention: A Unified Framework for Human Image Synthesis
- [CVPR 2021] Few-Shot Human Motion Transfer by Personalized Geometry and Texture Modeling
- [CVPR 2021] Neural Body: Implicit Neural Representations with Structured Latent Codes for Novel View Synthesis of Dynamic Humans
- [CVPR 2020] Dynamic Multiscale Graph Neural Networks for 3D Skeleton-Based Human Motion Prediction
- [CVPR 2020] TransMoMo: Invariance-Driven Unsupervised Video Motion Retargeting
- [CVPR 2020] Attention Mechanism Exploits Temporal Contexts: Real-time 3D Human Pose Reconstruction
- [SIGGRAPH 2020] XNect: Real-time Multi-Person 3D Motion Capture with a Single RGB Camera
- [IEEE FnG+TBIOM 2020] Head2Head: Video-based Neural Head Synthesis
- [NeurIPS 2020] Residual Force Control for Agile Human Behavior Imitation and Extended Motion Synthesis
- [CVPR 2021] We Are More Than Our Joints: Predicting How 3D Bodies Move
- [ICCV 2019] DeepHuman: 3D Human Reconstruction from a Single Image
- [ICCV 2019] PIFu: Pixel-Aligned Implicit Function for High-Resolution Clothed Human Digitization
- [CVPR 2019] Photo Wake-Up: 3D Character Animation from a Single Photo
- [CVPR 2019] Textured Neural Avatars
- [CVPR 2020] PIFuHD: Multi-Level Pixel-Aligned Implicit Function for High-Resolution 3D Human Digitization
- [CVPR 2020] CAPE: Clothed Auto-Person Encoding
- [NeurIPS 2020] Geo-PIFu: Geometry and Pixel Aligned Implicit Functions for Single-view Human Reconstruction
Third party implementations (if any) will be put in second-order bullet points
- [CVPR 2018] Synthesizing Images of Humans in Unseen Poses
- [CVPR 2019] Coordinate-based Texture Inpainting for Pose-Guided Human Image Generation
- [arXiv 2020] Human Motion Transfer with 3D Constraints and Detail Enhancement
- [CVPR 2021] SMPLicit: Topology-aware Generative Model for Clothed People
- [CVPR 2021] Face-Vid2Vid: One-Shot Free-View Neural Talking-Head Synthesis for Video Conferencing