Neural Pose Representation Learning for Generating and Transferring
Non-Rigid Object Poses

KAIST

Teaser. Results of motion sequence transfer (left) and shape variation generation (right) using the proposed neural pose representation. On the left, poses from source shapes (first and third rows) are transferred to target shapes (second and fourth rows), preserving intricate details like horns and antlers. On the right, new poses sampled from a cascaded diffusion model, trained with shape variations of the bunny (last column), are transferred to other animal shapes.

Abstract

We propose a novel method for learning representations of poses for 3D deformable objects, which specializes in 1) disentangling pose information from the object’s identity, 2) facilitating the learning of pose variations, and 3) transferring pose information to other object identities. Based on these properties, our method enables the generation of 3D deformable objects with diversity in both identities and poses, using variations of a single object. It does not require explicit shape parameterization such as skeletons or joints, point-level or shape-level correspondence supervision, or variations of the target object for pose transfer. We first design the pose extractor to represent the pose as a keypoint-based hybrid representation and the pose applier to learn an implicit deformation field. To better distill pose information from the object’s geometry, we propose the implicit pose applier to output an intrinsic mesh property, the face Jacobian. Once the extracted pose information is transferred to the target object, the pose applier is fine-tuned in a self-supervised manner to better describe the target object’s shapes with pose variations. The extracted poses are also used to train a cascaded diffusion model to enable the generation of novel poses. Our experiments with the DeformThings4D and Human datasets demonstrate state-of-the-art performance in pose transfer and the ability to generate diverse deformed shapes with various objects and poses.

Method


reformulation

Method Overview. Our framework extracts keypoint-based hybrid pose representations from Jacobian fields. These fields are mapped by the pose extractor $g$ and mapped back by the pose applier $h$. The pose applier, conditioned on the extracted pose, acts as an implicit deformation field for various shapes, including those unseen during training. A refinement module $\alpha$, positioned between $g$ and $h$, is trained in a self-supervised manner, leveraging the target's template shape. The compactness of our latent representations facilitates the training of a diffusion model, enabling diverse pose variations through generative modeling in the latent space.


Qualitative Results Using DeformingThings4D-Animals

Deform4d_Results Deform4d_Results Deform4d_Results Deform4d_Results




Qualitative Comparisons Using SMPL Human Body Shapes

SMPL_Results SMPL_Results SMPL_Results SMPL_Results


Qualitative Results Using Adobe Mixamo

Mixamo_Results Mixamo_Results Mixamo_Results Mixamo_Results



Citation

Please consider citing our work if you find it useful.


      @inproceedings{yoo2024neuralpose,
        title = {{Neural Pose Representation Learning for Generating and Transferring Non-Rigid Object Poses}},
        author = {Yoo, Seungwoo and Koo, Juil and Yeo, Kyeongmin and Sung, Minhyuk},
        booktitle = {NeurIPS},
        year = {2024},
      }