Next Article in Journal
Blind Source Separation with Strength Pareto Evolutionary Algorithm 2 (SPEA2) Using Discrete Wavelet Transform
Next Article in Special Issue
Fault Detection in Solar Energy Systems: A Deep Learning Approach
Previous Article in Journal
A Proposed Single-Input Multi-Output Battery-Connected DC–DC Buck–Boost Converter for Automotive Applications
Previous Article in Special Issue
Global Individual Interaction Network Based on Consistency for Group Activity Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deform2NeRF: Non-Rigid Deformation and 2D–3D Feature Fusion with Cross-Attention for Dynamic Human Reconstruction

1
School of Mathematics and Computer Science, Nanchang University, Nanchang 330031, China
2
School of Information Engineering, Nanchang University, Nanchang 330031, China
3
School of Software, Nanchang University, Nanchang 330031, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2023, 12(21), 4382; https://doi.org/10.3390/electronics12214382
Submission received: 19 September 2023 / Revised: 12 October 2023 / Accepted: 16 October 2023 / Published: 24 October 2023
(This article belongs to the Special Issue Advances of Artificial Intelligence and Vision Applications)

Abstract

:
Reconstructing dynamic human body models from multi-view videos poses a substantial challenge in the field of 3D computer vision. Currently, the Animatable NeRF method addresses this challenge by mapping observed points from the viewing space to a canonical space. However, this mapping introduces positional shifts in predicted points, resulting in artifacts, particularly in intricate areas. In this paper, we propose an innovative approach called Deform2NeRF that incorporates non-rigid deformation correction and image feature fusion modules into the Animatable NeRF framework to enhance the reconstruction of animatable human models. Firstly, we introduce a non-rigid deformation field network to address the issue of point position shift effectively. This network adeptly corrects positional discrepancies caused by non-rigid deformations. Secondly, we introduce a 2D–3D feature fusion learning module with cross-attention and integrate it with the NeRF network to mitigate artifacts in specific detailed regions. Our experimental results demonstrate that our method significantly improves the PSNR index by approximately 5% compared to representative methods in the field. This remarkable advancement underscores the profound importance of our approach in the domains of new view synthesis and digital human reconstruction.

1. Introduction

Recently, obtaining a three-dimensional model of the human body through dynamic video or multi-view images has become a popular research topic in the computer vision field. This technology has significant applications in the meta-universe, virtual reality, animation, and more. However, modeling the dynamic human body solely based on multi-view images [1,2,3,4] is challenging because human motion typically involves non-rigid transformations. Additionally, due to occlusion and limited views, accurately estimating human poses from multi-view images is an extremely difficult problem [5].
With the increasing popularity of NeRF research, one of the mainstream approaches is to render new perspectives and postures using NeRF [6,7,8]. Studies such as NT [9], NHR [10], and Animatable NeRF [1] utilize multi-view images to obtain rendered new-view images and create a rough human body model. Animatable NeRF also introduces the SMPL [11] model as an a priori model for the first time, which plays a role in constraining the movement of points and has achieved good results. However, the SMPL model has limited effectiveness in modeling large non-rigid deformations of the human body. Previous studies [1,2,3,12] also have certain artifacts and issues with view inconsistency. Moreover, these studies only utilized multi-view images for light sampling and did not effectively utilize 2D–3D feature fusions. Therefore, we aim to leverage 2D–3D feature fusion more effectively to eliminate artifacts and reduce view inconsistency.
We propose a method called Deform2NeRF, which builds upon Animatable NeRF by introducing two new modules: the non-rigid deformation field and the 2D–3D feature fusion field. The non-rigid offset correction field corrects the offset problem in non-rigid transformation by addressing large-scale motion. The 2D–3D feature fusion fusion field extracts features from each image and more effectively utilizes the features of the image, thereby alleviating problems with artifacts and view inconsistency. We tested and evaluated our designed network on ZJU-Mocap [2] and H36M [13] datasets, and the results demonstrate that our design is reasonable and has achieved remarkable results. The results are shown in Figure 1.
In summary, our contributions are as follows:
  • We propose a new method called Deform2NeRF to correct the offset in non-rigid deformation and extract more information.
  • We propose a 2D–3D feature fusion field to reduce artifacts and inconsistent views.
  • Our Deform2NeRF network can achieve good results without additional training for the synthesis of new postures, indicating that our model is robust.

2. Related Work

2.1. Neural Radiance Field

NeRF [6,7,8] proposes to render new perspectives by inputting the 5D coordinate points of an object and outputting the density and RGB values of the point cloud, thereby implicitly representing static objects or scenes. Thanks to the introduction of the volume rendering formula [14], NeRF has achieved excellent results and has become a hot research topic in computer vision. NeRF++ [15] extends NeRF to a wide range of boundaryless scenarios. Plenoxels, Instant-NGP, and other works [6,8,16] have optimized the sampling and training strategies of NeRF, greatly reducing the training time. Consistent-NeRF [17] employs depth-derived geometry information and a depth-invariant loss to concentrate on pixels that exhibit 3D correspondence and maintain consistent depth relationships. This significantly improves the performance of the model under sparse view conditions. Head NeRF [3] applies NeRF to the head and achieves real-time high-fidelity reconstruction of the human head. D-NeRF [18] designed a deep learning model to implicitly encode a scene and synthesize novel views at an arbitrary time. Other methods [1,9,10,19,20,21] extend NeRF from static scenes to moving human bodies, providing new ideas for the generation of digital human images and expanding the application range of NeRF.

2.2. Three-Dimensional Reconstruction of Human Bodies Based on NeRF

The 3D reconstruction of dynamic human bodies based on NeRF has recently gained significant attention in the 3D computer vision community. Several approaches [21,22,23] have been proposed to enhance the capability of NeRF in capturing the non-rigid deformation of human bodies. For instance, NHR [10] extracts 3D point cloud features through PointNet++ [24] to achieve human body rendering. Neural Human Performer [19] utilizes an attention mechanism to solve the non-rigid transformation of the human body. Neural Body [2] describes the motion state of the human body by introducing potential coding. Animatable NeRF [1] introduces the SMPL model as an a priori model to constrain the human bone structure to some extent. However, since human body movements are complex non-rigid deformations, these methods may generate artifacts in the details. The SLRF [23] method addresses this issue by introducing structured local radiance fields to better describe the details of the folds of clothing on human bodies. SelfNeRF [20] incorporates KNN and hash coding to constrain the movement of human point clouds and improve computational efficiency. Nevertheless, effectively computing the non-rigid deformation of the human body, and eliminating artifacts while preserving details, remains a challenging problem to be solved.

3. Methods

3.1. Neural Blend Weight Fields

Animatable NeRF [1] proposes the use of neural blend weight fields based on a three-dimensional human skeleton and skeleton-driven deformation framework [25] to solve the problem of under-constraint in human deformation.
Specifically, it defines the human skeleton as K parts [11] and generates the K transform matrix { G k } S E ( 3 ) . Using the linear hybrid peeling algorithm, the point x can in the canonical space [22] can be transformed into the observation space x obs . The specific formula is defined as:
x obs = k = 1 K w ( x can ) k G k x can ,
Similarly, we can also convert points in the observation space to points in the standard space. Our method is outlined in Figure 2. w o ( x ) is a hybrid weight function defined in the observation space.
x can = k = 1 K w o ( x obs ) k G k 1 x obs
However, training the mixed weight field with the NeRF network does not achieve good results. Therefore, Animatable NeRF [1] samples for any three-dimensional point by first assigning the initial mixed weights according to the body model, and then uses the residual vector learned by the network to correct the model. The residual network is represented by an MLP (Multi-Layer Perception Machine) network:
F Δ w ( x , ψ i ) Δ w i
where ψ is the potential code obtained by the inter-frame embedding layer, and the residual vector Δ w i R K . Thus, we can define the neural mixed weight field w i of the i-th image, where w s represents the initial neural blender weights and S i denotes the human body model:
w i ( x ) = norm F Δ w x , ψ i + w s x , S i ,
In this way, we obtain the blended weight corresponding to this picture, allowing us to realize the mutual mapping of sampling points in canonical space and observation space [26]. That is, we map the sampling point x in the observation space to the canonical space to obtain the point x , and input it into the NeRF networks F c and F σ to obtain the color c i ( x ) and opacity σ i ( x ) of the point x for view direction d as follows:
σ i ( x ) , z i ( x ) F σ γ x x c i ( x ) F c z i ( x ) , γ d ( d ) , i
where z i ( x ) denotes the shape feature of the human body, i denotes the latent code, and γ denotes the position encoding.
Then, we use the volume rendering formula to render [14]. In this manner, we derive the color information C ^ ( r ) for each pixel in the generated image from this view direction:
C ^ ( r ) = i = 1 N T i 1 exp σ i δ i c i , where T i = exp j = 1 i 1 σ j δ j
where δ i denotes the distance between successive points and N denotes the number of sampled point clouds in the ray sampling. c and σ are the RGB value and density of the point cloud, respectively.

3.2. The 2D–3D Feature Fusion Network with Cross-Attention

We observe that previous approaches in human NeRF only utilize image features in light sampling as ground truth, neglecting the potential of using image feature information to guide network training. We consider this as a missed opportunity to leverage valuable features. To address this limitation, we propose to extract feature information from the multi-view image I using a well-established ResNet [27] network:
f 2 D = R e s N e t ( I )
By extracting these features, we can effectively discriminate the actions depicted in each picture. It is important to note that each frame of the picture corresponds to a three-dimensional point cloud model, denoted as S. However, these models may exhibit roughness or contain certain interference points, thereby affecting the learning process of the NeRF network to a certain extent. To overcome this challenge, we propose the extraction of global features from the point cloud model through a dedicated point cloud feature network [24,28,29]. By capturing the global characteristics of the point cloud, we aim to represent the human body using these aggregated features. This approach enables the NeRF network to better handle the impact of local irregularities or interference points, thereby enhancing the overall reconstruction performance:
f 3 D = P o i n t N e t + + ( S )
The fusion of 2D and 3D features is achieved by merging their respective feature representations. Notably, N represents the number of 3D point clouds, g θ denotes the feature extraction network, and ⊗ represents the cross-attention and feature fusion method:
f = g θ ( [ f 3 D r e p e a t ( f 2 D , N ) ] )
Subsequently, the fused feature f is incorporated into the NeRF network, allowing for enhanced constraints on human motion. This fusion feature plays a crucial role in capturing both the spatial information from the 3D point clouds and the visual characteristics from the 2D images. By leveraging this combined feature representation, our approach can effectively constrain the modeling of human motion within the NeRF framework, leading to more accurate and realistic reconstructions.

3.3. Non-Rigid Deformation Field with NeRF

It is well-known that tracking non-rigid deformations in human motion presents a significant challenge. While methods such as the neural mixed weight field proposed in Animatable NeRF [1] have been developed, serious artifacts may still occur for non-rigid deformations with large amplitudes. Building upon the insights gained from SLRF [23], we propose the use of a non-rigid deformation field to correct for any offset induced by non-rigid deformation. Specifically, we introduce an MLP network, denoted as Δ x , defined as follows:
Δ x i = F Δ x ( P i , i , t i )
where P represents the SMPL [11] model parameters, t corresponds to the timestamp of the image, and is the corresponding potential coding.
To further improve the performance of NeRF, we aim to incorporate 2D–3D feature fusions into the rendering process. Specifically, we propose to extract 2D–3D feature fusions using a 2D–3D feature fusion extraction network; the 2D–3D feature fusions are then integrated into the NeRF network to address the issues of artifacts and inconsistent views. Currently, NeRF only utilizes multi-view images for light sampling, without exploiting their inherent features. Therefore, we propose to incorporate these 2D–3D feature fusions to further enhance the rendering process.
After extracting the 2D–3D feature fusions, we integrate them into the NeRF network to use the two-dimensional features of the image and fuse them with the three-dimensional point cloud features, as follows:
σ i ( x ) , z i ( x ) = F σ γ x x + Δ x c i ( x ) = F c z i ( x ) , γ d ( d ) , i , f
The above networks are optimized jointly through the NeRF network, and the blend weight field is continuously updated:
w new x , ψ new = norm F Δ w x , ψ new + w s x , S new
By jointly optimizing the above networks through the NeRF network, we are able to continuously track the deformation of human motion and update the parameters of the network. This enables us to effectively address the artifacts caused by non-rigid deformation and the blurring of details.
Following the acquisition of more precise color and opacity data for each point, we employ the well-established NeRF strategy to synthesize a composite view from this perspective using the volume rendering equation (Equation (6)). Subsequently, the synthesized image undergoes a comparative analysis with the ground truth to compute the loss function:
L rgb = r R C ˜ i ( r ) C i ( r ) 2 ,

4. Results

4.1. Datasets

To evaluate the effectiveness of our proposed network, we conducted experiments on the widely used H36M [13] and ZJU-Mocap datasets [2]. Following the standard evaluation strategy for human NeRF, to facilitate an equitable comparison, we have embraced a congruent evaluation strategy and utilized identical indicators as the predecessor baseline method. Additionally, the dataset employed holds status as a widely accepted benchmark for assessing the reconstruction of human dynamics. We used the ZJU-Mocap dataset (313, 315, 377, 386), which uses 21 cameras to capture multi-view videos. We selected four views as training data and the remaining views for evaluation. Similarly, for the H36M dataset, which includes complex human actions, we used S1, S5, S6, S7, S8, S9, and S11 as our data sets, where we used three camera views for training and the remaining views for testing. Two common metrics, PSNR and SSIM, were used to assess the performance of our model in new view synthesis and to compare it with previous works.

4.2. Evaluation

To evaluate the effectiveness of our approach in the field of 3D vision, we conducted evaluation and ablation experiments using two widely used datasets, H36M and ZJU-Mocap. Following the evaluation strategies used in previous studies [1] on human NeRF, we used ZJU-Mocap, which includes 21 cameras and records multi-view videos. We selected four viewpoints for training and used the remaining viewpoints for testing. Similarly, for the H36M dataset, we used multi-view videos recorded using four cameras, including a series of complex human actions. We used S1, S5, S6, S7, S8, S9, and S11 as datasets, where we selected images from three viewpoints for training [17] and used the remaining images for testing.
We compared our approach with previous methods such as Neural Body and Animatable NeRF, and conducted quantitative analysis using two common metrics, PSNR and SSIM. The results of our experiments are presented in Table 1 and demonstrate the effectiveness of our approach. As shown in Figure 3, the images rendered by our model have a more complete human body.
In the ZJU-Mocap dataset, we selected four characters (313, 315, 377, 386) to evaluate our approach for new viewpoint and new pose synthesis. Table 1 presents the quantitative analysis of the new viewpoint assessment, while Table 2 displays the results of the new pose assessment. Additionally, Table 3 shows the outcomes of the new view assessment. Similarly, we evaluated our approach on the H36M dataset and presented the results of image comparison, which show that our approach effectively eliminates artifacts.

4.3. Ablation Studies

To validate the effectiveness of our network, we conducted ablation experiments on the H36M dataset, specifically on the S11 tester. In these experiments, we analyzed the impact of the offset correction network and the feature extraction network. Additionally, we examined the influence of the time step parameter t on the results of synthesizing new perspectives and new postures of the human body. Furthermore, we explored the effects of different time steps and training rounds on the outcomes. The results of these ablation experiments are summarized in Table 4, Table 5 and Table 6.
Impact of the non-rigid deformation field network. In Table 4, we present a comparison of the performance of the non-rigid deformation correction network. The results clearly demonstrate that our proposed offset network is capable of accurately capturing and describing the complex non-rigid deformations of the human body. Furthermore, it effectively alleviates artifacts that may arise during the rendering process. This analysis confirms the effectiveness of our offset correction network in improving the quality of synthesized human body representations.
Impact of the 2D–3D feature fusion fusion network. To effectively utilize the features present in the images, we devised a 2D–3D feature fusion extraction network. The experimental results substantiate the benefits of incorporating 2D–3D feature fusion information, as it significantly improves the rendering quality. Table 5 provides a quantitative comparison, further supporting the superior performance achieved through the integration of 2D–3D feature fusions into our approach.
Impact of the time stamp. To enhance the temporal coherence and stability of the variables, particularly the residual position mapping of the sampled points, we incorporated the time step of the input images into our network. This addition enables a more accurate depiction of the image sequence. Table 6 visually illustrates the impact of different time steps on the rendering results, further highlighting the significance of considering temporal information in our approach.

5. Discussion

Previous methods in the field of human NeRF have demonstrated certain limitations in effectively tracking and estimating non-rigid deformations caused by human motion. To address this challenge, we introduced the non-rigid correction module into our framework. However, it is important to note that this problem requires further exploration and improvement, potentially through the incorporation of additional feature point information or a more accurate human reference SMPL model. While our proposed method alleviates the artifact problem to a certain extent, future research may benefit from the integration of techniques such as KNN [20] to further constrain human deformation and enhance the overall performance of human NeRF.
Furthermore, we observed that previous human NeRF approaches did not fully exploit the potential of 2D–3D feature fusion information and point cloud feature information in the 3D model. In our work, we designed a feature fusion network to extract and integrate two-dimensional and three-dimensional features, aiming to achieve better results. The under-constrained nature of human NeRF and the sparsity of input images highlight the importance of extracting as much information as possible from 2D images. Maximizing the utilization of feature information from both image and 3D point cloud sources may prove to be a crucial factor in improving the quality of reconstruction results.
In conclusion, while our proposed method represents a step forward in addressing the challenges of human NeRF, further advancements are needed to fully capture and model non-rigid deformations. Additionally, exploring more effective methods to leverage feature information and extracting comprehensive information from limited input images are areas that can contribute to improving the overall effectiveness of human NeRF in future research.

6. Conclusions

In conclusion, our proposed method, Deform2NeRF, represents a significant advancement in the field of computer vision, particularly in the context of dynamic human body modeling using multi-view images. We have addressed several key challenges associated with this technology and made notable contributions:
(1)
Offset correction for non-rigid deformation: We introduced a novel module, the non-rigid deformation field, which effectively corrects the offset problem inherent in non-rigid transformations. This module addresses large-scale motion and significantly improves the accuracy of dynamic human body modeling.
(2)
The use of 2D–3D feature fusion with a cross-attention network: Our 2D–3D feature fusion field method is a pioneering approach that extracts and utilizes features from each image more effectively. Instead of merely using multi-view images for light sampling, this module fuses 2D and 3D features to mitigate problems related to artifacts and view inconsistency.
(3)
Improved model generalization: Our experiments clearly demonstrate the effectiveness of our approach in eliminating artifacts and enhancing the generalization ability of the model. Notably, our method obviates the need for separate training to synthesize new poses, showcasing its robustness and versatility.
(4)
Outstanding results on benchmark datasets: We conducted rigorous evaluations on the ZJU-Mocap and H36M datasets, and our method consistently outperformed previous state-of-the-art approaches. This indicates the practicality and real-world applicability of our Deform2NeRF model.
However, it is important to acknowledge the following limitations of our method:
(1)
Challenges with non-rigid clothing: Our method may encounter difficulties in accurately reconstructing non-rigid clothing or fabrics, as it primarily focuses on modeling the human body itself.
(2)
Sensitivity to lighting conditions: Like many computer vision techniques, our model may be sensitive to variations in lighting conditions, potentially affecting the quality of the reconstructed models.
(3)
Statistical body model: Our study employs the SMPL model as the foundational framework for representing the human body—a prevalent approach in the realm of human body reconstruction. It is essential to acknowledge that alternative methodologies, including, but not limited to, STAR [30] and SMPL-X [31], have simpler and more efficient means of human body model representation, potentially yielding superior results.
Despite these limitations, Deform2NeRF represents a promising step forward in dynamic human body modeling. We believe that ongoing research and development can address these challenges and further enhance the capabilities of our approach, making it even more valuable for applications in the meta-universe, virtual reality, animation, and related fields.

Author Contributions

Conceptualization, X.X. and X.G.; methodology, X.G. and X.X.; software, X.G. and X.X.; validation, X.G. and X.X.; formal analysis, X.X., X.G. and W.L.; investigation, X.X. and X.G.; resources, J.L. and J.X.; data curation, X.G.; writing—original draft preparation, X.G. and X.X.; writing—review and editing, W.L. and X.G.; visualization, X.X.; supervision, J.L. and J.X.; project administration, W.L.; funding acquisition, W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Foundations of China (Grant No. 62361041) awarded to Wei Li, the National Natural Science Foundation of Jiangxi Province (NSFC) under Grant 20212BAB212012 awarded to Wei Li, the National Natural Science Foundations of China (Grant No. 62266032) awarded to Yanyang Xiao, the National Natural Science Foundation of China under Grant 62102174 awarded to Jianfeng Xu, and the Jiangxi Training Program for Academic and Technical Leaders in Major Disciplines-Leading Talents Project (Grant No. 20225BCI22016) awarded to Jianfeng Xu.

Data Availability Statement

In this study, the ZJU-Mocap dataset was downloaded from ZJU-Mocap dataset (https://chingswy.github.io/Dataset-Demo/, accessed on 25 November 2022) and the H36M dataset was downloaded from H36M dataset (http://vision.imar.ro/human3.6m/description.php, accessed on 25 November 2022).

Acknowledgments

I would like to express my heartfelt gratitude to my supervisor for their invaluable guidance and support throughout this research. Additionally, I extend my appreciation to all those who contributed to this study in any way. The authors would like to acknowledge the anonymous reviewers for their valuable comments.

Conflicts of Interest

I declare that there are no conflicts of interest regarding the research presented in this paper.

References

  1. Peng, S.; Dong, J.; Wang, Q.; Zhang, S.; Shuai, Q.; Zhou, X.; Bao, H. Animatable neural radiance fields for modeling dynamic human bodies. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Online, 11–17 October 2021; pp. 14314–14323. [Google Scholar]
  2. Peng, S.; Zhang, Y.; Xu, Y.; Wang, Q.; Shuai, Q.; Bao, H.; Zhou, X. Neural body: Implicit neural representations with structured latent codes for novel view synthesis of dynamic humans. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 9054–9063. [Google Scholar]
  3. Hong, Y.; Peng, B.; Xiao, H.; Liu, L.; Zhang, J. Headnerf: A real-time nerf-based parametric head model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 20374–20384. [Google Scholar]
  4. Zhao, F.; Yang, W.; Zhang, J.; Lin, P.; Zhang, Y.; Yu, J.; Xu, L. Humannerf: Efficiently generated human radiance field from sparse inputs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 7743–7753. [Google Scholar]
  5. Saito, S.; Huang, Z.; Natsume, R.; Morishima, S.; Kanazawa, A.; Li, H. Pifu: Pixel-aligned implicit function for high-resolution clothed human digitization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2304–2314. [Google Scholar]
  6. Barron, J.T.; Mildenhall, B.; Tancik, M.; Hedman, P.; Martin-Brualla, R.; Srinivasan, P.P. Mip-nerf: A multiscale representation for anti-aliasing neural radiance fields. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Online, 11–17 October 2021; pp. 5855–5864. [Google Scholar]
  7. Mildenhall, B.; Srinivasan, P.P.; Tancik, M.; Barron, J.T.; Ramamoorthi, R.; Ng, R. Nerf: Representing scenes as neural radiance fields for view synthesis. Commun. ACM 2021, 65, 99–106. [Google Scholar] [CrossRef]
  8. Fridovich-Keil, S.; Yu, A.; Tancik, M.; Chen, Q.; Recht, B.; Kanazawa, A. Plenoxels: Radiance fields without neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5501–5510. [Google Scholar]
  9. Xiang, F.; Xu, Z.; Hasan, M.; Hold-Geoffroy, Y.; Sunkavalli, K.; Su, H. Neutex: Neural texture mapping for volumetric neural rendering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 7119–7128. [Google Scholar]
  10. Wu, M.; Wang, Y.; Hu, Q.; Yu, J. Multi-view neural human rendering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1682–1691. [Google Scholar]
  11. Loper, M.; Mahmood, N.; Romero, J.; Pons-Moll, G.; Black, M.J. SMPL: A skinned multi-person linear model. ACM Trans. Graph. (TOG) 2015, 34, 1–16. [Google Scholar] [CrossRef]
  12. Pavlakos, G.; Zhu, L.; Zhou, X.; Daniilidis, K. Learning to estimate 3D human pose and shape from a single color image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 459–468. [Google Scholar]
  13. Ionescu, C.; Papava, D.; Olaru, V.; Sminchisescu, C. Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 1325–1339. [Google Scholar] [CrossRef] [PubMed]
  14. Kajiya, J.T.; Von Herzen, B.P. Ray tracing volume densities. ACM SIGGRAPH Comput. Graph. 1984, 18, 165–174. [Google Scholar] [CrossRef]
  15. Zhang, K.; Riegler, G.; Snavely, N.; Koltun, V. Nerf++: Analyzing and improving neural radiance fields. arXiv 2020, arXiv:2010.07492. [Google Scholar]
  16. Müller, T.; Evans, A.; Schied, C.; Keller, A. Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph. (ToG) 2022, 41, 1–15. [Google Scholar] [CrossRef]
  17. Hu, S.; Zhou, K.; Li, K.; Yu, L.; Hong, L.; Hu, T.; Li, Z.; Lee, G.H.; Liu, Z. ConsistentNeRF: Enhancing Neural Radiance Fields with 3D Consistency for Sparse View Synthesis. arXiv 2023, arXiv:2305.11031. [Google Scholar]
  18. Pumarola, A.; Corona, E.; Pons-Moll, G.; Moreno-Noguer, F. D-nerf: Neural radiance fields for dynamic scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 10318–10327. [Google Scholar]
  19. Kwon, Y.; Kim, D.; Ceylan, D.; Fuchs, H. Neural human performer: Learning generalizable radiance fields for human performance rendering. Adv. Neural Inf. Process. Syst. 2021, 34, 24741–24752. [Google Scholar]
  20. Peng, B.; Hu, J.; Zhou, J.; Zhang, J. SelfNeRF: Fast Training NeRF for Human from Monocular Self-rotating Video. arXiv 2022, arXiv:2210.01651. [Google Scholar]
  21. Dong, J.; Shuai, Q.; Zhang, Y.; Liu, X.; Zhou, X.; Bao, H. Motion capture from internet videos. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part II 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 210–227. [Google Scholar]
  22. Lewis, J.P.; Cordner, M.; Fong, N. Pose space deformation: A unified approach to shape interpolation and skeleton-driven deformation. In Proceedings of the 27th Annual Conference on Computer Graphics and Interactive Techniques, New Orleans, LA, USA, 23–28 July 2000; pp. 165–172. [Google Scholar]
  23. Zheng, Z.; Huang, H.; Yu, T.; Zhang, H.; Guo, Y.; Liu, Y. Structured local radiance fields for human avatar modeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 15893–15903. [Google Scholar]
  24. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  25. Park, K.; Sinha, U.; Barron, J.T.; Bouaziz, S.; Goldman, D.B.; Seitz, S.M.; Martin-Brualla, R. Nerfies: Deformable neural radiance fields. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Online, 11–17 October 2021; pp. 5865–5874. [Google Scholar]
  26. Gong, K.; Liang, X.; Li, Y.; Chen, Y.; Yang, M.; Lin, L. Instance-level human parsing via part grouping network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 770–785. [Google Scholar]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. Li, Y.; Bu, R.; Sun, M.; Wu, W.; Di, X.; Chen, B. Pointcnn: Convolution on x-transformed points. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; Volume 31. [Google Scholar]
  29. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  30. Osman, A.A.; Bolkart, T.; Black, M.J. Star: Sparse trained articulated human body regressor. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part VI 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 598–613. [Google Scholar]
  31. Pavlakos, G.; Choutas, V.; Ghorbani, N.; Bolkart, T.; Osman, A.A.; Tzionas, D.; Black, M.J. Expressive body capture: 3d hands, face, and body from a single image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10975–10985. [Google Scholar]
Figure 1. Given a multi-view dynamic human body image, we enable the synthesis of new perspectives and new poses, thereby implicitly reconstructing a three-dimensional representation of the human body.
Figure 1. Given a multi-view dynamic human body image, we enable the synthesis of new perspectives and new poses, thereby implicitly reconstructing a three-dimensional representation of the human body.
Electronics 12 04382 g001
Figure 2. Our network employs dynamic multi-view images of the human body as input to generate an a priori model using the SMPL framework. The mixed weight field predicts the positions of points in the canonical space. To address non-rigid deformations, we utilize a non-rigid deformation field to correct the offset resulting from such deformations. Additionally, our 2D–3D feature fusion network combines and integrates abundant feature information to enhance the effectiveness of human body reconstruction.
Figure 2. Our network employs dynamic multi-view images of the human body as input to generate an a priori model using the SMPL framework. The mixed weight field predicts the positions of points in the canonical space. To address non-rigid deformations, we utilize a non-rigid deformation field to correct the offset resulting from such deformations. Additionally, our 2D–3D feature fusion network combines and integrates abundant feature information to enhance the effectiveness of human body reconstruction.
Electronics 12 04382 g002
Figure 3. Qualitative results of novel view synthesis on the H36M dataset.
Figure 3. Qualitative results of novel view synthesis on the H36M dataset.
Electronics 12 04382 g003
Table 1. Quantitative comparison of novel view synthesis on the ZJU-Mocap dataset.
Table 1. Quantitative comparison of novel view synthesis on the ZJU-Mocap dataset.
PSNR↑SSIM↑
ANOursANOurs
31326.7727.300.9430.944
31520.0020.620.8670.851
37723.6925.710.9190.921
38626.6227.190.8860.898
Average24.2725.210.9040.904
Note: Bold formatting is used to emphasize the data in the table. The up arrow (↑) indicates that a larger indicator is better.
Table 2. Quantitative comparison of novel pose synthesis on the H36M dataset.
Table 2. Quantitative comparison of novel pose synthesis on the H36M dataset.
PSNR↑SSIM↑
NTNHRANOursNTNHRANOurs
20.4220.9322.4123.010.8420.8580.8760.881
Note: Bold formatting is used to emphasize the data in the table. The up arrow (↑) indicates that a larger indicator is better.
Table 3. Results of novel view synthesis on the H36M dataset in terms of PSNR and SSIM (higher is better).
Table 3. Results of novel view synthesis on the H36M dataset in terms of PSNR and SSIM (higher is better).
PSNR↑SSIM↑
NTNHRANOursNTNHRANOurs
S120.9821.0822.7623.900.8600.8720.8940.902
S519.8720.6423.3224.160.8550.8720.8910.902
S620.1820.4022.7723.890.8160.8300.8670.891
S720.4720.2921.9523.070.8560.8680.8890.905
S816.7719.1322.8823.270.8370.8710.8990.902
S922.9623.0424.6225.490.8730.8790.9040.918
S1122.9623.0424.6625.640.8590.8710.9030.911
Average20.4220.9323.2824.200.8510.8660.8920.904
Note: Bold formatting is used to emphasize the data in the table. The up arrow (↑) indicates that a larger indicator is better.
Table 4. Comparison with and without non-rigid offset correction network on subject “S11”.
Table 4. Comparison with and without non-rigid offset correction network on subject “S11”.
PSNR↑SSIM↑
with non-rigid deformation field25.640.911
without non-rigid deformation field24.650.903
Note: Bold formatting is used to emphasize the data in the table. The up arrow (↑) indicates that a larger indicator is better.
Table 5. Comparison with and without feature fusion network on subject “S11”.
Table 5. Comparison with and without feature fusion network on subject “S11”.
PSNR↑SSIM↑
with feature fusion25.640.911
without feature fusion24.680.903
Note: Bold formatting is used to emphasize the data in the table. The up arrow (↑) indicates that a larger indicator is better.
Table 6. Comparison with and without time stamp on subject “S11”.
Table 6. Comparison with and without time stamp on subject “S11”.
PSNR↑SSIM↑
with time stamp25.640.911
without time stamp25.340.908
Note: Bold formatting is used to emphasize the data in the table. The up arrow (↑) indicates that a larger indicator is better.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, X.; Guo, X.; Li, W.; Liu, J.; Xu, J. Deform2NeRF: Non-Rigid Deformation and 2D–3D Feature Fusion with Cross-Attention for Dynamic Human Reconstruction. Electronics 2023, 12, 4382. https://doi.org/10.3390/electronics12214382

AMA Style

Xie X, Guo X, Li W, Liu J, Xu J. Deform2NeRF: Non-Rigid Deformation and 2D–3D Feature Fusion with Cross-Attention for Dynamic Human Reconstruction. Electronics. 2023; 12(21):4382. https://doi.org/10.3390/electronics12214382

Chicago/Turabian Style

Xie, Xiaolong, Xusheng Guo, Wei Li, Jie Liu, and Jianfeng Xu. 2023. "Deform2NeRF: Non-Rigid Deformation and 2D–3D Feature Fusion with Cross-Attention for Dynamic Human Reconstruction" Electronics 12, no. 21: 4382. https://doi.org/10.3390/electronics12214382

APA Style

Xie, X., Guo, X., Li, W., Liu, J., & Xu, J. (2023). Deform2NeRF: Non-Rigid Deformation and 2D–3D Feature Fusion with Cross-Attention for Dynamic Human Reconstruction. Electronics, 12(21), 4382. https://doi.org/10.3390/electronics12214382

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop