Next Article in Journal
Spatial Dynamics and Drivers of Urban Growth in Thua Thien Hue Province, Vietnam: Insights for Urban Sustainability in the Global South
Previous Article in Journal
A Novel Evolutionary Deep Learning Approach for PM2.5 Prediction Using Remote Sensing and Spatial–Temporal Data: A Case Study of Tehran
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-View Three-Dimensional Reconstruction Based on Feature Enhancement and Weight Optimization Network

1
School of Surveying and Geo-Informatics, Shandong Jianzhu University, Jinan 250101, China
2
Shandong Provincial Institute of Land Surveying and Mapping, Jinan 250101, China
3
Shandong Zhengyuan Aerial Remote Sensing Technology Co., Ltd., Jinan 250101, China
*
Author to whom correspondence should be addressed.
ISPRS Int. J. Geo-Inf. 2025, 14(2), 43; https://doi.org/10.3390/ijgi14020043
Submission received: 26 October 2024 / Revised: 9 January 2025 / Accepted: 22 January 2025 / Published: 24 January 2025

Abstract

:
Aiming to address the issue that existing multi-view stereo reconstruction methods have insufficient adaptability to the repetitive and weak textures in multi-view images, this paper proposes a three-dimensional (3D) reconstruction algorithm based on Feature Enhancement and Weight Optimization MVSNet (Abbreviated as FEWO-MVSNet). To obtain accurate and detailed global and local features, we first develop an adaptive feature enhancement approach to obtain multi-scale information from the images. Second, we introduce an attention mechanism and a spatial feature capture module to enable high-sensitivity detection for weak texture features. Third, based on the 3D convolutional neural network, the fine depth map for multi-view images can be predicted and the complete 3D model is subsequently reconstructed. Last, we evaluated the proposed FEWO-MVSNet through training and testing on the DTU, BlendedMVS, and Tanks and Temples datasets. The results demonstrate significant superiorities of our method for 3D reconstruction from multi-view images, with our method ranking first in accuracy and second in completeness when compared to the existing representative methods.

1. Introduction

Multi-view stereo (MVS) reconstruction focuses on using two or more perspective photographs to recover the geometric surface structure information of a target scene [1]. It is a popular technique [2] used in the domains of actual 3D construction, historic heritage restoration and preservation, and other related areas [3]. However, the 3D reconstruction of multi-view images with weak textures and repetitive patterns is still challenging in both digital photogrammetry and computer vision fields [4].
Despite the significant progress made by traditional MVS reconstruction methods such as Gipuma [5] and COLMAP [6], the accuracy and completeness of reconstructed point clouds in complex scenes remain insufficient. We aim to address the limitations of traditional methods. In recent years, the development of 3D reconstruction has been strongly boosted with the introduction of convolutional neural networks (CNNs) into the field of scene depth estimation. Han et al. proposed MatchNet [7], which mainly consists of a metric model composed of three fully connected layers to compute the similarity between the features to be matched. To improve the automation of the method, Yao et al. constructed the end-to-end depth estimation network MVSNet [8], which significantly improves the accuracy of the target depth information, but it also consumes a large amount of memory space when constructing the cost model and calculating the feature information of multiple views. For this reason, Yao et al. further propose the recursive stereo visual network R-MVSNet [9], which regularizes the original 2D feature cost map by gated recursive units, effectively reducing the memory occupation of the model. Gu et al.’s study incorporates cascaded convolutional network layers and multi-scale strategies into the network and proposes a multi-scale depth estimation model called CasMVSNet [10], and the experimental results validate the advantages of this method in terms of computational accuracy and efficiency. Wang et al., inspired by the traditional PatchMatch algorithm, propose the deep learning matching network PatchMatchNet [11], which quickly finds the best matching region in an image by introducing global and local optimization strategies, and the test results show that it is able to generate depth information for complex environments. To enhance feature extraction in multi-scale space, ASPPMVSNet [12] introduces Atrous Spatial Pyramid Pooling (ASPP) into the model, using dilated convolutions for multi-scale feature information extraction. To improve the correlation of feature information in both the channel and spatial dimensions during the feature extraction process, DRI-MVSNet [13] builds upon CasMVSNet by introducing a Channel and Spatial Combined Processing (CSCP) module to capture relevant contextual information.
While traditional methods and CNN-based networks have significantly improved depth estimation in multi-view stereo (MVS) geometry, transformers [14] have a strong ability to capture global contextual features in MVS tasks, which significantly enhances the representation of various features and injects new impetus into the development of 3D reconstruction. Both MVSTER [15] and MVSTR [16] leverage transformer architectures to improve MVS reconstruction. However, MVSTR focuses on enhancing feature matching through global context and inter-view interactions with its transformer-based modules, while MVSTER emphasizes optimizing depth estimation through epipolar geometry, monocular depth enhancement, and cascade refinement. Both methods achieve good performance, with MVSTER being more efficient and MVSTR being more robust across benchmarks. However, we must consider the critical issues of weak textures and repetitive textures in the process of MVS 3D reconstruction for improving the accuracy and completeness of the model. To address this issue, Zhu et al. propose to construct DCNv2 [17], which is an approach that improves the original DCN (Deep Crossing Network) by utilizing the variable sensory field convolution, which effectively improves the loss problem of feature information propagation. Ding et al. constructed a feature matching converter TransMVSNet [18] using a transformer, inserting an Adaptive Receptive Field (ARF) module after the FPN. This approach improves the reconstruction quality of texture-deprived regions by adaptively expanding the receptive field. Weak textures create sparse features and matching ambiguity, while repetitive textures easily lead to mismatches, further degrading reconstruction quality. These challenges suggest an urgent need to enhance algorithms for improving the quality of reconstruction in these scenarios.
The above methods show great advantages and potentials in multi-view 3D reconstruction and can also provide effective references for higher-level 3D reconstruction. However, the tests of the above methods in various scenarios show that they often find it difficult to reconstruct a complete and reliable 3D model in repetitive- or weak-texture regions. Inspired by TransMVSNet, this paper proposes a 3D reconstruction algorithm based on Feature Enhancement and Weight Optimization MVSNet (FEWO-MVSNet). Our method firstly extracts global and local feature information from the multi-view images using the adaptive feature enhancement module, then introduces the adaptive optimization mechanism for weight allocation to estimate the fine depth map of the scenery with weak/repeated textures and finally completes the reconstruction of the dense 3D point cloud. Through the training and testing of multiple international standard datasets, the test results show that our FEWO-MVSNet achieves improvements in both accuracy and completeness.

2. Methodology

AS shown in Figure 1, we designed a 3D reconstruction network called FEWO-MVSNet based on adaptive feature enhancement and weight allocation, using CasMVSNet as the base network. The improved work in our method includes the following:
  • Introduce an Adaptive Spatial Feature Fusion (ASFF) module [19] on the basis of a Feature Pyramid Network (FPN) to enhance the capability of capturing feature information at different scales;
  • Adaptively expand the search range of features through the deformable feeler field convolution module DCNv2 [17] and combine it with the transformer positional encoder for the enhancement of global contextual feature information aggregation;
  • Design an Adaptive Space Weight Allocation (ASWA) module, which is integrated into SENet [20], to highlight the low-frequency information in the convolutional channel and color space and then realize the dense extraction of feature information of multi-view images containing weak- and repetitive-texture regions.
The algorithm of FEWO-MVSNet can be briefly summarized as follows: Firstly, the multi-scale features of each image at three resolutions are extracted using the FPN. Secondly, in view of inconsistent scales between different features, the feature maps are input into the Adaptive Spatial Feature Fusion (ASFF) module to achieve information enhancement and feature normalization. Thirdly, in order to improve the accuracy of the multi-view feature matching and feature aggregation, the deformable receptive field module DCNv2 and the transformer feature aggregation module are combined to aggregate the updated feature information (for the purpose of emphasizing the adequate feature information that has been captured, we spatially label the features using SENet and ASWA). Fourthly, using differentiable homography, we obtain the multi-view image for feature mapping, and the cost volume can be generated based on Spatial Feature Fusion Spatial Feature Fusion depth correlation weighting method and feature transform. Then, cost volume regularization is performed by the 3D CNN to generate the probability volume that can predict the depth maps. Lastly, we blend each depth map using a concatenated scheme and obtain dense point clouds of the target scenery. In all, we realize the end-to-end 3D reconstruction of the model based on FEWO-MVSNet.

2.1. Feature Extraction Networks

Deep learning-based MVS methods typically use downsampling in the feature extraction module to expand the receptive field while reducing image resolution to save memory space. Based on the most downsampling, this produces the final feature map, which is directly input into subsequent networks. This process often causes texture information loss, which negatively impacts the accuracy of reconstruction results. For this, we introduce an FPN to replace the original CNN to realize the feature extraction with different scales from the top to the bottom of the pyramid network. Based on many tests, we find that feature maps from different layers of the hierarchical pyramid vary in terms of resolution, receptive fields, and semantic representation. These scale differences cause information loss during top-down feature fusion. The loss of critical feature information causes inaccuracies in scene reconstruction and leads to missing local details or surface discontinuities. Therefore, the Adaptive Spatial Feature Fusion (ASFF) module is further introduced after the feature extraction network FPN to obtain the optimal weights of each channel through adaptive learning, which is conducive to maintaining the invariance of the pyramid image feature-extracting ratio to realize the intelligent extraction of complex texture features.
The improved feature extraction network is shown in Figure 2. The main content of this network can be presented as follows. Firstly, given an input image with W × H resolution as well as its elements of internal and external orientation, the image is input into the bottom-up feature extraction part, which reduces the resolution of the image using 2 × 2 downsampling to increase its semantic information. Secondly, the network uses the top-down path to accomplish the fusion between high-level images and low-level images based on 2 × 2 upsampling. Thirdly, the top-down feature fusion path is connected to the bottom-up feature extraction path through the lateral connections, and those lateral connections combine the high-resolution information of the left-part images with the rich semantic information of the right-part images. Fourthly, we incorporate the Adaptive Spatial Feature Fusion (ASFF) module with the FPN [21]. Finally, we have three image channels with different resolutions: W / 4 × H / 4 ,   W / 2 × H / 2 ,   W × H . The adaptive weighting coefficients ( α ,   β , and γ ), and the corresponding feature layer L i merge together by weighting. The fused feature Z i j l is then obtained based on the formula, as follows.
Z i j l = α i j l · L i j 1 l + β i j l · L i j 2 l + γ i j l · L i j 3 l
where L i j n l represents the feature vector at the position (i, j) on the feature maps resized from level n to level l. As shown in Equation (1), the fused feature Z i j l is computed using an additive approach. Therefore, the output features from levels 1 to 3 must have the same size and number of channels when being added. To achieve this, features from different levels require upsampling or downsampling, along with adjustments to the number of channels. For the weight parameters α, β and γ, they are obtained by applying a 1 × 1 convolution to the resized feature maps from levels one to three. After concatenation, the parameters α, β and γ undergo a SoftMax operation [22] using Equation (2), which normalizes their values to the range [0,1] and ensures that their sum equals one.
α = e λ α e λ α + e λ β + e λ γ
where α , β and γ are generated using three SoftMax control parameters: λ α , λ β and λ γ . This may adaptively fuse the features from all levels at each scale.

2.2. Aggregation and Enhancement for Features Based on Transformer

2.2.1. Adaptive Enlargement of Receptive Field

Due to the significant difference in the receptive field of contextual information between the FPN and transformer, the feature information can be lost easily in the process of transmission. For the purpose of correct transmission, we add the DCNv2 between the modules of the FPN and transformer. In the convolution operation, the DCNv2 can compensate adaptive offsets of texture, which are generated by the learning of the network. This network can adaptively adjust the receptive field according to the local features of the image; thus, it can extract texture information with different brightnesses and types. The adaptive offset matrix can be computed by Equation (3) as follows.
M x 0 = x n R ε x n · l x 0 + x n + x n · m k
where M x 0 represents the offset of sampled intensity, ε x n is the deformable convolution parameter, l is the input feature map, x 0 is the sampling point in the feature map, x n is the sampling point of the offset, x n is the offset matrix, and m k is the weight of the sampling point.

2.2.2. Feature Encoding Based on Transformer

Among the existing multi-view stereo reconstruction approaches, MVS reconstruction demands the comprehensive exploitation of information from the entire scene for depth estimation. However, traditional CNNs often encounter limitations when dealing with large-scale contextual information. The transformer mechanism is capable of establishing the relationships among pixels within a global scope; it can make full use of global contextual information and improve the distinctiveness of feature information. For the purpose of enhancing the accuracy and robustness of the feature map, we employ the FPN, followed by the transformer feature encoding module [14], including self-attention and cross-attention units. The extracted feature map L is flattened into a one-dimensional vector. Then, we calculate the encoding position based on the feature information of the flattened map by Equation (4).
P E = sin p o s / 10000 n / d m o d e l c o s p o s / 10000 n / d m o d e l
where PE denotes the position encoding, pos denotes the position information of each pixel, and d m o d e l represents the output dimension of the encoding–decoding model; this equation uses the sine and cosine functions to alternately express the encoding information at each position.
The transformer module applies positional encoding to the reference features Z 0 and input features Z i . This enhances positional consistency and robustness across different resolutions. The self-attention mechanism processes the encoded features to extract global dependencies, while the cross-attention mechanism enables interactions between the reference features and input features to capture contextual information. These processes ultimately update and refine the feature information. Suppose that Z i represents the initial feature information, the feature enhancement for Z i can be given as follows. Firstly, three attention vectors, namely Query (Q), Key (K) and Value (V), are input into the feature encoder simultaneously. The context information of V is retrieved from the corresponding Q and K based on the multi-layer stacking of the self-attention mechanism. At the same time, it captures the dependencies existing in the input sequence to produce new feature information Z i . Secondly, the cross-information between the reference image and the input image is captured by the cross-attention mechanism. Finally, the updating calculation for the new feature information Z i is performed. The updated result Z i comprises the dense features with the global and local context information.

2.3. Adaptive Allocation for Feature Weights

To obtain the key feature information with adaptive dimensions, we integrate the SENet with ASWA in our model to generate adaptive weights for features, which can be beneficial to deal with different view images with various types of textures. The adaptive allocation for feature weight is illustrated by Figure 3.
The aforementioned strategy of Section 2.2 can effectively enhance the perception of the model to recognize weak features as more as possible. However, it may also lead to the accumulation of redundant feature information, further resulting in high computational complexity and preventing it from capturing key features. To dynamically adjust the weights of various feature regions, we introduce SENet into our model to effectively extract weak-texture features. In this part, we optimize the channel attention network SENet. The correlation of feature channels is enhanced through squeeze, excitation, and adaptive weighting operations. Specifically, the key feature information z i is produced by Equation (5).
z i = Z i · s i g m o i d φ 2 · R e L u φ 1 · G A P L i
where s i g m o i d and R e L u are the activation functions, φ 1 and φ 2 denote the weight factors of the fully connected layers, G A P denotes the global average pooling function and L i represents the arbitrary feature layer.
The SENet is trained for the adaptive adjustment of the weight coefficients φ 1 and φ 2 through forward and backward propagation, which guarantee the reliability of feature information in all channels. Similarly, the weight allocation for 3D spatial feature information extraction is also important. To adaptively adjust the spatial feature weights according to feature textures, the ASWA module is incorporated after the SENet. Then, the activation function s i g m o i d is used to normalize the weight coefficients, which enables the robust capture of feature information from each spatial channel.

2.4. Correlation Volume Construction and Loss Function Estimation

2.4.1. Correlation Volume Construction

Similar to most existing deep learning-based MVSNet methods, all the input image features Z i can be aligned to the reference image features Z 0 through differentiable warping [23]. To start with, the transformation between point p i of the reference image and point p i of the input image can be expressed by Equation (6) under the depth hypothesis d i .
p i = K R K 0 1 p i d i + t
where R and t, respectively, denote the rotation and translation matrices between the input image and the reference image, and K 0 and K are the calibration matrices of the camera. Next, we calculate the correlation between corresponding points based on Equation (6) and the bilinear interpolation method. The correlation metric is expressed by Equation (7).
c d i p i = < F 0 p i , F d i p i >
where F 0 p i is the feature map of the reference image, F d i p represents the i-th feature map of the input image at depth d i , and c d i p i represents the correlation coefficient of the input image and the reference image at the pixel point p i . To decrease the memory consumption of the regularization of correlation volume, the channel number is reduced to one. To maintain the maximum correlation of data in the depth dimension, we introduce the depth correlation weighting method [18] to aggregate the correlation volume C p i by Equation (8).
C p i = i = 1 X 1 m a x c d i p i · c d i p i
We chose a 3D CNN to conduct the regularization of correlation volume. The 3D CNN is composed of the 3D convolutional operations and moves among all dimensions of the correlation volume. Through the 3D CNN, we can obtain the probability volume P, which can predict depth information. To reduce the memory consumption and computational complexity of the 3D CNN, we first reduce the number of channels in the feature map to one, thereby cutting down the feature information storage for each pixel. Additionally, we aggregate X-1 pairs of correlation volumes using Equation (8) to reduce redundant or mismatched information. The two measures are beneficial to optimize 3D CNN computation.

2.4.2. Loss Function Estimation

To highlight the key information in challenging texture regions and improve the accuracy of depth estimation, we adopt the combination of the focal loss function F L [24] and the smooth loss function S L [25] using Equation (9). This can adaptively adjust the weight between the background points and the model points and thus improve the learning rate of the key regions.
= F L + S L
F L = α F L 1 P d p i γ log P d p i
where P d p i denotes the predicted probability at a pixel point p i at depth hypothesis d , d denotes the depth value closest to the ground truth among all hypotheses, and γ is the focus parameter. α F L is a balancing parameter that is used to adjust the influence between the background and the model. According to experience, γ = 0 is suitable for relatively simple scenarios and γ = 2 fits more complicated scenarios.
S L = 1 2 s 2   , s 1 s 1 2   , s > 1
where s denotes the gaps between the predicted values and the target values. S L adjusts the model more robustly through s . This can restrict the gradient of outliers from two aspects, as follows. First, when s is large, the gradient value is prevented from being excessively large. Second, when s is small, the gradient value is sufficiently small. S L can also provide a smoother response to outliers during the training process. Finally, we use the combination of F L and S L as our loss function for model training to estimate depth information.

2.5. Depth Map Filtering and Fusion

In this part, we reconstruct a complete 3D point-cloud model as follows. Firstly, the depth map outliers and background points can be filtered through photometric and geometric constraints. The filtering parameters are adjusted based on dynamic operations [26]. Secondly, we convert the pixel coordinates of valid points from the filtered depth maps into world coordinates and extract the corresponding color information for each point. Finally, we fuse the depth maps to generate the final dense 3D point cloud.

3. Results and Discussion

3.1. Experimental Datasets

Three large open datasets—DTU, BlendedMVS and Tanks and Temples—are chosen to train our FEWO-MVSNet and the representative models. A detailed introduction to these datasets is shown in Table 1.

3.2. Experimental Details

We train the FEWO-MVSNet on the DTU dataset with PyTorch 1.11.0. In the training phase, we set the number of input data images to N = 5 and the resolution to 640 × 512 pixels. The depth hypotheses are sampled from 425 mm to 935 mm. The FEWO-MVSNet adopts a three-stage cascaded network. The number of plane sweeping depth hypotheses of each section is 48, 32, and 8, respectively. To test the generalization of FEWO-MVSNet, we employ BlenedMVS to fine-tune the model. The input image data are also set to N = 5, with a resolution of 768 × 576 pixels. The FEWO-MVSNet is trained using Adam for 16 epochs with an initial learning rate of 0.001. The learning rate is halved, respectively, after the 6, 8, and 12 epochs. We used the NVIDIA GeForce GTX 3090 to train FEWO-MVSNet. The batch size is one.

3.3. Result and Analysis

3.3.1. Comparative Experiments

The FEWO-MVSNet is validated for effectiveness on the DTU dataset. We evaluate the Accuracy (Acc), Completeness (Comp), and Overall scores of the reconstructed 3D point cloud using the MATLAB 2018 code provided by DTU. The Acc and Comp scores are employed to evaluate the average distance between the ground-truth point cloud and the reconstructed point cloud of the model. The Overall score is the average of Acc and Comp found by Equation (12).
O v e r a l l = A c c . + C o m p . 2
(1)
The lower values of the three metrics Acc, Comp, and Overall indicate that the reconstructed point cloud is closer to the real point cloud. We compare the FEWO-MVSNet with the traditional methods (Gipuma, Colmap), methods of deep learning (MVSNet, R-MVSNet, CasMVSNet, DRI-MVSNet, ASPPMVSNet, PatchMatchNet), and methods of deep learning using a transformer (MVSTR, MVSTER, TransMVSNet). Table 2 shows the quantitative comparison results of the DTU dataset. Compared to the classic deep learning algorithms MVSNet and CasMVSNet, FEWO-MVSNet improves accuracy by 8% and 1.2% while enhancing completeness by 15% and 4.3%. The accuracy increases by 2% and 3.7% compared to the TransMVSNet and MVSTER algorithms using a transformer.
(2)
Figure 4 shows the dense point-cloud reconstruction. The FEWO-MVSNet improves the weak/repetitive textures by combining weight optimization and feature enhancement. In the red box in Figure 4, we can see the oscilloscope dial and side in scan11, the left and right sides of the beer packaging in scan12, the top of the cup in scan48 and the top of the sculpture in scan118. We enhance the texture information in the blank areas of these scenes. FEWO-MVSNet can produce denser and complete point clouds while preserving more details.
(3)
We conduct comparative experiments between the typical TransMVSNet and our FEWO-MVSNet in four repetitive-/weak-texture scenes, as shown in Figure 5 and Table 3. In the black boxes of Figure 5, it is evident that FEWO-MVSNet significantly increases the point-cloud density across the four scenes. In Table 3, we compare the three metrics Acc, Comp, and Overall through experiments. FEWO-MVSNet surpasses TransMVSNet in all three metrics, and thus we effectively improve the reconstruction quality in repetitive-/weak-texture regions.

3.3.2. Generalization Experiments

To verify the generalization ability of the FEWO-MVSNet, we train the model by using the BlendedMVS dataset and evaluate it by using the Tanks and Temples dataset. We present the point clouds generated by FEWO-MVSNet to the official website of Tanks and Temples, which enables us to obtain the F-score. This metric indicates that a higher value means better reconstruction quality. To achieve quantitative comparison results for the intermediate set, we compare the FEWO-MVSNet with the traditional methods, methods of deep learning, and methods of deep learning using transformer.
(1)
The Tanks and Temples (Inter) dataset includes many complex real-world scenes with intricate geometries, repetitive textures, lighting variations, and occlusions, making the reconstruction highly challenging. Gipuma is designed with a focus on dense matching and depth estimation but lacks the modularity and adaptability of methods like COLMAP. Gipuma is more commonly used in standardized laboratory settings (e.g., the DTU dataset), whereas its practical effectiveness is limited when dealing with more complex outdoor scenes. Therefore, only COLMAP is quantitatively evaluated in Table 4.
(2)
Table 4 presents the quantitative testing results of different methods on the Tanks and Temples (Inter) dataset. Mean represents the average metric value across all scenes. A higher value of this metric indicates better reconstruction quality. We compare FEWO-MVSNet with 10 other methods, and the Mean ranks first among all methods. Compared to the classic deep learning algorithms MVSNet and CasMVSNet, FEWO-MVSNet improves Mean by 20% and 7.2%. The Mean increases by 0.16% and 2.7% compared to the TransMVSNet and MVSTER algorithms using the transformer.
(3)
The experiments show that the algorithm proposed in this article can achieve better reconstruction results in the case of multiple influencing factors such as outdoor scene light and image noises. Figure 6 presents the results for the intermediate dataset scenarios of Francis, Playground, Family, M60, Panther, and Train. The rich texture information has been successfully caught by FEWO-MVSNet, even in these weak-texture areas like the Francis surface, and large-scale scenes like the Playground. The F-scores of the method in this article are better than those of existing methods. The enhancement of the effect in repetitive- and weak-texture regions proves that FEWO-MVSNet has a certain generalization capability.

3.4. Ablation Experiment

To examine the effect of the modules added in FEWO-MVSNet on the reconstruction results, we employ DTU as the dataset for the ablation experiments. The baseline is built on the CasMVSNet cascade architecture, employing an FPN feature extraction network and variance-based metrics, without any additional modules. We conduct four ablation experiments, sequentially integrating the ASFF, DCNv2, and SENet-ASWA modules. The ablation experiments adopt the same evaluation metrics as DTU. This is done to demonstrate the effectiveness of FEWO-MVSNet.
The results of the ablation experiments in Table 5 show that each module contributes to improving point-cloud reconstruction. The ASFF and DCNv2 modules capture local and global feature information, which enhances the feature extraction ability of FEWO-MVSNet. The SE-ASWA module filters the extracted feature information. It extracts key information through adaptive weight allocation. These modules improve weak/repetitive textures and detailed features. The reconstruction results show improvements in both accuracy and completeness.
The introduction of the ASFF, DCNv2 and SE-ASWA modules and the original transformer increases the memory consumption and computational complexity of FEWP-MVSNet. We train the model on a single GPU. As shown in Table 6, under the same environment, the running speed of FEWO-MVSNet decreases by 14.5% compared to TransMVSNet, and memory usage increases by 1.6%. In the next step, we would optimize memory consumption and running speed by training the model on multiple GPUs and incorporating the technology of lightweight modules.

4. Conclusions

We construct a novel multi-view stereo network, FEWO-MVSNet, which is based on a FEWO strategy and a TransMVSNet framework. First, the Adaptive Spatial Feature Fusion (ASFF) module is used to improve the feature learning capability of the FPN, and then the DCNv2 and transformer mechanisms are combined to enhance the context information of features. Next, we employ the SENet-ASWA module to optimize the redundant elements in the extracted feature and finally produce a rich and accurate 3D dense point cloud. Extensive experiments on standard datasets, such as DTU, BlendedMVS, and Tanks and Temples, demonstrate that the proposed method is superior to many existing deep learning-based methods for weak- and repetitive-texture regions. Our algorithm still has limitations in terms of operational efficiency and reconstruction completeness. In future work, we aim to use lightweight modules to replace complex computations, further improving the performance and adaptability of the algorithm, making it more suitable for various scenarios.

Author Contributions

Conceptualization, Guobiao Yao and Guozhong Wei; methodology, Guobiao Yao and Ziheng Wang; software, Guobiao Yao and Ziheng Wang; validation, Guobiao Yao, Ziheng Wang and Fengqi Zhu; formal analysis, Guobiao Yao and Guozhong Wei; investigation, Ziheng Wang and Qingqing Fu; resources, Guozhong Wei and Fengqi Zhu; data curation, Ziheng Wang and Min Wei; writing—original draft preparation, Guobiao Yao and Ziheng Wang; writing—review and editing, Guobiao Yao; visualization, Ziheng Wang and Qian Yu; supervision, Guobiao Yao; funding acquisition, Qian Yu. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China with Project No. 42171435, the Shandong Provincial Natural Science Foundation with Project No. ZR2021MD006, the Postgraduate Education and Teaching Reform Foundation of Shandong Province with Project Province with Project No. SDYJG19115, and the Undergraduate Education and Teaching Reform Foundation of Shandong Province with Project No. Z2021014. This work was also funded by the high-quality graduate course of Shandong Province with Project No. SDYKC2022151.

Data Availability Statement

The case data can be downloaded from GitHub HuanHuanWZH/MVS-Date: MVS 3D reconstruction (accessed on 30 September 2024).

Acknowledgments

The authors would like to thank Yikang Ding, Xizhou Zhu, and Jie Hu for providing their key algorithms.

Conflicts of Interest

Author Min Wei was employed by the company Shandong Zhengyuan Aerial Remote Sensing Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Luo, H.; Zhang, J.; Liu, X.; Zhang, L.; Liu, J. Large-Scale 3D Reconstruction from Multi-View Imagery: A Comprehensive Review. Remote Sens. 2024, 16, 773. [Google Scholar] [CrossRef]
  2. Dong, Y.; Song, J.; Fan, D.; Ji, S.; Lei, R. Joint Deep Learning and Information Propagation for Fast 3D City Modeling. ISPRS Int. J. Geo-Inf. 2023, 12, 150. [Google Scholar] [CrossRef]
  3. Bi, J.; Wang, J.; Cao, H.; Yao, G.; Wang, Y.; Li, Z.; Sun, M.; Yang, H.; Zhen, J.; Zheng, G. Inverse distance weight-assisted particle swarm optimized indoor localization. Appl. Soft Comput. 2024, 164, 112032. [Google Scholar] [CrossRef]
  4. Gao, X.; Yang, R.; Chen, X.; Tan, J.; Liu, Y.; Wang, Z.; Tan, J.; Liu, H. A New Framework for Generating Indoor 3D Digital Models from Point Clouds. Remote Sens. 2024, 16, 3462. [Google Scholar] [CrossRef]
  5. Galliani, S.; Lasinger, K.; Schindler, K. Massively parallel multiview stereopsis by surface normal diffusion. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 873–881. [Google Scholar]
  6. Schonberger, J.L.; Frahm, J.-M. Structure-from-motion revisited. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4104–4113. [Google Scholar]
  7. Han, X.; Leung, T.; Jia, Y.; Sukthankar, R.; Berg, A.C. Matchnet: Unifying feature and metric learning for patch-based matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3279–3286. [Google Scholar]
  8. Yao, Y.; Luo, Z.; Li, S.; Fang, T.; Quan, L. Mvsnet: Depth inference for unstructured multi-view stereo. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 767–783. [Google Scholar]
  9. Yao, Y.; Luo, Z.; Li, S.; Shen, T.; Fang, T.; Quan, L. Recurrent mvsnet for high-resolution multi-view stereo depth inference. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5525–5534. [Google Scholar]
  10. Gu, X.; Fan, Z.; Zhu, S.; Dai, Z.; Tan, F.; Tan, P. Cascade cost volume for high-resolution multi-view stereo and stereo matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 2495–2504. [Google Scholar]
  11. Wang, F.; Galliani, S.; Vogel, C.; Speciale, P.; Pollefeys, M. Patchmatchnet: Learned multi-view patchmatch stereo. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 14194–14203. [Google Scholar]
  12. Saeed, S.; Lee, S.; Cho, Y.; Park, U. ASPPMVSNet: A high-receptive-field multiview stereo network for dense three-dimensional reconstruction. ETRI J. 2022, 44, 1034–1046. [Google Scholar] [CrossRef]
  13. Li, Y.; Li, W.; Zhao, Z.; Fan, J. DRI-MVSNet: A depth residual inference network for multi-view stereo images. PLoS ONE 2022, 17, e0264721. [Google Scholar] [CrossRef] [PubMed]
  14. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar] [CrossRef]
  15. Wang, X.; Zhu, Z.; Huang, G.; Qin, F.; Ye, Y.; He, Y.; Chi, X.; Wang, X. Mvster: Epipolar transformer for efficient multi-view stereo. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 573–591. [Google Scholar]
  16. Zhu, J.; Peng, B.; Li, W.; Shen, H.; Zhang, Z.; Lei, J. Multi-view stereo with transformer. arXiv 2021, arXiv:2112.00336. [Google Scholar]
  17. Zhu, X.; Hu, H.; Lin, S.; Dai, J. Deformable convnets v2: More deformable, better results. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9308–9316. [Google Scholar]
  18. Ding, Y.; Yuan, W.; Zhu, Q.; Zhang, H.; Liu, X.; Wang, Y.; Liu, X. Transmvsnet: Global context-aware multi-view stereo network with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8585–8594. [Google Scholar]
  19. Liu, S.; Huang, D.; Wang, Y. Learning spatial fusion for single-shot object detection. arXiv 2019, arXiv:1911.09516. [Google Scholar]
  20. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  21. Teng, J.; Sun, H.; Liu, P.; Jiang, S. An Improved TransMVSNet Algorithm for Three-Dimensional Reconstruction in the Unmanned Aerial Vehicle Remote Sensing Domain. Sensors 2024, 24, 2064. [Google Scholar] [CrossRef]
  22. Wang, G.; Wang, K.; Lin, L. Adaptively connected neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1781–1790. [Google Scholar]
  23. Kendall, A.; Martirosyan, H.; Dasgupta, S.; Henry, P.; Kennedy, R.; Bachrach, A.; Bry, A. End-to-end learning of geometry and context for deep stereo regression. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 66–75. [Google Scholar]
  24. Ross, T.-Y.; Dollár, G. Focal loss for dense object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2980–2988. [Google Scholar]
  25. Girshick, R. Fast r-cnn. arXiv 2015, arXiv:1504.08083. [Google Scholar]
  26. Jia, X.; De Brabandere, B.; Tuytelaars, T.; Gool, L.V. Dynamic filter networks. Adv. Neural Inf. Process. Syst. 2016, 29. [Google Scholar] [CrossRef]
  27. Aanæs, H.; Jensen, R.R.; Vogiatzis, G.; Tola, E.; Dahl, A.B. Large-scale data for multiple-view stereopsis. Int. J. Comput. Vis. 2016, 120, 153–168. [Google Scholar] [CrossRef]
  28. Yao, Y.; Luo, Z.; Li, S.; Zhang, J.; Ren, Y.; Zhou, L.; Fang, T.; Quan, L. Blendedmvs: A large-scale dataset for generalized multi-view stereo networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1790–1799. [Google Scholar]
  29. Knapitsch, A.; Park, J.; Zhou, Q.-Y.; Koltun, V. Tanks and temples: Benchmarking large-scale scene reconstruction. ACM Trans. Graph. (ToG) 2017, 36, 1–13. [Google Scholar] [CrossRef]
Figure 1. The proposed 3D reconstruction network, FEWO-MVSNet.
Figure 1. The proposed 3D reconstruction network, FEWO-MVSNet.
Ijgi 14 00043 g001
Figure 2. The improved feature extraction network by integrating the FPN with the ASFF module.
Figure 2. The improved feature extraction network by integrating the FPN with the ASFF module.
Ijgi 14 00043 g002
Figure 3. The adaptive allocation mechanism for feature weight.
Figure 3. The adaptive allocation mechanism for feature weight.
Ijgi 14 00043 g003
Figure 4. We chose three representative methods (MVSNet, CasMVSNet and TransMVSNet) for comparison. The reconstruction results in four scenes with weak/repeated textures are shown, with optimized details indicated in the red box.
Figure 4. We chose three representative methods (MVSNet, CasMVSNet and TransMVSNet) for comparison. The reconstruction results in four scenes with weak/repeated textures are shown, with optimized details indicated in the red box.
Ijgi 14 00043 g004
Figure 5. Comparison of the reconstructed 3D dense point clouds from repetitive- and weak-texture scenes based on TransMVSNet and our FEWO-MVSNet. The reconstruction results in four scenes with weak/repeated textures are shown, with optimized details indicated in the black box.
Figure 5. Comparison of the reconstructed 3D dense point clouds from repetitive- and weak-texture scenes based on TransMVSNet and our FEWO-MVSNet. The reconstruction results in four scenes with weak/repeated textures are shown, with optimized details indicated in the black box.
Ijgi 14 00043 g005
Figure 6. The reconstruction of scenes based on FEWO-MVSNet using the Tanks and Temples (Intermediate) dataset. The reconstruction results in six scenes with weak/repeated textures are shown, with optimized details indicated in the red box.
Figure 6. The reconstruction of scenes based on FEWO-MVSNet using the Tanks and Temples (Intermediate) dataset. The reconstruction results in six scenes with weak/repeated textures are shown, with optimized details indicated in the red box.
Ijgi 14 00043 g006
Table 1. Experimental datasets in detail.
Table 1. Experimental datasets in detail.
Data TypesData DescriptionThumbnails of Data
(1) DTU [27]DTU is a large indoor dataset that encompasses 128 scenes. It covers the scene by adopting 49 or 63 camera positions. We divide 27,097 training samples as a training set with 79 sceneries, an evaluation set with 18 sceneries, and a test set with 22 sceneries.Ijgi 14 00043 i001
(2) BlendedMVS [28]BlendedMVS includes 113 various types of scenes, for example, cities and buildings, with a total of 17,818 images. At present, the dataset does not provide evaluation tool. Therefore, it is only used for model training in the generalization experimentIjgi 14 00043 i002
(3) Tanks and Temples [29]Tanks and Temples is a large indoor and outdoor dataset that comprises 14 scenes of different scales. This dataset is adopted as test sets for the generalization experiments. We categorize it as an intermediate set with eight sceneries and an advanced set with six sceneries.Ijgi 14 00043 i003
Table 2. Quantitative comparison results for DTU. The bold values denote the best results, and the underlined values denote the second-best results.
Table 2. Quantitative comparison results for DTU. The bold values denote the best results, and the underlined values denote the second-best results.
MethodsAcc/mmComp/mmOverall/mm
Gipuma [5]0.2830.8730.578
Colmap [6]0.4000.6640.532
MVSNet [8]0.3960.5270.462
R-MVSNet [9]0.3830.4520.417
CasMVSNet [10]0.3250.3850.355
DRI-MVSNet [13]0.4320.3270.379
ASPPMVSNet [12]0.3340.3600.347
PatchMatchNet [11]0.4270.2770.352
MVSTR [16]0.3560.2950.326
MVSTER [15]0.3500.2760.313
TransMVSNet [18]0.3330.3010.317
Ours0.3130.3110.312
Table 3. Quantitative comparison results of four scenes from the DTU dataset. The bold values indicate the best results.
Table 3. Quantitative comparison results of four scenes from the DTU dataset. The bold values indicate the best results.
ScenesAcc/mmComp/mmOverall/mm
TranMVSNetFEWO-MVSNetTranMVSNetFEWO-MVSNetTranMVSNetFEWO-MVSNet
Scan110.3210.3350.3390.3110.3300.323
Scan1203290.3240.2090.2090.2690.266
Scan480.3690.3540.6270.4870.4980.420
Scan1180.2290.2290.3180.3170.2740.273
Table 4. Quantitative testing results of different methods on Tanks and Temples (Inter). The bold values denote the best results, and the underlined values denote the second-best results.
Table 4. Quantitative testing results of different methods on Tanks and Temples (Inter). The bold values denote the best results, and the underlined values denote the second-best results.
MethodIntermediate
MeanFam.Fran.HorseL.H.M60Path.P.G.Train
Colmap [5]42.1450.4122.2526.6356.4344.8346.9748.5342.04
MVSNet [8]43.4855.9928.5525.0750.7953.9650.8647.9034.69
R-MVSNet [9]50.5573.0154.4643.4243.8846.8046.6950.8745.25
CasMVSNet [10]56.4276.3658.4546.2055.5356.1154.0258.1746.56
DRI-MVSNet [13]52.7173.6453.4840.5753.9048.4846.4459.0946.10
ASPPMVSNet [12]54.0376.5047.7436.3455.1257.2854.2857.4347.54
PatchMatchNet [11]53.1566.9952.6443.2554.8752.8749.5454.2150.81
MVSTR [16]56.9376.9259.8250.1656.7356.5351.2256.5847.48
MVSTER [15]60.9280.2163.5152.3061.3861.4758.1658.9851.38
TransMVSNet [18]63.5280.9265.8356.8962.5463.0660.0060.2058.67
Ours63.6881.0965.0856.9262.1862.7961.2761.3458.75
Table 5. Comparison of quantitative results of ablation experiments.
Table 5. Comparison of quantitative results of ablation experiments.
MethodsAcc/mmComp/mmOverall/mm
Baseline Net0.3510.3390.345
+ASFF and DCNv20.3340.3220.328
+SE-ASWA0.3250.3150.320
FEWO-MVSNet0.3130.3110.312
Table 6. Comparison of GPU memory usage and runtime for a single epoch on the DTU dataset with the same input size.
Table 6. Comparison of GPU memory usage and runtime for a single epoch on the DTU dataset with the same input size.
MethodsInput Size (Pixels)Depth Map Size (Pixels)Time (s)Memory-Usage (MB)
TransMVSNet640 × 5121152 × 86422,32011,892
FEWO-MVSNet640 × 5121152 × 86425,56012,085
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yao, G.; Wang, Z.; Wei, G.; Zhu, F.; Fu, Q.; Yu, Q.; Wei, M. Multi-View Three-Dimensional Reconstruction Based on Feature Enhancement and Weight Optimization Network. ISPRS Int. J. Geo-Inf. 2025, 14, 43. https://doi.org/10.3390/ijgi14020043

AMA Style

Yao G, Wang Z, Wei G, Zhu F, Fu Q, Yu Q, Wei M. Multi-View Three-Dimensional Reconstruction Based on Feature Enhancement and Weight Optimization Network. ISPRS International Journal of Geo-Information. 2025; 14(2):43. https://doi.org/10.3390/ijgi14020043

Chicago/Turabian Style

Yao, Guobiao, Ziheng Wang, Guozhong Wei, Fengqi Zhu, Qingqing Fu, Qian Yu, and Min Wei. 2025. "Multi-View Three-Dimensional Reconstruction Based on Feature Enhancement and Weight Optimization Network" ISPRS International Journal of Geo-Information 14, no. 2: 43. https://doi.org/10.3390/ijgi14020043

APA Style

Yao, G., Wang, Z., Wei, G., Zhu, F., Fu, Q., Yu, Q., & Wei, M. (2025). Multi-View Three-Dimensional Reconstruction Based on Feature Enhancement and Weight Optimization Network. ISPRS International Journal of Geo-Information, 14(2), 43. https://doi.org/10.3390/ijgi14020043

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop