A deep learning-based framework for efficient and accurate 3D real-scene reconstruction
International Journal of Information Technology (Singapore), ISSN: 2511-2112, Vol: 16, Issue: 7, Page: 4605-4609
2024
- 3Citations
- 1Captures
Metric Options: Counts1 Year3 YearSelecting the 1-year or 3-year option will change the metrics count to percentiles, illustrating how an article or review compares to other articles or reviews within the selected time period in the same journal. Selecting the 1-year option compares the metrics against other articles/reviews that were also published in the same calendar year. Selecting the 3-year option compares the metrics against other articles/reviews that were also published in the same calendar year plus the two years prior.
Example: if you select the 1-year option for an article published in 2019 and a metric category shows 90%, that means that the article or review is performing better than 90% of the other articles/reviews published in that journal in 2019. If you select the 3-year option for the same article published in 2019 and the metric category shows 90%, that means that the article or review is performing better than 90% of the other articles/reviews published in that journal in 2019, 2018 and 2017.
Citation Benchmarking is provided by Scopus and SciVal and is different from the metrics context provided by PlumX Metrics.
Example: if you select the 1-year option for an article published in 2019 and a metric category shows 90%, that means that the article or review is performing better than 90% of the other articles/reviews published in that journal in 2019. If you select the 3-year option for the same article published in 2019 and the metric category shows 90%, that means that the article or review is performing better than 90% of the other articles/reviews published in that journal in 2019, 2018 and 2017.
Citation Benchmarking is provided by Scopus and SciVal and is different from the metrics context provided by PlumX Metrics.
Article Description
Normally, 3D scenes can be reconstructed with the help of several photo-imaging techniques that are used for 3D geo-visualization and scene analysis. However, the methods involved in the reconstruction of scenes are not accurate and time-consuming. Thus, the given paper proposes a 3D reconstruction framework from multi-view stereo (MVS) images based on deep learning. The proposed multi-view aggregation matching network (MVAMN) comprises pre-processing of the dataset, multi-scale feature extraction using convolutional neural network (CNN), cost volume regularization, and refinement of depth map inference to handle more challenging situations, including mutual occlusions, big-depth variations in oblique photos, and substantial viewpoint changes. The refinement phase involves the computation of probability distributions from the initial depth map as well as the refined depth map. After the refinement is done, the loss function is calculated to quantify the errors between the initial and refined depth maps, thereby optimizing and enhancing the performance of the proposed framework. Lastly, the performance of the model is validated and compared with existing deep learning MVS methods by taking open-source image dataset into consideration. The findings demonstrate that lower memory utilization and higher efficiency make the proposed model highly suitable for aerial images in large-scale and high-resolution 3D surface reconstruction tasks.
Bibliographic Details
Springer Science and Business Media LLC
Provide Feedback
Have ideas for a new metric? Would you like to see something else here?Let us know