research-article
- Authors:
- Hong Huang School of Computer Science and Technology, Guangdong University of Technology, China
School of Computer Science and Technology, Guangdong University of Technology, China
Search about this author
- Zhuowei Wang School of Computer Science and Technology, Guangdong University of Technology, China
School of Computer Science and Technology, Guangdong University of Technology, China
Search about this author
- Genping Zhao School of Computer Science and Technology, Guangdong University of Technology, China
School of Computer Science and Technology, Guangdong University of Technology, China
Search about this author
ICMLC '24: Proceedings of the 2024 16th International Conference on Machine Learning and ComputingFebruary 2024Pages 369–375https://doi.org/10.1145/3651671.3651732
- 0citation
- 0
- Downloads
Metrics
Total Citations0Total Downloads0Last 12 Months0
Last 6 weeks0
- Get Citation Alerts
New Citation Alert added!
This alert has been successfully added and will be sent to:
You will be notified whenever a record that you have chosen has been cited.
To manage your alert preferences, click on the button below.
Manage my Alerts
New Citation Alert!
Please log in to your account
- Publisher Site
- Get Access
ICMLC '24: Proceedings of the 2024 16th International Conference on Machine Learning and Computing
Efficient 3D Reconstruction of Multiple Plants from UAV Images with Deep Learning
Pages 369–375
PreviousChapterNextChapter
ABSTRACT
Acquiring the 3D structure of plants is a critical task in the agricultural industry. Existing methods of generating 3D point clouds for multiple plants require a long processing time. In this paper, a 3D reconstruction method for numerous plants is proposed. Firstly, camera parameters in different viewpoints are obtained from the aerial image of plants by incremental structure from motion. Subsequently, the learning-based multi-view stereo takes images and the corresponding camera parameters as inputs to acquire initial depth maps. Finally, the depth maps are filtered and fused to produce a complete and dense 3D point cloud. We conducted experiments on an agricultural orchard dataset to compare with other methods. Experimental results demonstrate that our method reconstructs point clouds of plants with good quality while having a lower running time.
References
- Henrik Aanæs, RasmusRamsbøl Jensen, George Vogiatzis, Engin Tola, and AndersBjorholm Dahl. 2016. Large-scale data for multiple-view stereopsis. International Journal of Computer Vision 120 (2016), 153–168.Google Scholar
Digital Library
- Connelly Barnes, Eli Shechtman, Adam Finkelstein, and DanB Goldman. 2009. PatchMatch: A randomized correspondence algorithm for structural image editing. ACM Trans. Graph. 28, 3 (2009), 24.Google Scholar
Digital Library
- EF Berra and MV Peppa. 2020. Advances and challenges of UAV SFM MVS photogrammetry and remote sensing: Short review. In 2020 IEEE Latin American GRSS & ISPRS Remote Sensing Conference (LAGIRS). IEEE, 533–538.Google Scholar
Cross Ref
- Zhen Chen, Hui Lv, Lu Lou, and JohnH Doonan. 2022. Fast and accurate 3D reconstruction of plants using mvsnet and multi-view images. In Advances in Computational Intelligence Systems: Contributions Presented at the 20th UK Workshop on Computational Intelligence, September 8-10, 2021, Aberystwyth, Wales, UK 20. Springer, 390–399.Google Scholar
- Xiaodong Gu, Zhiwen Fan, Siyu Zhu, Zuozhuo Dai, Feitong Tan, and Ping Tan. 2020. Cascade cost volume for high-resolution multi-view stereo and stereo matching. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2495–2504.Google Scholar
Cross Ref
- Xian-Feng Han, Hamid Laga, and Mohammed Bennamoun. 2019. Image-based 3D object reconstruction: State-of-the-art and trends in the deep learning era. IEEE transactions on pattern analysis and machine intelligence 43, 5 (2019), 1578–1604.Google Scholar
- Jakob Iglhaut, Carlos Cabo, Stefano Puliti, Livia Piermattei, James O’Connor, and Jacqueline Rosette. 2019. Structure from motion photogrammetry in forestry: A review. Current Forestry Reports 5 (2019), 155–168.Google Scholar
Cross Ref
- San Jiang, Wanshou Jiang, and Lizhe Wang. 2021. Unmanned aerial vehicle-based photogrammetric 3d mapping: A survey of techniques, applications, and challenges. IEEE Geoscience and Remote Sensing Magazine 10, 2 (2021), 135–171.Google Scholar
Cross Ref
- Chenhui Lin, Hong Wang, Chengliang Liu, and Liang Gong. 2020. 3D reconstruction based plant-monitoring and plant-phenotyping platform. In 2020 3rd World Conference on Mechanical Engineering and Intelligent Manufacturing (WCMEIM). IEEE, 522–526.Google Scholar
Cross Ref
- Tsung-Yi Lin, Piotr Dollár, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. 2017. Feature pyramid networks for object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition. 2117–2125.Google Scholar
Cross Ref
- Fusang Liu, Pengcheng Hu, Bangyou Zheng, Tao Duan, Binglin Zhu, and Yan Guo. 2021. A field-based high-throughput method for acquiring canopy architecture using unmanned aerial vehicle images. Agricultural and Forest Meteorology 296 (2021), 108231.Google Scholar
Cross Ref
- Elias Marks, Federico Magistri, and Cyrill Stachniss. 2022. Precise 3D reconstruction of plants from UAV imagery combining bundle adjustment and template matching. In 2022 International Conference on Robotics and Automation (ICRA). IEEE, 2259–2265.Google Scholar
Digital Library
- Paul Merrell, Amir Akbarzadeh, Liang Wang, Philippos Mordohai, Jan-Michael Frahm, Ruigang Yang, David Nistér, and Marc Pollefeys. 2007. Real-time visibility-based fusion of depth maps. In 2007 IEEE 11th International Conference on Computer Vision. Ieee, 1–8.Google Scholar
Cross Ref
- Seishi Ninomiya. 2022. High-throughput field crop phenotyping: current status and challenges. Breeding Science 72, 1 (2022), 3–18.Google Scholar
Cross Ref
- Niall O’Mahony, Sean Campbell, Lenka Krpalkova, Daniel Riordan, Joseph Walsh, Aidan Murphy, and Conor Ryan. 2019. Computer Vision for 3D Perception: A Review. In Intelligent Systems and Applications: Proceedings of the 2018 Intelligent Systems Conference (IntelliSys) Volume 2. Springer, 788–804.Google Scholar
- Yeping Peng, Mingbin Yang, Genping Zhao, and Guangzhong Cao. 2021. Binocular-vision-based structure from motion for 3-D reconstruction of plants. IEEE Geoscience and Remote Sensing Letters 19 (2021), 1–5.Google Scholar
- Jessada Phattaralerphong and Hervé Sinoquet. 2005. A method for 3D reconstruction of tree crown volume from photographs: assessment with 3D-digitized plants. Tree Physiology 25, 10 (2005), 1229–1242.Google Scholar
Cross Ref
- Worasit Sangjan, RebeccaJ McGee, and Sindhuja Sankaran. 2022. Optimization of UAV-based imaging and image processing orthom*osaic and point cloud approaches for estimating biomass in a forage crop. Remote Sensing 14, 10 (2022), 2396.Google Scholar
Cross Ref
- JohannesL Schonberger and Jan-Michael Frahm. 2016. Structure-from-motion revisited. In Proceedings of the IEEE conference on computer vision and pattern recognition. 4104–4113.Google Scholar
Cross Ref
- JohannesL Schönberger, Enliang Zheng, Jan-Michael Frahm, and Marc Pollefeys. 2016. Pixelwise view selection for unstructured multi-view stereo. In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part III 14. Springer, 501–518.Google Scholar
Cross Ref
- Sergio Vélez, Rubén Vacas, Hugo Martín, David Ruano-Rosa, and Sara Álvarez. 2022. High-Resolution UAV RGB Imagery Dataset for Precision Agriculture and 3D Photogrammetric Reconstruction Captured over a Pistachio Orchard (Pistacia vera L.) in Spain. Data 7, 11 (2022), 157.Google Scholar
Cross Ref
- Fangjinhua Wang, Silvano Galliani, Christoph Vogel, Pablo Speciale, and Marc Pollefeys. 2021. Patchmatchnet: Learned multi-view patchmatch stereo. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 14194–14203.Google Scholar
Cross Ref
- Xiang Wang, Chen Wang, Bing Liu, Xiaoqing Zhou, Liang Zhang, Jin Zheng, and Xiao Bai. 2021. Multi-view stereo in the deep learning era: A comprehensive review. Displays 70 (2021), 102102.Google Scholar
Cross Ref
- Huang Yao, Rongjun Qin, and Xiaoyu Chen. 2019. Unmanned aerial vehicle for remote sensing applications—A review. Remote Sensing 11, 12 (2019), 1443.Google Scholar
Cross Ref
- Yao Yao, Zixin Luo, Shiwei Li, Tian Fang, and Long Quan. 2018. Mvsnet: Depth inference for unstructured multi-view stereo. In Proceedings of the European conference on computer vision (ECCV). 767–783.Google Scholar
Digital Library
- Yao Yao, Zixin Luo, Shiwei Li, Tianwei Shen, Tian Fang, and Long Quan. 2019. Recurrent mvsnet for high-resolution multi-view stereo depth inference. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 5525–5534.Google Scholar
Cross Ref
- Yanchao Zhang, Hanxuan Wu, and Wen Yang. 2019. Forests growth monitoring based on tree canopy 3D reconstruction using UAV aerial photogrammetry. Forests 10, 12 (2019), 1052.Google Scholar
Cross Ref
- Genping Zhao, Weitao Cai, Zhuowei Wang, Heng Wu, Yeping Peng, and Lianglun Cheng. 2022. Phenotypic parameters estimation of plants using deep learning-based 3-D reconstruction from single RGB image. IEEE Geoscience and Remote Sensing Letters 19 (2022), 1–5.Google Scholar
Cited By
View all
Index Terms
Efficient 3D Reconstruction of Multiple Plants from UAV Images with Deep Learning
Computing methodologies
Artificial intelligence
Computer vision
Recommendations
- Time-of-Flight Camera Calibration for Improved 3D Reconstruction of Indoor Scenes
ISCID '14: Proceedings of the 2014 Seventh International Symposium on Computational Intelligence and Design - Volume 02
Calibration for Time-of-Flight (ToF) cameras are different from traditional color cameras, in that their calibration involves additional error compensation steps for the depth camera caused by complicated factors. Without this process, there may be ...
Read More
- Dense 3D reconstruction combining depth and RGB information
Dense 3D reconstruction has important applications in many fields. The existing depth information based methods are typically constrained in their effective camera-object distance which should be from 0.4m to 4m. We present a novel method that can ...
Read More
- Fixed-Lens camera setup and calibrated image registration for multifocus multiview 3D reconstruction
Abstract
Image-based 3D reconstruction or 3D photogrammetry of small-scale objects including insects and biological specimens is challenging due to the use of a high magnification lens with inherently limited depth of field, and the object’s fine ...
Read More
Login options
Check if you have access through your login credentials or your institution to get full access on this article.
Sign in
Full Access
Get this Publication
- Information
- Contributors
Published in
ICMLC '24: Proceedings of the 2024 16th International Conference on Machine Learning and Computing
February 2024
757 pages
ISBN:9798400709234
DOI:10.1145/3651671
Copyright © 2024 ACM
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [emailprotected].
Sponsors
In-Cooperation
Publisher
Association for Computing Machinery
New York, NY, United States
Publication History
- Published: 7 June 2024
Permissions
Request permissions about this article.
Author Tags
- 3D reconstruction
- Deep learning
- Plant
- Point cloud
- UAV images
Qualifiers
- research-article
- Research
- Refereed limited
Conference
Funding Sources
Other Metrics
View Article Metrics
- Bibliometrics
- Citations0
Article Metrics
- View Citations
Total Citations
Total Downloads
- Downloads (Last 12 months)0
- Downloads (Last 6 weeks)0
Other Metrics
View Author Metrics
Cited By
This publication has not been cited yet
PDF Format
View or Download as a PDF file.
eReader
View online with eReader.
eReader
Digital Edition
View this article in digital edition.
View Digital Edition
HTML Format
View this article in HTML Format .
View HTML Format
- Figures
- Other
Close Figure Viewer
Browse AllReturn
Caption
View Table of Contents
Export Citations
Your Search Results Download Request
We are preparing your search results for download ...
We will inform you here when the file is ready.
Download now!
Your Search Results Download Request
Your file of search results citations is now ready.
Download now!
Your Search Results Download Request
Your search export query has expired. Please try again.