Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/94278
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Land Surveying and Geo-Informaticsen_US
dc.creatorWang, Yen_US
dc.creatorZhu, Xen_US
dc.creatorWu, Ben_US
dc.date.accessioned2022-08-11T02:01:35Z-
dc.date.available2022-08-11T02:01:35Z-
dc.identifier.issn0143-1161en_US
dc.identifier.urihttp://hdl.handle.net/10397/94278-
dc.language.isoenen_US
dc.publisherTaylor & Francisen_US
dc.rights© 2018 Informa UK Limited, trading as Taylor & Francis Groupen_US
dc.rightsThis is an Accepted Manuscript of an article published by Taylor & Francis in International Journal of Remote Sensing on 6 Sep 2018 (Published online), available online: http://www.tandfonline.com/10.1080/01431161.2018.1513669en_US
dc.titleAutomatic detection of individual oil palm trees from UAV images using HOG features and an SVM classifieren_US
dc.typeJournal/Magazine Articleen_US
dc.identifier.spage7356en_US
dc.identifier.epage7370en_US
dc.identifier.volume40en_US
dc.identifier.issue19en_US
dc.identifier.doi10.1080/01431161.2018.1513669en_US
dcterms.abstractOil palm trees are important economic crops in tropical areas. Accurate knowledge of the number of oil palm trees in a plantation area is important to predict the yield of palm oil, manage the growing situation of the palm trees and maximise their productivity. In this study, we propose a novel automatic method for detection and enumeration of individual oil palm trees using images from unmanned aerial vehicles (UAVs). This method required three major steps. First, images from UAVs were classified as vegetation or non-vegetation by the support vector machine (SVM) classifier. Second, a feature descriptor based on the histogram of oriented gradient (HOG) was designed for palm trees and used to extract features for machine learning. Finally, a SVM classifier was trained and optimised using the HOG features from positive (i.e., oil palm trees) and negative samples (i.e., objects other than oil palm trees). The trained classifier was then applied to detect individual oil palm trees using adaptive moving windows that allowed it to also return the crown size of each oil palm tree. The method was trained at one site and validated independently at four other sites with different situations. The overall accuracy of palm tree detection was 99.21% at the training site and 99.39%, 99.06%, 99.90% and 94.63% at the four validation sites; the last one was for the most challenging site, in which palm trees were mixed with other trees. These tests confirm the effectiveness of the proposed method. The simplicity and great efficiency of the proposed method allow it to support oil palm tree counting for large areas using imagery from UAVs.en_US
dcterms.accessRightsopen accessen_US
dcterms.bibliographicCitationInternational journal of remote sensing, 2019, v. 40, no. 19, p. 7356-7370en_US
dcterms.isPartOfInternational journal of remote sensingen_US
dcterms.issued2019-
dc.identifier.scopus2-s2.0-85053238076-
dc.identifier.eissn1366-5901en_US
dc.description.validate202207 bckwen_US
dc.description.oaAccepted Manuscripten_US
dc.identifier.FolderNumbera1565; LSGI-0166-
dc.identifier.SubFormID45447-
dc.description.fundingSourceOthersen_US
dc.description.fundingTextHong Kong Polytechnic University; National Natural Science Foundation of Chinaen_US
dc.description.pubStatusPublisheden_US
dc.identifier.OPUS19751600-
Appears in Collections:Journal/Magazine Article
Files in This Item:
File Description SizeFormat 
Wang_UAV_HOG_SVM.pdfPre-Published version2.83 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Final Accepted Manuscript
Access
View full-text via PolyU eLinks SFX Query
Show simple item record

Page views

48
Last Week
0
Last month
Citations as of May 12, 2024

Downloads

93
Citations as of May 12, 2024

SCOPUSTM   
Citations

91
Citations as of May 17, 2024

WEB OF SCIENCETM
Citations

72
Citations as of May 16, 2024

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.