• Laser & Optoelectronics Progress
  • Vol. 57, Issue 24, 241102 (2020)
Xunqiang Gong1、2、*, Xinglei Liu1、2, Tieding Lu1、2, and Dan Liu2
Author Affiliations
  • 1Fundamental Science on Radioactive Geology and Exploration Technology Laboratory, East China University of Technology, Nanchang, Jiangxi 330013, China
  • 2Faculty of Geomatics, East China University of Technology, Nanchang, Jiangxi 330013, China
  • show less
    DOI: 10.3788/LOP57.241102 Cite this Article Set citation alerts
    Xunqiang Gong, Xinglei Liu, Tieding Lu, Dan Liu. Accuracy Assessment of Object-Oriented Classification Based on Regular Verification Points[J]. Laser & Optoelectronics Progress, 2020, 57(24): 241102 Copy Citation Text show less

    Abstract

    Remote sensing image classification is an important part of image analysis, and post-classification accuracy assessment is the main basis for determining the effect of image classification. At present, random verification points are often used as assessment parameters in object-oriented classification, which may easily lead to inaccurate classification results. An object-oriented classification accuracy assessment method based on regular verification points is proposed in this paper. Regular and random verification points are used to evaluate the classification accuracy by using support vector machine, CART (classification and regression tree ) decision tree, and K nearest neighbor classification. Experimental results show that the proposed method has higher classification accuracy than traditional methods based on random verification points. The optimal overall classification accuracy of the three classification methods based on the regular verification points reaches 87.92%, 91.94%, and 94.63%, respectively, which are better than the accuracy assessment results of random verification points based methods.
    Xunqiang Gong, Xinglei Liu, Tieding Lu, Dan Liu. Accuracy Assessment of Object-Oriented Classification Based on Regular Verification Points[J]. Laser & Optoelectronics Progress, 2020, 57(24): 241102
    Download Citation