03374naa a2200241 a 450000100080000000500110000800800410001902200140006002400570007410000190013124501710015026000090032152025540033065000150288465300190289965300330291865300180295170000190296970000180298870000260300670000150303277300850304721243652020-08-17 2020 bl uuuu u00u1 u #d a0924-27167 ahttps://doi.org/10.1016/j.isprsjprs.2020.08.0042DOI1 aMARTINS, V. S. aExploring multiscale object-based convolutional neural network (multi-OCNN) for remote sensing image classification at high spatial resolution.h[electronic resource] c2020 aConvolutional Neural Network (CNN) has been increasingly used for land cover mapping of remotely sensed imagery. However, large-area classification using traditional CNN is computationally expensive and produces coarse maps using a sliding window approach. To address this problem, object-based CNN (OCNN) becomes an alternative solution to improve classification performance. However, previous studies were mainly focused on urban areas or small scenes, and implementation of OCNN method is still needed for large-area classification over heterogeneous landscape. Additionally, the massive labeling of segmented objects requires a practical approach for less computation, including object analysis and multiple CNNs. This study presents a new multiscale OCNN (multi-OCNN) framework for large-scale land cover classification at 1-m resolution over 145,740 km2. Our approach consists of three main steps: (i) image segmentation, (ii) object analysis with skeleton-based algorithm, and (iii) application of multiple CNNs for final classification. Also, we developed a large benchmark dataset, called IowaNet, with 1 million labeled images and 10 classes. In our approach, multiscale CNNs were trained to capture the best contextual information during the semantic labeling of objects. Meanwhile, skeletonization algorithm provided morphological representation (?medial axis?) of objects to support the selection of convolutional locations for CNN predictions. In general, proposed multi-OCNN presented better classification accuracy (overall accuracy ~87.2%) compared to traditional patch-based CNN (81.6%) and fixed-input OCNN (82%). In addition, the results showed that this framework is 8.1 and 111.5 times faster than traditional pixel-wise CNN16 or CNN256, respectively. Multiple CNNs and object analysis have proved to be essential for accurate and fast classification. While multi-OCNN produced a high-level of spatial details in the land cover product, misclassification was observed for some classes, such as road versus buildings or shadow versus lake. Despite these minor drawbacks, our results also demonstrated the benefits of IowaNet training dataset in the model performance; overfitting process reduces as the number of samples increases. The limitations of multi-OCNN are partially explained by segmentation quality and limited number of spectral bands in the aerial data. With the advance of deep learning methods, this study supports the claim of multi-OCNN benefits for operational large-scale land cover product at 1-m resolution. aLand cover aAerial imagery aConvolutional neural network aDeep learning1 aKALEITA, A. L.1 aGELDER, B. K.1 aSILVEIRA, H. L. F. da1 aABE, C. A. tISPRS Journal of Photogrammetry and Remote Sensinggv. 168, p. 56-73, oct. 2020.