ISPRS Journal of Photogrammetry and Remote Sensing 169 (2020) 180–194 Available online 24 September 2020 0924-2716/© 2020 International Society for Photogrammetry and Remote Sensing, Inc. (ISPRS). Published by Elsevier B.V. All rights reserved. Developing a machine learning based cotton yield estimation framework using multi-temporal UAS data Akash Ashapure a , Jinha Jung a, * , Anjin Chang b , Sungchan Oh a , Junho Yeom c , Murilo Maeda d , Andrea Maeda d , Nothabo Dube e , Juan Landivar e , Steve Hague f , Wayne Smith f a Purdue University, USA b Texas A&M University Corpus Christi, USA c Gyeongsang National University, South Korea d Texas A&M AgriLife Extension, Lubbock, USA e Texas A&M AgriLife Research at Corpus Christi, USA f Texas A&M University, USA A R T I C L E INFO Keywords: Precision agriculture Cotton genotype selection UAS ANN ABSTRACT In this research a machine learning framework was developed for cotton yield estimation using multi-temporal remote sensing data collected from unmanned aircraft system (UAS). The proposed machine learning model was based on an artificial neural network (ANN) and used three types of crop features derived from UAS data to predict the yield, namely; multi-temporal features including canopy cover, canopy height, canopy volume, normalized difference vegetation index (NDVI), excessive greenness index (ExG); non-temporal features including cotton boll count, boll size and boll volume, and irrigation status as a qualitative feature. The model provided low residual values with predicted yield values close to the observed yield values (R 2 ~ 0.9). ANN model performance was compared with support vector regression (SVR) and random forest regression (RFR). Comparison results revealed that ANN model outperforms SVR and RFR. Additionally, redundant features were removed using correlation analysis, and an optimal subset of features was obtained that included canopy volume, ExG, boll count, boll volume and irrigation status. Moreover, the relative significance of each feature in the optimal input feature subset was determined using sensitivity analysis. It was found that canopy volume and ExG contributed around 50% towards the corresponding yield. Finally, further analysis was performed to investigate how early in the growing season the model can accurately predict yield. It was observed that even at 70 days after planting the model predicted yield with reasonable accuracy (R 2 of 0.72 over test set). This study revealed that UAS derived multi-temporal data along with non-temporal and qualitative data can be combined within a machine learning framework to provide a reliable estimation of crop yield and provide effective understanding for crop management. 1. Introduction With a total aerial coverage of approximately 6 million acres, cotton is amongst one of the leading cash crops in the state of Texas (Adhikari et al., 2017), which has led to an upsurge of the cotton breeding research in the state. One of the main objectives of cotton breeding research is to select genotypes suitable for specific environments. For example, in South Texas the focus is to develop cotton genotypes resistant to water stress due to the dry climate and high irrigation costs. Traditionally, cotton breeding research has focused on manual field-based evaluation approaches that require the entire cotton field to be harvested, and later the best performing genotypes are selected based on ranking the yield of the individual genotypes (Iqbal et al., 2008; Kazerani, 2012; Shaukat et al., 2013; Clement et al., 2014). As this process is cumbersome, the scale of the field experiment is constrained by the limited availability of resources. As a result, cotton breeding research is focusing on the development of automated genotype selection techniques which do not require the entire field to be harvested. Remote sensing-based crop yield estimation methods have the potential to help automate the genotype selection process. In the literature, satellite remote sensing data have been extensively utilized for crop yield estimation (Singh et al., 2002; Ferencz et al., 2004; Sayago and Bocco, 2018; Hunt et al., 2019; Meng * Corresponding author at: Lyles School of Civil Engineering, Purdue University, 550 W Stadium Ave, West Lafayette, IN 47907, USA. E-mail address: jinha@purdue.edu (J. Jung). Contents lists available at ScienceDirect ISPRS Journal of Photogrammetry and Remote Sensing journal homepage: www.elsevier.com/locate/isprsjprs https://doi.org/10.1016/j.isprsjprs.2020.09.015 Received 20 April 2020; Received in revised form 11 August 2020; Accepted 15 September 2020