Introduction

A regular monitoring of agricultural land and crop production is crucial to obtain critical data for food security and farmland management. For example, accurate crop yield estimation provides valuable information to guide national food management, trade and development strategies (Hunt et al., 2019; Weiss et al., 2020). Crop yield information is also relevant when making informed decisions on insurance, agriculture land management and market intervention (Atzberger, 2013). Nowadays, various large-scale agricultural monitoring systems offer near-real-time data on current crop conditions (Defourny et al., 2019; dos Santos Luciano et al., 2021; Rembold et al., 2019; Weiss et al., 2020; Whitcraft et al., 2015). However, mapping efforts to produce crop yield maps that describe both the within-field variability (needed by farmers for a better crop management) and the large-scale (e.g. regional or national) synoptic view remains technologically and practically challenging. A large part of this challenge is related to the cost for acquiring reference samples in the field, against which EO data can be calibrated (Emilien et al., 2021; Revill et al., 2020; Riihimäki et al., 2019). To respond to this gap, this study proposes a multi-scale workflow for yield mapping with an example for sugarcane in Thailand.

Sugarcane in Thailand

Sugarcane is a tall perennial grass in the genus Saccharum that grows in tropical and subtropical regions (Som-ard et al., 2021; Srijantr et al., 2002; Sumesh et al., 2021; Wang et al., 2020; Xu et al., 2020). Sugarcane is an economically important crop contributing to food, bio-energy production and chemical components extraction (Moraes et al., 2015; Renouf et al., 2008; Som-ard et al., 2018).

Thailand was the world’s fifth-largest exporter of sugarcane in 2020 (FAO, 2022) with excellent growing conditions contributing to significant revenue (Srijantr et al., 2002). Both the government and sugar mills have promoted sugarcane expansion and developed sugarcane varieties to increase production (Sriroth et al., 2016). In Thailand, the total harvested sugarcane area has increased from 0.06 million hectares (Mha) in 1961 to 1.83 Mha in 2020, while annual production increased by 73 million tons (Mt). During the same period, the average yields increased from 32 t/ha to 41 t/ha (FAO, 2022) and sugarcane production constituted a major economical income for farmers (Chunhawong et al., 2018).

Northeast Thailand reports the highest harvested sugarcane area and production in Thailand. Farmers in this region have adopted different crop management strategies to increase production and productivity (Office of the Cane & Sugar Board, 2020; Som-ard et al., 2018; Sumesh et al., 2021). In the past years, many critical sugarcane cultivation problems arose including inaccurate yield estimation, outdated agricultural technology, sugar productivity, lack of labour, uncontrolled crop treatment, unsatisfying price decision-making process, poor irrigation management and, above all, impact of climate change (Chunhawong et al., 2018; Silalertruksa et al., 2017; Som-ard et al., 2022; Sumesh et al., 2021). These issues are addressed by government policy and related agencies to increase the efficiency for the sugar mills (Pipitpukdee et al., 2020; Sriroth et al., 2016), but not all are yet successfully tackled and solved.

Sugarcane yield estimation

An accurate and cost-efficient sugarcane yield estimation system would offer benefits to many stakeholders and is deemed necessary for price setting, to allocate quota on production by sugar mills, as well as for marketing and for monitoring the effects of drought and weather impacts.

At field scale, recent studies by Shendryk et al. (2021), Sumesh et al. (2021), Xu et al. (2020) and Yu et al. (2020) indicated that remote sensing technology can provide key information to guide precision agricultural practices. For example, one of the important activities of sugarcane management is that farmers can use it as a cost-effective tool to better understand their field conditions and then rapidly manage water and fertilizer inputs for specific areas of the field.

Satellite-based Earth Observation (EO) data have over the past four decades strengthened crop yield estimation capabilities by providing enhanced yield mapping and up-to-date information at different spatial scales (Shendryk et al., 2021; Som-ard et al., 2021; Weiss et al., 2020). Widely used EO data includes MODIS (Aqua and Terra), Landsat, Sentinel-1 (S1), and Sentinel-2 (S2). Many studies have been published leveraging machine learning (ML) approaches for crop yield estimation (i.e., sugarcane, rice and wheat) at fine to course spatial resolutions (dos Santos Luciano et al., 2021; Franch et al., 2021; Hunt et al., 2019; Segarra et al., 2022; Son et al., 2020). Results have shown a good potential for accurate crop yield maps.

Since the launch of the first S2 in 2015, rich information (in terms of high spatial and temporal resolutions) of S2 imagery can be used to improve accurate crop yield mapping at the within-field level. Exemplary studies that have demonstrated a high potential for predicting sugarcane yield using the capabilities of S2 imagery together with ML approaches such as a random forest (RF), support vector machines (SVM) and linear regression approaches are listed in Table 1. More results and examples of EO applications for sugarcane yield estimation can be found in the recent review by Som-ard et al. (2021).

Table 1 Exemplary studies for sugarcane yield estimation using Sentinel-2 data

As a common denominator, all studies have reported that the lack of ground measurements is a major limitation (e.g., through crop cutting experiment), together with problems related to cloud cover conditions and difficulties to transfer models to different geographical locations (from field to national scale).

Integration of UAV data

Very high spatial resolution data derived from unmanned aerial vehicle (UAV) instruments are known to be useful for sugarcane monitoring at local scales (Sanches et al., 2018; Som-ard et al., 2021) (Table 2).

Table 2 Exemplary studies for sugarcane yield estimation at local scales using different sensors based on UAVs

The need for reference data

The application of EO data is currently mainly limited by the scant availability of reference data. UAV data is not able to cover cost-efficiently large areas. However, very high resolution UAV-RGB data can be used to derive yield maps at the field scale based on only a few reference samples. The idea is thus to use UAV-RGB derived yield maps to augment/enlarge the number of available ground reference observations so that more reference samples are available to construct satellite-based models (Daryaei et al., 2020; Som-ard et al., 2021; Yu et al., 2020). In this case, UAV data can be considered as a proxy of ground data to train EO-based models and therefore to achieve a larger spatial coverage (Emilien et al., 2021; Revill et al., 2020; Riihimäki et al., 2019). Till now, upscaling to provide data interoperability (i.e., UAV-RGB and S2 imagery) has not been applied for sugarcane yield estimation.

Objectives

This study attempts to estimate sugarcane yield for a complex landscape in Udon Thani Province, Thailand using a combination of UAV-RGB and S2 multispectral data. Field data were collected to calibrate UAV-based models so that detailed yield maps for 20 fields can be obtained. Subsequently, the augmented/enlarged yield dataset was used for a S2-based sugarcane yield modeling at the regional scale.

The specific objectives of the study are:

  • to explore the potential of a plant height information derived from UAV-RGB data to estimate sugarcane yield at smallholder fields using only a limited number of ground truth data;

  • to evaluate the optimal integration of UAV-RGB and S2 data to estimate sugarcane yield and within-field variability at the landscape scale.

Material and methods

Figure 1 illustrates the workflow used to achieve the specific objectives of the study. The input data consisted of:

  1. (i)

    UAV data to produce within-field sugarcane masks and maps of plant heights for 20 fields;

  2. (ii)

    41 yield measurements in 20 different fields collected through field work as well as farmer reported post-harvest yields for the same 20 fields;

  3. (iii)

    S2 imagery processed to construct monthly composites of various spectral features.

Fig. 1
figure 1

Workflow diagram of sugarcane yield estimation in Thailand using the random forest regression (RFR) algorithm and integrating UAV-derived RGB and Sentinel-2 (S2) imagery

Using these data, three yield models were calibrated using random forest regression (RFR):

  • YieldUAV: UAV-based yield map obtained through direct use of 41 field samples;

  • YieldS2: S2-based yield map obtained through direct use of 41 field samples;

  • YieldUAV_S2: S2-based yield map using the UAV yield map for model calibration.

The validation of each approach was carried out with yield data not used in the training process: (i) out-of-bag (OOB), and (ii) five-fold cross validation.

Study area and field measurement

Twenty conventional sugarcane fields were selected in Udon Thani Province, Northeast Thailand (between 16° 49′  38″  to 18° 03′  11″ N and 101° 59′  55″ to 103° 41′  29″ E). Udon Thani Province covers an area of 11 730 km2 (Fig. 2). The area comprises highlands with rolling plains of the Phu Phan range running north to south and plains along the Lam Pao River and the Mong River. Elevation ranges from 137 to 977 m above sea level. The climate is classified as tropical semi-humid dry-savanna (Köppen climate, classification: Aw) with a mean temperature of 27.2 °C and an average annual precipitation of 1423 mm. The main seasons are summer (mid-February to mid-May), rainy (mid-May to mid-October) and winter (mid-October to mid-February). This climate is highly suitable for sugarcane cultivation.

Fig. 2
figure 2

The study area in Udon Thani Province, Thailand. The background (a) shows a very high spatial resolution image (source: satellite image in QGIS version 3.8). The red polygons depict the locations of the 20 sugarcane fields selected for this study. b and c show sugarcane field 01 using Sentinel-2 (S2) data, respectively, the UAV- RGB imagery. The blue polygons are the three field sample locations of sugarcane field 01. Three sampling plots (20 m × 20 m) were used to collect the ground reference data in 2020

In this region, sugarcane is typically planted from September to November (rainy season) and from January to March (dry season), with a harvesting period from mid-December to March (Som-ard et al., 2022). The field survey for data collection was conducted from 15th November 2020 to 20th January 2021, with sugarcane crop cycles for only new plantations to fourth growth selected in 20 sugarcane fields. The sugarcane species, Khonkaen-3 (KK3) is popularly planted in northeast Thailand and was used for this study. Sugarcane planting patterns are 0.2 to 0.4 m for inter-row spacing with row spacing of 1.2 to 1.5 m.

For the 20 sugarcane fields, a total of 41 sampling plots with a size of 20 m × 20 m were delineated and probed to collect sugarcane data including height, stalk density and yield. For this purpose, each sampling plot was harvested and the mentioned characteristics were measured during the fieldwork. The geographical locations of each plot were recorded using a global navigation satellite system (GNSS) receiver of UniStrong G10 from Beijing UniStrong Science & Technology company and they are displayed in Fig. 2 below, along with an exemplary field with the sampling plot locations. In addition, the farmers also provided the total yield for each field after harvesting.

UAV data acquisition and preprocessing

The UAV data acquisition for all 20 fields was done with a Phantom 3 Professional UAV (DJI, Shenzhen, Guangdong, China) together with the field work from 15th November 2020 to 20th January 2021. Flight parameters were set as altitude 50 m, front and side overlap 80%, resulting in images with 2.1 cm spatial resolution using the DroneDeploy application (Sumesh et al., 2021). The camera operation was set for white balance as “sunny” and fixed exposure to stabilize the radiation fluctuation for RGB images (Sumesh et al., 2021). The UAV flights were performed at suitable times of the winter season (during the morning and at noon) avoiding adverse weather conditions (i.e., strong winds, rain and clouds). For each flight, 12 ground control points (GCPs) markers were used along the border of each field and geographical coordinates and elevation were recorded using a GNSS receiver together with a real-time kinematic (RTK) mode.

To create point clouds and ortho-mosaics, the UAV image data were processed in Agisoft Metashape Professional software version 1.7.5 (Agisoft LLC, 2021) using the Structure from Motion (SfM) algorithm. The ortho-mosaics and digital surface models (DSM) were exported as raster files with 2 cm pixel size, and the point clouds were exported to LAS format. The LAS file of each field was then used to generate the digital terrain models (DTM) using ArcGIS 10.8.1 software, following the processing steps of De Souza et al. (2017). Subsequently, plant height (PH) raster data—used later as predictor variables for estimating the sugarcane yield—were obtained by subtracting the DTM from the DSM. In a totally 12 GCPs, Eight GCPs were used for georeferencing, and four GCPs as checkpoints, following previous steps from Riihimäki et al. (2019) and Sumesh et al. (2021). A detailed accuracy assessment of the UAV-SfM point cloud is shown in Table A1 (see Appendix A).

Numerous studies have shown the high potential of five green-VIs from RGB images including ground-level index (GLI) (Louhaichi et al., 2001), green–red vegetation index (GRVI) (Tucker, 1979), excess green (ExG) (Woebbecke et al., 1995), shadow index (SI) (Zhou et al., 2021) and visible atmospherically resistant index (VARI) (Gitelson et al., 2002). We used these VIs to separate sugarcane/non-sugarcane for subsequent plant height estimations of the sugarcane plants.

For each of the 20 fields, 143 to 336 reference points (sugarcane/non-sugarcane) were obtained by visual interpretation of the UAV-RGB ortho-mosaics. The reference samples were split into 80% for training and 20% for validation. Subsequently, binary masks of sugarcane vs non-sugarcane were derived for each of the 20 fields using the random forest (RF) classifier (Breiman, 2001) based on training samples of (1) UAV-RGB, (2) UAV-PH and (3) UAV-green-VI datasets as inputs. To optimize the RF-classification model, a stepwise recursive feature selection process was used based on the “Mean Decrease in Accuracy” (MDA). This improve the computations and minimize overfitting (Immitzer et al., 2018, 2019; Som-ard et al., 2022). Hyperparameters were set to their default values: (i) number of trees (ntree) = 1000 and (ii) number of randomly selected variables for each node (mtry) = square root of the number of total input features. The accuracy assessment based on the validation dataset using a confusion matrix (Congalton, 1991) and the kappa statistic (Cohen, 1960) achieved overall accuracy values from 90.7 to 98.5% (kappa 84.5 to 96.6%) for the 20 fields (see Figure A1 in Appendix A). The map results provide detailed information on the spatial distribution of sugarcane pixels, respectively, non-sugarcane pixels (i.e., soil surface, grass, trees and shrubs) (see Figure A2). The sugarcane masks were used to remove non-sugarcane areas from the PH raster data for the yield modeling as shown in Figure A2 (right).

The masked PH dataset was extracted and for all 41 filed sampling plots nine metrics (Table 3) were calculated and used as predictor variables to build a RFR model (“YieldUAV” model). The sugarcane PH dataset was extracted using “extract” function of the “raster” (Hijmans et al., 2015) package in R software (R version 4.1.1).

Table 3 Plant height (PH) metrics used for (UAV) high resolution yield modeling

To consider intra-field heterogeneity, the RFR model prediction at field level was done by using a Moving Window (MW) approach. A moving window of the same size as the sampling plots was chosen. To apply the MW, the UAV data set was resampled to 1 m. For each MW the same nine PH metrics (Table 3) as for the sampling plots were calculated and the results were assigned to the center pixel resulting in layers with 1 m pixel size. The nine resulting rasters were used as input for the RFR-model to estimate sugarcane yield across the field.

Sentinel-2 dataset

In total, 38 S2 Level-2A scenes within seven S2 granules from November 2020 were used (cloud cover < 40%). The atmospheric correction was done using Sen2Cor Version 2.4 through the BOKU service (Vuolo et al., 2016). For cloud removal, the Scene Classification (SCL) band was used to exclude clouds, cirrus and cloud shadow (Zhu & Woodcock, 2012). Only the ten S2 bands with 10 m (B2, B3, B4 and B8), respectively, 20 m (B5, B6, B7, B8a, B11 and B12) spatial resolution were used for this study. All 20 m bands were resampled to 10 m.

Five different vegetation indices (VIs) were derived from the S2 images as yield predictor variables: (1) normalized difference vegetation index (NDVI) (Rouse et al., 1974), (2) normalized difference infrared index (NDII) (Klemas & Smart, 1983), (3) green normalized difference vegetation index (Gitelson et al., 1996), (4) enhanced vegetation index (EVI) (Huete et al., 1997) and (5) leaf area index (LAI) (Thenkabail et al., 2000). Similar indices were used in other sugarcane yield estimation studies (Abebe et al., 2022; Canata et al., 2021; dos Santos Luciano et al., 2021; Mulianga et al., 2013; Rahman & Robson, 2020; Sanches et al., 2018).

Due to cloud cover in the study area, the median compositing (MC) method was used to construct monthly composites (i.e., November 2020) for each S2 spectral band and the selected VI datasets. This method provides high efficiency to generate spatiotemporally consistent and gap-free image composites (Griffiths et al., 2019).

In addition to the MC method, minimum, mean and maximum compositing were also applied to the VI images (Griffiths et al., 2019; Som-ard et al., 2022; Vancutsem et al., 2007). All pixel-based compositing methods were used to produce for a given VI image to select the corresponding pixels to provide VI values of each image. These compositing methods reduced the impacts of clouds and shadows when generating image datasets.

The results of each pixel-based compositing method were manually compared and checked with a very high spatial resolution image (source: satellite image in Google Earth pro). The processed compositing datasets (i.e., spectral bands and the generated VIs) were used to construct the RFR-predictive model. Table 4 lists all predictor variables.

Table 4 The processed compositing datasets (i.e., spectral bands and indices) of Sentinel-2 imagery in November 2020 for all predictor variables as inputs, which were used for sugarcane yield estimation

For the RFR model training, only the S2 10-m-pixel falling entirely into the 20 m x 20 m sampling plot was used.

Sugarcane yield modeling

For the sugarcane yield estimation we used RFR, an ensemble learning method for multiple non-linear regression problems based on uncorrelated decision trees (DTs) (Breiman, 2001). Bootstrapping with replacement randomly selects many training sets, resulting in individual DTs that are later combined using the majority vote (Breiman, 2001). RFR analysis was conducted using the random forest package (Liaw & Wiener, 2002) in R software.

As advantages, the RFR method provides the mean squared generalization error (i.e., RMSE and R2) as OOB error of the RF model by managing a large number of input features and ranking their internal importance (Breiman, 2001). Recent studies have shown the high effectiveness of RFR to estimate sugarcane yield (Canata et al., 2021; dos Santos Luciano et al., 2021; Shendryk et al., 2021; Xu et al., 2020). To further improve model performance, the recursive feature elimination procedure method as “Increase of Mean Squared Error” (%IncMSE) based on the RF was used for feature selection, similar to previous studies (dos Santos Luciano et al., 2021; Immitzer et al., 2016).

The RFR model requires two hyperparameters: ntree and mtry. To find the best parameter settings we tested ntree values between 100 to 1,500 (15 different values) and obtained the best model performances for all models using the default value of 500. For mtry we used the default value of mtry = number of input variables divided by 3… In addition to the OOB result an additional five-fold-cross-validation was applied. The splitting was done on field level (in some fields two sampling plots were measured) and the random selection of fields and corresponding sampling plots are listed in Table 5. The same splitting was used for all model approaches.

Table 5 Detailed groups of a random sampling selection for the number of sugarcane fields together with the number of sampling plots for five-fold cross-validation (Every time one group is excluded)

For the sugarcane yield estimation three different model approaches were implemented:

  1. I.

    UAV-based model (YieldUAV): The modeling was done using the yield of the 41 sampling plots and nine predictor variables (Table 3) from the PH dataset. The predictions from this model were later used to train the S2-based model (YieldUAV_S2).

  2. II.

    S2-based model (YieldS2): The yield of the 41 sampling plots and 30 S2 features (Table 4) were used to train the RFR model. This would be the “standard approach” if no UAV data would have been available.

  3. III.

    S2-based model trained by UAV yield (YieldUAV_S2): this RFR model was based on the same 30 S2 features as in the YieldS2-model, but trained on the UAV yield map (YieldUAV) for the entire fields instead of the 41 sampling plots. To match the S2 pixel size, the UAV-derived yield map was aggregated from 1 m pixel size to 10 m (see Sect. 2.3). Note that the UAV based yields from all pixels in all fields (minimum distance to boundary: 20 m) were used to train the model.

The three RFR models were used to map sugarcane yields across 20 sugarcane fields in Udon Thani Province. The RFR-predictive models were assessed for efficiency using %IncMSE and OOB results. The optimized features for each model were ranked based on the RF method and are described in Appendix A (Table A2).

Performance of the three RFR-predictive models was assessed using the root mean square error (RMSE) and coefficient of determination (R2). These accuracy measures were computed using (i) OOB values, and (ii) the aggregated results of the five-fold cross validation on plot level, and (iii) the aggregated results of the five-fold cross validation on field level was compared against the yield values reported by the farmers.

The best model based on S2 input data was further applied to the entire province and the generated large-scale sugarcane yield map were compared to the official yield statistics of Udon Thani Province from OCSB (Office of the Cane & Sugar Board, 2020). Average yields were calculated for each of the 20 districts of this region and compared to the official values. The provincial average yield value was multiplied by the total sugarcane areas for comparison with S2-derived production data. For estimating yield across the region, this study used sugarcane field database of the year 2020 from OCSB. This database includes a sugarcane field map which was used to produce a yield map for the entire the region based on S2 datasets and the YieldUAV_S2 model.

Results

Estimated sugarcane yield from the UAV and Sentinel-2 datasets

Table 6 summarizes R2 and RMSE results when using (i) solely the field reference measurements for calibration (e.g., YieldS2 model) vs (ii) the results when using UAV data to augment/enlarge the reference dataset (e.g. YieldUAV_S2 model). Results clearly indicate that the upscaling of field measurements (i.e., using UAV data) to train S2 leads to a notable improvement compared to directly using the field observations and S2 data alone (YieldUAV_S2 vs YieldS2). More detailed results are reported in Table A3.

Table 6 Coefficient of determination (R2) and root mean square error (RMSE) using the random forest regression (RFR) approach based on different datasets

The scatterplots in Fig. 3 show the cross-validated estimated yields at plot- and field-level for the different prediction models. At plot level (a-c) (n = 41), RMSE and R2 values ranged between 5.25 to 9.36 t/ha for the RMSE and 0.65 to 0.90 for R2. As expected, the YieldUAV model gave the highest effectiveness for sugarcane yield estimation. Interestingly, the YieldUAV_S2 model was clearly superior to the YieldS2 model, which showed the lowest accuracy amongst the three RFR models.

Fig. 3
figure 3

Scatterplots of estimated yields obtained through five-fold cross-validation (plot level: a–c and yield by farmers (field level: d-f). Plot level: results from the UAV-based model (YieldUAV) (a), the S2-based model (YieldS2) (b) and the S2-based model trained by UAV yield result (Yield_S2) (c). Field level: results from YieldUAV (d), YieldS2 (e) and YieldUAV_S2 (f)

The validation results on field level (n = 20) are shown in Fig. 3d–f. The YieldUAV model gave again an outstanding performance against the farmer reported yields, but the YieldUAV_S2 model also provided high accuracy values. In comparison, the YieldS2 model trained directly on the field reference data obtained results with the lowest accuracy.

The three RFR models (i.e., YieldUAV, YieldS2 and YieldUAV_S2) were applied to map the intra-field sugarcane yield variability. Results are shown in Fig. 4 for two exemplary fields (id 01 and id 08). Estimated yields from the YieldUAV predictive model were smooth, with well-distributed spatial yield heterogeneity (left). The YieldUAV_S2 model also resulted in plausible yield maps thanks to the enlarged/augmented reference data (right). On the contrary, using solely the 41 sampling plots to train the YieldS2 model gave unsatisfactory results (middle).

Fig. 4
figure 4

Estimated yield maps (example: fields 01 and 08) from the optimized-random forest regression (RFR) models and the UAV-derived RGB and S2 image datasets. Yield maps were derived from the UAV-based model (YieldUAV) (a and d), the S2-based model (YieldS2) (b and e), and the S2-based model trained by UAV yield result (YieldUAV_S2) (c and f). As background UAV- RGB images of field 01 and 08 were used

Figure 5 depicts circular bar plots displaying the average estimated yield (t/ha) from the 20 sugarcane fields along a circle from the results of the three optimized-RFR-predictive models. Figure 5a and c show similar patterns because of the use of UAV yield results to train the YieldUAV_S2 model. Conversely, results from the YieldS2-predictive model (Fig. 5b) showed markable differences compared to the other RFR models.

Fig. 5
figure 5

Circular bar plots for the estimated yields for each of the 20 sugarcane fields. Results from the UAV-based model (YieldUAV) (a), the S2-based model (YieldS2) (b) and the S2-based model trained by UAV yield result (YieldUAV_S2) (c)

Mapping regional and within-field sugarcane yield variability across the region

The estimated sugarcane yield across the region is shown in Fig. 6. For the modeling, the YieldUAV_S2 model was used—this model was trained using all available samples (n = 1714). Highest yields were found in the southern areas, and lowest yield values in the northeast. Yields ranged from 41.0 to 89.9 t/ha and the mean yield value was 58.5 t/ha. The total production for the Udon Thani region was found 5.75 Mt, well in line with the provincial yield statistics from OCSB (at 5.64 Mt).

Fig. 6
figure 6

Spatial distribution map of estimated sugarcane yield for the entire Udon Thani Province based on the optimized-random forest regression (RFR)-predictive model involving UAV and S2 data (YieldUAV_S2 model)

Figure 7 shows a histogram of the frequency distribution of the estimated yield at 10 m spatial resolution across the region. A positively skewed distribution can be seen, splitting the region into (large) areas with relatively low yield and (smaller) areas with high yields.

Fig. 7
figure 7

Histogram showing the frequency distribution of estimated yield (t/ha) across the region from 10 m S2 datasets (bin width = 1) based on the optimized-random forest regression (RFR)-predictive model trained against the UAV yield (YieldUAV_S2 model)

The bar chart in Fig. 8 confirms high consistency in estimating sugarcane production from 20 districts using the YieldUAV_S2 predictive model.

Fig. 8
figure 8

Bar chart of statistical yield data (Megatonnes [Mt]) from 20 districts and estimated production based on the YieldUAV_S2-predictive model

Discussion

Sugarcane classification

The RF classification models achieved highly satisfactory results with excellent efficiency in sugarcane classification. The mapping results showed high overall accuracy by selecting important features. The RF classification model was also the fastest and most cost-efficient for crop mapping producers. Similarly, De Castro et al. (2018), Johansen et al. (2020) and Wang et al. (2022) indicated that UAV data together with the RF method yielded high benefits for mapping small-sized crop fields, offering efficient information as reference data to connect and train to large-scale EO-satellite data.

Estimated sugarcane yield

UAV yield results from the YieldUAV model were used to train the S2 datasets for constructing the YieldUAV_S2 model. In this two-step process, UAV data is used to enlarge/augment the labor-intensive and costly field work by generating highly accurate, high resolution yield maps for the probed fields, and this expanded reference data is used to train the large-scale S2-based models. The resulting YieldUAV_S2-predictive model showed high potential in estimating sugarcane yield with higher accuracies (lower RMSE and higher R2) compared to the YieldS2 model trained directly on the field samples, due to the augmented/expanded reference data set. This result underlines the high efficiency of UAV-derived reference data. The maps based on UAV data provide detailed information for producing vegetation or crop yield maps at multiple scales as also highlighted in recent studies (Daryaei et al., 2020; Jiang et al., 2022; Revill et al., 2020; Riihimäki et al., 2019; Zhang et al., 2019). Therefore, upscaling field measurements using UAV data showed high improvement over models directly trained on (fewer) field samples. Future studies should quantify the minimum number of field reference samples needed to generate useful UAV models.

When trained against 41 reference samples, the YieldUAV predictive model provided high efficiency and consistency in estimating sugarcane yield at several smallholder fields with low RMSE and high R2. The RFR predictive models achieved high accuracy (R2 > 0.89), compared to recent studies of (Sumesh et al., 2021; Xu et al., 2020), indicated UAV data are valuable information for sugarcane yield estimation at local scale. Results demonstrated different yield densities, strongly related to plant height. Similar to previous studies (De Souza et al., 2017; Sumesh et al., 2021; Xu et al., 2020; Yu et al., 2020), results successfully demonstrated that the PH dataset from the UAV-derived RGB improved sugarcane yield estimation compared to field measurements on a few plants. The RFR model addressed the main problems of overfitting and handling of input features for model performance (Xu et al., 2020). The estimated intra-field variability map from the YieldUAV model revealed high spatial yield heterogeneity (t/ha) and showed satisfactory result, which can be used to train the S2 model (YieldUAV_S2).

The S2 dataset trained against field measurements showed good yield estimation which could however be further improved by using the larger reference dataset derived from the UAV data. The RFR method—together with S2 imagery—was in this respect appropriate for estimating sugarcane yield, similar to results found by Canata et al. (2021) and Shendryk et al. (2021). Our study had however to deal with small field sizes and a complex landscape. The obtained accuracies demonstrate a good agreement with previous studies (Canata et al., 2021; Rahman & Robson, 2020).

Our study demonstrated the great potential of integrating UAV and S2 datasets to estimate sugarcane yield. Findings confirmed that the upscaling method overcame the small size of field measurement and improved the spatial heterogeneity of yield mapping. Other studies (Daryaei et al., 2020; Revill et al., 2020; Riihimäki et al., 2019; Zhang et al., 2019) also indicated that the key advantages of UAV data could be adopted as a reference to the EO dataset for vegetation upscaling. In future studies, our upscaling yield estimation approach can be applied to predict yields for sugarcane and other crops on regional or national scales.

Regional sugarcane yield mapping

Yield models based on S2 data have the potential to produce large scale (wall-to-wall) maps. Compared to the official yield data from the provincial statictics, the S2-derived sugarcane yield map for the Udon Thani Province indicated excellent efficiency based on the YieldUAV_S2 model (e.g., the YieldS2 model trained on the UAV maps). Our map show accurately the spatial distribution of inter- and intra-field yield variability. Such accurate sugarcane yield estimations could be used to complement statistical yield data with crop warning systems to inform national or international trade for Thai government policy, even for production management by local sugar mills. A spatial yield map is also a valuable information for sustainable agricultural management for sugar mills, authorities, local for increasing productivity and production. Such maps can also support farmers in better crop management practices for specific areas of the field. Moreover, such maps provide important evidence for various applications such as insurance, response and recovery after negative impacts from climate change and extreme weather events, etc. For very large regions, the additional use of meteorological variables is potentially helpful and should be further investigated [e.g., Cheng et al. (2022)]. To improve the model perfromance, future studies should consider other machine learning algorithms, as well as additional predictive variables such as topography, climate, or soil quality.

In this study, we obtained very good results with S2 datasets from the early harvest season (November/December), and our estimated yield map fully agreed with previous findings (Abebe et al., 2022; Mulianga et al., 2013; Shendryk et al., 2021). In future studies, EO time series data using our proposed “augmentation method” can be used to improve sugarcane yield estimations and also support for example crop stress detection [e.g., Berger et al. (2022)].

Conclusions

The technology with respect to unmanned aerial vehicles (UAVs) and satellite sensors is constantly improving, and we demonstrate that the combined data offer highly efficient means to estimate crop and crop yield across different scales. We specifically propose to integrate UAV and Sentinel-2 (S2) imagery to estimate crop yields so that costly and time-consuming field work can be minimized.

Important findings and further suggestions are:

  • in cases of insufficient field data, UAV data have a high potential to generate additional reference data to subsequently train satellite-based models (i.e. S1, S2 and Landsat data);

  • our proposed “augmentation methodology” confirmed that integration of UAV and S2 datasets improved estimation of sugarcane yield across scales;

  • future research should apply the proposed method to estimate yields in other regions and for other crop types;

  • the RFR method showed high potential for estimating sugarcane yield. However, future studies should assess this method against other machine learning algorithms;

  • our results offer valuable information to guide national food management, trade, sugar mills and government policy to promote sustainable agricultural management.