Application of geographically weighted regression to

0 downloads 0 Views 712KB Size Report
percentage vegetation index; GNDVI, green normalized difference vegetation index; ...... PLoS One 6(5): 1–10. ..... A decimal code for the growth stages of cereals. ..... 0.25. 0.31. 0.3. 0.33. 0.36. 0.32. 0.32. 0.36. BNDVI Mosaic,. March 10. -0.08.
Page 1 of 39

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

1

Application of geographically weighted regression to improve grain yield prediction from

2

unmanned aerial system imagery

3 4

Atena Haghighattalab1, Jared Crain1, Suchismita Mondal2, Jessica Rutkoski3 , Ravi Prakash

5

Singh2, and Jesse Poland1,4,*

6 7

1

Department of Plant Pathology, Kansas State University, Manhattan, KS, 66506, USA

8

2

International Maize and Wheat Improvement Center (CIMMYT), Int. Apdo. Postal 6-641,

9

06600, Mexico D.F.

10

3

International Rice Research Institute, DAPO Box 7777, Metro Manila, 1301, Philippines

11

4

Wheat Genetics Resource Center, Department of Plant Pathology and Department of

12

Agronomy, Kansas State University, Manhattan, KS 66506, USA

13

*Corresponing Author: [email protected]

14 15 16

Abbreviations:

17

AIC, Akaike’s information criterion; BNDVI, blue normalized difference vegetation index;

18

CIMMYT, International Maize and Wheat Improvement Center; DEM, digital elevation model;

19

DSM, digital surface model; DTM, digital terrain model; ENDVI, enhanced normalized difference

20

vegetation index; EYT, elite yield trials; GCP, ground control point; GIPVI, green infrared

21

percentage vegetation index; GNDVI, green normalized difference vegetation index; GW,

22

geographically weighted; GWR, geographically weighted regression; HTP, high-throughput

23

phenotyping; NDVI, normalized difference vegetation index; NIR, near infrared; PCR, principal

24

component regression; SD, standard deviation; UAV unmanned aerial vehicle; UAS, unmanned

25

aerial system; VI, vegetation index

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

26

ABSTRACT

27

Phenological data such as plant height are important assessments of the in-season growth of

28

crops, though this assessment is generally limited by the availability of spatial and temporal

29

data during the growing season. Unmanned aerial systems (UAS) have the potential to provide

30

high spatial and temporal resolution for phenotyping tens-of-thousands of small plots under

31

field conditions without requiring substantial investments in time, cost, and labor. Thus, there is

32

considerable interest in implementing UAS as a rapid screening tool in crop breeding programs.

33

The overall objective of this research was to determine whether an accurate remote sensing

34

based method could be developed to monitor plant growth and estimate grain yield using aerial

35

spatio-temporal imagery in small plot wheat yield evaluation nurseries. We used a modified

36

consumer grade camera mounted on a low-cost unmanned aerial vehicle (UAV) to obtain blue,

37

green and near-infrared imagery. The UAS was deployed multiple times throughout the growing

38

season in yield trials of advanced breeding lines with irrigated and drought-stressed

39

environments at the International Maize and Wheat Improvement Center (CIMMYT) station in

40

Ciudad Obregon, Sonora, Mexico. We assessed data quality and evaluated the potential to

41

predict grain yield on a plot level by examining the relationships between information derived

42

from UAS imagery and the final measured grain yield. Using geographically weighted (GW)

43

models, we predicted grain yield for both drought and irrigated environments. The relationship

44

between measured phenotypic traits derived from imagery and grain yield was highly

45

correlated (r=0.74 and r=0.46 [p < 0.001] for drought and irrigated environments respectively).

46

Residuals from GW models were much lower and less spatially dependent than methods using

47

principal component regression suggesting the superiority of spatially corrected models. These

48

results show that image based vegetation indices collected from high-throughput UAS imagery

49

can be used to predict grain yield and subsequently used for selection decisions and to improve

50

genomic prediction models.

Page 2 of 39

Page 3 of 39

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

51

Wheat is a primary staple crop throughout the world, with a total projected production of 727

52

million metric tons in 2016 (USDA, 2016). Due to the emergence of more aggressive pests and

53

diseases, shortage of available water resources, lack of fertile soil, and climate change, wheat

54

production is projected to fall behind the increasing demand (Ray et al., 2013). With limited

55

options to increase the amount of land under production, it is generally agreed that increasing

56

crop yield per unit area, rather than expanding cultivated land area, is needed to meet the

57

world demand for wheat by 2050 (Muir et al., 2010; Foley et al., 2011). A rate of gain per year

58

of 2% or more is needed to meet the projected targets (Ray et al., 2013), while most studies

59

estimate the current rate of genetic increase is estimated at 1% or less (Reynolds et al., 2012).

60

Thus, substantial investments and deployment of new tools in plant breeding programs are

61

needed.

62 63

Several studies have indicated that breeding programs must substantially increase or even

64

double the rate of variety development and improvement to increase crop yield at a sufficient

65

pace to meet growing demands (Reynolds et al., 2009; Tester and Langridge, 2010).

66

Theoretically, any gains in plant breeding can be summarized from the breeders equation which

67

is a function of the additive genetic variance, trait heritability, selection intensity, and breeding

68

cycle time (Falconer and Mackay, 1996; De La Fuente et al., 2013). A critical part of traditional

69

phenotypic selection involves the use of manual data collection that is time-consuming, labor-

70

intensive, expensive and often burdensome or limited when making selections in large

71

populations. The precision of the data collected directly relates to the accuacy of selection by

72

correctly partitioning environmental and genetic variance. While there have been little changes

73

in phenotypic tools and evaluation methods (White et al., 2012), astonishing advances over the

74

last 5-10 years have occurred in sequencing technology resulting in many advanced tools and

75

methods to assess genotypes and improve plant breeding techniques through genomics and

76

reducing the time of the breeding cycle (Elshire et al., 2011; White et al., 2012; Cobb et al.,

77

2013). The grand challenge remains in measuring plant phenotypes with the same speed and

78

precision as is possible for observing the genome. This has led to active innovation in field-

79

based high-throughput phenotyping (HTP) platforms.

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

80 81

High-throughput phenotyping platforms could provide the keys to connecting genotype to

82

phenotype by increasing the capacity, efficiency, and precision of evaluating large plant

83

populations. Due to potentially very rapid measurements, HTP can increase the population sizes

84

screened for desirable traits in breeding nurseries. An increase in population size enabled by

85

HTP increases the probability of identifying valuable traits and the genetic basis of those traits.

86

Related to the breeders equation, a breeding program can achieve a higher capacity through

87

genetic gain by increasing the intensity and accuracy of selection (Goggin et al., 2015; Poland,

88

2015; Reynolds and Langridge, 2016).

89 90

The precision and accuracy with which these traits can be measured is also expected to

91

increase using HTP compared to manual measurements (White et al., 2012; Goggin et al.,

92

2015). The emphasis on high dimensionality, the need for automation and more non-

93

destructive sampling techniques in general, necessitates that HTP platforms often rely on

94

imaging to measure plant phenotypes (Goggin et al., 2015). One emerging technology allowing

95

plant breeders to rapidly screen large numbers of genotypes is unmanned aerial systems (UAS).

96

The main advantage of UAS over traditional manual approaches is the ability to cover an entire

97

experiment area in a very short time, enabling researchers to monitor larger fields with higher

98

temporal resolution and higher spatial resolution during the growing season.

99 100

Different imaging systems are being used on unmanned aerial vehicle (UAV) phenotyping

101

platforms making them capable of generating various measurements useful for plant breeding

102

programs (White et al., 2012; Deery et al., 2014; Liebisch et al., 2015). High resolution

103

multispectral cameras can be used to produce 3D models of the experiment area resulting in

104

the extraction of plant height using a digital elevation model (DEM) that determines height as

105

the difference between the digital terrain model (DTM) and the digital surface model (DSM)

106

(Rosnell and Honkavaara, 2012; Feng et al., 2015). Other sensors include hyperspectral and

107

thermal cameras that can be used to investigate different vegetation indices (VIs) or canopy

108

temperatures and their relation to crop biomass and yield (Hruska et al., 2012).

Page 4 of 39

Page 5 of 39

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

109 110

Several spectral reflectance measurements and VIs have shown a strong relationship to plant

111

health and crop growth. From this, remote sensing has proven to be a valuable tool for

112

estimating crop yields in advance of harvest. One of the most common VIs, the normalized

113

difference vegetation index (NDVI), has been used by researchers for crop yield predictions

114

since the 1980’s (Singh et al., 2001; Teal et al., 2006; Sultana et al., 2014; Wang et al., 2014). In

115

a study conducted by Salazar et al. (2007) a strong correlation was found between winter

116

wheat yield and NDVI extracted from satellite imagery during the critical period of development

117

and productivity occurring during April to May in Kansas. Grain-yield prediction using remote

118

sensing has been extensively studied in cereal crops, such as wheat, using satellite and aerial

119

imagery (Prasad et al., 2006; Salazar et al., 2007; Mkhabela et al., 2011; Manjunath et al., 2016;

120

Rutkoski et al., 2016a) or ground-based platforms (Raun et al., 2001; Rischbeck et al., 2016).

121

However, little work has been conducted using UAV platforms.

122 123

The use of physiological measurements to predict grain yield is an indirect selection approach

124

proposed and implemented in wheat breeding (Babar et al., 2006a; Crain et al., 2016; Rutkoski

125

et al., 2016b). Vegetation indices have been widely studied as the potential tool to differentiate

126

genotypes for yield under different environmental growing conditions such as irrigated and

127

non-irrigated (Arnold et al., 1995; Babar et al., 2006b; Araya et al., 2013; You et al., 2013).

128

Additionally, several studies have examined how yield-enhancing traits, such as stay-green, can

129

be used as a predictor of grain yield using spectral responses. Lopes and Reynolds (2012)

130

studied temporal NDVI data from grain fill to maturity to improve understanding of genotypic

131

stay-green variation. In this study, they found that stay-green variables explained 30% of yield

132

variability in multiple regression analysis, although the cumulative effect that stay-green

133

variables provide, together with other traits may allow for additional insights to be gained

134

about plant health and final grain yield.

135 136

Data derived from HTP can provide deeper and broader insight into the crop growth,

137

development and performance than what is currently avaliable to breeders. With the addition

Page 6 of 39

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

138

of georeferenced and spatial data inherent in HTP and imagery, researchers can utilize tools not

139

traditionally used in the area of plant breeding including spatial modeling commonly applied in

140

geospatial sciences. The purpose of this research was to evaluate the performance of

141

geographically weighted models in predicting grain yields at plot level in breeding nurseries.

142

The foundation of geostatistical rationale is the understanding that spatial data, information

143

about the locations and shapes of geographic features (wheat plots) and the relationships

144

between them, stored as coordinates and topology, will vary across a field (Mitchell, 2005).

145

Regression-based models mostly ignore this assumption, which could potentially impact the

146

accuracy of prediction models.

147 148

In this scope, a typical multiple linear regression model follows:  =  + ∑ ,   + 

149

(1)

150

With  is the ith observation of the dependent variable,  the ith observation of the jth

151

independent variables, and  the residual that is identically and independently distributed.

152

When applied to spatial data, this model assumes a stationary spatial process (spatially

153

consistent relationships between the dependent variable and each explanatory variable across

154

the field); and produces one single model for all parts of the study area. Field trials, however,

155

are inherently spatial, typically having spatial trends across the field which will impact the

156

accuracy of these models.

157 158

Along with typical linear regression, other statistical measures have been proposed to alleviate

159

the challenges of spatial data. The application of mixed models to spatial data has generally

160

resulted in improved accuarcy and efficiency (Gilmour, et al. 1997). Even with these

161

improvements, geostatistical mixed model analyis like kriging still assume that spatial variaion

162

in the phenomenon is statistically homogeneous and the same pattern of variation can be

163

observed at all locations (Oliver and Webster, 2015).

164 165

Geographically weighted regression (GWR) is a local statistical approach to model spatially

166

heterogeneous processes, which could be applicable to field trials. This regression model is

Page 7 of 39

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

167

used to analyze spatial nonstationarity; when the extent of relationships among variables

168

differs from location to location (Brunsdon et al., 1996; Fotheringham et al., 2002; Lu et al.,

169

2014a). For example in field trials, the location refers to the distance between plots of one

170

experiment rather than to multi-site or multi-environment trials. In this approach, unlike other

171

regression models, the spatial non-stationary relationship is assumed and is tested, based on

172

Tobler's first law of geography: everything is related to everything else, but near things are

173

more related than distant things (Tobler, 1970). Continuing the above example, plots located

174

on the far north end of the experiment would be assumed more like other plots located (closer

175

distance) in the northern area of the experiment, than plots at the southern end.

176

Geographically weighted regression allows the relationships to vary over space, and extends

177

the traditional regression framework of equation (1) by allowing local variations in rates of

178

change so the coefficients in the model are no longer global estimates but are specific to a

179

location i. This is the core of GWR, in the linear form:  =  + ∑ ,   + 

180

(2)

181

where  is the value of the jth parameter at location i. Instead of remaining the same

182

everywhere,  now varies in terms of the locations (i) (Brunsdon et al., 1996; Fotheringham et

183

al., 2002; Mennis, 2006; Lu et al., 2014a). While GWR should allow more flexibility of modeling,

184

there has been limited application of the technique in agriculture and plant breeding. One

185

agricultural example of GWR, in comparison to spatial analysis of regression kriging showed

186

that GWR resulted in lower root mean squared error and higher accuracies than regression

187

kriging (Wang et al., 2012)

188 189

In the context of growing interest in UAS applied to precision agriculture studies, we evaluated

190

the relationship between the data extracted from UAS imagery, including green NDVI (GNDVI),

191

blue NDVI (BNDVI), enhanced NDVI (ENDVI), Green Infrared Percentage Vegetation Index

192

(GIPVI), and plant height during the growing season to grain yield. Using the extracted data, we

193

evaluated the application of spatio-temporal UAS imagery to predict within-season grain yield

194

in a large international wheat breeding program using GWR and compared those results to

195

principal component regression (PCR) of the same data. PCR is a popular technique that can

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

196

adequately account for multiple variables that may have varying levels of correlation between

197

variables (Mevik and Wehrens, 2007).

198 199

MATERIALS AND METHODS

200 201

Study Area

202

The study was conducted at Norman E. Borlaug Experiment Station at Ciudad Obregon, Sonora,

203

Mexico, in two different growing environments drought and irrigated. The genetic materials

204

were advanced breeding lines in the advanced yield trials (AYT) of the CIMMYT bread wheat

205

breeding program. The experiment consisted of 280 advanced wheat lines distributed in 10

206

trials, with each trial consisting of 28 entries and two checks. Each trial was arranged as an

207

alpha lattice design with three replications and six blocks per replication. The different

208

environments were based on mega-environments which are major wheat growing regions that

209

experience similar growing environments, with the tested environments of drought and

210

irrigated representing approximately 10 and 32 million ha world wide respectively (Braun et al.,

211

1996). The plot size for each experimental unit in irrigated trials was 2.8 m long and 1.6 m wide

212

(plot area = 4.48 m2) sown on two 0.8 m wide raised beds with two rows per bed, whereas plots

213

in drought trials were 4 m long and 1.3 m wide (plot area = 5.2 m2) sown on flat land as 6 rows

214

18 cm apart with 40 cm interplot spacing. The drought trial was sown at the optimum planting

215

date with minimal irrigation throughout the season using drip tape. The irrigated environment

216

represents an optimal planting date (mid-November) with optimal flood irrigation throughout

217

the growing season. During the crop season, the drought environment received 180 mm of

218

irrigation, and the irrigated environment received 500 mm of water. Fertilizer and pesticide

219

control were similar in both environments.

220 221

Platforms and Cameras

222

High-throughput phenotypic data from a UAS was collected as previously described by

223

Haghighattalab et. al (2016). Briefly, we used a consumer grade Canon S100 camera modified

224

by MaxMax (LDP LLC, Carlstadt, NJ 07072, USA, www.maxmax.com) to collect blue, green, and

Page 8 of 39

Page 9 of 39

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

225

near infrared (NIR) light. The camera was mounted on a low-cost UAV, (IRIS+, 3D Robotics, Inc,

226

Berkeley, CA 94710, USA) with a custom designed gimbal to allow for nadir image collection.

227

Supplementary Table 1 contains detailed specifications for instruments used to collect data.

228 229

Ground Control Points and Reflectance Calibration Panel

230

To geo-reference the aerial images, 13 ground control points (GCPs) were distributed across

231

each field at the beginning of the season. The GCPs coordinates were measured with a Trimble

232

R4 RTK GPS (Trimble, Sunnyvale, CA 94085, www.trimble.com) that has a horizontal accuracy of

233

2.5 cm and a vertical accuracy of 3.5 cm. These targets were maintained throughout the crop

234

season to enable the overlay of measurements from multiple dates. The GCPs were designed as

235

coded targets (printed as circles with black and white segments) on 25 cm x 25 cm square white

236

metal sheets and mounted on a 50 cm post to provide easy identification of GCPs in the image

237

processing software. Six white 30 cm x 30 cm square ceramic tiles mounted on 10 cm blocks

238

were also used to have a variety of heights in GCPs in order to increase the accuracy of height

239

extraction analysis.

240 241

To enable consistent analysis of extracted data from aerial imagery across multiple days, we

242

placed a calibration panel in the field before each flight. This foldable panel consisted of six

243

individual 7 mm thick lightweight plywood panels with levels of gray from 0% being white and

244

100% being black and were attached using metal hinges. Using a handheld spectroradiometer

245

(ASD Inc., Boulder, CO 80301, www.asdi.com) on a sunny day, the spectral reflectance of these

246

panels were measured at a fixed height of 50 cm from five different angles. We also used this

247

panel for white balance adjustment of the raw imagery captured by our consumer grade

248

camera.

249 250

Field Data Collection

251

At maturity, the whole plot was harvested to estimate grain yield in tons per hectare. For the

252

irrigated environment, manual measurements for plant height were recorded for the first

253

replicate in each trial (n=300), and for all plots in the drought environment (n=900). These

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

254

measurements were taken as the distance from the ground to the average top of the terminal

255

spikelet for a representative plant, usually from the middle of the plots. Before measurement,

256

culms were manually pulled up to the vertical position if they were lodged (Bell et al., 1994).

257 258

For each environment, aerial imagery was taken at multiple time points using the UAV

259

phenotyping platform (IRIS+ and a modified Canon S100 camera) during the growing season,

260

with observation days spaced at approximately seven-day intervals, depending on the weather

261

condition. All images were taken in RAW format (.CR2) to avoid loss of image information. An

262

overview of each day of UAS observation is summarized in Table 1.

263 264

HTP Image Processing

265

For each experiment we collected three sets of observations from heading through senescence

266

(Zadoks et al., 1974). The semi-automated data analysis pipeline presented in Haghighattalab et

267

al. (2016) was used to analyze each dataset and extract plot level phenomics values. Image

268

processing was primarily conducted using Python scripts described and availabe from

269

Haghighattalab, 2014; 2015, open source software QGIS (http://qgis.osgeo.org) and Agisoft

270

PhotoScan package (Agisoft LLC, St. Petersburg, Russia, 191144).

271 272

To generate plot level values for VIs and plant height, the image processing developed by

273

Haghighattalab et al. (2016) was used. In summary, HTP data processing, for each time point of

274

data collection, included the following steps:

275

1) Pre-process raw images.

276

2) Perform camera alignment.

277

3) Import and reference GCPs.

278

4) Optimize GCP alignment.

279

5) Construct a dense point cloud.

280

6) Classify the data within the dense point cloud as plant or soil.

281

7) Build a digital elevation model using the classified point cloud.

282

8) Form an orthomosaic image using two different blending methods.

Page 10 of 39

Page 11 of 39

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

283

9) Generate an orthomosaic and DEM.

284

10) Calculate the various VIs and plant height.

285

11) Extract the plot-level data.

286

Due to the available bandwidths using the modified Canon S100, the calculated VIs were limited

287

to GNDVI, BNDVI, ENDVI, and GIPVI. The following formulas were used to calculate each index:

288

BNDVI =

     

   !

; GNDVI =   

!

  !"∗

; ENDVI =   

!"∗



; GIPVI =   

!

.

289 290

Plant Height Extraction

291

PhotoScan generates a dense point cloud based on the estimated camera positions. The

292

program also calculates depth information for each photo to be combined into a single dense

293

point cloud (Agisoft LLC, 2016). Dense point clouds with high accuracy were generated, where

294

the user has the option to specify the level of detail. Using high level accuracy resulted in DEM

295

with accuracy of 1.83 and 1.79 cm/pixel for the drought and irrigated trials, respectively).

296 297

To extract plant height, both the DTM and the DSM were produced, with plant height being

298

calculated as the difference between these two models (equation 3):

299

Plant Height = DSM – DTM

300

The generated dense point cloud includes everything that can be seen in the images, pixels

301

from the ground, vegetation, and any other artifacts or noise. PhotoScan offers two options of

302

dense point cloud classification: 1) automatic classification of the points into two categories –

303

ground points and other. 2) manual selection of a group of points to be placed in a particular

304

class from the standard list known for LIDAR data, such as vegetation, water, or road surface.

305

Three parameters control automatic ground point classification: maximum angle, maximum

306

distance, and maximum cell size (Agisoft LLC, 2016). Point cloud classification provides ways to

307

customize the DEM generation step as the software allows choosing types of objects within the

308

scene to be reconstructed and indicate the corresponding point class as a source data for DEM

309

generation. For example, DEM reconstruction based on ground points only, allows users to

310

export DTM (as opposed to DSM, based on the complete dense point cloud, excluding the

311

artifacts and noises) (Agisoft LLC, 2016).

(3)

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

312 313

In the plot imagery, the default values for maximum angle and maximum distance parameter

314

values were reduced since the height for crops are considerably lower than heights of other

315

objects like buildings. To estimate plant height, automatic classification of ground points was

316

performed following by manual classification of vegetation points, generating DTM and DSM

317

for each dataset. Training sets for classes of ‘ground’, ‘low vegetation’, and ‘high vegetation’

318

were defined by sampling the points. Using the classified point cloud, the DTM was extracted

319

by only using ground points, and the DSM by using data assigned to the vegetation points. For a

320

small field, the generated classified point cloud can be used to calculate plant height directly.

321

However, the point cloud exported from PhotoScan are very large files and for large wheat

322

breeding nurseries, as used in our study, the generated DEM was used.

323 324

Using our custom field map based method, two plot boundaries for each field were extracted.

325

One plot boundary was buffered to avoid pixels outside the plot, and was used for the DSM

326

extraction, as we are only interested in values reflected from wheat plants. The second plot

327

boundary was extended to have the DTM values of the ground surrounding the wheat plots.

328

Mean, standard deviation (SD) for DSM, and lower quantile (25%) for DTM were calculated.

329

Height was calculated for each plot as the difference between the mean values of DSM and

330

lower quantile values of DTM as these values best represent the ground elevation point in the

331

DTM.

332 333

Using the classified DEM, orthomosaic using both blending methods (average and mosaic) were

334

generated. In the average mode, the brightness values of the final orthomosaic is the average

335

of all the overlapping images, while mosaic mode only uses the pixel value from the image

336

taken from the most nadir view of that given point (Agisoft LLC, 2016; Haghighattalab et al.,

337

2016).

338 339

As described in our developed pipeline, after calibrating each orthomosaic data set for each

340

environment using empirical line method, VI maps using open source software QGIS were

Page 12 of 39

Page 13 of 39

Crop Sci. Accepted Paper, posted 05/11/2017. doi:10.2135/cropsci2016.12.1016

341

generated (Smith and Milton, 1999; Haghighattalab et al., 2016). Statistical values for each plot

342

for each day of data were calculated that included the minimum, mean, maximum, and

343

standard deviation for both blending methods.

344 345

HTP Data Analysis

346

Broad-sense Heritability

347

Within environment and trial, broad-sense heritability were calculated on a line-mean basis

348

within environment, commonly known as repeatability (Piepho and Möhring, 2007), for

349

extracted VIs measured on individual dates and grain yield. To estimate heritability, a two steps

350

process were performed where we first calculated the plot-level value for HTP measurements

351

and then fit a mixed model for replicate, block nested within replicate, and genotype as random

352

effects. Broad-sense heritability for each trait within trial within environment was calculated

353

from equation 4 (Holland et al., 2003).

354

3" =

> 456789:;

? > 456789:;