Towards improved calibration of crop models – Where are we now and where should we go?
Introduction
Nowadays, crop models are widely used in agronomic research and are proposed as possible tools for multiple applications including crop management, advice to policy makers, an aid in crop breeding or climate change impact assessment and adaptation to climate change (Ewert et al., 2015, Rosenzweig et al., 2013, Rötter et al., 2015). Recently several studies have involved multi-model ensembles of crop models (Asseng et al., 2013, Bassu et al., 2014, Palosuo et al., 2011). These studies have all shown that there is a wide variability between crop model predictions and even among the model predictions (see e.g. Martre et al. (2015)) by the same model when used by different users (Confalonieri et al., 2016). Clearly there is major uncertainty in crop model predictions and it is of fundamental importance to improve these models and how they are used in order to allow practical applications.
Uncertainty in crop model simulation can arise from three sources: (1) model structure, (2) input data and (3) model parameters (van Oijen and Ewert, 1999). (Note that Walker et al. (2003), in his list of possible locations of uncertainty, includes the above three sources plus context uncertainty, related to choice of boundaries of the modelled system, and outcome uncertainty, which is the accumulated uncertainty in model predictions and which results from the other sources.) Crop model improvement can target any of these, alone or in combination. Model structural uncertainty is partly due to the fact that models do not include all relevant explanatory variables. No model can include exhaustively all of the explanatory variables that determine crop growth and development. Various studies have proposed adding additional explanatory variables to existing crop models in order to capture specific sources of variability, for example row spacing (Flenet et al., 1996), CO2 concentrations (Kersebaum et al., 2009, Tubiello et al., 2000) or ground level ozone concentrations (Cappelli et al., 2016). Structural uncertainty also arises from the fact that the model equations, which describe the effects of the explanatory variables on the output variables, may not have the correct or best functional form. There have been many comparisons of different possible functional forms, for example development rate as a function of temperature (Alderman and Stanfill, 2017, Kumudini et al., 2014) or photosynthesis (Seidel et al., 2016).
Crop model inputs are affected by measurement or sampling errors of observed variables, limitations in data availability and high spatial and temporal variability of, e.g. weather variables. There have been several studies describing how to characterize data and discussing data requirements for crop model calibration (Boote et al., 2016, Kersebaum et al., 2015).
Finally, the values of the parameters in a model may not be those that minimize prediction error. Partly this is again a data problem; measurement errors and insufficient data both degrade parameter estimation. In addition, the calibration technique may not be capable of finding the best parameter values given the data. There have been studies in the literature proposing or comparing calibration methods for crop models (He et al., 2010, Makowski et al., 2002, Wallach et al., 2014). The authors show that multiple calibration approaches are possible, and that the choice has a significant effect.
The three sources of error lead to very different approaches to crop model improvement. The first source of error is related to the behavior of the system. Reducing this source of error is mainly through experimentation targeted to obtaining new information about specific processes or conditions, followed by mathematically describing the experimental results. Reducing the second source of error involves obtaining more, and better, data. Reducing the third source of error on the other hand is mainly about making the best possible use of existing data. It is this last pathway to model improvement, namely improved model calibration, that is targeted here.
Calibration, or parameter estimation, has been very extensively studied in statistics. However, it is often difficult to apply standard statistical methods and software to crop models. Reasons for this include the fact that these models have complex mathematical structure generally involving non-linearities and multiple discontinuities, that these models have multiple outputs which lead to complex correlation structures of errors, and that the software for these models is often difficult to couple with existing calibration software (Wallach et al., 2014). The result is that there is no standard approach to crop model calibration. Different approaches have been taken including least squares (Ramirez-Villegas et al., 2017, Zeng et al., 2016), Bayesian parameter estimation (Wallach et al., 2012), generalized likelihood uncertainty estimation (GLUE) (Chisanga et al., 2015) and various ad hoc, often trial and error, approaches (Li et al., 2015).
In addition, the details of how a method is applied can have important repercussions on the estimated parameters. For example, in a study of the GLUE method, He et al. (2010) found that the choice of likelihood function and method of combining likelihood values had a strong influence on parameter estimation results for the CERES-Maize model. In a comparison of GLUE and Markov Chain Monte Carlo (MCMC) parameter estimation, Makowski et al. (2002) found that both approaches are sensitive to the prior assumptions made about parameter values. The choice of data to use for calibration, and the way that data is used, for example in multiple calibration steps, also strongly affects the results of calibration (Angulo et al., 2013, Guillaume et al., 2011). Strategies for calibration of several individual models have been proposed (Ahuja and Ma, 2011a). Often the suggested strategies focus on a minimum data set for calibration, and the specific parameters to adjust to fit each output variable (for example Ma et al. (2011)). In the conclusion to a book with chapters on calibration for several different models, Ahuja and Ma (2011b) state “… due to the complexity of system models, there has not been a standard method to parameterize a system model. Methods reported in this book and elsewhere are often model and user dependent.” A highlighted message is “The development of a systematic and hopefully common protocol is needed to help users”.
Recommendations for crop model calibration generally are proposed by model developers, as in Ahuja and Ma (2011a). It is not clear to what extent these methods are applied by modelers. There is in fact very little information on the actual calibration practices of modelers. What is the diversity of calibration situations that are encountered (models, data, objectives), what are the choices faced during crop model calibration, to what extent is there a consensus in the community of crop modeling practitioners on calibration, or is there a wide diversity of practices? It is important to answer these questions, as a first step towards improving crop model calibration practices. Indeed, recommendations for improved calibration must take into account current practices, and try to understand the reasons behind those practices.
The overall goal of this study was to obtain a detailed picture of current practices of crop model calibration, from a broad cross section of practitioners who have been involved in calibration activities. Information on current practices was obtained by conducting a web-survey among a large group of researchers involved in crop modeling. A survey was preferred to a literature search, because information on calibration is often described very succinctly, if at all, in publications, whereas the survey format allowed us to examine multiple aspects of calibration in detail. All together 211 usable questionnaires were returned and analyzed. The survey results were used to translate the information on current practices into an outline for guidelines for crop model calibration. Based on the survey we identified a number of questions that arise in practice, and the range of responses of modelers. This was the basis for identifying the problems that need to be addressed in future guidelines.
Section snippets
Participation
The survey was targeted widely for the global crop modelling community. The request to participate defined the target audience as anybody who has “been responsible for a crop model calibration activity” (quotation from the survey questionnaire). To avoid multiple responses by the same researcher, each respondent was requested to fill out the survey only once, “for just one specific study (one data set, one model, one set of calibrated parameters), the one that you feel could be most useful to
Results
All survey results other than comments are shown in Table 2 (Appendix). Note that percentages there are relative to the number of responses for that question.
Discussion
The difficulty of crop model calibration and the need for standardized procedures has been noted (Ahuja and Ma, 2011b). The survey results here support this need. They show that the time devoted to calibration is significant, that the range of approaches to crop model calibration is large, and that almost all modelers feel that there are major difficulties involved in crop model calibration.
Improvements to the situation would involve more or better data, easier to use calibration software, more
Conclusions
A large scale survey of crop model calibration practices (211 responses) conducted among the global crop modelling community, showed that there is a very large variability in approaches to crop model calibration. To some extent this reflects differences in the context of the calibration exercise (the specific model and available data). However, calibration also involves multiple choices, and different respondents made different choices. The survey covered many of the areas of choice, and showed
Acknowledgments
We would like to thank all respondents for answering the survey. This study is part of the BonaRes project (BOMA03037514) of the Federal Ministry of Education and Research (BMBF), Germany. We also acknowledge the AgMIP and MACSUR projects for creating the collaborative framework that led to this study. Taru Palosuo was supported by the Academy of Finland through the PLUMES project (decision nos. 277403 and 292836).
References (40)
- et al.
Quantifying model-structure and parameter-driven uncertainties in spring wheat phenology prediction with Bayesian analysis
Eur. J. Agron.
(2017) - et al.
Implication of crop model calibration strategies for assessing regional impacts of climate change in Europe
Agric. For. Meteorol.
(2013) - et al.
Uncertainty in crop model predictions: what is the role of users?
Environ. Model. Softw.
(2016) - et al.
Crop modelling for integrated assessment of risk to food production from climate change
Environ. Model. Softw.
(2015) - et al.
Methodological comparison of calibration procedures for durum wheat parameters in the STICS model
Eur. J. Agron.
(2011) - et al.
Influence of likelihood function choice for estimating crop model parameters using the generalized likelihood uncertainty estimation method
Agric. Syst.
(2010) - et al.
Analysis and classification of data sets for calibration and validation of agro-ecosystem models
Environ. Model. Softw.
(2015) - et al.
Simulation of winter wheat yield and its variability in different climates of Europe: a comparison of eight crop growth models
Eur. J. Agron.
(2011) - et al.
Assessing uncertainty and complexity in regional-scale crop model simulations
Eur. J. Agron.
(2017) - et al.
The Agricultural Model Intercomparison and Improvement Project (AgMIP): protocols and pilot studies
Agric. For. Meteorol.
(2013)