Cammarano, D., Rivington, M., Matthew, K. B., Miller, D. G., & Bellocchi, G. (2016). Uncertainties of different weather data input on three multi-models simulations of yield and water use.. Berlin (Germany).
|
Wallach, D., & Rivington, M. (2015). Standardised methods and protocols based on current best practices to conduct sensitivity analysis (Vol. 6).
Abstract: The purpose of this report is to propose a general procedure for sensitivity analysis when used to evaluate system sensitivity to climate change, including uncertainty information. While sensitivity analysis has been largely used to evaluate how uncertainties in inputs or parameters propagate through the model and manifest themselves in uncertainties in model outputs, there is much less experience with sensitivity analysis as a tool for studying how sensitive a system is to changes in inputs. This report should help make clear the differences between these two uses of sensitivity analysis, and provide guidance as to the procedure for using sensitivity analysis for evaluating system sensitivity to climate change. No Label
|
Rivington, M., & Wallach, D. (2015). Quantified Evidence of Error Propagation (Vol. 6).
Abstract: Error propagation within models is an issue that requires a structured approach involving the testing of individual equations and evaluation of the consequences of error creation from imperfect equation and model structure on estimates of interest made by a model. This report briefly covers some of the key issues in error propagation and sets out several concepts, across a range of complexity, that may be used to organise an investigation into error propagation. No Label
|
Bellocchi, G., Rivington, M., & Acutis, M. (2014). Protocol for model evaluation (Vol. 3).
Abstract: This deliverable focuses on the development of methods for model evaluation in order to have unambiguous indications derived from the use of several evaluation metrics. The information about model quality is aggregated into a single indicator using a fuzzy expert system that can be applied to a wide range of model estimates where suitable test data are available. This is a cross-cutting activity between CropM (C1.4) and LiveM (L2.2). No Label
|
Rivington, M., & Wallach, D. (2015). Information to support input data quality and model improvement (Vol. 6).
Abstract: Data quality is a key factor in determining the quality of model estimates and hence a models’ overall utility. Good models run with poor quality explanatory variables and parameters will produce meaningless estimates. Many models are now well developed and have been shown to perform well where and when good quality data is available. Hence a major limitation now to further use of models in new locations and applications is likely to be the availability of good quality data. Improvements in the quality of data may be seen as the starting point of further model improvement, in that better data itself will lead to more accurate model estimates (i.e. through better calibration), and it will facilitate reduction of model residual error by enabling refinements to model equations. This report sets out why data quality is important as well as the basis for additional investment in improving data quality. No Label
|