Overfitting and uncertainty in the presence of model structural error
Vincent Larson, Principal Investigator, University of Wisconsin --- Milwaukee
Benjamin Stephens, Co-Principal Investigator, University Corporation for Atmospheric Research
Introduction and description
The current standard method to evaluate a new physical parameterization in a global atmospheric model re-uses much of the same observational data both to train the parameterization and to test it. This practice of re-using training data favors more complex parameterizations and is prone to overfitting.
To avoid reuse of training data for the purpose of testing, this project will use cross-validation, a resampling method that draws sample points for tuning and sets the others aside for testing. Cross-validation requires fast re-tuning. This will be accomplished by an approximate tuner (“QuadTune”) developed by the PIs.
Objectives
- Calculate (out-of-sample) present-day prediction error and use it to assess overfitting.
- Develop an improved method to evaluate new parameterizations.
- Explore the relationship between present-day prediction error and cloud feedback strength.
Methods
- Cross-validation is used to estimate prediction error of a global atmospheric model.
- Monte-Carlo integration over the parameter space is used in order to assess the relationship between prediction error and cloud feedback strength.
Impact and benefits
The project is expected to lead to improved understanding of the effects of model structural error and overfitting on uncertainty in cloud feedback strength. It is also expected to lead to an improved method for evaluating atmospheric parameterizations.