“Prediction is very difficult, especially about the future.”
—Niels Bohr, Danish physicist (1885 – 1962)
Niels Bohr is not the only one who has had trouble predicting the future. This article questions the ability of long-term climate models to accurately predict the climate. By extension, this article questions the wisdom of implementing expensive and disruptive policies aimed at combating climate change, such as corporate average fuel economy (CAFE) standards, carbon cap and trade plans, the ban on incandescent light bulbs, and the Environmental Protection Agency’s (EPA) proposed tax and regulation of carbon dioxide under the 1970 Clean Air Act.
Climate models are composed of five basic components:
Let’s consider each of these components. How are historical temperatures known, other than through actual thermometer readings? Of course, no one has been keeping good thermometer readings of Hawaii’s temperature since 1700. We can, however, get historical temperatures from tree rings, which show cold and warm growing years. How are historical carbon dioxide levels known? Ice cores can provide this. How about old solar radiation levels? Surprisingly, fossilized seashells can provide this information. While not perfect, historical model input and output readings appear to be perfectly adequate.
How can we determine if a model truly has predictive ability?
Climate models start showing their weaknesses in the physical relationships they capture—or don’t capture. Consider clouds. Clouds should definitely be included in climate models because, as we know, clouds both reflect incoming and trap outgoing radiation. Clouds thus play a critically important role in the earth’s climate. If the sky is cloudy on a winter night, then the air will be warmer than on a night with clear skies. If the world really heats up, will there be more or less cloud cover? Does the climate drive clouds or do clouds drive the climate? It turns out that scientists don’t know.
Here’s what Henrik Svensmark, director of the Center for Sun-Climate Research at the Danish National Space Center in Copenhagen, wrote:
All we know about the effect of [carbon dioxide] is really based on climate models that predict how climate should be in 50 to 100 years, and these climate models cannot actually model clouds at all, so they are really poor. When you look at them, the models are off by many hundreds percent. It’s a well-known fact that clouds are the major uncertainty in any climate model. So the tools that we are using to make these predictions are not actually very good.
The Pew Center on Global Climate Change adds:
Uncertainty remains about how particular aspects of climate change will play out in the future, such as changes in cloud cover or the timing or magnitude of droughts and floods.
Even the Intergovernmental Panel on Climate Change (IPCC) concurs, saying:
Significant uncertainties, in particular, are associated with the representation of clouds, and in the resulting cloud responses to climate change.
It’s not that climate scientists have ignored clouds due to an oversight; it’s that scientists don’t understand how clouds are formed. That’s where Svensmark comes in. His research suggests that clouds are formed via deep space cosmic rays hitting the earth. These cosmic rays are, in turn, affected by the amount of radiant energy put out by our sun. The sun’s energy blocks these cosmic rays. When the sun’s energy output is high, as it has been recently, there are fewer cosmic rays hitting earth, fewer clouds being formed, and the earth heats up. This large effect from cloud creation overshadows the minor effect from additional solar radiation hitting our planet.
This is a hot topic and just last month CERN, the European Organization for Nuclear Research, released some preliminary results of an experiment designed to look into exactly this method of cloud formation. Are these cosmic ray/cloud relationships included in climate models? How could they be if the relationships are not fully understood?
Scientists don’t understand how clouds are formed.
Other questions to ask about climate models are: how accurately do they reflect reality and how well do they predict the future? In order to answer these questions, we have to validate the model. There are really three components to model validation: (1) calibration, (2) holdout testing, and (3) long-term testing.
Calibration means building the model and having it provide us with data we already know. We do that by comparing the modeled values—those that come out of the model—with actual historical data. Climate models do a good job in this respect. According to the Pew Center on Global Climate Change, “Most of the current models do a reasonable job of simulating past climate variability on time scales of decades to centuries…”
Holdout testing, or hindcasting, is the next step. We hold out, or hide, some of our historical data, and then rerun the model without those data. If the model’s results are close to the actual, held-out values, then we can start to believe that the model reflects reality with some degree of accuracy.
As good as holdout testing is, it isn’t perfect. Ask yourself the following question: What’s the chance of a meltdown in the residential housing market? In 2005, you might have assigned a low probability. Now that you’ve lived through the subprime fiasco and the housing market collapse, you will, whether consciously or subconsciously, assign higher probabilities to such a “rare” event. Whether modelers are honest with themselves or not, it is impossible to ignore what we know has already happened, and if a model fails to replicate known events, it will be tweaked until it does.
Just like statisticians, we modelers have ways to fool others and ourselves. Here are five ways to be bamboozled by deceptive models:
How can we determine if a model truly has predictive ability? There’s really only one way. We finish it, lock it down, and put it away like a good bottle of Bordeaux. The reason we lock it down is that each time model inputs are altered, even slightly, the model results might change dramatically. Each time we tweak a variable, we are effectively giving the model a “peak ahead” to events that have already happened and the clock needs to start anew. As economist David Friedman points out, “In order to test the predictions of a model you need to do it against actual predictions—information that didn't go into building the model.”
Just like statisticians, we modelers have ways to fool others and ourselves.
It is inevitable that our models will evolve as computing power, theory, and climate measurements progress. Once the models are stable, however, they should be stored for a length of time sufficient to “age” them and test their predictive ability. Climate models have had some success in this regard, such as NASA’s Goddard Institute for Space Studies’ model of 1988, in which actual measured temperatures matched reasonably well the model’s middle scenario for a decade and a half. But successes like this are just baby steps.
How long should climate models be put aside to test their predictive value? If we are trying to predict the climate for the next 100 years, the models should be put away for at least a few decades because it will always be easier for them to nail down the 10-year forecast than the 100-year forecast. That’s due to the fact that there’s less uncertainty in the near term than in the long term.
This uncertainty, or variability, in the weather is another reason to put the models aside for longer. Consider the historical snowfall chart for Donner Summit in the Sierra Nevada mountain range. There doesn’t seem to be any pattern; the data seems to be truly random, with the snowfall of one year having little relationship to the snowfall of another.
Data like the above make it hard to discern the underlying trend in the climate. Do two warm years mean the earth is warming or are those just a sign of the randomness of the weather? This requires us to test our climate models for longer periods to be sure we are picking up true climate trends rather than just “noise.”
Consider the Intergovernmental Panel on Climate Change’s following claim that climate models are accurate. “There is considerable confidence that climate models provide credible quantitative estimates of future climate change, particularly at continental scales and above. This confidence comes from the foundation of the models in accepted physical principles and from their ability to reproduce observed features of current climate and past climate changes.”
The IPCC is saying that these models are accurate because they model known physical principles properly and have the ability to replicate past events. What if there are unknown physical principles that can’t be modeled because they aren’t known, like the cosmic rays and cloud formation? And the ability to replicate past events covers only the first or, perhaps, the first and second, of the three requirements for validating models.
That climate models need to prove their predictive ability in the long-term is a given. The question is: What do we do in the meantime? Some, like climate change blogger Coby Beck, say we need to act now: “It is only long-term predictions that need the passage of time to prove or disprove them, but we don't have that time at our disposal.”
Other people, including myself, think that if long-term climate models are slated to dictate trillion-dollar decisions affecting the world economy, they must prove their accuracy. With such a fundamental omission as cloud formation, climate models may be more misleading than helpful. Predicting the future of something as complicated as the earth’s climate is indeed difficult with inadequate and unsubstantiated climate models.
Charles L. Hooper is president of Objective Insights, a company that consults for pharmaceutical and biotech companies, and a visiting fellow at the Hoover Institution.