VI. Method: Conceptual Understanding and Uncertainty
It is this dynamic—the cumulative development of a mainstream view—that gives the findings of climate science a high degree of reliability. Its brackets of uncertainty have narrowed over time.
They have done so in a culture and process that distinguishes established, rock-solid findings about the climate system from newest, best science that is just emerging and is still weighing competing explanations for their consistency with available evidence and predictive ability. Determining bedrock facts is mediated through a set of interactions of scientists, which are constructed so that they give appropriate weight to expert judgment.
Such judgment is built on the expert’s years of related research, evaluation of observational evidence, understanding of theory, deep knowledge of evolving research, interactions with colleagues and much more. That judgment is not just subjective, though it draws on subjective factors. But it draws most persuasively upon scientific argument, evidence, and what has been accepted earlier to support a finding.
New science, on the other hand, may be tentative, inchoate, or just plain incorrect. Though subject to peer review, that process by no means assures that the finding stands the tests of intellectual challenge and further research over time. Moreover, peer reviewers sometimes note that they are reviewing for logic and methodology and are not privy to the original data or do not check calculations, so new science may contain errors from the outset.
While these limitations apply to individual scientific research papers, the entire scientific community also may find that accumulated evidence does not eliminate uncertainty. The state of the knowledge may be insufficient to make a precise statement, in which case scientists express their uncertainty by bracketing their results in a range of possible values, in hopes that they later can say something with greater precision.
A classic example of this process involved estimates of “climate sensitivity.” For over 40 years the degree to which the climate is expected to warm with a doubling of the CO2 concentration in the atmosphere was reported to range between about 1.5 and 4.5 degrees Celsius of heating.
Additional data and analysis accumulated over decades culminated in an important synthesis published in 2020. In it, a large international team of scientists narrowed the range substantially by combining many different lines of evidence.
Drawing on this and other scientific literature, the IPCC Sixth Assessment Report (AR6) now provides a best estimate of the climate sensitivity as 3.0 degrees Celsius, within a reduced range of 2.5 to 4.0 degrees.
Uncertainty remains—there is still a range of warming possible from a doubling of atmospheric CO2, and statements about how the world will respond to such an increase must be understood to have uncertainty. But this refinement eliminated more modest temperature increases from the overall range of possibilities, increasing the likelihood of severe climate impacts from a given increase in atmospheric CO2.
Note it also reduced the high-end limit, which underscores the self-correction that science values as well as its integrity.
Qualitatively, climate scientists have no doubt in what direction the curve of temperature is going, nor that such increases will lead to significant impacts. The good news, if you count it so, is that humanity still has some control of the outcome in principle, depending on whether and when we choose to stop emitting greenhouse gases.
Climate scientists often think about risk in terms of probability and consequences. Climate risks include highly likely impacts with important consequences, such as more extreme and frequent heat waves and their impacts on public health. They also include impacts with a lower likelihood of occurrence but very significant consequences if they were to occur, such as the additional feet of sea-level rise due to Antarctic ice sheet collapse. Risks of both types motivate action to address climate change. We will treat climate risk in greater depth in a separate module.
VII. How Accurate Are Climate Models?
Because there is only one Earth, it is not possible to run a controlled experiment on the entire Earth system of air, sea, land, ice, and life. Climate scientists cannot compare a control planet that might exist without increased greenhouse gases with the planet we actually inhabit. Nor can they systematically vary the amount of greenhouse gases emitted to measure the sensitivity of the Earth’s climate to greenhouse warming.
This is where climate models must stand in. While there is only one Earth, there are infinitely many possible alternative Earth systems and scenarios for greenhouse gas emissions that can be simulated with computers. These models of the Earth system provide crucial simulations that advance scientific understanding of climate change and its consequences. Because of the complexity of the Earth’s climate system and the impossibility of precisely predicting the future, they cannot fully resolve all uncertainties in that understanding. But they have become increasingly accurate over time.
The practice of mathematical modelling of physical systems on digital computers dates from the early days of nuclear physics and is now considered a well-established, essential tool in science. Climate models have simulated the climate system, first crudely and then more finely, for more than a half century. The earliest models were useful mostly to estimate the magnitude of Earth-system warming at the global scale. As digital computing grew in power, climate scientists refined their models to simulate smaller scales and account for more complex phenomena such as the melting of polar ice and local risks of heat waves, drought, heavy downpours, and even forest fires.
Representing the dynamics of the Earth system in a model requires at least three things: 1. physical understanding of the Earth system as represented by mathematical equations of its dynamics, 2. sufficiently accurate measurements of initial conditions, such as concentrations of greenhouse gases in the atmosphere or temperatures of the surface of the oceans, to plug into these equations, and 3. enough computing capacity to simulate, at a desired precision or resolution, scenarios that develop out of those initial conditions.
Climate models consist of mathematical equations that describe the processes of mass and energy transfer throughout the Earth system of atmosphere, land, ocean, and ice. They are based on accepted physical and biological science, but the accuracy of their results is dependent on factors with a range of uncertainties. These include whether the equations fully capture the processes being modelled (including those that operate at smaller spatial scales than the model operates), and how accurately climate observations determine the initial conditions of the parameters in those equations.
One way to test the reliability of a model is to run a “hindcast,” starting with values of the model parameters at a given time in the past and then running the model to compare its results to observations of what took place over the historical period. If the results conform well to what actually happened, the modeler gains confidence that the model is reflecting the system well and so can be relied upon to project a future state of the Earth system. Such tests show that climate models are highly successful in reproducing observed conditions.
Climate models have been criticized for oversimplifying the complexity of Earth-system interactions, but simplification might not be a drawback. Though born of finite computer power and limited empirical data, simple models allow for extraction of essential behavior of a system. But models must meet the standard of prediction, including accurate hindcasts and projections into the future that prove out over time. In early global circulation models, overall global averages were much more accurate, but detailed outputs at the regional level were much less accurate. As historian Spencer Weart has written, this played into the hand of skeptics who cast doubt on the fundamental validity of models.
The most basic and important projection from a global policy point of view is the magnitude of global average temperature increase for a given increase in greenhouse gases in the atmosphere—the logical equivalent of “climate sensitivity.” With an accurate understanding of how much global temperature increase to expect from a given increase in atmospheric greenhouse gas concentration, policymakers can better understand the severity of climate risks at different levels of cumulative greenhouse gas emissions. Getting this right is crucial to setting targets for national and international action to reduce emissions.
Recognizing its importance, a multi-institution research team recently analyzed global mean temperature projections of 17 models developed between 1970 and 2007. They compared model predictions to observed temperatures over the period and found that 14 of the 17 models projected mean temperature to the present quite accurately. Two of the models got the fundamental relationship between emissions and temperature right more than 30 years ago.
With more precise inputs of better measurements of temperature from Earth’s surface and satellites, more refined, higher resolution models in time and space have grown more accurate at simulating climate trends at the regional and local scales as well.
It has been said that models are neither theory nor experiment but a third category of scientific thinking.
They do not measure the observable parameters of the climate and they do not provide an explanation of its mechanisms, but they make use of both. Their purpose is to mimic nature to be able to forecast its behavior. They are, in effect, that second “Earth” we do not have, except inside a computer. Their downside risk is to distance their users from specific, direct causal connections of the kind that, for example, explain unequivocally why a car collision causes damage. They are inherently probabilistic, representing unknown or uncertain contributing factors by probabilities and deriving from them distributions of possible outcomes rather than one determined outcome.
Their potential reward, however, is to fairly simulate a nature that is complex and itself not practically determinate. Averages of their results—like the global mean temperature—provide both projections of future outcomes and insight into the dynamics of the system. So, to the question of accuracy we might add how useful are models for understanding the climate system. Sometimes criticized as a weakness, their dependence on observed measurements to set initial conditions that are practically impossible to determine from first principles is really a strength. It is the mooring by which the model is tethered to the real world.
VIII. Issues of Reliability of Scientific Evidence: Zombie Theories, False Balance, Cherry-Picking
Even as models in the 1970s were emerging as a powerful tool for understanding the climate system, Earth scientists did not agree as to whether the Earth was warming or cooling. Early models led to both possibilities because of uncertainty in the relative strength of warming factors compared to cooling factors caused by human activities.
Scientific work on climate change predominantly documented the warming effects of the greenhouse effect, but some work examined the forces that could lead to cooling.
The popular press covered research on cooling as well, and in retrospect appeared to have overemphasized its likelihood, even after global warming had been accepted by most climate scientists.
Politically polarized discussions used early speculation that the Earth might be cooling to fuel doubt of climate change. Some climate skeptics claimed that climate scientists were continuing to predict global cooling long after the scientific consensus on warming had formed. A documented hoax that invokes the hypothesis of global cooling continues to circulate on the Internet.
It is a not-uncommon occurrence in science for such “zombie theories” to be perpetuated long after they have been disproved. (To be clear, we are not speaking of theories about zombies, which might be fun, but theories about real scientific questions that live on long after they have been answered.) What might once have been a legitimate scientific research question, like the possibility of global cooling, might be resurrected if keeping it alive serves some unscientific purpose.
Examples of zombie theories beyond climate change can be found in virology (vaccines cause autism), physics (nuclear fusion can be achieved at room temperature), and biochemistry (free radicals cause disease), among others. The reasons for their perpetuation vary and may include financial interest, political or ideological stance, or philosophical conviction, among others.
Personal biases in individuals or research groups, or indeed even fraudulent data as in the case of autism, have been known to taint scientific research findings. In the next section, we will look at scientific institutions that exist to deter or ferret these out.
In law, deeply ingrained norms of fairness and strict procedure guide how courts consider an issue brought before them. In court, there is a formal process by which the claimant and the respondent each is allotted an equal opportunity to present its side of the issue and rebut the other side. As a rule, there are only two sides and the decision is for one or the other. Science works differently. Science certainly insists on a fair hearing of new ideas or new evidence. But within the bounds of fairness lies the possibility that a scientific argument is so far from valid that it is not deserving of consideration. As Justice Breyer recounted in the Reference Manual on Scientific Evidence, physicist Wolfgang Pauli once declared acerbically to a colleague who asked if a certain scientific paper was wrong, “the paper wasn’t good enough to be wrong.”
Not every scientific argument deserves to be taken seriously.
One oft-cited example from climate science stands out. In the past, some scientists have acknowledged that the planet is getting warmer but doubted its human cause. Some of these have argued that the heating is coming from increased radiation from the sun. Indeed, the sun’s intensity, or irradiance, does vary over time, both on very long timescales and over shorter periods of about 11-years as sunspots wax and wane. It is also true that the Earth’s temperature varies in step with these solar variations. So, they argue, the global warming observed in recent decades can be attributed to intensification of solar irradiance.
But satellite measurements of total solar irradiance have long-since disproved this claim.
While the sun’s intensity does vary, its variation does not explain the increase in atmospheric heating over the last decades because average irradiance has not increased in that period.
Nevertheless, uninformed discussion of it has often received media coverage in the name of balance.
There is no obligation to give equal time to a debunked claim in climate science—a point that mainstream media has begun to acknowledge in the last decade by labelling it “false balance.” Corollary to this, widespread reporting in the press about an alleged but incorrect scientific fact does not necessarily bestow credibility on it, as we have come to see recently in other arenas as well.
In addition, the phenomenon of “cherry picking” is a well-known pitfall in analyzing scientific data. Unconscious or not, bias in selecting which data to consider will invalidate a scientific analysis as surely as inaccurate measurements. Efforts to refute climate science have frequently involved instances of cherry picking, for many reasons—confounding of science and politics, the high social and political stakes involved, and the inherent uncertainty of many climate parameters that gives room for motivated reasoning (to name but three).
A much-discussed instance is worth noting here. In the early part of the 2010s, climate scientists wondered why there appeared to have been little or no increase in the global average temperature as measured by satellites between about 1998 and 2015. Critics who were skeptical of global warming took this apparent “pause” in global warming as proof that predictions of global warming were wrong. Expert climate scientists noted that the period in question was much shorter than the conventional 30 years for analyzing climate trends and, assuming that the total Earth system was continuing to warm, they argued that the oceans most likely were absorbing excess heat.
A graph of atmospheric temperatures over the 43 years since satellite data became available, from 1978 to 2021 (see Figure 9 below), demonstrates what expert scientists have since concluded—that there was no pause at all. Rather, 1998 was an extraordinarily hot year whose cause (an El Niño) is a recurrent geophysical phenomenon that is largely unrelated to climate change. Thus, the period considered began with a very high deviation above average temperature, whose graphical effect was to artificially level the trend line only during the period of apparent pause. In the years since 2015, heating has risen in a manner consistent with the long-term upward trend. The full graph over those four-plus decades reveals a clear upward trendline, with statistical fluctuations that are consistent with the average upward trend, within the bounds of uncertainty.