Do CMIP5 models skillfully match actual warming?

From Climate Etc.

Nic Lewis

Why matching of CMIP5 model-simulated to observed warming does not indicate model skill

A well-known Dutch journalist, Maarten Keulemans of De Volkskrant, recently tweeted an open letter to the Nobel-prizewinning physicist Professor Clauser in response to his signing of the Clintel World Climate Declaration that “There is no climate emergency”, asking for his response to various questions. One of these was:

The CLINTEL Declaration states that the world has warmed “significantly less than predicted by (the) IPCC”. Yet, a simple check of the models versus observed warming demonstrates that “climate models published since 1973 have generally been quite skillful predicting future warming”, as Zeke Hausfather’s team at Berkeley Earth recently analysed.

The most recent such analysis appears to be that shown for CMIP5 models in a tweet by Zeke Hausfather, reproduced in Figure 1. While the agreement between modeled and observed global mean surface temperature (GMST) warming over 1970–2020 shown in the Figure 1 looks impressive, it is perhaps unsurprising given that modelers knew when developing and tuning their models what the observed warming had been over most of this period.

Figure 1. Zeke Hausfather’s comparison of global surface temperature warming in CMIP5 climate models with observational records. Simulations based on the intermediate mitigation RCP4.5 scenario of global human influence on ERF through emissions of greenhouse gases, etc. were used to extend the CMIP5 Historical simulations beyond 2005.

It is well-known that climate models have a higher climate sensitivity than observations indicate. Figure 2 compares equilibrium climate sensitivity (ECS) diagnosed in CMIP5 models and in the latest generation, CMIP6, models with the corresponding observational estimate on the same basis in Lewis (2022) of 2.16°C and (likely range 1.75–2.7°C). Only one model has an ECS below the estimate in Lewis (2022), and most models have ECS values exceeding the upper bound of its likely range. CMIP6 models are generally even more sensitive than CMIP5 models, with half of them having ECS values above the top of the 2.5–4°C likely range given in the IPCC’s 2021 Sixth Assessment Report: The Physical Science Basis (AR6 WG1).

Figure 2.  Red bars: equilibrium climate sensitivity in CMIP5 and CMIP6 models per Zelinka et al. (2020) Tables S1 & S2 estimated by the standard method (ordinary least squares regression over years 1–150 of abrupt4xCO2 simulations). Blue line and blue shaded band: best estimate and likely (17%-83% probability) range for ECS in Lewis (2022), derived from observational evidence over the ~150 year historical period but adjusted to correspond to that estimated using the aforementioned standard method  for models.

So, how is it possible that Hausfather gets an apparently good match between models and observations in the period 1970-2020? Does it imply that the models correctly represent the effects of changes in “climate forcers”, such as the atmospheric concentration of greenhouse gases and aerosols, on GMST, and accordingly that their climate sensitivities are correct?

The key question is this. Matching by CMIP5 climate models, in aggregate, with observed GMST changes would only be evidence that models correctly represent the effects of changes in “climate forcers”, such as the atmospheric concentration of greenhouse gases and aerosols, on GMST if resulting changes in their combined strength in models matched best estimates of the actual changes in those forcers. The standard measure of strength of changes in climate forcers, in terms of their effect on GMST, is their “effective radiative forcing” (ERF), which measures the effect on global radiative flux at the top of the Earth’s atmosphere once it and the land surface have adjusted to the changes in climate forcers (see IPCC AR6 WG1 Chapter 7, section 7.3)

It is therefore important to compare changes in total ERF as diagnosed in CMIP5 models during their Historical and RCP4.5 scenario simulations over 1970–2020 with the current best estimates of their actual changes, which I will take to be those per IPCC AR6 WG1 Annex III, extended from 2019 to 2020 using the almost identical Climate Indicator Project ERF time series.

Historical and RCP4.5 ERF (referred to as “adjusted forcing”) in CMIP5 models was diagnosed in Forster at al. (2013), for the 20 models with the necessary data. I take the mean ERF for that ensemble of models[1] as representing the ERF in the CMIP5 models used in Figure 1.

Figure 3 compares the foregoing estimates of mean ERF in CMIP5 models with the best estimates given in IPCC AR6. Between the early 1980s and the late 2000s CMIP5 and AR6 ERF estimates agreed quite closely, but they diverged both before and (particularly) after that period. The main reason for their divergence since 2007 appears to be that aerosol ERF, which is negative, is now estimated to have become much smaller over that period than was projected under the RCP4.5 scenario. Updated estimates of aerosol ERF also appears likely to account for about half of their lesser divergence prior to 1983, with the remainder mainly attributable to differences in ERF changes for land use and various other forcing agents.

Figure 3. Effective radiative forcing (ERF) over 1970–2020 as estimated in CMIP5 models (mean across 19 models) and the best estimate given in the IPCC Sixth Assessment Scientific Report (AR6 WG1). The ERF values are relative to their 1860–79 means.

The IPCC AR6 best estimate of the actual  ERF change between 1970 and 2020 is 2.53 Wm−2. The linear trend change over 1970–2020 given by ordinary least squares regression is 2.66 Wm−2, while the change between the means of the first and last decades in the period, scaled to the full 50 year period, is 2.59 Wm−2.

By comparison, the mean ERF change for CMIP5 models between 1970 and 2020 is 1.67 Wm−2. The linear trend change over 1970–2020 is 1.92 Wm−2, and the scaled change between the first to last decades’ means is 1.76 Wm−2.

It is evident that the AR6 estimate of the actual 1970–2020 ERF change is far greater than that in CMIP5 models. Based on the single years 1970 and 2020, the AR6-to-CMIP5 model ERF change ratio is 1.51. Based on linear trends that ratio is 1.39, while based on first and last decades’ means it is 1.46. The last of these measures is arguably the most reliable, since single year ERF estimates may be somewhat unrepresentative, and due to intermittent volcanism the ERF has large deviations from a linear relationship to time. As there is some uncertainty I will take the ratio as being in the range 1.4 to 1.5.

So, CMIP5 models matched the observed 1970–2020 warming trend, but the estimated actual change in ERF was 1.4 to 1.5 times greater than that in CMIP5 models. On the assumption that both the CMIP5 model ERF estimates and the IPCC AR6 best estimates of ERFs are accurate, it follows that:

  • CMIP5 models are on average 1.4 to 1.5 times as sensitive as the real climate system was to greenhouse gas and other forcings over 1970–2020[2]; and
  • CMIP5 models would have over-warmed by 40–50% if their ERF change over that period had  been in line with reality.

It seems clear that the ERF change in CMIP5 models over 1970–2020 was substantially less than the IPCC AR6 best estimate, and that CMIP5 models substantially overestimated the sensitivity of the climate system during that period to changes in ERF. Moreover, the divergence is increasing: the ratio of AR6 to CMIP5 model ERF changes is slightly higher if the comparison is extended to 2022.

In conclusion, Maarten Keulemans’ claim that “a simple check of the models versus observed warming demonstrates that “climate models published since 1973 have generally been quite skillful predicting future warming” is false.

Contrary to the impression given by Zeke Hausfather’s rather misleading graph, CMIP5 models have not been at all skillful in predicting future warming; they have matched the illustrated 1970–2020 observed warming (which was past rather than future warming until the late 2000s, when CMIP5 models were still being tuned) due to their over-sensitivity being cancelled out by their use of ERF that increased much less than the IPCC’s latest best estimates of the actual ERF increase.

Nic Lewis               5 September 2023


[1] ex FGOALS-s2, the Historical and RCP simulations of which were subsequently withdrawn from the CMIP5 archive.

[2] There are some caveats to the conclusion that CMIP5 models were oversensitive by a factor of 1.4 to 1.5 times:

  • the ensemble of CMIP5 models used in Forster et al. (2013) might not have been a representative subset of the entire set of CMIP5 model. However, there appears to be little or no evidence suggesting that is the case;
  • despite their careful compilation, the AR6 best estimates of the evolution of ERF might be inaccurate;
  • the CMIP5 model forcings derived by Forster et al. (2013) might be inaccurate. There are reasons to suspect that their method might produce ERF estimates that are up to about 10% lower than the methods used for IPCC AR6. However, Forster et al. present some evidence in favour of the accuracy of their method. Moreover, the agreement in Figure 2 between the CMIP5 and AR6 ERF time series between 1983 and 2007 (with divergences before and after then largely attributed to differences in particular forcing agents) is further evidence suggesting that the Forster et al. (2013) CMIP5 ERF estimates are fairly accurate; and
  • due to the heat capacity of the ocean mixed layer, GMST is more closely related to average ERF exponentially-decayed over a few years rather than to ERF in the same year. Using exponentially-decayed ERFs would somewhat reduce the 1.4 low end estimate given above for the ratio of AR6 to CMIP5 model ERF 1970–2020 increase estimates, perhaps by ~10%.
5 12 votes
Article Rating
93 Comments
Oldest
Newest Most Voted
Inline Feedbacks
View all comments
Joseph Zorzin
September 6, 2023 6:22 am

This is way over my head! But I do, as always, enjoy the AI generated image at the top.

Otherwise, just watched a new video with Dr. Stephen Koonin, on Dr. Brian Keating’s YouTube channel and in this video Koonin does discuss climate models:

Steven Koonin: Stop POLITICIZING Climate Science!
https://www.youtube.com/watch?v=80vH3sTbHL4&t=1318s

The headline is:

Is climate science being politicized? Are facts being misrepresented and distorted to fit a certain narrative? Are climate scientists trying to dictate policy instead of investigating the actual truth? And what does it mean to be accused of being a global warming denier? Here today to discuss this controversial topic with me is no other than Steven Koonin! Steven is a renowned theoretical physicist and has recently been working on urban studies and government policies. He has also published a very provocative book, Unsettled: What Climate Science Tells Us, What It Doesn’t, and Why It Matters. The book caused a lot of controversy as it challenged the dominant narrative on global warming, and today, he is here to state his case!

Richard Page
Reply to  Joseph Zorzin
September 6, 2023 6:35 am

Basically the climate models are still badly wrong in their basic assumptions but climate modellers are getting better and better at adding ‘fudge factors’ so they look like they’re following historical data faithfully. The implication is that the modellers know what they’re doing is fraudulent and immoral, but the money and acclaim makes it worthwhile.

ClimateBear
Reply to  Richard Page
September 10, 2023 6:14 pm

Basically the ‘climate models’ are not ‘models’ rather they arefudged approximations fitted to historical data and then extrapolated forward in time.

An actial model uses the Navier Stokes Equations (of fluid momentum) and otherknown mathematical realionships addressing the physics of what is going on via a suitably fine mesh model using the method of CFD (computational fluid dynamics)/FEM (finite element modelling). The bsic practical problem of this is that in order to ‘model’ a particular element of physical interaction, say a tripical storm system orsimilar you need a mesh several tiles smaller than the system in focus. The smaller the mesh the more iterations the model has to go through and in practical reality that just blows the end out of the time budget, even with the sort of computing power available. So? Use a fudge factors and a coarser mesh which can pump out results in time for the next publication cycle!

What could possobly be wrong with that? Its all science, after all. Well sufficiently sciency to get past the pals who review this stuff.

MarkW
Reply to  Joseph Zorzin
September 6, 2023 8:39 am

Is climate science being politicized?

I would say no. It’s been politicized since the beginning.

David Pentland
Reply to  MarkW
September 6, 2023 9:16 am
KevinM
Reply to  MarkW
September 6, 2023 7:47 pm

Anyone who’s spent 10 minutes clicking curve fit options in Exchel to make two lines look “same enough” on their chart for a presentation knows what the college students were doing to himdcast climate data.

Tom Abbott
Reply to  KevinM
September 7, 2023 2:34 am

Yeah, and they are hindcastig bogus data. What does that say about the models? They are matching bogus temperature readings to their models. That would make their models bogus, too.

Reply to  Tom Abbott
September 7, 2023 11:08 am

What is interesting is that Nick Stokes doesn’t show up to defend the CMIP5 modeling scenario fantasies.

Editor
September 6, 2023 6:39 am

Thanks for the post, Nic, showing yet another way that the CMIP5 models are trash. For those new to the topic, see the collection of WUWT posts about the trashy CMIP5 models.
https://wattsupwiththat.com/?s=cmip5

If you’re really new to the topic, CMIP stands for Climate Model Intercomparison Project, which were the climate models referred to by the IPCC for their 5th assessment report.

Regards,
Bob

Scissor
September 6, 2023 6:40 am

Alas, it’s just weather, but 10C in my area of Colorado this morning seems a bit cold for early September. Thankfully it wasn’t 9.95C.

Streetcred
Reply to  Scissor
September 6, 2023 7:27 am

I’m in San Francisco at them moment, temperature varies 5C between Golden Gate Bridge and Fisherman’s Wharf, a mere few kilometres apart.

Steve Keohane
Reply to  Scissor
September 6, 2023 8:12 am

Hit 38°F this am on the western slope at 6600′.

Scissor
Reply to  Steve Keohane
September 6, 2023 9:04 am

Was that 38.00? One of the alarmists that comments here says that 0.05C is significant.

lanceman
September 6, 2023 7:01 am

The models are constantly being modified and tuned. But when the modified models are run going forward for comparison purposes, the recent climate record is now known. In fact the models would not be released unless they were tuned to match the recent climate history. In effect, they have seen the answer key to the exam.

Shouldn’t the modified models have to start the comparison period anew beginning with the present?

KevinM
Reply to  lanceman
September 6, 2023 7:53 pm

Yes let’s see what that 1973 model says about 2023. To those who would say “not fair”: why should we think 2023 models can predict 2073 much better?

EastDevonOldie
Reply to  KevinM
September 8, 2023 6:51 am

Even the IPCC acknowledge the fat that:

“The climate system is a coupled non-linear chaotic system, and therefore the long-term prediction of future climate states is not possible.”
 
IPCC Third Assessment Report 2001

Steve Case
September 6, 2023 7:07 am

If you Google “how much warming do the models project by 2100” it comes up with “a likely increase of at least 2.7°F…”

First it should be noted that Google chooses to use Fahrenheit degrees since that produces a bigger number than 1.5 Celsius degrees. And really, would a world that’s 2.7 degrees warmer than today be a crisis? A crisis that requires us to destroy the economy?

Besides that, you can Google “The Little Ice Age” to find out that it was a

     climate interval that occurred from the early
     14th century through the mid-19th century

So there’s been a warming trend since 1850. You don’t need a model to figure out that it’s likely to continue for a while.

Then there’s this from the Nic Lewis article:

     it is perhaps unsurprising given that modelers knew when developing and tuning
     their models what the observed warming had been over most of this period.”

Uh, didn’t they know over ALL of this period? Well anyway how well did the AR one two three four and five do? Here’s the 2100 prediction chart from the IPCC AR1 :

IPCC AR1 Temp Chart.png
Steve Case
Reply to  Steve Case
September 6, 2023 7:19 am

You can find that chart on Page 73 of the FAR Executive Summary (pdf-12)

Scissor
Reply to  Steve Case
September 6, 2023 9:09 am

That’s man made warming for sure. Such a beautiful change in function around 1975.

Joseph Zorzin
Reply to  Steve Case
September 6, 2023 10:57 am

If you Google “how much warming do the models project by 2100” it comes up with “a likely increase of at least 2.7°F…”

“It’s tough to make predictions, especially about the future”.
Yogi Berra

PCman999
Reply to  Steve Case
September 6, 2023 11:59 am

The warming trend started in the 1600 or 1700’s – once the Thames stopped freezing over solid every year.

Steve Case
Reply to  PCman999
September 6, 2023 12:28 pm

Ok 400 years ago, and common sense says it will continue for a while, and that same common sense says it won’t go up forever.

Our good friends never say when it will stop going up, indeed some of them think the Earth will go the way Venus. At least the late Stephen Hawking thought so LINK

simonsays
Reply to  Steve Case
September 6, 2023 2:21 pm

Even Mann called out that Hawking for that stupid theory. At least he did politely,

“Hawking is taking some rhetorical license here,” Michael Mann,

KevinM
Reply to  Steve Case
September 6, 2023 8:01 pm

predictions that can’t be verified during the lifetime of the predictor are so easy to commit to.

Richard Page
Reply to  PCman999
September 6, 2023 5:02 pm

1814 was the last time. The winter of 1813-14 was the very last London Frost Fair when the Thames was frozen over from December to February.

Streetcred
September 6, 2023 7:24 am

Let’s be clear, Hausfather has been no more skilful in hindcasting past measured temperature than had he simply copied and pasted the graph.

PCman999
Reply to  Streetcred
September 6, 2023 12:03 pm

That’s essentially what tuning the models do – they aren’t improving the physics they are kust correcting the previous wild-ass-guesses.

That’s why when the model is run, even with 20 or more years of data built inside, the forecast still runs too hot to the same degree.

KevinM
Reply to  PCman999
September 6, 2023 8:08 pm

The math has gotten way ahead of the physics- a consequence of objective scoring systems and slowly passing time

gc
September 6, 2023 7:26 am

I think I understand the tuning point, but isn’t the worst thing about Mr. Hausfather’s claim the fact that the graph he uses shows the models conforming with known temperature, not with a “prediction”? How well have the models done at predicting temperature? I appreciate that the models could still be useless even if they have made accurate predictions, but I’m still curious as to the accuracy of the predictions they have made. It doesn’t look like Mr. Hausfather’s graph addresses that issue. Or am I missing something?

Richard Page
Reply to  gc
September 6, 2023 7:55 am

Using a much higher climate sensitivity than modern estimates means that no matter how much ‘tuning’ or ‘fudge factors’ you use to recreate historical data, all predictions will run hot and sail off into the wide blue yonder of fantasy land. Not accurate.

KevinM
Reply to  gc
September 6, 2023 8:10 pm

You missed nothing. Mr H is counting on you accepting that historical performance guarantees future returns.

Rud Istvan
September 6, 2023 7:59 am

All CMIP5 produce a tropical troposphere hotspot that does not exist in reality.
They produce ECS significantly higher than observational EBM.
Hausfalter’s claim that they did well predicting GMST is almost beside the point. At least now we know why—lower projected ERF than AR6 says was the actual case. “Right for the wrong reasons” is not ‘science’.

Reply to  Rud Istvan
September 6, 2023 8:34 am

Yeah, I keep looking for that increasingly mythical “Hot Sport” prediction too that was supposed to exist according to the modeling factory how long has the waiting been…. 25-30 years now?

Mark Luhman
Reply to  Sunsettommy
September 6, 2023 5:29 pm

To me at that point the CO2 climate change theory guess failed by
the data. As Feynman put it (I am paraphrasing)if the data does not support your
guess, your guess it wrong! It does not matter how smart you are or how elegant
your guess was, it is wrong, you need another guess. Somehow in their education they
missed that basic fact!  It should have been, game, set and match at that point. Yet over 35 years later the failed guess continues.  It looks to me lying pays rather well.  Sad, very Sad.

ToldYouSo
September 6, 2023 8:29 am

From the title block to Figure 1 in the above article:
“Zeke Hausfather’s comparison of global surface temperature warming in CMIP5 climate models with observational records. Simulations based on the intermediate mitigation RCP4.5 scenario of global human influence on ERF through emissions of greenhouse gases, etc. were used to extend the CMIP5 Historical simulations beyond 2005.”

IMHO, there is a rather sophomoric sleight-of-hand trick being played by Hausfather in his graph.

It shows the generally good agreement of CMIP5 “multimodel” average hindcasts from 1970 through 2005. But, as duly noted in the article, these models were tuned (aka forced to match the observations (data) over this time period. Hence we see the excellent fits for the relatively rapid temperature dips starting around 1982 and 1991.

However, since the models used the RCP4.5 scenario (not data!) to forecast the temperature trending from 2006 onward, we do not see any further matching of the mutimodel predictions to the dips observed in 2007, 2010 and 2017. In fact, the multimodel predictions from 2006 onward appear to be nothing more than an extension of the slope (albeit with some very slight added fluctuations) of the least-square curve-fit of data of 1970-2005.

So, is it really global human influence on ERF through emissions of greenhouse gases, etc.” using twenty or more $multimillion, supercomputer climate models as Hausfather claims, or is it instead nothing more than extending a simple curve fit of existing data?

You decide.

“The simplest explanation is preferable to one that is more complex.”
— Occam’s razor, itself simplified

Tim Gorman
Reply to  ToldYouSo
September 6, 2023 11:06 am

As Pat Frank has shown, the climate models can be emulated by a simple linear equation. Neither the models or the linear equation tracks with observed data. The uncertainty associated with the linear equation is at least as small as what should be given for the climate models – and it gets so large so quickly that there is no way to tell what is going to happen in the future. The future uncertainty interval defines an area of “UNKNOWN” – i.e. a cloudy crystal ball. Like a fortune teller you can pick anything you want out of that cloudy ball – and the CAGW advocates pick a future that maximizes their power and fortune.

DavsS
Reply to  ToldYouSo
September 7, 2023 1:27 am

That’s my take. If lots of people are fitting models to a rising trend it is hardly a surprise if the mean of those models looks about right. It says nothing whatsoever about skill of individual models (which is generally poor).

michael hart
September 6, 2023 8:42 am

You get zero points out of ten for predictive-modelling data that has already happened.

The climate modelling ‘community’ long ago abrogated any bonds of scientific trust they may have enjoyed.

Hausfather should ask himself whether he really wants to lumped in with them.

bnice2000
Reply to  michael hart
September 6, 2023 2:23 pm

Zeke most definitely wants to be one of the AGW gurus.

He will lie, con, distort… do basically anything to try to support the AGW meme.

His whole life and income depends on it.

dougsorensen
September 6, 2023 8:48 am

I’m a little confused about what Hausfather used for the model inputs. Model builders are always tweaking their models to better fit observed data. They’d be fools not to. The question is how much of model output in Fig. 1 is models that have had post hoc adjustment. A better comparison would have been older model outputs to current data.

Rud Istvan
Reply to  dougsorensen
September 6, 2023 11:40 am

For CMIP5, yearend 2005. Everything before is tuned to best hindcast. Everything after is a tuned model projection.

bnice2000
Reply to  Rud Istvan
September 6, 2023 1:33 pm

If they are hindcasting to ADJUSTED URBAN surface data then they already have a massive spurious, not-real, warming trend built in.

Anything they churn out after that is going to be just garbage.

KevinM
Reply to  bnice2000
September 6, 2023 8:22 pm

First week of class is probably data recently from an elderly professor’s paper photocopy notebook onto a computer.

KevinM
Reply to  KevinM
September 6, 2023 8:23 pm

…. To run a python script written by an uncredited foreign grad student.

KevinM
Reply to  dougsorensen
September 6, 2023 8:20 pm

“A better comparison would have been older model outputs to current data.”

Winner.Comment sums up the whole debate in a sentence.

Kevin Kilty
September 6, 2023 9:09 am

Couple of thoughts:

  1. There is nothing in these model vs reality comparisons that looks anything like a fair test. Compare a fair test of a medical treatment regimen or medicine, where there is blinding possibly on two sides, treatment/control group membership and also evaluation, against these comparisons where the modellers already have a clear target. I can’t get too excited about the so-called agreement.
  2. There are way too many means of adjusting observations and simulations to be at all surprised by these “matches”. For example, as the author says, “due to the heat capacity of the ocean mixed layer, GMST is more closely related to average ERF exponentially-decayed over a few years rather than to ERF in the same year….”

Looking back at Figure 1 doesn’t seem to indicate any delay at all in the biggest change in forcings (Pinatubo) by virtue of heat capacities. Nonetheless, one can bring out the exponentially-delayed horse argument and trot him around the track. Don’t the feedback models have already built into them these appropriate time constants? The cartoons of feedback don’t snow this explicitly, but the blocks actually operate in the “s” domain. The models, done in the time domain ought to show already this without saying.

ToldYouSo
September 6, 2023 9:51 am

In more honest times, the “Estimated Effective Radiative Forcing (ERF)” values graphed in Figure 3 would be simply called fudge factors used to tune the noted CMIP models to force them to match existing data.

If one were to politely ask what are the fundamental causes—due to mankind and/or due to nature—for the relatively rapid and high magnitude changes in ERF seen over the intervals of 1982–1986 and 1991–1995, no science-based answer would be forthcoming.

Are those two critical intervals of ERF change due to:
— Changes in solar irradiance at Earth TOA? No.
— Changes in Earth’s areal cloud coverage (albedo)? No.
— Changes resulting from periods of El Nino or La Nina? No. Very strong 1997–1998 and 2015-2016 El Ninos are not evident, and very strong La Ninas of 1998–2000 and 2007–2011 are not evident.
— Changes resulting from unusual global volcanic activity? Likely no. Major eruptions of El Chichón (1982, VEI5) and Mount Pinatubo (1991, VEI6) might be indicated, but major eruptions of Mount St. Helens (1980, VEI5) and Puyehue and Cordón Caulle (2011-2012, VEI5) are not evident.
— Changes resulting from comets or asteroid impacts? No.
— Changes in human emissions of CO2? Certainly no.

But those asserted ERF changes are certainly needed mathematically to “tune” climate models.

Go figure.

JCM
September 6, 2023 10:08 am

CMIP members exhibit wildly different proportions in feedback kernels. They are only constrained in historical temperature estimates. Otherwise, they are totally free.

Additionally, in spectral space, they have wildly different proportions of feedbacks in the FIR (water vapor band) v “window” spectra. Models are totally unconstrained in the “unknown” parameters of interest.

Evidently, there is still far too much freedom to glean information from GCMs. The lack of consistency suggests missing pieces of the puzzle. As a consequence, the puzzle is being arranged in wildly different ways. This should be the primary conclusion of the model intercomparison project to date. Why else run a MIP?

Simple models arrive at the same global temp estimates as complex models. MIP GCMs are supposed to test our knowledge of the mechanisms at work. The result to date is that knowledge is low.

Graemethecat
Reply to  JCM
September 7, 2023 6:00 am

The models are there merely to fool hoi polloi and give a “sciencey” veneer to a vicious Marxist powergrab.

More Soylent Green!
September 6, 2023 10:40 am

Nobody has ever shown the physical climate works the way the climate models say they do.

Gilbert K. Arnold
September 6, 2023 10:54 am

In other words the climate modelers are applying the: “UEFF” (Universal
Engineering Fudge Factor), aka Finagle’s Constant

ToldYouSo
Reply to  Gilbert K. Arnold
September 6, 2023 11:04 am

. . . except that it is NOT a constant, but varies as needed to match past observations, and as needed to support the “climate change” meme du jour.

Joseph Zorzin
September 6, 2023 11:01 am

Do any climate models honestly include unknown variables with unknown properties? If they dared to do that they’d be admitting predicting future climate is impossible.

Richard Page
Reply to  Joseph Zorzin
September 6, 2023 2:11 pm

Our understanding of the climate and the processes that are associated with it is incomplete – to a greater or lesser extent (opinions vary). Therefore, by definition, all climate models include ‘best guesses’ which are unknown variables and exclude some things, like clouds, that climate scientists don’t know how to model. All of these climate models are a complete fudge – they can’t represent reality because we don’t know all the factors and processes; they tell us nothing and are a waste of time and money.

Mark Luhman
Reply to  Richard Page
September 6, 2023 5:37 pm

I don’t think you will ever be able to model the climate with computer models, since the variables are nearly infinite, yet the run time of the model must be finite. Let alone the fact we don’t know most of the variables.

Nicholas McGinley
September 6, 2023 11:11 am

This is all a complete waste of time, since the so-called “observed” temperatures are fake data.
They are comparing their awful models with their fake temps.

bnice2000
Reply to  Nicholas McGinley
September 6, 2023 2:27 pm

yep ! Junk from start to finish !

thumbs.png
Gilbert K. Arnold
September 6, 2023 11:32 am

;; Perhap’s it should more properly named “Finagle’s Correction”?

Tom.1
September 6, 2023 12:47 pm

Is there another branch of physics where the average of a bunch of very inaccurate theories is held up as being a valid representation of the physical world?

Richard Page
Reply to  Tom.1
September 6, 2023 2:12 pm

String theory? (ducks).

buckeyebob
September 6, 2023 1:03 pm

CHIMP5 could do a better job, I’m sure. All the have to do is toss darts.

Chas
September 6, 2023 1:03 pm

Looking at anomalies it is easy to forget to the spread in in the average temperatures:
comment image

from Bob Tidale’s site
https://bobtisdale.wordpress.com/2018/11/27/what-was-earths-preindustrial-global-mean-surface-temperature-in-absolute-terms-not-anomalies/

ToldYouSo
Reply to  Chas
September 6, 2023 4:55 pm

Independent of their differences in absolute temperature, all of the plotted CMIP3 computer models have noticeable upward slope breaks that occur at around 1961.

Hmmmm . . . in 1961 the Soviet space probe “Venera 1” became the first man-made vehicle to reach Venus.

As is well known, Venus has the hottest sensible atmosphere of all planets in the Solar System. I’m not necessarily asserting Venusians are a vengeful species, but it does invite the question . . . 

Mark BLR
Reply to  Chas
September 7, 2023 3:01 am

The IPCC (WG-I) report actually included an updated version of this graph using CMIP6 “Historical Data” (to 2014) and the SSP1-2.6 extension to 2100.

See the top panel of Figure 1.11, from page 190, which should appear “automagically” below …

comment image

PS : Hat-tip to Javier Vinos for digging out the “/ar6/wg1/figures/” URL at the IPCC website allowing the above “auto-magic” to occur (URL string must end with something like “.png” or “.jpg”, no “?<options>” suffixes allowed …).

ToldYouSo
Reply to  Mark BLR
September 7, 2023 7:59 am

Again, these updated charts invite the question: What the heck happened to/on Earth circa 1960 to cause the noticeable upward increase in the rate of global warming?

At least one inquiring mind would like to know, finding no ready explanation anywhere on the Web.

If I had to hazard a guess, I would say that it was the beginning of the UHI-poisoning effect on global surface temperature measurements . . . but that’s just me.

JBP
September 6, 2023 1:17 pm

Question for ALL the knowledgeable folks here (I’ll leave the determination of who that might be to ‘consensus’, haha):

Why hasn’t the modelling of the Earth’s climate focused on total energy? In my inexpert opinion any representation using a temperature is inappropriate for characterizing the behavior of our climate. The ongoing argument has focused on the future value of temperature (and most times atmospheric temperature), but isn’t that a distraction from the true question that would lead to a determination (model) of climate:

‘Is the energy level of the earth increasing, decreasing or staying the same?’

In order to answer that question such variables as temperature must be measured, but just as, if not more importantly, thermal (energy) batteries such as the oceans, atmosphere and landmasses must be characterized properly. Energy transfer obviously must be characterized also.

Don’t the oceans represent the most massive thermal batteries (rechargeable, or capacitors) by far? For instance, how does a seemingly small shift in depth and/or value of the thermal gradient of the ocean compare to an equivalent shift in the atmosphere? It would seem that, given the very large energy capacity of water versus ‘air’, that determining the total level of ocean energy might be one of the the most important pieces of the puzzle.

For those purveyors of the climate doom scenario, I can understand the use of temperature as it provides an easily manipulated measure, that also can be used at the emotional level to scare laymen into belief of ‘settled science’.

But for those on the side questioning and refuting doomsayers predictions, I see you falling into the trap of letting your opponent define the terms and conditions of the argument.

Graemethecat
Reply to  JBP
September 7, 2023 6:02 am

Climate “Scientists” have never understood the fundamental distinction between temperature and enthalpy.

ToldYouSo
Reply to  JBP
September 9, 2023 9:00 am

“Why hasn’t the modelling of the Earth’s climate focused on total energy?”

Well, there are two fundamental answers to that question:

1) Indeed climate scientists have attempted to model Earth’s total “energy balance” in more practically-expressed units of power flux (W/m^2) balances between incoming power and outgoing power, and the individual components comprising such. Overall, when averaged over hundreds of years, Earth can be consider as being in a near-equilibrium condition, so it is much easier to discuss the perturbations in energy than it is to discuss the total quantity of energy involved (e.g., the huge numbers associated with the thermal energy contained in the world’s oceans). However, at a finer level, Earth is in a state of continuous fluctuation about the long-term average due to things like day vs. night, seasonal changes, solar irradiation variation with sunspot cycle, PDO and AMO cycle variations, etc. Considering both viewpoints, it is more practical (useful) to examine the balance/imbalance of power crossing some well-defined interfaces, using the metric expressed as energy/unit time/per cross-section area normal to the energy flow direction, or W/m^2. This is the basic parameter evaluated in the original Kiehl and Trenberth diagram (1997, see attachment) describing the beakdown of components comprising Earth’s average “energy balance” (yes, the incorrectly named parameter of interest) as well as the numerous subsequent updates and modifications of such.

2) As you alluded to, it is inappropriate to use temperature as a metric to discuss energy or its rate of flow over time (power). For example, liquid water can undergo phase change to vapor at a constant temperature even though a significant exchange of energy is required per pound mass of water; similarly water ice can melt to liquid water at a constant temperature if it absorbs a substantial amount of thermal energy per pound water. Within Earth’s climate system and its hydrological cycle, water exists in solid, liquid and gas phases and frequently converts between these phases with involvement of large amounts of power.
Also, a change in temperature of a given mass of solid, liquid or gas can only be interpreted as a change in the energy of that quantity of matter with knowledge of the specific heat of the substance and with understanding of the nature of the change process (e.g., for gases, is it a constant pressure or constant volume process?). The there is the whole issue of establishing (and quantifying) the degree of temperature uniformity within any mass being considered. Things get complicated very quickly when trying to use material temperature changes over time instead of just using power as the metric of interest.

K-T_diagram_1997.jpg
Tim Gorman
Reply to  ToldYouSo
September 9, 2023 10:03 am

Things get complicated very quickly when trying to use material temperature changes over time instead of just using power as the metric of interest.”

The problem is that the “metric” always falls back to being temperature, i.e. global average temperature.

The other big problem is attributing energy flows in and out properly to specific causes, e.g. CO2 growth.

Giving_Cat
September 6, 2023 1:25 pm

How can any “model” that supposedly predicted the eruption of Mt Pinatubo be taken seriously?

bnice2000
September 6, 2023 1:27 pm

And of course, the surface temperature fabrications used are URBAN/AIRPORT surface temperatures ADJUSTED in an attempt to meet the warming requirements of the models.

They are not representative of “global” temperatures.

Roy Clark
September 6, 2023 2:08 pm

The concepts of radiative forcing, feedbacks and climate sensitivity used in the climate models are pseudoscientific nonsense. An IR ‘greenhouse gas forcing’ does not change the energy balance of the earth, nor does it produce a measurable change in surface temperature. The climate models are empirically ‘tuned’ to match the mathematical construct of a global mean temperature record. One set of meaningless numbers has been ‘tuned’ to match another. Physical reality has been abandoned in favor of mathematical simplicity. 
 
Climate model ‘tuning’ can be traced back at least to the 1981 paper by Hansen et al, specifically Figure 5 of that paper. Here, a combination of an increase in CO2 concentration, changes in solar flux and ‘volcanic aerosols’ were used to tune a 1-D RC model to match a global mean temperature record. The model was also ‘tuned’ using ‘water vapor feedback’ to get a climate sensitivity of 2.8 °C. This provided the pseudoscientific foundation for all of the later climate models. There are at least nine fundamental scientific errors in this paper – that have been copied by about 50 modeling groups. The fraudulent concept of radiative forcing has been used by the IPCC since it was founded in 1988. This history was reviewed by Ramaswamy et al [2019].
 
The errors in the Hansen 81 paper are:
 
1) An IR radiative forcing (increase in atmospheric greenhouse gas concentration) does not change the energy balance of the earth. Any slight increase in heat content in the troposphere is just reradiated back to space by wideband LWIR emission.
2) There is no greenhouse effect temperature.
3) The LWIR flux is coupled to the turbulent convection in the tropospheric heat engine.
4) The mathematical artifacts created by the Manabe and Wetherald 1967 modeling assumptions are accepted without question as real surface temperature changes (see MW67 P. 242).
5) The surface energy transfer processes, in particular the coupling of the LWIR flux to the wind driven latent heat flux are ignored in their ‘slab’ ocean model.
6) The discussion of radiative perturbations to the 1-D RC model has nothing to do with the earth’s climate.
7) The role of the ocean oscillations, particularly the Atlantic Multi-decadal Oscillation (AMO) in setting the global mean temperature is ignored.
8) Any increase in surface temperature from a ‘CO2 doubling’ is too small to measure.
9) A contrived set of ‘radiative forcings’ is used to ‘tune’ the 1-D RC model so that the output artifacts appear to match the global mean temperature series.
 
The basic climate issue that needs to be addressed is as follows:
 
Since the start of the Industrial Revolution about 200 years ago, the atmospheric concentration of CO2 has increased by approximately 140 parts per million (ppm), from 280 to 420 ppm. Radiative transfer calculations show that this has produced a decrease near 2 W m-2 in the longwave IR (LWIR) flux emitted to space at the top of the atmosphere (TOA) within the spectral range of the CO2 emission bands. There has also been a similar increase in the downward LWIR flux from the lower troposphere to the surface. For a ‘CO2 doubling’ from 280 to 560 ppm, the decrease in outgoing longwave radiation (OLR) is estimated to be 3.7 W m-2. At present, the average annual increase in CO2 concentration is near 2.4 ppm. This produces an increase in the downward LWIR flux to the surface of approximately 0.034 W m-2 per year. How do these changes in LWIR flux alter the surface temperature of the earth?
 
The short answer is that thermal engineering calculations of the change in surface temperature using the time dependent flux terms coupled to the surface thermal reservoir show that any CO2 induced change in surface temperature is ‘too small to measure’. The whole concept of radiative forcings, feedbacks and climate sensitivity as discussed in Chapter 7 of the AR6 Working Group 1 Report is pseudoscientific nonsense. 
 
There are five parts to the engineering analysis.
 
1) The radiative transfer calculation of the change in LWIR flux at the top of the atmosphere (TOA) is incomplete. It has to be extended to include the change in the rate of cooling of the troposphere. When this is done, the maximum change for a ‘CO2 doubling’ is a decrease in the rate of cooling, or a slight warming of +0.08 K per day. At a lapse rate of -6.5 K km-1 an increase in temperature of +0.08 K is produced by a decrease in altitude of about 12 meters. This is equivalent to riding an elevator down four floors.
 
2) The upward and downward LWIR flux terms are decoupled by molecular line broadening. Almost all of the downward LWIR flux to the surface originates from within the first 2 km layer of the troposphere. Approximately half of this flux originates from the first 100 meter layer above the surface. This means that the small amount of tropospheric heating produced by a ‘greenhouse gas forcing’ is simply re-radiated to space as wideband LWIR emission (there may also be a change in altitude and therefore gravitational potential). THERE IS NO CHANGE TO THE ENERGY BALANCE OF THE EARTH. (The changes in cooling rates in the stratosphere require very small changes in flux because of the low air density).
 
3) At the surface, the penetration depth of the LWIR flux into the oceans is less than 100 micron (0.004 inches). Here it is fully coupled to the much larger and more variable wind driven evaporation (latent heat flux). Using long term zonal averages, the sensitivity of the latent heat flux to the wind speed within the ±30° latitude bands is at least 15 W m-2/m s-1. The 2 W m-2 increase in downward LWIR flux to the surface from 140 ppm CO2 is dissipated by an increase in wind speed of 13 centimeters per second. The annual increase of 0.034 W m-2 from 2.4 ppm CO2 is dissipated by an increase in wind speed of 2 mm s-1. Any CO2 induced ocean temperature changes are too small to measure.
 
4) Over land, all of the flux terms are absorbed by a thin surface layer. The surface temperature initially increases after sunrise as the solar flux is absorbed. This establishes a thermal gradient with both the cooler air above and the subsurface ground layers below. The surface-air gradient drives the evapotranspiration and the subsurface gradient conducts heat below the surface during the first part of the day after sunrise. Later in the day, as the surface cools, the subsurface gradient reverses and the stored heat is returned to the surface. As the land and air temperatures equalize in the evening, the convection stops and the surface cools more slowly by net LWIR emission. This convection transition temperature is reset each day by the local weather system passing through. Almost all of the absorbed solar heat is dissipated within the same diurnal cycle. The day to day changes in convection transition temperature are much larger than any temperature change produced by CO2.
 
5) When the global climate anomaly record, such as the HadCRUT4 data set is evaluated, the dominant term is found to be the Atlantic Multi-decadal Oscillation (AMO). The additional part of the recent warming may be explained as a combination of three factors. First there are urban heat islands related to population growth that were not part of the earlier record. Second, the mix of urban and rural weather stations use to create the global record has changed. Third, there are so called ‘homogenization’ adjustments that have been made to the raw temperature data. These include the ‘infilling’ of missing data and adjustments to correct for ‘bias’ related to changes in weather station location and instrumentation. It has been estimated that half of the warming in the ‘global record’ has been created by such adjustments.
 
The climate models are based on an invalid correlation between a contrived set of radiative forcings and an equally contrived ‘global average temperature’. Such models have no predictive capabilities over climate time scales because of Lorenz instabilities. The solutions to the large number of coupled non-linear equations is unstable and the errors increase over time. The models are simply ‘tuned’ to match the global average temperature record. Simple inspection of such records reveals the 1940 AMO peak. The IPCC climate fraud then continues by separating the contrived radiative forcings into ‘human’ and ‘natural’ factors. The models are then rerun with just the natural factors and this is used to ‘attribute’ climate change to ‘human’ or ‘anthropogenic’ causes. This is illustrated in Figure 1 using illustrations and data from IPCC AR6 WG1.
 
A more detailed discussion of climate energy transfer is provided in the recent book ‘Finding Simplicity in a Complex World – The Role of the Diurnal Temperature Cycle in Climate energy Transfer and Climate Change’ by Roy Clark and Arthur Rörsch. A summary and selected abstracts including references relevant to this discussion are available at http://www.clarkrorschpublication.com.
More information on climate pseudoscience is available at http://www.venturaphotonics.com.
.pdf summaries can be downloaded using the links:
https://venturaphotonics.com/files/VPCP_025.1_GreenhouseGasForcings.pdf
https://venturaphotonics.com/files/VPCP_026.1_TheCorruptionofClimateScience.pdf

Figure 1: Understanding the IPCC climate fraud: a) Changes in radiative forcings since 1750, b) simulated temperature increases from 1750 to 2019, based on a), c) time dependence of the radiative forcings and d) time dependence of the temperature changes derived from c), e) ‘tuned’ temperature record using a contrived set of radiative forcings that appear to simulate the global mean temperature record, f) the separation of the contrived forcings to create fraudulent ‘human’ and ‘natural’ temperature records, g) the contributions of the AMO, UHI etc. to the global mean climate record, h) the [pseudoscientific] equilibrium climate sensitivity (ECS) estimated from the CMIP6 models (IPCC AR6, WG1, figures 7.6, 7.7, 2.10, 7.8, 3.4b and FAQ 3.1 Fig. 1, ECS data from Table 7.SM.5).
 

Figure1.jpg
macha
September 6, 2023 3:15 pm

Patrick Frank demonstrates curve fitting models.
https://youtu.be/0-Ke9F0m_gw?si=ORNwZntfrKyg781V

B Zipperer
September 6, 2023 4:11 pm

More on “implausibly hot” models
https://www.sciencemag.org/news/2021/07/un-climate-panel-confronts-implausibly-hot-forecasts-future-warming

Here are two of Gavin Schmidt’s takes on hot CIMP6:
https://www.realclimate.org/index.php/archives/2021/12/making-predictions-with-the-cmip6-ensemble/
and a graph from his Twitter thread [IIRC]: note the subtle way it’s done. The color band is actually two colors. He arbitrarily lops-off all the “too hot” GCMs to take the average of the rest to make it look closer to the surface temps. [Seems misleading to me.]

So, according to the alarmists, “the GCMs are running too hot, but we know them to be valid anyway. Trust us! “

Schmidt_CMIP6_graph.png
bnice2000
Reply to  B Zipperer
September 6, 2023 4:19 pm

And remember, that black line is a fabrication of agenda adjusted urban and airport temperatures.

There is absolutely no possibility of it being remotely representative of global temperatures.

John in Oz
September 6, 2023 4:54 pm

Could someone smarter than I (most of you) tell me if these models use fudge factors throughout the entire period they are emulating in order to match reality?

If they do, then the model parameters at the start would be different to those at the end

Also, if they do, it would be interesting to see the end result using the last set of parameters but starting at the first year with no further ‘adjustments’.

Richard Page
Reply to  John in Oz
September 6, 2023 5:14 pm

Yes, yes and probably not. The fudge factors are necessary to make the model fit the historical data but, because they change to fit the data, I don’t think they are then used for the projections. The ecs used is about twice the amount of many modern estimates so the models immediately start to run hot, diverging from reality fairly rapidly. If you tried the experiment you suggest you’d just get a ski slope extending upwards from the start date – not very interesting or illuminating really.

gc
Reply to  Richard Page
September 7, 2023 10:54 am

Thanks John and Richard. This is interesting to me. Why bother with hindcasting if the projections are based on different physics?

Richard Page
Reply to  gc
September 7, 2023 4:59 pm

Hindcasting is used to sell the model as an accurate forecasting tool. The use of ‘fudge factors’ and specific tuning isn’t explicitly stated; the modellers are trying to fool the onlookers into believing that the model is accurate whilst trying to hide just by how much they’ve had to manipulate the data to make it fit.

ferdberple
September 6, 2023 5:21 pm

Any model that fails to deliver the expected result will be modified. Any model that delivers the expected result will remain unchanged.

g3ellis
September 6, 2023 6:03 pm

In Decision Support Systems, you generally accept that you use a model once. The tendency is to hang on to it because it worked, then start adding fine tuning as you go forward. In that, all you have done is curve fitting and wasted everyone’s time and money. When the model is wrong, you don’t fine tune it. You trash it and try to build another one with what you have learned. But if you inject your religion into it, you are just making things up.

Richard Page
Reply to  g3ellis
September 7, 2023 5:02 pm

The models have become the cornerstone of the new religion – to question their accuracy is to question holy writ, the very words of the Climate God.

Bob
September 6, 2023 6:24 pm

I’m sure Nic’s explanation is meaningful but I don’t understand it. Hoping someone can help me.

My understanding is that RCP (representative concentration pathway) is an indication of how much earth’s average global temperature will rise over the average global temperature of the mid to late 1800s. If I remember right there were four pathways 2.6, 4.5, 6.0 and 8.5. My understanding was that the 8.5 scenario was business as usual with CO2 emissions steadily rising, 6.0 was some effort to reduce CO2 emissions but CO2 still steadily rising, 4.5 was substantial efforts to reduce CO2 emissions where we would see a leveling off and some reduction in the future and 2.6 would be an all out serious reduction of CO2 emissions, an effort that would keep average global temperature increases below the 1.5-2.0C level.

If all of this is true then using the 4.5 RCP and claiming it matches observations is not honest. Although Europe and the US have made efforts to lower CO2 emissions the rest of the world has not, China and India are instead massively increasing emissions. Consequently earth’s total CO2 has been increasing rather than plateauing or decreasing. Therefore using the 4.5 RCP is inappropriate.

old cocky
Reply to  Bob
September 6, 2023 6:58 pm

It’s a bit complicated, but the RCPs are indeed Representative Concentration Pathways.
Each concentration pathway more or less ties in to different emissions scenarios.

As I understand it, the emissions and concentrations were assumed to map pretty much as you summarised.

Those assumptions seriously underestimated natural CO2 sinks (or overestimated CO2 residence), with the end result that subsequent emission rates give RCP 4.5 concentrations rather than RCP 8.5.

bdgwx has posted charts showing Hansen and IPCC projections vs observations, which are quite useful. Perhaps he can do the same in response to your comment.

Most current “Western” policies seem to be based on the emissions rather than concentrations.

Mark BLR
Reply to  Bob
September 7, 2023 4:10 am

My understanding is that RCP (representative concentration pathway) is an indication of how much earth’s average global temperature will rise over the average global temperature of the mid to late 1800s.

Almost, but not quite …

This is one of the very few occasions where the AR6 assessment report from Working Group Three (WG-III) actually contains something “useful / interesting”.

Annex III is titled “Scenarios and modelling methods”, and can be downloaded as a PDF file from a link near the bottom of the following webpage.
URL : https://www.ipcc.ch/report/ar6/wg3/

From your post I think the sections the most … “helpful” (?) when it comes to answering your current set of questions are A.III.II.1.3.1, “History of scenario frameworks used by the IPCC”, and A.III.II.1.3.2, “Current scenario framework and SSP-based emission scenarios”, on pages 1872 and 1873 respectively.

From the end of section A.III.II.1.3.1 :

To shorten development times, a parallel approach was chosen (Moss et al. 2010) and representative concentration pathways (RCPs) were developed (van Vuuren et al. 2011b) to inform the next generation of climate modelling for [AR5]. RCPs explored four different emissions and atmospheric composition pathways structured to result in different levels of radiative forcing in 2100: 2.6, 4.5, 6.0 and 8.5 W m-2.

Notes

– The RCPs defined emission inputs to the (CMIP5) climate models that produced, on average, set targets of radiative forcing in 2100. GMST is related to RF, but they are different things.

– For the CMIP5 (/ AR5) modelling round the IPCC didn’t bother with how “feasible”, either technically or politically, a given target was. They just plugged the “GHG emissions levels” needed to reach the chosen “How much RF in 2100 ?” number into the model input files.

It was only with AR6 that the “socio-economic” aspects of the modelling process were (finally) included, as explained in section A.III.II.1.3.2 :

The current scenario framework for climate change research (van Vuuren et al. 2014; O’Neill et al. 2014; Kriegler et al. 2014c) is based on the concept of Shared Socio-Economic Pathways (SSPs) (Kriegler et al. 2012; O’Neill et al. 2014). Unlike their predecessor scenarios from the SRES (IPCC 2000), their underlying narratives are motivated by the purpose of using the framework for mitigation and adaptation policy analysis. Hence the narratives are structured to cover the space of socio-economic challenges to both adaptation and mitigation. They tell five stories of sustainability (SSP1), middle of the road development (SSP2), regional rivalry (SSP3), inequality (SSP4) and fossil-fuelled development (SSP5) (O’Neill et al. 2017). SSP1, SSP2, and SSP3 were structured to explore futures with socio-economic challenges to adaptation and mitigation increasing from low to high with increasing number of SSP. SSP4 was structured to explore a world with high socio-economic challenges to adaptation but low socio-economic challenges to mitigation, while SSP5 explored a world with low challenges to adaptation but high challenges to mitigation.

AR6 (WG-III) also assessed the “feasibility” of each “SSP + Radiative forcing in 2100” combination, as summarised in Annex III, Figure 4 (on page 1874, copied below).

Note that they got “feasible” combinations for SSP5 all the way down to 1.9 W/m² (in 2100).

NB : This post only partially answers your questions.
In the words of “old cocky”, if you’re looking for a complete response “It’s a bit complicated” …

AR6-WG-III_Annex-III-Figure-4.png
Bob
Reply to  Mark BLR
September 7, 2023 4:31 pm

Thanks Mark, I am aware of the SSPs but my comment is specifically meant to point out that Zeke used a graph of RCP 4.5 to justify that IPCC models matched observations. My point is that if he is going to use an RCP scenario it should be the one that matches reality. The reality is that regardless of what Europe and the US have done CO2 emissions have not plateaued or retreated but have increased steadily. That is not the pathway that 4.5 represents. RCP 6.0 or 8.5 would better represent today’s reality and graphs of either of those two pathways would not match observations.

Mark BLR
Reply to  Bob
September 7, 2023 4:28 am

If all of this is true then using the 4.5 RCP and claiming it matches observations is not honest.

As so often in the domain of “climate science”, Mark Twain’s famous quote about “lies,damned lies and statistics” applies.

Looking at fossil-fuel (not “total, including LULUCF / AFOLU” …) CO2 (not “all greenhouse gases” …) emissions, then an argument can be made that “the RCP 4.5 emissions ‘pathway’ is indeed the closest one to current observations”, and may even continue to be the closest up to (at least ?) 2030.

FF-CO2-emissions_2000-2030_thin.png
KevinM
September 6, 2023 7:37 pm

Each successive round of misleading curve fitting makes it more difficult for a genuine model to compete. If actual error is reasonably large, anyone who tries to develop an honest physical model will be destroyed by mathematicians with computers. I would be so upset if I were on a clean team – lost in obscurity like names that aren’t famous because of Barry Bonds and Lance Armstrong.

cwright
September 7, 2023 2:49 am

The match between prediction and actual shown in Fig. 1 is so good as to be totally unbelievable. The explanation is obvious. The computer runs are probably quite recent, so the modellers knew the answer already before doing the runs. Any fool can predict the past.

A modeller called it a “dirty secret”: the models are filled with “parameterisation” adjustments – otherwise known as fudge factors – so it’s easy to match the models against past observations by adjusting the fudge factors. Because of this, using adjusted model runs when the answer was already known to “prove” CO2 climate change is fraudulent.

The only way to assess a model’s prediction success is to compare it with observed temperatures AFTER the date of the computer runs. So, to check the period 1970 to 2020, computer runs from 1970 should only be used. The resulting graph would be *very* different to Fig. 1!

John Christy has done this. It shows that the models have been exaggerating the warming by almost a factor of three. They have zero predictive skill. Like the Covid graphs of doom, their sole purpose is to instill fear. They have little to do with science.

Chris

beng135
September 7, 2023 6:13 am

Zeke Hausenpheifer is an ilk of Stokes — a paid dupe.

JASchrumpf
September 7, 2023 4:12 pm

“The agreement between modeled and observed global mean surface temperature (GMST) warming over 1970–2020 shown in the Figure 1” does not look “impressive” to me. The range between the high and the low values is around 0.13. The average of the models looks to be around 0.53.

Who’s impressed by a precision of 0.53 +/- 12%?

Tim Gorman
Reply to  JASchrumpf
September 7, 2023 4:35 pm

It puts the unknown area at about +/- 0.06. Meaning it would be difficult to establish the difference between min and max to anything lower than the tenths digit. You certainty can’t establish a difference in the hundredths digit.

Mark BLR
September 8, 2023 3:54 am

The most recent such analysis appears to be that shown for CMIP5 models in a tweet by Zeke Hausfather, reproduced in Figure 1.

I instinctively actively avoid Twitter (/ now “X” ?), so didn’t actually check your link until today.

Nic’s “most recent” graph is almost three years old …

Real Climate, where Zeke Hausfather posted as “zeke” (the website is now basically mothballed, but guess who “mike” and “gavin” are / were …), updated their “analysis” in January of this year, which included the following for the CMIP5 modelling cycle.

URL : https://www.realclimate.org/index.php/climate-model-projections-compared-to-observations/

The link to RC’s “CMIP5 historical simulations (using the RCP4.5 projection post-2005)” PNG image file should “automagically” get inserted here …

comment image

NB : “B Zipperer” already posted a copy of RC’s CMIP6 “adjustments” graph, but I will include a link to the original at RC here as well to avoid having to scroll up and down this comments section …

comment image

Note that for CMIP5 (RCP 4.5 from 2006) the “forcing adjustments” had the effect of “bending” the entire range of the IPCC’s “projections”, while for CMIP6 (SSP2-4.5 from 2015) they just “filtered out” the evidence that “the most recent models are simply running too hot”.

PS : For the CMIP3 (SRES A1B from 2001) graph they decided no “adjustments” whatsoever were required …

comment image

%d
Verified by MonsterInsights