Buy

Books
Click images for more details

Twitter
Support

 

Recent comments
Recent posts
Currently discussing
Links

A few sites I've stumbled across recently....

Powered by Squarespace
« Von Storch on the pause | Main | The Economist continues to waver »
Friday
Jun212013

Brown out

Robert G. Brown (rgbatduke) has posted another devastating comment at WUWT, which I am again taking the liberty of reproducing in full here. For the counter-view, see Matt Briggs here.

Sorry, I missed the reposting of my comment. First of all, let me apologize for the typos and so on. Second, to address Nick Stokes in particular (again) and put it on the record in this discussion as well, the AR4 Summary for Policy Makers does exactly what I discuss above. Figure 1.4 in the unpublished AR5 appears poised to do exactly the same thing once again, turn an average of ensemble results, and standard deviations of the ensemble average into explicit predictions for policy makers regarding probable ranges of warming under various emission scenarios.

This is not a matter of discussion about whether it is Monckton who is at fault for computing an R-value or p-value from the mish-mosh of climate results and comparing the result to the actual climate — this is, actually, wrong and yes, it is wrong for the same reasons I discuss above, because there is no reason to think that the central limit theorem and by inheritance the error function or other normal-derived estimates of probability will have the slightest relevance to any of the climate models, let alone all of them together. One can at best take any given GCM run and compare it to the actual data, or take an ensemble of Monte Carlo inputs and develop many runs and look at the spread of results and compare THAT to the actual data.

In the latter case one is already stuck making a Bayesian analysis of the model results compared to the observational data (PER model, not collectively) because when one determines e.g. the permitted range of random variation of any given input one is basically inserting a Bayesian prior (the probability distribution of the variations) on TOP of the rest of the statistical analysis. Indeed, there are many Bayesian priors underlying the physics, the implementation, the approximations in the physics, the initial conditions, the values of the input parameters. Without wishing to address whether or not this sort of Bayesian analysis is the rule rather than the exception in climate science, one can derive a simple inequality that suggests that the uncertainty in each Bayesian prior on average increases the uncertainty in the predictions of the underlying model. I don’t want to say proves because the climate is nonlinear and chaotic, and chaotic systems can be surprising, but the intuitive order of things is that if the inputs are less certain and the outputs depend nontrivially on the inputs, so are the outputs less certain.

I will also note that one of the beauties of Bayes’ theorem is that one can actually start from an arbitrary (and incorrect) prior and by using incoming data correct the prior to improve the quality of the predictions of any given model with the actual data. A classic example of this is Polya’s Urn, determining the unbiased probability of drawing a red ball from an urn containing red and green balls (with replacement and shuffling of the urn between trials). Initially, we might use maximum entropy and use a prior of 50-50 — equal probability of drawing red or green balls. Or we might think to ourselves that the preparer of the urn is sneaky and likely to have filled the urn only with green balls and start with a prior estimate of zero. After one draws a single ball from the urn, however, we now have additional information — the prior plus the knowledge that we’ve drawn a (say) red ball. This instantly increases our estimate of the probability of getting red balls from a prior of 0, and actually very slightly increases the probability of getting a red ball from 0.5 as well. The more trials you make (with replacement) the better your successive approximations of the probability are regardless of where you begin with your priors. Certain priors will, of course, do a lot better than others!

I therefore repeat to Nick the question I made on other threads. Is the near-neutral variation in global temperature for at least 1/8 of a century (since 2000, to avoid the issue of 13, 15, or 17 years of “no significant warming” given the 1997/1999 El Nino/La Nina one-two punch since we have no real idea of what “signficant” means given observed natural variability in the global climate record that is almost indistinguishable from the variability of the last 50 years) strong evidence for warming of 2.5 C by the end of the century? Is it even weak evidence for? Or is it in fact evidence that ought to at least some extent decrease our degree of belief in aggressive warming over the rest of the century, just as drawing red balls from the urn ought to cause us to alter our prior beliefs about the probable fraction of red balls in Polya’s urn, completely independent of the priors used as the basis of the belief?

In the end, though, the reason I posted the original comment on Monckton’s list is that everybody commits this statistical sin when working with the GCMs. They have to. The only way to convince anyone that the GCMs might be correct in their egregious predictions of catastrophic warming is by establishing that the current flat spell is somehow within their permitted/expected range of variation. So no matter how the spaghetti of GCM predictions is computed and presented — and in figure 11.33b — not 11.33a — they are presented as an opaque range, BTW, — presenting their collective variance in any way whatsoever is an obvious visual sham, one intended to show that the lower edge of that variance barely contains the actual observational data.

Personally, I would consider that evidence that, collectively or singly, the models are not terribly good and should not be taken seriously because I think that reality is probably following the most likely dynamical evolution, not the least likely, and so I judge the models on the basis of reality and not the other way around. But whether or not one wishes to accept that argument, two very simple conclusions one has little choice but to accept are that using statistics correctly is better than using it incorrectly, and that the only correct way to statistically analyze and compare the predictions of the GCMs one at a time to nature is to use Bayesian analysis, because we lack an ensemble of identical worlds.

I make this point to put the writers of the Summary for Policy Makers for AR5 that if they repeat the egregious error made in AR4 and make any claims whatsoever for the predictive power of the spaghetti snarl of GCM computations, if they use the terms “mean and standard deviation” of an ensemble of GCM predictions, if they attempt to transform those terms into some sort of statement of probability of various future outcomes for the climate based on the collective behavior of the GCMs, there will be hell to pay, because GCM results are not iid samples drawn from a fixed distribution, thereby fail to satisfy the elementary axioms of statistics and render both mean behavior and standard deviation of mean behavior over the “space” of perturbations of model types and input data utterly meaningless as far as having any sort of theory-supported predictive force in the real world. Literally meaningless. Without meaning.

The probability ranges published in AR4′s summary for policy makers are utterly indefensible by means of the correct application of statistics to the output from the GCMs collectively or singly. When one assigns a probability such as “67%” to some outcome, in science one had better be able to defend that assignment from the correct application of axiomatic statistics right down to the number itself. Otherwise, one is indeed making a Ouija board prediction, which as Greg pointed out on the original thread, is an example deliberately chosen because we all know how Ouija boards work! They spell out whatever the sneakiest, strongest person playing the game wants them to spell.

If any of the individuals who helped to actually write this summary would like to come forward and explain in detail how they derived the probability ranges that make it so easy for the policy makers to understand how likely to certain it is that we are en route to catastrophe, they should feel free to do so. And if they in fact did form the mean of many GCM predictions as if GCMs are some sort of random variate, form the standard deviation of the GCM predictions around the mean, and then determine the probability ranges on the basis of the central limit theorem and standard error function of the normal distribution (as it is almost certain they did, from the figure caption and following text) then they should be ashamed of themselves and indeed, should go back to school and perhaps even take a course or two in statistics before writing a summary for policy makers that presents information influencing the spending of hundreds of billions of dollars based on statistical nonsense.

And for the sake of all of us who have to pay for those sins in the form of misdirected resources, please, please do not repeat the mistake in AR5. Stop using phrases like “67% likely” or “95% certain” in reference to GCM predictions unless you can back them up within the confines of properly done statistical analysis and mere common wisdom in the field of predictive modeling — a field where I am moderately expert — where if anybody, ever claims that a predictive model of a chaotic nonlinear stochastic system with strong feedbacks is 95% certain to do anything I will indeed bitch slap them the minute they reach for my wallet as a consequence.

Predictive modeling is difficult. Using the normal distribution in predictive modeling of complex multivariate system is (as Taleb points out at great length in The Black Swan) easy but dumb. Using it in predictive modeling of the most complex system of nominally deterministic equations — a double set of coupled Navier Stokes equations with imperfectly known parameters on a rotating inhomogeneous ball in an erratic orbit around a variable star with an almost complete lack of predictive skill in any of the inputs (say, the probable state of the sun in fifteen years), let alone the output — is beyond dumb. Dumber than dumb. Dumb cubed. The exponential of dumb. The phase space filling exponential growth of probable error to the physically permitted boundaries dumb.

In my opinion — as admittedly at best a well-educated climate hobbyist, not as a climate professional, so weight that opinion as you will — we do not know how to construct a predictive climate model, and will never succeed in doing so as long as we focus on trying to explain “anomalies” instead of the gross nonlinear behavior of the climate on geological timescales. An example I recently gave for this is understanding the tides. Tidal “forces” can easily be understood and derived as the pseudoforces that arise in an accelerating frame of reference relative to Newton’s Law of Gravitation. Given the latter, one can very simply compute the actual gravitational force on an object at an actual distance from (say) the moon, compare it to the actual mass times the acceleration of the object as it moves at rest relative to the center of mass of the Earth (accelerating relative to the moon) and compute the change in e.g. the normal force that makes up the difference and hence the change in apparent weight. The result is a pseudoforce that varies like (R_e/R_lo)^3 (compared to the force of gravity that varies like 1/R_lo^2 , R_e radius of the earth, R_lo radius of the lunar orbit). This is a good enough explanation that first year college physics students can, with the binomial expansion, both compute the lunar tidal force and compute the nonlinear tidal force stressing e.g. a solid bar falling into a neutron star if they are a first year physics major.

It is not possible to come up with a meaningful heuristic for the tides lacking a knowledge of both Newton’s Law of Gravitation and Newton’s Second Law. One can make tide tables, sure, but one cannot tell how the tables would CHANGE if the moon was closer, and one couldn’t begin to compute e.g. Roche’s Limit or tidal forces outside of the narrow Taylor series expansion regime where e.g. R_e/R_lo << 1. And then there is the sun and solar tides making even the construction of an heuristic tide table an art form.

The reason we cannot make sense of it is that the actual interaction and acceleration are nonlinear functions of multiple coordinates. Note well, simple and nonlinear, and we are still a long way from solving anything like an actual equation of motion for the sloshing of oceans or the atmosphere due to tidal pseudoforces even though the pseudoforces themselves are comparatively simple in the expansion regime. This is still way simpler than any climate problem.

Trying to explain the nonlinear climate by linearizing around some set of imagined “natural values” of input parameters and then attempting to predict an anomaly is just like trying to compute the tides without being able to compute the actual orbit due to gravitation first. It is building a Ptolemaic theory of tidal epicycles instead of observing the sky first, determining Kepler’s Laws from the data second, and discovering the laws of motion and gravitation that explain the data third, finding that they explain more observations than the original data (e.g. cometary orbits) fourth, and then deriving the correct theory of the tidal pseudoforces as a direct consequence of the working theory and observing agreement there fifth.

In this process we are still at the stage of Tycho Brahe and Johannes Kepler, patiently accumulating reliable, precise observational data and trying to organize it into crude rules. We are only decades into it — we have accurate knowledge of the Ocean (70% of the Earth’s surface) that is at most decades long, and the reliable satellite record is little longer. Before that we have a handful of decades of spotty observation — before World War II there was little appreciation of global weather at all and little means of observing it — and at most a century or so of thermometric data at all, of indifferent quality and precision and sampling only an increasingly small fraction of the Earth’s surface. Before that, everything is known at best by proxies — which isn’t to say that there is not knowledge there but the error bars jump profoundly, as the proxies don’t do very well at predicting the current temperature outside of any narrow fit range because most of the proxies are multivariate and hence easily confounded or merely blurred out by the passage of time. They are Pre-Ptolemaic data — enough to see that the planets are wandering with respect to the fixed stars, and perhaps even enough to discern epicyclic patterns, but not enough to build a proper predictive model and certainly not enough to discern the underlying true dynamics.

I assert — as a modest proposal indeed — that we do not know enough to build a good, working climate model. We will not know enough until we can build a working climate model that predicts the past — explains in some detail the last 2000 years of proxy derived data, including the Little Ice Age and Dalton Minimum, the Roman and Medieval warm periods, and all of the other significant decadal and century scale variations in the climate clearly visible in the proxies. Such a theory would constitute the moral equivalent of Newton’s Law of Gravitation — sufficient to predict gross motion and even secondary gross phenomena like the tides, although difficult to use to compute a tide table from first principles. Once we can predict and understand the gross motion of the climate, perhaps we can discern and measure the actual “warming signal”, if any, from CO_2. In the meantime, as the GCMs continue their extensive divergence from observation, they make it difficult to take their predictions seriously enough to condemn a substantial fraction of the world’s population to a life of continuing poverty on their unsupported basis.

Let me make this perfectly clear. WHO has been publishing absurdities such as the “number of people killed every year by global warming” (subject to a dizzying tower of Bayesian priors I will not attempt to deconstruct but that render the number utterly meaningless). We can easily add to this number the number of people a year who have died whose lives would have been saved if some of the half-trillion or so dollars spent to ameliorate a predicted disaster in 2100 had instead been spent to raise them up from poverty and build a truly global civilization.

Does anyone doubt that the ratio of the latter to the former — even granting the accuracy of the former — is at least a thousand to one? Think of what a billion dollars would do in the hands of Unicef, or Care. Think of the schools, the power plants, the business another billion dollars would pay for in India, in central Africa. Go ahead, think about spending 498 more billions of dollars to improve the lives of the world’s poorest people, to build up its weakest economies. Think of the difference not spending money building inefficient energy resources in Europe would have made in the European economy — more than enough to have completely prevented the fiscal crisis that almost brought down the Euro and might yet do so.

That is why presenting numbers like “67% likely” on the basis of gaussian estimates of the variance of averaged GCM numbers as if it has some defensible predictive force to those who are utterly incapable of knowing better is not just incompetently dumb, it is at best incompetently dumb. The nicest interpretation of it is incompetence. The harshest is criminal malfeasance — deliberately misleading the entire world in such a way that millions have died unnecessarily, whole economies have been driven to the wall, and worldwide suffering is vastly greater than it might have been if we had spent the last twenty years building global civilization instead of trying to tear it down!

Even if the predictions of catastrophe in 2100 are true — and so far there is little reason to think that they will be based on observation as opposed to extrapolation of models that rather appear to be failing — it is still not clear that we shouldn’t have opted for civilization building first as the lesser of the two evils.

I will conclude with my last standard “challenge” for the warmists, those who continue to firmly believe in an oncoming disaster in spite of no particular discernible warming (at anything like a “catastrophic” rate” for somewhere between 13 and 17 years), in spite of an utterly insignificant rate of SLR, in spite of the growing divergence between the models and reality. If you truly wish to save civilization, and truly believe that carbon burning might bring it down, then campaign for nuclear power instead of solar or wind power. Nuclear power would replace carbon burning now, and do so in such a way that the all-important electrical supply is secure and reliable. Campaign for research at levels not seen since the development of the nuclear bomb into thorium burning fission plants, as the US has a thorium supply in North Carolina alone that would supply its total energy needs for a period longer than the Holocene, and so does India and China — collectively a huge chunk of the world’s population right there (and thorium is minded with rare earth metals needed in batteries, high efficiency electrical motors, and more, reducing prices of all of these key metals in the world marketplace). Stop advocating the subsidy of alternative energy sources where those sources cannot pay for themselves. Stop opposing the burning of carbon for fuel while it is needed to sustain civilization, and recognize that if the world economy crashes, if civilization falls, it will be a disaster that easily rivals the worst of your fears from a warmer climate.

Otherwise, while “deniers” might have the blood of future innocents on their hands if your future beliefs turn out to be correct, you’ll continue to have the blood of many avoidable deaths in the present on your own.

[Updated to add link to Briggs.]

 

PrintView Printer Friendly Version

Reader Comments (79)

It's about time more physicists speak out as RGB has done. What a magnificent screed.

Jun 21, 2013 at 3:17 PM | Unregistered CommenterNZ Willy

Roger Longstaff

I was referring to carbon trading / Soros / global finance, not 'irrelevant science peasants'. Nobody cares about science or scientists as long as they do what they are told. They generally do. Apart from Doug Proctor, Pielke (cleverly), Spencer and a few others.

Tongue in cheek. No personal offence intended to anyone. We all have to make a living.

Jun 21, 2013 at 3:18 PM | Unregistered CommentereSmiff

Michael Mann, Kevin Treberth, Gavin Schmidt, Eric Steig?

How about Karoly?

etc etc etc

Jun 21, 2013 at 3:21 PM | Unregistered Commenternot banned yet

So, in order to get the ensemble mean closer to predictions they just have to add more to the ensemble that have a climate sensitivity to CO2 much closer to reality and reject those that are miles off.

So what can be the problem?

/:)

Jun 21, 2013 at 3:21 PM | Unregistered Commenterssat

Briggs is once again just making stuff up. Ok he knows about how they do things for weather forecasting. But there are good reasons why weather forecasting is suitable for ensemble modeling: The individual models are regularly tested against real life and nobody expects them to be accurate more than a few days out. Yes Briggs is right to claim that ensemble modeling sometimes works. This result did not arrive by any stats theory because to a real modeler like Brown (and me) but not Briggs the idea is absurd. It was merely discovered by happy accident. The reasons why it sometimes works and sometimes doesn't has bu**er all to do with any stats method; it is because each individual model has to be reasonably close to reality prior to even thinking about doing an ensemble.

Most of the rest of what Briggs then writes is pure semantic argumentation with no actual maths involvement. Of course Brown is correct to talk about Monte Carlo selection of input parameters as the only way you can apply frequentist stats to models but Brown did not say it was smart to do so. All he is saying is that frequentist stats terminology cannot be used at all - regardless of what Briggs thinks. This is well established in the literature and Brown is on solid ground. Model selection is at best a hopelessly biased Bayesian process, where inexpert people have decided that mere gut feeling can replace expertise. That the models do not reflect reality, however combined, is a reflection that such amateur Baysianism only works when each model is reasonable in the first place. To summarise, the problem is threefold:
a) The models have not been individually validated - in terms of percentage error. Hindcasting is not validation! Without proper validation they cannot reasonably go into any ensemble: Expecting good results from rubbish input is plain stupid!
b) The prior assumption of high climate sensitivity is merely circular reasoning with a computational cloak, that is then fudged by the aerosol parameter to achieve the consequently faked hindcasting results. The models run too hot because of this circular reasoning.
c) But the real problem is that the task is just too difficult until more work is done to establish exactly what natural variation consists of. Even after that is done, the models cannot be fine enough to represent the myriad of eddies and randomness that exist in nature.

Hence the whole exercise is a waste of time and money as I'm sure both Briggs and Brown agree. What Nick Stokes thinks is, as always, totally irrelevant as he is known for repeatedly defending the indefensible to the point of arguing black is white.

Jun 21, 2013 at 3:39 PM | Unregistered CommenterJamesG

simon abingdon, you just neatly illustrated the dangers of trying to simplify an idea past the point where the idea is lost :)

Jun 21, 2013 at 3:42 PM | Unregistered CommenterTheBigYinJames

@ simon abingdon (3:06 PM)

Answering my own question TBYJ, the average of the children's guesses cannot bear any relationship to the teacher's choice since she makes it independently of them.

And if the children's guesses are like the (necessarily flawed) GCM's and the teacher's (secret) choice represents the best GCM, then the average of the children's guesses (the ensemble of flawed GCM's) clearly bears no relationship whatsoever to the teacher's prior choice (the preferred but unknown GCM whose identity only reveals itself with the passage of time).

Jun 21, 2013 at 4:12 PM | Unregistered Commentersimon abingdon

Duhhh. Strike 'predictions' insert 'obsevations'.

Anyway, IPCC and reality are divergent. Waste of poor typing really.

Jun 21, 2013 at 5:18 PM | Unregistered Commenterssat

This is interesting and worthy of discussion but i think what matters is Hans von Storchs' admission that only 2 % of his model runs can accomadate the 15 year " pause in warming" we are currently in and that by his estimation another 5 years of no " warming" is game over for all the models. Dont let the warmist continually play the bait and switch game.
This is what matters now and for AR5.
Observation trumps models.
Von Storch and Spiegel Online should be applauded and supported for stating the inescapable. The rest deserve hard questions and contempt.

Jun 21, 2013 at 7:08 PM | Unregistered CommenterNick in Vancouver

simon, probably a better example would be the various "systems" used by gamblers to win at the horses. They're all flawed one way or another, but averaging all of them is not going to give you a better system than any of the individual ones.

Jun 21, 2013 at 8:20 PM | Unregistered CommenterTheBigYinJames

@TheBigYinJames (8:20 PM)

Absolutely so. Averaging flawed GCM's in the pious hope that maybe some sort of Wisdom of the Crowds effect will suddenly come into play looks like utter Botox to me.

Jun 21, 2013 at 9:13 PM | Unregistered Commentersimon abingdon

What this means is that as far as the models go, the vast consensus is worthless.
What is entertaining is how Trenberth and other defenders of the true faith are coming up with excuses, aspersions, distractions, etc. to avoid dealing with their failures.

Jun 21, 2013 at 10:51 PM | Unregistered Commenterlurker, passing through laughing

JamesG
Thanks for that post.

Because Brown talks about how the WHO calculates mortality statistics from climate change, I would encourage readers to google 'Pride Chigwedere', and follow the science, wherever it takes.

Jun 21, 2013 at 11:02 PM | Registered Commentershub

Briggs suggests that the averaging of models originated because it sometimes works. Brown, on the other hand, points to the real reason I think it started in most applications, which is the Central Limit Theorem. As a first-year statistics student knows, take enough samples of any distribution, and the average of those samples will approach to a Normal distribution, which would make the statistics a lot easier. Ooh, Ooh, how tantalizing.

But oh my, how tangled up people get with this concept. If you mis-apply it, it is worse than useless. What it refers to is the AVERAGES of INDEPENDENT samples from the ENTIRE population distribution. Otherwise we have several names for the miserable mistake you have just made, such as auto-correlation. But the point is the rule doesn't apply if you start by violating its assumptions and everything it stands for.

If you do that, you might possibly come up with a number closer to reality than any one of the samples, but you most certainly did NOT improve your statistical chances. You just got lucky.

Jun 21, 2013 at 11:12 PM | Unregistered CommenterRoberto

Has nobody mentioned another reason that you can't do this with model runs. Model runs are weeded before the data ever gets to a graph. Some are curtailed mid-run, when they go wild. Some are rejected immediately because the result doesn't fit the narrative. There's no way the survivors represent anything but the prejudices of the modellers.

Jun 21, 2013 at 11:17 PM | Unregistered Commenterrhoda

NotBannedYet: "In your scientific endeavours have you ever worked in modelling?"

I am working with multi-million cell stochastic models every day, every week, every year for the last 20 years (of a professional earth science career spanning 28 years). Difference is my models are tested by experiment - drilling oil wells. I make a good living from success, not failure - and a healthy dose of humility when making predictions against Mother Nature and being on the losing side often enough to know the limitations of my expertise, skill and the geological and geophysical models we use.

Jun 21, 2013 at 11:45 PM | Unregistered CommenterThinkingScientist

CheshireRed: "Thinking Scientist's final line summarises it nicely."

Do you mean "in a nutshell, the climate models are bollocks."

or:

"Model...meet real world."?

Jun 21, 2013 at 11:49 PM | Unregistered CommenterThinkingScientist

probably a better example would be the various "systems" used by gamblers to win at the horses. They're all flawed one way or another,

Oi, I've made my living betting on the nags for 10 years now. I have loads of hunches (theories), back test them on historical data, those that survive I then run forward on real live results. Any that don't survive that are culled, those that show promise I run with and put money on, consistently monitoring to check they're still working. Losing real cash doesn't half focus the mind. No averaging in sight, survival of the fittest.

That process doesn't seem hard to understand to me, the hard part is getting the theory right in the first place.

Very interesting debate this one though as Briggs is a bit of a hero but then RGB's essays are also superb. I'm not into the technical stats as much as some of you guys/gals but I can't see how keeping in unvalidated crap models can improve any analysis. You've got to cull surely.

I'm hoping it is just some semantical disagreement that can be resolved but (and I like to put my neck on the line, best way to learn) I'd put my money on RGB.

Jun 21, 2013 at 11:52 PM | Registered CommenterSimonW

A note of clarification. I suspect that the actual driver for a lot of the averaging is the Central Limit Theorem. How do I get from one to the other?

I have observed a lot of people who had heard something like the Central Limit Theorem in some form. This got mushed up and distorted in their minds into a general expectation that averaging cures a multitude of statistical ills. If in doubt, average it. This will improve your calculations, your chances, your information content, and anything else. Averaging has magic to cure the wild uncertainty.

But in reality, it all depends on what you average. For instance if you take the average of 20 quick temperature readings in the refrigerator for one sample and the average of 20 quick temperature readings by the fireplace for another sample, that is no better at all than 1 of each. It is no more information, no improvement in calculations, no better distribution, and so forth. These are called correlated samples, and they are no use at all. (At least not unless they swing wildly within the same sample. That might be worth knowing.)

Jun 21, 2013 at 11:59 PM | Unregistered CommenterRoberto

Thinking scientist - sounds fun :-)

Did you see the WG1 explanation of terms for their work? Do you see the distinction between a scenario and the resulting projection?

Jun 22, 2013 at 12:09 AM | Unregistered Commenternot banned yet

"Briggs is once again just making stuff up."

Not quite. But I don't think people are understanding what Briggs was trying to say - it's a lot less significant an objection than people seem to think.

"Most of the rest of what Briggs then writes is pure semantic argumentation with no actual maths involvement."

That's probably closer to it.

What I think Briggs is saying is that it's wrong to describe a prediction based on the mean and spread of a model ensemble as "meaningless". It has a meaning: it's the interval you're predicting the value will fall into. It doesn't matter whether you calculated it from running climate models, or by picking phone numbers out of a book. It's a calculable number, which you are making a prediction with.

What Brown meant when he said "meaningless" was that there is no rational reason to expect such a method to give an accurate forecast of the climate. The spread does not represent the uncertainty that you rationally ought to have. It's an arbitrary, calculated number, a mish-mash of concepts that says more about the prejudices of climate modellers than it does about the future climate

Briggs agrees with this, but is saying it has nothing to do with the point he was making, which is that wrong, stupid, irrational forecasts are still *meaningful* in statistics, like the sentence "the sky is green" is meaningful. We know what it means. They're just wrong.

It's just nit-picking about the semantics. Unless you are deep into the Bayesian/Frequentist philosophy debate, it's not important.

--
Just as another minor nit-pick, the central limit theorem only works if the mean and variance are finite (there are some distributions where they're not), and there are versions of it that apply to adding variables that are not independent or identically distributed, so long as certain other conditions apply. Just because the distributions are not identical doesn't mean that adding them won't tend towards a Gaussian distribution. But it can be a delicate point, that ought not to be assumed. But again, this is not something the layman needs to worry about.

Jun 22, 2013 at 12:33 AM | Unregistered CommenterNullius in Verba

Sure, the average of model ensembles can be useful, but not this ensemble of models.
===============

Jun 22, 2013 at 5:37 AM | Unregistered Commenterkim

Yes, understanding the behaviours of tides globally seems pretty fundamental to any theory of climate evolution. Maybe an outline explanation of how they're assimilated into the GCM's would be interesting.

Jun 22, 2013 at 8:57 AM | Unregistered Commentersimon abingdon

A couple of people have touched on what is another really basic abuse of stats by apparently clever maths-oriented people. Briggs, Schmidt, Santer, Annan and many others have made this basic error in their arguments: You just cannot take a handful of samples of anything (most especially preselected samples) and a priori assume any statistical distribution whatsoever. You have to take hundreds or thousands; ie enough to actually plot out the correct distribution; be it normal, skewed or otherwise. NB: The distribution is found, not assumed! Without proof of a distribution then all you have is a spread of results each of which is as likely as the other. If you circumvent this process then you have made a judgement based on your assumed expertise. That is why it is pseudo-Baysianism at best. Talk about standard deviations or meaningful spreads for preselected models is just wrong. Do not perpetuate other peoples basic errors, or even entertain them in the first place! Wrong is wrong!

Frequentist stats are for experimental data alone. And no matter how hard people pretend otherwise to themselves; model runs are not experiments, they are attempts to mimic nature. If nature disagrees with your model (using percentage error or simple averages or even basic eyeballing) then the model needs fixed. At no point should anyone assume that models trump reality. Sure they can sometimes point out some flaws in the observational equipment but that is rare. Most of the time the data is the final arbiter as Feynmann so eloquently reminded us. To adjust data to fit models as per Rahmstorf and Foster - words alone cannot express enough disgust: They need a good kick up the arse.

Jun 22, 2013 at 10:22 AM | Unregistered CommenterJamesG

These comments by Hans von Storch in 2012 ("We always make models for something, not of something") seem relevant to this discussion, suggesting that the models are far from independent:

Q: If the models are different in terms of the parametrization—are they nevertheless giving the same predictions about the climate?

Von Storch: In principle, yes. Predictions in climate science are mostly conditioned predictions, scenarios. The emission of greenhouse gases, for example, is something you cannot predict but it is a crucial parameter for the evolution of the climate. You can make an assumption about the level of emissions and then calculate how the climate evolves under that assumption.

But taking this into account, all the models yield quite similar results. There might be small differences, maybe the equilibrium temperature for a doubling of the CO2 concentration in the atmosphere is 3 deg C for one model and 4 deg C for another model. But the general trends are reproduced with all models. Of course that is something, one could also be critical about. The scientists making these models know each other more or less. And if somebody then finds very unusual results, he might become shaky and say: Well, maybe my model is not as good as the other 17 models that are around. And then he tries to adjust his model to agree with the other 17 models. There is also a social process that leads to the agreement between all the different climate models. [my bold – RD]

The whole interview is here (pdf).

Jun 22, 2013 at 12:31 PM | Registered CommenterRuth Dixon

Could we have another opinion about this? Briggs is an expert, I'm not. Is Brown?
Tol or who knows who?
This is important.

Jun 22, 2013 at 10:53 PM | Unregistered Commentercdc

Bayes' Theorem is a source of much confusion in debates about climate science. That confusion can be removed if we recognize that application of Bayes' Theorem is highly sensitive to the context in which it is applied. Bayes' Theorem can be applied with confidence in contexts that are highly ramified. In contexts that are not highly ramified, Bayes' Theorem can do no more that help us learn our mistakes in betting.

Bayes' Theorem is used in the microprocessors that drive our computers. That context is highly ramified. To see the degree of ramification, list the physical laws, the principles of chip design, the principles of electrical engineering, and the laws of signal theory that go into the design and implementation of microprocessors in computers. The list will extend into the hundreds though each item is either a highly confirmed law of physical theory or a principle of engineering that is well understood within that context. Measuring relative ramification on a scale of 1-10, the context rates an eight or nine. Compare this context to that of climate science.

In climate science, there is no physical theory that comprehends and explains all of the forcings and feedbacks that are believed to have an effect on the global average temperature or whatever measure that we choose to focus upon. If we attempt to construct a list of physical laws and various other principles for climate science then our list will be incomplete. On a scale of 1-10, the relative ramification of this context is zero. In this context, Bayes' Theorem applies only to collections of data because there is no theory to serve as context.

Things could have been different. The early Trenberth believed that radiation theory alone would comprehend and explain the phenomena of climate science. If he had been correct, applications of Bayes' Theorem would have been successful in this highly ramified context. But the later Trenberth says that "the missing heat" is hiding in the deep oceans. In doing so, he acknowledges that radiation theory must be supplemented with physical laws that describe natural processes apart from radiation theory. We are back to applying Bayes' Theorem to collections of data. The context rates a zero on the scale of relative ramification.

Bayes' Theorem can be applied to the collections of data that models use. Applying it to the models themselves, to model projections, and expecting some scientific benefit requires assuming that there is a unified context that allows a fair characterization of the differences among the models. A comprehensive climate theory is required. (Please note that "compehensive" does not mean final.)

Jun 23, 2013 at 5:21 PM | Unregistered CommenterTheo Goodwin

Take a look at this "spegetti graph" by Dr Spencer. http://www.drroyspencer.com/wp-content/uploads/CMIP5-73-models-vs-obs-20N-20S-MT-5-yr-means1.png

Notice that all the climate forecasts are "high" with regards to predictions/projections/whatever you want to call them as compared to actual observations. Any kind of averaging/whatever you want to call it is still using models that are wrong.

In Dr Browns first comment on this he was suggesting to use a few of the "wrong" models that were closest to reality and work with those and drop/dump/disregard/whatever you want to call it. In a nut shell let's get rid of those that were the most wrong. All the straw man arguments and sidetrack distraction still doesn't change the fact that the climate models are simply wrong and most much more so than others.

Damn, if the apples in your basket are mostly fully rotten just keep the ones that you may be able to stomach if you are not willing to dump the entire basket.

Jun 24, 2013 at 1:09 AM | Unregistered Commentereyesonu

I made an omission of a couple of words in my comment above:

" ... use a few of the "wrong" models that were closest to reality and work with those and drop/dump/disregard/whatever you want to call it" [add text: that are the most in error.] In a nut shell let's get rid of those that were the most wrong.

Proof reading is a good habit. ;-)

Jun 24, 2013 at 3:52 AM | Unregistered Commentereyesonu

PostPost a New Comment

Enter your information below to add a new comment.

My response is on my own website »
Author Email (optional):
Author URL (optional):
Post:
 
Some HTML allowed: <a href="" title=""> <abbr title=""> <acronym title=""> <b> <blockquote cite=""> <code> <em> <i> <strike> <strong>