SI
SI
discoversearch

We've detected that you're using an ad content blocking browser plug-in or feature. Ads provide a critical source of revenue to the continued operation of Silicon Investor.  We ask that you disable ad blocking while on Silicon Investor in the best interests of our community.  If you are not using an ad blocker but are still receiving this message, make sure your browser's tracking protection is set to the 'standard' level.
Politics : Formerly About Advanced Micro Devices -- Ignore unavailable to you. Want to Upgrade?


To: Land Shark who wrote (633111)10/25/2011 1:08:07 PM
From: Brumar892 Recommendations  Read Replies (2) | Respond to of 1578134
 
One good thing about the BEST PR campaign is it has established that the warming trend we're in goes back to at least 1800, long before mankind began using fossil fuels on a large scale.



To: Land Shark who wrote (633111)10/25/2011 1:08:37 PM
From: Brumar892 Recommendations  Read Replies (1) | Respond to of 1578134
 
A mathematician’s response to BEST

It's blistering!

... Considering the second paper, on Urban Heat Islands, the conclusion there is that there has been some urban cooling. That conclusion contradicts over a century of research as well as common experience. It is almost certainly incorrect.

... The data analysis in the BEST papers would not pass in a third-year undergraduate course in statistical time series
.

.. it is not merely my opinion that the BEST statistical analysis is seriously invalid; rather, the BEST statistical analysis is seriously invalid


Posted on October 21, 2011 by Anthony Watts

Doug Keenan in 2009

Doug Keenan, who readers may remember doggedly pursued and won some tree ring data that Queens University held back, was asked to comment of the BEST papers by the Economist. He posted up the full correspondence, including his critiques. There’s some interesting things in there. Since Dr. Muller and BEST want full transparency, in that interest, I’m making this available here. Start from the bottom up to maintain the timeline. h/t to Bishop Hill

He writes:

The Economist asked me to comment on four research papers from the Berkeley Earth Surface Temperature (BEST) project. The four papers, which have not been published, are as follows.

Below is some of the correspondence that we had. (Note: my comments were written under time pressure, and are unpolished.)

From: D.J. Keenan
To: Richard Muller [BEST Scientific Director]; Charlotte Wickham [BEST Statistical Scientist]
Cc: James Astill; Elizabeth Muller
Sent: 17 October 2011, 17:16
Subject: BEST papers
Attach: Roe_FeedbacksRev_08.pdf; Cowpertwait & Metcalfe, 2009, sect 2-6-3.pdf; EmailtoDKeenan12Aug2011.pdf

Charlotte and Richard,

James Astill, Energy & Environment Editor of The Economist, asked Liz Muller if it would be okay to show me your BEST papers, and Liz agreed. Thus far, I have looked at two of the papers.

  • Decadal Variations in the Global Atmospheric Land Temperatures
  • Influence of Urban Heating on the Global Temperature Land Average Using Rural Sites Identified from MODIS Classifications
Following are some comments on those.
In the first paper, various series are compared and analyzed. The series, however, have sometimes been smoothed via a moving average. Smoothed time series cannot be used in most statistical analyses. For some comments on this, which require only a little statistical background, see these blog posts by Matt Briggs (who is a statistician):
Do not smooth times series, you hockey puck!
Do NOT smooth time series before computing forecast skill

Here is a quote from those (formatting in original).

Unless the data is measured with error, you never, ever, for no reason, under no threat, SMOOTH the series! And if for some bizarre reason you do smooth it, you absolutely on pain of death do NOT use the smoothed series as input for other analyses! If the data is measured with error, you might attempt to model it (which means smooth it) in an attempt to estimate the measurement error, but even in these rare cases you have to have an outside (the learned word is “exogenous”) estimate of that error, that is, one not based on your current data.

If, in a moment of insanity, you do smooth time series data and you do use it as input to other analyses, you dramatically increase the probability of fooling yourself! This is because smoothing induces spurious signals—signals that look real to other analytical methods.

This problem seems to invalidate much of the statistical analysis in your paper.

There is another, larger, problem with your papers. In statistical analyses, an inference is not drawn directly from data. Rather, a statistical model is fit to the data, and inferences are drawn from the model. We sometimes see statements such as “the data are significantly increasing”, but this is loose phrasing. Strictly, data cannot be significantly increasing, only the trend in a statistical model can be.

A statistical model should be plausible on both statistical and scientific grounds. Statistical grounds typically involve comparing the model with other plausible models or comparing the observed values with the corresponding values that are predicted from the model. Discussion of scientific grounds is largely omitted from texts in statistics (because the texts are instructing in statistics), but it is nonetheless crucial that a model be scientifically plausible. If statistical and scientific grounds for a model are not given in an analysis and are not clear from the context, then inferences drawn from the model should be regarded as unfounded.

The statistical model adopted in most analyses of climatic time series is a straight line (usually trending upward) with noise (i.e. residuals) that are AR(1). AR(1) is short for “first-order autoregressive”, which means, roughly, that this year (only) has a direct effect on next year; for example, if this year is extremely cold, then next year will have a tendency to be cooler than average.

That model—a straight line with AR(1) noise—is the model adopted by the IPCC (see AR4: §I.3.A). It is also the model that was adopted by the U.S. Climate Change Science Program (which reports to Congress) in its analysis of “ Statistical Issues Regarding Trends”. Etc. An AR(1)-based model has additionally been adopted for several climatic time series other than global surface temperatures. For instance, it has been adopted for the Pacific Decadal Oscillation, studied in your work: see the review paper by Roe [2008], attached.

Although an AR(1)-based model has been widely adopted, it nonetheless has serious problems. The problems are actually so basic that they are discussed in some recent introductory (undergraduate) texts on time series—for example, in Time Series Analysis and Its Applications (third edition, 2011) by R.H. Shumway & D.S. Stoffer (see Example 2.5; set exercises 3.33 and 5.3 elaborate).

In Australia, the government commissioned the Garnaut Review to report on climate change. The Garnaut Review asked specialists in the analysis of time series to analyze the global temperature series. The report from those specialists considered and, like Shumway & Stoffer, effectively rejected the AR(1)-based statistical model. Statistical analysis shows that the model is too simplistic to cope with the complexity in the series of global temperatures.

Additionally, some leading climatologists have strongly argued on scientific grounds that the AR(1)-based model is unrealistic and too simplistic [ Foster et al., GRL, 2008].

To summarize, most research on global warming relies on a statistical model that should not be used. This invalidates much of the analysis done on global warming. I published an op-ed piece in the Wall Street Journal to explain these issues, in plain English, this year.

http://www.informath.org/media/a42.htm

The largest center for global-warming research in the UK is the Hadley Centre. The Hadley Centre employs a statistician, Doug McNeall. After my op-ed piece appeared, Doug McNeall and I had an e-mail discussion about it. A copy of one of his messages is attached. In the message, he states that the statistical model—a straight line with AR(1) noise—is “simply inadequate”. (He still believes that the world is warming, primarily due to computer simulations of the global climate system.)

Although the AR(1)-based model is known to be inadequate, no one knows what statistical model should be used. There have been various papers in the peer-reviewed literature that suggest possible resolutions, but so far no alternative model has found much acceptance.

When I heard about the Berkeley Earth Surface Temperature project, I got the impression that it was going to address the statistical issues. So I was extremely curious to see what statistical model would be adopted. I assumed that strong statistical expertise would be brought to the project, and I was trusting that, at a minimum, there would be a big improvement on the AR(1)-based model. Indeed, I said this in an interview with The Register last June.

BEST did not adopt the AR(1)-based model; nor, however, did it adopt a model that deals with some of the complexity that AR(1) fails to capture. Instead, BEST chose a model that is much more simplistic than even AR(1), a model which allows essentially no structure in the time series. In particular, the model that BEST adopted assumes that this year has no effect on next year. That assumption is clearly invalid on climatological grounds. It is also easily seen to be invalid on statistical grounds. Hence the conclusions of the statistical analysis done by BEST are unfounded.

All this occurred even though understanding the crucial question—what statistical model should be used?—requires only an introductory level of understanding in time series. The question is so basic that it is discussed by the introductory text of Shumway & Stoffer, cited above. Another text that does similarly is Introductory Time Series with R by P.S.P. Cowpertwait & A.V. Metcalfe (2009); a section from that text is attached. (The section argues that, from a statistical perspective, a pure AR(4) model is appropriate for global temperatures.) Neither Shumway & Stoffer nor Cowpertwait & Metcalfe have an agenda on global warming, to my knowledge. Rather, they are just writing introductory texts on time series and giving students practical examples; each text includes the series of global temperatures as one of those examples.

There are also textbooks that are devoted to the statistical analysis of climatic data and that discuss time-series modeling in detail. My bookshelf includes the following.
Climate Time Series Analysis (Mudelsee, 2010)
Statistical Analysis in Climate Research (von Storch & Zwiers, 2003)
Statistical Methods in the Atmospheric Sciences (Wilks, 2005)
Univariate Time Series in Geosciences (Gilgen, 2006)

Considering the second paper, on Urban Heat Islands, the conclusion there is that there has been some urban cooling. That conclusion contradicts over a century of research as well as common experience. It is almost certainly incorrect. And if such an unexpected conclusion is correct, then every feasible effort should be made to show the reader that it must be correct.

I suggest an alternative explanation. First note that the stations that your analysis describes as “very rural” are in fact simply “places that are not dominated by the built environment”. In other words, there might well be, and probably is, substantial urbanization at those stations. Second, note that Roy Spencer has presented evidence that the effects of urbanization on temperature grow logarithmically with population size.
The Global Average Urban Heat Island Effect in 2000 Estimated from Station Temperatures and Population Density Data

Putting those two notes together, we might expect that the UHI effect will be larger at the sites classified as “very rural” than at the sites classified as urban. And that is indeed what your analysis shows. Of course, if this alternative explanation is correct, then we cannot draw any inferences about the size of UHI effects on the average temperature measurements, using the approach taken in your paper.

There are other, smaller, problems with your paper. In particular, the Discussion section states the following.

We observe the opposite of an urban heating effect over the period 1950 to 2010, with a slope of -0.19 ± 0.19 °C/100yr. This is not statistically consistent with prior estimates, but it does verify that the effect is very small….

If the two estimates are not consistent, then they contradict each other. In other words, at least one of them must be wrong. Hence one estimate cannot be used “verify” an inference drawn from the other. This has nothing to do with statistics. It is logic.
Sincerely, Doug
* * * * * * * * * * * *
Douglas J. Keenan
http://www.informath.org

From: Richard Muller
To: James Astill
Cc: Elizabeth Muller
Sent: 17 October 2011, 23:33
Subject: Re: BEST papers

Dear James,

You’ve received a copy of an email that DJ Keenan wrote to me and Charlotte. He raises lots of issues that require addressing, some that reflect misunderstanding, and some of which just reflect disagreements among experts in the field of statistics. Since these issues are bound to arise again and again, we are preparing an FAQ that we will put on our web site.

Keenan states that he had not yet read our long paper on statistical methods. I think if he reads this he is more likely to appreciate the sophistication and care that we took in the analysis. David Brillinger, our chief advisor on statistics, warned us that by avoiding the jargon of statistics, we would mislead statisticians to think we had a naive approach. But we decided to write in a more casual style, specifically to be able to reach the wider world of geophysicists and climate scientists who don’t understand the jargon. Again, if Keenan reads the methods paper, he will have a deeper appreciation of what we have done.

It is also important to recognize that we are not creating a new field of science, but are adding to one that has a long history. In the past I’ve discovered that if you avoid using the methods of the past, the key scientists in the field don’t understand what you have done. As my favorite example, I cite a paper I wrote in which I did data were unevenly spaced in time, so I did a Lomb periodogram; the paper was rejected by referees who argued that I was using an “obscure” approach and should have simply done the traditional interpolation followed by Blackman-Tukey analysis. In the future I did it their way, always being careful however to also do a Lomb analysis to make sure there were no differences.

His initial comment is on the smoothing of data. There are certainly statisticians who vigorously oppose this approach, but there have been top statisticians who support it. Included in that list are David Brillinger, and his mentor, the great John Tukey. Tukey revolutionize the field of data analysis for science and his methods dominate many fields of physical science.

Tukey argued that smoothing was a version of “pre-whitening”, a valuable way to remove from the data behavior that was real but not of primary interest. Another of his methods was sequential analysis, in which the low frequency variations were identified, fit using a maximum likelihood method, and then subtracted from the data using a filter prior to the analysis of the frequencies of interest. He showed that this pre-whitening would lead to a more robust result. This is effectively what we did in the Decadal variations paper. The long time scale changes were not the focus of our study, so we did a maximum-likelihood fit, removed them, and examined the residuals.

Keenan quotes: “If, in a moment of insanity, you do smooth time series data and you do use it as input to other analyses, you dramatically increase the probability of fooling yourself! This is because smoothing induces spurious signals—signals that look real to other analytical methods.” Then he draws a conclusion that does not follow from this quote; he says: “This problem seems to invalidate much of the statistical analysis in your paper.”

He is, of course, being illogical. Just because smoothing can increase the probability of our fooling ourselves doesn’t mean that we did. There is real value to smoothing data, and yes, you have to beware of the traps, but if you are then there is a real advantage to doing that. I wrote about this in detail in my technical book on the subject, “Ice Ages and Astronomical Causes.” Much of this book is devoted to pointing out the traps and pitfalls that others in the field fell into.

Keenan goes on to say, “In statistical analyses, an inference is not drawn directly from data. Rather, a statistical model is fit to the data, and inferences are drawn from the model.” I agree wholeheartedly! He may be confused because we adopted the language of physics and geophysics rather than that of statistics. He goes on to say that “This invalidates much of the analysis done on global warming.” If we are to move ahead, it does no good simply to denigrate most of the previous work. So we do our work with more care, using valid statistical methods, but write our papers in such a way that the prior workers in the field will understand what we say. Our hope, in part, is to advance the methods of the field.

Unfortunately, Keenan’s conclusion is that there has been virtually no valid work in the climate field, that what is needed is a better model, and he does not know what that model should be. He says, “To summarize, most research on global warming relies on a statistical model that should not be used. This invalidates much of the analysis done on global warming. I published an op-ed piece in the Wall Street Journal to explain these issues, in plain English, this year.”

Here is his quote basically concluding that no analysis of global warming is valid under his statistical standards: “Although the AR(1)-based model is known to be inadequate, no one knows what statistical model should be used. There have been various papers in the peer-reviewed literature that suggest possible resolutions, but so far no alternative model has found much acceptance.”

What he is saying is that statistical methods are unable to be used to show that there is global warming or cooling or anything else. That is a very strong conclusion, and it reflects, in my mind, his exaggerated pedantry for statistical methods. He can and will criticize every paper published in the past and the future on the same grounds. We might as well give up in our attempts to evaluate global warming until we find a “model” that Keenan will approve — but he offers no help in doing that.

In fact, a quick survey of his website shows that his list of publications consists almost exclusively of analysis that shows other papers are wrong. I strongly suspect that Keenan would have rejected any model we had used.

He gives some specific complaints. He quotes our paper, where we say, “We observe the opposite of an urban heating effect over the period 1950 to 2010, with a slope of -0.19 ± 0.19 °C/100yr. This is not statistically consistent with prior estimates, but it does verify that the effect is very small….”
He then complains,

If the two estimates are not consistent, then they contradict each other. In other words, at least one of them must be wrong. Hence one estimate cannot be used “verify” an inference drawn from the other. This has nothing to do with statistics. It is logic.

He is misinterpreting our statement. Our conclusion is based on our analysis. We believe it is correct. The fact that it is inconsistent with prior estimates does imply that one is wrong. Of course, we think it is the prior estimates. We do not believe that the prior estimates were more than back-of-the-envelope “guestimates”, and so there is no “statistical” contradiction.

He complains,

Considering the second paper, on Urban Heat Islands, the conclusion there is that there has been some urban cooling. That conclusion contradicts over a century of research as well as common experience. It is almost certainly incorrect. And if such an unexpected conclusion is correct, then every feasible effort should be made to show the reader that it must be correct.

He is drawing a strong a conclusion for an effect that is only significant to one standard deviation! He never would have let us claim that -0.19 ± 0.19 °C/100yr indicates urban cooling. I am surprised that a statistician would argue that such a statistically insignificant effect indicates cooling.

Please be careful whom you share this email with. We are truly interested in winning over the other analysts in the field, and I worry that if they were to read portions of this email out of context that they might interpret it the wrong way.
Rich

From: D.J. Keenan
To: James Astill
Sent: 18 October, 2011 17:53
Subject: Re: BEST papers

James,

On the most crucial point, it seems that Rich and I are in agreement. Here is a quote from his reply.

Keenan goes on to say, “In statistical analyses, an inference is not drawn directly from data. Rather, a statistical model is fit to the data, and inferences are drawn from the model.” I agree wholeheartedly!

And so the question is this: was the statistical model that was adopted for their analysis a reasonable choice? If not, then–since their conclusions are based upon that model–their conclusions must be unfounded.

In fact, the statistical model that they adopted has been rejected by essentially everyone. In particular, it has been rejected by both the IPCC and the CCSP, as cited in my previous message. I know of no work that presents argumentation to support their choice of model: they have just adopted the model without any attempt at justification, which is clearly wrong.

(It has been known for decades that the statistical model that they adopted should not be used. Although the statistical problems with the model were clear, for a long time, no one knew the physical reason. Then in 1976, Klaus Hasselmann published a paper that explained the reason. The paper is famous and has since been cited more than 1000 times.)

We could have a discussion about what statistical model should be adopted. It is certain, though, that the model BEST adopted should be rejected. Ergo, their conclusions are unfounded.

Regarding smoothing, the situation here requires only little statistics to understand. Consider the example given by Matt Briggs at
Do NOT smooth time series before computing forecast skill
We take two series, each entirely random. We compute the correlation of the two series: that will tend to be around 0. Then we smooth each series, and we compute the correlation of the two smoothed series: that will tend to be greater than before. The more we smooth the two series, the greater the correlation. Yet we started out with purely random series. This is not a matter of opinion; it is factual. Yet the BEST work computes the correlation of smoothed series.

The reply uses rhetorical techniques to avoid that, stating “Just because smoothing can increase the probability of our fooling ourselves doesn’t mean that we did”.
The statement is true, but it does not rebut the above point.

Considering the UHI paper, my message included the following.

There are other, smaller, problems with your paper. In particular, the Discussion section states the following.

We observe the opposite of an urban heating effect over the period 1950 to 2010, with a slope of -0.19 ± 0.19 °C/100yr. This is not statistically consistent with prior estimates, but it does verify that the effect is very small….

If the two estimates are not consistent, then they contradict each other. In other words, at least one of them must be wrong. Hence one estimate cannot be used “verify” an inference drawn from the other. This has nothing to do with statistics. It is logic.

The reply claims “The fact that [their paper's conclusion] is inconsistent with prior estimates does imply that one is wrong”. The claim is obviously absurd.

The reply also criticizes me for “drawing a strong a conclusion for an effect that is only significant to one standard deviation”. I did not draw that conclusion, their paper suggested it: saying that the effect was “opposite in sign to that expected if the urban heat island effect was adding anomalous warming” and that “natural explanations might require some recent form of “urban cooling””, and then describing possible causes, such as “For example, if an asphalt surface is replaced by concrete, we might expect the solar absorption to decrease, leading to a net cooling effect”.

Note that the reply does not address the alternative explanation that my message proposed for their UHI results. That explanation, which is based on the analysis of Roy Spencer (cited in my message), implies that we cannot draw any inferences about the size of UHI effects on the average temperature measurements, using the approach taken in their paper.

I has a quick look at their Methods paper. It affects none of my criticisms.

Rich also cites his book on the causes of the ice ages. Kindly read my op-ed piece in the Wall Street Journal, and especially consider the discussion of Figures 6 and 7. His book claims to analyze the data in Figure 6: the book’s purpose is to propose a mechanism to explain why the similarity of the two lines is so weak. In fact, to understand the mechanism, it is only necessary to do a simple subtraction–as my piece explains. In short, the analysis is his book is extraordinarily incompetent–and it takes only an understanding of subtraction to see this.

This person who did the data analysis in that book is the person in charge of data analysis at BEST. The data analysis in the BEST papers would not pass in a third-year undergraduate course in statistical time series.

Lastly, a general comment on the surface temperature records might be appropriate. We have satellite records for the last few decades, and they closely agree with the surface records. We also have good evidence that the world was cooler 100-150 years ago than it is today. Primarily for those reasons, I think that the surface temperature records–from NASA, NOAA, Hadley/CRU, and now BEST–are probably roughly right.

Cheers, Doug

From: James Astill
To: D.J. Keenan
Sent: 18 October 2011, 17:57
Subject: Re: BEST papers

Dear Doug

Many thanks. Are you saying that, though you mistrust the BEST methodology to a great degree, you agree with their most important conclusion, re the surface temperature record?

BEST
James

James Astill
Energy & Environment Editor

From: D.J. Keenan
To: James Astill
Sent: 18 October 2011, 18:41
Subject: Re: BEST papers

James,

Yes, I agree that the BEST surface temperature record is very probably roughly right, at least over the last 120 years or so. This is for the general shape of their curve, not their estimates of uncertainties.

Cheers, Doug

From: D.J. Keenan
To: James Astill
Sent: 20 October, 2011 13:11
Subject: Re: BEST papers

James,

Someone just sent me the BEST press release, and asked for my comments on it. The press release begins with the following statement.

Global warming is real, according to a major study released today. Despite issues raised by climate change skeptics, the Berkeley Earth Surface Temperature study finds reliable evidence of a rise in the average world land temperature of approximately 1°C since the mid-1950s.

The second sentence may be true. The first sentence, however, is not implied by the second sentence, nor does it follow from the analyses in the research papers.

Demonstrating that “global warming is real” requires much more than demonstrating that average world land temperature rose by 1°C since the mid-1950s. As an illustration, the temperature in 2010 was higher than the temperature in 2009, but that on its own does not provide evidence for global warming: the increase in temperatures could obviously be due to random fluctuations. Similarly, the increase in temperatures since the mid 1950s could be due to random fluctuations.

In order to demonstrate that the increase in temperatures since the mid 1950s is not due to random fluctuations, it is necessary to do valid statistical analysis of the temperatures. The BEST team has not done such.

I want to emphasize something. Suppose someone says “2+2=5?. Then it is not merely my opinion that what they have said is wrong; rather, what they have said is wrong. Similarly, it is not merely my opinion that the BEST statistical analysis is seriously invalid; rather, the BEST statistical analysis is seriously invalid.

Cheers, Doug

From: James Astill
To: D.J. Keenan
Sent: 20 October 2011, 13:19
Subject: Re: BEST papers

Dear Doug

Many thanks for all your thoughts on this. It’ll be interesting to see how the BEST papers fare in the review process. Please keep in touch.

BEST

james
James Astill
Energy & Environment Editor

A story about BEST was published in the October 22nd edition of The Economist. The story, authored by James Astill, makes no mention of the above points. It is subheaded “A new analysis of the temperature record leaves little room for the doubters. The world is warming”. Its opening sentence is “For those who question whether global warming is really happening, it is necessary to believe that the instrumental temperature record is wrong”.

Douglas J. Keenan

www.informath.org/apprise/a5700.htm was last updated on 2011-10-21.

wattsupwiththat.com



To: Land Shark who wrote (633111)10/25/2011 1:09:38 PM
From: Brumar892 Recommendations  Respond to of 1578134
 
Anthony Watts on BEST agreements and disagreements so far:

BEST: What I agree with and what I disagree with – plus a call for additional transparency to prevent “pal” review
Posted on October 21, 2011 by Anthony Watts
There’s lots of hay being made by the usual romminesque flaming bloggers, some news outlets and the like, over my disagreement with the way data was handled in one of the Berkeley Earth Surface Temperature (BEST) papers, the only one I got to review before yesterday’s media blitz. Apparently I’m not allowed to point out errors, and BEST isn’t allowed to correct any before release, such as the six incorrectly spelled citations of the Fall et al 2011 paper I pointed out to BEST a week earlier, which they couldn’t be bothered to fix.

And then there’s the issue of doing a 60 year study on siting, when we only guaranteed 30. Even NOAA’s Menne et al paper knew not to make such a stupid mistake. Making up data where there isn’t any is what got Steig et al into trouble in Antarctica and they got called on it by Jeff Id, Steve McIntyre, and Ryan O’Donnell in a follow on peer-reviewed paper.

But I think it’s useful to note here (since I know some other bloggers will just say “denier” and be done with it) what I do in fact agree with and accept, and what I don’t. They wanted an instant answer, before I had a chance even to read the other three papers. Media outlets were asking for my opinion even before the release of these papers, and I stated clearly that I had only seen one and I couldn’t yet comment on the others. That didn’t matter, they lumped that opinion on one I had seen into an opinion on all four.

What I agree with:


  1. The Earth is warmer than it was 100-150 years ago. But that was never in contention - it is a straw man argument. The magnitude and causes are what skeptics question.
  2. From the BEST press release “Global Warming is real” …see point one. Notably, “man-made global warming” was not mentioned by BEST, and in their findings they point out explicitly they didn’t address this issue as they state in this screencap from the press release:
  3. As David Whitehouse wrote: “The researchers find a strong correlation between North Atlantic temperature cycles lasting decades, and the global land surface temperature. They admit that the influence in recent decades of oceanic temperature cycles has been unappreciated and may explain most, if not all, of the global warming that has taken place, stating the possibility that the “human component of global warming may be somewhat overstated.”. Here’s a screencap from that paper:





4. The unique BEST methodology has promise. The scalpel method used to deal with station discontinuity was a good idea and I’ve said so before.

5. The findings of the BEST global surface analysis match the finding of other global temperature metrics. This isn’t surprising, as much of the same base raw data was used. There’s a myth that NASA GISS, HadCRUT, NOAA’s, and now Berkeley’s source data are independent of one another. That’s not completely true. They share a lot of common data from GHCN, administered by NOAA’s National Climatic Data. So it isn’t surprising at all they would match.



What I disagree with:

1. The way they dealt with my surfacestation data in analysis was flat-out wrong, and I told them so days ahead of this release. They offered no correction, nor even an acknowledgement of the issue. The issue has to do with the 60 year period they used. Both peer-reviewed papers on the subject, Menne et al 2010, and Fall et al 2011 used 30 year periods. This is a key point because nobody knows (not me, not NOAA, not BEST) what the siting quality of weather stations was 30-60 years ago. Basically they did an analysis on a time period for which metadata doesn’t exist. I’ve asked simply for them to do it on 30 years as the two peer reviewed papers did, an apples-to-apples comparison. If they do that and the result is the same, I’m satisfied. OTOH, they may find something new when done correctly, we all deserve that opportunity.

Willis Eschenbach points out this quote from the paper:

We evaluate the effect of very-rural station siting on the global average by applying the Berkeley Earth Surface Temperature averaging procedure to the very-rural stations. By comparing the resulting average to that obtained by using all the stations we can quantify the impact of selecting sites not subject to urbanization on the estimated average land temperature.


He adds: That seems crazy to me. Why compare the worst stations to all stations? Why not compare them to the BEST stations?

2. The UHI study seems a bit strange in its approach. They write in their press release that:



They didn’t adequately deal with that 1% in my opinion, by doing a proper area weighting. And what percentage of weather stations were in that 1%? While they do have some evidence of the use of a “kriging” technique, I’m not certain is has been done properly. The fact that 33% of the sites show a cooling is certainly cause for a much harder look at this. That’s not something you can easily dismiss, though they attempt to. This will hopefully get sorted out in peer review.

3. The release method they chose, of having a media blitzkrieg of press release and writers at major MSM outlets lined up beforehand is beyond the pale. While I agree with Dr. Muller’s contention that circulating papers among colleagues for wider peer review is an excellent idea, what they did with the planned and coordinated (and make no mistake it was coordinated for October 20th, Liz Muller told me this herself) is not only self-serving grandiosity, but quite risky if peer review comes up with a different answer.

The rush to judgment they fomented before science had a chance to speak is worse than anything I’ve ever seen, and from my early dealings with them, I can say that I had no idea they would do this, otherwise I would not have embraced them so openly. A lie of omission is still a lie, and I feel that I was not given the true intentions of the BEST group when I met with them.

So there you have it, I accept their papers, and many of their findings, but disagree with some methods and results as is my right. It will be interesting to see if these survive peer review significantly unchanged.

One thing we can count on that WON’T normally be transparent is the peer review process, and if that process includes members of the “team” who are well versed enough to but already embracing the results such as Phil Jones has done, then the peer review will turn into “pal review”.

The solution is to make the names of the reviewers known. Since Dr. Muller and BEST wish to upset the apple cart of scientific procedure, putting public review before peer review, and because they make this self-assured and most extraordinary claim in their press release:



That’s some claim. Four papers that have not been peer-reviewed yet, and they KNOW they’ll pass peer review and will be in the next IPCC report? Is it just me or does that sound rigged? Or, is it just the product of an overactive ego on the part of the BEST group?

I say, if BEST and Dr. Muller truly believes in a transparent approach, as they state on the front page of their website…



…let’s make the peer review process transparent so that there is no possibility of “pal review” to ramrod this through without proper science being done.

Since Dr. Muller claims this is “one of the most important questions ever”, let’s deal with it in an open a manner as possible. Ensuring that these four papers get a thorough and non-partisan peer review is the BEST way to get the question answered.

Had they not made the claim I highlighted above of it passing peer review and being in the next IPCC report before any of that even is decided, I would never think to ask for this. That overconfident claim is a real cause for concern, especially when the media blitzkrieg they launched makes it difficult for any potential review scientists to not notice and read these studies and news stories ahead of time, thus becoming biased by media coverage.

We can’t just move the “jury pool” of scientists to the next county to ensure a fair trial now that is been blathered worldwide can we?

Vote on it:

Should BEST call for an open named peer review process to prevent the possibility of "pal" review?


Yes
No

Vote on it:

Thank you for voting!

Yes 95.26% (1,146 votes)
No 4.74% (57 votes)

Total Votes: 1,203

wattsupwiththat.com




  • To: Land Shark who wrote (633111)10/25/2011 1:10:21 PM
    From: Brumar891 Recommendation  Respond to of 1578134
     
    Steve McIntyre's first thoughts on BEST:

    First Thoughts on BEST

    Oct 22, 2011 – 9:34 AM
    Rich Muller sent me the BEST papers about 10 days ago so that I would have an opportunity to look at them prior to their public release. Unfortunately, I’ve been very busy on other matters in the past week and wasn’t able to get to it right away and still haven’t had an opportunity to digest the methods paper. (Nor will I for a week or two.)

    As a disclaimer, Rich Muller is one of the few people in this field who I regard as a friend. In 2004, he wrote an article for MIT Review that drew attention to our work in a favorable way. I got in touch with him at the time and he was very encouraging to me. I started attending AGU at his suggestion and we’ve kept in touch from time to time ever since. While people can legitimately describe Phil Jones as not being a “nuclear physicist”, the same comment cannot be made of Rich Muller in either sense of the turn of phrase.

    The Value of Independent Analysis
    The purpose of audits in business is not to overturn the accounts prepared by management, but to provide reassurance to the public. 99% of all audits support management accounts. I’ve never contested the idea that it is warmer now than in the 19th century. If nothing else, the recession of glaciers provides plenty of evidence of warming in the last century.

    Nonetheless, it is easy to dislike the craftsmanship of the major indices (GISS, CRU and NOAA) and the underlying GHCN and USHCN datasets. GISS, for example, purports to adjust for UHI through a “two legged adjustment” that seems entirely ad hoc and which yields counterintuitive adjustments in most areas of the world other than the US. GISS methodology also unfortunately rewrites its entire history whenever it is updated. CRU notoriously failed to retain its original separate data sets, merging different stations (ostensibly due to lack of “storage” space, though file cabinets have long provided a low-technology method of data storage. GHCN seems to have stopped collecting many stations in the early 1990s for no good reason (the “great dying of thermometers”) though the dead thermometers can be readily located on the internet.

    Even small changes in station history can introduce discontinuities. Over the years, USHCN has introduced a series of adjustments for metadata changes (changes in observation times, instrumentation), all of which have had the effect of increasing trends. Even in the US where metadata is good, the record is still plagued by undocumented discontinuities. As a result, USHCN recently introduced a new method that supposedly adjusts for these discontinuities. But this new method has not been subjected to thorough scrutiny by external statisticians.

    The US has attempted to maintain a network of “rural” sites, but, as Anthony Watts and his volunteers have documented, these stations all too often do not adhere to formal standards of station quality.

    The degree to which increased UHI has contributed to observed trends has been a longstanding dispute. UHI is an effect that can be observed by a high school student. As originally formulated by Oke, UHI was postulated to be more or less a function of log(population) and to affect villages and towns as well as large cities. Given the location of a large proportion of stations in urban/town settings, Hansen, for example, has taken the position that an adjustment for UHI was necessary while Jones has argued that it isn’t.

    Unlike the statistical agencies that maintain other important indices (e.g. Consumer Price Index), the leaders of the temperature units (Hansen, Jones, Karl) have taken strong personal positions on anthropogenic global warming. These strong advocacy and even activist positions are a conflict of interest that has done much to deter acceptance of these indices by critics.

    This has been exacerbated by CRU’s refusal to disclose station data to critics, while readily providing the same information to fellow travellers, a refusal. Nonetheless, as I reminded CA readers during CRU’s refusal of even FOI requests, just because they were acting like jerks, didn’t mean that the indices themselves were in major error. Donna Laframboise’s “spoiled child” metaphor is apt.

    The entry of the BEST team into this milieu is therefore welcome on any number of counts. An independent re-examination of the temperature record is welcome and long overdue, particularly when they have ensured that their team included not only qualified statistical competence, but eminent (Brillinger).

    Homogeneity
    They introduced a new method to achieve homogeneity. I have not examined this method or this paper and have no comment on it.

    Kriging
    A commenter at Judy Curry’s rather sarcastically observed that, with my experience in mineral exploration, I would undoubtedly endorse their use of kriging, a technique used in mineral exploration to interpolate ore grades between drill holes.

    His surmise is correct.

    Indeed, the analogies between interpolating ore grades between drill holes and spatial interpolation of temperatures/ temperature trends has been quite evident to me since I first started looking at climate data.

    Kriging is a technique that exists in conventional statistics. While I haven’t had an opportunity to examine the details of the BEST implementation, in principle, it seems far more logical to interpolate through kriging rather than through principal components or RegEM (TTLS).

    Dark Areas of the Map
    In the 19th century, availability of station data is much reduced. CRU methodology, for example, does not take station data outside the gridcell and thus leaves large portions of the globe dark throughout the 19th century.

    BEST takes a different approach. They use available data to estimate temperatures in dark grid cells while substantially increasing the error bars of the estimates. These estimates have been roundly condemned by some commenters on threads at Judy Curry’s and Anthony Watts’.

    After thinking about it a little, I think that BEST’s approach on this is more logical and that this is an important and worthwhile contribution to the field. The “dark” parts of the globe did have temperatures in the 19th century and ignoring them may impart a bias. While I haven’t examined the details of their kriging, my first instinct is in favor of the approach.

    The Early Nineteenth Century
    A second major innovation by BEST has been to commence their temperature estimates at the start of the 19th century, rather than CRU’s 1850/1854 or GISS’s 1880. They recognize the increased sparsity of station data with widely expanded error bars. Again, the freshness of their perspective is helpful here.(They also run noticeably cooler than CRU between 1850 and 1880.) Here is their present estimate:



    The differences between BEST and CRU have an important potential knock-on impact in the world of proxy reconstructions – an area of technical interest for me. “Justification” of proxy reconstructions in Mannian style relies heavily on RE statistics in the 19th century based on CRU data. My guess is that the reconstructions have been consciously or subconsciously adapted to CRU and that RE statistics calculated with BEST will deteriorate and perhaps a lot. For now, that’s just a dig-here.

    It’s also intriguing that BEST’s early 19th century is as cold as it is.

    BEST’s estimate of the size of the temperature increase since the start of the 19th century is much larger than previous estimates
    . (Note- I’ll update this with an example.)

    The decade of the 1810s is shown in their estimates as being nearly 2 degrees colder than the present. Yes, this was a short interval and yes, the error bars are large. The first half of the 19th century is about 1.5 degrees colder than at present.

    At first blush, these are very dramatic changes in perspective and, if sustained, may result in some major reinterpretations. Whereas Jones, Bradley and others attempted to argue the non-existence of the Little Ice Age, BEST results point to the Little Ice Age being colder and perhaps substantially colder than “previously thought”.


    [ Mann's, Bradley's, Hughes' 1000 year hockey stick is really dead. Don't let Rat read this. ]


    It’s also interesting to interpret these results from the context of “dangerous climate change”, defined by the UN as 2 deg C. Under BEST’s calculations, we’ve already experienced nearly 2 deg C of climate change since the early 19th century. While the authors of WG2 tell us that this experience has been entirely adverse, if not near catastrophic, it seems to me that we have, as a species, not only managed to cope with these apparently very large changes, but arguably even flourished in the last century. This is not to say that we would do equally well if faced with another 2 deg C. Only that if BEST estimates are correct, the prior 2 degrees do not appear to have been “dangerous” climate change.

    Comparison to SST
    They do not compare their land results to SST results. These two data sets have been said to be “independent” and mutually reinforcing, but I, for one, have had concerns that the results are not truly independent and that, for example, the SST bucket adjustments have, to some extent, been tailored, either consciously or subconsciously, so that the SST data cohere with the land data.

    Here is a plot showing HadSST overlaid onto the Berkeley graphic. In the very early portion, the shape of the Berkeley series coheres a little better to HadSST than CRUTem. Since about 1980, there has been a marked divergence between HadSST and the land indices. This is even more marked with the Berkely series than with CRUTem.



    Station Quality
    I have looked at some details of the Station Quality paper using a spreadsheet of station classification sent to me by Anthony in August 2011 and cannot replicate their results at all. BEST reported a trend of 0.388 deg C/decade from “bad” stations (CRN 4/5) and 0.309 deg C/decade from “good” stations” (CRN1/2). Using my archive of USHCN raw data (saved prior to their recent adjustments), I got much lower trends, with trends at good stations being lower than at bad stations in a coarse average. The station counts for good and bad stations don’t match to the information provided to me.

    Watts Count

    Rohde Count

    Rohde Trend

    Trend(1950)

    Trend (1979)

    506

    705

    0.388

    0.16

    0.27

    185

    88

    0.509

    0.15

    0.22

    As was observed in very early commentary on surface stations results, there is a strong gradient in US temperature trends (more negative in the southeast US and more positive in the west). The location of good and bad stations is not spatially random, so some care has to be taken in stratification.

    In my own quick exercises on the topic, I’ve experimented with a random effects model, allowing a grid cell effect. I’ve also experimented with further stratification for rural-urban (using the coarse USHCN classification) and for instrumentation.

    On this basis, for post-1979 trends, “rural bad” had a trend 0.08 deg C/decade greater than “rural good”; “small town bad” was 0.07 deg C decade greater than “small-town good” and “urban bad” was the opposite sign to “urban good” : 0.01 deg C/decade cooler.

    Stratifying by measurement type, “CRS bad” was 0.05 deg C/dec warmer than “CRS good” while “MMTS bad” was 0.15 deg C warmer than “MMTS good”.

    Combining both stratifications, “MMTS rural good” had a post-1979 trend of 0.11 deg C/decade while “CRS urban bad” had a corresponding trend of 0.42 deg C/decade.

    Details of the BEST calculation on these points are not yet available, though they’ve made a commendable effort to be transparent and I’m sure that this lacuna will be remedied. I’ve placed my script for the above results online here. (The script is not turnkey as it relies on a spreadsheet on station quality that has not been released yet, but the script shows the structure of the analysis.)

    Conclusion
    As some readers have noticed, I was interviewed by Nature and New Scientist for their reports on BEST. In each case, perhaps unsurprisingly, the reporters chose to emphasize criticisms. For example, my nuanced criticism of the analysis of the effect of station quality was broadened by one reporter into a sweeping claim about overall replicability that I didn’t make.

    Whatever the outcome of the BEST analysis, they have brought welcome and fresh ideas to a topic which, despite its importance, has had virtually no intellectual investment in the past 25 years. I am particularly interested in their 19th century conclusions.

    http://climateaudit.org/2011/10/22/first-thoughts-on-best/



    To: Land Shark who wrote (633111)10/25/2011 1:11:44 PM
    From: Brumar891 Recommendation  Respond to of 1578134
     
    How much of the warming is due to humans and what will be the likely effects? We made no independent assessment of that.

    I wish to point out the article you posted included the above admission .... belatedly, but it's there and its VERY important.



    To: Land Shark who wrote (633111)10/25/2011 1:44:34 PM
    From: joseffy  Respond to of 1578134
     
    The idiot Landshirk quotes is from Berkeley.

    LOL



    To: Land Shark who wrote (633111)10/25/2011 2:26:18 PM
    From: Brumar891 Recommendation  Read Replies (1) | Respond to of 1578134
     
    Ministry of Truth rewrites history:

    In 1975, NCAR generated this graph of global cooling. Temperatures plummeted from 1945 to at least 1970.





    In 2011, Richard Muller published this graph, showing that it never happened.



    Below is an overlay at the same scale. The cooling after 1950 has disappeared. Winston Smith would be proud!



    real-science.com



    To: Land Shark who wrote (633111)10/30/2011 4:29:11 PM
    From: Brumar892 Recommendations  Read Replies (1) | Respond to of 1578134
     
    Curry drops bomb on Muller. She's the second lead author for BEST:

    Uh oh: It was the BEST of times, it was the worst of times
    Posted on October 29, 2011 by Anthony Watts

    Alternate title: Something wonky this way comes

    I try to get away to work on my paper and the climate world explodes, pulling me back in. Strange things are happening related to the BEST data and co-authors Richard Muller and Judith Curry. Implosion might be a good word.

    Popcorn futures are soaring. BEST Co-author Judith Curry drops a bombshell:

    Her comments, in an exclusive interview with The Mail on Sunday, seem certain to ignite a furious academic row. She said this affair had to be compared to the notorious ‘Climategate’ scandal two years ago.

    Here’s the short timeline.

    1. The GWPF plots a flat 10 year graph using BEST data:



    2. The Mail on Sunday runs a scathing article comparing BEST’s data plotted by GWPF and the data presented in papers. They print this comparison graph:



    Note: timescales don’t match on graphs above, 200 years/10 years. A bit naughty on the part of the Sunday Mail to put them together as many readers won’t notice.

    3. Dr. Judith Curry, BEST co-author, turns on Muller, in the Mail on Sunday article citing “hide the decline”:

    In Prof Curry’s view, two of the papers were not ready to be published, in part because they did not properly address the arguments of climate sceptics.

    As for the graph disseminated to the media, she said: ‘This is “hide the decline” stuff. Our data show the pause, just as the other sets of data do. Muller is hiding the decline.

    ‘To say this is the end of scepticism is misleading, as is the statement that warming hasn’t paused. It is also misleading to say, as he has, that the issue of heat islands has been settled.’

    Prof Muller said she was ‘out of the loop’. He added: ‘I wasn’t even sent the press release before it was issued.’



    But although Prof Curry is the second named author of all four papers, Prof Muller failed to consult her before deciding to put them on the internet earlier this month, when the peer review process had barely started, and to issue a detailed press release at the same time.

    He also briefed selected journalists individually. ‘It is not how I would have played it,’ Prof Curry said. ‘I was informed only when I got a group email. I think they have made errors and I distance myself from what they did.

    ‘It would have been smart to consult me.’ She said it was unfortunate that although the Journal of Geophysical Research had allowed Prof Muller to issue the papers, the reviewers were, under the journal’s policy, forbidden from public comment.

    4. Ross McKittrick unloads:

    Prof McKittrick added: ‘The fact is that many of the people who are in a position to provide informed criticism of this work are currently bound by confidentiality agreements.

    ‘For the Berkeley team to have chosen this particular moment to launch a major international publicity blitz is a highly unethical sabotage of the peer review process.’

    5. According to BEST’s own data, Los Angeles is cooling, fast:

    BEST data graph from Steve McIntyre

    Steve McIntyre emailed me the graph above tonight as part of a larger discussion.

    But compare it to the GISS station record, and you get a whole different story:



    Overlay: Combined to fit scale and time:



    [ Wow. That's a great illustration of how GISS has been cooking their temperature records. What's it going to take to get James Hansen fired? }

    Here’s the GWPF article:

    Best Confirms Global Temperature Standstill

    Saturday, 29 October 2011 23:55 Dr. David Whitehouse

    Contrary to claims being made by the leader of the Best global temperature initiative their data confirms, and places on a firmer statistical basis, the global temperature standstill of the past ten years as seen by other groups.

    Many people have now had some time to read the papers issued in preprint form from the Best project. My strong impression is that they are mostly poorly written, badly argued and at this stage unfit for submission to a major journal. Whilst I have made some comments about Best’s PR and data release strategy, I want to now look at some aspects of the data.

    When asked by the BBC’s Today programme Professor Richard Muller, leader of the initiative, said that the global temperature standstill of the past decade was not present in their data.

    “In our data, which is only on the land we see no evidence of it having slowed down. Now the evidence which shows that it has been stopped is a combination of land and ocean data. The oceans do not heat as much as the land because it absorbs more of the heat and when the data are combined with the land data then the other groups have shown that when it does seem to be leveling off. We have not seen that in the land data.”

    My first though would be that it would be remarkable if it was. The global temperature standstill of the past decade is obvious in HadCrut3 data which is a combination of land and sea surface data. Best is only land data from nearly 40,000 weather stations. Professor Muller says they “really get a good coverage of the globe.” The land is expected to have a fast response to the warming of the lower atmosphere caused by greenhouse gas forcing, unlike the oceans with their high thermal capacity and their decadal timescales for heating and cooling, though not forgetting the ENSO and la Nina.

    Fig 1 shows the past ten years plotted from the monthly data from Best’s archives. Click on the image to enlarge.



    It is a statistically perfect straight line of zero gradient. Indeed, most of the largest variations in it can be attributed to ENSO and la Nina effects. It is impossible to reconcile this with Professor Muller’s statement. Could it really be the case that Professor Muller has not looked at the data in an appropriate way to see the last ten years clearly?

    Indeed Best seems to have worked hard to obscure it. They present data covering more almost 200 years is presented with a short x-axis and a stretched y-axis to accentuate the increase. The data is then smoothed using a ten year average which is ideally suited to removing the past five years of the past decade and mix the earlier standstill years with years when there was an increase. This is an ideal formula for suppressing the past decade’s data.

    When examined more objectively Best data confirms the global temperature standstill of the past decade. That the standstill should be present in land only data is remarkable. There have been standstills in land temperature before, but the significance of the past decade is that it is in the era of mankind’s postulated influence on climate through greenhouse gas forcing. Predictions made many times in the past few years suggest that warming should be the strongest and fastest in the land data.

    Only a few years ago many scientists and commentators would not acknowledge the global temperature standstill of the past decade. Now that it has become unarguable there has emerged more explanations for it than can possibly be the case.

    To explain the combined sea-land temperature hiatus some have suggested that the oceans are sucking up the heat, as professor Muller outlines in his radio interview. This explanation is strained in my view if the land temperature stays constant. Could we really have the very special situation whereby the oceans sequester just enough heat at just the right time to keep the land temperature flat? Aerosols, postulated by some to be coming from China, don’t provide an explanation for the land temperature hiatus either. In fact, the constant land temperature puts a strain on all of the explanations offered for why the land-sea combination hasn’t warmed in the past decade or so.

    We make a big deal of the temperature going up. In my view we should make a bigger scientific deal about temperature flatlining for a decade or more in the face of rising CO2 levels. If further scrutiny of the Best dataset confirms this finding we will have new questions about the nature and balance of oceanic and land warming.

    The fact that Best confirms the global temperature hiatus and shows that it is apparent in land only data is significant, and in my view its major scientific finding, so far. It is puzzling that they missed it.

    ============================================================

    UPDATE: 10/30/2011 7AM PST Judith Curry says the “climategate” comparison was indirectly attributed to her, she gives her take on the story: http://judithcurry.com/2011/10/30/mail-on-best/

    She does reiterate that: ‘I agree that the way the data is presented in the graph “hides the decline.” and adds, “I thought the project was a great idea, and I still do, but it currently has a tarnish on it. Lets see what we can do about this.”

    Jeff Id has written a critique of the data processing algorithms:

    Overconfidence Error in BEST

    He writes in WUWT comments: “To be clear, I believe I have identified a specific mathematical error which will require a re-write of the CI portion of the methods paper.”

    ...http://wattsupwiththat.com/2011/10/29/uh-oh-it-was-the-best-of-times-it-was-the-worst-of-times/#more-50286

    .....



    Al Gored says:

    October 29, 2011 at 10:12 pm

    “to have chosen this particular moment to launch a major international publicity blitz is a highly unethical sabotage of the peer review process.’

    With the best informed critics gagged, “unethical”barely describes it.

    Who would buy a used car from such people, let alone a used climate theory?

    Judith Curry is a hero.

    David Ball says:

    October 29, 2011 at 10:17 pm

    Must have been a large window in Mullers office for all that credibility to fly out at once, ……..
    Malice or incompetence, neither is acceptable.

    .....

    Rosco says:

    October 29, 2011 at 10:22 pm

    The fact that the second lead author claims the lead author is – well – fabricating the results – tends to destroy the consensus.

    Plus 0ne for Judith.

    ....