In the 2007 analysis of the GISS dataset, Detroit Lakes was used as a test case. (See prior posts on this station here). I’ve revisited it in the BEST data set, comparing it to the older USHCN data that I have on hand from a few years ago.
First, here is a simple plot of USHCN raw and BEST versions. The BEST version is neither an anomaly series (like CRU) nor a temperature series (like USHCN). It is described as “seasonally adjusted”. The mechanism for seasonal adjustment is not described in the covering article. I presume that it’s somewhere in the archived code. The overall mean temperature for USHCN raw and Berkeley are very close. The data availability matches in this case – same starting point and same gaps (at a quick look). So no infilling thus far.
Figure 1. Simple plot of USHCN Raw and BEST versions of Detroit Lakes
The Berkeley series is not, however, the overall average plus an anomaly as one might have guessed. Here is a barplot comparing monthly means of the two versions. While the Berkeley version obviously has much less variation than the observations, it isn’t constant either (as it would be if it were overall average plus monthly anomaly). I can’t figure out so far where the Berkeley monthly normals come from.
Figure 2. Monthly Averages of two versions.
If one does a simple scatter plot of USHCN raw vs Berkeley, one gets a set of 12 straight lines with near identical slope, one line for each month:
Figure 3. Scatter plot of USHCN raw vs Berkeley
I then tried the following. I subtracted the Berkeley monthly average from each Berkeley data point and added back the USHCN monthly average. This yielded the following:
Figure 4. USHCN raw versus Berkeley (renormalized for each month)
The Berkeley data seems to be virtually identical to USHCN raw data less monthly normals that are different from normals of USCHN raw data plus annual average. The implied monthly averages in the BEST normalized data are shown below. The range of difference is from -2.27 to 1.41 deg C.
My original examination of Detroit Lakes and other stations was directed at whether NASA GISS had software to detect changes – a point that had been then been raised in internet debates by Josh Halpern as a rebuttal to the nascent surface stations project. I used Detroit Lakes as one of a number of type cases to examine this, accidentally observing the Y2K discontinuity. One corollary was that GISS software did not, after all, have the capability of detecting the injected Y2K discontinuity.
It would be interesting to test the BEST algorithm against the dataset with the Y2K discontinuity to see if they can pick it up with their present methodology. At first blush, it looks as though USHCN data is used pretty much as is, other than the curious monthly normals.
[Update: it looks like this data is prior to homogenization.]
13 Comments
Steve,
As near as I can tell, the BEST seasonal normalization contains a 4 month period oscillation. Third harmonic ringing in their low pass filter?
Steve: no idea. I’m starting with their results and observing properties.
For anyone who has trouble on Windows…
Lines that look like this — change the quotes…
load(“d:/climate/data/berkeley/details.tab”)
To this:
load(“d:/climate/data/berkeley/details.tab”)
Straight Upsy-downsy Quotes…
There are quite a few to change — in the plot lines (which are thickening — sorry couldn’t resist…)
I also moved the “details,tab” to the directory you referenced
load(“d:/climate/data/station/ushcn/details.tab”)
Presumably it was the same “details.tab” we downloaded previously…
At least I duplicated your graphs… fwiw So I am guessing I was correct…
Next?
And now I see it is WordPress –FIXING the quotes… argghhh!
Steve: There’s a command in wordpress to block off the text. Pete holzman knows the command. I’ll try to locate it.
You could try <pre> </pre> … not sure how that will come through … but here’s a test (works in a post, so, at least in theory, it should work in a comment:
Steve: Thanks, Hilary. That’s what I wanted.
You’re most welcome, Steve. Your usage of this Helpful Hint from Hilary™ gives me confidence to grant myself a brownie point for my (very minor) contribution to the advancement of understanding BEST 😉
I don’t have the USHCN data (yet, here) as I just started with BEST. The BEST data contains a record for #samples in each monthly record at each station. Generally these seem to be daily – e.g. the one for the Detroit Lakes station above averages just over 29 samples per record when the field isn’t null (-99). Is this typical of the underlying dataset (daily records)?
Steve – I have an old but relevant USCHN collation online at http://www.climateaudit.info/data/station/ushcn/
thx got it
Steve, have they released the raw data (as opposed to the “seasonally adjusted” data) yet? I find it frustrating to use data that’s been pre-munged …
w.
Steve: there ‘s a very large file of original data. It was too large for my computer to read. Much of it will probably be the same as GHCN.
Kinda sorta related to the “too large for my computer…” – Oracle is now supporting R in the database – I’m hoping to get some time on an Exadata to push it… 😀
info – http://www.oracle.com/technetwork/database/options/odm/oracle-r-enterprise-oow11-517498.pdf
What is the data (Steve supplied) from BEST actually describing? I looked at my home town (the years were right).
The first 4 months:
BEST: 8.892 8.109 8.241 7.421
Environment Canada Mean: 2.6 2.3 4.5 6.4
Environment Canada Max: 6.5 6.8 10.1 11.6
Steve,
If you don’t mind a little amateurish work, I done a quick writeup of my examination of BEST data for my town. I was careful to use only the simplest manipulation of data so as to preserve as much of its original content as practical. I am not impressed with the quality of the work from BEST.
Click to access BEST_data_for_Lebanon_Missouri.pdf
Gary Wescom
For those who want to follow the R Discussions but need a quick course or refresher… I am going to suggest the Andrew Robinson documents as they seem to bring you along fairly rapidly…
Just click on this link:
http://www.ms.unimelb.edu.au/~andrewpr/
The go to the R-Users group directory to download his notes and data. There are a few quirks to entwrig the listings — but that’s all..
Start a new edit window, then just cut and paste from the book and remove the chevrons “>” and the Pluses “+” from the beginning of each line when you paste — that should do it.. The “Select and Execute”…
You can continue to add to a script and just execute the new part using the “Edit” menu — “Run Line or Selection”..
Then save the script at appropriate points and carry on with the lessons.
2 Trackbacks
[…] release was not even the raw data. It was processed by removing the monthly averages … but we don’t know what those averages were, or how they were […]
[…] Source: https://climateaudit.org/2011/10/29/detroit-lakes-in-best/ […]