Non-centring in the Forest 2006 study

This is a cautionary tale, about a mystery that had an unexpected explanation. It’s not intended as a criticism of the scientists involved, and the problem involved, although potentially serious, actually had little impact on the results of the study concerned. However, I am hopeful that mathematically and computing orientated readers will find it of interest. But first I need to give some background information.

Forest et al. 2006 (F06), here, was a high profile observationally-constrained Bayesian study that estimated equilibrium climate sensitivity (Seq) simultaneously with two other key climate system parameters / properties, ocean effective vertical diffusivity (Kv) and aerosol forcing (Faer). Both F06 and its predecessor Forest 2002 had their climate sensitivity PDFs featured in Figure 9.20 of the AR4 WG1 report. I started investigating F06 in 2011, with a view to using its data to derive estimated climate system parameter PDFs using an objective Bayesian method.  That work eventually led to my paper ‘An objective Bayesian, improved approach for applying optimal fingerprint techniques to estimate climate sensitivity’, recently published in Early online release form by Journal of Climate, here.

Readers may recall that I found some basic statistical errors in the F06 code, about which I wrote a detailed article at Climate Audit, here. But those errors could not have affected any unrelated studies.  In this post, I want to focus on an error I have discovered in F06 that is perhaps of wider interest. The error has a source that will probably be familiar to CA readers – failure to check that data is zero mean before undertaking principal components analysis (PCA) / singular value decomposition (SVD).

The Forest 2006 method
First, a recap of how F06 works. It uses three ‘diagnostics’  (groups of variables whose observed values are compared to model-simulations): surface temperature averages from four latitude zones for each of the five decades comprised in 1946–1995; deep ocean 0–3000 m global temperature trend over 1957–1993; and upper air temperature changes from 1961–80 to 1986–95 at eight pressure levels for each 5-degree latitude band (8 bands being without data). AOGCM unforced long control run data is used to estimate natural, internal variability in the diagnostic variables. The MIT 2D climate model, which has adjustable parameters calibrated in terms of  Seq , Kv and Faer, was run several hundred times at different settings of those parameters, producing sets of model-simulated temperature changes on a coarse, incomplete, grid of the three climate system parameters.

A standard optimal fingerprint method, as used in most detection and attribution (D&A) studies to deduce anthropogenic influence on the climate, is employed in F06. The differences between changes in model-simulated and observed temperatures are ‘whitened’, with the intention of making them independent and all of unit variance. Then an error sum-of-squares, r2, is calculated from the whitened diagnostic variable differences and a likelihood function is computed from r2, on the basis of an appropriate F-distribution. The idea is that, the lower r2 is, the higher the likelihood that the model settings of Seq, Kv and Faer correspond to their true values. Either the values of the model-simulated diagnostic variables are first interpolated to a fine regular 3D grid (my approach) or the r2 values are so interpolated (the F06 approach). A joint posterior PDF for Seq, Kv and Faer is then computed, using Bayes’ rule, from the multiplicatively-combined values of the likelihoods from all three diagnostics and a prior distribution for the parameters. Finally, a marginal posterior PDF is computed for each parameter by integrating out (averaging over) the other two parameters.

The whitening process involves a truncated inversion of the estimated (sample) control-run data covariance matrix. First an eigendecomposition of that covariance matrix is performed. A regularized inverse transpose square root of that matrix is obtained as the product of the eigenvectors and the reciprocal square roots of the corresponding eigenvalues, only the first k eigenvector–eigenvalue pairs being used. The raw model-simulation – observation differences are then multiplied by that covariance matrix inverse square root to give the whitened differences. As many readers will know, by setting the number of retained eigenfunctions or EOFs (eigenvector-patterns), k affects how much detail is retained upon the inversion of the covariance matrix. The higher the truncation parameter k, the more detail is retained and the better is discrimination between differing values of Seq, Kv and Faer. However, if k is too high then the likelihood values will be heavily affected by noise and potentially very unrealistic. There is a standard test, detailed in Allen and Tett 1999 (AT99), here, that can be used to guard against k being too high.

The Forest 2006 upper air diagnostic: effect of EOF truncation and mass weighting choices
My concern in this article is with the F06 upper air (ua) diagnostic. F06 weighted the upper air diagnostic variables by the mass of air attributable to each, which is proportional to the cosine of its mean latitude multiplied by the pressure band allocated to the relevant pressure level. The weighting used only affects the EOF patterns; without EOF truncation it would have no effect. It seems reasonable that each pressure level’s pressure band should be treated as extending halfway towards the adjacent pressure levels, and to surface pressure (~1000 hPa) at the bottom. But where to treat the top end of the pressure band attributable to the highest, 50 hPa pressure level, as being is less clear. One choice is halfway towards the top of the atmosphere (0 hPa). Another is halfway towards 30 hPa, on the grounds that data for the 30 hPa level exists – although it was excluded from the main observational dataset due to excessive missing data. The F06 weighting was halfway towards 30 hPa. The weighting difference is minor: 4.0% for the 50 hPa layer on the F06 weighting, 5.6% on the alternative weighting.

However, it turns out that the fourteenth eigenvector – and hence the shape of the likelihood surface, given the F06 choice of kua = 14 –  is highly sensitive to which of these two mass-weighting schemes is applied to the diagnostic variables, as is the result from the AT99 test.  Whatever the physical merits of the two 50 hPa weighting bases, the F06 choice appears to be an inferior one from the point of view of stability of inference. It results, at kua = 14, in failure of  the recommended, stricter, version of the AT99 test, and a likelihood surface that is completely different from that when kua = 14 and the alternative weighting choice is made (which well satisfies the AT99 test). Moreover, if  kua is reduced to 12 then whichever weighting choice is made the AT99 test is satisfied and the likelihood surface is similar to that at kua = 14 when using the higher alternative, non-F06, 50 hPa level weighting.

AT99 test results
The below graph plots the AT99 test values – the ratio of the number of degrees of freedom in the fit to the r2 value at the best fit point, r2min, for the two 50 hPa weightings. To be cautious, the value should lie within the area bounded by the inner, dotted black lines (which are the 5% and 95% points of the relevant chi-squared distribution). The nearer it is to the unity, the better the statistical model is satisfied.

Using the F06 50 hPa level weighting, at kua = 13 the AT99 test is satisfied (although less well than at kua = 12), and the likelihood surface is more similar to that – using the same weighting – at kua = 12 than to that at kua = 14.

Upper air diagnostic likelihood surfaces
The following plots show what the upper air diagnostic likelihood consistency surface looks like in {Seq, Kv} space using the F06 50 hPa level weighting, at successively kua = 12, kua = 13 and kua = 14. Faer  has been integrated out, weighted by its marginal PDF as inferred from all diagnostics. The surface is for the CDF, not the PDF. It shows how probable it is that the upper air diagnostic r2 value at each {Seq, Kv} point could have arisen by chance, given the estimated noise covariance matrix. Note that the orientation of the axes is non-standard.

Notice that the combination of high climate sensitivity but fairly low ocean diffusivity is effectively ruled out by the kua = 12 likelihood surface, but not by the kua = 14 surface nor (except somewhere below sqrt(Kv) = 2) by the kua = 13 surface. It is the combination of high climate sensitivity but moderately low ocean diffusivity that the F06 surface and deep ocean diagnostics, acting together, have difficulty well-constraining. So the failure of the upper air diagnostic to do so either fattens the upper tail of the climate sensitivity PDF.

Why didn’t the Forest 2006 upper air r2 values reflect kua = 14, as used?
I couldn’t understand why, although F06 used kua = 14, the pattern of its computed r2 values, and hence the likelihood surface, were quite different from those that I computed in R from the same data. Why should the F06 combination of kua = 14 and 50 hPa level weighting produce one answer when I computed the r2 values in R, and a completely different one when computed by F06’s code? Compounding the mystery, the values produced by the F06 code seemed closely related to the kua = 13 case.

I could explain why each F06 r2 would only be only 80% of its expected value, because the code F06 used was designed for a different situation and it divides all the r2 values by 1.25. The same unhelpful division by 1.25 arises in F06’s computation of the r2 values for its surface diagnostic.  But even after adjusting for that, there were large discrepancies in the upper air r2 values. I thought at first that it might be something to do with IDL, the rather impenetrable language in which all the F06 code is written, having vector and array indexing subscripts that start at 0 rather than, as in R, at 1. But the correct explanation was much more interesting.

A missing data mask is generated as part of the processing of the observational data, based on a required minimum proportion of data being extant. That mask may have been used to mark what points should be treated as missing in the (initially complete) control data, when processing it to give changes from the mean of each twenty year period to the mean of a ten year period starting 25 years later. In any event, the values of the 80–85°N latitude band, 150 and 200 hPa level diagnostic variables, along with the variables for a lot of other locations, are marked as missing in the processed control data temperature changes matrix, by being given an undefined data marker value of ‑32,768°C. Such use of an undefined data marker value is common practice for external data, although not being an IDL acolyte I was initially uncertain why it was employed within IDL. It turns out that, when the code was written, IDL did not have a NaN value available.

Rogue undefined data maker values
Variables in the processed control data are then selected using a missing values mask derived from the actual processed observational data, which should eliminate all the control data variables with ‘undefined data’ marker values . Unfortunately, it turns out that the two 80–85°N latitude band control data variables mentioned, unlike all the other control data variables marked as missing by having ‑32,768°C values, aren’t in fact marked as missing in the processed observational data. So they get selected from the control data along with the valid data. I think that the reason why those points aren’t marked as missing in the observational data could possibly be linked to what looks to me like a simple coding error in a function called ‘vltmean’, but I’m not sure.  (All the relevant data and IDL code can be downloaded as large (2 GB) file archive GRL06_reproduce.tgz here.)

So, the result is that the ‑32,768°C control data marker values for the 80–85°N latitude band, 150 and 200 hPa pressure levels get multiplied by the cosine of 82.5° (the mid-point of the latitude band) and then by pressure-level weighting factors of respectively 0.05 and 0.075, to give those variables weighted values of ‑213.854°C and ‑320.781°C for all rows of the final, weighted, control data matrix.  Here’s an extract, at the point it is used to compute the whitening transformation, from the first row of the weighted control data matrix, CT1WGT, resulting from running the F06 IDL code, with the rogue data highlighted:

-0.0013 -213.8540   0.0127    0.0094     0.0051     0.0047     0.0056     0.0024     0.0005          0.0011
0.0038       0.0115     0.0125     0.0058     0.0060    0.0046      0.0030    0.0016     0.0000        -0.0002
0.0036       0.0077    0.0069     0.0001   -0.0047    -0.0060    -0.0067    -0.0050    -0.0030    -320.7810

I should say that I have been able to find this out thanks to considerable help from Jonathan Jones, who has run various modified versions of the F06 IDL code for me. These output relevant intermediate data vectors and matrices as text files that I can read into R and then manipulate.

Why does the rogue data have any effect?
Now, the control data contamination with missing data marker values doesn’t look good, but why should it have any effect? A constant value in all rows of a column of a matrix gives rise to zero entries in its covariance matrix, and a corresponding eigenvalue of zero, which – since the eigenvalues are ordered from largest to smallest – will result in that eigenfunction being excluded upon truncation.

But that didn’t happen. The reason is as follows. F06 used existing Detection and Attribution (D&A) code in order to carry out the whitening transformation – an IDL program module called ‘detect’.  That appears to be a predecessor of the standard module ‘gendetect’, Version 2.1, available at The Optimal Detection Package Webpage, which I think will have been used for a good number of published Detection and Attribution studies. Now, neither detect nor gendetect v2.1 actually carries out an eigendecomposition of the weighted control data covariance matrix.  Instead, they both compute the SVD of the weighted control data matrix itself. If all the columns of that matrix had been centred, then the SVD eigenvalues would be the square roots of the weighted control data sample covariance matrix eigenvalues, and the right singular vectors of the SVD decomposition would match that covariance matrix’s eigenvectors.

However, although all the other control data columns are pretty well centred (their means are within ± 10% or so of their – small – standard deviations), the two columns corresponding to the constant rogue data values are nothing like zero-mean. Therefore, the first eigenfunction  almost entirely represents a huge constant value for those two variables, and has an enormous eigenvalue. The various mean values of other variables, tiny by comparison, will also be represented in the first eigenfunction. The reciprocal of the first eigenvalue is virtually zero, so that eigenfunction contributes nothing to the r2 values. The most important non-constant pattern, which would have been the first eigenfunction of the covariance matrix decomposition, thus becomes the second SVD eigenfunction. There is virtually perfect equivalence between the eigenvalues and eigenvectors, and thus the whitening factors derived from them, of the SVD EOFs 2–14 and (after taking the square root of the eigenvalues) those of the covariance matrix eigendecomposition EOFs 1–13. So, although F06 ostensibly uses kua=14, it effectively uses kua =13. Mystery solved!

Concluding thoughts
It’s not clear to me whether, in F06’s case, the combination of arbitrary marker values incorrectly getting through the mask and then dominating the SVD had any further effects. But it is certainly rather worrying that this could happen. Is it possible that, without centring, the means of a control data matrix could give rise to a rogue EOF that did affect the r2 values, or otherwise materially distort the EOFs, in the absence of any masking error? If that did occur, might the results of some D&A studies be unreliable? Very probably not in either case. However, it does show how careful one needs to be with coding, and importance of making code as well as data available for investigation.

There was actually a good reason for use of the SVD function (from the related PV-wave language, not from IDL itself) rather than the IDL PCA function, which despite its name does actually compute the covariance matrix. Temperature changes in an unforced control run have a zero mean expectation, provided drifting runs are excluded. Therefore, deducting the sample mean is likely to result in a less accurate estimate of the control data covariance matrix than not doing so (by using the SVD eigendecomposition, or otherwise). However, the downside is that if undefined data marker values are used, and something goes wrong with the masking, the eigendecomposition will be heavily impacted. Version 3.1 of gendetect v3.1 does use the PCA function, so the comments in this post are not of any possible relevance to recent studies that use the v3.1 Optimal Detection Package code. However, I suspect that Version 2.1 is still in use by some researchers.

So, the lesson I draw is that use within computer code of large numbers as marker values for undefined data is risky, and that an SVD, rather than a eigendecomposition of the covariance matrix, should only be used to obtain eigenvectors and eigenvalues if great care is taken to ensure that all variables are in fact zero mean (in population, expectation terms).

Finally, readers may like to note that I had a recent post at Climate Etc, here, about another data misprocessing issue concerning the F06 upper air data. That misprocessing appears to account for the strange extended shoulder between 4°C and 6°C  in F06’s climate sensitivity PDF, shown below.

Nic Lewis

F06 Fig9.20



  1. Ian
    Posted Apr 22, 2013 at 11:05 AM | Permalink


    Your links aren’t active.

    Impressive review, as always.


  2. Lance Wallace
    Posted Apr 22, 2013 at 11:06 AM | Permalink

    Last paragraph: Is “here” meant to be a link to Climate Etc?

  3. Posted Apr 22, 2013 at 11:26 AM | Permalink

    Thanks, Ian and Lance. I’ve fixed the inactive links now.

  4. Posted Apr 22, 2013 at 12:56 PM | Permalink

    However, it does show how careful one needs to be with coding, and importance of making code as well as data available for investigation.

    Absolutely. Without this one might from a distance imagine that any languages used by now have a NaN value. The real world of programming is almost always messier than such optimistic imaginings. Hence the need for code.

  5. James Shutiak
    Posted Apr 22, 2013 at 3:09 PM | Permalink

    Great article Steve! Keep up your exceptionally good work in finding and exposing errors and worse by climate warmers.

    Steve: You mean Nic.

    • Andrew S.
      Posted Apr 22, 2013 at 4:02 PM | Permalink

      I think you mean Nic.

  6. JasonScando
    Posted Apr 22, 2013 at 4:10 PM | Permalink

    Nic, is there a way to practically identify who may be using Version 2.1 and forward this post to them? I’d hate to see this discovery go to waste

  7. Posted Apr 22, 2013 at 4:19 PM | Permalink


    I doubt that it is practicable to find out who is using Version 2.1. I have drawn the attention of the author of the original code to the point involved. He thinks that it is most unlikely to be an issue in other cases.

  8. Jean S
    Posted Apr 22, 2013 at 4:47 PM | Permalink

    The error has a source that will probably be familiar to CA readers – failure to check that data is zero mean before undertaking principal components analysis (PCA) / singular value decomposition (SVD).

    Hmmm, Mann did not use non-centered PCA, but a short-centered PCA. And the use of that method well-established in the statistical literature was no failure nor an error, just ask Mann or Tamino. 😉

    Nice post! Btw, Jolliffe is also recommending uncentered/non-centered PCA (i.e., your SVD case) for this type of a situation.

  9. Steven Mosher
    Posted Apr 22, 2013 at 4:55 PM | Permalink

    don’t ask me how many times i have run a computation on the Missing value of -9999.

    I hope folks take your lead in moderating their responses

    ‘This is a cautionary tale, about a mystery that had an unexpected explanation. It’s not intended as a criticism of the scientists involved, and the problem involved, although potentially serious, actually had little impact on the results of the study concerned. However, I am hopeful that mathematically and computing orientated readers will find it of interest. ”

    extra diligence is required when using OPD ( other peoples data0 especially in the cases where folks use things like -999 as “missing data” indicators..

  10. Jeff Condon
    Posted Apr 22, 2013 at 5:37 PM | Permalink

    It is truly amazing that you have parsed that study to this extent. I’m a little surprised that Forest didn’t see the problem. Pretty wild stuff.

  11. Geoff Sherrington
    Posted Apr 22, 2013 at 7:16 PM | Permalink

    I have a problem with cosine weighting of temperature with latitude, which I think is what you are doing (you might have a compensation step later).
    Analogy. I can understand rainfall being weighted because its can be expressed in units of mm per square mm of ground (or similar).
    However, temperature in the atmosphere is not particularly related to the area of the ground below. Its units are K, not K per mm squared.
    One can collect daily rainfall sample and pour them together at the end of a week to confirm that they are additive, but not so with temperature as conventionally instrumented.
    If a cosine transform of temperatures induces heating or cooling on approach to the Poles, is this a real effect or a mathematical artefact?
    Second question. When observations are converted to standard deviations and scaled, how do you preserve the process of propagation of error terms? If the answer is lengthy but the result OK, please don’t go into it.
    Nice work, but I shudder at the hours of detective work you must have spent. I do get a little worried when terms calculated 14 truncation parameters out from the start can have such an influence, when intuition would be to drop them as increasingly insignificant. I wonder whether the initial data quality can support this discrimination, because it places much reliance on Allen & Tett’s test.

  12. Craig Loehle
    Posted Apr 22, 2013 at 7:56 PM | Permalink

    In the beginning of the Royal Society, scientists would bring their discoveries and demonstrate them on the bench top (microscope, chemistry, fossils…). Hence their motto Nullius in verba because you could see the results yourself. Even today in many studies you can understand what was done (drug x given to 1000 people…) and examine the table or graph of results. But when there are large data sets and fancy transformations of that data, just as when there are models, the errors become hard to see. In my first computer class I had about 10 lines of code (yes, FORTRAN IV) for a homework assignment, but it was acting strange. I took it to the teacher who could not find the mistake either and gave me full credit. 10 lines. Which is what makes Steve’s scripts so very special and useful as a tool and example.

  13. Posted Apr 22, 2013 at 8:32 PM | Permalink

    I have a colleague who does work using APL. He is an exceptionally careful guy, but I wonder what would happen if someone tried to audit his stuff. My brief college exposure to that language convinced me that there was nothing else quite so gnomic and obscure (although certainly compact).

    Steve: I remember APL from the 1970s or early 1980s. It was, as you say, very compact.

    • Posted Apr 23, 2013 at 12:48 AM | Permalink

      APL was an excellent language. S was based on it (specifically linear algebra), which led on to R.

    • Posted Apr 23, 2013 at 12:55 AM | Permalink

      APL was invented by Iverson, a Canadian, and substantially developed in Toronto.

      • Posted Apr 23, 2013 at 6:31 AM | Permalink

        I’d forgotten Iverson was Canadian. APL and I.P. Sharp became influential in fund management but APL lacked a congruous and speedy database. Mike Caruso of Insyte in Boston fixed that in the mid 80s, influenced by APL and Smalltalk, in a time series database called Vision, now owned by FactSet. Prototype-based like Self (and thus JavaScript) with persistence built in. Little known outside analysts in outfits like Putnams but a pointer to a better future.

    • Frank Knarf
      Posted Apr 23, 2013 at 3:50 PM | Permalink

      Bill Gates was an APL fan back in the mid 70s. He wanted to add it to the Microsoft language set.

    • MrPete
      Posted Apr 23, 2013 at 10:07 PM | Permalink

      Re: stevepostrel (Apr 22 20:32),
      Heh. Obscure trivia: the programming language for the first retail microcomputer (“sold at Macy’s near you!”) was APL. Our CEO kicked young Bill Gates out the door when he tried to sell us his Basic.

      (Helping create the APL cartridge was one of my first commercial consulting projects 😉 )

  14. Posted Apr 23, 2013 at 12:51 AM | Permalink

    Same problem for all of Safari, Chrome and Firefox on my Mac.

    [Jean S: Something is broken (in my Firefox both left and right panels are gone). I can get into the admin panel, but I don’t see anything unusual in the areas I have an access. We have to wait someone more knowledgeable to fix the problem.]

  15. kim
    Posted Apr 23, 2013 at 3:28 AM | Permalink

    Read my fine manual,
    Early, much is not centered.
    Have a cuppa cha.

  16. johanna
    Posted Apr 23, 2013 at 4:24 AM | Permalink

    While we wait for the restoration of normal service, here is an (indented) couplet:

    Inducing from a proxy
    Requires a lot of moxie!

  17. Alan Bates
    Posted Apr 23, 2013 at 5:27 AM | Permalink

    IE9.0.8 64 bit. Highly hillarious – “Non-centring in Climate Audit”

  18. andyd
    Posted Apr 23, 2013 at 6:45 AM | Permalink

    Hopefully it won’t become de-trended as well.

  19. Posted Apr 23, 2013 at 11:07 AM | Permalink

    Geoff Sherrington
    ” I have a problem with cosine weighting of temperature with latitude”

    I’m following Forest here, who is using a mass weighting scheme, as seems commonly employed in this sort of study. I think the mass weighting by pressure level makes more difference than by latitude, as temperature changes in the tropospheric and stratospheric layers have inconsistent implications about climate sensitivity (highish and very low respectively, broadly speaking).

    “Second question. When observations are converted to standard deviations and scaled, how do you preserve the process of propagation of error terms? If the answer is lengthy but the result OK, please don’t go into it.”

    Measurement errors are relatively small compared with internal variability except for the univariate deep ocean diagnostic – where their variance is added to that from internal variability – and are ignored in the F06 method.

    “Nice work, but I shudder at the hours of detective work you must have spent.”
    Yes, unfortunately very time consuming.

    “I do get a little worried when terms calculated 14 truncation parameters out from the start can have such an influence, when intuition would be to drop them as increasingly insignificant. I wonder whether the initial data quality can support this discrimination, because it places much reliance on Allen & Tett’s test.”

    I think the answer may well be no in the case of the upper air diagnostic. And the highest retained EOFs, which have the smallest eigenvalues but the lowest signal-to-noise ratios, may well have the most influence on the r-squared value from which the likelihood is derived, because the whitening process involves dividing by the eigenvalues.

    • Geoff Sherrington
      Posted Apr 24, 2013 at 6:20 AM | Permalink

      Nic, thank you for the detailed reply. Correct process trumps my belief again.
      Re cosine weighting, I guess the question goes away for pressure which is measured over an area, psi in old terms if you like. But do you have a comment on area weighting of temperatures as used by others when creating grid cells and filling with values? My thoughts are going around in circles. Temperature seems a different class of number to (say) rainfall.

  20. EdeF
    Posted Apr 23, 2013 at 11:47 AM | Permalink

    -32,768 deg C is the exact temperature of my house right after my wife discovers I have eaten the last piece of key lime pie!

  21. MrPete
    Posted Apr 23, 2013 at 3:14 PM | Permalink

    Re: OldUnixHead (Apr 23 00:10),
    Many apologies. Don’t know what is wrong. We have an urgent call in to WordPress tech support…

    • MrPete
      Posted Apr 23, 2013 at 3:15 PM | Permalink

      Re: MrPete (Apr 23 15:14),
      NOTE: if anyone has a saved copy of a recent CA web page, please zip it up and email to me at webbed dot pete at gmail — perhaps I can discover the difference between what you saved and what is there now!

  22. MrPete
    Posted Apr 23, 2013 at 9:54 PM | Permalink

    Thanks everyone for your patience. (Obviously) the site display issue is solved. (Turns out a new setting was added. They guessed wrong in setting the default for our site, and didn’t inform about the change.)

    • ianl8888
      Posted Apr 24, 2013 at 7:07 PM | Permalink

      Whenever a software organisation announces an “upgrade” to some programme I’m using, I back everything up to a non-local disk then run home and lock the door 🙂

  23. Paul Vincelli
    Posted Jun 16, 2013 at 3:35 PM | Permalink

    Several months ago, the University of Kentucky hosted of forum on climate change with three excellent speakers who were all self-described conservatives…

    [Snip – this blog is not about liberal/conservative perspectives. And the link you provided began with yet another illustration of shoddy scientific thinking. Let’s not go there.]

%d bloggers like this: