The Real Story of Climategate

A year ago today, an unidentified hacker published a zipped folder in several locations online. In this folder were approximately one thousand emails and three thousand files which had been stolen from the backup server of the Climatic Research Unit in the UK, a top centre for global temperature analysis and climate change studies. As links to the folder were passed around on blogs and online communities, a small group of people sorted through the emails, picking out a handful of phrases that could be seen as controversial, and developing a narrative which they pushed to the media with all their combined strength. “A lot is happening behind the scenes,” one blog administrator wrote. “It is not being ignored. Much is being coordinated among major players and the media. Thank you very much. You will notice the beginnings of activity on other sites now. Here soon to follow.”

This was not the work of a computer-savvy teenager that liked to hack security systems for fun. Whoever the thief was, they knew what they were looking for. They knew how valuable the emails could be in the hands of the climate change denial movement.

Skepticism is a worthy quality in science, but denial is not. A skeptic will only accept a claim given sufficient evidence, but a denier will cling to their beliefs regardless of evidence. They will relentlessly attack arguments that contradict their cause, using talking points that are full of misconceptions and well-known to be false, while blindly accepting any argument that seems to support their point of view. A skeptic is willing to change their mind. A denier is not.

There are many examples of denial in our society, but perhaps the most powerful and pervasive is climate change denial. We’ve been hearing the movement’s arguments for years, ranging from illogic (“climate changed naturally in the past, so it must be natural now“) to misrepresentation (“global warming stopped in 1998“) to flat-out lies (“volcanoes emit more carbon dioxide than humans“). Of course, climate scientists thought of these objections and ruled them out long before you and I even knew what global warming was, so in recent years, the arguments of deniers were beginning to reach a dead end. The Copenhagen climate summit was approaching, and the public was beginning to understand the basic science of human-caused climate change, even realize that the vast majority of the scientific community was concerned about it. A new strategy for denial and delay was needed – ideally, for the public to lose trust in researchers. Hence, the hack at CRU, and the beginning of a disturbing new campaign to smear the reputations of climate scientists.

The contents of the emails were spun in a brilliant exercise of selective quotation. Out of context, phrases can be twisted to mean any number of things – especially if they were written as private correspondence with colleagues, rather than with public communication in mind. Think about all the emails you have sent in the past decade. Chances are, if someone tried hard enough, they could make a few sentences you had written sound like evidence of malpractice, regardless of your real actions or intentions.

Consequently, a mathematical “trick” (clever calculation) to efficiently analyse data was reframed as a conspiracy to “trick” (deceive) the public into believing the world was warming. Researchers discussed how to statistically isolate and “hide the decline” in problematic tree ring data that was no longer measuring what it used to, but this quote was immediately twisted to claim that the decline was in global temperatures: the world is cooling and scientists are hiding it from us!

Other accusations were based not on selective misquotation but on a misunderstanding of the way science works. When the researchers discussed what they felt were substandard papers that should not be published, many champions of the stolen emails shouted accusations that scientists were censoring their critics, as if all studies, no matter how weak their arguments, had a fundamental right to be published. Another email, in which a researcher privately expressed a desire to punch a notorious climate change denier, was twisted into an accusation that the scientists threatened people who disagreed with them. How was it a threat if the action was never intended to materialize, and if the supposed target was never aware of it?

These serious and potentially damaging allegations, which, upon closer examination, are nothing more than grasping at straws, were not carefully examined and evaluated by journalists – they were repeated. Early media reports bordered on the hysterical. With headlines such as “The final nail in the coffin of anthropogenic global warming” and “The worst scientific scandal of our generation“, libelous claims and wild extrapolations were published mere days after the emails were distributed. How could journalists have possibly had time to carefully examine the contents of one thousand emails? It seems much more likely that they took the short-cut of repeating the narrative of the deniers without assessing its accuracy.

Even if, for the sake of argument, all science conducted by the CRU was fraudulent, our understanding of global warming would not change. The CRU runs a global temperature dataset, but so do at least six other universities and government agencies around the world, and their independent conclusions are virtually identical. The evidence for human-caused climate change is not a house of cards that will collapse as soon as one piece is taken away. It’s more like a mountain: scrape a couple of pebbles off the top, but the mountain is still there. For respected newspapers and media outlets to ignore the many independent lines of evidence for this phenomenon in favour of a more interesting and controversial story was blatantly irresponsible, and almost no retractions or apologies have been published since.

The worldwide media attention to this so-called scandal had a profound personal impact on the scientists involved. Many of them received death threats and hate mail for weeks on end. Dr. Phil Jones, the director of CRU, was nearly driven to suicide. Another scientist, who wishes to remain anonymous, had a dead animal dumped on his doorstep and now travels with bodyguards. Perhaps the most wide-reaching impact of the issue was the realization that private correspondence was no longer a safe environment. This fear only intensified when the top climate modelling centre in Canada was broken into, in an obvious attempt to find more material that could be used to smear the reputations of climate scientists. For an occupation that relies heavily on email for cross-national collaboration on datasets and studies, the pressure to write in a way that cannot be taken out of context – a near-impossible task – amounts to a stifling of science.

Before long, the investigations into the contents of the stolen emails were completed, and one by one, they came back clear. Six independent investigations reached basically the same conclusion: despite some reasonable concerns about data archival and sharing at CRU, the scientists had shown integrity and honesty. No science had been falsified, manipulated, exaggerated, or fudged. Despite all the media hullabaloo, “climategate” hadn’t actually changed anything.

Sadly, by the time the investigations were complete, the media hullabaloo had died down to a trickle. Climategate was old news, and although most newspapers published stories on the exonerations, they were generally brief, buried deep in the paper, and filled with quotes from PR spokespeople that insisted the investigations were “whitewashed”. In fact, Scott Mandia, a meteorology professor, found that media outlets devoted five to eleven times more stories to the accusations against the scientists than they devoted to the resulting exonerations of the scientists.

Six investigations weren’t enough, though, for some stubborn American politicians who couldn’t let go of the article of faith that Climategate was proof of a vast academic conspiracy. Senator James Inhofe planned a McCarthy-like criminal prosecution of seventeen researchers, most of whom had done nothing more than occasionally correspond with the CRU scientists. The Attorney General of Virginia, Ken Cuccinelli, repeatedly filed requests to investigate Dr. Michael Mann, a prominent paleoclimatic researcher, for fraud, simply because a twelve-year-old paper by Mann had some statistical weaknesses. Ironically, the Republican Party, which prides itself on fiscal responsibility and lower government spending, continues to advocate wasting massive sums of money conducting inquiries which have already been completed multiple times.

Where are the politicians condemning the limited resources spent on the as yet inconclusive investigations into who stole these emails, and why? Who outside the scientific community is demanding apologies from the hundreds of media outlets that spread libelous accusations without evidence? Why has the ongoing smear campaign against researchers studying what is arguably the most pressing issue of our time gone largely unnoticed, and been aided by complacent media coverage?

Fraud is a criminal charge, and should be treated as such. Climate scientists, just like anyone else, have the right to be presumed innocent until proven guilty. They shouldn’t have to endure this endless harassment of being publicly labelled as frauds without evidence. However, the injustice doesn’t end there. This hate campaign is a dangerous distraction from the consequences of global climate change, a problem that becomes more difficult to solve with every year we delay. The potential consequences are much more severe, and the time we have left to successfully address it is much shorter, than the vast majority of the public realizes. Unfortunately, powerful forces are at work to keep it that way. This little tussle about the integrity of a few researchers could have consequences millennia from now – if we let it.

Update: Many other climate bloggers are doing Climategate anniversary pieces. Two great ones I read today were Bart Verheggen’s article and the transcript of John Cook’s radio broadcast. Be sure to check them out!

Advertisement

Odds and Ends

I must thank Michael Tobis for two pieces of reading that his blog recently pointed me to. First, a fantastic article by Bill McKibben, which everyone should print out and stick to their fridge. Here’s a taste:

Read the comments on one of the representative websites: Global warming is a “fraud” or a “plot.” Scientists are liars out to line their pockets with government grants. Environmentalism is nothing but a money-spinning “scam.” These people aren’t reading the science and thinking, I have some questions about this. They’re convinced of a massive conspiracy.

The odd and troubling thing about this stance is not just that it prevents action. It’s also profoundly unconservative. If there was ever a radical project, monkeying with the climate would surely qualify. Had the Soviet Union built secret factories to pour carbon dioxide into the atmosphere and threatened to raise the sea level and subvert the Grain Belt, the prevailing conservative response would have been: Bomb them. Bomb them back to the Holocene—to the 10,000-year period of climatic stability now unraveling, the period that underwrote the rise of human civilization that conservatism has taken as its duty to protect. Conservatism has always stressed stability and continuity; since Burke, the watchwords have been tradition, authority, heritage. The globally averaged temperature of the planet has been 57 degrees, give or take, for most of human history; we know that works, that it allows the world we have enjoyed. Now, the finest minds, using the finest equipment, tell us that it’s headed toward 61 or 62 or 63 degrees unless we rapidly leave fossil fuel behind, and that, in the words of NASA scientists, this new world won’t be “similar to that on which civilization developed and to which life on earth is adapted.” Conservatives should be leading the desperate fight to preserve the earth we were born on.

Read the rest of the article here. Highly recommended to all.

The other link I wanted to share was a new publication entitled “Science and the Media”, just released by the American Academy of Arts and Sciences (not to be confused with the American Association for the Advancement of Science – why all the acronym duplication?)

With contributions from everyone from Donald Kennedy to Alan Alda, and essays with titles from “The Scientist as Citizen” to “Civic Scientific Literacy: The Role of the Media in the Electronic Era”, I’m virtually certain that I will enjoy this one (sorry, I can’t bring myself to say things like “certain” without caveats any more). The 109-page pdf is available free of charge and can be accessed from this page, which also includes information on ordering hard copies.

In other news, the La Niña conditions in the eastern Pacific (see anomaly map above) have bumped this year’s temperatures down a bit, so January-September 2010 is now tied for the warmest on record, rather than being a clear winner. This analysis is from NCDC, however, and I’m not sure how they deal with sparse data in the Arctic (for background, see this post – a summary of one of the most interesting papers I’ve read this year). Does anyone know if GISS has an up-to-date estimate for 2010 temperatures that we could compare it to? All I can find on their website are lines and lines of raw data, and I’m not really sure how to process it myself.

That’s all for today. Enjoy the week, everyone.

Global Surface Temperature Change

I really enjoyed reading “Global Surface Temperature Change“, by James Hansen and his team at GISS. Keep in mind that it’s still in the draft stages – they haven’t submitted to a journal yet, but they certainly plan to, and it’s a very credible team of scientists that will almost definitely get it published.

The paper is mostly about the methods of global temperature analysis. It’s more of a review paper than an account of a single experiment. However, their main discussion point was that even by using the same data, problems can be addressed in different ways. The two main problems with temperature analysis are:

  • “incomplete spatial and temporal coverage” (sparse data)
  • “non-climatic influences on measurement station environment” (urban heat island effect).

The authors explain the methods they use and why, and explore the impacts that different methods have on their results.

GISS measures anomalies in the temperatures, largely because they are much smoother and more consistent, geographically, than absolute temperatures. In 1987, they determined that anomalies could be safely extrapolated for a radius of 1200 km from a station and still be accurate. GISS smooths the whole map out by extrapolating everything and averaging the overlapping bits.

Extrapolating is also very useful in areas with very few stations, such as the polar regions and parts of Africa. In this map, grey indicates missing data:



The Arctic is particularly problematic, not only because its data is so sparse, but also because it has the largest anomaly of any region in the world. If you have incomplete coverage of an area that is warming so dramatically, it won’t pull its full weight in the global trend, and your result will almost certainly be too low.

This difficulty with the Arctic is the reason that GISS says 2005 is the warmest year on record, while HadCRUT, the team in England, says that 1998 is. GISS extrapolates from the stations they have, and end up getting pretty good coverage of the Arctic:

They’re assuming that areas with missing data have the same anomaly as whatever temperature stations are within 1200 km, which, as they determined in 1987, is a pretty fair assumption.

However, HadCRUT doesn’t do this extrapolating thing. When they don’t have data for an area, they just leave it out:

This might sound safer, in a way, but this method also makes an assumption. It assumes that the area has the same anomaly as the global average. And as we all know, the Arctic is warming a lot more and a lot faster than the global average. So it’s quite possible that GISS is right on this one.

Another adjustment that NASA makes is for local, anthropogenic, non-climatic effects on temperature data. The most obvious of these is the urban heat island effect. As an area becomes more urban, it gets more pavement, less vegetation, and its albedo goes down – it absorbs more heat. This often makes cities substantially warmer than the surrounding rural areas, which can obviously contaminate the temperature record. However, there are ways of eliminating urban influences from the data so we can see what the real trend is.

The first step is determining what stations are considered urban. The obvious way to do this is through population, but that’s actually not very accurate. Think of somewhere like Africa, where, even if there are thousands of people living in a small area, the urban influences such as concrete, absence of vegetation, or exhaust aren’t usually present. A much better indication is energy use, and a good proxy for energy use, that’s easy to measure, is lights at night-time.

So GISS put a bit of code into their analysis that singles out stations where nightlight brightness is greater than 32 µW/m2/sr/µm, and adjusts their trends to agree with rural stations within 1200 km. If there aren’t enough rural stations within that radius, they’ll just exclude the station from the analysis.

They did an even more rigorous test for this paper, to test just how much urban influences were contaminating the long-term trend, and it was pretty interesting.

There were enough stations considered “pitch-dark” at night, where they couldn’t detect any light, to run a global analysis all by themselves. The trend that came out was <0.01 °C/century smaller than GISS’s normal calculation, an amount of error that they described as “immeasurably small”.

The result of all this temperature analysis is a graph, with one new point every year, that is “eagerly awaited by some members of the public and the media”:

However, this graph isn’t actually as useful as this one – the 12-month running mean:

“From a climate standpoint there is nothing special about the time  of year at which the calendar begins”, so instead of only measuring January-December, you can also do February-January, March-February, and so on. This way, you get a data point every month instead of every year, and more data means more accuracy. It also solves problems with short-term influences, such as El Nino, La Nina, and volcanic eruptions, that the annual graph was having. These fleeting, but fairly substantial, influences can fall completely into one calendar year or be split between two – so their influence on global temperature could be overestimated or underestimated, depending on the starting month of the calendar. The 12-month running mean is much less misleading in this fashion.

As it is, we just set a new record for the 12-month running mean, and unless La Nina really takes off, 2010 will likely set a new record for the annual graph as well. But the authors argue that we need to start moving away from the annual graph, because it isn’t as useful.

The authors also discuss public perception of climate change, and media coverage of the issue. They say, “Our comments here about communication of this climate science to the public are our opinion…[We offer it] because it seems inappropriate to ignore the vast range of claims appearing in the media and in hopes that open discussion of these matters may help people distinguish the reality of global change sooner than would otherwise be the case.”

They make the very good point that “Lay people’s perception tends to be strongly influenced by the latest local fluctuation”, and use this winter as a case study, where a strongly negative Arctic Oscillation index caused significantly cooler-than-normal conditions across the United States and Europe. Consequently, a lot of people, especially in the US, began to doubt the reality of global warming – even though, in the world as a whole, it was the second warmest winter on record:

The authors also talk about data sharing. GISS likes to make everything freely available to the public – temperature station data, computer code, everything. However, putting it out there immediately, so that anyone can help check for flaws, has “a practical disadvantage: it allows any data flaws to be interpreted and misrepresented as machinations.” Multiple times in the past few years, when there have been minor errors that didn’t actually change anything, GISS was widely accused of making these mistakes deliberately, to “intentionally exaggerate the magnitude of global warming”. They realized this wasn’t working, so they changed their system: Before releasing the data to everyone, they first put it up on a private site so that only select scientists can examine it for flaws. And, of course, this “has resulted in the criticism that GISS now “hides” their data”.

Personally, I find the range and prevalence of these accusations against scientists absolutely terrifying. Look at what has become mainstream:

Scientific fraud is a very serious allegation, and it’s one thing for citizens to make it without evidence, but it’s another thing altogether for the media to repeat such claims without first investigating their validity:

I have been disgusted by the media coverage of climate science, especially over the past year, especially in the United States, and I worry what this will mean for our ability to solve the problem.

However, there is still fantastic science going on that is absolutely fascinating and essential to our understanding of global climate change. This paper was a very interesting read, and it helped me to better understand a lot of aspects of global temperature analysis.

So What Happened with ClimateGate?

Remember back in December, when the news was buzzing each day about the stolen emails from top climate researchers? They were described as “the final nail in the coffin of anthropogenic global warming”, or worse. Apparently, the scientists had written things that severely compromised the underpinnings for the idea that human activity was causing the Earth to warm. We could now all stop worrying and forget about cap-and-trade.

But that wasn’t the end of the story. There were no less than four independent investigations into the contents of these emails – conducted by scientists, universities, and governments, not general reporters rushing off a story about an area of science with which they were unfamiliar, and trying to make it sound interesting and controversial in the process.

So what did these investigations find? Is the Earth still warming? Are humans still responsible? Can we trust the scientific process any more, or should we throw peer-review out the window and practice Blog Science instead?

Actually, all four of the investigations concluded that absolutely no science was compromised by the contents of the emails. The CRU scientists weren’t as good as they should have been about making data easily accessible to others, but that was the only real criticism. These scientists are not frauds, although they are accused of it on a daily basis.

Pennsylvania State University, over a series of two reports, investigated the actions of their employee, Dr. Michael Mann, who is arguably at the top of the field of paleoclimatology. They found that, contrary to most accounts in the mainstream media, he did not hide or manipulate any data to exaggerate global warming, delete any emails that might seem suspicious and be subject to Freedom of Information requests, or unjustly suppress skeptical papers from publication. After a second investigation, following up on the catch-all accusation of “seriously deviating from accepted practices within the academic community”, Penn State exonerated Mann. They criticized him for occasionally sharing unpublished manuscripts with his colleagues without first obtaining the express permission of the authors, but besides that minor (and somewhat unrelated) reprimand, they found absolutely nothing wrong.

The British House of Commons investigated the actions of CRU director Phil Jones, and came to a similar conclusion. They found that his “actions were in line with common practice in the climate science community”, that he was “not part of a systematic attempt to mislead” or “subvert the peer review process”, and that “the focus on CRU….has been largely misplaced”. They criticized CRU’s lack of openness with their data, but said that the responsibility should lie with the University of East Anglia, which CRU is a part of. So these scientists should really catch up to the climate research team at NASA, for example, which publishes all of their raw data, methodologies, and computer codes online, with impeccable archives.

The University of East Anglia conducted their own investigation into the actions of CRU as a whole. They found no hint of tailoring results to a particular agenda”, and asserted thatallegations of deliberate misrepresentation and unjustified selection of data are not valid”. They also explored the lack of transparency in CRU, but were more sympathetic. “CRU accepts with hindsight”, they write, “that they should have devoted more attention in the past to archiving data and algorithms and recording exactly what they did. At the time the work was done, they had no idea that these data would assume the importance they have today and that the Unit would have to answer detailed inquiries on earlier work.” They also note that CRU should not have had to respond to Freedom of Information requests for data which they did not own (such as weather station records).

Just last week, the final investigation, headed by Sir Muir Russell on behalf of UEA, found that “their rigour and honesty as scientists are not in doubt.” Is this starting to seem a bit repetitive? To illustrate their point, over the course of two days, they independently reconstructed the global temperature record using publicly available data, and came to the same conclusion as CRU. Again, there was the criticism that CRU was not as open as it should have been. They also noted that an obscure cover figure for a 1999 World Meteorological Organization report, constructed by Phil Jones, did not include enough caveats about what was proxy data and what was instrumental data. However, the more formally published, and much more iconic, graphs in Mann 98 and the IPCC TAR, were fine.

There have been some great comments on the results of these investigations since they were released, especially by scientists. Here are some samples:

[The CRU researchers] are honest, hardworking scientists whose reputations have been unjustifiably smeared by allegations of unscrupulous behaviour…I hope that the media will devote as much attention to this comprehensive dismissal of the allegations as it did to promoting the hysteria surrounding the email theft in the first place. Will the Daily Telegraph now retract its claim that the emails revealed “the greatest scientific scandal of our age” and apologize unreservedly to Phil Jones? Will there now be a public inquiry about the erroneous, shallow and repetitive nonsense promulgated in the media over this affair? If there is a scandal to be reported at all, it is this: the media stoked a controversy without properly investigating the issues, choosing to inflate trivialities to the level of an international scandal, without regard for the facts or individuals affected. This was a shameful chapter in the history of news reporting. -Raymond Bradley, director of the Climate System Research Center at the University of Massachusetts

The call for greater transparency and openness among scientists and their institutions is necessary and welcomed, but certainly they aren’t the only ones who deserve that reminder. What institution on the planet would pass muster under such intense scrutiny? Certainly not the U.S. government agencies, which often deny or impede FOIA requests, or global corporations like BP, Massey Energy and Koch Industries, which seem to revel in hiding information from the public all the time. More transparency is needed everywhere, not just among scientists in lab coats. -Brendan DeMelle, freelance journalist, DeSmogBlog

[The Muir-Russell report] makes a number of recommendations for improvements in processes and practices at the CRU, and so can be taken as mildly critical, especially of CRU governance. But in so doing, it never really acknowledges the problems a small research unit (varying between 3.5 to 5 FTE staff over the last decade) would have in finding the resources and funding to be an early adopter in open data and public communication, while somehow managing to do cutting edge research in its area of expertise too. -Steve Easterbrook, computer science professor at the University of Toronto

I agree with these statements. I think that we are holding scientists in general, but especially climate scientists, to a far higher standard than any other group of people in the world. We need to relax a bit and realize that scientists make mistakes, and that innocent mistakes are not evidence of fraud that will bring a long-standing theory tumbling down. We need to realize that scientists are employees like any others, who don’t always follow ideal actions in every professional situation, especially when they are under intense pressure that includes death threats and accusations of criminal activity.

However, at the same time, we need to start holding other groups of people, especially journalists, to a higher standard. Why has the media been able to get away with perpetuating serious allegations without first investigating the what really happened, and without publishing explicit retractions and apologies when the people whose reputations they smeared are found innocent? Why haven’t there been four official investigations into who stole these emails, and why?

2010 On Track for the Warmest Year on Record

The data is in from both NASA and NCDC, and NASA’s prediction of 2010 being the warmest year on record is well on its way to coming true, unless La Niña conditions rapidly develop (see page 15 of the NASA document). It has been:

  • the warmest March on record
  • the warmest January-March on record
  • the warmest April on record
  • the warmest January-April on record

Read NCDC’s reports on the March and April global temperatures, NASA’s maps of the January-April temperatures in 2010 as compared to 2005 and 1998 (the warmest years on record, at least until now), and Joe Romm’s excellent summaries.

As Romm says, “After the endless disinformation-based global cooling stories of the past few years, it’s time for the media to start do some serious fact-based global warming stories.” I fully agree. Everyone keep your eyes open, and see whether or not these record-breaking global temperatures are actually covered.

A Better Credibility Spectrum

It’s been over a year since I wrote The Credibility Spectrum, my first post ever. Since then I’ve learned a lot, and have altered the credibility spectrum in my own mind – so I thought I’d alter it here, too.

This credibility spectrum is sort of split into two: the scientific community, and the non-scientific community. The scientific community starts with scientists, and I want to stress that this category only includes scientists with experience in the issue at hand. Just because someone has a PhD in one area of science doesn’t mean that they are an expert in all areas. For example, it’s very easy for a computer scientist to go through ten years of university without studying any biology at all.  Treating them as an expert in evolution, therefore, would be illogical.

These scientists write peer-reviewed papers, published in journals like Nature and Science, which are another step up the credibility spectrum. Instead of just having the name of an expert attached to them, their methods and conclusions have been evaluated for robustness and accuracy. This is the minimum level of credibility from which I recommend citing scientific claims.

However, as thousands of papers are published every month, and they’re generally studying the frontier of their field, it’s inevitable that some of them will be proven wrong later. As Gavin Schmidt and Michael Mann wisely said, peer review is a necessary but not sufficient condition.

That’s why there are scientific organizations and assessment reports, like NASA or the IPCC, which compile peer-reviewed knowledge which has stood the test of time into consensus statements. Even the top level of the credibility spectrum isn’t infallible, but it sure has a low error rate compared to other sources.

Everyone who isn’t a scientist, which is most of us, falls into the lower half of the credibility spectrum. The category I refer to as “communicators” includes the mainstream media, projects like Manpollo or 350.org, high school teachers, politicians…….They’re not part of the scientific community, so you should always always always check their citations, but they’re held more accountable for what they say than just any random person on the street. If they make glaring errors, people will be more upset than if the same errors were made by individuals – comments on YouTube, discussions with your neighbours – which make up the lowest rung of our credibility spectrum.

Something that I found really interesting  when I put this together was the general flow of information between different sources. In the scientific community, research starts with scientists, and the best research is published in journals, and the best journal articles are picked up by major organizations. As the scientific knowledge progresses through the different sources, the weaker assertions are weeded out along the way. The flow of information is going up the pyramid, towards the narrower part of the pyramid, so that only the best is retained.

However, in the non-scientific community, the flow of information goes the other way. Communicators present information to individuals, which is a much larger group. Information travelling down the pyramid, instead of up, allow rumours and misconceptions to flourish much more easily.

This isn’t to say that, when they come head-to-head, organizations are always right and individuals are always wrong. But given the history of such disagreements, and the levels of credibility involved, you’ll know where to place your bets.

We Have Slides!

After a marathon PowerPoint-session yesterday I finally got my 63 slides out of the way. Here is the presentation for anyone who is interested. The script is written in the notes beneath the slides.

I like to have things fading in and out of my slides, so sometimes the text boxes and images are stacked on top of each other and it won’t make sense until you view the animation.

Researching the median lethal dose of arsenic during my spare at school was really awkward. I had to do a lot of hasty explaining to my friends about how it was a metaphor for small concentrations having large effects, and no, I wasn’t planning to poison anyone.

Anyway, enjoy.

Mind the Gap (12 MB)

A Better Term

The past week or so, there’s been a lot of discussion on the blogs I read about whether or not we should use the term “global warming”. Simon Donner covered it, then Michael Tobis put in his two cents, and Peter Sinclair discussed it on live TV. It got me thinking.

The term “global warming” doesn’t cover the implications of a rising temperature, just the fact that the average global temperature is rising. “Climate change” covers everything.

However, the public is far more acquainted with the term “global warming”, so it makes sense to keep it in use so we are understood. However, this term has led to many misconceptions. To the public, it’s often interpreted as “everywhere-in-the-world-all-the-time warming”. So every time it’s cold in Europe, or it snows on the Atlantic Coast, there are comments of “where’s global warming” from the usual suspects, and even reasonable people start to doubt the phenomenon.

Because we’ll never be able to completely replace the term “global warming” with something more accurate, I think it would be best to slightly alter it, in a way that will reinforce the idea of averages rather than variability, even to people who don’t use it. I think we should call it “net global warming“.

Really, this is exactly what the folks at NASA, NCDC, or any of the other global temperature monitoring stations calculate. They add up all the regional anomalies from thousands of weather stations to find a net anomaly for the entire planet. This way, we don’t know precisely what the absolute global temperature is, but we know how much it has changed overall. Some places are warmer, some places are colder, but this single annual value shows us at a glance what the planet’s temperature is doing.

Even that value doesn’t consistently warm every year – we don’t set a new temperature record every year – because the global climate has enough internal variability to slightly change, from year to year, the percentage of the planet’s heat that is stored at the surface, which is what NASA et al measures. La Niña, which made 2008 cooler than previous years, doesn’t actually change the total heat energy stored on earth. It just sucks some of it out of the atmosphere and into the oceans.

Luckily, La Niña has a friend, El Niño, which does the opposite to net global surface temperatures, so they basically cancel each other out every five years or so. All these short-term cycles of variability – ENSO, sunspots – average out over the years, and as time goes on we stop seeing the noise and start to see the trend. Look at the red line in NASA’s graph, rather than the jagged black line, to see less noise and more trend.

If we can get the idea of “net global warming”, rather than “everywhere-in-the-world-all-the-time warming”, into the public consciousness, our collective understanding of this problem will shoot up. Plus, we’ll be less prone to misleading statements like “It’s cold today, so there is no global warming” or “Warming stopped in 1998“.

NASA Speaks on 2009

I’d been watching the GISS page closely, to no avail, but it turns out that the annual summation for 2009 global temperatures was posted on James Hansen’s Columbia page.

(Click on the graph for a better resolution in the Columbia document.)

2009 is the second warmest on record, “but it is so close to 1998, 2002, 2003, 2006, and 2007 that we must declare these years as being in a virtual tie as the second warmest year,” says Dr  Hansen. 2009 is probably the warmest among those years, and 2006 is probably the coolest.

2009, especially the spring and summer, was quite chilly in my area, so this is a perfect example of the difference between regional and global temperature. Global temperature reflects the amount of infrared radiation in the atmosphere; regional temperature reflects how that energy is distributed.

NCDC’s preliminary calculations, however, show 2009 as the fifth warmest on record. What are the reasons for this discrepancy? Do they monitor Arctic temperatures differently than NASA, like the British Met Office does?

Hopefully, this will be the end of “global warming stopped in 1998” comments. But maybe the usual suspects will keep saying it and hope nobody notices, or say “global warming stopped in 2005” instead.