Next year’s International Conference on Software Engineering (ICSE), to be held in Zurich, has an interesting conference slogan: Sustainable Software for a Sustainable World

In many ways, ICSE is my community. By that I mean, this is the conference where I have presented my research most often, and is generally my first choice of venue for new papers. This is an important point: one of the most crucial pieces of advice I give to new PhD students is to “find your community”. To be successful as a researcher (and especially as an academic) you have to build a reputation for solid research within an existing research community. Which means figuring out early which community you belong to: who will be the audience for your research results? who will understand your work well enough to review your papers? And eventually, which community will you be looking to for letters of support for job applications, tenure reviews, and so on? And once you’ve figured out which community you belong to, you have to attend the conferences and workshops run by that community, and present your work to them as often as you can, and you have to get to know the senior people in that community. Or rather, they have to get to know you.

The problem is, in recent years, I’ve gone off ICSE. Having spent a lot of time in the last few years mixing with a different research community (climate science, and especially geoscientific model development), I come back to the  ICSE community with a different perspective, and what I see now (in general) it is a rather insular community, focussed on a narrow, technical set of research questions that seem largely irrelevant to anything that matters, and a huge resistance to inter-disciplinary research. This view crystallized for me last fall, when I attended a two-day workshop on “the Future of Software Engineering”, but came away very disappointed (my blog post from the workshop captured this very well).

I should be clear, I don’t mean to write off the entire community – there’s some excellent people in the ICSE community, doing fascinating research – many of them I regard as good friends. But the conference itself seems ever less relevant. The keynote talks always suck. And the technical program tends to be dominated by a large number of dull papers: incremental results on unimaginative research problems.

Perhaps this is a result of the way conference publication works. Thomas Anderson sets out a fascinating analysis of why this might be so for computer systems conferences, in his 2009 paper “Conference Reviewing Considered Harmful“. Basically, the accept/reject process for conferences that use a peer-review system creates a perverse incentive to researchers to write papers that are just good enough to get accepted, but no better. His analysis is consistent with my own observations – people talk about “the least publishable unit” of research. The net result is a conference full of rather dull papers, where nobody takes risks on more exciting research topics.

There’s an interesting contrast with the geosciences community here, where papers are published in journals rather than conferences. For example, at the AGU and EGU conferences, you just submit an abstract, and various track chairs decide whether to let you present it as a talk in their track, or whether it should appear as a poster. Researchers are only allowed to submit one abstract as first author, which means the conference is really a forum for each researcher to present her best work over the past year, with no strong relationship to the peer-reviewed publication process. This makes for big conferences, and very variable quality presentations. Attendees have to do a little more work in advance to figure out which talks might be worth attending. But the perverse incentive identified by Anderson is missing all together – each presenter is incentivized to present her best work, no matter what stage the research is at.

Which brings me back to ICSE. Next year’s conference chairs have chosen the slogan “Sustainable Software for a Sustainable World” for the conference. An excellent rallying call, but I sincerely hope they can do more with this than most conferences do – such conference slogans are usually irrelevant to the actual conference program, which is invariably business as usual. Of course, the term sustainability has been wildly overused recently, to the point that its in danger of becoming meaningless. So, how could ICSE make it something more than a meaningless slogan?

First, one has to acknowledge that an understanding of sustainability requires some systems thinking, and the ability to analyze multiple interacting systems. The classic definition, due to the Bruntland Commission, is that it refers to humanity’s ability to meet its needs, without compromising the needs of future generations. As Garvey points out, this is entirely inadequate, as it’s impossible to figure out how to balance our resource needs with those of an unknown number of potential future earthlings. A better approach is to break the concept down into sustainability in different, overlapping systems. Sverdrup and Svensson do this by breaking it down to three inter-related concepts: natural sustainability, social sustainability, and economic sustainability. Furthermore, they are hierarchically related: sustainability of social and economic activity are constrained by physical limits such as thermodynamics and mass conservation (e.g. forget a sustained economy if we screw the planet’s climate), and economic sustainability is constrained by social limits such as a functioning civil society.

How does this apply to ICSE? Well, I would suggest applying the sustainability concept to a number of different systems:

  • sustainability of the ICSE community itself, which would include nurturing new researchers, and fixing the problems of perverse incentives in the paper review processes. But this only makes sense within:
  • sustainability of scientific research as a knowledge discovery process, which would include analysis of the kinds of research questions a research community ought to tackle, and how should it engage with society. Here, I think ICSE has some serious re-assessment to do, especially with respect to it’s tendency to reject inter-disciplinary work.
  • sustainability of software systems that support human activity, which would suggest a switch in attention by the ICSE community away from the technical processes of de novo software creation, and towards questions of how software systems actually make life better for people, and how software systems and human activity systems co-evolve. An estimate I heard at the CHASE workshop is that only 20% of ICSE papers make any attempt to address human aspects.
  • sustainability of software development as an economic activity, which suggests a critical look at how existing software corporations work currently, but perhaps more importantly, exploration of new economic models (e.g. open source; end-user programming; software startups; mashups, etc)
  • the role of software in social sustainability, by which I mean a closer look at how software systems help (or hinder) the creation of communities, social norms, social equity and democratic processes.
  • the role of software in natural sustainability, by which I mean green IT topics such as energy-aware computing, as well as the broader role of software in understanding and tackling climate change.

A normal ICSE would barely touch on any of these topics. But I think next year’s chairs could create some interesting incentives to ensure the conference theme becomes more than just a slogan. At the session on SE for the planet that we held at ICSE 2009, someone suggested that in light of the fact that climate change will make everything else unsustainable, ICSE should insist that all submitted papers to future conferences demonstrate some relevance to tackling climate change (which is brilliant, but so radical that we have to shift the Overton window first). A similar suggestion at the one of the CHASE meetings was that all ICSE papers must demonstrate relevance to human & social aspects, or else prove that their research problem can be tackled without this. For ICSE 2012, perhaps this should be changed to simply reject all papers that don’t contribute somehow to creating a more sustainable world.

I think such changes might help to kick ICSE into some semblance of relevancy, but I don’t kid myself that they are likely. How about as a start, a set of incentives that reward papers that address sustainability in one of more of the senses above? Restrict paper awards to such papers, or create a new award structure for this purpose. Give such papers prominence in the program, and relegate other papers to the dead times like right after lunch, or late in the evening. Or something.

But a good start would be to abolish the paper submission process all together, to decouple the conference from the process of publishing peer-reviewed papers. That’s probably the biggest single contribution to making the conference more sustainable, and more relevant to society.

I’m a bit of an Information Visualization junkie. I love good well presented data (I’m a fan of Tufte) and I dislike visualizations that are badly presented and/or misleading. I posted last week about various graphs showing relationships between urban density and transportation fuel consumption, some of which were hideous, some elegant, and some possibly misleading. I bemoaned the lack of access to the raw data, and a lively discussion followed about the believability of the relationship plotted on the graphs.

Yesterday I came across an interesting case, in the leaflet distributed to Torontonians from the city council, showing revenue and expenditure data. From a data visualization point of view, it looks like a series of poor choices were made, and I’m glad someone cared enough to point them out. But when you interpret these choices in the context of a right-wing Mayor who was elected on a tax-cutting, pro-car, anti-transit platform, it would appear these weren’t just mistakes – they were part of deliberate (if subtle) attempt to mislead:

  • The leaflet shows a pie chart of revenue sources (in $billions) along side a pie chart of capital expenditure (in $millions), setting up a false impression that transit projects gobble up the majority of the city’s budget. The deception is enhanced by the fact that the largest segments in each pie are the same colour, and of a similar size. A quick glance therefore leaves the impression that nearly all our property taxes go to the Toronto Transit Commision.
  • The leaflet fails to distinguish between gross and net expenditure. So a bar chart of budget items shows that the TTC (at $1.5 billion) is by far the most expensive item, followed by employment and social services. But the net cost of the TTC to the city is only about $0.5 billion, because most of its costs come from fares, while employment and social services are largely funded by the province. If you look at net costs (which is what most homeowners expect in answer to the question “how does the city spend our property taxes?”), the Police Service is by far the biggest item.

It’s the steady drip drip drip of this kind of misinformation that allows certain politicians to generate support for cutting budgets for transit and social services. Surely we should be investing in the kinds of community programs that reduce crime, so that we can trim that massive policing budget?

Here’s the chart on (gross) expenditures that they used:

20110625fordgraphs4.png

and here’s the chart they should have used:

20110625fordgraphs5.png

Over the past month, we’ve welcomed a number of new researchers to the lab. It’s about time I introduced them properly:

  • Kaitlin Alexander is with us for the summer on a Centre for Global Change Science (CGCS) undergraduate internship. She’s doing her undergrad at the University of Manitoba in applied math, and is well know to many of us already through her blog, ClimateSight. Kaitlin will be studying the software architecture of climate models, and she’s already written a few blog posts about her progress.
  • Trista Mueller is working in the lab for the summer as a part-time undergrad intern – she’s a computer science major at U of T, and a regular volunteer at Hot Yam. Trista is developing a number of case studies for our Inflo open calculator, and helping to shake out a few bugs in the process.
  • Elizabeth Patitsas joins us from UBC, where she just graduated with a BSc honours degree in Integrated Science – she’s working as a research assistant over the summer and will be enrolling in grad school here in September. Elizabeth has been studying how we teach (or fail to teach) key computer science skills, both to computer science students, and to other groups, such as physical scientists. Over the summer, she’ll be developing a research project to identify which CS skills are most needed by scientists, with the eventual goal of building and evaluating a curriculum based on her findings.
  • Fabio da Silva is a professor of computer science at the Federal University of Pernambuco (UFPE) in Brazil, and is joining us this month for a one year sabbatical. Fabio works in empirical software engineering and will be exploring how software teams coordinate their work, in particular the role of self-organizing teams.

I spent a little time this afternoon trying to get hold of data. I guess I have high expectations that the web should deliver what I want instantly; in the old days it would have taken a few days in the library to track down the data sets I needed, and then a few weeks waiting for it on inter-library loan. In some respects, things haven’t changed much, although now it just means you hit the paywall faster. Here’s today’s tale…

It began with a post by George Monbiot on how we’ll have to make cities much more dense if we are to cut down their energy needs. George then tweeted about a fabulous graph from the UNEP which illustrates the point nicely:

In which Toronto holds an interesting position compared to other North American cities. Anyway, someone then pointed out that this data is a little old – it’s based on a classic study by Newman and Kenworthy from the 1980’s. So now the hunt begins: is there an updated version of this anywhere, and if not, can I get hold of the data to create it?

Luke Devlin tweeted out a newer version, published in 2009, based on data from the UITP Mobility in Cities Database, which has data from around the year 2001:

However, this graph is pretty ugly, and has none of the cities labelled. So, methinks that would be easy to fix – all I need is the data. Unfortunately the database (on CD-ROM – how quaint!) costs €1,200. And I’d have to wait for it to arrive. Surely someone has this online for free? No? After all, I only want to use one indicator…

Okay, so the data hunt is on. Population density data is easy to get hold of – wikipedia has plenty of it. In exploring this a little, I find some wikified concerns expressed about the original graph, and a whole can of worms about how exactly you compute population density for a city (tl;dr: it depends where you think the city boundaries are).

A little more googling turns up a fascinating 2003 paper “Transport Energy Use and Greenhouse Gases in Urban Passenger Transport Systems: A Study of 84 Global Cities” (by the same Kenworthy), which has a graph of exactly the data I need:

But of course, it points me back at the same UITP dataset for the actual numbers. Darn.

Then there’s a UNEP report dated March 2011, “Technologies for Climate Change Mitigation – Transport Sector“, which uses the same data, but actually does plot the graph I’m after:

It’s a little better than the previous version, but still doesn’t label the individual cities (which one is Toronto??). And of course, although the report is dated 2011, it’s still the same 2001 dataset from UITP.

So where else might I get data like this? A little more googling and I hit what looks like the jackpot: An extensive list of resources on transportation statistics. Unfortunately, the only one that seems to have the transport data by city is the UITP dataset. Back to that paywall again.

In the meantime, I seem to have launch George Monbiot off into an investigation of the academic publishing racket, exploring why the results of publicly funded research is invariably behind a paywall:

I look forward to reading his blog post on that topic. Meanwhile, I’m off to track down someone on campus who might already have the UITP CD-ROM…

Update 4-Jul-2011: Chris Kennedy sent me his 2009 paper in which he did a detailed analysis for 10 cites, with an update of the density vs transport energy consumption curve. He tells me he has the energy data for more cities, but not the density data, as this is very hard to do consistently. Oh, and silly me – I’d already blogged this, together with Chris’ graph last year. Here’s Chris’s graph. He says “The logarithm of urbanized density has a statistically significant fit (t stat ) -10.26) against the logarithm of GHG emissions from transportation fuels with an R2 of 0.93 (Table 2). The logarithm of average personal income is statistically insignificant (t stat ) -0.35).” (p7299)

Chris also tells me the IEA report on the world’s energy, due out later this year, will chapter on cities, with an update of the graph.

As a followup to my post earlier this week about how dangerous cycling is in Toronto, I decided to take my camera with me on my daily commute. Over two days, I managed to take snaps of the many wonderful uses of bike lanes – it turns out these are incredibly versatile strips of land. Which means the city would be crazy not to maintain them properly, right? Eh? Oh:

Well, anyway. Here’s my ABC of the many wonderful uses of bike lanes.

Bike lanes are for: Ambulances, in case of accidents:

Bike lanes are for: Bi-modal parking, because sidewalks just aren’t big enough:

Bike lanes are for: Council vans, because there’s nowhere to park the fleet:

Bike lanes are for: Deliveries, so every store should have one:

Bike lanes are for: Excavating, to save us digging up car lanes:

Bike lanes are for: Free parking, just right for a quick bit of shopping:

Bike lanes are for: Going around, because we like the scenic route:

Bike lanes are for: Hydro vans, because mobile workshops are cool:

Bike lanes are for: Idiots, who swim against the flow:

Bike lanes are for: Junk piles, because trash is expensive to haul:

Bike lanes are for: Kerb repairs, a safety margin for the crew:

Bike lanes are for: Lorries, although Canadians call them trucks:

Bike lanes are for: Manhole covers, spaced carefully across the lane:

Bike lanes are for: “No stopping” signs, though nobody knows they’re there:

Bike lanes are for: On-kerb parking, which the police just happen to ignore:

Bike lanes are for: Patching practice, because road crews have to learn:

Bike lanes are for: Quantities of dirt, which are just too big for elsewhere:

Bike lanes are for: Rails. Streetcar rails. You never know when you’ll need them:

Bike lanes are for: Spillovers, because building sites are so small:

Bike lanes are for: Timber piles – look how much will fit:

Bike lanes are for: Unexpected doors, that open in your face:

Bike lanes are for: Very large scoops, just waiting to make more holes:

Bike lanes are for: Washrooms, because even cyclists need to pee:

Bike lanes are for: TaXis, they’re out there cruising for fares:

Bike lanes are for: Yellow diggers, and yes, that’s the second one today:

Bike lanes are for: Zooming along, on the few occasions they’re clear:

Note: All photos were taken by me, this week, on my commute to work, except for the taxi, as the one I was trying to snap drove away too quick (see: Taxi photo credit). Click the photos for bigger versions on Flickr.

I thought I’d write a little on cycling in the Toronto this month, partly because is Bike Month, partly because I’ve been reading up on idea of complete streets, and partly because Toronto has just introduced a bike share system, Bixi bikes, and I fear for the consequences of more casual riders on Toronto’s dangerous streets. And because I bike to work everyday and spend most of my ride thinking about how things could be better. Oh, and because getting rid of our car-centred transport system is a key climate mitigation strategy.

Toronto is an awful city to bike in. The city streets are entirely car-centric, with a few bike lanes added in as an afterthought. The thing is, I enjoy biking to work for much the same reasons I enjoy dangerous sports. I like the adrenalin rush, and I like the hyper-focus that’s necessary to survive as a cyclist in Toronto. I’ve been doing it for many, many years, ever since I used to bike across central London as a grad student. I’ve been knocked off my bike twice in 25 years of inner city cycling (once in London, and once in Toronto), with no serious injuries either time, but then I’ve had good training in defensive biking. I ride fast and furious, I know how to own the road, and I know how to anticipate and avoid risks.

But dangerous sports are dangerous. I certainly don’t like the thought of my kids cycling on Toronto’s streets, even though I’d love them to be able to bike to school. The problem is that while the city has been developing a network of cycle lanes, the whole design is wrong. Worse than wrong – I think Toronto’s cycle lanes are more dangerous than the roads that don’t have them. The problem is that they’re formed just by drawing a white line a few feet from the curb. Which makes them an idea space for taxis to stop, delivery vehicles to park, construction crews to dump materials, and so on:

Where parking is allowed, there’s no buffer zone between the parking spaces and the bike line, which means that car doors are a hazzard – a neighbour of mine spent months off work with a broken shoulder recently because a car door opened in front of him.

On streets with bike lanes but no parking, there are “no stopping” signs everywhere, which are universally ignored. This makes cycling much more dangerous – whenever the bike lane is blocked, cyclists have to weave into the main traffic lane, which is now only just wide enough for cars and trucks. Such weaving is more dangerous than cycling along in the main traffic lane all the time. I used to get angry about this, and swear at people who stop their vehicles in the bike lane. But eventually I realized this isn’t the fault of the drivers – it’s the fault of the bike lane design. Delivery vehicles and taxis have to stop frequently, and will always pull to the curb to do so. The bike lane actually encourages this – it’s the ideal space, just out of the main traffic lane, perfect for temporary stops. It’s like the people who designed the bike lanes have no idea about the theory of affordances: what they’ve done is create a strip of road that’s just perfect for drivers to pull over into when they need to stop to pick up passengers or to drop off deliveries.

What we need are physically separated bike lanes. Ones that cannot be used by motor vehicles, because there is a physical barrier to stop them. It’s great to see the Toronto Cyclists Union taking this up as a serious project, and the city has even commissioned a feasibility study, due to be completed later this month.

But that still leaves us with a few other problems to solve. One is that most schools in Toronto are not reachable by bike, because nobody ever considered how to support safe biking to school for kids. The other is that, no matter how the streets are designed, there’s no accounting for stupidity. I’ll leave you with this wonderful video to explain what I mean by stupidity:


3-Way Street from ronconcocacola on Vimeo.

Fabio sent me some pointers to upcoming conferences on IT and climate change:

Pity that two of them coincide! I guess this complements my earlier post on readings in Green IT.

28. May 2011 · 4 comments · Categories: psychology

This xkcd cartoon made me laugh out loud, because it beautifully captures my attitude towards sports. The mouse-over comment on xkcd points out that it applies to financial analysis too, and (more directly), Dungeons and Dragons:

But then I got thinking. Huge numbers of people are fascinated by the narratives that these commentators produce. There’s something compelling about declaring allegiance to one of the weighted random number generators (sports teams, stock picks, etc), selecting which of the narratives to believe based on that allegiance, and then hoping for (and perhaps betting on) which numbers it will produce. Sometimes the numbers turn out the way you’d hoped, and sometimes they don’t, but either way people prefer to believe in the narratives, rather than acknowledge the randomness. There’s a thrill to the process, but at the same time, a strong sense that everyone else’s narratives are delusional.

What if that’s how most people view science? What if they think that scientists are no different from sports commentators and financial analysts, sounding knowledgeable, but mostly just making up explanations based on wishful thinking? What if people believe that scientists’ explanations are just as unreliable as those of sports commentators, and that therefore you can pick what to believe based on tribal affiliations, rather than on, say, the weight of evidence?

Certainly, that’s how many non-scientist commentators approach climate science. Each new paper published is like a new game outcome. The IPCC team may have won a lot of games in the past, put it’s very unpopular among some groups of fans. Results that suggest some other teams are on the rise can be spun into fabulous narratives about how the IPCC team is past its prime, and cruising towards a big upset.

For those who have no idea what science does, the caption for the cartoon might just as well read “All climate science”.

I picked up a fascinating book today – “The Handbook of Sustainability Literacy” edited by Arran Stibbe. It’s a set of short essays on each of a long list of skills needed for thinking about and achieving sustainability. The contents listing makes worthwhile reading on it’s own, covering many of the things I’ve been reading up on for the last few months. I wonder if it’s possible to design an education program that fosters all these skills:

  • ECOCRITICISM – the ability to investigate cultural artefacts from an ecological perspective.
  • OPTIMISATION – the art of personal sufficiency.
  • GROUNDED ECONOMIC AWARENESS – economic awareness grounded in ecological and ethical values.
  • ADVERTISING AWARENESS – the ability to expose advertising discourses that undermine sustainability, and resist them.
  • TRANSITION SKILLS – skills for transition to a post fossil-fuel age.
  • COMMONS THINKING – the ability to envisage and enable a viable future through connected action.
  • EFFORTLESS ACTION – the ability to fulfil human needs effortlessly through working with nature.
  • PERMACULTURE DESIGN – designing our lives with nature as the model.
  • COMMUNITY GARDENING – skills for building community and working within environmental limits.
  • ECOLOGICAL INTELLIGENCE – viewing the world relationally.
  • SYSTEMS THINKING – the ability to recognize and analyse the inter-connections within and between systems.
  • GAIA AWARENESS – awareness of the animate qualities of the Earth.
  • FUTURES THINKING – the ability to envision scenarios for the future and work towards bringing desirable ones into being.
  • VALUES REFLECTION AND THE EARTH CHARTER – the ability to critique the values of an unsustainable society and consider alternatives.
  • SOCIAL CONSCIENCE – the ability to reflect on deeply-held opinions about social justice and sustainability.
  • NEW MEDIA LITERACY – communication skills for sustainability.
  • CULTURAL LITERACY – understanding and skills for culturally appropriate communication.
  • CARBON CAPABILITY – understanding, ability and motivation for reducing carbon emissions.
  • GREENING BUSINESS – the ability to drive environmental and sustainability improvements in the workplace.
  • MATERIALS AWARENESS – the ability to expose the hidden impact of materials on sustainability.
  • APPROPRIATE TECHNOLOGY AND APPROPRIATE DESIGN – the ability to design systems, technologies and equipment in an appropriate way.
  • TECHNOLOGY APPRAISAL – the ability to evaluate technological innovations.
  • COMPLEXITY, SYSTEMS THINKING AND PRACTICE – skills for managing complexity.
  • COPING WITH COMPLEXITY – the ability to manage complex sustainability problems.
  • EMOTIONAL WELLBEING – the ability to research and reflect on the roots of emotional wellbeing.
  • FINDING MEANING WITHOUT CONSUMING – the ability to experience meaning, purpose and satisfaction through non-material wealth.
  • BEING-IN-THE-WORLD – the ability to think about the self in interconnection and interdependence with the surrounding world.
  • BEAUTY AS A WAY OF KNOWING – the redemption of knowing through the experience of beauty.

There’s a few things I’m might add (social networking and social justice spring to mind), and I see they’ve added some additional chapters on the website. But phew, this looks like an extremely valuable book.

Bad news today – we just had a major grant proposal turned down. It’s the same old story – they thought the research we were proposing (on decision support tools for sustainability) was excellent, but criticized, among other things, the level of industrial commitment and our commercialization plans. Seems we’re doomed to live in times where funding agencies expect universities to take on the role of industrial R&D. Oh well.

The three external reviews were very strong. Here’s a typical paragraph from the first review:

I found the overall project to be very compelling from a “need”, potential “payoff’, technical and team perspective. The linkage between seemingly disparate technology arenas–which are indeed connected and synergistic–is especially compelling. The team is clearly capable and has a proven track record of success in each of their areas and as leaders of large projects, overall. The linkage to regional and institutional strengths and partners, in both academic and industrial dimensions, is well done and required for success.

Sounds good huh? I’m reading it through, nodding, liking the sound of what this reviewer is saying. The problem is, this is the wrong review. It’s not a review of our proposal. It’s impossible to tell that from this paragraph, but later on, mixed in with a whole bunch more generic praise, are some comments on manufacturing processes, and polymer-based approaches. That’s definitely not us. Yet I’m named at the top of the form as the PI, along with the title of our proposal. So, this review made it all the way through the panel review process, and nobody noticed it was of the wrong proposal, because most of the review was sufficiently generic that it passed muster on a quick skim-read.

It’s not the first time I’ve seen this happen. It happens for paper reviews for journals and conference. It happens for grant proposals. It even happens for tenure and promotion cases (including both of the last two tenure committees I sat on). Since we started using electronic review systems, it happens even more – software errors and human errors seem to conspire to ensure a worrying large proportion of reviews get misfiled.

Which is why every review should start with a one paragraph summary of whatever is being reviewed, in the reviewer’s own words. This acts as a check that the reviewer actually understood what the paper or proposal was about. It allows the journal editor / review panel / promotions committee to immediately spot cases of mis-filed reviews. And it allows the authors, when they receive the reviews, to get the most important feedback of all: how well did they succeed in communicating the main message of the paper/proposal?

Unfortunately, in our case, correcting the mistake is unlikely to change the funding decision (they sunk us on other grounds). But at least I can hope to use it as an example to improve the general standard of reviewing in the future.

Previously I posted on the first two sessions of the workshop on A National Strategy for Advancing Climate Modeling” that was held at NCAR at the end of last month:

  1. What should go into earth system models;
  2. Challenges with hardware, software and human resources;

    The third session focussed on the relationship between models and data.

    Kevin Trenberth kicked off with a talk on Observing Systems. Unfortunately, I missed part of his talk, but I’ll attempt a summary anyway – apologies if it’s incomplete. His main points were that we don’t suffer from a lack of observational data, but from problems with quality, consistency, and characterization of errors. Continuity is a major problem, because much of the observational system was designed for weather forecasting, where consistency of measurement over years and decades isn’t required. Hence, there’s a need for reprocessing and reanalysis of past data, to improve calibration and assess accuracy, and we need benchmarks to measure the effectiveness of reprocessing tools.

    Kevin points out that it’s important to understand that models are used for much more than prediction. They are used:

    • for analysis of observational data, for example to produce global gridded data from the raw observations;
    • to diagnose climate & improve understanding of climate processes (and thence to improve the models);
    • for attribution studies, through experiments to determine climate forcing;
    • for projections and prediction of future climate change;
    • for downscaling to provide regional information about climate impacts;

    Confronting the models with observations is a core activity in earth system modelling. Obviously, it is essential for model evaluation. But observational data is also used to tune the models, for example to remove known systematic biases. Several people at the workshop pointed out that the community needs to do a better job of keeping the data used to tune the models distinct from the data used to evaluate them. For tuning, a number of fields are used – typically top-of-the-atmosphere data such as net shortwave and longwave radiation flux, cloud and clear sky forcing, and cloud fractions. Also, precipitation and surface wind stress, global mean surface temperature, and the period and amplitude of ENSO. Kevin suggests we need to do a better job of collecting information about model tuning from different modelling groups, and ensure model evaluations don’t use the same fields.

    For model evaluation, a number of integrated score metrics have been proposed to summarize correlation, root-mean-squared (rms) error and variance ratios – See for example, Taylor 2001Boer and Lambert 2001Murphy et al, 2004Reichler & Kim 2008.

    But model evaluation and tuning aren’t the only ways in which models and data are brought together. Just as important is re-analysis, where multiple observational datasets are processed through a model to provide more comprehensive (model-like) data products. For this, data assimilation is needed, whereby observational data fields are used to nudge the model at each timestep as it runs.

    Kevin also talked about forward modelling, a technique in which the model used to reproduce the signal that a particular instrument would record, given certain climate conditions. Forward modelling is used for comparing models with ground observations and satellite data. In much of this work, there is an implicit assumption that the satellite data are correct, but in practice, all satellite data have biases, and need re-processing. For this work, the models need good emulation of instrument properties and thresholds. For examples, see: Chepfer, Bony et al, 2010Stubenrauch & Kinne 2009.

    He also talked about some of the problems with existing data and models:

    • nearly all satellite data sets contain large spurious variability associated with changing instruments and satellites, orbital decay/drift, calibration, and changing methods of analysis.
    • simulation of the hydrological cycle is poor, especially in the intertropical convergence zone (ITCZ). Tropical transients are too weak, runoff and recycling is not correct, and the diurnal cycle is poor.
    • there are large differences between datasets for low cloud (see Marchand at al 2010)
    • clouds are not well defined. Partly this is a problem of sensitivity of instruments, compounded by the difficulty of distinguishing between clouds and aerosols.
    • Most models have too much incoming solar radiation in the southern oceans, caused by too few clouds. This makes for warmer oceans and diminished poleward transport, which messes up storm tracking and analysis of ocean transports.

    What is needed to support modelling over the next twenty years? Kevin made the following recommendations:

    • Support observations and development into climate datasets.
    • Support reprocessing and reanalysis.
    • Unify NWP and climate models to exploit short term predictions and confront the models with data.
    • Develop more forward modelling and observation simulators, but with more observational input.
    • Targeted process studies such as GEWEX and analysis of climate extremes, for model evaluation.
    • Target problem areas such as monsoons and tropical precipitation.
    • Carry out a survey of fields used to tune models.
    • Design evaluation and model merit scoring based on fields other than those used in tuning.
    • Promote assessments of observational datasets so modellers know which to use (and not use).
    • Support existing projects, including GSICS, SCOPE-CM, CLARREO, GRUAN,

    Overall, there’s a need for a climate observing system. Process studies should not just be left to the observationists – we need the modellers to get involved.

    The second talk was by Ben Kirtman, on “Predictability, Credibility, and Uncertainty Quantification“. He began by pointing out that there is ongoing debate over what predictability means. Some treat it as an inherent property of the climate system, while others think of it as a model property. Ben distinguished two kinds of predictability:

    • Sensitivity of the climate system to initial conditions (predictability of the first kind);
    • Predictability of the boundary forcing (predictability of the second kind).

    Predictability is enhanced by ensuring specific processes are included. For example, you need to include the MJO if you want to predict ENSO. But model-based estimates of predictability are model dependent. If we want to do a better job of assessing predictability, we have to characterize model uncertainty, and we don’t know how to do this today.

    Good progress has been made on quantifying initial condition uncertainty. We have pretty of good ideas for how to probe this (stochastic optimals, bred vectors, etc.) using ensembles with perturbed initial conditions. But from our understanding of chaos theory (e.g. see the Lorenz attractor), predictability depends on which part of the regime you’re in, so we need to assess the predictability for each particular forecast.

    Uncertainty in external forcing include uncertainties in both the natural and anthropogenic forcing; however this is becoming less of an issue in modelling, as these forcings are better understood. Therefore, the biggest challenge is in quantifying uncertainties in model formulation. These arise because of the discrete representation of climate system, the use of parameterization of subgrid processes, and because of missing processes. Current approaches can be characterized as:

    • a posteriori techniques, such as the multi-model ensembles of opportunity used in IPCC assessments, and perturbed parameters/parameterizations, as used in climateprediction.net.
    • a priori techniques, where we incorporate uncertainty as the model evolves. The idea is that uncertainty is in subscale processes and missing physics. Model this non-locally and stochastically. E.g. backscatter, interactive ensembles to incorporate uncertainty in the coupling.

    The term credibility is even less well defined. Ben asked his students what they understood by the term, and they came up with a simple answer: credibility is the extent to which you use the best available science [which corresponds roughly to my suggestion of what model validation ought to mean]. In the literature, there are a number of other way of expressing credibility:

    • In terms of model bias. For example, Lenny Smith offers a Temporal (or spatial) credibility ratio, calculated as the ratio of the smallest timestep in the model to the smallest duration over which a variable has to be averaged before it compares favourably with observations. This expresses how much averaging over the temporal (or spatial) scale you have to do to make the model look like the data.
    • In terms of whether the ensembles bracket the observations. But the problem here is that you can always pump up an ensemble to do this, and it doesn’t really tell you about probabilistic forecast skill.
    • In terms of model skill. In numerical weather prediction, it’s usual to measure forecast quality using some specific skill metrics.
    • In terms of process fidelity – how well the processes represented in the model capture what is known about those processes in reality. This is a reductionist approach, and depends on the extent to which specific processes can be isolated (both in the model, and in the world).
    • In terms of faith – for example, the modellers’ subjective assessment of how good their model is.

    In the literature, credibility is usually used in a qualitative way to talk about model bias. Hence, in the literature, model bias is roughly synonymous with inverse of credibility. However, in these terms, the models currently have a major credibility gap. For example, Ben showed the annual mean rainfall from a long simulation of CESM1, showing bias with respect to GPCP observations. These show the model struggling to capture the spatial distribution of sea surface temperature (SST), especially in equatorial regions.

    Every climate model has a problem with equatorial sea surface temperatures (SST). A recent paper, Anagnostopoulos et al 2009 makes a big deal of this, and is clearly very hostile to climate modelling. They look at regional biases in temperature and precipitation, where the models are clearly not bracketing observations. I googled the Anagnostopooulos paper while Ben was talking – The first few pages of google hits are dominated by denialist website proclaiming this as a major new study demonstrating the models are poor. It’s amusing that this is treated as news, given that such weaknesses in the models are well known within the modelling community, and discussed in the IPCC report. Meanwhile the hydrologists at the workshop tell me that it’s a third-rate journal, so none of them would pay any attention to this paper.

    Ben argues that these weaknesses need to be removed to increase model credibility. This argument seems a little weak to me. While improving model skill and removing biases are important goals for this community, they don’t necessarily help with model credibility in terms of using the best science (because often replacing an empirically derived parameterization with one that’s more theoretically justified will often reduce model skill). More importantly, those outside the modeling community will have their own definitions of credibility, and they’re unlikely to correspond to those used within the community. Some attention to the ways in which other stakeholders understand model credibility would be useful and interesting.

    In summary, Ben identified a number of important tensions for climate modeling. For example, there are tensions between:

    • the desire to measure prediction skill vs. the desire to explore the limits of predictability;
    • the desire to quantify uncertainty, vs. the push for more resolution and complexity in the models;
    • a priori vs. a posteriori methods of assessing model uncertainty.
    • operational vs. research activities (Many modellers believe the IPCC effort is getting a little out of control – it’s a good exercise, but too demanding on resources);
    • weather vs climate modelling;
    • model diversity vs critical mass;

    Ben urged the community to develop a baseline for climate modelling, capturing best practices for uncertainty estimation.

    During a break in the workshop last week, Cecilia Bitz and I managed to compare notes on our undergrad courses. We’ve both been exploring how to teach ideas about climate modeling to students who are not majoring in earth sciences. Cecilia’s course on Weather and Climate Prediction is a little more advanced than mine, as she had a few math and physics pre-requisites, while mine was open to any first year students. For example, Cecilia managed to get the students to run CAM, and experiment with altering the earth’s orbit. It’s an interesting exercise, as it should lead to plenty of insights into connections between different processes in the earth’s system. One of the challenges is that earth system models aren’t necessarily geared up for this kind of tinkering, so you need good expertise in the model being used to understand which kinds of experiments are likely to make sense. But even so, I’m keen to explore this further, as I think the ability to tinker with such models could be an important tool for promoting a better understanding of how the climate system works, even for younger kids

    Part of what’s missing is elegant user interfaces. EdGCM is better, but still very awkward to use. What I really want is something that’s as intuitive as Angry Birds. Okay, so I’m going to have to compromise somewhere – nonlinear dynamics are a bit more complicated than the flight trajectories of avian slingshot.

    But that’s not all – someone needs to figure out what kinds of experiments students (and school kids) might want to perform, and provide the appropriate control widgets, so they don’t have mess around with code and scripts. Rich Loft tells me there’s a project in the works to do something like this with CESM – I’ll looking forward to that! In the meantime, Rich suggested two examples of simple simulations of dynamical systems that get closer to what I’m looking for:

    • The Shodor Disease model that lets you explore the dynamics of epidemics, with people in separate rooms, where you can adjust how much they can move between rooms, how the disease works, and whether immunization is available. Counter-intuitive lesson: crank up the mortality rate to 100% and (almost) everyone survives!
    • The Shodor Rabbits and Wolves simulation, which lets you explore population dynamics of predators and prey. Counter-intuitive lesson: double the lifespan of the rabbits and they all die out pretty quickly!

    In the last post, I talked about the opening session at the workshop on “A National Strategy for Advancing Climate Modeling”, which focussed on the big picture questions. In the second session, we focussed on the hardware, software and human resources challenges.

    To kick off, Jeremy Kepner from MIT called in via telecon to talk about software issues, from his perspective working on Matlab tools to support computational modeling. He made the point that it’s getting hard to make scientific code work on new architectures, because it’s increasingly hard to find anyone who wants to do the programming. There’s a growing gap between the software stacks used in current web and mobile apps, gaming, and so on, and that used in scientific software. Programmers are used to having new development environments and tools, for example for developing games for Facebook, and regard scientific software development tools as archaic. This means it’s hard to recruit talent from the software world.

    Jeremy quipped that software is an evil thing – the trick is to get people to write as little of it as possible (and he points out that programmers make mistakes at the rate of one per 1,000 lines of code). Hence, we need higher levels of abstraction, with code generated automatically from higher level descriptions. Hence, an important question is whether it’s time to abandon Fortran. He also points out that programmers believe they spend most of their time coding, but in fact, coding is a relatively small part of what they do. At least half of their time is testing, which means that effort to speed up the testing process gives you the most bang for the buck.

    Ricky Rood, Jack Fellows, and Chet Koblinsky then ran a panel on human resources issues. Ricky pointed out that if we are to identify shortages in human resources, we have to be clear about whether we mean for modeling vs. climate science vs. impacts studies vs. users of climate information, and so on. The argument can be made that in terms of absolute numbers there are enough people in the field, but the problems are in getting an appropriate expertise mix / balance, having people at the interfaces between different communities of expertise, a lack of computational people (and not enough emphasis on training our own), and management of fragmented resources.

    Chet pointed out that there’s been a substantial rise in the number of job postings using the term “climate modelling” over the last decade. But there’s still a widespread perception is that there aren’t enough jobs (i.e. more grad students being trained than we have positions for). There are some countervailing voices – for example Pielke argues that universities will always churn out more than enough scientists to support their mission, and there’s a recent BAMs article that explored the question “are we training too many atmospheric scientists?“. The shortage isn’t in the number of people being trained, but in the skills mix.

    We covered a lot of ground in the discussions. I’ll cover just some of the highlights here.

    Several people observed that climate model software development has diverged from mainstream computing. Twenty years ago, academia was the centre of the computing world. Now most computing is in commercial world, and computational scientists have much less leverage than we used to. This means that some tools we rely on might no longer be sustainable. E.g. fortran compilers (and autogenerators?) – fewer users care about these, and so there is less support for transitioning them to new architectures. Climate modeling is a 10+ year endeavour, and we need a long-term basis to maintain continuity.

    Much of the discussion focussed on anticipated disruptive transitions in hardware architectures. Whereas in the past, modellers have relied on faster and faster processors to deliver new computing capacity, this is coming to an end. Advances in clock speed have tailed off, and now its  massive parallelization that delivers the additional computing power. Unfortunately, this means the brute force approach of scaling up current GCM numerical methods on a uniform grid is a dead-end.

    As Bryan Lawrence pointed out, there’s a paradigm change here: computers no longer compute, they produce data. We’re entering an era where CPU time is essentially free, and it’s data wrangling that forms the bottleneck. Massive parallelization of climate models is hard because of the volume of data that must be passed around the system. We can anticipated 1-100 exabyte scale datasets (i.e. this is the size not of the archive, but of the data from a single run of an ensemble). It’s unlikely than any institution will have the ability to evolve their existing codes into this reality.

    The massive parallelization and data volumes also bring another problem. In the past, climate modellers have regarded bit-level reproducibility of climate runs to be crucial, partly because reproducing a run exactly is considered good scientific practice, and partly because it allows many kinds of model test to be automated. The problem is, at the scales we’re talking about, exact bit reproducibility is getting hard to maintain. When we scale up to millions of processors, and terabyte data sets, bit-level failures are frequent enough that exact reproducibility can no longer be guaranteed – if a single bit is corrupted during a model run, it may not matter for the climatology of the run, but it will mean exact reproducibility is impossible. Add to this the fact that in the future, CPUs are likely to be less deterministic, then, as Tim Palmer argued at the AGU meeting, we’ll be forced to fundamentally change our codes, and therefore, maybe we should take the opportunity to make the models probabilistic.

    One recommendation that came out of our discussions is to consider a two track approach for the software. Now that most modeling centres have finished their runs for the current IPCC assessment (AR5), we should plan to evolve current codes towards the next IPCC assessment (AR6), while starting now on developing entirely new software for AR7. The new codes will address i/o issues, new solvers, etc.

    One of the questions the committee posed to the workshop was the potential for hardware-software co-design. The general consensus was that it’s not possible in current funding climate. But even if the funding was available, it’s not clear this is desirable, as the software has much longer useful life than any hardware. Designing for specific hardware instantiations tends to bring major liabilities, and (as my own studies have indicated) there seems to be an inverse correlation between availability of dedicated computing resources and robustness/portability of the software. Things change in climate models all the time, and we need the flexibility to change algorithms, refactor software, etc. This means FPGAs might be a better solution. Dark silicon might push us in this direction anyway.

    Software sharing came up as an important topic, although we didn’t talk about this as much as I would have liked. There seems to be a tendency among modelers to assume that making the code available is sufficient. But as Cecelia Deluca pointed out, from the ESMF experience, community feedback and participation is important. Adoption mandates are not constructive – you want people to adopt software because it works better. One of the big problems here is understandability of shared code. The learning curve is getting bigger, and code sharing between labs is really only possible with a lot of personal interaction. We did speculate that auto-generation of code might help here, because it forces the development of higher level language to describe what’s in a climate model.

    For the human resources question, there was a widespread worry that we don’t have the skills and capacity to deal with anticipated disruptive changes in computational resources. There is a shortage of high quality applicants for model development positions, and many disincentives for people to pursue such a career: the long publication cycle, academic snobbery, and the demands of the IPCC all make model development an unattractive career for grad students and early career scientists. We need a different reward system, so that contributions to the model are rewarded.

    However, it’s also clear that we don’t have enough solid data on this – just lots of anecdotal evidence. We don’t know enough about talent development and capacity to say precisely where the problems are. We identified three distinct roles, which someone amusingly labelled: diagnosticians (who use models and model output in their science), perturbers (who explore new types of runs by making small changes to models) and developers (who do the bulk of model development). Universities produce most of the first, a few of the second, and very few of the third. Furthermore, model developers could be subdivided between people who develop new parameterizations and numerical analysts, although I would add a third category: developers of infrastructure code.

    As well as worrying about training of a new generation of modellers, we also worried about whether the other groups (diagnosticians and perturbers) would have the necessary skillsets. Students are energized by climate change as a societal problem, even if they’re not enticed by a career in earth sciences. Can we capitalize on this, through more interaction with work at the policy/science interface? We also need to make climate modelling more attractive to students, and to connect them more closely with the modeling groups. This could be done through certificate programs for undergrads to bring them into modelling groups, and by bringing grad students into modelling centres in their later grad years. To boost computational skills, we should offer training in earth system science to students in computer science, and expand training for earth system scientists in computational skills.

    Finally, let me end with a few of the suggestions that received a very negative response from many workshop attendees:

    • Should the US be offering only one center’s model to the IPCC for each CMIP round? Currently every major modeling center participates, and many of the centres complain that it dominates their resources during the CMIP exercise. However, participating brings many benefits, including visibility, detailed comparison with other models, and a pressure to improve model quality and model documentation.
    • Should we ditch Fortran and move to a higher level languages? This one didn’t really even get much discussion. My own view is that it’s simply not possible – the community has too much capital tied up in Fortran, and it’s the only language everyone knows.
    • Can we incentivize a mass participation in climate modeling, like the “develop apps for the iphone”? This is an intriguing notion, but one that I don’t think will get much traction, because of the depth of knowledge needed to do anything useful at all in current earth system modeling. Oh, and we’d probably need a different answer to the previous question, too.

    This week I’ve been attending a workshop at NCAR in Boulder, to provide input to the US National Acadamies committee on “A National Strategy for Advancing Climate Modeling”. The overall goal is to:

    “…develop a strategy for improving the nation’s capability to accurately simulate climate and related Earth system changes on decadal to centennial timescales. The committee’s report is envisioned as a high level analysis, providing a strategic framework to guide progress in the nation’s climate modeling enterprise over the next 10-20 years”

    The workshop has been fascinating, addressing many of the issues I encountered on my visits to various modelling labs last year – how labs are organised, what goes into the models, how they are tested and validated, etc. I now have a stack of notes from the talks and discussions, so I’ll divide this into several blog posts, corresponding to the main themes of the workshop:

    The full agenda is available at the NRC website, and they’ll be posting meeting summaries soon.

    The first session of the workshop kicked off with the big picture questions: what earth system processes should (and shouldn’t) be included in future earth system models, and what relationship should there be with the models used for weather forecasting?To get the discussion started, we had two talks, from Andrew Weaver (University of Victoria) and Tim Palmer (U Oxford and ECMWF).

    Andrew Weaver’s talk drew on lessons learned from climate modelling in Canada to argue that we need flexibility to build different types of model for different kinds of questions. Central to his talk was a typology of modelling needs, based on the kinds of questions that need addressing:

    • Curiosity-driven research (mainly done in the universities). For example:
      • Paleo-climate (e.g. what are the effects of a Heinrich event on African climate, and how does this help our understanding of the migration of early humans from Africa?)
      • Polar climates (e.g. what is the future of the greenland ice sheet?)
    • Policy and decision making questions (mainly done by the government agencies). These can be separated into:
      • Mitigation questions (e.g. how much can we emit and still stabilize at various temperature levels?)
      • Adaptation questions (e.g. infrastructure planning over things like water supplies, energy supply and demand)

    These diverse types of question place different challenges on climate modelling. From the policy-making point of view, there is a need for higher resolution and downscaling for regional assessments, along with challenges of modelling sub-grid scale processes (Dave Randall talked more about this in a later session). On the other hand, for paleo-climate studies, we need to bring additional things into the models, such as representing isotopes, ecosystems, and human activity (e.g. the effect on climate of the switch from hunter-gatherer to agrarian society).

    This means we need a range of different models for different purposes. Andrew argues that we should design a model to respond to a specific research question, rather than having one big model that we apply to any problem. He made a strong case for a hierarchy of models, describing his work with EMICs like the UVic model, which uses a simple energy/moisture balance model for the atmosphere, coupled with components from other labs for sea ice, ocean, vegetation, etc. He raised a chuckle when he pointed out that neither EMICs nor GCMs are able to get the greenland icesheet right, and therefore EMICS are superior because they get the wrong answer faster. There is a serious point here – for some types of question, all models are poor approximations, and hence, their role is to probe what we don’t know, rather than to accurately simulate what we do know.

    Some of problems faced in the current generation of models will never go away: clouds and aerosols, ocean mixing, precipitation. And there are some paths we don’t want to take, particularly the idea of coupling general equilibrium economic models with earth system models. The latter may be very sexy, but it’s not clear what we’ll learn. The uncertainties in the economics models are so large that you can get any answer you like, which means this will soak up resources for very little knowledge gain. We should also resist calls to consolidate modelling activities at just one or two international centres.

    In summary, Andrew argued we are at a critical juncture for US and international modelling efforts. Currently the field is dominated by a race to build the biggest model with the most subcomponents for the SRES scenarios. But a future priority must be on providing information needed for adaptation planning – this has been neglected in the past in favour of mitigation studies.

    Tim Palmer’s talk covered some of the ideas on probabilistic modeling that he presented at the AGU meeting in December. He began with an exploration of why better prediction capability is important, adding a third category to Andrew’s set of policy-related questions:

    • mitigation – while the risk of catastrophic climate change is unequivocal, we must reduce the uncertainties if we are ever to tackle the indifference to significant emissions cuts; otherwise humanity is heading for utter disaster.
    • adaptation – spending wisely on new infrastructure depends on our ability to do reliable regional prediction.
    • geo-engineering – could we ever take this seriously without reliable bias-free models?

    Tim suggested two over-riding goals for the climate modeling community. We should aspire to develop ab initio models (based on first principles) whose outputs do not differ significantly from observations; and we should aspire to improve probabilistic forecasts by reducing uncertainties to the absolute minimum, but without making the probabilistic forecasts over-confident.

    The main barriers to progress are limited human and computational resources, and limited observational data. Earth System Models are complex – there is no more complex problem in computational science – and great amounts of ingenuity and creativity are needed. All of the directions we wish to pursue – great complexity, higher resolution, bigger ensembles, longer paleo runs, data assimilation – demand more computational resources. But just as much, we’re hampered by an inability to get the maximum value from observations. So how do we use observations to inform modelling in a reliable way?

    Are we constrained by our history? Ab initio climate models originally grew out of numerical weather prediction, but rapidly diverged. Few climate models today are viable for NWP. Models are developed at the institutional level, so we have diversity of models worldwide, and this is often seen as a virtue – the diversity permits the creation of model ensembles, and rivalry between labs ought to be good for progress. But do we want to maintain this? Are the benefits as great as they are often portrayed? This history, and the institutional politics of labs, universities and funding agencies provide an unspoken constraint on our thinking. We must be able to identify and separate these constraints if we’re going to give taxpayers best value for money.

    Uncertainty is a key issue. While the diverisity of models often cited as measure of confidence in ensemble forecasts, this is very ad hoc. Models are not designed to span uncertainty in representation of physical processes in any systematic way, so the ensembles are ensembles of opportunity. So if the models agree, how can we be sure this is a measure of confidence? An obvious counter-example is in economics, where the financial crisis happened despite all economics models agreeing.

    So can we reduce uncertainty if we had better models? Tim argues that seamless prediction is important. He defines it as “bringing the constraints and insights of NWP into the climate arena”. The key idea is to be able to use the same modeling system to move seamlessly between forecasts at different scales, both temporally and spatially. In essence, it means unifying climate and weather prediction models. This unification brings a number of advantages:

    • It accelerates model development and reduces model biases, by exploring model skill on shorter, verifiable timescales.
    • It helps to bridge between observations and testing strategies, using techniques such as data assimilation.
    • It brings the weather and climate communities together
    • It allows for cross-over of best practices.

    Many key climate amplifiers are associated with fast timescale processes, which are best explored in NWP mode. This approach also allows us to test the reliability of probabilistic forecasts, on weekly, monthly and seasonal timescales. For example, reliability diagrams can be used to explore the reliability of a whole season of forecasts. This is done by subsampling the forecasts, taking, for example, all the forecasts that said it would rain with 40% probability, and checking that it did rain 40% of the time for this subsample.

    Such a unification also allows a pooling of research activity in universities, labs, and NWP centres to make progress on important ideas such as stochastic parameterizations and data assimilation. Stochastic parameterizations are proving more skilful in NWP than multi-model ensembles on monthly timescales. The ideas are still in their infancy, but there is potential to pool research activity in universities, labs, and NWP centres. Data assimilation provides a bridge between modelling and observational world (As an example, Rodwell and Palmer were able to rule out high climate sensitivities in the climateprediction.net data using assimilation with the 6-hour ECMWF observational data).

    In contrast to Andrew’s argument to avoid consolidation, Tim argues that consolidation of effort at the continental (rather than national) level would bring many benefits. He cites the example of the Aerobus, where European countries pooled their efforts because aircraft design had become too complex for any individual nation to go it alone. The Aerobus approach involves a consortium, allowing for specialization within each country.

    Tim closed with a recap of an argument he made in a recent Physics World article. If a group of nations can come together to fund the Large Hadron Collider, then why can’t we come together to fund a computing infrastructure for climate computing? If climate is the number one problem facing the world, then why don’t we have number one supercomputer, rather than being around 50 in the TOP500.

    Following these talks, we broke off into discussion groups to discuss a set of questions posed by the committee. I managed to capture some of the key points that emerged from these discussions – the committee report will present a much fuller account.

    • As we expand earth system models to include ever more processes, we should not lose sight of the old unsolved problems, such as clouds, ice sheets, etc.
    • We should be very cautious about introducing things into the models that are not tied back to basic physics. Many people in the discussions commented that they would resist human activities being incorporated into the models, as this breaks the ab initio approach. However, others argued that including social and ecosystem processes is inevitable, as some research questions demand it, and so we should  focus on how it should be done. For example, we should start with places where 2-way interaction is the largest. E.g. land use, energy feedbacks, renewable energy.
    • We should be able to develop an organized hierarchy of models, with traceable connections to one another.
    • Seamless prediction means same within the same framework/system, but not necessarily same model.
    • As models will become more complex, we need to be able to disentangle different kinds of complexity – for example the complexity that derives from emergent behaviour vs. number of processes vs. number of degrees of freedom.
    • The scale gap between models and observations is disappearing. Model resolution is increasing exponentially, while observational capability increasing only  polynomially. This will improve our ability to test models against observations.
    • There is a great ambiguity over what counts as an Earth System Model. Do they include regional models? What level of coupling defines an ESM?
    • There are huge challenges in bringing communities together to consolidate efforts, even on seemingly simple things like agreeing terminology.
    • There’s a complementary challenge to the improvement of ESMs: how can we improve the representation of climate processes in other kinds of model?
    • We should pay attention to user-generated questions. In particular, the general public doesn’t feel that climate modelling is relevant to them, which partly explains problems with lack of funding.

    Next up: challenges arising from hardware, software and human resource limitations…

    19. April 2011 · 2 comments · Categories: politics

    For those outside Canada, in case you haven’t heard, we’re in the middle of a general election. Canada has a parliamentary system, modelled after the British one, with a first-past-the-post system for electing representatives (members of parliament), where party with the most seats after the election is invited to form a government, and its leader to become Prime Minister. For the last few parliaments we’ve had minority governments, first Liberal, then Conservative.

    Somewhere along the way, many people just stopped voting: from turnouts in the high 70s back in the 60’s, we’ve had 64.7% and then 58.8% turnout respectively in the last two elections – the last being the lowest turnout ever. There maybe many different reasons for this lack of enthusiasm, although listening to the main parties whining about each other during this election, it’s not hard to see why so many people just don’t bother. But one thing is clear: young people are far less likely to vote than any other age group.

    So it was great to see last week Rick Mercer with a brilliant call for young voters to use their votes to “scare the hell out of the people who run this country”:

    And his message seems to have resonated. Students on campuses across the country have been using social networking to organise vote mobs, making videos along the way as they challenge others to do the same. But here’s the interesting thing. The young people of this country have a very different set of preferences to the general population:

    Just look at how the projected composition of parliament would look it it were up to the youngsters: the Liberals and the Green Party virtually neck-and-neck for most votes, and instead of the greens being shut out of parliament, they’d hold 43 seats! Of course, the projected seat count also throws into sharp focus what’s wrong with our current voting system: the Bloq, with lowest share of the vote of any of the parties would still hold 60 seats. And the Liberals with just 2% more of the votes than the greens would still get more than twice as many seats. Nevertheless, I like this picture much more than the parliaments we’ve had in the last few elections.

    So, if you’re eligible to vote, and you’re anywhere around half my age, make my day – help change our parliament for the better!