This week, Ontario’s new Feed-in Tariff (FIT) program kicks in. The program sets specific prices that the province will pay to people who develop their own renewable power sources and sell the energy back to the grid. The key idea is that it sets up a guaranteed return on investment for people to build renewable capacity, and at a premium price, too.

The prices are set at different levels for different types of power generation and for different sizes of installations, with each price point designed to make it attractive for people to invest (with presumably some weighting in favour of the power mix the province would like to aim for). For example, a homeowner who puts solar panels on the roof will be paid 80c per kilowatt hour ($0.80/kWh), with the price guaranteed for 20 years. That’s for installations lower than 10kW; the price goes down for bigger installations (e.g. for 44c/kWh for rooftop solar larger than 500kW).

Current electricity prices in Ontario are around $0.08/Kwh, so the province is paying 10 times the current market rate for small-scale solar generation. Which makes is a pretty major subsidy. However, the entire program is intended to be revenue neutral. The creation of a large network of small suppliers may prevent the province having to build so many new power stations (the province recently turned down bids of $26 billion for new nuclear plants), and allow it phase out the existing coal plants within the next few years.

So what does this mean for the homeowner? A typical household solar installation will be well below 10kW. I grabbed a few ballpark figures from the web. A small household solar installation might generate about 12kWh per day, i.e. about $10 per day, or about $3,500 per year at the FiT rate; while the average household consumption is about 12,000kWh per year, or about $1,000 at current market prices. So the panels will pay for themselves within a few years, and then become a source of revenue!

The idea of a Feed-in Tariff program isn’t new – they’ve worked well in Europe for a number of years, and indeed the province of Ontario has had one in place since 2006. However the old program was criticised for setting rates too low, especially for small-scale generation; the new program increases the rates dramatically – for example the new small scale solar rate is twice the old rate.

Full details of the new program are at the Ontario Power Authority’s site.

I’m teaching our introductory software engineering course this term, for which the students will be working on a significant software development project over the term. The main aim of the course is to get the students thinking about and using good software development practices and tools, and we organise the term project as an agile development effort, with a number of small iterations during the term. The students have to figure out for  themselves what to build at each iteration.

For a project, I’ve challenged the students to design new uses for the Canadian Climate Change Senarios Network. This service makes available the data on possible future climate change scenarios from the IPCC datasets, for a variety of end users. The current site allows users to run basic queries over the data set, and have the results returned either as raw data, or in a variety of visualizations. The main emphasis is on regional scenarios for Canada, so the service offers some basic downscaling, and ability to couple the scenarios with other regional data sources, such as data from weather monitoring stations in the region. However, to use the current service, you need to know quite a bit about the nature of the data: it asks you which models you’re interested in; which years you want data for (assumes you know something about 30-year averages); which scenarios you want (assumes you know something about the standard IPCC scenarios); which region you want (in latitude and longitude); and which variables you want (assumes you know something about what these variables measure). The current design reflects the needs of the primary user group for which the service was developed – (expert) researchers working on climate impacts and adaptation.

The challenge for the students on my course is to extend the service for new user groups. For example, farmers who want to know something about likely effects of climate change on growing seasons, rainfall and heat stress in their local area. High school students studying climate and weather. Politicians who want to understand what the latest science tells us about the impacts of climate change on the constituencies they represent. Activists who want to present a simple clear message to policymakers about the need for policy changes. etc.

I have around 60 students on the course, working in teams of 4. I’m hoping that the various teams will come up with a variety of ideas for how to make this dataset useful to new user groups, and I’ve challenged them to be imaginative. But more suggestions are always welcome…

This post by Paul Gilding sums up my experience very well:

Some days my head hurts, as I shift between what feels like two parallel universes in the climate change debate. First I have these conversations with world-class scientists who calmly lay out the scientific view of the various risks posed by climate change and their relative scale and likelihoods. They tell me the science says it is almost certain the impacts will be serious and destabilising for our society and our economy. The science also describes a lower level of risk – which they find hard to quantify but generally say between 10% and 50% – that the impacts of climate change will be catastrophic, perhaps even civilisation threatening. This could include widespread famine, war and economic collapse. Not certain, but a reasonable possibility.

It is very clear when you listen to these scientists and read their peer-reviewed reports that, on any calm and rational analysis, we should be preparing for a carbon reduction war. Yes, a war – with all that implies about focus, effort and sacrifice. The threat posed is, after all, a “clear and present danger” and the response should be strong, global and immediate. This should be a ‘whatever it takes’ moment.

Then I shift into the parallel universe.

I spend time in corporate boardrooms and listen to the analysis of business executives who explain how we mustn’t damage the economy by “over-reacting”….

Go read the rest.

The Global Campaign for Climate Action is an umbrella organisation, based in Montreal, which aims to coordinate many diverse environmentalist and science groups (including Greenpeace, WWF, Union of Concerned Scientists, Oxfam, 350.org, and many others) to focus attention on the need for an ambitious, fair and binding climate treaty at the Copenhagen talks in December. Their campaign leading up to the Copenhagen meeting is called TckTckTck, and it promises a bold series of actions over the next few months.

The first of these is next week: Global Wake Up Call (which nicely fits with the “sleepwalking into disaster” idea), and ties in with the premier of The Age of Stupid on Sept 21. The idea is to coordinate the sound of bells and telephones ringing around the world at 12:18pm on Monday, as the wake up call. There’s quite a few in Toronto – including one at Dundas Square. I’ve no idea if this flashmob style protesting works, but I guess there’s one way to find out. Anyone fancy a walk down to Dundas Square on Monday lunchtime?

Update: The Age of Stupid is being screened at the Royal Theatre on Monday night, 7pm.

Here’s an interesting competition (with cash prizes), organized by the Usability Professionals’ Association, to develop a new concept or product, with user-centred design principles, that aims to cut energy consumption or reduce pollution.

And here’s another: A Video Game Creation Contest to create a playable video game that uses earth observations to help address environmental problems.

I’ve just been browsing the sessions for the AGU Fall Meeting, to be held in San Francisco in December. Abstracts are due by September 3. The following sessions caught my attention:

Plus some sessions that sound generally interesting:

I’ve been tasked with identifying people and initiatives across campus that are involved in Digital Media and Climate Change/Environment. It’s part of a push by the University for greater funding for digital media research. And as everyone seems to interpret the term digital media differently, I’m going to give it the broadest possible interpretation: if it involves doing things with computers (either as a primary research tool or as an object of study), it counts as digital media. Here’s my list of faculty across the University who are doing relevant research. Feel free to suggest more people, or to rearrange my categories…

Understanding Climate Change through Earth Systems Modeling

Impacts of Climate Change and Adaptation

Earth Systems Management (as in: how we manage forests, water supplies, land use, etc)

Sustainable design and energy management (e.g. architectural design, urban planning, etc)

Sustainable Transportation Systems

Geographical Information Systems (GIS) and Environmental Informatics

Policy and Decision Making

For sociologists, a strong call to action in the report of an NSF sponsored workshop on Sociological Perspectives on Global Climate Change. Like the APA report I wrote about earlier, it’s a call to action, covering the key research challenges for the field, and addressing the barriers that might prevent researchers participating in such research. Among the recommendations are better data collection on organisational and community behaviour relevant to climate change, and better inter-disciplinary links:

“…social scientists are seldom consulted except as an afterthought in natural science and engineering research projects […and…] social scientists tend not to seek out collaborations with natural scientists and engineers and often are uninformed about major research programs on climate change. The result is that the research of each community does not tend to be informed by the insights and resources available from the others. This is true not only between the social sciences and the natural sciences, but among the social sciences themselves. For instance, sociological research projects seldom incorporate spatial processes, behavioral analyses, or economic models.

For a short summary, read the article “The Wisdom of Crowds” in this week’s Nature Reports, and indeed, the editorial that goes with it.

…is a section heading on page 23 of this new report, “Psychology and Global Climate Change: Addressing a Multi-faceted Phenomenon and Set of Challenges” from the APA on how the field of psychology can contribute to the climate crisis. It’s a very good report, covering many of the core issues in the psychology of climate change, and laying out a research agenda for the field. Let me just quote the final paragraph of the report:

“a psychological perspective is crucial to understanding the probable effects of climate change, to reducing the human drivers of climate change, and to enabling effective social adaptation.  By summarizing the relevant psychological research, we hope not only to enhance recognition of the important role of psychology by both psychologists and non-psychologists, but also to encourage psychologists to be more aware of the relevance of global climate change to our professional interests and enable them to make more of the contributions the discipline can offer.”

Or, if you’re short of time, just read the press release.

Now, where’s the equivalent task force for the computer science community?

I’ve spent some time pondering why so many people seem unable or unwilling to understand the seriousness of climate change. Only half of all Americans understand that warming is happening because of our use of fossil fuels. And clearly many people still believe the science is equivocal. Having spent many hours arguing with denialists, I’ve come to the conclusion that they don’t approach climate change in a scientific way (even those who are trained as scientists), even though they often appear to engage in scientific discourse. Rather than assessing all the evidence and trying to understand the big picture, climate denialists start from their preferred conclusion and work backwards, selecting only the evidence that supports the conclusion.

But why? Why do so many people approach global warming in this manner? Previously I speculated that the Dunning-Kruger effect might explain some of this. This effect occurs when people at the lower end of the ability scale vastly overestimate their own competence. Combine this with the observation that few people really understand the basic system dynamics, for example that concentrations of greenhouse gases in the atmosphere will continue to rise even if emissions are reduced, as long as the level of emissions (burning fossil fuels) exceeds the removal processes (e.g. sequestration by the oceans). The Dunning-Kruger effect suggests that people whose reasoning is based on faulty mental models are unlikely to realise it.

While incorrect mental models and overconfidence might explain some of the problem that people have in accepting the scale and urgency of the problem, it doesn’t really explain the argumentation style of climate denialists, particularly the way in which they latch onto anything that appears to be a weakness or an error in the science, while ignoring the vast majority of the evidence in the published literature.

However, a series of studies by Kahan, Braman and colleagues explain this behaviour very well. In investigating a key question in social epistemology, Kahan and Braman set out to study why strong political disagreements seem to persist in many areas of public policy, even in the face of clear evidence about the efficacy of certain policy choices. These studies reveal a process they term cultural cognition, by which people filter (scientific) evidence according to how well it fits their cultural orientation. The studies explore this phenomenon for contentious issues such as the death penalty, gun control and environmental protection, as well as issues that one might expect would be less contentious, such as immunization and nanotechology. It turns out that not only do people care about how well various public policies cohere with their existing cultural worldviews, but their beliefs about the empirical evidence are also derived from these cultural worldviews.

For example, in a large scale survey, they tested people’s attitudes to the perception of risks from global warming, gun ownership, nanotechnology and immunization. They assessed how well these perceptions correlate with a number of characteristics, including gender, education, income, political affiliation, and so on. While political party affiliation correlates well with attitudes on some of these issues, there was a generally stronger correlation across the board with the two dimensions of cultural values identified by Douglas and Wildavsky: ‘group’ and ‘grid’. The group dimension assesses whether people are more oriented towards individual needs (‘individualist’) or the needs of the group (‘communitarian’); and the grid dimension assesses whether people tend to believe societal roles should be well defined and differentiated (‘hierarchical’) or those who believe in more equality and less rigidity (‘egalitarian’).

The most interesting part of the study, for me, is a an experiment on how perceptions change depending on how the risk of global warming is presented. About 500 subjects were given one of two different newspaper articles to read, both of which summarized the findings of a scientific report about the threat of climate change. In one version, the scientists were described as calling for anti-pollution regulations, while in the other, they were calling for investment in more nuclear power. Both these were compared with a control group who saw neither version of the report. Here are the results (adapted from Kahan et al, with a couple of corrections supplied by the authors):

Kahan-etal-fig3aKahan-etal-fig3b

In all cases, the mean risk assessment of the subjects correlates with their position on these dimensions: individualists and hierarchs are much less worried about global warming than communitarians and egalitarians. But more interestingly, the two different newspaper articles affect these perceptions in different ways. For the article that described scientists as calling for anti-pollution measures, people had quite opposite reactions: for communitarians and egalitarians, it increased their perception of the risk from global warming, but for individualists and hierarchs, it decreased their perception of the risk. When the same facts about the science are presented in an article that calls for more nuclear power, there is almost no effect. In other words, people assessed the facts in the report about climate change according to how well the policy prescription fits with their existing worldview.

There are some interesting consequences of this phenomenon. For example, Kahan and Braman argue that there is really no war over ideology in the US, just lots of people with well-established cultural worldviews, who simply decide what facts (scientific evidence) to believe based on these views. The culture war is therefore really a war over facts, not ideology.

The studies also suggest that certain political strategies are doomed to failure. For example, a common strategy when trying to resolve contentious political policy issues is to attempt to detach the policy question from political ideologies, and focus on the available evidence about the consequences of the policy. Kahan and Braman’s studies show this won’t work, because different cultural worldviews prevent people from agreeing what the consequences of a particular policy will be (no matter what empirical evidence is available). Instead, they argue that policymakers must find ways of framing policy so that affirm the values of diverse cultural worldviews simultaneously.

As an example, for gun control, they suggest offering a bounty (e.g. a tax rebate) for people who register handguns. Both pro- and anti- gun control groups might view this as beneficial to them, even though they disagree on the nature of the problem. For climate change, the equivalent policy prescriptions include tradeable emissions permits (which appeal to individualists and hierarchists), and more nuclear power (which egalitarians and hierarchists tend to view as less risky when presented as a solution to global warming).

Update: There’s a very good opinion piece by Kahan in the January 21, 2010 issue of Nature.

The recording of my Software Engineering for the Planet talk is now available online. Having watched it, I’m not terribly happy with it – it’s too slow, too long, and I make a few technical mistakes. But hey, it’s there. For anyone already familiar with the climate science, I would recommend starting around 50:00 (slide 45) when I get to part 2 – what should we do?

[Update: A shorter (7 minute) version of the talk is now available]

The slides are also available as a pdf with my speaking notes (part 1 and part 2), along with the talk that Spencer gave in the original presentation at ICSE. I’d recommend these pdfs rather than the video of me droning on….

Having given the talk three times now, I have some reflections on how I’d do it differently. First, I’d dramatically cut down the first part on the climate science, and spend longer on the second half – what software researchers and software engineers can do to help. I also need to handle skeptics in the audience better. There’s always one or two, and they ask questions based on typical skeptic talking points. I’ve attempted each time to answer these questions patiently and honestly, but it slows me down and takes me off-track. I probably need to just hold such questions to the end.

Mistakes? There are a few obvious ones:

  • On slide 11, I present a synoptic view of the earth’s temperature record going back 500 million years (it’s this graph from wikipedia). I use it to put current climate change into perspective, but also also to make the point that small changes in the earth’s temperature can be dramatic – in particular, the graph indicates that the difference between the last ice age and the current inter-glacial is about 2°C average global temperature. I’m now no longer sure this is correct. Most textbooks say it was around 8°C colder in the last ice age, but these appear to be based on an assumption that temperature readings taken from ice cores at the poles represent global averages. The temperature change at the poles is always much greater than the global average, but it’s hard to compute a precise estimate of global average temperature from polar records. Hansen’s reconstructions seem to suggest 3°C-4°C. So the 2°C rise shown on the wikipedia chart is almost certainly an underestimate. But I’m still trying to find a good peer-reviewed account of this question.
  • On slide 22, I talk about Arrhenius’s initial calculation of climate sensitivity (to doubling of CO2) back in the 1880’s. His figure was 4ºC-5ºC, whereas the IPCC’s current estimates are 2ºC-4.5ºC. And I need to pronounce his name correctly.

What’s next? I need to turn the talk into a paper…

This afternoon, I’m at the science 2.0 symposium, or “What every scientist needs to know about how the web is changing the way they work”. The symposium has been organised as part of Greg’s Software Carpentry course. There’s about 120 people here, good internet access, and I got here early enough to snag a power outlet. And a Timmie’s just around the corner for a supply of fresh coffee. All set.

1:05pm. Greg’s up, introducing the challenge: for global challenges (e.g. disease control, climate change) we need two things: Courage and Science. Most of the afternoon will be talking about the latter. Six speakers, 40 minutes each, wine and cheese to follow.

1:08pm. Titus Brown, from Michigan State U. Approaching Open Source Science: Tools Approaches. Aims to talk about two things: how to suck people into your open source project, and automated testing. Why open source? Ideologically: for reproducibility and open communication. Idealistically: can’t change the world by keeping what you do secret. Practical reason: other people might help. Oh and “Closed-source science” is an oxymoron. First, the choice of license probably doesn’t matter, because it’s unlikely anyone will ever download your software. Basics: every open source project should have a place to get the latest release, a mailing list, and an openly accessible version control system. Cute point: a wiki and issue tracker are useful if you have time and manpower, but you don’t, so they’re not.

Then he got into a riff about whether or not to use distributed version control (e.g. git). This is interesting because I’ve heard lots of people complain that tools like git can only be used by ubergeeks (“you have to be Linus Torvolds to use it). Titus has been using it for 6 months, and says it has completely changed his life. Key advantages: decouples developers from the server, hence ability to work offline (on airplanes), but still do version control commits. Also, frees you from “permission” decisions – anyone can take the code and work on it independently (as long as they keep using the same version control system). But there are downsides – creates ‘effective forks’, which might then lead to code bombs – someone who wants to remerge a fork that has been developed independently for months, and which then affects large parts of the code base.

Open development is different to open source. The key question is do you want to allow others to take the code and do their own things with it, or do you want to keep control of everything (professors like to keep control!). Oh, and you open yourself up to “annoying questions” about design decisions, and frank (insulting) discussion of bugs. But the key idea is that these are the hallmarks of a good science project – a community of scientists thinking and discussing design decisions and looking for potential errors.

So, now for some of the core science issues. Titus has been working on Earthshine – measuring the albedo of the earth by measuring how much radiation from the earth lights up the (dark side of the) moon. He ended up looking though the PVwave source code, trying to figure out what the grad student working on the project was doing. By wading through the code, he discovered the student had been applying the same correction to the data multiple times, to try and get a particular smoothing. But the only people who understood how the code worked were the grad student and Titus. Which means there was no way, in general, to know that the code works. Quite clearly, “code working” should not be judged by whether it does what the PI thinks it should do. In practice the code is almost never right – more likely that the PI has the wrong mental model. Which lead to the realization that we don’t teach young scientists how to think about software – including being suspicious of their code. And CS programs don’t really do this well either. And fear of failure doesn’t seem to be enough incentive – there are plenty of examples where software errors have lead to scientific results being retracted.

Finally, he finished off with some thoughts about automated testing. E.g. regression testing is probably the most useful thing scientists can do with their code: run the changed code and compare the new results with the old ones. If there are unexpected changes, then you have a problem. Oh, and put assert statements in to check that things that should never occur don’t ever occur. Titus also suggests that code coverage tools can be useful for finding dead code, and continuous integration is handy if you’re building code that will be used on multiple platforms, so an automated process builds the code and tests it on multiple platforms, and reports when something broke. Bottom line: automated testing allows you to ‘lock down’ boring code (code that you understand), and allows you to focus on ‘interesting’ code.

Questions: I asked whether he has ever encountered problems with the paranoia among some scientific communities, for example, fear of being scooped, or journals who refuse to accept papers if any part has already appeared on the web. Titus pointed out that he has had a paper rejected without review, because when he mentioned that many people were already using the software, the journal editor then felt this means it was not novel. Luckily, he did manage to publish it elsewhere. Journals have to take the lead by, for example, refusing to publish paper unless the software is open, because it’s not really science otherwise.

1:55pm. Next up Cameron Neylon, “A Web Native Research Record: Applying the Best of the Web to the Lab Notebook”. Cameron’s first slide is a permission to copy, share, blog, etc. the contents of the talk (note to self – I need this slide). So the web is great for mixing, mashups, syndicated feeds, etc. Scientists need to publish, subscribe, syndicate (e.g. updates to handbooks), remix (e.g. taking ideas from different disciplines and pull them together to get new advances). So quite clearly, the web is going to solve all our problems, right?

But our publication mechanisms is dead, broken, disconnected. A PDF of a scientific paper is a deadend, when really it should be linked to data, sources, citations, etc. It’s the links between things that matter. Science is a set of loosely coupled chunks of knowledge, they need to be tightly wired to each other so that we understand their context, we understand their links. A paper is too big a piece to be thought of as a typical “chunk of science”. A tweet (example was of MarsPhoenix team announcing they found ice on Mars) is too small, and too disconnected. A blog post seems about right. It includes embedded links (e.g. to detailed information about the procedures and materials used in an experiment). He then shows how his own research group is using blogs as online lab notebooks. Even better, some blog posts are generated automatically by the machines (when dealing with computational steps in the scientific process). Then if you look at the graph of the ‘web of objects’, you can tell certain things about them. E.g. an experiment that failed occupies a certain position in the graph; a set of related experiments appear as a cluster; a procedure that wasn’t properly written up might appear as a disconnected note; etc.

Now, how do we get all this to work? Social tagging (folksonomies) don’t work well because of inconsistent use of tagging, not just across different people, but over time by the same person. Templates help, and the evolution of templates over time tells you a lot about the underlying ontology of the science (both the scientific process and the materials used). Cameron even points out places where their the templates they have developed don’t fit well with established taxonomies of materials developed (over many years) within his field, and that these mismatches reveal problems in the taxonomies themselves, where they have ignored how materials are actually used.

So, now everything becomes a digital object: procedures, analyses, materials, data. What we’re left with is the links between them. So doing science becomes a process of creating new relationships, and what you really want to know about someone’s work is the (semantic) feed of relationships created. The big challenge is the semantic part – how do we start to understand the meaning of the links. Finally, a demonstration of how new tools like Google Wave can support this idea – e.g. a Wave plugin that automates the creation of citations within a shared document (Cameron has an compelling screen capture of someone using it).

Finally, how do we measure research impact? Eventually, something like pagerank. Which means scientists have to be wired into the network, which means everything we create has to be open and available. Cameron says he’s doing a lot less of the traditional “write papers and publish” and much more of this new “create open online links”). But how do we persuade research funding bodies to change their culture to acknowledge and encourage these kinds of contribution? Well, 70% of all research is basically unfunded – done on a shoestring.

2:40pm. slight technical hitch getting the next speaker (Michael) set up, so a switch of speakers: Victoria Stodden, How Computational Science is Changing the Scientific Method. Victoria is particularly interested in reproducibility in scientific research, and how it can be facilitated. Massive computation changes what we can do in science, e.g. data mining for subtle patterns in vast databases, and large scale simulations of complex processes. Examples: climate modeling, high energy physics, astrophysics. Even mathematical proof is affected – e.g. use of a simulation to ‘prove’ a mathematical result. But is this really a valid proof? Is it even mathematics?

So, effectively this might be a third branch of science. (1) deductive method for theory development – e.g. mathematics and logic (2) inductive/empirical – the machinery of hypothesis testing. And now (3) large scale extrapolation and prediction. But there’s lots of contention about this third branch. E.g. Anderson “The End of Theory“, Hillis rebuttal – we look for patterns first, and then create hypotheses, just as we always have. Weinstein points out that simulation underlies the other branches – tools to build intuitions, and tools to test hypotheses. Scientific approach is primarily about the ubiquity of error, so that the main effort is to track down and understand sources of error.

Although computational techniques being widely used now (e.g. in JASA, over the last decade, grown to more than half the papers using them), but very few make their code open, and very little validation going on, which means that there is increasingly a credibility crisis. Scientists make their papers available, but not their complete body of research. Changes are coming (e.g. Madagascar, Sweave,…), and the push towards reproducibility pioneered by Jon Claerbout.

Victoria did a study of one particular subfield: Machine Learning. Surveyed academics attending one of the top conferences in the field (NIPS). Why did they not share? Top reason: time it takes to document and clean up the code and data. Then, not receiving attribution, possibility of patents, legal barriers such as copyright, and potential loss of future publications. Motivations to share are primarily communitarian (for the good of science/community), while most of the barriers are personal (worries about attribution, tenure and promotion, etc).

Idea: take the creative commons license model, and create a reproducible research standard. All media components get released under as CC BY license, code gets released under some form of BSD license. But what about data? Raw facts alone are not generally copyrightable, so this gets a little complicated. But the expression of facts in a particular way is.

So, what are the prospects for reproducibility? Simple case: small scripts and open data. But harder case: inscrutible code and organic programming. Really hard case: massive computing platforms and streaming data. But it’s not clear that readability of the code is essential, e.g. Wolfram Alpha – instead of making the code readable (because in practice nobody will read it), make it available for anyone to run it in any way they like.

Finally, there’s a downside to openness, in particular, a worry that science can be contaminated because anyone can come along, without the appropriate expertise, and create unvalidated science and results, and they will get cited and used.

3:40pm. David Rich. Using “Desktop” Languages for Big Problems. David starts of with an analogy of different types of drill – e.g. a hand drill – trivially easy to use, hard to hurt yourself, but slow; up to big industrial drills. He then compares these to different programming languages / frameworks. One particular class of tools, cordless electric drills, are interesting because they provide a balance between power and usability/utility. So what languages and tools do scientific programmers need? David presented the results of a survey of their userbase, to find out what tools they need. Much of the talk was about the need/potential for parallelization via GPUs. David’s company has a tool called Star-P which allows users of Matlab and NumPy to transform their code for parallel architectures.

4:10pm. Michael Nielsen. Doing Science in the Open: How Online Tools are Changing Scientific Discovery. Case study: Terry Tao‘s use of blogs to support community approaches to mathematics. In particular, he deconstructs one particular post: Why global regularity for Navier-Stokes is hard, which sets out a particular problem, identifies the approaches that have been used, and has attracted a large number of comments from some of the top mathematicians in the field, all of which helps to make progress on the problem. (similar examples from other mathematicians, such as the polymath project), and a brand new blog for this: polymathprojects.org.

But these examples couldn’t be published in the conventional sense. They are more like the scaling up of a conversation that might occur in a workshop or conference, but allowing the scientific community to continue the conversation over a long period of time (e.g. several years in some cases), and across geographical distance.

These examples are pushing the boundaries of blog and wiki software. But blogs are just the beginning. Blogs and open notebooks enable filtered access to new information sources and new conversations. Essentially, they are restructuring expert attention – people focus on different things and in a different way than before. And this is important because expert attention is the critical limiting factor in scientific research.

So, here’s a radically different idea. Markets are a good way to efficiently allocate scarce resources. So can we create online markets in expert attention. For example Innocentive. One particular example: need in India to get hold of solar powered wireless routers to support a social project (ASSET India) helping women in india escape from exploitation and abuse. So this was set up as a challenge on Innocentive. A 31-yr old software engineering from Texas designed a solution, and it’s now being prototyped.

But, after all, isn’t all this a distraction? Shouldn’t you be writing papers and grant proposals rather than blogging and contributing to wikipedia? When Galileo discovered the rings of Saturn (actually, that Saturn looked like three blobs), he sent an anagram to Kepler, which then allowed him to claim credit. The modern scientific publishing infrastructure was not available to him, and he couldn’t conceive of the idea of open sharing of discoveries. The point being that these technologies (blogs etc) are too new to understand the full impact and use, but we can see ways in which they are already changing the way science is done.

Some very interesting questions followed about attribution of contribution, especially for the massive collaboration examples such as polymath. In answer, Michael pointed to the fact that the record of the collaboration is open and available for inspection, and that letters of recommendation from senior people matter a lot, and junior people who contributed in a strong way to the collaboration will get great letters.

[An aside: I’m now trying to follow this on Friendfeed as well as liveblogging. It’s going to be hard to do both at once]

4:55pm. Last but not least, Jon Udell. Collaborative Curation of Public Events. So, Jon claims that he can’t talk about science itself, because he’s not qualified, but will talk about other consequences of the technologies that we’re talking about. For example, in the discussions we’ve been having with the City of Toronto on it’s open data initiative, there’s a meme that governments sit on large bodies of data, and people would like to get hold of. But in fact, citizens themselves are owners and creators of data, and that’s a more interesting thing to focus on than governments pushing data out to us. For example, posters advertising local community events on lampposts in neighbourhoods around the city. Jon makes the point that this form of community advertising is outperforming the web, which is shocking!

Key idea: syndication hubs. For example, an experiment to collate events in Keene, NH, in the summer of 2009. Takes in datafeeds from various events websites, calendar entries etc. Then aggregates them, and provides feeds out to various other websites. But not many people understand what this is yet – it’s not a destination, but a broker. Or another way of understanding it is as ‘curation’ – the site becomes a curator looking after information about public events, but in a way that distributes responsibility for curation to the individual sources of information, rather than say a person looking after an events diary.

Key principles: syndication is a two way process (need to both subscribe to things and publish your feeds).But tagging and data formating conventions become critical.  The available services form an ecosystem, and they co-evolve, and we’re now starting to understand the eco-system around RSS feeds – sites that are publishers, subscribers, and aggregators. Similar eco-system growing up around iCalendar feeds, but currently missing aggregators. iCalendar is interesting because the standard is 10 years old, but it’s only recently become possible to publish feeds from many tools. And people are still using RSS feeds to do this, when they are the wrong tool – an RSS feed doesn’t expose the data (calendar information) in a usable way.

So how do we manage the metadata for these feeds, and how do we handle the issue of trust (i.e. how do you know which feeds to trust for accuracy, authority, etc)? Jon talks a little about uses of tools like Delicious to bookmark feeds with appropriate metadata, and other tools for calendar aggregation. And the idea of guerilla feed creation – how to find implicit information about recurring events and making them explicit. Often the information is hard to scrape automatically – e.g. information about a regular square dance that is embedded in the image of a cartoon. But maybe this task could be farmed out to a service like mechanical turk.

And these are great examples of computational thinking. Indirection – instead of passing me your information, pass me a pointer to it, so that I can respect your authority over it. Abstraction – we can use any URL as a rendezvous for social information management, and can even invent imaginary ones just for this purpose.

Updates: The twitter tag is tosci20. Andrew Louis also blogged (part of) it, and has some great photos; Joey DeVilla has detailed blog posts on several of the speakers; Titus reflects on his own participation; and Jon Udell has a more detailed write up of the polymath project. Oh, and Greg has now posted the speakers’ slides.

23. July 2009 · 6 comments · Categories: advocacy

Here’s a simple parable for climate change:

A large group of kids has congregated out on the sidewalk in front of their school. It started with just a few friends, showing off their latest video game. But the crowd grew, and now completely blocks the sidewalk. A guy in a wheelchair wants to pass, but can’t. The kids are so wrapped up in their own interests that they don’t even notice that together they have completely blocked the sidewalk.

Further along the street there is a busy pub. The lunchtime crowd has spilled out on to the sidewalk, and now has become so big that again the sidewalk is blocked. When the guy in the wheelchair wants to pass, quite a few people in the crowd recognize the problem, and they try to squeeze out of the way. But individually, none of them can make much difference to the blockage – there are just too many people there. They shrug their shoulders and apologise to the guy in the wheelchair.

In both cases, the blockages are not caused by individuals, and cannot be solved by individuals. The blockage is an emergent property of the crowd of people as a whole, and only occurs when the crowd grows to a certain size. In the first case, the members of the crowd remain blissfully unaware of the problem. In the second case, many people do recognise the problem, but cannot, on their own, do much about it. It would take concerted, systematic action by everyone in the crowd to clear a suitable passage. Understanding the problem and wanting to do something about it is not sufficient to solve it – the entire crowd has to take coordinated action.

And if some members of the crowd are more like the kids, unable to recognise the problem, no solution is possible.

When I was at the EGU meeting in Vienna in April, I attended a session on geoengineering, run by Jason Blackstock. During the session I blogged the main points of Jason’s talk, the key idea of which is that it’s time to start serious research into the feasibility and consequences of geoengineering, because it’s now highly likely we’ll need a plan B, and we’re going to need a much better understanding of what’s involved before we do it. Jason mentioned a brainstorming workshop, and the full report is now available: Climate Engineering Responses to Climate Emergencies. The report is an excellent primer on what we know currently about geoengineering, particularly the risks. It picks out stratospheric aerosols as the most likely intervention (from the point of view of both cost/feasibility, and current knowledge of effectiveness).

I got the sense from the meeting that we have reached an important threshold in the climate science community – previously geoengineering was unmentionable, for fear that it would get in the way of the serious and urgent job of reducing emissions. Alex Steffen explains this fear very well, and goes over the history of how the mere possibility of geoengineering has been used as an excuse by the denialists for inaction. And of course, from a systems point of view, geoengineering can only ever be a distraction if it tackles temperature (the symptom) rather than carbon concentrations (the real problem).

But the point made by Jason, and in the report, is that we cannot rule out the likelihood of climate emergencies – either very rapid warming triggered by feedback effects, or sudden onset of unanticipated consequences of (gradual) warming. In other words, changes that occur too rapidly for even the most aggressive mitigation strategies (i.e. emissions reduction) to have an effect on. Geoengineering then can be seen as “buying us time” to allow the mitigation strategies to work – e.g slowing the warming by a decade or so, while we get on and decarbonize our energy supplies.

Now, maybe it’s because I’m looking out for them, but I’ve started to see a flurry of research interest in geoengineering. Oliver Morton’s article “Great White Hope” in April’s Nature gives a good summary of several meetings earlier this year, along with a very readable overview of some of the technology choices available. In June, the US National Academies announced a call for input on geoengineering which yielded a treasure trove of information – everything you’ve ever wanted to know about geoengineering. And yesterday, New Scientist reported that geoengineering has gone mainstream, with a lovely infographic illustrating some of the proposals.

Finally, along with technical issues of feasibility and risk, the possibility of geoengineering raises major new challenges for world governance. Who gets to decide which geoengineering projects should go ahead, and when, and what will we do about the fact that, by definition, all such projects will have a profound effect on human society, and those effects will be distributed unequally?

Update: Alan Robock has a brilliant summary in the Bulletin of the Atomic Scientists entitled 20 reasons why geo-engineering might be a bad idea.