Thanks for the link, John.
When most of SI (and a few others) discussed the issue for 10+ hours, we tentatively concluded that WBE progress should not be accelerated. (Context: see differential technological development.)
Edit: This appears to be a working link to the discussion Luke is referencing.
Some thoughts:
It doesn't look like any of the people in your discussion were academic neuroscientists.
I don't know how much we can read into the fact that the explicit aim of the project is to emulate a brain, not make brain inspired AI.
At this point, I'm pretty uncertain, and it doesn't look like there is much interest from less wrong users in constructing a better estimate. And since the discussion you describe is something of a black box, it seems like it would be awkward to improve upon.
This is the team responsible for simulating the rat cortical column.
http://www.nature.com/news/2011/110308/full/news.2011.143.htm
The team is one of 6 that is being considered for at least 2 "FET Flagship" positions, which comes with all that funding. Each of the six competing teams is proposing to work on some kind of futuristic technology:http://cordis.europa.eu/fp7/ict/programme/fet/flagship/6pilots_en.html
Of course, word on the street is that academic neuroscientists don't think much of the project:
I think trying to influence the committee's decision potentially represents very low hanging fruit in politics as charity.
Even if academic neuroscientists don't think much of the project in its current state, it seems likely that $1.4 billion would end up attracting a lot of talent to this problem, and get us the first upload significantly sooner.
It's true that Less Wrong doesn't have a consensus position on whether to speed development of cell modeling and brain scanning technology or not. But I think if we have a discussion and a vote, we're significantly more likely than the committee to come up with the right decision for humanity. As far as I can tell, the committee will essentially be choosing at random. It shouldn't be hard for us to beat that.
Edit: But that's not to say that our estimate should be quick and dirty. In the spirit of holding off on proposing solutions, I discourage anyone from taking a firm public position on this topic for now.
In terms of avenues for influence, here are a few ideas off the top of my head:
We, and the folks at the Future of Humanity Institute, SI, and other groups, seem to spend a lot of time thinking about what would happen in the ideal scenario in terms of the order in which technologies are developed and how they are deployed. I think there is a good case for also investing in the complementary good of trying to actually influence the world towards a more ideal scenario.