While lying awake late at night worrying about what kind of world my children will inherit, I find it helpful to come up with schemas for the most obvious and inevitable of the large societal problems. It makes them seem slightly more manageable to place them in order of importance, or time. Further, being clear on what are the biggest and most important problems is an essential prerequisite to thinking about solutions: these problems all interact, and solutions to the smaller of them may not be radical enough to address the larger of them.
In this post, I would like to argue for the above ordering of problems. I mean the '>' symbol in two senses: "A > B" meaning both "The main impact of A will fall later in time than the main impact of B", and also "A is a more serious and fundamental threat to humanity than B". While a full explication of the arguments would occupy a number of books, today you are going to have to make do with a single measly blog post, albeit longer than usual.
Peak Oil > Financial Crisis
I think it's indisputable that we are in the middle of the financial crisis right now. We had the main private sector crisis in 2008, now we are finding out which sovereign governments are going to end up needing to default, and perhaps that will induce a second wave of private sector defaults. By contrast, I think the arguments for a very near-term peak in oil supply have started to look quite weak. Firstly, work on the expansion of Iraqi oil supply continues apace. But even if Iraqi production couldn't expand much, it's likely the 2008 peak of oil supply will be exceeded.
However, I've never been able to get too terribly excited about the financial crisis as a massive long-term threat to humanity. At the end of the day, debt is not a physical quantity. The fact that humanity, collectively, has written too many debt instruments means that we have been too optimistic about the future, and created more promises than can be actually serviced. However, that doesn't create any fundamental physical constraint on our activities: it just means that the excessive promises need to be renegotiated to be more in line with the our actual future capabilities. This process will be painful and difficult in the short term, but I can't see how it poses any fundamental difficulty to the continued operation of civilization. There have been financial crises and sovereign debt defaults for many centuries, and we have survived them: we will very likely survive this one too.
By contrast, peak oil, when it does come, represents a significant physical constraint, and will require a large scale transformation of a number of important infrastructure elements, one way or another, over the course of a few decades. It's also an unprecedented situation - nobody has written papers about the many previous oil peaks over the centuries. Since infrastructures like transportation are critical to the operation of society, if we were to blow the handling of peak oil, it could be quite dangerous.
Thus I argue that peak oil is both probably later than the financial crisis, but ultimately more important and threatening.
Climate Change > Peak Oil
At the same time, I don't believe peak oil by itself is really a mortal threat, unless we seriously mishandle it. Fundamentally, our society is so wealthy and creates such a large economic surplus relative to our actual bodily needs that we have tremendous amounts of waste and potential for conservation. We can perfectly well get to work, or the grocery store, on scooters instead of in SUVs if we have to, or electric hybrid bicycles if it comes to that. We can take fewer foreign vacations, and we can video-conference instead of flying off on business trips. We may not want to, we may bitch and moan, but it's certainly possible to do these things rather than experience the end of the world. Fundamentally, what the post-peak period requires is that we get a few percent more efficient in our oil usage each year. We've already proven our ability to do that in the late seventies and early eighties.
So the real threat of peak oil is that we'll refuse to accept what needs to be done, and instead start fighting resource wars and the like (see Cheney, Dick, Iraq 2003). Converting ever larger fractions of our food supply to fuel is also a spectacularly dumb approach that we have a choice whether or not to do. The issue is thus more social/political/cultural than it is technological.
However, the great thing about peak oil, especially compared to climate change, is that when it occurs (and even in the run-up as we saw in 2005-2007) it will call attention to itself in an impossible-to-ignore way. When you have not quite enough oil, the price goes through the roof, and everyone immediately realizes that there is a major not-quite-enough oil problem, and begins thinking in terms of smaller cars, hybrid cars, skipping the family vacation this year, etc. The price feedback is immediate and relentless until enough conservation has been achieved for however much oil is available this particular year. And the same thing will be true all the way down - if at some point we need to transition from gas scooters to electric scooters, there will $40/gallon gas pointing out to us the immediate urgency of that step.
By contrast, climate change lacks the immediate personal feedback. We all pump the greenhouse gases up into the air, and there is no immediate consequence to us - no equivalent to paying a high price at the fuel pump. Instead, the gases gradually accumulate and the climate slowly warms, and the oceans slowly warm, and over the decades, the floods, the deadly heatwaves, the hurricanes, the forest fires, the disease outbreaks, the coastal cliff collapses, the species extinctions and bizarre ecological catastrophes all get more and more common and more and more serious. But because both the climate system and the major infrastructure in the economy have decades and decades of lag built in, by the time the problem is so incontrovertible that everyone can agree on its true urgency, it will be very late, and we will already be locked into an enormous amount of damage. Thus ultimately, I think climate change poses the more serious threat because it doesn't map so naturally to the way human incentive structures work.
And by the same token, I think the bulk of the climate change problem will occur later than peak oil. I think it's pretty much inconceivable that peak oil will occur later than around 2030, but climate change will just really be getting going then, and will gradually turn into a more and more furious rolling catastrophe, with the worst impacts in the second half of this century, and getting more and more severe until we finally get the situation under control (or succumb to it).
Singularity > Climate Change
But still, as enormous as the transformation required by climate change is, I don't think it's the biggest monster under the bed. True, most of our industrial society has been powered by fossil fuels for a couple of centuries, and it's all we've ever known as individuals. But perfectly good high-net-energy alternatives are known in modern wind, solar, and nuclear, and there's no in-principle reason we can't transition to them provided we make a serious all-out effort.
And the good thing is, although it's the largest such transition we've ever attempted, it's fundamentally a kind of challenge that we in Western Civilization recognize. We've created a problem, and the solution requires technological innovation. We need better renewables, smarter grids, more efficient electric cars. This stuff is fun for techies to think about, and Silicon Valley is already on the job with massive investment in clean-tech in the last five years. A defining feature of western civilization is that we value innovation highly. The patent was invented and became widespread in medieval Europe; with one minor exception for the Greek city state of Sybaris, there is no evidence of earlier civilizations using patents. Braudel, in comparing medieval European civilization to other world civilizations at the time points out that Europe was the only place with a clear concept of fashion. Other civilizations simply didn't have the concept that it was valuable to dress or decorate in a new and different way than your forebears. So while other civilizations have certainly created many innovations, none has institutionalized and glorified the innovation process in the way that ours has. And to this day, across the contemporary political spectrum, innovation is valued and seen as the way to solve all our problems.
However, the problem that we are gradually innovating more and more of our fellow citizens out of jobs, and thus out of any meaningful stake in our society, is not one that obviously lends itself to innovation. In that case, innovation looks to me like the problem, not the solution.
I have not yet talked much about the singularity on this blog, so let me briefly recap the main ideas of the techno-optimist computer scientists who have mostly thought about this, and then give my own take.
The basic idea was first formulated by computer scientist and science fiction author Vernor Vinge, but in recent years the most visible exponent has been computer scientist and inventor Ray Kurzweil. I was first introduced to these ideas via the latter's book The Singularity is Near, which I take to be still a reasonably current exposition of Kurzweil's thinking.
The argument goes like this:
Obviously, in compressing an entire long book into seven bullet points, I am leaving out a variety of nuances and supporting lines of evidence, which readers are welcome to explore on their own, or we can discuss further on this blog in future. My overall reaction to the book was to first feel incredibly depressed for a few weeks, and then spend a number of months in research gradually sorting out which parts I believed and which I didn't. Herewith my take:
The choice of "singularity" as metaphor for this all this is interesting, I think. Falling into a black hole would not normally be considered a pleasant or desirable prospect, but this was the metaphor picked by people who think this will be a good thing.
I think the key issue to analyzing the goodness, or otherwise, of these trends is to look at why/how technologies get adopted. I've spent the last decade working in technology startups building and selling new computer technologies, so I think I have a decent understanding of this. There are basically two main selling points to any new technology:
1) This is whizz-bang cool and will enhance your social status to own.
and then, either:
2a) This will enhance your productivity and allow you to do more/better of whatever it is you do in less time (for a utilitarian or business proposition).
2b) This will be fun to do (for a consumer/entertainment proposition)
Usually points 1) and either 2a) or 2b) points are required to close the sale. Technologies do not get adopted because they generally improve human welfare, they get adopted because they work on these points for specific sets of customers. Thus corporate voice recognition phone systems did not come around because all our lives are improved by it, but because it allows companies to handle more customers with fewer staff, thereby providing a return on investment to the project. If in the future, we get automated vehicle driving systems, the reasons will have nothing to do with overall human welfare, but with the specific profitability equations of trucking firms, or taxi-firms.
Likewise, new consumer technologies do not get deployed to the benefit of poor people, they get deployed to the benefit of people who can and will pay for them (aka customers).
And this will continue to be the case in the future.
So in my view, the dominant social effect of all these ever-accelerating-integrated-globalized-world-stuff is to create a divide: if you are sufficiently smart/skilled/creative to stay employed and take advantage of all the new toys, life gets better and more fun and you feel more powerful and enabled. But if you are not sufficiently smart/skilled/creative (or you just decide the game is fucked), then you are out of a job and now you can't afford most of the fun stuff and not only that, you live in the midst of a society that you know doesn't have any use for you and doesn't value you.
And unfortunately, we have this:
So the fraction of people below the divide is growing steadily. Slowly, but steadily. In fact, if you look at the trend, it's pretty close to linear. Here is the employment/population ratio for men 25-54 again, with a linear and quadratic fit. They barely differ:
Given the way technology is adopted by businesses, one of the effects of a singularity that I would expect is that the employment-population ratio would fall to zero. Machine intelligence, if equally good, will always displace human intelligence from the workplace because it doesn't insist on going home at night, doesn't have any human rights, and can be replicated in arbitrary quantities (meaning it will cost much less) and be customized to the task at hand. If we extend the employment population ratio out to 2100, and extrapolate the trends, we get this:
The orange circle shows Kurzweil's approximate singularity date with my contention that it will be signalled economically by an employment/population of zero. I don't find it plausible that that curve is going to bend over that far, but maybe. I wouldn't necessarily have a great deal of confidence in a linear or quadratic extension either - it's hard to believe that another few decades of Moore's law won't have some pretty massive effects on the economy, and on employment in particular.
My best guess is that a singularity of some kind is probably coming, but slower than folks like Kurzweil think. However, just as the gravitational tidal forces from a black hole tear you apart some time before you reach the singularity, I think the tidal forces of increasing automation are already at work and will become gradually more stressful over the coming century.
Finally, I note that there are complex interplays between these various threats.
For example, consider the effects of peak oil in the context of increasing automation. Reversalist thinkers like the Archdruid, or Jim Kunstler, believe that one effect of peak oil will be to cause replacement of machines by human labor again (on the theory that with less energy, we'll need to use more human labor, just like we did back in the day, when we had less energy before). However, this certainly isn't evident in the data to date, and it's not at all obvious that it follows. Consider, for example, the decision as to whether to lay off the security guards, and replace them with security robots. This is a decision that companies will take individually based on the economics and quality of the security robots, but presumably over time, these will improve and there will be more robots and fewer humans doing physical security, and the employment-population ratio will drift a little lower.
However, consider the implications for oil demand. An employed security guard will be driving to work in his (or her!) pickup truck, and if you pay him enough, he may even fly off for a yearly vacation, or to his daughter's wedding on the other side of the country. If he's not employed, he probably will use far less oil like that, and the robot that replaces him will run on a little electricity and no oil.
So it's not at all clear that less oil implies less automation. Similarly, addressing climate change by switching to renewable power doesn't prevent continued automation either, because machines can be fundamentally more efficient than humans at converting sunlight to work.
And so I think, in the longer term, the threat to most of the nine or ten billion of us that there will soon be is that the economy simply won't need us to operate. And if we address the financial crisis, solve peak oil, and fix climate change with ever increasing amounts of cleverness and innovation, we will probably simply hasten that day.