Singularity > Climate Change > Peak Oil > Financial Crisis
While lying awake late at night worrying about what kind of world my children will inherit, I find it helpful to come up with schemas for the most obvious and inevitable of the large societal problems. It makes them seem slightly more manageable to place them in order of importance, or time. Further, being clear on what are the biggest and most important problems is an essential prerequisite to thinking about solutions: these problems all interact, and solutions to the smaller of them may not be radical enough to address the larger of them.
In this post, I would like to argue for the above ordering of problems. I mean the '>' symbol in two senses: "A > B" meaning both "The main impact of A will fall later in time than the main impact of B", and also "A is a more serious and fundamental threat to humanity than B". While a full explication of the arguments would occupy a number of books, today you are going to have to make do with a single measly blog post, albeit longer than usual.
Peak Oil > Financial Crisis
I think it's indisputable that we are in the middle of the financial crisis right now. We had the main private sector crisis in 2008, now we are finding out which sovereign governments are going to end up needing to default, and perhaps that will induce a second wave of private sector defaults. By contrast, I think the arguments for a very near-term peak in oil supply have started to look quite weak. Firstly, work on the expansion of Iraqi oil supply continues apace. But even if Iraqi production couldn't expand much, it's likely the 2008 peak of oil supply will be exceeded.
However, I've never been able to get too terribly excited about the financial crisis as a massive long-term threat to humanity. At the end of the day, debt is not a physical quantity. The fact that humanity, collectively, has written too many debt instruments means that we have been too optimistic about the future, and created more promises than can be actually serviced. However, that doesn't create any fundamental physical constraint on our activities: it just means that the excessive promises need to be renegotiated to be more in line with the our actual future capabilities. This process will be painful and difficult in the short term, but I can't see how it poses any fundamental difficulty to the continued operation of civilization. There have been financial crises and sovereign debt defaults for many centuries, and we have survived them: we will very likely survive this one too.
By contrast, peak oil, when it does come, represents a significant physical constraint, and will require a large scale transformation of a number of important infrastructure elements, one way or another, over the course of a few decades. It's also an unprecedented situation - nobody has written papers about the many previous oil peaks over the centuries. Since infrastructures like transportation are critical to the operation of society, if we were to blow the handling of peak oil, it could be quite dangerous.
Thus I argue that peak oil is both probably later than the financial crisis, but ultimately more important and threatening.
Climate Change > Peak Oil
At the same time, I don't believe peak oil by itself is really a mortal threat, unless we seriously mishandle it. Fundamentally, our society is so wealthy and creates such a large economic surplus relative to our actual bodily needs that we have tremendous amounts of waste and potential for conservation. We can perfectly well get to work, or the grocery store, on scooters instead of in SUVs if we have to, or electric hybrid bicycles if it comes to that. We can take fewer foreign vacations, and we can video-conference instead of flying off on business trips. We may not want to, we may bitch and moan, but it's certainly possible to do these things rather than experience the end of the world. Fundamentally, what the post-peak period requires is that we get a few percent more efficient in our oil usage each year. We've already proven our ability to do that in the late seventies and early eighties.
So the real threat of peak oil is that we'll refuse to accept what needs to be done, and instead start fighting resource wars and the like (see Cheney, Dick, Iraq 2003). Converting ever larger fractions of our food supply to fuel is also a spectacularly dumb approach that we have a choice whether or not to do. The issue is thus more social/political/cultural than it is technological.
However, the great thing about peak oil, especially compared to climate change, is that when it occurs (and even in the run-up as we saw in 2005-2007) it will call attention to itself in an impossible-to-ignore way. When you have not quite enough oil, the price goes through the roof, and everyone immediately realizes that there is a major not-quite-enough oil problem, and begins thinking in terms of smaller cars, hybrid cars, skipping the family vacation this year, etc. The price feedback is immediate and relentless until enough conservation has been achieved for however much oil is available this particular year. And the same thing will be true all the way down - if at some point we need to transition from gas scooters to electric scooters, there will $40/gallon gas pointing out to us the immediate urgency of that step.
By contrast, climate change lacks the immediate personal feedback. We all pump the greenhouse gases up into the air, and there is no immediate consequence to us - no equivalent to paying a high price at the fuel pump. Instead, the gases gradually accumulate and the climate slowly warms, and the oceans slowly warm, and over the decades, the floods, the deadly heatwaves, the hurricanes, the forest fires, the disease outbreaks, the coastal cliff collapses, the species extinctions and bizarre ecological catastrophes all get more and more common and more and more serious. But because both the climate system and the major infrastructure in the economy have decades and decades of lag built in, by the time the problem is so incontrovertible that everyone can agree on its true urgency, it will be very late, and we will already be locked into an enormous amount of damage. Thus ultimately, I think climate change poses the more serious threat because it doesn't map so naturally to the way human incentive structures work.
And by the same token, I think the bulk of the climate change problem will occur later than peak oil. I think it's pretty much inconceivable that peak oil will occur later than around 2030, but climate change will just really be getting going then, and will gradually turn into a more and more furious rolling catastrophe, with the worst impacts in the second half of this century, and getting more and more severe until we finally get the situation under control (or succumb to it).
Singularity > Climate Change
But still, as enormous as the transformation required by climate change is, I don't think it's the biggest monster under the bed. True, most of our industrial society has been powered by fossil fuels for a couple of centuries, and it's all we've ever known as individuals. But perfectly good high-net-energy alternatives are known in modern wind, solar, and nuclear, and there's no in-principle reason we can't transition to them provided we make a serious all-out effort.
And the good thing is, although it's the largest such transition we've ever attempted, it's fundamentally a kind of challenge that we in Western Civilization recognize. We've created a problem, and the solution requires technological innovation. We need better renewables, smarter grids, more efficient electric cars. This stuff is fun for techies to think about, and Silicon Valley is already on the job with massive investment in clean-tech in the last five years. A defining feature of western civilization is that we value innovation highly. The patent was invented and became widespread in medieval Europe; with one minor exception for the Greek city state of Sybaris, there is no evidence of earlier civilizations using patents. Braudel, in comparing medieval European civilization to other world civilizations at the time points out that Europe was the only place with a clear concept of fashion. Other civilizations simply didn't have the concept that it was valuable to dress or decorate in a new and different way than your forebears. So while other civilizations have certainly created many innovations, none has institutionalized and glorified the innovation process in the way that ours has. And to this day, across the contemporary political spectrum, innovation is valued and seen as the way to solve all our problems.
However, the problem that we are gradually innovating more and more of our fellow citizens out of jobs, and thus out of any meaningful stake in our society, is not one that obviously lends itself to innovation. In that case, innovation looks to me like the problem, not the solution.
I have not yet talked much about the singularity on this blog, so let me briefly recap the main ideas of the techno-optimist computer scientists who have mostly thought about this, and then give my own take.
The basic idea was first formulated by computer scientist and science fiction author Vernor Vinge, but in recent years the most visible exponent has been computer scientist and inventor Ray Kurzweil. I was first introduced to these ideas via the latter's book The Singularity is Near, which I take to be still a reasonably current exposition of Kurzweil's thinking.
The argument goes like this:
- Throughout pre-history and history, the rate of evolution has accelerated. Physical evolution proceeds slowly at inventing new forms, once humanity came on the scene prehistoric cultural evolution proceeded faster, then with writing faster still, then a globally integrated society, now Internet connected, has innovated faster and faster still. The rate of change is now breathtaking.
- In particular, computers are now getter faster and faster and more capable. The rate of physical computation speed is increasing exponentially according to Moore's law, in which the amount of computation that can be done by one chip doubles every two years (roughly stated). There are no major physical limitations that will prevent Moore's law continuing decades into the future.
- Computers are now slowly getting smarter and smarter as we solve particular problems one at a time. Eg, the voice recognition problem is slowly coming under control, face recognition is starting to be a commercial proposition, driving a vehicle is now at the pretty-decent-research-prototype stage, etc.
- By the mid 2030s, available computers produced each year will exceed the information processing capacity of the entire human population, and by the mid 2040s it will exceed it by a factor of a billion.
- With ever-increasing progress in neuroscience, we will be able by then to reverse-engineer the human brain, and use computers to create general human-quality intelligence, except it will be able to go faster and faster than the basic version that biology has managed to produce to date.
- Once machine intelligence surpasses human intelligence by a sufficient margin, it will continue to accelerate the rate of intellectual and technical progress at rates massively higher than at present, and this acceleration will continue at ever greater rates. The name singularity comes from the idea of a comparison with a black hole where there is an event horizon beyond which one cannot see. Once ever-accelerating machine intelligence takes over, things will happen so fast and furiously that we today have no idea what will happen and cannot predict the consequences.
- Nonetheless, you shouldn't worry about any of this, because we will be able to use all this amazing technology to augment and transform human consciousness. We will be able to reverse engineer how we work, modify and improve on it, download ourselves into new/better/synthetic bodies, cure world hunger, etc, etc, etc... So life will be better overall.
Obviously, in compressing an entire long book into seven bullet points, I am leaving out a variety of nuances and supporting lines of evidence, which readers are welcome to explore on their own, or we can discuss further on this blog in future. My overall reaction to the book was to first feel incredibly depressed for a few weeks, and then spend a number of months in research gradually sorting out which parts I believed and which I didn't. Herewith my take:
- (Accelerating technical progress). I think, with some caveats, this is broadly qualitatively true. Our society does innovate faster, than, say the Ubaid proto-civilization, and in turn, the Ubaid were undoubtedly innovating faster than Homo Erectus. However, you can, and Kurzweil does, overstate the issue. Probably the best metric of overall rate of technical progress in the industrial era is growth in the global world product, and that has been a few percent a year throughout that time - it is not dramatically accelerating recently, though industrial growth is faster than pre-industrial growth.
- (Moore's Law). The historical rate of progress in computer speed is incontrovertible and general technical consensus is that there is no near term barrier to it continuing (eg see this book for a summary of the issues). I don't see a reasonable basis for questioning this, though of course future events are always somewhat uncertain.
- (Computers getting smarter). This is incontrovertibly true. While the speed with which this would happen was overstated in the early years of AI, it is the case that with each passing decade more and more things traditionally requiring human expertise can now be done by computers (chess, voice recognition, driving, etc).
- (Available computer power will exceed human information processing capability in 25 years or so). As a statement of raw computing power, this seems plausible, though of course the further we try to extrapolate Moore's law, the less certain the result.
- (Ability to reverse engineer human intelligence and produce a synthetic equivalent also within about 25 years). Here I think Kurzweil gets onto thinner and thinner ice. At the moment, we computer scientists have no real idea how to produce anything like the generality and flexibility of human reasoning, so it's going to take major conceptual breakthroughs before we are able to do so. So I think the time estimates here get much less certain than he indicates - raw computation power alone doesn't make for a general flexible intelligence, and fundamental breakthroughs are hard/impossible to predict in advance. At the same time, it does seem reasonable to project that the general direction at present will continue, with more and more tasks being automatable, and as neuroscience and cognitive science and computer science all continue, it seems likely there will be major progress in these areas. However, I think the timing is much harder to project.
- (Further acceleration and singularity). The further out we go, the more speculative it all gets. However, I don't dispute the general proposition that machine intelligence comparable to or better than human intelligence will be a complete game-changer and things will never be the same again and we can't predict what the world would look like afterwards. That much seems elementary.
- (This will be good for the average human). Here I found Kurzweil's reasoning very fantastical... In particular, I think he seriously neglects the fact that medical progress is far slower than computer progress and there are excellent unavoidable reasons for this: new medical technologies require massive clinical trials to establish their safety and efficacy, while new generations of chips and computers don't. Therefore, progress on silicon intelligence will continue to be much faster than progress on biological intelligence. This might not be so good for many of the biological types...
The choice of "singularity" as metaphor for this all this is interesting, I think. Falling into a black hole would not normally be considered a pleasant or desirable prospect, but this was the metaphor picked by people who think this will be a good thing.
I think the key issue to analyzing the goodness, or otherwise, of these trends is to look at why/how technologies get adopted. I've spent the last decade working in technology startups building and selling new computer technologies, so I think I have a decent understanding of this. There are basically two main selling points to any new technology:
1) This is whizz-bang cool and will enhance your social status to own.
and then, either:
2a) This will enhance your productivity and allow you to do more/better of whatever it is you do in less time (for a utilitarian or business proposition).
2b) This will be fun to do (for a consumer/entertainment proposition)
Usually points 1) and either 2a) or 2b) points are required to close the sale. Technologies do not get adopted because they generally improve human welfare, they get adopted because they work on these points for specific sets of customers. Thus corporate voice recognition phone systems did not come around because all our lives are improved by it, but because it allows companies to handle more customers with fewer staff, thereby providing a return on investment to the project. If in the future, we get automated vehicle driving systems, the reasons will have nothing to do with overall human welfare, but with the specific profitability equations of trucking firms, or taxi-firms.
Likewise, new consumer technologies do not get deployed to the benefit of poor people, they get deployed to the benefit of people who can and will pay for them (aka customers).
And this will continue to be the case in the future.
So in my view, the dominant social effect of all these ever-accelerating-integrated-globalized-world-stuff is to create a divide: if you are sufficiently smart/skilled/creative to stay employed and take advantage of all the new toys, life gets better and more fun and you feel more powerful and enabled. But if you are not sufficiently smart/skilled/creative (or you just decide the game is fucked), then you are out of a job and now you can't afford most of the fun stuff and not only that, you live in the midst of a society that you know doesn't have any use for you and doesn't value you.
And unfortunately, we have this:
So the fraction of people below the divide is growing steadily. Slowly, but steadily. In fact, if you look at the trend, it's pretty close to linear. Here is the employment/population ratio for men 25-54 again, with a linear and quadratic fit. They barely differ:
Given the way technology is adopted by businesses, one of the effects of a singularity that I would expect is that the employment-population ratio would fall to zero. Machine intelligence, if equally good, will always displace human intelligence from the workplace because it doesn't insist on going home at night, doesn't have any human rights, and can be replicated in arbitrary quantities (meaning it will cost much less) and be customized to the task at hand. If we extend the employment population ratio out to 2100, and extrapolate the trends, we get this:
The orange circle shows Kurzweil's approximate singularity date with my contention that it will be signalled economically by an employment/population of zero. I don't find it plausible that that curve is going to bend over that far, but maybe. I wouldn't necessarily have a great deal of confidence in a linear or quadratic extension either - it's hard to believe that another few decades of Moore's law won't have some pretty massive effects on the economy, and on employment in particular.
My best guess is that a singularity of some kind is probably coming, but slower than folks like Kurzweil think. However, just as the gravitational tidal forces from a black hole tear you apart some time before you reach the singularity, I think the tidal forces of increasing automation are already at work and will become gradually more stressful over the coming century.
Finally, I note that there are complex interplays between these various threats.
For example, consider the effects of peak oil in the context of increasing automation. Reversalist thinkers like the Archdruid, or Jim Kunstler, believe that one effect of peak oil will be to cause replacement of machines by human labor again (on the theory that with less energy, we'll need to use more human labor, just like we did back in the day, when we had less energy before). However, this certainly isn't evident in the data to date, and it's not at all obvious that it follows. Consider, for example, the decision as to whether to lay off the security guards, and replace them with security robots. This is a decision that companies will take individually based on the economics and quality of the security robots, but presumably over time, these will improve and there will be more robots and fewer humans doing physical security, and the employment-population ratio will drift a little lower.
However, consider the implications for oil demand. An employed security guard will be driving to work in his (or her!) pickup truck, and if you pay him enough, he may even fly off for a yearly vacation, or to his daughter's wedding on the other side of the country. If he's not employed, he probably will use far less oil like that, and the robot that replaces him will run on a little electricity and no oil.
So it's not at all clear that less oil implies less automation. Similarly, addressing climate change by switching to renewable power doesn't prevent continued automation either, because machines can be fundamentally more efficient than humans at converting sunlight to work.
And so I think, in the longer term, the threat to most of the nine or ten billion of us that there will soon be is that the economy simply won't need us to operate. And if we address the financial crisis, solve peak oil, and fix climate change with ever increasing amounts of cleverness and innovation, we will probably simply hasten that day.
What do you think? Leave a comment below.
Sign up for regular Resilience bulletins direct to your email.