by nef

The New Economics Foundation

25 January 2010

from NewEconomics Website

 

NEF is an independent think-and-do tank that inspires and demonstrates real economic well-being.

We aim to improve quality of life by promoting innovative solutions that challenge mainstream thinking on economic, environmental and social issues. We work in partnership and put people and the planet first.

NEF (the new economics foundation) is a registered charity founded in 1986 by the leaders of The Other Economic Summit (TOES), which forced issues such as international debt onto the agenda of the G8 summit meetings. It has taken a lead in helping establish new coalitions and organizations such as the Jubilee 2000 debt campaign; the Ethical Trading Initiative; the UK Social Investment Forum; and new ways to measure social and economic well-being.

Schumacher College is the UK’s international green college. We are delighted to be partnering with NEF in publishing Growth isn't possible: Why we need a new economic direction. As with NEF, Schumacher College owes its inspiration to the work of the radical economist and development educator Fritz Schumacher. For twenty years we have been running a wide range of short courses attracting the world's leading thinkers on new economics.

 

Over the past two years we have hosted two think tanks around sustainable economics and intelligent growth, leading to a declaration on sustainable economics, the 'E4 Declaration' and our strong support for the Green New Deal. This year we took the decision to develop with help from NEF colleagues a post-graduate MSc course in New, or 'Transition' Economics, the first in the country.

 

We will also be opening a second campus in 2011 focusing upon the 'great re-skilling', with the provision of a range of courses on sustainable food and farming, renewal energy, eco design and sustainable business development, together with launching an E Learning program.

 

For further information on the courses available at Schumacher College please see our website www.schumachercollege.org.uk.

 

Schumacher College is an initiative of The Dartington Hall Trust.

 

 

Contents

  1. Greenhouse gas emissions and current climate change

  2. Scenarios of growth and emission reductions

  3. Peak Oil, Gas and Coal?

  4. Carbon capture and storage - The nuclear fusion of the 2010s?

  5. The limits to nuclear

  6. The hydrogen economy

  7. Biofuels

  8. Geoengineering - Technological saviour or damaging distraction?

  9. How much can energy efficiency really improve?

  10. Equity considerations

  11. If not the economics of global growth, then what? - Getting an economy the right size for the planet

 

 

 

 

 

 

 

 

 

 


 

 

Foreword

If you spend your time thinking that the most important objective of public policy is to get growth up from 1.9 per cent to 2 per cent and even better 2.1 per cent we’re pursuing a sort of false god there. We’re pursuing it first of all because if we accept that, we will do things to the climate that will be harmful, but also because all the evidence shows that beyond the sort of standard of living which Britain has now achieved, extra growth does not automatically translate into human welfare and happiness.
Lord Adair Turner1 Chair of the UK Financial Services Authority

Anyone who believes exponential growth can go on forever in a finite world is either a madman or an economist.
Kenneth E. Boulding Economist and co-founder of General Systems Theory

In January 2006, nef (the new economics foundation) published the report Growth isn’t working.2 It highlighted a flaw at the heart of the general economic strategy that relies upon global economic growth to reduce poverty.

 

The distribution of costs and benefits from economic growth, it demonstrated, are highly unbalanced. The share of benefits reaching those on the lowest incomes was shrinking. In this system, paradoxically, in order to generate ever smaller benefits for the poorest, it requires those who are already rich and ‘over-consuming’ to consume ever more.

The unavoidable result under business as usual in the global economy is that, long before any general and meaningful reduction in poverty has been won, the very life-support systems that we all rely on are almost certain to have been fundamentally compromised.

Four years on from Growth isn’t working, this new publication, Growth isn’t possible goes one step further and tests that thesis in detail in the context of climate change and energy. It argues that indefinite global economic growth is unsustainable. Just as the laws of thermodynamics constrain the maximum efficiency of a heat engine, economic growth is constrained by the finite nature of our planet’s natural resources (biocapacity).

 

As economist Herman Daly once commented, he would accept the possibility of infinite growth in the economy on the day that one of his economist colleagues could demonstrate that Earth itself could grow at a commensurate rate.3

Whether or not the stumbling international negotiations on climate change improve, our findings make clear that much more will be needed than simply more ambitious reductions in greenhouse gas emissions.

 

This report concludes that a new macro economic model is needed, one that allows the human population as a whole to thrive without having to relying on ultimately impossible, endless increases in consumption.

Andrew Simms
Victoria Johnson
January 2010


Back to Contents

 

 

 

 

Introduction

We really have to come up with new metrics and new measures by which we look at economic welfare in a much larger context than just measuring GDP, which I think is proving to be an extremely harmful way of measuring economic progress.
R K Pachauri Ph.D,4 Chairman, Intergovernmental Panel on Climate Change Director-General, The Energy and Resources Institute, Director, Yale Climate and Energy Institute

Towards what ultimate point is society tending by its industrial progress? When the progress ceases, in what condition are we to expect that it will leave mankind?
John Stuart Mill (1848) 5

From birth to puberty a hamster doubles its weight each week. If, then, instead of levelling-off in maturity as animals do, the hamster continued to double its weight each week, on its first birthday we would be facing a nine billion tonne hamster. If it kept eating at the same ratio of food to body weight, by then its daily intake would be greater than the total, annual amount of maize produced worldwide.6

 

There is a reason that in nature things do not grow indefinitely.

The American economist Herman Daly argues that growth’s first, literal dictionary definition is,

‘…to spring up and develop to maturity. Thus the very notion of growth includes some concept of maturity or sufficiency, beyond which point physical accumulation gives way to physical maintenance’.7

In other words, development continues but growth gives way to a state of dynamic equilibrium - the rate of inputs are equal to the rate of outputs so the composition of the system is unchanging in time.8

 

For example, a bath would be in dynamic equilibrium if water flowing in from the tap escapes down the plughole at the same rate. This means the total amount of water in the bath does not change, despite being in a constant state of flux.

In January 2006, nef (the new economics foundation) published the report Growth isn’t working.9

 

It highlighted a flaw at the heart of the economic strategy that relies overwhelmingly upon economic growth to reduce poverty. The distribution of costs and benefits from global economic growth, it demonstrated, are highly unbalanced. The share of benefits reaching those on the lowest incomes was shrinking. In this system, paradoxically, in order to generate ever smaller benefits for the poorest, it requires those who are already rich and ‘over-consuming’ to consume ever more.

The unavoidable result, the report points out, is that, with business as usual in the global economy, long before any general and meaningful reduction in poverty has been won, the very life-support systems we all rely on are likely to have been fundamentally compromised.

Four years on from Growth isn’t working, Growth isn’t possible goes one step further and tests that thesis in detail in the context of climate change and energy. It argues that indefinite global economic growth is unsustainable. Just as the laws of thermodynamics constrain the maximum efficiency of a heat engine, economic growth is constrained by the finite nature of our planet’s natural resources (biocapacity).

 

As Daly once commented, he would accept the possibility of infinite growth in the economy on the day that one of his economist colleagues could demonstrate that Earth itself could grow at a commensurate rate.10

The most recent data on human use of biocapacity sends a number of unfortunate signals for believers in the possibility of unrestrained growth. Our global ecological footprint is growing, further overshooting what the biosphere can provide and absorb, and in the process, like two trains heading in opposite directions, we appear to be actually shrinking the available biocapacity on which we depend.

Globally we are consuming nature’s services - using resources and creating carbon emissions - 44 per cent faster than nature can regenerate and reabsorb what we consume and the waste we produce. In other words, it takes the Earth almost 18 months to produce the ecological services that humanity uses in one year.

 

The UK’s footprint has grown such that if the whole world wished to consume at the same rate it would require 3.4 planets like Earth.11

Growth forever, as conventionally defined (see Box 1), within fixed, though flexible, limits isn’t possible. Sooner or later we will hit the biosphere’s buffers. This happens for one of two reasons. Either a natural resource becomes over-exploited to the point of exhaustion, or because more waste is dumped into an ecosystem than can be safely absorbed, leading to dysfunction or collapse. Science now seems to be telling us that both are happening, and sooner, rather than later.

Yet, for decades, it has been a heresy punishable by career suicide for economists (or politicians) to question orthodox economic growth.

 

As the British MP Colin Challen quipped in 2006,

‘We are imprisoned by our political Hippocratic oath: we will deliver unto the electorate more goodies than anyone else.’12

 

 

Box 1: What is growth?

 

The question is deceptive, because the word has many applications. They range from the description of biological processes to more abstract notions of personal development. But, when used to describe the economy, growth has a very specific meaning. This often causes confusion.

Growth tends to be used synonymously with all things that are good. Plants grow, children grow, how could that be bad? But, of course, even in nature, growth can be malign, as in the case of cancer cells.

In economics ‘growth’, or the lack of it, describes the trajectory of Gross Domestic Product and Gross National Product, two slightly different measures of national income (they differ, basically, only in that one includes earnings from overseas assets). The value of imports is deducted and the value of exports added.

Hence, an economy is said to be growing if the financial value of all the exchanges of goods and services within it goes up. The absence of growth gets described, pejoratively, as recession. Prolonged recessions are called depressions.

Yet, it is not that simple. An economy may grow, for example, because money is being spent on clearing up after disasters, pollution, to control rising crime or widespread disease. You may also have ‘jobless growth,’ in which the headline figure for GDP rises but new employment is not generated, or environmentally destructive growth in which a kind of false monetary value is created by liquidating irreplaceable natural assets on which livelihoods depend.

The fact that an economy is growing tells you nothing about the ‘quality’ of economic activity that is happening within it. Conversely, history shows that in times of recession, life expectancy can rise, even as livelihoods are apparently harmed. This happens in rich countries probably due to force of circumstances, as people become healthier by consuming less and exercising more, using cheaper, more active forms of transport such as walking and cycling.

It is possible, in other words, to have both ‘economic’ and ‘uneconomic’ growth and we should not assume that growth per se is a good thing, to be held on to at all costs.

 

 

The growth debate: historical context
There is a kind of reverse political correctness that prevents growth being debated properly.

 

Yet this has not always been true. Historically, there have been vigorous debates on the optimal scale for the economy, which we survey briefly towards the end of this report (also summarized in Box 2 below).

More familiarly, the 1960s and early 1970s saw a vigorous debate on the environmental implications of growth. But this was sometimes hampered by insufficient data. Scientists at the Massachusetts Institute of Technology (MIT) were commissioned by the Club of Rome to research and publish the controversial Limits to growth, which came out in 1972. Since then, the original report has been successively revised and republished.

Matthew Simmons, founder of the world’s largest energy investment banking firm, commented on publication of the 2004 update that its message was more relevant than ever and that we, ‘wasted 30 valuable years of action by misreading the message of the first book’.13

 

Originally dismissed and criticized for ‘crying wolf’, the report has, in fact, stood the test of time. A study in 2008 by physicist Graham Turner from CSIRO (Commonwealth Scientific and Industrial Research Organization), Australia’s leading scientific research institute, compared its original projections with 30 years of subsequent observed trends and data.14 His research showed that they ‘compared favorably’.

Less well known is that in this fairly recent period, there was also a significant debate on the desirability of economic growth from the point of view of social and individual, human well-being.15,16,17 It is disciplines other than economics that have seemed able to view the issue of growth less dogmatically, asking difficult questions and making inconvenient observations, their views apparently less constrained by hardened doctrine.

For example, the implications of ‘doubling’, graphically represented by our voracious hamster, were addressed in May 2007 by Roderick Smith, Royal Academy of Engineering Research Professor at Imperial College, London. The physical view of the economy, he said, ‘is governed by the laws of thermodynamics and continuity’, and so, ‘the question of how much natural resource we have to fuel the economy, and how much energy we have to extract, process and manufacture is central to our existence.’18

Engineers must deal every day with the stuff, the material, the ‘thingyness’ of the world around them, the stresses and strains that make things stand up, fall down, last or wear out. Because of this, they are perhaps more in tune with the real world of resources than the economist working with abstract mathematical simplifications of life.

Hence Smith honed in on one of the economy’s most important characteristics - its ‘doubling period’, by which its bulk multiplies in proportion to its current size. Even low growth rates of around 3 per cent, he points out, lead to ‘surprisingly short doubling times’. Hence, ‘a 3 per cent growth rate, which is typical of the rate of a developed economy, leads to a doubling time of just over 23 years. The 10 per cent rates of rapidly developing economies double the size of the economy in just under 7 years.’

But then, if you are concerned about humanity’s ecological debt, comes what Smith quaintly calls the ‘real surprise’. Because, according to Smith, ‘each successive doubling period consumes as much resource as all the previous doubling periods combined’, just as 8 exceeds the sum of 1, 2 and 4.

 

Adding, almost redundantly, as jaws in the room fall open,

‘this little appreciated fact lies at the heart of why our current economic model is unsustainable.’
 

Why do economies grow?
We should ask the simple question, why do economies grow? And, why do people worry that it will be a disaster if they stop? The answers can be put reasonably simply.

For most countries in much of human history, having more stuff has given human beings more comfortable lives. Also, as populations have grown, so have the economies that housed, fed, clothed and kept them.

Yet, there has long been an understanding in the quiet corners of economics, as well as louder protests in other disciplines, that growth cannot and need not continue indefinitely.

 

As John Stuart Mill put it in 1848,

‘the increase of wealth is not boundless: that at the end of what they term the progressive state lies the stationary state.’19

The reasons for growth not being ‘boundless’ too, have been long known.

 

Even if the modern reader has to make allowances for the time in which Mill wrote, his meaning remains clear:

‘It is only in the backward countries of the world that increased production is still an important object: in those most advanced, what is economically needed is a better distribution.’20

 

 

Box 2. No-growth economics: a select chronology of books and papers
 

In contemplating any progressive movement, not in its nature unlimited, the mind is not satisfied with merely tracing the laws of the movement; it cannot but ask the further question, to what goal? Towards what ultimate point is society tending by its industrial progress? When the progress ceases, in what condition are we to expect that it will leave mankind?

It must always have been seen, more or less distinctly, by political economists, that the increase of wealth is not boundless: that at the end of what they term the progressive state lies the stationary state, that all progress in wealth is but a postponement of this, and that each step in advance is an approach to it.21
John Stewart Mill, 1848

1821 On the principles of political economy and taxation (3rd edition) by David Ricardo (on the ‘Stationary State’)

1848 Principles of political economy by John Stuart Mill (on the ‘Stationary State’, in Book IV, Chapter VI)

1883 Human labour and the unit of energy by Sergei Podolinsky

1922 Cartesian economics by Frederick Soddy

1967 The costs of economic growth by E J Mishan

1971 The entropy law and the economic process by Nicholas Georgescu-Roegen

1972 Limits to growth: A report for the Club of Rome’s project on the predicament of mankind by Donella Meadows

1973 Small is beautiful: A study of economics as if people mattered by E F Schumacher Toward a steady state economy by Herman E Daly (ed)

1977 The economic growth debate: An assessment by E J Mishan Social limits to growth by Fred Hirsch

1978 The economic growth debate: Are there limits to growth? By Lawrence Pringle

1982 Overshoot by William R Catton

1987 Our common future by the World Commission on Environment and Development

1989 Beyond the limits to growth: A report to the Club of Rome by Eduard Pestel

1992 The growth illusion: How economic growth has enriched the few, impoverished the many, and endangered the planet by Richard Douthwaite and Edward Goldsmith

1995 Our ecological footprint: Reducing human impact on the Earth by William Rees and Mathis Wackernagel

1996 Beyond growth by Herman E Daly

1997 Sustainable development: Prosperity without growth by Michael J Kinsley

2004 Limits to growth: The 30 year update by Donella Meadows, Jorgen Randers and Dennis Meadows Growth fetish by Clive Hamilton

2005 Ecological debt: The health of the planet and the wealth of nations by Andrew Simms

2006 Growth isn’t working: The unbalanced distribution of benefits and costs from economic growth by David Woodward and Andrew Simms

2008 Managing without growth by Peter Victor

2009 Prosperity without growth by Tim Jackson

2010 Growth isn’t possible by Andrew Simms, Victoria Johnson and Peter Chowla.



So why is it, that over 160 years after Mill wrote those words, rich nations are more obsessed than ever with economic growth?

Countries like the UK are decades past the point where increases in national income, measured by GNP and GDP lead to similar increases in human well-being and life expectancy.22 Yet no mainstream politician argues against the need for economic growth.

The reasons are partly to do with policy habits, partly political posturing, and partly because we have set our economic system up in such a way that it has become addicted to growth.

Growth-based national accounting became popular in the 1930s as a guide to quantify the value of government interventions to rescue economies from the depression, and also later as a tool to aid increased production as part of the war planning effort. But the new measurement came with a very big health warning attached.

One of the indicator’s key architects, the economist Simon Kuznets, was explicit about its limitations. Growth did not measure quality of life, he made clear, and it excluded vast and important parts of the economy where exchanges were not monetary. By this he meant, family, care and community work - the so-called ‘core economy’ which makes society function and civilization possible.23

 

So, for example, if the money economy grows at the expense of, and by cannibalizing the services of the core economy - such as in the way that profit driven supermarkets grow at the expense of communities - it is a kind of false growth. Similarly if the money economy grows simply by liquidating natural assets that are treated as ‘free income’ this, too, is a kind of ‘uneconomic growth’.

Also, it was repeatedly observed that growth in aggregate national income couldn’t tell you anything about the nature of the economy, whether activity was good or bad. Spending on prisons, pollution and disasters pushed up GDP just as surely as spending on schools, hospitals and parks. But growth nevertheless became the eclipsing indicator of an economy’s virility and success.

 

Even though, in 1968, Robert Kennedy pointed out that growth measured everything apart from ‘that which makes life worthwhile’.24

The problem with our economic system is now threefold. First, governments plan their expenditure assuming that the economy will keep growing. If it then didn’t grow, there would be shortfalls in government income with repercussions for public spending. The same is true for all of us; for example, when we plan for old age by putting our savings into pensions.

Today, though, many economies like the UK are facing this problem in any case. Ironically, however, it comes as a direct consequence of the economic damage caused by the behavior of weakly regulated banks, which were busy chasing maximum rates of growth through financial speculation.

Secondly, neo-liberal economies typically put legal obligations on publicly listed companies to grow. They make the maximization of returns to shareholders the highest priority for management.

 

As major investors are generally footloose, they are free to take their money wherever the highest rates of return and growth are found.

 

 

Box 3. Climate change is not the only limit
 

This report focuses mainly on how the need to preserve a climate system that is conducive to human society puts a limit on orthodox economic growth. But climate change is not the only natural parameter. Other limits of our biocapacity also need respecting if we are to maintain humanity’s environmental life support system. Two important areas of research, described below, provide examples of attempts to define some of those limits and raise questions for economists and policy makers.

The Ecological Footprint 25
From a methodology first developed by the Canadian geographer William Rees in the early 1980s, the ecological footprint is now a well-established technique being constantly refined as available data and understanding of ecosystems improves. It compares the biocapacity available to provide, for example, farmland, fisheries and forestry, as well as to absorb waste from human economic activity, with the rate at which humanity consumes those resources and produces waste, for example in the form of greenhouse gas emissions.

The 2009 set of Global Footprint Accounts reveal that the human population is demanding nature’s services, using resources and generating CO2 emissions, at a rate that is 44 per cent faster than what nature can replace and reabsorb. That means it takes the Earth just under 18 months to produce the ecological services humanity needs in one year. Very conservatively, for the whole world to consume and produce waste at the level of an average person in the United Kingdom, we would need the equivalent of at least 3.4 planets like earth. Most worryingly there are signs that available biocapacity is actually reducing, being worn out, by current levels of overuse, setting up a negative spiral of over-consumption and weakening capacity to provide.

Planetary boundaries
A much more recent approach, published in science journal Nature in September 2009, uses the notion of ‘planetary boundaries.’26 The work, co-authored by 29 leading international scientists, identifies nine processes in the biosphere for which the researchers considered it necessary to ‘define planetary boundaries’.

 

They are:

  • climate change

  • rate of biodiversity loss (terrestrial and marine)

  • interference with the nitrogen and phosphorus cycles

  • stratospheric ozone depletion

  • ocean acidification

  • global freshwater use

  • change in land use

  • chemical pollution

  • atmospheric aerosol loading

Of these nine, the authors found that three boundaries had already been transgressed: climate change, interference with the nitrogen cycle, and biodiversity loss (see Table 1).

Setting boundaries is complex. Earth systems change and react in often non-linear ways. The erosion or overburdening of one system can affect the behavior and resilience of another. As the research points out, ‘If one boundary is transgressed, then other boundaries are also under serious risk. For instance, significant land-use changes in the Amazon could influence water resources as far away as Tibet.’

Nevertheless, and even though with caveats, the authors identify boundaries for seven of the nine processes leaving the safe thresholds for atmospheric aerosol loading and chemical pollution still ‘to be identified.’

The work on planetary boundaries complements (although unusually doesn’t reference) the ecological footprint method. The latter, due to a lack of previous research on safe rates of harvest and waste dumping, merely produces a best assessment of full available biocapacity and compares it to human rates of consumption and waste generation. This conservatively, or rather generously, creates the impression that all biocapacity might be available for human use.

 

The attempt to define more nuanced planetary boundaries concerning different earth systems, is set to produce more realistic, and almost inevitably smaller assessments of the share of the earth’s resources and services available for safe human economic use.

 

 
Table 1.

Identifying planetary boundaries that should not be crossed.

Limits for earth processes in grey have already been transgressed.




Thirdly, in the modern world, money is lent into existence by banks with interest rates attached.

 

Because for every pound, dollar, yen or Euro borrowed, more must be paid back, economies that function largely on interest-bearing money have a built-in growth dynamic. The problem extends beyond the economy. Our increasingly consumerist society demands ever higher consumption to demonstrate social status - conspicuous consumption.27

 

To see how advanced, industrialized nations might escape from a locked-in growth dynamic, see the conclusion to this report.
 


First principles - the laws of thermodynamics

The first law says you can’t win, the second law says you can’t even break even
C.P. Snow

The physicist and novelist C.P. Snow became famous for trying to bridge the gap between the ‘two cultures’, science and the arts. When he described the alleged division, he made reference to the failure of those in the humanities to understand the Second Law of Thermodynamics.

 

While delivering The Rede Lecture in 1959, Snow observed,

‘Once or twice I have been provoked and have asked the company how many of them could describe the Second Law of Thermodynamics. The response was cold: it was also negative. Yet I was asking something which is about the scientific equivalent of: “Have you read a work of Shakespeare’s?” ’ 28

Yet, 50 years after delivering his lecture, while scientists are still thought to be illiterate if they haven’t read Shakespeare, how many experts in the arts would be able to explain the laws of thermodynamics?

 

This is not simple point-scoring between disciplines. Politicians and civil servants tend to be drawn from the fields of economics, politics, history and the arts.29 This could go some way to explain why, on one level, the mainstream political and economic establishment have little comprehension about the finiteness of the planet’s resources and the limits to efficiency.

One representative from a conservative economic think tank was questioned on where the resources to fuel infinite economic growth would come from. It was at a public debate in the Dana Centre, part of the Science Museum in London.

 

After thinking for a moment, his answer was confidently asserted,

‘We could mine asteroids,’ he said.
 

 

The First Law
The First Law of Thermodynamics, formalized by nineteenth-century German physicist, Rudolf Clausius, is a generalization of the universal law of energy conservation.30 The First Law states that within a closed system, energy can neither be created nor destroyed.

 

For example, energy within the Universe is constant, or the amount of energy lost in a steady-state process cannot be greater than the amount of energy gained. Thus, a measure of heat transferred into a system will result in an increase in temperature and in the system’s ability to do work.
 


The Second Law
The Second Law of Thermodynamics applies a direction to the conservation of energy described by the First Law. It says that not all heat input into a system can be converted into useful work.

 

Put simply, transferring heat into work with 100 per cent efficiency is impossible. Some heat will always escape into the surrounding environment as wasted energy. Ultimately, therefore, all energy tends to heat or disorder (entropy) and no transaction of energy from one type to another is completely reversible.

Because the laws of thermodynamics imply that entropy will always increase, Clausius imagined that sometime in the distant future - the universe would eventually fall fate to a ‘heat death’. Entropy will have increased to its maximum level and no more work could be done.

As entropy increases - ‘free energy’ or exergy decreases. This describes the maximum useful work obtainable from an energy system at a given state in a specified environment.

 

In other words, it represents the thermodynamic ‘quality’ of an energy carrier based on the Second Law. For example, electricity has a high degree of exergy and is widely regarded as an efficient carrier of energy. Low-temperature hot water, however, has low exergy and whilst it is also a carrier of energy, can generally only be used for heating purposes.

According to the second law of thermodynamics, order (sometimes called negative-entropy, neg-entropy) can be increased only at the expense of generating more disorder (entropy) elsewhere.31 This means importantly that human-created order - the emergence of structured civilization and latterly that of advanced industrialized society - will also result in large quantities of entropy in/on the surrounding environment.32

From this, the potential for environmental damage from economic activity becomes clear. Industrial activities cannot continue without energy, nor can they be generated without some environmental impact.

This observation was the basis of Herman Daly’s ‘steady-state economy’ which, building on the work of economist Nicholas Georgescu-Roegen, challenges humanity’s failure to notice the entropic nature of the economic process (although, it is more fairly, a specific a failure of mainstream economics).33, 34

While the Second Law means that energy efficiency in any process can never in reality be 100 per cent, the practical limits of energy efficiency approached in the real world is much lower.

 

This is discussed in more detail later on in the report.


Why the ‘unthinkable’ must be debated
The meaning of sustainability has been blurred since the flurry of activity that led up to the United Nation’s 1992 Earth Summit in Brazil. Today it is applied as much to merely sustaining economic growth as it is to preserving a livable planet for future generations.

This mainstream view of sustainable development is quite different from definitions of so-called ‘strong sustainability’ (Box 4 below). The ‘mainstream’ view tends to emphasize decoupling economic growth from environmental degradation (including climate change). And, to drive that dynamic it relies heavily on market-based initiatives - the ‘ecological modernization’ of the economy, defined by German sociologist Joseph Huber as a twin process of ‘ecologising the economy’ and ‘economising ecology’.35

Ecological modernization assumes that already existing political, economic and social institutions can adequately deal with environmental problems - focusing, almost exclusively on industrialism, with much less consideration (if any at all) being given to the accumulative process of capitalism, military power or the nation-sate system, even though all contribute in different ways to environmental degradation by being instrumental to growth and international competitiveness.36

Policies of environmental or ecological modernization include: the ‘polluter pays’ principle, eco-taxes, government purchasing initiatives, consumer education campaigns and instituting voluntary eco-labeling schemes. Such a strategy relies on small acts of individual consumer sovereignty (sustainable consumption) to change the market.37

 

The growing emphasis on the individual to practice sustainable consumption as a cure-all, however, is awkwardly juxtaposed against the systemic nature of the problems. There is now a growing view and body of evidence that ecological modernization has not been effective in reducing carbon emissions. In fact, some would argue it has acted in the opposite direction, driving emissions upwards.

Environmental debates, therefore, seem caught between paralyzing catastrophe scenarios, and ill-thought-out technological optimism. We are told that either the planet would like to see the back of us, or that we can have the planet and eat it. The truth, as ever, is more complex and interesting.

The point of this report, Growth isn’t possible, is to remove an obstacle to exploring the possibilities in that more nuanced reality.

 

Mainstream economics is frozen in its one-eyed obsession with growth. Across the political spectrum of governments, pursuing international competitiveness and a rising GDP is still seen as panacea for social, economic and environmental problems. Unfortunately, a combination of the science of climate change, natural resource accounting, economic realities and the laws of physics tell us that this assertion has become quite detached from reality.

 

Our earlier report, Growth isn’t working, showed that global economic growth is a very inefficient way to reduce poverty, and is becoming even less so.
 


Why growth isn’t working
Between 1990 and 2001, for every $100 worth of growth in the world’s income per person, just $0.60, down from $2.20 the previous decade, found its target and contributed to reducing poverty below the $1-a-day line.38

 

A single dollar of poverty reduction took $166 of additional global production and consumption, with all its associated environmental impacts. It created the paradox that ever smaller amounts of poverty reduction amongst the poorest people of the world required ever larger amounts of conspicuous consumption by the rich.

Growth wasn’t (and still isn’t) working.39 Yet, so deeply engrained is the commitment to growth, that to question it is treated as a challenge to the whole exercise of economics. Nothing could be further from the truth. This report is a companion volume to nef’s earlier and ongoing research. It is written in the hope that we can begin to look at the fascinating opportunities for economics that lie beyond the doctrine - it could be called dogma - of growth.

One of the few modern economists to have imagined such possibilities in any depth is Herman Daly.40

 

The kind of approach called for in a world constrained by fuzzy but fundamental limits to its biocapacity is one, according to Daly, that is:

‘…a subtle and complex economics of maintenance, qualitative improvements, sharing frugality, and adaptation to natural limits. It is an economics of better, not bigger’.41

 

 

Box 4. Sustainable development?
 

Civil servant and environmental economist, Michael Jacobs described six core ideals and themes within sustainable development.

 

These include:42

  • The integration of the economy and environment: economic decisions to have regard to their environmental consequences

  • Intergenerational obligation: current decisions and practices to take account of their effect on future generations

  • Social justice: all people have the equal right to an environment in which they can flourish (or have their basic human needs met)

  • Environmental protection: conservation of resources and protection of the non-human world

  • Quality of life: a wider definition of human well-being beyond narrowly defined economic prosperity

  • Participation: institutions to be restructured to allow all voices to be heard in decision-making (procedural justice)

The core ideals cover three fields - the environment, economy and society - the three pillars of sustainability. A view of sustainable development that encompasses all three dimensions can be defined as ‘strong sustainability’.

According to Andrew Dobson, Professor of Politics at Keele University, ‘strong sustainability’ will require, ‘radical changes in our relation with the non-human natural world, and in our mode of social and political life’.43


 

Relying on the wished-for trickle-down of income from global growth as the main economic strategy to meet human needs, maximize well-being and achieve poverty reduction appears ineffective, frequently counter-productive and is in all practical senses, impossible.

Given current, highly unequal patterns of the distribution of benefits from growth, to get everyone in the world onto an income of at least $3 per day - the level around which income stops having an extreme effect on life expectancy - implies, bizarrely, the need for 15 planets’ worth of resources to sustain the requisite growth. Even then, environmental costs would fall disproportionately, and counter-productively, on the poorest - the very people the growth is meant to benefit.44

So, globally, including in relatively rich countries, there is a danger of locking in a self-defeating spiral of over-consumption by those who are already wealthy, justified against achieving marginal increases in wealth amongst the poorest members of society.

Another assault on the doctrine of growth stems from the large but still emerging field of studying life-satisfaction and human well-being. It presents a critique of how, in industrialized countries, patterns of work and rising consumption are promoted and pursued that repeatedly fail to deliver the expected gains in life satisfaction. At the same time, these patterns of (over)work potentially erode current well-being by undermining family relationships and the time needed for personal development.45

The assumption that by increasing efficiency, whether it is energy efficiency or resource efficiency, will allow us to continue along the same, ever expanding consumption path is wrong. It does, however, allow us to skirt around the bigger issue relating to work-and-spend lifestyles that developed nations have become so accustomed to, and which are unquestioningly assumed to be the correct and best development models for developing nations.

In fact, a growing body of literature shows that once people have enough to meet their basic needs and are able to survive with reasonable comfort, higher levels of consumption do not tend to translate into higher levels of life satisfaction, or well-being.46

 

Instead, people tend to adapt relatively quickly to improvements in their material standard of living, and soon return to their prior level of life satisfaction. This is known as becoming trapped on the ‘hedonic treadmill’, whereby ever higher levels of consumption are sought in the belief that they will lead to a better life, whilst simultaneously changing expectations leave people in effect having to ‘run faster’, consuming more, merely to stand still.

National trends in subjective life satisfaction (an important predictor of other hard, quantitative indicators such as health) stay stubbornly flat once a fairly low level of GDP per capita is reached.47 And, importantly, only around 10 per cent of the variation in subjective happiness observed in western populations is attributable to differences in actual material circumstances, such as income and possessions.48

Figure 1 shows the results of an online survey of life satisfaction and consumption in Europe, gathered by nef. The web-based survey contained questions about lifestyle - consumption patterns, diet, health, family history - as well as subjective life satisfaction. Using this data, estimates of footprint and life expectancy could be calculated.

 

Over 35,000 people in Europe completed the survey.

Figure 1:

Life satisfaction compared to levels of material consumption in Europe.49

 

The blue line represents the distribution of ecological footprints across the total sample, expressed in terms of the number of planets’ worth of resources that would be required if everyone on the planet were to live the same way.

 

To the right end of the distribution are those people with high consumption lifestyles, approaching ‘seven planet living’. To the left are those whose lifestyles have the least environmental impact, approaching the planetary fair share ‘one planet living’.

 

The arrows depict the nature of the transition that is required both to level and lower the consumption playing field towards equitable and sustainable use of the Earth’s resources.

This data represents both a challenge and an opportunity. It is challenging because it shows starkly the extent of European over-use of planetary resources. Not only is the distribution of footprint extremely unequal in this sample, it is also far too high in absolute terms.

 

But, Figure 1 also suggests that well-being has little to do with consumption; which, in turn, allows for the possibility that our collective footprint could be reduced significantly without leading to widespread loss in well-being.

 

As one analyst put it, an initial reduction in energy use of around one-quarter ‘would call for nothing more than a return to levels that prevailed just a decade or no more than a generation ago’, adding rhetorically:

‘How could one even use the term sacrifice in this connection? Did we live so unbearably 10 or 30 years ago that the return to those consumption levels cannot be even publicly contemplated by serious policymakers?’58

 

 

Box 5. Life rage

Economic growth is indeed triumphant, but to no point. For material prosperity does not make humans happier: the ‘triumph of economic growth’ is not a triumph of humanity over material wants; rather it is the triumph of material wants over humanity.50
Professor Richard Layard, London School of Economics

Studies over the past decade, using both qualitative and quantitative methods, reveal levels of anger and moral anxiety about changes in society that were not apparent 30 years ago.51 Whilst these studies mainly focused on the UK, the USA and Australia, the findings are, to varying degrees, applicable to other high-consuming industrialized nations. In other words, our levels of well-being are being eroded. But why?

Research shows that the strong relationship between life expectancy and income levels-off at a remarkably low level. The influence of rising income on life satisfaction levels-off at higher levels, but not much higher.52,53 Life expectancy continues to rise in most countries and this is only partly due to greater wealth; happiness has not increased in recent decades in rich nations, despite on average, people have become much wealthier.54

Social epidemiologist, Professor Richard Wilkinson argues in his book Impact of inequality: how to make sick societies healthier that poorer nations with lower wealth inequality tend to have higher levels of well-being (physical and mental) than more wealthy but more unequal nations.55

 

For example, life expectancy in rich nations shows a strong correlation with relative equality. His more recent work with co-author Professor Kate Pickett, The Spirit Level, makes an even stronger case.56 Here they demonstrate that more equal societies almost always do better against a wide range of social and environmental indicators.

In Impact of inequality Wilkinson compared various social indicators in Greece to those in the USA. He found that while Greece has almost half the per capita GDP, citizens have a longer life expectancy than the USA. While globally, the USA is the wealthiest nation, it has one of the highest levels inequality and lowest life expectancy in the global North.

 

Furthermore, Wilkinson demonstrates that crime rates are most strongly correlated to a nation’s level of inequality, rather than its aggregated wealth. Given this, Wilkinson concludes that the most equal countries tend to have the highest levels of trust and social capital.

As Nicholas Georgescu-Roegen, one of the fathers of ecological economics argues, as we have become caught up in our obsession with consumption and material throughput, we have failed to recognise the ‘immaterial flux of the enjoyment of life’.57



Despite this, high-consuming lifestyles seem ‘locked-in’ by our economic, technological and cultural context, which fails to address equality and instead drives relative poverty.

 

As the gap between the ‘haves’ and ‘have-nots’ widens, there tends to be a concomitant loss of life satisfaction, sense of community and, ultimately, a rise in social disequilibrium.

For example, in an update to the infamous Whitehall Study led by Professor Michael Marmot at the Department of Epidemiology and Public Health at University College London, researchers found that subjective socio-economic status was a better predictor of health status and decline in health status over time than more objective measures.59,60

This work implies the health impacts of relative poverty are more likely to be determined by an individual’s perception of his or her socio-economic status than, beyond a certain level of sufficient consumption, their actual socio-economic circumstances.

 

Therefore perceived socio-economic barriers can act as a barrier to progressive improvements in overall well-being, as the physical and mental well-being of those in the lowest strata is undermined, creating domino effects throughout society.

There are questions to be asked of growth, of its science-based limits, and more generally of its effectiveness today in meeting human needs and maximizing well-being. This report suggests that we are reaching the point at which the doctrine of global economic growth as a central policy objective and primary strategy for meeting society’s various needs is becoming redundant.

Later in this report we will argue that focusing only on improvements in carbon and energy intensity of the economy, as a strategy to combat climate change, means only that we are buying time, and even then very little. In a best-case scenario, delaying arrival at critical concentrations of greenhouse gases by 10-20 years, and in a worst-case scenario, not delaying at all.

 

So let us first address the question, what is, and what should be accepted as ‘safe’ levels of greenhouse gases in the atmosphere?
 

Back to Contents

 

 

 

 

Greenhouse gas emissions and current climate change


The Earth’s climate system is currently changing at greater rates and in patterns that are beyond the characteristics of natural variation.

 

The concentration of carbon dioxide (CO2) in the atmosphere today, the most prevalent anthropogenic greenhouse gas, far exceeds the natural range of 180-300 ppm. The present concentration is the highest during the last 800,000 years and probably during the last 20 million years.61,62,63

In the space of just 250 years, as a result of the Industrial Revolution and changes to land use, such as the growth of cities and the felling of forests, we have released cumulatively more than 1800 gigatonnes (Gt) of CO2 into the atmosphere.64 Global atmospheric concentrations of CO2 are now a record 390 ppm, almost 40 per cent higher than they were at the beginning of the Industrial Revolution.65, 66

The primary source of the increased concentration of CO2 is unequivocally due to the burning of fossil fuels such as coal, oil, and natural gas.67 Annual fossil fuel CO2 emissions have increased year on year from an average of 23.4 Gt CO2 per year in the 1990s to 30 Gt CO2 per year today.

 

To put this in perspective, the increase in annual emissions over the past 20 years is almost double the total emissions produced by EU27 nations each year.68 Changes in land use have also contributed significantly to increasing rates of CO2 emissions, contributing around 5.5 Gt CO2 per year to the atmosphere.

We now release just over 1000 tonnes of CO2 into the Earth’s atmosphere every second.

In 2007, the Intergovernmental Panel on Climate Change (IPCC) Fourth Assessment Report - a synthesis of peer-reviewed research on climate change, its causes and effects (including socio-economic consequences) involving over 2500 scientists worldwide - stated that if fossil fuels continued to be burnt at the current rate, global average surface temperatures could rise by 4°C by the end of the century, with an uncertainty range of 2.4-6.4°C.69

A more recent study published in the American science journal Proceedings of the National Academy of Sciences found that the ‘committed’ level of warming by the end of the century is 2.4°C (1.4-4.3°C) - if atmospheric concentrations of greenhouse gases are held at 2005 levels. This value is based on past emissions and includes the warming already observed of 0.76°C plus 1.6°C of additional warming which is yet to occur due to the thermal inertia of the climate system and the ‘masking’ by cooling aerosols.70

Although 2008 may have been the coolest year of the current decade, it was still the tenth warmest year since instrumental records began in 1850.71 While observations actually suggest that global temperature rise has slowed during the last decade, analyses of observations and modeling studies show that this is due to internal climate variability and that the warming trend will resume in the next few years.72,73

One of the studies by atmospheric scientists Professors Kyle Swanson and Anastasios Tsonis ends with the following cautionary note:

‘…there is no comfort to be gained by having a climate with a significant degree of internal variability, even if it results in a near-term cessation of global warming…If the role of internal variability in the climate system is as large as this analysis would suggest, warming over the 21st century may well be larger than that predicted by the current generation of models’.74

Indeed, over the course of 2008 and 2009 numerous scientific papers were published revealing that climate change was far more serious even than reported in the most recent review of the science by the IPCC.75,76

 

The long-term warming trend has had a large impact on mountain glaciers and snow cover worldwide, and also changes in rainfall patterns and intensity, ocean salinity, wind patterns and aspects of extreme weather including droughts, heavy precipitation, heat waves and the intensity of tropical cyclones. Such changes to the biophysical world are already having harmful impacts on society, which will worsen with time.

As Professor Stefan Rahmstorf of the Potsdam Institute for Climate Impact Research reflected in 2007:

‘As climatologists, we’re often under fire because of our pessimistic message, and we’re accused of overestimating the problem… But I think the evidence points to the opposite - we may have been underestimating it.’77

Two years on, at International Scientific Congress on Climate Change in March 2009, Rahmstorf confirmed this view.

‘What we are seeing now is that some aspects are worse than expected’, he said speaking at a plenary session of the Congress. He continued: ‘I’m frustrated, as are many of my colleagues, that 30 years after the US National Academies of Science issued a strong warning on CO2 warming, the full urgency of this problem hasn’t dawned on politicians and the general public.’78


Dangerous climate change

Science on its own cannot give us the answer to the question of how much climate change is too much.
Margaret Beckett speaking at the Avoiding Dangerous Climate Change Conference (February 2005)

Margaret Beckett’s comments highlight the ethical and political dilemma of what constitutes a tolerable degree of climate change. Science can tell us what may happen as the temperature rises, but only we can decide what is tolerable and how far climate change should be allowed to go.

The United Nations Framework Convention on Climate Change (UNFCCC) was signed by over 160 countries at the United Nations Conference on Environment and Development held in Rio de Janeiro in June 1992, and came into force in 1994. The objective of the Convention was to slow and stabilize climate change by establishing an overall framework for intergovernmental efforts to respond to climate change.

 

It recognizes the significance of climate change and the uncertainties associated with future projections. But it also states that despite uncertainties, mitigating action should be taken - namely a ‘no regrets’ approach. Furthermore, it recognizes the responsibility of developed nations to take the lead due to their historical emissions, and therefore responsibility.

The long-term objective of the Convention, outlined in Article 2, is to achieve:

…stabilization of greenhouse gas concentrations in the atmosphere at a level that would prevent dangerous anthropogenic interference with the climate system. Such a level should be achieved within a time frame sufficient to allow ecosystems to adapt naturally to climate change, to ensure that food production is not threatened and to enable economic development to proceed in a sustainable manner.79

 

The burning embers diagram
An important part of the international climate change debate relates to the interpretation of dangerous climate change. This is of growing importance and of particular relevance to post-Kyoto negotiations.

In order to codify what ‘dangerous anthropogenic interference’ might mean, authors of the Third Assessment Report of the IPCC identified ‘five reasons for concern’.

 

These are listed below:80

  1. Risks to unique and threatened systems - e.g., coral reefs, tropical glaciers, endangered species, unique ecosystems, biodiversity hotspots, small island states and indigenous communities.
     

  2. Risk of extreme weather events - e.g., the frequency and intensity, or consequences of heat waves, floods, droughts, wildfires, or tropical cyclones.
     

  3. Distribution of impacts - some regions, countries and populations are more at risk from climate change than others.
     

  4. Aggregate impacts - e.g., the aggregation of impacts into a single metric such as monetary damages, lives affected or lost.
     

  5. Risks of large scale discontinuities - e.g., tipping points within the climate system such as partial or complete collapse of the West Antarctic or Greenland ice sheet, or collapse/reduction in the North Atlantic Overturning Circulation.

Figure 2, also known as the ‘burning embers diagram’ is an illustration of the IPCC’s five reasons for concern.

 

It shows that the most potentially serious climate change impacts (arrow heads) - expected to be experienced due to a range of equilibrium warming temperatures projected from stabilization levels between 400 ppm and 750 ppm of carbon dioxide equivalent (CO2e) - typically occur after only a few degrees of warming.81

In April 2009, a team of researchers, many of whom were lead authors of the most recent IPCC report, revised the burning embers diagram. While the diagram was rejected from the IPCC’s Forth Assessment Report because the artwork was also thought to be too unnerving, it was later published it in the peer-reviewed journal Proceedings of the National Academy of Sciences.

 

The updated diagram showed that an even smaller increase in global average surface temperature could lead to significant consequences for all five elements in the ‘reasons for concern’ framework.82

Figure 2:

Burning embers diagram 83

The solid horizontal lines indicate the 5-95 per cent range based on climate sensitivity

estimates from the IPCC 2001 and a study by one of the UK’s leading climate research unit, Hadley Centre study.84

The vertical line indicates the mean of the 50th per centile point.

The dashed lines show the 5-95 per cent range based on 11 recent studies.85

The bottom panel illustrates the range of impacts expected at different levels of warming.

 

 

Box 6: Sea-level rise
 

Rising sea levels will be one of the most significant impacts of climate change over the next century. This is because coastal zones are home to a significant proportion of humanity. These regions often have average population densities three times the global mean density.86

Tidal gauge and satellite data shows a global average sea-level rise of 1.8mm per year between 1961 and 2003.87 In recent years, however, this rate has increased to around 3.3 ± 0.4mm per year over the period 1993 to 2006.88 This observation is 40 per cent above the IPCC projected best-estimate rise of less than 2mm per year.

The main contribution to rising sea levels has been through thermal expansion of the oceans, but also a contribution from melting land-based ice (e.g. glaciers, and the Greenland and Antarctic ice sheets).

Due to a number of uncertainties about the way that ice-sheets behave, an accurate picture of future sea level rise is difficult to predict. Nevertheless, melt-water from Antarctica, Greenland and small ice caps could lead to a global sea level rise (the mean value of local sea level taken across the ocean) of between 0.75-2 m by the end of the century.89,90,91

However, recent research published by NASA’s James Hansen and a team of researchers warned that destabilization of the Greenland ice sheet is possible before global surface temperatures reach 2°C.92,93 This could lead to a sea-level rise of seven meters or more. While this rise may occur over a number of centuries, a mechanism of ‘albedo-flip’ could result in a much more rapid sea-level rise.94

 

The albedo-flip is a key feedback mechanism on large ice sheets, and occurs when snow and ice begin to melt. While snow cover has a high albedo (i.e. reflects back to space most of the sunlight striking it), melting ‘wet’ ice is darker and absorbs much more sunlight. A proportion of the melt water burrows through the ice sheet and lubricates its base, accelerating the release of icebergs to the ocean.

Such an extreme rise in sea level would have catastrophic implications for humanity. For example one study estimates that currently roughly 410 million people (or about 8 per cent of global population) live within five meters of present high tide.95

 

Allowing for population growth, this figure could well double over the course of the twenty-first century. Densely-populated Nile and Asian ‘mega-deltas’ may disappear in addition to large areas around the southern North Sea.

 

 

 

Aiming for 2°C
Historically, an increase in equilibrium temperature of Earth’s atmosphere by 2°C has been considered a ‘safe’ level of warming.

 

James Hansen’s warning that global temperatures should not be allowed to exceed 1.7°C, however, strongly suggests that a warming of 2°C cannot be described as ‘safe’.

 

As Professor Rahmstorf says:

‘If we look at all of the impacts, we’ll probably decide that two degrees is a compromise number, but it’s probably the best we can hope for’.

In 2007, NASA’s James Hansen argued in 2007 that temperatures should not go beyond 1.7°C (or 1°C above 2000 temperatures) if we are to avoid aiming to avoid practically irreversible ice sheet and species loss.96

 

For example, collapse of the Greenland ice sheet is more than likely to be triggered by a local warming of 2.7°C, which could correspond to a global mean temperature increase of 2°C or less.97, 98 The disintegration of the Greenland ice sheet could correspond to a sea-level rise by up to 7m in the next 1000 years, not to mention the positive climate feedback effects due to changes in land-surface reflective properties (see Box 6).

 

This would act to increase the warming as darker surfaces absorb more heat. Coral reef, alpine and Arctic ecosystems will also potentially face irreversible damage below a global average surface temperature rise of 2°C.99

In terms of the social impacts of climate change, what is manageable for some is actually catastrophic for others. For example, at the climate change conference in Copenhagen in late 2009, the Alliance of Small Island States - a grouping of 43 of the smallest and most vulnerable countries - rejected the 2°C target. They argued that 1.5°C is a better target, as many of their islands will disappear with warming beyond this point.100

Climate policy, therefore, needs to redefine what is described as a ‘safe’ level of warming or redefine its definitions from an acceptable level of warming decided by those who bear the least impact. Additionally, recent research (see Box 10) shows that real temperature outcomes are unlikely to be related to concentrations of greenhouse gases but rather a cumulative carbon budget.101,102

 

In other words, not only is 2°C unsafe, it is unhelpful when defining targets for climate policy. But, given that a 2°C target is now firmly established within the policy context, it is worth examining what it will mean should this temperature be exceeded.

 

The inter-agency report Two degrees, one chance published by Tearfund, Oxfam, Practical Action, Christian Aid states:

Once temperature increase rises above 2°C up to 4 billion people could be experiencing growing water shortages. Agriculture will cease to be viable in parts of the world and millions will be at risk of hunger. The rise in temperature could see 40-60 million more people exposed to malaria in Africa. The threshold for the melting of the Greenland ice-sheet is likely to have been passed and sea-level rise will accelerate.

 

Above 2°C lies the greater danger of ‘tipping points’ for soil carbon release and the collapse of the Amazon rainforest.103


Abrupt climate change: tipping points in the climate system
The Earth’s geological history is full of examples of abrupt climate change, when the climate system has undergone upheaval, shifting from one relatively stable state to another.

 

Transition to a new state is triggered when a critical threshold is crossed. When this happens, the rate of change becomes determined by the climate system itself, occurring at faster rate than the original forcing. For example, until 6000 years ago the Sahara Desert was a covered by vegetation and wetlands.

 

While the transition was driven by subtle and smooth changes in incoming solar radiation, at a critical point there was a regime shift in the rainfall patterns causing the landscape to switch from lush vegetation to desert, at a rate far greater than the original solar forcing.104

In 2008, Tim Lenton, Professor of Earth System Science and a team of researchers at the University of East Anglia, concluded that because of these critical thresholds in the climate system ‘society may have been lulled into a false sense of security’ by the projections of apparently ‘smooth’ climate change.105

 

The research suggested that that a variety of tipping elements of the climate system, such as the melting of ice sheets or permafrost could reach their critical point (tipping point) within this century under current emission trajectories. Tipping elements describe subsystems of the Earth’s system that are at least sub-continental in scale and can be switched - under certain circumstances - into a qualitatively different state by small perturbations. The tipping point is the corresponding critical point.

Tipping elements identified by the study include: collapse of the Greenland ice sheet; drying of the Amazon rainforest; collapse of the West Antarctic ice sheet; dieback of Boreal forests; greening of the Sahara/Sahel due to a shift in the West African monsoon regime; collapse of the North Atlantic ocean circulation; and changes to the El Niño-Southern Oscillation amplitude.

Whether or not these highly unpredictable factors are made part of decision-making is a political choice. But, given the existence of tipping points in the climate system, it is hard to reconcile the assumption that we may be able to stabilize the climate or even CO2 concentrations once a certain level of threshold of temperature or concentration of CO2 is reached.

 

But, the authors of the assessment identified a significant gap in research into the potential of tipping elements in human socio-economic systems, especially into whether and how a rapid societal transition towards sustainability could be triggered.106

If the impacts of climate change are non-linear then our response both in mitigating against and adapting to climate change also has to be non-linear.

 

 

 

Box 7. Time is running out 107
 

In August 2008 nef calculated that 100 months from 1 August 2008, atmospheric concentrations of greenhouse gases will begin to exceed a point whereby it is no longer likely we will be able to avert potentially irreversible climate change. ‘Likely’ in this context refers to the definition of risk used by the IPCC to mean that, at that particular level of greenhouse gas concentration, there is only a 66-90 per cent chance of global average surface temperatures stabilizing at 2°C above pre-industrial levels.

In December 2007, the likely CO2e concentration is estimated to be just under 377ppm, based on a CO2 concentration of 383 ppm. This seemingly counter-intuitive measure is explained by the proper inclusion in the CO2e figure of all emissions effecting radiative forcing - in other words, both those with cooling and warming effects.

If stabilization occurs at 400 ppm, there is a 10-34 per cent chance of overshooting a 2°C warming. Beyond this point, the probability of stabilizing global surface temperatures at less than 2°C decreases.

 

It would seem that if policy-makers are at all serious about avoiding dangerous climate change at a threshold of 2°C or less, emissions need to be reduced significantly.

 

 

 

What is the risk of overshooting 2°C under various stabilization scenarios?

We wouldn’t fly in a plane that had more than a 1 per cent chance of crashing. We should be at least as careful with the planet. Current climate policies provide us with far less than a 99 per cent chance of avoiding catastrophic climate change.108
Paul Sutton, Carbon Equity

When the Kyoto Protocol was established in 1997, the best scientific understanding implied that a 50 per cent reduction in emissions below 1990 levels by 2050 would be sufficient to avoid dangerous climate change.

 

Thirteen years on, the understanding of what constitutes safe climate change has improved significantly. Now, there is a growing consensus that at least an 80 per cent reduction in CO2 emissions below 1990 levels will be required by 2050 globally if we are to have a greater than 60 per cent chance of not exceeding 2°C.109

 

A recent analysis by the Tyndall Centre for Climate Change Research demonstrated what this means for the UK. Incorporating all sectors of the economy, the UK is required to reduce its carbon dioxide emissions by some 70 per cent by 2030, and around 90 per cent by 2050.110

Not only is the safe level of temperature rise misleading as described earlier, a number of assessments exploring the probability of exceeding various temperature thresholds have been published. These studies demonstrate that the stabilization of atmospheric concentrations of greenhouse gases at anything above 400 ppm is too high to avoid a temperature rise of 2°C.111,112

Research led by Malte Meinhausen, a climate modeler based at the Potsdam Institute for Climate Impact Research in Germany, has shown that stabilization of greenhouse gas concentrations (defined as CO2e) at 550 ppm is accompanied by the risk of overshooting 2°C warming by 68-99 per cent.113 According to the IPCC, this is defined as ‘likely’ to ‘very likely’.114 Meinhausen’s work also suggests that only by stabilising emissions at 400 ppm is it ‘likely’ that the climate will stabilise at 2°C.

In early 2009, however, James Hansen and colleagues at Columbia University contended that current atmospheric concentrations of CO2 need to be reduced to 350 ppm.115 Hansen’s analysis for the first time used a climate sensitivity parameter (temperature change due to an instant doubling of CO2) that included slower surface albedo feedbacks.

Traditionally, the climate sensitivity parameter only includes fast-feedbacks (i.e. changes to water vapor, clouds and sea-ice) whilst keeping slow changing planetary surface conditions constant (i.e., forests and ice sheets). In addition, long-lived non-CO2 forcings (other gases and aerosols) are also kept constant over time. It is worth noting, to avoid any confusion, that Hansen and his team were specifically referring to CO2 only - not CO2e which also includes non-CO2 forcings.

The paper concluded with the harrowing warning:

‘If humanity wishes to preserve a planet similar to that on which civilization developed and to which life on Earth is adapted, paleoclimate evidence and ongoing climate change suggest that CO2 will need to be reduced from its current 385 ppm to at most 350 ppm, but likely less than that.’116


Questioning climate policy assumptions
Certain assumptions underlie scenarios for the future stabilization of greenhouse gas emissions and of their accumulation in the atmosphere.

 

These include that historical rates for both energy efficiency improvements and declining energy intensity will continue and accelerate into the future. In turn, it is assumed that these will result in an absolute decrease in energy consumption.

 

Yet, these assumptions are hugely dependent on three questions that are not so much unanswered, as barely even asked:

  1. Is the stabilization of greenhouse gases through long-term targets the most effective response to climate change?

  2. What are the theoretical and practical limits to energy efficiency of the economy?

  3. Do increases in energy efficiency actually result in decreases in the demand for energy services?

Under this questioning, current climate change policies appear seriously flawed, worsening the prognosis for future climate change and our ability to deal with it.

For example, there are theoretical limits to efficiency governed by the laws of thermodynamics. There are practical limits to efficiency, relating to economic, social and political barriers, and the speed at which we can replace current energy systems.

Observations in the real world suggest that increases in energy efficiency can have perverse consequences, resulting in rises in the demand for energy services - the so called ‘rebound effect’ (see Box 8).

Technological optimists believe that technical innovations will reduce the demand for energy.117 But, in fact, technological improvements have tended to push demand for high levels of ‘service use’ and greater consumption. The history of fuel efficiency in cars is one such example (see Box 8).118,119

Before these questions are addressed, it is necessary to be clear about what is meant by energy efficiency, energy intensity and ‘carbon intensity’.

 

All three terms are described in more detail below, but fundamentally, they all represent ratios. This means they place more emphasis on outputs, rather than inputs. And, as long as the growth rate in consumption of the input increases at a greater rate than efficiency (intensity) increases (decreases) any improvements to the system are effectively ‘eaten up’. In other words, no absolute reduction to energy consumption or carbon (in the case of carbon intensity) would be observed.

For example, at the global level, even if technological energy efficiency and the uptake of new, more efficient devices increased by 50 per cent over the next 20-30 years with GDP rising at a conservative 2.5 per cent, within 25 years, we’d be back where we are now.120,121,122

 

 

 

Box. 8: The rebound effect

It is a confusion of ideas to suppose that the economical use of fuel is equivalent to diminished consumption. The very contrary is the truth.
William Stanley Jevons (1865)123

There is no evidence that using energy more efficiently reduces the demand for it.
Brookes (1990)124

Despite the recognition that consumption levels need to decline in developed nations, governments and businesses are reluctant to address the restriction of consumption. Yet, without limits to consumption, improvements in efficiency are often offset by the ‘rebound effect’.125

For example, a recent report published by the European Commission’s Joint Research Centre (JRC) showed an increase in energy use across all sectors - residential, service and industry - in recent years, despite improvement in energy efficiency.126

 

For example, in the domestic sector while new measures have led to some improvements, particularly in the case of ‘white goods’ (e.g. refrigerators, washing machines, dishwashers), the increasing use of these products and other household appliances, such as tumble driers, air conditioning and personal computers, has more than offset savings.

The ‘rebound effect’ was an observation made by William Stanley Jevons in his book The Coal Question, published in 1865.127 Here, Jevons contended that although technological advancement improves the overall efficiency (E) with which a resource is used, efficiency gains rebound or even backfire, causing higher production and consumption rather than stabilization or reduction. Since improvements generally reduce the cost of energy per unit, economic theory predicts that this has the effect of triggering an overall increase in consumption.

If a car, for instance, can drive more kilometers on a liter of petrol, the fuel costs per kilometer fall, and so will the total costs per kilometer. The price signal acts to increase consumption and, thus, part of the efficiency gains is lost.

One area where the rebound effect is prominent is domestic energy consumption. An analysis of energy consumption before and after installation of energy savings measures found that only half of the efficiency gains translate into actual reductions in carbon emissions.128 This is supported by more recent analysis of the effectiveness of England’s Home Energy Efficiency Scheme (Warm Front).

 

While there are appreciable benefits in terms of use of living space, comfort, quality of life, physical and mental well-being, the analysis found that there was little evidence of lower heating bills.129 This has also been observed in Northern Ireland.130 In other words, improvements in energy efficiency are offset by increased levels of thermal comfort.

An more in-depth economy-wide assessment of the rebound effect carried out on behalf of the UK Energy Research Council in 2007 found that rebound effects are not exclusive to domestic energy consumption.131 They can be both direct (e.g., driving further in a fuel-efficient car) and indirect (e.g., spending the money saved on heating on an overseas holiday).

 

Findings from the research suggest that while direct rebound effects may be small - less than 30 per cent for households for example, much less is known about indirect effects. Additionally, the study suggests that in some cases, particularly where energy efficiency significantly decreases the cost of production of energy intensive goods, rebounds may be larger.

A further rebound effect is caused by ‘time-saving devices’.132 With the current work-and-spend-lifestyle implicit to industrialized societies, there is an increase in the demand for time-saving products. Although these devices save time, they also tend to require more energy, for example, faster modes of transport.
 


How large is the rebound effect?
How much energy savings are eaten up by the rebound effect is surrounded by lively debate. Estimates range from almost nothing in the energy services 133 to being of sufficient strength to completely offset any energy efficient savings.134,135 There are a number of empirical analyses, however, that suggest that the rebound effect may be real and significant (Table 2).136

The majority of work investigating the rebound effect has focused on a few goods and services.137 However, the few studies that explore the macroeconomic impact of the rebound effect, find it to be significant. For example, using a general equilibrium model, one study by environmental economist Toyoaki Washida assessed the Japanese Economy.139

 

On testing a variety of levels of CO2 tax, the rebound effect was found to be significant (between 35-70 per cent of the efficiency savings).

 

Table 2.

Summary of empirical evidence for rebound effects138

 

Policy implications
The policy implications of the rebound effect are that energy/carbon needs to be priced so the price remains relatively constant while efficiencies improve. Surprisingly, however, rebound effects have often been neglected by both experts and policymakers.

 

For example, they do not feature in the recent Stern and IPCC reports or in the UK Government’s Energy White Paper.140

 

According to Steve Sorrell, a senior fellow at the Science and Policy Research Unit at the University of Sussex:

‘This is a mistake. If we do not make sufficient allowance for rebound effects, we will overestimate the contribution that energy efficiency can make to reducing carbon emissions. This is especially important given that the Climate Change Bill (now Act) proposes legally binding commitments to meet carbon emissions reduction targets.’141

 

 

 

Box 9. The history of fuel efficiency in cars
 

Technological improvements in fuel efficiency have largely been offset by traffic growth and low occupancy rates. The increase in traffic has affected the ability of drivers to utilize the maximum vehicle efficiency speed, but the increase in traffic also means that demand for safer vehicles has significantly increased the weight of vehicles. 

 

This phenomenon is termed ‘cocooning’, and is due to the fact that we now spend so much time in cars. Vehicles now have more and more gadgets to provide greater levels of comfort as people spend more and more time sitting in traffic jams or traveling further distances. This has had the effect of increasing the weight of the vehicle and also the energy required to power the gadgets in them.

Table 3 compares the fuel consumption of the Volkswagen Golf (a reference case for all compact family cars) over the period 1975-2003. Since 1975, fuel consumption has improved by a measly 5 per cent. When compared with the weight of the vehicle, it is clear that the reason for the modest improvements in fuel consumption is due to a greater than 50 per cent increase in the weight of the vehicle.142
 

Table 3.

Fuel consumption of the Volkswagen Golf 1975-2003.

 

 

 

Are long-term stabilization targets the correct policy response?
Recent modeling studies suggest that stabilization of greenhouse gas emissions is far from the most effective policy response to climate change.

 

For example, using a coupled climate carbon-cycle model, one study found that from a suite of nine IPCC stabilization scenarios, eight showed that temperatures did not appear to stabilize over the next several centuries, but rather continued to increase well beyond the point of CO2 stabilization at around 2400.143

 

The continuation of temperature increase beyond atmospheric CO2 stabilization is due to the long thermal memory (i.e., long-term changes in planetary albedo, due to loss of ice caps, changes in cloud cover, etc.) and equilibration time of the climate system.144

Given this, many are now calling for a policy response of a ‘peak and decline’, not just in carbon emissions but also atmospheric concentrations of CO2. The faster and further that greenhouse gas concentrations can be lowered below their peak, the lower the maximum temperature reached will be.

 

 

 

Box 10. The trillionth tonne
 

Due to uncertainties in the carbon-cycle (see Box 11), the final equilibrium temperature change associated with a given stabilization concentration of greenhouse gases is poorly understood. In order to address this uncertainty, a number of studies have begun to quantify cumulative greenhouse gas emissions that would limit warming to below 2°C.145

Meinshausen, also the lead author of one of the studies, found that in order to stand a 75 per cent chance of keeping temperatures below 2°C, the world has to limit the cumulative emissions of all greenhouse gases to approximately 1.5 trillion tonnes of CO2e. To reduce the risk by another 5 per cent, this means capping total emissions to just over 1 trillion tonnes of CO2e. 146

Myles Allen, Head of the Climate Dynamics group at University of Oxford’s Atmospheric, Oceanic and Planetary Physics Department and lead author of another study comes to similar a conclusion. Allen and his colleagues argue that if humans can limit cumulative emissions of carbon to one trillion tonnes of carbon, there is a good chance not exceeding the 2°C. They estimate that we could follow our current emissions pathway for another 40 years, and then would have stop emitting carbon in to the atmosphere altogether.147

But, this doesn’t mean we’ve got 40 years; far from it. Meinshausen argues that if emissions are still 25 per cent above 2000 levels in 2020, the risk of exceeding 2°C shifts to more likely than not.148 That’s a reduction in global emissions by 2.5 per cent year on year, starting now. Given that emissions are currently growing at approximately 3.5 per cent per year - this represents a phenomenal challenge, and requires unprecedented action.

 

 

 

Box 11. Global carbon-cycle feedbacks
 

Less than half of fossil fuel carbon emitted remains in the atmosphere. For the period 2000-2005, the fraction of total anthropogenic CO2 emissions remaining in the atmosphere has been around 0.48.

 

This is because over half has been sequestered by the global carbon cycle.149 This fraction, however, has increased slowly with time, implying that there is a weakening of the carbon sinks relative to emissions.150

 

A very real and immediate concern is what effect both increasing concentrations of CO2 in the atmosphere and the subsequent temperature affect will have on the global carbon cycle.
 

Figure 3:

Global flows of Carbon 151
 

Weakening of terrestrial and oceanic sinks could accelerate climate change and result in a greater warming. This will therefore mean even lower levels of anthropogenic greenhouse gas emissions than currently imagined may be necessary to achieve a given stabilization target.

Positive terrestrial carbon-cycle feedbacks result from a combination of increased soil respiration and decreased vegetation production due to climate change. Positive oceanic carbon-cycle feedbacks, however, result from decreased CO2 solubility with increasing ocean temperature, as well as changes in ocean buffering capacity, ocean circulation and the solubility pump (the mechanism that draws atmospheric CO2 into the ocean’s interior). The net effect is to amplify the growth of atmospheric CO2.

Observations suggest that the carbon sinks may already be weakening. For example a paper published in the journal Science in 2007 found evidence to suggest that the Southern Ocean sink of CO2 had weakened over the period 1981-2004.152 This is significant because the Southern Ocean is one of the largest carbon sinks, absorbing 15 per cent of all carbon emissions. The study found that the proportion of CO2 absorbed by the Southern Ocean had remained the same for 24 years, even though emissions have increased by around 40 per cent during the same period.

The Southern Ocean is now absorbing 5-30 per cent less CO2 than previously thought. It is believed that a strengthening of Southern Ocean winds caused by man-made climate change has reduced the efficiency of transportation of CO2 to the deep ocean. Rather than CO2 finding its way into the deep ocean, where it stays, it is being released by the increase in ocean mixing caused by strengthening winds. Worryingly, climate models did not predict this would happen for another 40 years.153

Whilst carbon-cycle feedbacks have been recognized for a number of years, it is only recently, with the aid of coupled carbon-cycle climate models, that these CO2 feedbacks have been quantified. A recent study that compared a number of different carbon-cycle climate models suggested that by the end of the century, the additional concentration of CO2 in the atmosphere due to carbon-cycle feedbacks could be between 20 and 200 ppm, with the majority of the models lying between 50 and 100 ppm.154

Recent real time observations suggest an already increased rate of atmospheric CO2 growth and reduced efficiency of ocean sinks, leads us to fear that the additional CO2 input into the atmosphere is more likely to reside near the upper-most figure.


 

The latest IPCC report also acknowledges that anthropogenic warming and sea-level rise could continue for up to 1000 years after stabilization of atmospheric greenhouse gas concentrations.155

 

This, of course, makes the very notion of stabilization of the climate untenable due to the complex push-pull relationship between temperature and CO2 concentration.
 


The relationship between economics, growth and carbon emissions
There was a time when the relationship between carbon emissions and economic growth seemed so simple.

 

Until recently, it was often argued that the relationship between income and CO2 emissions followed the Environmental Kuznets Curve (EKC) model. The EKC evolved from Simon Kuznets’s original thesis on economic growth and income inequality.156 Kuznets postulated that with economic growth, income inequality first increases over time, and then at a certain point begins to reverse. In theory, then, the relationship between economic growth and income inequality placed on a graph takes on the shape of an inverted-U.

In environmental economics, the EKC proposes a relationship between environmental pollution and economic activity.157 The theory again suggests an early rise in pollution that later reverses its relationship with growth. Several attempts have been made to determine whether the EKC paradigm can be applied to per capita emissions of CO2 in the form of a Carbon Kuznets Curve.158

Some early literature on the subject does suggest that there is a relationship between per capita income in a country and the per capita or gross emissions in the country.159,160 There is now unequivocal evidence, however, that in the case of carbon emissions, the EKC simply represents idiosyncratic correlations and holds no predictive power.161

For example a recent study published in the journal Proceedings of the National Academy of Sciences found that income was the biggest driver of ever increasing emissions.162 Of nine regions, which included developed regions such as the USA, Europe, Japan, and developing regions such as China, India, all showed a strong correlation of increasing emissions and income.

The problems of directly applying the EKC paradigm to greenhouse gases are twofold.

First, key greenhouse gases have a long atmospheric lifetime 163 compared to other environmental pollutants, such as particulates. Their long atmospheric lifetime means that their environmental impact is transboundary, i.e., their effect on the climate is not restricted to the region within which they are produced.

 

Given the asymmetries of the stages of economic development between nations, in principle the EKC model for global climate change cannot work, and the connection between control of domestic emissions in higher-income countries and the benefits to their citizens is very weak. Calculations based on direct national emissions are also misleading because they fail to account for the ‘embedded’ carbon of goods manufacture abroad and consumed domestically.

 

For example, the effect of much of Britain’s heavy industry and manufacturing having been ‘outsourced’ to less wealthy countries creates the impression that Britain pollutes less, now that it is richer. In fact, the pollution has largely been outsourced too. It still exists, but not on Britain’s official inventory of emissions (see Box 12).

Second, we are constrained by the arrow of time. There is clear evidence to suggest that both developed and developing economies would begin the decline on the inverted-U curve well beyond concentrations of greenhouse gases that are classed as safe.164

 

In other words, by the time we got to the less polluting slope of the curve, we would already have gone over the cliff of irreversible global warming; it would be too late to be green.

 

 

Box 12. Carbon laundering:

the real driver of falling carbon and energy intensity in developed nations
 

As economies develop, historically there is a move away from heavy industry towards service-driven economies that are less energy intensive, so-called ‘post-industrialism’. The crude method of national reporting of carbon emissions, and therefore carbon intensity, further reinforces the impression of declining environmental impact.

Yet, in fact, nothing could be further from the truth. In a global economy, it’s not just about how the majority of a nation’s population earn their living, it’s also about how they consume. High incomes have conventionally led to high consumption. So rather than declining carbon emissions, high end-service economies actually increase global energy and material throughput, outsourcing production to other nations rather than decarbonizing and dematerializing the economy.

In 2001, over five billion tonnes of CO2 were embodied in the international trade of goods and services, most of which flowed from developing nations (non-Annex 1 nations of the UNFCCC) to developed nations (Annex 1 nations of the UNFCCC) - i.e., five billion tonnes excluded from developed nations emissions inventories.165 This is greater than total annual CO2 emissions from all EU25 nations combined.166 This means, in effect, the economies of countries like the UK and the USA are ‘laundering carbon’ to offshore carbon inventories.

This was illustrated by the report from City firm, Henderson Global Investors, The Carbon 100.167 It suggested that the UK may be responsible for more than the ‘official’ 2.13 per cent responsibility often claimed by politicians.

The Carbon 100 suggested that the UK was actually responsible for between six to eight times more than this (around 12-15 per cent). Tracing the worldwide activities of the UK’s leading companies listed on the UK stock exchange paints a more accurate picture of the UK’s real emissions responsibility.

While establishing the ‘embodied emissions’ of trade is notoriously difficult, a recent study published by researchers from Lancaster University’s Environment Centre explored the carbon embodied within trade flows between the UK and China.168 The study showed that imports from China to the UK were embodied with 555 million tonnes of CO2 in 2004.

 

Put another way - the carbon embodied in trade reduces the apparent CO2 emissions of UK consumers by 11 per cent, but increases the real carbon footprint of UK consumers by 19 per cent and global emissions by 0.4 per cent. This is due to the carbon inefficiencies of Chinese industrial processes compared to those in the UK. Furthermore, the study estimated that the shipping of goods from China to the UK in 2004 resulted in the emission of perhaps a further 10 Mt CO2.

 

This estimate falls towards the high end of earlier estimates for embodied carbon for all of the UK’s trade partners. It also means that the UK’s progress towards its Kyoto emission targets of 12.5 per cent below 1990 levels vanishes into the global economic atmosphere.

This suggests that carbon or energy intensity is an indicator that is grossly misleading at the national level. As a country moves towards post-industrialism, the goods demanded by the high-consumption society are simply produced elsewhere, resulting in a displacement of emissions.

 

A more accurate indicator of changes in domestic emissions would be on a per capita basis based on the average individual ecological/carbon footprint. See for example nef’s report Chinadependence: The second UK Interdependence Day report.169

 

 

 

Energy efficiency, energy intensity and carbon intensity
There are two types of energy efficiency improvements. The first relates to the development or exploration of more sustainable conversion technologies ranging from renewable technology, to improved efficiency of electricity generation. This will be referred to as supply-side efficiency or εss.

The second relates to the improvement in energy efficiency of demand-side applications or end-use efficiency -εeu. For example, εeu can be improved by increasing the efficiency of light bulbs, fridges, televisions, and so on.

The overall efficiency (Ε) of converting primary energy into GDP can therefore be defined as the product of εss and end-use efficiency εeu. Energy intensity (energy use per unit of GDP) is the inverse of Ε, this is shown in Equation 1.

 

Equation 1

= (useful energy/primary energy) × (GDP/useful energy)

= (GDP/primary energy)

= 1 / Energy intensity of the economy

 

 

Box 13. The Kaya Identity
 

The Kaya Identity, developed by the Japanese energy economist Yoichi Kaya plays a core role in the development of future emissions scenarios in the IPCC Special Report on Emissions Scenarios (SRES).170,171

 

It shows that total (anthropogenic) emission levels depend on the product of four variables: population, Gross Domestic Product (GDP) per capita, energy use per unit of GDP (energy intensity) and emissions per unit of energy consumed (carbon intensity of energy).

 

The Kaya Identity is shown in Equation 2. It has, been adapted to take into account natural carbon sinks.172

Equation 2

Where:

Net F is the magnitude of net carbon emissions to the atmosphere
F is global CO2 emissions from human sources
P is global population
G is world GDP and g = is global per-capita GDP,
E is global primary energy consumption and e = is the energy intensity of world GDP,
and f = is the carbon intensity of energy,
S is the natural (or induced) carbon sink.

Climate policy so far has dealt with the second half of the equation - energy intensity of the economy and carbon intensity of energy. For the former, the ratio is expected to decline over time through improvements in efficiency of both supply and demand. Carbon intensity (f) of energy relates to improvements in efficiency of carbon-based energy supply and decarbonisation of the energy supply through diffusion of renewable energy technology (wind, hydro, solar, geothermal and biomass) nuclear fission.

In terms of population growth, it is noteworthy that fertility rates in the developed world have fallen dramatically in recent decades.173

 

In terms of both social justice and effectiveness, the education of women is the only viable option for the long-term stabilization of population growth. This, in turn, is dependent on the progress of human development. Nevertheless, by 2050, the global population is expected to reach nine billion.174 Economic growth, however, defined by changes in g remains, by and large and as discussed earlier in this report; unchallenged.

Recent evidence suggests that the surge in emissions growth is primarily due to increases in economic activity. The growth rate of CO2 emissions includes carbon-cycle feedbacks (see Box 11) as well as direct anthropogenic emissions.

 

Of the 3.3 per cent average annual growth rate of emissions between 2000 and 2006, 18 ± 15 per cent of the annual growth rate is due to carbon-cycle feedbacks, while 17 ± 6 per cent is due to the increasing carbon intensity of the global economy (ratio of carbon per unit of economic activity).175

 

The remaining 65 ± 16 per cent is due to the increase in the global economic activity.

While it is often argued that technological innovation could in theory improve resource and energy efficiency and lead to decarbonisation of the economy, recent evidence challenges this view. This is discussed later on in this report.

 

 

Energy intensity, the amount of primary energy required to generate economic activity (GDP), is a standard for energy use per unit of productivity. While carbon intensity refers to the carbon produced for each unit of productivity (see Box 13 on the Kaya Identity). Since carbon emissions and energy consumption are currently so strongly coupled, at present the two terms can effectively be viewed as roughly analogous.176

Logically, therefore, energy intensity improvements will only reduce emissions if improvements are made at a greater rate than increases in GDP.

 

For example, the International Energy Agency’s (IEA) World Energy Outlook 2009 projects increases in global average economic growth by 3.1 per cent per annum between 2007 and 2030. In order to observe absolute reductions in carbon emissions of 1 per cent per year, the carbon intensity of the economy would need to improve by 4.1 per cent per year assuming energy intensity remains the same.177

Historically, global carbon intensity of energy has declined at a average rate of around 1.3 per cent per year since the mid-1800s.178 However, disaggregating these data over the past 40 years gives a more much more detailed picture (see Table 4).

Since 1971 global carbon intensity of the energy has fallen, on average fallen by just 0.15 per cent each year, with a maximum annual decline of 0.41 per cent between 1980 and 2000. However, in recent years the carbon intensity of energy has increased at a rate of 0.33 per cent between 2000-2007. This increase in carbon intensity of energy is due to the increased use of coal in recent years.

 

While coal use grew less rapidly than all other sources of energy between 1971-2002 over the past four years this trend has been reversed.

 

Coal use is now growing by 6.1 per cent each year, more than double the rate of all other energy sources.179 This rise in carbon intensity of energy has more than offset the small improvements in energy intensity of the economy - bringing improvements to carbon intensity of the economy to a standstill and causing total carbon emissions to soar.

Even in developed nations, carbon intensity of the economy and energy have never managed to reach the levels required to stop total carbon emissions rising year on year. Table 5 shows changes in carbon intensity in the United States. Since the 1950s, carbon intensity in the USA declined at an average rate of around 1.6 per cent per year, with a maximum annual decline in carbon intensity of 2.7 per cent between 1980 and 1990.

 

Current rates of carbon intensity fall are now around 1.6 per cent annually.

At a time where never before has there been so much financial and intellectual capital directed towards innovation to improve the carbon and energy intensity of the economic system, this slowdown of improvements implies that we may be reaching the practical limits of efficiency.

 

Table 4:

Change in global carbon intensity180

 

 


Box 14: The UK: Leading by example?
 

The UK ‘dash for gas’ was largely responsible for the relative ease with which the UK reached its Kyoto Protocol targets.

 

For example, the Royal Commission for Environmental Pollution (RCEP) states that the UK’s emission reductions are ‘largely fortuitous’.181 The ‘dash for gas’ was due to the rapid shift in electricity generation from coal to gas in the early 1990s.182 This was an unintended consequence of the Conservative government’s liberalization of the energy market. Although this has been supplemented by changes to industrial processes, waste management and the outsourcing of production to developing nations such as China and India (see Box 12).183

A Defra (Department for the Environment, Food and Rural Affairs) spokeswoman said the UK had already beaten its 2012 emissions target of 12.5 per cent under the Kyoto protocol and that the figures for 2005 showed a reduction of 15.3 per cent on 1990 levels. ‘The action we have taken to cut our greenhouse gas emissions at the same time as maintaining economic growth makes us an exemplar,’ she said.184 In reality, the majority of the UK’s emissions reductions have simply been achieved through this fuel switch (and outsourcing of production).

For example, simply by displacing 1400GW of base load coal-fired power stations with 1400GW of energy efficient combined cycle gas turbine (CCGT) power stations could save approximately 1 billion tonnes of carbon (3.67 billion tonnes of CO2) per year.185 Indeed, this has been proposed as one such method of reducing global emissions.

 

But, as we shall see later, like its fuel-cousin oil, natural gas too is facing production limits.

The UK Climate Change Program (2006) suggests that 25 per cent of emissions reductions in the UK were due to fuel switching in 1990s from coal to gas.

 

A further 35 per cent of reductions were thought to be due to energy efficiency (but could equally be due to outsourcing of production), and a further 30 per cent of reductions were due to the reduction of non-CO2 greenhouse gases (comparatively less reduction compared to CO2 due to the higher global warming potential of, for example, methane, nitrous oxide and fluorinated gases).

 

 

Table 5:

Change in carbon intensity in the United States186



Globally, stabilization of CO2 to a safe level would require an 80-90 per cent reduction in current anthropogenic CO2 emissions.

 

Worldwide, they are actually growing by 3.4 per cent a year (average over 2000-2008 period).187 At the same time, carbon intensity of energy is increasing by on average 0.33 per cent per annum. This trend is unlikely to change at least for the remaining 3-year term of the Kyoto protocol.

So is growth really possible?

 

Back to Contents
 

 

 


Scenarios of growth and emission reductions

If humanity wishes to preserve a planet similar to that on which civilization developed and to which life on Earth is adapted… CO2 will need to be reduced from its current 385 ppm to at most 350 ppm CO2, but likely much less than that… If the present overshoot of this target CO2 is not brief, there is a possibility of seeding irreversible catastrophic effects.188
James Hansen NASA/Goddard Institute for Space Studies

Since changes to both carbon intensity of energy and the economy are assumed to play such a major role in mitigating strategies, we ask - what declines in carbon intensity are necessary to meet a number of emissions scenarios ranging from the high risk, to what current science implies is necessary?

To address this question, we performed a number of analyses that examine the relationship between growth, carbon intensity of the economy, energy intensity of the economy (efficiency), carbon intensity of energy and emissions reductions. Focusing specifically on CO2 rather than other greenhouse gas emissions, we modeled future consumption of fossil fuels.

 

Since CO2 produced by burning fossil fuels is approximately 70 per cent of all anthropogenically produced greenhouse gases, has a long atmospheric lifetime and is the best studied and modeled of the greenhouse gases, just focusing on CO2 is a good starting point.

Unless otherwise stated, conversion and emissions factors and historical data on carbon emissions have been taken from the Carbon Dioxide Information and Analysis Centre, a section of the US Department of Energy.189,190

 

Data from the World Resources Institute Climate Analysis Indicators Tool were also used.191
 


The scenarios
Although the IPCC has produced a suite of scenarios that describe possible future emissions pathways, they are non-mitigating (i.e., they do not consider climate-related policy), so will not include the impacts of current climate policies.192

 

They also incorporate a wide range of possible technical, social, and economic factors that are difficult to break down into their component parts.

Figure 4.

Comparison of the RS and AP scenarios presented in the World Energy Outlook, 2006

to the updated World Energy Outlook, 2008 RS and AP scenarios.

 

Given this, more recent scenarios constructed by the IEA form the basis of our analysis, particularly as they do include mitigation policies.

 

In this report, we explore the implications of the World Energy Outlook - 2006 Reference (RS) and Alternative Policy (AP) scenarios. These scenarios are primarily driven by four parameters: economic growth, demographics, international fossil fuel prices and technological developments.

At the time of publication, the IEA had produced a further three World Energy Outlook reports since 2006, each containing a revision of these scenarios. However, on comparing the emission pathways for the 2007 and 2008 editions (shown in Figure 4) we find little divergence over the 50-year timeframe - the period that is the focus of our analysis.

The 2008 RS and AP-550 scenario are not dissimilar to the 2006 RS and AP scenarios. While there is a large difference between the AP-550 and AP-450, assumptions made about the latter scenario are arguably questionable and unrealistic (see Box 15). Given that the RS and AP-550 scenarios are both similar to the World Energy Outlook 2006 RS and AP scenarios, we base our analysis on the 2006 scenarios.

 

In the 2009 edition (not shown), again there is minimal divergence from the RS. However, the only AP analyzed is AP-450, which follows a similar trajectory to the World Energy Outlook 2008 AP-450 scenario.


 

 

Box 15. WEO - 2008 Scenarios
 

Reference scenario
The reference scenario (RS) includes the effect of government policies up until mid-2008, but not new ones. For the RS, CO2 is expected to have doubled by 2100, reaching 700 ppm (CO2 only) and 1000 ppm (CO2e). World primary energy demand increases at a slower rate than in previous RSs, due to the recent economic slowdown and implementation of new climate policies. This translates into annual carbon emissions that are just 1GtC less than the WEO 2007 RS. The RS also assumes decrease in CO2 intensity by 1.7 per cent per annum (pa.)
 


Alternative scenarios
The alternative scenarios (APs) assume negotiations for the next phase of the Kyoto Protocol agree stabilization targets of either 550 ppm CO2e or 450 ppm CO2e, which are achieved by 2200. These are both peak and decline scenarios. In other words, the target atmospheric concentration of CO2e is overshot, and subsequently reduced.

 

Given this, in the AP-450 scenario, atmospheric concentrations of CO2 peak between 2075 and 2085, and then begin a long-term decline to 2200. For the AP-550 scenario, atmospheric concentrations of CO2 peak in the middle of the next century and slowly decline to 550 ppm CO2e by 2200. Scenarios are met through three key climate policy mechanisms - cap-and-trade, sectoral agreements, and national policies and measures.

In order to meet these targets, the scenarios assume significant growth in low-carbon energy such as: hydroelectric power, nuclear, biomass, other renewables and carbon capture and storage.
 


AP-550-AS
AP-550 assumes that while world primary energy demand increases by 32 per cent over the period 2006-2030 (0.4 per cent slowdown in growth rate compared to the RS), emissions would rise by no more than 32,900 MtCO2 in 2030 and decline thereafter. This requires a $4.1 trillion investment into low carbon energy related infrastructure or 0.2 per cent of annual GDP.

The change in energy mix results in a decline in CO2 intensity by 2.6 per cent pa. This is due to both the increase in low carbon energy and the decrease in the average quantity of CO2 emitted per tonne of fossil fuel energy. By 2030, low carbon energy would account for 26 per cent of the primary energy mix compared to 19 per cent in 2006.

 

This level of decarbonisation of the power sector is equivalent to seven coal-fired plants and three gas-fired plants with carbon capture and storage (CCS), 11 nuclear plants, 12,000 wind turbines each year and the equivalent of three, Three Gorges Dams every two years. In addition, emissions from fossil fuel energy fall from 2.94 tonnes/toe in 2006 to 2.83 tonnes/toe in 2030. This is due to a falling share of coal in the primary energy mix.
 


AP-450-AS
AP-450 requires that CO2e emissions fall dramatically in 2020 from 35,000 MtCO2e to 27,500 MtCO2e by 2030 and 14,000 MtCO2e by 2050. Energy efficient improvements at both production and end-use levels result in a low growth rate in energy demand (0.8 per cent pa). By 2030, low carbon energy will account for 36 per cent of the global primary energy mix (including CCS), costing $9.3 trillion or 0.6 per cent of annual world GDP.

 

Fossil fuels still account for 67 per cent of the primary energy demand in 2030; however, there is an assumption that CCS technologies will be more widespread in the power generation sector and will also be introduced into industry. Additionally, it is assumed 13 nuclear power stations need to be built each year and biofuels are more widespread the transport sector.

 

Beyond 2030, the power sector becomes ‘virtually decarbonised’ with a strong emphasis on CCS in the power and industrial sectors and electric, hybrid and biofuels in the transport sectors (private and goods vehicles, shipping and aviation).

Between 2006 and 2030, there is:

  • a tenfold increase in wind, solar and other renewablesan increase in modern biomass (modern bioenergy plants that use organic waste or cultivated feedstocks) by almost 80 per cent

  • a near doubling of nuclear energy

There are three fundamental critiques of these scenarios, however. First, it is noteworthy that recent research by Lowe et al. has stated that in order to have less than a 50 per cent change of not exceeding 2°C, emissions need to peak by 2015 and fall by 3 per cent each year thereafter.

 

Neither the RS nor the AP scenarios achieve such an early and dramatic peak and decline scenario. Lowe et al. also note that even if emissions peak in 2015, there is still a one-in-three chance that near-surface temperatures will rise by more than 2°C in 100 years’ time.193

 

The IEA, however, dismisses a scenario that does not achieve overshoot stating:

‘A 450 stabilization trajectory without overshoot would need to achieve substantially lower emissions in the period up to 2020 and, realistically, this could be done only by scrapping very substantial amounts of existing capital across all energy-related industries. In any case, given the scale of new investment required, it is unlikely that the necessary new equipment and infrastructure could be built and deployed quickly enough to meet demand.’194

Wigley et al. also note that a policy that allows emissions to follow an overshoot pathway means that in order to recover to lower temperatures within a century timescale, we may, for a period, require negative global emissions of CO2.195

Second, the assumptions about growth in capacity of CCS are also overly optimistic. The consensus view is that CCS may be commercially viable by 2020; however, a number of analysts believe even this is an optimistic scenario suggesting that 2030 may be more realistic.

Third, given the optimism attached to CCS as a viable technology in the near future, the assumption that CO2 intensity can feasibly decline by 2.6 per cent per year can also be viewed as over optimistic. Figure 5 produced by Pielke et al. compares predicted (IPCC scenarios) and observed changes in energy intensity the economy carbon intensity of energy.

 

Observations (2000-2005) imply both an increase in energy intensity of the economy and carbon intensity of energy by approximately 0.25 and 0.3 per cent pa respectively.
 

Figure 5.

Assumed decarbonisation in the 35 IPCC scenarios for 2000-2010

compared to observations between 2000 and 2005.196

 

 


World Energy Outlook 2006 Scenarios
As already discussed, The World Energy Outlook 2006 provides two scenarios: RS and AP.197 The RS (business-as-usual case) provides a baseline projection of energy usage, or carbon emissions, in the absence of further policy changes from mid-2006.

 

As such, by 2030, the RS projects global primary energy demand increases of 53 per cent, with over 70 per cent of this coming from developing nations, led by China and India.

Conversely, the AP scenario estimates the impact of implementing all currently proposed policy changes on energy use/carbon emissions, such as speeding up efficiency improvements or shifting to renewable energy sources. By 2030, global energy demand is reduced by 10 per cent, mainly due to the improved efficiency of energy use. Twenty-nine per cent of the decrease in emissions is expected to be achieved by electricity end-use efficiency and 36 per cent by fossil fuel end-use efficiency.

These scenarios are both based on average GDP growth of 3.4 per cent between 2004 and 2030 as well as average population growth of 1 per cent. We assume that beyond 2030 to 2050, GDP would grow at the rate of 2.9 per cent a year, the 2030 growth rate.

In analyzing the data, we separate two important components of the carbon intensity of an economy: efficiency in usage of energy (energy intensity of the economy - see Equation 1) and the carbon intensity of energy. The interaction of these elements is described in Equation 3.198
 


Equation 3
Where F is global CO2 emissions from human activity (in tCO2), E is total primary energy supply (in tonnes of oil equivalent) and G is world GDP (in ‘000s $US).

Using the two WEO scenarios of total consumption of energy, including electricity supply, we calculated energy intensity of the economy and carbon intensity of energy.

It is noteworthy that the emissions calculations made here only used primary energy supply projections (coal, oil, and natural gas). Other forms of energy usage, such as biomass, nuclear and hydroelectric power were assumed to have no carbon emissions.199 This assumption was made for ease of calculation, as sufficiently detailed projections for the type of biomass in use were not available to allow emissions projections.

 

Additionally, land-use changes from hydroelectric power projects were not included. Given this, we expect all our calculations to be conservative.

In extending the scenario to 2050, we have projected the energy supply growth for each fuel type using the annual growth rates estimated for 2015-2030, employing the same method to project total final consumption (TFC).

In describing the RS, the IEA states that the energy intensity of the economy will decline by 1.4 per cent on average until 2030.200 Again we have used this figure to project forward until 2050. No specific growth rate is suggested for the AP scenario, but it does imply a faster rate of improvements in energy intensity of the economy.

 

Based on historical precedents we assume an ambitious 2.0 per cent decline in energy intensity between 2004 and 2015, 2.2 per cent between 2015 and 2030 and 2.6 per cent between 2030 and 2050.
 


Model assumptions
We used a globally aggregated Earth system model - the Integrated Science Model (ISAM) global carbon model to predict the effect of emissions on atmospheric concentrations of CO2.

 

The ISAM model is available online and has been used widely in the IPCC assessment reports and climate policy analyses related to greenhouse gas emissions.201 The carbon-cycle component is representative of current carbon-cycle models.202 Model iterations were run with the IPCC B scenario for carbon emissions from land-use changes.203

 

Emissions of other greenhouse gases besides CO2 were also assumed to follow the IPCC B scenario.

Even though the model provides a projection of median temperature increases, these have not been reported due to the uncertainty in projecting temperature changes with increasing greenhouse gas concentrations.204 We have, therefore, confined ourselves to demonstrating the necessary improvements in carbon intensity to meet various CO2 emissions targets.

To test whether the projections correspond to a sustainable economy, we examine the potential for overshooting of CO2 emission targets, with a given level of energy intensity of the economy improvements, energy demand and GDP growth.

 

We have used the SIMCAP modeling platform developed by Malte Meinshausen to generate potential target emissions pathways.205 The model uses an Equal Quantile Walk (EQW) method to create more plausible scenarios for emissions paths out of the infinite combinations of yearly emissions that might achieve the targets.206

We have reported the results for target peaks of atmospheric CO2 concentrations of 350 ppm, 400 ppm, 450 ppm, 500 ppm and 550 ppm CO2. Note that we have confined our analysis in this section to actual CO2 emissions, ignoring the effect of other greenhouse gases. This was necessary because of the limits of the model in converting other emissions into CO2e emissions.

 

Thus, the actual warming effect is greater than that created by the CO2 emissions. Based on current proportions, CO2e (Kyoto gases only) would be around 50 ppm greater; for example, 385 ppm CO2 is around 435 ppm CO2e.

The EQW method was used to create the emission scenarios required to meet the target, with emissions reductions starting in 2007 for the OECD and 2010 for other regions of the world. Using this scenario and the previously defined rates of GDP growth, we have calculated what the necessary energy intensity and/or carbon intensity improvements would have to be to remain below the CO2 targets.

 

The EQW method was also used to create the post-2050 emissions pathways that would be necessary under the RS and AP scenarios to meet the targets.

Recent evidence and modeling has brought further clarity to the debate over feedback considerations. In the carbon-cycle, faster rates of emissions growth and accumulation of CO2 in the atmosphere will weaken the rate at which it can be absorbed into the oceans or terrestrial carbon sinks (see Box 11).

 

While we have excluded such feedbacks from the main analysis, we have provided estimates using these data separately.
 


Peak Oil
Although increasingly warning of production capacity constraints, the IEA makes no detailed mention of the possible physical limits to continuing exploitation of fossil fuels to drive the global economy.

 

That is, with the single exception in one media interview, when Fatih Birol, the IEA’s chief economist, said,

‘In terms of the global picture, assuming that OPEC will invest in a timely manner, global conventional oil can still continue, but we still expect that it will come around 2020 to a plateau.’207

In other words, a peak and long-term decline in the global production of oil. Evidence is presented later in this report on the likely onset of Peak Oil.

Projections for oil and gas production were obtained from Colin Campbell and the Association for the Study of Peak Oil (ASPO).208 Given the constraints in building and developing alternative sources of energy, such as nuclear or hydroelectric power stations, we have assumed that the energy requirements left unfilled because of the shortage of oil and gas will be filled by replacing those fuels with coal - a phenomenon that appears to be occurring already.209

 

This has significant effects on the carbon intensity of energy. While the rate of supply side efficiency improvements to the energy intensity of the economy are also dependent on the fuel mix, this substitution serves as a first order estimate of the effects of Peak Oil on anthropogenic greenhouse gas emissions.

As CCS is still an immature technology, yet to be proven at scale, we do not assume that it plays a role in reducing the carbon intensity of the economy.210 The future role of CCS is discussed in more detail later in this report.

We have also erred on the side of caution by not factoring in the declining net energy gains from fossil fuel extraction as more marginal stocks of oil, gas and coal are exploited. Increasing amounts of energy must be used to exploit heavy oils and tar sands which would have deleterious effects on the energy intensity of the economy.211

 

But without a very comprehensive and detailed global energy model, predicting such effects would be difficult. Additionally, using coal that is higher in moisture or otherwise less efficient for electricity production would have similar negative effects on the energy intensity of the economy.

 

We have not modeled this here for lack of data.
 


Results
As shown in Figure 6, the scenarios developed by the IEA would lead to extremely high concentrations of atmospheric CO2, with the RS breaching the upper limit of our most generous target range in 2047.

 

Even the optimistic AP scenario, would lead to atmospheric concentrations of CO2 of 487 ppm by 2050.

A possible emissions scenario that would seek to stabilize atmospheric CO2 concentration at 500 ppm after 2050 is shown in Figure 7. Given the pre-2050 emissions pathway of the alternative policy scenario, it is impossible to prevent an overshoot of the target. The changes in emissions levels needed to even bring about stabilization after an overshoot are quite dramatic.

 

As Figure 7 shows, if the alternative policy scenario is followed until 2050, immediately thereafter carbon emissions would still have to be curtailed by roughly 1.1 per cent annually to even stabilize atmospheric CO2 below 550 ppm.

 

This does not account for the impact of carbon-cycle feedbacks, however.

Figure 6.

IEA scenario emissions and resulting atmospheric CO2 concentrations


 


Figure 7.

Possible post-2050 emissions scenarios


 


Figure 8.

The impact of carbon cycle feedbacks on atmospheric concentration of CO2

 




Figure 9.

Effect of declining oil production on emissions
 

If we take into account the effects of carbon-cycle feedback mechanisms, the atmospheric concentrations of CO2 corresponding to a given level of emissions increases over time.

 

As climate models disagree about the magnitude of the feedback effect, we have demonstrated the range of possible CO2 concentrations in Figure 8. Data on the potential carbon-cycle feedbacks were take from the C4MIP Model Intercomparison.212 In the worst-case scenario, the atmospheric concentration of CO2 is about 10 per cent larger than previously modeled.

The situation becomes much worse when the Peak Oil projections are combined with the possible efficiency improvements described in the IEA scenarios (see Figure 9).

 

In the AP scenario, resulting emissions from the projected change in the fuel mix would be nearly 17 per cent higher than the IEA projections. This would bring projected atmospheric CO2 concentration to 501 ppm in 2050 (note, concentrations are not shown on the graph).

 

Peak Oil, therefore implies that proceeding with every proposed improvement to energy intensity and adoption of cleaner fuels will not be sufficient to prevent a breach of even the most generous target and thus potentially disastrous climate change.
 


Emissions scenarios with target CO2 concentrations
The second phase of our analysis compared possible emissions scenarios with target pathways that would generate specified levels of atmospheric CO2 concentrations.

 

Using the EQW method, emissions scenarios were created to match the targets of 350 ppm, 400 ppm, 450 ppm, 500 ppm and 550 ppm.

 

Figure 10 shows the emissions pathways as compared to the IEA pathways.


Figure 10.

Possible emissions scenarios to meet various atmospheric CO2 concentration targets.

 



Figure 11.

The growing gap in the carbon intensity of energy.

 

We then examined the gap that would have to be plugged by changes in carbon intensity of energy to meet the targets.

 

Maintaining the assumptions in the alternative policy scenario about the improvements to the energy intensity of the global economy and using the stylized emissions pathway that would meet the target atmospheric concentrations of CO2, we can find a typical pathway of improvement in the fuel mix that would enable growth at the rate specified in the IEA scenarios.

 

As shown in Figure 11, the aggressive advancement of renewable energy in the AP scenario does not meet the needs of an emissions pathway that could mitigate climate change.

Figure 12.

Improvements needed to meet the target emissions pathway at different levels of growth.

 

The projection for a decline in oil production and substitution by dirtier coal energy sources counterbalances the other improvements in the fuel mix.

 

Despite the scenario assuming about 25 per cent greater use of nuclear power and non-hydroelectric renewable energy sources than in the RS, which already includes almost 10 per cent per annum average increases in renewables, the effects of declining oil production mean that the carbon intensity of energy remains about the same over time.

 

This demonstrates that without radical changes in lifestyle in terms of energy usage or even faster moves towards non-fossil-fuel energy sources, it will not be possible to have economic growth at the rate indicated.

Looking at the overall carbon intensity of the economy, meaning that we allow variable improvements in both the carbon intensity of energy and energy intensity of the economy, Figure 12 shows that kind of improvement that would be needed to meet the target emissions pathway at different levels of growth.

Even at 1.5 per cent growth, the global economy would need to reduce its carbon intensity by 71 per cent between 2006 and 2050, equivalent to a 1.3 per cent average annual decline. But, this assumes a steady improvement, since following a different trajectory - for example, with delayed measures to improve the carbon intensity - would cause cumulative emissions to increase, and an overshoot of the target.

Any delay in improvements would have to be paid for with even greater improvements in the future to ensure that atmospheric carbon concentrations do not peak above the maximum of the target range, namely 500 ppm CO2.

 

The question remains as to whether such improvements could be made.

 

 

Box 16. Historical precedents for rapid changes in carbon intensity
 

An absolute annual reduction in CO2 emissions greater than 3 per cent is rarely considered to be a viable option.213 Worse still, where mitigation policies are more developed, emissions from international aviation and shipping are not included. For example, Anderson et al. note that the UK’s CO2 emissions are, on average, 10 per cent greater than official records for this reason.214

In the Stern Review, historical precedents of reductions in carbon emissions were examined. Their analysis found that annual reductions of greater than 1 per cent have ‘been associated only with economic recession or upheaval’.215 Stern points to the collapse of the former Soviet Union’s economy, which brought about annual emission reductions of over 5 per cent for a decade. While France’s 40-fold increase in nuclear capacity in just 25 years and the UK’s ‘dash for gas’ in the 1990s both corresponded, respectively, with annual CO2 and greenhouse gas emission reductions of only 1 per cent.

In 1990, the Dutch government proposed to increase the rate of energy efficiency from 1 per cent per year to 2 per cent per year. The pledge was considered a ‘real test of strength’, by the Ministry of Economic Affairs. This was against the backdrop for what was actually achieved generally during the last century of 1.2 per cent per year. However, the target up to 2010 was later revised to 1.3-1.4 per cent per year.216



Would the global economy manage to lower its carbon intensity by 2.7 per cent per year on average (the necessary improvement to meet the 500 ppm target with 3 per cent growth levels) while maintaining growth?

 

Historical indicators are not positive, with the average annual declines of carbon intensity between 1965 and 2002 being just 1.2 per cent. Worse still, between 2000 and 2007, improvements in energy intensity of the economy slowed to an average of just 0.4% each year. Over the same period, carbon intensity of the economy increased by on average 0.37% per year.

Figure 13 again highlights the gaps between the AP scenario and the targets, much as Figure 12. The scenarios are clearly way above even the riskiest target level.

Figure 13.

The gap between scenarios and targets.
 

 


Figure 14.

Target carbon intensities with no economic growth.

 

Even if growth were to cease, implying a decline in global per capita incomes because of population growth, we could not be complacent on the technology and energy front, as shown in Figure 14.

 

Maintaining a low risk profile and keeping ambient CO2 concentrations below 400 ppm would require similar levels of investment in energy efficiency and emissions reductions as described in the AP scenario, all without any increase in overall economic activity.

As a final analysis, we looked at the effect of carbon-cycle feedbacks on the need for carbon intensity improvements and emissions reductions. To meet the same 450 ppm target for atmospheric CO2 concentrations in a coupled carbon-cycle model, the actual emissions pathway must correspond to a concentration of between 410 ppm and 445 ppm in an uncoupled carbon-cycle model.

 

The results are shown in Figure 15, and demonstrate that the effect of carbon-cycle feedbacks can be significant.

Figure 15.

Potential effects of carbon cycle feedbacks.

 

The following sections explore some of the factors that may modify these scenarios. They seek to indicate the relative likelihoods of the range of different possible outcomes - better or worse - are more probable.

Since our main work was completed, Professor Kevin Anderson of the Tyndall Centre for Climate Change Research at Manchester University also looked at a range of scenarios for growth, greenhouse gas concentration levels and global warming.217

Assuming that growth continued, he looked at the rate of emissions reductions that would be needed to achieve greenhouse gas concentration levels commensurate with a 2, 3 or 4°C temperature rise. Most, of course, agree that temperature rise above two degrees represents unacceptable, dangerous warming.

 

Anderson’s conclusion was stark:

‘Economic growth in the OECD cannot be reconciled with a 2, 3 or even 4°C characterization of dangerous climate change.’218

Back to Contents
 

 

 


Peak Oil, Gas and Coal?

If we could spend the oil age in an Irish pub…the glass was more or less full in 1900, just about half full in 2000 and there are a few little dregs left at the end of this century.
Dr Colin Campbell (1 February, 2007)

My father rode a camel. I drive a car. My son flies a jet plane. His son will ride a camel.219
Saudi saying

Supplying the world with all the crude oil and natural gas it wants is about to become much harder, if not impossible.

 

For oil, the horizon of the global peak and decline of production appears close and that for gas not much further behind. When demand exceeds production rates, the rivalry for what remains is likely to result in dramatic economic and geopolitical events that could make the financial chaos of 2008 in Europe and the USA seem light-hearted.

 

Ultimately, it may become impossible for even a single major nation to sustain an industrial model as we have known it during the twentieth century.220

Counter-intuitively, the imminent global onset globally of the peak, plateau and decline of the key fossil fuels, oil and gas, will not help arrest climate change. If anything, it could be a catalyst for worse emissions and accelerating warming. For example, in October 2009, the UK Energy Research Centre (UKERC) reviewed the current state of knowledge on oil depletion.221

 

The study argued as we advance through peak oil:

…there will be strong incentives to exploit high carbon non-conventional fuels. Converting one third of the world’s proved coal reserves into liquid fuels would result in emissions of more than 800 million tonnes of CO2, with less than half of these emissions being potentially avoidable through carbon capture and storage.

In other words, with the analyses by Meinshausen and Allen discussed earlier in this report in mind, without extensive investment in low carbon alternatives to conventional oil, and policies that encourage demand reduction, Peak Oil is likely to drive emissions further towards a threshold of dangerous climate change.
 

 

 

Box 17. Peak Oil and food production
 

Increased fossil energy prices will in turn cause the price of food to increase significantly. On average, 2.2 kilocalories of fossil fuel energy are needed to extract 1 kilocalorie of plant-based food.222 In the case of meat, the average amount of kcal fossil energy used per kcal of meat is much greater, with an input/output ratio of 25.223

In early 2008, the UN World Food Program had to reassess its agreed budget for the year after identifying a $500 million shortfall. It found that the $2.1 billion originally allocated to food aid for 73 million people in 78 countries would prove to be inadequate because of the rising costs of food.

 

Higher oil and gas prices have contributed to this by increasing the costs of using farm vehicles and machinery, transporting food and manufacturing fossil-fuel-dependent input such as fertilizer. The move to grow biofuel crops has also exerted upward pressure on food prices by leaving less productive land available to grow crops.

 


The global economy is still heavily dependent on fossil fuels. Oil remains the world’s most important fuel largely because of its role in transport and agriculture and the ease with which it can be moved around.

The historical pattern has been for industrial societies to move from low-quality fuels (coal contains around 14-32.5MJ per kg) to higher quality fuels (41.9 MJ/kg for oil and 53.6 MJ per kg), and from a solid fuel easily transported and therefore well suited to a system of global trade in energy resources.224

Now, almost all aspects of our economy are dependent on a constant and growing supply of cheap oil, from transport to farming, to manufacturing and trade. In the majority world, where too many people live close to, or below the breadline, the long tail of green revolution agriculture depends on pesticides and fertilizers that need large amounts of fossil fuels.

 

The implication of any interruption to that supply, either in terms of price or simple availability, means a significant shock to the global economy.

 

Everyone will be affected, but some more than others.

 

 

 

Box 18. We’ve been here before
 

The world oil crises in the 1970s provide some idea of how the effects of Peak Oil may ricochet through the economy. The two world oil crises in the 1970s (the most significant occasions when demand exceeded supply due to politically caused interruptions) caused widespread panic that the economy would fall into a global depression. During the first oil embargo in 1973, oil supplies only fell by 9 per cent. The second oil crisis caused by the Iranian oil cut-off resulted in a fall in oil production by 4 per cent.225 Both world oil crises were followed by recession, resulting in economic hardship, unemployment and social unrest around the world.226

Interestingly, the first and second oil crises are the only recorded times in the industrial epoch where energy efficiency improvements have actually resulted in a decrease in demand for energy. This shows how a strong price signal, aggressive government policy and awareness can work together to decrease energy demand.


 


The price burden of crude oil
Recent research explored the price burden of crude oil on French households in 2006.227

 

This is the first analysis of this type. Other analyses have only focused on direct domestic energy consumption (electricity and gas).228,229,230 This study, however, explores the contribution of indirect or ‘embodied’ energy within goods and services. The results and can be taken to be broadly consistent with other developed nations.

The analysis found that in 2006, the average burden of crude oil was equivalent to 4.4 per cent of the total budget of a typical French household. This figure, however, varied significantly depending on income, age or the size of their city of residence. The results are presented in Figure 16. This provides some indication of the vulnerability to oil price rises.

 

In general, Figure 16 shows the largest burden is likely to be experienced by the elderly and low-income groups. This illustrates that changes in oil prices are an acute social justice issue.

Figure 16.

Dependence of the contribution of crude oil to household’s budget as a function of

per capita income, age of the household’s reference person, and the type of residential area.231
 

In an international context, different government responses to oil price rises can also radically alter the consequences for developing countries.

 

Following the 1973 oil price shock, relaxed monetary policy in rich countries caused low to negative real interest rates on hard currencies. As well as maintaining demand for poor countries’ exports this also laid the foundations for the Latin American debt crisis. But following the 1979 oil price shock, rich countries’ fear of inflation created a triple blow for their poorer relations.

 

Economist David Woodward describes the consequences of tightening monetary policy,

‘demand contracted, developing countries’ export prices collapsed and real interest rates increased dramatically to historically high levels’.232

Consequently, the price of oil imports doubled ‘overnight’ and interest rates on commercial foreign debts doubled over the next three years. Even at oil prices prevailing in early 2004, the IEA believed that oil-importing developing countries were being seriously disadvantaged.233

 

As the International Monetary Fund (IMF) observes, although the so-called Heavily Indebted Poor Countries (HIPCs) ‘account for only a small share of global GDP, many of them are among the most seriously affected by higher oil prices’.234

The IMF points out that 30 of the 40 HIPCs are net oil importers, making them particularly sensitive to price fluctuations. Their problems are compounded by several interconnected economic factors including: low per capita incomes, high level of oil imports relative to GDP, large current account deficits, high external debt, and limited access to global capital markets.

 

Altogether, according to the IMF, this means that,

‘the impact of higher oil prices on output is relatively large, as it will have to be met primarily through a reduction in domestic demand’.235

This is economists’ speak for the poor getting poorer.
 


Timing of Peak Oil

We may argue about when the peak is, but it doesn’t change the argument that it is coming.236
Robert Kaufmann, Energy Economist at Boston University

The actual global peak year will only be known when it has passed, but most estimates suggest that we are either at, or very close to this point.

 

At most it is one or, less likely, two decades away. Against a background of rising demand, ‘peaking’ will result in a major shock to the global economy. But, even before then, an opening gap between production and demand is already driving prices up.

The recent review published by UKERC warned that, almost unequivocally, peak production will occur before 2030, with a significant risk that this will occur before 2020.237 Estimates of the precise onset of Peak Oil range from 2006 to 2030 (Table 6). The higher-end estimates are by and large due to exaggeration of technical reserves. A constant flow of new studies and industry leaks, however, point towards a downward revision of potential reserves.

Actual technical reserves of oil are often very different from published reserves, the former rarely changing and the latter being related to political circumstance (often overestimated because of poor data, to bolster financial investment, political and institutional self interest, and other complicating factors).

 

But, despite the variety of different estimates, many credible analysts have recently become much more pessimistic about the possibility of finding the huge new reserves needed to meet growing world demand, and even the most optimistic forecasts suggest that world oil peaking will occur in less than 25 years.

A central problem in the estimation of ‘real’ oil reserves is that not all oil companies work to the same standards of reporting. Whilst the US Securities and Exchange Commission sets rules for how to report reservoir estimates, only US and major international companies generally abide by those standards, reporting is not always performed reliably..238,239

 

Jeremy Leggett, an expert on Peak Oil, reports in his book Half Gone that reporting by Organization of Petroleum Exporting Countries (OPEC) is usually particularly dubious:

‘Middle East official reserves jumped 43 per cent in just three years [during the 1980s] despite no new major finds.’240

Additionally, Saudi Arabia has posted a constant level of reserves (260 billion barrels) over the past 15 years, despite the fact that it has produced over 100 billion barrels in the same period.241
 

Table 6.

Projected dates of reaching ‘Peak Oil’.242

 

The North Sea is the only place where a significant new discovery has been made outside of OPEC nations, Russia and Alaska in the past four decades.

 

Both Norway and the UK are seeing decreases in the production from the region to the extent that the UK no longer exports oil. Furthermore, no new giant oilfields are replacing those which have already passed their peaks.

Of all the oil resources remaining:243

  • 62 per cent is in the Persian Gulf

  • 10 per cent is in Africa, mostly Angola, Libya, and Nigeria

  • 10 per cent is in the former Soviet Union (FSU), mostly Russia, Kazakhstan, and Azerbaijan

  • 10 per cent in Latin America, mostly Venezuela

A failure to grasp the problems associated with Peak Oil was, until recently, a serious blind spot in many official government policies and reviews.

 

For example, ASPO commented on the 2006 Stern Review: ‘It fails to take note that oil and gas, which drive the modern economy, are close to peak, and will decline over most of this century to near exhaustion. The coal resources are indeed large, but the coal-burning airliner has yet to take off.’244

Whilst there is considerable uncertainty surrounding future oil reserves, and the field is surrounded by intense debate, the current view appears to be converging towards the view that Peak Oil is a very real and impending problem that could have catastrophic implications for the global economy, to the extent that it is gradually filtering into the A-list of political concerns with then Secretary of State for Environment, Food and Rural Affairs, David Miliband addressing an audience at the University of Cambridge in March 2007 stating:

‘The time is right to look at what it would mean for the UK over the period of 15 to 20 years to create a post-oil economy - a declaration less of ‘oil independence’ and more the end of oil dependence.’245

More recently, the IEA has begun to identify the problems of Peak Oil.

 

The Medium Term Oil Market Report published by the IEA (an official advisor to most of the major economic powers) reported in 2008 that:

‘there will be a narrowing of spare capacity to minimal levels by 2013’.

Since the previous year alone it had made, ‘significant downward revisions’ on ‘both non-OPEC supplies and OPEC capacity forecasts’.246

 

The fuel price volatility of the last two years looks to be a foretaste of a far more massive crunch that will follow as the graph lines for global oil demand and supply head in opposite directions.247 The IEA’s motto - ‘energy security, growth and sustainability’ - appears to be the antithesis of the situation that it surveys.

Since UK North Sea production peaked around 1999, hopeful eyes have been focused on the major producers like Saudi Arabia to keep the economy’s arteries full of oil.248 But, looking ahead, Saudi Arabia appears to have other ideas.

 

Over the next 12 years it intends to spend around $600 billion (about the same staggering figure that the USA earmarked for propping up its financial system) on a massive domestic infrastructure program, including power stations, industrial cities, aluminium smelters and chemical plants. And, while doubts persist that its reserves are a lot less than publicly stated, guess what: all these new developments will be powered with Saudi oil.

 

The rest of the world should not hold its breath waiting to be rescued.249

Figure 17.

The rise in the price of Light Crude (NYMEX) between January 2004 and December 2009 (current $US per barrel).

 

Already the cost of a barrel of oil has risen almost 14-fold in the last decade reaching $147 a barrel in July 2008 (Figure 17).

 

While the price dropped in late 2008 to $40 a barrel, they have doubled again since. Oil prices are becoming increasingly volatile due to declining indigenous production and growing reliance on international markets.

 

It is noteworthy that that several analysts forecast oil prices could rise to $200 to $300 a barrel in the near future.250,251
 


The energy return on investment

The first half of the total oil resource is easy to extract, the second half is hard. We will transition from oil fields that are shallow, big, onshore, safe, and close, to fields that are deep, dispersed, offshore, remote, and unsafe.252
Professor Michael Klare, author of Blood and Oil (2004)

[It] takes vast quantities of scarce and valuable potable water and natural gas to turn unusable oil into heavy low-quality oil…In a sense, this exercise is like turning gold into lead.253
Matthew Simmons, leading expert on Peak Oil

The Energy Return on Investment (EROI) is the ratio between the useful energy obtained from a source divided by all the direct and indirect energy inputs needed to obtain it.

 

For example, a fuel with an EROI of 10:1 means 10 joules would need to be invested to yield 100 Joules of useful energy, resulting in a net energy (or energy surplus) of 90. If, however, the EROI was 3:1 (in the case of unconventional oil), to net the same 90 Joules around 45 Joules would need to be invested.

As ecological systems with a large energy surplus have a competitive advantage, so does the economy. Indeed, the huge growth in the global economy can be attributed to the switch from low EROI wood (30:1) to coal (80:1) and finally to oil (100:1). Our economy thrives on high EROI energy sources.

Not only is the discovery rate of oil falling, oil production is experiencing diminishing returns. This is clearly illustrated by the evolution of EROI for oil in the US over time.254

  • 1930s, EROI = 100:1

  • 1970s, EROI = 25:1

  • 1990s, EROI = 11-18:1

Another study found that the global average EROI for oil in the first half of the 2000s, was approximately 20:1.255

 

And, if current trends continue the ratio will change to 1:1 in the next 20 to 30 years. In other words, at this point, oil will cease to be a net energy source of energy.

With declining conventional oil reserves, it will be necessary to increasingly rely on unconventional oil reserves, such as Canadian tar sands and Venezuela’s Orinoco tar belt. Whilst many estimates of the unconventional oil resource indicate that it may well substantially exceed those of conventional oil, increasing amounts of energy will be required to extract that resource.256

 

Unconventional oil is estimated to have an EROI of around 3:1 - bearing in mind that once EROI approaches a ratio of 2:1, the oil might as well be left in the ground, given the additional energy required to refine it into a useful fuel.257

The techno-optimistic belief holds that when Peak Oil arrives, we will be able to deal with it.

 

This outlook is generally not held by the majority of Peak Oil experts, many of whom hold the view that no combination of existing and emerging technologies will provide industrial nations with the energy necessary to sustain current consumption rates and exorbitant lifestyles.258

Figure 18.

Global supply of liquid hydrocarbons from all fossil fuel resources and associated costs

in dollars (top) and GHG emissions (bottom).

EOR = enhanced oil recovery.259

 

In the past, higher prices led to increased estimates of conventional oil reserves worldwide, since oil reserves are dependent on price.

 

In other words, reserves are defined as the amount of oil that is considered economically feasible to recover. Geology, however, places an upper limit on what is actually recoverable from conventional oil. Effectively, there is an upper limit to the price of oil - beyond this point additional conventional oil will not be recoverable at any realistic price.

The high price of oil over the past decade has provided an incentive for oil companies to conduct extensive exploration over that period. The results, however, have been disappointing.
 


Alternatives
What are the potential alternatives to oil if the Peak Oil experts are wrong about a technofix, such as liquid and gas synthetic fuels (synfuels) produced from coal, or the widespread use of biofuels?

Coal has an EROI ratio of around 80:1. Therefore, coal could be transformed into synthetic oil through the Fischer-Tropsch process.260 However, synthetic transport fuels emit even more carbon on a well-to-wheels basis than conventional crude; and when the feedstock is coal, the emissions are double.261

 

Even if the process producing synfuels included CCS, CO2 emissions would still be greater than those associated with conventional diesel and petrol. According to one study even if 85 per cent of the carbon emitted from the processing of coal were captured (bearing in mind this is the upper limit of what most CCS experts believe is possible), emissions from end-use of these synthetic fuels would produce on average 19-25 per cent as much CO2 as petroleum derived fuels.262

Much of the literature focuses on the availability of oil as a result of Peak Oil.

 

But some analysts have raised concerns about the transition from conventionally produced oil, highlighting that synthetic liquid fuels are generally higher capital, higher energy intensive and have higher carbon to hydrogen ratios, and therefore produce more CO2 than conventional crude oil.263 Figure 18 shows that the oil transition is not necessarily a shift from abundance to scarcity, but a transition from high quality resources to lower quality resources that have potentially higher levels of environmental damage.

Investment into synthetic fuels will tend to cause world oil prices to fall, benefiting consumers, with potentially the impact of increasing demand even more.

 

Therefore, the management of the oil transition may not be necessarily focused on dealing with global economic collapse, but rather dealing with the environmental problems associated with synthetic liquid fuels derived from other fossil fuels, such as coal and tar sands.
 


Peak Gas

‘Peak gas is an entirely unheard of and unwelcome spectre’ 264
Andrew McKillop, energy analyst

Less discussed, but equally real is the prospect for the global peak and decline in the production of natural gas. Peak Gas is analogous to Peak Oil, but refers to the maximum rate of the production of natural gas.

 

For example, in the context of the UK the Digest of UK Energy Statistics reports:

‘The UK oil and natural gas production peaked in 1999 and 2000. Since then they have declined at an average rate of 7 per cent per annum (pa) and 3 per cent pa respectively (to 2004).’ 265

In 2007, Defra reported that emissions from industry in the UK increased during 2006 as power stations had to switch from gas to coal due to high gas prices.266

 

This implies rising gas prices connected to geopolitics or decline in production could also result in an increase in carbon emissions. Additionally, because a significant proportion of domestic dwellings are dependent on gas for space heating, declining gas supply and subsequent price increases could have a significant impact on fuel poverty.

UK gas fields have already peaked, and it’s expected that most of the UK’s gas will eventually come from Russia, Iran and Qatar. Figure 19 shows the changes in the UK’s indigenous production and consumption of natural gas between 1998 and 2008.267 Since 1998, demand (white) of natural gas shows an inter-annual variability of approximately 5 per cent.

 

At the same time, indigenous gas production showed a slow decline from 2000 (light grey). In 2004, in order to meet demand for the first time since 1997, the UK began importing gas. This reduced the UK’s energy independence significantly.

The ‘energy dependence’ factor is the ratio of net energy imports to demand, and multiplied by 100 to produce a scalable figure. When it becomes ‘positive’, it means that we are obliged to import energy to meet our demand. In other words, our independence declines. Between 2004, when the UK first lost its energy independence, and 2008, the energy dependence factor has risen almost 5-fold.268

More recently Shell’s vice president, John Mills, told delegates at the Abu Dhabi International Petroleum Exhibition and Conference (ADIPEC) on 5 November 2008 that:

‘Globally, what people have woken up to is that there is a prospect for the gas industry that its supply-demand crunch could come earlier than anticipated.’269

Figure 19.

Natural gas production, net exports/ imports and consumption 1998-2008.270

Consumption plus net exports will differ from production plus net imports

because of stock changes, losses and the statistical difference item.

 

Many energy policies have no concept of Peak Gas being imminent.

 

This is largely due to poor reporting of gas reserves. Whilst estimates of gas reserves succumb to the same problems and lack of accurate disclosure as the oil industry, unlike oil, the gas market is regional.271

 

For example, oil can be transported from the other side of the world for consumption in the UK, but the UK gas market is generally restricted to Europe and Russia. In short, gas is very difficult and expensive to move around, and infrastructure is necessary before a gas reserve can have a market (i.e., storage and pipelines).

If we consider an energy market under Peak Oil/Gas conditions, we would expect the UK to be able to afford to outbid poorer countries in the global oil market. In the Euro-Russian gas market, however outbidding all other equally wealthy European countries would be extremely costly resulting in large increases in gas prices.

 

This suggests that for developed nations like the UK, Peak Gas may pose a greater threat to the economy than Peak Oil, and naturally both will present significant problems to developing nations following a similar carbon intensive development pathway.

 

 

 

Box 19. The feasibility of saving 1 Gt of Carbon by switching to gas
 

A program to displace 1400GW of coal-fired power stations with, for example, 1400 1GW 70 per cent fuel efficient CCGT plants, would require an additional 0.7 per cent annual increase in natural gas production on top of the business-as-usual annual increases in demand of 2.3 per cent projected by the IEA.272

Whilst an additional increase of global gas demand by 0.7 per cent may not seem huge, in the context of known gas reserves and current production from fields, this rate of increase is unlikely to be sustainable for long. For example, gas fields in large developed economies are declining, while regional natural gas constraints are already being observed, primarily in North America (the most intensive consumer of the resource), as well as Russia and Europe.273 ,274


 

Overall, any carbon emissions savings made through fuel switching from coal or oil to gas will be undermined by the onset of Peak Gas. Equally, our assumptions about how gas will be able to carry us through to a low carbon economy are seriously flawed.

For example, in 2006, carbon emissions from British industry covered by the EU ETS (Emissions Trading System) rose by 3.5 per cent during 2006.275

 

These rising emissions were due to power generators switching from gas to coal in response to high gas prices during 2006. The rise in emissions from these power stations cancelled out all improvements across those sectors that actually reduced their emissions.

Natural gas is also important for many plastics, fabrics, even plastic bags. It provides the heat necessary for cement production, and is also indispensable for making synthetic oils from tar sands (see previous section on Peak Oil).276

 

Additionally, natural gas is ‘absolutely indispensable’ for the production of industrial fertiliser.277
 


Unconventional gas
Unconventional gas is defined by the International Gas Union as: ‘methane from tight (very low permeability) formations, methane from coal seams, methane from geo-pressured brine, methane from biomass (onshore and offshore), and methane from hydrates’.278

 

But, the fundamental problem with unconventional gas is that its recovery is more energy intensive and expensive compared to oil, and the production process can be much slower.

 

While technology may help to overcome some of these problems, a very real problem will be transportation, and the significant reduction of the EROI.
 


Peak Coal?
A scenario seldom discussed is the peaking of coal production. Global consumption of coal is growing rapidly.

 

From 2000 to 2007, world coal extraction grew by a rate of 4.5 per cent compared to 1.06 per cent for oil (oil production actually fell by 0.2 per cent between 2006 and 2007).279 This is opposite to the trend observed over the past two decades. In particular, as China rapidly industrializes, the use of coal is increasing dramatically. In 2005, China was responsible for 36.1 per cent of world coal consumption, the USA 9.6 per cent, and India 7.3 per cent.280

Global coal production is expected to peak around 2025 at 30 per cent above present production in the best-case scenarios. Geographically, coal reserves are concentrated in just a handful of nations. Approximately 85 per cent of global coal reserves are concentrated in six countries (in descending order of reserves): USA, Russia, India, China, Australia, and South Africa.

 

Furthermore, coal consumption generally takes place in the country of extraction - around 85 per cent of coal is used domestically, with around 15 per cent exported.281

 

Again, the concentration of coal in a small number of nations increases energy insecurity.
 


Coal’s contribution to the economy
Currently, coal provides over 25 per cent of the world’s primary energy and generates around 40 per cent of electricity.

 

For a number of reasons - including the cost of mining, transport and the lower energy density of coal, and the more inefficient process of electricity generation - its primary energy yield is only around one-third of the economic productivity of the primary energy in oil.282

While coal may be able to provide some buffer to Peak Oil and Gas, it is one of the most environmentally damaging fossil fuels. For example, while it produces a quarter of the world’s energy, it is responsible for almost 40 per cent of the greenhouse gases. Since 1750, the burning of coal has released around 150 gigatonnes of carbon into the atmosphere.283

Although carbon sequestration could in theory reduce the carbon burden of coal, coal is problematic for other reasons. For example, sulphur, mercury and radioactive elements are released into the air when coals is burned. These are particularly difficult to capture at source.

 

The mining of coal also destroys landscapes, and very fine coals dust originating in China and containing arsenic and other toxic elements has been detected drifting around the globe in increasing amounts.284
 


Clean coal?
Clean coal technology refers to some form of CCS but, there is something rather peculiar about the phrase ‘clean coal’. Despite the environmental burden from the mining of coal, stick the word clean in front of it, and suddenly it becomes palatable.

In his keynote speech at the Labour Party conference in 2008, the Prime Minister, Gordon Brown, called for a new generation of ‘clean coal’ plants.

 

Speaking almost simultaneously in the USA, former Vice President and Nobel Prize winner Al Gore stated explicitly:

‘Clean coal does not exist.’285

More recently, the Subcommittee on Investigations and Oversight of the US Committee on Science and Technology, which is responsible for overseeing all non-defense research and development programs at a number of federal agencies published a report examining the recent abandonment of FutureGen by the Department of Energy.286


FutureGen was a 10-year long $1 billion government/private partnership program to build a 275MW CCS power plant in Mattoon, Illinois.

 

The report argued:

‘Creating ‘clean coal’ is an extremely complex task involving not only the development of reliable and economical technology to capture CO2 and other pollutants, and integrating it into electricity-producing coal plants, but also the acceptance of higher electricity prices and unknown liability for carbon dioxide sequestration sites by the public and their elected officials worldwide.’

In other words, clean coal is further away than we are being led to believe.

We discuss the potential of ‘clean coal’ in the context of carbon capture and storage in the following section.
 

Back to Contents

 

 

 


Carbon capture and storage - the nuclear fusion of the 2010s?

‘… carbon sequestration is irresponsibly portrayed as an imminently useful large-scale option for solving the challenge. But to sequester just 25 per cent of CO2 emitted in 2005 by large stationary sources of the gas […] we would have to create system whose annual throughout (by volume) would be slightly more than twice that of the world’s crude-oil industry, an undertaking that would take many decades to accomplish.’ 287
Professor Vaclav Smil (2008)

By 2015, the European Union aims to have 12 large CCS demonstration projects in place, requiring an investment of e5 billion.

 

The expectation is that this development will cause significant cost reductions, making the technology affordable by 2020. There are, however, many drawbacks; for example, it will keep costing large sums of money to make sure the CO2 stays where it is supposed to, and the process is energy intensive.

CCS - capturing CO2 and storing it indefinitely - is one of the key technologies expected to contribute to the stabilization of atmospheric concentrations of CO2. The IPCC has now endorsed its use, Nicholas Stern concludes that it will be a crucial technology in the 2006 Stern Review, and the UK Climate Policy Program places significant emphasis on this as a plausible technological response.

Despite this optimism, many still highlight that it is still by no means clear that it will work or that it will become commercially viable in time to have a significant impact on the mitigation of climate change.288

 

For example, a recent editorial in the journal Nature Geoscience argued:

‘Capacities for geological storage are uncertain, pilot projects for deep ocean sequestration have been halted, and public acceptance of both options is at best questionable - not least because full risk assessments based on solid scientific data are scarce’.289


A short overview of CCS
CCS can involve a number of different processes which are heavily reliant on advanced and unproven engineering.

 

There are three types of CCS processes currently under consideration. And all three processes are already being applied in several industries on smaller scales, but most without storage.

  • Post-combustion - the mixture of CO2 and flue gases after combustion is separated by using a liquid solvent.

  • Pre-combustion - the fuel is processed prior to combustion resulting in a mixture of mainly CO2 and hydrogen. Both gas streams are subsequently separated, so that the hydrogen can be combusted for electricity production and the CO2 for storage.

  • Oxyfuel combustion - using pure oxygen instead of air when combusting resulting in flue gas that contains mainly water vapor and CO2. Both streams can easily be separated and treated further if necessary.

According to MIT’s Carbon Dioxide Capture and Storage Project Database, there are approximately 40 carbon storage demonstration projects in various scales running at present.290

 

But CCS is still an experimental technology, or rather a collection of technologies which has yet to be proven at scale. Such optimism in a technology is worrying, particularly as yet, not a single coal plant has been built anywhere in the world that uses complete capture and storage.

The first US pilot plant that can capture CO2 from coal burning, FutureGen, was due online in 2012.

 

FutureGen began in 2003 by testing safety, permanence and the economic feasibility of storing large volumes of CO2 in geological structures at 22 test sites. A decision made by the Bush Administration, however, appears to have stalled the progress of the project.291

Disposal of CO2 under sea-beds is still at the research phase according to the IPCC, who also states that pre- and post-combustion capture of the gas has passed research and demonstration stages and is now ‘economically feasible under specific conditions’.292
 


The cost of CCS
IPCC estimates that installing CCS at a coal-fired power plant could raise the cost of generating power from 4-5 ¢/ kWh to between 6-10 ¢/ kWh.

 

So, CCS could effectively double the cost of electricity from coal at worst and increase the cost by a third at best. If the captured gas is used for enhanced oil recovery (EOR), revenue could decrease to between 5-8 ¢/ kWh. In the case of EOR, however, whilst CO2 is being stored deep underground, more fossil fuels are being burned at its expense.

Natural gas can also be used with CCS technology. Gas can be transformed into hydrogen by reacting high temperature steam with natural gas in a process called ‘steam methane reformation’. When burned, hydrogen is considered to be a clean fuel and can also be used in fuel cells. The carbon within the natural gas is captured and pumped underground.

How quickly is CCS likely to become commercially viable?

Proponents of CCS claim that ‘all technology is proven at the desired scale; we are only demonstrating the ability to integrate technology’.293 While a number of CCS projects are underway, and have been for some time, there is a plethora of serious concerns about this technology.294

 

It has been claimed that all the necessary steps required for underground storage of CO2 have been commercially proven, yet at a recent hearing of the Senate Energy and Natural Resources Committee in 2007, the Director of the US Geological Survey laid out a timeline of commercialization of workable CCS schemes after 2012.295 He argued that the first commercial deployment would be around 2020, with widespread CCS by 2045.

So what does this mean in terms of emission reductions?

 

One estimate by the Natural Resources Defense Council’s Climate Centre suggests that if the total number of new coal plants that analysts think will be built around the world over the next 25 years were built without CCS, these new plants will emit around 30 per cent more CO2 than all previous human uses of coal.296

 

But, if the first pilot plant for coal CCS is not going to be online until 2012 - this means the recent trend of increasing carbon intensity of the economy is very likely to continue well into the new decade.

Is CCS the magic bullet?

If artificial carbon storage in the twenty-first century becomes the main route of carbon emission reductions, the total carbon storage by the end of the century could exceed 600GtC.297

 

Since this may be an unrealistic level of artificial carbon sequestration, in Box 20, we examine the potential implications of capturing 1-3 GtC per year.

 

 

 

Box 20: Achieving emissions reductions through CCS
 

Assuming a rate of increase in CCS of 70 Sleipner-sized* geological storage formations per year over the next 50 years, providing a total artificial sink capacity of 1 GtC per year would result in the cumulative storage of 27GtC of carbon dioxide by 2050.

 

If this annual carbon capture rate was kept constant over following 50 years until 2100, the cumulative carbon dioxide stored would reach approximately 80 GtC. If this was increased to 3GtC naturally the cumulative carbon stored would be three times this amount (240 GtC).

 

* Sleipner is the first operational carbon reserve. It is located in the North Sea amd captures around 0.3 MtC every year.

By capturing this volume of carbon, it is reasonable to assume some leakage would be unavoidable. It would be impossible to detect, monitor and control all potential escape routes of CO2 for hundreds if not thousands of years - therefore, geological storage cannot be viewed as truly permanent.298

If we consider, on average a 1 per cent global leakage rate from the cumulative reserves, the amount of carbon dioxide leaked from the storage of just 1 GtC per year could, by the end of the century, be of comparable size as the amount of carbon captured and stored (0.8 GtC leaked per annum) - i.e., recapture would be 80 per cent of the emissions captured.

 

Whilst the annual leakage rate of 1 per cent is arbitrary (the IPCC believes that a 99 per cent retention of CO2 is ‘very likely’ and ‘likely’ over 1,000 years) we must accept that the more carbon dioxide we decide to capture and store in geological reservoirs, the more energy intensive it will be to keep it there, and monitor that it is still there, transferring the responsibility to future generations.

Therefore, artificial carbon sequestration in geological reserves should only be viewed as temporary relief, if at all.

 

 

 

Risk of leakage

People often ask; ‘is geological storage [of carbon dioxide] safe’… it’s a very difficult question to answer. Is driving safe...You might say yes or no, but what makes driving something we’re willing to do…You get automakers to build good cars, we have driver training, we don’t let children drive, we have laws against drunk driving-we implement a whole system to ensure that the activity is safe.299
Sally Benson, Executive Director, Global Climate and Energy Project

As journalist Jeff Goodell writes in his book Big Coal, tens of thousands of people may be destined to live above a giant bubble of CO2 and since,

‘CO2 is buoyant underground it can migrate through cracks and faults in the earth, pooling in unexpected places.’300

A sudden release of large amounts of CO2 due to, for example, an earthquake resulting in the fracturing or pipeline failure could result in the immediate death of both people and animals, since asphyxiation can result from inhalation of CO2 at just a 20 per cent concentration.

 

Because CO2 is a colorless, odorless and tasteless gas; a large leak would be undetected. An example of just how catastrophic a leak could be is the natural limnic eruption of CO2 in 1986 from Lake Nyos in Cameroon. The sudden release of 1.6 Mt CO2 resulted in the asphyxiation of around 1,700 people and 3,500 livestock.

If this rules out the storage of CO2 in land-based geological sites, let us consider sequestration in ocean saline aquifers, such as Sleipner in Norway.

 

Slow, gradual leakage of CO2 could result in the dissolution of CO2 in shallow aquifers, causing the acidification of groundwater and undesirable change in geochemistry (i.e., mobilization of toxic metals), water quality (leaching of nutrients) and ecosystem health (e.g., pH impacts on organisms).301

Transportation of captured carbon could also be problematic. CCS involves a process of converting CO2 to something else, or moving it somewhere else. Taking the transport of natural gas as an example, we can estimate how secure CO2 transportation might be.

 

The world’s largest gas transport system, 2,400km long running through Russia (the Russian gas transport system), is estimated to lose around 1.4 per cent (a range of 1.0-2.5 per cent).302 This is comparable to the amount of methane lost from US pipelines (1.5 ± 0.5 per cent). Therefore, it is reasonable to assume that CO2 leakage from transport through pipelines could be in the order of 1.5 per cent.

 

Furthermore, it is noteworthy that around 9 per cent of all natural gas extracted is lost in the process of extraction, distribution and storage.
 


Storage capacity
A detailed analysis (rather than an estimate) of known US geological sequestration sites undertaken by the US Department of Energy revealed that only 3GtC could be stored in abandoned oil and gas fields.303 This estimate, however, does exclude saline aquifers (very little is known about potential US saline aquifers).

Assuming that the USA took responsibility for CO2 emissions that were directly proportional to its share of global emissions, the USA’s capacity to store its own carbon in known geological sequestration sites would be exhausted in 12 years. Similarly, a recent analysis explored the potential storage capacity in Europe.

 

The study found that based on Europe’s current annual emission rate of 4.1 GtCO2 per year in the EU 25, the medium-range estimate of storage capacity is only 20 times this.304 In other words, CCS is clearly not a long-term solution, as ‘peak storage’ could be reached relatively quickly.

Further sequestration would require expensive and potentially unsafe pipelines directing CO2 to sequestration sites further a field. This would be an energy-intensive process which is why CCS not only poses significant future risks in terms of leakage, but also reduces the net energy gained from a particular fuel - what has been called the ‘energy penalty’.305 Given these problems, to put such faith in schemes which are operationally immature, instead of decreasing our carbon emissions, seems outrageously risky.

 

Surely it would be better not to produce the emissions in the first place?

One further limitation of CCS is that, only one-third of emissions in industrialized countries are actually produced in fossil-fuelled power stations. A significant proportion comes from the transport sector (around 30 per cent), and as yet CCS has only been developed for static CO2 sources.

By pursuing a CCS pathway, we are encouraging our continued reliance on fossil fuels delivering energy through a centralized system. Should CCS become economically viable, it could act to undermine initiatives to move towards a more efficient distributed energy system with diverse arrays of low carbon energy sources.

Could CCS be another ‘just around the corner’ technology like nuclear fusion? Will small-scale pilot projects ever realistically be scaled up to make a significant impact on ever growing global emissions?

For over 50 years, physicists have been promising that power from nuclear fusion (see Box 21) is on the horizon. While fusion has been achieved, in the JET (Joint European Torus) reactor, the experimental rector did not break even, i.e., it consumed more energy that it generated, but managed to produce 16MW of energy for a few seconds.

 

In a Nature news feature, science journalist Geoff Brumfiel commented that,

‘…the non-appearance [of nuclear fusion] should give us some insight into how attempts to predict the future can go wrong’.306

Back to Contents
 

 

 


The limits to nuclear

‘So the big question about nuclear ‘revival’ isn’t just who’d pay for such a turkey, but also… why bother? Why keep on distorting markets and biasing choices to divert scarce resources from the winners to the loser - a far slower, costlier, harder, and riskier niche product - and paying a premium to incur its many problems?

 

Nuclear advocates try to reverse the burden of proof by claiming it’s the portfolio of non-nuclear alternatives that has an unacceptably greater risk of non-adoption, but actual market behavior suggests otherwise.’ 307
Amory Lovins, Chief Scientist, Rocky Mountain Institute

nef’s 2005 report Mirage and Oasis, made the case that nuclear power faced insurmountable problems in living up to expectations placed upon the sector to help deliver both energy security and an answer to climate change.308

 

The report made the case that, if anything, an expanding nuclear program would increase insecurity and, by distracting skills and other resources, delay more effective solutions.

In his book - The lean guide to nuclear energy: a life-cycle in trouble - David Fleming introduces the term ‘energy bankruptcy’, referring to a point in the nuclear energy life cycle where more energy is used in the life cycle than can be supplied as electricity.309 Fleming illustrates that whilst emissions of CO2 from nuclear energy superficially look ‘rather good’ at approximately 60g/kWh (cf. 190g/kWh for natural gas), scratch the surface and it becomes very clear that this comparison is very misleading.

Fleming identifies that the long-term disposal solution for nuclear waste has been deferred, resulting in a ‘back log’ of emissions neither realised nor accounted for yet.

 

Not only will we eventually have to face the challenge of a long-term storage solution of nuclear waste, which will be a very energy-intensive process due to the necessary over-engineering to safeguard future generations from the hazardous waste, but emissions from nuclear energy will grow relentlessly as uranium ores used progressively turn to low-grade.

 

 

 

Box 21. Nuclear fusion
 

Nuclear fusion is technology that produces energy by mimicking the Sun. The fusion of two hydrogen nuclei (a hydrogen atom stripped of its electron) results in the formation of a single Helium nucleus. Since the mass of a single helium nucleus is less than the combined masses of the two hydrogen nuclei, energy is released based on Einstein’s mass-energy equivalence formula E = mc². Initiating the process of fusion requires extremely high temperatures (hundreds of millions of °C), as the positively charged nuclei need to overcome their natural repulsion. This can only be achieved when the nuclei are moving very fast or are closely packed together. As has often been commented, any practical, large-scale application of fusion technology remains decades away, as it has done for decades.


Back to Contents

 

 

 


The hydrogen economy

It is often argued that the next evolutionary step in the global energy system is the substitution of natural gas with hydrogen - often assumed to be a zero-carbon fuel.

 

Whilst this is true at the point of end use, it ignores carbon embedded within the fuel.

Hydrogen itself is not a source of energy, but a carrier. Because of this, hydrogen first has to be produced from the reaction between carbon monoxide (CO) and methanol, through steam reactions (steam reformation) with natural gas, oil or even coal or by the electrolysis of water (efficiencies of fuel cells and hydrogen production are discussed later). But there are two problems here.

Hydrogen will only be truly zero carbon if it is produced through zero-carbon electricity generation. A life-cycle assessment by the National Renewable Energy Laboratory estimates the carbon emissions associated with hydrogen production from the steam reformation of natural gas without CCS, would equal just under 12kg of CO2e for every kg of H2 - one kg of H2 has a similar energy content to 3m3 of natural gas, or the same amount of energy required to drive a 2003 Golf Edition approximately 30km.310

A hydrogen economy, promoted as a zero-carbon energy source, based on the energy system we have at present (i.e., dominance of fossil fuels) relies heavily on the assumption that CCS is safe and secure. And, we have already argued that CCS is by no means guaranteed to work, and there are limited gas reserves.

Other alternatives to steam reforming include the electrolysis of water into hydrogen by using a renewable energy source, such as wind.

 

Yet the process of electrolyzing water to hydrogen, and then burning it as a clean fuel to use in a fuel cell to produce electricity introduces two additional inefficiencies. Why introduce these inefficiencies if there is zero-carbon electricity generation in the first place? Secondly transportation of hydrogen is expensive (both cost and energy).311

Whilst hydrogen may become an effective way of storing energy from renewables to cope with intermittency of electricity supply from renewables, such as wave, solar and wind (an issue often raised by those not in favor of renewable energy), it doesn’t seem likely that the hydrogen economy will be upon us any time soon.

 

 

 

Box 22: Hydrogen economy for the UK’s transport system: is it possible? 312
 

If we decided to run Britain’s road transport system, say, on cleanly produced hydrogen - electrolyzing water using non-CO2-emitting forms of generation - our options would be:

  • solar array covering every inch of Norfolk and Derbyshire combined

  • a wind farm bigger than the entire southwest region of England

All very well, but we’d also need space for renewable energy technology for use in our homes, offices and industries.


Back to Contents

 

 

 


Biofuels

Whilst, biofuels can be produced sustainably and with real CO2 reductions …in the industrialized world there simply isn’t the land.313
David Strahan, author of The Last Oil Shock (2007)

Concern for climate change and the rising price of oil has resulted in new policies that aim to substitute petrol and diesel with biofuels.314 There are, however, a number of unintended consequences of the agro-industrial scaling out biofuels.

Last year the impact of the US’s significant drive for increasing production of bioethanol had a significant impact on the food market because of the diversion of cereals, specifically Maize away from animal feed.315

 

For example, in its 2008 World Development Report, The World Bank stated:

‘Biofuel production has pushed up feedstock prices. The clearest example is maize, whose price rose by over 60 per cent from 2005 to 2007, largely because of the US ethanol program [sic] combined with reduced stocks in major exporting countries. Feedstock supplies are likely to remain constrained in the near term.’ 316

The report then goes on to state:

‘The grain required to fill the tank of a sports utility vehicle with ethanol…could feed one person for a year; this shows how food and fuel compete. Rising prices of staple crops can cause significant welfare losses for the poor, most of whom are net buyers of staple crops’

In other words, the rise in popularity of biofuels is creating competition for land and water between crops grown for food and those grown to make biofuels.

 

This has led to civil unrest around the world. For example, the ‘Tortilla Riots’ in Mexico in 2007 followed the dramatic rise in price of corn (a staple food for poor households) as more land was given over for biofuel production.

In terms of climate change, new calculations looking at the full lifecycle of palm oil production concluded that under a range of fairly typical circumstances vastly more carbon was released into the atmosphere as a result of growing palm oil, than results from burning fossil fuels. 317 In the context of bioethanol, research has also shown that biofuels produced from corn, wheat or barley all contain less energy than the energy required to produce them.318

Research published earlier in 2007 showed that the growth of palm oil for biodiesel for the European market is now the main cause of deforestation in Indonesia.319 Because of deforestation and drainage of peat-lands necessary to grow the crop, every tonne of palm oil created in South East Asia resulted in up to 33 tonnes of carbon dioxide emissions - ten times as much as conventional petroleum.

 

Separately, an estimate by a coalition of aid and environment groups including Greenpeace, Oxfam, the RSPB, WWF and Friends of the Earth, suggests that soya grown for biodiesel grown on deforested land would take 200 years before it could be considered carbon neutral.320

In light of the seemingly unsustainable nature of biofuels, in 2008 the UK government commissioned Edward Gallagher to examine the indirect impact of biofuels on climate change and food security.321 The review confirmed growing concerns of the negative impacts of UK and EU biofuels policy on land use, greenhouse gas emissions and food security.

 

In light of the review, the UK government has agreed to reconsider its policy on biofuels.

 

 

 

Box 23:

Is the complete or even partial substitution of diesel and petrol fuels with biofuels possible?

  • If the UK directly substituted all its diesel and petrol fuels (by energy not volume) to rapeseed oil biodiesel and corn bioethanol, the amount of agricultural land required would be approximately 36 million hectares. To put this figure in context, the total land area in the UK is just over 24 million hectares. Furthermore, less than 20 per cent of the UK’s land is suitable for agriculture.

  • To meet President Bush’s goal of increasing bioethanol production from the five billion gallons currently produced to 35 billion gallons by 2017 would require more corn than the USA currently produces.322

  • To replace 10 per cent of global petrol production with bioethanol, Brazil would have to increase its ethanol production by a factor of 40, and would result in the destruction of around 35 per cent of the remaining Amazon Rainforest.

  • By increasing the consumption of bioethanol to around 34 million barrels per year by 2050, 1GtC of carbon could be offset, due to the substitution of mineral liquid fuels.323 We find, however, that coupled to population growth; this would require a 25 per cent increase in cultivated land by 2050. This will clearly mean claiming a vast amount of land from the already stressed natural environment.


Back to Contents

 

 

 


Geoengineering - technological saviour or damaging distraction?

‘There is a suspicion, and I have that suspicion myself, that a large number of people who label themselves “green” are actually keen to take us back to the 18th or even the 17th century. [Their argument is] “Let’s get away from all the technological gizmos and developments of the 20th century”…And I think that is utter hopelessness ... What I’m looking for are technological solutions to a technologically driven problem, so the last thing we must do is eschew technology.’
Professor Sir David King, former Chief Scientific Advisor to the UK government 324

As we have shown earlier in this report, even modest changes to the work-and-spend lifestyle of the global North would be hugely beneficial, yet David King’s comments imply that the political consensus is that changes in lifestyle should not be necessary and would be largely unwelcome. As a result more novel solutions to climate change are beginning to receive more and more interest.

Once an idea limited to the realms of a James Bond film, human manipulation of climate - geoengineering - is increasingly being discussed by some of the most respected climate scientists in the world.

 

From giant mirrors in space reflecting sunlight away from Earth, to pumping aerosols into the stratosphere, or large-scale cloud-seeding (releasing aerosols in the lower atmosphere is thought to initiate the formation of clouds), geoengineering could, in the not too distant future, become a reality.

In its current context,325 geoengineering technologies can be divided into two categories: those that remove greenhouse gases from the atmosphere, and those that reduce incoming solar radiation - with the intention of offsetting the changes to Earth’s radiation budget caused by anthropogenic greenhouse gases. The debate about the role geoengineering can and should play in dealing with the impacts resulting from climate change is one which is already beginning to gain momentum.326

In most cases, geoengineering schemes are viewed as a stopgap between now and some point in the future where mitigation technology is cheaper and more widespread. There are, however, large technical and scientific uncertainties.

 

For example Professor David Victor, Director of the Laboratory on International Law and Regulation at Stanford University argues:

‘…real-world geoengineering will be a lot more complex and expensive than currently thought because simple interventions—such as putting reflective particles in the stratosphere—will be combined with many other costlier interventions to offset nasty side effects.’327

The large majority of academics working in the field of geoengineering research have been clear that their research and technical propositions are not intended to distract from the efforts of reducing greenhouse gas emissions as the first priority for controlling climate change.

 

However, many now argue that a technological intervention may be required parallel to current mitigation efforts.328

The Royal Society’s recent report Geoengineering the climate: Science, governance and uncertainty assessed both technical and social aspects of geoengineering options.329 With respect to the technical level, two approaches are identified: Carbon Dioxide Removal (CDR) techniques and Solar Radiation Management (SRM) techniques.

The objective of CDR methods is to remove CO2 from the atmosphere by; enhancing uptake and storage by terrestrial biological systems, enhancing uptake and storage of oceanic biological systems or using engineered systems (physical, chemical, biochemical). In contrast to this, SRM techniques focus on changing the Earth’s radiation budget by reducing shortwave radiation absorbed by the Earth.

 

Both techniques have the ultimate aim of reducing global temperatures, however they differ in their modes of action, timescales over which they work, and costs. There is a general preference towards CDR methods as a way to augment continuing mitigation action in the long term, whilst SRM could provide short-term back-up for rapid reduction in global temperatures.330

Of the two techniques, The Royal Society report found SRM to have the least potential. This is due to high levels of uncertainty associated with large-scale modification of the climate. In particular, climate scientists have raised concerns about the potential impact SRM may have on rainfall patterns.331 While temperatures may return to those of the pre-industrial era, rainfall patterns would not.332

 

There is also particular concern about the impact of SRM interventions on the Asian and African summer monsoons on which billions depend.333

 

Furthermore, beyond non-invasive laboratory/computer modeling and analogue case studies - the first phase of research and development of SRM technologies - research would necessarily involve intentional interventions with the climate system.

 

Because it is a technology with many uncertainties, field experiments beyond limited duration, magnitude and spatial-scale could involve some risk of unintended climate consequences. Yet, the collection of direct empirical evidence from large-scale field experiments would be a necessary part of any research programme.334

Researchers have also highlighted that should any SRM intervention stop abruptly or fail, global temperatures could rise rapidly.335 As the concentration of CO2 in the atmosphere increases, carbon sinks would be weakened with possible carbon-cycle feedbacks accelerating the increase in CO2 concentration in the atmosphere.

 

Termination of the climate modulation provided by a geoengineering scheme, could result in a temperature change of 2-4°C per decade (there is no evidence that global temperature changes have approached this rate at any time over the last several glacial cycles).336 This rate of temperature change is 20 times faster than the rate predicted under a business-as-usual scenario. Clearly such a rapid change in climate would have devastating impacts on humans and the environment.

The Royal Society’s viewed CDR as having the most potential and as some mimic natural processes (e.g. ecosystem-based CDR and some engineered CDR) they may involve fewer risks compared to SRM. However, this category of geoengineering is likely to be less effective in reducing global temperatures quickly.

Both CDR and SRM are relatively under researched technologies.337 Specifically with respect to SRM, there has been limited consideration in these proposals on the impact of continued increases in CO2 - this is the most worrying.

 

For example, the direct effect of elevated CO2 could have significant effects on the hydrological cycle. For example, a recent modelling study showed that in the absence of climate warming and with elevated CO2, changes to plant water use efficiency resulted in a decrease in precipitation over vegetated areas in the Tropics. 338

However, one of the most critical reasons for making absolute cuts in CO2 emissions is due to acidification of ocean waters.339 As CO2 is absorbed by the oceans, it forms a weak acid, called carbonic acid. Part of this acidity is neutralized by the buffering effect of seawater, but the overall impact is to increase the acidity.

 

According to a report by the Royal Society, apart from global climate change, this should be the second largest motivation for reducing CO2 emissions.340

 

So far, the acidity of the ocean surface has increased by 0.1 units. General circulation models show that if CO2 emissions from fossil fuels continue to rise, a reduction of 0.77 units could occur by 2300.

To put this in perspective, over the past 300 million years, there is no evidence that the pH of the ocean has ever declined by more than 0.6 units. While there is limited research into the impact of pH decline on marine biota, organisms which have calcium carbonate skeletons or shells, such as molluscs, coral and calcareous plankton, may be particularly affected, especially as a large proportion of marine life resides in surface water.341

Given that techniques for reducing acidity are unproven on a large scale and could have additional negative impacts on the marine environment, it is clear that a technical solution that only partially deals with controlling the climate will not address anthropogenic interference of the carbon-cycle.

Whilst none of the current geoengineering methods currently offer immediate solutions to the problems of climate change, nor do they replace the need to continually reduce emissions- a growing group of academics now argue that they could be a potential option to actively engineer the climate on a planetary-scale to curb and control the impacts associated with a global temperature rise of 2ºC or more.

Although our understanding of the climate system continues to improve, and the forecasting skill of climate models improves, there is still no guarantee that we’d be able to predict the implications of manipulating the delicate energy balance of the climate system that has already been hurled out of equilibrium.

As well as the technical feasibility of geoengineering, its application must also be socially and ethically permissible.

 

The unknown factors associated with manipulating climate change heighten the need for any decisions to be mutually agreed upon and accepted. The language of ‘risk’ cannot be disassociated from this debate as the changes created by geoengineering, may, in the long term be irreversible.

 

So, if the effects of geoengineering were to be irreversible, then those who made the decision to undertake these technologies would be choosing one climate path for future generations rather than another.342, 343
 

Back to Contents

 

 

 


How much can energy efficiency really improve?

One hundred years ago, electricity production, at best was only 5-10 per cent efficient. For every unit of fuel used, between 0.05 and 0.1 units of electricity were produced. Today, the global average efficiency for electricity generation is approaching 35 per cent and has remained largely unchanged for the past 40 years.344

 

This may come as a surprise given the often-held view that technology has continued to improve ess and will continue to do so in the future.

 

Whilst this is largely due to the current mix of the global energy system, rather than individual technologies, it highlights two problems associated with the assumption that we can expect a steady increase in energy efficiency/decline in carbon intensity of the global economy. First, as a general rule of thumb, in a given technology class, efficiency normally starts low, grows for decades to centuries and levels-off at some fraction of its theoretical peak.345

 

As described earlier in the report, the second law of thermodynamics, is one of the most fundamental physical laws; it states that energy conversion always involves dissipative losses (an increase in entropy). As such, any conversion can never be 100 per cent efficient.

The results of our analyses have shown, future stabilization pathways are dependent on assumptions about energy intensity and, therefore, energy efficiency. These assumptions fail to acknowledge, however, that in many cases of ess, engineers have already expended considerable effort to increase the energy efficiency.346

Second, we are built into and are still building ourselves into a centralized energy system. Such systems favor fossil and nuclear fuels over renewable energy, do not exploit the maximum efficiency possible (i.e., do not favor a system where an exergy cascade, such as combined heat and power, can be utilized), and the energy system is subject to large distribution loses.

 

This is likely to continue into the future if energy policies rely heavily on nuclear and CCS schemes.

 

Particularly given that CCS reduces the efficiency of the energy system, and nuclear fission is a mature technology, already approaching its efficiency limit, and is far from being carbon neutral, as is often claimed. If nuclear fusion ever becomes a viable option, it is likely to have the same thermal efficiencies as nuclear fission.347

 

In other words, many of the technologies that make up the global energy system are mature technologies and their current efficiencies are at or almost at their practical maximums.

Figure 20:

Average lifespans for selected energy/related capital stock348

 

The slow capital stock turnover for large energy infrastructure - shown in Figure 20 also means that energy decisions made now will influence the trajectory of emissions over the next 25-60 years, with obvious implications for the speed at which a transition to a low carbon economy can take place.

Amongst the most efficient technologies are large electric generators (98-99 per cent efficient) and motors (90-97 per cent). This is followed by rotating heat engines that are limited by the Carnot efficiency limits (35-50 per cent), diesel (30-35 per cent) and internal combustion engines (15-25 per cent).

 

Improvements in these areas are, therefore, small. In fact, the energy efficiency of steam boilers and electrical generators has been close to maximum efficiency for more than half a century. 349

 

Similarly, the most efficient domestic hot water and home heating systems have been close to maximum efficiency for a few decades.350

Whilst hydrogen fuel cells are often pursued as future sources of ‘clean, zero-carbon, highly efficient sources of energy’, there are also upper limits to the energy efficiency achievable. Fuel cells are currently 50-55 per cent efficient, and are believed to reach a maximum at around 70 per cent. This is due to limits imposed by electrolytes, electrode materials and catalysts within the fuel cell system. Additionally the production of hydrogen from oil or methanol is has a maximum efficiency of 75-80 per cent.351

In terms of renewable energy, photovoltaic (PV) cells currently have efficiencies between 15 and 20 per cent (in commercial arrays) with a theoretical peak of around 24 per cent (highest recorded efficiency = 24.7 per cent).

 

This maximum is higher for multi-band cells and lower for more cost-effective amorphous thin films. Wind turbines have commercial limits are around 30-40 per cent, with a maximum efficiency limit of 59.6 per cent - the Betz limit.352 Hydroelectric power is already at its maximum average efficiency of around 85 per cent.353

Photosynthesis is highly inefficient in converting sunlight into chemical energy with the most productive ecosystems being around 1-2 per cent efficient and a theoretical peak of around 8 per cent. The extent of bioenergy is also restricted by the volume of biomass necessary versus land available which is possibly not greater than 30 per cent of the Earth’s land-surface.

In the case of lighting, high pressure sodium vapor has an energy efficiency of around 15-20 per cent, whilst fluorescent (10-12 per cent) and incandescent (2-5 per cent) illumination generate more heat than light.

For transport systems, specifically road transport, improvements in private vehicle efficiencies are largely due to vehicle mass (see Box 9), driving patterns and aerodynamic drag and the use of technology such as regenerative breaking (electric power recovery from mechanical energy otherwise lost). The efficiency of the internal combustion and diesel engine are largely at their maximum.

 

Further improvements could be made by hybrid, electric (dependent on the central power plant efficiency) or fuel cell vehicles.

 

Box 24 shows a similar leveling of in aviation efficiency gains.

 

 

 

Box 24. Aviation eating up efficiency gains
 

Some are optimistic that technological improvements will allow air travel to continue to grow into the future while keeping emissions under control - and eventually reducing them overall. This kind of optimism was embodied by the strap line that heralded the new Airbus A380 on its maiden flight from Singapore to Sydney in 2007: ‘cleaner, greener, quieter, smarter’.

Overall fuel efficiency gains of 70 per cent between 1960 and 2000 are often cited as evidence for continued improvements in efficiency.

 

For example, the Air Transport Action Group has said:

‘Building on its impressive environmental record, which includes a 70 per cent reduction in… emissions at source during the past 40 years, the aviation industry reaffirmed its commitment to… further develop and use technologies and operational procedures aimed at minimizing noise, fuel consumption and emissions.’354

There is little evidence, however, that major improvements will be made in the near future. Despite technological achievements so far, absolute growth in fuel use by aircraft has grown by at least 3 per cent per year.355 Quite simply, the efficiency improvements of 0.5 to 1.3 per cent a year that have been achieved are being dwarfed by the industry’s annual growth of 5-6 per cent.356 The time it takes to pension off and replace commercial aircraft is long, and any additional efficiency gains anticipated are likely to be wiped out by a continuing increase in flights.357

The Advisory Council for Aeronautics Research in Europe (ACARE) has established ambitious goals for improvements to aircraft efficiency. By 2020, it wants the industry to achieve a 50 per cent reduction in CO2 per passenger kilometer. Of this, 15-20 per cent will be from improvements to engines, 20-25 per cent from airframe improvements and a further 5-10 per cent from air traffic management.358

 

But to achieve these targets, the industry would need to improve its efficiency by over 2.5 per cent per year. In reality, efficiency gains of just 1 per cent have been described as ‘rather optimistic’ given that the jet engine is now regarded as mature technology, and annual efficiency improvements are already falling.359

An analysis of projected aviation growth and anticipated improvements in aircraft efficiency suggests that if growth in Europe continues at 5 per cent, traffic will double by 2020 (relative to 2005). With an ‘ambitious’ 1 per cent annual improvement in fleet efficiency, CO2 emissions would rise by 60 per cent by 2020 (and 79 per cent if emission trading did not affect growth).

 

Even if a 10 per cent reduction in CO2 per passenger kilometer were to be achieved, CO2 emissions would rise by 45 per cent.360

Figure 21 shows long-haul aircraft efficiency gains since 1950 as an index based on the De Havilland DH106 Comet 4 (the least efficient long-haul jet airliner that ever flew). It shows a sharp improvement in efficiency between 1960 and 1980 but a steady slowing of efficiency gains since then. Further efficiency gains between 2000 and 2040 are likely to be in the order of 20-25 per cent.361

 

Even the performance of the new Airbus A380 fits neatly into the regression, indicating that the 50 per cent more efficient aircraft that some have predicted by 2020 are highly unlikely.

 

 

Figure 21.

Long-haul aircraft efficiency gains since 1950 as an index based on the De Havilland DH106 Comet 4.362

 


Figure 22.

EROI for various electric power generators.363

 

One way of comparing efficiencies of different technologies is through an EROI assessment.

 

Figure 22 shows various EROI ratios for a number of electric power generators. It shows that wind turbines compares favorably with other power generation systems. Base load coal-fired power generation has an EROI between 5 and 10:1. Nuclear power is probably no greater than 5:1, although there is considerable debate regarding how to calculate its EROI.

 

The EROI for hydropower probably exceeds 10, but in most places in the world the most favorable sites have already been developed.
 


Practical limitations to the improvements in supply-side energy efficiency

An increase in resource efficiencies alone leads to nothing, unless it goes hand in hand with an intelligent restraint of growth.364
Wolfgang Sachs (1999)

In terms of work generation from a heat engine (where heat is converted to work), the Carnot efficiency, named after the French Physicist Nicolas Léonard Sadi Carnot, determines the maximum efficiency in which this can be achieved.

The thermal efficiency of gas and steam turbines is a function of the temperature difference between the inlet temperature and the outlet temperature. In a perfect Carnot cycle, the maximum efficiency that can be achieved is around 85 per cent. In reality, the most efficient combined-cycle gas turbine (CCGT) plants have efficiencies in the range of 59-61 per cent.

 

In a CCGT, gas is used to drive a turbine and the exhaust gases are used to raise steam to drive a second turbine. The high efficiency of this type of turbine is due to the use of both the gas and the ‘waste’ exhaust gases. Currently, however, the average fossil-fuelled power plant is approximately 33 per cent efficient.365

 

With the potentially imminent peaking in production of gas, it seems unlikely this will change significantly in the future.

An Integrated Gasification Combined Cycle plant (IGCC), is a similar technology to CCGT, but uses coal as a feedstock. Coal is converted into a synthetic gas and then used in a CCGT.

 

The efficiency of an IGCC is in the range of 30-45 per cent. Obviously, without CCS, this process would act to increase the carbon intensity of the economy, but with CCS the efficiency of the plant declines. Biomass could be used as a feedstock however, which could have a significant impact on the level of carbon emissions.

Fuel cell technology converts the chemical energy of fuels directly (electrochemically rather than through combustion) and therefore, is not restricted by the Carnot efficiency limit. Therefore, considerably higher efficiencies can be met. There are a number of different types of fuel cells entering the market. Generally all fuel cells run on hydrogen, although some can run on fuels such as CO, methanol, natural gas or even coal if externally converted to hydrogen.366

 

The advantage of fuel cells is that emissions at point of use are simply water vapor and therefore could significantly contribute to a reduction in urban pollution.

 

But, as described earlier in the report, hydrogen is not a fuel; it is a carrier of energy. And, if the hydrogen is produced from a hydrocarbon fuel, then the benefits as a low carbon solution are reduced. Furthermore, scaled up significantly, fuel cell technology will hit other limiting factors, such as the availability of the metal platinum - a catalyst in the fuel cell.

It is useful at this point to return to the term ‘exergy’. This describes the maximum useful work obtainable from an energy system at a given state in a specified environment.

By and large, any attempt to increase the overall efficiency of a supply-side energy process could be achieved by making use of low exergy products as well as high exergy products of energy generation. An example of this is CCGT (described above) or a combined heat and power station (co-generation). Co-generation involves the recovery of thermal energy that is normally lost or wasted. Both electricity and the low-grade waste heat are used for both powering appliances and heating.

 

By adding district heating capacity to a CCGT, efficiency can increase to almost 80 per cent.
 


Distributed generation?

An area that is strongly associated with efficiency of the energy industry is distributed generation. While its main benefits are cleaner and more efficient generation and location of generation closer to demand, distributed generation also has an effect on losses.

 

In simple terms, locating generation closer to demand will reduce distribution losses as the distance electricity is transported will be shortened, the number of voltage transformation levels this electricity undergoes is lessened and the freed capacity will reduce utilization levels.367
Ofgem (2003)

Using an economic model developed by the World Alliance for Decentralized Energy (WADE),368 it has been repeatedly shown that the pursuit of a decentralized renewable energy system with cogeneration is becoming increasingly economically attractive; not only for mitigating climate change, but also in the face of dwindling fossil fuel reserves.369

Centralized energy systems, such as the UK’s on average lose 9.3 per cent (global average is 7.5 per cent) of all electricity generated through transmission and distribution losses.370

 

Ofgem estimated that the UK could achieve approximately 4 per cent of the UK government’s domestic target of a 20 per cent reduction in CO2e by 2010 through simply reducing distribution losses by 1 per cent. When these distribution losses are considered, the argument against a new nuclear age or large-scale CCS is strengthened further.

Distributed energy is a favorable pathway for developing nations. This is because a centralized energy system using a transmission network like the National Grid requires a high capital transmission and distribution network. Once in place, the network will also have high operation and maintenance costs as well as significant energy losses.

The challenges to decentralized energy are fourfold, however:

  1. Policy and regulatory barriers to decentralized energy.

  2. Lack of awareness and effectiveness of decentralized energy.

  3. Failure of industrial end-users to accept and adapt to decentralized energy agenda.

  4. Concerns regarding the dependence of decentralized/cogeneration of fossil fuels. Indeed, the decentralized system proposed by Ken Livingstone, is based on combined heat and power from CCGT.

Cogeneration lends itself to specific types of generation, generally small scale (less efficient), close to where the low-grade heat can be used.

 

This, therefore excludes nuclear. It is also difficult to obtain large and/or consistent benefits from cogeneration, since the normally lost or waste heat cannot be stored until needed. Thus, it is necessary to try to balance the amount and timing of the loads between electricity generation and heat utilization.371

Given this,

‘cogeneration is likely to remain a relatively minor contributor to improved energy efficiency.’372

Nevertheless, a decentralized energy system is still more efficient in terms of transmission and distribution losses.

The absolute theoretical efficiency that can be achieved assumes that energy operations experience no losses. It is estimated that ess is currently 37 per cent at the global level and that a two-fold increases may be possible, i.e., a 200 per cent improvement in ess.373

 

But the assumption that the types of technology that could lead to such a significant change will become commercially available and installed at a rate concomitant with within the timescales necessary to stabilize greenhouse gas concentrations at a ‘safe level’ is questionable.

Whilst the limits of thermodynamics only apply to the heat engine (thermal) generation of electricity, there are also theoretical and practical limits to the use of renewable energy, also based on the second law of thermodynamics.
 


The limits to a renewable energy fix
There are numerous reasons for a rapid transition to a global energy system based on renewable technologies: wind, water and solar.

 

As described throughout this report, these include climate change, energy security in the face of Peak Oil, cost-effective conversion and flexible and secure supply. Several studies have shown that, although not without a few difficulties to overcome, it is both practical and possible to meet the global demand for energy from these sources.374

One recent study published in Scientific American in late 2009 outlined a plan to achieve just this - the complete decarbonisation of the global energy system - by the year 2030.375

 

Based only on existing technology that can already be applied on a large scale, it called for the building of 3.8 million large wind turbines, 90,000 solar plants and a combination of geothermal, tidal and rooftop solar-PV installations globally.

 

The authors point out that while this is undeniably a bold scheme, the world already produces 73 million cars and light trucks every year. And, for comparison, starting in 1956 the US Interstate Highway System managed to build 47,000 miles of highway in just over three decades, ‘changing commerce and society’.

But, even plentiful supplies of renewable energy are not a ‘get out of jail free’ card for economic growth. The reasons are few and straightforward.

  • First, growth has a natural resource footprint that goes far beyond energy and we have to learn to live within the waste-absorbing and regenerative capacity of the whole biosphere.

     

  • Secondly, even under the most ambitious program of substituting new renewable energy for old fossil fuel systems, it will take time and, in climate terms, we are, according at least to James Hansen, already beyond safe limits of greenhouse gas concentrations.376

    More global growth will take us even further beyond, with few guarantees that in the space of a few short years the chances of avoiding runaway climate change become unacceptably small.

     

  • Thirdly, we also have to take into account the fact that, at least until renewable energy achieves a scale whereby its own generated energy becomes self-reproducing in terms of the energy needed for manufacture, even renewable energy systems have a resource footprint to account for.

     

    For example, recent research by the Tyndall Centre for Climate Change Research suggests that embodied energy in new energy infrastructure means that it would be approximately eight years before a decarbonisation plan would have a meaningful impact on emissions.377

Renewable technologies are rightly regarded as a potential source of future employment and have a large economic contribution to make, and tend to be seen as carbon neutral or potentially negative.378

 

Despite this, their overall environmental impact is not entirely benign, and this is particularly evident when renewable technologies are considered on a large-scale, something that is regularly assumed in future emission/ economic growth scenarios.

Renewable energy supply is still constrained by the laws of thermodynamics, since energy is being removed from a system; the natural system of the Earth. Whilst this refers to the theoretical limits of energy from renewable sources, there are also practical limits; for example,

‘…large enough interventions in [these] natural energy flows and stocks can have immediate and adverse effects on environmental services essential to human well-being’.379

This is most obviously the case where biomass (e.g. biofuels) are concerned. It has been suggested that given that 30-40 per cent of the terrestrial primary productivity is already appropriated by humans; any major increase could cause the collapse of critical ecosystems.380

In the IEA AP scenario, it is assumed that biofuels, such as biodiesel and bioethanol will replace mineral oil for use in transport. Without encouraging more land-use change, a major anthropogenic contributor to CO2 emissions, relying on energy biomass to provide a natural replacement to gasoline (petrol) would mean competition of agricultural land for food and fuel. Yet, with increasing population and increasing energy requirements is this physically possible without causing widespread ecosystem collapse?

 

This is one of the key reasons why Jacobsen and Delucchi, authors of the study published in Scientific American, do not rely on biofuels in their plan.381

Not all biofuels, though, are reliant on a primary resource feedstock, such as sugarcane and corn (bioethanol) or rapeseed and soya (biodiesel). Cellulosic ethanol can potentially be produced from agricultural plant wastes, such as corn stover, cereal straws, sugarcane bagasse, paper, etc. The technology, however, requires aggressive research and development as it is not yet commercially viable.

At present the energy intensity of this type of ethanol production means that the overall energy value of the product is negative, or only marginally positive, although it is hoped that this will improve as technology develops.382

 

However, a number of experts feel less positive.383

 

For example: according to Eric Holt-Giménez, the executive director of FoodFirst/Institute for Food and Development Policy:

‘The fact is that with cellulosic ethanol, we don’t have the technology yet. We need major breakthroughs in plant physiology. We might have to wait for cellulosic for a long time.’384

Elsewhere, approximately one-half of the global available hydro power has already been harnessed. Little efficiency improvement, also, can be expected from wind turbines, which are at about 80 per cent of the maximum theoretical efficiency.385

 

The efficiency of solar PV cells could, however, increase from the present 15 per cent to between 20 per cent and 28 per cent in unconcentrated sunlight.386

To be unequivocal, renewable energy is a very good thing and has enormous potential to expand. Something like the Jacobson and Delucchi plan for 2030 is an urgent necessity at a global level if we are to avoid catastrophic global warming.387 As we have shown, zero-carbon or low carbon energy sources are not infinite.

 

Therefore there is no excuse to avoid addressing the waste of energy.
 


Practical limits to energy efficiency (demand and supply side)
In general, energy efficiency improves at a slow rate of around 1 per cent per year.

 

This rate is not policy-induced and is entirely due to technological developments the Autonomous Energy Efficiency Improvement parameter (AEEI). This global figure, however, has a regional signature. For example, evidence in 1990 suggested that the pace of AEEI in the USA slowed or stopped.388

Overall an energy efficiency improvement rate of 2 per cent (AEEI plus policy intervention) per year is often considered achievable. Higher energy-efficiency improvement rates in the range of 3-3.5 per cent are also thought to be possible due to continuous innovation in the field of energy efficiency.389

 

For industrialized countries, this means a reduction of primary energy use by 50 per cent in 50 years compared to current levels. This means that in spite of the doubling of energy use under business-as-usual conditions, the energy use could be as low as 50 per cent of the current level.390

But given the limitations discussed above, significant improvements to efficiency increases are only likely to be due to improvements in chemical processes rather than fuel combustion and increases in end-user energy efficiency.391

In terms of end-user efficiency, there is a long way to go. ‘Unrealised’ energy conservation measures in OECD countries may amount to 30 per cent of total energy consumption.392,393

 

Some suggest that if there are no economic, social or political barriers, an instantaneous replacement of current energy systems by the best available technology would result in an overall efficiency improvement of 60 per cent.394 This is contrary to the forecast improvement of efficiency by 270 per cent if historical efficiency improvement rate of 1 per cent continues and is maintained over the next 100 years.

 

And, even if this was possible, the improvement rate of 1 per cent would be unlikely to continue beyond 100 years.395
 


Demand side barriers

When energy efficiency promoters claim that we can get more out of less, we must conclude that the focus so far has been to get more out, period! 396
Nakicenovic and Gruebler (1993)

Throughout this report, we have shown that observations of changes to carbon intensity and energy intensity of the economy over the past decade have failed to improve at a rate necessary to slow the increase in greenhouse gas concentrations, and in recent years appear to be heading in the opposite direction.

 

This is supported by a recent report by the IEA on trends in energy consumption between 1973 and 2004.397

 

The report found that while energy intensity had fallen by over 30 per cent since 1973 - it now takes one third less energy to produce a unit of GDP in IEA economies - the rate of change has slowed.398 Improvements in energy intensity have slowed in all sectors of the economy since the 1980s.

 

As such, projecting forward historical rates of energy efficiency are misleading. But this is the basic assumption made by most the future emissions scenarios.

While we have shown that improvements to supply-side efficiency is limited by practical limits to energy conversion and technological, what are the drivers of demand side energy efficiency? Earlier in the report we discussed the significance of the rebound effect, whereby efficiency savings are offset by increases in consumption (see Box 8). There are a number of additional barriers to demand side efficiency - eeu (see Equation 1). These are shown in Box 25.

All these factors contribute to the failure of energy efficiency to drive absolute emissions downward. The main reason, however, relates to the market imperfection. For example the IEA found that the price signals in the 1970s did more to increase efficiency than improved technology has done since the 1980s.399

 

In other words, the cost of energy is currently too low. Because of subsidies or the externalization of the environmental cost, the wasteful use of energy is encouraged.
 


Limits to the speed of technological uptake

The magnitude of implied infrastructure transition suggest the need for massive investments in innovation energy research.400
Hoffert et al. (1998)

Based on historical evidence, what is the capacity for social and institutional organizations to rapidly change? Is there a limit to our ability to produce knowledge and new technology to deal with a problem? Surprisingly, this is a vastly under researched field.

 

For example, Tim Lenton and colleagues conclude in their paper on tipping points with the following statement:

‘A rigorous study of potential tipping elements in human socioeconomic systems would also be welcome, especially to address whether and how a rapid societal transition toward sustainability could be triggered, given that some models suggest there exists a tipping point for the transition to a low-carbon-energy system.’401
 

 

Box 25. Barriers for energy efficiency improvements 402, 403
 

Technical barriers: Options may not yet be available, or actors may consider options not sufficiently proven to adopt them.

Knowledge/information barriers: Actors may not be informed about possibilities for energy-efficiency improvement. Or they know certain technologies, but they are not aware to what extent the technology might be applicable to them.

Economic barriers: The standard economic barrier is that a certain technology does not satisfy the profitability criteria set by firms. Another barrier can be the lack of capital for investment. Also the fact that the old equipment is not yet depreciated can be considered as an economic barrier.

Institutional barriers: Especially in energy-extensive companies there is no well-defined structure to decide upon and carry out energy-efficiency investments.

The investor-user or landlord-tenant barrier: This barrier is a representative of a group of barriers that relate to the fact that the one carrying out an investment in energy efficiency improvement (e.g., the owner of an office building) may not be the one who has the financial benefits (in this example the user of the office building who pays the energy bill).

Lack of interest in energy-efficiency improvement: May be considered as an umbrella barrier. For the vast majority of actors, the costs of energy are so small compared to their total (production or consumption) costs that energy-efficiency improvement is even not taken into consideration. Furthermore, there is a tendency that companies, organizations and households focus on their core activities only.


 

While there is a growing awareness of the urgency with which the transition to a low carbon economy must be made, identification of potential tipping elements in human systems is still a largely under-researched area.

A recent report to the US Department of Energy has noted that it takes decades to remake energy infrastructures.404 This is further supported by Figure 20 which maps capital stock turnover rates for energy related infrastructure. Decisions made now in terms of transport and energy infrastructure and the built environment will determine the capability of a nation to reduce its carbon footprint.

 

Highly centralized energy systems, inefficient buildings and poor planning will make a difficult task even more challenging.

Climate change has long been viewed as a pollution problem. This has led to the interpretation of climate change in predominantly scientific terms by policy makers, the media and environmental NGOs resulting in technocentric responses gaining more interest than any more systemic change.

 

However, the growing emphasis on the technological or market-based initiatives as a cure-all ignores what we have shown in this report - that the challenges we currently face, have their roots in a faulty economic system. So, with the vast majority of efficiencies realized, it appears restructuring of the economic system may be the only route by which we can achieve the emission cuts necessary.

In the context of energy systems, the findings of this report only add to the desirability of carefully considered low carbon planning, and other prompt actions to slow down the use of energy and resources.

 

Such solutions can also improve inter alia resilience to exogenous shocks such as volatile food or energy prices, local economic regeneration, social cohesion, physical and mental well-being, employment opportunities and the increased individual and community capacity to reduce emissions and resource use.

 

For example, investment into renewable energy can create new jobs often in areas where they are needed the most. If installed at the local level, renewable energy schemes can also contribute to local economic regeneration, social cohesion (an important factor for adaptive capacity) and improve environmental literacy.

 

Energy efficiency and decentralized or low carbon energy production targeted at low-income households also has the potential to reduce fuel poverty or access to energy caused by poor living standards and low-incomes.

 

Back to Contents
 

 

 


Equity considerations

So far, in the growth and emissions scenarios, we have abstracted from national differences to look solely at globally aggregate data.

 

Unfortunately, detailed national projections for fuel mix and fuel usage are not readily available, not to mention the difficulty of making assumptions about national technology levels and adoption speeds. It is possible, however, to look at national level GDP and growth data, as this is more easily available. Additionally we have been abstracting from actual predictions of growth to look at the energy and emissions possibilities given varying levels of growth.

The scenarios presented earlier indicate that even with very optimistic assumptions about energy and carbon intensity improvements and technology adoption, the world will not meet the target for emissions reductions.

 

To meet that target will require aggressive technological improvements combined with a slowing of our use of resources and a reduced demand for energy-intensive goods and services. That implies lower growth. Yet the world is not an equal place, with income and emissions levels varying by orders of magnitude from one country to the next.

 

Expecting reductions in growth along with carbon/energy intensity improvements may seem reasonable for industrialized economies, where additional income does little to increase well-being in society.

Clearly the situation is different in low-income countries, some of which have incredibly low income levels along with their high mortality rates, low life expectancy and low measures of well-being.

 

These countries could not be expected to bear equal measures of growth reduction, especially since they were not responsible for the historical emissions which have brought us to this critical threshold of rapid climate change.

Allowing some low-income countries to grow rapidly and offsetting that with further reductions of growth in the industrialized world would not be very costly for most cases, as the low-income countries start with low bases of economic size. Ten per cent growth in Malawi, for example, would require little offsetting growth reductions in the UK.

 

But this is not uniformly the case. Leaving aside the problems of domestic inequality, fast-growing economies such as India and China have large bases of economic activity, despite their comparatively lower per capita incomes. Faster growth in those two economies, which could help eliminate global poverty if well distributed, would need to be accompanied by off-setting reductions in the industrialized world.405

 

Consumption in the North simply cannot continue at its current level if society is to address both the poverty and climate change problems.

 

Back to Contents
 

 

 


If not the economics of global growth, then what? Getting an economy the right size for the planet
 


The stationary state
The lineage of the notion of ‘one planet living’ can be traced at least as far back as the early nineteenth century. Philosopher and political economist John Stuart Mill was shaped by the human and environmental havoc of the voracious Industrial Revolution.

In reaction to it, he argued that, once certain conditions had been met, the economy should aspire to exist in a ‘stationary state’. It was a hugely radical notion for the time. Mill thought that an intelligent application of technology, family planning, equal rights, and a dynamic combination of a progressive workers movement with the growth of consumer cooperatives could tame the worst excesses of capitalism and liberate society from the motivation of conspicuous consumption.

He prefigured Kropotkin’s analysis that economics could learn from the success of cooperation, or ‘mutual aid’ as he coined it, in ecological systems, itself a riposte to the fashionable misappropriation of Darwinism to social and economic problems.406

 

The latter economic folk wisdom remains nevertheless strong. And even today, the Anglo Saxon economic model is commonly defended with similar misappropriations of Darwin that emphasize the ‘law of the jungle’ and ‘survival of the fittest.’

 

This view suggests that competition in economics, as in nature, should be the natural, dominant mode of operation. Yet, actual evolutionary biology has moved far beyond this caricature, identifying a wide range of different and equally successful strategies in evolution alongside competition.407

These include symbiosis (an example of which is the bacteria which fix nitrogen in plant roots consequently making life possible), collaboration (as was the case with primeval slime mould), co-evolution (the pollinating honey bee responsible for about one in three mouthfuls of the food we eat), and even reason (as with problem solving animals - like elephants, dogs, cats, rats, sperm whales and, sometimes, humans).

 

Optimal diversity too is considered a key condition - nature’s insurance policy against disaster - suggesting that economic systems which allow clone towns to be dominated by massive global chain stores, are probably a bad idea.

Mill also prefigured Keynes’s hope, and similar faith in technology, that once the ‘economic problem’ was solved, we would all be able to turn to more satisfying pursuits, and put our feet up more.

 

He also prepared the ground for the emergence of ecological economics.
 


The Steady state
In a fairly direct line of intellectual descent, economist Herman Daly has done perhaps more than anyone to popularize the notion of what he calls ‘steady state’ economics.

 

His comprehensive critique, worked-up over decades, decries the absence of any notion of optimal scale in macro-economics, and the persistent, more general refusal of the economics profession to accept that it, too, like the rest of life on the planet, is bound by the laws of physics (see Introduction).

As he wrote in Beyond Growth:

‘Since the earth itself is developing without growing, it follows that a subsystem of the earth (the economy) must eventually conform to the same behavioral mode of development without growth.’408

Of course the big question concerns when, precisely, the ‘eventually’ moment comes. Daly borrows a public safety analogy from the shipping industry to demonstrate what is needed ecologically at the planetary level.

The introduction of the ‘Plimsoll line’ was, so to speak, a watershed to do with a watermark. When a boat is too full, rather obviously it is more likely to sink. The problem used to be that, without any clear warning that a safe maximum carrying capacity had been reached, there was always an economic incentive to err on the incautious side by overfilling.

 

The Plimsoll line solved the problem with elegant simplicity: a mark painted on the outside of the hull that indicates a maximum load once level with the water.

Daly’s challenge to economics is to adopt or design an equivalent, ‘To keep the weight, the absolute scale, of the economy from sinking our biospheric ark’.409 But Daly is not a crude environmental determinist; for any model to work he insists that alongside optimal scale, equally important is a mechanism for optimal distribution based on equity and sufficiency.

To date, the nearest, in fact, only, leading contender to provide the environmental Plimsoll line is the Ecological Footprint. Before the Contraction and Convergence model, which is designed to manage safely greenhouse gas emissions, was ever thought of, Daly identified its basic mechanism as the way to manage the global environmental commons.

 

First, he said, you need to identify the limit of whichever aspect of our natural resources and biocapacity concerns you, then within that, allocate equitable entitlements and, in order to allow flexibility, make them tradable. Such an approach could be applied to the management of the world’s forests and oceans as much as CO2.

 

Daly credits the innovative American architect and polymath Richard Buckminster Fuller for first suggesting the approach. At a fundamental level, this is the primary mechanism to avoid the tragedy of the commons.

In addition, an indicator such as the Happy Planet Index 410 which incorporates the Ecological Footprint helps to reveal the degree of efficiency with which precious natural resources are converted into the meaningful human outcomes of long and happy lives.

At the ‘eventually’ moment, or rather well before, these other ways of organizing and measuring the economy become vital. In one sense it has already passed. According to the Ecological Footprint, the world has been over-burdening its biocapacity - consuming too many natural resources and producing more waste than can be safely absorbed - since the mid-1980s.

 

We’ve been living beyond our ecological means. But, at what point does the damage become irreversible? This will be different for different ecosystems. But, where climate change is concerned, we have drawn a line in the atmospheric sand at the end of 2016.

 

Based on current trends and several conservative assumptions, at that point, greenhouse gas concentrations will begin to push a new, more perilous phase of global warming.411
 


Dynamic equilibrium
‘Stationary’, ‘steady’, up to a point these words communicate the message that, logically, a subset of a system (the economy) cannot outgrow the system itself (the planet), and the need to establish a balance. Why suggest yet another term for an essential characteristic of true sustainability?

Yet, the terms ‘stationary’, and ‘steady’, are unattractive for our purposes. They fail to capture sufficiently the dynamism of the interactions between human society, the economy and the biosphere. They wrongly appear to suggest for economics, what was once famously, and with epic error announced for history, namely its end.

But, on the contrary, writes Daly, it is just that a very different economics is needed, one that is,

‘a subtle and complex economics of maintenance, qualitative improvements, sharing frugality, and adaptation to natural limits, It is an economics of better, not bigger’.412

‘Dynamic equilibrium’, is both a more accurate description of the condition we have to find and manage, and a more attractive term.

 

Found typically in discussions of population biology and forest ecology, it captures a mirror of nature for society, in which, within ecosystem limits, there is constant change, shifting balances and, evolution. ‘Dynamic’ in the sense that little is steady or stationary, but ‘equilibrium’ in that the vibrant, chaotic kerfuffle of life, economics and society must organize its affairs within the parent-company boundaries of available biocapacity.

In his parting address from the World Bank, where he worked for six years, Daly left his colleagues with a formula for sustainability:

  • stop counting the consumption of natural capital as income

  • tax labour and income less, and resource extraction

  • maximize the productivity of natural capital in the short run and invest in increasing its supply in the long run

  • most contentiously, abandon the ideology of global economic integration through free trade, free capital mobility, and export-led growth

nef’s report, The Great Transition, explores how best to organize an economy that exists in a state of dynamic equilibrium with the biosphere.

 

That and other research underway seeks to address all the usual questions such as ensuring livelihoods, security in youth and old age, maximizing well-being and social justice.

 

The point of this report has been simply to establish the case, as far as possible beyond question, that such an economy is needed.
 


The challenge: How to create good lives and flourishing societies that do not rely on infinite orthodox growth
This report set out to examine the physical and environmental constraints to unlimited global economic growth as measured by GDP.

 

Taking climate change and fossil fuel use as a particular focus, we find that these constraints at the global level are real and immediate. This means, that in order to allow economic growth in low per capita income countries where, for example, rising income has a strong relationship to greater life expectancy, there will need to be less growth in those high-income countries where the relationship to increasing life expectancy and satisfaction has already broken down.

It is not the purpose of this report to explore in detail what the latter might look like in practice. This is the focus of a large amount of work by nef that is unnecessary for us to duplicate.

 

We refer the reader, for example, to the book produced by nef and the Open University, titled Do Good Lives Have to Cost the Earth?, and to recent nef reports including:

  • The Happy Planet 2.0 (2009), which provides a new compass to set society on the path to real progress by measuring what matters to people - living a long and happy life - and what matters to the planet - our rate of resource consumption.

  • The National Accounts of Well-Being (2009), proposes nations should directly measure people’s well-being in a regular and thorough way, and that policy is shaped to ensure high, equitable and sustainable well-being.

  • The Great Transition (2009), which is a bold and broad plan for the UK that demonstrates how, even with declining GDP, it is possible to see rises in both social and environmental value. The plan envisages a pathway of rapid decarbonisation for the economy and significant increases in equality in society.

It is possible, though, to say something briefly here about why the things that lock economies like the UK into GDP growth are not immutable. In Box 1 at the beginning of the report we summarized those reasons as being mainly threefold.

First, governments plan their expenditure assuming that the economy will keep growing. If it then didn’t, there would be shortfalls in government income with repercussions for public spending. Secondly, listed companies are legally obliged to maximize returns to shareholders, and investors generally take their money wherever the highest rates of return and growth are found. Thirdly, nearly all money is lent into existence bearing interest. For every pound lent, more must be repaid, demanding growth.

Encouragingly, however, none of these three conditions is a given, unchangeable ‘state of nature’.

 

Economic rules and habits are not like the laws of physics. Today’s fiduciary duties on company management are not on a par with the force of gravity. These things are the result of cultural and political choices, which can, if necessary, be changed in the light of necessary and urgent circumstances.

In terms of government spending on essential services, governments have more room for maneuver than they like to admit. When the financial crisis hit, in the UK alone over £1 trillion was found to support the banks, apparently from nowhere. It can be done.

 

Through so-called ‘quantitative easing’ money really was conjured from thin air (the dirty little secret of banking is that this is practically what happens all the time when people borrow, for example, to buy a house).

Governments can also change priorities, spending less on unproductive military expenditure and more on schools, hospitals and support for those who need care. New techniques employing greater reciprocity with the users of public services can also radically reduce the upfront cash-cost of services by making them more effective (through so-called ‘co-production’).

 

There’s also no reason why fairer taxation and greater redistribution, coupled with better services cannot provide security for all in old age, removing the insecurity that makes us all worry about having a private pension with a high interest rate.

Herman Daly makes the point that in a non-growing, steady state (or dynamic equilibrium) economy it might actually be easier to approach full employment. With lower levels of material throughput and lower levels of fossil fuel energy use, the proportion of human energy input (labour) is likely to increase. Generations of having people made redundant by machines largely powered by coal, oil and gas could be reversed.

 

He writes:

‘There are several reasons for believing that full employment will be easier to attain in a SSE [steady state economy] than in our failing growth economies… the policy of limiting the matter-energy throughput would raise the price of energy and resources relative to the price of labour. This would lead to the substitution of labor for energy in production processes and consumption patterns, thus reversing the historical trend of replacing labour with machines and inanimate energy, whose relative prices have been declining.’413

Such a new economy implies the need for a great ‘re-skilling’, for example in the food economy, and the growth of urban agriculture.

 

Other adaptations could bring a range of social, environmental, and economic benefits. A redistribution of paid employment via a shorter working week, tackling the twin problems of overwork and unemployment, would free up time for people to do more things for themselves, each other and the community, and reduce their dependence on paid-for services.

At the corporate level, there are many other forms of governance that could reduce or remove the pressure to service shareholders who have a one-eyed obsession with maximum growth and returns.

 

Cooperatives, mutuals, publicly owned companies and social enterprises all have broader or simply different objectives.

Finally, when it comes to monetary systems, there is a whole world of alternatives, and a long history of innovation, some of it explored in the Green New Deal, published by nef in 2008, and widely written about in the works of people like Bernard Lietaer, David Boyle, Ann Pettifor and James Robertson.414

 

There are different forms of exchange, such as Time Banks,415 and different kinds of local and regional currencies, each with their own characteristics. Not all money need be interest bearing.

 

Low- or no-cost credit can be created by Central Banks for the purpose of achieving particular tasks - such as building new infrastructures for energy, transport, farming and buildings - for the environmental transformation of the economy. Such money can have special conditions attached to prevent it becoming inflationary.

Unending global economic growth, it would seem therefore is not possible, but also neither desirable nor necessary. If you have any doubts, ask a hamster.

 

Back to Contents
 


 


Video


What the impossible hamster has to teach us about economic growth. A new animation from nef (the new economics foundation), scripted by Andrew Simms, numbers crunched by Viki Johnson and pictures realized by Leo Murray.

We wanted to confront people with the meaning and logical conclusion of the promise of endless economic growth.

 

We used a hamster to illustrate what would happen if there were no limits to growth because they double in size each week before reaching maturity at around 6 weeks. But if a hamster grew at the same rate until its first birthday, wed be looking at a nine billion tonne hamster, which ate more than a years worth of world maize production every day.

 

There are reasons in nature, why things don't grow indefinitely. As things are in nature, sooner or later, so they must be in the economy. As economic growth rises, we are pushing the planet ever closer to, and beyond some very real environmental limits.

 

With every doubling in the global economy we use the equivalent in resources of all of the previous doublings combined.

 


The Impossible Hamster
by onehundredmonths
January 24, 2010

from YouTube Website

 

 

 

Back to Contents