Back to Results

EFTA00625129.pdf

Source: DOJ_DS9  •  Size: 29135.7 KB  •  OCR Confidence: 85.0%
PDF Source (No Download)

Extracted Text (OCR)

Why Money Trickles Up Geoff Willis gwillis@econodynamics.org The right of Geoffrey Michael Willis to be identified as the author of this work has been asserted by him in accordance with the Copyright, Designs and Patents Act 1988. EFTA00625129 0.0 Abstract This paper combines ideas from classical economics and modern finance with Lotka-Volterra models, and also the general Lotka-Volterra models of Levy & Solomon to provide straightforward explanations of a number of economic phenomena. Using a simple and realistic economic formulation, the distributions of both wealth and income are fully explained. Both the power tail and the log-normal like body are fully captured. It is of note that the full distribution, including the power law tail, is created via the use of absolutely identical agents. It is further demonstrated that a simple scheme of compulsory saving could eliminate poverty at little cost to the taxpayer. Such a scheme is discussed in detail and shown to be practical. Using similar simple techniques, a second model of corporate earnings is constructed that produces a power law distribution of company size by capitalisation. A third model is produced to model the prices of commodities such as copper. Including a delay to capital installation; normal for capital intensive industries, produces the typical cycle of short- term spikes and collapses seen in commodity prices. The fourth model combines ideas from the first three models to produce a simple Lotka-Volterra macroeconomic model. This basic model generates endogenous boom and bust business cycles of the sort described by Minsky and Austrian economists. From this model an exact formula for the Bowley ratio; the ratio of returns to labour to total returns, is derived. This formula is also derived trivially algebraically. This derivation is extended to a model including debt, and it suggests that excessive debt can be economically dangerous and also directly increases income inequality. Other models are proposed with financial and non-financial sectors and also two economies trading with each other. There is a brief discussion of the role of the state and monetary systems in such economies. The second part of the paper discusses the various background theoretical ideas on which the models are built. This includes a discussion of the mathematics of chaotic systems, statistical mechanical systems, and systems in a dynamic equilibrium of maximum entropy production. There is discussion of the concept of intrinsic value, and why it holds despite the apparent substantial changes of prices in real life economies. In particular there are discussions of the roles of liquidity and parallels in the fields of market-microstructure and post-Keynesian pricing theory. 2 EFTA00625130 0. Zeroth Section 0.0 Abstract 0.1 Contents 0.2 Introduction 0.3 Structure of Paper Part A — Some Models Part A.I — Heavy Duty Models 1. Wealth & Income Models 1.1 Wealth & Income Data — Empirical Information 1.2 Lotka-Volterra and General Lotka-Volterra Systems 1.3 Wealth & Income Models - Modelling 1.4 Wealth & Income Models - Results 1.5 Wealth & Income Models - Discussion 1.6 Enter Sir Bowley - Labour and Capital 1.7 Modifying Wealth and Income Distributions 1.8 A Virtual 40 Acres 1.9 Wealth & Income Distributions - Loose Ends 2. Companies Models 2.1 Companies Models - 2.2 Companies Models - 2.3 Companies Models - 2.4 Companies Models - 3. Commodity models 3.1 Commodity models 3.2 Commodity models 3.3 Commodity models 3.4 Commodity models Background Modelling Results Discussion - Background - Modelling - Results - Discussion 4. Minsky goes Austrian a la Goodwin — Macroeconomic Models 4.1 Macroeconomic Models - Background 4.2 Macroeconomic Models - Modelling 4.3 Macroeconomic Models - Results 4.4 Macroeconomic Models - Discussion 4.5 A Present for Philip Mirowski? — A Bowley-Polonius Macroeconomic Model EFTA00625131 Part A.II - Speculative Building 4.6 Unconstrained Bowley Macroeconomic Models 4.7 A State of Grace 4.8 Nirvana Postponed 4.9 Bowley Squared 4.10 Siamese Bowley - Mutual Suicide Pacts 4.11 Where Angels Fear to Tread - Governments & Money 4.12 Why Money Trickles Up Part B - Some Theory 5. Theory Introduction Part B.I — Mathematics 6. Dynamics 6.1 Drive My Car 6.2 Counting the Bodies - Mathematics and Equilibrium 6.3 Chaos in Practice — Housing in the UK 6.4 Low Frequency / Tobin Trading 6.5 Ending the Chaos 7. Entropy 7.1 Many Body Mathematics 7.2 Statistical Mechanics and Entropy 7.3 Maximum Entropy Production 7.4 The Statistical Mechanics of Flow Systems Part &II — Economic Foundations 8. Value 8.1 The Source of Value 8.2 On the Conservation of Value 8.2.1 Liquidity 8.2.2 On the Price of Shares 9. Supply and Demand 9.1 Pricing 9.2 An Aside on Continuous Double Auctions 9.3 Supply — On the Scarcity of Scarcity, or the Production of Machines by Means of Machines EFTA00625132 9.4 Demand Part &III — The Logic of Science 10. The Social Architecture of Capitalism 11. The Logic of Science Part C — Appendices 12. History and Acknowledgements 13. Further Reading 14. Programmes 15. References 16. Figures 5 EFTA00625133 0.2 Introduction "The produce of the earth -- all that is derived from its surface by the united application of labour, machinery, and capital, is divided among three classes of the community; namely, the proprietor of the land, the owner of the stock or capital necessary for its cultivation, and the labourers by whose industry it is cultivated To determine the laws which regulate this distribution, is the principal problem in Political Economy..." On The Principles of Political Economy and Taxation - David Ricardo [Ricardo 1817] "We began with an assertion that economic inequality is a persistent and pressing problem; this assertion may be regarded by many people as tendentious. Differences in economic status - it might be argued - are a fact of life; they are no more a 'problem' than are biological differences amongst people, or within and amongst other species for that matter. Furthermore, some economists and social philosophers see economic inequality, along with unfettered competition, as essential parts of a mechanism that provides the best prospects for continuous economic progress and the eventual elimination of poverty throughout the world. These arguments will not do. There are several reasons why they will not do However there is a more basic but powerful reason for rejecting the argument that dismisses economic inequality as part of the natural order of things. This has to do with the scale and structure of inequality " Economic Inequality and Income Distribution — DG Champernowne [Champernowne & Cowell 1998] "Few if any economists seem to have realized the possibilities that such invariants hold for the future of our science. In particular, nobody seems to have realized that the hunt for, and the interpretation of, invariants of this type might lay the foundations for an entirely novel type of theory." Schumpeter (1949, p. 155), discussing the Pareto law — via [Gabaix 2009] This paper introduces some mathematical and simulation models and supports these models with various theoretical ideas from economics, mathematics, physics and ecology. The models use basic economic variables to give straightforward explanations of the distributions of wealth, income and company sizes in human societies. The models also explain the source of macroeconomic business cycles, including bubble and crash behaviour. The models give simple formulae for wealth distributions, and also for the Bowley ratio; the ratio of returns to labour and capital. Usefully, the models also provide simple effective methods for eliminating poverty without using tax and welfare. The theoretical ideas provide a framework for extending this modelling approach systematically across economics. 6 EFTA00625134 The models were produced firstly by taking basic ideas from classical economics and basic finance. These ideas where then combined with the mathematics of chaotic systems and dynamic statistical mechanics, in a process that I think can be well summed up as 'econodynamics' as it parallels the approaches of thermodynamics, and ultimately demonstrates that economics is in fact a subset of thermodynamics. This makes the process sound planned. It wasn't. It was a process of common sense and good luck combined with a lot of background reading. It was suggested to me in 2006 That the generalised Lotka-Volterra (GLV) distribution might provide a good fit for income data. The suggestion proved to be prescient. The fit to real data proved to be better than that for other previously proposed distributions. At this point, in 2006, I used my limited knowledge of economics to propose two alternative models that might fit the simplest economically appropriate terms into two different generating equations that produce the (GLV). I passed these ideas forward to a number of physicists. The history of this is expanded in section 12. After that, nothing very much happened for three years. This was for three main reasons. Firstly, I didn't understand the detailed mathematics, or indeed have a strong feel for the generalised Lotka-Volterra model. Secondly, my computer programming, and modelling skills are woeful. Thirdly, the academics that I wrote to had no interest in my ideas. In 2009/2010 I was able to make progress on the first two items above, and in early 2010 I was able, with assistance from George Vogiatzis and Maria Chli, to produce a GLV distribution of wealth from a simulation programme with just nine lines of code, that included only a population of identical individuals, and just the variables of individual wealth (or capital), a single uniform profit rate and a single uniform (but stochastic) consumption (or saving) rate. This simple model reproduced a complex reality with a parsimony found rarely even in pure physics. After a brief pause, the rest of the modelling, research and writing of this paper was carried out between the beginning of May 2010 and the end of March 2011. This was done in something of a rush, without financial support or academic assistance; and I would therefore ask forbearance for the rough and ready nature of the paper. From the first wealth-based model, and with greater knowledge of finance and economics; models for income, companies, commodities and finally macroeconomics dropped out naturally and straightforwardly. The models are certainly in need of more rigorous calibration, but they appear to work well. The wealth and income models appear to be powerful, both in their simplicity and universality, and also in their ability to advise future action for reducing poverty. The macroeconomic models are interesting, as even in these initial simple models, they give outcomes that accord closely with the qualitative descriptions of business and credit cycles in the work of Minsky and the Austrian school of economics. These descriptions describe well the actual behaviour of economies in bubbles and crashes from the Roman land speculation of 33AD through tulipomania and the South Sea bubble up to the recent credit crunch. Part A of this paper goes through these various models in detail, discussing also the background and consequences of the models. 7 EFTA00625135 The agents in the initial models were identical, and painfully simple in their behaviour. They worked for money, saved some of their money, spent some of their money, and received interest on the money accumulated in their bank accounts. Because of this the agents had no utility or behavioural functions of the sort commonly used in agent-based economic modelling. As such the models had no initial underlying references to neoclassical economics, or for that matter behavioural economics. There simply was no need for neoclassicism or behaviouralism. As the modelling progressed, somewhat to my surprise, and, in fact to my embarrassment, it became clear that the models were modelling the economics of the classical economists; the economics of Smith, Ricardo, Marx, von Neumann (unmodified) and Sraffa. With hindsight this turned out to be a consequence of the second of the two original models I had proposed in 2006. In this model wealth is implicitly conserved in exchange, but created in production and destroyed in consumption. Ultimately total wealth is conserved in the long term. This model denies the premises of neoclassicism, and adopts an updated form of classical economics. Despite the rejection of neoclassicism, the models work. Classical economics works. Where the classical economists were undoubtedly wrong was in their belief in the labour theory of value. They were however absolutely correct in the belief that value was intrinsic, and embodied in the goods bought, sold and stored as units of wealth. Once intrinsic wealth, and so the conservation of wealth is recast and accepted, building economic models becomes surprisingly easy. The re-acceptance of intrinsic wealth; and so the abandonment of neoclassicism, is clearly controversial. Given the wild gyrations of the prices of shares, commodities, house prices, art works and other economic goods, it may also seem very silly. Because of this a significant section of part B of this paper discusses these issues in detail, and the economic and finance background in general. The other main aim of part B of this paper is to introduce the ideas of chaotic systems, statistical mechanics and entropy to those that are unfamiliar with them. Partly because of these theoretical discussions this paper is somewhat longer than I initially expected. This is mainly because I have aimed the paper at a much larger audience than is normal for an academic paper. In my experience there are many people with a basic mathematical background, both inside and outside academia, who are interested in economics. This includes engineers, biologists and chemists as well as physicists and mathematicians. I have therefore written the paper at a level that should be relatively easy to follow for those with first year undergraduate mathematics (or the equivalent of a UK A-level in maths). Although the numbers are much smaller, I believe there is also a significant minority of economists, especially younger economists, who are acutely aware that the theory and mathematical tools of economics are simply not adequate for modelling real world economies. This paper is also aimed at these economists. 8 EFTA00625136 I would not be particularly surprised if every single model in this paper has to be reworked to make them describe real world economies. It may even be the case that many of the models have to be superseded. This would be annoying but not tragic, but is beside the point. The main point of this paper is the power of the mathematical tools. The two main tools used in this paper are chaotic differential equation systems and statistical mechanics. In both cases these tools are used in systems that are away from what are normally considered equilibrium positions. It is these tools that allow the production of simple effective economic models, and it is these tools that economists need in order to make progress. Comparative statics may be intellectually satisfying and neat to draw on a blackboard, but it doesn't work in dynamic multi-body systems. For a dynamic system you need dynamic differential equation models. For systems with large numbers of interacting bodies you need statistical mechanics and entropy. Although a minority of economists have toyed with chaos theory, and many economists claim to use 'dynamic' models, I have only encountered one economist; Steve Keen, who truly 'gets' dynamic modelling in the way that most physicists, engineers and mathematical modellers use dynamic modelling. Indeed the macroeconomic model in this paper shares many ideas with, and certainly the approaches of, Steve Keen who has used dynamical mathematical models to follow the ideas of Goodwin, Minsky and others; and who has used the Lotka-Volterra dynamics in particular. Although Keen's models are certainly heterodox he is almost unique in being an economic theoretician who predicted the credit crunch accurately and in depth. While other economists predicted the credit crunch, almost all the others who did so did this from an analysis of repeating patterns of economic history. That is, they could spot a bubble when they saw one. Steve Keen is unusual in being a theoretical economist who is able to model bubbles with a degree of precision. The use of statistical mechanics in economics is even more frustrating. Merton, Black and Scholes cherry-picked the diffusion equation from thermodynamics while completely ignoring its statistical mechanical roots and derivation. They then sledge-hammered it into working in a neoclassical framework. Tragically, a couple of generations of physicists working in finance have not only accepted this, but they have built more and more baroque models on these flimsy foundations. The trouble with Black-Scholes is that it works very well, except when it doesn't. This basic flaw has been pointed out from Mandlebrot onwards, to date with no notice taken. This is most frustrating. If physicists were doing their jobs properly, finance would be one of the simplest most boring parts of economics. The only economist I have encountered who truly 'gets' statistical mechanics is Duncan Foley. He is uniquely an economist who has fully realised not only the faults with the mathematics used by most economists, but also dedicated considerable effort to applying the correct mathematics, statistical mechanics, to economics. Although primarily modelled in a static environment, Foley's work is profoundly insightful, and demonstrates very clearly how statistical mechanical approaches are more powerful than utility based approaches, and how statistical mechanics approaches naturally lead to the market failures seen in real economies. Despite this visionary insight he has ploughed a somewhat lonely furrow, with the relevant work largely ignored by economists, and more embarrassingly also by physicists. 9 EFTA00625137 Because chaos and statistical mechanics are unfamiliar in economics, I have spent some effort in both the modelling sections and the theory sections in explaining how the models work in detail, how these concepts work in general, and why these mathematical approaches are not just relevant but essential for building mathematical models in economics. This extra explanation for less mathematical scientists and economists may mean that the paper is over-explained and repetitive for many physicists and mathematicians. For this I can only offer my apologies. However, even for physicists some of the background material in the discussions on entropy contains novel and powerful ideas regarding non-equilibrium thermodynamic systems. This is taken from recent work in the physics of planetary ecology and appears not to have percolated into the general physics community despite appearing to have general applicability. The ideas of Paltridge, Lorenz, Dewar and others, along with the mathematical techniques of Levy & Solomon, may not be familiar to many physicists, and I believe may be very powerful in the analysis of complex 'out of equilibrium' systems in general. In fact, although I was trained as a physicist, I am not much of a mathematician, and by emotional inclination I am more of an engineer. My skills lie mostly in seeing connections between different existing ideas and being able to bolt them together in effective and sometimes simpler ways. Part of the reason for the length of this paper is that I have taken a lot of ideas from a lot of different fields, mainly from classical economics, finance, physics, mathematics and ecology, and fitted them together in new ways. I wish to explain this bolting together in detail, partly because very few people will be familiar with all the bits I have cherry-picked, but also I suspect that my initial bolting together may be less than ideal, and may need reworking and improving. I feel I should also apologise in advance for a certain amount of impatience displayed in my writing towards traditional economics. From an economics point of view the paper gets more controversial as it goes along. It also gets increasingly less polite with regard to the theories of neoclassical economics. In the last two years I have read a lot of economics and finance, a significant proportion of which was not profoundly insightful. Unfortunately, reading standard economics books to find out how real economies work is a little like reading astrology books to find out how planetary systems work. Generally I have found the most useful economic ideas in finance or heterodox economics, areas which are not usually well known to physicists, or indeed many economists. These ideas include recent research in market microstructure, liquidity, post-Keynesian pricing theory as well as the work of Foley, Keen, Smithers, Shiller, Cooper, Pettis, Pepper & Oliver, Mehrling, Lyons and others. Neoclassical economics, while forming an intellectually beautiful framework, has proved of limited use to me as a source of knowledge. Partly this is because the mathematics used, comparative statics, is simply inappropriate. Partly it is because some of the core suppositions used to build the framework; such as diminishing returns and the importance of investment and saving, are trivially refutable. 10 EFTA00625138 The only defence I can make for my impoliteness is a very poor one; that I am considerably more polite than others. If any of my comments regarding neoclassical economics cause offence, I advise you to read the work of Steve Keen and Phillip Mirowski with some caution. Both are trained economists who have the mathematical and historical skills to realise the inappropriateness of neoclassicism. Their writing has the polemical edge of a once devout Christian who has recently discovered that the parish priest has been in an intimate liaison with his wife for the last fifteen years. Finally I would like to comment on the work of Ian Wright, Makoto Nirei & Wataru Souma and others. Throughout this paper comparisons are made to the work of Ian Wright who describes simulated economic models in two notable papers [Wright 2005, 2009]. Wright's models are significantly different to my own, most notably in not involving a financial sector. Also, unlike the present paper, Wright takes a 'black box' and 'zero intelligence' approach to modelling which eschews formal fitting of the models to mathematical equations. Despite these profound differences, at a deeper level Wright's models share fundamental similarities with my own, sharing the basic conservation of value of the classical economists, as well as using a dynamic, stochastic, statistical mechanical approach. More significantly, the models are striking in the similarities of their outputs to my own work. Also it is important to note that Wright's models have a richness in some areas, such as unemployment which are missing from my own models. In relevant sections I discuss detailed differences and similarities between the models of Wright and myself. In two papers Souma & Nirei [Souma & Nirei 2005, 2007] build a highly mathematical model that produces a power tail and an exponential distribution for income. Their approach also builds ultimately on the work of Solomon & Levy. However their approach is substantially more complex than my own. Their models do however share a number of similarities to my own models. Firstly, the models of Souma & Nirei use consumption as the negative balancing term in their model in a manner almost identical to the role of consumption in my own model. Secondly, their models ascribe a strong positive economic role to capital as a source of wealth, however this is ascribed to the process of capital growth, not the dividends, interest, rent, etc that is used in my own models. Both Wright's work and that of Souma & Nirei predate this paper. Their work also predates my original models produced in 2006. Given the process by which I came to produce the models below, I believe I did so independently of Wright, Souma & Nirei. However, I would be very foolish to discount that possibility that I was subconsciously influenced by these authors, and so I do not discount this. It is certainly clear to me that Wright, Souma & Nirei have made very substantial inroads in the same directions as my own research, and that if I had not had lucky breaks in advancing my own research, then one or other of them would have produced the models below within the near future. Given that the work of Wright, Souma & Nirei predates my own, and so gives rise to questions of originality, I have included a brief history of the gestation of the present paper in section 12, History and Acknowledgements. With regard to precedence, I would like to note that the general approach for the macroeconomic models in section 4 were partly inspired by the work of Steve Keen, though the 11 EFTA00625139 models themselves grew straight out of my company and commodity models; and ultimately out of my income models. More importantly, not a word of this paper would have been written without the work of Levy & Solomon and their GLV models. Manipulation of the GLV is beyond my mathematical ability. Although Levy & Solomon's economic explanations are naive, their gut feeling of the applicability of the GLV to economics in particular, and complex systems in general, was correct. I believe their work is of profound general importance. In later sections of this paper I quote extensively from the work of Ian Wright, Duncan Foley and Steve Keen, as their explanations of the importance of statistical mechanics and chaos in economics are difficult to improve on. 0.3 Structure of the Paper Part A of this paper discusses a number of economic models in detail, Part A.I discusses a number of straightforward models giving results that easily accord with the real world and also with the models of Ian Wright. Part A.II discusses models that are more speculative. Part B discusses the background mathematics, physics and economics underlying the models in Part A. The mathematics and physics is discussed in Part B.I, the economics in part B.II, the conclusions are in part B.III. Finally, Part C gives appendices. Within Part A; section 1 discusses income and wealth distributions; section 1.1 gives a brief review of empirical information known about wealth and income distributions while section 1.2 gives background information on the Lotka-Volterra and General Lotka-Volterra models. Sections 1.3 to 1.5 gives details of the models, their outputs and a discussion of these outputs. Section 1.6 discusses the effects that changing the ratio of waged income to earnings from capital has on wealth and income distributions. Sections 1.7 and 1.8 discuss effective, low-cost options for modifying wealth and income distributions and so eliminating poverty. Finally, section 1.9 looks at some unexplained but potentially important issues within wealth and income distribution. Sections 2.1 to 2.4 go through the background, creation and discussion of a model that creates power law distributions in company sizes. Sections 3.1 to 3.4 use ideas from section 2, and also the consequences of the delays inherent in installing physical capital, to generate the cyclical spiking behaviour typical of commodity prices. Sections 4.1 to 4.4 combine the ideas from sections 1, 2 and 3 to provide a basic macroeconomic model of a full, isolated economy. It is demonstrated that even a very basic model can endogenously generate cyclical boom and bust business cycles of the sort described by Minsky and Austrian economists. 12 EFTA00625140 In section 4.5 it is demonstrated that an exact formulation for the Bowley ratio; the ratio of returns to labour to total returns, can easily be derived from the basic macroeconomic model above, or indeed from first principles in a few lines of basic algebra. In section 4.6 and 4.7 the above modelling is extended into an economy with debt. From this a more complex, though still simple, formulation for the Bowley ratio is derived. This formulation suggests that excessive debt can be economically dangerous and also directly increases income inequality. The more general consequences of the Bowley ratio for society are discussed in more depth in section 4.8. In section 4.9 two macroeconomic models are arranged in tandem to discuss an isolated economy with a financial sector in addition to an ordinary non-financial sector. In section 4.10 two macroeconomic models are discussed in parallel as a model of two national economies trading with each other. To conclude Part A, section 4.11 introduces the role of the state and monetary economics, while section 4.12 briefly reviews the salient outcomes of the modelling for social equity. In Part B, section 6.1 discusses the differences between static and dynamic systems, while section 6.2 looks at the chaotic mathematics of differential equation systems. Examples of how this knowledge could be applied to housing markets is discussed in section 6.3, while applications to share markets are discussed in section 6.4. A general overview of the control of chaotic systems is given in section 6.5. Section 7.1 discusses the theory; 'statistical mechanics', which is necessary for applying to situations with many independent bodies; while section 7.2 discusses how this leads to the concept of entropy. Section 7.3 discusses how systems normally considered to be out of equilibrium can in fact be considered to be in a dynamic equilibrium that is characterised as being in a state of maximum entropy production. Section 7.4 discusses possible ways that the statistical mechanics of maximum entropy production systems might be tackled. Moving back to economics; in section 8.1 it is discussed how an intrinsic measure of value can be related to the entropy discussed in section 7 via the concept of 'humanly useful negentropy'. Section 8.2 discusses the many serious criticisms of a concept of intrinsic value in general, with a discussion of the role of liquidity in particular. Section 9.1 looks at theories of supply and pricing, the non-existence of diminishing returns in production, and the similarities between the market-microstructure analysis and post-Keynesian pricing theory. Section 9.3 looks for, and fails to find, sources of scarcity, while section 9.4 discusses the characteristics of demand. In section 10 both the theory and modelling is reviewed and arranged together as a coherent whole, this is followed by brief conclusions in section 11. Sections 12 to 16 are appendices in Part C. Section 12 gives a history of the gestation of this paper and an opportunity to thank those that have assisted in its formation. 13 EFTA00625141 Section 13 gives a reading list for those interested in learning more about the background maths and economics in the paper. Section 14 gives details of the Matlab and Excel programmes used to generate the models in Part A of the paper. Sections 15 and 16 give the references and figures respectively. 14 EFTA00625142 Part A — Some Models Section A.I — Heavy Duty Models 1. Wealth & Income Models 1.1 Wealth & Income Data — Empirical Information "Endogeneity of distribution Neoclassical economics approaches the problem of distribution by positing a given and exogenous distribution of ownership of resources. The competitive market equilibrium then determines the relative value of each agent's endowment (essentially as rents). I think there are problems looming up with this aspect of theory as well. One reason to doubt the durability of the assumption of an exogenous distribution of ownership of resources is that income and wealth distributions exhibit empirical regularities that are as stable as any other economic relationships. I think there is an important scientific payoff in models that explain the size distributions of wealth and income as endogenous outcomes of market interactions." Duncan K. Foley [Foley 1990] Within theoretical economics, the study of income and wealth distributions is something of a backwater. As stated by Foley above, neo-classical economics starts from given exogenous distributions of wealth and then looks at the ensuing exchange processes. Utility theory assumes that entrepreneurs and labourers are fairly rewarded for their efforts and risk appetite. The search for deeper endogenous explanations within mainstream economics has been minimal. This is puzzling, because, as Foley states, it has been clear for a century that income distributions show very fixed uniformities. Vilfredo Pareto first showed in 1896 that income distributions followed the power law distribution that now bears his name [Pareto 1896]. Pareto studied income in Britain, Prussia, Saxony, Ireland, Italy and Peru. At the time of his study Britain and Prussia were strongly industrialised countries, while Ireland, Italy and Peru were still agricultural producers. Despite the differences between these economies, Pareto discovered that the income of wealthy individuals varied as a power law in all cases. Extensive research since has shown that this relationship is universal across all countries, and that not only is a power law present for high income individuals, but the gradient of the power law is similar in all the different countries. Typical graphs of income distribution are shown below. This is data for 2002 from the UK, and is an unusually good data set [ONS 2003]. Figure 1.1.1 here 15 EFTA00625143 Figure 1.1.1 above shows a probability density function. A probability distribution function (pdf) is basically a glorified histogram or bar chart. Along the x-axis are bands of wage. The y-axis shows the number of people in each wage band. As can be seen this shape has a large bulge towards the left-hand side, with a peak at about £300 per week. To the right hand side there is a long tail showing smaller and smaller numbers of people with higher and higher earnings. Also included in this chart is a log-normal distribution fitted to the curve. The log-normal distribution is the curve that economists normally fit to income distributions (or pretty much anything else that catches their attention). On these scales the log-normal appears to give a very good fit to the data. However there are problems with this. Figure 1.1.2 here Figure 1.1.2 above shows the same data, but this time with the y-axis transformed into a log scale. Although the log-normal gives a very good fit for the first two thirds of the graph, somewhere around a weekly wage level of £900 the data points move off higher than the log- normal fit. The log-normal fit cannot describe the income of high-earners well. Figure 1.1.3 here Figure 1.1.3 above shows the same data but organised in a different manner. This is a 'cumulative density function' or cdf. In this graph the wealth is still plotted along the x-axis, but this time the x-axis is also a log scale. This time the y-axis shows the proportion of people who earn more than the wage on the x-axis. In figure 1.1.3 about 10% of people, a proportion of 0.1, earn more than £755 per week. It can be seen that the curve has a curved section on the left-hand side, and a straight line section on the right-hand side. This straight section is the 'power-tail' of the distribution. This section of the data obeys a 'power-law' as described by Pareto 100 years ago. The work of Pareto gives a remarkable result. An industrial manufacturing society and an agrarian society have very different economic systems and societal structures. Intuitively it seems reasonable to assume that income would be distributed differently in such different societies. What the data is saying is that none of the following have an effect on the shape of income distribution in a country: • Whether wealth is owned as industrial capital or agricultural land • Whether wealth is owned directly or via a stock market • What sort of education system a country has 16 EFTA00625144 • What sort of justice system a country has • Natural endowments of agricultural land or mineral wealth • And so on with many other social and economic factors Intuitively it seems reasonable that any or all of the above would affect income distribution, in practice none of them do. Income distributions are controlled by much deeper and basic processes in economics. The big unexpected conclusion from the data of Pareto and others is the existence of the power tail itself. Traditional economics holds that individuals are fairly rewarded for their abilities, a power tail distribution does not fit these assumptions. Human abilities are usually distributed normally, or sometimes log-normally. The earning ability of an individual human being is made up of the combination of many different personal skills. Logically, following the central limit theorem, it would be reasonable to expect that the distribution of income would be a normal or log-normal distribution. A power law distribution however is very much more skewed than even a log-normal distribution, so it is not obvious why individual skills should be overcompensated with a power law distribution. While Pareto noted the existence of a power tail in the distribution, it should be noted that more recently various authors have suggested that there may be two or even three power tail regions, with a separation between the 'rich' and 'super-rich', see for example [Borges 2002, Clementi & Gallegati 2005b, Souma, Nirei & Souma 2007]. While the income earned by the people in the power tail of income distribution may account for approximately 50% of total earnings, the Pareto distribution actually only applies to the top 10%-20% of earners. The other 80%-90% of middle class and poorer people are accounted for by a different 'body' of the distribution. Going back to the linear-linear graph in figure 1.1.1 it can be seen that, between incomes of £100 and £900 per week, there is a characteristic bulge or hump of individuals, with a skew in the hump towards the right hand side. In the days since Pareto the distribution of income for the main 80%-90% of individuals in this bulge has also been investigated in detail. The distribution of income for this main group of individuals shows the characteristic skewed humped shape similar to that of the log-normal distribution, though many other distributions have been proposed. These include the gamma, Weibull, beta, Singh-Maddala, and Dagum. The last two both being members of the Dagum family of distributions. Bandourian, McDonald & Turley [Bandourain et al 2002] give an extensive overview of all the above distributions, as well as other variations of the general beta class of distributions. They carry out a review of which of these distributions give best fits to the extensive data in the Luxembourg Income Study. In all they analyse the fit of eleven probability distributions to twenty-three different countries. They conclude that the Weibull, Dagum and general-beta2 distributions are the best fits to the data depending on the number of parameters used. 17 EFTA00625145 For more information, readers are referred to 'Statistical Size Distributions in Economics and Actuarial Sciences' [Kleiber & Kotz 2003] for a more general overview of probability distributions in economics, and also to Atkinson and Bourguignon [Atkinson & Bourguignon 2000] for a very detailed discussion of income data and theory in general. The author has analysed a particularly good set of income data from the UK tax system, one example is shown in figures 1.1.1-3 above. This data suggests that a Maxwell-Boltzmann distribution also provides a very good fit to the main body of the income data that is equal to that of the log-normal distribution [Willis & Mimkes 2005]. The reasons for the split between the income earned by the top 10% and the main body 90% has been studied in more detail by Clementi and Gallegati [Clementi & Gallegati 2005a] using data from the US, UK, Germany and Italy. This shows strong economic regularities in the data. In general it appears that the income gained by individuals in the power tail comes primarily from income gained from capital such as interest payments, dividends, rent or ownership of small businesses. Meanwhile the income for the 90% of people in the main body of the distribution is primarily derived from wages. These conclusions are important, and will be returned to in the models below. This view is supported, though only by suggestion, by one intriguing high quality income data set. This data set comes from the United States and is from a 1992 survey giving proportions of workers earning particular wages in manufacturing and service industries. The ultimate source of the data is the US Department of Labor; Bureau of Statistics, and so the provenance is believed to be of the good quality. Unfortunately, enquiries by the author has failed to reveal the details of the data, such as sample size and collection methodology. The data was collected to give a comparison of the relative quality of employment in the manufacturing and service sectors. Although the sample size for the data is not known, the smoothness of the curves produced suggest that the samples were large, and that the data is of good statistical quality. The data for services is shown in figures 1.1.4 & 1.1.5 below, the data for manufacturing is near identical. Figure 1.1.4 here Figure 1.1.5 here Like the UK data, there appears to be a clear linear section in the central portion of the data on a log-linear scale in figure 1.1.5, indicating an exponential section in the raw data. Again this data can be fitted equally well with a log-normal or a Maxwell-Boltzmann distribution. What is much more interesting is that, beyond this section, the data heads rapidly lower on the logarithmic scale. This means it is heading rapidly to zero on the raw data graph. With these two distributions there is no sign whatsoever of the 'power tail' that is normally found in income distributions. 18 EFTA00625146 It is the belief of the author that the methodology for this US survey restricted the data to 'earned' or 'waged' income, as the interest in the project was in looking at pay in services versus manufacturing industry. It is believed income from assets and investments was not included as this would have been irrelevant to the investigation. This US data set has been included for a further reason, a reason that is subtle; but in the belief of the author, important. Looking back at figure 1.1.1 for the UK income data, there is a very clear offset from zero along the income axis. That is the curve does not start to rise from the income axis until a value of roughly £100 weekly wage. The US data shows an exactly similar offset, with income not rising until a weekly wage of $100. This is important, as the various curves discussed above (log-normal, gamma, Weibull, beta, Singh-Maddala, Dagum, Maxwell-Boltzmann, etc) all normally start at the origin of the axis, point (0,0) with the curve rising immediately from this point. While it is straightforward enough to put an offset in, this is not normally necessary when looking at natural phenomena. In the 1930s Gibrat, an engineer, pioneered work in economics that studied work on proportional growth processes that could produce log-normal or power law distributions depending on the parameters. His work primarily looked at companies, and was the first attempt to apply stochastic processes to produce power law distributions. Following the work of Pareto, the details of income and wealth distributions have rarely been studied in mainstream theoretical economics, a notable and important exception being Champernowne. Champernowne was a highly gifted mathematician who was diverted into economics, he was the first person to bring a statistical mechanical approach to income distribution, and also noted the importance of capital as a major creator of inequality, though his approach concentrated on generational transfers of wealth [Champernowne & Cowell 1998]. Despite the lack of interest within economics, this area has had a profound attraction to those outside the economics profession for many years, a review of this history is provided by Gabaix [Gabaix 2009]. In recent years, the study of income distributions has gone through a small renaissance with new interest in the field shown by physicists with an interest in economics, and has become a significant element of the body of research known as 'econophysics'. Notable papers have been written in this field by Bouchaud & Mezard, Nirei & Souma, Dragulescu & Yakovenko, Chatterjee & Chakrabarti, Slanina, Sinha and many, many, others [Bouchaud & Mezard 2000, Dragulescu & Yakovenko 2001, Nirei & Souma 2007, Souma 2001, Slanina 2004, Sinha 2005]. The majority of these papers follow similar approaches; inherited either from the work of Gibrat, or from gas models in physics. Almost all the above models deal with basic exchange processes, with some sort of asymmetry introduced to produce a power tail. Chatterjee et al 2007, Chatterjee & Chakrabarti 2007 and Sinha 2005 give good reviews of this modelling approach. The approaches above have been the subject of some criticism, even by economists who are otherwise sympathetic to a stochastic approach to economics, but who are concerned that a 19 EFTA00625147 pure exchange process is not appropriate for modelling modern economies [Gallegati et al 2006]. An alternative approach to stochastic modelling has been taken by Moshe Levy, Sorin Solomon, and others [Levy & Solomon 1996]. They have produced work based on the 'General Lotka-Volterra' model. Unsurprisingly, this is a generalised framework of the 'predator-prey' models independently developed for the analysis of population dynamics in biology by two mathematicians/physicists Alfred Lotka and Vito Volterra. A full discussion of the origin and mathematics of GLV distributions is given below in section 1.2. These distributions are interesting for a number of reasons; these include the following: • the fundamental shape of the GLV curve • the quality of the fit to actual data • the appropriateness of the GLV distribution as an economic model Figure 1.1.6 here Figure 1.1.7 here With regard to the fundamental shape of the GLV curve, figures 1.1.6 and 1.1.7 above show plots of the UK income data against the GLV on a linear-linear and log-log plot. The formula for this distribution is given by: P(w) = K(e' r""•)/((w/L)"+"1) (1.1a) and it has three parameters; K is a general scaling parameter, L is a normalising constant for w, and a relates to the slope of the power tail of the distribution. It should firstly be noted that the GLV has both a power tail and a 'log-normal'-like main body. That is to say it can model both the main population and the high-end earners at the same time. This is a very significant advantage over other proposed distributions. The second and more subtle point to note is that the GLV has a 'natural' offset from zero. It is in the nature of the GLV that the rise from zero probability on the y-axis starts at a non-zero value on the x-axis, this is discussed further in section 1.2 Below. Finally the detailed fit of the GLV appears to be equivalent or better than the log-normal distribution. 20 EFTA00625148 Figure 1.1.8 Reduced Chi Squared Full Data Set Reduced Data Set Boltzmann Fit 3.27 1.94 Log Normal Fit 2.12 3.02 GLV Fit 1.21 1.83 Figure 1.1.8 above gives results from a basic statistical analysis using the GLV, log-normal and Maxwell-Boltzmann distributions. (The values in the table are the reduced chi-squared values, using an assumed standard measurement error of 100. The actual measurement error is not known, so the values above are not absolute, however, changing the measurement value will change the values in the table by equal proportions, so the relative sizes of the values in the table will stay the same.) It can be seen from the figures in the first column that the GLV, with the lowest value of chi- squared, gives the best fit. In itself this is not altogether surprising, as it is known that the log- normal and the Maxwell-Boltzmann have exponential tails, and so are not able to fit power tails. More remarkably, the figures in the second column show the same analysis carried out using a truncated data set with an upper limit of £800 per week. This limit was taken to deliberately exclude the data from the power tail. Again it can be seen that the GLV still just gives the best fit to the data. This in itself suggests that the GLV should be preferred to the log-normal or the Maxwell-Boltzmann distributions. It is also of note that in parallel to the work of Solomon et al, Slanina has also proposed an exchange model that produces the same output distribution as the GLV [Slanina 2004]. Unfortunately the modelling approaches of Solomon et al, and Slanina use economic models that are not wholly convincing, and as such have significant conceptual shortcomings. It is the belief of the author that an alternative economic analysis, using more appropriate analogies allows a much more effective use of GLV distributions in an intuitive and simple economic formulation. This is the third main reason for preferring the GLV distribution, and forms the key content of the initial sections of this paper. As previously noted Souma & Nirei have also pursued research in this direction. Before discussing the GLV distribution in detail I would firstly like to review some background on power law distributions. Power laws are deeply beloved of theoretical physicists, and there are many different ways to produce power laws. Most theoretical physicists tend to have a particular affection for their pet process and it's particular mathematical derivation, and then proceed to fit their pet equations to any model that happens to have a power tail with gay abandon. Also, as is usually necessary, this requires the sledgehammer of many pages of complex mathematical derivation, in an attempt to fit a square peg into a round hole. An unfortunate consequence of this is that most of 21 EFTA00625149 the very extensive literature on power laws is confusing, apparently conflicting, and to a great extent simply incoherent. This is a shame, as most power laws distributions are actually produced very simply, in a restricted number of ways. For those who want more background on the formation of power laws, log-normal laws and related processes, there are three very good background papers by Newman [Newman 2005], Mitzenmacher [Mitzenmacher 2004] and Simkin & Roychowdhury [Simkin & Roychowdhury 2006]. The papers by Newman and Mitzenmacher give very good overviews of what make power law and log-normal normal distributions without being mathematically complex. One basic point from the papers is that there are many different ways of producing power law distributions, but the majority fall into three main classes. The first class gives a power law distribution as a function of two exponential distributions; of two growth processes. The second class gives power law distributions as an outcome of multiplicative models. This is the route that Levy and Solomon have followed in their work, and forms the basis for the GLV distribution discussed in detail in the next section. The third class for producing power laws uses concepts of 'self-organised criticality' or 'SOC'. A second basic point, discussed in Mitzenmacher, is that the difference between a log-normal distribution and a power law distribution is primarily dependent on the lower barrier of the distribution, if the lower barrier is at zero, then you get a log-normal distribution, if the barrier is above zero, then the distribution gives a power tail. A non-zero barrier, provided by wage income, is an essential part of the GLV model discussed in section 1.2 below. The paper of Simkin and Roychowdhury is illuminating and entertaining. It shows that the same basic mechanisms for producing power laws, and branching processes in general, have been rediscovered dozens of times, and that most power law / branching processes are in fact analogous. As an example, the models of Levy & Solomon follow processes previously described by Champernowne in economics, and ultimately by Yule and Simon almost a century ago. This is not to devalue the work of Solomon and Levy; their approach allows for dynamic equilibrium formation, this includes an element missing from most branching models that in my opinion makes the Solomon and Levy model much more powerful as a general model. This is returned to in section 1.2 below. It is however my belief that reading Simkin and Roychowdhury by all those involved in modelling power laws would make their lives a lot easier. Finally it is important to note the difference between income and wealth. Income data is relatively easy to collect from income tax returns. Pareto's original work and almost all subsequent analysis of data is based on that from income data. Wealth data of any quality is very difficult to find. Where this data has been collected it almost exclusively pertains to the richest portion of society, and suggests that wealth is also distributed as a power law for these people. I am not aware of any data of sufficient quality to give any conclusions about the distribution of wealth amongst the bottom 90% of individuals. This has led to some very unfortunate consequences within the econophysics community. Without exception all the exchange models by all the various authors above, including those of Solomon and Slanina, are wealth exchange models. I have not yet seen a model where income (trivially the time derivative of wealth) is measured. 22 EFTA00625150 Despite this, the output distributions from these wealth models are often judged to be successful when they map well onto data derived from income studies. Wealth and income (and sometimes money) are used interchangeably in econophysics papers. This is most unfortunate. A paper on physics; written by an economist, that used energy and power interchangeably would be greeted with considerable scorn by physicists. An explanation for why wealth models can give outputs that can then define income data successfully is given in section 1.4.4 below. Before moving on to the modelling of income and wealth distributions, I would first like to discuss the derivation and mechanics of the Lotka-Volterra distribution and the GLV distribution in more detail. 23 EFTA00625151 1.2 Lotka-Volterra and General Lotka-Volterra Systems 1.2.1 Lotka-Volterra systems Lotka-Volterra systems were independently discovered by Alfred Lotka [Lotka 1925] and Vito Volterra [Volterra 1926] and are used to describe the dynamics of populations in ecological systems. Ultimately this dynamic approach goes back directly to the economic growth equations of Malthus and Sismondi. A basic Lotka-Volterra system consists of a population of prey (say rabbits) whose size is given by x, and a population of predators (say foxes) given by y. Not explicitly given in this simple case, it is further assumed that there is a steady supply of food (eg. grass) for the prey. When no predators are present this means that the population of the rabbits is governed by: dx — = ax dt (1.2.1a) where a is the population growth rate. Left to their own business, this would give exponential, Malthusian growth in the population of the rabbits. In the absence of any rabbits to eat, it is assumed that there is a natural death rate of the foxes: dy — —cx dt (I.2.1b) where c is the population die-off rate, and the negative sign indicates a decline in the population. This would give an exponential fall in the fox population. When the foxes encounter the rabbits, two further effects are introduced, firstly the rate at which rabbits are killed is proportional to the number of rabbits and the number of foxes (ie the chance of foxes encountering rabbits), so: dx dt = —ocx y (1.2.1c) where a is a constant, and the —ve sign indicates that such encounters are not good for the rabbits. However these interactions are good for the foxes, giving: 24 EFTA00625152 dy dt = yxy Where y is again a fixed constant. (I.2.1d) Taken together, the results above give a pair of differential equations: dx = ax — axy dt = x(a — ay) for the rabbits, and: for the foxes. dy = dt yxy — cy = y(yx — c) (1.2.1 e) = y( —c + yx) (1.2.1f) The most important point about this pair of equations is that x depends on y, while at the same time, y depends on x. The dependency goes in both directions, this make things fun. While it is possible for these equations to have a single stable solution, this is often not the case. Commonly the populations of both rabbits and foxes fluctuates wildly. An example is given in figure 1.2.1.1 below for lynx preying on arctic hares [BBC]: Figure 1.2.1.1 here The data for the graph above comes from long-term records of pelts collected by the Hudson Bay Company. The graph shows very closely the recurrent booms and busts in population of the two types of animals. In the short term the population and total biomass of both lynx and hares can increase or decrease substantially. The population of lynx can be large or small in proportion to that of the hares. The populations of both are highly unstable. A subtlety to note is that the population of the lynx follows, 'lags', the population of the hares. It is also worth considering, even at this early stage, the behaviour, or indeed the 'behaviouralism' of the lynx in particular. 25 EFTA00625153 Following a previous collapse, the population of hares can expand rapidly as there are very few lynx to hunt them. As the population of hares increases rapidly, the lynx behave 'rationally' (at least given the absence of long-term, liquidly tradable, hare futures) in both eating lots of hares, and also giving birth to lots of new lynx to feed on the excess of hares. Eventually, of course there are too many lynx for the population of hares, and ultimately there are too many lynx and hares for the underlying amount of grass available. At the peaks of hare and lynx populations there is simply too much biomass wandering around for the land to support. Despite the substantial fluctuations seen in figure 1.2.1.1 above, the populations of both lynx and hares show stable fluctuations around long term averages; roughly 40,000 or so for the hares and 20,000 or so for the lynx, though note that the populations pass through these average values very quickly. In fact the values of the two populations are confined to a band of possible values. The population can move round in a limited set of possible options, this is shown for example in the two figures from simulations below. Figure 1.2.1.2 here Note also the figure 1.2.1.2 shows the same leads and lags in predator and prey populations as the real data. The populations of wolves and rabbits can be displayed on one graph, this then produces the phase diagram in figure 1.2.1.3 below showing how the population of wolves and rabbits vary with each other, and how they are constrained to a particular set of paths. Figure 1.2.1.3 here These diagrams are taken from the website of Kumar, [Kumar 2006], which gives a very good brief introduction to the maths and modelling of Lotka-Volterra systems. It can be seen that the simulated population of wolves and rabbits wanders continuously around average values of approximately seventeen rabbits and six wolves. In contrast, figures 1.2.1.4 & 5 below show the same system with minor changes to the rates of growth. In this model the oscillations slowly die down to stable long-term values. Another alternative is that the oscillations can grow in size unstably and explode to infinity. Figure 1.2.1.4 here 26 EFTA00625154 Figure 1.2.1.5 here One of the important things to note about non-linear dynamic systems such as these is that relatively minor changes in parameters can result in dramatic differences in system behaviour. All the talk of predators and prey can give rise to emotive, and wholly inappropriate, language and modelling. It is an easy, but foolish, course to represent one group of actors (financiers say) as predators, and others (workers) as prey. This is flawed for two reasons. Sometimes the mathematics works the other way, so for example, the Marxian inspired models of Goodwin actually model workers as predators. More importantly, the maths and models are impersonal; they are totally unconnected to the motives of the actors. In fact you don't need both predators and prey, a solitary animal population that grows too quickly can also suffer from population booms and crashes. An example is that of Soay sheep on the island of Soay (in this case the grass can be considered to be the prey, though a better solution would be to use the logistic equation or a similar carrying capacity based approach). 1.2.2 General Lotka-Volterra (GLV) systems As the name implies, the General Lotka-Volterra system (GLV) is a generalisation of the Lotka- Volterra model to a system with multiple predators and prey. This can be represented as: dx, dt = x,r, + Ea xx z.j 1 (1.2.2a) = x,(r, + ah,x,) (1.2.2b) here, dx,/dt is the overall rate of change for the i-th particular species, out of a total of N species. This is made up of two terms. The first term is the natural growth (or death) rate, r4 for the species, where xi is the population of species i. This rate r, is equivalent to the growth rate 'a' in equation (1.2.1e) or the death rate '-c' in equation (1.2.1f). The second term gives the sum of all the interactions with the j number of other species. Here aw is the interaction rate defining the relationship between species i and j. aw is negative if species j is a predator feeding on species i, positive if species i is a predator feeding on species j, or can be of either sign for a heterotroph. a,,; is equivalent to the a of equation (1.2.1e) or the y of equation (1.2.1f). Hopefully it is clear that equations (1.2.2a) and (1.2.2b) are generalisations of equations (1.2.1e) and (1.2.1f) for many interacting species. 27 EFTA00625155 For each species in the system, potentially N-1 interaction rates aw are needed, while N! separate differential equations are needed to describe the whole system. This makes direct solution of the equations for the system somewhat problematic. Fortunately in many systems it is possible to make simplifying assumptions. As an example Solomon [Solomon 2000] proposes the following difference equation as a possible explanation for the power law distribution of city population sizes. This equation describes changes in the distribution in terms of discrete time-steps from time t to time t+1: wht — t t.t (1.2.2c) The terms on the right hand side, in say the year 2003, the year t, add up to give the population w of city i in the year 2004 on the left hand side, which is at time t+1. Such equations are typically used in simulations, one after the other, to give a model of how populations change. Sometimes, though often not, clever mathematicians can derive output population distributions from the underlying difference equations. In equation (1.2.2c), X is the natural growth rate of the population w of city i, but is assumed that X is the same for each city. at is the arrival rate of population from other cities, which is multiplied by the average population w of all the cities. The final term gives the rate of population leaving each city, which is due to the probability ct of an individual meeting a partner from another city. This is given by multiplying the average population w with the population of city i. Leaving aside the detail of the model, important generalisations have been made to produce a more tractable model. In this case X, a and c are universal rates, applicable to all members of the system. X and a both give 'positive autocatalytic' (positive feedback) terms which increase the population w of each city. While the negative value of c ensures that the population of each city has an element of decrease. In the absence of the negative feedback term, the populations of the cities can increase indefinitely to infinity without reaching a stable solution. In the absence of the positive autocatalytic growth of the X in the first term on right hand side, the second and third terms will cause all of the population to end up in a single city. Normally one or more variables are assumed to be stochastic; that is they can vary randomly. In Solomon's example above, all three of A, a and c are assumed to be stochastic. This stochasticity need not be large; it can be small fluctuations around a long-term mean, but it ensures that a locally stable solution is not reached, and that the system evolves into a single long term equilibrium solution. While the above may seem complex, it will be argued later in section 7.3 that this model can be seen as a very general model across many different real world complex systems. 28 EFTA00625156 It is possible to show (though not by me) that the above system can give a stable resultant probability distribution function of populations over the various cities of the form: P(w) (1.2.2d) Which is the general form of the GLV distribution. Or more specifically: P(w) = K(e-1"-" I"11')/((w/L)"+"') (1.2.2e) As has been shown above in section 1.1 this formula gives a very good fit to income data. As well as the quality of fit there are three other reasons that suggest that the GLV may be appropriate for wealth and income distributions. The first two reasons are technical and are discussed below, the third is more subjective and forms the core of this paper. A first reason for preferring the GLV is that this distribution is notable in that the distribution has a main body that is similar to a Maxwell-Boltzmann distribution or log-normal Maddala etc distribution, while the tail follows a power law distribution. While other theories, from both economics and physics, are able to explain one part of the distribution well, it is generally necessary to invoke complex assumptions to explain the remaining part of the distribution, if such an explanation is even attempted. The GLV kills both the birds of income distribution with a single theoretical stone. The second reason for preferring the GLV is that the autocatalytic terms in the GLV give the GLV an automatic offset from zero. As noted above in section 1.1 both the UK and US income data show this offset. While it is perfectly straightforward to put an offset into a log-normal or Maxwell-Boltzmann and other distributions, systems commonly found in nature modelled by the above distributions typically have their origin at zero. The third reason is that the GLV naturally describes complex dynamic flow systems that have reached a maximum entropy production equilibrium. Economics is such a complex dynamic flow system, and it will be seen that the straightforward models described below model real economic outcomes surprisingly well. Solomon further proposes a similar model as an explanation for income distribution: µht+1=A l wi.l +aW —cWw I I t 3.t (1.2.2f) In this case X is proposed to be positive gains by individuals with origins on the stock market, 'a' is assumed to represent wealth received in the form of 'subsidies, services and social benefits, while 'c' is assumed to represent competition for scarce resources, or 'external limiting factors (finite amount of resources and money in the economy, technological inventions, wars, disasters, etc.) as well as internal market effects (competition between investors, adverse 29 EFTA00625157 influence of bids on prices such as when large investors sell assets to realize their value and prices fall as a result. While it is the author's belief that a form of the GLV is appropriate for modelling wealth and income distributions, it is believed that the above economic mechanisms are not realistic. At heart the models of Levy & Solomon remain pair exchange models, with random movements of wealth between individuals. As a realistic description of an economic system this falls short of reasonable requirements. As noted previously, Souma & Nirei [Souma & Nirei 2005, Nirei & Souma 2007] have uniquely moved forward from Levy & Solomon's work in a way that gets closer to meaningful economic fundamentals, however their models include a high degree of complexity. It is also noteworthy that Slanina has produced a pair exchange model that generates an identical output distribution to the GLV output, again it is contended that simple pair exchange is not appropriate as an economic model [Slanina 2004]. In the next section an economic model is proposed that I believe much more closely represents real life economic mechanisms. 30 EFTA00625158 1.3 Wealth & Income Models - Modelling Figure 1.3.1 here Figure 1.3.1 above shows a simple macroeconomic model of an economy. This model is taken from figure 1 of chapter 2 of 'Principles of Economics', by Gregory Mankiw [Mankiw 2004]. Figure 1.3.2 below shows a modified version of the diagram. The two 'markets' between the firms and households have been removed, investment and saving streams have been added, as well as the standard economics symbols for the various flows. Figure 1.3.2 here All standard economics textbooks use similar diagrams to figures 1.3.1 and 1.32 for macroeconomic flows; I have chosen that of Mankiw as his is one of the most widely used. Flows of goods and services are shown in the black lines. The lighter broken lines show the flows of money. (As a simple-minded engineer I prefer diagrams that include flows of goods as well as cash, as I find them easier to follow.) Note that Mankiw shows households owning 'factors of production' such as land and capital, which the households are then shown as selling to firms. This is indicated as a flow of land and capital (along with labour) from households to firms. I personally have never actually sold any machine tools to a manufacturing company, and I have never met any householder who has done so. We will return to this particular 'flow' later. Note also that the total system shows a contained circularity of flow, with balances between supply and demand of goods and services. In this circular flow model economic textbooks assume some basic equalities: C=G C = Y (1.3a) (1.3b) Equation (1.3b) state that the total income gained from firms adding value is equal to the total consumption of goods and services. 31 EFTA00625159 [Nb. In writing this paper I have attempted to use standard notation from economics wherever possible. This occasionally results in confusion. It should be noted that the capital letter Y is used as standard in (macro) economics for income, while small y is used as standard in (micro) economics for outputs from companies. This is not normally a problem, as the two are rarely discussed at the same time in standard economic models. In the discussions of income that follows y is not actually necessary for the analysis, and Y invariably refers to income in the equations of the mathematical model and is normally subscripted.] Figure 1.3.3 here In figure 1.3.3 above I have modified this standard model to reflect what I believe is something closer to reality. Firstly in this model households have been changed to individuals, this is simply to bring the model more in line with the standard analysis of statistical physics and agent based, modelling techniques. This amounts to little more than pickiness. This distinction can be made irrelevant by simply assuming that all households consist of a single individual. Much more importantly, the flow pattern has been changed and the circularity has been disturbed. In the real world most goods and services are consumed in a relatively short period of time. To show this, Consumption C, has been changed to represent the actual consumption of goods. This is a real flow of goods, and represents a destruction of value. Note that this is a change from the standard use of C in economics textbooks. That which was previously shown as consumption is now shown as 'y' the material output of goods and services, which are provided to consumers from the firms operating in the economy. The money paid for these goods and services is shown by My. As can be seen in figure 1.3.3 above, the income stream Y has been split into two components, one, e is the earnings; the income earned from employment as wages and salaries, in return for the labour supplied. is the 'profit' and represents the payments made by firms to the owners of capital, this can be in the form of dividends on shares, coupons on bonds, interest on loans, rent on land or other property, etc. 32 EFTA00625160 The flow of capital has been shown as a dotted line. This is because, as pointed out previously, capital doesn't flow. Householders do not hold stocks of blast furnaces in their backyards in the hope of selling them to firms in exchange for profit or interest on their investments. Capital, such as machine tools and blast furnaces, is normally bought in by firms from other firms, sometimes using money provided by households, but mostly by retained earnings. In fact in all the various models that follow in this paper we are going to ignore both investment I, and saving S. In the income models it is always assumed that the overall economy is in a steady state and so, firstly, that all funds required for wear & repair are taken from internal flows. More importantly, in later models; both for companies and macroeconomic modelling, it is also assumed that all new capital is produced from retained earnings within companies. For many economists, somewhat oddly, this will be seen as a serious flaw. Since at least the time of Keynes, investment and saving have been at the heart of macroeconomic modelling, and this is true of neo-classical and other heterodox modelling, not just that in the Keynesian tradition. The reasons for this are not understood by the author; given that: "Most corporations, in fact, do not finance their investment expenditure by borrowing from banks."[Miles & Scott 2002, 14.2] As examples, Miles & Scott give the following table for proportions of investment financing for four countries averaged over the years 1970-1994. Figure 1.3.4 here [Miles & Scott 2002 / Corbett & Jenkinson 1997] As can be seen the maximum possible proportion of external financing (the IS so beloved of economists) is 36.8% for Japan. For the UK it doesn't even reach 20%. This financing is small to negligible in importance. Most financing is taken from cash flow. Companies that have spare cash buy new toys to play with. Companies that don't, don't. In the whole of this paper the economic models follow reality rather than hypothesis. They are built by modelling capital created and destroyed through imbalances in cash flow. External investment is ignored as the sideshow that it is. Why the whole of macroeconomics should build their models directly contrary to observed data evidence remains a profound mystery. Going back to capital; real capital, in the form of land, machine tools, computers, buildings, etc will be represented in the diagram as fixed stocks of real capital K, held by the companies. All of this real capital is assumed to be owned by households, in the form of paper assets, W, representing claims on the real assets in the form of stocks or shares. In the following discussions bonds and other more complex assets will be ignored, and it will be assumed that all the wealth of K is owned in the form of shares (stocks) in the various firms. This paper wealth will be represented as W in total, or wi for each of i individuals. 33 EFTA00625161 For the income models in the first part of this paper it will further be assumed that the paper wealth of the households accurately represents the actual physical capital owned by the companies, so: total W = total K (1.3e) or: w, = W = K (lid) the total real capital invested in the firms is equal to the total value of financial assets held by individuals. The dotted line in the figure 1.3.3 indicates the assumed one to one link between the financial assets W and the real assets K. It is dotted to show that it is not a flow, it simply indicates ownership. This mapping of real and financial assets assumes that the financial assets are 'fairly' priced, and can be easily bought and sold in highly liquid markets. In the models below it is assumed that there is a steady state, so the totals of W and K are both constant. This means that the model has no growth, and simply continues at a steady equilibrium of production and consumption. There is no change in population, no change in technology, no change in the efficiencies of the firms. The example of Japan over the last two decades has shown that economies can continue to function in a normal manner with extended periods of negligible growth. For a modern economy the difference between the creation and the destruction is economic growth of the GDP, and at 2%-4% or so per annum is pretty close to being stable. This assumption of equality between W and K will be relaxed in later models, with interesting results; but for the moment we will assume the market operates efficiently with regard to asset pricing. It is important to note that the capital discussed here is only the capital vested in productive companies. Other personal capital is excluded, the most important of these is housing. I have ignored the role of housing in these early models, though clearly this is a major simplification. This is discussed further in section 1.9.1 below. For the moment all wealth held is assumed to be financial assets. All other personal assets such as housing, cars, jewellery, etc are ignored. There are some other important base assumptions of the model. These are discussed briefly below: The economy is isolated; there are no imports or exports. There is no government sector, so no taxation or welfare payments, government spending, etc. 34 EFTA00625162 There is no unemployment; all individuals are employed, with a given wage, either from a uniform distribution or a normal distribution depending on the model. Labour and capital are assumed to be complementary inputs and are not interchangeable at least in the short term. It turns out, much later, that this assumption is not only true, but of profound importance, this is discussed at some length later in this paper. There is no investment and saving, the economy is stationary, and depreciation is made good from earned profits. The role of money is ignored in these models, for the sake of argument, it can be assumed that payments are made in the form of units SW of the paper assets held by the individuals, say in units of WI or FTSE all share trackers. Finally there is no debt included in the income models. Figure 1.3.5 below shows some of the assumptions above, it also adds in some more flows to help bring the model closer to the real world. Figure 1.3.5 here There are two main reasons for changing the diagram in this manner. One reason is to bring the diagram into line with the ideas of the classical economists such as Smith, Ricardo, Marx and Sraffa. The second is to help the model comply with some of the more basic laws of physics. Starting with the classical economics. It has previously been defined that consumption by the individuals means the destruction of value in the form of using up resources. This consumption could be food eaten in a few days, clothes which wear out in a few months or cars and furniture that take years to wear out, but which ultimately need to be replaced periodically. The consumption can also be services such as meals in restaurants, going to see films, receiving haircuts, going on holiday, etc. All value destruction is assumed to take place within households as consumption. In physics terms, this destructive process is characterised as a local increase in entropy. To balance this destruction, it is assumed that all value is created in the processes of production, and that all this value is created within firms. I am going to follow in SchrOdinger's footsteps and describe this increase in value as the creation of something called 'negentropy'. For physicists a better term might be 'humanly useful free energy'. For non-physicists, it is asked that detailed understanding of the meanings of 'entropy', 'negentropy' or 'humanly useful free energy' are postponed to part B, where it is discussed at 35 EFTA00625163 length. For the moment the important thing to grasp is that negentropy is equivalent to economic value, the more negentropy something has, the more you are willing to pay for it. Although the discussions in these models use production of manufactured goods as an easily understandable example; it should be noted that 'production' is any process that adds value, and produces higher value inputs than the outputs. So agriculture, mining, power generation, as well as distribution, retail, personal and financial services are all forms of production. Indeed, almost any process that is done within a company is production. That is why companies exist, so that the value added is kept securely within the company. In general, exchange processes don't create value, they are simply a means for swapping goods from different points along the supply chain leading up to the final point of consumption. Exchanges are simply a result of the division of labour between different companies or individuals who have particular sets of skills and abilities. Whether it is the sale of 'lemon' used cars, or the manipulative momentum trading of high- frequency traders, if value is created for one party during an exchange process then this is usually a consequence of an inadequately regulated market that lacks proper informational transparency. The model in figure 1.3.5 above essentially goes back to the ideas of the classical economists; of Smith, Ricardo, Marx, Sraffa and others. It assumes that goods and services have meaningful, long term, intrinsic values, and that long-term prices reflect these values. Short-term prices may move away from these values, primarily to allow generation of new capital. In the models in this paper it is always assumed that value is created in production and that normally exchanges are 'fair' and so there is not net gain of value for either party in an exchange process, again this discussed at more length later in the paper. This paper explicitly rejects the marginalist view that value is exogenously set by the requirements and beliefs of individuals, and that exchange between such individuals creates value. Figure 1.3.6 here Figure 1.3.6 above figure demonstrates these assumptions for a more complex model of linear flows of value added. In figure 1.3.6, all the horizontal flows (flows through the side walls) are direct exchanges of actual goods for monetary tokens. Assuming a free market with fair pricing, and that the currency is a meaningful store of value, then all the horizontal exchange flows have zero net value. xl + Mxl = 0 or: xl = —Mxl, x2 = —Mx2, xk = —Mxk, etc Vertical flows, through the top and bottom of the boxes, involve changes; increases or decrease in negentropy. 36 EFTA00625164 In economic terms this is stated as value being added or wealth being created. In figure 1.3.6 above the values of the final output y and the series of inputs x are related by: y > x3 > x2 > x I and clearly My > Mx3 > Mx2 > Mx I The differences between these values represents the wealth created by the employees and capital of the firm acting on the inputs to create the outputs. The employees are rewarded for this wealth creation via their wage earnings, while the owners of the capital are rewarded with returns on their capital. Figure 1.3.7 here Figure 1.3.7 above gives another layout that shows that the whole system doesn't have to be linear, but that the same assumptions regarding adding value still hold. Finally to satisfy the physicists reading; waste streams are included so that the 2ntl law is not violated. The total entropy created by the waste streams from the firms, principally low grade heat, is greater than the negentropy created in the products of the firms. Essentially figures 1.3.5 to 1.3.7 bring together the economic and physical diagrams discussed in Ayres & Nair [Ayres & Nair 1984]; so that the circulation of wealth and money complies with the laws of physics as well as the laws of finance. The discussions of Ayres & Nair clearly have strong antecedents in the theories of Georgescu-Roegen [Georgescu-Roegen 1971]. Figure 1.3.5 here So, going back to figure 1.3.5, we are now at a point where we can move into the detail of the mathematical model. Firstly we will assume that x = Mx and that both are irrelevant to the rest of the debate. We will also assume that L = e, ie that labour is fairly rewarded for the value of its input. In later sections this is discussed in more depth, but becoming bogged down in a tedious Marxist debate at this stage of the modelling would be particularly unhelpful. Next we will assume y = My, ie that 'fair' prices are being paid for the goods sold to the consumers. We will eventually relax this assumption in later models. 37 EFTA00625165 In this model it will further be assumed that: total C = total y = total My at steady state equilibrium. It will be seen later that this is actually a natural outcome of the model, and doesn't need to be forced. Note that although the totals of C and y are the same, they may not be the same for individuals. Some individuals may consume less than they earn, or vice versa. In these earlier models, we are not interested in the detail of the firms so we are going to ignore the difference between the capital K and it's financial equivalent W. We will assume that total K = total W, and so assume that companies are fairly and accurately priced in the financial markets. These assumptions will be relaxed later, again with interesting consequences. The paper wealth W will be split between N individuals, so from individual i = 1 to individual i = N. Going back to figure 1.3.5 and equation 1.3d above; although the total capital and wealth is fixed, individual wealth is allowed to vary, so: wi., = E = NV = K = constant (I.3e) Where w, is the wealth of individual i. This is economics at a statistical level; a level below microeconomics, nanoeconomics perhaps. Looking at a single individual in the box on the right of figure 1.3.5, in one time unit, from t to ti-1, the change in wealth is given by the following equation: = w... y,.. — MY... + e. . ni t — labour, , — capital,., (1.3f ) This equation states that the wealth for a single individual at time t+1, on the left hand side, is equal to the wealth at time t, plus the contributions of the seven arrows going into or out of the box on the right hand side of figure 1.3.5. However equation (1.3f) is not meaningful as it is trying to add apples and oranges. The items y, C, labour and capital are real things, while w, My, e and a are all financial quantities. Adding the non-financial things is not appropriate, however all the financial flows must ultimately add up. So looking then at the financial flows, we have the following equation: 38 EFTA00625166 whin = whi + Chi + (1.3g) This now counts things that are the same (remember that the currency used for our cash flows were units of SW ). As stated above, although the totals of My = y = C some individuals can consume less than y, and so accumulate more wealth W, others can consume more than y and so reduce their total W. To make this process clearer, I am going to use — Co in place of — My;,, in equation (1.3g). In this case CA is now a monetary unit, and effectively reverts to standard economics usage. To keep the units correct, it is assumed that in practice heavy consumers exchange part of their wealth W with some heavy savers, in return for some of the savers real goods y. This may seem a little confusing but is hoped this will become clearer as the model is more fully explained. Substituting and rearranging, this then leaves us with the following equation: = wi., + e,., + rr,., - Ca., (1.311) This then is the difference equation for a single agent in this model. In a single iteration, the paper wealth w of an individual i increases by the wages earned e plus the profits received it. The individual's wealth also reduces by the amount spent on consumption C. A moment's reflection suggests that this is trivially obvious. We now need to investigate the mechanics of this in more detail. Looking at the second, third and fourth terms on the right hand side of (1.3h) in order, we start with earned income; e. In the first model, Model 1; it is assumed that all agents are identical, and unchanging in their abilities in time, so: e, = e = constant; (1.3i) for all i agents. The assumption above effectively assumes that the economy as a whole is in dynamic equilibrium (the difference between static and dynamic equilibria is discussed at length in section 6 below), there is no technological advancement, no education of employees, etc. It assumes that all individuals have exactly the same level of skills and are capable of producing the exact same level of useful output as one another; and that this is unchanging through time. 39 EFTA00625167 We move next to a, the income from returns. We assume that the economy consists of various companies all with identical risk ratings, all giving a uniform constant return; r on the investments owned, as paper assets, by the various individuals. Here r represents profits, dividends, rents, interest payments, etc to prevent confusion with other variables, r will normally be referred to as the profit rate. This gives: TT = W IJ LI Given r as constant, then: ( I.3j) for each of the i agents. E Tr, = rE w, (1.3k) so: = = r E n t w, Fr w and (1.31) giving: where it and w are the average values of it and W respectively. Note that r, w and Tr are all fixed constants as a consequence of the definitions. So for an individual: R Tit., = W — a. (I.3m) For the final term consumption; C is assumed to be a simple linear function of wealth. As wealth increases, consumption increases proportionally according to a fixed rate n (a suggested proof that this might be reasonable a assumption is given in Burgstaller [Burgstaller 1994], the constancy of n is discussed in depth in section 4.5). So: 40 EFTA00625168 C,., = wa.,12 ( I.3n) This final assumption gives the conceptual reason for using C rather than My for this final term. Clearly a linear consumption function is not realistic, and a concave consumption function would reasonably be expected, with the rate of consumption declining as wealth increased. For most of the modelling, this simple consumption function is sufficient to demonstrate the required results, this is examined further in section 1.9.1 below. In model 1A, Q is made to be stochastic, with a base value of 30% multiplied by a sample from a normal distribution which has a variance of 30% of this base value. By stochastic it is meant that the value can vary randomly up and down about a central average. Consumption is chosen as the stochastic element, as being realistic in a real economy. While earnings are usually maximised and fixed as salaries, choosing to save or spend is voluntary. It should be noted that all agents remain fully identical. While the proportion consumed by each agent changes in the model in each iteration, on average each agent spends exactly 30% of its wealth. This is critically important, and I will not tire of repeating it, in model 1A all the agents are identical and have the same long-term average saving propensity, as well as earning ability. Taken together and substituting into (1.3h) this gives the difference equation for each agent as follows: whin = w,., + e + w,.1 — — w,.112 or simply: wi.,+, = w,., + e + wa.,r — (1.3o) Equation (1.3o) is the base equation for all the income models. Although this is a little different to the standard GLV equations quoted in section 1.2 above, it shares the same basic functions. Firstly it is worth noting how simple this equation is. Here w is the only variable. e, r and n are all constants of one form or another, depending on the modelling used. Note that equation (1.3o) is for a single individual in the model. In future models e, r and n may be different constants for different individuals. However, in this first model, e and r are constant, and the same for all individuals. n is slightly different. It is the same for all individuals, and is constant over the long term, but varies slightly bigger and smaller over the short term due to stochastic variation. The second term on the RHS, the earned income e, provides a constant input that prevents individual values of wealth collapsing to zero. Note that this is additive, where in the models of Levy & Solomon in section 1.2 above this term was multiplicative. 41 EFTA00625169 The third term on the RHS is a multiplicative term and gives a positive feedback loop. The fourth term is also multiplicative and gives negative feedback. In all the income models studied, the total income Y per time unit was fixed, and unless otherwise specified, the earned income was fixed equal to the returns income. So: Y = 1 + = TT% = Tr % = Y 2 constant, always (1.3p) and usually (I.3q) So unless otherwise specified, the total returns to labour are equal to the total returns to capital. This last relationship; that total payments in salaries and total profits are similar in size is not outlandish. Depending on the level of development of an economy, the share of labour earnings out of total income can vary typically between 0.75 and 0.5. Although the value appears to vary cyclically, in developed economies the value tends to be very stable in the region of 0.65 to 075. This was first noted by a statistician, Arthur Bowley a century ago, and is known as Bowley's Law, and represents as close to a constant as has ever been found in economics, figure 1.3.8 below gives an example for the USA. In developing economies, with pools of reserve subsistence labour, values can vary more substantially. Young gives a good discussion of the national income shares in the US, noting that the overall share is constant even though sector shares show long-term changes [Young 2010]. Gollin gives a very thorough survey of income shares in more than forty countries [Gollin 2002]. Figure 1.3.8 here [St Louis Fed 2004] We will come back to Bowley's Law in some depth in sections 1.6 and 4.5-4.8 as it turns out that Bowley's law is of some importance. Because of this importance, it is useful to define some ratios. We already have: Profit rate r = Tr W (1.3r) Where profit can refer to any income from paper assets such as dividends, rent, coupons on bonds, interest, etc. To this we will add: 42 EFTA00625170 Income rate = EY E w (I.3s) which is the total earnings over the total capital. Here total earnings is all the income from wages and all the income from financial assets added together. To these we add the following: E e Bowley ratio /3 = Y (1.3t) En Profit ratio p = Y (1.3u) These two define the wages and profit respectively as proportions of the total income. Following from the above, the following are trivial: /3 + p = 1 (1.3v) Profit ratio p = —r ( I.3w) Finally, in most of the following models, unless otherwise stated (3 = p = 0.5 Going back to equation (1.3o), at equilibrium, total income is equal to total consumption, so: E w2.,+, = E E Y,.,+, = ,f2E SO: where EY. is the total income from earnings and profit. w = (1.3x) so the average wealth is defined by the average total income and the consumption rate. There is an important subtlety in the discussion immediately above. In the original textbook economic model the total income and consumption are made equal by definition. In the models in this paper, income is fixed, but consumption varies with wealth. The negative feedback of the 43 EFTA00625171 final consumption term ensures that total wealth varies automatically to a point where consumption adjusts so that it becomes equal to the income. This automatically brings the model into equilibrium. If income is greater than consumption, then wealth, and so consumption, will increase until C=Y. If income is less than consumption, the consumption will decrease wealth, and so consumption, until again, C=Y. 1.4 Wealth & Income Modelling - Results 1.4.1 Model 1A Identical Waged Income, Stochastic on Consumption In the first model, Model 1A, the model starts with each agent having an identical wealth. The distribution of earning power, that is the wages received e, is completely uniform. Each agent is identical and earns exactly 100 units of wealth per iteration. The split between earnings to labour and earnings to capital are fifty-fifty, ie half to each. The consumption of each agent is also identical, at an average of 30% of wealth. So 70% of wealth is conserved by the agent on average through the running of the model. However the consumption of the agents is stochastic, selected from a normal range so that almost all the agents have a consumption rate between zero and 60% on each iteration. So although the consumption of each agent is identical on average, consumption varies randomly from iteration to iteration. So an agent can consume a large amount on one iteration, followed by a small amount of consumption on the next iteration. It is restated, in the very strongest terms, that all these agents are identical and indistinguishable. The models were run for 10,000 iterations, the final results were checked against the half-way results, and this confirmed that the model quickly settled down to a stable distribution. The results in figure 1.4.1.1 show the probability density function, showing the number of agents that ended up in each wealth band. This is a linear-linear plot. Also shown is the fit for the GLV function. Figure 1.4.1.1 here 44 EFTA00625172 It can be seen that the data has the characteristic shape of real world wealth and income distributions, with a large body at low wealth levels, and a long declining tail of people with high levels of wealth. As expected, the GLV distribution gives a very good fit to the modelling data. Figure 1.4.1.2 shows the cumulative distribution for wealth for each of the agents in the model on a log-log plot. The x-axis gives the amount of wealth held by the agent, the y-axis gives the rank of the agents with number 1 being the richest and number 10,000 Being the poorest. So the poorest agent is at the top left of the graph, while the richest is at the bottom right. Figure 1.4.1.3 shows the top end of the cumulative distribution. It can be seen from figure 1.4.1.3 that there is a very substantial straight-line section to the graph for wealth levels above 1000 units. It can also be seen that this section gives a very good fit to a power law, approximately 15% of the total population follow the power law. Figures 1.4.1.2 here Figures 1.4.1.3 here The earnings distribution for this model is uniform, so the Gini coefficient for the earnings is strictly zero. The Gini coefficient for wealth however is 0.11. In this wealth distribution, the wealth of the top 10% is 1.9 times the wealth of the bottom 10%. The wealthiest individual has slightly more than four times the wealth of the poorest individual. So the workings of a basic capitalist system have created an unequal wealth distribution out of an absolutely equal society. This model, gives probably the most important result in this paper. A group of absolutely identical agents, acting in absolutely identical manners, when operating under the standard capitalist system, of interest paid on wealth owned, end up owning dramatically different amounts of wealth. The amount of wealth owned is a simple result of statistical mechanics; this is the power of entropy. The fundamental driver forming this distribution of wealth is not related to ability or utility in any way whatsoever. In the first model, the random nature of changes in consumption / saving ensure that agents are very mobile within the distribution; individual agents can go from rags to riches to rags very quickly. As a consequence, income changes are very rapid as they depend on the amount of wealth owned. So individual incomes are not stable. For this reason the distribution for income is not shown for model 1A. 45 EFTA00625173 1.4.2 Model 1B Distribution on Waged Income, Identical Consumption, Non- stochastic In model 1B, the characteristics of the agents are changed slightly. Firstly, the agents are assumed to have different skills and abilities, and so different levels of waged income (it is also assumed the are being fairly rewarded for their work). It is still assumed that all agents has an average earning power of 100, and the total split of earnings to capital is still 50%-50%. However, prior to starting the model, each agent is allotted an earnings ability according to a normal distribution so earning ability varies between extremes of about 25 units and 175 units. The worker retains exactly the same working ability throughout the model. Meanwhile the saving propensity in this model is simplified. Throughout the running of the model, each agent consumes exactly 20% of its wealth. There is no longer a stochastic element for the saving, and all agents are identical when it comes to their saving propensity. It should be noted that, although there is a random distribution of earning abilities prior to running the model, because this distribution is fixed and constant throughout the simulation, the model itself is entirely deterministic. This is not a stochastic model. It turns out this model is in fact very dull. With equal savings rates the output distributions for wealth and income are exactly identical in shape to the input earnings distribution. All three distributions have exactly the same Gini coefficient. 1.4.3 Model 1C Identical Waged Income, Distribution on Consumption, Non- stochastic In model 1C, the characteristics of the agents are reversed to those in model 1B. As with model 1A, the agents are assumed to have absolutely identical skills and abilities, and so identical levels of waged income. It is again assumed that each agent has an earning power of exactly 100, and the total split of earnings to capital is still 50%-50%. However, prior to starting the model, each agent is allotted a consumption propensity according to a normal distribution so average consumption rates are 20%, but vary between extreme values of 12% and 28%, while 95% of values fall between 16% and 24%. This is a much narrower range of consumption rates than model 1A with rates only varying plus or minus 20% from the normal rate for the vast majority of people. The big difference to model 1A is that each worker retains exactly the same saving propensity throughout the model, from beginning to end. Again it should be noted that, although there is a random distribution of saving propensity prior to running the model, because this distribution is fixed and constant throughout the simulation, the model itself is entirely deterministic. This is not a stochastic model. 46 EFTA00625174 Figures 1.4.3.1 here Figures 1.4.3.2 here Figure 1.4.3.1 and 1.4.3.2 show the distributions of the wealth data. Figure 1.4.3.1 is the probability density function in linear-linear space while figure 1.4.3.2 is the cumulative density function in log-log space. Again it can be seen that the GLV distribution fits the whole distribution, and that the tail of the distribution gives a straight line, a power law. The fit to the GLV distribution is now less good, especially when compared with figure 1.4.1.1 for model 1A. This is because model 1C is not a 'true' GLV distribution. In the original GLV model described in sections 1.2 and 1.3, and modelled in model 1A, the consumption function was stochastic, and balanced out to a long-term average value. All the agents were truly identical. In model 1C the distribution of consumption is fixed at the outset and held through the model, the agents are no longer identical. As a result the underlying consumption distribution can influence the shape of the output GLV distribution. This is explored in more detail in section 1.4.4 and 1.9.1. In this model, because the consumption ratios are fixed and constant throughout, the hierarchy of wealth is strictly defined. The model comes to an equilibrium very quickly, and after that wealth, and so income, remain fixed for the remainder of the duration of the modelling run. This allows a meaningful sample of income to be taken from the last part of the modelling run. Figures 1.4.3.3 and 1.4.3.4 below show the pdf and cdf for the income earned by the agents in model 1C. Figures 1.4.3.3 here Figures 1.4.3.4 here Figure 1.4.3.4 shows a very clear power law distribution for high earning agents. However figure 1.4.3.3 shows that a fit of the GLV distribution to this model distribution for income is very poor. This income distribution does not match the real life income distributions seen in section 1.1 above. There is a very good reason for this. This is most easily explained by going on to model 1D. 47 EFTA00625175 Not withstanding this, it is worth looking at some of the outputs of the model, compared to the inputs. The inputs are exactly equal earning ability; so a Gini index of zero, and a consumption propensity that varied between 0.16 and 0.24 for 95% of the population — hardly a big spread. The outputs are a Gini index of 0.06 for income and 0.12 for wealth. The top 10% of the population have double the wealth of the bottom 10%, and the richest individual has more than six times the wealth of the poorest individual. As with model 1A, near equality of inputs results in gross wealth differences on outputs. 1.4.4 Model 1D Distribution on Consumption and Waged Income, Non-stochastic In model 1D the distribution of wages is a normal distribution as in model 1B, however the distribution is narrower than that for model 1B. The average wage is 100 and the extremes are 62 and 137. 95% of wages are between 80 and 120. The Gini coefficient for earnings is 0.056 and the earnings of the top 10% is 1.43 times the earnings of the bottom 10%. The distribution of consumption is exactly as model 1C. Importantly the distributions of wages and consumption propensity are independent of each other. Some agents are high earners and big savers, some are high earners and big spenders, similarly, low earners can be savers or spenders. As in models 16 & 1C, the earning and consumption abilities are fixed at the beginning of the model run and stay the same throughout. Again the model is deterministic, not stochastic. Figures 1.4.4.1 here Figures 1.4.4.2 here Figures 1.4.4.1 and 1.4.4.2 show the distributions of the wealth data. Figure 1.4.4.1 is the probability density function in linear-linear space while figure 1.4.4.2 is the cumulative density function in log-log space. Again it can be seen that the GLV distribution fits the whole distribution, and that the tail of the distribution gives a power law section. Again, as with model 1C, there are small variations from the GLV due to the influence of the input distributions. In this model the hierarchy of wealth is strictly defined. The model comes to an equilibrium very quickly, and after that wealth, and so income, remain fixed for the remainder of the duration of the modelling run. Figure 1.4.4.3 and 1.4.4.4 below show the pdf and cdf for the income earned by the agents in model 1D. 48 EFTA00625176 Figures 1.4.4.3 here Figures 1.4.4.4 here It can be that the GLV distribution gives a good fit to the curve, much better than that for model 1C. On the face of it the curve for income distribution appears to be a GLV and the power law tail is also evident. (In fact it is possible that two power tail sections are present, this will be returned to in section 1.9.1 below.) However these assumptions are not quite correct. The power law tail is a direct consequence of the income earned from capital. For the individuals who are in the power tail the amount of income earned from capital is much higher than that earned from their own labour, and the capital income dominates the earned income. So the power tail for income is directly proportional to the power tail for capital. In the main body, things are slightly different. This is not in fact a GLV distribution. The income distribution is actually a superposition of two underlying distributions. The first element of the income distribution is the investment income. This is proportional to the wealth owned. The wealth owned is a GLV distribution; as found above, so the distribution of investment income is also a GLV distribution. The second element of income distribution is just the original distribution of earned income. This input was defined in the building of the model as a normal distribution. By definition the graph is a sum of the two components of Y that is e for wage earnings, and a for payments from investments. The full distribution of income is the sum of these two components. This then explains why the income graph in model 1C fitted reality so badly. In model 1C the underlying earnings distribution was a flat, uniform distribution. This is highly unrealistic, so reality shows a different distribution. In fact there are reasons to believe that the underlying distribution is a 'pseudo-Maxwell- Boltzmann' or 'additive GLV' distribution, which would show a longer, exponential, fall. This is discussed in section 1.9.2 below. Finally this model represents a more realistic view of the real world, with variations in both earning ability and consumption propensity. It is again worth looking at the outcomes for different individuals. Earnings ability varies by only plus or minus 20% for 95% of individuals in this model. Similarly consumption propensity only varies by plus or minus 20% for 95% of people. Despite this the top ten percent of individuals earn more than twice as much as the poorest 10% and the most wealthy individual has 11 times the wealth of the poorest. The outputs give a Gini index of 0.082 for income and 0.131 for wealth. 49 EFTA00625177 1.5 Wealth & Income Modelling - Discussion To start a discussion of the results above, it is worth firstly looking back at figure 1.4.4.2 above. There is a changeover between two groups in this distribution. The bottom 9000 individuals, from 1000 to 10,000 (the top quarter of the graph) are included in the main, curved, body of the distribution. The top 1000 individuals are included in the straight-line power tail. In this, very simple model, class segregation emerges endogenously. The distribution has a 'middle class' which includes middle income and poor people; 90% of the population. This group of individuals are largely dependent on earnings for their income. Above this there is an 'upper class' who gain the majority of their income from their ownership of financial assets. As discussed in 1.4.1 above, the rewards for this group are disproportionate to their earnings abilities, this is most obvious in model 1A where earnings abilities are identical. In economic terms this is a very straightforward 'wealth condensation model'. The reason for this wealth condensation is due to the unique properties of capital. In the absence of slavery, labour is owned by the labourer. Even with substantial differences in skill levels, assuming approximately fair market rewards for labour, there is a limit to how much any single person can earn. In practice only a very limited number of people with special sporting, musical, acting or other artistic talents can directly earn wages many times the average wage, and in fact, such people can be seen as 'owning' monopolistic personal capital in their unique skills. Capital however is different. Crucially, capital can be owned in unlimited amounts. And with capital, the more that is owned, the more that is earned. The more that is earned, then the more that can be owned. So allowing more earning, and then more ownership. Indeed, in the absence of the labour term providing new wealth each cycle, the ownership of all capital would inevitably go to just one individual. (Trivially, this is demonstrated in the game of Monopoly, where there is negligible consumption and insufficient provision of new income (via passing Go, etc) to prevent one agent accumulating all the capital.) In the various income models above, the new wealth input at the bottom (due solely to earnings not capital) prevents the condensation of all wealth to one individual, and results in a spread of wealth from top to bottom. But this still results in a distribution with a large bias giving most of the wealth to a minority of individuals. Going back to the Lotka-Volterra and GLV models discussed in section 1.2, it is better to abandon the predator-prey model of foxes killing rabbits, and instead think in terms of a 'grazing' model where the 'predators' are sheep and the 'prey' is grass. In this model the prey is not killed outright, but is grazed on, with a small proportion of its biomass being removed. The wealth condensation process can then be thought of in terms of a complex multi-tier grazing model, a little analogous to the tithing model in medieval Europe. 50 EFTA00625178 In a simple tithing system, the peasants don't own the land, but are tied to the land-owners. They are allowed to work the land and keep a proportion of the crops grown. However they are obliged to pay a portion of the tithes to the lord of the manor, and also some to the church. The tithes form the rent payable for being allowed to use the land. The lord of the manor may be obliged to pay taxes to the local noble. The noble will be obliged to pay taxes to the king. As national institutions the church and king can gain substantial wealth, even with a relatively low tax, as they can tax a lot more people. In a modern capitalist system things are similar but the payments are now disintermediated. People supply their labour to employers, and receive payments in wages as compensation. Payments to capital are returned in the form of interest on the owners of the capital. The more capital you have, the more return you get. The more capital you have, the bigger grazer you are in a near infinite hierarchy of grazers. The higher up you get the grazers get bigger but fewer. So, to take an example, Rupert Murdoch is a fairly high level grazer as he owns many national newspapers and television stations, so many people make use of his business, and reward him with a small percentage of profit. At the time of writing, Bill Gates is the apex grazer, because even Rupert Murdoch's companies use lots of computers with Windows software. The more capital you have got, the more grazing you get to do. That capital causes wealth to condense at high levels in this way is in fact a simple statement of the obvious. To the man on the street it is clear that the more money you have, the easier it is to make more, and the question of whether money that is gained by investment is 'earned' or justified remains open to debate. The fact that paying interest unfairly benefits the rich has of course been noted by Proudhon, Marx, Gesell and other economists and philosophers. For the same reasons usury was also condemned by the writers of Exodus, Leviticus and Deuteronomy. Other critics of usury include Allah, Thomas Aquinas, and all the popes from Alexander III (1159 to 1181) to Pope Leo XII (1823 to 1829); not to mention writers in Hinduism and Buddhism. In these circumstances, the failure of mainstream economists to notice this basic problem with capitalism is puzzling. As an aside, this may explain the common emergence of civilisation in river valleys that run through deserts; such as Mesopotamia and Egypt. What these areas have in common is good fertile land, but land that is limited in supply. If there is a bad year, a farmer with excess food, due say to different balance of crops, could offer assistance to another farmer with no food, in return for a portion of land. After a while, some farmers will end up with excess land, others with insufficient land. Those with insufficient land will be obliged to labour for those with excess. This then starts off the multiplicative process of accumulation that ends up with Pharohs who own very large amounts of land, and can afford to luxuriate in the arts. For evidence of the existence of power laws in ancient Egypt see [Abul- Magd 2002]. This would not have worked in for example the Rhine or Danube valleys, because while both these rivers have fertile land, there is also plenty of surrounding, rain-fed land, which is also available. A person who became landless would simply move up the side of the valley and create some new personal capital by changing forests into fields with an axe. 51 EFTA00625179 The actual details of how the wealth is shared out is a consequence of entropy. An understanding of entropy provides standard methodologies of counting possible states that a multi-body system can occupy. In the case of the GLV, this appears to be a consequence of 'path entropy' the number of different routes through a system that can be taken. One of the profound things about entropy, and one of the reasons why it can be so useful, is that the statistical power of entropy can make microscopic interactions irrelevant. So important macroscopic properties of multi-body systems can be calculated without a knowledge of detailed microscopic interactions. It is not proposed to discuss this in detail here; the second part of this paper discusses the concept and consequences of entropy in much more detail. The essential point that needs to be understood at this point is that the GLV distribution is the only possible output distribution in this model because of simple statistical mechanical counting. No other output distribution is possible given the restraints on the system. The invisible hand in this system is the hand of entropy. As has been repeatedly noted, a GLV, complete with power tail, and gross inequality, can be produced from model IA which uses absolutely identical agents. In this regard, it is worth noting; and this is extremely important, some of the many things which are not needed to produce a wealth distribution model that closely models real life. It is clear that to produce such a model, you don't need any of the following: • Different initial endowments • Different saving/consumption rates • Savings rates that change with wealth • Different earning potentials • Economic growth • Expectations (rational or otherwise) • Behaviouralism • Marginality • Utility functions • Production functions In this equilibrium, utility theory is utterly irrelevant. In fact there is no need for utility in any form whatsoever; and, sadly, in an act of gross poetic injustice; you don't need Pareto efficiency to produce a Pareto distribution. The GLV distribution is a direct consequence of the power of entropy combined with the simple concept of a rate of return on capital. It is a full equilibrium solution, a dynamic equilibrium, but an equilibrium nonetheless. 52 EFTA00625180 In economic systems utility is not maximised. In fact it appears that there is an alternate maximisation process controlling economics, the maximisation of entropy production, and that this is of profound importance, this is discussed in 7.3 below. The non-maximisation of utility of course has important consequences; the distributions of wealth and income dictated by the GLV are neither efficient or rational, never mind fair. In real life human beings are not rewarded proportionally for their abilities or efforts. I would like to end this discussion by noting the similarities and differences between my own models and those of Ian Wright. Superficially Wright's models are very different to my own. Wright does not include a financial sector, or interest rate payments. So clearly Wright's models can not follow my own mathematical definitions. (Wright's approach does not discuss mathematical modelling formally in general.) In Wright's models, the workforce is split into owner manager 'capitalists' who each own an individual company, and 'workers' who are employed by the capitalists. Importantly, Wright allows movement between the capitalist and worker class, through new company formation and dissolution. In practice this results in the same fundamentals as my own models. The capitalists pay the workers for their labour, which is identical to my own models. The capitalists are then rewarded with income according to the size of their own company. So although wealth is not disintermediated, stochastic effects allow wealth to concentrate in the hands of individual capitalists to form a power law identical to my own models. As a result the distributions of wealth and income are similar in Wright's models to my own. While I believe that my own models are more realistic in using the disintermediation of interest/dividend payments. Wright's models are 'purer' and demonstrate the fundamental power of statistical mechanics. Wright demonstrates that you don't even need a financial sector to produce the same income distributions that are seen in the real world. 1.6 Enter Sir Bowley - Labour and Capital All the income models above were carried out using a 50%/50% split in the earnings accrued from capital and labour. So in all the previous models the profit ratio p and the Bowley ratio 13 are both equal to 0.5. In this section the effects of changing these ratios is investigated. It was noted in model 1B that the input wage distribution, of itself, has no effect on the output distribution. That is to say; the input wage distribution is copied through to the output distribution. It is the consumption/savings ratios that generate the power tails and make things interesting. To keep things clearer, model 1C was therefore chosen, as this has a uniform wage distribution. This is less realistic, but makes analysis of what is happening in the model easier. 53 EFTA00625181 Reruns of the simulations were carried out for model 1C with varying proportions of returns to capital and labour. The profit ratio p; the ratio of returns to capital over total returns, was varied from 0 to 1, ie from all returns to labour to all returns to capital. From the resulting distributions it was possible to calculate the Gini coefficients and the ratio of wealth/income between the top 10% and the bottom 10%. The poverty ratio, the proportion of people below half the average wealth/income is also shown. The data for this model is included in figure 1.6.1. The variation of Gini coefficients and poverty ratios with profit ratio are shown in figure 1.6.2. Figure 1.6.3 shows how the ratio of the top 10% to the bottom 10% changes with profit ratio. The results are dramatic. Figure 1.6.1 Profit Ratio 0.00 0.10 0.20 0.30 0.40 0.50 0.60 0.70 0.80 0.90 1.00 Bowley Ratio 1.00 0.90 0.80 0.70 0.60 0.50 0.40 0.30 0.20 0.10 0.00 Gini coefficient wealth 0.06 0.06 0.07 0.08 0.10 0.12 0.15 0.37 0.63 0.84 1.00 Gini coefficient total income 0.00 0.01 0.01 0.02 0.04 0.06 0.09 0.26 0.50 0.75 1.00 decile ratio wealth 1.43 1.49 1.57 1.68 1.84 2.09 2.58 7.81 22.68 67.31 Mil decile ratio income 1.00 1.04 1.10 1.17 1.28 1.45 1.78 4.60 12.46 36.04 Inf poverty ratio wealth 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.07 0.76 0.99 1.00 poverty ratio income 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.37 0.99 1.00 Figure 1.6.2 here Figure 1.6.3 here The model used is model 1C In which the earnings potential is a uniform distribution and so is equivalent for all individuals, that is all the agents have equal skills. However in model 1C savings rates are different for different agents. Clearly when all earnings are returned as wages p = 0, 13 = 1, and the Gini index is zero. In contrast, when all earnings are returned as capital, one individual, the one with the highest saving propensity, becomes the owner of all the wealth, and the Gini index goes to 1. 54 EFTA00625182 (From a profit ratio of 0.65 upwards, the Gini coefficient for wealth appears to vary linearly with the profit ratio, though the mathematics of this were not investigated.) Figures 1.6.4 and 1.6.5 show the variation of the power exponent (which describes the power tail of the distribution) with the profit ratio. Figure 1.6.4 Bowley Ratio 1.00 0.90 0.80 0.70 0.60 0.50 0.40 Profit Ratio 0.00 0.10 0.20 0.30 0.40 0.50 0.60 Power Tail Slope Wealth na -17.42 -14.81 -12.20 -9.59 -6.97 -4.23 Figure 1.6.5 here For very low and very high values of the profit ratio the power tail is not well defined, but for a range of values in the middle the results are mathematically very interesting. For model 1C The relationship between alpha and the profit ratio p is strikingly linear. If the plot is limited to the thirteen data points between 0.05 and 0.65 the R2 value is 0.9979. If the plot is further restricted to the eleven points between 0.1 and 0.6 the R2 value rises to 0.9999. It appears that in this case there is a direct mathematical relationship between the Bowley Ratio and the a that defines the power tail in the GLV equation. This relationship was investigated further by rerunning the model and varying the various parameters in the model systematically. The value of a was calculated in the model using the top 400 data points and the formula: a = I + ni In(xix„„„) where n is 400, and the sum is from 1 to n. (1.6a) The parameters available to change are as follows. Firstly the ratio of total income to total capital; that is the total income to both labour and capital (wages plus dividends) as a proportion of total capital, this was defined as the income rate, 1, in equation (1.3s). Secondly relative returns to labour and capital; that is either the profit ratio p or the Bowley ratio (3. Either can be used as they sum to unity. Thirdly the average value of the consumption rate Q, and fourthly the variance of this consumption rate. The first interesting thing to come out of this analysis was that the income rate, the ratio of total returns to total capital r had no effect on a whatsoever. Indeed the author reran the models a number of times believing an error had been made in the coding — eventually the presence of 55 EFTA00625183 very small differences at multiple decimal places demonstrated that the models were in fact working correctly. The second attribute to drop out of the model was that seen in figure 1.6.5 above; for fixed values of the other parameters there was a substantial central section of the profit ratio p for which (absolute) a declined linearly with increasing p. Like the total returns, varying the absolute value of the consumption rate s2 had no effect whatsoever on the value of a. Although the absolute value of Q had no effect on a, changing the variance of Q had a significant effect. In this model Q is distributed normally, and v is used to denote the matlab variance (O2) parameter compared to the total value of Q. In this model the value of a appears to vary as a power law of v. It should be noted that the value of v could only be increased from 0 to roughly 0.25. Around this value of 0.25 the outliers in the distribution of s2 become similar to the average size of Q. This creates negative values of s-2 for some individuals which results in no consumption, and so hyper-saving for these individuals. This is both unrealistic and results in an unstable model. (a better model would treat this as a new boundary condition.) A first attempt at fitting of the data gave very good fits across the range of p and v using the following equation for (absolute) a: a = 1.5 1.9p V 1.30 V 1.07 (1.66) The presence of power laws for v under both terms, with similar powers, was too tempting. So a second fit was attempted using a common denominator. This gave the equation below which gave a fit to the data almost as good as equation (1.6b): (1.37 — 1.44p) = IAS V (I.6c) now the two constants had moved suspiciously close together, so a further fit was carried out using a common constant, again this gave a data fit almost as good as (1.6b) and (1.6c): CY = 1.36(1 — p) VMS (1.6d) Of course (1.6d) can more simply be written as: 56 EFTA00625184 = 1.360 I 'S V (1.6e) Where is of course the Bowley ratio. Equations (1.6d) and (1.6e) are deceptively simple and appealing, and their meaning is discussed below in more detail. Before this is done, it is worth stressing some caveats. Firstly the two equations (1.6d) and (1.6e) have been extracted empirically from a model. They have not been derived mathematically. Neither have they been extracted from real data. Although it is the belief of the author that the equations are important and are sound reflections of economic reality, this remains solely a belief until either the equations are derived exactly or supporting evidence is found from actual economic data; or, ideally, both. Secondly the nature of the two variables 13 and v are different. The Bowley ratio is well known in economics and is an easily observed variable in national accounts. In contrast v is the variance in an assumed underlying distribution of consumption saving propensity. In real economics the shape of such a distribution is highly controversial and is certainly not settled. Thirdly, the two equations are limited by the parameters included in a highly simplified model. In real economies it is likely that other parameters will also effect a. Finally, the two equations are for wealth, and do not fit the income data. A similar investigation was carried out to look at the variation of the a for the income distribution power tails. The results were much more complex, and beyond this authors mathematical abilities to reduce to a single equation. As with the wealth distributions, neither the total returns or the average value of the consumption ratio s2 had any effect on the value of a for income. For any fixed value of v, the absolute value of a declined with increasing p, however the decline appeared to be exponential rather than linear. Similarly for any fixed value of p the value of a appeared to decline exponentially with v. Attempts to combine these facts together necessitated introductions of increasing numbers of terms and proved fruitless. Hopefully somebody with greater mathematical skills than myself should be able to illuminate this. Despite this failure to extract a meaningful formulation, it is clear that increasing the value of the profit ratio p, or reducing the Bowley ratio 13 has a direct causal relationship on a resulting in reducing the absolute value of a for income, just as it does for the a for wealth. This is of the utmost importance for the welfare of human beings in the real world. It is of course trivially obvious that decreasing the Bowley ratio and increasing the profit ratio is bad for wealth and income distribution. If more income is moved to the small numbers of capital holders, at the expense of the much larger number of wage earners, then income distribution as a whole is going to get worse. But equation (1.6d) shows that it is in fact much worse than that. 57 EFTA00625185 The a of the GLV defines the log law of differences in wealth for people in the power tail. As the absolute value of a decreases, inequality increases. Because a is the 'slope' of an inverse law curve (rather than say the slope of a straight line), small changes in a produce very large changes in distribution of wealth. Also by moving wealth around in the main body of the GLV, the a has a profound effect on the wealth and income of all people, not just the rich. The clear link between the Bowley ratio and the a's of the wealth and income distributions means that the changing value of the Bowley ratio has profound effects on the Gini index, relative poverty levels etc. Increasing returns to capital, at the expense of labour produces substantial feedback loops that increases poverty dramatically. All of this of course begs the question of what exactly controls the values of the profit ratio p, the Bowley ratio 13 and the shape of the consumption rate distribution, so giving v. I intend to return to the source of the Bowley ratio in detail in sections 4.5 to 4.8 below with what appears to be a straightforward derivation. My answer to the source of v is more tentative and more subjective, this will be introduced briefly below, but will be returned to in more depth in section 7.3 under the theoretical part below. Before discussing the source of the consumption rate distribution, I would first like to return to equations (1.6d) and (1.6e): = 1.36(1 — p) 1.15 V 1.360 = (1.6e) v1.15 (1.6d) Although equation (1.6e) is simpler, equation (1.6d) is the key equation here. Indeed the more diligent readers; those who boned up on their power law background material, may have noted the strong resemblance of equation (1.6d) with the exponent produced from equation (45) in Newman [Newman 2005], which gives a general formula for a as: a= I — a/b (I.6f) Where a and b are two different exponential growth rate constants. This is of course exactly what we have in equation (1.6d) where p is the ratio of two different growth constants, r and r. Going all the way back to equations (1.3h, 1.3p, 1.3v, 1.3s and 1.3w) p is the ratio of the different components of Y, which are e and it. 58 EFTA00625186 The total income produced by capital, the amount of value created in each cycle, is given by the sum of wages and profits: Total Income E Y = Ee + LIT (I.3p) EY Income rate = (1.3s) E w The direct returns to capital; that is the returns to the owners of the capital, is given by the profit rate: Profit rate r = E Tr L w (I.3r) but p is defined by: Profit ratio p = direct returns to capital total income from capital E rr/Ew Profit rate r = so: EY/Ew Profit ratio p = —r (1.3w) The value of p is simply the growth rate that capitalists get on capital, divided by the growth rate that everybody (capitalists and workers) gets on capital. It is the combination of these two growth rates that creates and defines the power law tail of the wealth and income distributions. This is the first, and simplest class of ways to generate power laws discussed in Newman [Newman 2005]. And a curious thing has happened here. There are many different ways to produce power laws, but most of them fall into three more fundamental classes; double exponential growth, branching/multiplicative models, and self- organised criticality. The models in this paper were firmly built on the second group. The GLV of Levy and Solomon is a multiplicative model built along the tradition of random branching models that go back to Champernowne in economics and ultimately to Yule and Simon [Simkin & Roychowdhury 2006]. 59 EFTA00625187 Despite these origins we have ended up with a model that is firmly in the first class of power law production, the double exponential model. It is the belief of the author that this is because the first two classes are inherently analogous, and are simply different ways of looking at similar systems. Much more tentatively, it is also the belief of the author that both the first two classes are incomplete descriptions of equilibrium states, and further input is need for most real systems to bring them to the states described by the third class; that of self organised criticality (SOC). Going back to the wealth and income distributions, equation (1.6d) can define many different possible outcomes for a. Even with a fixed Bowley ratio of say 0.7, it is possible to have many different values for a depending, in this case, on the value of v. It is worth noticing that there is a mismatch between the values for ce given by the models and economic reality. The models give values of a of 4 and upwards for both wealth and income. In real economies the value of alpha can vary in extreme cases can between 1 and 8, but is typically close to a value of 2 see for example Ferrero [Ferrero 2010]. While the model clearly needs work to be calibrated against real data, it is the belief of the author that the relationship between a and p or (3 is valid and important. It is the belief of the author that in a dynamic equilibrium, the value of a naturally tends to move to a minimum absolute value, in this case by maximising v to the point where the model reaches the edge of instability. At this point, with the minimum possible value of a (for any given value of p or (3) there is the most extreme possible wealth/income distribution, which, it is the belief of the author is a maximum entropy, or more exactly a maximum entropy production, equilibrium. This belief; that self-organised criticality is an equilibrium produced by maximum entropy production, is discussed in more detail in section 7.3 below. It is the suspicion of the author that the unrealistic distribution for S2 used in the modelling approach above results in a point of SOC, that is artificially higher than that in real economies. Indeed, it is a suspicion that movement towards SOC may of itself help to define underlying distributions of earnings and consumption. This is returned to in section 7.4. 1.7 Modifying Wealth and Income Distributions The modelling above shows that grossly unequal distributions of wealth and income are produced as a natural output of statistical mechanics and entropy in a free market society. In particular, the ownership of capital and the function of saving are key to the formation of inequality in wealth and income distributions. In communist states strict, and active, microeconomic control was the normal way of attempting to prevent large discrepancies in wealth. In democratic countries this has generally been avoided, partly because of the stunting effects on economic growth, but also because of the restrictions on liberty. Instead these countries have instituted substantial systems of taxation and 60 EFTA00625188 welfare in an attempt to transfer income from the rich to the poor. Meanwhile trade unions and professional societies also attempt to modify wealth distributions for their own members. From an econodynamics point of view the above methods of attempting to influence income distribution are deeply flawed. In a system of a large number of freely interacting agents the GLV distribution is inevitable and methods of exchange, even ones such as tax and welfare, are largely irrelevant. One approach that does make some sense is that of the trade unionists and professional societies. By tying together the interests of thousands, or even millions, of individuals their members are no longer "freely interacting" and are able to release themselves from the power of entropy to a limited extent. (Monopolistic companies attempt to subvert entropy by similar means). Traditional methods of taxation and welfare have much less justification. This solution attacks the income flows directly, and does not address the issues of capital. Also by attempting to directly micromanage the income distribution, taxation and welfare attempts to impose a non- equilibrium statistical outcome at a microscopic level. This approach is doomed to failure. It is common experience that such transfers give little long-term benefit to the poor. Transfers need to be massive and continuous to be effective, and there is a wealth of data to suggest that many welfare programmes result in the giving of benefit to those of medium and high incomes, rather than to the poor, see section 1.8 below for a discussion of this. This is of course exactly what an econodynamic analysis would predict. Given the power of entropy to force the overall distribution regardless of different sorts of microeconomic interactions, it would initially seem that attempts to modify income distribution will be futile. This is not necessarily the case. As discussed above trying to fight entropy head on is a pointless task. However in the following two sections alternative approaches look at how wealth and income distributions might be modified, given the knowledge that these distributions are formed in a statistical mechanical manner. The first approach looks at imposing boundary conditions on a model of society, the second looks at modifying the saving mechanism feedback loop. 1.7.1 Maximum Wealth The author has previously proposed that the imposition of a maximum wealth level should, by symmetry, produce a symmetrical distribution of wealth and income [Willis 2005]. This proposed solution was based on the (mistaken) assumption that wealth and income distributions were formed in a static exchange equilibrium. Model 1D was rerun to test this theory. Two different versions were rerun, a lazy version and a greedy version. Both versions included an additional rule that came into play when any agent reached a wealth level of more than 50% greater than the average wealth level. In the first rerun; the lazy version, any agent that reached the maximum wealth level duly had their incentives reduced, and reduced their work rate by 5% (5% of its current value). If the agent repeatedly hit the maximum wealth limit, then they repeatedly had their work rate reduced. 61 EFTA00625189 In the second rerun; the greedy version, any agent that reached the maximum wealth increased their consumption by 5% of current. Figure 1.7.1.1 shows the cdf outcome for the increasing consumption model, the graph for the decreasing work model is almost identical. Figure 1.7.1.1 here Contrary to the expectations of the author, the maximum wealth model fails dismally in achieving it's hoped for aims. The resulting distribution merely flattens off the unconstrained distribution. This has the effect of bunching a large minority of agents at near equal wealth levels close to the maximum permitted wealth. It is worth noting that, in the real world, this particular group of agents would include most of the ambitious, clever, innovative, entrepreneurial, well educated and politically well connected. This model also has the notable non-effect of not assisting the impoverished at the bottom of the distribution in any noticeable way. This model makes the rich poorer, but doesn't make the poor richer. Taken together, this social model would seem to present a highly effective way of precipitating a coup While the author remains romantically attached to the concept of maximum wealth limits, and believes that they may form the basis for interesting future research, this approach is not currently proposed as a basis for tackling inequality in a real economy. 1.7.2 Compulsory Saving The second approach for changing income distributions focuses on the crucial role of saving in the GLV equation. From models 1B and 1C it appears that rates of consumption and saving are critical to the formation of the power tail and so large wealth inequalities. If saving is the problem, it seems sensible to use saving as the solution. Again model 1D was used as the base model. In this model a simple rule was introduced. If any agent's current wealth was less than 90% of the average wealth, that agent was obliged to decrease their consumption rate by 20 percent. This could be thought of an extra tax on these individuals, which is automatically paid into their own personal savings plan. It should be noted that this increase, though significant, is not enormous, and is comparable say to the rate of VAT/income tax in many European countries. Figure 1.7.2.1 here 62 EFTA00625190 Figure 1.7.2.2 here Figures 1.7.2.1 and 1.7.2.2 show the log-log and log-linear cumulative distributions for the model, with and without the compulsory saving rule. It can be clearly seen in figures 1.7.2.1 and 1.7.2.2 that the number of poor people is much smaller with compulsory saving. For the bottom half of the agents (the top half of figure 1.7.2.2), the distribution is very equal, though it retains a continual small gradient of wealth difference. The top half of society retains a very pronounced power-law distribution, with approximately the same slope. Each individual in the top half is less wealth by an amount that varies from roughly 5% for those in the middle to roughly 10% for those at the top. Despite this they remain far richer than the average. This drop in wealth seems a very slight price to pay for the elimination of poverty, and the likely associated dramatic reduction in crime and other social problems. The power tail structure would leave in place the opportunity for the gifted and entrepreneurial to significantly better themselves. Retaining the group of high earners in the power tail would also have the useful secondary effect of providing an appropriate source of celebrity gossip and target for quiet derision for the remaining, now comfortable bottom half. Figure 1.7.2.3 shows various measures of equality with and without the saving rule. Figure 1.7.2.3 No Compulsory Saving Compulsory Saving Gini Earnings 0.056 0.056 Gini Wealth 0.131 0.077 Gini Income 0.082 0.058 Earnings Deciles Ratio 1.429 1.429 Wealth deciles ratio 2.268 1.617 Income deciles ratio 1.686 1.451 The results are dramatic and also very positive. Without compulsory saving the input earnings distribution was magnified through saving in the GLV into a more unequal distribution for wealth and income. This can be seen in both the Gini indices and also the ratio of the wealth or income of the top 10% to the bottom 10%. With compulsory saving the output distribution for income has almost the same inequality values as the original earnings distribution for both the Gini index and deciles ratio. Wealth is more unequal, but much less so than in the model without compulsory saving. In fact the shapes of this output income distribution (in figures 1.7.2.1 & 2 above) is significantly different in shape to the input earnings distribution, which in this case is a normal distribution. But by smoothing out the rough edges of the GLV, compulsory saving provides an output that is similar in fairness to the skill levels of the inputs. This is probably a distribution that society could live with. 63 EFTA00625191 In practice poverty has been eliminated for all except those that combine a very poor earnings ability with a very poor savings rate — individuals who in real life would be necessarily be candidates for intervention by the social services. Rather than being purely equitable distributions, the output distributions could be better described as pre-Magrathean: "Many men of course became extremely rich, but this was perfectly natural and nothing to be ashamed of because no one was really poor " It is also worth noting the form in which this transfer of wealth takes place. In this model the rich are not taxed. In this model the poor are compelled to save. The rich would only notice this form of financial redistribution in the form of increased competition for the purchase of financial assets. In practice a compulsory saving scheme would be highly effective once the new, more equal, distribution was in place. However expecting people who are currently very poor to save their way out of poverty is not reasonably realistic. Section 1.8 below discusses extensions of these ideas in more detail. 1.8 A Virtual 40 Acres In this section more detailed proposals are made for modifying wealth and income distributions; based on the outcomes of the models above. It is hoped that these proposals will provide solutions that are more practical, effective and far less costly than current mechanisms such as welfare and subsidised housing. Before continuing with these discussions, I believe it is worth stating some of my own personal political beliefs. This paper uses theoretical ideas from Marx, though the classical economics is equally attributable to Adam Smith. In addition the discussion below is substantially about the reallocation of capital. However I emphasise that I disagree in the very strongest terms with Marx's proposed methods for redistributing capital. I strongly believe that the creation of wealth by market capitalism, within a democratic state, must remain at the core of any effective economic system. I believe that redistribution of capital can be achieved in an effective manner within a democratic, capitalist state, in ways that are much cheaper and more effective than methods currently used in democracies. My aim is not to take from the rich and give to the poor. My aim is to achieve a property owning democracy, where all members own sufficient property to guarantee a basic standard of living (and where the word property does not refer just to housing). In sentiment, though not in policy particulars, I am much closer emotionally to the followers of binary economics and their Capitalist Manifesto, than I am to the ideas in the Communist Manifesto. In the previous section, I proposed that redistribution is carried out by forcing the poor to save, rather than taxing the rich. It is hoped that this makes clear that, while I am very sympathetic to some Marxian insights into economic theory, I am wholly opposed to traditional Marxist proposals to deal with inequality. 64 EFTA00625192 In many ways I believe the ideas represented in this section are improvements on ideas first proposed by Milton Friedman. Although staunchly right wing, unlike most laissez-faire free- market economists, Milton Friedman recognised that capitalist economies did not ensure a distribution of income that allowed all citizens to meet their basic needs. In his book 'Capitalism and Freedom' [Friedman 1962], he proposed the introduction of a 'negative income tax', a policy that now exists in the form of 'earned income tax credit' in the USA, and which has been copied successfully in other countries. As a form of income redistribution, Friedman's ideas suffer from needing continuous flows. I believe my own proposals achieve the same aims of those of Milton Friedman, at much less cost. I would ask that readers consider these proposals to be more neo—Friedmanite than neo- Marxian. If, however, my ideas are incorrect, then I would rather live with freedom and inequality than equality and injustice. Civil rights are more important than economic rights. To briefly review the conclusions on income models discussed above in sections 1.3 to 1.5, it is possible to conclude the following: Income and wealth distributions are defined by entropy. Income and wealth distributions are not defined by utility, marginality, ability in general or entrepreneurial ability in particular. Income and wealth is gained in a reinforcing circular flow, the more money you have, the more money you will receive. Income and wealth distributions are strongly skewed, giving disproportionate wealth to a small minority. Income and wealth distributions are strongly biased in favour of those who inherit wealth. Despite the above conclusions there is still a question, that needs to be answered, as to why it is felt necessary to change income distributions at all. Some of the arguments are discussed briefly below. The first thing to note is that recognising that wealth and income distributions are caused by entropy, rather than say utility or ability, changes the whole nature of the political debate on redistribution. At present, it is normally assumed within economics that income and wealth distributions are 'natural' and caused by maximisation of utility and/or rewards for entrepreneurial or other ability. It is further assumed that moving away from this 'natural' equilibrium will have bad effects; interfering with the market, reducing overall utility, removing incentives for wealth creation, etc. Under these assumptions, economists and many politicians take the view that any case for changing existing income distributions must be very strong, and movement from the 'natural' position must be justified. Once it is realised that income and wealth distributions are caused by entropy, then things become very different. The entropy equilibrium position may be 'natural' in the scientific sense, but it does not maximise utility. It specifically punishes hardworking people, the majority of individuals, who are effectively debarred from the ownership of capital. This is despite the fact 65 EFTA00625193 that the labour of these people form the main supply of new wealth that allows capital formation. In this sense the current system of ownership of capital works as a private taxation system acting on the majority of individuals, transferring the majority of the wealth to a small minority of individuals. This 'taxation' is far more iniquitous then any standard taxation system used in a normal democracy. Under these circumstances, failing to modify income distributions becomes a highly political decision. It becomes a decision to support and entrench a system that takes from the poor and middle classes to reward the rich. If this is what the public in a democracy choose to do, then that is fine; but the political debate needs to be made absolutely clear. Two recent papers suggest that the understanding of the deep seated nature of this injustice is very deep. In their paper, Griffiths and Tenenbaum [Griffiths & Tenenbaum 2006] demonstrate that ordinary people, lacking in a mathematical education, are capable of accurately judging whether data fit different mathematical distributions such as the normal or power law. Given that most skills are based on a normal or log-normal distribution, and that wealth is distributed as a power law, this would suggest that people intuitively, and reasonably, realise that distributions of wealth are unfair. In another paper Norton and Ariely [Norton & Ariely 2010], show that Americans, even rich Americans, believe that the United States would ideally have a distribution of wealth more like that of Sweden. Given the political nature of the decision discussed above, the first obvious reason for modifying income distributions is simply common decency. Or, alternatively, basic obedience to spiritual teachings. All major religions recognised the inequities of usury; the bible clearly prohibits usury in Deuteronomy 23:20. For many, particularly the wealthy and those that remain wedded to neo-classical ideas, an appeal to common decency or divine guidance may not be sufficient. So it is worth considering two other, more selfish, reasons for modifying income distributions. The first issue to consider is that strongly skewed income distributions negatively affect rich people as well as poor people, though clearly they affect poor people more than the rich. There are two main reasons that the rich are disadvantaged by skewed wealth distributions, the obvious one is crime, the other, less obviously, is in overall health levels. I will review these very briefly below, for more information; the arguments are discussed at length, in great detail, with much supporting evidence, in the book 'The Spirit Level' by Richard Wilkinson and Kate Pickett [Wilkinson & Pickett 2009]. The issues of crime are easily understood. More unequal societies have much more crime, and higher general levels of aggression and violence. In unequal societies rich people have material wealth, but may have their quality of life significantly reduced through fear of crime. This includes the fear of being attacked in the street or having their homes broken into, and may result in not being able to move about freely or being obliged to live in isolated, highly secure accommodation. 66 EFTA00625194 The data on health is much more counter-intuitive. It is of course obviously plausible that average life expectancy and health outcomes correlate very closely with fairer wealth distributions, and the statistical data supports this. Critically, and quite surprisingly, these statistical benefits are not just due to outcomes in the poorer parts of the populations. Rich people live longer and are healthier in countries like Sweden or Japan that have more equal wealth distributions. In fact often poor people in more equal countries have better health outcomes than rich people in countries with unequal wealth distributions, see for example figures 1.8.1, 1.8.2 and 1.8.3 below from Wilkinson & Pickett. The reasons are not fully clear, but appear to be due to increased levels of stress throughout the whole of society. Figure 1.8.1 here Figure 1.8.2 here Figure 1.8.3 here The second 'selfish' reason for using statistical theory for changing income distribution is that in practice all democracies attempt to carry out income redistribution. Such efforts, by fighting entropy head on, are normally expensive and of limited effectiveness. Ultimately such efforts must be paid for out of taxation, whether they are effective or not. In Europe of course, the welfare state and high taxation are used in attempts to redistribute income. The workings are obvious, as is the expense. Such systems are generally looked down on by individuals from 'free-market' countries such as Hong Kong, Singapore and the US. In fact, even in the most avowedly free-market democracies, leaving things completely to the market has never been acceptable. All democracies put in some sort of support for the poor. Hong Kong famously has very poor benefits for unemployment, but few people realise that about half of the population of Hong Kong live in subsidised public housing. Those that are purchasing property are allowed to offset up to 100% of home loan interest payments against tax up to a maximum of $100,000 per year. The proportion of the population living in subsidised housing in Singapore is even higher than that in Hong Kong [Telegraph 2010b] The US of course publicly repudiates the horrors of providing public housing. Instead for many years they have given covert subsidies to housing of the poor and middle classes indirectly. Americans, though presumably not particularly poor ones, can receive mortgage tax relief on up to $1,000,000 worth of debt on their homes. Also, very large housing subsidies have been provided through the underwriting policies of the GSE's primarily Freddie Mac and Fannie May. The effects of these gross distortions to the market have been disastrous, not just to the US but to the whole world, as the credit-crunch was triggered by the sub-prime mess created by these back door subsidies. Remarkably, the US appears not to have understood the lessons of this recent disaster. I don't know of any country in 'socialist' Europe that uses government backed mortgage insurance, but in the US the future of the GSE's is still under discussion. 67 EFTA00625195 The big problem with all the current forms of welfare, whether overt or covert, is that they don't work. The welfare systems currently used by states around the world fall into one of two classes, either they provide income, in the form of benefits, or they provide subsidies to housing. What poor people actually need is capital. If they had capital, they would have their own income, and if they had sufficient income they would be able to provide their own housing. Simply providing income directly doesn't work. This is because the income will be spent immediately and so the income stream needs to be continuous, and even then will not lift people out of poverty. It is also iniquitous. As the British MP Frank Field has pointed out; effectively, in the UK, welfare claimants are stuck in a poverty trap because the income streams they receive mean they 'own' the equivalent of very substantial capital which amount to 'lifetime pensions'. Subsidising housing is better, but is not ideal. Housing is not real capital (see discussions below in section 6.3) and does not give good long-term gains, and again providing housing at less than its cost means that subsidies are continuous. Housing provided by the state also badly affects freedom of choice, allows social stratification and creates ghettos for the poor with associated problems of crime and restricted economic opportunities. The aim of the proposals in this section is to make the process of aiding the poor much easier, by understanding and so using the statistical mechanics of the economic system. The main aim is to transfer capital to poorer people and ensure that they retain that capital. This would make transfers one-offs rather than continuous. In the longer term this in itself would reduce taxes significantly. If secondary effects include less crime and better health, then total tax takes should reduce even further. From the analysis and modelling in sections 1.4 to to 1.6 above it is clear that there is a fundamental near-fixed nature to the ratio of returns to labour and capital (this is discussed in much greater depth in section 4.5 below). This fixed ratio of returns to labour and capital then gives fixed parameters for the GLV distribution, which in turn gives a fixed proportion of people in poverty, as discussed in sections 1.5 and 1.6. The fixed nature of the ratio of labour/capital returns, and the fixed shape of the GLV distribution necessarily mean that the only way that the elimination of poverty can be achieved is by moving capital into the hands of poorer people. Without changes of ownership of capital, poverty will remain fixed. Other methods of attempting to alleviate poverty will necessarily fail. If these methods involve taxation, then they will fail expensively. As discussed above, I believe the key to eliminating poverty is increasing the amount of capital owned by poorer individuals. One solution to this problem would be to encourage employee ownership much more strongly. For example it would be possible to increase the use of employee share ownership plans (Esops) by giving greater tax advantages to them. A better alternative is to encourage full-scale ownership of companies. In the UK employee- owned organisations currently include companies such as John Lewis, a major retailer and Arup 68 EFTA00625196 and Mott-Macdonald, both of which are major engineering consultancies. Such companies have been very successful in the service sector where capital costs are relatively low and quality of service is key to success. In these companies, profits are normally distributed to employees as bonuses, which are typically paid out in proportion to annual salaries. In 2010 John Lewis staff received bonuses equal to 15% of basic salary, in 2009 they received 13%, in 2008, pre- recession, it was 20%. Although this still results in an unequal distribution of capital, it is a much more equal distribution than that found through the normal pattern of distribution via shares owned by private individuals, which of course is a GLV distribution. Stronger encouragement of employee owned organisations, by the use of tax advantages might in itself be very successful in producing a more equal distribution of wealth. In practice though, it is difficult to see how such organisations could easily raise the capital needed for extractive industries, heavy manufacturing industry, or for that matter companies involved in scientific research or large-scale finance. (Clearly, if such companies use external debt financing for capital investment this just recreates the problem of paying out profits to external capital owners, so recreating the GLV). There can also be very severe problems when people's personal capital is tied up in their employer. In the case of bankruptcy, individuals lose twice over, losing their investments as well as their jobs. Additionally, employee owned organisations do not solve the problem of balancing saving of individuals over the lifecycle. If all companies were employee owned, middle-aged people would not have suitable places to invest their savings for their pensions. (And Robert Maxwell showed that investing your pension in your employer is a profoundly unwise thing to do.) Realistically, for much of the economy there will need to remain a separation of ownership of capital from employment. In practice, I believe the target must be to create a 'virtual 40 acres' of capital for all members of society. The phrase '40 acres and a mule' is 150 years old. In 1865 at the end of the American Civil War, it was the policy of the Northern army to provide freed slaves with 40 acres of fertile land and an ex-army mule to provide a draft animal. At the time it was recognised that this combination was enough to provide a family with a self-sufficient homestead. In practice the policy was not carried out except in parts, and was mostly rescinded even then. As shown in the model in 1.7.2 above, one way of ensuring that people have extra capital is simply to introduce compulsory saving. The main reason for using compulsory saving in this model is simply because it is very easy to model mathematically. In real life such a model would have a big problem starting up. Once it was up and running, and income was already well distributed, then it would be easy to enforce compulsory saving. However trying to enforce compulsory saving, which will feel like an extra tax, on people who are currently poor would be very difficult. It would also have the perverse short-term effect of making people significantly poorer in terms of day-to-day income. A more realistic model for starting the system up would be to introduce assisted saving, where governments allowed tax rebates and/or paid subsidies to people who were saving money. To make such a scheme work effectively, the easy bit is giving assistance to poorer people. The difficult bit is ensuring that the money is not spent as income; to ensure that it is in fact saved. 69 EFTA00625197 Fortunately there are well-established precedents for schemes of this type, most notably pension systems. In most democracies, people who save for pensions are given tax relief and even assistance with their savings. As a quid pro quo for this assistance, governments lay down strict rules as to when and how the money can be withdrawn in old age. From country to country many other forms of government assistance are given, such as tax relief on mortgage payments, tax-free savings accounts or tax-free share ownership (ISA's in the UK), and even assisted savings such as Government Gateway in the UK. Unfortunately such schemes tend to have grown up historically on an ad hoc basis, without any theoretical underpinnings. As such the results have been, at best, haphazard. Taking the UK as an example, a review of who benefits from such schemes, is quite enlightening. Firstly in the UK, individuals are allowed to invest in tax free savings accounts or 'ISA's'. Any individual is allowed to pay in £5,100 per year if the investment is in cash, or £10,200 per year if the investment is in shares. Money can be left in as long as is wanted. If money is removed, it can't be put back in; the ISA allowance is lost. Any dividends or capital growth achieved are completely tax-free. It is rumoured that some successful stock-pickers have managed to accumulate millions of pounds in their ISA's, and are allowed to receive income from these investments tax-free. It is not clear exactly how this contributes to social equity and cohesion. Clearly the ISA system is much more advantageous to the rich who can both save regularly, and are less likely to need to raid their ISA's in the short term. Also tax-free savings are of no benefit to people who are so poor that they pay little taxes. Policy on pensions provision in the UK is even more interesting, though profoundly confusing. (UK pension and tax policy is very complex, if I have made errors in the brief summary below, I would welcome correction.) Individuals in the UK can pay income into a personal pension fund free of tax. If you are a basic rate taxpayer (a poor person), the maximum you can save is 20%. If you are a higher rate taxpayer (a rich person), then the amount of tax relief you can earn can increase up to a maximum of 40% total. Contributions to your private pension scheme are capped each year to your maximum income. So if you are a poor person, you are only allowed to put a small amount in, and receive a small amount of tax relief. If you are a rich person, you are allowed to put a lot in to your pension, and earn a lot of tax relief. This is an important restriction, as it prevents people with variable income from paying money saved from a good year in during a bad earnings year. Sensibly, there is a maximum limit to how much you can save in your pension tax-free each year. The current maximum is £50,000 per year. (This was recently reduced from £255,000 - I am not making this up.) So the maximum subsidy, per rich person, is nearly £20k per year. The average salary in the UK is approximately £25,000 per year. In addition to the above, there is also a 'lifetime allowance' on the total notional size of the pension fund, and pension receipts from the part of the fund above this allowance are subject to income tax. The lifetime allowance is currently £1.5 million. Even on an interest only basis, assuming no draw down of the fund, at 3% real interest rates this would allow a tax-free pension of roughly double the average UK salary. 70 EFTA00625198 The 'aim' of all these subsidies to the rich is to avoid people being dependent on state pensions in their old age. The current maximum UK basic state pension is £97.65 a week, so if a person retired at an age of 65 and lived for thirty years, the cost to the state would be roughly £150k. Even including for housing benefit in rented accommodation the cost would be less than £300k. It is not clear to me that the 'aim' of saving money for the state is being successfully achieved. All the above system was put into place and managed under the Labour government of 1997 to 2010, notionally a social democratic, if not socialist party. Perhaps due to a concern with the above largesse lavished on the rich, the same government also introduced an assisted saving scheme called the Savings Gateway. To qualify for the Savings Gateway you must earn less than £16,040 per year, and must also be claiming some sort of benefit. The maximum payment into the scheme is £25 per month. For every £1 that a participant saves, the government will add a further 50p. So the maximum subsidy, per poor person, per year is £150. Whether the Savings Gateway proves to be successful in helping to reduce poverty remains to be seen. I, for one, am not holding my breath. This disparity in assistance for the rich and the poor is not restricted to the UK, this from the Economist in 2005: Politicians' main method for boosting thrift is a swathe of tax-advantaged retirement accounts. This year these accounts will cost some $150 billion in foregone tax revenue. Most of this subsidy goes to richer Americans, who have higher marginal tax rates and who are more likely to save anyway. Only one saving incentive—the Saver's Credit—is targeted at poorer Americans. It is worth only about $1 billion in forgone tax revenue and is due to expire in 2006. And even that offers no incentive to the 50m households who pay no income tax. [Economist 2005]. The report 'Upside Down' gives a detailed analysis of how the majority of assistance given to working families in the USA ends up in the hands of the rich [Woo et al 2010]. While the efficacy of the many different policies used above can rightly be questioned, the important point is that the financial tools and institutions needed for creating private capital for all members of society are already available. Interestingly perhaps the best example of such a system is one initiated by a group of radically right-wing free market economists. The Chilean pension system that the 'Chicago boys' created for dictator Augusto Pinochet in Chile works in exactly this manner. In Chile, all salaried workers are forced to pay 10% of their salary into one of a number of strongly regulated pension funds. The pension funds in turn invest in private companies through the stock market, bond purchases, etc. The pension funds are strictly regulated, and individuals are allowed to switch easily between different suppliers. 71 EFTA00625199 The major difference between the Chilean pension scheme, and my proposed 'virtual 40 acres' (henceforth 'v40') is that part or all of the interest from the capital, and some of the capital, would be made available during the normal working life of an adult. A rough outline of the 'v40' is as follows. The v40 would consist of a pot of money, held with an officially sanctioned investment fund exactly like those that operate in Chile. The funds would have controls on appropriate investments and proportions of investment in different assets, as is normal with regulated pension funds. At any one time there would be a maximum amount that could be held in the v40, for the present discussions the maximum amount will be assumed to be £50k. This is approximately twice the average annual wage, and as an investment sum it is not particularly large. There is an important reason for this small proposed size, this is discussed later. All people who are in paid employment would be obliged to pay into their v40 at a minimum rate of say 10% of salary. This would apply to all people who had not got a full pot of £50k invested in their v40. Note that people who had the full £50k invested would not be obliged to pay into their v40 pot; in fact people with a full v40 pot would be specifically prohibited from paying further into their v40. To make this compulsory saving more palatable, all payments into the v40 would be before tax and any other payments such as social security. Similarly all interest payments, and eventually capital repayments out of the v40 would also be free of income and capital gains or any other taxes, provided they had been invested for a minimum period of say five years. There would be no limit to the amount paid into the pot each year, up to the total limit of £50k, and all payments up to this amount would be tax-free. (In the UK for example, all current ISA holdings, up to £50k, could be transferred over into the v40 tax-free. ISA's would then be discontinued as a tax-free vehicle.) For poorer people, two further regimes are proposed. Here poorer can mean one of two things. Firstly it can mean people who have low levels of savings in their v40, and so low income from the v40. Secondly it can mean people who have poor employment income, either through low skills level or because of intermittent employment. In practice either or both of these definitions may apply to the 'poor' and 'very poor' discussed below. For 'poor' people further assistance can be given by allowing payments to the v40 account to be counted as an alternative to taxation. So if a poor person is paying 10% of their salary into a v40, then they would have their 'normal' taxation reduced by the same amount of money. For 'very poor' people the government would follow the ideas of the 'Savings Gateway' and other similar schemes, and pay matching amounts to give assisted saving, so helping the very poor move into the category of simply poor. With regard to withdrawals, a portion of interest payments could be withdrawn immediately, but on a sliding scale with strict rules. So the percentage of interest earned that could be withdrawn each year would vary as the percentage of the total v40 allowance held. To take some examples. Assume that the real interest is at 3% per annum (halfway between long-term US and UK rates, see section 4.5 below). Assume also that the v40 limit is £50k. 72 EFTA00625200 If somebody had a full pot of £50k invested in their v40, then they would earn £1,500 interest per year, and would be allowed to take the full amount out each year as tax-free income. In fact they would be obliged to remove this interest, and any capital accumulation above the £50k, from the account. If somebody had saved half of their v40 allowance or £25k, then they would earn £750 in a year, and would be allowed to remove half of this interest, or £375. The remaining £375 would be automatically reinvested as capital in the v40. Clearly there would be no compulsion to remove any of the interest. If somebody had only £10k in their v40, or 20% of the allowance, then they would earn £300 interest. They would only be allowed to remove 20% of this interest, or £60, with the remaining £240 of interest being reinvested as capital. Finally, to further discourage early removal of interest, a minimum five-year period should be included with punishment of taxation if the interest is removed within five years of it being earned, for 'normal' investors. Or, in the case of 'poor' investors, a reward if the accrued interest is held in the account for a minimum of five years, similar to the 'Savings Gateway' scheme. Note that this punitive taxation would not apply to those who have reached the maximum of the v40 pot. While the above may seem somewhat complex, the aim of all the detail is the same. All the incentives, for rich or poor, are to encourage people to save as much money in their v40 as they can, as quickly as possible. It is hoped that in this manner the v40 will be seen as a sensible way to build up capital by all members of society, even the poorest. While the v40 is being built up, a portion of the accrued interest will be available for removal, as emergency funding, in the case of a financial crisis. But the incentives should encourage such use only in genuine emergency. Once the v40 allowance has been fully reached, then the fund becomes a useful additional income support. At this point, removal of interest and capital gains would become compulsory, and would need to be spent as consumption or moved into private investments that do not attract tax exemption. With regard to removal of capital, it is suggested that rules along the lines of the following are used. Firstly no capital can be withdrawn until a minimum age of say forty years. After that age, capital can be withdrawn according to a set rate depending on the notional length of time that the v40 account will be held. A notional date for the end of the account is assumed, this effectively being a notional date of decease of the account holder. This could be say the age of 80 years old, or ten years older than the current age, whichever is the larger. The amount of capital that could then be withdrawn would be the reciprocal of the number of years between the current age and the notional end date. So if the owner of the v40 was forty, and the notional end date was 80, the difference would be 40 years, and the holder would be allowed to remove 1/40th of the value of the v40's capital, in addition to the allowed interest. At sixty years old the holder would be allowed to remove 1/20th of the value of the v40. From age seventy onwards the holder would be allowed to remove 1/10th of the value of the v40. This 73 EFTA00625201 would be the maximum amount of capital that could be removed from the account at any time. Removal of capital would not be compulsory. Following the decease of the v40 account holder, all the value of capital would be inheritable. This would be fully tax free, including free of death duties, providing that the v40 money was passed to other individuals, with sufficient spare allowance, for transfer into their own v40's. If the capital was brought out of the protection of the v40 system, it would be taxed, and subject to death duties, as normal capital. Finally there is one subtlety that needs to be controlled if the v40 scheme is to be effective. It is not sufficient simply to prevent people running down the capital in the scheme and using it as income. It is also essential that people be fully prevented from using the capital in the v40 as collateral against which they can borrow money. This would destroy the v40 scheme by allowing savings to be converted into income. The best way to do this is to allow relatively lax personal bankruptcy laws and to specifically exempt money invested in a v40 from being included in bankruptcy cases. That is, even a person who has been made bankrupt is allowed to keep the full value of their v40 intact. If this is put into place, then it will not be possible to secure loans made to an individual against their v40, as such loans will be extinguished in the bankruptcy. In such circumstances individuals should not be able to get loans against v40's. Protection in this manner will also have the advantage of encouraging use of the v40 as a savings vehicle. The net result of this is to have something that works in very similar manner to a pension scheme, but also has characteristics similar to that of an employment insurance scheme. It is aimed to meet basic and/or emergency needs throughout a working life. As such it can be seen as a 'personalised' welfare scheme, and at least in part, can form an effective 'personalisation' of welfare. By handing the main responsibility for management of this 'welfare' to individuals it should be much more effective than state run welfare schemes that lose the link between contributions and benefits. Despite this 'personalisation' it has to be stated in the strongest possible terms that the iron law of the GLV means that some form of government action will always be necessary if such a personalised form of welfare is to succeed. As an absolute minimum, a government would need to strictly enforce compulsory saving to ensure that such schemes operate. It seems more realistic that general tax advantages, assistance for the poorest and a backstop of enforcement will be the most effective policy mix to ensure the v40 operates effectively. To give an example of how this could work, I would like to take Norway as an example, though as will be seen later this is not quite a reasonable choice. Norway is of course very rich. Not only does it have a very well run Scandinavian social and political system, it has also enjoyed four decades of oil production. Despite this, Norwegians still have problems of relative poverty, where depending on definitions, between 4% and 10% of the population have less than 60% of median earnings [EWCO 2010]. Given the very high costs of living in Norway this relative poverty can be debilitating. Poverty in Norway was seen as a priority for the incoming government in 2005. 74 EFTA00625202 As a result of careful saving, by successive governments, Norway now has a sovereign wealth fund of more than three trillion Norwegian crowns, equivalent to about 500 billion US dollars. The population of Norway is 4.7 million, which must mean there are roughly 3.5 million adults. Using these figures the sovereign wealth fund is worth about $130,000 per person. So trivially, the Norwegian government could simply create 3.5 million v40 accounts tomorrow and give each Norwegian adult $140,000 worth of assets to hold in the account. This isn't actually very sensible, as many Norwegians are already quite wealthy and don't need to be given all that money. Let's assume that say 20% of Norwegians are quite rich and have many assets to hand which they will be happy to transfer into a tax free v40 account given the opportunity. Let's assume 20% of Norwegians are comparatively poor and need to be given their full v40 allowance by the state. Finally we will assume that the remaining 60% of Norwegians are middle income and that they will only need an incentive to transfer their savings and/or income to their v40's. Suppose this is a tax-free incentive of equivalent to 30% of the v40 investment. This means the Norwegian government can make its sovereign wealth fund go much farther, actually more than two and a half times farther. So now the v40 allowance can be set at about $375,000 per head. If we again assume that long-term real interest rates are 3%, then this gives each and every Norwegian adult an independent income of $11,000 per year. Just for comparison, a quick look on the internet suggests that rents in Oslo for a 3-bed apartment are currently about $1000 per month, so such an income would pay most housing costs. But then if you lived in a beautiful country like Norway, and you had an independent income, why on earth would you live in Oslo. From my own limited knowledge of Scandinavian culture, a surprising proportion of Scandinavians have second homes hidden away as rural retreats. With private income like this, if Norwegians moved to the countryside; apart from childcare, hospital care and care for the elderly; the whole of Norway could pretty much retire, and live, a little frugally, on their investment income. There is, of course, no reason to stop at this point. The Norwegian government could still oblige all Norwegians to continue investing a portion of their earnings in their v40's. By enforcing some short term frugality, and maybe even working a couple of days a week, Norwegians could be forced to further increase the value of their v40's, making the whole country richer and richer. Although this should work for Norway, there is a significant problem with expanding such schemes on a global basis. Going back to the UK example given above, I set the v40 allowance at £50k per year. Using long-term UK interest rates, this gives an investment income of £1500 per year. A typical rent in the midlands of the UK would be in the region of £500 per month for a two bed flat. Even with two adults, £3000 a year would only cover half a year's rent, never mind other living costs. While this money would be very helpful, it would fall far short of being truly a 'virtual forty acres'. Even sharing housing costs, and living very frugally, it is not possible to survive in the UK on £1500 per year. In fact £30 a week would hardly cover food and utility costs even if you owned your own home. 75 EFTA00625203 I chose the value of £50k for an important reason. The stock market capitalisation of the top forty UK companies is in the region of £1000 billion, if we assume the total capitalisation is double this, a brave assumption, then the total wealth available for investment in the UK is £2000 billion. The population of the UK is 61 million, or say roughly 50 million adults. So the available capital on the UK stock market for investment in v40's is about £40k per head. This assumes no other investment use for this capital, such as, for example pensions. Alternatively in 2009 UK gdp per head was roughly $35,000 per head [Economist 2010c]. Assuming that total non-residential capital per head is roughly 2.5 times gdp per head [Miles & Scott 2002, 5.1 or 14.1], this gives $88,000 capital per person, or roughly £57k per person. Another calculation; the Halifax Building Society [BBC 2010a] estimates that total UK personal wealth amounts to £6.3 trillion or £237,000 per household, however more than a third of this is in the form of housing. A large part of the rest will be in pension funds. If one third is in housing, that leaves £158k per household. Assuming 2 adults per household this gives £80k per adult, which gives ball-park agreement with the figures for stock market capitalisation above. This leaves us with a basic problem. If UK capital is used for UK savings, there simply isn't enough wealth per person, even if it is shared out absolutely equally, to give a modest investment income for every person. And of course a major part of the current capitalisation is already tied up in pension funds and is committed to future retirement needs. This actually is obvious if you go back to Bowley's rule as discussed in sections 1.3 and 1.6 above. Historically, in capitalist societies, total returns on capital are roughly equal to half of the total returns to labour. So even if capital was shared absolutely equally to all individuals, it would only be equivalent to half their wages. With present levels of capital it would not be enough money to live on. Norway's sovereign wealth fund represents a special case. Most of the investments in Norway's sovereign wealth fund are invested in companies outside Norway. So most of the investment income accruing to Norway comes from other countries. Interestingly this means that egalitarian, liberal Norway, with it's generous high per capita spending on foreign aid, is probably the world's most effective, and most discrete, neo-colonialist nation. This general problem of insufficiency of capital will be returned to in depth in section 4.8 below. 1.9 Wealth & Income Distributions - Loose Ends Before leaving discussions of income modelling I would like to briefly discuss two areas of income distribution that I have not been able to model successfully, but which I think are of importance. 1.9.1 Double Power Laws 76 EFTA00625204 Back in section 1.1 above, it was noted that some researchers have noted that there appears to be a split in the power tail of income distribution into two or even three separate sections. This appears to give a split between the 'rich' and the 'super-rich'. Some models have been proposed for this, of varying plausibility. It is possible that this arises simply from the basic models above. Figure 1.9.1.1 here For example, figure 1.9.11 above for model 1E is simply a rerun of model 1D but with larger spreads on the normal distributions for consumption. Figure 1.9.1.1 is a log-log graph, with a long power tail that shows two or possibly three different straight line zones. It is likely that a more realistic log-normal distribution would exaggerate this effect. Another possible source of different power laws is the consumption function. All the models in this paper have used a savings/consumption function that is strictly proportional to wealth. This has the value of simplicity, but may not be realistic. Common sense suggests that the more wealth people have the smaller the proportion of their wealth they consume and the greater the proportion they will save. Note that rich people are assumed to spend more as they get richer, just that the extra spend is not as big as the extra wealth. It should be noted however that this assumption is controversial, though recent research findings tend to support this assumption [Dynan et al 2004]. The idea that consumption functions are concave in this manner seems so obvious that it has in fact been proposed as a source of wealth condensation effects. Clearly this paper has demonstrated that this mechanism is not necessary. During modelling for this paper, an attempt was made to run income models that included concave consumption functions. The results suggested that concave consumption functions did indeed produce a two-section power law. However the results were highly unstable; small change in parameters could result either in a return to a single power law, or collapse of the distribution to a single wealthy individual. The results were not sufficiently strong to justify presentation here, but they do suggest that this is a possibly useful area for future research, given access to better data to calibrate the models with. Finally, while discussing the role of consumption and savings functions, it is worth noting that there is little role for being judgemental with regard to savings. It is very easy to suggest that it is the fault of poor people for being poor if they do not save for the future. But as has been seen in previous income models the rewards for saving are disproportionate. 77 EFTA00625205 While it the form of savings functions are still up for debate, it is clearly easier to save a portion of your income if your income is higher. Indeed, in the exact opposite of the '40 acres' model, in normal life people face a 'compulsory spending' world. People are obliged to spend a minimum amount of money on food, clothing, housing, heating costs, transport, etc. This compulsory spending will have exactly the reverse effect of the compulsory saving of section 1.7.2 above; it will make inequality worse. Rich people have more discretionary spending, which makes saving easier. On top of this, as Champernowne pointed out, the role of inherited wealth gives an enormous advantage to the better off. 1.9.2 Waged Income The second loose end is potentially much more interesting, and relates to the payment of income in the form of wages and salaries. In all the models in this paper, wage distributions have assumed to be either uniform or normal distributions. The uniform distributions are clearly very unrealistic. They were used primarily for simplicity, and also to demonstrate very clearly that gross inequalities of wealth could be produced with absolutely identical individuals. The normal distribution was used in the more realistic models primarily to avoid controversy, and to provide a useful comfort blanket to any economists still reading the paper. In fact a log- normal would probably have been a more realistic choice, as per figures 1.1.1, 1.1.2, 1.1.4 & 1.1.5. The author has looked at a comparison of the log-normal and the Maxwell-Boltzmann distribution for describing income distributions applied to high quality data sets from the UK and US [Willis & Mimkes 2005]. From this I am firmly of the belief that waged income is distributed as a Maxwell-Boltzmann, or rather a Maxwell-Boltzmann like distribution. The main reason for this is that the Maxwell-Boltzmann distribution is inherently a two- parameter distribution, unlike the log-normal which is a three parameter distribution. So the Maxwell-Boltzmann is inherently simpler than the log-normal. Another way of thinking about this is that the log-normal can take many different shapes, the Maxwell-Boltzmann only has one. It is an extraordinary coincidence that two completely separate sets of data from the US and UK can be fitted by the only log-normal, out of all possible log-normals, that can fit a Maxwell-Boltzmann distribution exactly. There is however one small fly in the ointment for these Maxwell-Boltzmann distributions (and also for the equivalent log-normal distributions). The Maxwell-Boltzmann distributions in income distribution show a significant offset from zero, something that is not normally seen in physics applications. Or indeed in physics theory; which in these models usually uses pure exchange processes subject to conservation principles (much more on this below in section 7.3). With their offsets and their exponential mid-sections, these 'Maxwell-Boltzmann' distributions in fact look very like GLV distributions, but of course without the power tails. 78 EFTA00625206 It is my belief that these distributions are in fact the product of a dynamic equilibrium process that produces an 'additive GLV' distribution, in contrast to the normal 'multiplicative GLV' distributions, that have been seen throughout this paper. A possible explanation for this is discussed in section 7.4 below, though this is highly speculative. Although speculative, I believe that this might be an important line of research. It also raises some important philosophical questions on the nature of inequality. If the distribution of income is a log-normal, then it could reasonably be suggested that the distribution arises from the inherent skills possessed by the individuals, which following the central limit theorem, could reasonably be distributed as a log-normal. This would make the distribution of wages exogenous to the models, as in fact they have been modelled in this paper. I personally am not convinced that the log-normal found in income distributions is exogenous. My personal experience of human skills is that the majority of human beings fall into a narrow band of skills and abilities; more like a normal than a log-normal, with a very large offset from zero. Fig 1.9.2.1 below shows my assumption of how skills might reasonably be distributed. Figure 1.9.2.2 gives the example of height. Figure 1.9.2.1 here Figure 1.9.2.2 here [Newman 2005] Intuitively, intelligence and other employment skills seem likely to be distributed in a similar manner. If the distribution of income is in fact a Maxwell-Boltzmann-like 'additive GLV', this would put a very different light on things. Such a GLV would be an outcome of a dynamic equilibrium process and would be created endogenously within the economic model. The consequences of income distribution being an endogenously created GLV are simple. It means that poor people are being underpaid for the labour, and better off people are being overpaid. It means that capitalism doesn't reward people fairly, even at the level of waged income. Clearly before such a bold statement can be made, it would be appropriate to produce a meaningful model for producing an 'additive GLV'. Notwithstanding these loose ends, we have effectively dealt with the problems of poverty. Time now to investigate some other problems in economics. 79 EFTA00625207 2. Companies Models Going back to figure 1.3.5, having looked in detail at the wealth and income distributions, we will now move our interest from the wealth owning individuals on the right hand side of the figure 1.3.5 over to companies, the source of wealth, on the left hand side. 2.1 Companies Models - Background The theory of the firm has long been recognized as a weak point of neoclassical theory. The paradigmatic case for neoclassical theory is the competitive industry, in which a large number (how large is open to considerable discussion) of similar firms coexist. Neoclassical theory roots its explanations in properties of resources, technology and preferences that are independent of the organization of economic activity itself (that is, are exogenous from the point of view of economic theory). What technology could give rise to the coexistence of many similar firms in an industry with free entry? If there are diminishing returns to scale, the industry should be atomized by the entry of ever-smaller rivals. If there are constant returns to scale, the theory cannot explain the actual size distribution of firms except as an historical or institutional datum. If there are increasing returns to scale the theory predicts the emergence of a few large firms, not the competitive market originally posited. [Foley 1990] As discussed previously, it is the belief of the author that firms exist to protect their value- increasing property, their sources of negentropy. Firms buy goods that have well defined prices such as raw materials, components, electricity and labour. They then use these inputs to go through a series of intermediate goods stages with, at best, indeterminate prices, at worst, very low prices. As an obvious example think of a car body shell, which has its engine and transmission installed, but hasn't yet had its electrics, glassware, finishes etc, installed. To the manufacturer it probably has more than two-thirds of its true value installed, in terms of components and labour supplied. However if it were sold on the open market it would have very low value, even to another car manufacturer, as the cost to completion for another company, or an individual, would be very high. To complete the process of production successfully, a company has to finish the goods to a well- defined point, where they can be easily priced in the market and sold to consumers or to other companies as intermediate goods. The company, with its plant, trained workforce, patents, designs and trademarks, exists to protect this wealth creation process. In neo-classical economic theory, as discussed above by Foley, the sizes of the companies should either be very small if entry to markets is easy, or very big and monopolistic, depending on the returns to scale. In fact it is well documented that company sizes, whether measured by number of employees or capitalisation follow well defined power law distributions. For background see Gabaix [Gabaix 2009] or [Gaffeo et al 2003]. 80 EFTA00625208 These power law distributions are of course similar to the power law distributions of wealth for property owning individuals that we have seen in the discussions of wealth and income above. The model for companies in this paper builds on the income models introduced in section 1.3 above. The modelling looks at company sizes in terms of total capitalisation K of the companies. To extend these models, three basic assumptions are made. Firstly, in a break with the previous models, it is no longer defined that the valuation of the paper assets W matches the real capital of the company K. That is to say the short-term stock- market price W is allowed to vary significantly from the 'fundamental' value of a company's real capital K. As well as introducing this degree of freedom, three further important assumptions are introduced. Firstly, it is assumed that shareholders are myopic, and judge expected company results simplistically on previous dividend returns. Secondly, it is assumed that managers of companies act to preserve the stability of dividend payouts, Thirdly, and more importantly it is assumed that managers act to preserve the capital of their companies. Justifications for these assumptions are given below. Until a few years ago, despite the wealth accumulated by Warren Buffet and other acolytes of the Benjamin Graham school of investing, the concept of companies having fundamental value was highly controversial. In recent years, these views have become more acceptable for discussion, firstly following the dramatic changes in value during the dotcom and housing booms of the last decade, and secondly because of the detailed research of Shiller, Smithers and others that both disprove a purely stochastic basis for stock market movements and also give substantial evidence for long term reversion to mean for stock market prices when measured by Tobin's q or by CAPE; the 'Cyclically Adjusted Price to Earnings ratio'. This is discussed at length in Smithers [Smithers 2009] for example, and is looked at in more detail in section 8.2.1 on liquidity, below. Following the credit crunch and the dramatic changes in prices associated with liquidity problems, ideas of fundamental values have become more acceptable. Following the recent work of Smithers, Shiller and others, and also the beliefs of the classical economists, this section takes as it's starting point the viewpoint that economic companies do have 'fundamental' values, and that these are frequently at odds with their stock market valuations. With regard to myopic behaviour the book 'Flow Markets Fail' by John Cassidy [Cassidy 2009], gives an extensive discussion of data that gives evidence for short term pricing behaviour. This is discussed in depth in chapter 14. It appears that this naive behaviour is not restricted to naive investors. Recent work by Baquero and Verbeek for example [Baquero & Verbeek 2009] suggests that pension funds, private banks and wealth individuals all commonly invest based on short term returns. 81 EFTA00625209 In their paper 'The Cross-Section of Expected Stock Returns' [Fama & French 1992] Fama and French, originators of the efficient market hypothesis, carried out econometric analysis that confirmed that four empirical factors appear to be involved in the pricing of stocks. The first of these is the risk associated with stocks, in line with the original capital asset pricing model (CAPM). The second is the size of the company. The third is the book to market value of the company. The fourth factor identified to fully explain stock market valuation is the presence of short-term momentum in pricing based on recent returns of the stock. The work of Korajczyk and Sadka [Korajczyk & Sadka 2005] also suggests that momentum is important in company valuations and arises from liquidity considerations. Recent academic work suggests that both size and book to market effects can be explained by changes in liquidity. This is potentially a very important topic, and is discussed at some length in section 8.2.1 below. For the companies model, liquidity, and so company size and book to market values are assumed to be irrelevant. It is assumed that liquidity is constant throughout the modelling process. As modelled by the CAPM, risk is peculiar to individual companies. In this model it is assumed that risk is identical, and in fact zero, for all companies in the model. Given the above assumptions of zero risk and high liquidity; following Fama & French, this leaves short term returns as the only factor that investors use to value companies. So, using basic finance theory, then the present value of a company is given simply by: Present Value = Dividend1 r Where r is the relevant market interest/profit rate; Dividends is the latest dividend payment, and capital growth is ignored. See for example [Brealey et al 2008, chapter 5]. This is the naïve neo-classical approach to valuing capital for aggregation; simply divide by the profit rate. We will simply take this naïve approach as it stands and follow the consequences through the model. With regard to management behaviour, research from Bray, Graham, Harvey and Michaely [Bray et al 2005] support the contention that maintenance of a constant dividend stream is an important priority for managers of corporations. Finally, with regard to the retention of capital within companies, the history of the defence company General Dynamics, gives a very interesting case study. General Dynamics (GD) are interesting in that GD formed a casebook example of how companies are supposed to behave, according to finance textbooks, by working solely to enhance the value of shareholder's stock. In the real world, GD are notable in their exceptionalism, in that their deliberate downsizing to enhance profitability was not only unique in the defence industry, but pretty much unique in corporate history. 82 EFTA00625210 In contrast to GD, other defence contractors in the 1990s followed deliberate policies of acquisition or diversification in order to maintain their size. This despite the obvious collapse of the defence market following the end of the Cold War. The following are quotations from 'Incentives, downsizing and value creation at General Dynamics' by Dial and Murphy: In the post-Cold War era of 1991, defense contractor General Dynamics Corporation (GD) faced declining demand in an industry saddled with current and projected excess capacity. While other contractors made defense-related acquisitions or diversified into non defense areas, GD adopted an objective of creating shareholder value through downsizing, restructuring, and partial liquidation. Facilitating GD's new strategy were a new management team and compensation plans that closely tied executive pay to shareholder wealth creation, including a Gain/Sharing Plan that paid large cash rewards for increases in the stock price. As GD's executives reaped rewards amid announcements of layoffs and divestitures, the plans became highly controversial, fueling a nationwide attack on executive compensation by politicians, journalists, and shareholder activists. Nonetheless, GD managers credit the incentive plans with helping to attract and retain key managers and for motivating the difficult strategic decisions that were made and implemented: GD realized a dividend-reinvested three year return of 553% from 1991 to 1993—generating $4.5 billion in shareholder wealth from a January 1991 market value of just over $1 billion.1 In the process, GD returned more than $3 billion to shareholders and debtholders through debt retirement, stock repurchases, and special distributions. [Dial & Murphy 1994] In contrast to the explicit strategy of creating shareholder value initiated by General Dynamics, this was the behaviour followed by their competitors: Table 7 summarizes the strategies selected by GD and eight other defense contractors from 1990 through 1993, based on an analysis of quantitative financial data as well as our qualitative interpretation of annual reports, press releases, and news articles. The table includes the nine largest domestic defense contractors (ranked by cumulative 1989-1992 defense contracts). Exceptions are General Electric and Boeing, excluded because their defense operations account for less than 10% of total firm revenues. Some of the strategic options adopted by these firms include: Acquisitions to achieve critical mass; diversification into non defense areas, or converting defense operations to commercial products and services; globalization, i.e., finding international markets for defense operations; downsizing and consolidation; and exit Diversification and commercialization. A 1992 survey of 148 defense companies sponsored by a defense/aerospace consulting firm found that more than half of the respondents report past attempts to "commercialize" (i.e., applying defense technologies to commercial products) and more than three-quarters predict future commercialization. Martin Marietta CEO Norman Augustine, however, cautioned his industry counterparts about wandering too far from their areas of expertise: "Our industry's record at defense conversion is unblemished by success. Why is it rocket scientists can't sell toothpaste? Because we don't know the market, or how to research, or how to market the product. Other than that, we're in good shape." ...Globalization. A number of firms are retaining a defense focus, attempting to bolster sales through globalization, selling U.S. built weapons abroad. This strategy is unlikely to yield 83 EFTA00625211 dramatic growth, since the demand for weapons is declining world-wide and many foreign countries have their own national producers who are also faced with excess capacity Downsizing, consolidation and exit. Table 7 shows that while most contractors adopted a combination of strategies, all adopted some form of downsizing or consolidation to reduce excess capacity. However, while a few contractors (including GM Hughes, Grumman, and McDonnell Douglas) have divested unprofitable non core businesses where they had little chance of building strategically competitive positions, only General Electric (not included in table 7) followed GD in exiting key segments of the defense industry. Interestingly, it was General Electric (where Anders held his first general management position) that pioneered the "#1 or #2" criterion as a strategic assessment for the composition of its portfolio of business units... ...Goyal, Lehn, and Rack (1993) also analyze investment policies in the defense industry. They report evidence that defense contractors began transferring resources from the industry as early as 1989-1990 through increased leverage, dividends, and share repurchases. Our complementary evidence suggests that although other contractors also espoused and eventually adopted consolidation and downsizing, GD's response in moving resources out of the industry was quicker and more dramatic. To draw an analogy: While other defense contractors engaged in a high-stakes game of musical chairs—hoping to be seated when the music stopped—GD pursued a strategy of offering its chair to the highest bidder. [Dial & Murphy 1994] Despite the obvious and dramatic decline of the defence industry following the end of the Cold War, and even despite the example of General Dynamics, the managers and directors of some of the largest and most important companies in the world's largest economy followed a clear pattern of attempting to maintain the size of their companies, without regard to the value of their shareholders investments. It is the belief of the author that this pattern is widespread throughout the management of limited companies, and so this will be used as a base assumption of the companies model that follows. 2.2 Companies Models - Modelling Figure 2.2.1 here Figure 2.2.1 above is a slightly modified version of figure 1.3.5. A few changes have been made, though the overall process is the same. We are now looking at the financial assets from a company point of view, and we are not interested in the individuals. So we now have a total of N companies, which we count from j=1 to j=N. The big difference with previous models is that we removed the assumption that K = W or that k, = 1/45. 84 EFTA00625212 So here we differentiate between the fundamental value of the real capital k, formed of the firms buildings, plant, patents, etc and the market valuation of the company w3. w) represents the sum of the stock market value of paper share certificates held by the owners of the company. (Note here that w; is the total wealth represented by all the shares in company j held by various different individuals — is not the same as w,.) At the beginning of each simulation we start with Ek, = K for all the companies, and also Zw, = K initially. That is, to start with, all the companies are the same size, and all are valued fairly by the stock market, with the fundamental value of each company equal to its market capitalisation. It is assumed that each of the j companies has a standard rate of growth r). The average ri will be 0.1, that is each company produces value roughly equal to 10% of its capital each year. So each of the companies is identically efficient in the use of their capital. However, to introduce a stochastic element, we will allow a normal distribution in the values of r, with a variance which is 20% of the value of r. So r varies typically between 6% and 14%. Effectively this assumes that although companies return the same on capital over the long term, they may have short-term good and bad years which allow returns to fluctuate slightly around the long term average. It is assumed that the market is not well informed about the fundamental value of individual companies. Following the research of Fama & French and others, it is assumed that investors simply use the average market rate of returns (0.1 or 10%) as their guide for valuing companies. So the new market capitalisation w) for each iteration of the model will simply be the last actual real returns Tem divided by the long-term rate of returns. so: w 3.1+1 r Then the expected returns for the next year will be the market capitalisation Wj multiplied by the average market rate of return. so: Ti = W j.14-1 1•I Which is an unnecessarily complicated way of saying that next year's expected returns will be the same as the previous years actual returns. As in the previous models, we will assume that labour is fairly rewarded for the amount of added value that it is supplies. So L = e exactly, and both L and e can be ignored in the mathematical model. 85 EFTA00625213 The loop of the simulation was carried out as follows: The amount of production is calculated by multiplying the capital of each company by the relevant production rate, so: production = k j., rj., After a round of production all of the companies will receive cash from purchasers of its manufactured goods. This cash value will represent the value added in the production process. Each of the companies will have a value of expected returns ( ) based on its current market capitalisation. In the simulations carried out actual payouts of profit t were varied by using different payout ratios. If the value added; the production, is greater than the expected returns then the managers might pay out 90% of the earnings, retaining 10% of the extra value, so allowing a buffer to be built up against future problems, also to allow expansion of the company, empire building, etc. This extra value is added to the total capital. If the managers only pay out 90% of the earnings, this is defined from now on as an 'payout ratio' of 90%. The model allows different payout ratios on the upside and downside. So managers may have an upside payout ratio of 90% and a downside payout ratio of 80%. This would mean that the management would pay out 90% of the earnings if earnings were greater than market expectations, but would only pay out 80% of earnings if earnings were less than market expectations. For example in model 26 both the upside and downside payout ratios were 90%. These actual payouts then give the market its new information for resetting the market value w, of the various companies. The capital k, of each company is then recalculated as follows: 1( 3.1+1 = k j., + production - actual_returns Finally at the end of each round the values of the company capitalisations have to be normalised. The reasons for this are as follows. This model assumes a stationary economy with a fixed total amount of capital K. This capital can be bought and sold between different companies, as they are required to give earnings in requirements of market expectations. All of the companies will receive cash from purchasers of its manufactured goods. This cash value will represent the value added in the production purpose. 86 EFTA00625214 Some companies will receive more cash than they are expected to payout, some will receive less. It is assumed that the cash rich companies will purchase real capital off the cash poor companies, so allowing the cash rich to expand, and the cash poor to pay their earnings. At each round of the modelling process, the sum of the capital is renormalised to the original K. This is because asymmetric retention of funds allows excess growth or decline for the whole economy. Ideally a more realistic model would automatically adjust these processes. However, this is problematic, there are deeper, and interesting, instabilities at work, these are the subject of models in section 4 below. 2.3 Companies Models - Results 2.3.1 Model 2A Fully Stochastic on Production, No Capital Hoarding Model 2A is the simplest model, so simple that it inevitably fails. Firstly the model is completely stochastic. Each company produces output worth exactly 10% of its capital on a long-term average. However the value of 10% varies up and down stochastically according to a normal distribution. In model 2A the payout ratio is deliberately set at 1. This means that the managers of the companies payout the full amount expected by the market. They do this no matter how well, or how badly the companies perform. Figure 2.3.1.1 shows the full log-log distribution of all the (non-negative) companies. Figure 2.3.1.2 shows the power tail with the trend line fit for the power tail. Figure 2.3.1.1 here Figure 2.3.1.2 here Companies that lose money, due to poor production, still pay out to market expectations, so they slowly drain their capital and lose it to other companies that have above average production. Because of this the model is not stable, and the distribution changes as the model progresses. Despite this, it is noticeable that the model quickly generates a stable power tail with an exponent close to —1; close to the value seen in real life. The power tail remains stable from 10k to 50k iterations. Above 50k iterations the number of companies being eliminated (going negative) becomes very large and the transfer of capital to the larger companies starts to change the exponent of the power tail. The important thing to note here is that a very simple model, using the standard valuation system of capitalism, quickly generates a power tail of companies of vastly different sizes. In the 87 EFTA00625215 50k, run power tail companies vary in their capital between 80k units and 80,000k units. But all the companies are absolutely identical in their earning ability, effectively the companies have identical managements making identical products with identical inputs. The differentiation in size has only occurred through the stochastic forces of chance. 2.3.2 Model 2B Fully Stochastic on Production, Capital Hoarding Model 2B is identical to model 2A in that the companies are identical in average earnings, but these earnings vary stochastically from model to model. Model 2B is different in that the payout ratios were changed in an attempt to create a stable model. Unfortunately this proved difficult. The only values that prevented 'washout' of smaller companies were payout ratios of 0.9 on both the upside and downside. Initial investigations suggest that this is related to the production rate of 0.1. The results are shown in figures 2.3.2.1 and 2.3.2.2. Figure 2.3.2.1 here Figure 2.3.2.2 here Unfortunately this model is a bit too stable. Although it shows a very clear power law, still with identical companies, the exponent of the power law is very different to that seen in the real world. It appears that the retention is too great and is forcing a high minimum value for companies, so preventing the formation of the power tails with slopes seen in model 1A. 2.3.3 Model 2C Deterministic on Production, Capital Hoarding In model 2C the production rates of the companies was set prior to running the model, and were again drawn from a normal distribution. So in this model some companies produced more than 10% all the way through the model, some produced less than 10% all the way through the model. Note that model 2C is not stochastic, it is deterministic. In this model some companies are more efficient than others with their use of capital. Again the payout ratios were adjusted to prevent elimination of companies from the bottom of the distribution. It was found that any downside payout ratio of less than 0.5 or so prevented this washout. Figures 2.3.3.1 and 2.3.3.2 below are for a downside payout ratio of 0.5 and an upside payout ratio of 0.9. 88 EFTA00625216 Figure 2.3.3.1 here Figure 2.3.3.2 here Intriguingly the power law exponent of -0.68 is close to the value of —1 seen in real life. However the fit is poor, and it turns out that the value of the exponent is highly sensitive to the value of the upside payout ratio and can change to high tens or low decimals for small changes in the upside payout ratio. Initial modelling suggests that the value of 0.9 is closely related to the production ratio of 0.1. As the production ratio is changed, an upside payout ratio of one minus the production ratio gives a power tail close to one. Again, the important thing to note is that relatively small changes in relative efficiency of the companies produces a power tail with very large, multiple factors of ten, differences in size for the companies. 2.4 Companies Models - Discussion As can be seen from the results, using a very simple combination of classical economics and dynamic statistical mechanics allows the building of simple models that give power law distributions for company sizes similar to those found in real life economies. As with the income models it noticeable that there are many things that are not needed to produce such a model, these include: • Economic growth • Population changes • Technology changes • Different initial endowments (of capital) • Shocks (exogenous or endogenous) • Marginality • Utility functions • Production functions The issue of marginality, utility, production functions will be returned to in a moment, before that I would like to discuss the roles of shocks, expectations and behaviouralism. It is notable that the models do not include for exogenous shocks, which are often found in explanations of company size. Models 2A and 2B are stochastic, and do therefore model minor endogenous shocks to productivity. These could be issues such as a variation in breakdown rates of machinery, management efficiency, etc from period to period. What is notable about models 2A and 2B is 89 EFTA00625217 that the average productivity of all companies over the long term is identical; and yet a power law still results. Model 2C is effectively deterministic. The initial productive efficiencies of the companies are determined prior to the simulation. The simulation then rapidly reaches an equilibrium with a power law distribution. There are no shocks in model 2C; external or internal. Expectations and behaviouralism do enter into the model in two different ways, firstly with regard to the pricing of stocks, and secondly with regard to the retention of capital within companies. In both cases these are very obvious forms of behaviour and are supported by economic research. With regard to returns, the assumption is simply to take the pricing of financial assets as strictly based on their recent returns. This is in fact the "traditional" naive neo-classical form of pricing capital and is supported by the research of Fama & French and other work discussed in section 2.1 above. This assumption that prices of assets are defined by simplistic projections of present earnings is also at the heart of Minsky's theories. The assumptions on capital retention are more subjective than the assumptions on returns, and more arbitrary in the specific amounts of returns chosen, and is the weakest part of my company modelling. This is discussed in more detail below, when comparing with the work of Ian Wright. However the work of Dial & Murphy regarding General Dynamics and other companies make the assumptions very plausible. What is important to note is that the above assumptions on expectations are the only assumptions needed. No detailed assumptions about the understanding of the economy, interest rates, growth, technology, etc are needed. The only 'behaviourism' that we need to assume is that, firstly investors are deeply short sighted, and secondly that managers don't like sacking themselves. It is clear from the models that neither utility nor marginality are relevant. Much more importantly, the output distribution for the models is demonstrably not 'efficient' in the normal neo-classical usage. To take models 2A and 2B as examples, capital is rapidly shifted between companies according to short-term results, and companies with equal long-term efficiencies end up being sized very differently. In a neo-classical version of model 2A of 2B, either one company would dominate, or all companies would be equally sized. Model 2C is far more realistic, and much more interesting. It also shows how profoundly free markets fail to allocate capital effectively. Model 2C has a range of production efficiencies. Some companies make better use of their capital than others. In a neo-classical outcome (or indeed in the classical models of Smith, Ricardo, etc) the outcome of such a model should be crystal clear. The most efficient company should continually be 90 EFTA00625218 rewarded with more capital until it ends up being a monopolist, owning all the capital in the economy. Despite the best efforts of managers to cling on to their capital, investors should continually remove their capital from all the less efficient companies until these companies have no capital left and go out of business. This is not what happens. In model 2C, and as Graham, Buffet and others have discovered, also in real life, poorly performing capital is simply written down. Companies are allowed to retain some of their real, book value, capital K. But part of their financial wealth is written off. Once an under-performing company's financial wealth W is small enough to make the (poor) returns from the actual K equal to the normal market rate, then the company is allowed to continue under-performing, and under-utilising its capital, indefinitely. So it is noticeable that moderately bad companies are only downgraded, they are not driven out of business as economic theory suggests they should be. This represents an enormous misallocation of real capital. In model 2C the top company has a capitalisation/capital ratio of 1.37, the bottom company has a capitalisation/capital ratio of 0.62. The bottom company is half as efficient as the top company, but once it has been written down, it is allowed to limp on inefficiently. That this happens in real life is supported by the effective long term investing models of Benjamin Graham, Warren Buffet and others. The accumulated wealth of Warren Buffet has always been one of the most pertinent criticisms of the efficient market hypothesis. In an economy such as model 2C above, the Graham/Buffet approach is straightforward. Finding companies with under-valued physical assets is straightforward; you simply look at the book value of assets compared to the stock price. Generally it is poorly performing human capital that has driven companies into under- performance. The quality of human capital is something that can change very quickly. As General Dynamics showed, a change of CEO can be sufficient. The Graham/Buffet approach uses various measures to identify increases in the efficiency of human capital. These include qualities such as paying down debt and good recent dividend history. By this process, investors such as Graham and Buffet can identify companies that are undervalued, with under-performing capital, and that are also likely to move quickly to over- valuation. In practice this failure of capitalism may not be as bad as painted above. Firstly it is likely that other processes will ensure that capital gets redeployed more quickly. Despite the best efforts of capital retaining managers, many companies do go bankrupt; many more get merged or taken over. Newer, more efficient companies also enter the market and take market share from existing non-performing companies. It may also be the case that the power law distribution is, accidentally, highly effective in preventing monopoly or oligopoly in the market place. Indeed, looking at deviations from power law distributions, in industry sectors as well as whole economies, may well be a very useful way of identifying monopolistic behaviour. If a company is 91 EFTA00625219 bigger than its place on a power law suggests, then it is probably behaving in a monopolistic or oligopolistic manner and should either be split up or subject to a super tax of some sort. It is the belief of the author that this modelling approach is generally applicable. Although the model focuses specifically on dividends, a simplistic Modigliani & Miller assumption of the irrelevance of forms of payout would allow that the model would work when capital growth was substituted for, or used in addition to, dividend payments. Even in the non-listed sector the same basic arguments hold. If a small business goes to a local bank for a loan, the bank may look at the size of the business assets as collateral for the loan, but the calculations of loan size will be based on estimates of the future revenue streams of the business, based on recent historic revenue streams. The general applicability of this type of model can be seen by looking at the shortcomings of my own model, and also by comparing the model with those of Wright. The workings of the model above are straightforward, and similar to the other GLV models. The companies have a positive feedback loop which means that the more companies earn, the more capital they get. There is also a negative feedback loop, so the bigger companies get the more income they have to pay to investors. If these were the only two rules, then the most efficient company would grow explosively into a monopoly. A true power law distribution can not go down to zero, so to be stable, a power law always needs some other distribution to 'support' it. That is why power law distributions are normally 'tails' to other distributions. As Levy & Solomon make clear, there needs to be a 'reflective barrier' above zero. The assumption of retention of capital assures a continuous, if minimal income to all companies, however small. This prevents collapse of the distribution to a single point, and allows the generation of the power tail distribution. This is the weakest part of the model above, with factors 'selected' (fixed, if you prefer) to ensure the distribution does not collapse. While these assumptions are somewhat contrived, the work of Wright shows that different, but similar assumptions are just as effective. In the modelling of companies the models of Ian Wright are significantly different to, and significantly better than, my own, but detailed analysis shows strong similarities. Wright does not model a financial sector, and the mathematical modelling above is not therefore relevant. 92 EFTA00625220 In Wright's models, each company is owned by a single 'capitalist', and there is no distinction between the capital of the company and the wealth of the owner. Wright models the expenditure of the capitalist and the income of the company as both being stochastic, and crucially, independent of each other. So the capitalist spends at a set, but stochastic, rate, which depends only on the wealth of the capitalist. So the capitalist is spending his 'expectation' of the future wealth of his company, which is implicitly assumed to be the same as the present wealth of his company (which is identical to his personal wealth). Meanwhile the income of the capitalist's company is set stochastically in the market, and may not match the expenditure of the company. Any mismatch then results in an expansion or contraction of the wealth of company/capitalist. This consequently results in a power law of company sizes that is analogous to my own model. It should be noted that in at least two ways Wright's models of companies are superior to my own. Firstly, Wright models employment directly which my own models ignore, substituting capitalisation. Secondly, Wright allows for the extinguishing of companies as they become too small to trade, and the creation of new start-up companies as individuals become sufficiently wealthy to employ other individuals. This avoids the somewhat artificial 'capital hoarding' approach that is used in my own model, which maintains all companies as operational entities, however severe their losses. In real life clearly both mechanisms operate, with bankruptcy and new company formation happening alongside poorly performing companies that limp on for years without giving good returns on their capital. A third mechanism of corporate takeover, divestment and splitting of companies also takes place. Detailed research would be needed to determine the relative importance of the different mechanisms. Personally I believe that Wright has identified the most important factor in new company formation and extinction. The main point is that, as long as you have a means of supporting the base of the distribution, the basic pricing mechanisms of capitalism produce a power law tail as seen in reality. The differences between the models of Wright, and my own, underline a much more important point. If you use the basic ideas of the classical economists, combined with statistical mechanics, it is in fact very easy to get the same power law distributions that are seen in real life. If you use neoclassical theory, efficient markets, and static equilibria, it is pretty much impossible to give convincing reasons for power law distributions. Neither Wright's or my own models may be fully correct, but they are both clearly closer to the truth than anything produced by neoclassical theory. Another area that needs further investigation is the exponent of the power tail. Data from real economies suggest that this has a value close to 1 in all cases whether measured by employees, capitalisation or other variables. This suggests that a deeper underlying equilibrium is being formed, with a 'self-organising criticality' (SOC) as previously suggested for income distribution. My first model produces this exponent well, but is not stable over the long term. My stable models can reproduce this value, but only by 'fixing' the parameters of the model, a solution that is neither universal nor acceptable. Wright's model does produce this exponent, and without any apparent 'tuning'. As such Wright's model appears to be superior to my own, but as a non- mathematised model, it is not fully clear why his model does this. This is a suitable area for further investigation. 93 EFTA00625221 3. Commodity models The following is a brief model, mainly to introduce some concepts and demonstrate the importance of a dynamic modelling approach to markets. This paper has taken a classical economics approach that assumes that all goods and services have a meaningful intrinsic value that ultimately relates through to basic concepts of entropy in physics and biology. It is immediately obvious that the prices of some goods; land, housing, gold, artworks, cabbage- patch dolls, etc, show wild fluctuations in price that appear to contradict the assumptions of fundamental value in classical economics. To investigate this further a simple dynamic model of a commodity market is constructed, largely following the lines of the previous company model. The intention is to model the behaviour of a commodity such as copper, platinum or coffee. For such commodities prices can fluctuate wildly, and this is often blamed on external factors such as demand, weather, war etc. In the model below it is demonstrated that the main source of price fluctuations are endogenous and relate to the provision of capital by financial markets. 3.1 Commodity Models - Background The model aims to model the behaviour of mining or agricultural commodities such as copper, aluminium, nickel, platinum, coffee, tea, cocoa, sugar, etc. Such commodities have wildly fluctuating prices, normally characterised by long periods of low prices punctuated by severe spikes. The figure 3.1.1 below for copper shows a typical example. Figure 3.1.1 here This pattern is also seen in other commodities such as oil or natural gas, land, housing, etc. While it is believed that similar forces operate in the markets for oil and houses, these commodities are sufficiently important that they can in turn have large impacts on the economy as a whole. For simplicity the model below chooses to model something like copper or sugar that can have large price spikes without having a significant effect on the economy as a whole. This allows important simplifying assumptions to be made in the model. Although at first glance copper, aluminium, nickel, platinum, coffee, cocoa and sugar would seem to have little in common; in fact they share three important factors. Firstly, in a stable economy demand for these things is quite stable and relatively insensitive to price. 94 EFTA00625222 Cables are made from copper, and if you build a house you need cables and you pay the price necessary. Similarly, most planes are made from aluminium. Even in poor countries people tend to drink a certain number of cups of tea or coffee each day, with their usual number of spoons of sugar. The total costs are small compared to other outgoings such as food or rent, and the pleasure obtained, so people tend not to cut back even if prices increase significantly. The second factor these commodities have in common are non-substitutability. Copper is both an excellent conductor and corrosion free, and is also relatively cheap compared to other metals with these properties. It is slowly being displaced by plastics for plumbing and aluminium for electrical use, but the substitution process is very slow. While Boeing are beginning to build airliners out of composites, the process has not been easy and demand for aluminium seems likely to remain high for decades. While some people swap between tea and coffee, most have a favourite brew, and there is no other easy substitute for hot caffeinated drinks. I don't know of anything that can effectively substitute for chocolate. The third factor is that all the above commodities take a long time to increase their output by installing new capital. Mines are large, complicated, and often isolated. To bring a new mine into production can easily take three to five years, even expanding an existing mine can take two to three years. Unlike say wheat or rice; coffee, tea and cocoa grow on trees or bushes, and there is a limit to how much you can rush nature. For commodities such as these, price signals take a long time to result in increased output. It is this delay that changes the problem from one of comparative statics to one of dynamics, so it a dynamic model that is needed. 3.2 Commodity Models - Modelling This model follows on from the companies model above, and in one way is much simpler. So simple that the model was moved to a spreadsheet. For anybody who is interested this can be copied and installed into Excel from appendix 14.8. Although the same basic model is used as that in the companies model above, in this case one section of the economy is modelled as a single unit, so there is only a single set of equations running in the model. For the sake of the argument, assume the commodity is copper. In this model, along the lines of classical economics, the production cost of copper is fixed and related directly to its inputs, a mix of energy, machines and various types of labour. We have assumed that the price of copper, even if it varies dramatically, has very little effect on the economy as a whole. This means that the prices of the inputs of energy, machines, labour and any other inputs vary negligibly with the price of copper. So the cost of producing copper is a simple linear function of the amount of copper produced. As with the companies and incomes models, the total amount produced is a fixed ratio of the capital installed. Taken together this means that the marginal price of extra copper is zero. This model ignores marginality, because its importance is marginal, to the point of irrelevance. 95 EFTA00625223 The price of copper is a different matter. It is assumed that total demand for copper is almost constant with a 'normal' amount required in the market place. In this model 100 units of copper. When this amount, or more, is available, copper companies charge the costs of production. Also they lower their output by closing down excess capacity. This gives a base price, a classical economics price, for copper of 1.0 in this model. If production drops below that required, then price increases very rapidly and demand is choked off very slowly, the demand is highly inelastic. Figure 3.2.1 below shows the price volume curve used in the model. Figure 3.2.1 here This is of course a completely unrealistic, hypothetical demand curve of the type beloved by economists. In a comparative statics analysis an economist would then draw one or more hypothetical supply curves across the same graph and predict a static equilibrium based on marginal outputs of the different mines. This is not a meaningful approach. The effects of delays in installing capital, and/or the retention of wealth by companies mean that a static equilibrium is not possible. In this model, just as in the companies model, the standard market interest rate defines the expected returns, based on the previous market capitalisation w. Again, as in the previous model, payouts are predicated on the expected returns using payout ratios, with companies hoarding capital or returning it to shareholders as appropriate. When supply is low, and prices jump up, the mining companies find themselves with much higher receipts than costs. In these circumstances the excess cash is used to provide more capital. As discussed above, this capital is added to the productive capital, but only after a lag of a number of iterations. This lag can be adjusted in the model from zero to ten cycles. Once the new capital has been added after the lag in time, then production can be increased. Eventually this allows supply to meet demand and prices can drop again. 3.3 Commodity Models - Results The results are fairly straightforward. Figure 3.3.1 below gives the output for Model 3A, this shows the prices for copper with no lag on capital installation and payout factors of one; ie no capital hoarding. 96 EFTA00625224 Figure 3.3.1 here Even with this very simple model the system is unstable and produces wide cyclical variations in prices (this was something of a surprise, I had thought the model might be stable with instant installation of capital and no capital hoarding). The real price of copper, based on inputs, should be 1 unit; note that the system is only at its true input price for short periods of time. Left to itself the market charges an average price slightly over 50% of the input cost price. The extra 50% being caused by the cyclical over-production and destruction of capital, and consequent rent taking. Figure 3.3.2 below for model 36 shows a capital lag of two periods, but still with a payout ratio of one. Figure 3.3.2 here This shows a pattern closer to reality; long periods at 'classical' prices are interrupted with intermittent spikes. Even in this simple model it is notable that the spikes have a variable pattern showing chaotic (not stochastic) behaviour. With this capital lag, the average price is raised to 1.7 times input cost, as the cycles of capital creation and destruction become more aggressive, and rent taking becomes larger. Finally figure 3.3.3 shows model 3C with zero capital lag, but with up and downside payout ratios of 0.9. Figure 3.3.3 here This figure demonstrates that capital hoarding alone can produce complex cyclical chaotic behaviour. As with figure 3.3.1, cycling only results in 50% price gouging. 3.4 Commodity Models - Discussion I intend to keep the discussion of the commodity model quite brief. The main issues raised are dealt with in more depth elsewhere. Some of the main points of note are as follows. Very simple dynamic economic models can result in complex chaotic behaviour. Behaviour that mimics real life surprisingly well. 97 EFTA00625225 The behaviour is chaotic, not stochastic. The random changes are generated endogenously. There is no stochastic generator in this model. This distinction is very important, and is discussed at length in section 5 below. This is a Lotka-Volterra model, not a General Lotka-Volterra model. This model is very similar to the lynx and hares model first discussed back in section 1.2, in fact it is closer to the Soay sheep and grass model. The build up of excess capital in the mining companies is analogous to the build up of excess sheep biomass on the island of Soay. The build up of capital is too much for the economy to support, as the build up of sheep is too much for the island to support. While the GLV models were stable, like many Lotka-Volterra models, the build up of capital in the commodity sector is inherently unstable. The problems are deep in the maths of the system. Blaming investors or speculators for misjudging their investments is as sensible as blaming the sheep for procreating. Diminishing returns and marginality are conspicuous by their absence. Diminishing returns are not needed for the model to work. Neither is marginality, and any costs associated with marginality are of an order smaller than those associated with dynamic effects. Using comparative statics to analyse a dynamic process is simply not appropriate. It is the wrong tool for the job. Using comparative statics to analyse dynamic problems is about as sensible as trying to do long division with roman numerals. Using classical economics within a dynamic framework works. It produces output prices that can be at substantial variance with input prices, and can vary substantially with time. It should also be noted that the model does not average to the correct input prices even over the long term. The correct input prices are instead associated with the bottoms of the cycles, and are only touched for short periods of time. Due to problems associated with the way assets are priced, the time taken to install capital, and (financial) capital hoarding by companies, the market is profoundly inefficient. Average prices are substantially higher than they would be if they had the opportunity to settle to long-term static equilibrium prices. The form of this over-pricing is interesting. Above I referred to it as associated with capital appreciation and destruction, but the process is more subtle than this. In a boom period, customers are substantially overcharged compared to the input costs. Extra capital is created, but the nominal capitalisation increases much faster than the real value of the capital installed. In short the companies become grossly overvalued. As a consequence they pay excessive dividends. In a boom most of the over-pricing passes straight through to shareholders as excess profits. In the following crash, the company is still expected to match dividends at the market rate. It does so by drawing down capital to pay dividends. Over the cycle as a whole customers are forced to overpay, with the payments transferred direct to excess profits. Allowing dynamic cycling of economic variables in this way allows large-scale rent-taking by the owners of resources. 98 EFTA00625226 For most markets these effects are not so important, with the very notable exception of oil, commodities are not a critical price input to the world economy. The price of manufactures and services are much less prone to bubble behaviour, partly due to the speed with which ordinary factories and offices can be built, and also to the fungibility of most non-commodity goods. The problems with oil have been largely mitigated in Europe with very high taxation of petroleum products. This makes the variable element much smaller, and also encourages the reduction of oil energy intensity in the economy. There are two other commodities for which these effects are of great importance. The first is housing, which seems particularly prone to destructive bubbles, this is returned to later in section 6.3. The other commodity is much more interesting, and is unique and of great importance to the analysis of the economy as whole. This commodity is labour. 4. Minsky goes Austrian a la Goodwin — Macroeconomic Models 4.1 Macroeconomic Models - Background So far in this paper three basic models have been developed using the tools of classical economics and the mathematics of the Lotka-Volterra and General Lotka-Volterra models (GLV's). The first set of models looked at the consumption side of the economy and the resulting distribution of income, the second series of models looked at the production side, and the resultant distribution of company sizes. The third, looking at commodities, introduced a very simple supply and demand based model. Although the GLV has not previously been used significantly in economics, some non-linear modelling work has been carried out at a macroeconomic level by Kalecki, Kaldor, Desai and others. Most notably Goodwin used the Lotka-Volterra predator-prey system to model a qualitative cycle described by Marx (though true-blooded Marxists will be disappointed to learn that in these models the workers are modelled as predators; the capitalists are the prey). Keen has extended the Goodwin model to model a Minskian business cycle [Keen 1995]. Despite (or possibly because of) these heterodox Marxian origins there is significant evidence to suggest that these cycles exist in real economies. Barbosa-Filho & Taylor [Barbosa-Filho Taylor 2006] have carried out a detailed study of business cycles in the US. Harvie [Harvie 2000] has carried out a similar study for ten OECD countries. In both cases the evidence is qualitatively strongly suggestive of cyclical changes in labour share of return and employment that match the patterns predicted by Goodwin. In both case though there are significant difficulties in fitting the data quantitatively. 99 EFTA00625227 In addition to the work above there have also been substantial qualitative studies of business cycles in other schools of non-orthodox economics. In the Austrian school, it has long been proposed that the build up of excess capital has been a fundamental cause of business cycles, with the blame for this generally put on government mishandling of credit availability. In parallel with this Minsky, coming primarily from the post-Keynesian school, but also following the work of Fisher, has also studied the build up of economic cycles, though with the blame being primarily placed with speculation and the unsustainable endogenous creation of debt. The Austrian and Minskian models share significant common features, the most obvious being their beliefs that booms and busts are natural features of economics. Another, unfortunately, is their shared disdain for formal mathematical modelling. In the modelling that follow a very simple macroeconomic model is built, that combines the Lotka-Volterra approach of Goodwin with the basic ideas of the Austrian / Minskian business cycles. The main ingredients for this model, including many simplifications, are already available in the proceeding models above. 4.2 Macroeconomic Models - Modelling In this section a simple macroeconomic model is introduced, based on most of the same variables as the company and income models above. The main assumptions of this model are as follows: In line with classical economic theory, produced goods have real values, but market prices can vary from these values in short time periods due to insufficient or excess demand. Consumption is a fixed proportion of consumers' perceived wealth, held in the form of paper assets, as in the income models above. Companies have real capital which can produce a fixed amount of output, and needs a proportional supply of labour, as in all the models above. The price of paper wealth assets is defined by the preceding revenue stream; as in the myopic companies model above. The management in companies can be capital preserving, as in the companies model above. There can be delays in installing capital as seen in the commodities model above. The price of labour is non-linear according to supply. That is real wage rates go up when there is a shortage of labour, and go down when there is a surplus of labour. Labour is a genuinely scarce resource. It should be noted that, unlike the Goodwin models, both population and technology are fixed. Although this macroeconomic model will be more complex, as it has more variables, in other ways it will be simpler, as we will not look at individual consumers or companies, but look at the aggregated whole of supply and demand, in the same manner as the commodities model. 100 EFTA00625228 With the macro economic model there will also be a much stronger interest in the behaviour of the model as a function of time. The big new assumption in this model is that labour costs vary with employment and unemployment. It is assumed that labour costs vary as a concave function of employment, ie labour costs will increase as the employment ratio increases, and will increase at an increasing rate. Figure 4.2.1 here In this model I have used a simple square law function, shown in figure 4.2.1 above. This is not a particularly realistic function, more realistically it should be asymptotic to the vertical on the right hand side as there is a realistic maximum somewhere around 6000 hours per year. However this basic function is sufficient for the needs of the model. It is also worth noting, this is not an inflation Phillips curve. This curve is a simple supply-price Phillips curve for labour in real terms. In this model, prices of goods and labour both go up and down, just as they did in the commodities model, but they move around stable long-term values. The analogy is with the cyclical price changes seen in a Victorian economy with a gold standard. There is no long-term monetary inflation. For a pithy study of the misinterpretation of the Phillips curve see Hussman [Hussman 2011]. Again, an element of marginality has been introduced. Over short to medium terms, the supply of labour is fixed, while demand can change. Because of this labour prices can change significantly through business cycles. In these models, it is assumed that individuals always spend 40% of their income at all times, SI = 0.4. It is possible that the consumption spending will exactly balance the amount of production capacity available in the companies, however this will not always be the case. It is also possible that there will be too much or too little capital available to match the consumption demand. Looking firstly at the case of too little demand; if the 40% spending provides insufficient demand, then excess capital will be available and some of that capital will be unused. As a consequence of this there will also be a reduction in labour employed. Also, following exactly the same logic as the companies models above, if companies create insufficient wealth to meet the payout targets set by their market capitalisation, then they will be obliged to convert some of their capital to wealth for payout. Clearly in this model such a conversion of capital to returns is less realistic than the companies model. In the companies model capital was swapped for cash between the successful and unsuccessful companies. In this macroeconomic model, all companies are shrinking in size at the same time. This would mean that first stocks of goods and then fixed capital would need to be converted into payouts. This would normally mean substantial losses on the value of the capital, especially the fixed 101 EFTA00625229 capital. In this simple model, this problem is ignored, and capital is assumed to be converted into payments at par. This assumption is returned to in the discussion in section 4.4. It is also possible that there may be insufficient capital available. In these circumstances it is assumed that consumption is still maintained at the full 40% of current wealth, even though insufficient capital available, and so insufficient goods are produced. In this case the consumption funds available for purchasing are simply divided amongst the goods that are available to be purchased, so increasing the nominal market price of the goods above their long- term natural prices. Consequently this results in short-term consumer price inflation. It is implicitly assumed that consumers judge value by price and continue to spend a fixed proportion of their wealth, even though they actually receive less real value for that wealth. When this happens super-profits are then earned by the corporate sector. If employment and so wage levels are low, then the income retained by the companies is converted into new capital to allow the production of more commodities. In this manner, super-profits are converted into new capital and new production until supply rises to meet the new demand, and the prices of consumer goods then drop back to their 'natural' values based on input costs. This is closely analogous to the commodities model. It is important to note that, in the company models, the total amount of capital was fixed; however in this macroeconomic model, the amounts of capital and labour employed can vary, though labour is still needed in a fixed proportion to capital used. In this macroeconomic model the capital and labour are still used in a fixed ratio to give a given output. The amount of capital can vary freely, in line with the demand of goods from consumers. The total supply of labour is fixed however, with the amount of the labour pool employed varying in fixed proportion to the amount of capital. Labour costs vary non-linearly with the amount of labour employed, which means that labour costs vary non-linearly with the amount of capital employed. So returns to labour and capital can vary. It is still assumed that the proportion of labour required to capital is fixed over the whole period of time being modelled. This means that there is no technological progress, and also that it is not possible to substitute capital for labour. Each iteration of the model operates as follows: The expected returns are defined as 10% of the current market capitalisation. The consumption, and so the payments made for consumer goods are defined as 40% of total wealth. If these payments are less than 20% of the available capital, then the amount of goods produced is equal to the value of the consumer payments. If the payments for consumer goods are greater than 20% Of the available capital, then the goods produced are equal to 20% of the total capital, ie, the maximum production possible is 0.2 times the capital K that is in existence. The income accruing to labour is calculated, according to the amount of capital used, and so the proportion of labour employed, according to the square law. 102 EFTA00625230 The surplus revenue that the company generates is then the value of the consumer payments received, less the earnings income paid out. The new value of the total real capital is then the old capital, plus the payments received for goods, less the labour earnings paid out, less the actual returns paid out. Finally, the consumers receive their dividends from the companies and revalue the market capitalisation according to the actual returns paid out. At this point, the cycle starts again. As in the companies model, the actual returns paid to the owners (shareholders) that is the payout ratios can depend on whether the surplus revenue generated is greater than the expected returns or less than the expected returns. For example in model 1D the actual returns paid out are always 70% of the revenue generated. However in models 4A to 4C the actual returns paid out are equal to the real returns produced. It is noted that these payout factors are different to the ones in the companies model above, clearly these models are preliminary and in need of future calibration to real economies. As with the commodities model, it is also possible to put a variable lag in to model the time it takes to install capital. A further important ingredient in this model is the existence of a 'cash balance' for the householders. This is needed in their role as owners of capital and spenders of money. This cash balance can result as an imbalance of spending outgoing against income received as a consequence of these being dynamic models. If the cash balance is positive then this represents spare cash in the bank. The householders have received more in wages and dividends than they have spent in consumption. If the cash balance is negative, then this represents a debt to the bank, due to the consumers spending more than they earn. In the notes following, the cash balance is referred to as H to differentiate it from the capital owned which is now labelled Q. The consumers are assumed to be sensible, so they carry out their consumption based on their total wealth W which is the sum of Q and H, so: C= (4.2a) or: C = (Q+H).O (4.2b) So, for example, if H is negative because the consumers have net debt, then consumption is reduced below that judged by the size of Q only. 103 EFTA00625231 This model was carried out in Excel, those who wish to go through the maths in detail can past the model into Excel from appendix 14.9. 4.3 Macroeconomic Models - Results As expected this model can show different sorts of behaviour, some examples are given below: Model 4A is the base model, with all the numbers designed to be nice and round. This model has payout ratios of 1 for both the upside and downside. It also allows capital to be added instantly, without any lags. It can be seen from figure 4.3.1 that the output is very stable, and so very dull. Figure 4.3.1 here Model 46, shown in figure 4.3.2 has exactly the same parameters as model 4A, the only difference is that the initial values were different. Figure 4.3.2 here This shows just how stable this model is, with the model quickly settling down to equilibrium values. Though even in this stable model it is notable that model 4B needs to go through a number of fluctuations before it arrives at stability (cf figure 1.2.1.4). But there is a more important difference to note between model 4A and 4B. The parameters of the model are exactly the same, but the equilibrium points are very different. Model 4A started with real capital of 100 units, and settled to an equilibrium at 100 units. Model 4B started with real capital of 400 units, and settled to an equilibrium at about 184 units. As a consequence, total capital employed at equilibrium in model 4B is much higher than that in model 4A, and more importantly, total employment is higher in model 46 than model 4A. Also the ratio of returns to labour to returns to capital is significantly higher in model 4A. This is Keynes writ large. Unlike static equilibria, dynamic equilibria can have multiple points of stability. The point of equilibrium that is reached depends on the parameters of the model, but also on the initial conditions. Different initial conditions can give different equilibria even with the same parameters. Once it has reached its equilibrium, the model can stay at that point indefinitely. To change the equilibrium an exogenous force is needed. The model will not rebalance itself to a 104 EFTA00625232 particular point; a point such as full employment for example. Mass unemployment can continue indefinitely without positive external action. Model 4C is the most interesting, and most realistic, model. In this model a time lag has been introduced between capital being purchased and being brought into use. This is identical to the way capital is installed in the commodities models in section 3. Note that the payout ratios are still at unity. Figure 4.3.3 shows the long term behaviour of the model. Figure 4.3.3 here As can be seen the model shows regular cycles of capital being created and destroyed. Again it is important to note that this is a chaotic model, not a stochastic one. There is no stochasticity in this model. All fluctuations in the model are created endogenously from the Lotka-Volterra like differential equations in the model. Figure 4.3.4 shows the detail of couple of cycles. Figure 4.3.4 here These are real live Minskian / Austrian business cycles. But with one big exception. It can be seen that real capital K builds up in advance of the total wealth (in this simple model paper wealth; capitalisation is constant), this build up of capital is unsustainable, and so leads to a fall in real capital. Interestingly, although debt (negative cash wealth) is present, this is a lagging variable. In this model debt creation is fuelled by capital growth, not the other way round. The chaotic, bubbly behaviour is not caused by excess credit, it caused by the basic pricing system of capitalism. Model 4D, shown in figure 4.3.5 below has no lag in the installation of capital. Instead this model has payout ratios of 0.7 on both the upside and the downside. Figure 4.3.5 here It is believed that this is a less realistic model, however it does demonstrate how highly chaotic behaviour can be generated in even a very simple model. 105 EFTA00625233 Finally model 4E is shown in figure 4.3.6 below. This has a just a small lag of 1 unit for the installation of capital, and payout ratios of 0.8. Figure 4.3.6 here Interestingly, it seems that similar results can be achieved without a lag. If both interest rates and payout factors are reduced, an explosive result is also seen. As can be seen these minor changes in the model are sufficient to create explosive behaviour. This is a true bubble, similar to that of Japan in the 1980s, or the US in the 1920s or in the last decade. Again the cash wealth (debt) is a lagging indicator. It is possible to create explosive bubbles just from the basic pricing system of capitalism. There is finally one important thing worth noting about the models. The value of the Bowley ratios, (3, for the first four models were as follows: Figure 4.3.7 13 Model 4A 0.75 (exactly) Model 46 0.92 Model 4C 0.78 Model 4D 0.85 The Bowley ratio is the ratio of returns to labour to the total returns. The values for models 4C and 4D are averages; the Bowley ratio varies wildly over the course of a cycle in these models. The numbers above are close to the 'stylised facts' for the Bowley ratio, and are of considerable importance. This is returned to at length in section 4.5 onwards. 4.4 Macroeconomic Models - Discussion As with the previous models, the results above show that a simple combination of classical economics and a dynamic analysis gives interesting results that mirror real economies. The author expected that such a model would be easily capable of producing boom and bust business cycles, and this is discussed in some detail in this section. The production of a suitable Bowley ratio was a surprise, though a pleasant and very important one. This is discussed further in sections 4.5 to 4.7. 106 EFTA00625234 Leaving aside the Bowley ratio, the most interesting result of this model is that the booms and busts are generated internally via an endogenous spiral of creation of wealth. In the model real capital is installed, which generates more paper wealth, which generates more consumption, so feeding into another cycle of wealth creation. The upswing is finally constrained by rising wages making the capital unproductive. This then generates a downswing of declining wealth, consumption and wages. This is the normal cycle of capitalism as described by Minsky and the Austrians. Booms and busts are endogenous. Free markets are not inherently stable. Again, as with the income and company models it noticeable that there are many things that are standard elements of neo-classical or Keynesian economic theory which are simply not needed to produce this macroeconomic model, these include: • Economic growth • Population changes • Technology changes • Productivity growth • Investment • Saving • Accelerators • Multipliers • Shocks (exogenous or endogenous) • Stochasticity (in any form) • Different initial endowments (of capital or wealth) • Utility functions • Production functions It has been noted that marginality has worked it's way into the modelling in the form of the pricing curve for labour, this is a reasonable argument, as labour is a commodity that is truly unchangeable in it's supply. Although marginality might be a mathematically useful way to address this, the history of entropy and information suggests there may be better ways to address this. More importantly, the results of the model show that the detailed form of curve are completely irrelevant to the model. The curve simply needs to be concave, to ensure that labour costs eventually choke the growth. Within reason, any concave curve will do this. So the actual detail of the calculations of marginality are irrelevant and do not have any influence on the long- term equilibrium, the cycle frequency or the distributions of wealth and income. This is discussed further in section 4.7 below. It is also worth considering the 'efficiency' of the economy in this model. This model again creates chaotic behaviour endogenously. There is no stochastic noise in this model. It is politely suggested by the author that a system that endogenously creates booms and busts, with short term creation of excess capital, and far worse; short term destruction of the very same capital, may not, in fact, be allocating capital in a particularly efficient manner. 107 EFTA00625235 Investment and saving has been deliberately ignored in this model, as it has been in all previous models. This is because, as the data given from Miles and Scott in section 1.3 show, saving and investment is a minor part of the economic cycle. The core driver of business investment is the availability of cash streams. When firms have more money coming in from revenue than they need to pay out as dividends they use it for investment. When they don't have spare money they don't invest. The mechanics of saving and investment are a side-show and diversion from the base model of macroeconomics. Similarly the general public is assumed to simply consume a fixed proportion of their wealth. In the real world it seems much more reasonable to assume that people who gain more wealth will divert a greater portion of this to saving, particularly in an environment, as here, in which companies appear to be showing increasing profits on their capital. I believe this is a simplification rather than a flaw. The point of the model is that endogenous business cycles arise at the heart of the system of pricing financial assets. Allowing transfers of excess savings in booms to investment rather than consumption would clearly exacerbate these booms. Indeed it is possible that the effects of saving and investment multipliers might be significant, but that is not the issue, the issue is that saving and investment is a multiplier rather than the root cause of the instability. In identical fashion to the companies models above, expectations and behaviouralism do enter into the model in two different ways, firstly with regard to the pricing of stocks, and secondly with regard to the retention of capital within companies. Again these are obvious forms of behaviour and are supported by economic research as discussed in section 2.1 above. It can be seen from the model results that economies can behave very differently according to relatively small changes in input parameters. This is because a system like this can show different regions of behaviour, a general property of Lotka-Volterra and other similar non-linear differential equation models. Depending on the settings of the variables in the model, there can be three different cases for the outputs. Firstly, the outputs can be completely stable, quickly going to constant values, this was seen in models 4A and 4B. Secondly, the outputs can be locally unstable with values constantly varying, but hunting round within a prescribed range of values, this is similar to the lynx and hares Lotka-Volterra model discussed back in section 1.2. This appears to be the way that most normal economies behave. This effect can be caused by the behaviour of capital, either by deliberate hoarding of capital by company managers, or by the time it takes for capital to be installed. The cyclical rise and fall of capital in business cycles is analogous to the cyclical rise and fall of biomass in a biological Lotka- Volterra system. Just as the hares and lynx respond rationally to the available grass, so business investors and speculators react rationally to the opportunities in the economy. Finally, the outputs can be explosive, moving quickly off to ± infinity. 108 EFTA00625236 In models 4A and 4B these values were 'fixed' to ensure a stable model, in 4C and 4D the parameters were fixed to give a quasi-stable cyclical model, in model 4E they were changed to get explosive models. In the real world it appears that economies operate largely in zones 4C/D, with occasional excursions into zone 4E. Model 4E suggests that if both interest rates and payout rates are too low then the company sector is too profitable and capital expands exponentially before finally wrecking the whole economy in a glut of capital, see figure 4.3.6 above. It seems plausible to argue that this reflects what actually happened in the US during the late nineteen twenties, and Japan in the late eighties. Following each of these bubbles the respective economies failed to return to a self-regulating pattern of booms and busts, but appear to have been moved to new equilibrium with much less productive economic patterns. So the economies moved very quickly from a 4E to a poorly performing 4A/B. It is the belief of the author that keeping interest rates and payout ratios too low allows a second common form of macroeconomic suicide. (The first form of economic suicide is introduced in section 4.6. Both forms of suicide are discussed in more detail in section 4.10.) A very important point to emphasise in the models above is the absolute lack of stochasticity. While there is certainly a significant element of stochasticity in real markets, the macroeconomic model above contains no stochasticity. The model is not stochastic, it is merely chaotic. Chaotic models like this are common in physics, astronomy, biology, engineering, and in fact all of the sciences other than economics, where determinism has hunkered down for a very effective last stand. The failure of these models to penetrate into mainstream economics, given the obvious turbulence of stock, commodity, housing and other financial markets, is puzzling. This endogeneity of chaos in business cycles is of profound importance. Standard economic theory, whether Keynesian lack of demand or the impacts of technology in 'Real Business Cycle theory', never mind neoclassical economics, seems incapable of believing that chaotic short term behaviour can be anything but externally driven. Exogenous drivers are simply not needed for quasi-cyclical, or explosive chaotic behaviour; all that is needed is the use of the correct modern mathematics, where 'modern' means post 1890. This mathematics, and chaotic systems in general, is discussed in section 6 below. As discussed above, Lotka-Volterra models have been used in Marxian analysis by Goodwin and others, though the models can be somewhat complex. The models presented above seem more efficacious than the Goodwin type Lotka-Volterra models, as they don't need: • population change • growth in labour force • technology change • productivity growth • inflation (long-term) • accelerators 109 EFTA00625237 all of which are used as standard in the Goodwin and descendant models. A central problem in the thinking of Goodwin and the researchers that followed Goodwin is in the idea of growth. It appears to have been assumed that to model short-term cycles of growth and decline it was necessary to include long-term economic growth rates. So these models include growth in the labour force, productivity, money supply, etc. This is a bit like trying to model waves on the ocean's surface by including things that cause changes in sea level, such as the tidal effects of the sun and moon, evaporation, precipitation, glacier melt rates, etc. This brings in a lot of irrelevancies into the basic model, and make it very hard to build the basic model. Even without any of the things listed above, natural cycles can occur that build up too much capital. That is not to argue against the secondary importance of any of the above factors, especially in long-term economic cycles. Going back to the evidence of Harvie [Harvie 2000] and Barbosa-Filho & Taylor [Barbosa-Filho Taylor 2006], the cycles for the mainland European countries appear to be long term, on a decadal scale; which would suggest a strong role for technology change and productivity growth (though very little for population change). However the cycles for the US and UK appear to show much faster oscillations; of only two to three years. Intuitively it is difficult to see how technology change could operate significantly on such short timescales, and this is more suggestive of the operation of the normal business cycle modelled above. Indeed the simple model proposed above may be more appropriate for modelling the regular short period cycles of booms and crashes seen in Victorian times. The important thing to note is that the basic instability in financial markets is much deeper than that proposed by Goodwin. Goodwin style feedbacks may exaggerate this basic cycle, or add longer super-cycles, however in this regard it appears that the basic insight of Minsky and the Austrians with regard to the essential instability of capitalism was correct. However, although I believe this basic Minskian/Austrian insight is valuable, it is also notable that to build model 4A to E, and create dramatic business cycles, you don't actually need any of the following: • governments • fiat money • fractional reserve banking • speculators • Ponzi finance • debt deflation or other common elements of the Austrian school or the work of Fisher and Minsky. 110 EFTA00625238 Debt, in the form of a negative cash balance, certainly does appear in the cyclical and explosive models. But models 4C and 4D show that the debt follows the cyclical instability of capital rather than the other way round. I would not wish to understate the importance of debt in exacerbating business cycles, indeed the role of debt appears to be very interesting and important, and is discussed further in 4.6 below. However debt itself is not the prime cause of the business cycles. Again, it is not suggested that any of the factors listed above are unimportant, however it appears that all the other factors are just potential magnifiers of an underlying inherent instability The instability is very basic, and, in the short term at least, perfectly rational. The instability arises, as Minsky noted, from the fundamental fact that paper prices of assets are based on projected future cash flows, not on costs of production. This is Minsky's crucial insight, of much greater importance than his analysis of the debt cycle. This is the same assumption originally proposed in the companies model in section 2.2 above. This instability naturally produces a growing cycle of apparent wealth, which is turned into excess capital as predicted by Hayek [Hayek 1931] in Austrian business cycles. But contrary to the Austrians, and in line with research data [Kydland & Prescott 1990], the liquidity or excess paper wealth is initially generated within the valuation system of capitalism, not by lax government policy. Creation of liquidity and monetary growth are endogenous to the basic pricing mechanisms of the finance system. Endogenous creation of financial wealth then feeds back into the creation of more real capital, so creating more financial wealth. This endogenous creation of financial wealth then gives apparently secure paper assets against which debt can be secured, and of course this debt allows yet more capital creation. Clearly, if the underlying system is unstable, with endogenous liquidity production a la Minsky; then other factors such as excessive debt, speculation, fractional reserve banking and inappropriate central bank intervention policies will all magnify the size and damage of the underlying cycles. But it is not excessive debt, speculation, fractional reserve banking or poor central bank policy that causes the boom and bust cycles. The cycles are caused by the basic pricing system of capitalism. Governments may of course fail to calm the markets by extracting liquidity in a timely manner, but it is scarcely the fault of governments that most investors are momentum chasers rather than fundamental analysts. Just as central banks are expected to control changes in the money supply caused by fractional reserve banking, it seems appropriate that they also need to control liquidity growth caused by Minskian asset pricing. This is discussed in more depth in section 8.2.1 on liquidity below. As noted previously, Minsky, although a follower of Fisher and Keynes, shared the Austrians' disdain for mathematics. It is the author's belief that bringing in a dynamic mathematical approach, on the lines of Lotka-Volterra modelling, to Minskian and Austrian ideas might not only give more weight to both these approaches, but also show them to be very comfortable bedfellows. 111 EFTA00625239 Essentially the company, commodity and macroeconomic models are all simple composites of ideas from Minsky and the Austrian school, though my producing them in this way happened more by accident than design. The models have Minsky's basic split between 'normal' assets such as goods and services that are priced on a mark-up basis, and financial assets which are priced on the basis of expected future cash flow. Following Minsky, and ultimately Keynes, the expectations of future flows are simplistic projections of present flows [Keen 1995]. Unlike Minsky the models use simple known behaviour of capital to explain the source of instability. In the companies model this was company managers hoarding incoming spare cash, and using it to build more capital. In the commodity model the instability was caused by the time actually taken to build and install new capital. In the macroeconomic models, either or both of these factors could cause instability. In this sense the models follow Austrian ideas. This has the advantage over the Minsky models that you don't need a complex financial system; speculators, Ponzi finance, etc, to form the instability. You can get the instability in pretty much any system where financial assets can be overvalued; this can be Industrial Victorian Britain with its savage business cycles, or even the Roman Empire (see 4.10 below). The critical insight of Minsky, in contrast to the Austrians, and seen in these models is that liquidity and new credit are generated endogenously in even the most basic of financial systems. You don't need governments to create excess credit, though certainly they can make things worse. In fact, faced with endogenous credit creation, you do need governments to actively remove credit and liquidity when financial assets become overpriced. In defining this macroeconomic model, a number of assumptions were made. I would like to briefly review these here: Note that the assumption of conversion of capital to equity at par in a downturn does not undermine the arguments. The losses incurred in a fire sale of assets to meet investor demands would simply exaggerate the viciousness of the cycles downwards. It was assumed that the ratio of capital to labour is fixed over the time of the business cycle, and that it is not possible to substitute capital for labour. There are two parts to discuss with this assumption. Firstly, in the short term, going into a boom, replacing labour with capital would simply allow further excess capital to be installed before wage inflation would kick in, so making the booms even larger. The resultant larger overhang of capital would then make the following slump more severe. So relaxing this assumption would simply make the business cycles worse. More importantly, the model shows that, in the long term, at the level of the economy as a whole, there is in fact a fixed ratio of capital to labour at any given set of market conditions. So it is not actually possible to substitute one for the other. Much more on this in sections 4.6 to 4.8 below. Note that allowing the market interest rate to float, say by making it the moving average of real returns over the previous few periods, would also have a large magnifying effect. As more capital was employed, overall interest rates would go down, making previously unprofitable capital investment profitable. Again this would encourage further excess capital creation in the booms. Finally I would like to return to a major assumption of the companies model in section 2.2. In this model capital was deliberately, and artificially, renormalised in each of the model iterations 112 EFTA00625240 to keep a constant value of K. As I hope is now clear, this was a necessary fix in the company model to prevent the introduction of severe cycling in the model output. Directly comparing my own macroeconomic models with those of Wright is not straightforward. My own models include a financial sector which is clearly more realistic, as Wright acknowledges in the Social Architecture of Capitalism [Wright 2005], where he notes that conflation of capital concentration with firm ownership may distort modelling results. So clearly Wright's models can not show cycles of debt build up and draw down. Despite this Wright's models do show recurrent booms and recessions, with much more complex behaviour than my own. Although Wright's business cycles are debt free, he models individual companies/owners, where my own model models the business sector as a whole. As a result recessions in Wright's models are of differing length and are quasi-periodic, this is clearly superior to my own models. Wright's models are also superior to my own in that they include for unemployment. My models just measure total over-employment and under-employment against a nominal full employment. Despite these substantial differences both Wright's and my own models produce cyclical endogenous business cycles from simple models based on statistical mechanics and classical economics. 4.5 A Present for Philip Mirowski? — A Bowley-Polonius Macroeconomic Model "I mean the stability of the proportion of national dividend accruing to labour, irrespective apparently of the level of output as a whole and of the phase of the trade cycle. This is one of the most surprising, yet best-established, facts in the whole range of economic statistics Indeed...the result remains a bit of a miracle. "[Keynes 1939] "...no hypothesis as regards the forces determining distributive shares could be intellectually satisfying unless it succeeds in accounting for the relative stability of these shares in the advanced capitalist economies over the last 100 years or so, despite the phenomenal changes in the techniques of production, in the accumulation of capital relative to labour and in real income per head."[Kaldor 1956] "FUTURE ISSUES - Theory 1. Is there a deep explanation for the coefficient of 1/3 capital share in the aggregate capital stock? This constancy is one of the most remarkable regularities in economics. A fully satisfactory explanation should not only generate the constant capital share, but some reason why the exponent should be 1/3 (see Jones 2005 for an interesting paper that generates a Cobb-Douglas production function, but does not predict the 1/3 exponent). With such an answer, we might understand more deeply what causes technological progress and the foundations of economic growth."[Gabaix 2009] Whenever economists hit a bad patch, it is inevitable that outsiders will begin to sneer how it is not a science and proceed to prognosticate how "real science" would make short work of the crisis. This is such a tired Western obsession that it is astounding that it has not occurred to critics that such proleptic emotions must have occurred before, and are thus themselves a part of a chronic debility in our understanding of economic history. As I have shown elsewhere in 113 EFTA00625241 detail, neoclassical economics was born of a crude attempt to directly imitate physics in the 1870s, and American orthodoxy was the product of further waves of physicists cascading over into economics in the Great Depression and WWII... ...Actually, it is understood among the cognoscenti that physicists have again been tumbling head over heels into economics since the 1980s, as their own field experienced severe contraction at the cessation of the Cold War. And where did most of them end up? Why, in the banks, of course, inventing all those ultra-complex models for estimating and parceling out risk. Some troubled to attain some formal degree in economics, while others felt it superfluous to their career paths. In any event, the exodus of natural scientists into economics was one of the (minor) determinants of the crisis itself—without "rocket scientists" and "quants," it would have been a lot harder for banks and hedge funds to bamboozle all those gullible investors. So much for the bracing regimen of a background in the natural sciences. If anything, responses to critics that tended to pontificate upon the nature of "science" were even more baffling than the original calls for deliverance through natural science in the first place. Economists were poorly placed to lecture others on the scientific method; although they trafficked in mathematical models, statistics, and even "experimentation," their practices and standards barely resembled those found in physics or biology or astronomy. Fundamental constants or structural invariants were notable by their absence. Indeed, one would be hard pressed to find an experimental refutation of any orthodox neoclassical proposition in the last four decades, so appeals to Popper were more ceremonial than substantial. Of course, sometimes the natural sciences encountered something commensurable to a crisis in their own fields of endeavor—think of dark matter and dark energy, or the quantum breakdown of causality in the 1920s—but they didn't respond by evasive manoeuvres and suppressing its consideration, as did the economists. In retrospect, science will be seen to have been a bit of a red herring in coming to terms with the current crisis. In the heat of battle, economists purported to be defending "science," when in fact, they were only defending themselves and their minions. [Mirowski 2010] As a physicist myself, I am somewhat embarrassed to admit that physicists as a class stand guilty as charged when accused of unnecessarily increasing the complexity and opacity of finance. This is the more embarrassing as the behaviour is so far from the norm in physics, where careful investigation and gaining of understanding is the general aim, and true kudos is gained by discovering neat and beautiful solutions to seemingly complex and insoluble problems. The entry of quants into finance seems not only to have been marked by a joy in the deliberately complex, but also a wilful desire to avoid any understanding of what is really happening in an economic or financial system. As previously noted, physicists seem very comfortable in using wealth and income interchangeably, some even conflate these two concepts with money. From my own conversations, I am led to doubt whether a majority of physicists working in finance could successfully define the difference between a real and a financial asset. As a penitence, on behalf of a profession behaving badly; I had hoped in this section to present to Philip Mirowski the explanation of a basic 'constant' in economics. Sadly for me, the constant turns out not to be constant at all but merely a humble ratio; an indicator of an underlying equilibrium. Unfortunately it cannot be described as either 'fundamental' or 'invariant'. On the bright side this at least allows for changing of the 'constant', and indeed it is one of the aims of later sections to change this 'constant' to the benefit of the population in general. Even more worryingly this constant may simply be seen by many as a trivial accounting identity, a red herring at best. 114 EFTA00625242 I do not believe this is the case and, however humble this ratio may be, I believe it is the first 'constant' to be explained in economics, and as such is worthy of note. The constant in question is the ratio of earnings received by labour to those received by labour and capital, the Bowley ratio 13 that was first introduced in section 1.3 above. Before looking at the derivation of the Bowley ratio, it is worth considering this 'constant' in more detail. For most mature economies the constant varies between about two-thirds and three-quarters and can be very stable, as discussed in section 1.3 above. Young gives a good discussion of the national income shares in the US, while Gollin gives a very thorough survey of income shares in more than forty countries [Young 2010, Gollin 2002]. In emerging economies 13 can be much lower, as low as 0.5. Currently, and exceptionally, in China it may be as low as 42% [Bai et al 2006, Subramanian 2008]. Arthur Lewis [Lewis 1954] has explained this as being due to wages being artificially depressed by the reserve of subsistence workers simultaneously with the wealthy being able to save more due to low living costs caused by low wage rates. Once economies absorb this spare rural labour, and pass their 'Lewisian turning point', then the ratio of returns to labour to total income stabilises and moves only slightly. In the UK, the first country in the world to absorb its rural labour force, the ratio has been fairly stable for a century and a half. The thing about this stability is that the more you consider it, the more bizarre it seems. In the last 150 years Britain has changed from a nation of factories powered by steam engines to a modern service economy. The amount of capital currently installed in the UK is many times greater than that of 150 years ago, labour intensive industry has all but disappeared. Wealth levels have changed incredibly. In the 1850s gdp in the UK was comparable to current gdp in Indonesia or the Philippines, however life expectancy in the UK in the 1850s was roughly half that of Indonesia or the Philippines today [gapminder]. It is quite extraordinary that the Bowley ratio has remained roughly constant throughout this period. In fact it is counter-intuitive. For somebody in Victorian Britain, as in modern day Indonesia, the majority of income would have been spent on food and basic housing, with little left over for anything else, most money is paid to other people carrying out labouring duties. As incomes rise it would naturally be expected that more money would be spent on manufactures and property, and that more spare cash would be available for investing in capital of one form or another, so increasing the returns to capital. Also, as wages rise it would also seem sensible for capital to substitute for labour, and again for returns to capital to increase at the expense of labour. In the long-term total factor productivity should increase, reducing the returns to labour and increasing those to capital. Indeed futurologists have been predicting for most of a century that as capital gets more efficient and productive the need for labour should slowly decline to nothing. To date these predictions have been conspicuously wrong. Working weeks have barely declined in the last forty years, huge numbers of women have entered the labour markets and people continue to complain of the problems of the work/life balance. Indeed at the time of writing this section France is currently paralysed by strikes trying to prevent an increase in retirement ages. In the long run it seems logical that mechanisation and the increasing use of capital would result in the Bowley ratio slowly moving towards zero. 115 EFTA00625243 In fact if you analyse the data on a sectoral basis, this is exactly what is happening. Young [Young 2010] shows clearly that for agriculture and manufacturing, returns to labour have declined significantly while returns to capital have increased. In the US returns to labour in agriculture have dropped from nearly 0.8 of total income in 1958 to less than 0.6 by 1996. In manufacturing, the change has been from 0.75 to two-thirds. This has happened because labour has been slowly displaced by machines in these industries. The fascinating thing is that despite the changes in the Bowley ratios for these two (very large) sectors, the national value of the Bowley ratio has stayed near constant between 0.69 and 0.66 using the same measures. The reason for this is that the labour intensive service sector has grown dramatically in size through the same period, and this has kept the national balance of returns to labour and capital very nearly constant. In the discussions that follow it is hoped that these puzzles will be explained. As shown in section 4.3 above, the output from a fairly randomly chosen model 4A produced an output with a Bowley Ratio, of waged earnings to total earnings, of exactly 0.75 with zero debt (It is to be noted, that Wright found similar results with (3 equal to 0.6 and 0.55 in his two papers). This was the subject of further modelling. A first problem with the models used in section 4.3 above is that they have too many degrees of freedom. Depending on the parameters and the starting values of a model run, different zones of stability can be encountered, and even if the model is restricted to options that end in stable, stationary outputs, different end points can be reached with the same parameters, but different starting positions. A second problem is the role of the 'cash balance', H, which can either be a positive surplus or a negative debt. In many of the models the stable output can have very large positive or negative cash balances, with an order of size of the capital wealth Q. As is often the way with debt, an item that was used as a minor temporary convenience ends up taking on a major unlooked for negative role. Having been introduced as a simple method of ensuring that the sums add up; the role of this cash balance is not clear, and it is not obvious that it is a meaningful item. There are problems as to exactly who or what this money is borrowed from / lent to, and also why interest is not charged on the lending or borrowing. Firstly, to remove these problems, the models were rerun in Excel, deliberately choosing parameters that stabilised into stationary outputs. A second condition used was that the payout ratios, both positive and negative, were set to 1.0. This makes for an immediate simplification of the model, as company payouts are just to the market expectations and make no reference to the profits produced by the companies. In this model payout ratios are not necessary, because although the total capital can increase and decrease, other mathematical limitations prevent the capital from shrinking to zero, at least in the stationary and periodic zones. 116 EFTA00625244 Thirdly, using 'solver', the range of stationary outputs was then restrained to the single solution that satisfied the requirement that there be no net borrowing or lending, ie the cash balance was always constrained to zero. This gives the single 'Bowley-Polonius' equilibrium point. With net borrowing and lending fixed to zero, the philosophical problem of what exactly the cash balance is becomes irrelevant. By changing the parameters of the model systematically some very interesting results arose. The first interesting thing was the role of the pricing of labour. As discussed in section 4.2 above, this model assumes that labour can be a scarce supply, and that the price of labour depends on the amount required. As such the concept of marginality has introduced it's way into the modelling in the form of a pricing curve for labour, this is a reasonable argument, as labour is a commodity that is truly unchangeable in it's supply. However, investigating the model shows that the actual form of the curve is not relevant to the model. If you change the parameters of the labour curve, then the model values change, with an offsetting increase or decrease in the cash balance. But if you reoptimise the model and force the cash balance back to zero, then the model returns to an equilibrium point with exactly the same value for the Bowley ratio. This is looked at again in section 4.7. Within reason, the parameters of the labour supply curve are simply not relevant to the ratio of wages to profits. The curve simply needs to be concave, to ensure that labour costs eventually choke the growth of the economy with higher costs. Any reasonable concave curve will do this. So the actual detailed calculations of marginality are utterly irrelevant and do not have any influence on the long-term equilibrium. ('Within reason' means that there are some labour curves that prevent the model coming to an appropriate equilibrium; that is they don't allow an equilibrium at zero cash balance. But as long as the curve allows an equilibrium, the parameters of the curve do not effect the location of the equilibrium). The second interesting thing is that, at the B-P equilibrium, the Bowley ratio is influenced by only two things; the consumption rate and the profit rate. Moreover, the ratio is given by the very simple form as follows: = Bowley Ratio waged income total income f2 — r f2 = I — (r/f2) (4.5a) It is straightforward to check equation (4.5a) against reality. A suitable long-term profit rate could be anywhere between long-term interest rates and long-term real stock-market returns. Long-term real interest rates are generally in the region of 2% to 5% [Homer & Sylla 1996, 117 EFTA00625245 Measuring Worth], see also figure 4.5.1 below. Long-term stock-market returns appear to be in the region of 7% to 8% [Campbell 2003, Ward 2008] see also figure 4.5.2 below. Consumption is typically about 60% of gdp [Miles & Scott 2002, section 2.2, fig 2.3]. While non- residential capital stock is typically 2.5 to 3 times gdp [Miles & Scott 2002, section 5.1 & 14.1]. Taken together this would give ,Q, the consumption rate as a proportion of capital a range of about 0.2 to 0.25. Substituting into equation (4.5a) this then gives a possible range of values for the Bowley ratio of between 0.60 and 0.92 Clearly this range is a little on the high side when compared with the 'stylised facts' of observed Bowley Ratios in the real world varying between the values of 0.5-0.75. We are however in the right ballpark. (The figures also confirms the common sense notion that stock-market returns are more appropriate than interest rates for 'r'.) As discussed above, intuitively it is not obvious why Bowley's law holds and the ratios of returns to capital are not much higher than the returns to labour. Using the basic ideas of classical economics we would expect the returns to have increased significantly as machines have got steadily more productive over the last two hundred years. Neoclassical ideas of utility and marginality have no theory to explain this. What equation (4.5a) says clearly is that Bowley's ratio will always be less than one, and given that rates of return are generally much lower than consumption rates, the value will be closer to one than zero. This agrees in general with the stylised facts, if not in detail. In section 4.6 below possible reasons for the mismatch between the values produced in the model and the real world models are discussed. These reasons are speculative, so before moving on to this I would first like to discuss the equation (4.5a) and its consequences in a little more detail. Firstly it should be noted that this equation was discovered by experimenting with the parameters of the model. The results from the simulations give results that match the formula above to multiple decimal places. With a little playing it turns out that it is in fact quite straightforward to derive formula (4.5a) from first principles. Firstly, when the model is at equilibrium, all values of flows and stocks are constant (in this part of the modelling, only models giving stable time outputs were used, the models suggest that the periodic models move around this point on average, as would be expected in a Lotka-Volterra model). At this equilibrium point, if the total capital Q is to be constant, then the total income must equal the total outgoings, so the algebra works as follows (note that for simplicity the summations have been dropped, all variables are assumed to be summed over the whole economy): 118 EFTA00625246 Consumption = Income C = Y = e + rr (4.5b) Here, at the Bowley-Polonius equilibrium, H = 0 and W = Q. Also, the consumption ratio SI is defined by: S2 =Q (4.5c) Trivially, the profit rate is defined by: (4.5d) If we multiply equation (4.5b) by equation (4.5d), then we get: TIC = rY (4.5e) Substituting from (4.5c) into the left hand side gives: rrf2 = rY (4.5f) Rearranging gives: TT = — Y 12 (4.5g) substituting from (1.3u) gives the profit ratio: r P = f2 (4.5h) Subtracting both sides from unity gives: 119 EFTA00625247 I — p = I — (4.5j) or, substituting from (1.3v): = Bowley ratio = I — — 12 (4.5k ) The base equation here is (4.5h) which is the ratio of returns from capital, to total returns. This equation looks suspiciously like an equation of state, discussion of which will be postponed to section 4.7. Whether equations (4.5h) and (4.5k) are sufficiently 'fundamental' to satisfy Phillip Mirowski remains to be seen; I would ask judgement to be reserved until the end of section 4.7. Multiplying consumption by interest rates isn't an 'obvious' thing to do, and clearly I discovered this derivation by reverse engineering my model output. At this point, more observant readers may have noticed something familiar about equation (4.5k). Equation (4.5k) gives: = I — i (4.5k) while back in section 1.3 equations (1.3v) and (1.3w) defined the Bowley ratio as: = I - (4.51) This is made simpler by looking at the profit ratio p, then (4.5h) and (1.3w) give: (4.5m) which clearly means: 12 = F (4.5n) from the definitions of .S2 and f it then follows that: 120 EFTA00625248 (4.5o) Where C is the consumption and Y is the total income from wage earnings and profits/dividends, etc. From which trivially we arrive at: C = Y (4.5p) which we have seen a very long time ago as (1.3b). This is of course a basic assumption of all traditional macroeconomics, and so is something of an anticlimax; like setting out across the Atlantic to find the Indies, and instead discovering Rockall. It is however firstly worth noting that while this identity is an assumed equality in traditional economics, it is a self-balancing outcome of the GLV and L-V models used in this paper. Consumption is not defined as equal to income or vice versa, consumption of individuals rises and falls with wealth, wealth changes with income and consumption, income depends on consumption. In the models in this paper the dependencies go round in circles, hence the Lotka- Volterra outputs, the equality of total income and consumption naturally falls out at the equilibrium of the model. This leads to a much simpler derivation of the Bowley ratio: 13 = I — p and p= so = I - by definition, also: 12 = — and r = - by definition, but C = Y so: 0 = r so t3 = I - f2 by definition, and so: QED. Of course the definition above does not require a single line of my modelling, theorising or pontificating. And for most economists it will appear to be a trivial and unimportant accounting identity. 121 EFTA00625249 But it isn't. It is all a question of directionality. Of cause and effect. For most people it is 'obvious' that consumption follows income, ie that people earn then spend, or that: C = Y Actually it is the other way round: Y = C r = f2 or more accurately: It is the consumption rate SI that defines 1; the ratio of total income to capital. Trivially this is the case in my models, where r and SI are fixed and r is allowed to float. But of course this is not sufficient justification. The problem with the economic literature with regard to the Bowley ratio is that economists have first defined the profit ratio and Bowley ratio as: P = /3 = 1 — r They have then spent the last hundred years or so trying to explain the two ratios above by attempting to look at the microeconomic structure of industry that could affect r and 1. This has almost entirely revolved around the analysis of 'production functions', the supposed microeconomic relations between capital and labour. The Cobb-Douglas production function has become a particular focus of attention, as its form gives rise to constant shares of returns to labour and capital. (I am somewhat reluctant to criticise Gabaix, as he is one of the few economists who has recognised the importance of power-laws and other 'anomalous' invariants in economics. However his quote at the start of this section shows how deeply ingrained within economics this approach has become. Gabaix defines the solution to the problem of the Bowley ratio as the finding of a theory that not only produces the Cobb-Douglas production function, but also gives certain fixed exponents for the Cobb- Douglas function). 122 EFTA00625250 There are however very major problems with this approach. Firstly, real analysis of companies suggests that any meaningful production function needs to be based on high fixed costs and increasing returns, and is far away from the Cobb-Douglas or other standard production functions used in neoclassical economics. Secondly, as the data from Young [Young 2010] shows the relative shares accruing to labour and capital can change quite significantly within individual sectors such as agriculture and manufacturing. This shows that production functions are not giving the required output on a sector-by-sector basis. (Casual inspection of company accounts shows that returns to labour and capital can vary dramatically from company to company.) The third and most important reason is the problems following the logical steps. Firstly, traditional economics states that production functions define the relationship between r, the rate of return to capital, and 1, the rate of total income to capital. Secondly, traditional economics states that total income is equal to total consumption, so, logically, ,S2 = 1. Putting these two statements together logically means that production functions, the microeconomic structure of the commercial sector, define the saving rate Si (This leaves aside r for the moment, we will return to r shortly.) This is very difficult to swallow. Squirrels save. As do beavers. And also some woodpeckers and magpies. Laplanders build up their reindeer herds as a form of saving, as also Arab pastoralists build up their herds of camels and goats, and the Masai and BaKgalakgadi build up their cattle herds. Almost all agricultural societies store grains and other foods to tide them from one harvest to the next. And whether you live in the tropics with alternating wet and dry seasons, or a temperate climate with warm and cold seasons, saving is a biological necessity genetically selected in human beings for its beneficial outcomes. From a behavioural point of view saving is a deeply ingrained human behaviour that borders on the compulsive. Most people put money away for a rainy day. While Bill Gates and Warren Buffet have shown extraordinary benevolence, they both continue to hoard wealth far beyond their possible needs. Leaving biology aside, traditional economics has well-established logical theories for saving. Lifetime cycles make it logical for young, and especially middle-aged people to save to ensure support in their old age. Whether you look at biology or economics, savings rates are largely exogenous to the economic system. They are defined by people's assessment of, and fear of, an unknown future. Clearly my use of 12 as a consumption function is simplistic. S2 uses only total wealth as a definer of consumption. In reality consumption and saving decisions are going to depend on current income and projected earnings in a complex manner. In particular, individual consumption and spending decisions will vary significantly with age and family circumstances. 123 EFTA00625251 Indeed an interesting paper by Lettau and Ludvigson [Lettau & Ludvigson 2001] suggests that there is a constant rebalancing of asset wealth to ensure long-term consumption, and that this feeds back predictably into asset prices. In reality, as people are born and die at roughly the same rates, the total pattern is relatively fixed, and over the long-term national consumption rates are relatively steady. Clearly consumption and savings rates are affected by economic fundamentals. Savings rates go down, and consumption goes up in booms, when returns look good and fear of unemployment is low. In recessions savings rates go up, and consumption goes down, as returns go down and fear of unemployment is high. But these reasons simply reinforce the hypothesis of exogenous drivers of biology and economic lifetime planning for consumption and saving. Despite the changes with economic cycles, over the long-term, savings rates show consistent trends linked to the relative wealth of a society, as originally described by Lewis [Lewis 1954]. The point here is that SI can be explained by long-term societal trends such as age, sex, family size, amounts of spare labour in a society and the state of a country's social-security system. Short-term trends can be explained by return rates of investments, unemployment rates, etc. While n is not an absolutely fixed exogenous variable, it is a slow-changing variable that can be calculated from mostly long-term variables. It stretches credulity to breaking point, to believe that saving and consumption behaviour is ultimately defined by the microeconomic production functions of commercial companies. The causality works the other way, the systems of capitalism are set up in such a manner that the consumption rate n defines r, the rate of total income to capital. When viewed in this way the data of Young makes sense [Young 2010]. In the period Young analysed, consumption rates stayed approximately constant, as did rates of return. During the same period, both agriculture and manufacturing increased their returns to capital and reduced returns to labour. Given fixed n, to keep things balanced, the economy as a whole was obliged to create new, labour-intensive, industries to ensure that returns to labour were maintained as a whole. All those cappuccino bars and hairdressers were created by the economy; by entropy, to ensure that the Bowley ratio remained equal to 1-(r/f/). In fact the consumption rate n, the Bowley ratio (3, and the profit rate p are not very interesting pieces of economics at all. S2 is already well defined by life-time planning and/or behaviouralism. The Bowley ratio and profit ratio are trivial outcomes from n and r. I find it difficult to believe that I am the first researcher to propose that the Bowley ratio should be defined by: 124 EFTA00625252 r 0 = I — f2 rather than : 0 = I — r However, I have not been able to find any other proposal of this relationship, and the recent writings of Gabaix, Young and others suggest that this is the case. If I am the first to do so I am happy to take the credit. If not I would be happy to update this manuscript appropriately. The interesting economics is in r; the rate of returns. To date I have generally been vague about the meaning of r and have included dividends and interest payments as well as rents in r. In fact there are three near economic constants which all show very stable long-term behaviour. In all three cases the behaviour is counter-intuitive and I believe likely to be related. The three variables are long-term real interest rates, long-term stock returns and long-term gdp growth rates. Figure 4.5.1 below shows the long-term cumulative returns due to real interest rates for the UK and the US. For the UK this starts with a value of 1.0 in 1729, for the US the start is at a value of 1.0 in 1798. The returns are calculated by multiplying the successive value from each year by the interest rate less the inflation rate. Data for these graphs, and also for the gdp graphs below were taken from the website 'Measuring Worth', for a very full discussion of historic interest rates see Homer and Sylla [Homer & Sylla 1996, Measuring Worth]. Figure 4.5.1 As can be seen, although there is significant variation around the trend, there is a very clear long-term trend, which is slightly over 2% for the UK and slightly over 4% for the US. Figure 4.5.2 below shows long-term stock-market returns for the USA, from 1800 to 2008. Figure 4.5.2 [Ward 2008] Again, although there are significant short-term variations, the long-term trend of 7% is clear. Finally figure 4.5.3 below shows real GDP in 2005 dollar for the United States from 1790 and 2005 pounds for the United Kingdom from 1830. The same long-term trend can be seen. This 125 EFTA00625253 time the trend is slightly below 2% for the UK and slightly below 4% for the US. The match of long-term gdp growth trends to long term interest rates is striking. Figure 4.5.3 In the discussions above, I have chosen r as an exogenously given constant. I have been vague about whether r should be the 2-4% of interest rates or the 7% of stock-market returns, or somewhere in between. This is, of course, because I don't know. I suspect it is somewhere between the two. I do think the assumption of exogeneity, at least for the level of discussions in this paper, are reasonable. Like the Bowley ratio, both interest rates and stock-market returns show long-term constancy. The Bowley ratio is the dull one, as it is simply a result of the regularity of returns r and consumption propensity CI. (As an aside, a quick note on the changes of the Bowley ratio in recessions. It is well known that returns to labour increase in recessions, and so that the value of R increases. It is also well known that saving increases and consumption decreases in recessions. If consumption decreases, then equation (4.5k) would mean that R would decrease, which appears to be a contradiction. However in recessions both interest rates and stock-market returns also decrease, and the proportional decrease in interest-rates and stock-market returns is usually much larger than the decrease in consumption. So, overall, 13 does increase in recessions despite falling consumption.) The interesting thing is where the constancy of interest rates, stock-market returns and gdp growth all come from. Traditional economics has tended to look at technology change and microeconomic factors as the drivers, again this seems difficult to justify. Firstly, technology tends to come in bursts; steam power, electrification, motorised transport, electronics, the internet, etc. This would suggest that both gdp growth and stock-market returns would come in bursts, and not necessarily bursts with the same rate of growth. Secondly, the rate of change of technology, from casual observation, appears to be accelerating, with the bursts of new technology becoming more frequent and wide-ranging. Thirdly, the growth of economies appears to be back to front. For the UK, growth started with the industrial revolution somewhere around 1800 and has continued at a regular rate of 2-2.5% for the last two centuries. Almost all the other rich countries have followed a different path. In the first phase of the catch- up they generally had high rates of growth; typically between 5% and 10%. Until they caught- up or slightly over-took the UK. From that point on they then slowed down to a similar 2-4% rate as the UK. For a very good visualisation of the process go to gapminder [gapminder]. 126 EFTA00625254 This is counter-intuitive, as common sense says that as countries get wealthier they should be able to devote more and more capital to investment, and so they should be able to grow more rapidly, not less. The constancy of the values of interest rates, returns and gdp suggest a much deeper equilibrium is present, a simple mathematical equilibrium. An equilibrium that is actually restraining growth significantly below that possible as a consequence of technology. It is the source of these three constants, and the relations of the three to each other, that is the most pressing mystery of economics. A possible, though highly speculative, proposal for the source of this equilibrium is suggested in section 7.4. Before moving on, I would like to discuss the parallels with Wright's models. In the Social Architecture of Capitalism Wright's model produces a value of p of 0.55, while in Implicit Microfoundations for Economics p is 0.6. Wright's models are not formally mathematical, so it is not fully clear how these values are generated. In both these papers the expenditure is drawn randomly from a uniform distribution of an agent's wealth, which I believe makes S2 equal to 0.5 in both models. The way that excess wealth is generated in Wright's models is much more complex, and possibly recursive, and it is not clear (at least to me) how the equivalent to interest rate in these models would be calculated. If equation (4.5a) proves to be correct then Wright appears to have defined the interest rates for the two papers above at 22.5% and 20% respectively. Finally, it should be noted that equation (1.6d) for the exponent of the wealth distribution power-law tail should now read as: OC = 1.36(1 - (r/12)) 1.15 V (4.5q) 127 EFTA00625255 Part A.II - Speculative Building At this point in the discussion of the modelling, I believe it is appropriate to give a clear and unambiguous health warning. Up to this point in the paper; although both the economics and the mathematical approaches of the modelling have been heterodox, I believe that the models built accord with basic common sense, most notably with the various variables and constants matching, at least approximately, measurable quantities in real life economics. In the remainder of the first section of this paper, this no longer remains the case. For one reason or another the models and policy proposals in the rest of this section are speculative. The models have been included because they give results which may be interesting or plausible, and that may allow future building of alternate, more realistic, models in the future. The conclusions produced from these models must also therefore be presumed to be highly speculative. I fully expect that some or all of the models and conclusions below will prove to be wrong. It is my hope that they will however prove to be informative for further work. 4.6 Unconstrained Bowley Macroeconomic Models In section 4.5 above, we looked at Bowley models that deliberately constrained the net cash / debt balance to zero. In this section these models are explored further by changing the net value of the cash balance so it is positive or negative and seeing what happens. As previously discussed, I have a profound philosophical problem with this approach. It is not clear to me who is holding this balance or debt, where it is held, etc. Because of this no interest is paid on the balance, or interest charged on the debt, for the simple reason that I do not know where in the model I should debit the interest from, or pay the interest to. Despite this I am presenting the results because, firstly they are mathematically interesting, and secondly the outcomes are beguilingly plausible. I find this worrying, as it characterises some of the attitudes I have found most frustrating in my reading of much mainstream economics; the triumph of interesting equations and common sense over meaningful models related to underlying data. The first model run was simply to put in typical parameters, from real economies of: Returns rate r 0.03 Consumption rate a 0.2 Bowley ratio 13 0.7 Along with a Capital Wealth, Q 100 And let the model reach an equilibrium, the resulting cash balance is: Cash Wealth H -50 128 EFTA00625256 There are two things to note here. Firstly, allowing a negative cash balance; that is allowing the use of debt, allows the Bowley ratio to drop. This means that the returns to labour are reduced and the returns to capital are increased. So, in short, allowing the use of debt allows more returns to capital. It should be noted however that using an returns rate of 0.07, based on stock market returns, gives a positive cash balance of +17. To investigate this further, the parameters of the cash/debt balance were changed systematically, along with changes to other variables, to investigate the results on the model. As with the Bowley-Polonius model, the model was surprisingly easy to parameterise, and gives an equation as follows: = + n(H/Q) - r + O(H/Q) — 1 + (4/Q) — (110) (4.611) 1 + (H/Q) where H is the cash balance (wealth held in the form of cash or negative debt) and Q is the wealth held as capital. Again, this equation has been derived 'experimentally' by investigating the model, but the equation fits the modelling exactly. As in the previous section it is fairly trivial to derive equation (4.6a) from first principles. As before, when the model is at equilibrium, all values of flows, stocks and debts are constant. At this point, if the values of capital Q and cash H are to be constant, then the total income must equal the total outgoings, so, as before: C = Y = e + rr (4.6b) However this time, in the original model, in equation (4.2b), we defined the consumption ratio SI as: 129 EFTA00625257 12 = so, (Q + H) .O(Q + H) = C or, substituting from (4.6b): .O(Q + H) = Y (4.6d) again, the profit rate is defined by: rr = r() 4.6 c If we multiply equation (4.6d) by equation (4.6e), then we get: rrf2(Q + H) = iQY (4.6f) Rearranging gives: rr rQ Y 12(Q+ H) rQ P 12(Q + H) Subtracting both sides from unity gives: or: (4.6g) 1 p - 1 12(Q H) (4.6h) or from (1.3v): + 12(Q + H) — rQ 12(Q + H) .OQ + OH — rQ lags + 1 + (H/Q) — (rid?) = 1 + (H/Q) or, dividing by 0 and Q; (4.6a) Once again the base equation here is (4.6g) which is the ratio of returns from capital, to total returns. In the next section I would like to discuss the overall meaning of equation (4.6g) in more detail, but before that I would like to look at some consequences of varying the debt value H. 130 EFTA00625258 It can be seen from equation (4.6a) that the Bowley ratio can be manipulated by changing the value of the cash balance H. If the cash balance is positive and increasing, Bowley's ratio just heads closer and closer to unity, good for workers, bad for capitalists. More interestingly, if H is negative, a debt, and the size of the debt is increased, then the size of both the numerator and denominator reduce, however the value of the numerator reduces more rapidly than the size of the denominator, and the Bowley ratio slowly decreases. At least at first. If debt is allowed to continue increasing, then a rather dull function suddenly becomes more interesting. Firstly the Bowley ratio drops rapidly to zero, and then shortly afterwards heads off to negative infinity. In the model itself it isn't possible to reach these points; as the Bowley ratio heads to zero the model becomes unstable, and explosive — the economy blows up in an entertaining bubble of excess real capital and even more excess debt. This may sound familiar. This brings us to the first, more traditional, form of macroeconomic suicide; allowing too much debt in an economy. Again this is discussed in more detail later in the international model in section 4.10 below. Unfortunately the model gives no indication of the policies to be followed post explosion, though it does suggest that sensible limits on total debt (or debt ratios) in a well run economy might be a good idea. There is a further consequence of this model that is intriguing. In this model the role of debt gives a direct output to the Bowley ratio. As was found in section 1.6 above, the Bowley ratio in turn gives a direct output to the parameters of the GLV income distribution. So, if the above models hold, there is a direct link from levels of debt in the economy to the levels of inequality. Specifically, increased levels of debt lead to increased levels of inequality. Intuitively this seems plausible. Looking back over the last century, especially at the US, the first part of the century was associated with high levels of inequality, and high levels of leverage, which ultimately resulted in the Wall Street crash and the depression. In reaction to this, from the 40's to the 70's, leverage was strictly controlled, and also income distribution was much more equitable. From the 70's to the end of the 201° century, increased financial deregulation, and increased leverage, went hand in hand with increased inequality. Given the mathematical simplicity of equations (4.6g) and (4.6a) it should be straightforward to check these relationships both historically for individual countries as well as across different countries. It seems highly likely that the complexity of economics means that there are other factors that need to be included in equation (4.6g), for example, all the above has been carried out with payout factors fixed at one. However, with luck the errors might be systematic and relationships may appear. 131 EFTA00625259 As a minimum it should be noted that a more realistic version of (4.6g) would include net returns based on returns from investments, based on say 7% [Ward 2008] less returns on debt at 3%; representing long term interest rates. I would guess that this would give something like: (rk — ri)Q f2(Q + H) (4.6i) Where rk is the typical return on investments in companies and rf is a long term risk free interest rate. I emphasise that equation (4.6i) is merely a supposition and has neither been derived nor modelled. If actual economic data give support for the relationship in (4.6g) above, then this would give some support to the fact that the debt in equation (4.6g) was in fact a meaningful value. If economic data does support equation (4.6g), or a variant of it, then this raises interesting discussions on the role of debt in a national economy. The history of the last forty years has been one in which neoclassical economists have argued forcefully for the liberalisation of financial markets under the assumption that deregulation would allow deeper and cheaper financial markets and that self-regulation would ensure a natural balancing of an equilibrium. Equation (4.6g) begs to differ. Equation (4.6g) dictates that persuading governments to allow greater leverage merely allows benefits to the owners of capital, while simultaneously moving towards a more unstable equilibrium that coincidentally increases overall wealth inequalities. In fact this is the second form of rent-seeking we have seen exposed. If they were true to the core values of their religion, neoclassical economists would condemn this rent-seeking for what it is, and support strict controls on leverage. In practice neoclassical economists have consistently supported the 'freeing' of credit markets in the mistaken belief that greater access to funding will reduce prices and increase overall 'welfare'. In the real world any practical cost benefits are negligible compared to the disadvantages. The disadvantages are a substantial shift of funds from the productive sector of the economy to rent-seeking financiers, and a large transfer of 'welfare' from the poor to the rich. Equation (4.6g) suggests that control of the national level of leverage can provide three separate economic benefits. Firstly for the working of the economy there will be a optimum level of debt that allows liquidity and provides capital for genuine economically productive investment. Secondly, by preventing extreme levels of debt financial instability can be prevented. Thirdly, the level of debt may be reduced to achieve reduced levels of inequality. If the third item above is tackled successfully then the second becomes irrelevant, so the debate regarding the appropriate level of debt becomes a trade off between the first and third items. While the income distribution requirements suggest an elimination of debt, this is clearly not practical for a well functioning economic system. While much investment is funded directly from cashflow, if the economy is to grow successfully non-financial firms clearly need access to debt financing for major capital investments. 132 EFTA00625260 Similarly, while it is always fashionable to attack 'speculation' a significant proportion of speculation is clearly useful. Neither farmers nor bakers are experts at predicting weather patterns. Both use derivatives on grain production to hedge their prices. It is the entrance of speculators into the grains futures markets, speculators who are able to look at weather patterns across the different grain producing countries of the world, who keep these markets working effectively, so benefiting both farmers and bakers. The same is true of speculators in any derivative market when they are functioning correctly. However there are clearly points where derivative markets fail to be efficient finders of future prices and start to be used by uninformed momentum chasers as apparent sources of financial growth in their own right. Although the work of Minsky is not quantitative in nature, his characterisation of the phases of debt build up is clear and easy to relate to real economic cycles. If equation (4.6g) above is found to be applicable, it should be possible to look through past economic cycles and note where debt moved from a useful point; of providing funds for investment and price finding speculation, to turning into a self-sustaining provider of bubble finance. This would then provide central banks with a guide to controlling financial markets for the benefit of the economy as a whole. I would now like to look at the character of equations (4.5h) and (4.6g) in more detail. 4.7 A State of Grace It has been previously stated that equation (4.5h): r P = S2 (4.5h) for non-debt economies, and equation 4.6g: rQ P = (2(Q + H) (4.6g) for economies with debt, look suspiciously akin to what physicists call 'equations of state'. This is a very brave statement and time will tell if this proposition is accepted. However it is clear that the equations work in ways similar to equations of state, and this is important for understanding what these equations signify, especially with regards to economic equilibrium. Firstly I would like to give a little background of other equations of state in physics. Historically, the study of thermodynamics; things such as the expansion of gases, heat engines, heat production from chemical reactions, etc, was problematic because there were large numbers of macroscopic and microscopic variables. Changing one of the variables generally resulted in simultaneous changes in many other variables and it was very difficult to work out what was 133 EFTA00625261 actually happening. In this regard, classical thermodynamics was similar to present day economics. In the study of gases a series of pioneering scientists carried out various carefully controlled experiments that resulted in various relationships being established. So Boyle's law states that, at constant temperature, the volume of a gas varied inversely with the pressure. Charles law states that, at constant pressure, volume is proportional to temperature, and so on. Finally it was found that all the different laws could be put together to give the 'ideal gas law' in the form of an equation: PV = nRT (4.7a) where P is the pressure, V is the volume, T is the Temperature, n is the amount of substance in moles, and R is a fundamental constant of the sort wished for by Mirowski. In fact the 'fundamental' nature of R is an accident of history. The concepts and measurement units of pressure, volume and temperature were generated independently with idiosyncratic units. Here R is just a method of adjusting the different measurement systems so that the units fit together. Later microscopic theory showed that that the equation could be changed to a more fundamental form of: PV = NkT (4.7b) where N is the number of molecules, and k is another much more fundamental constant (Boltzmann's constant) that once again mops up all the different unit systems. If physicists were allowed to start from scratch they would change all the units so that the constants were all dimensionless '1's, which would make things easier for physicists but harder for butchers, bakers and shoppers. The point about equation (4.7a) is that for an ideal gas (and the 'ideal' is very important) equation (4.7a) defines all possible equilibrium points for the volume of gas you are looking at. With the three variables of p, V and T there are an infinite number of points of equilibrium on a two-dimensional sheet in a three-dimensional space that can be occupied. However, any equilibrium must be on this sheet. So if you double the pressure of the gas, you will either halve the volume or double the temperature, or simultaneously change both volume and temperature so that equation (4.7a) balances. Other thermodynamic systems are characterised by similar equations They are interesting for a number of reasons. 134 EFTA00625262 Firstly, despite the complexity of the underlying system, equations of state are often surprisingly simple. Secondly, the way the variables fit together can be non-obvious or even counterintuitive. Familiarity with equation (4.7a) means that people are used to it, but for the pioneers in the field, there was no obvious reason why these three variable should fit together in this way, and in fact it wasn't until many years later that the equation was independently explained at an atomic level by Maxwell and Boltzmann. Thirdly, the equations do not refer to underlying microscopic mechanisms or variables. In equation (4.7a) there are no references to elasticities of collision, the masses of the gas molecules, etc, in fact the equation should be the same for any perfect gas. Fourthly, it is common to find that many of the variables in an equation of state are intensive, that is the properties do not depend on the amount of material present. So in equation (4.7a) pressure and temperature are both intensive parameters, you can measure pressure and temperature locally at different points throughout the system as long as it is at equilibrium. Volume on the other hand is an extensive parameter that depends on the amount of stuff present. Finally, by reducing a complex system to a simple equation, equations of state are extraordinarily useful for defining and analysing systems. Going back to equation (4.6g): 1-Q P = 12(Q + H) (4.6g) this equation appears to fill all the above characteristics fully. Firstly it can be noted that both p (returns/total-returns) and (Q/(Q+H)) can be seen as macroeconomic ratios. Then equation (4.6g) becomes a formula incorporating just four intensive variables and could be expressed as: 14(1 + G) = r (4.7c) Where p is the profit ratio and G is a cash-debt gearing ratio H/Q, and none of 12, p , G or r depend on the size of an economy. This meets conditions one and four. Condition three is certainly met; there are none of the microscopic foundations beloved of economists in equation (4.6g). Condition two would appear to be the case, given that this equation has followed Bowley's original discovery by over a century. 135 EFTA00625263 The fifth condition remains to be proved. Just as an aside, an accident of history means that I am unable to present Phillip Mirowski with his fundamental constant, something similar to the R of (4.7a) or the k of (4.7b). Luckily for economists almost all variables in economics have been defined in terms of money, people or money per person. As a result the equations of state fit together automatically and the balancing constant is simply unity. Unfortunately for naming conventions, persuading people that the dimensionless number 'one' is a fundamental constant rather than a lucky accident is a little tricky. Why equation 4.6g (or (4.7c)) is important is that it says that you can't change the Bowley ratio without changing the savings ratio, the gearing ratio or long term returns. Or vice versa for any of the savings ratio, gearing ratio or long term returns. Which means that you can't change the Bowley ratio by changing things like the tax system, the education system, trade union bargaining rights, monopolistic behaviour, reducing friction in capital markets, affirmative action, inheritance laws, or a thousand and one other things that people believe will make incomes better for ordinary folks. None of the above will have any effect on the Bowley ratio unless they change one of the other factors in equation (4.6g). In extremis, as the Russians discovered and the Chinese are discovering, you can't even get more money into the pockets of the workers by introducing state ownership and a workers paradise. Ultimately, if your economy becomes technologically advanced, the factories become informally 'owned' by a nomenklatura or similar business class linked to the elite, and Bowley's law and the appropriate matching unequal GLV distribution reasserts itself. Sadly for Marx, his perceptive insights prove so powerful that they work their wonders even in 'Marxist' economies. It is for these reasons that my own proposals for solving poverty look at redistributing wealth rather than redistributing earnings. Going back to equation (4.6g), it is worth focusing again on the underlying model in section 4.6. There are very important economic factors in the model that do not appear in equation (4.6g). This includes the amount of physical capital K, or the proportion of this capital that is used. It includes the productivity of this capital. It also includes the function of the compensation of the workers, and so in a real economy, the level of employment and unemployment. All of these things have no relevance to the overall, macroeconomic balance of the model. All these things have secondary functions in the model. The overall model has an infinite number of equilibrium points that balance to equation (4.6g) even when the solutions are stationary. This is the prime equilibrium that is being sustained. The equilibrium that the system automatically and inevitably returns to. When the model moves into unstable zones, the equilibrium hunts around an equilibrium with the parameters in (4.6g) changing cyclically. There is an infinite number of points the cycles can pass through, but within a constrained zone, much like the foxes and rabbits of the original Lotka-Volterra model. Within each of these infinite solutions the values of capital, capital productivity and waged earnings all adjust to a give a solution that satisfies equation (4.6g). 136 EFTA00625264 To take a trivial example, suppose that the amount of labour needed to service the real capital K is exactly halved for all values of K. This can be modelled in model 4A, or the other models, in appendix 14.9. by changing the parameter 'labour_required' from 1 to 0.5. If you simply change the value of labour_required from 1 to 0.5 then all the various parameters in equation (4.6g) will change to new values. Most notably the value of the cash/debt balance will change. If the model is then returned to it's original overall parameters, by using solver to return the debt to its original value by adjusting K, then a new equilibrium is achieved, with a higher value of K. A comparison is shown below, column A is the first equilibrium, column B shows the result of changing the value of labour_required, finally column C shows the result of returning the cash balance to zero. Figure 4.7.1 A B C interest rate 0.10 0.10 0.10 production_rate 0.20 0.20 0.20 consumption rate (52) 0.40 0.40 0.40 labour_required 1.00 0.50 0.50 goods_payments 40.00 32.39 40.00 earnings_income 30.00 22.39 30.00 actual_returns 10.00 10.00 10.00 capital (K) 100.00 119.03 135.61 capital_wealth (Q) 100.00 100.00 100.00 cash_wealth (H) 0.00 -19.03 0.00 total_wealth (W) 100.00 80.97 100.00 total_returns 40.00 32.39 40.00 Bowley Ratio ((3) 0.75 0.69 0.75 Halved A to B Forced to zero B to C Reverts to 0.75 A to C In this case an increase in labour productivity has been balanced by decreasing employment. A new equilibrium has been achieved, and at this point there is no need for any further adjustment in the model. In the case of the change of labour_required from 1 to 0.5, the new equilibrium at zero cash balance is 136 units of capital. The requirements of labour per unit of capital has halved, but the amount of capital has increased by only a third. The actual labour required to be employed has reduced by nearly a third. The new equilibrium has rebalanced by sacking workers. The marginality of labour is not relevant to the model, the model simply moves to ensure that equation (4.6g) is balanced, it does this without any reference to the underlying labour supply curve. Model 4A, and all the other models, can create mass unemployment as a consequence of improved technology, and can then sustain that mass unemployment indefinitely. Indeed one of the main conclusions of models of section 4 and equation (4.6g) is that labour and capital, because of their different forms of ownership are not substitutable at a macroeconomic level. This is discussed at length in section 4.8 below. 137 EFTA00625265 There are many different ways that the model can be rebalanced, and many different ways that the equilibrium can be achieved. The key for the model and equation (4.6g) is that the total earnings; wages plus dividends, must balance the total consumption, which must be SI times the wealth. Which equilibrium point will be achieved will depend on other factors, but the model won't naturally rebalance to full employment of its own volition. To get a clearer understanding, I urge readers to load the model in excel from appendix 14.9 and experiment for themselves. This demonstrates that Keynes' fundamental insight was correct; that such a system could be stable even though it was not at the level of full employment, and that deliberate demand management would be needed to move it back to full employment. Unfortunately, Keynes avoided detailed mathematics in his main works, also his theories have been developed almost exclusively using the concepts of saving and investment as drivers, even when, as discussed in section 1.3 above, it has become clear that the IS paradigm is a secondary part of the economic cycle. Returning to the discussions of an equation of state it is worth noting that equation (4.6g) does not mean that other relationships can not affect the variables in equation (4.6g), just that if one factor of (4.6g) is changed, then the others must vary to compensate. Similarly it is possible that other relationships could cause one variable in 4.6g to affect another variable. It is also worth noting that the original gas model, shown in equation (4.7a) was that for an 'ideal' gas. While some gases, such as the noble gases, are close to ideal, most gases divert from the behaviour of (4.7a) under certain circumstances, most notably as temperatures drop. Water vapour, for example, obeys (4.7a) fairly closely at atmospheric pressure above 100C. However if water vapour is cooled to 100C at atmospheric pressure, the volume of the gas drops dramatically as the gas condenses into a liquid. To cope with such problems, instead of using equations of state, scientists and engineers use phase diagrams that show the relations between the state variables (p, V, T, etc) as the substance under observation changes between different states. Sometimes changes in state can be large and instantaneous. For example, superheated liquid can suddenly boil off explosively, or supercooled water can freeze instantaneously. Both these changes can be precipitated by for example a minor contaminant, or small movement. Casual observation suggests that similar phase changes may be encountered with national economies. Looking at the bubble behaviour in Japan in 1989 or the US in 1929 or 2008, in all three cases it looks like a superheated, apparently stable, system suddenly made a dramatic shift to another, very distant equilibrium point accompanied by dramatic changes in debt level, consumption level and the ratios of nominal capital (Q) to real capital (K). The example of Argentina between 2000 and 2005 suggests that income distributions can also change dramatically in the short term during major economic shocks [Ferrero 2010]. Such system changes also typically involve hysteresis so it is not possible to simply reverse conditions and return to the start point. Such phase change behaviour can be modelled within non-linear dynamics and chaotic systems, see Strogatz for example [Strogatz 2000]. It remains the case that claims that equations (4.5h) and (4.6g) are equations of state, rather than simple accounting conventions, could merely be an act of pretension. It is of course possible that the modelling, and so the equation is simply wrong. However the models and equations remain the only ever effective attempt to model theoretically the stylised facts that 138 EFTA00625266 Bowley observed a century ago, and the values produced are uncannily close to the observed data. If this approach is in fact wrong it does suggest that a similar approach may be one that finally clarifies this mystery of economics. 4.8 Nirvana Postponed In the previous section it was explained how a Bowley type model could produce an equilibrium that resulted in persistent long-term unemployment. This in itself gives severe poverty problems for the least able in society, as well as a significant tax burden for those in employment, who have to provide the welfare. A second problem for a Bowley type model is that, with interest rates, consumption rates and debt ratio generally stable over the long term; equation (4.6g) (shown again below), gives a fixed value for the Bowley ratio, and so, as we saw in section 1.5 a fixed value for alpha in the GLV distribution. The fixed value of alpha then gives a fixed ratio of inequality and means that a significant minority of the population receives substantially below the average income. Taken together these two elements mean that the bottom third or so of society in a modern economy can get a very raw deal; moving between long-term unemployment and intermittent low wage employment. There are however deeper and much more important reasons why all individuals, including the rich, suffer from poor life quality in a Bowley type economy. Going back to equation (4.6G): rQ P = 12(Q + H) (4.6g) Again given that the profit rate, consumption rate, and debt gearing are all fairly constant in a mature economy, then the Bowley ratio tends to be close to constant, and the stylised facts show that the returns to labour are typically two-thirds to three-quarters, while the returns to capital are one third to a quarter. To all intents and purposes, at the level of the economy as a whole, this means that the ratio of returns to capital and labour is pretty much close to invariant. At a macro level at least, the basic neo-classical, Walrasian assumption of substitutability of labour and capital is simply wrong. In this respect, the Austrian school is fundamentally correct, there is a 'natural balance' between capital and labour. And, in the absence of severe epidemics or genocide, the quantity of labour cannot easily be changed. 139 EFTA00625267 While it is possible to build up capital in the short term this is not sustainable, and a boom in capital above the long-term trend is followed by a bust, with at best stagnation in capital growth. If too much capital has built up, then there is the danger of capital destruction. Interestingly, in the models in section 4, the amount of financial capital Q can increase dramatically for small increases in actual capital K, especially when debt is allowed to increase. In these circumstances, the Austrian remedies for bubbles seem very sensible. As well as reducing debt back to sensible levels, the nominal value of capital, Q, needs to be reduced quickly via bankruptcies, wiping out the value of share and bond holders, etc. If this is done quickly then the economy can rebalance financial flows easily so that employment can be maintained and the fullest use of the real capital can be achieved. This was the approach used successfully in the 1990's by Sweden and other Nordic countries. In recent crises in Japan and the US, fear of hurting owners of financial assets; ultimately mostly politically important holders of pension funds, has resulted in deliberate government policies of attempting to maintain the value of financial assets in 'zombie' institutions, or to bail out asset holders altogether by nationalising debts. While this may seem sensible in the short term, the effect of delaying a return to the natural equilibrium of equation (4.6g) above may result in unexpected consequences of deflation or inflation, and the long-term destruction of real (as against financial) capital. Clearly a much better plan is simply to prevent excess debt, and so inappropriate capital building up in the first place. One thing that should be clear from a fixed ratio of returns to capital and labour, is that attempting to 'rebalance' the economy by cutting wages and 'pricing workers back into jobs' is a course of great foolishness, and would guarantee a spiral of reducing returns to both labour and capital, so reducing employment and utilisation of capital. This was one of Keynes's central insights. In one sense this 1/3"1 — 2/3rd split of returns to capital and labour can be seen as a good thing. It is caused by the shortage of surplus labour past a Lewisian turning point, and prevents Marx's prediction of ever increasing returns to capitalists and ever further impoverishment of workers. However, in a deeper sense this is also a very negative thing. As has been discussed above in section 4.7, when the productivity of machines increases, one way the system can reach equilibrium is simply by using less human input. As capital becomes more productive, to get the same returns you just use less of it. What equation (4.6g) means, in fact what any formulation of Bowley's law means, is that because the balance of returns to labour and capital is fixed, to get any progress, to get any growth in gdp; to get more wealth, you must get more returns to labour. Historically this generally been achieved by increasing the output from labour. If the returns ratio of labour to capital is fixed at 2:1, then it is the amount and efficiency of labour that has to be improved to get gdp growth. Progress is constrained by the amount and productivity of labour, not capital. Increasing the amount and efficiency of capital is relatively easy. But doing this alone has no useful effect. 140 EFTA00625268 Although Western economies are now highly mechanised, the workings of the financial system dictate that two-thirds of the earnings that are produced by capitalism are paid directly to people in the form of wages. Also, as discussed in section 1 of the paper, for 80% of people, payment for labour forms almost all their income. This necessarily demands the full time presence of people at work. We have been enslaved by the machines. In the second half of the 20th century, for most Western countries, increasing the amount of production provided by labour was very easy. It was achieved very simply by moving women out of the home and into the workforce. This one change in itself was probably the most important source of economic growth through the fifties to the seventies. Once this step has been completed, increasing the size of human capital becomes much more problematic. So the next stage is to increase the efficiency of human capital, however this is also problematic. Human capital is primarily restricted to the skills and abilities that human beings have, and carry around with them in their brains. There are a few obvious skills such as driving, using basic word processing software, or other basic computer skills that can be easily learnt by almost all people. But beyond that things get difficult. Information Technology is a good example. Computers are generally owned by companies, so returns on their wealth generated are taken by the companies. As we have seen above, if this improves returns to companies, it just results in less capital being needed overall. By replacing many basic clerking and administrative duties computers have actually taken skills that used to be in the hands of human beings and moved them to the owners of capital. Some people of course have made a great deal of money out of their personal capital in the IT revolution. Computer programmers and mathematical modellers are two examples. But to get the returns to the humans, the human capital needed is knowledge of VBA, C++, Excel, etc as well as advanced mathematics. This is human capital that is only available to a minority of people with the requisite logical and mathematical abilities. Another way to benefit from IT is to be a good and effective manager. However most would agree that this is also a minority skill. This may explain some of the apparent problems of the modern world. Firstly it might account for the non-visibility of IT in productivity despite the amount spent on it. It might also account for the imbalance in work requirements between different skill groups. Unskilled labour is now of marginal assistance to serving machines, and has been largely replaced by the machines themselves. This is as true for clerking and administrative work as it is for labour. Spreadsheets and stock control systems have replaced the clerks. Forklift trucks and containers have replaced the labourers. In contrast skilled professionals, from plumbers and 141 EFTA00625269 technicians to programmers and managers, people who have the abilities to serve the machines, find themselves under continuous pressure to increase their working hours. Taken all together this might account for the fairly acrid taste that is seen in political debate in most Western societies. On one side there is large population of the unskilled who find it difficult to find and hold decent work of any sort. These people face unemployment, poor wages, no opportunities for advancement and semi-permanent dependence on welfare. They often have stretches of involuntary inactivity. Despite their subsidies and enforced leisure, for these people hard work is not rewarded and life lacks hope of betterment. On the other side there are skilled trades people, professionals and managers, who work longer hours and pay higher taxes than their parents, primarily, as they see it, to support the idle poor. This is not a happy recipe. Futurologists have been predicting for decades that once basic needs have been satisfied, human beings would be able to relax into a life of leisure. To date, futurologists have been wrong. And it is not for the want of suitable capital, the progress of automated technology continues at an extraordinary rate. In section 9.3 examples such as fruit picking machines, automated hospitals and personal rapid transport systems are discussed. All of these examples share the common features of being able to replace large amounts of unskilled labour and also being technologies that are being brought into use. Despite this, in real life, almost the opposite is happening, working weeks have been steady, and in some cases increasing. In Europe and the US retirement ages are being revised upwards rather than downwards. In the west we have achieved enormous personal wealth, but through an accident of mathematics, we have been required to sacrifice our time to the mechanism of wealth production. Nirvana has been postponed. As an amateur futurologist, it is possible to conceive of a world where the main inputs of human labour could be reduced to direct care for the young, the sick, the elderly and the provision of entertainment and spiritual needs. Which is what, biologically, human beings are designed to do. Other animals that dance, sing and make art works; such as birds of paradise for example, are generally animals that do not face significant predation and that have more than enough resources available, and so time on their hands. In the absence of predators to compete with, or resources to fight over, they turn to competition in the arts. Almost certainly prior to the agricultural revolution, human beings fell into this class of animal. Human beings were simply not designed to work forty hours a day five days a week. Both hunter-gatherers and most agricultural societies are characterised by underemployment. Historically this was true in the West until recently. 142 EFTA00625270 The second half of the twentieth century is almost unique in being one in which the well off are characterised by having full time employment. In the past the rich were notable by not working, they lived off their capital and looked down on paid work. This labour capital split of the Bowley ratio might also explain the bizarre behaviour of growth. As has been discussed in section 4.5 above, when they start growing, economies typically follow a path of rapid expansion to use up surplus subsistence labour. Casual observation suggests that this can be associated with growth rates of up to 10%. The 10% restraint appears to be due to the difficulties of building infrastructure fast enough. China has been following this path for the last two decades, the Asian tigers did so before this; now India appears to be following the same route. Once the surplus labour has been used up then growth generally drops to a slow continuous growth rate of about 2-4%. The UK has been expanding like this for over 200 years, the US for over 150 years see figure 4.5.3 above, or gapminder for some very pretty graphics [gapminder]. In theory this is very odd, once economies are mature, why do they just not continue increasing the capital stock at 10% per annum to provide for all people's needs and eliminate the need for labour? This should be easy, as countries, and people, get richer, more of their basic needs should be provided for, so diverting revenue (in the most general sense — not just public taxation) for provision of capital should become easier to do. If however, growth is restrained by the productivity of labour, then a growth rate of 2-4% seems more sensible. Once reserves of subsistence labour have been exhausted, human capital cannot quickly be increased in the same way that physical capital can be. I suspect that this might be only part of the explanation. As discussed previously, I find the growth rate of 2-4% suspiciously regular. It also goes hand in hand with suspiciously constant real interest rates at 3% or so, and suspiciously regular stock market returns, at 7%, see figures 4.5.1 to 4.5.3 in section 4.5. The 'stylised facts' of these three growth rates are very suggestive of a deeper underlying process equilibrium. The presence of a fixed ratio of returns to capital and labour also gives a very big problem that there is a general shortage of 'real' assets. As we have seen in section 1.8 above, there simply aren't enough real assets available to provide even for everybody's retirement needs. This in itself could be a source of the search in the finance industry to create new and exotic assets that appear to solve this problem. Unfortunately, Bowley's law dictates that the underlying 'real' economy is fixed, so the total real returns are fixed. Trying to create new assets out of old is no more possible than other more traditional forms of alchemy. You can't create real new revenue streams simply by repackaging assets. Similarly, this may explain the hunger for government bonds in the financial markets, especially given their apparent safety. But ultimately, government bonds are dependent, via taxation, on revenue earned in the private sector. The most obvious example in the shortfall of capital is the example of housing. Other public goods such as health, education and pensions have obvious market failure reasons for not being provided fully. 143 EFTA00625271 Housing should be simple to provide for in a wealthy society. Simply build enough of it for everyone, then all you need to do is maintain it. In practice many societies have attempted to do this, through mass council (public) housing in the UK to the recent disaster of state subsidised mortgages in the USA. The problem of course has always been that the poor have rarely been able to afford the maintenance of the housing, never mind the capital payments. So, the key question here is whether this system can be changed so that more capital can be accumulated to carry out more work on behalf of labour. Interestingly, history suggests that the system can be changed significantly, and especially as a result of the scarcity of labour. The trick is not to change the efficiency of labour but to fully remove the surplus labour and turn it into an increasingly scarce resource that is over compensated for its efforts. Back in section 1.3 I made the assumption that labour was 'fairly' paid for its inputs to the production process. I kept this assumption through all the income models, though it was then discretely abandoned in the macroeconomic modelling. Actually, because labour is a uniquely non-adjustable factor input, it is the only truly scarce, non- substitutable resource. Also, because of Bowley's law, labour is very rarely paid it's true worth. It is usually significantly under or overpaid. Following the theories of WA Lewis [Lewis 1954], or for that matter Marx, in a society with excess subsistence labour, capital can 'under-pay' labour employed in the commercial sector, as pay rates are held down at subsistence level by the presence of under-utilised rural labour. This has been the normal state for most countries for most of history, and has provided the main critique of capitalism until at least the end of the Second World War. In such an economy, with surplus labour, the economy doesn't reach a true equilibrium for the Lotka-Volterra / GLV approaches described above. The subsistence farmers are outside the equilibrium, and they also hold down the wages of those employed. In such a society the rich are overcompensated for their ownership of capital, and also have low living costs due to the low labour costs. In these societies the Bowley ratio can be as low as 0.5, this can be seen in China today, even as it approaches its Lewisian turning point. Things are much more interesting in a 'normal' industrialised country; one that has passed it Lewisian turning point and has absorbed the majority of its cheap labour. In such an economy labour is generally over-rewarded; returns to labour are in excess of the value actually provided by labour. This was actually the case in the macroeconomic model in section 4 Where labour generally gained through the economic cycle, being 'overpaid' in exactly the same way that suppliers of commodities were overpaid in the commodity cycle in section 3. In this case the employees are successfully extracting 'rents' from the capitalists. And a good thing too. I believe that, in the second half of the 20th century, parts of the world moved, for a period, fully into the zone described in this model. 144 EFTA00625272 Following the Second World War; all the communist countries, most of the de-colonised countries, and most of Latin America voluntarily withdrew from the world trade system. The communists followed their own socialist paths; almost all of the rest followed a route of import substitution behind high tariff barriers. Following rapid post war growth, most of Western Europe and North America went through a period in the fifties and sixties with full employment and ongoing labour shortages. Meanwhile the few poor or poorer countries that remained in the world trading system; countries such as Japan, Italy, South Korea, Taiwan, Hong Kong, Singapore and Malaysia, saw breakneck growth, moving from subsistence agriculture to industrialisation in a generation. In the West full employment artificially increased returns to labour. Through the Bowley ratio this then forced investment in capital to increase returns to capital. Over the longer term, expensive labour forced investment in labour saving production, so increasing the efficiency of capital. This period resulted in a virtuous circle with high wages and full employment forcing rapid growth. Returns to both labour and capital kept increasing in lockstep. It is worth remembering that labour was so scarce in this period that large-scale immigration was allowed into the UK, and guest workers were invited to Germany, to do the menial work that Britons and Germans were unwilling to do. From the nineteen-seventies onwards many poorer countries, most notably China, re-entered the world economic system, providing alternate supplies of cheap labour, and competition for labour in industrialised countries. The portion of the world's economy that is integrated into the trade system moved back to a pre-Lewisian state with excess subsistence labour in Asia, Africa and South America competing with Western labour. It is the belief of the author that, at the time of writing, the richer, industrialised, countries are currently simultaneously in a complex pre- and post-Lewisian state. Pre-Lewisian for unskilled labour, and post-Lewisian for skilled labour. This is due to an accident of history caused by the third world's absence from, and then re-entry to, the global economy. These conclusions appear to have some support from data. As well as showing smaller cycles, many of the country graphs in Harvie [Harvie 2000] show a much longer term cycle of change in the compensation to labour, starting with lows in 1956 going to high points in the 1970s, then returning to lower points by 1994 (the last points in the data sets, all of which were for industrial economies). It will be interesting to see what happens in the near future. China appears to be passing through it's Lewisian turning point. Already China's low-cost manufacturing base is relocating to poorer countries such as Vietnam and Bangladesh. That is the manufacturing base that supplies cheap toys, shoes and clothes to richer countries. This in itself will spread wealth, and labour shortages, to these countries as they start exporting to the West. Simultaneously China will also need to start importing cheap manufactures from poorer countries to supply its own population. Given that India is already close to peak expansion rates, primarily through providing information services to the West, the worldwide supply of surplus cheap labour could dwindle very quickly. 145 EFTA00625273 It is possible that we are close to seeing a repeat of the full employment boom of the 50s and 60s, but this time repeated on a worldwide scale. Even without waiting for this process to happen naturally, it is possible that the proposed '40 acres' compulsory saving process proposed in section 1.8 above might also be able to produce the same effect artificially in single countries. Although some people are natural workaholics, most would choose to 'downsize' and have more leisure time if they could. But they can't. It is common in neoclassical economics to see discussions of individuals choosing between spending and leisure. Because of the workings of the GLV, most individuals have no such choice. To seriously consider reducing working hours; a family needs to own their own house, have a good pension plan in place, have enough money coming in to cover day-to-day expenses and be sure of access to a decent health service and a good education system for their children. Even in the richest of Western countries few people have all, or even most, of these things. Primarily because they have insufficient capital. If a 'forty acres' style system is used it would give more returns from capital to all members of society, it would reduce reliance on earned income. It could slowly start a virtuous circle like that seen in the 50s and 60s. By ensuring that all individuals move up to the point that they have sufficient wealth and income to meet their day to day needs, compulsory saving would allow people to move into voluntary saving and allow faster investment in decent housing and sufficient pensions. This would then allow a much more genuine choice between work and leisure. As individuals begin to withdraw from the labour market, this would then start a virtuous circle of rising labour costs and full employment. In the longer term this would then also encourage a drive to more labour saving capital. Probably it would start with middle class families choosing to keep a partner at home when children are young. But even such a small withdrawal would tighten the labour market in the skills removed and so push up wages. As people withdraw from the labour market, this will force wages up, and will also increase the share of returns to those still in the labour market. With labour tight, and wages rising this will also encourage adoption of more efficient, labour saving technology. With Bowley's ratio holding, returns to both labour and capital will go up, while more and more of the actual work done by the machines. The aim would be to create mass underemployment, or even unemployment, but not, as presently happens by accidentally creating unemployment at the bottom of society. Instead, the aim would be to create voluntary underemployment at the top of society, as people choose to live more on their investment income and less on their wages. As this then forces wages up, the process will then work its way down to poorer people. 146 EFTA00625274 The aim is to create underemployment at the top end of society, so creating full employment throughout society. So increasing wages for all, so increasing returns to labour, so, via the Bowley ratio, forcing up returns to capital. The aim would be to build up the v40 so that it would consist of shares in companies owning machines carrying out fruit-picking, hospital-cleaning and personal rapid transport. Meanwhile the people who used to be agricultural labourers, cleaners and taxi drivers would get more rewarding and better paid jobs with shorter hours. They would be helped by the income from their own v40's. A good aim would be to get the v40 sufficiently large for everybody that dividend payments pay the equivalent of two working days per week of total living costs, while people still work three days a week for their remaining income. On retirement, the additional drawdown of capital would provide for five working days per week of income. A three day working week seems a sensible aim. There will always be a need for human beings to provide education, caring and entertainment. Three days a week would be sufficient to give structure and integration in society, but would leave ample time for family, friendship and leisure. To many the above will seem ridiculously naïve, but the example of Norway given previously shows that the numbers can add up, and a three day week is feasible. As long as enough capital is available. Futurologists' predictions have gone wrong because of the workings of the Bowley Ratio. Understanding how the Bowley Ratio works may allow the future to be changed. 4.9 Bowley Squared Going back again to the base model shown in figure 1.3.5, this shows financial wealth W being held by households in the form of stocks and shares as claims on the real wealth K in the productive companies. Figure 1.3.5 here In one important way, this is very unrealistic. I personally don't own any shares. In reality very few people own shares directly. In fact, aside from housing, most people do not own any capital directly. Most peoples' wealth is in the form of bank deposits, pension funds, insurance policies, mutual funds, etc. All of these investments form financial claims on companies within the financial sector. 147 EFTA00625275 The companies in the financial sector then own the claims on the real assets of the non-financial sector. When it works correctly this is just a sensible way of dividing labour. Most people who have money to invest do not want to spend their spare time investigating possible investments. Also they would prefer to spread their investments across different companies to spread their risk. It makes lots of sense to lend their money to professional experts who can save costs by analysing investments on the behalf of lots of different investors at the same time. This then results in a model of the form shown in figure 4.9.1: Figure 4.9.1 here While it might seem very sensible to set up a specialist finance sector in this manner, from a control systems point of view this is something of a nightmare. This repeats the feed back loop of the simple macro economic model a second time. Instead of one simple feedback loop capable of creating endogenous cyclical behaviour, you now have two feedback loops both capable of creating endogenous cyclical behaviour, and more importantly, capably of interacting with each other to give even bigger more complicated endogenous cycles. The original macroeconomic model can be considered to be a very simple unstable model on the lines of the Soay sheep model discussed briefly in section 1.2.1 In this model the companies grow too rapidly for the base level of labour that can support them, in the same way that Soay sheep breed too quickly for the grass to support them. Introducing a financial sector, installs a second population on top of the first. It is similar to adding wolves to predate on the sheep of the first model. I have not attempted to construct this model mathematically. The models discussed in section 4 above already have sufficient loose parameters and dynamic complexity to produce confusing patterns of behaviour. They really need pinning down with real data before being expanded to the model in figure 4.9.1. But even without modelling, some of the behaviour is easy to predict. In fact we have returned back to something very similar to the original fox and rabbits Lotka-Volterra model discussed back in section 1.2. In this case, the rabbits are the non-financial sector and the foxes are the financial sector. Typically a boom would start with a small financial sector and a growing productive sector. As the productive sector grows the financial sector grows more and more rapidly taking up an increasing proportion of the economy. Then the productive sector will start to decline slowly. A short, but significant time after that, the financial sector will show a sudden and much more rapid decline. The operation of the two business sectors is analogous to the fluctuations of biomass in a Lotka- Volterra model. First biomass builds up in the rabbits then in the foxes, then it declines in the 148 EFTA00625276 rabbits and then the foxes. Similarly capital should build up in the productive and then financial sectors, followed by declines, in turn for each sector. So a prediction of this model is that over the next five to ten years, the proportional size of the financial sector in countries such as the USA and UK should decline back significantly towards proportional sizes seen in say the 1980s or early 90s. One other outcome of this model is that the two sectors can follow their own paths to a significant extent. In such a model, the secondary feedback loop, that of the finance system can vary much more dramatically than the underlying population, see figure 1.2.1.1, showing the original Hudson Bay lynx and hare populations. This makes control of such a dual speed economy very difficult when you are only using the single weapon of inflation targeting and interest rates. While the underlying economy may respond reasonably to interest rates, the liquidity generated in this productive economy can generate much larger changes in liquidity in the finance sector, which are harder to control. Also the fluctuations in the financial sector will not be in the same time phase as the main economy. To take an analogy this model can be likened to an air-conditioning system. The main economy can be imagined as a large office block somewhere in the temperate northern hemisphere. Depending on the time of year or time of day this main block will need a certain amount of heating or cooling. The financial sector can be seen as similar to a large atrium on the south aspect of the building, full of hothouse flowers. The two buildings will be connected together, and will be roughly aligned through the seasons and days, but will vary greatly in the amount of cooling and heating needed. The atrium will need more heating in winter and more cooling in summer. This will depend on the amount and direction of sun and the external air temperature. On some spring and autumn days, the atrium might need cooling when the building needs heating or vice versa. The Bowley squared model is a complex system and needs full understanding to control effectively. The topic of financial sector liquidity and how to control it is revisited in some depth in section 8.2.1 below. Despite the complexity of the model in figure 4.9.1, it remains the case that control of such a system should be straightforward using standard controls systems feedback theory. 4.10 Siamese Bowley - Mutual Suicide Pacts In the previous section one Bowley model was placed on top of another, in a way that was multiplicative. 149 EFTA00625277 An alternative model would be to put two Bowley models side by side and allow individuals in one half of the model to own capital in the other half of the model. This is illustrated in figure 4.10.1 below. Figure 4.10.1 here This gives an international model, with international trade. The discussion that follows borrows heavily from the work of Michael Pettis [Pettis 2001], whose writing I have found highly illuminating, in contrast to much standard economic work on international economics and finance. Pettis's work takes a financial framework for analysis, and concentrates heavily on flows and stocks of capital and debt. As such it fits well with the analytical models described in this paper. Pettis's work also fits closely with the known facts of repeated booms and busts triggered in poorer nations by investment booms and financial crises initiated by capital investment typically from London or New York; a process documented beautifully by Reinhart and Rogoff in 'This Time is Different' [Reinhart & Rogoff 2009]. One aside with regard to the use of the word capital, which in international economics is used in a markedly different way to that in normal macroeconomics, or the preceding sections of this paper. In this paper capital can refer to K, the stock of physical assets that produce real wealth in the form of goods and services. It can also mean W (or Q), the stocks of paper financial assets that are held as claims on those productive physical assets, such as stocks, shares and company bonds. In international finance a 'capital flow' is used to refer to a flow of money in return for a stream of paper financial assets; sometimes financial assets of companies, but these can also be assets such as government bonds. So a capital inflow from Britain to Brazil would indicate purchase of Brazilian financial assets by institutions in Britain. The ownership of these financial assets would then give the right of the British owners to receive a stream of financial income based on the wealth produced by the underlying real physical capital. In theory such a capital inflow should be used to invest in physical capital goods in the recipient country so allowing the country to become more productive and pay the interest on the loans. Unfortunately it is all too common for the 'capital flow' to be used as payments for imports into the country receiving the 'capital flow', eg, Brazil paying for imports from the UK. When this is the case, the original meaning of the word 'capital' is lost altogether, and the 'capital inflow' is simply a way of describing lending money as a form of debt, often effectively unsecured. And as can be seen from the analysis of Pettis or the research of Reinhardt and Rogoff, it is this quick and natural split of countries into creditors and debtors that is symptomatic of financial trade. 150 EFTA00625278 International finance can be very confusing, with a large number of variables, especially when currency flows and exchange rates are taken into account. Much analysis of international finance concentrates on the role of currency, along with control of interest rates and the role of inflation. Actually, history suggests that different currencies are in fact something of a red herring. To get the basic model for analysis you don't need currencies. Throughout history there are many examples of international trade, and gross trade imbalances occurring when countries shared a common currency. Pettis gives the first such well documented example as that of different parts of the Roman empire in a speculative property boom in 33AD. In this case the metropolis of Rome was the debtor, while the grain producing provinces were the creditors. History is replete with currency unions or fixed exchange rate pegs coming to grief through trade imbalances. Many of the imbalances of the depression, when the US was a creditor and most of the rest of the world were debtors, were exacerbated by the fixed exchange rates of the gold standard. Most of the countries involved in the Asian financial crises of the 1997 were on fixed pegs to the US dollar. Mexico was forced off its fixed exchange rate during the tequila crisis of 1994 And Argentina suffered severe economic problems until it abandoned it's currency board in 2002. At the time of writing Greece, Ireland, Portugal and Spain are suffering major structural problems while Germany and it's near neighbours simultaneously enjoy good growth. The common currency of the euro is currently magnifying trade problems, not reducing them. Another factor that can be ignored in a base model is relative wealth. Although it is most common for the rich nation to be the creditor nation and the poorer nation to be the debtor nation it is sometimes the other way round. Ancient Rome provides one example, where the rich metropolis was in hock to the poor provinces. A much better example is the current one of the rich USA being a very substantial debtor balanced by a much poorer China as a very substantial creditor. In fact, when looking at trade imbalances, it is my belief that it is debt, or more particularly, savings rates, that are key. In Europe rich Germany has a high savings rate while Ireland and the Mediterranean countries have lower savings rates and higher debt. On a bigger scale poorer China has one of the highest savings rates ever seen, and America has moved, in less than a century, from the world's creditor to the world's debtor. It is unfortunate that this is often seen in moralistic terms, especially by creditor nations. In fact, though cultural reasons are clearly important, savings rates are often driven by deeper fundamentals. As Lewis [Lewis 1954] pointed out lucidly, newly industrialising countries tend to have high savings rates as the newly rich elite have access to cheap land and cheap labour, and have little else to do with their money but save it. 151 EFTA00625279 The US complains bitterly about China's 'currency manipulation' causing an imbalance of trade, but the US made the same complaints about France and Germany in the 50's and 60's, about Japan in the 70's and 80's, and about the Asian tigers in the 90's. The common denominator here is the US; the exceptionalism of the US in this case is their ability to issue the world's reserve currency. As issuers of the reserve currency, the US is able to borrow at cheaper rates than other countries, so it is hardly surprising that they have become the world's biggest debtor. An identical process happened in the UK in the 19th century. In fact there appears to be a cycle in reserve countries over the last half a millennium. Reserve currency status has been held in turn by Portugal, Spain, Holland, France, the UK and now the US, with each country holding the status for a roughly a century. In each case it appears that a country starts with a solid productive base that put it at the heart of trade. This trade and creditor role then allowed its currency to become dominant in trade. Reserve currency status then allowed cheap borrowing and increased debt. The increasing debt, allied with 'imperial over-reach' defending trade routes, then caused a crisis and loss of reserve status to the next upstart. So going back to figure 4.10.1 below; Figure 4.10.1 here We have two countries, Chermany with a high savings rate, and Medimerica with a lower savings rate. The two countries could start with the same population and the same amounts of capital K and wealth W per head, but the situation is naturally unstable. Chermany, with its higher saving rate will consume less than Medimerica and will accumulate more capital. After the first iteration, Medimerica will have a little less capital, but will still have a thirst to consume rather than save. In the short term the flows can be balanced by an unholy trade off. Chermany can supply funds; 'capital outflow' to Medimerica in return for financial assets belonging to Medimerica. Medimerica can then use this cash to buy imports from Chermany, mopping up the extra production that Chermany's high savers don't need. Unfortunately, although this balances the flows in the short term, it results in a grave problem with stocks. Chermany keeps on building up capital that it doesn't need. Meanwhile Medimerica increases it's financial debt to Chermany while simultaneously running down it's badly needed capital to pay for imports from Chermany. This system is inherently unstable and can only end in tears. Eventually there will come a point where Medimerica simply can not pay the interest on it's debt. It no longer has sufficient real capital to generate the real income to do so. At this point Medimerica has to default one way or another. This can be by straight repudiation of debt, or by devaluation and inflation to reduce the value of the debts. For Chermany this then gives two problems. Firstly the loss of value of the foreign assets owned. Secondly, and more importantly, the loss of markets for the exported goods produced by the excess capital that has been built up. 152 EFTA00625280 This was most dramatically demonstrated in the run up to the depression of the 30's when almost the whole world used the gold standard. During the 20's as the world's creditor, the USA (and also France) slowly built up their proportion of the world's gold reserves until Germany, the UK and other nations ran low on gold and were forced off the gold standard. They were also forced to partially default on their debts to the USA. The US was left with a large productive capacity and no buyers for its goods and also sank into depression. The US cried foul, but with a large portion of the world's gold in the US it was not clear what the Europeans were supposed to use to buy American goods. This bilateral instability goes back to the two forms of economic suicide introduced previously. One form of economic suicide is to run up too much debt as discussed in section 4.6 which eventually becomes unsustainable. Running up debt can be very appealing, as it allows consumption to run ahead of real growth, and also inflates the values of financial assets. Until the party ends and the hangover kicks in, this feels good for public and politicians alike. The second form of economic suicide is to allow capital to build up too quickly as discussed in section 4.4 above. Again in the short term this feels good because the rapidly expanding capital base increases employment and wages. (It can also have the unfortunate side effect of increasing pride in supposed national industriousness and thrift.) While it is possible to carry out each form of suicide independently, this is not so easy. In a single isolated economy the results of too much debt or too much manufacturing capacity are difficult to ignore. It is difficult to keep increasing debt in a home market beyond a certain point, and it is also difficult to build up capital and carry out a mercantilist export policy without people to export to. It is much easier to carry this out as a form of mutual suicide pact where one country takes on the role of debtor and the other of creditor, as described in the model above. The debtor country is able to borrow more and more at easy rates, the creditor country is able to sell more and more of its exports. Unfortunately neither of these processes can go on forever. In the thirties it was the debtor countries that first collapsed one by one. In the Plaza accord of 1985 the debtor countries laid down the law with over-exporting Germany and Japan. Germany took heed and rebalanced its economy (at least until the launch of the euro). Japan continued to push export led growth and imploded in 1989; to date it has not recovered. From 2006 onwards the American economy started to sputter, stalled by too much debt. In 2008 the American economy imploded in the credit crunch taking other debtor countries such as the UK and Spain with it. At the time of writing, the creditors, primarily China and Germany, have rebounded, but with a world full of excess industrial capacity it isn't clear who they are going to keep exporting too. In Europe the need for rebalancing is obvious, Ireland and the Mediterranean members of the EU are moving into outright depression and are likely to default. In the world as a whole it remains to be seen whether China can rebalance in time to prevent a Japan style bust. The big problem for China is that easing back on its export machine will result in mass unemployment and serious political unrest. A possible solution is to move capital into the hands of the workers, as discussed in section 1.8 above, so that workers would have more to spend, 153 EFTA00625281 and would not be reliant on wages alone. All in all it would make sense for the Chinese and Germans to consume more of the goods that they make. As with the Bowley squared model in the previous section I have not attempted to create a mathematical model on the lines of figure 4.10.1 above. Again there are a lot of different variables and the base models need first to be benchmarked against real data. Conceptually however, the models should be straightforward to build. Again, this sort of system is common in control systems engineering, and should be familiar to most office dwellers. To take the example of air conditioning systems again, an analogous system is where two large air-conditioning units are installed on an open office floor, each with its own independent control loop, set to control at exactly the same temperature. Common sense suggests that two identical systems like this should move up and down together in tandem. However in this case, common sense is wrong. Unfortunately, although the two units may be wired separately, the flows of air from one part of the building to another mean that the two units are actually influencing each other in what is called a 'coupled system'. Such a system can very easily become unbalanced, for example if their settings are slightly different or if part of the office is in shade and the other is receiving sunlight. In the second example the a/c unit in the shady part will provide a little cooling, while the a/c unit in the sunny part will provide a lot of cooling. Unfortunately, the cold air can then flow from the sunny part of the office to the shady part, while the warmer air from the shady part can flow to the sunny part. In fact convection will make this inevitable. When this happens the a/c unit in the shady part reduces its cooling, while the a/c unit in the sunny part ramps up its supply of cold air, and the two units end up in an ever-increasing battle to control the temperature. Ultimately, the a/c unit in the shady part may even convert to heating mode. This results in stratified air, bad draughts, general discomfort and very expensive utility bills. In this case the two a/c units are coupled but end up working in anti-phase; working in opposite directions. This is a common outcome in this type of control system. The same can happen with national economies, though it doesn't have to be the case. For example, where a large country has good economic links with a smaller country, the smaller tends to move into phase with the larger. This is true for example with Canada and the United States. Although Canada can be influenced by external events such as commodity prices, its economy usually moves closely with that of the US. The same is true of the many smaller countries around Germany, not only does this include euro users such as Austria, Finland, the Netherlands and Belgium, it also includes others such as the Czech Republic, Denmark, Sweden and the Baltic states. Together, these countries form a linked bloc, with all countries moving closely in phase with Germany. 154 EFTA00625282 In contrast, due to their size and different economic fundamentals, Italy, Spain, Portugal, Ireland and Greece have moved into anti-phase with the problems discussed in the model above. France remains uneasily stuck between the two conditions. The model described in this section is analogous to a competitive Lotka-Volterra model (in contrast to the predator-prey Lotka-Volterra models we have discussed previously). A competitive L-V model consists of, for example, sheep and rabbits living side by side eating grass on the same island. Depending on the different growth rates and breeding rates animals in these situations can come to different equilibria. If the animals are similar, say sheep and horses, an equilibrium can be reached with fixed proportions of the two groups of animals. If the animals are different the equilibrium is unstable and moves to one extreme or the other. So with say sheep and rabbits, depending on the start point, one or other group will dominate and drive the other group to extinction. One group of animals will take over all the biomass, just as in international trade it is possible for one company to take over all the real capital. Clearly the above model could be adapted in many ways, most obviously by introducing different currencies. Empirical data from the history of failed monetary unions and fixed currencies suggests that independent currencies have a significant effect, largely beneficial. If managed correctly devaluation generally allows beneficial adjustment. Obviously to introduce currency in international trade models, it first needs to be introduced in domestic economies, this is discussed in brief in section 4.11 below. 4.11 Where Angels Fear to Tread - Governments & Money I move into a discussion of the theory of money, and the role of governments, with some trepidation. Of all the areas of economics, this seems to be the one in which a religious belief in theory unfounded on empirical fact seems to be most widespread. And discussions in this sphere seem to take on the character of arguments between religious zealots. Exceptionally, Perry Mehrling writes on this field with great clarity and insight [Mehrling 2000]. It is my belief that an understanding based on flows and stocks, as followed in the rest of this paper could be productive. It would be possible first to start by looking at commodity money as an actual commodity in line with section 3 above. Using a commodity, such as gold, in the real world is problematic, because, as Robert Triffin noted, the supply of gold is insufficient to allow expansion of the money supply to keep pace with the size of the economy. To get around this problem all modern economies have moved to systems of fiat money, generally with inflation targeting or some other control system. 155 EFTA00625283 While I have many grave reservations regarding 'Modern Money' theory (see for example [Wray 1998]) I find their central insight of treating money as an artificially created commodity flow as appealing. Diagram 4.11.1 below shows a typical treatment. Figure 4.11.1 here The big problems with modern money theorists is their almost religious belief that governments can expand public debt without limit when the economy is below full output capacity. A brief review of [Bernholz 2003], [Reinhart & Rogoff 2009] or [Pettis 2001], shows that the empirical data demonstrates that this is emphatically not true. As Perry Mehrling [Mehrling 2000] points out very lucidly, the problem with the approach of Wray and others is that the state's ability to pay coupons on government bonds ultimately depends on the states ability to raise taxes, and also on the good use that the state puts those taxes to. In the simplistic examples of Modern Money, a colonial governor in a undeveloped rural economy raises hut taxes to pay for new roads and schools, and this clearly results in substantial economic improvements. That this can be translated into a modern western economy is not obvious. In fact, in industrialised countries, much money raised, whether by taxation or borrowing from private markets, is not invested in infrastructure but instead passed straight through to consumption as transfers. In this light the relationship between government and the private economy would appear to resemble the relationship between a debtor nation and a creditor nation in the Siamese-Bowley models above. The modern money theorists are surely correct in their belief that a significant amount of government debt is good for the economy as it provides a secure asset that gives needed liquidity for effective private markets. To believe that this debt can be expanded indefinitely is to undermine the most important value this debt has; that of security. In similar vein I find much of Milton Friedman's monetary theory terrifyingly naive. However I have found the blogging of 'kitchen-sink' monetarists such as Simon Ward [Ward] and John Hussman [Hussman] enormously insightful and surprisingly able in their predictive power. Friedman's theories, though simplistic were also of course based on flows, and assumed delays in action. So although his formulation was not dynamic, his underlying model, and the data it was based on was. I am insufficiently skilled to be able to judge whether either or both of the modern money and the monetarist approaches can be synthesised effectively into the modelling framework described into this paper. But I believe it may be an approach worth pursuing. Another problem with monetary theory is that 'money' can be artificially created by at least two dynamic feedback mechanisms. The first is the loop of fractional reserve banking that can allow a large multiplier of debt to be created for each sum of reserves pushed into the economy by the reserve bank. A second multiplier is the endogenous creation of liquidity within the finance system this was seen in the models in section 4, and is discussed at length in section 8.2.1 of this paper. 156 EFTA00625284 Taking all the above together, this then ends up with a basic model of the financial system that works something like the diagram below. Figure 4.11.2 here This simple model, includes at least two amplification loops and two feedback loops with positive feedback. If housing were included in the diagram, with the leverage of mortgages, there would be more feedback and amplification. With my control engineer's hat on, the only thing I can say about this as a control system is that if I was trying to design an effective control system, it definitely wouldn't look like the diagram above. It is about as sensible as trying to control a steam engine with a system made out of cheap rusty shower mixer valves and some lengths of garden hose. In democratic countries, central bankers are expected to control the whole of the country effectively by controlling the variables on the left hand side. Whatever they are paid, it is not enough. 4.12 Why Money Trickles Up Before finishing this section on modelling, and moving on to a discussion of background theory, I would first like to revisit the premise of this paper. At this point I am forced to confess to having committed a major offence that I have accused others of. I used the phrase 'Why Money Trickles Up' as the title for this paper to give an emotional impact; the title should really have read 'Why Wealth Trickles Up' or perhaps 'Why Income Trickles Up'. I have only discussed monetary theory as a passing aside. I believe however that I have given an authoritative explanation of both how and why wealth trickles up from the poor to the rich, as well as a detailed description of the mechanisms. In brief, macroeconomic factors including interest rates, saving/consumption rates and debt define the Bowley ratio; the proportions of wealth returned as wages and profits. The Bowley ratio then defines the parameters of the General Lotka-Volterra distribution that defines the distribution of wealth between individuals. This distribution of wealth then defines the majority of the shape of the distribution of income. 157 EFTA00625285 That is why money trickles up. 158 EFTA00625286 Part B - Some Theory 5. Theory Introduction Section A introduced a range of possible models to look at some of the basic interactions of economics. Though they may have had inspiration from other sources, the models are my own work. In many ways the models are naive and simplistic. Time will tell whether they prove useful or not. If the models survive unchanged I will be pleased, but also surprised. If the models are trashed and replaced I will be disappointed, but not particularly surprised. The accuracy of the models is beside the point. The point of the models is that by using a set of tools selected from other areas of science in combination with ideas primarily from classical economics and finance, it is possible to create simple effective models that address basic, fundamental regularities in economics. This is the main point of the models. If the approaches of the models above are taken further, but the models themselves are superseded, then I will have achieved the main aim of this paper. The scientific tools come primarily from physics, biology and pure mathematics. For almost all economists these tools; ideas such as chaotic mathematics, statistical physics, and entropy will be unfamiliar to the point of being quite alien. Even for most physicists, ideas such as the GLV and maximum entropy production will be unfamiliar, and I believe these will be of interest to many working in the field of complex systems whether this includes economics or not. As to the economics, of course almost all scientists will be ignorant of the basics of economics. Sadly, with vary rare exceptions, even most physicists, mathematicians and modellers researching in economics seem to take a perverse delight in not knowing anything at all about basic economists. This attitude seems to be something along the lines of "we know all about steel plate, diesel engines, turbo-chargers, power steering, inertial guidance systems, etc — why on earth should we spend our time learning about sailing boats?" However; although sailors take a lot of time and effort tacking backwards and forwards without getting anywhere particularly fast, some of their knowledge is quite useful; for example, where the shoals and reefs are, how to use a compass and sextant, why you should carry a fog-horn, not to mention lifeboats and life-jackets. And why it is a good idea to know how to swim. In fact many of the economic ideas in this paper will be unfamiliar to many economists. The economic ideas come largely from finance, economic historians and classical and other heterodox economics; including, somewhat to my own surprise, Marxian economics. All of these ideas are outside the canon of mainstream neo-classical economics and so are not just ignored but are politely rubbished, in the case of economic history and finance, or very impolitely rubbished in the case heterodox economics. None of these ideas are included in undergraduate economics courses other than at the most maverick of universities. As this paper is largely based on non-standard economics, I have gone to some efforts, not just to explain this background, but also to justify it to sceptical economists steeped in marginality and utility theory. 159 EFTA00625287 This is firstly to explain unfamiliar ideas to both economists and non-economic scientists. Also, for the economists, it is to explain how many other things, such as liquidity and dynamic scarcity, explain large apparent diversions from the idea of intrinsic value which is inherent in classical economics but absent in neo-classical economics. Once these diversions are understood and correctly modelled, classical economics becomes a much more powerful theoretical method than neoclassicism. The economic historians such as Reinhart & Rogoff, Shiller, Smithers, Harrison, Napier and Bernholz have the advantage of the long sweep of history to prevent them from accepting high- faluting theory that disagrees with reality. This research shows clear patterns in economics, such as strong cyclical and mean reversion behaviour, that clearly supports Austrian, Minskian and similar views. This clearly supports the theory of intrinsic value, and discredits orthodox economics. Similarly the inclusion of ideas from finance was not particularly surprising, people working in finance do not have the option of embracing intellectually beautiful ideas that don't describe reality; at least if they wish to remain working in finance. They are obliged to adopt rules of thumb that work. Some of the more thoughtful financiers, people such as Pettis, Shiller, Smithers, Cooper, Pepper & Oliver have then made insightful attempts to explain why these rules of thumb work in practice. In the field of market-microstructure in particular these approaches have been researched systematically and are both close to regularisation, and are also close to melding with the work of the more insightful financial econophysicists, despite the fact that the econophysicists have approached these problems from a completely different direction. Like econophysics, market-microstructure is highly mathematised, and very difficult to comprehend on a first reading. Perhaps because of this combination of complex mathematics and inscrutability, most curiously, market-microstructure appears to have been accepted as mainstream economics. This suggests most mainstream economists have never read any market-microstructure, as its rejection of marginality is, though very discreet, absolute. Which brings us to heterodox economics. Firstly the parallels between market-microstructure and post-Keynesian pricing seem, to this author, both obvious, and of considerable practical importance. Though I stand to be corrected, this parallel does not appear to have been noted previously, presumably because post-Keynesians don't read market-microstructure papers and vice-versa. The main reason for adopting classical economics was almost accidental. I had previously rejected the dabblings of both Foley and Wright into Marxian economics as misguided foolishness. I was wrong, they were right. My first reason for rejecting Marxian economics was because the labour theory of value is so obviously wrong-headed, the second was because I had believed that Marxian economics had been systematically disproved by neoclassical economics. More reading of economics quickly proved the second assumption to be false, Sraffa was the victor of the Cambridge capital controversies. The labour theory of value is indeed nonsense. However the concept of absolute value is not nonsense, it is in fact very powerful. The concept of 'negentropy' as value, as articulated by Ayres & Nair [Ayres & Nair 1984] for example, is not just basic common sense; it works as a theoretical approach, as evidenced by the models in part A. Once the labour theory of value is replaced by a "negentropy theory of value", not only does classical economics make perfect sense, it also allows economics to become a self-consistent theory that is an obvious subset of the natural sciences. A very large, very interesting and very important subset; but a subset nonetheless. 160 EFTA00625288 In contrast, the fundamental innovation of neoclassical economics; that value is not inherent, but is set in the collective sub-conscious of buyers and sellers has proved to be a spectacular non- achiever. This assumption also has the worrying theoretical feel that one somehow has to believe in fairies; that the value of a brick or a ham sandwich can dramatically change overnight just because a lot of people believe its value should change. That is not to say that I have an intrinsic problem with believing in fairies. When studying quantum mechanics or information theory, I find the explanations seem to depend on a worrying existence of an intelligent external observer. Given the assumed existence of quantum mechanics and systems described by information prior to humanity's descent from the trees, I find this worrying. However I feel obliged to accept both quantum mechanics and information theory because the maths works well, unbelievably well, in describing the characteristics of real world systems. In contrast, neoclassical economics, despite 140 years of theoretical effort has singularly failed to achieve a single macroeconomic model of the slightest usefulness. Neoclassical theory failed spectacularly to predict the credit crunch of 2008; as it failed to predict the crash in Argentina in 2002 before that, or the failure of LTCM (despite the Nobels) in 1998, the multiple crashes in Asia in 1997, of Mexico in 1994, of the collapse of the European monetary system in 1992, or the collapse of Japan into deflation in the early 1990's. At the time of writing it is clear that the central banks of the USA, the Eurozone, Japan, the UK, Switzerland, Sweden and others are all following their own significantly different policies, based primarily on experience and intuition. This is because they have no meaningful macroeconomic models. The ones they did have in 2008 have been quietly abandoned, and they are now largely flying by the seat of their pants with a finger in the air to check the weather conditions. Such is the legacy of a century and a half of neoclassical economics. It is the belief of the author, that the movement instigated by neo-classical economics to subjective value, remains the biggest and most damaging wrong turn ever made in the history of the sciences. The teaching of chaos, statistical mechanics and entropy is famously difficult. The concepts of liquidity and market microstructure are similarly opaque when first encountered. Despite this, once the ideas are grasped they are actually quite simple and can become easily understood and then become very powerful tools to understand problems. I have neither the teaching skills nor the space in a paper of this length to do justice in explaining these ideas. What I have attempted to do in Part B is to give a basic feel for the ideas, with very simplistic models and almost no mathematics. I have than also pointed to other authors, authors more skilled than myself, who can give greater depth and clarity than I can. Finally in section 13 I have included a reading list to point the way forwards into these subjects for mathematicians, economists and other scientists. In the sections that follow I have included some lengthy quotes from some authors, primarily Duncan Foley, Steve Keen and Ian Wright. This is mainly because they explain some of the points I wish to make very eloquently. In most cases I have then attempted to explain the ideas in alternative ways in my own words. Some readers may not find the extracts easy to follow on first reading. If this is the case I suggest that readers skim these extracts and read my own 161 EFTA00625289 words, then reread the extracts. It is hoped that the two different descriptions will help illuminate the underlying theories. It goes without saying that the basic ideas in part B are not my own. The ideas of mathematical chaos, statistical mechanics and basic entropy are centuries old, as are the ideas of classical economics. Other concepts such as maximum entropy production, market-microstructure, liquidity and post- Keynesian pricing theory are relatively recent; recent enough to be largely unknown in wider physics and economics circles. My own limited input includes, firstly, occasionally suggesting possible practical examples and uses that emerge from the theory — the ideas are speculative, and whether they actually prove to be useful remains to be seen. The intention of these proposals is to encourage a new way of tackling problems in economics and finance. More importantly, I believe I have pulled together an apparent rag-bag of ideas, from seemingly unconnected fields, that may allow a systematic approach to be put together that gives economics a strong, coherent, mathematically rigorous basis that transcends the petty boundaries of the many current competing economic models. Part B.I — Mathematics 6. Dynamics 6.1 Drive My Car Before moving into the ideas of non-linear dynamics and chaotic mathematics I would like to briefly start with a discussion of the difference between statics and dynamics. Imagine that you own a car, or better a pick-up truck, a small vehicle with an open space at the back for carrying loads. For the moment we will discuss what happens when the truck is parked, this is the case were the mathematics of statics is relevant. If the truck is unloaded it will be high up on its springs, with a big space between the top of the back wheels and the top of the wheel-arch on the body. This is a particular static equilibrium, the force of the gravity and the force of the spring come to a balance at a particular point. If you then put a dozen bags of cement in the back of the pick up truck, the truck will move down on its springs and the body will move closer to the wheels. This is a new static equilibrium at a different point where the new greater weight due to gravity balances with a new bigger force from the more compressed spring. Now the truck will also have dampers; shock-absorbers fitted. In a normal pick-up truck these dampers will be quite beefy, and will slow down the movement from one static equilibrium point to another. These dampers provide a frictional force, and from the point of view of the static equilibria beloved of economists, they are very inefficient. They physically prevent rapid movement from one static equilibrium to another. From this line of thinking it would be better to reduce the size of the dampers or just remove the dampers altogether. Then, following a point change in the weight, the truck would move to its new equilibrium much faster. 162 EFTA00625290 Using this line of thinking, a neoclassical economist could also point out that, once you started driving you won't be changing the load anyway so you don't need to worry about the dampers as you won't be moving away from whichever static equilibrium you started at. More thoughtful people will realise that this is not a sensible line of argument. A moving truck is in a dynamic situation. When you set off driving you will need to turn corners and you will sometimes hit bumps in the road, this will set of bouncing in the truck, and you need dampers to slow the up and down movements of the truck. Obviously if you drive down a dirt road, with a lot of bumps, you will need dampers or the truck will bounce about all over the place. What very few people realise, even very thoughtful people, is that dynamic systems are much more difficult to control than that. If you take the dampers off a car, and then you drive the car very carefully, down an absolutely flat, absolutely straight road (an airport runway say), within a few tens of seconds the car will start bucking like a bronco and will be almost undriveable. It doesn't matter how carefully you drive the car, the car will rapidly move into a strongly vibrating mode. The problem is that as soon as you start driving the car, you introduce extra time based equations into the system of mathematics that describes the car. This new system of mathematics, the dynamic model, is completely different to the static solution. It is not an extension of the static model, it is not a modification of the static model. It is a different system with different solutions. For a car without dampers the solution is similar to the Lotka-Volterra model seen in figures 1.2.1.2 and 1.2.1.3 in section 1.2 above. This solution is naturally unstable and rotates around a central point indefinitely. Even if you deliberately start the car off with conditions at the central point (which would be the solution to the static system), the car's movements will quickly spiral out to the circle of dynamic points. That is because this circle is the solution to the dynamic equations. The central point is not a solution to the dynamic system, so the car cannot stay at this point. The car will have a natural 'resonant frequency' and will move into this form of vibration. Like the Lotka-Volterra system, this vibrational mode is the equilibrium solution for this physical model. In this case the equilibrium is dynamic, it has constantly variable parameters. If you put the dampers back on the car, then the central point is a solution to the dynamic system, the behaviour of the car then becomes similar to that seen in figures 1.2.1.4 and 1.2.1.5 in section 1.2 above, or to that seen in some of the commodity models of section 3 and the macroeconomic models in section 4. Even if the car hits a bump and starts bouncing, its movements will be damped and will quickly move back to the stable point. That is why cars have dampers, they automatically and very simply change an unstable dynamic equilibrium into a stable dynamic equilibrium. In a static framework dampers are inefficient, they prevent rapid movement to a new equilibrium. In a dynamic framework, dampers are essential, they move the system from an ever-changing cyclical dynamic equilibrium close to the static dynamic equilibrium. Similar problems are found in many other systems, a famous example is the Tacoma Narrows suspension bridge ("Gallopin' Gertie") in the United States that was destroyed by the wind (for a little entertainment do an internet search for videos of 'Tacoma Narrows'). Common sense 163 EFTA00625291 suggests that the wind should not be strong enough to destroy a bridge built of steel. But the wind blew around the suspension cables and induced vibrations in the cables at their resonant frequencies. These then induced vibrations in the bridge at its natural frequency, which eventually built up enough to destroy the whole bridge. Nowadays suspension bridges are normally built with dampers installed on the cables to prevent vibrations building up, as well as vanes to prevent alternate vortex shedding (similar vanes can usually be seen on tall steel chimneys). More recently a similar problem occurred with the Millennium footbridge near St Paul's in London. This time the vibrations were induced in the bridge by pedestrians. In this case the pedestrians started movements in the bridge at the natural frequency of the bridge. The movements of the bridge then forced the pedestrians to walk at this natural frequency, so a feedback process built up that caused large movements in the bridge. The bridge had to be closed the day it opened, and stayed closed for some months until dampers could be installed. Another very elegant example of how dynamic systems can behave in unexpected ways is the example of traffic flows. A video of a beautiful example of a system moving into a stable but chaotic zone of behaviour is given at [New Scientist 2008]. Here a number of drivers were asked to drive in a circle at a constant 30km/h. They signally failed to achieve these very simple instructions An alternative system quickly set itself up with a clear and stable wave pattern of blocked vehicles moving around the system at a steady speed. This system of flows being blocked and forcing rhythmical patterns of fast and slow is exactly analogous to the flows of goods, and changes in prices in economic systems. For 140 years economists have treated economics as a static system. A Walrasian auctioneer compares all bids and offers in the market and then closes out all purchases and sales at a market clearing price. To compare two different economic points economists use 'comparative statics'. They look at one static point, say 'stationary truck unloaded', and then look at another point, 'stationary truck loaded', and then calculate the locus of movement from one point to the other. From this view economists conclude that economic systems will quickly and naturally come to an equilibrium, they also conclude that frictional forces are bad and prevent rapid movement to the equilibrium. In recent years economists have started using what they call 'dynamic' models. With the notable exception of the Goodwin models, these are lots of small stationary comparative static analyses carried out one after the other. This might be better described as 'high-frequency statics', and are equivalent to loading and unloading the truck rapidly with lots of small bags of cement. Even the Goodwin model is highly confused, attempting to model growth process, presumably long term exponentials, via the Lotka-Volterra model, which although it shows short term growth and decline, is most certainly a long-term stable model, not a growth model. Certainly none of the 'dynamic' models proposed in recent years have made it into the mainstream textbooks, for the simple reason that the models don't work and don't effectively model anything. To take the two mainstream economic texts cited in this paper, Mankiw [Mankiw 2004] has a dozen or so time based graphs, but all show actual data, not theoretical modelling. There are lots of theoretical graphs in Mankiw, but all are static or comparative static; almost all of them being variations of price versus quantity. Similarly Miles & Scott [Miles & Scott 2002], a much better book, has many dozens of time based data graphs but only one theoretical time based graph; their figure 7.2. There is no discussion of dynamic equilibrium in Miles & Scott, all theory is discussed in a comparative static framework. 164 EFTA00625292 A century and a half of neoclassicism has prevented economists getting in the car, turning on the ignition and releasing the handbrake. Economics is a dynamic system. Whether it is a trader selling shares on a stock market or shopper buying groceries in a supermarket, traditional auctions are notable by their absence. Prices are never formally closed, prices are settled dynamically in real time. They are set by price setters; market-makers or books in stock-markets, by suppliers in retail markets. These prices are set by people who look at the prices of competitors, the rate of purchase of goods, the inventory of goods in the shops, the prices of raw materials, etc. The values of all these items are historic, they are functions of past time. With a shop, the competitor's prices may have been collected the previous day. For a stock trader the competitor's prices may only be seconds old. But with high-frequency trading, seconds old is definitely pre-historic. So the most important variable in the functions that are used for setting prices is that of time. Price setting is a dynamic process, with a lot more equations than a static process. These dynamic systems give feedback loops and often give unstable equilibrium solutions just as with biological Lotka-Volterra systems and car suspension systems. This is painfully obvious to see in the cyclical behaviour of stock-markets, house prices, commodity prices, currency fluctuations, etc. These fluctuations are inherent in economics. Because economies are dynamic systems. The fluctuations of stock-markets, house prices, commodity prices are a result of natural dynamic equilibria. Neoclassical economics states that the fluctuations shouldn't exist, and if they do it is a result of frictional inefficiencies. As a result the policy recommendations of neoclassical economists make the fluctuations in dynamic economies worse. If neoclassical economists genuinely believe that comparative statics is a sensible way to analyse and manage dynamic systems like economies, they should prove it by taking the shock- absorbers off their cars. 6.2 Counting the Bodies - Mathematics and Equilibrium In his book, Debunking Economics [Keen 2004], Steve Keen puts his finger on the problem at the heart of economics. Economists are using the wrong sort of mathematics when they attempt to solve their problems: Economics remains perhaps the only area of applied mathematics that still believes in Laplace's dictum that, with an accurate enough model of the universe and accurate enough measurement today, the future course of the universe could be predicted. For mathematicians, that dictum was dashed in 1899 by Poincare's proof of the existence of chaos. PoincarE showed that not only was it impossible to derive a formula which could predict 165 EFTA00625293 the future course of a dynamical model with three or more elements to it, but even any numerical approximation of this system would rapidly lose accuracy.... The more appropriate starting point for mathematical models of the economy are dynamic equations, in which the relationships between variables cannot be reduced to straight lines. These are known as nonlinear differential equations. The vast majority of these cannot be solved, and once three or more such equations interact, they are impossible to solve. Table 1 summonses the situation. Economic theory attempts to analyse the economy using techniques appropriate to the upper left-hand part of Table 1 (with boldface text), when in fact the appropriate methods are those in the lower right-hand part (with cells shaded gray). Table 1 The solvability of mathematical models (adapted from Constanza 1993) Equations Linear Non-linear One Equation Several Equations Many Equations One Equation Several Equations Many Equations Algebraic Trivial Easy Possible Very difficult Very difficult Impossible Ordinary Differential Easy Difficult Essentially Impossible Very difficult Impossible Partial Differential Difficult Essentially Impossible Impossible Or alternatively, as Wright puts it: The state-space of a system is the set of all possible configurations of the DOF [degrees of Freedom]. A particular configuration is a 'point' in state space. In general we find that many neat systems, if they enter equilibrium, tend toward a point or trajectory in state-space. A canonical example is a set of weighing scales. Place some weights on each arm and the scales will tend toward an equilibrium point in which the internal forces balance and the system is at rest. This is a simple kind of deterministic equilibrium, in which the equilibrium configuration is a subset of state-space. The classical mechanics concept of equilibrium was a founding metaphor of the 19th Century marginal revolution in economics (e.g., see Mirowski (1989)). And it appears in a more developed form in 20th Century neoclassical general equilibrium models (e.g., Debreu (1959)). But most messy systems, if they enter equilibrium, do not tend toward a subset of state-space. [Wright 2009] And, of course, economics is not a neat system; economics is a messy system, economics is a multibody system. Foley gives this background in more detail: The concept of equilibrium states has played a decisive role in the development of quantitative sciences. The study of mechanical equilibrium, conceived as a balancing of forces in a static system, clarified the fundamental notions of force and mass in the course of the 17th century development of Newtonian physics. The 19th century saw the emergence of characteristically statistical descriptions and theories of mass phenomena (see Stephen Stigler, 1986; Theodore Porter, 1986) which migrated from the social sciences to physics, where they blossomed into the 166 EFTA00625294 marvelously successful and equally marvelously puzzling methods of statistical mechanics (see Lawrence Sklar, 1993). These statistical theories eschew the goal of describing in detail the situation of all the subsystems that constitute a large system with many degrees of freedom in favor of drawing strong conclusions about the observable macro behavior of the system based on statistical considerations. As Edwin T. Jaynes (1978), following the approach of 3. Willard Gibbs, realized, statistical equilibrium in all its various applications occurs when the appropriately defined entropy of the system is maximized subject to the appropriate constraints. The entropy is a strictly concave function of the probability distributions describing the system, and the constraints are typically linear or convex functions, so that this maximization implicitly calculates shadow prices (Lagrange multipliers) for each of the constraints, which are uniform over the subsystems and characterize its important properties in equilibrium. One might have expected that these statistical methods would be a natural basis for the attempt to put social theory, and particularly economic theory, on firm mathematical and quantitative foundations. It is a commonplace of social and economic methodology to point out that human behavior, no matter how idiosyncratic and unpredictable it is in individual human beings, is subject to statistical regularity and predictability in the aggregate. The Maxwell-Boltzmann-Gibbs methods of statistical mechanics, furthermore, are based on the calculation of dual variables that have the dimension of prices, and effectively view the establishment of physical equilibrium as a kind of economizing process. Thus it would not have been surprising had economic theory developed a statistical concept of equilibrium. By a curious turn of the history of thought, however, economic theory, despite an almost obsessive fixation on physical models and analogies (see Philip Mirowski, 1989), gave birth to an idiosyncratic conception of equilibrium fashioned more on the mechanical analogy, in the work of Leon Walras, Vilfredo Pareto, Irving Fisher, and Francis Y. Edgeworth (to name a few of the more important figures). In Walras' equilibrium each subsystem (firm or household) deterministically maximizes profit or utility facing uniform prices "cried out" by an "auctioneer': The auctioneer experiments until she has determined an equilibrium price system at which the offers to sell and buy each good in each market are exactly balanced. Because this theory assumes as an axiom that no transactions take place until the equilibrium prices are determined, households with the same preferences and endowment will always receive the same bundle of consumption goods in the equilibrium: horizontal equity (or equal treatment) is guaranteed by this a priori assumption. The Walrasian conception of equilibrium is in sharp contrast to the statistical thermodynamic conception in which the equilibrium energy distribution of subsystems (say, molecules) is achieved by their exchange of energy as they interact during the transient approach to equilibrium. In a thermodynamic context we would be astonished to find that two molecules that started in the same energy state generally end up in the same energy state. Apparently physicists tried to alert Walras to the peculiar nature of the conception of equilibrium he was proposing, but without success, either because Walras did not understand the statistical point of view very well, or because he considered it and rejected it on other grounds. J. W. Gibbs served as Irving Fisher's thesis adviser at Yale apparently without raising questions about the non-statistical conception of the equilibrium systems Fisher was studying. Francis Edgeworth distrusted Walras' conception of the auctioneer enough to propose an abstract combinatorial model of exchange, based on the idea of recontracting among coalitions of traders (which has developed into the modern theory of the core). The recontracting feature of Edgeworth's theory, however, implies equal treatment of agents with the same preferences and endowments, thus reproducing the key elements of Walras' system. One aim of Walras' and Edgeworth's theories was to explain the emergence of coherent market price systems from the decentralized interaction of atomistic traders. Unfortunately, both Walras and Edgeworth resort to strong and unrealistic assumptions to address this issue: Walras 167 EFTA00625295 invented a fictional information centralizing auctioneer, and Edgeworth posited costless recontracting among agents. The statistical approach offers an elegant alternative in this respect: market prices can be regarded as the shadow prices or Lagrange multipliers arising inherently from entropy maximization. In this view the system constraints (market clearing conditions) give rise to global prices just as the constraints of volume and energy in a physical system give rise to the emergent properties of pressure and temperature in a confined gas. The atomistic agents in a market "feel" the effects of these global constraints combinatorially as the relative difficulty of changing their holdings of goods, just as individual molecules "feel" the global constraints on energy and volume in terms of the likelihood of reaching any given energy state. [Foley 1996b] Few physicists read economics books. Even the physicists who are profoundly interested in economics, and produce papers on economics, rarely read economics books. The main reason; for the scientifically trained, is the extraordinarily unscientific approach that they have. Statements such as 'Assume a demand curve ', 'assume a budget line etc simply inculcate an overriding feeling of 'why?'. Where on earth do these assumptions come from, and why should they be assumed. For more intrepid physicists who persevere, it comes as something of shock to discover that utility theory was directly copied from the field theory of physics in the 1870's, and copied with gross errors. More extraordinarily, having absorbed field theory and adopted it as the core of economics, economics has studiously ignored the majority of mathematics developed since the 1870's (game theory being a notable exception) even though this mathematics would be much more appropriate for the analysis of economics. In this regard economics resembles a tenacious terrier, unable to eat the plates of meat set down in front of it, due to its inability to let go of the very well chewed bone it has firmly gripped in its teeth. The full horror of this calamity is recounted at length, in very entertaining detail, in Mirowski's book 'More Heat than Light'; a book that, contrary to its title, many economists might find enlightening reading. [Mirowski 1989] The central point of Mirowski's book is that utility was copied from field theory, but in doing so economists threw away the basic conservation principles that give field theory any meaning. If fields are not conservative, then there is little point in drawing curves and lines to visualise them. Without conservation laws, two different paths between the same two points will give different values, and so the curves and lines do not have values that can be meaningfully represented; neither graphically nor mathematically. The second problem with field theory as a basis for economics, is that it is simply, and absolutely, not appropriate for multibody systems. In their different ways; gravity, electromagnetism, relativity and quantum mechanics are all varieties of field theory. But in the application of their mathematics interactions are limited to two bodies, so for example an electric current can be seen as a unified flow of the separate electrons, moving at the same speed in the same direction. Newton's theory of gravity was the first and the classic description of field theory, and with two bodies; the sun and a single planet for example, Newton's theories work perfectly. 168 EFTA00625296 But even with a very simple multibody planetary system, Newton's theories break down, and fail to explain behaviour exactly. The errors are small, but the errors are there. As soon as you get to three bodies; for example the sun, earth and moon, it becomes impossible to find exact solutions for the motions of the bodies. Even in a three body system the motions of the bodies become chaotic and unpredictable at a detailed level. In 1890 Poincare demonstrated that it is actually impossible to solve the equations for a three body system in a simple field system, so even a system as simple as the sun, moon and earth is chaotic, and can not be accurately predicted over the long term. This, and a full history of analysing the motions of the planets is written up in the very enjoyable book by [Peterson 1993], Poincar4's work is discussed in chapter seven. It is important to note that this chaotic motion is noticeable in objects as large as planets. This is not simply the chaos of quantum effects or the stochasticity found in Black-Scholes. This is 'deterministic chaos' or usually 'chaos theory'. The chaos is present even in problems that can be described in exact mathematics and are completely free from random exogenous or microscopic behaviour. The original Lotka-Volterra model is just such a mathematical system. In practice the meeting of foxes and rabbits will have a stochastic element, but the system at a macroscopic scale is described very well by deterministic equations. In deterministic chaos, the behaviour of the system can change dramatically according to very small changes in initial conditions, as described in the analogy of butterflies causing tornados a continent away. However it is of course obvious, that although the positions of the earth, sun and moon can not be predicted exactly, they can be predicted to a very high degree of accuracy, and that their paths follow strongly constrained bands. This is a different type of equilibrium, a constrained chaotic equilibrium, that never stabilises at a fixed point, and so never becomes a static equilibrium. The Lotka-Volterra equilibriums (but not the General Lotka-Volterra's) fall into this class of equilibrium. So in a simple eco-system, the number of rabbits and foxes can vary significantly, but a peak in the population of either will be followed by a trough; and the long term average values of both populations will be very stable. In economics, Minskian, Austrian and Goodwin type systems fall into these categories, and the commodity and macroeconomic models discussed in sections 3 and 4 above attempt to model such systems. Such systems can show different behaviour depending on their underlying characteristics. The systems can be very stable staying close to the long-term averages, they can oscillate strongly, or they can grow explosively to infinite positive or negative values. And of course, real economies clearly follow the same patterns empirically. Business cycles have been evident and documented for at least two centuries. The periodicity may have changed as economies have changed, but the fluctuations remain. These can be short term cycles of building up and drawing down inventories, they can be the 15-20 year land cycles documented by Harrison [Harrison 2005], they can be the decadal mean-reversions of stock prices documented by Smithers and Shiller [Smithers 2009], they can also be the once per lifetime financial crises such as the great crash or the credit crunch caused by the retirement of all the people who remember the reasons why strict controls were imposed on the financial system after the last such crisis [Napier 2007]. And in the great crashes the system moves out of periodicity into explosive behaviour. 169 EFTA00625297 Fortunately in the last half-century or so there has been a great deal of progress in analysing such systems in the field known as 'nonlinear dynamics' and there are many standard ways of solving such problems. In fact the Lotka-Volterra system is one of the simplest such systems and strictly is not necessarily non-linear, though in my models a little non-linearity has been introduced. There are two big reasons, and one small one, why economics needs to use the mathematics of non-linear dynamics. The first reason is the inclusion of time as a variable. In comparative statics prices change with supply, and prices change with demand. Equilibrium is reached when the prices match each other and supply equals demand. The mathematical derivatives for the equilibrium relate the prices and the quantities. In the real world prices cannot change instantaneously, the main derivatives of prices are with respect to time. The economy is constantly moving with a continuous series of trades, the economy rarely formally 'clears' prices. This is true even for goods such as cheap manufactures that show strong price stability, this is equivalent to a car moving smoothly down a motorway at constant velocity, it is not equivalent to a parked car. If you put a brick under the wheel of a parked car, a new equilibrium point will be reached in a couple of seconds, if you drive over a brick while doing 70mph, it might take a little longer for a new equilibrium to be reached. In real economies the most important derivatives are the time derivatives, and the mathematical framework for economics must be cast in these derivatives. Adding in the time derivatives allows extra degrees of freedom and complexity, and normally moves the real equilibrium away from the static equilibrium, it also allows oscillating and explosive solutions that do not have a short-term or any equilibrium respectively. The analogy between stock-market crashes and normal (eg car) crashes is a mathematically exact one. Comparative statics states that a temporary liquidity crisis should not bring an economy to its knees, in the same way that putting a brick under the wheel of a parked car should not destroy the car. However if the car is doing 70mph, it is quite likely that the car will end up wrapped around a lamp-post. Similarly a liquidity crisis in a debt-laden economy can turn into a general solvency crisis. The most obvious way that time is important to the economy is with the delay of installation of capital in capital intensive sector and also with housing and office building. But time delays can be much shorter and still have strong effects, the research of Milton Friedman showed that monetary effects had delays of six months or more. Inventory stocking cycles operate on similar timescales. In financial markets time delays allow momentum effects on the scale of seconds. The second big reason that economics needs non-linear dynamics is that the variables in economics have two-way effects (and as discussed above, the effects are fed back with time delays). These mutual feedback loops are legion. For example: Increasing prices of company shares creates new apparent wealth - new apparent wealth allows people to invest in companies, so pushing up share prices. Increasing wealth in the productive sector allows more consumption — more consumption allows increased investment in the productive sector. Increasing debt allows more liquidity and rising asset prices - rising asset prices gives more apparent capital against which more debt can be secured. 170 EFTA00625298 A decrease in saving propensity gives a boost to consumption and the productive sector — more earnings from the productive sector allows a decrease in saving propensity. In all these cases, and many, many, more, economics has mutually reinforcing feedback loops. And in all these cases the feedback can reverse and work in the opposite direction. In all these conditions you have coupled systems with feedback, where: dx/dt = f(x,y) and also dy/dt = f(x,y) In these systems y gives feedback to x, and x gives feedback to y. Even with linear systems this can give periodic and explosive behaviour. All of these are analogous to the lynx and hares in the original model discussed in section 1.2 the populations of both can expand or contract over long periods before an external limit changes the direction of growth. The imposition of limits brings us to the third reason for using non-linear dynamics. Some functions in economics are non-linear. The most obvious ones are when you have genuine scarcity such as a fixed supply of labour or urban land suitable for house-building. Minerals such as gold, copper, platinum or oil also have scarcity, at least in the short-term, as installing capital is expensive and takes time. In finance, access to credit and other financing can be limited beyond a certain point and can lead to highly non-linear functions. A very good text explaining these approaches, with lots of practical examples, is 'Nonlinear dynamics and Chaos: with Applications to Physics, Biology, Chemistry and Engineering', by Strogatz [Strogatz 2000], a good alternative is Hirsch, Smale & Devaney [Hirsch et al 2003]. Prior to either of these books, chapter eight of Keen gives a very good brief introduction to chaotic systems, Ruhla also gives an excellent introduction with a little more maths [Keen 2004, Ruhla 1992]. Although the approach may seem very new to most economists, actually the techniques are extensions of techniques familiar from basic economics. Most non-linear systems are not directly solvable, so mathematicians often resort to graphical representation in 'phase space' to resolve the problems. This ends up with intersecting lines and curves not dissimilar (and a bit more fun) than the diagrams found in comparative statics. Jacobian matrices, for example, appear a third of the way through Strogatz. Although dynamic systems can be very complex and are often mathematically insoluble, there are standard approaches to analysing these systems, and it is usually possible to produce important mathematical conclusions out of such analysis. It is usually possible to identify the controlling variables and the different zones of stability and instability. Indeed one of the interesting things about complex systems is that while they can be very difficult to analyse and describe, they are usually very easy to control. Usually it is just a question of installing suitable damping or time delays in the system. In engineering such systems are commonly encountered within control systems where problems of feedback can be highly deleterious. 171 EFTA00625299 On the plus side, control system engineering, and systems dynamics, have investigated the problems of such systems in detail, and when the underlying characteristics of the system are understood, relatively minor changes in the system can result in dramatic changes to the stability of the system. See for example Control Systems Engineering [Nise 2000]. In the following two sections, and also in section 9.2 I take a qualitative look at house prices and share trading and ideas of how the natural cycles in these markets could be damped out. Section 9.2 is somewhat out of order in the paper, this is because it is necessary to introduce some ideas of market microstructure first. The ideas in these sections are pretty much common sense on the issue of housing, the ideas regarding share trading are much more speculative and contentious. The main point of the discussion is to make it clear that, counter-intuitively, just as with shock- absorbers in cars, introducing damping can create a better system. 6.3 Chaos in Practice — Housing in the UK It is a common aphorism of economics that it is a difficult science to progress, as it is not possible to carry out suitable experiments. This is tosh. Experiments are regularly carried out in economics, though usually by accident. The problem is that economists ignore the results, even when the damage to the public is substantial. The example of housing provides one of the clearest and most important experiments ever carried out in economics in the UK. Figure 6.3.1 here Figure 6.3.1 above shows the prices of housing in the UK from 1953 to 2010, divided by the average wage, prepared using data from the Nationwide Building Society and the UK Office of National Statistics. The high house prices immediately following the Second World War were a consequence of substantial loss of housing during the war and a suspension of house construction for the six-year duration of the war. During the 1950s and 60s access to mortgages in the UK was tightly regulated and controlled by government micro-management of financial institutions, with direct lending ceilings imposed on banks and building societies; resulting in strict rules on eligibility, deposit sizes, etc. During this period house prices showed remarkable stability at a cost of roughly 3.0 to 3.5 times average salary. It is very important to note that, despite the strong state controls on access to housing finance, the 50's and 60's were a time of substantial private house building in the UK, as the post war generation, including large sections of the working class, fled their city terraces for suburban semis. Despite the restrictions imposed by the state, even at these regulated 'low' prices, demand created lots of supply. As can be seen in figure 6.3.2 below UK private house building reached a prolonged peak in the mid 1960s. 172 EFTA00625300 Figure 6.3.2 [ONS 2004] Access to mortgages was liberalised in 1971 under the policy of 'Competition and Credit Control', which, despite its title, pretty much abandoned credit control; in line with neoclassical theory. This resulted in the 'Barber boom', starkly clear in figure 6.3.1, stimulated by the resulting rise in liquidity, and the first, of many, UK house price bubbles. From the 1970's onwards, the UK housing market has been characterised by vicious cyclic booms and busts, with a very clear reversion to the pre-Barber long-term trend or 3 to 3.5 at the bottoms of the cycles. These cycles are identical in form to the ones discussed in the commodity models in section 3 and the macroeconomic models in section 4. Compare figure 6.3.1 (or 6.3.3 below for the US) with the outputs in figures 3.3.2 and 4.3.3 in previous sections. These are exactly the outputs you would expect from a non-linear differential system that is showing quasi-periodic cyclical stability. In fact, if you look at the pre-1971 section it is possible to see the same cyclical fluctuations, just that the amplitude of the cycles is very much smaller. It is important to note that at the bottom of both the actual housing data, and the commodities models, prices reach their 'real', 'fundamental', Sraffian values. At these prices the value of housing represents the cost of the inputs. The same can be seen even more clearly in data from the United States (this time deflated for cpi); see figure 6.3.3 below. Figure 6.3.3 here [Shiller 2010] Supply is capable of balancing demand at these Sraffian prices. Any increase above these prices is pure speculation and rent-taking. Indeed the persistence of these cycles is deep within the economy of the UK. In his book 'Boom, Bust, House Prices, Banking and the Depression of 2010' [Harrison 2005] (first published in 2005) Fred Harrison not only confirms how trivially easy economic forecasting is if you are willing to believe in fundamentals and cyclical behaviour, but also shows that the cycles in the UK go back to at least the middle of the eighteenth century. As an experiment, you could scarcely ask for clearer data output. The basic system dynamics are substantially and dramatically changed following a point change in policy. Not only that, but this experiment has controls, Germany and Switzerland for example, have retained strict controls on mortgages for house purchases and don't suffer from strong cyclical booms and busts in house prices. 173 EFTA00625301 The consequences of this experiment are of some considerable importance to the welfare of all people living in the UK. Figure 6.3.4 below has the average value of house prices included for the two periods. Figure 6.3.4 here On the scales used, average house prices from 1955 to 1970 were 3.3 times average salary. During the last thirty years, from 1971 to 2009, average house prices in the UK have cost an average of 4.0 times average salary. In the latest boom, prices have gone to even higher levels, though a meaningful average can't be given until the cycle has bottomed. The net result of the liberalisation of credit in 1971 was the increase in average cost of housing for all Britons by roughly 23%. In the last cycle, from 1996 to 2010, prices were fully 40% higher than the '55-70 baseline rate. This represents a very significant reduction in welfare for residents of the UK. It also has many secondary negative effects. Many more poorer people are unable to afford housing, and are forced to rely on social housing and subsidies paid from taxation. This then helps to create ghettos of poorer people, which exacerbate employment and crime problems, which again requires more social spending and higher taxation. Even for the well-off that can still afford to buy houses, on average they must spend more money on housing, reducing that available for saving, pensions, or simply enjoying life. The beneficiaries here are the financial companies that issue the mortgages, or rather the investors and savers with these companies. Once again, exactly like the commodity cycles in section 3, We have a case of unjustified rent-taking on a massive scale. Given that private sector rents are substantially set by house prices, some of the rent-taking is literal. Taken as a whole, this represents a large transfer of wealth from the poor and middle income individuals to the rich. Housing suffers from the same problem as capital-intensive commodities, as modelled in section 3 above. Construction of housing takes a finite time, and so house prices can go up significantly before market mechanisms have time to work. Unfortunately, housing also has the same problems of endogenous liquidity creation that is seen in the macroeconomic model. As house prices go up, people feel richer, and also as with shares 'momentum' kicks in, and house prices, and the economy as a whole keeps rising, until finally house prices become unaffordable for new entrants in the market, and the bubble bursts. As a capital-intensive industry, housing is naturally cyclical. Although this conclusion is based on casual observation, it seems that housing seems to be much more dangerous to the overall economy than other asset classes. Booms in commodities and shares seem to be survivable when they turn into busts. Normally such collapses are followed by recessions and rebalancing for a couple of years, and then the economy picks up again. Housing crashes seem often seem to morph into financial crises, threatening the stability 174 EFTA00625302 of the whole economy, and recovery from such crises normally takes much longer. It seems likely that this is because housing is the only highly-leveraged asset generally available to the public. This again shows that the contrast between the comparative statics of neoclassical economics, and the real world of dynamic differential equations is stark. With comparative statics it is easy to 'prove' that credit controls and other government interventions 'must' increase the price of goods, and so reduce the welfare of the public. So neoclassical economists always push for removal of such controls. In the real world, where speculative cycles can be endogenously created within the economic system; credit controls and other 'interferences' in the market work beneficially by 'damping' the cyclical behaviour. It may be counterintuitive, but in the right circumstances, applying controls and apparent 'costs' to the market actually reduces the price of goods. And reduces them substantially. In the area of UK housing, the experimental data shows that the reduction would be over 20% if strict credit controls were reimposed tomorrow as they were in the '50s and '60s. It is essential to understand that the logic of this argument is supported by the experimental data of figures 6.3.1 and 6.3.3. It also happens to be supported by the mathematical models, if you understand the right maths, but that is a secondary issue. The experimental data is clear; credit controls reduce the cost of houses, by very helpfully damping, and largely removing, the cyclical nature of house price movements. If you reject this experimental data, and hold on to a theory that states, purely on theoretical logical grounds, that removing credit controls must make house prices cheaper, then you are not following science. You are following a religious dogma. Again neoclassical economists, by failing to understand basic dynamic systems, accidentally support massive rent-taking by insisting on deregulation of markets in search of nebulous market efficiencies. The 'Barber Boom' of the early 1970's ended with a spectacular crash and the 'secondary banking crisis' in which the Bank of England had to launch the 'lifeboat' to rescue thirty or so banks in the UK's very own dry run of the credit crunch. Despite this early warning, deregulation was not rolled back, but instead was systematically pursued in all areas of UK finance and economics. The results can be seen in figure 6.3.1, recurring housing bubbles in UK housing of increasing size and ferocity. The strength of this religious dogma is quite profound. Since 1971 the UK has had ten chancellors and eight prime ministers, all advised by what must be many hundreds of the most intelligent economists that work in the UK. Despite this the 'reforms' of 1971 have never been questioned, never mind reversed. The citizens of the UK are consequently still obliged to spend their lives paying off their expensive mortgages. The worst economic experiment carried out in the UK in modern times continues. The damage that this dogma has done to Britain is writ large in figure 6.3.4. From the early 1970s onwards, the liberalisation of credit has increased house prices in the UK by 23%. Another more subtle problem can be seen in figure 6.3.2. Private sector house-building continued at a roughly constant rate from the 1960s to the present. The liberalisation of finance failed spectacularly in encouraging new house-building; presumably because its main effect was to make houses more expensive. 175 EFTA00625303 What did change in the 1970s was the collapse of the provision of social housing. From the mid- 1970s onwards the government reduced funding for social housing, primarily because, from the 1970s onwards, the UK has had ongoing severe budget problems. This was due to a dramatic increase in the need for welfare payments compared to the 1950s and 60s. The need for welfare payments was needed to cope with the dramatic rise in unemployed and the poorly paid in the 1970s, a problem that has never gone away. The blame for the steep rise in the poor in the 1970s has been blamed variously on oil price shocks, de-industrialisation, union power, foreign competition, etc. While all of these factors may have had contributions, it is the belief of the author that the main factor was the ongoing deregulation starting in the Barber era. This increased overall debt levels and changed the Bowley ratio and so the GLV distribution. This not only created the poor, but forced higher taxes on the rich. It is perhaps time to end this experiment. Unfortunately the political drive for deregulation is powerful. The biggest problem, at least in Anglo-Saxon countries, is that many people believe that housing is a good long term investment. Going back to figure 6.3.1 or 6.3.3 for the UK and US it is clear that the 'investment' value of housing is a chimera. Over the long term, growth in the value of houses is derisory and barely keeps up with the growth in earnings. Stock market growth is typically 5% higher than this. Smithers discusses the dual properties of housing as both a form of consumption and investment in Wall Street Revalued p 107-108 [Smithers 2009]. The fact that housing is fundamentally consumption is demonstrated by the continuous reversion to a fixed proportion of wages. Equally this demonstrates that, for all the apparent growth in the booms, housing is a lousy investment, which over the full business cycle only manages to match the increase in wages. Figures 6.3.1 and 6.3.3 show clearly that in the long-term housing is a proportion of wages, and behaves as consumption. Governments should treat it as such, and actively prevent houses being treated as investments, and most certainly should prevent them being treated as speculative investments. Despite this the booms are usually longer than the crashes, and inflation often masks real falls in house prices. Both of these effects may explain the visceral attachment of the public, and worse politicians, to housing as investments. Historically, politicians have invented many ways of subsidising housing purchase; so assisting bubbles to form, and so unintentionally, and perversely, making housing more unaffordable. In the recent credit crunch the US did this so effectively as to put the financial system of the whole world at risk of collapse. Politicians are a very big part of this problem. They seem profoundly addicted to housing booms. Encouraging home ownership is always popular, though if people don't have the wealth or income to maintain the homes they purchase, home ownership alone doesn't solve any problems. More worryingly politicians seem to enjoy the public's enjoyment of rising house prices. Very few politicians seem to be able to comprehend that house prices cannot rise above gdp growth rates over the long term, neither do they seem to appreciate that long-term rising house prices necessarily produces high, and ultimately unaffordable house prices. This is puzzling. Whether you are a dyed in the wool socialist or a radical free marketeer, it should surely be the aim of any politician to ensure decent affordable housing for all. In addition to the problem of the housing cycle causing over priced-houses, there are other very major issues. Firstly the diversion of resources to the housing sector that would be better used 176 EFTA00625304 elsewhere, secondly and more importantly, as Harrison has shown, the cycles in housing appear to be the main driver of the cycles of boom and bust in economy as a whole. One of the central themes of this paper is that governments should assist in the transfer of capital to poorer people. But housing is not productive capital, and it is the wrong target for such transfers. Of course, housing can be a very good short-term investment if you get your timing right. Anybody who bought in the UK in 1970, 1978, 1983 or 1996 will almost certainly make a substantial unearned profit when they sell. But this of course is simply speculation, and speculation in it's non-healthy form. This represents a transfer of wealth to the well informed, and usually already wealthy. This is wealth that is removed from the hands of ordinary people. And this gives another big problem with allowing cyclical behaviour in economic systems. Most people buy without addressing the timing of booms and busts. If you are lucky and buy at the bottom you win, if you are unlucky and buy at the top you lose. As such allowing this cyclical behaviour in the housing market allows massive inter-generational transfers of wealth on a completely arbitrary basis. Looking both at the UK data and the US data in figures 6.3.1 and 6.3.3, a very worrying development is that in both countries the size of the booms is steadily rising, though the falls back to normal are the same. From a controls point of view this is very worrying, it suggests that the cycles could be even more dramatic and dangerous in the future — as if the last two years were not traumatic enough. Faced with a dynamic, cyclical system, standard control systems knowledge can be used to control the system. There are two ways to remove cycling (what engineers call 'hunting') in a control system. One is to use deliberate counter-cyclical feedback; most central banks try to do this using interest rates to control the economy as a whole. As central bankers are only too aware, this is not an easy way to control anything. A good example of such a feedback loop is a domestic shower system. A combination of a difficult to use mixer valve, and the delay between making the change at the tap and feeling the change in the water temperature often results in alternating flows of water that is too hot or too cold . Wherever possible, a much better solution is to use damping of the cycle. When done successfully this can result in a dramatic drop in oscillations with fairly minor, adjustments to the system. This is like the example of using shock absorbers with a car's wheels to prevent the car vibrating wildly on its springs every time it hits a bump. The strict credit controls used in the UK prior to 1971 provided just such an effective damping system. If all else fails it is imperative that such controls are reintroduced in the UK. However it may be possible that less draconian measures may be just as effective. As a rule of thumb, to be effective, damping measures need to have a time span of a similar order to that of the natural cycle time of the system, as a minimum they should be of a length of half a cycle or so. For the UK Harrison [Harrison 2005] shows strong evidence for a fifteen to 177 EFTA00625305 twenty year cycle for house prices. Sensibly, damping measures need to be of the order of ten years or so. Looking closely at the US data in figure 6.3.3; there is the same flat trend as the UK at the bottoms of the cycles; showing the same reversion to real, non-speculative, prices. It is also clear that the booms are a relatively new phenomenon. A subtly different experiment has been carried out in the US. The change in behaviour of the housing market appears to be correlated with the rise in non-standard mortgage products. Historically the US has used fixed-rate mortgages, only moving to adjustable rate mortgages comparatively recently. In the UK adjustable, or short term fixed mortgages have been the norm for many years, and it is very difficult to get fixed rate mortgages of more than five years. The finance industry does not like fixed-rate mortgages. It leaves the issuers holding interest rate and inflation risk. Moving to adjustable rates gives the appearance of moving the risk to individual mortgage holders. This in itself is a practice to be questioned in a democratic society. Why sophisticated finance companies should be allowed to offload complex financial risk onto individuals with little mathematical, let alone financial, training is not clear. In reality, offloading risk in systemic fashion like this simply creates systemic risk. As has been made abundantly clear in recent years; ultimately the only realistic holder of systemic risk is the taxpayer. Allowing financial companies to issue variable rate mortgages is to give the financial companies government subsidised one-way bets. Figure 6.3.5 below gives a comparison of mortgage types issued in various different countries in Europe. 6.3.5 here [Hess & Holzhausen 2008] The mainly variable countries are Greece, Spain, Ireland, Luxembourg, Portugal, Finland and the UK. This pretty much speaks for itself. The solution to this is trivially straightforward. All loans that are secured against domestic property should be limited to a ten-year minimum and a thirty year maximum. They should also be fixed rate, or, as a minimum, be a fixed percentage above rpi or cpi, throughout the period of the mortgage. This would move interest rate risk back on to the shoulders of the finance industry. Where it belongs. Variable rate mortgages should be strictly illegal in any self-respecting democracy. There are other sensible mechanisms to reduce the use of houses as investments, especially as speculative investments. The most obvious one is to have a capital gains tax that is more punitive than that for other investments. The tax should be charged on all houses, including first homes, without exception. Sensibly this would be a tapered tax; starting at say 20% for the first year, then drop by two percentage points per year, so reaching zero after ten years. 178 EFTA00625306 A much better approach would be to have a sales tax on all houses. This should be applied to the seller of all houses, whether they have increased or decreased in value. Again, sensibly, the tax should be tapered over the years. A tapered capital-gains tax or house sales tax, with a ten-year taper should bring in the damping of the sort required to deal with a 15 to 20 year endogenous property cycle. People buying houses to live in would not be punished, speculators would be. In addition annual property taxes, or land taxes, should be charged on the value of houses or on the value of the underlying land, rather than on the occupants, as many local taxes are. Another sensible policy would be to have compulsory mortgage indemnity guarantee (MIG). House purchasers would be obliged to take out insurance to cover full potential losses against potential negative equity, ie the difference between mortgage loan value and likely sale value of house. Such insurance would be cheap if the purchaser had a large deposit and prices were below the long-term trend. The insurance would be very expensive if the deposit was small and it was the height of a boom. As such, compulsory MIG should act in a strongly counter-cyclical manner. (For an off topic discussion of a different sort of deposit protection, refer also to the endnote 6.3.1 below.) Many countries enforce minimum deposit requirements [Hess & Holzhausen 2008]. This seems a very sensible policy, as those with small deposits are far more likely to default, see for example figure 6.3.6 below. 6.3.6 here [FT/M 2010] It can be seen that arrears rates increase dramatically as deposit sizes reduce. As with variable rate mortgages, when governments allow financial institutions to offer low deposit rates; that is highly leveraged asset purchases, they allow financial institutions to offload their risk onto the state. There is a more sophisticated and better way of addressing this particular risk problem. Rather than prescribe laws on deposits, a more effective law would define a maximum limit of say 80% of the sale value of a house that could be repaid to pay off debt secured on the property. So if a homeowner was foreclosed on, and their property was sold off, a minimum of 20% of the sale proceeds would go to the homeowner, and the other 80% would be shared by all the creditors who have loans secured on the property. This would have a number of advantages. It would have the same effect as a minimum deposit requirement of 20%. Banks would generally be reluctant to supply a mortgage of greater than 80% of the value of the house. It would also make it much more difficult to evade the minimum deposit rules by taking out secondary loans secured on the house. More subtly it would also act in a counter-cyclical manner. When house prices were at historical lows, banks might be willing to lend 90% mortgages, confident that house price were likely to 179 EFTA00625307 rise. Conversely, when house prices were significantly above their long-term averages banks would require larger and larger deposits due to their fears that house prices might drop in the future. Similarly they would be very reluctant to allow mortgage equity withdrawal. In addition to the passive management techniques discussed above, there is also a strong case for active counter-cyclical monitoring and management of the economy by central banks and other monetary authorities. Despite protestations to the opposite, housing bubbles are very easy to spot. The first obvious measure is that shown in figures 6.3.1 and 6.3.3 for the US and UK. The ratio of house prices to median wages shows very strong patterns of reversion to mean. Similar patterns are also seen in ratios of housing costs to rental costs. When house prices are correctly valued, housing costs (mortgage payments, etc) are close to rents on equivalent properties [FT 2010]. If either of these ratios increases significantly above the long-term trend then you are moving into a housing bubble. At this point the central bank should intervene to prick the bubble as early as possible. This could be by increasing the sales tax or capital gains tax on houses, increasing deposit and MIG requirements or by imposing a tax on mortgage debt. Finally, if none of the above work effectively to damp markets then the necessary solution is to simply bring back the same credit controls that the UK had prior to 1971. It would also be wise to impose similar controls on commercial property, especially office accommodation, which also seems to be subject to dramatic fluctuations with the business cycle. Of course, many economists, banks, building societies, estate agents, and most politicians will believe, and argue vociferously, that bringing in control measures such as those above will slow the economy and make homeownership available only to the few. These people are wrong. The economic theories are wrong. Experimental data confirms that these theories are wrong. When listening to these people it is important to bear in mind that it was the very same economists, financiers and real estate professionals that created the recent housing booms, and the consequent crashes in the US, UK, Ireland and Spain. Both housing and commercial building are very important as candidates for effective damping for two very big reasons. Firstly as leveraged assets the busts following the booms can be very financially damaging. Secondly housing and commercial construction have very big impacts on employment in the construction industry and so have large effects on the economy as a whole. [6.3.1 An Aside on Deposit Insurance Talking about deposit insurance, but wandering completely off-topic; it has puzzled me as to why compulsory default insurance is not instituted for bank deposits. 180 EFTA00625308 This would not be intended as a realistic way of insuring the deposits, but as a way of introducing market pricing into the risk of government bank deposit insurance. If done correctly this would also reduce the moral hazard element of public assurance of bank deposits. Realistically, in a democratic capitalist society, a government run central bank will always need to be the lender of last resort and will need to guarantee the deposits of members of the general public to a basic level. However, such guarantees remove all risk for all but the richest members of the public. It encourages them to move their deposits to the highest interest payers without any need to worry about whether the bank is well run or in danger of collapse. This then encourages all banks, even the well run, to compete on interest paid while ignoring the risk taken. Indeed the well run banks are forced to match the foolishness of their badly run competitors if they wish to stay in business. A way to resolve this is to insist that all deposit-taking banks apply compulsory deposit insurance on their deposits. The insurance would be strictly in the form of a percentage charged on the deposits, and this would be displayed in parallel to the interest rate paid by the bank. It would be illegal for a particular bank to offer its own insurance on its own accounts, and it would be compulsory for banks to offer all alternative insurance from all alternative deposit taking banks. Bank customers would be able to swap their insurance simply and electronically at any time they wished, from a visible list of alternatives available via the account. All deposit taking banks would be obliged to offer a price for insurance for all their competitors. They may wish to price their insurance at a high level, but they would be obliged to price, and would be obliged to take on the insurance at the price offered. In the event of a bank failing, the insuring banks would be obliged to pay the deposits of the insured depositors from their own bank's funds (to avoid spreading systemic risk, reinsurance of this risk would be prohibited; banks would be obliged to carry a portion of funds against these risks on their balance sheets). The central bank would remain the ultimate insurer of the deposits but would only step in if there was a pattern of systemic risk, and even then only after bank shareholders and all bondholders were wiped out. In the event of a single bank failure due to poor management, the other banks, the insurers, would carry the costs by themselves. Further rules would apply even in the event of systemic failure. Government deposit guarantee would apply up to a maximum limit (say £100,000), but this maximum guarantee would apply across all deposits for a single person, no matter how many accounts failed at any number of banks. The maximum paid out would be £100,000 even if the person invested £l0k in each of 20 different accounts, all of which failed simultaneously. Similarly the government deposit guarantee would only cover £100,000 maximum over any 10-year rolling period. Individual bank customers would only be able to waive the compulsory bank insurance where they could demonstrate that they already had £100,000 deposited in insured accounts. Although the above may sound complex, it would be trivial to put in place in a modern electronic retail banking system. The net effect of this would be to create a market in retail bank deposit insurance. While the Bank of England may have been surprised by the collapse of Northern Rock, Bradford & Bingley and HBOS; the author was not. The rumours of all these impending bank failures were wandering around internet forums from early 2007 onwards. Banking insiders knew that the funding models for these banks were unsustainable and dangerous. Forcing banks to insure each other's deposits would force banks to price the risk on badly run banks like Northern Rock at higher rates than better run banks such as FISBC and Barclays. By pricing this risk strictly as a percentage rate, the general public would gain direct visibility of the default risk. Under this regime, a well-run bank might still pay lower interest rates, but would be compensated with even lower insurance rates. This should make the net interest rate; interest less insurance, of the low risk bank better than that of the risky bank. Competition would no longer be on interest rates alone. With the best will in the world, such a system would not be capable of insuring all deposits in the event of a systemic bubble. But that is not the point. The point is; that by introducing effective market based pricing of risk, the general public and the banks would be penalised for indulging in the risk-taking that encourages bubbles in the first place. 181 EFTA00625309 Additionally, the general rates of insurance should act as both an early warning system for the monetary authorities and even as a counter-cyclical assistance in popping bubbles in the first place. In normal times, insurance rates for all but the most foolish of banks should be ridiculously low. In the event of the economy moving into bubble conditions, insurance rates would start to creep up on the riskiest banks. This would then start to pass on the infection, via the insurance, to other banks, but at a much earlier stage than normally happens when entering a financial bubble. Faced with the obligation of holding more reserves on their balance sheets to cover the deposit failure of others, all banks would be obliged to cut back on credit in general. All banks would be affected, but with the strongest effects on the worst run and most highly leveraged banks. Monitoring of individual and overall insurance rates would give the central banks live data on the perceived risks of the banks in their charge, as well as the financial system as a whole.] 6.4 Low Frequency / Tobin Trading THE spectacular collapse of so many big financial firms during the crisis of 2008 has provided new evidence for the belief that stockmarket capitalism is dangerously short-termist Shareholders can no longer with a straight face cite the efficient-market hypothesis as evidence that rising share prices are always evidence of better prospects, rather than of an unsustainable bubble. If the stockmarket can get wildly out of whack in the short run, companies and investors that base their decisions solely on passing movements in share prices should not be surprised if they pay a penalty over the long term. But what can be done to encourage a longer-term perspective? In the early 1980s shares traded on the New York Stock Exchange changed hands every three years on average. Nowadays the average tenure is down to about ten months. That helps to explain the growing concern about short-termism. Last year a task force of doughty American investors (Warren Buffett, Felix Rohatyn and Pete Peterson, among others) convened by the Aspen Institute, a think-tank, published a report called "Overcoming Short-Termism". It advocated various measures to encourage investors to hold shares for longer, including withholding voting rights from new shareholders for a year. [Economist 2010a] Warren Buffet is of course a value investor, the sort of investor who intuitively understands the workings of the companies models in section 2 of this paper. The sort of investor that the efficient market hypothesis states cannot exist. Value investors also intuitively understand that the short-term liquidity and momentum effects seen in the commodity and macroeconomic models in sections 3 and 4 not only make value investing difficult, but also add no value to the process of creating wealth that capitalism aspires to. The proposals of the Aspen Institute were pretty much stillborn for a number of reasons. Firstly, because orthodox economics assumes, erroneously, that any cost imposed on market transactions must increase costs to the consumer. Secondly because such a tax would destroy a substantial part of the finance industry, which makes the majority of its profits by charging rents on the very volatility they create in the first place. And thirdly, and more reasonably, if such a tax were imposed in one country, trading would simply move to an alternative jurisdiction. To understand just how short-term the finance industry has become, it is worth noting that stock-trading is now dominated by 'high-frequency trading' (HFT). In the major stock-markets supercomputers trade billions of dollars of trades in seconds using automated algorithms. Individual bids and offers may be held open for fractions of a second. High frequency trading 182 EFTA00625310 systems are now being co-located within stock-exchange buildings as the speed of light now means that companies trading from a few blocks away are at a significant disadvantage. To anybody who has actually worked in a real company, the idea that the real market value of a normal company can change from millisecond to millisecond is bizarre; it is palpable nonsense. A full discussion of high-frequency trading is postponed to section 9.2 below. It is my belief that Buffet, Shiller, Smithers et al are correct, and that the unnecessary volatility is induced endogenously in share markets, causing excessive movements away from real value on timescales from seconds to decades. It is my belief that the decadal movements are caused by liquidity at a macroeconomic scale, a problem that will need tackling at a macroeconomic level — this is discussed in detail in section 8.2.1 below. Other timescales are much shorter and give the appearance of being quasi-periodic momentum effects. Although the evidence is controversial, typical time-scales for the periodicity appear to be on the order of fifty and two hundred trading days, with other shorter time scales also present. A system is proposed below that would dampen the fluctuations on these timescales. The solution proposed is a private-sector approach, independent of government. Following the same logic as housing in the previous section, it is proposed to introduce damping with losses imposed on early retrading on the lines of those proposed by Buffet et al. This would be done by introducing a new class of shares, or special investment certificates, in the companies. These shares would have different rules as to their trading. The issuing of such shares would be voluntary, at the choice of the companies involved. In the same way as housing, damping would be imposed with a haircut of say 10% imposed on anybody who sold a share within the specified time period. The haircut would be paid back to the company in which the share is held at the time of sale, as such it would be effectively a 'negative dividend' on the share, paid by the owner to the company. The haircut would automatically be deducted from the sale proceeds. In extremis the haircut would be imposed for a period of say three years. However unlike housing it is not proposed that the haircut on all shares be imposed for the full term of three years. This would present great problems for pricing of the shares. If a large purchase was made of a company's shares this would kill the market in that company's shares for years at a time, which would make price discovery for the company almost impossible. Instead it is proposed that all shares that have been sold are marked as 'locked'. This would be in contrast to all the remaining shares that would be 'unlocked'. Every trading day a random selection would be made across all the currently 'locked' shares and 1% of all the currently locked shares would be unlocked. The owners of these newly unlocked shares would then be able to sell the shares immediately without penalty. Assuming 250 days of trading per year, then this release of 1% of shares per trading day would give a half-life for locked shares of roughly six months. This means that if every single sha

Document Preview

PDF source document
This document was extracted from a PDF. No image preview is available. The OCR text is shown on the left.

Document Details

Filename EFTA00625129.pdf
File Size 29135.7 KB
OCR Confidence 85.0%
Has Readable Text Yes
Text Length 500,000 characters
Indexed 2026-02-11T23:08:00.967580
Ask the Files