EH.net is owned and operated by the Economic History Association
with the support of other sponsoring organizations.

British Mail Steamers to South America, 1851-1965: A History of the Royal Mail Steam Packet Company and Royal Mail Lines

Author(s):Forrester, Robert E.
Reviewer(s):Kaukiainen, Yrjö

Published by EH.Net (December 2014)

Robert E. Forrester, British Mail Steamers to South America, 1851-1965: A History of the Royal Mail Steam Packet Company and Royal Mail Lines. Farnham, UK: Ashgate Publishing, 2014. xiii + 248 pp. £70 (hardcover), ISBN: 978-1-4724-1661-2.

Reviewed for EH.Net by Yrjö Kaukiainen, Department of History, University of Helsinki.

Royal Mail ships played a prominent role in the early history of steam shipping. Around the middle of the nineteenth century, in particular, the subventions paid by the British government were of vital importance for the success of the companies in question; gradually, however, the economic importance of mail contracts diminished and RMS vessels rather became just a special — albeit prestigious — category of passenger liners.

Forrester’s book is of special interest because it describes the history of a company which started in 1840 and lived until 1965 — a period which almost perfectly coincides with the rise and decline of what we could call traditional liner shipping. Thus the RMSP/Royal Mail Lines lived through most of the great revolutions which totally changed the outlook of international shipping in the course of a “long century.”

Overall, the book is a sound example of the British business history tradition. Its significant strength is the practical expertise of the author who — before moving to the publishing business and finally to maritime history studies — worked for two decades as a ship officer. Accordingly he can competently describe and discuss the technical and navigational aspects of his story. In this sense, Chapter 2 is particularly valuable for vividly describing the teething troubles of early steam shipping. In the 1830s and 1840s, there was no infrastructure to support power-driven vessels and the pioneering companies had to build it by themselves. This included coaling stations — which were particularly important in the period before compound engines started to decrease fuel consumption — as well as networks of foreign offices and agents, and sometimes even harbor development. I think this is something which every scholar working with cliometric analysis concerning the transition from sail to steam should read. However, to put things into a perspective, quite similar lack of infrastructure was faced by early container shipping.

The author presents a rather detailed picture of the development of the fleet as well as of passenger and freight volumes, which were particularly fast before First World War. This period of growth even included the beginning of refrigerated transportation of meat.  A good background for these developments is provided by overviews of relevant markets: imports and exports of South American countries, as well as the competition with other shipping lines. One of the author’s main arguments is that the mail steamers were important in instituting and maintaining the considerable British financial, commercial and industrial presence in Latin America.

As the book can be characterized as narrative rather than analytical it may not fully meet the expectations of scholars who are interested in a macro perspective. Its outline consists of strictly chronological chapters which makes it difficult for a reader to perceive an overall picture of different long-terms trends. For example the role of transoceanic postal services in information transmission — which is specifically referred to in the back-cover blurb — as well as their improvement, receives no systematic analysis. Regarding the continuous decline of postal subventions for the company’s economy this, however, cannot be regarded a fatal omission. Moreover, such aspects, at least as far as the period until 1875 is concerned, have already been quite sufficiently analyzed in a Finnish doctoral dissertation published in 2007.

The scope of the book is fairly Britain-centered, which can even be seen in the bibliography. Thus, comparisons with French, German or American mail lines (or other competing shipping) are quite infrequent. One more limitation is already indicated by the title of book: while the Royal Mail lines sailed both to the West Indies and South America, the former area has in practice been excluded. This obviously was a personal choice of the author — he served for ten years, until 1965, as a deck officer on the company’s South American ships. The limitation, however, involves practical problems. Since the company’s accounts did not differentiate between various destinations (which is not really surprising because a number of ships sailed both to the West Indies and South America) the specific economic returns of South American, or Brazil and Argentine traffic cannot be found in the relevant source material. In business history terms, a sounder alternative would have been to deal with the entire RMSP-company right from its beginning (as the subtitle of the book actually suggests).

It is clear that a book with a wide chronological scope is able to offer interesting data for further research. A good example of valuable information can be found in a quotation of freight rates to and from South America in the 1850s (p. 22) which fully confirms the surmise that shipping goods on early steamers was expensive. The description of the company’s economic decline, collapse and reorganization in the late 1920s presents a rather typical case of the difficulties faced by ocean shipping after the First World War. Even the final downturn and demise of the Royal Mail Lines in the 1960s is a good example of a general trend. When the United Nations Conference of Trade and Development (UNCTAD) was instituted in 1964 it almost immediately became the organ of developing countries requiring more equal rules of the game (vis-a-vis the industrial world) for liner shipping. While the actual liner code was adopted only after twenty years, South American countries were very active in developing, with subventions, their own merchant navies already in the 1960s. This harmed not only British lines but even the South America lines of other West and North European countries. Finally, I would like to point out that the Fleet list published as an appendix, offers valuable data of technical development across 120 years.

Yrjö Kaukiainen, Professor Emeritus of European History, University of Helsinki, recently published “The Role of Shipping in the ‘Second Stage of Globalisation,’” International Journal of Maritime History (2014).

Copyright (c) 2014 by EH.Net. All rights reserved. This work may be copied for non-profit educational uses if proper credit is given to the author and the list. For other permission, please contact the EH.Net Administrator (administrator@eh.net). Published by EH.Net (December 2014). All EH.Net reviews are archived at http://www.eh.net/BookReview

Subject(s):Transport and Distribution, Energy, and Other Services
Geographic Area(s):Europe
Latin America, incl. Mexico and the Caribbean
Time Period(s):19th Century
20th Century: Pre WWII
20th Century: WWII and post-WWII

The History of Bankruptcy: Economic, Social and Cultural Implications in Early Modern Europe

Editor(s):Safley, Thomas Max
Reviewer(s):Hansen, Bradley A.

Published by EH.Net (June 2014)

Thomas Max Safley, editor, The History of Bankruptcy: Economic, Social and Cultural Implications in Early Modern Europe. New York: Routledge, 2013. xi + 250 pp. $130 (hardcover), ISBN: 978-0-415-68730-0.

Reviewed for EH.Net by Bradley A. Hansen, Department of Economics, University of Mary Washington.

The papers in this volume demonstrate that bankruptcy provides an avenue to explore cultural, as well as economic, dimensions of early modern Europe. The book makes available the work of several scholars that had not been previously available in English. The papers are grouped into three categories: social context, business practice, and institutional change. The categorization is a largely matter of emphasis since most of the papers contain information about each of these areas. In general, the authors take two different approaches to bankruptcy. The first approach uses evidence from bankruptcy cases to provide a window into business practices in early modern Europe. The second approach explores how the methods of resolving bankruptcy cases evolved over time.

Many of the papers in the book use the failures of particular merchants to examine business practices, while others examine large numbers of failures. Mark Häberlein examines the failure of the Sulzer brothers in late sixteenth century Augsburg, with particular emphasis on the ways in which bankruptcy was connected to issues of honor, kinship and gender. He notes, for instance, that bankrupt merchants “had to take their seats among the women during weddings and funerals.” Similar to Haberlein, Mina Ishizu uses the records from the failure of John Leigh, to examine British finance of Atlantic trade, and Dana Stefanova examines how one Viennese bank dealt with the failures of two businesses. Several papers examine the failures of public, or at least semi-public, firms. Paulo Avallone examines public banks in Naples, Mauro Carboni and Massimo Fornasari examine the Monte de pieta in Bologna, and Andre Wakefield examines the failure of Celle’s Zuchthaus.  Thomas Brennan examines court records of over 750 defaults in Champagne between 1769 and 1772, mapping the network of debtors and creditors to explore the development of markets in the region. Similarly, Natacha Coquery examines the bankruptcy records of 120 Parisian shopkeepers during the eighteenth century, and Klas Nyberg and Häkan Jakobsson use merchant bankruptcy cases to reconstruct financial networks in eighteenth century Stockholm.

The remaining papers examine the evolution of institutions, both formal and informal, for dealing with bankruptcy. Paul Fischer compares German legal texts from 1447 and 1749 to document changes in bankruptcy law in the early modern period, noting the decreased emphasis on “shame and infamy” and increased emphasis on efficient administration.  John MacLeod provides a detailed review of changes in laws governing fraudulent transfers by debtors in seventeenth century Scotland. Dave De ruysscher chronicles the changes in the ways that Antwerp merchants dealt with bankruptcy between 1490 and 1540; he finds that Antwerp was an early mover in the shift away from a focus on punishing debtors toward a focus on efficient handling of insolvency. Jerome Sgard provides a comparative analysis of bankruptcy and debt renegotiation in France and England. He finds that, although English bankruptcy law, unlike French bankruptcy law, did not provide mechanisms for renegotiation, English merchants used other legal procedures to facilitate private renegotiation.

The wide variety of the approaches, in terms of geography, time period, and central questions, is both the strength and the weakness of the book. Anyone interested in any aspect of bankruptcy, business practices, institutional change or early modern Europe will be able to find one or more papers they should read. On the other hand, even the scholar most interested in bankruptcy in early modern Europe may not be interested in every chapter.  The papers on the Italian banks and the German Zuchthaus, for instance, are in many ways more concerned with issues of public finance and administration than with business insolvency.

A wide variety of perspectives are represented in the book. Methodologically, however, the approach is fairly uniform.  There is some quantification, but explicit economic theory and statistical analysis are not employed. The methods employed are appropriate for the questions that are asked, but the book raises questions that will probably require more explicit use of theory and statistical analysis. To what extent were the credit networks of bankrupt firms representative of all firms? What consequences did differences in institutions, from one country to another or over time, have for economic performance? To what extent, were different institutional arrangements associated with different rates of business formation, default, or economic growth?

Although Sgard’s paper is the only one that is explicitly comparative, there are several recurring themes throughout the essays. Credit transactions were an essential part of commercial life. Merchants were embedded in dense networks of credit, in which they typically acted as both borrowers and lenders. Credit transactions and insolvency were not just economic issues, narrowly construed, but involved issues of kinship, religion, honor and gender. Changing economic conditions led to changes in the institutions, both public and private, that governed the resolution of insolvency.  In general, the institutions that governed the resolution of insolvency became less focused on the punishment and more focused on efficiency, especially resolving collective action problems among creditors. This book should provide an essential starting point for the comparative historical analysis of one of the central features of capitalism. In addition, it suggests the rich sources that the records of economic failure can provide for historians to answer a wide array of questions.

Bradley A. Hansen is the author (with Mary Eschelbach Hansen) of “Crisis and Bankruptcy: The Mediating Role of State Law, 1922-1932,” Journal of Economic History (2012); “Religion, Social Capital and Business Bankruptcy in the United States, 1921-1932,” Business History (2008), and “The Role of Path Dependence in the Development of U.S. Bankruptcy Law, 1880-1938,” Journal of Institutional Economics (2007).

Copyright (c) 2014 by EH.Net. All rights reserved. This work may be copied for non-profit educational uses if proper credit is given to the author and the list. For other permission, please contact the EH.Net Administrator (administrator@eh.net). Published by EH.Net (June 2014). All EH.Net reviews are archived at http://www.eh.net/BookReview

Subject(s):Business History
Markets and Institutions
Social and Cultural History, including Race, Ethnicity and Gender
Geographic Area(s):Europe
Time Period(s):16th Century
17th Century
18th Century

An Economic History of Nineteenth-Century Europe: Diversity and Industrialization

Author(s):Berend, Ivan T.
Reviewer(s):Voth, Hans-Joachim

Published by EH.Net (March 2014)

Ivan T. Berend, An Economic History of Nineteenth-Century Europe: Diversity and Industrialization. Cambridge: Cambridge University Press. 2013.  xviii + 521 pp. $45 (paperback), ISBN: 978-1-107-68999-2.

Reviewed for EH.Net by Hans-Joachim Voth, Department of Economics, University of Zurich.

Like a mythical animal out of an old fable, Ivan Berend’s latest book is a two-headed creature. On the one hand, it is an erudite, readable, and insightful overview of nineteenth-century economic history, written by a scholar who commands encyclopedic knowledge with ease, and more often than not displays an ability to package all this wisdom intelligently. On the other hand, there are serious lapses of judgment, oversights, and missing key concepts and references in the literature; and there are also serious issues with the writing, structure, and analysis.

The first chapters bring the reader up to speed about how the author thinks economic history should be done, and what happened before the nineteenth century (part I). Part II looks at success (“Successful Industrial Transformations of the West”), and Part III describes the causes of failure in the periphery. The part on success contains chapters that look at science and education, agriculture, transport, business and finance, demography, and the role of the state. A full section examines Europe’s interactions with the rest of the world. The part on failure examines mechanisms for the transmission of the “Western spark,” at the advantages of being on the periphery, and the state as predator (Balkans and the borderlands of Austria-Hungary). Throughout, the author weaves together an enormous literature with a light touch. The description of how in Bosnia-Herzegovina in the period before World War I, illiterate teachers taught prayers to children, with neither understanding what the words meant, is a perfect way to start the section on stagnation in the periphery. Oddly, the same anecdote is repeated, almost word for word, later in the book.

The Industrial Revolution takes center stage – as is only to be expected in a book on nineteenth century European economic history. The author moves from the description of (in the words of an anonymous schoolboy cited by T.S. Ashton [1]) the “wave of gadgets” that washed over England in the eighteenth century to broader explanations of why industrialization occurred and how it spread. Some parts here feel like a throwback to the days of W.W. Rostow’s stages of economic growth [2], with a mechanical summary of how many boxes on the list of requirements for take-off a country has managed to tick. Berend also approvingly cites Robert Allen’s recent work on the role of high wages (and cheap coal) in explaining why the Industrial Revolution was British [3] – the argument being that dear labor and abundant energy gave strong incentives to create technologies that economized on the former by using the latter. This is an example of what economists call directed technological change, and it is typically invoked to explain, say, the rising college premium in an age of falling energy prices. There are several issues with this view (not least the fact that English wages were not very high once one factors in high levels of productivity – as pointed out by Cormac O’Grada, Morgan Kelly, and Joel Mokyr [4] – and not cited by Berend), and the book does not quite rise to the level of even-handed and insightful discussion that this reviewer would have liked to see.

The conclusion at the end of Chapter 1 illustrates some of the issues and problems in this book: “The reasons behind the British Industrial Revolution are extremely complex (my emphasis). … The British Industrial Revolution … was not a deus ex machina, but rather the outcome of six hundred years of gradual progress and change inside Europe… Industrialization is a complex civilizing process, ‘a phenomenon which exists only on a certain level in the cultural, economic, social … evolution of man’ (Zamorski 1998, 36). Socio-economic and cultural developments occurred in a very symbiotic way, inspiring and engendering one another” (p. 77). We never learn what these symbiotic relationships actually are. A lack of clarity pervades the analysis here; there is no ordering of factors, no discussion of what mattered more or less, no description of what produces complex interactions. The book itself has actually surprisingly little to say on cultural and social antecedents of industrial development.

There are significant lacunae in several sections. Institutional approaches, especially when married to rigorous economic analysis, are unevenly covered. For example, this must be one of the few books on economic development published in the last decade or so that does not cite a single book or paper by Daron Acemoglu. Important work on the “Rise of Atlantic Europe” gets no billing [5], nor does the attempt to identify the effects of exogenous institutional change through the French Revolution. Similarly, demographic forces and their interaction with living standards are not given much room (though Greg Clark’s theory that differential reproductive process is at the heart of development makes a cameo appearance). Given that in the mind of many, the nineteenth century marks the point in history when Malthusian forces started to weaken, this is more than just an oversight. While citing such obscure works as Joseph Love’s Crafting the Third World: Theorizing Underdevelopment in Rumania and Brazil (1996), the book builds no bridges to current influential work in economics on the determinants of long-run growth. Inexplicably, it misses “unified growth theory” (and all the associated empirical analysis) entirely [6, 7, and 8].

One of the early chapters makes it clear where the author positions himself in the profession – firmly in favor of “old-style” economic history in the tradition of Fernand Braudel and Carlo Cipolla, and opposed to new-fangled Cliometrics. This chapter cites some negative remarks about quantitative economic history by Nobel Laureate Robert Solow, before then moving on to note that some of new economic history is perhaps not all bad. The book also makes bold claims about how economics-inspired economic history can never take culture and social dynamics seriously: “Socio-political and cultural factors, and all of those other phenomena that are impossible to quantify or incorporate into a few factors of an economic model, are evidently marginalized in the analysis. In many ways, however, it is precisely those qualitative features, those knowledge, social-cultural, and behavioral patterns and those historically determined characteristics that are the most profound factors influencing economic growth” (p. 16).  It is hard to take the confident verdict (“impossible to quantify or incorporate…”) seriously. The author arrives at this conclusion by ignoring a vibrant, rapidly growing literature that thinks about precisely these questions – with seminal contributions by Alesina, Guiso, Sapienze, Zingales, Nunn, Becker and Woessmann, to name only a few [9, 10, 11, and 12]. Similarly startling is the absence of any reference to the recent work by Besley, Persson, and others on state capacity [13, and 14] – a topic the book touches on but never quite gets to grip with.

To sum up, inside the actual book, there is an ambitious, erudite, wide-ranging, and often highly insightful manuscript that is struggling to get out – but failing. Apparently, the project did not receive the tender, loving care from the editorial process that could have helped. The publisher (Cambridge University Press) did not produce professional graphs; the book is adorned with numerous charts straight out of Excel that would look poor in a fifth-grader’s class report, with 3-D lines thrusting upwards at weird angles to illustrate that three numbers increased over time, pointless legends (“S1″) right next to graphs with a single line, and pie charts drawn from such an angle as to distort the proportions completely. The index is almost comically bad. For example, the entry on “Western Europe” comes with no less than 111 individual page entries (“272, 273, 274, 276, 277…”), but only six sub-headings – which virtually guarantees that no reader will be able to find anything in this book other than by playing search-and-cite.

Despite the numerous shortcomings, this book contains a broad-ranging overview of key developments in the European economy between 1800 and 1900. As such, it can usefully serve as a textbook for an introductory class on European economic history – provided it is supplemented with all the essential readings the author left out.

References:
1. T.S. Ashton, The Industrial Revolution, 1760-1830 (CUP Archive, 1948).
2. W.W. Rostow, The Stages of Economic Growth: A Non-Communist Manifesto (University Press, Cambridge, 1960).
3. R.C. Allen, The British Industrial Revolution in Global Perspective (Cambridge University Press Cambridge, 2009).
4. M. Kelly, J. Mokyr, and C.Ó. Gráda, “Precocious Albion: Factor Prices, Technological Change and the British Industrial Revolution” (2012).
5. D. Acemoglu, S. Johnson, and J. Robinson, “The Rise of Europe: Atlantic Trade, Institutional Change, and Economic Growth,” American Economic Review 95, 546–579 (2005).
6. O. Galor, From Stagnation to Growth: Unified Growth Theory (Elsevier, 2005; http://ideas.repec.org/h/eee/grochp/1-04.html), pp. 171–293.
7. O. Galor, and O. Moav, “Natural Selection and the Origin of Economic Growth,” Quarterly Journal of Economics 117, 1133–1191 (2002).
8. Q. Ashraf, and O. Galor, “Dynamics and Stagnation in the Malthusian Epoch,” American Economic Review 101, 2003–41 (2011).
9. L. Guiso, P. Sapienza, and L. Zingales, “Long Term Persistence” (NBER Working Paper 14278, 2008).
10. S. O. Becker, and L. Woessmann, “Was Weber Wrong? A Human Capital Theory of Protestant Economic History,” Quarterly Journal of Economics 124, 531–596 (2009).
11. N. Nunn, and L. Wantchekon, “The Slave Trade and the Origins of Mistrust in Africa,” American Economic Review 101, 3221–3252 (2011).
12. N. Fuchs-Schundeln, and A. Alesina, “Good-Bye Lenin (Or Not?),” American Economic Review 97, 1507–1528 (2007).
13. T. Besley, and T. Persson, “The Origins of State Capacity: Property Rights, Taxation, and Politics,” American Economic Review 99, 1218–1244 (2009).
14. T. Besley, and T. Persson, “State Capacity, Conflict, and Development,” Econometrica 78, 1–34 (2010).

Hans-Joachim Voth is the author (with Mauricio Drelichman) of Lending to the Borrower from Hell: Debt, Taxes, and Default in the Age of Philip II (Princeton University Press, 2014).

Copyright (c) 2014 by EH.Net. All rights reserved. This work may be copied for non-profit educational uses if proper credit is given to the author and the list. For other permission, please contact the EH.Net Administrator (administrator@eh.net). Published by EH.Net (March 2014). All EH.Net reviews are archived at http://www.eh.net/BookReview

Subject(s):Economic Development, Growth, and Aggregate Productivity
Economywide Country Studies and Comparative History
Industry: Manufacturing and Construction
Geographic Area(s):Europe
Time Period(s):19th Century

Women Workers in the British Industrial Revolution

Joyce Burnette, Wabash College

Historians disagree about whether the British Industrial Revolution (1760-1830) was beneficial for women. Frederick Engels, writing in the late nineteenth century, thought that the Industrial Revolution increased women’s participation in labor outside the home, and claimed that this change was emancipating. 1 More recent historians dispute the claim that women’s labor force participation rose, and focus more on the disadvantages women experienced during this time period.2 One thing is certain: the Industrial Revolution was a time of important changes in the way that women worked.

The Census

Unfortunately, the historical sources on women’s work are neither as complete nor as reliable as we would like. Aggregate information on the occupations of women is available only from the census, and while census data has the advantage of being comprehensive, it is not a very good measure of work done by women during the Industrial Revolution. For one thing, the census does not provide any information on individual occupations until 1841, which is after the period we wish to study.3 Even then the data on women’s occupations is questionable. For the 1841 census, the directions for enumerators stated that “The professions &c. of wives, or of sons or daughters living with and assisting their parents but not apprenticed or receiving wages, need not be inserted.” Clearly this census would not give us an accurate measure of female labor force participation. Table One illustrates the problem further; it shows the occupations of men and women recorded in the 1851 census, for 20 occupational categories. These numbers suggest that female labor force participation was low, and that 40 percent of occupied women worked in domestic service. However, economic historians have demonstrated that these numbers are misleading. First, many women who were actually employed were not listed as employed in the census. Women who appear in farm wage books have no recorded occupation in the census.4 At the same time, the census over-estimates participation by listing in the “domestic service” category women who were actually family members. In addition, the census exaggerates the extent to which women were concentrated in domestic service occupations because many women listed as “maids”, and included in the domestic servant category in the aggregate tables, were really agricultural workers.5

Table One

Occupational Distribution in the 1851 Census of Great Britain

Occupational Category Males (thousands) Females (thousands) Percent Female
Public Administration

64

3

4.5

Armed Forces

63

0

0.0

Professions

162

103

38.9

Domestic Services

193

1135

85.5

Commercial

91

0

0.0

Transportation & Communications

433

13

2.9

Agriculture

1788

229

11.4

Fishing

36

1

2.7

Mining

383

11

2.8

Metal Manufactures

536

36

6.3

Building & Construction

496

1

0.2

Wood & Furniture

152

8

5.0

Bricks, Cement, Pottery, Glass

75

15

16.7

Chemicals

42

4

8.7

Leather & Skins

55

5

8.3

Paper & Printing

62

16

20.5

Textiles

661

635

49.0

Clothing

418

491

54.0

Food, Drink, Lodging

348

53

13.2

Other

445

75

14.4

Total Occupied

6545

2832

30.2

Total Unoccupied

1060

5294

83.3

Source: B.R. Mitchell, Abstract of British Historical Statistics, Cambridge: Cambridge University Press, 1962, p. 60.

Domestic Service

Domestic work – cooking, cleaning, caring for children and the sick, fetching water, making and mending clothing – took up the bulk of women’s time during the Industrial Revolution period. Most of this work was unpaid. Some families were well-off enough that they could employ other women to do this work, as live-in servants, as charring women, or as service providers. Live-in servants were fairly common; even middle-class families had maids to help with the domestic chores. Charring women did housework on a daily basis. In London women were paid 2s.6d. per day for washing, which was more than three times the 8d. typically paid for agricultural labor in the country. However, a “day’s work” in washing could last 20 hours, more than twice as long as a day’s work in agriculture.6 Other women worked as laundresses, doing the washing in their own homes.

Cottage Industry

Before factories appeared, most textile manufacture (including the main processes of spinning and weaving) was carried out under the “putting-out” system. Since raw materials were expensive, textile workers rarely had enough capital to be self-employed, but would take raw materials from a merchant, spin or weave the materials in their homes, and then return the finished product and receive a piece-rate wage. This system disappeared during the Industrial Revolution as new machinery requiring water or steam power appeared, and work moved from the home to the factory.

Before the Industrial Revolution, hand spinning had been a widespread female employment. It could take as many as ten spinners to provide one hand-loom weaver with yarn, and men did not spin, so most of the workers in the textile industry were women. The new textile machines of the Industrial Revolution changed that. Wages for hand-spinning fell, and many rural women who had previously spun found themselves unemployed. In a few locations, new cottage industries such as straw-plaiting and lace-making grew and took the place of spinning, but in other locations women remained unemployed.

Another important cottage industry was the pillow-lace industry, so called because women wove the lace on pins stuck in a pillow. In the late-eighteenth century women in Bedford could earn 6s. a week making lace, which was about 50 percent more than women earned in argiculture. However, this industry too disappeared due to mechanization. Following Heathcote’s invention of the bobbinet machine (1809), cheaper lace could be made by embroidering patterns on machine-made lace net. This new type of lace created a new cottage industry, that of “lace-runners” who emboidered patterns on the lace.

The straw-plaiting industry employed women braiding straw into bands used for making hats and bonnets. The industry prospered around the turn of the century due to the invention of a simple tool for splitting the straw and war, which cut off competition from Italy. At this time women could earn 4s. to 6s. per week plaiting straw. This industry also declined, though, following the increase in free trade with the Continent in the 1820s.

Factories

A defining feature of the Industrial Revolution was the rise of factories, particularly textile factories. Work moved out of the home and into a factory, which used a central power source to run its machines. Water power was used in most of the early factories, but improvements in the steam engine made steam power possible as well. The most dramatic productivity growth occurred in the cotton industry. The invention of James Hargreaves’ spinning jenny (1764), Richard Arkwright’s “throstle” or “water frame” (1769), and Samuel Crompton’s spinning mule (1779, so named because it combined features of the two earlier machines) revolutionized spinning. Britain began to manufacture cotton cloth, and declining prices for the cloth encouraged both domestic consumption and export. Machines also appeared for other parts of the cloth-making process, the most important of which was Edmund Cartwright’s powerloom, which was adopted slowly because of imperfections in the early designs, but was widely used by the 1830s. While cotton was the most important textile of the Industrial Revolution, there were advances in machinery for silk, flax, and wool production as well.7

The advent of new machinery changed the gender division of labor in textile production. Before the Industrial Revolution, women spun yarn using a spinning wheel (or occasionally a distaff and spindle). Men didn’t spin, and this division of labor made sense because women were trained to have more dexterity than men, and because men’s greater strength made them more valuable in other occupations. In contrast to spinning, handloom weaving was done by both sexes, but men outnumbered women. Men monopolized highly skilled preparation and finishing processes such as wool combing and cloth-dressing. With mechanization, the gender division of labor changed. Women used the spinning jenny and water frame, but mule spinning was almost exclusively a male occupation because it required more strength, and because the male mule-spinners actively opposed the employment of female mule-spinners. Women mule-spinners in Glasgow, and their employers, were the victims of violent attacks by male spinners trying to reduce the competition in their occupation.8 While they moved out of spinning, women seem to have increased their employment in weaving (both in handloom weaving and eventually in powerloom factories). Both sexes were employed as powerloom operators.

Table Two

Factory Workers in 1833: Females as a Percent of the Workforce

Industry Ages 12 and under Ages 13-20 Ages 21+ All Ages
Cotton 51.8 65.0 52.2 58.0
Wool 38.6 46.2 37.7 40.9
Flax 54.8 77.3 59.5 67.4
Silk 74.3 84.3 71.3 78.1
Lace 38.7 57.4 16.6 36.5
Potteries 38.1 46.9 27.1 29.4
Dyehouse 0.0 0.0 0.0 0.0
Glass 0.0 0.0 0.0 0.0
Paper - 100.0 39.2 53.6
Whole Sample 52.8 66.4 48.0 56.8

Source: “Report from Dr. James Mitchell to the Central Board of Commissioners, respecting the Returns made from the Factories, and the Results obtained from them.” British Parliamentary Papers, 1834 (167) XIX. Mitchell collected data from 82 cotton factories, 65 wool factories, 73 flax factories, 29 silk factories, 7 potteries, 11 lace factories, one dyehouse, one “glass works”, and 2 paper mills throughout Great Britain.

While the highly skilled and highly paid task of mule-spinning was a male occupation, many women and girls were engaged in other tasks in textile factories. For example, the wet-spinning of flax, introduced in Leeds in 1825, employed mainly teenage girls. Girls often worked as assistants to mule-spinners, piecing together broken threads. In fact, females were a majority of the factory labor force. Table Two shows that 57 percent of factory workers were female, most of them under age 20. Women were widely employed in all the textile industries, and constituted the majority of workers in cotton, flax, and silk. Outside of textiles, women were employed in potteries and paper factories, but not in dye or glass manufacture. Of the women who worked in factories, 16 percent were under age 13, 51 percent were between the ages of 13 and 20, and 33 percent were age 21 and over. On average, girls earned the same wages as boys. Children’s wages rose from about 1s.6d. per week at age 7 to about 5s. per week at age 15. Beginning at age 16, and a large gap between male and female wages appeared. At age 30, women factory workers earned only one-third as much as men.

Figure One

Distribution of Male and Female Factory Employment by Age, 1833

Figure 1

Source: “Report from Dr. James Mitchell to the Central Board of Commissioners, respecting the Returns made from the Factories, and the Results obtained from them.” British Parliamentary Papers, 1834 (167) XIX.

The y-axis shows the percentage of total employment within each sex that is in that five-year age category.

Figure Two

Wages of Factory Workers in 1833

Figure 2

Source: “Report from Dr. James Mitchell to the Central Board of Commissioners, respecting the Returns made from the Factories, and the Results obtained from them.” British Parliamentary Papers, 1834 (167) XIX.

Agriculture

Wage Workers

Wage-earners in agriculture generally fit into one of two broad categories – servants who were hired annually and received part of their wage in room and board, and day-laborers who lived independently and were paid a daily or weekly wage. Before industrialization servants comprised between one-third and one-half of labor in agriculture.9 For servants the value of room and board was a substantial portion of their compensation, so the ratio of money wages is an under-estimate of the ratio of total wages (see Table Three). Most servants were young and unmarried. Because servants were paid part of their wage in kind, as board, the use of the servant contract tended to fall when food prices were high. During the Industrial Revolution the use of servants seems to have fallen in the South and East.10 The percentage of servants who were female also declined in the first half of the nineteenth century.11

Table Three

Wages of Agricultural Servants (£ per year)

Year Location Male Money Wage Male In-Kind Wage Female Money Wage Female In-Kind Wage Ratio of Money Wages Ratio of Total Wages
1770 Lancashire

7

9

3

6

0.43

0.56

1770 Oxfordshire

10

12

4

8

0.40

0.55

1770 Staffordshire

11

9

4

6

0.36

0.50

1821 Yorkshire

16.5

27

7

18

0.42

0.57

Source: Joyce Burnette, “An Investigation of the Female-Male Wage Gap during the Industrial Revolution in Britain,” Economic History Review 50 (May 1997): 257-281.

While servants lived with the farmer and received food and lodging as part of their wage, laborers lived independently, received fewer in-kind payments, and were paid a daily or a weekly wage. Though the majority of laborers were male, some were female. Table Four shows the percentage of laborers who were female at various farms in the late-18th and early-19th centuries. These numbers suggest that female employment was widespread, but varied considerably from one location to the next. Compared to men, female laborers generally worked fewer days during the year. The employment of female laborers was concentrated around the harvest, and women rarely worked during the winter. While men commonly worked six days per week, outside of harvest women generally averaged around four days per week.

Year Location Percent Female
1772-5 Oakes in Norton, Derbyshire

17

1774-7 Dunster Castle Farm, Somerset

27

1785-92 Dunster Castle Farm, Somerset

40

1794-5 Dunster Castle Farm, Somerset

42

1801-3 Dunster Castle Farm, Somerset

35

1801-4 Nettlecombe Barton, Somerset

10

1814-6 Nettlecombe Barton, Somerset

7

1826-8 Nettlecombe Barton, Somerset

5

1828-39 Shipton Moyne, Gloucestershire

19

1831-45 Oakes in Norton, Derbyshire

6

1836-9 Dunster Castle Farm, Somerset

26

1839-40 Lustead, Norfolk

6

1846-9 Dunster Castle Farm, Somerset

29

Sources: Joyce Burnette, “Labourers at the Oakes: Changes in the Demand for Female Day-Laborers at a Farm near Sheffield During the Agricultural Revolution,” Journal of Economic History 59 (March 1999): 41-67; Helen Speechley, Female and Child Agricultural Day Labourers in Somerset, c. 1685-1870, dissertation, Univ. of Exeter, 1999. Sotheron-Estcourt accounts, G.R.O. D1571; Ketton-Cremer accounts, N.R.O. WKC 5/250

The wages of female day-laborers were fairly uniform; generally a farmer paid the same wage to all the adult women he hired. Women’s daily wages were between one-third and one-half of male wages. Women generally worked shorter days, though, so the gap in hourly wages was not quite this large.12 In the less populous counties of Northumberland and Durham, male laborers were required to provide a “bondager,” a woman (usually a family member) who was available for day-labor whenever the employer wanted her.13

Table Five

Wages of Agricultural Laborers

Year Location Male Wage (d./day) Female Wage (d./day) Ratio
1770 Yorkshire 5 12 0.42
1789 Hertfordshire 6 16 0.38
1797 Warwickshire 6 14 0.43
1807 Oxfordshire 9 23 0.39
1833 Cumberland 12 24 0.50
1833 Essex 10 22 0.45
1838 Worcester 9 18 0.50

Source: Joyce Burnette, “An Investigation of the Female-Male Wage Gap during the Industrial Revolution in Britain,” Economic History Review 50 (May 1997): 257-281.

Various sources suggest that women’s employment in agriculture declined during the early nineteenth century. Enclosure increased farm size and changed the patterns of animal husbandry, both of which seem to have led to reductions in female employment.14 More women were employed during harvest than during other seasons, but women’s employment during harvest declined as the scythe replaced the sickle as the most popular harvest tool. While women frequently harvested with the sickle, they did not use the heavier scythe.15 Female employment fell the most in the East, where farms increasingly specialized in grain production. Women had more work in the West, which specialized more in livestock and dairy farming.16

Non-Wage-Earners

During the eighteenth century there were many opportunities for women to be productively employed in farm work on their own account, whether they were wives of farmers on large holdings, or wives of landless laborers. In the early nineteenth century, however, many of these opportunities disappeared, and women’s participation in agricultural production fell.

In a village that had a commons, even if the family merely rented a cottage the wife could be self-employed in agriculture because she could keep a cow, or other animals, on the commons. By careful management of her stock, a woman might earn as much during the year as her husband earned as a laborer. Women also gathered fuel from the commons, saving the family considerable expense. The enclosure of the commons, though, eliminated these opportunities. In an enclosure, land was re-assigned so as to eliminate the commons and consolidate holdings. Even when the poor had clear legal rights to use the commons, these rights were not always compensated in the enclosure agreement. While enclosure occurred at different times for different locations, the largest waves of enclosures occurred in the first two decades of the nineteenth century, meaning that, for many, opportunities for self-employment in agriculture declined as the same time as employment in cottage industry declined. 17

Only a few opportunities for agricultural production remained for the landless laboring family. In some locations landlords permitted landless laborers to rent small allotments, on which they could still grow some of their own food. The right to glean on fields after harvest seems to have been maintained at least through the middle of the nineteenth century, by which time it had become one of the few agricultural activities available to women in some areas. Gleaning was a valuable right; the value of the grain gleaned was often between 5 and 10 percent of the family’s total annual income.18

In the eighteenth century it was common for farmers’ wives to be actively involved in farm work, particularly in managing the dairy, pigs, and poultry. The diary was an important source of income for many farms, and its success depended on the skill of the mistress, who usually ran the operation with no help from men. In the nineteenth century, however, farmer’s wives were more likely to withdraw from farm management, leaving the dairy to the management of dairymen who paid a fixed fee for the use of the cows.19 While poor women withdrew from self-employment in agriculture because of lost opportunities, farmer’s wives seem to have withdraw because greater prosperity allowed them to enjoy more leisure.

It was less common for women to manage their own farms, but not unknown. Commercial directories list numerous women farmers. For example, the 1829 Directory of the County of Derby lists 3354 farmers, of which 162, or 4.8%, were clearly female.20 While the commercial directories themselves do not indicate to what extent these women were actively involved in their farms, other evidence suggests that at least some women farmers were actively involved in the work of the farm.21

Self-Employed

During the Industrial Revolution period women were also active businesswomen in towns. Among business owners listed in commercial directories, about 10 percent were female. Table Seven shows the percentage female in all the trades with at least 25 people listed in the 1788 Manchester commercial directory. Single women, married women, and widows are included in these numbers. Sometimes these women were widows carrying on the businesses of their deceased husbands, but even in this case that does not mean they were simply figureheads. Widows often continued their husband’s businesses because they had been active in management of the business while their husband was alive, and wished to continue.22 Sometimes married women were engaged in trade separately from their husbands. Women most commonly ran shops and taverns, and worked as dressmakers and milliners, but they were not confined to these areas, and appear in most of the trades listed in commercial directories. Manchester, for example, had six female blacksmiths and five female machine makers in 1846. Between 1730 and 1800 there were 121 “rouping women” selling off estates in Edinburgh. 23

Table Six

Business Owners Listed in Commercial Directories

Date City Male Female Unknown Gender Percent Female
1788 Manchester

2033

199

321

8.9

1824-5 Manchester

4185

297

1671

6.6

1846 Manchester

11,942

1222

2316

9.3

1850 Birmingham

15,054

2020

1677

11.8

1850 Derby

2415

332

194

12.1

Sources: Lewis’s Manchester Directory for 1788 (reprinted by Neil Richardson, Manchester, 1984); Pigot and Dean’s Directory for Manchester, Salford, &c. for 1824-5 (Manchester 1825); Slater’s National Commercial Directory of Ireland (Manchester, 1846); Slater’s Royal National and Commercial Directory (Manchester, 1850)

Table Seven

Women in Trades in Manchester, 1788

Trade Men Women Gender Unknown Percent Female
Apothecary/ Surgeon / Midwife

29

1

5

3.3

Attorney

39

0

3

0.0

Boot and Shoe makers

87

0

1

0.0

Butcher

33

1

1

2.9

Calenderer

31

4

5

11.4

Corn & Flour Dealer

45

4

5

8.2

Cotton Dealer

23

0

2

0.0

Draper, Mercer, Dealer of Cloth

46

15

19

24.6

Dyer

44

3

18

6.4

Fustian Cutter / Shearer

54

2

0

3.6

Grocers & Tea Dealers

91

16

12

15.0

Hairdresser & Peruke maker

34

1

0

2.9

Hatter

45

3

4

6.3

Joiner

34

0

1

0.0

Liquor dealer

30

4

14

11.8

Manufacturer, cloth

257

4

118

1.5

Merchant

58

1

18

1.7

Publichouse / Inn / Tavern

126

13

2

9.4

School master / mistress

18

10

0

35.7

Shopkeeper

107

16

4

13.0

Tailor

59

0

1

0.0

Warehouse

64

0

14

0.0

Source: Lewis’s Manchester Directory for 1788 (reprinted by Neil Richardson, Manchester, 1984)

Guilds often controlled access to trades, admitting only those who had served an apprenticeship and thus earned the “freedom” of the trade. Women could obtain “freedom” not only by apprenticeship, but also by widowhood. The widow of a tradesman was often considered knowledgeable enough in the trade that she was given the right to carry on the trade even without an apprenticeship. In the eighteenth century women were apprenticed to a wide variety of trades, including butchery, bookbinding, brush making, carpentry, ropemaking and silversmithing.24 Between the eighteenth and nineteenth centuries the number of females apprenticed to trades declined, possibly suggesting reduced participation by women. However, the power of the guilds and the importance of apprenticeship were also declining during this time, so the decline in female apprenticeships may not have been an important barrier to employment.25

Many women worked in the factories of the Industrial Revolution, and a few women actually owned factories. In Keighley, West Yorkshire, Ann Illingworth, Miss Rachael Leach, and Mrs. Betty Hudson built and operated textile mills.26 In 1833 Mrs. Doig owned a powerloom factory in Scotland, which employed 60 workers.27

While many women did successfully enter trades, there were obstacles to women’s employment that kept their numbers low. Women generally received less education than men (though education of the time was of limited practical use). Women may have found it more difficult than men to raise the necessary capital because English law did not consider a married woman to have any legal existence; she could not sue or be sued. A married woman was a feme covert and technically could not make any legally binding contracts, a fact which may have discouraged others from loaning money to or making other contracts with married women. However, this law was not as limiting in practice as it would seem to be in theory because a married woman engaged in trade on her own account was treated by the courts as a feme sole and was responsible for her own debts.28

The professionalization of certain occupations resulted in the exclusion of women from work they had previously done. Women had provided medical care for centuries, but the professionalization of medicine in the early-nineteenth century made it a male occupation. The Royal College of Physicians admitted only graduates of Oxford and Cambridge, schools to which women were not admitted until the twentieth century. Women were even replaced by men in midwifery. The process began in the late-eighteenth century, when we observe the use of the term “man-midwife,” an oxymoronic title suggestive of changing gender roles. In the nineteenth century the “man-midwife” disappeared, and women were replaced by physicians or surgeons for assisting childbirth. Professionalization of the clergy was also effective in excluding women. While the Church of England did not allow women ministers, the Methodists movement had many women preachers during its early years. However, even among the Methodists female preachers disappeared when lay preachers were replaced with a professional clergy in the early nineteenth century.29

In other occupations where professionalization was not as strong, women remained an important part of the workforce. Teaching, particularly in the lower grades, was a common profession for women. Some were governesses, who lived as household servants, but many opened their own schools and took in pupils. The writing profession seems to have been fairly open to women; the leading novelists of the period include Jane Austen, Charlotte and Emily Brontë, Fanny Burney, George Eliot (the pen name of Mary Ann Evans), Elizabeth Gaskell, and Frances Trollope. Female non-fiction writers of the period include Jane Marcet, Hannah More, and Mary Wollstonecraft.

Other Occupations

The occupations listed above are by no means a complete listing of the occupations of women during the Industrial Revolution. Women made buttons, nails, screws, and pins. They worked in the tin plate, silver plate, pottery and Birmingham “toy” trades (which made small articles like snuff boxes). Women worked in the mines until The Mines Act of 1842 prohibited them from working underground, but afterwards women continued to pursue above-ground mining tasks.

Married Women in the Labor Market

While there are no comprehensive sources of information on the labor force participation of married women, household budgets reported by contemporary authors give us some information on women’s participation.30 For the period 1787 to 1815, 66 percent of married women in working-class households had either a recorded occupation or positive earnings. For the period 1816-20 the rate fell to 49 percent, but in 1821-40 it recovered to 62 percent. Table Eight gives participation rates of women by date and occupation of the husband.

Table Eight

Participation Rates of Married Women

 

High-Wage Agriculture

Low-Wage Agriculture

Mining

Factory

Outwork

Trades

All

1787-1815

55

85

40

37

46

63

66

1816-1820

34

NA

28

4

42

30

49

1821-1840

22

85

33

86

54

63

62

Source: Sara Horrell and Jane Humphries, “Women’s Labour Force Participation and the Transition to the male-Breadwinner Family, 1790-1865,” Economic History Review 48 (February 1995): 89-117

While many wives worked, the amount of their earnings was small relative to their husband’s earnings. Annual earnings of married women who did work averaged only about 28 percent of their husband’s earnings. Because not all women worked, and because children usually contributed more to the family budget than their mothers, for the average family the wife contributed only around seven percent of total family income.

Childcare

Women workers used a variety of methods to care for their children. Sometimes childcare and work were compatible, and women took their children with them to the fields or shops where they worked.31 Sometimes women working at home would give their infants opiates such as “Godfrey’s Cordial” in order to keep the children quiet while their mothers worked.32 The movement of work into factories increased the difficulty of combining work and childcare. In most factory work the hours were rigidly set, and women who took the jobs had to accept the twelve or thirteen hour days. Work in the factories was very disciplined, so the women could not bring their children to the factory, and could not take breaks at will. However, these difficulties did not prevent women with small children from working.

Nineteenth-century mothers used older siblings, other relatives, neighbors, and dame schools to provide child care while they worked.33 Occasionally mothers would leave young children home alone, but this was dangerous enough that only a few did so.34 Children as young as two might be sent to dame schools, in which women would take children into their home and provide child care, as well as some basic literacy instruction.35 In areas where lace-making or straw-plaiting thrived, children were sent from about age seven to “schools” where they learned the trade.36

Mothers might use a combination of different types of childcare. Elizabeth Wells, who worked in a Leicester worsted factory, had five children, ages 10, 8, 6, 2, and four months. The eldest, a daughter, stayed home to tend the house and care for the infant. The second child worked, and the six-year-old and two-year-old were sent to “an infant school.”37 Mary Wright, an “over-looker” in the rag-cutting room of a Buckinghamshire paper factory, had five children. The eldest worked in the rag-cutting room with her, the youngest was cared for at home, and the middle three were sent to a school; “for taking care of an infant she pays 1s.6d. a-week, and 3d. a-week for the three others. They go to a school, where they are taken care of and taught to read.”38

The cost of childcare was substantial. At the end of the eighteenth century the price of child-care was about 1s. a week, which was about a quarter of a woman’s weekly earnings in agriculture.39 In the 1840s mothers paid anywhere from 9d. to 2s.6d. per week for child care, out of a wage of around 7s. per week.40

For Further Reading

Burnette, Joyce. “An Investigation of the Female-Male Wage Gap during the Industrial Revolution in Britain.” Economic History Review 50 (1997): 257-281.

Davidoff, Leonore, and Catherine Hall. Family Fortunes: Men and Women of the English Middle Class, 1780-1850. Chicago: University of Chicago Press, 1987.

Honeyman, Katrina. Women, Gender and Industrialisation in England, 1700-1870. New York: St. Martin’s Press, 2000.

Horrell, Sara, and Jane Humphries. “Women’s Labour Force Participation and the Transition to the Male-Breadwinner Family, 1790-1865.” Economic History Review 48 (1995): 89-117.

Humphries, Jane. “Enclosures, Common Rights, and Women: The Proletarianization of Families in the Late Eighteenth and Early Nineteenth Centuries.” Journal of Economic History 50 (1990): 17-42.

King, Peter. “Customary Rights and Women’s Earnings: The Importance of Gleaning to the Rural Labouring Poor, 1750-1850.” Economic History Review 44 (1991): 461-476

Kussmaul, Ann. Servants in Husbandry in Early Modern England. Cambridge: Cambridge University Press, 1981.

Pinchbeck, Ivy. Women Workers and the Industrial Revolution, 1750-1850, London: Routledge, 1930.

Sanderson, Elizabeth. Women and Work in Eighteenth-Century Edinburgh. New York: St. Martin’s Press, 1996.

Snell, K.D.M. Annals of the Labouring Poor: Social Change and Agrarian England, 1660-1900. Cambridge: Cambridge University Press, 1985.

Valenze, Deborah. Prophetic Sons and Daughters: Female Preaching and Popular Religion in Industrial England. Princeton University Press, 1985.

Valenze, Deborah. The First Industrial Woman. Oxford: Oxford University Press, 1995.

1 “Since large-scale industry has transferred the woman from the house to the labour market and the factory, and makes her, often enough, the bread-winner of the family, the last remnants of male domination in the proletarian home have lost all foundation – except, perhaps, for some of that brutality towards women which became firmly rooted with the establishment of monogamy. . . .It will then become evidence that the first premise for the emancipation of women is the reintroduction of the entire female sex into public industry.” Frederick Engels, The Origin of the Family, Private Property and the State, in Karl Marx and Frederick Engels: Selected Works, New York: International Publishers, 1986, p. 508, 510.

2 Ivy Pinchbeck (Women Workers and the Industrial Revolution, Routledge, 1930) claimed that higher incomes allowed some women to withdraw from the labor force. While she saw some disadvantages resulting from this withdrawal, particularly the loss of independence, she thought that overall women benefited from having more time to devote to their homes and families. Davidoff and Hall (Family Fortunes: Man and Women of the English Middle Class, 1780-1850, Univ. of Chicago Press, 1987) agree that women withdrew from work, but they see the change as a negative result of gender discrimination. Similarly, Horrell and Humphries (“Women’s Labour Force Participation and the Transition to the Male-Breadwinner Family, 1790-1865,” Economic History Review, Feb. 1995, XLVIII:89-117) do not find that rising incomes caused declining labor force participation, and they believe that declining demand for female workers caused the female exodus from the workplace.

3 While the British census began in 1801, individual enumeration did not begin until 1841. For a detailed description of the British censuses of the nineteenth century, see Edward Higgs, Making Sense of the Census, London: HMSO, 1989.

4 For example, Helen Speechley, in her dissertation, showed that seven women who worked for wages at a Somerset farm had no recorded occupation in the 1851 census See Helen Speechley, Female and Child Agricultural Day Labourers in Somerset, c. 1685-1870, dissertation, Univ. of Exeter, 1999.

5 Edward Higgs finds that removing family members from the “servants” category reduced the number of servants in Rochdale in 1851. Enumerators did not clearly distinguish between the terms “housekeeper” and “housewife.” See Edward Higgs, “Domestic Service and Household Production” in Angela John, ed., Unequal Opportunities, Oxford: Basil Blackwell, and “Women, Occupations and Work in the Nineteenth Century Censuses,” History Workshop, 1987, 23:59-80. In contrast, the censuses of the early 20th century seem to be fairly accurate; see Tim Hatton and Roy Bailey, “Women’s Work in Census and Survey, 1911-1931,” Economic History Review, Feb. 2001, LIV:87-107.

6 A shilling was equal to 12 pence, so if women earned 2s.6d. for 20 hours, they earned 1.5d. per hour. Women agricultural laborers earned closer to 1d. per hour, so the London wage was higher. See Dorothy George, London Life in the Eighteenth-Century, London: Kegan Paul, Trench, Trubner & Co., 1925, p. 208, and Patricia Malcolmson, English Laundresses, Univ. of Illinois Press, 1986, p. 25. .

7 On the technology of the Industrial Revolution, see David Landes, The Unbound Prometheus, Cambridge Univ. Press, 1969, and Joel Mokyr, The Lever of Riches, Oxford Univ. Press, 1990.

8 A petition from Glasgow cotton manufactures makes the following claim, “In almost every department of the cotton spinning business, the labour of women would be equally efficient with that of men; yet in several of these departments, such measures of violence have been adopted by the combination, that the women who are willing to be employed, and who are anxious by being employed to earn the bread of their families, have been driven from their situations by violence. . . . Messrs. James Dunlop and Sons, some years ago, erected cotton mills in Calton of Glasgow, on which they expended upwards of [£]27,000 forming their spinning machines, (Chiefly with the view of ridding themselves of the combination [the male union],) of such reduced size as could easily be wrought by women. They employed women alone, as not being parties to the combination, and thus more easily managed, and less insubordinate than male spinners. These they paid at the same rate of wages, as were paid at other works to men. But they were waylaid and attacked, in going to, and returning from their work; the houses in which they resided, were broken open in the night. The women themselves were cruelly beaten and abused; and the mother of one of them killed; . . . And these nefarious attempts were persevered in so systematically, and so long, that Messrs. Dunlop and sons, found it necessary to dismiss all female spinners from their works, and to employ only male spinners, most probably the very men who had attempted their ruin.” First Report from the Select Committee on Artizans and Machinery, British Parliamentary Papers, 1824 vol. V, p. 525.

9 Ann Kussmaul, Servants in Husbandry in Early Modern England, Cambridge Univ. Press, 1981, Ch. 1

10 See Ivy Pinchbeck, Women Workers and the Industrial Revolution, Routledge, 1930, Ch. 1, and K.D.M. Snell, Annals of the Labouring Poor, Cambridge Univ. Press, 1985, Ch. 2.

11 For the period 1574 to 1821 about 45 percent of servants were female, but this fell to 32 percent in 1851. See Ann Kussmaul, Servants in Husbandry in Early Modern England, Cambridge Univ. Press, 1981, Ch. 1.

12 Men usually worked 12-hour days, and women averaged closer to 10 hours. See Joyce Burnette, “An Investigation of the Female-Male Wage Gap during the Industrial Revolution in Britain,” Economic History Review, May 1997, 50:257-281.

13 See Ivy Pinchbeck, Women Workers and the Industrial Revolution, Routledge, 1930, p. 65.

14 See Robert Allen, Enclosure and the Yeoman, Clarendon Press, 1992, and Joyce Burnette, “Labourers at the Oakes: Changes in the Demand for Female Day-Laborers at a Farm near Sheffield During the Agricultural Revolution,” Journal of Economics History, March 1999, 59:41-67.

15 While the scythe had been used for mowing grass for hay or cheaper grains for some time, the sickle was used for harvesting wheat until the nineteenth century. Thus adoption of the scythe for harvesting wheat seems to be a response to changing prices rather than invention of a new technology. The scythe required less labor to harvest a given acre, but left more grain on the ground, so as grain prices fell relative to wages, farmers substituted the scythe for the sickle. See E.J.T. Collins, “Harvest Technology and Labour Supply in Britain, 1790-1870,” Economic History Review, Dec. 1969, XXIII:453-473.

16 K.D.M. Snell, Annals of the Labouring Poor, Cambridge, 1985.

17 See Jane Humphries, “Enclosures, Common Rights, and Women: The Proletarianization of Families in the Late Eighteenth and Early Nineteenth Centuries,” Journal of Economic History, March 1990, 50:17-42, and J.M. Neeson, Commoners: Common Rights, Enclosure and Social Change in England, 1700-1820, Cambridge Univ. Press, 1993.

18 See Peter King, “Customary Rights and Women’s Earnings: The Importance of Gleaning to the Rural Labouring Poor, 1750-1850,” Economic History Review, 1991, XLIV:461-476.

19 Pinchbeck, Women Workers and the Industrial Revolution, Routledge, 1930, p. 41-42 See also Deborah Valenze, The First Industrial Woman, Oxford Univ. Press, 1995

20 Stephen Glover, The Directory of the County of Derby, Derby: Henry Mozley and Son, 1829.

21 Eden gives an example of gentlewomen who, on the death of their father, began to work as farmers. He notes, “not seldom, in one and the same day, they have divided their hours in helping to fill the dung-cart, and receiving company of the highest rank and distinction.” (F.M. Eden, The State of the Poor, vol. i., p. 626.) One woman farmer who was clearly an active manager celebrated her success in a letter sent to the Annals of Agriculture, (quoted by Pinchbeck, Women Workers and the Industrial Revolution, Routledge, 1930, p. 30): “I bought a small estate, and took possession of it in the month of July, 1803. . . . As a woman undertaking to farm is generally a subject of ridicule, I bought the small estate by way of experiment: the gentlemen of the county have now complimented me so much on having set so good and example to the farmers, that I have determined on taking a very large farm into my hands.” The Annals of Agriculture give a number of examples of women farmers cited for their experiments or their prize-winning crops.

22 Tradesmen considered themselves lucky to find a wife who was good at business. In his autobiography James Hopkinson, a cabinetmaker, said of his wife, “I found I had got a good and suitable companion one with whom I could take sweet council and whose love and affections was only equall’d by her ability as a business woman.” Victorian Cabinet Maker: The Memoirs of James Hopkinson, 1819-1894, 1968, p. 96.

23 See Elizabeth Sanderson, Women and Work in Eighteenth-Century Edinburgh, St. Martin’s Press, 1996.

24 See K.D.M. Snell, Annals of the Labouring Poor, Cambridge Univ. Press, 1985, Table 6.1.

25 The law requiring a seven-year apprenticeship before someone could work in a trade was repealed in 1814.

26 See Francois Crouzet, The First Industrialists, Cambridge Univ. Press, 1985, and M.L. Baumber, From Revival to Regency: A History of Keighley and Haworth, 1740-1820, Crabtree Ltd., Keighley, 1983.

27 First Report of the Central Board of His Majesty’s Commissioners for inquiry into the Employment of Children in Factories, with Minutes of Evidence, British Parliamentary Papers, 1833 (450) XX, A1, p. 120. \

28 For example, in the case of “LaVie and another Assignees against Philips and another Assignees,” the court upheld the right of a woman to operate as feme sole. In 1764 James Cox and his wife Jane were operating separate businesses, and both went bankrupt within the space of two months. Jane’s creditors sued James’s creditors for the recovery of five fans, goods from her shop that had been taken for James’s debts. The court ruled that, since Jane was trading as a feme sole, her husband did not own the goods in her shop, and thus James’s creditors had no right to seize them. See William Blackstone, Reports of Cases determined in the several Courts of Westminster-Hall, from 1746 to 1779, London, 1781, p. 570-575.

29 See Deborah Valenze, Prophetic Sons and Daughters: Female Preaching and Popular Religion in Industrial England, Princeton Univ. Press, 1985.

30 See Sara Horrell and Jane Humphries, “Women’s Labour Force Participation and the Transition to the male-Breadwinner Family, 1790-1865,” Economic History Review, Feb. 1995, XLVIII:89-117.

31 In his autobiography James Hopkinson says of his wife, “How she laboured at the press and assisted me in the work of my printing office, with a child in her arms, I have no space to tell, nor in fact have I space to allude to the many ways she contributed to my good fortune.” James Hopkinson, Victorian Cabinet Marker: The Memoirs of James Hopkinson, 1819-1894, J.B. Goodman, ed., Routledge & Kegan Paul, 1968, p. 96. A 1739 poem by Mary Collier suggests that carrying babies into the field was fairly common; it contains these lines:

Our tender Babes into the Field we bear,

And wrap them in our Cloaths to keep them warm,

While round about we gather up the Corn;

. . .

When Night comes on, unto our Home we go,

Our Corn we carry, and our Infant too.

Mary Collier, The Woman’s Labour, Augustan Reprint Society, #230, 1985, p. 10. A 1835 Poor Law report stated that in Sussex, “the custom of the mother of a family carrying her infant with her in its cradle into the field, rather than lose the opportunity of adding her earnings to the general stock, though partially practiced before, is becoming very much more general now.” (Quoted in Pinchbeck, Women Workers and the Industrial Revolution, Routledge, 1930, p. 85.)

32 Sarah Johnson of Nottingham claimed that she ” Knows it is quite a common custom for mothers to give Godfrey’s and the Anodyne cordial to their infants, ‘it is quite too common.’ It is given to infants at the breast; it is not given because the child is ill, but ‘to compose it to rest, to sleep it,’ so that the mother may get to work. ‘Has seen an infant lay asleep on its mother’s lap whilst at the lace-frame for six or eight hours at a time.’ This has been from the effects of the cordial.” [Reports from Assistant Handloom-Weavers’ Commissioners, British Parliamentary Papers, 1840 (43) XXIII, p. 157] Mary Colton, a lace worker from Nottingham, described her use of the drug to parliamentary investigators thus: ‘Was confined of an illegitimate child in November, 1839. When the child was a week old she gave it a half teaspoonful of Godfrey’s twice a-day. She could not afford to pay for the nursing of the child, and so gave it Godfrey’s to keep it quiet, that she might not be interrupted at the lace piece; she gradually increased the quantity by a drop or two at a time until it reached a teaspoonful; when the infant was four months old it was so “wankle” and thin that folks persuaded her to give it laudanum to bring it on, as it did other children. A halfpenny worth, which was about a teaspoonful and three-quarters, was given in two days; continued to give her this quantity since February, 1840, until this last past (1841), and then reduced the quantity. She now buys a halfpenny worth of laudanum and a halfpenny worth of Godfrey’s mixed, which lasts her three days. . . . If it had not been for her having to sit so close to work she would never have given the child Godfrey’s. She has tried to break it off many times but cannot, for if she did, she should not have anything to eat.” [Children’s Employment Commission: Second Report of the Commissioners (Trades and Manufactures), British Parliamentary
Papers
, 1843 (431) XIV, p. 630].

33 Elizabeth Leadbeater, who worked for a Birmingham brass-founder, worked while she was nursing and had her mother look after the infant. [Children’s Employment Commission: Second Report of the Commissioners (Trades and Manufactures), British Parliamentary Papers, 1843 (431) XIV, p. 710.] Mrs. Smart, an agricultural worker from Calne, Wiltshire, noted, “Sometimes I have had my mother, and sometimes my sister, to take care of the children, or I could not have gone out.” [Reports of Special Assistant Poor Law Commissioners on the Employment of Women and Children in Agriculture, British Parliamentary Papers, 1843 (510) XII, p. 65.] More commonly, though, older siblings provided the childcare. “Older siblings” generally meant children of nine or ten years old, and included boys as well as girls. Mrs. Britton of Calne, Wiltshire, left her children in the care of her eldest boy. [Reports of Special Assistant Poor Law Commissioners on the Employment of Women and Children in Agriculture, British Parliamentary Papers, 1843 (510) XII, p. 66] In a family from Presteign, Wales, containing children aged 9, 7, 5, 3, and 1, we find that “The oldest children nurse the youngest.” [F.M. Eden, State of the Poor, London: Davis, 1797, vol. iii, p. 904] When asked what income a labourer’s wife and children could earn, some respondents to the 1833 “Rural Queries” assumed that the eldest child would take care of the others, leaving the mother free to work. The returns from Bengeworth, Worcester, report that, “If the Mother goes to field work, the eldest Child had need to stay at home, to tend the younger branches of the Family.” Ewhurst, Surrey, reported that “If the Mother were employed, the elder Children at home would probably be required to attend to the younger Children.” [Report of His Majesty’s Commissioners for Inquiry in the Administration and Practical Operation of the Poor Law, Appendix B,
“Rural Queries,” British Parliamentary Papers, 1834 (44) XXX, p. 488 and 593]

34 Parents heard of incidents, such as one reported in the Times (Feb. 6, 1819):

A shocking accident occurred at Llandidno, near Conway, on Tuesday night, during the absence of a miner and his wife, who had gone to attend a methodist meeting, and locked the house door, leaving two children within; the house by some means took fire, and was, together with the unfortunate children, consumed to ashes; the eldest only four years old!

Mothers were aware of these dangers. One mother who admitted to leaving her children at home worried greatly about the risks:

I have always left my children to themselves, and, God be praised! nothing has ever happened to them, though I thought it dangerous. I have many a time come home, and have thought it a mercy to find nothing has happened to them. . . . Bad accidents often happen. [Reports of Special Assistant Poor Law Commissioners on the Employment of Women and Children in Agriculture, British Parliamentary Papers, 1843 (510) XII, p. 68.]

Leaving young children home without child care had real dangers, and the fact that most working mothers paid for childcare suggests that they did not consider leaving young children alone to be an acceptable option.

35 In 1840 an observer of Spitalfields noted, “In this neighborhood, where the women as well as the men are employed in the manufacture of silk, many children are sent to small schools, not for instruction, but to be taken care of whilst their mothers are at work.”[ Reports from Assistant Handloom-Weavers’ Commissioners, British Parliamentary Papers, 1840 (43) XXIII, p. 261] In 1840 the wife of a Gloucester weaver earned 2s. a week from running a school; she had twelve students and charged each 2d. a week. [Reports from Assistant Handloom Weavers’ Commissioners, British Parliamentary Papers, 1840 (220) XXIV, p. 419] In 1843 the lace-making schools of the midlands generally charged 3d. per week. [Children’s Employment Commission: Second Report of the Commissioners (Trades and Manufactures), British Parliamentary Papers, 1843 (431) XIV, p. 46, 64, 71, 72]

36 At one straw-plaiting school in Hertfordshire,

Children commence learning the trade about seven years old: parents pay 3d. a-week for each child, and for this they are taught the trade and taught to read. The mistress employs about from 15 to 20 at work in a room; the parents get the profits of the children’s labour.[ Children’s Employment Commission: Second Report of the Commissioners (Trades and Manufactures), British Parliamentary Papers, 1843 (431) XIV, p. 64]

At these schools there was very little instruction; some time was devoted to teaching the children to read, but they spent most of their time working. One mistress complained that the children worked too much and learned too little, “In my judgment I think the mothers task the children too much; the mistress is obliged to make them perform it, otherwise they would put them to other schools.” Ann Page of Newport Pagnell, Buckinghamshire, had “eleven scholars” and claimed to “teach them all reading once a-day.” [Children’s Employment Commission: Second Report of the Commissioners (Trades and Manufactures), British Parliamentary Papers, 1843 (431) XIV, p. 66, 71] The standard rate of 3d. per week seems to have been paid for supervision of the children rather than for the instruction.

37 First Report of the Central Board of His Majesty’s Commissioners for Inquiring into the Employment of Children in Factories, British Parliamentary Papers, 1833 (450) XX, C1 p. 33.

38 Children’s Employment Commission: Second Report of the Commissioners (Trades and Manufactures), British Parliamentary Papers, 1843 (431) XIV, p. 46.

39 David Davies, The Case of Labourers in Husbandry Stated and Considered, London: Robinson, 1795, p.14. Agricultural wages for this time period are found in Eden, State of the Poor, London: Davis, 1797.

40 In 1843 parliamentary investigator Alfred Austin reports, “Where a girl is hired to take care of children, she is paid about 9d. a week, and has her food besides, which is a serious deduction from the wages of the woman at work.”[ Reports of Special Assistant Poor Law Commissioners on the Employment of Women and Children in Agriculture, British Parliamentary Papers,1843 (510) XII, p.26] Agricultural wages in the area were 8d. per day, so even without the cost of food, the cost of child care was about one-fifth a woman’s wage. One Scottish woman earned 7s. per week in a coal mine and paid 2s.6d., or 36 percent of her income, for the care of her children.[ B.P.P. 1844 (592) XVI, p. 6] In 1843 Mary Wright, a “over-looker” at a Buckinghamshire paper factory, paid even more for child care; she told parliamentary investigators that “for taking care of an infant she pays 1s.6d. a-week, and 3d. a-week for three others.” [Children’s Employment Commission: Second Report of the Commissioners (Trades and Manufactures), British Parliamentary Papers, 1843 (431) XIV, p. 46] She earned 10s.6d. per week, so her total child-care payments were 21 percent of her wage. Engels put the cost of child care at 1s. or 18d. a week. [Engels, [1845] 1926, p. 143] Factory workers often made 7s. a week, so again these women may have paid around one-fifth of their earnings for child care. Some estimates suggest even higher fractions of women’s income went to child care. The overseer of Wisbech, Cambridge, suggests a higher fraction; he reports, “The earnings of the Wife we consider comparatively small, in cases where she has a large family to attend to; if she has one or two children, she has to pay half, or perhaps more of her earnings for a person to take care of them.” [Report of His Majesty’s Commissioners for Inquiry in the Administration and Practical Operation of the Poor Law, Appendix B, “Rural Queries,”
British Parliamentary Papers, 1834 (44) XXX, p. 76
]

The Economics of the American Revolutionary War

Ben Baack, Ohio State University

By the time of the onset of the American Revolution, Britain had attained the status of a military and economic superpower. The thirteen American colonies were one part of a global empire generated by the British in a series of colonial wars beginning in the late seventeenth century and continuing on to the mid eighteenth century. The British military establishment increased relentlessly in size during this period as it engaged in the Nine Years War (1688-97), the War of Spanish Succession (1702-13), the War of Austrian Succession (1739-48), and the Seven Years War (1756-63). These wars brought considerable additions to the British Empire. In North America alone the British victory in the Seven Years War resulted in France ceding to Britain all of its territory east of the Mississippi River as well as all of Canada and Spain surrendering its claim to Florida (Nester, 2000).

Given the sheer magnitude of the British military and its empire, the actions taken by the American colonists for independence have long fascinated scholars. Why did the colonists want independence? How were they able to achieve a victory over what was at the time the world’s preeminent military power? What were the consequences of achieving independence? These and many other questions have engaged the attention of economic, legal, military, political, and social historians. In this brief essay we will focus only on the economics of the Revolutionary War.

Economic Causes of the Revolutionary War

Prior to the conclusion of the Seven Years War there was little, if any, reason to believe that one day the American colonies would undertake a revolution in an effort to create an independent nation-state. As apart of the empire the colonies were protected from foreign invasion by the British military. In return, the colonists paid relatively few taxes and could engage in domestic economic activity without much interference from the British government. For the most part the colonists were only asked to adhere to regulations concerning foreign trade. In a series of acts passed by Parliament during the seventeenth century the Navigation Acts required that all trade within the empire be conducted on ships which were constructed, owned and largely manned by British citizens. Certain enumerated goods whether exported or imported by the colonies had to be shipped through England regardless of the final port of destination.

Western Land Policies

The movement for independence arose in the colonies following a series of critical decisions made by the British government after the end of the war with France in 1763. Two themes emerge from what was to be a fundamental change in British economic policy toward the American colonies. The first involved western land. With the acquisition from the French of the territory between the Allegheny Mountains and the Mississippi River the British decided to isolate the area from the rest of the colonies. Under the terms of the Proclamation of 1763 and the Quebec Act of 1774 colonists were not allowed to settle here or trade with the Indians without the permission of the British government. These actions nullified the claims to land in the area by a host of American colonies, individuals, and land companies. The essence of the policy was to maintain British control of the fur trade in the West by restricting settlement by the Americans.

Tax Policies

The second fundamental change involved taxation. The British victory over the French had come at a high price. Domestic taxes had been raised substantially during the war and total government debt had increased nearly twofold (Brewer, 1989). Furthermore, the British had decided in1763 to place a standing army of 10,000 men in North America. The bulk of these forces were stationed in newly acquired territory to enforce its new land policy in the West. Forts were to be built which would become the new centers of trade with the Indians. The British decided that the Americans should share the costs of the military buildup in the colonies. The reason seemed obvious. Taxes were significantly higher in Britain than in the colonies. One estimate suggests the per capita tax burden in the colonies ranged from two to four per cent of that in Britain (Palmer, 1959). It was time in the British view that the Americans began to pay a larger share of the expenses of the empire.

Accordingly, a series of tax acts were passed by Parliament the revenue from which was to be used to help pay for the standing army in America. The first was the Sugar Act of 1764. Proposed by England’s Prime Minister the act lowered tariff rates on non-British products from the West Indies as well as strengthened their collection. It was hoped this would reduce the incentive for smuggling and thereby increase tariff revenue (Bullion, 1982). The following year Parliament passed the Stamp Act that imposed a tax commonly used in England. It required stamps for a broad range of legal documents as well as newspapers and pamphlets. While the colonial stamp duties were less than those in England they were expected to generate enough revenue to finance a substantial portion of the cost the new standing army. The same year passage of the Quartering Act imposed essentially a tax in kind by requiring the colonists to provide British military units with housing, provisions, and transportation. In 1767 the Townshend Acts imposed tariffs upon a variety of imported goods and established a Board of Customs Commissioners in the colonies to collect the revenue.

Boycotts

American opposition to these acts was expressed initially in a variety of peaceful forms. While they did not have representation in Parliament, the colonists did attempt to exert some influence in it through petition and lobbying. However, it was the economic boycott that became by far the most effective means of altering the new British economic policies. In 1765 representatives from nine colonies met at the Stamp Act Congress in New York and organized a boycott of imported English goods. The boycott was so successful in reducing trade that English merchants lobbied Parliament for the repeal of the new taxes. Parliament soon responded to the political pressure. During 1766 it repealed both the Stamp and Sugar Acts (Johnson, 1997). In response to the Townshend Acts of 1767 a second major boycott started in 1768 in Boston and New York and subsequently spread to other cities leading Parliament in 1770 to repeal all of the Townshend duties except the one on tea. In addition, Parliament decided at the same time not to renew the Quartering Act.

With these actions taken by Parliament the Americans appeared to have successfully overturned the new British post war tax agenda. However, Parliament had not given up what it believed to be its right to tax the colonies. On the same day it repealed the Stamp Act, Parliament passed the Declaratory Act stating the British government had the full power and authority to make laws governing the colonies in all cases whatsoever including taxation. Policies not principles had been overturned.

The Tea Act

Three years after the repeal of the Townshend duties British policy was once again to emerge as an issue in the colonies. This time the American reaction was not peaceful. It all started when Parliament for the first time granted an exemption from the Navigation Acts. In an effort to assist the financially troubled British East India Company Parliament passed the Tea Act of 1773, which allowed the company to ship tea directly to America. The grant of a major trading advantage to an already powerful competitor meant a potential financial loss for American importers and smugglers of tea. In December a small group of colonists responded by boarding three British ships in the Boston harbor and throwing overboard several hundred chests of tea owned by the East India Company (Labaree, 1964). Stunned by the events in Boston, Parliament decided not to cave in to the colonists as it had before. In rapid order it passed the Boston Port Act, the Massachusetts Government Act, the Justice Act, and the Quartering Act. Among other things these so-called Coercive or Intolerable Acts closed the port of Boston, altered the charter of Massachusetts, and reintroduced the demand for colonial quartering of British troops. Once done Parliament then went on to pass the Quebec Act as a continuation of its policy of restricting the settlement of the West.

The First Continental Congress

Many Americans viewed all of this as a blatant abuse of power by the British government. Once again a call went out for a colonial congress to sort out a response. On September 5, 1774 delegates appointed by the colonies met in Philadelphia for the First Continental Congress. Drawing upon the successful manner in which previous acts had been overturned the first thing Congress did was to organize a comprehensive embargo of trade with Britain. It then conveyed to the British government a list of grievances that demanded the repeal of thirteen acts of Parliament. All of the acts listed had been passed after 1763 as the delegates had agreed not to question British policies made prior to the conclusion of the Seven Years War. Despite all the problems it had created, the Tea Act was not on the list. The reason for this was that Congress decided not to protest British regulation of colonial trade under the Navigation Acts. In short, the delegates were saying to Parliament take us back to 1763 and all will be well.

The Second Continental Congress

What happened then was a sequence of events that led to a significant increase in the degree of American resistance to British polices. Before the Congress adjourned in October the delegates voted to meet again in May of 1775 if Parliament did not meet their demands. Confronted by the extent of the American demands the British government decided it was time to impose a military solution to the crisis. Boston was occupied by British troops. In April a military confrontation occurred at Lexington and Concord. Within a month the Second Continental Congress was convened. Here the delegates decided to fundamentally change the nature of their resistance to British policies. Congress authorized a continental army and undertook the purchase of arms and munitions. To pay for all of this it established a continental currency. With previous political efforts by the First Continental Congress to form an alliance with Canada having failed, the Second Continental Congress took the extraordinary step of instructing its new army to invade Canada. In effect, these actions taken were those of an emerging nation-state. In October as American forces closed in on Quebec the King of England in a speech to Parliament declared that the colonists having formed their own government were now fighting for their independence. It was to be only a matter of months before Congress formally declared it.

Economic Incentives for Pursuing Independence: Taxation

Given the nature of British colonial policies, scholars have long sought to evaluate the economic incentives the Americans had in pursuing independence. In this effort economic historians initially focused on the period following the Seven Years War up to the Revolution. It turned out that making a case for the avoidance of British taxes as a major incentive for independence proved difficult. The reason was that many of the taxes imposed were later repealed. The actual level of taxation appeared to be relatively modest. After all, the Americans soon after adopting the Constitution taxed themselves at far higher rates than the British had prior to the Revolution (Perkins, 1988). Rather it seemed the incentive for independence might have been the avoidance of the British regulation of colonial trade. Unlike some of the new British taxes, the Navigation Acts had remained intact throughout this period.

The Burden of the Navigation Acts

One early attempt to quantify the economic effects of the Navigation Acts was by Thomas (1965). Building upon the previous work of Harper (1942), Thomas employed a counterfactual analysis to assess what would have happened to the American economy in the absence of the Navigation Acts. To do this he compared American trade under the Acts with that which would have occurred had America been independent following the Seven Years War. Thomas then estimated the loss of both consumer and produce surplus to the colonies as a result of shipping enumerated goods indirectly through England. These burdens were partially offset by his estimated value of the benefits of British protection and various bounties paid to the colonies. The outcome of his analysis was that the Navigation Acts imposed a net burden of less than one percent of colonial per capita income. From this he concluded the Acts were an unlikely cause of the Revolution. A long series of subsequent works questioned various parts of his analysis but not his general conclusion (Walton, 1971). The work of Thomas also appeared to be consistent with the observation that the First Continental Congress had not demanded in its list of grievances the repeal of either the Navigation Acts or the Sugar Act.

American Expectations about Future British Policy

Did this mean then that the Americans had few if any economic incentives for independence? Upon further consideration economic historians realized that perhaps more important to the colonists were not the past and present burdens but rather the expected future burdens of continued membership in the British Empire. The Declaratory Act made it clear the British government had not given up what it viewed as its right to tax the colonists. This was despite the fact that up to 1775 the Americans had employed a variety of protest measures including lobbying, petitions, boycotts, and violence. The confluence of not having representation in Parliament while confronting an aggressive new British tax policy designed to raise their relatively low taxes may have made it reasonable for the Americans to expect a substantial increase in the level of taxation in the future (Gunderson, 1976, Reid, 1978). Furthermore a recent study has argued that in 1776 not only did the future burdens of the Navigation Acts clearly exceed those of the past, but a substantial portion would have borne by those who played a major role in the Revolution (Sawers, 1992). Seen in this light the economic incentive for independence would have been avoiding the potential future costs of remaining in the British Empire.

The Americans Undertake a Revolution

1776-77

British Military Advantages

The American colonies had both strengths and weaknesses in terms of undertaking a revolution. The colonial population of well over two million was nearly one third of that in Britain (McCusker and Menard, 1985). The growth in the colonial economy had generated a remarkably high level of per capita wealth and income (Jones, 1980). Yet the hurdles confronting the Americans in achieving independence were indeed formidable. The British military had an array of advantages. With virtual control of the Atlantic its navy could attack anywhere along the American coast at will and would have borne logistical support for the army without much interference. A large core of experienced officers commanded a highly disciplined and well-drilled army in the large-unit tactics of eighteenth century European warfare. By these measures the American military would have great difficulty in defeating the British. Its navy was small. The Continental Army had relatively few officers proficient in large-unit military tactics. Lacking both the numbers and the discipline of its adversary the American army was unlikely to be able to meet the British army on equal terms on the battlefield (Higginbotham, 1977).

British Financial Advantages

In addition, the British were in a better position than the Americans to finance a war. A tax system was in place that had provided substantial revenue during previous colonial wars. Also for a variety of reasons the government had acquired an exceptional capacity to generate debt to fund wartime expenses (North and Weingast, 1989). For the Continental Congress the situation was much different. After declaring independence Congress had set about defining the institutional relationship between it and the former colonies. The powers granted to Congress were established under the Articles of Confederation. Reflecting the political environment neither the power to tax nor the power to regulate commerce was given to Congress. Having no tax system to generate revenue also made it very difficult to borrow money. According to the Articles the states were to make voluntary payments to Congress for its war efforts. This precarious revenue system was to hamper funding by Congress throughout the war (Baack, 2001).

Military and Financial Factors Determine Strategy

It was within these military and financial constraints that the war strategies by the British and the Americans were developed. In terms of military strategies both of the contestants realized that America was simply too large for the British army to occupy all of the cities and countryside. This being the case the British decided initially that they would try to impose a naval blockade and capture major American seaports. Having already occupied Boston, the British during 1776 and 1777 took New York, Newport, and Philadelphia. With plenty of room to maneuver his forces and unable to match those of the British, George Washington chose to engage in a war of attrition. The purpose was twofold. First, by not engaging in an all out offensive Washington reduced the probability of losing his army. Second, over time the British might tire of the war.

Saratoga

Frustrated without a conclusive victory, the British altered their strategy. During 1777 a plan was devised to cut off New England from the rest of the colonies, contain the Continental Army, and then defeat it. An army was assembled in Canada under the command of General Burgoyne and then sent to and down along the Hudson River. It was to link up with an army sent from New York City. Unfortunately for the British the plan totally unraveled as in October Burgoyne’s army was defeated at the battle of Saratoga and forced to surrender (Ketchum, 1997).

The American Financial Situation Deteriorates

With the victory at Saratoga the military side of the war had improved considerably for the Americans. However, the financial situation was seriously deteriorating. The states to this point had made no voluntary payments to Congress. At the same time the continental currency had to compete with a variety of other currencies for resources. The states were issuing their own individual currencies to help finance expenditures. Moreover the British in an effort to destroy the funding system of the Continental Congress had undertaken a covert program of counterfeiting the Continental dollar. These dollars were printed and then distributed throughout the former colonies by the British army and agents loyal to the Crown (Newman, 1957). Altogether this expansion of the nominal money supply in the colonies led to a rapid depreciation of the Continental dollar (Calomiris, 1988, Michener, 1988). Furthermore, inflation may have been enhanced by any negative impact upon output resulting from the disruption of markets along with the destruction of property and loss of able-bodied men (Buel, 1998). By the end of 1777 inflation had reduced the specie value of the Continental to about twenty percent of what it had been when originally issued. This rapid decline in value was becoming a serious problem for Congress in that up to this point almost ninety percent of its revenue had been generated from currency emissions.

1778-83

British Invasion of the South

The British defeat at Saratoga had a profound impact upon the nature of the war. The French government still upset by their defeat by the British in the Seven Years War and encouraged by the American victory signed a treaty of alliance with the Continental Congress in early 1778. Fearing a new war with France the British government sent a commission to negotiate a peace treaty with the Americans. The commission offered to repeal all of the legislation applying to the colonies passed since 1763. Congress rejected the offer. The British response was to give up its efforts to suppress the rebellion in the North and in turn organize an invasion of the South. The new southern campaign began with the taking of the port of Savannah in December. Pursuing their southern strategy the British won major victories at Charleston and Camden during the spring and summer of 1780.

Worsening Inflation and Financial Problems

As the American military situation deteriorated in the South so did the financial circumstances of the Continental Congress. Inflation continued as Congress and the states dramatically increased the rate of issuance of their currencies. At the same time the British continued to pursue their policy of counterfeiting the Continental dollar. In order to deal with inflation some states organized conventions for the purpose of establishing wage and price controls (Rockoff, 1984). With its currency rapidly depreciating in value Congress increasingly relied on funds from other sources such as state requisitions, domestic loans, and French loans of specie. As a last resort Congress authorized the army to confiscate property.

Yorktown

Fortunately for the Americans the British military effort collapsed before the funding system of Congress. In a combined effort during the fall of 1781 French and American forces trapped the British southern army under the command of Cornwallis at Yorktown, Virginia. Under siege by superior forces the British army surrendered on October 19. The British government had now suffered not only the defeat of its northern strategy at Saratoga but also the defeat of its southern campaign at Yorktown. Following Yorktown, Britain suspended its offensive military operations against the Americans. The war was over. All that remained was the political maneuvering over the terms for peace.

The Treaty of Paris

The Revolutionary War officially concluded with the signing of the Treaty of Paris in 1783. Under the terms of the treaty the United States was granted independence and British troops were to evacuate all American territory. While commonly viewed by historians through the lens of political science, the Treaty of Paris was indeed a momentous economic achievement by the United States. The British ceded to the Americans all of the land east of the Mississippi River which they had taken from the French during the Seven Years War. The West was now available for settlement. To the extent the Revolutionary War had been undertaken by the Americans to avoid the costs of continued membership in the British Empire, the goal had been achieved. As an independent nation the United States was no longer subject to the regulations of the Navigation Acts. There was no longer to be any economic burden from British taxation.

THE FORMATION OF A NATIONAL GOVERNMENT

When you start a revolution you have to be prepared for the possibility you might win. This means being prepared to form a new government. When the Americans declared independence their experience of governing at a national level was indeed limited. In 1765 delegates from various colonies had met for about eighteen days at the Stamp Act Congress in New York to sort out a colonial response to the new stamp duties. Nearly a decade passed before delegates from colonies once again got together to discuss a colonial response to British policies. This time the discussions lasted seven weeks at the First Continental Congress in Philadelphia during the fall of 1774. The primary action taken at both meetings was an agreement to boycott trade with England. After having been in session only a month, delegates at the Second Continental Congress for the first time began to undertake actions usually associated with a national government. However, when the colonies were declared to be free and independent states Congress had yet to define its institutional relationship with the states.

The Articles of Confederation

Following the Declaration of Independence, Congress turned to deciding the political and economic powers it would be given as well as those granted to the states. After more than a year of debate among the delegates the allocation of powers was articulated in the Articles of Confederation. Only Congress would have the authority to declare war and conduct foreign affairs. It was not given the power to tax or regulate commerce. The expenses of Congress were to be made from a common treasury with funds supplied by the states. This revenue was to be generated from exercising the power granted to the states to determine their own internal taxes. It was not until November of 1777 that Congress approved the final draft of the Articles. It took over three years for the states to ratify the Articles. The primary reason for the delay was a dispute over control of land in the West as some states had claims while others did not. Those states with claims eventually agreed to cede them to Congress. The Articles were then ratified and put into effect on March 1, 1781. This was just a few months before the American victory at Yorktown. The process of institutional development had proved so difficult that the Americans fought almost the entire Revolutionary War with a government not sanctioned by the states.

Difficulties in the 1780s

The new national government that emerged from the Revolution confronted a host of issues during the 1780s. The first major one to be addressed by Congress was what to do with all of the land acquired in the West. Starting in 1784 Congress passed a series of land ordinances that provided for land surveys, sales of land to individuals, and the institutional foundation for the creation of new states. These ordinances opened the West for settlement. While this was a major accomplishment by Congress, other issues remained unresolved. Having repudiated its own currency and no power of taxation, Congress did not have an independent source of revenue to pay off its domestic and foreign debts incurred during the war. Since the Continental Army had been demobilized no protection was being provided for settlers in the West or against foreign invasion. Domestic trade was being increasingly disrupted during the 1780s as more states began to impose tariffs on goods from other states. Unable to resolve these and other issues Congress endorsed a proposed plan to hold a convention to meet in Philadelphia in May of 1787 to revise the Articles of Confederation.

Rather than amend the Articles, the delegates to the convention voted to replace them entirely with a new form of national government under the Constitution. There are of course many ways to assess the significance of this truly remarkable achievement. One is to view the Constitution as an economic document. Among other things the Constitution specifically addressed many of the economic problems that confronted Congress during and after the Revolutionary War. Drawing upon lessons learned in financing the war, no state under the Constitution would be allowed to coin money or issue bills of credit. Only the national government could coin money and regulate its value. Punishment was to be provided for counterfeiting. The problems associated with the states contributing to a common treasury under the Articles were overcome by giving the national government the coercive power of taxation. Part of the revenue was to be used to pay for the common defense of the United States. No longer would states be allowed to impose tariffs as they had done during the 1780s. The national government was now given the power to regulate both foreign and interstate commerce. As a result the nation was to become a common market. There is a general consensus among economic historians today that the economic significance of the ratification of the Constitution was to lay the institutional foundation for long run growth. From the point of view of the former colonists, however, it meant they had succeeded in transferring the power to tax and regulate commerce from Parliament to the new national government of the United States.

TABLES
Table 1 Continental Dollar Emissions (1775-1779)

Year of Emission Nominal Dollars Emitted (000) Annual Emission As Share of Total Nominal Stock Emitted Specie Value of Annual Emission (000) Annual Emission As Share of Total Specie Value Emitted
1775 $6,000 3% $6,000 15%
1776 19,000 8 15,330 37
1777 13,000 5 4,040 10
1778 63,000 26 10,380 25
1779 140,500 58 5,270 13
Total $241,500 100% $41,020 100%

Source: Bullock (1895), 135.
Table 2 Currency Emissions by the States (1775-1781)

Year of Emission Nominal Dollars Emitted (000) Year of Emission Nominal Dollars Emitted (000)
1775 $4,740 1778 $9,118
1776 13,328 1779 17,613
1777 9,573 1780 66,813
1781 123.376
Total $27,641 Total $216,376

Source: Robinson (1969), 327-28.

References

Baack, Ben. “Forging a Nation State: The Continental Congress and the Financing of the War of American Independence.” Economic History Review 54, no.4 (2001): 639-56.

Brewer, John. The Sinews of Power: War, Money and the English State, 1688- 1783. London: Cambridge University Press, 1989.

Buel, Richard. In Irons: Britain’s Naval Supremacy and the American Revolutionary Economy. New Haven: Yale University Press, 1998.

Bullion, John L. A Great and Necessary Measure: George Grenville and the Genesis of the Stamp Act, 1763-1765. Columbia: University of Missouri Press, 1982.

Bullock, Charles J. “The Finances of the United States from 1775 to 1789, with Especial Reference to the Budget.” Bulletin of the University of Wisconsin 1 no. 2 (1895): 117-273.

Calomiris, Charles W. “Institutional Failure, Monetary Scarcity, and the Depreciation of the Continental.” Journal of Economic History 48 no. 1 (1988): 47-68.

Egnal, Mark. A Mighty Empire: The Origins of the American Revolution. Ithaca: Cornell University Press, 1988.

Ferguson, E. James. The Power of the Purse: A History of American Public Finance, 1776-1790. Chapel Hill: University of North Carolina Press, 1961.

Gunderson, Gerald. A New Economic History of America. New York: McGraw- Hill, 1976.

Harper, Lawrence A. “Mercantilism and the American Revolution.” Canadian Historical Review 23 (1942): 1-15.

Higginbotham, Don. The War of American Independence: Military Attitudes, Policies, and Practice, 1763-1789. Bloomington: Indiana University Press, 1977.

Jensen, Merrill, editor. English Historical Documents: American Colonial Documents to 1776 New York: Oxford university Press, 1969.

Johnson, Allen S. A Prologue to Revolution: The Political Career of George Grenville (1712-1770). New York: University Press, 1997.

Jones, Alice H. Wealth of a Nation to Be: The American Colonies on the Eve of the Revolution. New York: Columbia University Press, 1980.

Ketchum, Richard M. Saratoga: Turning Point of America’s Revolutionary War. New York: Henry Holt and Company, 1997.

Labaree, Benjamin Woods. The Boston Tea Party. New York: Oxford University Press, 1964.

Mackesy, Piers. The War for America, 1775-1783. Cambridge: Harvard University Press, 1964.

McCusker, John J. and Russell R. Menard. The Economy of British America, 1607- 1789. Chapel Hill: University of North Carolina Press, 1985.

Michener, Ron. “Backing Theories and the Currencies of Eighteenth-Century America: A Comment.” Journal of Economic History 48 no. 3 (1988): 682-692.

Nester, William R. The First Global War: Britain, France, and the Fate of North America, 1756-1775. Westport: Praeger, 2000.

Newman, E. P. “Counterfeit Continental Currency Goes to War.” The Numismatist 1 (January, 1957): 5-16.

North, Douglass C., and Barry R. Weingast. “Constitutions and Commitment: The Evolution of Institutions Governing Public Choice in Seventeenth-Century England.” Journal of Economic History 49 No. 4 (1989): 803-32.

O’Shaughnessy, Andrew Jackson. An Empire Divided: The American Revolution and the British Caribbean. Philadelphia: University of Pennsylvania Press, 2000.

Palmer, R. R. The Age of Democratic Revolution: A Political History of Europe and America. Vol. 1. Princeton: Princeton University Press, 1959.

Perkins, Edwin J. The Economy of Colonial America. New York: Columbia University Press, 1988.

Reid, Joseph D., Jr. “Economic Burden: Spark to the American Revolution?” Journal of Economic History 38, no. 1 (1978): 81-100.

Robinson, Edward F. “Continental Treasury Administration, 1775-1781: A Study in the Financial History of the American Revolution.” Ph.D. diss., University of Wisconsin, 1969.

Rockoff, Hugh. Drastic Measures: A History of Wage and Price Controls in the United States. Cambridge: Cambridge University Press, 1984.

Sawers, Larry. “The Navigation Acts Revisited.” Economic History Review 45, no. 2 (1992): 262-84.

Thomas, Robert P. “A Quantitative Approach to the Study of the Effects of British Imperial Policy on Colonial Welfare: Some Preliminary Findings.” Journal of Economic History 25, no. 4 (1965): 615-38.

Tucker, Robert W. and David C. Hendrickson. The Fall of the First British Empire: Origins of the War of American Independence. Baltimore: Johns Hopkins Press, 1982.

Walton, Gary M. “The New Economic History and the Burdens of the Navigation Acts.” Economic History Review 24, no. 4 (1971): 533-42.

An Economic History of Weather Forecasting

Erik D. Craft, University of Richmond

Introduction

The United States Congress established a national weather organization in 1870 when it instructed the Secretary of War to organize the collection of meteorological observations and forecasting of storms on the Great Lakes and Atlantic Seaboard. Large shipping losses on the Great Lakes during the 1868 and 1869 seasons, growing acknowledgement that storms generally traveled from the West to the East, a telegraphic network that extended west of the Great Lakes and the Atlantic Seaboard, and an eager Army officer promising military discipline are credited with convincing Congress that a storm-warning system was feasible. The United States Army Signal Service weather organization immediately dwarfed its European counterparts in budget and geographical size and shortly thereafter created storm warnings that on the Great Lakes alone led to savings in shipping losses that exceeded the entire network’s expenses.

Uses of Weather Information

Altering Immediate Behavior

The most obvious use of weather information is to change behavior in response to expected weather outcomes. The motivating force behind establishing weather organizations in England, France, Germany, and the United States was to provide warnings to ships of forthcoming storms, so that the ships might remain in harbor. But it soon became obvious that agricultural and commercial interests would benefit from weather forecasts as well. Farmers could protect fruit sensitive to freezes, and shippers could limit spoilage of produce while en route. Beyond preparation for severe weather, weather forecasts are now created for ever more specialized activities: implementing military operations, scheduling operation of power generation facilities, routing aircraft safely and efficiently, planning professional sports teams’ strategies, estimating demand for commodities sensitive to weather outcomes, planning construction projects, and optimizing the use of irrigation and reservoir systems’ resources.

Applying Climatological Knowledge

Climatological data can be used to match crop varieties, construction practices, and other activities appropriately to different regions. For example, in 1947 the British Government planned to grow groundnuts on 3.2 million acres in East and Central Africa. The groundnut was chosen because it was suited to the average growing conditions of the chosen regions. But due a lack of understanding of the variance in amount and timing of rainfall, the project was abandoned after five years and initial capital outlays of 24 million British pounds and annual operating costs of 7 million pounds. The preparation of ocean wind and weather charts in the 1850s by Matthew Fontaine Maury, Superintendent of the U.S. Navy’s Depot of Charts and Instruments, identified better routes for vessels sailing between America and Europe and from the United States East Cost to United States West Coast. The reduced sailing durations are alleged to have saved millions of dollars annually. Climatological data can also be used in modern environmental forecasts of air quality and how pollution is dispersed in the air. There are even forensic meteorologists who specialize in identifying weather conditions at a given point in time after accidents and subsequent litigation. Basic climatological information is also one reason why the United States cinema industry became established in Southern California; it was known that a high percentage of all days were sunny, so that outdoor filming would not be delayed.

Smoothing Consumption of Weather-Sensitive Commodities

An indirect use of weather forecasts and subsequent weather occurrences is their influence on the prices of commodities that are affected by weather outcomes. Knowledge that growing conditions will be poor or have been poor will lead to expectations of a smaller crop harvest. This causes expected prices of the crop to rise, thereby slowing consumption. This is socially efficient, since the present inventory and now smaller future harvest will have to be consumed more slowly over the time period up until the next season’s crop can be planted, cultivated, and harvested. Without an appropriate rise in price after bad weather outcomes, an excessive depletion of the crop’s inventory could result, leading to more variability in the consumption path of the commodity. People generally prefer consuming their income and individual products in relatively smooth streams, rather than in large amounts in some periods and small amounts in other periods. Both improved weather forecasts and United State Department of Agriculture crop forecasts help buyers more effectively consume a given quantity of a crop.

The History Weather Forecasts in the United States

An important economic history question is whether or not it was necessary for the United States Federal Government to found a weather forecasting organization. There are two challenges in answering that question: establishing that the weather information was socially valuable and determining if private organizations were incapable of providing the appropriate level of services. Restating the latter issue, did weather forecasts and the gathering of climatological information possess enough attributes of a public good such that private organizations would create an insufficiently large amount of socially- beneficial information? There are also two parts to this latter public good problem: nonexcludability and nonrivalry. Could private producers of weather information create a system whereby they earned enough money from users of weather information to cover the costs of creating the information? Would such a weather system be of the socially optimal size?

Potential Organizational Sources of Weather Forecasts

There were many organizations during the 1860s that the observer might imagine would benefit from the creation of weather forecasts. After the consolidation of most telegraphic service in the United States into Western Union in 1866, an organization with employees throughout the country existed. The Associated Press had a weather-reporting network, but there is no evidence that it considered supplementing its data with forecasts. One Ebenezer E. Merriam began supplying New York newspapers with predictions in 1856. Many years later, astronomer turned Army Signal Service forecaster Cleveland Abbe concluded that Merriam made his predictions using newspaper weather reports. The Chicago Board of Trade declined an invitation in 1869 to support a weather forecasting service based in Cincinnati. Neither ship-owners nor marine insurers appear to have expressed any interest in creating or buying weather information. Great Lakes marine insurers had even already overcome organizational problems by forming the Board of Lake Underwriters in 1855. For example, the group incurred expenses of over $11,000 in 1861 inspecting vessels and providing ratings on behalf of its members in the annual Lake Vessel Register. The Board of Lake Underwriters even had nine inspectors distributed on the Great Lakes to inspect wrecks on behalf of its members. Although there was evidence that storms generally traveled in a westerly direction, none of these groups apparently expected the benefits to itself to exceed the costs of establishing the network necessary to provide useful weather information.

Cleveland Abbe at the Cincinnati Observatory began the most serious attempt to establish a quasi-private meteorological organization in 1868 when he sought financial support from the Associated Press, Western Union, local newspapers, and the Cincinnati Chamber of Commerce. His initial plan included a system of one hundred reporting stations with the Associated Press covering the $100 instrument costs at half of the stations and the dispatch costs. In the following year, he widened his scope to include the Chicago Board of Trade and individual subscribers and proposed a more limited network of between sixteen and twenty-two stations. The Cincinnati Chamber of Commerce, whose president published the Cincinnati Commercial, funded the experiment from September through November of 1869. Abbe likely never had more than ten observers report on any given day and could not maintain more than about thirty local subscribers for his service, which provided at most only occasional forecasts. Abbe continued to receive assistance from Western Union in the collection and telegraphing of observations after the three-month trial, but he fell short in raising funds to allow the expansion of his network to support weather forecasts. His ongoing “Weather Bulletin of the Cincinnati Observatory” was not even published in the Cincinnati Commercial.

Founding of the Army Signal Service Weather Organization

Just as the three-month trial of Abbe’s weather bulletin concluded, Increase A. Lapham, a Milwaukee natural scientist, distributed his second list of Great Lakes shipping losses, entitled “Disaster on the Lakes.” The list included 1,164 vessel casualties, 321 deaths, and $3.1 million in property damaged in 1868 and 1,914 vessel casualties, 209 lives lost, and $4.1 million in financial losses in 1869. The number of ships that were totally destroyed was 105 and 126 in each year, respectively. According to a separate account, the storm of November 16-19, 1869 alone destroyed vessels whose value exceeded $420,000. Lapham’s list of losses included a petition to establish a weather forecasting service. In 1850, he had prepared a similar proposal alongside a list of shipping of losses, and twice during the 1850s he had tracked barometric lows across Wisconsin to provide evidence that storms could be forecast.

Recipients of Lapham’s petitions included the Wisconsin Academy of Sciences, the Chicago Academy of Sciences, the National Board of Trade meeting in Richmond, a new Chicago monthly business periodical entitled The Bureau, and Congressman Halbert E. Paine of Milwaukee. Paine had studied meteorological theories under Professor Elias Loomis at Western Reserve College and would introduce storm-warning service bills and eventually the final joint resolution in the House that gave the Army Signal Service storm-warning responsibilities. In his book Treatise on Meteorology (1868), Loomis claimed that the approach of storms to New York could be predicted reliably given telegraphic reports from several locations in the Mississippi Valley. From December 1869 through February 1870, Lapham’s efforts received wider attention. The Bureau featured nine pieces on meteorology from December until March, including at least two by Lapham.

Following the Civil War, the future of a signaling organization in the Army was uncertain. Having had budget requests for telegraph and signal equipment for years 1870 and 1871 cut in half to $5000, Colonel Albert J. Myer, Chief Signal Officer, led a small organization seeking a permanent existence. He visited Congressman Paine’s office in December of 1869 with maps showing proposed observation stations throughout the United Stations. Myer’s eagerness for the weather responsibilities, as well as the discipline of the Army organization and a network of military posts in the West, many linked via telegraph, would appear to have made the Army Signal Service a natural choice. The marginal costs of an Army weather organization using Signal Service personnel included only instruments and commercial telegraphy expenses. On February 4, 1870, Congress approved the Congressional Joint Resolution which “authorizes and requires the Secretary of War to provide for taking of meteorological observations . . . and for giving notice on the northern lakes and on the sea-coast of the approach and force of storms.” Five days later, President Grant signed the bill.

Expansion of the Army Signal Service’s Weather Bureau

Observer-sergeants in the Signal Service recorded their first synchronous observations November 1, 1870, 7:35 a.m. Washington time at twenty-four stations. The storm-warning system began formal operation October 23, 1871 with potential flag displays at eight ports on the Great Lakes and sixteen ports on the Atlantic seaboard. At that time, only fifty general observation stations existed. Already by June 1872, Congress expanded the Army Signal Service’s explicit forecast responsibilities via an appropriations act to most of the United States “for such stations, reports, and signal as may be found necessary for the benefit of agriculture and commercial interests.” In 1872, the Signal Service also began publication of the Weekly Weather Chronicle during the growing seasons. It disappeared in 1877, reemerging in 1887 as the Weather Crop Bulletin. As the fall of 1872 began, confidence in the utility of weather information was so high that 89 agricultural societies and 38 boards of trade and chambers of commerce had appointed meteorological committees to communicate with the Army Signal Service. In addition to dispensing general weather forecasts for regions of the country three times a day, the Signal Service soon sent special warnings to areas in danger of cold waves and frosts.

The original method of warning ships of dangerous winds was hoisting a single red flag with a black square located in the middle. This was known as a cautionary signal, and Army personnel at Signal Service observation stations or civilians at display stations would raise the flag on a pole “whenever the winds are expected to be as strong as twenty-five miles per hour, and to continue so for several hours, within a radius of one hundred miles from the station.” In the first year of operation ending 1 September 1872, 354 cautionary signals were flown on both the Great Lakes and the Atlantic Seaboard, approximately 70% of which were verified as having met the above definition. Such a measure of accuracy is incomplete, however, as it can always be raised artificially by not forecasting storms under marginal conditions, even though such a strategy might diminish the value of the service.

The United States and Canada shared current meteorological information beginning in 1871. By 1880, seventeen Canadian stations reported meteorological data to the United States at least twice daily by telegraph. The number of Army Signal Service stations providing telegraphic reports three times a day stabilized at 138 stations in 1880, dipped to 121 stations in 1883, and grew to approximately 149 stations by 1888. (See Table 1 for a summary of the growth of the Army Signal Service Meteorological Network from 1870 to 1890.) Additional display stations only provided storm warnings at sea and lake ports. River stations monitored water levels in order to forecast floods. Special cotton-region stations, beginning in 1883, comprised a dense network of daily reporters of rainfall and maximum and minimum temperatures. Total Army Signal Service expenditures grew from a $15,000 supplemental appropriation for weather operations in fiscal year 1870 to about one million dollars for all Signal Service costs around 1880 and stabilized at that level. Figure 1 shows the extent geographical extent of the Army Signal Service telegraphic observation network in 1881.

Figure 1: Army Signal Service Observation Network in 1881
Click on the image for the larger, more detailed image (~600K)Source: Map between pages 250-51, Annual Report of the Chief Signal Officer, October 1, 1881, Congressional Serial Set Volume 2015. See the detailed map between pages 304-05 for the location of each of the different types of stations listed in Table 1.

Table 1: Growth of the United States Army Signal Service Meteorological Network

Budget (Real 1880 Dollars)

Stations of the Second Order

Stations of the Third Order

Repair Stations

Display Stations

Special River Stations

Special Cotton-Region Stations

1870

32,487

25

1871

112,456

54

1872

220,269

65

1873

549,634

80

9

1874

649,431

92

20

1875

749,228

98

20

1876

849,025

106

38

23

1877

849,025

116

29

10

9

23

1878

978,085

136

36

12

11

23

1879

1,043,604

158

30

17

46

30

1880

1,109,123

173

39

49

50

29

1881

1,080,254

171

47

44

61

29

87

1882

937,077

169

45

3

74

30

127

1883

950,737

143

42

27

7

30

124

1884

1,014,898

138

68

7

63

40

138

1885

1,085,479

152

58

8

64

66

137

1886

1,150,673

146

33

11

66

69

135

1887

1,080,291

145

31

13

63

70

133

1888

1,063,639

149

30

24

68

78

116

1889

1,022,031

148

32

23

66

72

114

1890

994,629

144

34

15

73

72

114

Sources: Report of the Chief Signal Officer: 1888, p. 171; 1889, p. 136; 1890, p. 203 and “Provision of Value of Weather Information Services,” Craft (1995), p. 34.

Notes: The actual total budgets for years 1870 through 1881 are estimated. Stations of the second order recorded meteorological conditions three times per day. Most immediately telegraphed the data. Stations of the third order recorded observations at sunset. Repair stations maintained Army telegraph lines. Display stations displayed storm warnings on the Great Lakes and Atlantic seaboard. Special river stations monitored water levels in order to forecast floods. Special cotton-region stations collected high temperature, low temperature, and precipitation data from a denser network of observation locations

Early Value of Weather Information

Budget reductions in the Army Signal Service’s weather activities in 1883 led to the reduction of fall storm-warning broadcast locations on the Great Lakes from 80 in 1882 to 43 in 1883. This one-year drop in the availability of storm-warnings creates a special opportunity to measure the value of warnings of extremely high winds on the Great Lakes (see Figure 2). Many other factors can be expected to affect the value of shipping losses on the Great Lakes: the level of commerce in a given season, the amount of shipping tonnage available to haul a season’s commerce, the relative composition of the tonnage (steam versus sail), the severity of the weather, and long-term trends in technological change or safety. Using a statistical technique know as multiple regression, in which the effect of these many factors on shipping losses are analyzed concurrently, Craft (1998) argued that each extra storm-warning location on the Great Lakes lowered losses by about one percent. This implies that the storm-warning system reduced losses on the Great Lakes by approximately one million dollars annually in the mid 1870s and between $1 million and $4.5 million dollars per year by the early 1880s.

Source: The data are found in the following: Chicago Daily Inter Ocean (December 5, 1874 p. 2; December 18, 1875; December 27, 1876 p. 6; December 17, 1878; December 29, 1879 p. 6; February 3, 1881 p. 12; December 28, 1883 p. 3; December 5, 1885 p. 4); Marine Record (December 27, 1883 p. 5; December 25, 1884 pp. 4-5; December 24, 1885 pp. 4-5; December 30, 1886 p. 6; December 15, 1887 pp 4-5); Chief Signal Officer, Annual Report of the Chief Signal Officer, 1871- 1890.

Note: Series E 52 of the Historical Statistics of the United States (U.S. Bureau of the Census, 1975) was used to adjust all values to real 1880 dollars.

There are additional indirect methods with which to confirm the preceding estimate of the value of early weather information. If storm-warnings actually reduced the risk of damage to cargo and ships due to bad weather, then the cost of shipping cargo would be expected to decline. In particular, such reductions in shipping prices due to savings in losses caused by storms can be differentiated from other types of technological improvements by studying how fall shipping prices changed relative to summer shipping prices. It was during the fall that ships were particularly vulnerable to accidents caused by storms. Changes is shipping prices of grain from Chicago to Buffalo during the summers and falls from the late 1860s to late 1880s imply that storm-warnings were valuable and are consistent with the more direct method estimating reductions in shipping losses. Although marine insurance premia data for shipments on the Great Lakes are limited and difficult to interpret due the waning and waxing of the insurance cartel’s cohesion, such data are also supportive of the overall interpretation.

Given Army Signal Service budgets of about one million dollars for providing meteorological services to the entire United States, a reasonable minimum bound for the rate of return to the creation of weather information from 1870 to 1888 is 64 percent. The figure includes no social benefits from any weather information other than Great Lakes storm warnings. This estimate of nineteenth century information implies that the creation and distribution of storm warnings by the United States Federal Government were a socially beneficial investment.

Transfer of Weather Services to the Department of Agriculture

The Allison Commission hearings in 1884 and 1885 sought to determine the appropriate organization of Federal agencies whose activities included scientific research. The Allison Commission’s long report included testimony and discussion relating to the organization of the Army Signal Service, the United States Geological Survey, the Coast and Geodetic Survey, and the Navy Hydrographic Office. Weather forecasting required a reliable network of observers, some of whom were the sole Army personnel at a location. Advantages of a military organizational structure included a greater range of disciplinary responses, including court-martials for soldiers, for deficient job performance. Problems, however, of the military organization included the limited ability to increase one’s rank while working for the Signal Service and tension between the civilian and military personnel. In 1891, after an unsuccessful Congressional attempt at reform in 1887, the Weather Bureau became a civilian organization when it joined the young Department of Agriculture.

Aviation and World War I

Interest in upper air weather conditions grew rapidly after the turn of the century on account of two related events: the development of aviation and World War I. Safe use of aircraft depended on more precise knowledge of weather conditions (winds, storms, and visibility) between takeoff and landing locations. Not only were military aircraft introduced during World War I, but understanding wind conditions was also crucial to the use of poison gas on the front lines. In the most important change of the Weather Bureau’s organizational direction since transfer to the Department of Agricultural, Congress passed the Air Commerce Act in 1926, which by 1932 led to 38% of the Weather Bureau’s budget being directed toward aerology research and support.

Transfer of the Weather Bureau to the Department of Commerce

Even though aerological expenditures by the Weather Bureau in support of aviation rivaled funding for general weather services by the late 1930s, the Weather Bureau came under increasing criticism from aviation interests. The Weather Bureau was transferred to the Department of Commerce in 1940 where other support for aviation already originated. This transition mirrored the declining role of agriculture in the United States and movement toward a more urban economy. Subsequently known as the United States Weather Service, it has remained there since.

World War II

During World War II, weather forecasts assumed greater importance, as aircraft and rapid troop movements became key parts of military strategy. Accurate long-range artillery use also depended on knowledge of prevailing winds. For extensive use of weather forecasts and climatological information during wartime, consider Allied plans the strike German oil refineries in Ploesti, Romania. In the winter of 1943 military weather teams parachuted into the mountains of Yugoslavia to relay weather data. Bombers from North Africa could only reach the refineries in the absence of headwinds in either direction of the sortie. Cloud cover en route was important for protection, clear skies were helpful for identification of targets, and southerly winds permitted the bombers to drop their ordinance on the first pass on the south side of the area’s infrastructure, allowing the winds to assist in spreading the fire. Historical data indicated that only March or August were possible windows. Though many aircraft were lost, the August 1 raid was considered a success.

Tide, wind, and cloud conditions were also crucial in the planning of the invasion of Normandy (planned for June 5 and postponed until June 6 in 1944). The German High Command had been advised by its chief meteorologist that conditions were not opportune for an Allied invasion on the days following June 4. Dissention among American and British military forecasters nearly delayed the invasion further. Had it been deferred until the next date of favorable tide conditions, the invasion would have taken place during the worst June storm in twenty years in the English Channel.

Forecasting in Europe

A storm on November 14, 1854 destroyed the French warship Henri IV and damaged other British and French vessels on the Black Sea involved in the Crimean War. A report from the state-supported Paris Observatory indicated that barometric readings showed that the storm has passed across Europe in about four days. Urban Leverrier, director of the Paris Observatory, concluded that had there been a telegraph line between Vienna and the Crimea, the British and French fleets could have received warnings. Although the United States weather network was preceded by storm-warning systems in the Netherlands in 1860, Great Britain in 1861, and France in 1863, the new United States observation network immediately dwarfed the European organizations in both financial resources and geographical magnitude.

Robert FitzRoy, captain of the Beagle during Darwin’s famous voyage, was appointed director of the Meteorological Department established by the British Board of Trade (a government organization) in 1854. The wreck of the well-constructed iron vessel Royal Charter in a storm with much loss of life in October of 1859 provided another opportunity for a meteorological leader to argue that storms could be tracked and forecast. With support from the Prince Consort, FitzRoy and the Meteorological Department were granted approval to establish a storm-warning service. On February 6, 1861 the first warnings were issued. By August 1861 weather forecasts were issued regularly. By 1863, the Meteorological Department had a budget of three thousand English pounds. Criticism arose from different groups. Scientists wished to establish meteorology on a sound theoretical foundation and differentiate it from astrology. At the time, many publishers of weather almanacs subscribed to various theories of the influence of the moon or other celestial bodies on weather (This is not as outlandish one might suppose; in 1875, well-known economist William Stanley Jevons studied connections between sunspot activity and meteorology with business cycles). Some members of this second group supported the practice of forecasting but were critical of FitzRoy’s technique, perhaps hoping to become alternative sources of forecasts. Amidst the criticism, FitzRoy committed suicide in 1865. Forecasts and warnings were discontinued in 1866 until the warnings resumed two years later. General forecasts were suspended until 1877.

In 1862, Leverrier wrote the French Ministry of Public Education that French naval and commercial interests might be compromised by their dependence on warnings from the British Board of Trade. A storm-warning service in France commenced in July of 1863. Given the general movement of storms westward, neither France nor Britain had the luxury of tracking storms well before they arrived, as would have been possible with the November 1854 storm in the Crimea and as the Army Signal Service soon would be able to do in America. On account of administrative difficulties that were to hinder effective functioning of the service until 1877, French warnings ceased in October 1865 but resumed in May the next year. The French Central Bureau Meteorology was founded only in 1878 with a budget of only $12,000.

After the initiation of storm warning systems that preceded the Army Signal Service weather network, Europe would not achieve meteorological prominence again until the Bergen School of meteorology developed new storm analysis techniques after World War I, which incorporated cold and warm fronts. In the difficult days in Norway during the conclusion of the Great War, meteorological information from the rest of Europe was unavailable. Theoretical physicist turned meteorological researcher Wilhelm Bjerknes appealed to Norway’s national interests in defense, in the development of commercial aviation, and in increased agricultural output to build a dense observation network, whose data helped yield a new paradigm for meteorology.

Conclusion

The first weather forecasts in the United States that were based on a large network of simultaneous observations provided information to society that was much more valuable than the cost of production. There was discussion in the early winter of 1870 between the scientist Increase Lapham and a businessman in Chicago of the feasibility of establishing a private forecasting organization in Wisconsin or Illinois (see Craft 1999). But previous attempts by private organizations in the United States had been unsuccessful in supporting any private weather-forecasting service. In the contemporary United States, the Federal government both collects data and offers forecasts, while private weather organizations provide a variety of customized services.

Weather Forecasting Timeline

1743

Benjamin Franklin, using reports of numerous postmasters, determined the northeastward path of a hurricane from the West Indies.

1772-1777

Thomas Jefferson at Monticello, Virginia and James Madison at Williamsburg, Virginia collect a series of contemporaneous weather observations.

1814

Surgeon General Tilton issues an order directing Army surgeons to keep a diary of the weather in order to ascertain any influences of weather upon disease.

1817

Josiah Meigs, Commission of the General Land Office, requests officials at their land offices to record meteorological observations.

1846-1848

Matthew F. Maury, Superintendent of the U.S. Naval Observatory, publishes his first charts compiled from ships’ log showing efficient sailing routes.

1847

Barometer used to issue storm warnings in Barbadoes.

1848

J. Jones of New York advertises meteorological reports costing between twelve and one half and twenty-five cents per city per day. There is no evidence the service was ever sold.

1848

Publication in the British Daily News of the first telegraphic daily weather report.

1849

The Smithsonian Institution begins a nearly three decade long project of collecting meteorological data with the goal of understanding storms.

1849

Captain Joseph Brooks, manager of the Portland Steamship Line, receives telegraphic reports three times a day from Albany, New York, and Plattsburg in order to determine if the line’s ships should remain in port in Maine.

1853-1855

Ebenezer E. Merriam of New York, using newspaper telegraphic reports, offers weather forecasts in New York’s newspapers on an apparently irregular basis.

1858

The U.S. Army Engineers begin collecting meteorological observations while surveying the Great Lakes.

1860

Christoph Buys Ballot issues first storm warnings in the Netherlands.

1861

Admiral Robert FitzRoy of the British Meteorological Office begins issuing storm-warnings.

1863

Urban Leverrier, director of the Paris Observatory, organizes a storm-warning service.

1868

Cleveland Abbe of the Cincinnati Observatory unsuccessfully proposes a weather service of one hundred observation stations to be supported by the Cincinnati Chamber of Commerce, Associated Press, Western Union, and local newspapers.

1869

The Cincinnati Chamber of Commerce funds a three-month trial of the Cincinnati Observatory’s weather bulletin. The Chicago Board of Trade declines to participate.

1869

Increase A. Lapham publishes a list of the shipping losses on the Great Lakes during the 1868 and 1869 seasons.

1870

Congress passes a joint resolution directing the Secretary of War to establish a meteorological network for the creation of storm warnings on the Great Lakes and Atlantic Seaboard. Storm-warnings are offered on November 8. Forecasts begin the following February 19.

1872

Congressional appropriations bill extends Army Signal Service duties to provide forecasts for agricultural and commercial interests.

1880

Frost warnings offered for Louisiana sugar producers.

1881-1884

Army Signal Service expedition to Lady Franklin Bay in support of international polar weather research. Only seven of the twenty-five member team survives.

1881

Special cotton-region weather reporting network established.

1891

Weather Bureau transferred to the Department of Agriculture.

1902

Daily weather forecasts sent by radio to Cunard Line steamships.

1905

First wireless weather report from a ship at sea.

1918

Norway expands its meteorological network and organization leading to the development of new forecasting theories centered on three-dimensional interaction of cold and warm fronts.

1919

American Meteorological Society founded.

1926

Air Commerce Act gives the Weather Bureau responsibility for providing weather services to aviation.

1934

First private sector meteorologist hired by a utility company.

1940

The Weather Bureau is transferred from the Department of Agriculture to the Department of Commerce.

1946

First private weather forecast companies begin service.

1960

The first meteorological satellite, Tiros I, enters orbit successfully.

1976

The United States launches its first geostationary weather satellites.

References

Abbe, Cleveland, Jr. “A Chronological Outline of the History of Meteorology in the United States.” Monthly Weather Review 37, no. 3-6 (1909): 87-89, 146- 49, 178-80, 252-53.

Alter, J. Cecil. “National Weather Service Origins.” Bulletin of the Historical and Philosophical Society of Ohio 7, no. 3 (1949): 139-85.

Anderson, Katharine. “The Weather Prophets: Science and Reputation in Victorian Meteorology.” History of Science 37 (1999): 179-216.

Burton, Jim. “Robert Fitzroy and the Early History of the Meteorological Office.” British Journal for the History of Science 19 (1986): 147-76.

Chief Signal Officer. Report of the Chief Signal Officer. Washington: GPO, 1871-1890.

Craft, Erik. “The Provision and Value of Weather Information Services in the United States during the Founding Period of the Weather Bureau with Special Reference to Transportation on the Great Lakes.” Ph.D. diss., University of Chicago, 1995.

Craft, Erik. “The Value of Weather Information Services for Nineteenth-Century Great Lakes Shipping.” American Economic Review 88, no.5 (1998): 1059-1076.

Craft, Erik. “Private Weather Organizations and the Founding of the United States Weather Bureau.” Journal of Economic History 59, no. 4 (1999): 1063- 1071.

Davis, John L. “Weather Forecasting and the Development of Meteorological Theory at the Paris Observatory.” Annals of Science 41 (1984): 359-82.

Fleming, James Rodger. Meteorology in America, 1800-1870. Baltimore: Johns Hopkins University Press, 1990.

Fleming, James Rodger, and Roy E. Goodman, editors. International Bibliography of Meteorology. Upland, Pennsylvania: Diane Publishing Co., 1994.

Friedman, Robert Marc. Appropriating the Weather: Vilhelm Bjerknes and the Construction of a Modern Meteorology. Ithaca: Cornell University Press, 1989.

Hughes, Patrick. A Century of Weather Service. New York: Gordon and Breach, 1970.

Miller, Eric R. “The Evolution of Meteorological Institutions in United States.” Monthly Weather Review 59 (1931): 1-6.

Miller, Eric R. “New Light on the Beginnings of the Weather Bureau from the Papers of Increase A. Lapham.” Monthly Weather Review 59 (1931): 65-70.

Sah, Raaj. “Priorities of Developing Countries in Weather and Climate.” World Development 7 no. 3 (1979): 337-47.

Spiegler, David B. “A History of Private Sector Meteorology.” In Historical Essays on Meteorology, 1919-1995, edited by James Rodger Fleming, 417- 41. Boston: American Meteorological Society, 1996.

Weber, Gustavus A. The Weather Bureau: Its History, Activities and Organization. New York: D. Appleton and Company, 1922.

Whitnah, Donald R. A History of the United States Weather Bureau. Urbana: University of Illinois Press, 1961.

Citation: Craft, Erik. “Economic History of Weather Forecasting”. EH.Net Encyclopedia, edited by Robert Whaples. October 6, 2001. URL http://eh.net/encyclopedia/an-economic-history-of-weather-forecasting/

An Overview of the Economic History of Uruguay since the 1870s

Luis Bértola, Universidad de la República — Uruguay

Uruguay’s Early History

Without silver or gold, without valuable species, scarcely peopled by gatherers and fishers, the Eastern Strand of the Uruguay River (Banda Oriental was the colonial name; República Oriental del Uruguay is the official name today) was, in the sixteenth and seventeenth centuries, distant and unattractive to the European nations that conquered the region. The major export product was the leather of wild descendants of cattle introduced in the early 1600s by the Spaniards. As cattle preceded humans, the state preceded society: Uruguay’s first settlement was Colonia del Sacramento, a Portuguese military fortress founded in 1680, placed precisely across from Buenos Aires, Argentina. Montevideo, also a fortress, was founded by the Spaniards in 1724. Uruguay was on the border between the Spanish and Portuguese empires, a feature which would be decisive for the creation, with strong British involvement, in 1828-1830, of an independent state.

Montevideo had the best natural harbor in the region, and rapidly became the end-point of the trans-Atlantic routes into the region, the base for a strong commercial elite, and for the Spanish navy in the region. During the first decades after independence, however, Uruguay was plagued by political instability, precarious institution building and economic retardation. Recurrent civil wars with intensive involvement by Britain, France, Portugal-Brazil and Argentina, made Uruguay a center for international conflicts, the most important being the Great War (Guerra Grande), which lasted from 1839 to 1851. At its end Uruguay had only about 130,000 inhabitants.

“Around the middle of the nineteenth century, Uruguay was dominated by the latifundium, with its ill-defined boundaries and enormous herds of native cattle, from which only the hides were exported to Great Britain and part of the meat, as jerky, to Brazil and Cuba. There was a shifting rural population that worked on the large estates and lived largely on the parts of beef carcasses that could not be marketed abroad. Often the landowners were also the caudillos of the Blanco or Colorado political parties, the protagonists of civil wars that a weak government was unable to prevent” (Barrán and Nahum, 1984, 655). This picture still holds, even if it has been excessively stylized, neglecting the importance of subsistence or domestic-market oriented peasant production.

Economic Performance in the Long Run

Despite its precarious beginnings, Uruguay’s per capita gross domestic product (GDP) growth from 1870 to 2002 shows an amazing persistence, with the long-run rate averaging around one percent per year. However, this apparent stability hides some important shifts. As shown in Figure 1, both GDP and population grew much faster before the 1930s; from 1930 to 1960 immigration vanished and population grew much more slowly, while decades of GDP stagnation and fast growth alternated; after the 1960s Uruguay became a net-emigration country, with low natural growth rates and a still spasmodic GDP growth.

GDP growth shows a pattern featured by Kuznets-like swings (Bértola and Lorenzo 2004), with extremely destructive downward phases, as shown in Table 1. This cyclical pattern is correlated with movements of the terms of trade (the relative price of exports versus imports), world demand and international capital flows. In the expansive phases exports performed well, due to increased demand and/or positive terms of trade shocks (1880s, 1900s, 1920s, 1940s and even during the Mercosur years from 1991 to 1998). Capital flows would sometimes follow these booms and prolong the cycle, or even be a decisive force to set the cycle up, as were financial flows in the 1970s and 1990s. The usual outcome, however, has been an overvalued currency, which blurred the debt problem and threatened the balance of trade by overpricing exports. Crises have been the result of a combination of changing trade conditions, devaluation and over-indebtedness, as in the 1880s, early 1910s, late 1920s, 1950s, early 1980s and late 1990s.

Population and per capita GDP of Uruguay, 1870-2002 (1913=100)

Table 1: Swings in the Uruguayan Economy, 1870-2003

Per capita GDP fall (%) Length of recession (years) Time to pre-crisis levels (years) Time to next crisis (years)
1872-1875 26 3 15 16
1888-1890 21 2 19 25
1912-1915 30 3 15 19
1930-1933 36 3 17 24-27
1954/57-59 9 2-5 18-21 27-24
1981-1984 17 3 11 17
1998-2003 21 5

Sources: See Figure 1.

Besides its cyclical movement, the terms of trade showed a sharp positive trend in 1870-1913, a strongly fluctuating pattern around similar levels in 1913-1960 and a deteriorating trend since then. While the volume of exports grew quickly up to the 1920s, it stagnated in 1930-1960 and started to grow again after 1970. As a result, the purchasing power of exports grew fourfold in 1870-1913, fluctuated along with the terms of trade in 1930-1960, and exhibited a moderate growth in 1970-2002.

The Uruguayan economy was very open to trade in the period up to 1913, featuring high export shares, which naturally declined as the rapidly growing population filled in rather empty areas. In 1930-1960 the economy was increasingly and markedly closed to international trade, but since the 1970s the economy opened up to trade again. Nevertheless, exports, which earlier were mainly directed to Europe (beef, wool, leather, linseed, etc.), were increasingly oriented to Argentina and Brazil, in the context of bilateral trade agreements in the 1970s and 1980s and of Mercosur (the trading zone encompassing Argentina, Brazil, Paraguay and Uruguay) in the 1990s.

While industrial output kept pace with agrarian export-led growth during the first globalization boom before World War I, the industrial share in GDP increased in 1930-54, and was mainly domestic-market orientated. Deindustrialization has been profound since the mid-1980s. The service sector was always large: focusing on commerce, transport and traditional state bureaucracy during the first globalization boom; focusing on health care, education and social services, during the import-substituting industrialization (ISI) period in the middle of the twentieth century; and focusing on military expenditure, tourism and finance since the 1970s.

The income distribution changed markedly over time. During the first globalization boom before World War I, an already uneven distribution of income and wealth seems to have worsened, due to massive immigration and increasing demand for land, both rural and urban. However, by the 1920s the relative prices of land and labor changed their previous trend, reducing income inequality. The trend later favored industrialization policies, democratization, introduction of wage councils, and the expansion of the welfare state based on an egalitarian ideology. Inequality diminished in many respects: between sectors, within sectors, between genders and between workers and pensioners. While the military dictatorship and the liberal economic policy implemented since the 1970s initiated a drastic reversal of the trend toward economic equality, the globalizing movements of the 1980s and 1990s under democratic rule didn’t increase equality. Thus, inequality remains at the higher levels reached during the period of dictatorship (1973-85).

Comparative Long-run Performance

If the stable long-run rate of Uruguayan per capita GDP growth hides important internal transformations, Uruguay’s changing position in the international scene is even more remarkable. During the first globalization boom the world became more unequal: the United States forged ahead as the world leader (nearly followed by other settler economies); Asia and Africa lagged far behind. Latin America showed a confusing map, in which countries as Argentina and Uruguay performed rather well, and others, such as the Andean region, lagged far behind (Bértola and Williamson 2003). Uruguay’s strong initial position tended to deteriorate in relation to the successful core countries during the late 1800s, as shown in Figure 2. This trend of negative relative growth was somewhat weak during the first half of the twentieth century, improved significantly during the 1960s, as the import-substituting industrialization model got exhausted, and has continued since the 1970s, despite policies favoring increased integration into the global economy.

 Per capita GDP of Uruguay relative to four core countries,  1870-2002

If school enrollment and literacy rates are reasonable proxies for human capital, in late 1800s both Argentina and Uruguay had a great handicap in relation to the United States, as shown in Table 2. The gap in literacy rates tended to disappear — as well as this proxy’s ability to measure comparative levels of human capital. Nevertheless, school enrollment, which includes college-level and technical education, showed a catching-up trend until the 1960’s, but reverted afterwards.

The gap in life-expectancy at birth has always been much smaller than the other development indicators. Nevertheless, some trends are noticeable: the gap increased in 1900-1930; decreased in 1930-1950; and increased again after the 1970s.

Table 2: Uruguayan Performance in Comparative Perspective, 1870-2000 (US = 100)

1870 1880 1890 1900 1910 1920 1930 1940 1950 1960 1970 1980 1990 2000
GDP per capita

Uruguay

101 65 63 27 32 27 33 27 26 24 19 18 15 16

Argentina

63 34 38 31 32 29 25 25 24 21 15 16

Brazil

23 8 8 8 8 8 7 9 9 13 11 10
Latin America 13 12 13 10 9 9 9 6 6

USA

100 100 100 100 100 100 100 100 100 100 100 100 100 100
Literacy rates

Uruguay

57 65 72 79 85 91 92 94 95 97 99

Argentina

57 65 72 79 85 91 93 94 94 96 98

Brazil

39 38 37 42 46 51 61 69 76 81 86

Latin America

28 30 34 37 42 47 56 65 71 77 83

USA

100 100 100 100 100 100 100 100 100 100 100
School enrollment

Uruguay

23 31 31 30 34 42 52 46 43

Argentina

28 41 42 36 39 43 55 44 45

Brazil

12 11 12 14 18 22 30 42

Latin America

USA

100 100 100 100 100 100 100 100 100
Life expectancy at birth

Uruguay

102 100 91 85 91 97 97 97 95 96 96

Argentina

81 85 86 90 88 90 93 94 95 96 95
Brazil 60 60 56 58 58 63 79 83 85 88 88
Latin America 65 63 58 58 59 63 71 77 81 88 87
USA 100 100 100 100 100 100 100 100 100 100 100

Sources: Per capita GDP: Maddison (2001) and Astorga, Bergés and FitzGerald (2003). Literacy rates and life expectancy; Astorga, Bergés and FitzGerald (2003). School enrollment; Bértola and Bertoni (1998).

Uruguay during the First Globalization Boom: Challenge and Response

During the post-Great-War reconstruction after 1851, Uruguayan population grew rapidly (fueled by high natural rates and immigration) and so did per capita output. Productivity grew due to several causes including: the steam ship revolution, which critically reduced the price spread between Europe and America and eased access to the European market; railways, which contributed to the unification of domestic markets and reduced domestic transport costs; the diffusion and adaptation to domestic conditions of innovations in cattle-breeding and services; a significant reduction in transaction costs, related to a fluctuating but noticeable process of institutional building and strengthening of the coercive power of the state.

Wool and woolen products, hides and leather were exported mainly to Europe; salted beef (tasajo) to Brazil and Cuba. Livestock-breeding (both cattle and sheep) was intensive in natural resources and dominated by large estates. By the 1880s, the agrarian frontier was exhausted, land properties were fenced and property rights strengthened. Labor became abundant and concentrated in urban areas, especially around Montevideo’s harbor, which played an important role as a regional (supranational) commercial center. By 1908, it contained 40 percent of the nation’s population, which had risen to more than a million inhabitants, and provided the main part of Uruguay’s services, civil servants and the weak and handicraft-dominated manufacturing sector.

By the 1910s, Uruguayan competitiveness started to weaken. As the benefits of the old technological paradigm were eroding, the new one was not particularly beneficial for resource-intensive countries such as Uruguay. International demand shifted away from primary consumption, the population of Europe grew slowly and European countries struggled for self-sufficiency in primary production in a context of soaring world supply. Beginning in the 1920s, the cattle-breeding sector showed a very poor performance, due to lack of innovation away from natural pastures. In the 1930’s, its performance deteriorated mainly due to unfavorable international conditions. Export volumes stagnated until the 1970s, while purchasing power fluctuated strongly following the terms of trade.

Inward-looking Growth and Structural Change

The Uruguayan economy grew inwards until the 1950s. The multiple exchange rate system was the main economic policy tool. Agrarian production was re-oriented towards wool, crops, dairy products and other industrial inputs, away from beef. The manufacturing industry grew rapidly and diversified significantly, with the help of protectionist tariffs. It was light, and lacked capital goods or technology-intensive sectors. Productivity growth hinged upon technology transfers embodied in imported capital goods and an intensive domestic adaptation process of mature technologies. Domestic demand grew also through an expanding public sector and the expansion of a corporate welfare state. The terms of trade substantially impacted protectionism, productivity growth and domestic demand — the government raised money by manipulating exchange rates, so that when export prices rose the state had a greater capacity to protect the manufacturing sector through low exchange rates for capital goods, raw material and fuel imports and to spur productivity increases by imports of capital, while protection allowed industry to pay higher wages and thus expand domestic demand.

However, rent-seeking industries searching for protectionism and a weak clienteslist state, crowded by civil servants recruited in exchange for political favors to the political parties, directed structural change towards a closed economy and inefficient management. The obvious limits to inward looking growth of a country peopled by only about two million inhabitants were exacerbated in the late 1950s as terms of trade deteriorated. The clientelist political system, which was created by both traditional parties while the state was expanding at the national and local level, was now not able to absorb the increasing social conflicts, dyed by stringent ideological confrontation, in a context of stagnation and huge fiscal deficits.

Re-globalization and Regional Integration

The dictatorship (1973-1985) started a period of increasing openness to trade and deregulation which has persisted until the present. Dynamic integration into the world market is still incomplete, however. An attempt to return to cattle-breeding exports, as the engine of growth, was hindered by the oil crises and the ensuing European response, which restricted meat exports to that destination. The export sector was re-orientated towards “non-traditional exports” — i.e., exports of industrial goods made of traditional raw materials, to which low-quality and low-wage labor was added. Exports were also stimulated by means of strong fiscal exemptions and negative real interest rates and were re-orientated to the regional market (Argentina and Brazil) and to other developing regions. At the end of the 1970s, this policy was replaced by the monetarist approach to the balance of payments. The main goal was to defeat inflation (which had continued above 50% since the 1960s) through deregulation of foreign trade and a pre-announced exchange rate, the “tablita.” A strong wave of capital inflows led to a transitory success, but the Uruguayan peso became more and more overvalued, thus limiting exports, encouraging imports and deepening the chronic balance of trade deficit. The “tablita” remained dependent on increasing capital inflows and obviously collapsed when the risk of a huge devaluation became real. Recession and the debt crisis dominated the scene of the early 1980s.

Democratic regimes since 1985 have combined natural resource intensive exports to the region and other emergent markets, with a modest intra-industrial trade mainly with Argentina. In the 1990s, once again, Uruguay was overexposed to financial capital inflows which fueled a rather volatile growth period. However, by the year 2000, Uruguay had a much worse position in relation to the leaders of the world economy as measured by per capita GDP, real wages, equity and education coverage, than it had fifty years earlier.

Medium-run Prospects

In the 1990s Mercosur as a whole and each of its member countries exhibited a strong trade deficit with non-Mercosur countries. This was the result of a growth pattern fueled by and highly dependent on foreign capital inflows, combined with the traditional specialization in commodities. The whole Mercosur project is still mainly oriented toward price competitiveness. Nevertheless, the strongly divergent macroeconomic policies within Mercosur during the deep Argentine and Uruguayan crisis of the beginning of the twenty-first century, seem to have given place to increased coordination between Argentina and Brazil, thus making of the region a more stable environment.

The big question is whether the ongoing political revival of Mercosur will be able to achieve convergent macroeconomic policies, success in international trade negotiations, and, above all, achievements in developing productive networks which may allow Mercosur to compete outside its home market with knowledge-intensive goods and services. Over that hangs Uruguay’s chance to break away from its long-run divergent siesta.

References

Astorga, Pablo, Ame R. Bergés and Valpy FitzGerald. “The Standard of Living in Latin America during the Twentieth Century.” University of Oxford Discussion Papers in Economic and Social History 54 (2004).

Barrán, José P. and Benjamín Nahum. “Uruguayan Rural History.” Latin American Historical Review, 1985.

Bértola, Luis. The Manufacturing Industry of Uruguay, 1913-1961: A Sectoral Approach to Growth, Fluctuations and Crisis. Publications of the Department of Economic History, University of Göteborg, 61; Institute of Latin American Studies of Stockholm University, Monograph No. 20, 1990.

Bértola, Luis and Reto Bertoni. “Educación y aprendizaje en escenarios de convergencia y divergencia.” Documento de Trabajo, no. 46, Unidad Multidisciplinaria, Facultad de Ciencias Sociales, Universidad de la República, 1998.

Bértola, Luis and Fernando Lorenzo. “Witches in the South: Kuznets-like Swings in Argentina, Brazil and Uruguay since the 1870s.” In The Experience of Economic Growth, edited by J.L. van Zanden and S. Heikenen. Amsterdam: Aksant, 2004.

Bértola, Luis and Gabriel Porcile. “Argentina, Brasil, Uruguay y la Economía Mundial: una aproximación a diferentes regímenes de convergencia y divergencia.” In Ensayos de Historia Económica by Luis Bertola. Montevideo: Uruguay en la región y el mundo, 2000.

Bértola, Luis and Jeffrey Williamson. “Globalization in Latin America before 1940.” National Bureau of Economic Research Working Paper, no. 9687 (2003).

Bértola, Luis and others. El PBI uruguayo 1870-1936 y otras estimaciones. Montevideo, 1998.

Maddison, A. Monitoring the World Economy, 1820-1992. Paris: OECD, 1995.

Maddison, A. The World Economy: A Millennial

Citation: Bertola, Luis. “An Overview of the Economic History of Uruguay since the 1870s”. EH.Net Encyclopedia, edited by Robert Whaples. March 16, 2008. URL http://eh.net/encyclopedia/article/Bertola.Uruguay.final

Sweden – Economic Growth and Structural Change, 1800-2000

Lennart Schön, Lund University

This article presents an overview of Swedish economic growth performance internationally and statistically and an account of major trends in Swedish economic development during the nineteenth and twentieth centuries.1

Modern economic growth in Sweden took off in the middle of the nineteenth century and in international comparative terms Sweden has been rather successful during the past 150 years. This is largely thanks to the transformation of the economy and society from agrarian to industrial. Sweden is a small economy that has been open to foreign influences and highly dependent upon the world economy. Thus, successive structural changes have put their imprint upon modern economic growth.

Swedish Growth in International Perspective

The century-long period from the 1870s to the 1970s comprises the most successful part of Swedish industrialization and growth. On a per capita basis the Japanese economy performed equally well (see Table 1). The neighboring Scandinavian countries also grew rapidly but at a somewhat slower rate than Sweden. Growth in the rest of industrial Europe and in the U.S. was clearly outpaced. Growth in the entire world economy, as measured by Maddison, was even slower.

Table 1 Annual Economic Growth Rates per Capita in Industrial Nations and the World Economy, 1871-2005

Year Sweden Rest of Nordic Countries Rest of Western Europe United States Japan World Economy
1871/1875-1971/1975 2.4 2.0 1.7 1.8 2.4 1.5
1971/1975-2001/2005 1.7 2.2 1.9 2.0 2.2 1.6

Note: Rest of Nordic countries = Denmark, Finland and Norway. Rest of Western Europe = Austria, Belgium, Britain, France, Germany, Italy, the Netherlands, and Switzerland.

Source: Maddison (2006); Krantz/Schön (forthcoming 2007); World Bank, World Development Indicator 2000; Groningen Growth and Development Centre, www.ggdc.com.

The Swedish advance in a global perspective is illustrated in Figure 1. In the mid-nineteenth century the Swedish average income level was close to the average global level (as measured by Maddison). In a European perspective Sweden was a rather poor country. By the 1970s, however, the Swedish income level was more than three times higher than the global average and among the highest in Europe.

Figure 1
Swedish GDP per Capita in Relation to World GDP per Capita, 1870-2004
(Nine year moving averages)
Swedish GDP per Capita in Relation to World GDP per Capita, 1870-2004
Sources: Maddison (2006); Krantz/Schön (forthcoming 2007).

Note. The annual variation in world production between Maddison’s benchmarks 1870, 1913 and 1950 is estimated from his supply of annual country series.

To some extent this was a catch-up story. Sweden was able to take advantage of technological and organizational advances made in Western Europe and North America. Furthermore, Scandinavian countries with resource bases such as Sweden and Finland had been rather disadvantaged as long as agriculture was the main source of income. The shift to industry expanded the resource base and industrial development – directed both to a growing domestic market but even more to a widening world market – became the main lever of growth from the late nineteenth century.

Catch-up is not the whole story, though. In many industrial areas Swedish companies took a position at the technological frontier from an early point in time. Thus, in certain sectors there was also forging ahead,2 quickening the pace of structural change in the industrializing economy. Furthermore, during a century of fairly rapid growth new conditions have arisen that have required profound adaptation and a renewal of entrepreneurial activity as well as of economic policies.

The slow down in Swedish growth from the 1970s may be considered in this perspective. While in most other countries growth from the 1970s fell only in relation to growth rates in the golden post-war ages, Swedish growth fell clearly below the historical long run growth trend. It also fell to a very low level internationally. The 1970s certainly meant the end to a number of successful growth trajectories in the industrial society. At the same time new growth forces appeared with the electronic revolution, as well as with the advance of a more service based economy. It may be the case that this structural change hit the Swedish economy harder than most other economies, at least of the industrial capitalist economies. Sweden was forced into a transformation of its industrial economy and of its political economy in the 1970s and the 1980s that was more profound than in most other Western economies.

A Statistical Overview, 1800-2000

Swedish economic development since 1800 may be divided into six periods with different growth trends, as well as different composition of growth forces.

Table 2 Annual Growth Rates in per Capita Production, Total Investments, Foreign Trade and Population in Sweden, 1800-2000

Period Per capita GDP Investments Foreign Trade Population
1800-1840 0.6 0.3 0.7 0.8
1840-1870 1.2 3.0 4.6 1.0
1870-1910 1.7 3.0 3.3 0.6
1910-1950 2.2 4.2 2.0 0.5
1950-1975 3.6 5.5 6.5 0.6
1975-2000 1.4 2.1 4.3 0.4
1800-2000 1.9 3.4 3.8 0.7

Source: Krantz/Schön (forthcoming 2007).

In the first decades of the nineteenth century the agricultural sector dominated and growth was slow in all aspects but in population. Still there was per capita growth, but to some extent this was a recovery from the low levels during the Napoleonic Wars. The acceleration during the next period around the mid-nineteenth century is marked in all aspects. Investments and foreign trade became very dynamic ingredients with the onset of industrialization. They were to remain so during the following periods as well. Up to the 1970s per capita growth rates increased for each successive period. In an international perspective it is most notable that per capita growth rates increased also in the interwar period, despite the slow down in foreign trade. The interwar period is crucial for the long run relative success of Swedish economic growth. The decisive culmination in the post-war period with high growth rates in investments and in foreign trade stands out as well, as the deceleration in all aspects in the late twentieth century.

An analysis in a traditional growth accounting framework gives a long term pattern with certain periodic similarities (see Table 3). Thus, total factor productivity growth has increased over time up to the 1970s, only to decrease to its long run level in the last decades. This deceleration in productivity growth may be looked upon either as a failure of the “Swedish Model” to accommodate new growth forces or as another case of the “productivity paradox” in lieu of the information technology revolution.3

Table 3 Total Factor Productivity (TFP) Growth and Relative Contribution of Capital, Labor and TFP to GDP Growth in Sweden, 1840-2000

Period TFP Growth Capital Labor TFP
1840-1870 0.4 55 27 18
1870-1910 0.7 50 18 32
1910-1950 1.0 39 24 37
1950-1975 2.1 45 7 48
1975-2000 1.0 44 1 55
1840-2000 1.1 45 16 39

Source: See Table 2.

In terms of contribution to overall growth, TFP has increased its share for every period. The TFP share was low in the 1840s but there was a very marked increase with the onset of modern industrialization from the 1870s. In relative terms TFP reached its highest level so far from the 1970s, thus indicating an increasing role of human capital, technology and knowledge in economic growth. The role of capital accumulation was markedly more pronounced in early industrialization with the build-up of a modern infrastructure and with urbanization, but still capital did retain much of its importance during the twentieth century. Thus its contribution to growth during the post-war Golden Ages was significant with very high levels of material investments. At the same time TFP growth culminated with positive structural shifts, as well as increased knowledge intensity complementary to the investments. Labor has in quantitative terms progressively reduced its role in economic growth. One should observe, however, the relatively large importance of labor in Swedish economic growth during the interwar period. This was largely due to demographic factors and to the employment situation that will be further commented upon.

In the first decades of the nineteenth century, growth was still led by the primary production of agriculture, accompanied by services and transport. Secondary production in manufacturing and building was, on the contrary, very stagnant. From the 1840s the industrial sector accelerated, increasingly supported by transport and communications, as well as by private services. The sectoral shift from agriculture to industry became more pronounced at the turn of the twentieth century when industry and transportation boomed, while agricultural growth decelerated into subsequent stagnation. In the post-war period the volume of services, both private and public, increased strongly, although still not outpacing industry. From the 1970s the focus shifted to private services and to transport and communications, indicating fundamental new prerequisites of growth.

Table 4 Growth Rates of Industrial Sectors, 1800-2000

Period Agriculture Industrial and Hand Transport and Communic. Building Private Services Public Services GDP
1800-1840 1.5 0.3 1.1 -0.1 1.4 1.5 1.3
1840-1870 2.1 3.7 1.8 2.4 2.7 0.8 2.3
1870-1910 1.0 5.0 3.9 1.3 2.7 1.0 2.3
1910-1950 0.0 3.5 4.9 1.4 2.2 2.2 2.7
1950-1975 0.4 5.1 4.4 3.8 4.3 4.0 4.3
1975-2000 -0.4 1.9 2.6 -0.8 2.2 0.2 1.8
1800-2000 0.9 3.8 3.7 1.8 2.7 1.7 2.6

Source: See Table 2.

Note: Private services are exclusive of dwelling services.

Growth and Transformation in the Agricultural Society of the Early Nineteenth Century

During the first half of the nineteenth century the agricultural sector and the rural society dominated the Swedish economy. Thus, more than three-quarters of the population were occupied in agriculture while roughly 90 percent lived in the countryside. Many non-agrarian activities such as the iron industry, the saw mill industry and many crafts as well as domestic, religious and military services were performed in rural areas. Although growth was slow, a number of structural and institutional changes occurred that paved the way for future modernization.

Most important was the transformation of agriculture. From the late eighteenth century commercialization of the primary sector intensified. Particularly during the Napoleonic Wars, the domestic market for food stuffs widened. The population increase in combination with the temporary decrease in imports stimulated enclosures and reclamation of land, the introduction of new crops and new methods and above all it stimulated a greater degree of market orientation. In the decades after the war the traditional Swedish trade deficit in grain even shifted to a trade surplus with an increasing exportation of oats, primarily to Britain.

Concomitant with the agricultural transformation were a number of infrastructural and institutional changes. Domestic transportation costs were reduced through investments in canals and roads. Trade of agricultural goods was liberalized, reducing transaction costs and integrating the domestic market even further. Trading companies became more effective in attracting agricultural surpluses for more distant markets. In support of the agricultural sector new means of information were introduced by, for example, agricultural societies that published periodicals on innovative methods and on market trends. Mortgage societies were established to supply agriculture with long term capital for investments that in turn intensified the commercialization of production.

All these elements meant a profound institutional change in the sense that the price mechanism became much more effective in directing human behavior. Furthermore, a greater interest in information and in the main instrument of information, namely literacy, was infused. Traditionally, popular literacy had been upheld by the church, mainly devoted to knowledge of the primary Lutheran texts. In the new economic environment, literacy was secularized and transformed into a more functional literacy marked by the advent of schools for public education in the 1840s.

The Breakthrough of Modern Economic Growth in the Mid-nineteenth Century

In the decades around the middle of the nineteenth century new dynamic forces appeared that accelerated growth. Most notably foreign trade expanded by leaps and bounds in the 1850s and 1860s. With new export sectors, industrial investments increased. Furthermore, railways became the most prominent component of a new infrastructure and with this construction a new component in Swedish growth was introduced, heavy capital imports.

The upswing in industrial growth in Western Europe during the 1850s, in combination with demand induced through the Crimean War, led to a particularly strong expansion in Swedish exports with sharp price increases for three staple goods – bar iron, wood and oats. The charcoal-based Swedish bar iron had been the traditional export good and had completely dominated Swedish exports until mid-nineteenth century. Bar iron met, however, increasingly strong competition from British and continental iron and steel industries and Swedish exports had stagnated in the first half of the nineteenth century. The upswing in international demand, following the diffusion of industrialization and railway construction, gave an impetus to the modernization of Swedish steel production in the following decades.

The saw mill industry was a really new export industry that grew dramatically in the 1850s and 1860s. Up until this time, the vast forests in Sweden had been regarded mainly as a fuel resource for the iron industry and for household heating and local residential construction. With sharp price increases on the Western European market from the 1840s and 1850s, the resources of the sparsely populated northern part of Sweden suddenly became valuable. A formidable explosion of saw mill construction at the mouths of the rivers along the northern coastline followed. Within a few decades Swedish merchants, as well as Norwegian, German, British and Dutch merchants, became saw mill owners running large-scale capitalist enterprises at the fringe of the European civilization.

Less dramatic but equally important was the sudden expansion of Swedish oat exports. The market for oats appeared mainly in Britain, where short-distance transportation in rapidly growing urban centers increased the fleet of horses. Swedish oats became an important energy resource during the decades around the mid-nineteenth century. In Sweden this had a special significance since oats could be cultivated on rather barren and marginal soils and Sweden was richly endowed with such soils. Thus, the market for oats with strongly increasing prices stimulated further the commercialization of agriculture and the diffusion of new methods. It was furthermore so since oats for the market were a substitute for local flax production – also thriving on barren soils – while domestic linen was increasingly supplanted by factory-produced cotton goods.

The Swedish economy was able to respond to the impetus from Western Europe during these decades, to diffuse the new influences in the economy and to integrate them in its development very successfully. The barriers to change seem to have been weak. This is partly explained by the prior transformation of agriculture and the evolution of market institutions in the rural economy. People reacted to the price mechanism. New social classes of commercial peasants, capitalists and wage laborers had emerged in an era of domestic market expansion, with increased regional specialization, and population increase.

The composition of export goods also contributed to the diffusion of participation and to the diffusion of export income. Iron, wood and oats meant both a regional and a social distribution. The value of prior marginal resources such as soils in the south and forests in the north was inflated. The technology was simple and labor intensive in industry, forestry, agriculture and transportation. The demand for unskilled labor increased strongly that was to put an imprint upon Swedish wage development in the second half of the nineteenth century. Commercial houses and industrial companies made profits but export income was distributed to many segments of the population.

The integration of the Swedish economy was further enforced through initiatives taken by the State. The parliament decision in the 1850s to construct the railway trunk lines meant, first, a more direct involvement by the State in the development of a modern infrastructure and, second, new principles of finance since the State had to rely upon capital imports. At the same time markets for goods, labor and capital were liberalized and integration both within Sweden and with the world market deepened. The Swedish adoption of the Gold Standard in 1873 put a final stamp on this institutional development.

A Second Industrial Revolution around 1900

In the late nineteenth century, particularly in the 1880s, international competition became fiercer for agriculture and early industrial branches. The integration of world markets led to falling prices and stagnation in the demand for Swedish staple goods such as iron, sawn wood and oats. Profits were squeezed and expansion thwarted. On the other hand there arose new markets. Increasing wages intensified mechanization both in agriculture and in industry. The demand increased for more sophisticated machinery equipment. At the same time consumer demand shifted towards better foodstuff – such as milk, butter and meat – and towards more fabricated industrial goods.

The decades around the turn of the twentieth century meant a profound structural change in the composition of Swedish industrial expansion that was crucial for long term growth. New and more sophisticated enterprises were founded and expanded particularly from the 1890s, in the upswing after the Baring Crisis.

The new enterprises were closely related to the so called Second Industrial Revolution in which scientific knowledge and more complex engineering skills were main components. The electrical motor became especially important in Sweden. A new development block was created around this innovation that combined engineering skills in companies such as ASEA (later ABB) with a large demand in energy-intensive processes and with the large supply of hydropower in Sweden.4 Financing the rapid development of this large block engaged commercial banks, knitting closer ties between financial capital and industry. The State, once again, engaged itself in infrastructural development in support of electrification, still resorting to heavy capital imports.

A number of innovative industries were founded in this period – all related to increased demand for mechanization and engineering skills. Companies such as AGA, ASEA, Ericsson, Separator (AlfaLaval) and SKF have been labeled “enterprises of genius” and all are represented with renowned inventors and innovators. This was, of course, not an entirely Swedish phenomenon. These branches developed simultaneously on the Continent, particularly in nearby Germany and in the U.S. Knowledge and innovative stimulus was diffused among these economies. The question is rather why this new development became so strong in Sweden so that new industries within a relatively short period of time were able to supplant old resource-based industries as main driving forces of industrialization.

Traditions of engineering skills were certainly important, developed in old heavy industrial branches such as iron and steel industries and stimulated further by State initiatives such as railway construction or, more directly, the founding of the Royal Institute of Technology. But apart from that the economic development in the second half of the nineteenth century fundamentally changed relative factor prices and the profitability of allocation of resources in different lines of production.

The relative increase in the wages of unskilled labor had been stimulated by the composition of early exports in Sweden. This was much reinforced by two components in the further development – emigration and capital imports.

Within approximately the same period, 1850-1910, the Swedish economy received a huge amount of capital mainly from Germany and France, while delivering an equally huge amount of labor to primarily the U.S. Thus, Swedish relative factor prices changed dramatically. Swedish interest rates remained at rather high levels compared to leading European countries until 1910, due to a continuous large demand for capital in Sweden, but relative wages rose persistently (see Table 5). As in the rest of Scandinavia, wage increases were much stronger than GDP growth in Sweden indicating a shift in income distribution in favor of labor, particularly in favor of unskilled labor, during this period of increased world market integration.

Table 5 Annual Increase in Real Wages of Unskilled Labor and Annual GDP Growth per Capita, 1870-1910

Country Annual real wage increase, 1870-1910 Annual GDP growth per capita, 1870-1910
Sweden 2.8 1.7
Denmark and Norway 2.6 1.3
France, Germany and Great Britain 1.1 1.2
United States 1.1 1.6

Sources: Wages from Williamson (1995); GDP growth see Table 1.

Relative profitability fell in traditional industries, which exploited rich natural resources and cheap labor, while more sophisticated industries were favored. But the causality runs both ways. Had this structural shift with the growth of new and more profitable industries not occurred, the Swedish economy would not have been able to sustain the wage increase.5

Accelerated Growth in the War-stricken Period, 1910-1950

The most notable feature of long term Swedish growth is the acceleration in growth rates during the period 1910-1950, which in Europe at large was full of problems and catastrophes.6 Thus, Swedish per capita production grew at 2.2 percent annually while growth in the rest of Scandinavia was somewhat below 2 percent and in the rest of Europe hovered at 1 percent. The Swedish acceleration was based mainly on three pillars.

First, the structure created at the end of the nineteenth century was very viable, with considerable long term growth potential. It consisted of new industries and new infrastructures that involved industrialists and financial capitalists, as well as public sector support. It also involved industries meeting a relatively strong demand in war times, as well as in the interwar period, both domestically and abroad.

Second, the First World War meant an immense financial bonus to the Swedish market. A huge export surplus at inflated prices during the war led to the domestication of the Swedish national debt. This in turn further capitalized the Swedish financial market, lowering interest rates and ameliorating sequential innovative activity in industry. A domestic money market arose that provided the State with new instruments for economic policy that were to become important for the implementation of the new social democratic “Keynesian” policies of the 1930s.

Third, demographic development favored the Swedish economy in this period. The share of the economically active age group 15-64 grew substantially. This was due partly to the fact that prior emigration had sized down cohorts that now would have become old age pensioners. Comparatively low mortality of young people during the 1910s, as well as an end to mass emigration further enhanced the share of the active population. Both the labor market and domestic demand was stimulated in particular during the 1930s when the household forming age group of 25-30 years increased.

The augmented labor supply would have increased unemployment had it not been combined with the richer supply of capital and innovative industrial development that met elastic demand both domestically and in Europe.

Thus, a richer supply of both capital and labor stimulated the domestic market in a period when international market integration deteriorated. Above all it stimulated the development of mass production of consumption goods based upon the innovations of the Second Industrial Revolution. Significant new enterprises that emanated from the interwar period were very much related to the new logic of the industrial society, such as Volvo, SAAB, Electrolux, Tetra Pak and IKEA.

The Golden Age of Growth, 1950-1975

The Swedish economy was clearly part of the European Golden Age of growth, although Swedish acceleration from the 1950s was less pronounced than in the rest of Western Europe, which to a much larger extent had been plagued by wars and crises.7 The Swedish post-war period was characterized primarily by two phenomena – the full fruition of development blocks based upon the great innovations of the late nineteenth century (the electrical motor and the combustion engine) and the cementation of the “Swedish Model” for the welfare state. These two phenomena were highly complementary.

The Swedish Model had basically two components. One was a greater public responsibility for social security and for the creation and preservation of human capital. This led to a rapid increase in the supply of public services in the realms of education, health and children’s day care as well as to increases in social security programs and in public savings for transfers to pensioners program. The consequence was high taxation. The other component was a regulation of labor and capital markets. This was the most ingenious part of the model, constructed to sustain growth in the industrial society and to increase equality in combination with the social security program and taxation.

The labor market program was the result of negotiations between trade unions and the employers’ organization. It was labeled “solidaristic wage policy” with two elements. One was to achieve equal wages for equal work, regardless of individual companies’ ability to pay. The other element was to raise the wage level in low paid areas and thus to compress the wage distribution. The aim of the program was actually to increase the speed in the structural rationalization of industries and to eliminate less productive companies and branches. Labor should be transferred to the most productive export-oriented sectors. At the same time income should be distributed more equally. A drawback of the solidaristic wage policy from an egalitarian point of view was that profits soared in the productive sectors since wage increases were held back. However, capital market regulations hindered the ability of high profits to be converted into very high incomes for shareholders. Profits were taxed very low if they were converted into further investments within the company (the timing in the use of the funds was controlled by the State in its stabilization policy) but taxed heavily if distributed to share holders. The result was that investments within existing profitable companies were supported and actually subsidized while the mobility of capital dwindled and the activity at the stock market fell.

As long as the export sectors grew, the program worked well.8 Companies founded in the late nineteenth century and in the interwar period developed into successful multinationals in engineering with machinery, auto industries and shipbuilding, as well as in resource-based industries of steel and paper. The expansion of the export sector was the main force behind the high growth rates and the productivity increases but the sector was strongly supported by public investments or publicly subsidized investments in infrastructure and residential construction.

Hence, during the Golden Age of growth the development blocks around electrification and motorization matured in a broad modernization of the society, where mass consumption and mass production was supported by social programs, by investment programs and by labor market policy.

Crisis and Restructuring from the 1970s

In the 1970s and early 1980s a number of industries – such as steel works, pulp and paper, shipbuilding, and mechanical engineering – ran into crisis. New global competition, changing consumer behavior and profound innovative renewal, especially in microelectronics, made some of the industrial pillars of the Swedish Model crumble. At the same time the disadvantages of the old model became more apparent. It put obstacles to flexibility and to entrepreneurial initiatives and it reduced individual incentives for mobility. Thus, while the Swedish Model did foster rationalization of existing industries well adapted to the post-war period, it did not support more profound transformation of the economy.

One should not exaggerate the obstacles to transformation, though. The Swedish economy was still very open in the market for goods and many services, and the pressure to transform increased rapidly. During the 1980s a far-reaching structural change within industry as well as in economic policy took place, engaging both private and public actors. Shipbuilding was almost completely discontinued, pulp industries were integrated into modernized paper works, the steel industry was concentrated and specialized, and the mechanical engineering was digitalized. New and more knowledge-intensive growth industries appeared in the 1980s, such as IT-based telecommunication, pharmaceutical industries, and biotechnology, as well as new service industries.

During the 1980s some of the constituent components of the Swedish model were weakened or eliminated. Centralized negotiations and solidaristic wage policy disappeared. Regulations in the capital market were dismantled under the pressure of increasing international capital flows simultaneously with a forceful revival of the stock market. The expansion of public sector services came to an end and the taxation system was reformed with a reduction of marginal tax rates. Thus, Swedish economic policy and welfare system became more adapted to the main European level that facilitated the Swedish application of membership and final entrance into the European Union in 1995.

It is also clear that the period from the 1970s to the early twenty-first century comprise two growth trends, before and after 1990 respectively. During the 1970s and 1980s, growth in Sweden was very slow and marked by the great structural problems that the Swedish economy had to cope with. The slow growth prior to 1990 does not signify stagnation in a real sense, but rather the transformation of industrial structures and the reformulation of economic policy, which did not immediately result in a speed up of growth but rather in imbalances and bottle necks that took years to eliminate. From the 1990s up to 2005 Swedish growth accelerated quite forcefully in comparison with most Western economies.9 Thus, the 1980s may be considered as a Swedish case of “the productivity paradox,” with innovative renewal but with a delayed acceleration of productivity and growth from the 1990s – although a delayed productivity effect of more profound transformation and radical innovative behavior is not paradoxical.

Table 6 Annual Growth Rates per Capita, 1971-2005

Period Sweden Rest of Nordic Countries Rest of Western Europe United States World Economy
1971/1975-1991/1995 1.2 2.1 1.8 1.6 1.4
1991/1995-2001/2005 2.4 2.5 1.7 2.1 2.1

Sources: See Table 1.

The recent acceleration in growth may also indicate that some of the basic traits from early industrialization still pertain to the Swedish economy – an international attitude in a small open economy fosters transformation and adaptation of human skills to new circumstances as a major force behind long term growth.

References

Abramovitz, Moses. “Catching Up, Forging Ahead and Falling Behind.” Journal of Economic History 46, no. 2 (1986): 385-406.

Dahmén, Erik. “Development Blocks in Industrial Economics.” Scandinavian Economic History Review 36 (1988): 3-14.

David, Paul A. “The Dynamo and the Computer: An Historical Perspective on the Modern Productivity Paradox.” American Economic Review 80, no. 2 (1980): 355-61.

Eichengreen, Barry. “Institutions and Economic Growth: Europe after World War II.” In Economic Growth in Europe since 1945, edited by Nicholas Crafts and Gianni Toniolo. New York: Cambridge University Press, 1996.

Krantz, Olle and Lennart Schön. Swedish Historical National Accounts, 1800-2000. Lund: Almqvist and Wiksell International (forthcoming, 2007).

Maddison, Angus. The World Economy, Volumes 1 and 2. Paris: OECD (2006).

Schön, Lennart. “Development Blocks and Transformation Pressure in a Macro-Economic Perspective: A Model of Long-Cyclical Change.” Skandinaviska Enskilda Banken Quarterly Review 20, no. 3-4 (1991): 67-76.

Schön, Lennart. “External and Internal Factors in Swedish Industrialization.” Scandinavian Economic History Review 45, no. 3 (1997): 209-223.

Schön, Lennart. En modern svensk ekonomisk historia: Tillväxt och omvandling under två sekel (A Modern Swedish Economic History: Growth and Transformation in Two Centuries). Stockholm: SNS (2000).

Schön, Lennart. “Total Factor Productivity in Swedish Manufacturing in the Period 1870-2000.” In Exploring Economic Growth: Essays in Measurement and Analysis: A Festschrift for Riitta Hjerppe on Her Sixtieth Birthday, edited by S. Heikkinen and J.L. van Zanden. Amsterdam: Aksant, 2004.

Schön, Lennart. “Swedish Industrialization 1870-1930 and the Heckscher-Ohlin Theory.” In Eli Heckscher, International Trade, and Economic History, edited by Ronald Findlay et al. Cambridge, MA: MIT Press (2006).

Svennilson, Ingvar. Growth and Stagnation in the European Economy. Geneva: United Nations Economic Commission for Europe, 1954.

Temin, Peter. “The Golden Age of European Growth Reconsidered.” European Review of Economic History 6, no. 1 (2002): 3-22.

Williamson, Jeffrey G. “The Evolution of Global Labor Markets since 1830: Background Evidence and Hypotheses.” Explorations in Economic History 32, no. 2 (1995): 141-96.

Citation: Schön, Lennart. “Sweden – Economic Growth and Structural Change, 1800-2000″. EH.Net Encyclopedia, edited by Robert Whaples. February 10, 2008. URL http://eh.net/encyclopedia/sweden-economic-growth-and-structural-change-1800-2000/

The Economics of the American Revolutionary War

Ben Baack, Ohio State University

By the time of the onset of the American Revolution, Britain had attained the status of a military and economic superpower. The thirteen American colonies were one part of a global empire generated by the British in a series of colonial wars beginning in the late seventeenth century and continuing on to the mid eighteenth century. The British military establishment increased relentlessly in size during this period as it engaged in the Nine Years War (1688-97), the War of Spanish Succession (1702-13), the War of Austrian Succession (1739-48), and the Seven Years War (1756-63). These wars brought considerable additions to the British Empire. In North America alone the British victory in the Seven Years War resulted in France ceding to Britain all of its territory east of the Mississippi River as well as all of Canada and Spain surrendering its claim to Florida (Nester, 2000).

Given the sheer magnitude of the British military and its empire, the actions taken by the American colonists for independence have long fascinated scholars. Why did the colonists want independence? How were they able to achieve a victory over what was at the time the world’s preeminent military power? What were the consequences of achieving independence? These and many other questions have engaged the attention of economic, legal, military, political, and social historians. In this brief essay we will focus only on the economics of the Revolutionary War.

Economic Causes of the Revolutionary War

Prior to the conclusion of the Seven Years War there was little, if any, reason to believe that one day the American colonies would undertake a revolution in an effort to create an independent nation-state. As apart of the empire the colonies were protected from foreign invasion by the British military. In return, the colonists paid relatively few taxes and could engage in domestic economic activity without much interference from the British government. For the most part the colonists were only asked to adhere to regulations concerning foreign trade. In a series of acts passed by Parliament during the seventeenth century the Navigation Acts required that all trade within the empire be conducted on ships which were constructed, owned and largely manned by British citizens. Certain enumerated goods whether exported or imported by the colonies had to be shipped through England regardless of the final port of destination.

Western Land Policies

Economic incentives for independence significantly increased in the colonies as a result of a series of critical land policy decisions made by the British government. The Seven Years’ War had originated in a contest between Britain and France over control of the land from the Appalachian Mountains to the Mississippi River. During the 1740s the British government pursued a policy of promoting colonial land claims to as well as settlement in this area, which was at the time French territory. With the ensuing conflict of land claims both nations resorted to the use of military force which ultimately led to the onset of the war. At the conclusion of the war as a result of one of many concessions made by France in the 1763 Treaty of Paris, Britain acquired all the contested land west of its colonies to the Mississippi River. It was at this point that the British government began to implement a fundamental change in its western land policy.

Britain now reversed its long-time position of encouraging colonial claims to land and settlement in the west. The essence of the new policy was to establish British control of the former French fur trade in the west by excluding any settlement there by the Americans. Implementation led to the development of three new areas of policy. 1. Construction of the new rules of exclusion. 2. Enforcement of the new exclusion rules. 3. Financing the cost of the enforcement of the new rules. First, the rules of exclusion were set out under the terms of the Proclamation of 1763 whereby colonists were not allowed to settle in the west. This action legally nullified the claims to land in the area by a host of individual colonists, land companies, as well as colonies. Second, enforcement of the new rules was delegated to the standing army of about 7,500 regulars newly stationed in the west. This army for the most part occupied former French forts although some new ones were built. Among other things, this army was charged with keeping Americans out of the west as well as returning to the colonies any Americans who were already there. Third, financing of the cost of the enforcement was to be accomplished by levying taxes on the Americans. Thus, Americans were being asked to finance a British army which was charged with keeping Americans out of the west (Baack, 2004).

Tax Policies

Of all the potential options available for funding the new standing army in the west, why did the British decide to tax their American colonies? The answer is fairly straightforward. First of all, the victory over the French in the Seven Years’ War had come at a high price. Domestic taxes had been raised substantially during the war and total government debt had increased nearly twofold (Brewer, 1989). In addition, taxes were significantly higher in Britain than in the colonies. One estimate suggests the per capita tax burden in the colonies ranged from two to four percent of that in Britain (Palmer, 1959). And finally, the voting constituencies of the members of parliament were in Britain not the colonies. All things considered, Parliament viewed taxing the colonies as the obvious choice.

Accordingly, a series of tax acts were passed by Parliament the revenue from which was to be used to help pay for the standing army in America. The first was the Sugar Act of 1764. Proposed by England’s Prime Minister the act lowered tariff rates on non-British products from the West Indies as well as strengthened their collection. It was hoped this would reduce the incentive for smuggling and thereby increase tariff revenue (Bullion, 1982). The following year Parliament passed the Stamp Act that imposed a tax commonly used in England. It required stamps for a broad range of legal documents as well as newspapers and pamphlets. While the colonial stamp duties were less than those in England they were expected to generate enough revenue to finance a substantial portion of the cost the new standing army. The same year passage of the Quartering Act imposed essentially a tax in kind by requiring the colonists to provide British military units with housing, provisions, and transportation. In 1767 the Townshend Acts imposed tariffs upon a variety of imported goods and established a Board of Customs Commissioners in the colonies to collect the revenue.

Boycotts

While the Americans could do little about the British army stationed in the west, they could do somthing about the new British taxes. American opposition to these acts was expressed initially in a variety of peaceful forms. While they did not have representation in Parliament, the colonists did attempt to exert some influence in it through petition and lobbying. However, it was the economic boycott that became by far the most effective means of altering the new British economic policies. In 1765 representatives from nine colonies met at the Stamp Act Congress in New York and organized a boycott of imported English goods. The boycott was so successful in reducing trade that English merchants lobbied Parliament for the repeal of the new taxes. Parliament soon responded to the political pressure. During 1766 it repealed both the Stamp and Sugar Acts (Johnson, 1997). In response to the Townshend Acts of 1767 a second major boycott started in 1768 in Boston and New York and subsequently spread to other cities leading Parliament in 1770 to repeal all of the Townshend duties except the one on tea. In addition, Parliament decided at the same time not to renew the Quartering Act.

With these actions taken by Parliament the Americans appeared to have successfully overturned the new British post war tax agenda. However, Parliament had not given up what it believed to be its right to tax the colonies. On the same day it repealed the Stamp Act, Parliament passed the Declaratory Act stating the British government had the full power and authority to make laws governing the colonies in all cases whatsoever including taxation. Legislation not principles had been overturned.

The Tea Act

Three years after the repeal of the Townshend duties British policy was once again to emerge as an issue in the colonies. This time the American reaction was not peaceful. It all started when Parliament for the first time granted an exemption from the Navigation Acts. In an effort to assist the financially troubled British East India Company Parliament passed the Tea Act of 1773, which allowed the company to ship tea directly to America. The grant of a major trading advantage to an already powerful competitor meant a potential financial loss for American importers and smugglers of tea. In December a small group of colonists responded by boarding three British ships in the Boston harbor and throwing overboard several hundred chests of tea owned by the East India Company (Labaree, 1964). Stunned by the events in Boston, Parliament decided not to cave in to the colonists as it had before. In rapid order it passed the Boston Port Act, the Massachusetts Government Act, the Justice Act, and the Quartering Act. Among other things these so-called Coercive or Intolerable Acts closed the port of Boston, altered the charter of Massachusetts, and reintroduced the demand for colonial quartering of British troops. Once done Parliament then went on to pass the Quebec Act as a continuation of its policy of restricting the settlement of the West.

The First Continental Congress

Many Americans viewed all of this as a blatant abuse of power by the British government. Once again a call went out for a colonial congress to sort out a response. On September 5, 1774 delegates appointed by the colonies met in Philadelphia for the First Continental Congress. Drawing upon the successful manner in which previous acts had been overturned the first thing Congress did was to organize a comprehensive embargo of trade with Britain. It then conveyed to the British government a list of grievances that demanded the repeal of thirteen acts of Parliament. All of the acts listed had been passed after 1763 as the delegates had agreed not to question British policies made prior to the conclusion of the Seven Years War. Despite all the problems it had created, the Tea Act was not on the list. The reason for this was that Congress decided not to protest British regulation of colonial trade under the Navigation Acts. In short, the delegates were saying to Parliament take us back to 1763 and all will be well.

The Second Continental Congress

What happened then was a sequence of events that led to a significant increase in the degree of American resistance to British polices. Before the Congress adjourned in October the delegates voted to meet again in May of 1775 if Parliament did not meet their demands. Confronted by the extent of the American demands the British government decided it was time to impose a military solution to the crisis. Boston was occupied by British troops. In April a military confrontation occurred at Lexington and Concord. Within a month the Second Continental Congress was convened. Here the delegates decided to fundamentally change the nature of their resistance to British policies. Congress authorized a continental army and undertook the purchase of arms and munitions. To pay for all of this it established a continental currency. With previous political efforts by the First Continental Congress to form an alliance with Canada having failed, the Second Continental Congress took the extraordinary step of instructing its new army to invade Canada. In effect, these actions taken were those of an emerging nation-state. In October as American forces closed in on Quebec the King of England in a speech to Parliament declared that the colonists having formed their own government were now fighting for their independence. It was to be only a matter of months before Congress formally declared it.

Economic Incentives for Pursuing Independence: Taxation

Given the nature of British colonial policies, scholars have long sought to evaluate the economic incentives the Americans had in pursuing independence. In this effort economic historians initially focused on the period following the Seven Years War up to the Revolution. It turned out that making a case for the avoidance of British taxes as a major incentive for independence proved difficult. The reason was that many of the taxes imposed were later repealed. The actual level of taxation appeared to be relatively modest. After all, the Americans soon after adopting the Constitution taxed themselves at far higher rates than the British had prior to the Revolution (Perkins, 1988). Rather it seemed the incentive for independence might have been the avoidance of the British regulation of colonial trade. Unlike some of the new British taxes, the Navigation Acts had remained intact throughout this period.

The Burden of the Navigation Acts

One early attempt to quantify the economic effects of the Navigation Acts was by Thomas (1965). Building upon the previous work of Harper (1942), Thomas employed a counterfactual analysis to assess what would have happened to the American economy in the absence of the Navigation Acts. To do this he compared American trade under the Acts with that which would have occurred had America been independent following the Seven Years War. Thomas then estimated the loss of both consumer and produce surplus to the colonies as a result of shipping enumerated goods indirectly through England. These burdens were partially offset by his estimated value of the benefits of British protection and various bounties paid to the colonies. The outcome of his analysis was that the Navigation Acts imposed a net burden of less than one percent of colonial per capita income. From this he concluded the Acts were an unlikely cause of the Revolution. A long series of subsequent works questioned various parts of his analysis but not his general conclusion (Walton, 1971). The work of Thomas also appeared to be consistent with the observation that the First Continental Congress had not demanded in its list of grievances the repeal of either the Navigation Acts or the Sugar Act.

American Expectations about Future British Policy

Did this mean then that the Americans had few if any economic incentives for independence? Upon further consideration economic historians realized that perhaps more important to the colonists were not the past and present burdens but rather the expected future burdens of continued membership in the British Empire. The Declaratory Act made it clear the British government had not given up what it viewed as its right to tax the colonists. This was despite the fact that up to 1775 the Americans had employed a variety of protest measures including lobbying, petitions, boycotts, and violence. The confluence of not having representation in Parliament while confronting an aggressive new British tax policy designed to raise their relatively low taxes may have made it reasonable for the Americans to expect a substantial increase in the level of taxation in the future (Gunderson, 1976, Reid, 1978). Furthermore a recent study has argued that in 1776 not only did the future burdens of the Navigation Acts clearly exceed those of the past, but a substantial portion would have borne by those who played a major role in the Revolution (Sawers, 1992). Seen in this light the economic incentive for independence would have been avoiding the potential future costs of remaining in the British Empire.

The Americans Undertake a Revolution

1776-77

British Military Advantages

The American colonies had both strengths and weaknesses in terms of undertaking a revolution. The colonial population of well over two million was nearly one third of that in Britain (McCusker and Menard, 1985). The growth in the colonial economy had generated a remarkably high level of per capita wealth and income (Jones, 1980). Yet the hurdles confronting the Americans in achieving independence were indeed formidable. The British military had an array of advantages. With virtual control of the Atlantic its navy could attack anywhere along the American coast at will and would have borne logistical support for the army without much interference. A large core of experienced officers commanded a highly disciplined and well-drilled army in the large-unit tactics of eighteenth century European warfare. By these measures the American military would have great difficulty in defeating the British. Its navy was small. The Continental Army had relatively few officers proficient in large-unit military tactics. Lacking both the numbers and the discipline of its adversary the American army was unlikely to be able to meet the British army on equal terms on the battlefield (Higginbotham, 1977).

British Financial Advantages

In addition, the British were in a better position than the Americans to finance a war. A tax system was in place that had provided substantial revenue during previous colonial wars. Also for a variety of reasons the government had acquired an exceptional capacity to generate debt to fund wartime expenses (North and Weingast, 1989). For the Continental Congress the situation was much different. After declaring independence Congress had set about defining the institutional relationship between it and the former colonies. The powers granted to Congress were established under the Articles of Confederation. Reflecting the political environment neither the power to tax nor the power to regulate commerce was given to Congress. Having no tax system to generate revenue also made it very difficult to borrow money. According to the Articles the states were to make voluntary payments to Congress for its war efforts. This precarious revenue system was to hamper funding by Congress throughout the war (Baack, 2001).

Military and Financial Factors Determine Strategy

It was within these military and financial constraints that the war strategies by the British and the Americans were developed. In terms of military strategies both of the contestants realized that America was simply too large for the British army to occupy all of the cities and countryside. This being the case the British decided initially that they would try to impose a naval blockade and capture major American seaports. Having already occupied Boston, the British during 1776 and 1777 took New York, Newport, and Philadelphia. With plenty of room to maneuver his forces and unable to match those of the British, George Washington chose to engage in a war of attrition. The purpose was twofold. First, by not engaging in an all out offensive Washington reduced the probability of losing his army. Second, over time the British might tire of the war.

Saratoga

Frustrated without a conclusive victory, the British altered their strategy. During 1777 a plan was devised to cut off New England from the rest of the colonies, contain the Continental Army, and then defeat it. An army was assembled in Canada under the command of General Burgoyne and then sent to and down along the Hudson River. It was to link up with an army sent from New York City. Unfortunately for the British the plan totally unraveled as in October Burgoyne’s army was defeated at the battle of Saratoga and forced to surrender (Ketchum, 1997).

The American Financial Situation Deteriorates

With the victory at Saratoga the military side of the war had improved considerably for the Americans. However, the financial situation was seriously deteriorating. The states to this point had made no voluntary payments to Congress. At the same time the continental currency had to compete with a variety of other currencies for resources. The states were issuing their own individual currencies to help finance expenditures. Moreover the British in an effort to destroy the funding system of the Continental Congress had undertaken a covert program of counterfeiting the Continental dollar. These dollars were printed and then distributed throughout the former colonies by the British army and agents loyal to the Crown (Newman, 1957). Altogether this expansion of the nominal money supply in the colonies led to a rapid depreciation of the Continental dollar (Calomiris, 1988, Michener, 1988). Furthermore, inflation may have been enhanced by any negative impact upon output resulting from the disruption of markets along with the destruction of property and loss of able-bodied men (Buel, 1998). By the end of 1777 inflation had reduced the specie value of the Continental to about twenty percent of what it had been when originally issued. This rapid decline in value was becoming a serious problem for Congress in that up to this point almost ninety percent of its revenue had been generated from currency emissions.

1778-83

British Invasion of the South

The British defeat at Saratoga had a profound impact upon the nature of the war. The French government still upset by their defeat by the British in the Seven Years War and encouraged by the American victory signed a treaty of alliance with the Continental Congress in early 1778. Fearing a new war with France the British government sent a commission to negotiate a peace treaty with the Americans. The commission offered to repeal all of the legislation applying to the colonies passed since 1763. Congress rejected the offer. The British response was to give up its efforts to suppress the rebellion in the North and in turn organize an invasion of the South. The new southern campaign began with the taking of the port of Savannah in December. Pursuing their southern strategy the British won major victories at Charleston and Camden during the spring and summer of 1780.

Worsening Inflation and Financial Problems

As the American military situation deteriorated in the South so did the financial circumstances of the Continental Congress. Inflation continued as Congress and the states dramatically increased the rate of issuance of their currencies. At the same time the British continued to pursue their policy of counterfeiting the Continental dollar. In order to deal with inflation some states organized conventions for the purpose of establishing wage and price controls (Rockoff, 1984). With few contributions coming from the states and a currency rapidly losing its value, Congress resorted to authorizing the army to confiscate whatever it needed to continue the war effort (Baack, 2001, 2008).

Yorktown

Fortunately for the Americans the British military effort collapsed before the funding system of Congress. In a combined effort during the fall of 1781 French and American forces trapped the British southern army under the command of Cornwallis at Yorktown, Virginia. Under siege by superior forces the British army surrendered on October 19. The British government had now suffered not only the defeat of its northern strategy at Saratoga but also the defeat of its southern campaign at Yorktown. Following Yorktown, Britain suspended its offensive military operations against the Americans. The war was over. All that remained was the political maneuvering over the terms for peace.

The Treaty of Paris

The Revolutionary War officially concluded with the signing of the Treaty of Paris in 1783. Under the terms of the treaty the United States was granted independence and British troops were to evacuate all American territory. While commonly viewed by historians through the lens of political science, the Treaty of Paris was indeed a momentous economic achievement by the United States. The British ceded to the Americans all of the land east of the Mississippi River which they had taken from the French during the Seven Years War. The West was now available for settlement. To the extent the Revolutionary War had been undertaken by the Americans to avoid the costs of continued membership in the British Empire, the goal had been achieved. As an independent nation the United States was no longer subject to the regulations of the Navigation Acts. There was no longer to be any economic burden from British taxation.

THE FORMATION OF A NATIONAL GOVERNMENT

When you start a revolution you have to be prepared for the possibility you might win. This means being prepared to form a new government. When the Americans declared independence their experience of governing at a national level was indeed limited. In 1765 delegates from various colonies had met for about eighteen days at the Stamp Act Congress in New York to sort out a colonial response to the new stamp duties. Nearly a decade passed before delegates from colonies once again got together to discuss a colonial response to British policies. This time the discussions lasted seven weeks at the First Continental Congress in Philadelphia during the fall of 1774. The primary action taken at both meetings was an agreement to boycott trade with England. After having been in session only a month, delegates at the Second Continental Congress for the first time began to undertake actions usually associated with a national government. However, when the colonies were declared to be free and independent states Congress had yet to define its institutional relationship with the states.

The Articles of Confederation

Following the Declaration of Independence, Congress turned to deciding the political and economic powers it would be given as well as those granted to the states. After more than a year of debate among the delegates the allocation of powers was articulated in the Articles of Confederation. Only Congress would have the authority to declare war and conduct foreign affairs. It was not given the power to tax or regulate commerce. The expenses of Congress were to be made from a common treasury with funds supplied by the states. This revenue was to be generated from exercising the power granted to the states to determine their own internal taxes. It was not until November of 1777 that Congress approved the final draft of the Articles. It took over three years for the states to ratify the Articles. The primary reason for the delay was a dispute over control of land in the West as some states had claims while others did not. Those states with claims eventually agreed to cede them to Congress. The Articles were then ratified and put into effect on March 1, 1781. This was just a few months before the American victory at Yorktown. The process of institutional development had proved so difficult that the Americans fought almost the entire Revolutionary War with a government not sanctioned by the states.

Difficulties in the 1780s

The new national government that emerged from the Revolution confronted a host of issues during the 1780s. The first major one to be addressed by Congress was what to do with all of the land acquired in the West. Starting in 1784 Congress passed a series of land ordinances that provided for land surveys, sales of land to individuals, and the institutional foundation for the creation of new states. These ordinances opened the West for settlement. While this was a major accomplishment by Congress, other issues remained unresolved. Having repudiated its own currency and no power of taxation, Congress did not have an independent source of revenue to pay off its domestic and foreign debts incurred during the war. Since the Continental Army had been demobilized no protection was being provided for settlers in the West or against foreign invasion. Domestic trade was being increasingly disrupted during the 1780s as more states began to impose tariffs on goods from other states. Unable to resolve these and other issues Congress endorsed a proposed plan to hold a convention to meet in Philadelphia in May of 1787 to revise the Articles of Confederation.

Rather than amend the Articles, the delegates to the convention voted to replace them entirely with a new form of national government under the Constitution. There are of course many ways to assess the significance of this truly remarkable achievement. One is to view the Constitution as an economic document. Among other things the Constitution specifically addressed many of the economic problems that confronted Congress during and after the Revolutionary War. Drawing upon lessons learned in financing the war, no state under the Constitution would be allowed to coin money or issue bills of credit. Only the national government could coin money and regulate its value. Punishment was to be provided for counterfeiting. The problems associated with the states contributing to a common treasury under the Articles were overcome by giving the national government the coercive power of taxation. Part of the revenue was to be used to pay for the common defense of the United States. No longer would states be allowed to impose tariffs as they had done during the 1780s. The national government was now given the power to regulate both foreign and interstate commerce. As a result the nation was to become a common market. There is a general consensus among economic historians today that the economic significance of the ratification of the Constitution was to lay the institutional foundation for long run growth. From the point of view of the former colonists, however, it meant they had succeeded in transferring the power to tax and regulate commerce from Parliament to the new national government of the United States.

TABLES
Table 1 Continental Dollar Emissions (1775-1779)

Year of Emission Nominal Dollars Emitted (000) Annual Emission As Share of Total Nominal Stock Emitted Specie Value of Annual Emission (000) Annual Emission As Share of Total Specie Value Emitted
1775 $6,000 3% $6,000 15%
1776 19,000 8 15,330 37
1777 13,000 5 4,040 10
1778 63,000 26 10,380 25
1779 140,500 58 5,270 13
Total $241,500 100% $41,020 100%

Source: Bullock (1895), 135.
Table 2 Currency Emissions by the States (1775-1781)

Year of Emission Nominal Dollars Emitted (000) Year of Emission Nominal Dollars Emitted (000)
1775 $4,740 1778 $9,118
1776 13,328 1779 17,613
1777 9,573 1780 66,813
1781 123.376
Total $27,641 Total $216,376

Source: Robinson (1969), 327-28.

References

Baack, Ben. “Forging a Nation State: The Continental Congress and the Financing of the War of American Independence.” Economic History Review 54, no.4 (2001): 639-56.

Baack, Ben. “British versus American Interests in Land and the War of American Independence.” Journal of European Economic History 33, no. 3 (2004): 519-54.

Baack, Ben. “America’s First Monetary Policy: Inflation and Seigniorage during the Revolutionary War.” Financial History Review 15, no. 2 (2008): 107-21.

Baack, Ben, Robert A. McGuire, and T. Norman Van Cott. “Constitutional Agreement during the Drafting of the Constitution: A New Interpretation.” Journal of Legal Studies 38, no. 2 (2009): 533-67.

Brewer, John. The Sinews of Power: War, Money and the English State, 1688- 1783. London: Cambridge University Press, 1989.

Buel, Richard. In Irons: Britain’s Naval Supremacy and the American Revolutionary Economy. New Haven: Yale University Press, 1998.

Bullion, John L. A Great and Necessary Measure: George Grenville and the Genesis of the Stamp Act, 1763-1765. Columbia: University of Missouri Press, 1982.

Bullock, Charles J. “The Finances of the United States from 1775 to 1789, with Especial Reference to the Budget.” Bulletin of the University of Wisconsin 1, no. 2 (1895): 117-273.

Calomiris, Charles W. “Institutional Failure, Monetary Scarcity, and the Depreciation of the Continental.” Journal of Economic History 48, no. 1 (1988): 47-68.

Egnal, Mark. A Mighty Empire: The Origins of the American Revolution. Ithaca: Cornell University Press, 1988.

Ferguson, E. James. The Power of the Purse: A History of American Public Finance, 1776-1790. Chapel Hill: University of North Carolina Press, 1961.

Gunderson, Gerald. A New Economic History of America. New York: McGraw- Hill, 1976.

Harper, Lawrence A. “Mercantilism and the American Revolution.” Canadian Historical Review 23 (1942): 1-15.

Higginbotham, Don. The War of American Independence: Military Attitudes, Policies, and Practice, 1763-1789. Bloomington: Indiana University Press, 1977.

Jensen, Merrill, editor. English Historical Documents: American Colonial Documents to 1776 New York: Oxford university Press, 1969.

Johnson, Allen S. A Prologue to Revolution: The Political Career of George Grenville (1712-1770). New York: University Press, 1997.

Jones, Alice H. Wealth of a Nation to Be: The American Colonies on the Eve of the Revolution. New York: Columbia University Press, 1980.

Ketchum, Richard M. Saratoga: Turning Point of America’s Revolutionary War. New York: Henry Holt and Company, 1997.

Labaree, Benjamin Woods. The Boston Tea Party. New York: Oxford University Press, 1964.

Mackesy, Piers. The War for America, 1775-1783. Cambridge: Harvard University Press, 1964.

McCusker, John J. and Russell R. Menard. The Economy of British America, 1607- 1789. Chapel Hill: University of North Carolina Press, 1985.

Michener, Ron. “Backing Theories and the Currencies of Eighteenth-Century America: A Comment.” Journal of Economic History 48, no. 3 (1988): 682-92.

Nester, William R. The First Global War: Britain, France, and the Fate of North America, 1756-1775. Westport: Praeger, 2000.

Newman, E. P. “Counterfeit Continental Currency Goes to War.” The Numismatist 1 (January, 1957): 5-16.

North, Douglass C., and Barry R. Weingast. “Constitutions and Commitment: The Evolution of Institutions Governing Public Choice in Seventeenth-Century England.” Journal of Economic History 49 No. 4 (1989): 803-32.

O’Shaughnessy, Andrew Jackson. An Empire Divided: The American Revolution and the British Caribbean. Philadelphia: University of Pennsylvania Press, 2000.

Palmer, R. R. The Age of Democratic Revolution: A Political History of Europe and America. Vol. 1. Princeton: Princeton University Press, 1959.

Perkins, Edwin J. The Economy of Colonial America. New York: Columbia University Press, 1988.

Reid, Joseph D., Jr. “Economic Burden: Spark to the American Revolution?” Journal of Economic History 38, no. 1 (1978): 81-100.

Robinson, Edward F. “Continental Treasury Administration, 1775-1781: A Study in the Financial History of the American Revolution.” Ph.D. diss., University of Wisconsin, 1969.

Rockoff, Hugh. Drastic Measures: A History of Wage and Price Controls in the United States. Cambridge: Cambridge University Press, 1984.

Sawers, Larry. “The Navigation Acts Revisited.” Economic History Review 45, no. 2 (1992): 262-84.

Thomas, Robert P. “A Quantitative Approach to the Study of the Effects of British Imperial Policy on Colonial Welfare: Some Preliminary Findings.” Journal of Economic History 25, no. 4 (1965): 615-38.

Tucker, Robert W. and David C. Hendrickson. The Fall of the First British Empire: Origins of the War of American Independence. Baltimore: Johns Hopkins Press, 1982.

Walton, Gary M. “The New Economic History and the Burdens of the Navigation Acts.” Economic History Review 24, no. 4 (1971): 533-42.

Citation: Baack, Ben. “Economics of the American Revolutionary War”. EH.Net Encyclopedia, edited by Robert Whaples. November 13, 2001 (updated August 5, 2010). URL http://eh.net/encyclopedia/the-economics-of-the-american-revolutionary-war/

Economic History of Portugal

Luciano Amaral, Universidade Nova de Lisboa

Main Geographical Features

Portugal is the south-westernmost country of Europe. With the approximate shape of a vertical rectangle, it has a maximum height of 561 km and a maximum length of 218 km, and is delimited (in its north-south range) by the parallels 37° and 42° N, and (in its east-west range) by the meridians 6° and 9.5° W. To the west, it faces the Atlantic Ocean, separating it from the American continent by a few thousand kilometers. To the south, it still faces the Atlantic, but the distance to Africa is only of a few hundred kilometers. To the north and the east, it shares land frontiers with Spain, and both countries constitute the Iberian Peninsula, a landmass separated directly from France and, then, from the rest of the continent by the Pyrenees. Two Atlantic archipelagos are still part of Portugal, the Azores – constituted by eight islands in the same latitudinal range of mainland Portugal, but much further west, with a longitude between 25° and 31° W – and Madeira – two islands, to the southwest of the mainland, 16° and 17° W, 32.5° and 33° N.

Climate in mainland Portugal is of the temperate sort. Due to its southern position and proximity to the Mediterranean Sea, the country’s weather still presents some Mediterranean features. Temperature is, on average, higher than in the rest of the continent. Thanks to its elongated form, Portugal displays a significant variety of landscapes and sometimes brisk climatic changes for a country of such relatively small size. Following a classical division of the territory, it is possible to identify three main geographical regions: a southern half – with practically no mountains and a very hot and dry climate – and a northern half subdivided into two other vertical sub-halves – with a north-interior region, mountainous, cool but relatively dry, and a north-coast region, relatively mountainous, cool and wet. Portugal’s population is close to 10,000,000, in an area of about 92,000 square kilometers (35,500 square miles).

The Period before the Creation of Portugal

We can only talk of Portugal as a more or less clearly identified and separate political unit (although still far from a defined nation) from the eleventh or twelfth centuries onwards. The geographical area which constitutes modern Portugal was not, of course, an eventless void before that period. But scarcity of space allows only a brief examination of the earlier period, concentrating on its main legacy to future history.

Roman and Visigothic Roots

That legacy is overwhelmingly marked by the influence of the Roman Empire. Portugal owes to Rome its language (a descendant of Latin) and main religion (Catholicism), as well as its primary juridical and administrative traditions. Interestingly enough, little of the Roman heritage passed directly to the period of existence of Portugal as a proper nation. Momentous events filtered the transition. Romans first arrived in the Iberian Peninsula around the third century B.C., and kept their rule until the fifth century of the Christian era. Then, they succumbed to the so-called “barbarian invasions.” Of the various peoples that then roamed the Peninsula, certainly the most influential were the Visigoths, a people of Germanic origin. The Visigoths may be ranked as the second most important force in the shaping of future Portugal. The country owes them the monarchical institution (which lasted until the twentieth century), as well as the preservation both of Catholicism and (although substantially transformed) parts of Roman law.

Muslim Rule

The most spectacular episode following Visigoth rule was the Muslim invasion of the eighth century. Islam ruled the Peninsula from then until the fifteenth century, although occupying an increasingly smaller area from the ninth century onwards, as the Christian Reconquista started repelling it with growing efficiency. Muslim rule set the area on a path different from the rest of Western Europe for a few centuries. However, apart from some ethnic traits legated to its people, a few words in its lexicon, as well as certain agricultural, manufacturing and sailing techniques and knowledge (of which the latter had significant importance to the Portuguese naval discoveries), nothing of the magnitude of the Roman heritage was left in the peninsula by Islam. This is particularly true of Portugal, where Muslim rule was less effective and shorter than in the South of Spain. Perhaps the most important legacy of Muslim rule was, precisely, its tolerance towards the Roman heritage. Much representative of that tolerance was the existence during the Muslim period of an ethnic group, the so-called moçárabe or mozarabe population, constituted by traditional residents that lived within Muslim communities, accepted Muslim rule, and mixed with Muslim peoples, but still kept their language and religion, i.e. some form of Latin and the Christian creed.

Modern Portugal is a direct result of the Reconquista, the Christian fight against Muslim rule in the Iberian Peninsula. That successful fight was followed by the period when Portugal as a nation came to existence. The process of creation of Portugal was marked by the specific Roman-Germanic institutional synthesis that constituted the framework of most of the country’s history.

Portugal from the Late Eleventh Century to the Late Fourteenth Century

Following the Muslim invasion, a small group of Christians kept their independence, settling in a northern area of the Iberian Peninsula called Asturias. Their resistance to Muslim rule rapidly transformed into an offensive military venture. During the eighth century a significant part of northern Iberia was recovered to Christianity. This frontier, roughly cutting the peninsula in two halves, held firm until the eleventh century. Then, the crusaders came, mostly from France and Germany, inserting the area in the overall European crusade movement. By the eleventh century, the original Asturian unit had been divided into two kingdoms, Leon and Navarra, which in turn were subdivided into three new political units, Castile, Aragon and the Condado Portucalense. The Condado Portucalense (the political unit at the origin of future Portugal) resulted from a donation, made in 1096, by the Leonese king to a Crusader coming from Burgundy (France), Count Henry. He did not claim the title king, a job that would be fulfilled only by his son, Afonso Henriques (generally accepted as the first king of Portugal) in the first decade of the twelfth century.

Condado Portucalense as the King’s “Private Property”

Such political units as the various peninsular kingdoms of that time must be seen as entities differing in many respects from current nations. Not only did their peoples not possess any clear “national consciousness,” but also the kings themselves did not rule them based on the same sort of principle we tend to attribute to current rulers (either democratic, autocratic or any other sort). Both the Condado Portucalense and Portugal were understood by their rulers as something still close to “private property” – the use of quotes here is justified by the fact that private property, in the sense we give to it today, was a non-existent notion then. We must, nevertheless, stress this as the moment in which Portuguese rulers started seeing Portugal as a political unit separate from the remaining units in the area.

Portugal as a Military Venture

Such novelty was strengthened by the continuing war against Islam, still occupying most of the center and south of what later became Portugal. This is a crucial fact about Portugal in its infancy, and one that helps one understand the most important episode in Portuguese history , the naval discoveries, i.e. that the country in those days was largely a military venture against Islam. As, in that fight, the kingdom expanded to the south, it did so separately from the other Christian kingdoms existing in the peninsula. And these ended up constituting the two main negative forces for Portugal’s definition as an independent country, i.e. Islam and the remaining Iberian Christian kingdoms. The country achieved a clear geographical definition quite early in its history, more precisely in 1249, when King Sancho II conquered the Algarve from Islam. Remarkably for a continent marked by so much permanent frontier redesign, Portugal acquired then its current geographical shape.

The military nature of the country’s growth gave rise to two of its most important characteristics in early times: Portugal was throughout this entire period a frontier country, and one where the central authority was unable to fully control the territory in its entirety. This latter fact, together with the reception of the Germanic feudal tradition, shaped the nature of the institutions then established in the country. This was particularly important in understanding the land donations made by the crown. These were crucial, for they brought a dispersion of central powers, devolved to local entities, as well as a delegation of powers we would today call “public” to entities we would call “private.” Donations were made in favor of three sorts of groups: noble families, religious institutions and the people in general of particular areas or cities. They resulted mainly from the needs of the process of conquest: noblemen were soldiers, and the crown’s concession of the control of a certain territory was both a reward for their military feats as well as an expedient way of keeping the territory under control (even if in a more indirect way) in a period when it was virtually impossible to directly control the full extent of the conquered area. Religious institutions were crucial in the Reconquista, since the purpose of the whole military effort was to eradicate the Muslim religion from the country. Additionally, priests and monks were full military participants in the process, not limiting their activity to studying or preaching. So, as the Reconquista proceeded, three sorts of territories came into existence: those under direct control of the crown, those under the control of local seigneurs (which subdivided into civil and ecclesiastical) and the communities.

Economic Impact of the Military Institutional Framework

This was an institutional framework that had a direct economic impact. The crown’s donations were not comparable to anything we would nowadays call private property. The land’s donation had attached to it the ability conferred on the beneficiary to a) exact tribute from the population living in it, b) impose personal services or reduce peasants to serfdom, and c) administer justice. This is a phenomenon that is typical of Europe until at least the eighteenth century, and is quite representative of the overlap between the private and public spheres then prevalent. The crown felt it was entitled to give away powers we would nowadays call public, such as those of taxation and administering justice, and beneficiaries from the crown’s donations felt they were entitled to them. As a further limit to full private rights, the land was donated under certain conditions, restricting the beneficiaries’ power to divide, sell or buy it. They managed those lands, thus, in a manner entirely dissimilar from a modern enterprise. And the same goes for actual farmers, those directly toiling the land, since they were sometimes serfs, and even when they were not, had to give personal services to seigneurs and pay arbitrary tributes.

Unusually Tight Connections between the Crown and High Nobility

Much of the history of Portugal until the nineteenth century revolves around the tension between these three layers of power – the crown, the seigneurs and the communities. The main trend in that relationship was, however, in the direction of an increased weight of central power over the others. This is already visible in the first centuries of existence of the country. In a process that may look paradoxical, that increased weight was accompanied by an equivalent increase in seigneurial power at the expense of the communities. This gave rise to a uniquely Portuguese institution, which would be of extreme importance for the development of the Portuguese economy (as we will later see): the extremely tight connection between the crown and the high nobility. As a matter of fact, very early in the country’s history, the Portuguese nobility and Church became much dependent on the redistributive powers of the crown, in particular in what concerns land and the tributes associated with it. This led to an apparently contradictory process, in which at the same time as the crown was gaining ascendancy in the ruling of the country, it also gave away to seigneurs some of those powers usually considered as being public in nature. Such was the connection between the crown and the seigneurs that the intersection between private and public powers proved to be very resistant in Portugal. That intersection lasted longer in Portugal than in other parts of Europe, and consequently delayed the introduction in the country of the modern notion of property rights. But this is something to be developed later, and to fully understand it we must go through some further episodes of Portuguese history. For now, we must note the novelty brought by these institutions. Although they can be seen as unfriendly to property rights from a nineteenth- and twentieth-century vantage point, they represented in fact a first, although primitive and incomplete, definition of property rights of a certain sort.

Centralization and the Evolution of Property

As the crown’s centralization of power proceeded in the early history of the country, some institutions such as serfdom and settling colonies gave way to contracts that granted fuller personal and property rights to farmers. Serfdom was not exceptionally widespread in early Portugal – and tended to disappear from the thirteenth century onwards. More common was the settlement of colonies, a situation in which settlers were simple toilers of land, having to pay significant tributes to either the king or seigneurs, but had no rights over buying and selling the land. From the thirteenth century onwards, as the king and the seigneurs began encroaching on the kingdom’s land and the military situation got calmer, serfdom and settling contracts were increasingly substituted by contracts of the copyhold type. When compared with current concepts of private property, copyhold includes serious restrictions to the full use of private property. Yet, it represented an improvement when compared to the prior legal forms of land use. In the end, private property as we understand it today began its dissemination through the country at this time, although in a form we would still consider primitive. This, to a large extent, repeats with one to two centuries of delay, the evolution that had already occurred in the core of “feudal Europe,” i.e. the Franco-Germanic world and its extension to the British Isles.

Movement toward an Exchange Economy

Precisely as in that core “feudal Europe,” such institutional change brought a first moment of economic growth to the country – of course, there are no consistent figures for economic activity in this period, and, consequently, this is entirely based on more or less superficial evidence pointing in that direction. The institutional change just noted was accompanied by a change in the way noblemen and the Church understood their possessions. As the national territory became increasingly sheltered from the destruction of war, seigneurs became less interested in military activity and conquest, and more so in the good management of the land they already owned land. Accompanying that, some vague principles of specialization also appeared. Some of those possessions were thus significantly transformed into agricultural firms devoted to a certain extent to selling on the market. One should not, of course, exaggerate the importance acquired by the exchange of goods in this period. Most of the economy continued to be of a non-exchange or (at best) barter character. But the signs of change were important, as a certain part of the economy (small as it was) led the way to future more widespread changes. Not by chance, this is the period when we have evidence of the first signs of monetization of the economy, certainly a momentous change (even if initially small in scale), corresponding to an entirely new framework for economic relations.

These essential changes are connected with other aspects of the country’s evolution in this period. First, the war at the frontier (rather than within the territory) seems to have had a positive influence on the rest of the economy. The military front was constituted by a large number of soldiers, who needed constant supply of various goods, and this geared a significant part of the economy. Also, as the conquest enlarged the territory under the Portuguese crown’s control, the king’s court became ever more complex, thus creating one more demand pole. Additionally, together with enlargement of territory also came the insertion within the economy of various cities previously under Muslim control (such as the future capital, Lisbon, after 1147). All this was accompanied by a widespread movement of what we might call internal colonization, whose main purpose was to farm previously uncultivated agricultural land. This is also the time of the first signs of contact of Portuguese merchants with foreign markets, and foreign merchants with Portuguese markets. There are various signs of the presence of Portuguese merchants in British, French and Flemish ports, and vice versa. Much of Portuguese exports were of a typical Mediterranean nature, such as wine, olive oil, salt, fish and fruits, and imports were mainly of grain and textiles. The economy became, thus, more complex, and it is only natural that, to accompany such changes, the notions of property, management and “firm” changed in such a way as to accommodate the new evolution. The suggestion has been made that the success of the Christian Reconquista depended to a significant extent on the economic success of those innovations.

Role of the Crown in Economic Reforms

Of additional importance for the increasing sophistication of the economy is the role played by the crown as an institution. From the thirteenth century onwards, the rulers of the country showed a growing interest in having a well organized economy able to grant them an abundant tax base. Kings such as Afonso III (ruling from 1248 until 1279) and D. Dinis (1279-1325) became famous for their economic reforms. Monetary reforms, fiscal reforms, the promotion of foreign trade, and the promotion of local fairs and markets (an extraordinarily important institution for exchange in medieval times) all point in the direction of an increased awareness on the part of Portuguese kings of the relevance of promoting a proper environment for economic activity. Again, we should not exaggerate the importance of that awareness. Portuguese kings were still significantly (although not entirely) arbitrary rulers, able with one decision to destroy years of economic hard work. But changes were occurring, and some in a direction positive for economic improvement.

As mentioned above, the definition of Portugal as a separate political entity had two main negative elements: Islam as occupier of the Iberian Peninsula and the centralization efforts of the other political entities in the same area. The first element faded as the Portuguese Reconquista, by mid-thirteenth century, reached the southernmost point in the territory of what is today’s Portugal. The conflict (either latent or open) with the remaining kingdoms of the peninsula was kept alive much beyond that. As the early centuries of the first millennium unfolded, a major centripetal force emerged in the peninsula, the kingdom of Castile. Castile progressively became the most successful centralizing political unit in the area. Such success reached a first climatic moment by the middle of the fifteenth century, during the reign of Ferdinand and Isabella, and a second one by the end of the sixteenth century, with the brief annexation of Portugal by the Spanish king, Phillip II. Much of the effort of Portuguese kings was to keep Portugal independent of those other kingdoms, particularly Castile. But sometimes they envisaged something different, such as an Iberian union with Portugal as its true political head. It was one of those episodes that led to a major moment both for the centralization of power in the Portuguese crown within the Portuguese territory and for the successful separation of Portugal from Castile.

Ascent of John I (1385)

It started during the reign of King Ferdinand (of Portugal), during the sixth and seventh decades of the fourteenth century. Through various maneuvers to unite Portugal to Castile (which included war and the promotion of diverse coups), Ferdinand ended up marrying his daughter to the man who would later become king of Castile. Ferdinand was, however, generally unsuccessful in his attempts to tie the crowns under his heading, and when he died in 1383 the king of Castile (thanks to his marriage with Ferdinand’s daughter) became the legitimate heir to the Portuguese crown. This was Ferdinand’s dream in reverse. The crowns would unite, but not under Portugal. The prospect of peninsular unity under Castile was not necessarily loathed by a large part of Portuguese elites, particularly parts of the aristocracy, which viewed Castile as a much more noble-friendly kingdom. This was not, however, a unanimous sentiment, and a strong reaction followed, led by other parts of the same elite, in order to keep the Portuguese crown in the hands of a Portuguese king, separate from Castile. A war with Castile and intimations of civil war ensued, and in the end Portugal’s independence was kept. The man chosen to be the successor of Ferdinand, under a new dynasty, was the bastard son of Peter I (Ferdinand’s father), the man who became John I in 1385.

This was a crucial episode, not simply because of the change in dynasty, imposed against the legitimate heir to the throne, but also because of success in the centralization of power by the Portuguese crown and, as a consequence, of separation of Portugal from Castile. Such separation led Portugal, additionally, to lose interest in further political adventures concerning Castile, and switch its attention to the Atlantic. It was the exploration of this path that led to the most unique period in Portuguese history, one during which Portugal reached heights of importance in the world that find no match in either its past or future history. This period is the Discoveries, a process that started during John I’s reign, in particular under the forceful direction of the king’s sons, most famous among them the mythical Henry, the Navigator. The 1383-85 crisis and John’s victory can thus be seen as the founding moment of the Portuguese Discoveries.

The Discoveries and the Apex of Portuguese International Power

The Discoveries are generally presented as the first great moment of world capitalism, with markets all over the world getting connected under European leadership. Albeit true, this is a largely post hoc perspective, for the Discoveries became a big commercial adventure only somewhere half-way into the story. Before they became such a thing, the aims of the Discoveries’ protagonists were mostly of another sort.

The Conquest of Ceuta

An interesting way to have a fuller picture of the Discoveries is to study the Portuguese contribution to them. Portugal was the pioneer of transoceanic navigation, discovering lands and sea routes formerly unknown to Europeans, and starting trades and commercial routes that linked Europe to other continents in a totally unprecedented fashion. But, at the start, the aims of the whole venture were entirely other. The event generally chosen to date the beginning of the Portuguese discoveries is the conquest of Ceuta – a city-state across the Straits of Gibraltar from Spain – in 1415. In itself such voyage would not differ much from other attempts made in the Mediterranean Sea from the twelfth century onwards by various European travelers. The main purpose of all these attempts was to control navigation in the Mediterranean, in what constitutes a classical fight between Christianity and Islam. Other objectives of Portuguese travelers were the will to find the mythical Prester John – a supposed Christian king surrounded by Islam: there are reasons to suppose that the legend of Prester John is associated with the real existence of the Copt Christians of Ethiopia – and to reach, directly at the source, the gold of Sudan. Despite this latter objective, religious reasons prevailed over others in spurring the first Portuguese efforts of overseas expansion. This should not surprise us, however, for Portugal had since its birth been, precisely, an expansionist political unit under a religious heading. The jump to the other side of the sea, to North Africa, was little else than the continuation of that expansionist drive. Here we must understand Portugal’s position as determined by two elements, one that was general to the whole European continent, and another one, more specific. The first is that the expansion of Portugal in the Middle-Ages coincides with the general expansion of Europe. And Portugal was very much a part of that process. The second is that, by being part of the process, Portugal was (by geographical hazard) at the forefront of the process. Portugal (and Spain) was in the first line of attack and defense against Islam. The conquest of Ceuta, by Henry, the Navigator, is hence a part of that story of confrontation with Islam.

Exploration from West Africa to India

The first efforts of Henry along the Western African coast and in the Atlantic high sea can be put within this same framework. The explorations along the African coast had two main objectives: to have a keener perception of how far south Islam’s strength went, and to surround Morocco, both in order to attack Islam on a wider shore and to find alternative ways to reach Prester John. These objectives depended, of course, on geographical ignorance, as the line of coast Portuguese navigators eventually found was much larger than the one Henry expected to find. In these efforts, Portuguese navigators went increasingly south, but also, mainly due to accidental changes of direction, west. Such westbound dislocations led to the discovery, in the first decades of the fifteenth century, of three archipelagos, the Canaries, Madeira (and Porto Santo) and the Azores. But the major navigational feat of this period was the passage of Cape Bojador in 1434, in the sequence of which the whole western coast of the African continent was opened for exploration and increasingly (and here is the novelty) commerce. As Africa revealed its riches, mostly gold and slaves, these ventures began acquiring a more strict economic meaning. And all this kept on fostering the Portuguese to go further south, and when they reached the southernmost tip of the African continent, to pass it and go east. And so they did. Bartolomeu Dias crossed the Cape of Good Hope in 1487 and ten years later Vasco da Gama would entirely circumnavigate Africa to reach India by sea. By the time of Vasco da Gama’s journey, the autonomous economic importance of intercontinental trade was well established.

Feitorias and Trade with West Africa, the Atlantic Islands and India

As the second half of the fifteenth century unfolded, Portugal created a complex trade structure connecting India and the African coast to Portugal and, then, to the north of Europe. This consisted of a net of trading posts (feitorias) along the African coast, where goods were shipped to Portugal, and then re-exported to Flanders, where a further Portuguese feitoria was opened. This trade was based on such African goods as gold, ivory, red peppers, slaves and other less important goods. As was noted by various authors, this was somehow a continuation of the pattern of trade created during the Middle Ages, meaning that Portugal was able to diversify it, by adding new goods to its traditional exports (wine, olive oil, fruits and salt). The Portuguese established a virtual monopoly of these African commercial routes until the early sixteenth century. The only threats to that trade structure came from pirates originating in Britain, Holland, France and Spain. One further element of this trade structure was the Atlantic Islands (Madeira, the Azores and the African archipelagos of Cape Verde and São Tomé). These islands contributed with such goods as wine, wheat and sugar cane. After the sea route to India was discovered and the Portuguese were able to establish regular connections with India, the trading structure of the Portuguese empire became more complex. Now the Portuguese began bringing multiple spices, precious stones, silk and woods from India, again based on a net of feitorias there established. The maritime route to India acquired an extreme importance to Europe, precisely at this time, since the Ottoman Empire was then able to block the traditional inland-Mediterranean route that supplied the continent with Indian goods.

Control of Trade by the Crown

One crucial aspect of the Portuguese Discoveries is the high degree of control exerted by the crown over the whole venture. The first episodes in the early fifteenth century, under Henry the Navigator (as well as the first exploratory trips along the African coast) were entirely directed by the crown. Then, as the activity became more profitable, it was, first, liberalized, and then rented (in totu) to merchants, whom were constrained to pay the crown a significant share of their profits. Finally, when the full Indo-African network was consolidated, the crown controlled directly the largest share of the trade (although never monopolizing it), participated in “public-private” joint-ventures, or imposed heavy tributes on traders. The grip of the crown increased with growth of the size and complexity of the empire. Until the early sixteenth century, the empire consisted mainly of a network of trading posts. No serious attempt was made by the Portuguese crown to exert a significant degree of territorial control over the various areas constituting the empire.

The Rise of a Territorial Empire

This changed with the growth of trade from India and Brazil. As India was transformed into a platform for trade not only around Africa but also in Asia, a tendency was developed (in particular under Afonso de Albuquerque, in the early sixteenth century) to create an administrative structure in the territory. This was not particularly successful. An administrative structure was indeed created, but stayed forever incipient. A relatively more complex administrative structure would only appear in Brazil. Until the middle of the sixteenth century, Brazil was relatively ignored by the crown. But with the success of the system of sugar cane plantation in the Atlantic Isles, the Portuguese crown decided to transplant it to Brazil. Although political power was controlled initially by a group of seigneurs to whom the crown donated certain areas of the territory, the system got increasingly more centralized as time went on. This is clearly visible with the creation of the post of governor-general of Brazil, directly respondent to the crown, in 1549.

Portugal Loses Its Expansionary Edge

Until the early sixteenth century, Portugal capitalized on being the pioneer of European expansion. It monopolized African and, initially, Indian trade. But, by that time, changes were taking place. Two significant events mark the change in political tide. First, the increasing assertiveness of the Ottoman Empire in the Eastern Mediterranean, which coincided with a new bout of Islamic expansionism – ultimately bringing the Mughal dynasty to India – as well as the re-opening of the Mediterranean route for Indian goods. This put pressure on Portuguese control over Indian trade. Not only was political control over the subcontinent now directly threatened by Islamic rulers, but also the profits from Indian trade started declining. This is certainly one of the reasons why Portugal redirected its imperial interests to the south Atlantic, particularly Brazil – the other reasons being the growing demand for sugar in Europe and the success of the sugar cane plantation system in the Atlantic islands. The second event marking the change in tide was the increased assertiveness of imperial Spain, both within Europe and overseas. Spain, under the Habsburgs (mostly Charles V and Phillip II), exerted a dominance over the European continent which was unprecedented since Roman times. This was complemented by the beginning of exploration of the American continent (from the Caribbean to Mexico and the Andes), again putting pressure on the Portuguese empire overseas. What is more, this is the period when not only Spain, but also Britain, Holland and France acquired navigational and commercial skills equivalent to the Portuguese, thus competing with them in some of their more traditional routes and trades. By the middle of the sixteenth century, Portugal had definitely lost the expansionary edge. And this would come to a tragic conclusion in 1580, with the death of the heirless King Sebastian in North Africa and the loss of political independence to Spain, under Phillip II.

Empire and the Role, Power and Finances of the Crown

The first century of empire brought significant political consequences for the country. As noted above, the Discoveries were directed by the crown to a very large extent. As such, they constituted one further step in the affirmation of Portugal as a separate political entity in the Iberian Peninsula. Empire created a political and economic sphere where Portugal could remain independent from the rest of the peninsula. It thus contributed to the definition of what we might call “national identity.” Additionally, empire enhanced significantly the crown’s redistributive power. To benefit from profits from transoceanic trade, to reach a position in the imperial hierarchy or even within the national hierarchy proper, candidates had to turn to the crown. As it controlled imperial activities, the crown became a huge employment agency, capable of attracting the efforts of most of the national elite. The empire was, thus, transformed into an extremely important instrument of the crown in order to centralize power. It has already been mentioned that much of the political history of Portugal from the Middle Ages to the nineteenth century revolves around the tension between the centripetal power of the crown and the centrifugal powers of the aristocracy, the Church and the local communities. Precisely, the imperial episode constituted a major step in the centralization of the crown’s power. The way such centralization occurred was, however, peculiar, and that would bring crucial consequences for the future. Various authors have noted how, despite the growing centralizing power of the crown, the aristocracy was able to keep its local powers, thanks to the significant taxing and judicial autonomy it possessed in the lands under its control. This is largely true, but as other authors have noted, this was done with the crown acting as an intermediary agent. The Portuguese aristocracy was since early times much less independent from the crown than in most parts of Western Europe, and this situation accentuated during the days of empire. As we have seen above, the crown directed the Reconquista in a way that made it able to control and redistribute (through the famous donations) most of the land that was conquered. In those early medieval days, it was, thus, the service to the crown that made noblemen eligible to benefit from land donations. It is undoubtedly true that by donating land the crown was also giving away (at least partially) the monopoly of taxing and judging. But what is crucial here is its significant intermediary power. With empire, that power increased again. And once more a large part of the aristocracy became dependent on the crown to acquire political and economic power. The empire became, furthermore, the main means of financing of the crown. Receipts from trade activities related to the empire (either profits, tariffs or other taxes) never went below 40 percent of total receipts of the crown, until the nineteenth century, and this was only briefly in its worst days. Most of the time, those receipts amounted to 60 or 70 percent of total crown’s receipts.

Other Economic Consequences of the Empire

Such a role for the crown’s receipts was one of the most important consequences of empire. Thanks to it, tax receipts from internal economic activity became in large part unnecessary for the functioning of national government, something that was going to have deep consequences, precisely for that exact internal activity. This was not, however, the only economic consequence of empire. One of the most important was, obviously, the enlargement of the trade base of the country. Thanks to empire, the Portuguese (and Europe, through the Portuguese) gained access to vast sources of precious metals, stones, tropical goods (such as fruit, sugar, tobacco, rice, potatoes, maize, and more), raw materials and slaves. Portugal used these goods to enlarge its comparative advantage pattern, which helped it penetrate European markets, while at the same time enlarging the volume and variety of imports from Europe. Such a process of specialization along comparative advantage principles was, however, very incomplete. As noted above, the crown exerted a high degree of control over the trade activity of empire, and as a consequence, many institutional factors interfered in order to prevent Portugal (and its imperial complex) from fully following those principles. In the end, in economic terms, the empire was inefficient – something to be contrasted, for instance, with the Dutch equivalent, much more geared to commercial success, and based on clearer efficiency managing-methods. By so significantly controlling imperial trade, the crown became a sort of barrier between the empire’s riches and the national economy. Much of what was earned in imperial activity was spent either on maintaining it or on the crown’s clientele. Consequently, the spreading of the gains from imperial trade to the rest of the economy was highly centralized in the crown. A much visible effect of this phenomenon was the fantastic growth and size of the country’s capital, Lisbon. In the sixteenth century, Lisbon was the fifth largest city in Europe, and from the sixteenth century to the nineteenth century it was always in the top ten, a remarkable feat for a country with such a small population as Portugal. And it was also the symptom of a much inflated bureaucracy, living on the gains of empire, as well as of the low degree of repercussion of those gains of empire through the whole of the economy.

Portuguese Industry and Agriculture

The rest of the economy did, indeed, remain very much untouched by this imperial manna. Most of industry was untouched by it, and the only visible impact of empire on the sector was by fostering naval construction and repair, and all the accessory activities. Most of industry kept on functioning according to old standards, far from the impact of transoceanic prosperity. And much the same happened with agriculture. Although benefiting from the introduction of new crops (mostly maize, but also potatoes and rice), Portuguese agriculture did not benefit significantly from the income stream arising from imperial trade, in particular when we could expect it to be a source of investment. Maize constituted an important technological innovation which had a much important impact on the Portuguese agriculture’s productivity, but it was too localized in the north-western part of the country, thus leaving the rest of the sector untouched.

Failure of a Modern Land Market to Develop

One very important consequence of empire on agriculture and, hence, on the economy, was the preservation of the property structure coming from the Middle Ages, namely that resulting from the crown’s donations. The empire enhanced again the crown’s powers to attract talent and, consequently, donate land. Donations were regulated by official documents called Cartas de Foral, in which the tributes due to the beneficiaries were specified. During the time of the empire, the conditions ruling donations changed in a way that reveals an increased monarchical power: donations were made for long periods (for instance, one life), but the land could not be sold nor divided (and, thus, no parts of it could be sold separately) and renewal required confirmation on the part of the crown. The rules of donation, thus, by prohibiting buying, selling and partition of land, were a major obstacle to the existence not only of a land market, but also of a clear definition of property rights, as well as freedom in the management of land use.

Additionally, various tributes were due to the beneficiaries. Some were in kind, some in money, some were fixed, others proportional to the product of the land. This process dissociated land ownership and appropriation of land product, since the land was ultimately the crown’s. Furthermore, the actual beneficiaries (thanks to the donation’s rules) had little freedom in the management of the donated land. Although selling land in such circumstances was forbidden to the beneficiaries, renting it was not, and several beneficiaries did so. A new dissociation between ownership and appropriation of product was thus introduced. Although in these donations some tributes were paid by freeholders, most of them were paid by copyholders. Copyhold granted to its signatories the use of land in perpetuity or in lives (one to three), but did not allow them to sell it. This introduced a new dissociation between ownership, appropriation of land product and its management. Although it could not be sold, land under copyhold could be ceded in “sub-copyhold” contracts – a replication of the original contract under identical conditions. This introduced, obviously, a new complication to the system. As should be clear by now, such a “baroque” system created an accumulation of layers of rights over the land, as different people could exert different rights over it, and each layer of rights was limited by the other layers, and sometimes conflicting with them in an intricate way. A major consequence of all this was the limited freedom the various owners of rights had in the management of their assets.

High Levels of Taxation in Agriculture

A second direct consequence of the system was the complicated juxtaposition of tributes on agricultural product. The land and its product in Portugal in those days were loaded with tributes (a sort of taxation). This explains one recent historian’s claim (admittedly exaggerated) that, in that period, those who owned the land did not toil it, and those who toiled it did not hold it. We must distinguish these tributes from strict rent payments, as rent contracts are freely signed by the two (or more) sides taking part in it. The tributes we are discussing here represented, in reality, an imposition, which makes the use of the word taxation appropriate to describe them. This is one further result of the already mentioned feature of the institutional framework of the time, the difficulty to distinguish between the private and the public spheres.

Besides the tributes we have just described, other tributes also impended on the land. Some were, again, of a nature we would call private nowadays, others of a more clearly defined public nature. The former were the tributes due to the Church, the latter the taxes proper, due explicitly as such to the crown. The main tribute due to the Church was the tithe. In theory, the tithe was a tenth of the production of farmers and should be directly paid to certain religious institutions. In practice, not always was it a tenth of the production nor did the Church always receive it directly, as its collection was in a large number of cases rented to various other agents. Nevertheless, it was an important tribute to be paid by producers in general. The taxes due to the crown were the sisa (an indirect tax on consumption) and the décima (an income tax). As far as we know, these tributes weighted on average much less than the seigneurial tributes. Still, when added to them, they accentuated the high level of taxation or para-taxation typical of the Portuguese economy of the time.

Portugal under Spanish Rule, Restoration of Independence and the Eighteenth Century

Spanish Rule of Portugal, 1580-1640

The death of King Sebastian in North Africa, during a military mission in 1578, left the Portuguese throne with no direct heir. There were, however, various indirect candidates in line, thanks to the many kinship links established by the Portuguese royal family to other European royal and aristocratic families. Among them was Phillip II of Spain. He would eventually inherit the Portuguese throne, although only after invading the country in 1580. Between 1578 and 1580 leaders in Portugal tried unsuccessfully to find a “national” solution to the succession problem. In the end, resistance to the establishment of Spanish rule was extremely light.

Initial Lack of Resistance to Spanish Rule

To understand why resistance was so mild one must bear in mind the nature of such political units as the Portuguese and Spanish kingdoms at the time. These kingdoms were not the equivalent of contemporary nation-states. They had a separate identity, evident in such things as a different language, a different cultural history, and different institutions, but this didn’t amount to being a nation. The crown itself, when seen as an institution, still retained many features of a “private” venture. Of course, to some extent it represented the materialization of the kingdom and its “people,” but (by the standards of current political concepts) it still retained a much more ambiguous definition. Furthermore, Phillip II promised to adopt a set of rules allowing for extensive autonomy: the Portuguese crown would be “aggregated” to the Spanish crown although not “absorbed” or “associated” or even “integrated” with it. According to those rules, Portugal was to keep its separate identity as a crown and as a kingdom. All positions in the Portuguese government were to be attributed to Portuguese persons, the Portuguese language was the only one allowed in official matters in Portugal, positions in the Portuguese empire were to be attributed only to Portuguese.

The implementation of such rules depended largely on the willingness of the Portuguese nobility, Church and high-ranking officials to accept them. As there were no major popular revolts that could pressure these groups to decide otherwise, they did not have much difficulty in accepting them. In reality, they saw the new situation as an opportunity for greater power. After all, Spain was then the largest and most powerful political unit in Europe, with vast extensions throughout the world. To participate in such a venture under conditions of great autonomy was seen as an excellent opening.

Resistance to Spanish Rule under Phillip IV

The autonomous status was kept largely untouched until the third decade of the seventeenth century, i.e., until Phillip IV’s reign (1621-1640, in Portugal). This was a reign marked by an important attempt at centralization of power under the Spanish crown. A major impulse for this was Spain’s participation in the Thirty Years War. Simply put, the financial stress caused by the war forced the crown not only to increase fiscal pressure on the various political units under it but also to try to control them more closely. This led to serious efforts at revoking the autonomous status of Portugal (as well as other European regions of the empire). And it was as a reaction to those attempts that many Portuguese aristocrats and important personalities led a movement to recover independence. This movement must, again, be interpreted with care, paying attention to the political concepts of the time. This was not an overtly national reaction, in today’s sense of the word “national.” It was mostly a reaction from certain social groups that felt a threat to their power by the new plans of increased centralization under Spain. As some historians have noted, the 1640 revolt should be best understood as a movement to preserve the constitutional elements of the framework of autonomy established in 1580, against the new centralizing drive, rather than a national or nationalist movement.

Although that was the original intent of the movement, the fact is that, progressively, the new Portuguese dynasty (whose first monarch was John IV, 1640-1656) proceeded to an unprecedented centralization of power in the hands of the Portuguese crown. This means that, even if the original intent of the mentors of the 1640 revolt was to keep the autonomy prevalent both under pre-1580 Portuguese rule and post-1580 Spanish rule, the final result of their action was to favor centralization in the Portuguese crown, and thus help define Portugal as a clearly separate country. Again, we should be careful not to interpret this new bout of centralization in the seventeenth and eighteenth centuries as the creation of a national state and of a modern government. Many of the intermediate groups (in particular the Church and the aristocracy) kept their powers largely intact, even powers we would nowadays call public (such as taxation, justice and police). But there is no doubt that the crown increased significantly its redistributive power, and the nobility and the church had, increasingly, to rely on service to the crown to keep most of their powers.

Consequences of Spanish Rule for the Portuguese Empire

The period of Spanish rule had significant consequences for the Portuguese empire. Due to integration in the Spanish empire, Portuguese colonial territories became a legitimate target for all of Spain’s enemies. The European countries having imperial strategies (in particular, Britain, the Netherlands and France) no longer saw Portugal as a countervailing ally in their struggle with Spain, and consequently promoted serious assaults on Portuguese overseas possessions. There was one further element of the geopolitical landscape of the period that aggravated the willingness of competitors to attack Portugal, and that was Holland’s process of separation from the Spanish empire. Spain was not only a large overseas empire but also an enormous European one, of which Holland was a part until the 1560s. Holland, precisely, saw the Portuguese section of the Iberian empire as its weakest link, and, accordingly, attacked it in a fairly systematic way. The Dutch attack on Portuguese colonial possessions ranged from America (Brazil) to Africa (Sao Tome and Angola) to Asia (India, several points in Southeast Asia, and Indonesia), and in the course of it several Portuguese territories were conquered, mostly in Asia. Portugal, however, managed to keep most of its African and American territories.

The Shift of the Portuguese Empire toward the Atlantic

When it regained independence, Portugal had to re-align its external position in accordance with the new context. Interestingly enough, all those rivals that had attacked the country’s possessions during Spanish rule initially supported its separation. France was the most decisive partner in the first efforts to regain independence. Later (in the 1660s, in the final years of the war with Spain) Britain assumed that role. This was to inaugurate an essential feature of Portuguese external relations. From then on Britain became the most consistent Portuguese foreign partner. In the 1660s such a move was connected to the re-orientation of the Portuguese empire. What had until then been the center of empire (its Eastern part – India and the rest of Asia) lost importance. At first, this was due to the renewal in activity in the Mediterranean route, something that threatened the sea route to India. Then, this was because the Eastern empire was the part where the Portuguese had ceded more territory during Spanish rule, in particular to the Netherlands. Portugal kept most of its positions both in Africa and America, and this part of the world was to acquire extreme importance in the seventeenth and eighteenth centuries. In the last decades of the seventeenth century, Portugal was able to develop numerous trades mostly centered in Brazil (although some of the Atlantic islands also participated), involving sugar, tobacco and tropical woods, all sent to the growing market for luxury goods in Europe, to which was added a growing and prosperous trade of slaves from West Africa to Brazil.

Debates over the Role of Brazilian Gold and the Methuen Treaty

The range of goods in Atlantic trade acquired an important addition with the discovery of gold in Brazil in the late seventeenth century. It is the increased importance of gold in Portuguese trade relations that helps explain one of the most important diplomatic moments in Portuguese history, the Methuen Treaty (also called the Queen Anne Treaty), signed between Britain and Portugal in 1703. Many Portuguese economists and historians have blamed the treaty for Portugal’s inability to achieve modern economic growth during the eighteenth and nineteenth centuries. It must be remembered that the treaty stipulated tariffs to be reduced in Britain for imports of Portuguese wine (favoring it explicitly in relation to French wine), while, as a counterpart, Portugal had to eliminate all prohibitions on imports of British wool textiles (even if tariffs were left in place). Some historians and economists have seen this as Portugal’s abdication of having a national industrial sector and, instead, specializing in agricultural goods for export. As proof, such scholars present figures for the balance of trade between Portugal and Britain after 1703, with the former country exporting mainly wine and the latter textiles, and a widening trade deficit. Other authors, however, have shown that what mostly allowed for this trade (and the deficit) was not wine but the newly discovered Brazilian gold. Could, then, gold be the culprit for preventing Portuguese economic growth? Most historians now reject the hypothesis. The problem would lie not in a particular treaty signed in the early eighteenth century but in the existing structural conditions for the economy to grow – a question to be dealt with further below.

Portuguese historiography currently tends to see the Methuen Treaty mostly in the light of Portuguese diplomatic relations in the seventeenth and eighteenth centuries. The treaty would mostly mark the definite alignment of Portugal within the British sphere. The treaty was signed during the War of Spanish Succession. This was a war that divided Europe in a most dramatic manner. As the Spanish crown was left without a successor in 1700, the countries of Europe were led to support different candidates. The diplomatic choice ended up being polarized around Britain, on the one side, and France, on the other. Increasingly, Portugal was led to prefer Britain, as it was the country that granted more protection to the prosperous Portuguese Atlantic trade. As Britain also had an interest in this alignment (due to the important Portuguese colonial possessions), this explains why the treaty was economically beneficial to Portugal (contrary to what some of the older historiography tended to believe) In fact, in simple trade terms, the treaty was a good bargain for both countries, each having been given preferential treatment for certain of its more typical goods.

Brazilian Gold’s Impact on Industrialization

It is this sequence of events that has led several economists and historians to blame gold for the Portuguese inability to industrialize in the eighteenth and nineteenth centuries. Recent historiography, however, has questioned the interpretation. All these manufactures were dedicated to the production of luxury goods and, consequently, directed to a small market that had nothing to do (in both the nature of the market and technology) with those sectors typical of European industrialization. Were it to continue, it is very doubtful it would ever have become a full industrial spurt of the kind then underway in Britain. The problem lay elsewhere, as we will see below.

Prosperity in the Early 1700s Gives Way to Decline

Be that as it may, the first half of the eighteenth century was a period of unquestionable prosperity for Portugal, mostly thanks to gold, but also to the recovery of the remaining trades (both tropical and from the mainland). Such prosperity is most visible in the period of King John V (1706-1750). This is generally seen as the Portuguese equivalent to the reign of France’s Louis XIV. Palaces and monasteries of great dimensions were then built, and at the same time the king’s court acquired a pomp and grandeur not seen before or after, all financed largely by Brazilian gold. By the mid-eighteenth century, however, it all began to falter. The beginning of decline in gold remittances occurred in the sixth decade of the century. A new crisis began, which was compounded by the dramatic 1755 earthquake, which destroyed a large part of Lisbon and other cities. This new crisis was at the root of a political project aiming at a vast renaissance of the country. This was the first in a series of such projects, all of them significantly occurring in the sequence of traumatic events related to empire. The new project is associated with King Joseph I period (1750-1777), in particular with the policies of his prime-minister, the Marquis of Pombal.

Centralization under the Marquis of Pombal

The thread linking the most important political measures taken by the Marquis of Pombal is the reinforcement of state power. A major element in this connection was his confrontation with certain noble and church representatives. The most spectacular episodes in this respect were, first, the killing of an entire noble family and, second, the expulsion of the Jesuits from national soil. Sometimes this is taken as representing an outright hostile policy towards both aristocracy and church. However, it should be best seen as an attempt to integrate aristocracy and church into the state, thus undermining their autonomous powers. In reality, what the Marquis did was to use the power to confer noble titles, as well as the Inquisition, as means to centralize and increase state power. As a matter of fact, one of the most important instruments of recruitment for state functions during the Marquis’ rule was the promise of noble titles. And the Inquisition’s functions also changed form being mainly a religious court, mostly dedicated to the prosecution of Jews, to becoming a sort of civil political police. The Marquis’ centralizing policy covered a wide range of matters, in particular those most significant to state power. Internal police was reinforced, with the creation of new police institutions directly coordinated by the central government. The collection of taxes became more efficient, through an institution more similar to a modern Treasury than any earlier institutions. Improved collection also applied to tariffs and profits from colonial trade.

Centralizing power by the government had significant repercussions in certain aspects of the relationship between state and civil society. Although the Marquis’ rule is frequently pictured as violent, it included measures generally considered as “enlightened.” Such is the case of the abolition of the distinction between “New Christians” and Christians (new Christians were Jews converted to Catholicism, and as such suffered from a certain degree of segregation, constituting an intermediate category between Jews and Christians proper). Another very important political measure by the Marquis was the abolition of slavery in the empire’s mainland (even if slavery kept on being used in the colonies and the slave trade continued to prosper, there is no way of questioning the importance of the measure).

Economic Centralization under the Marquis of Pombal

The Marquis applied his centralizing drive to economic matters as well. This happened first in agriculture, with the creation of a monopolizing company for trade in Port wine. It continued in colonial trade, where the method applied was the same, that is, the creation of companies monopolizing trade for certain products or regions of the empire. Later, interventionism extended to manufacturing. Such interventionism was essentially determined by the international trade crisis that affected many colonial goods, the most important among them gold. As the country faced a new international payments crisis, the Marquis reverted to protectionism and subsidization of various industrial sectors. Again, as such state support was essentially devoted to traditional, low-tech, industries, this policy failed to boost Portugal’s entry into the group of countries that first industrialized.

Failure to Industrialize

The country would never be the same after the Marquis’ consulate. The “modernization” of state power and his various policies left a profound mark in the Portuguese polity. They were not enough, however, to create the necessary conditions for Portugal to enter a process of industrialization. In reality, most of the structural impediments to modern growth were left untouched or aggravated by the Marquis’ policies. This is particularly true of the relationship between central power and peripheral (aristocratic) powers. The Marquis continued the tradition exacerbated during the fifteenth and sixteenth centuries of liberally conferring noble titles to court members. Again, this accentuated the confusion between the public and the private spheres, with a particular incidence (for what concerns us here) in the definition of property and property rights. The act of granting a noble title by the crown, on many occasions implied a donation of land. The beneficiary of the donation was entitled to collect tributes from the population living in the territory but was forbidden to sell it and, sometimes, even rent it. This meant such beneficiaries were not true owners of the land. The land could not exactly be called their property. This lack of private rights was, however, compensated by the granting of such “public” rights as the ability to obtain tributes – a sort of tax. Beneficiaries of donations were, thus, neither true landowners nor true state representatives. And the same went for the crown. By giving away many of the powers we tend to call public today, the crown was acting as if it could dispose of land under its administration in the same manner as private property. But since this was not entirely private property, by doing so the crown was also conceding public powers to agents we would today call private. Such confusion did not help the creation of either a true entrepreneurial class or of a state dedicated to the protection of private property rights.

The whole property structure described above was kept, even after the reforming efforts of the Marquis of Pombal. The system of donations as a method of payment for jobs taken at the King’s court as well as the juxtaposition of various sorts of tributes, either to the crown or local powers, allowed for the perpetuation of a situation where the private and the public spheres were not clearly separated. Consequently, property rights were not well defined. If there is a crucial reason for Portugal’s impaired economic development, these are the things we should pay attention to. Next, we will begin the study of the nineteenth and twentieth centuries, and see how difficult was the dismantling of such an institutional structure and how it affected the growth potential of the Portuguese economy.

Suggested Reading:

Birmingham, David. A Concise History of Portugal. Cambridge: Cambridge University Press, 1993.

Boxer, C.R. The Portuguese Seaborne Empire, 1415-1825. New York: Alfred A. Knopf, 1969.

Godinho, Vitorino Magalhães. “Portugal and Her Empire, 1680-1720.” The New Cambridge Modern History, Vol. VI. Cambridge: Cambridge University Press, 1970.

Oliveira Marques, A.H. History of Portugal. New York: Columbia University Press, 1972.

Wheeler, Douglas. Historical Dictionary of Portugal. London: Scarecrow Press, 1993.

Citation: Amaral, Luciano. “Economic History of Portugal”. EH.Net Encyclopedia, edited by Robert Whaples. March 16, 2008. URL http://eh.net/encyclopedia/economic-history-of-portugal/