Claude Hillinger: Science and Ideology in Economic, Political, and Social Thought

1. INTRODUCTION This paper has two sources: One is my own research in three broad areas: business cycles, economic measurement and social choice. In all of these fields I attempted to apply the basic precepts of the scientific method as it is understood in the natural sciences. I found that my effort at using natural science methods in economics was met with little understanding and often considerable hostility. I found economics to be driven less by common sense and empirical evidence, then by various ideologies that exhibited either a political or a methodological bias, or both. This brings me to the second source: Several books have appeared recently that describe in historical terms the ideological forces that have shaped either the direct areas in which I worked, or a broader background. These books taught me that the ideological forces in the social sciences are even stronger than I imagined on the basis of my own experiences. The scientific method is the antipode to ideology. I feel that the scientific work that I have done on specific, long standing and fundamental problems in economics and political science have given me additional insights into the destructive role of ideology beyond the history of thought orientation of the works I will be discussing. How serious is the problem of ideology? While the entire paper is an argument to the effect that it is very serious, I will try to give a preliminary answer in terms of an anecdote. A few years ago in a conversation with a prominent colleague, an economics professor at an outstanding American university, he told me that he was unable to have much concern about future generations, since these would be infinitely better off than we are. I hasten to add that this was a man whom I had experienced to be highly intelligent as well as considerate and helpful in his personal relationships. How could it be that a man, considered to be a scholar and scientist by himself as well as by society, would pronounce such a complete absurdity concerning an aspect of reality that is central to his own discipline? Depending on how one interprets this question, the answer is either simple, or complex. The simple answer is that this colleague was stating the implication of a standard model of economic growth. According to this class of models, the material output of an economy can grow …


Introduction
This paper has two sources: One is my own research in three broad areas: business cycles, economic measurement and social choice. In all of these fields I attempted to apply the basic precepts of the scientific method as it is understood in the natural sciences. I found that my effort at using natural science methods in economics was met with little understanding and often considerable hostility. I found economics to be driven less by common sense and empirical evidence, then by various ideologies that exhibited either a political or a methodological bias, or both. This brings me to the second source: Several books have appeared recently that describe in historical terms the ideological forces that have shaped either the direct areas in which I worked, or a broader background. These books taught me that the ideological forces in the social sciences are even stronger than I imagined on the basis of my own experiences.
The scientific method is the antipode to ideology. I feel that the scientific work that I have done on specific, long standing and fundamental problems in economics and political science have given me additional insights into the destructive role of ideology beyond the history of thought orientation of the works I will be discussing.
The idea that the scientific method, or more generally rationalism, the unfettered use of human reason, could ban irrationality in human affairs, leading to a sane and prosperous world in which humans would develop their full potential, has occurred several times in history. It was evident in the Golden Age of Greece and again became the moving force of the Enlightenment. But the Enlightenment ended in the totalitarianisms and wars of the Twentieth Century. Following the Second World War each of the competing political systems led by the United States and the Soviet Union held out the promise of a golden future based on the advance of science and a rational social organization, if only the correct model advocated by either side were adopted. Both sides vastly expanded higher education to train armies of scientists and technicians charged with achieving the stated objective.
Following the collapse of communism, capitalist democracy exemplified by the United States and the associated neo-liberal ideology, appeared to have achieved the final triumph. Francis Fukuyama (1992) announced The End of History. However, as the present article is being written, the shine is already off the neo-liberal ideology. The principal protagonist of that ideology, the United States, is in comparative decline. The populations of the traditional western industrialized nations have lost the élan that derived from an optimistic view of the future. They look towards the future with reduced expectations and they question the moral and intellectual capacities of their politicians to lead them wisely.
I believe that a key to understanding the situation in which the world presently finds itself lies in an understanding of the evolution and current state of the social sciences. They have experienced an explosive growth in the number of university and other professionals, in the volume of journal and book publications and, particularly in the case of economics, an enormous influence on policy. This expansion was the main vehicle by means of which, it was hoped, a more rational and reasonable social order would be attained. This hope is an example of a failed belief in the ideology of scientism. The nature of this ideology and how it evolved in the social sciences is a major topic of this paper.
The key concepts of this paper are 'science' and 'ideology'. The meaning that is often attached to these terms is that 'science' is what we do; 'ideology' is the opinion of those we disagree with. By 'ideology' I mean a rather broad set of beliefs held by a group of people that involves a partial or distorted view of reality, such that the omissions or distortions serve the interests of those adhering to the ideology. An ideology, to be believable, must describe at least a part of reality fairly accurately. Most adherents to an ideology are sincere; they are consciously unaware of the distortions involved or of the self-interested motivation.
The ideologies most often considered are political ideologies and these are central also to the present paper. Political ideologies have existed in great variety, but for many purposes it has proven useful to classify them along a Left/Right political spectrum. Accordingly, I will also refer to the political Left, or political Right, fully aware of the fact that this is a crude classification. Generally, the ideologies of the Left and the political parties propagating these ideologies have stood for a strong state actively pursuing redistributive and egalitarian policies. The ideologies of the Right have emphasized the importance of individual freedom, limited government and the efficiency of markets. Naturally, the more marginalized segments of society have tended to the Left, the dominant elements to the Right. At times I will use an alternative terminology, by referring to the broad spectrum of the Left as 'socialism' and to that of the Right as 'neoliberalism'; again recognizing that many on either side of the spectrum would refer to themselves differently.
Another ideology that is central to the present discussion is scientism. It has two related aspects: One is an exaggerated faith in the ability of science to solve virtually all individual or social problems and an associated devaluation of all forms of discourse that are not considered to be scientific. At the same time, the conception of science involved is often superficial, focusing on surface appearances such as the formation of university departments, professional journals, a specialized language or, particularly in the case of economics, the use of mathematics.
To define science, I will use a negative definition due to the philosopher and historian of science Ravetz. He defines immature or ineffective fields of inquiry as those that do not have agreed upon criteria for establishing factual truth. He views the social sciences as falling in this category, an assessment that I share. 1 The implied criterion for a genuine or functioning science is that it does have such criteria and hence is able to acquire a growing stock of factual knowledge.
I have some hesitation in using the terms 'political science' and 'social science' since I don't believe that the fields usually designated in this manner are scientific by the definition just given. I decided to continue using them, since it is awkward to change well established terms. I use 'social science' in an inclusive sense to refer to all fields that study aspects of society, also economics that is sometimes excluded from this definition.
The paper is unusual both in content and style. I have worked within what I consider to be the economic mainstream, at least from a long run perspective. Within that framework I have tried to make strong hypotheses that explain much on the basis of a few simple assumptions. I was never interested in descriptive detail for its own sake. In this paper I use a similar approach in dealing with the history of ideas. My aim is to identify central tendencies, not to give a full account of the thought of any of the authors mentioned.
Analyzing the role played by ideologies is necessarily an exercise in the history of ideas. When I began my study of economics, the history of economic thought was still a required item in most degree programs; it has since disappeared from these. Quite generally, the social sciences have become ahistoric while their ideological content has increased. The two trends are related; if one does not reflect where current, assumptions came from; one does not realize their ideological character. I will illustrate this with an example that is important in the context of this paper: Today's students of collective choice learn that preferences are formalized as orderings over alternatives; that the central result obtained is Arrow's (1951) 'impossibility theorem', which states that no rational aggregation of preferences is possible; and that a large literature has found no convincing escape from that stark conclusion. The ordering assumption appears to be even more fundamental than the axioms of Arrow, or of other theorists, since it provides the very formalism by means of which they are expressed. What the student does not learn is that the germ of the assumption can be found in the ideological battle waged by the marginalists against classical economics and its philosophical backdrop of utilitarianism. It was resurrected by a later generation of economists to provide ammunition in their ideological struggle against socialism. Arrow's work, the culmination of this line of argumentation, was formulated at the RAND Corporation in the context of the Cold War ideological battles. Neither is the student told why collective choice theory has attempted to build on the negative result of Arrow rather than on Harsanyi's (1955) demonstration of the possibility of a consistent cardinal theory.
A final caveat concerns the basic sources on the history of ideas that I have used. In no case have I tried to review these works. Instead I have taken from them what I found useful for the purpose of my argument. I have tried to state the relevant opinions of these authors correctly and also to distinguish them clearly from my own. I have refrained from repeating here the further references given in the sources that I am citing. The reader interested in pursuing a given topic further should consult the given references first.
The paper has the following structure: In Section 2 I relate my own experiences in attempting to apply the scientific method to core problems of economics and politics. Subsequent sections deal in a roughly chronological order with aspects of the history of ideas, with the emphasis always on the extent to which the ideas were shaped by either ideological or empirical considerations. This question is closely related to how the human agents, intersecting in society and economy, were conceptualized. Accordingly, Section 3 discusses how the enlightenment conception of man narrowed to that of economic man.
Section 4 reviews empirical traditions in economics. I conclude that these were generally marginal and tended to relate to established theory with mutual hostility. Empirical and theoretical traditions were briefly joined first in the era of Petty, later in that of Keynes; subsequently the tradition of abstract theorizing always regained dominance. The joining of empirical and theoretical traditions on the basis of equality that was characteristic of the rise of natural science never took place in economics.
Section 5 is devoted to the Chicago School, particularly the concerted effort by Stigler and Friedman to move economics into conformance with the neoliberal ideology. Section 6 turns to the cold war origins of rational choice theory at the Rand Corporation. It was part of the effort to develop an ideology supporting capitalist democracy against the claims of socialism.
Section 7 describes how rational choice moved from the military/government sector to academia. I concentrate on the single most famous and most influential result in academic rational choice theory, Arrow's impossibility theorem. This theorem is based on assumptions intended to refute socialism, instead, it refutes capitalist democracy; another Arrow paradox! The paper contains much evidence of ideology; is there any science to be found in social 'science'? In Section 2 I had described my own efforts at scientific work, particularly in relation to business cycles. Unfortunately, this work directly contradicted the dominant ideological tendencies in macroeconomics and therefore did not motivated independent investigators to test it. The key requirement of independent confirmation is thus still missing. In Section 8 I discuss the quantity theory of money that I view as having been confirmed by strict scientific standards.
Having discussed various ideologies related to economics and to the political Right, I turn in Section 9 to a consideration of ideologies of the political Left and their impact on the social sciences and humanities. The philosophies and sociologies of science associated with Thomas Kuhn and Steve Fuller are discussed in this connection.
Little confirmed knowledge is generated in the social sciences and even confirmed knowledge is widely disregarded. In Section 10 I attribute this to a misguided search for novelty and a misunderstanding of the nature of and the need for replication.
In the concluding Section 11 I argue that the problems of the social sciences reflect those of the larger society both with respect to culture and institutions.

Motivation
The dominant goal of my career as an economist has been the application of the scientific method as it is understood in the natural sciences in order to solve central problems of economics. The three major areas of research that I engaged in were selected out of this motivation. Two of these areas I identified early, as a graduate student. They are business cycles and economic measurement. A third, voting theory came later. In the following I describe how the scientific problem in each of these areas presented itself to me, as well as my subsequent experiences in solving the problem and publicizing the results. This involved inevitable clashes between what I regarded as the proper scientific method and dominant ideological commitments within the profession.

Business Cycles
I began my graduate studies at the University of Chicago in 1959. At that time it was widely accepted that the macroeconomic fluctuations in the United States were to a large extent inventory cycles. It had been observed that inventory disinvestment, i.e., the drawing down of stocks, accounted for between 50 and 100 percent of the typical US recession. Furthermore, it was seen that the typical length of the cycle, including both a phase of negative and of positive inventory investment, was fairly regular at about 3.2 years. The phenomenon had attracted much attention including extensive hearings in the US Senate, chaired by Senator Douglas, a former University of Chicago economics 6 professor. A famous theoretician in the economics department at Chicago at that time was Lloyd Metzler. He had proposed a formal model, that I found very plausible, of how the decisions of producers generated the cycle. In spite of all this activity, I noticed that there was no systematic effort at determining the quantitative characteristics of the inventory cycle, using the best statistical methods available for the purpose. Furthermore, no effort had been made to test Metzler's model against the data. It seemed to me that the scientific method was not being applied.
I made the quantification of inventory cycle stylized facts and the test of Metzler's model against these the subject of my dissertation. I had come to Chicago without a substantial background in mathematics and statistics and these subjects did not have a strong tradition in the Chicago economics department of that time. Outside of Chicago, the dominant style of doing econometrics in macroeconomics was to construct largescale models involving hundreds of equations. The emphasis was on a good fit, not on the explanation of any independently observed stylized facts. I regarded this as a perversion of the scientific method. At Chicago an entirely different approach to macroeconometric modeling had evolved in the context of monetarism as it was being shaped by Milton Friedman. He believed in a monocausal influence from changes in the stock of money to changes in both nominal and real GDP. Accordingly, in empirical studies a single equation approach was used in which changes in GDP followed, with appropriate lags, from changes in the money stock. The need for any structural modeling and with it the possibility of an endogenous dynamic of the macro economy was denied.
Thus, from the beginning, my purposes clashed with various commitments on the part of dominant groups within the economics profession. The matter might have rested there with my turning to some other less controversial project, if there had not been a particular development in the department. It was the establishment of a chair in econometrics to which R. L. Baseman was appointed. Baseman shared my interest in the philosophy of science and was equally critical of the econometric mainstream, particularly the practice of building large scale macroeconometric models. His support enabled me to carry out my intended project. 2 Advancing and publishing my work on the inventory cycle proved difficult for several reasons. One is that powerful movements in macroeconomics, with regard to both substance and method moved in diametrically opposite directions to that in which I was moving. Keynesians argued that active stabilization policies made business cycles a phenomenon of the past and hence no longer of interest. 3 The monetarists reached a similar conclusion, though from a contrary understanding of macroeconomic reality. They thought that cycles were the result of erratic monetary policies and lacked any interesting own dynamic; they would disappear following the adoption of a steady monetary policy. Similarly, with respect to method, I ran afoul of both Keynesians and Monetarists, though for opposite reasons. Keynesians believed that a macroeconometric model had to consist of hundreds of equations to be considered 'realistic'; the monetarist, at the other extreme, believed in single-equation monocausal explanations.
The sort of small explanatory model that I constructed to explain the inventory cycle was ruled out by both sides.
Another major difficulty was the need to acquire the econometric tools that I felt I needed and that differed sharply from those in common use in macroeconomics. Such tools were needed both for the description of the inventory cycle and for the construction of explanatory models. This was a daunting task! The standard tool used in the natural sciences to describe processes that contain both random and periodic components is spectral analysis. Classical spectral analysis, the only kind available at that time, is a nonparametric and hence highly data intensive method. It has been estimated that the precise identification of a cycle requires an observation length that includes about 30 cycles. Not surprisingly, the method has yielded unimpressive results with the short available economic time series. An application to the identification of the inventory cycle that I made was also unimpressive.
The other econometric problem concerned the estimation of explanatory models. Here as elsewhere, the standard econometric practice departs from that used in the natural sciences. There, dynamic models, both in theory and in applications are generally formulated in continuous time. The reason is both that reality is itself continuous, at least at the macro level, and also that continuous models tend to be simpler and more elegant. In macroeconomics, the assumption that the data are generated by a discrete process is unrealistic. Economic agents do not all make a decision at the beginning of a period, be it the quarter or the year; decisions, adjustments and random disturbances affecting these occur more nearly continuously. There exists by now a large literature that demonstrates that the fact that observations are discrete does not obviate the necessity to model the continuous data generating process 4 . An important aspect is the following: Suppose that the data are generated by the continuous process ( ) Ax t and are modeled by x B . Then and theoretical restrictions, that apply to , particularly zero restrictions on the elements of , do not apply to This argument alone shows that structural modeling in macroeconomics in discrete time does not make sense; this is particularly true of large-scale macroeconometric models with hundreds of zero restrictions.
The discussion to this point makes clear that what I had in mind called for a large scale research project, including people with strong backgrounds in econometrics and good computing skills. The fact that I saw no possibility of organizing such a project, combined with the fact that interest in business cycles had vanished within the economics profession led me to essentially abandon the project by 1972.
A strong new impetus to continue the research came in that year in connection with my move to the University of Munich. The city is also the home of the large IFO-Institute for empirical economic research where I gave a talk on my work on the US inventory cycle including the three-equation model of the cycle that I had constructed. A member of the institute, Klaus Schüler, decided to fit the model to German data. It turned out that the exact model that I had fitted to the US data gave even better results when fitted to the German data. This was a consequence of the fact that the cycle itself appeared stronger and less erratic in German than in US data.
A few words about the significance of this result are in order. The large-scale models then in fashion where not stable even from one quarterly forecast to the next within the same country. In this case the simplest model that could be devised to replicate the stylized facts was fitted without modification to data for another country, data that I had been unaware of when originally devising the model. I felt that this was a very strong confirmation that I was on the right path, at least from a scientific point of view and I was motivated to resume the research with renewed vigor.
I recommitted myself to a research program on business cycles that was wider both conceptually and empirically than the previous one. I had become aware of a long tradition of work on investment cycles, in which the inventory cycle was only one kind, others being the equipment cycle, and the building cycle. For post-WWII data, I found that genuine building cycles had become localized to metropolitan areas, or regions, and that at the national level a fixed investment cycle could be identified that was distinct from the inventory cycle. The other extension was to internationalize the data set so as to include the major industrialized economies. I never had any major problem in formulating the research problems and all of the principal decisions that I made were ultimately validated by the empirical results. The reason is the basic ideas were already available; they had to be found and identified, rather than invented from scratch. The basic issue was one of deciding on relevance. Thus I committed early to the scientific method of the natural sciences as against the then current econometric practices. There was also a substantial, empirically oriented tradition of business cycle theory. The main findings of that tradition were summarized in a fine book by R. C. O. Matthews (1959), just before the tradition disappeared from economists' radar screen. My work here was basically one of refinement and formalization as well as the carrying out of the appropriate econometric program. Here three specific methods were essential, all coming from the natural sciences: maximum entropy spectral analysis to determine the business cycle stylized facts; continuous-time econometrics for estimating the explanatory models formulated as differential equations; finally, parametric resonance to explain the aggregation of cycles from the level of firms to the macro-economy.
Publishing the results proved to be very difficult, as far as major journals are concerned, virtually impossible. A number of publications appeared in applied, or more specialized journals, or in conference proceedings. The most comprehensive publications were the conference proceedings of two conferences that I organized. I was fortunate in being able to secure Oxford UP and Cambridge UP as the publishers 5 .
Regarding the research on business cycles, a relatively complete summary is Hillinger (2005a). The principal empirical findings, based on several studies of the major industrialized economies are: a. Inventory investment is the most volatile component of GDP, fixed investment the second most volatile. The contributions of these components to the fluctuations of aggregate GDP is substantially greater then their average shares. b. Each type of investment has a periodic as well as a random component, with the random component being more prominent in inventory investment. Roughly, the period length is about 4 years for inventory investment and about 8 years for fixed investment.
c. The cycles can be explained by partial adjustment processes that result in overshooting, the excessive accumulation or decummulation of stocks, and ultimately in directional reversals. I have formalized these ideas as second order accelerator equations. d. The aggregation of cycles from the level of individual firms to the aggregate level can be explained by means of the theory of parametric resonance taken from physics. This theory explains what happens when several mechanism capable of similar oscillations are brought into contact with each other. If the intrinsic periods of the mechanisms and the initial phases of their fluctuations are sufficiently close to each other, then the cycles will 'agree' on an average period and an average phase and the amplitudes of the individual oscillations will increase. The cycles of the firms are coordinated and accentuated in this manner and become visible at the aggregate level.
These empirical findings and explanatory theory are a confirmation of the theory of investment cycles that was dominant in the first half of the Twentieth Century. It sharply contradicts contemporary theories that regard the economy as being passively pushed around by unexplained 'shocks'.

Economic Measurement
The second major unsolved problem for a scientific approach to economics of which I became aware as a graduate student was that of economic measurement. In the natural sciences, most impressively in physics, measurement is intimately tied to theory; if this were not the case the measurements could not be used to test the theories. I first became aware of the fact that this close connection between theory and measurement was lacking in economics in relation to real GDP. From a practical point of view it is the measure of all things economic; the overall success of an economy is measured by the growth rate of its real GDP. No theory existed to justify the significance attached to this measure.
As I read up on various theories dealing with economic measurement, I became aware of a profound, but little reflected paradox: The most common characterization of an advanced economy is as monetary; diverse activities are all made comparable by means of this measure. Yet economic theorists have regarded the monetary measure with skepticism, if not hostility. The origin of this attitude is the utilitarian postulate that subjective utility, not money income, is the ultimately relevant measure. The Utilitarians proposed maximization of the sum of utilities, but did not know how to measure these. Much later, Bergson and Samuelson proposed maximization of a social welfare function, but could specify neither the function, nor measure the utilities that they postulated as arguments of the function. Samuelson severely criticized various monetary measures used in applied welfare analysis. The end result of these developments was that theoretical and applied welfare economics parted ways. There resulted a theoretical welfare economics without applications and an applied welfare economics without an adequate theoretical foundation. Economic measurement came to be viewed as a problem to be dealt with by statistical agencies, no longer a concern of academic economists. 6 I decided to take on the task of providing a theoretical foundation for those economic measures that were expressed in monetary terms. This turned out to be a lengthy process in the course of which I was searching for the most adequate conceptualization of the problem. To explain the problems that I faced I have to discuss briefly the existing theories.
I will discuss the most important approaches to theory construction in this area: One is the axiomatic theory of index numbers. Apart from the fact that there has been no agreement on which set of axioms should be adopted, there is a fundamental objection to this approach: it does not provide any economic inference from the measured values of the indexes that obey some set of axioms. 7 There is also the economic theory of index numbers. The bulk of this theory deals with a single utility maximizing consumer with homothetic preferences. There is general agreement that consumer preferences are not homothetic. Moreover, price and quantity indexes that are of interest are computed with aggregate data, even for total GNP that includes business, government and foreign sectors. This theory too I found inapplicable. A third approach is the econometric theory. It involves the estimation of the parameters of a system of aggregate consumer demand functions assumed to be generated by a representative consumer with homothetic preferences. Here the problem is that: a. A representative consumer does not exist (see Section 6) and if he did exist he would not have homothetic preferences. b. The data employed are not the prices and quantities of individual commodities and services, as consumer theory would require, but rather price and quantity indexes.
Another important area of measurement is consumer surplus. The formula , or related formulas, are widely employed in cost/benefit analysis.
Many economists, among them such famous names as Hicks, Hotelling and Harberger have attempted to give rigorous derivations, but all of their derivations are flawed. 8 The final approach to be discussed here is associated with the names of Bennet (1920) and Divisia (1925). To motivate it some preliminary remarks are in order. All of the monetary measures discussed here attempt to do essentially the same thing: to divide a change in nominal expenditure into a real value change associated with the changes in quantities and an inflation associated with the changes in prices. Fixed price weights are used to measure the quantity changes and fixed quantities are used to measure the price changes. The problem is that there are many different ways in which these weights, and the formulas in which they appear, can be specified.
The preceding applies to formulas that decompose the change in value over an interval between two discrete price/quantity situations. The insight of Bennet and Divisia was that there is a natural and unique decomposition of a value at a point, obtained by differentiation. Given a value v = pq , the Divisia price and quantity differentials P, Q are implicitly defined in terms of the differentials 7 I once found precisely this criticism made in a book on the philosophy of science; along with the comment that this shows that economists do not understand the scientific method. Unfortunately, I have not been able to recover the reference. 8 Each of the three authors mentioned implicitly assume a constant marginal utility of income. Samuelson has shown that this assumption cannot be valid. The matter is discussed in detail in Hillinger (2001 Of course, ultimately a formula is needed that allows the comparison of distinct price/quantity situations. The natural attempt at obtaining such formulas is to integrate the Divisia differentials over an interval. However, this does not lead to a closed form solution and the integrals are in fact path dependent. Törnqvist (1936) suggested the indexes bearing his name as approximations to the Divisia integrals, but neither he nor subsequent authors gave convincing derivations. The Törnqvist indexes are: My most advanced contributions have utilized the Divisia/Törnqvist framework. I provided more convincing derivations of the Törnqvist indexes. I also elaborated the theory in the context of explicitly utility maximizing consumers and heterogeneous groups of such consumers. This work is reported in Hillinger (2003aHillinger ( , 2007.
The problem that initially motivated me to get involved with the theory of economic measurement, how to meaningfully define real GDP and its components has been strangely ignored by economic theorists and was left to the statisticians of the agencies producing the data. The result is that different countries do it fundamentally differently and, as I have argued, they all do it incorrectly. The problem is that the statisticians believe that they must compute each sector as well as the total as separate quantity indexes. This causes the sectors not to add up to the total. I have argued that instead all components and the total should be deflated by means of the same deflator, preferably the deflator for consumption expenditures. This is discussed in Hillinger (2003bHillinger ( , 2007.

Voting
My third major field of interest is voting theory. Although this interest began early, I turned to it in a concentrated fashion quite late, when I was nearly 10 years into retirement from active teaching. I no longer recall exactly how this interest developed, but I think that there were three different sources: The first was my interest in measurement. Most economic measurements have to do with determining in which of two situations people are better off, i.e. which provides higher levels of utility. Voting also has to do with the determination of which of two or more situation is best in the voters' judgment. In my early paper on voting (Hillinger, 1971), I showed that when parties bundle their stands on several issue, then the platform that combines the most popular stands on individual issues will not in general be the one most preferred by the voters.
Gradually I became aware of other problems connected with voting. The most common voting procedure used in general elections is plurality voting. Two major defects of this method are: a. A candidate strongly opposed by a majority of the voters may be elected. This will happen when there are several candidates acceptable to the majority who divide the majority vote among themselves with the consequence that a minority candidate, disliked by the majority, is elected.

b.
In an otherwise fairly even election, a minority candidate may take votes away from the major party candidate that he is ideologically closest to and cause him to lose the election. In the 2000 US election the minority candidate was the consumer advocate Ralph Nader, who was ideologically closer to the Democrat Gore than to the Republican Bush. It is likely that by taking votes from Gore, he tipped the scales in favour of Bush. c. The plurality rule provides a strong incentive for strategic voting. If a voter feels that the candidate he favours has no chance of winning, he is likely to regard a vote for that candidate as wasted and will vote instead for a candidate with a chance of winning. Plurality voting faces the voter with difficult strategic decisions and leads to a vote that does not reflect the true preferences of the voters.
Voting is the central institution of democracy. Given that plurality voting is so highly defective, it seemed to me that there was no more important problem in social science then to devise a better method.
A third source of my interest was the literature on collective choice and in particular Arrow's (1951) impossibility theorem that implied that a voting procedure satisfying reasonable conditions could not exist. I felt a strong resistance to accepting the theorem at face value because intellectual consistency would then require abandonment of the idea of democracy as a real possibility. Engagement for democracy would then be as foolish as trying to construct a perpetuum mobile. I found this alternative unacceptable. Furthermore, I felt that democracies, imperfect though they were, actually existed and that the task was to improve, not to deny them.
If I was to be right, Arrow had to be wrong. There was clearly nothing wrong with his mathematics that had been subjected to a huge amount of scrutiny. Furthermore, his assumptions, stated as axioms, had also been intensively studied and their modification did not point to a way out of the dilemma. I found the basic idea for the solution in another area of social choice theory. John Harsanyi (1955) had shown that a plausible and consistent set of axioms for social choice could be constructed if one adopted a cardinal framework for the description of preferences. Instead of the weaker formalism of orderings, preferences were now represented by numbers obeying the rules of arithmetic, thereby allowing an expression of intensity of preference.
Harsanyi's axioms were abstract and his discussion moved on an equally abstract philosophical level. He did not discuss any application. To me the application to voting was clear. The voter must be given a numerical scale such that the numbers on that scale could be used to evaluate the alternatives. The voting scale could be utilized as a normalized utility scale used to 'measure' the utility the voter attached to each alternative. As in any ordinary measurement process, there should be no restriction on how the voter distributes the values permitted by the scale over the alternatives; specifically, he must be free to assign the same value to alternatives that he values equally.
Having come this far, it also became clear to me that this is in fact how preferences are being universally aggregated everywhere except in political voting. Businesses measuring customer satisfaction, or pollsters measuring the popularity of policies or politicians, all proceed in this manner.
The startling conclusion is that the central problem of political theory, described as insoluble in a large theoretical literature, has in fact had a pragmatic solution that is widely practiced everywhere except where it would matter most, in political elections. www.economics-ejournal.org The disciplines of collective choice and voting theory are perfect examples of disciplines that have managed to completely insulate themselves from that reality for the understanding of which they were originally created. To publish in their journals, one must not offer a solution to the voting problem, but rather follow certain conventions such as providing a game-theoretical analysis of voter behavior. After existing for more than 100 years and publishing hundreds of articles, the profession has not come one step closer to agreement on a superior voting method.
My analysis of voting appeared in Hillinger (2005b). To the method of voting with an unrestricted cardinal scale I gave the name utilitarian voting, a reference to the utilitarian tradition in economics which goes back to Jeremy Bentham and classical economics. More specifically I proposed for general elections a three-valued voting scale, allowing the voter to vote for or against any alternative, or abstain. I referred to voting with this scale as evaluative voting.
The mathematician Warren D. Smith has also proposed voting with an unrestricted scale and refers to it as range voting. 9 Range voting advocates a scale of 1-10, or 1-100. Also a form of utilitarian voting, but with a scale having only two values is approval voting. 10

Conclusion
In all three areas discussed above I found a strong resistance to the publication of my results. I necessarily became aware of the ideological resistances to my work. My understanding of these ideological forces became much clearer after reading the books discussed in the following sections. Had I had this full awareness at the start of my career, I likely would not have embarked on it. As it turned out, I believe that I made substantial contributions to the core problems of each of these areas.

From the Enlightenment Conception of Man to Economic Man
I had already started on this paper when I chanced across a forgotten, but now highly relevant book in my library; da Fonseca's (1991) Beliefs in Action: Economic Philosophy and Social Change. The book is a history of ideas about ideas; their generation, transmission, distortion and finally their social impact. In spite of the obvious importance of the subject and the fact that it has been of concern to philosophers and economists from antiquity to the present, I am aware of no other systematic and historical treatment. My aim here is not to review the book, but rather to extract those findings that are relevant to the present paper. I will also indicate where my interpretation or emphasis differs.
The principal insights that emerge from da Fonseca's book and that he richly documents with quotations from the original works are the following: a. Before the advent of the neoclassical concept of 'economic man' (EM) 11 , philosophers from antiquity through the Enlightenment, most importantly David Hume, Adam Smith and John Stuart Mill, had a different conception of human nature. This conception is most easily explained by contrasting it with the by now more familiar EM concept. Under the EM concept individual preferences are taken as primitive givens, not subject to further scrutiny. The distinction between purely selfish motives and ethical motives for behavior disappears. For the philosophers this distinction was the principal concern. After all, Smith held a chair of moral philosophy. While it is true that adherents to the EM conception have allowed altruistic motives to be part of the individual's preference system, this concession is but a pale ghost of moral philosophy. Specifically, the question of the kind of altruistic elements that should be a part of individual preference is not considered. Altruism, along with all other decisions of the individual are regarded as purely personal issues, not subject to any rational debate. b. Under the EM conception the individual is regarded as the best judge of his own welfare. The philosophers questioned this. They regarded the opinions of the bulk of mankind as being feeble and unstable structures, largely determined by non-rational or even irrational factors. Though neither the philosophers nor da Fonseca used the term, I would refer to many of these opinions as ideologies. To give a currently relevant example: An element of the neoliberal ideology that is transported by advertising, the popular media, and the pronouncements of political leaders is that the ever increasing consumption of material goods is the key to happiness. Modern happiness research sides with the philosophers in refuting this belief. c. Another major issue that has concerned philosophers and later economists from antiquity to the present is the question of how dominant ideas in a society change and the effect that ideas have on society. Most of the writers considered by da Fonseca both desired and expected their writings to have a positive effect. Accordingly, the prevailing view was that moral philosophers can improve the moral quality of the population, economist the economic understanding and that all of this would contribute to the general improvement of society. This belief system is related to the idealistic tradition in philosophy. Not surprisingly, men of ideas emphasize the autonomy and importance of ideas. This can be contrasted with the views tending in the opposite direction, of men of affairs. In each case an element of ideology is involved, that aggrandizes the own role.
The most prominent challenge to idealism in social thought is the doctrine of historical determinism due to Hegel and Marx. According to Marx, the material conditions of society determine the ideas that people hold and the dynamics of social change. The ideological aspect of their theories is rather obvious. Hegel, a philosophy professor at the University of Berlin, claimed that history had found its ultimate perfection in the Prussian state. Marx attempted to motivate the proletarian revolution by claiming that the desired outcome was in any case inevitable. Neither Hegel nor Marx was completely consistent, since each was expecting to have an influence with his writings. Marx in fact expected Das Kapital to become the 'bible of the proletariat' and an important instrument in organizing the revolution.
The relation between ideas and social reality will be a continuing theme of this paper. The question of the priority of one over the other is in my view akin to the question of the priority of the chicken or the egg. A more useful general framework is that of the three worlds of Sir Karl Popper: The first world of material objects; the second world of subjective ideas; the third world of objective interpersonal knowledge. In Popper's view, that I share, the three worlds interact in a dynamic process; there is no priority of one over the others. 12 The first chapter of da Fonseca's book, as sketched above, deals in a general way with the role of ideas in relation to society. The remainder of the book narrows the discussion to a specific topic, albeit a very important one, namely the transformation of the concept of man held by Enlightenment philosophers and classical economists, to homo oeconomicus. This reduction is itself part of wider intellectual movements that began with ancient Greek philosophy and continued in relation to the rising natural sciences. This is the subject of da Fonseca's second chapter. The concepts that I would stress in discussing the currents of thought that are involved are: reductionism, materialism, rationalism and scientism. 13 The Ionian philosophers taught that all phenomena resulted from the combinations of the four elements fire, air, water and earth. More radical and closer to modern science was Democritus who taught that all phenomena result from the motions of atoms. He regarded mental phenomena as caused by the motions of particularly fine atoms. Democritus was thus an early representative of materialism, the doctrine that stands in opposition to idealism, which holds that the origin of all phenomena is mental. Both reductionism and materialism have motivated and continue to inspire modern science as well as fields that have merely laid claim to being scientific.
With the rise of science in the Eighteenth Century these philosophical traditions revived, but in a somewhat changed form. Since physics now recognized forces in addition to matter as fundamental entities, the doctrine of materialism was replaced by that of physicalism, according to which physical phenomena are behind all appearances. Rationalism in its pure form persisted in philosophy, but there also developed the related doctrine of scientism according to which the methods of the natural sciences, above all those of physics, were appropriate in all areas of inquiry. 14 Particularly important as a forerunner of the later economic man was the French physician and philosopher de la Mettrie who in 1747 published L'Homme Machine. Da Fonseca, starting with a quotation from de la Mettrie describes his philosophy as follows: 'Man is a machine, and there is nothing in the entire universe but a single substance diversely modified'. The basic idea is to see human beings and their actions as objects and events in the physical world, that is there to be described and accounted for, like all the rest, in purely objective terms; in terms in which moral thinking and mental processes in general -the agent's subjective experiences -have no place. (p. 26).
The rise of the EM concept is described in da Fonseca's Chapter 3. This development was both gradual and controversial. da Fonseca points out that the Eighteenth Century founders of economics, Quesnay and Smith, both rejected the mechanical model of human behavior, arguing instead for the reality of human freedom and the importance of moral choices. A tendency to limit the discussion to motives that we would today call economic was first evident in the work of Bentham and was subsequently elaborated by Ricardo. Ricardo's work also involved a substantial increase in the degree of formalism, an important aspect of scientism. da Fonseca points out that this reductionism to purely economic motives was controversial. Thus, J. S. Mill, in his mature work distanced himself from the reductionism of both Bentham and Ricardo and refused to draw a sharp line between economic and non-economic motives. The most prominent economist of the early Twentieth Century, George Marshall, still expressed the (failed) hope that economists would pay increasing attention to non-economic motives.
The conquest of economics by the EM concept began in the early Nineteenth Century with the advent of Marginalism on the continent and with the work of Jevons and a little late of Edgeworth in England. The basic conceptualization in terms of the decision problem facing a utility maximizing agent was the same in both cases, but the English school is more interesting in the present context because of their emphasis on the use of mathematics, specifically the calculus, for the analysis of the problem. With the work of these economists, economics assumed a semblance of physics. This was fully intended. Edgeworth (quoted by da Fonseca, 1991, p. 45) wrote: 'Mecanique Sociale' may one day take her place along with 'Mecanique Celeste', throned each upon the double-sided height of one maximum principle, the supreme pinnacle of moral as of physical science. As the movements of each particle, constrained or loose, in a material cosmos are continually subordinated to one maximum sum-total of accumulated energy, so the movements of each soul, whether selfishly isolated or linked sympathetically, may continually be realizing the maximum energy of pleasure, the Divine love of the universe.
At the end of the chapter da Fonseca reaches the following conclusion: So, the mechanization of the economic picture advanced by Jevons, Edgeworth and some of their contemporaries can be seen as combining two divergent tendencies: 1. There is the simplification of economic action so as to make it independent of moral ends. The central feature of the metamorphosis of economic agents into 'pleasure-machines' is that they cease being moral persons and become players, each one computing his way -always within the legal framework -towards the means that shall enable him to fulfill his desires and commitments once he steps out of the purely economic sphere of action. 2. There is the metamorphosis of the economic agents into 'pleasure-machines' as constituting in itself an end. Economic actions, the argument runs, ought to be purely instrumental rather than ends in themselves. And it is a good thing, given the nature and properties of a pure market economy, that agents should suspend their moral beliefs and opinions as they try to amass at lowest cost the maximum of the means necessary for the satisfaction of their highest preferences. The disengagement of the economic life from morality turns out to be not only a promising starting-point for abstract analysis, it becomes a moral end-point too, that is, a desirable state of affairs. This section has so far been based on the first three chapters of da Fonseca's book. The remaining chapters are not as directly relevant to my purposes and I omit a consideration of them. At this point I give my own interpretation of what has been discussed, an interpretation that I hope to deepen in the following sections. My interpretation differs significantly both from that of da Fonseca as well as of most other critics of the economic mainstream.
I do completely agree with da Fonseca that the concerns of the enlightenment philosophers with moral judgments as well as with the general uncertainty and limited rationality of beliefs is essential to any broad understanding of social issues. Where I disagree is in evaluating the analytical separation of economic and non-economic motives and their consequences. I believe that the separation was natural and essential for the development of economics. Smith himself largely separated the analysis, treating the one part in his Theory of Moral Sentiments and the other in Wealth of Nations. Let us look at Smith's central analytical argument concerning the 'invisible hand'. He argued that under conditions that we would today refer to as 'perfect competition' those goods that are most desired by consumers would be produced. If this were not the case, then some producers could bid resources away from competitors producing the less desired goods and produce the more desired ones. This argument is purely economic and analyses the implications of given tastes and technologies. Adding moral issues or uncertainty complicates the analysis without adding to our understanding. Later in this paper I will argue that as contemporary economists are turning in large numbers to the themes neglected in neoclassical economics they are also abandoning the solution of long standing problems of economics.
Subsequent to Adam Smith the subject matter of the Wealth of Nations was steadily elaborated and evolved into the respected academic discipline of economics. Moral philosophy, the subject of The Theory of Moral Sentiments largely stagnated. Some aspects of the social thought of the enlightenment philosophers were elaborated in sociology, as well as in Marxism and other non-orthodox strands of economics. But this did not lead to a complex and codified structure of thought as in orthodox economics. That economics was able to impress more with its sophistication was one reason for the difference in social acceptance and prestige. Another reason is that economics could be more easily aligned with conservative ideologies, sociology with revolutionary ones. Where revolutions seemed possible or actually took place, sociological thought dominated; with the world-wide collapse of revolutionary regimes and their ideologies economics began to rule supreme. An ideological fixation on isolated aspects of reality at the expense of other equally important aspects has prevented the emergence of a comprehensive and balanced system of social thought.

The Empirical Tradition at the Birth of Economics
Ideology and empiricism may be thought of as opposites; empiricism is the earnest study of reality to find out what it is actually like; ideology is the attempt to make believe that reality is as we would like it to be. Of course, this is a bold, but I believe illuminating assertion. It provides the motive of the present section. I argue that the source of the strength of ideology in social thought is the weakness of the empiricist tradition. More specifically I will argue that the train of empiricism was derailed at the very beginning of modern economics, the dominant social discipline. I begin the discussion by looking at the key ideas contributed by three founding fathers of modern economics and their subsequent fate. The founding fathers are Sir William Petty, François Quesnay and Adam Smith. The extensive writings of the three authors on what may be termed social philosophy are not the subject here. Instead, I focus on the one idea that each author is primarily identified with and that has retained its relevance to the present day.
Petty was a typical universal genius of the Enlightenment era. Rising from a humble background he became at a young age a doctor in physics, a professor of anatomy and a professor of 'music', a term designating at that time a wide field of cultural studies. Drawn to practical pursuits, Petty soon terminated his academic career and enlisted as a physician with Cromwell's army that invaded Ireland. He was given the task of surveying the land and assessing the riches some of which were to be distributed among the victors. As a consequence of these endeavors he published in 1672 The Political Anatomy of Ireland. In this and in subsequent publications he estimated population, national income and national wealth for Ireland and England. He is thus the father of quantitative economics, or to put it differently, of measurement in economics. The title of his last, posthumously published book Discourse on Political Arithmetic contained the title he gave to this new subject of study.
Petty, a founding member of the Royal Society was personally involved in the scientific revolution occurring in his day. He believed in the methodological unity of the natural and the social sciences and took them to rest on a foundation of measurement. In the introduction to the Discourse he expressed his famous credo: The Method I take to do this, is not yet very usual; for instead of using only comparative and superlative Words, and intellectual Arguments, I have taken the course (as a Specimen of the Political Arithmetick I have long aimed at) to express my self in Terms of Number, Weight, or Measure; to use only Arguments of Sense, and to consider only such Causes, as have visible Foundations in Nature; leaving those that depend upon the mutable Minds, Opinions, Appetites, and Passions of particular Men, to the Consideration of others. (Cited in Roncaglia, 1987, p. 855).
Petty's work inspired some further work in political arithmetic particularly by Charles Davenant who supplied the essentially modern definitions of national income and product.
Quesnay was like Petty a physician, in fact the court physician to Louis XV. Under the label of physiocracy he established a school of social and economic thought. Much of physiocratic thought is antiquated, particularly the idea that all wealth originates in agriculture. The idea that Quesnay is famous for and that has retained its relevance is that of a circular flow of goods and services through the sectors of an economy that could be measured in terms of the associated money flows in the opposite direction; an idea that suggested itself to the physician in analogy to the circulation of the blood. He gave a numerical example of such a system in his Tableau Economique (1758), a forerunner of the modern national income and product accounts.
With the work of Petty and Quesnay the basic ideas on which quantitative macroeconomics is based were in place. Thereafter the subject languished until the early Twentieth Century when the rise of socialism and of the econometric movement generated a strong interest in social and economic statistics. The modern equivalents of Petty's and Quesnay's measures, the national income and product statistics are generally regarded as the most important of all economic statistics and enter the national economic and political debates on an almost daily basis.
Adam Smith, like the other two authors, fully participated in the spirit of the enlightenment and of the beginning scientific and industrial revolution. The structure of his two principal works was in fact inspired by Newton's theory of universal gravitation. Just as Newton had postulated a single force to explain, among other phenomena, the motions of the heavenly bodies, so Smith postulated in each of his works a single force to explain an aspect of human behavior. In the 1759 Theory of Moral Sentiments this force was sympathy; derived from an empathic concern for the feelings of others. In the 1776 An Inquiry into the Nature and Causes of the Wealth of Nations he postulated self interest as the driving economic force.
Smith differs from Petty and Quesnay in having never been active in a natural science. Instead, his principal occupation was as a professor of moral philosophy. Nor did Smith share the interest in measurement and quantitative analysis of the other two. Famously he declared 'I have no great faith in political arithmetic'. Undoubtedly he was expressing a valid skepticism relative to the very rough and uncertain estimates of the political arithmeticians; mere skepticism however does not help a field to advance. 15 The academic economic mainstream as it evolved subsequent to Smith remained essentially in the mold of philosophy, reasoning from intuitively perceived first principles. This was true of classical economics, of the marginalists, of the French mathematical school and, as I shall argue, also of contemporary economics. It is instructive to contrast the beginning of economics with the beginning of the natural sciences. The universities, of the early Renaissance, dominated by their theological faculties, were hostile to the rise of science and particularly to the experimentalists who generally came from the class of artisans of lower social standing. The early organization of science took place in scientific organizations outside the universities. It was only when the prestige of science began to replace that of religion that scientists established themselves at the universities, with experimentalists and theorists enjoying equal prestige. The marriage of empiricism with speculative theory, so characteristic of natural science, never occurred in economics or later social sciences.

The Empirical Tradition from Smith to the Early Twentieth Century
The evolution of economics after Smith illustrates the clash of various ideologies with empiricism. Broadly speaking, academic economics concerned itself with the elaboration of the properties of a competitive equilibrium, at first primarily verbally and conceptually, later increasingly in mathematical terms. Since the principal result is the optimality of the competitive equilibrium, there developed a natural affinity between economics and a conservative laissez faire political ideology. The analysis of what happens outside of equilibrium was both difficult and would have stood in the way of this conservative alignment of the profession. Ideologies tend to extreme positions; in this case Jean-Baptiste Say's law of markets, which ruled out on theoretical grounds the very possibility of a general overproduction.
Irrespective of Say's law, periods of economic depression and evidently lacking aggregate demand where a fairly common experience of the rising industrial nations. They were labeled at first as 'crises' and later, when a certain regularity of the occurrence was noted, as 'cycles'. The question of the possibility of a general glut due to insufficient demand was the subject of a long debate between Ricardo and Malthus. Ricardo, the abstract theoretician represented the economic mainstream, upholding the validity of Say's law. The empiricist Malthus took the modern position that income could be hoarded with the consequence that demand would be insufficient to absorb capacity output. On this and other issues the two could in the end only agree to disagree, but Ricardo's views remained the standard of academic economics, while those of Malthus, as observed by Keynes, persisted only in an economic 'underground'.
It is relevant in the present context that Malthus was the founder of demography as an essentially empirical discipline. Demography evolved subsequently outside of economics, to the great loss of the latter.
A great divide developed between the academic mainstream and those who wrote about economic fluctuations, referred to by the earlier writers as 'crises' and later as the 'trade cycle', or business cycle. Prior to Keynes, writers on this subject were, almost without exception, not academic economists. They were ignored not only by their academic contemporaries, but also by later and contemporary writers on the history of economic thought. Of the half dozen books on the subject in my library, none has an explicit treatment of the subject. Surprisingly, this is true even of Schumpeter's history; even though he had written a book on the business cycle. The treatment of the subject in Spiegel (1971) is typical. He has a short section on Schumpeter that contains the following passage: Bunches of innovation, reinforced by imitators and speculators, would make for cyclical movements, with the economy pulsating to the threefold rhythm of the threeyear Kitchins, nine-year Juglars, and fifty-five-year Kondratieffs, so named by Schumpeter after their discoverers. (p. 545).
The first two authors are not mentioned elsewhere in Spiegel's book. Kondratieff is briefly mentioned elsewhere under the topics 'Soviet Economics' and 'Econometrics'. The book contains no section explicitly devoted to business cycles.
From entries in The New Palgrave it can be learned that Kitchin was a journalist and businessman. Juglar was by training a physician who similarly to Quesnay drew an analogy to human physiology, discovering in the business cycle the 'heartbeat' of the economy. Kondratieff, finally, was an economist and statistician in the Marxist tradition.
The most extensive discussion of business cycle theories of which I am aware is Gottfried Haberler's Prosperity and Depression. Originally commissioned by the League of Nations, the first edition appeared in 1937. Haberler, himself a prominent academic, focuses on contribution to theory, mainly empirical work such as that of Kitchin or Juglar is not mentioned. Also, he focuses on contributions that were recent at the time of his writing; a time with more academic contributions than earlier periods, since the Great Depression could not be ignored even by the academic mainstream. Given this caveat, I will use the book to convey an idea of the main contributors and their ideas. A valid remark that Haberler makes prior to introducing his classification of theories is that they are not mutually exclusive; rather each emphasizes a particular aspect without necessarily denying the others.
Under the heading of purely monetary theories of the cycle only one author is prominently mentioned: R. G. Hawtrey, who for most of his career was an official at the UK Treasury. In this theory the cycle is caused by fluctuation in the supply of money that are in turn caused by fluctuations in the willingness of banks to extend credit.
A broad range of theories are discussed under the heading of over-investment theories: The central theme of all these theories is the over-development of industries which produce producers' goods or capital goods in relation to industries producing consumers' goods. They all start from the universally admitted fact that the capitalgoods industries are much more severely affected by the business cycle than industries which produce for current consumption. During the upward phase of the cycle the output of producers' goods rises much more, and during the downward phase is much more curtailed, than the output of perishable consumers' goods. Durable consumers' goods, such as houses and automobiles, are in a special position approximating to that of capital goods. (p. 29).
Two sub-groups of theories are distinguished. The first group is that of monetary overinvestment theories. The authors discussed under this heading are: F. A. Hayek, F. Machlup, L. Mises, L. Robbins, W. Röpke and R. Strigl. Under the heading of nonmonetary over-investment theories the principal authors are A. Spiethoff and G. Cassel whose line of thought can be traced back to Tugan-Baranowski and Marx. Prominent in these lists are authors from the German speaking area, not associated with the classical to marginalist mainstream.
The following chapter in Haberler's book is titled: Changes in cost, horizontal maladjustments and over-indebtedness as causes of crises and depressions. I limit myself to listing some of the prominent authors listed under this heading: W. C. Mitchell, A. C. Pigou, F. W. Taussig, Sir W. Beveridge, I. Fisher and others.
Under-consumption theories are the subject of Ch. 5. I quote from the first two paragraphs: The under-consumption theories have a long history. In fact, they are almost as old as the science of economics itself. Lord Lauderdale, Malthus. and Sismondi are prominent among the early adherents of this school of thought. The authors who have done most in recent times to re-state and propagate the under-consumption theory in a scientific way are Mr. J. A. Hobson in England, Messrs. W. T. Foster and W. Catchings in the United States, and Professor Emil Lederer in Germany. The cruder versions of the theory, which exist in innumerable varieties in all countries, will not be considered here, as their fallacy has been clearly demonstrated on various occasions. It is difficult to summarize these theories because, with some notable exceptions, their scientific standard is lower than the standard of those reviewed earlier in this volume. They cannot be reviewed as systematically as the over-investment and monetary explanations, for it is only in regard to certain phases of the cycle that these theories have anything original to contribute. The under-consumption theory is a theory of the crisis and depression rather than a theory of the cycle.
The discussion thus far has covered the principal theories and authors contributing to what may be called traditional business cycle theories. It was a largely empirical tradition that existed apart from the economic mainstream and observed a phenomenon that the mainstream asserted could not exist. The situation changed as a result of the Great Depression that even the mainstream could not ignore. But, the perception changed with the rise of Keynesianism. Keynes thought that economic conditions had altered fundamentally so that instead of traditional business cycles there would now be a permanently depressed situation due to insufficient demand. Ironically, he was inspired by the under-consumptionists whom Haberler above characterized as more primitive than other business cycle theorists.
Keynes was clearly wrong in assuming that permanent depression instead of business cycles would henceforth characterize the capitalist economies. His prestige largely contributed to the disappearance of the traditional business cycle theory. It is an example of the pathological nature of the transmission, or non-transmission, of ideas in economics.

The Early Econometric Movement
Early econometrics, even before that term came into use with the founding of the Econometric Society in 1930, had the analysis of business cycles as its principal object of study. The early studies were essentially time series analyses with the purpose of determining periodicities. 16 These studies were not successful for several reasons: a. The periodogram analysis originally employed is not a good method and was later abandoned in favour of spectral analysis which involves a smoothing of the periodogram. b. The statistical material available was very limited, particularly in relation to investment, which is the key variable in the business cycle theories discussed above. c. Uninterrupted economic time series tend to be short, whereas the traditional methods of time series analysis require lengthy series. 17 I turn to the econometric efforts at macroeconomic modeling, specifically with the purpose of explaining business cycles. In this connection I need to discuss a methodological issue. In the natural sciences, when it is said that the predictions of a theory have been confirmed, what is meant is that a characteristic pattern implied by the theory has been observed. In the natural sciences such patterns are referred to as empirical regularities in economics as stylized facts. In econometrics, prediction is taken to mean predicting the future values of some variables, which is entirely different. A further difference between natural science methodology and that of econometrics is the emphasis in the latter on statistical testing which plays hardly any role in natural science. 18 To illustrate: One of the most famous episodes in the history of science is Newton's deduction, from the postulate of universal gravitation, of Kepler's laws of planetary motion, in particular, that the planets follow elliptical orbits. The determination of Kepler's laws required complex three dimensional geometry, but no statistical testing. That the orbits were nearly elliptical was, given the proper calculations, obvious so that a statistical test was superfluous. A statistical test though would have rejected the hypothesis since the orbits are not perfectly elliptical.
An economic example can be given from the business cycle stylized facts. Short-run macroeconomic fluctuations are to a very large extent fluctuations of inventory investment or disinvestment. Many studies made at my institute and elsewhere, using a wide variety of different methods have found that the typical short-run fluctuation in aggregate output is accounted for to an extent of between 50 and 100 percent by 16 Excerpts from some of these are reprinted in Part II: Early Time-Series Analysis, of Hendry and Morgan (1995). 17 Maximum entropy spectral analysis is a modern method that can be employed with much shorter series. For a discussion of the application of this method to business cycle analysis see Hillinger (2005a). 18 I do not mean to imply that the many applications of statistics are in some sense unscientific. Instead, I mean that the typical application of statistics does not involve the testing of scientific laws. This point has also been made by Heckman (2001). fluctuations in inventory investment. Given that average inventory investment is only 1 or 2 percent of GDP, this is a startling and robust result. This is the type of stylized fact that any business cycle model should be expected to replicate. Again, this stylized fact is so massive and robust that statistical testing is superfluous. Such a test would also be difficult since the share of inventory investment in fluctuations, while large, fluctuates substantially itself. 19 The methodological position that I have just described will now be used to evaluate the evolution of econometrics and of Keynesian economics. The question is: To what extent did these developments advance our empirical knowledge.
'Cobweb' models of cycles in the prices and quantities of agricultural commodities began to appear around 1930. 20 Inspired by these Tinbergen (1959Tinbergen ( [1931) made the first application to an industrial cycle, namely the 'shipbuilding cycle' involving freight rates and shipping capacity for the British, American and German merchant marines. The paper begins with a figure showing graphs of these variables as deviations from trend. A cycle is evident, especially clearly in freight rates. This is the stylized fact that the model is intended to explain. The assumptions incorporated into the model are plausible in terms of our general understanding of how markets function as well as of characteristics of shipbuilding that can be verified by talking to the people involved. The basic assumptions are: a. When shipping capacity is low/high, freight rates are high/low. b. When freight rates are high/low, the rate of orders for new ships will be high/low and so will be the increase in tonnage or ship launchings θ years later, θ being the gestation lag. The fundamental equation obtained by Tinbergen is θ − is the tonnage θ years ago and f'(t) is current investment (increase in tonnage). He specifies the gestation lag for building a ship as two years and shows that, depending on the adjustment speed, a variety of cyclical and non-cyclical solutions are possible. Specifically, he obtains an 8 year cycle that he takes to be typical.
Both Frisch (1933) and Kalecki (1933Kalecki ( , [1966) took the decisive next step of constructing business cycle .models meant to be applicable to a national economy rather than an isolated market. The key variable of their models is the volume of fixed business investment.
The basic assumptions entering Frisch's model are as follows: a. Capital is required in fixed proportion for the production of consumer and producer goods. 19 The business cycle research conducted at my institute did produce a statistical test of business cycle stylized facts reported in Reiter and Woitek (1999). Their conclusion: …the data support the predictions of the classical writers: fixed investment tends to have more spectral mass in the frequency range of 7-10 years, inventory behavior in the range of 3-5 years. Looking at each country separately, these regularities are often not significant in the statistical sense, which is not too surprising given the shortness of the available time series. The significance comes mainly from the similarity of results across countries. 20 For a discussion see Pashigian (1987). The most prominent of these, the hog cycle is apparently still alive and well; see Stearns and Petry (1996).
www.economics-ejournal.org b. Orders for new capital goods have two components, a replacement component assumed proportional to total production and an acceleration component proportional to the rate of change of the production of consumption goods. c. Following an order, investment takes place at a constant rate over an intervalε , the gestation lag, until the completion of the capital good. d. Consumption is a constant, modified by a relationship between the fixed money supply and the transaction demand for money.
Frisch demonstrates that for plausible parameter values his model produces two cycles, regarding which he writes: The primary cycle of 8.57 years corresponds nearly exactly to the well-known long business cycle…Furthermore; the secondary cycle obtained is 3.50 years, which corresponds nearly exactly to the short business cycle. (p. 170).
The cyclical properties of the model are shown to depend sensitively on only one model parameter, the gestation lag ε . In part V of his paper, entitled 'Erratic shocks as a source of energy in maintaining oscillations', he gives the first precise definition of the type of irregular cyclical movements which I refer to as 'quasi-cycles'. He gives the example of a damped (frictional) pendulum which can be modeled by differential equations with complex roots, giving the period and damping of the oscillations. If the pendulum is subject to erratic shocks, they can be added to the deterministic equation. After analyzing the effect of the shocks, he concludes that: The result of this analysis is...a curve that is moving more or less regularly in cycles, the length of the period and also the amplitude being to some extent variable, their variations taking place, however, within such limits that it is reasonable to speak of an average period and average amplitude. In other words, there is created just the kind of curves which we know from actual statistical observations. (p. 183).
Unfortunately, much of the discussion of economic fluctuations to the present day suffers from a failure to use this clear-cut concept of a quasi-cycle.
In contrast to Frisch, who made no further contribution to this subject, macroeconomics was the main interest of Kalecki throughout his professional career. Despite his many articles and books on economic cycles, beginning in 1933 in Polish and after 1936 in English, his basic approach and model did not change very much. The basic formulation is contained in his 1933 paper 21 . In that paper Kalecki cites no references, and it is likely that he was unaware of the cobweb models mentioned earlier.
After moving to England, he published in 1935 a more elaborate paper in Econometrica in which he cited Frisch and Tinbergen and used their solution methods to analyze a highly sophisticated and carefully specified lag structure of the investment process.
Kalecki also makes a considerable effort to obtain range estimates of the parameters. He shows that, for plausible parameter values, the model generates cycles which lie in the observed range of 8-12 years. In subsequent publications (particularly Kalecki 1954(particularly Kalecki , [1966) he tested his model against different data-sets and obtained generally plausible results.
Regarding the investment process, Kalecki distinguishes between orders for investment goods, actual investment, which takes place subsequently for the duration of the gestation lag, and additions to the capital stock, which occur at the end of the gestation period.
The rest of Kalecki's model is rather unsatisfactory. He assumes that workers do not save and that they consume a constant amount. Capitalists base their investment decision on their gross profit without an accelerator effect or the consideration of capacity utilization. Gross profits in turn are in his model a simple accounting consequence of the level of investment.
Each of the three early econometricians whose work I have just discussed attempted to use the methodology of the natural sciences to explain macroeconomic fluctuations. Their attempts remained isolated. Part of the reason is the rise of Keynesianism that will be discussed next, but that is not the only reason. Some other aspects will be mentioned here.
The only one of the three authors who had a substantial impact on the subsequent evolution of macro-econometrics is Tinbergen. In turning from the original paper on the shipbuilding cycle to general macro-econometric modeling, Tinbergen abandoned his original methodology. He now concentrated on the statistical testing of the individual equations entering his models, rather than on the dynamic properties of the complete model. This became the style of subsequent macro-econometrics.
In the case of Frisch, his style was to make an important contribution to some problem and then to move on to other problems. His model of the interaction of the long and short cycles remained an isolated effort.
Kalecki, a Polish Jew and a Marxist, without any formal training in economics was the consummate outsider, a position that he retained after returning to Poland to participate in economic planning. It is not surprising that he was eclipsed by Keynes, the consummate insider.
An important element in the neglect of the work of the three authors is the difficulty of the mathematics that they used: mixed difference-differential equations. The mathematics of these is substantially more difficult than that of pure difference or differential equations. I surmise that at the time the three authors were the only ones in economics to understand the mathematics; nor was this ever a topic to engage economists subsequently. An ironic aspect of the situation is that there was no plausible reason for adopting the mathematics of Tinbergen's shipbuilding paper to macroeconomic models. The many different firms in the economy will have gestation lags of different length and they will start capital projects at different times; consequently, aggregate investment will be more nearly continuous than discrete.
Above all, I believe that the lack of response to this work is an instance of a more general problem, namely, the failure to establish in economics the natural science tradition of intensively studying stylized fact so as to determine them with a maximum of precision and to secure a collective agreement with regard to them. This theme will come up again in Section 8 in relation to the quantity theory of money.

Keynesian Economics
A vast literature deals with the 'Keynesian Revolution' most of it devoted to doctrinal controversies. I will emphasize here aspects that have received little discussion.
The evolution of classical and neoclassical economics had been driven by the desire to elaborate the formal apparatus of the theory of markets in perfect competitive equilibrium. This required no empirical input. Empiricists were attracted to the more negative aspects of competitive markets: crises, business cycles, social inequality and the cultural impoverishment of the working class. These problems were marginalized by the representatives of the mainstream. Keynes changed that; his 1936 General Theory of Employment Interest and Money (GT) was directly inspired by the Great Depression, the dominant empirical phenomenon of his time. He believed that depression due to insufficient aggregate demand would henceforce be characteristic for advanced capitalist economies unless compensated by government expenditure. He regarded structurally determined insufficient demand as an essentially static problem with the consequence that the main part of the GT, intended to deal with the problem, is a static theory. Had Keynes, the empiricist, lived longer into the postwar era he would undoubtedly have realized that the basic assumption underlying his GT, the indefinite continuation of the conditions of the Great Depression, was invalid. What kind of theory is Keynes likely to have produced, given that the rhythm of the business cycle, with its alternation of periods of prosperity and recession reestablished itself. At first glance the question seems absurd; how can anyone tell what Keynes would have thought had he lived much longer and maintained his intellectual vigor? In fact, a reasoned answer can be given because it is outlined in neglected chapters of the GT.
I have stressed that the GT was motivated by Keynes's vision of a post-Depression world of mature capitalism, characterized by stagnation and chronically deficient demand. Most of the GT elaborates a model of such an economy. However, in Chapter 22, 'Notes on the Trade Cycle', Keynes cast a backward glance at a world which he thought had ceased to exist. He attempted a brief explanation of economic cycles which he explicitly described as a 'nineteenth-century phenomenon'.
The building blocks for Keynes' dynamic theory are contained in Chapters 5 and 12 on short-and long-run expectations. The entrepreneur is described as making his current decisions on the basis of his expectations regarding the future. Future expectations are based on extrapolations from the past, but these may be strongly influenced by irrational or volatile factors of individual or mass psychology. Keynes discusses two fundamental decisions of the firm. In Chapter 5 it is the decision on how much to produce, which is related to short-term expectations regarding the demand for the product and also to current inventory levels. Chapter 12 is devoted to the firm's decision to invest in fixed capital. Here the decisive consideration is the relationship between the current cost of capital and its expected long-term yield. The current cost depends on the price of capital goods and the rate of interest, the expected long-run evolution of market demand and on costs of producing with old or new capital.
Keynes' attempt in Chapter 22 to construct an endogenous dynamic theory of economic cycles, based on these building blocks, has remained rudimentary. The main reason appears to be that he devoted little effort to the task, since he regarded the chapter as no more than an historical aside. Also relevant is the fact that a logically tight description of the dynamics of an oscillatory process is virtually precluded by the rather diffuse verbal style of Keynes and earlier writers on economic cycles. 22 Considering the three chapters together, Keynes made a substantial contribution to the explanation of economic fluctuations that was firmly in the tradition of the theories of investment cycles. Keynes' reflections on business cycles were a major influence on my own work.
Keynesian economics, particularly in the IS-LM formalization given to it by Hicks, dominated academic economics for several decades. The exaggerated claims connected with latter day Keynesianism, would hardly have been supported by Keynes himself. This is particularly true of the claim that the large-scale macroeconometric models would enable policy makers to ban the business cycle and usher in the age of perpetual equilibrium growth. 23 The theory of equilibrium growth became a prime occupation of macro theorists. The turn away from Keynesianism towards monetarism and subsequently real business cycle theory and representative agent models will be the subject of subsequent sections.
The turn away from Keynesian theory in academia cannot alter the profound and lasting impact that Keynes has had on macroeconomics. He played a key role in the creation of the national income and product accounts. 24 The structure of the accounts reflects that of the GT both in terms of the definition of the major sectors of the economy as well as in the emphasis on flows rather then on stocks. In the postwar period efforts were made to expand the accounts so as to integrate real and financial stocks with the flow accounts. As these efforts lacked a patron with the influence of Keynes, and as the economics profession generally moved away from an interest in measurement, the efforts remained in limbo.

Measurement
The above narrative touched on measurement at many points. The importance of measurement in science and the exceedingly strange history of measurement first in economics, later also in other social sciences, make it advisable to pull the threads together in this subsection.
The crucial event in this story is the invention of the utility concept by the Utilitarians. They stressed that utility is a subjective magnitude, different from money income. In this they were correct. The importance of the distinction is confirmed by modern happiness research, where it is demonstrated that above a certain minimum income, factors other than money income become important determinants of happiness. 25 Equally important was the Utilitarians idea that the goal of economic policy should be the maximization of aggregate utility. This marked the change from mercantilism, where the aim of economic policy was seen as the maximization of the ruler's treasure, to the modern conception that the purpose of economic activity is to benefit the society.
The difficulty with the utilitarian position was that they did not know how to measure utility, though they were convinced that it was measurable in principle. They also thought that when a way to measure utility was found, it would be done as in the natural sciences by means of a cardinal scale. This would allow the total utility to be determined by simple addition of the numerical values obtained for the individual utilities. Lacking both a mathematical model of the economy as well as precise quantitative measurements, the utilitarian maxim of 'the greatest happiness for the greatest number' remained a vague exhortation rather than a precise guide to action.
The nature and implication of utility as a subjective magnitude has remained the central and unresolved problem of economics throughout its history. Since the economic approach came to dominate collective choice theory, it has also become the central and unresolved problem of political theory.
The marginalists, in their endeavor to replace classical economics, rejected the utilitarian analytical apparatus. They showed that the consumer's optimum is defined by the condition that the rates of substitution between commodities that keep utility constant is equal to the inverse of the ratio of their prices. In this formulation there is no need to consider any utility increments. They argued further that such increments could not be observed and that they were therefore a useless metaphysical construct. This pointed the way towards utility as an ordinal concept, with indifference curves indicating rates of substitution, and utility levels regarded as irrelevant.
This argumentation on the part of the early marginalists was not entirely fair, since they were able to do without cardinal utility only because they ignored the problem for the solution of which cardinal utility was invented: the determination of a social optimum. With the rise of general equilibrium theory this problem came again to the fore. It was solved in the spirit of marginalism by Pareto. He showed that the equilibrium that results under perfect competition has the property that not all individuals can be made better off. This condition of 'Pareto optimality' also makes no use of any utility increments. The Pareto optimum demonstrates the efficiency of a competitive economy. It is weaker than the optimum optimorum envisaged by the Utilitarians, since the income distribution generated by a competitive economy may be unacceptable. On this the Pareto principle is silent.
The next crucial development was the publication in 1932 of Lionel Robbins An Essay on the Nature and Significance of Economic Science. The essay was directed against the socialist claim that economic and social policy could and should be based on science. Robbins' argument was based on the marginalist position according to which interpersonal utility comparisons are impossible. Actually, Robbins made the more modest claim that such comparisons lacked a scientific basis. Since the determination of a social optimum apparently requires such comparisons, it follows that policy aiming at the attainment of such a state cannot be entirely scientific.
What Robbins actually said should not have been very controversial. A reasoned reply came from Myrdal (1969) who argued that those who propose a policy should make their values explicit. Unfortunately, Robbins' essay did not lead to a reasoned debate among economists. Instead, the stronger assertion that 'interpersonal comparisons of utility are impossible!' usually preceded by 'Of course!' became a mantra by means of which economists signaled that they were members in good standing of the academic economic mainstream. This strong form of impossibility assumption became the foundation of Arrow's work and, as discussed in Sections 2 and 7, of modern collective choice theory generally.
The conviction of the impossibility of interpersonal utility comparisons also impacted economic measurement. The pervasive economic measure is money. The Utilitarians had made the distinction between money and utility and had given priority www.economics-ejournal.org to the latter. This amounted to a de-emphasis of the money measure. As already described, the first half of the Twentieth Century saw a great expansion of social and economic statistics, produced with the aim of promoting a scientific approach to the solution of social problems. Many of these measures were expressed in monetary form, most importantly GDP. For economic theorists there arose the problem of relating the monetary measures to changes in welfare that in turn had to be defined in relation to utility.
For the GDP, the solution was sought via the Pareto criterion and various compensation measures. It was argued that an increase of GDP signified that the winners could compensate the losers, so that, at least in principle, the Pareto criterion would signify an improvement.
The other measure that theorists paid much attention too was Dupuit's measure of consumer surplus. It was typically used to calculate if the benefit of a project that would result in a lower price to consumers for some service would result in a greater benefit then the cost involved.
Around 1950, Paul Samuelson demonstrated that all of these attempts were irreparably flawed. The consequence was that economic theorists largely abandoned problems of measurement. This did not in any way reduce their practical importance; instead, theory and practice simply parted ways.
It is highly ironic that cardinal measurement was abandoned by theorists in economics and political science, precisely the two fields that have natural cardinal measures in the form of money and the ballot. The opposite path was taken in psychology. Here central concepts such as intelligence, or various emotions, have no natural metric, but immense efforts were made to construct such metrics. Sophisticated statistical methods such as factor analysis and multi-dimensional scaling were developed for the purpose. Where scientism in economics concentrated on the construction of mathematical models, in psychology it concentrated on measurement.
As already described, my own work on measurement and on voting essentially involved the rejection of the self-defeating and ideologically motivated commitment to ordinalism. Instead, I elaborated the theory of the naturally available cardinal measures.

Representative Agents and the decline of the Empirical Tradition
The representative agent concept has come to dominate applied economics. The reason is that applied economics almost always deals with aggregate data, while the central concept of economic theory is the maximizing agent. The representative agent combines these two aspects in the simplest possible manner by assuming that the aggregate data behave as though they were the outcome of the decision making of a single maximizing agent.
For my discussion in this section I rely on Hartley (1997) The Representative agent in Macroeconomics as well as on other sources and my own relevant work. As the title of his book indicates, Hartley is primarily concerned with the use of representative agent models in the contemporary macroeconomic mainstream that is referred to as new classical macroeconomics or real business cycle theory. This is somewhat limited, since the representative agent concept is also central to modern public economics and the difficulties connected with it are the same, regardless of which application is being considered. The book does have an excellent and detailed discussion of the historical origins of the concept.
The representative agent is a very strange concept because it has been rejected by virtually everyone who has seriously considered the question of its justification; this rejection however has done nothing to curtail its popularity. The representative agent model would be justified if it were possible, starting from the assumption of individually maximizing agents, to deduce that observable aggregate variables would behave as though they were the outcome of the maximizing behavior of a single agent. That such a drastic reduction in complexity should be possible is implausible. Aggregation theory confirms this judgment and shows that such a result is possible only under implausible conditions.
Hartley (1997) writes: Under what conditions will we be able to derive a consistent representative agent (or equivalently a macro model)? The most important conditions are in Gorman (1953): If all agents have parallel, linear Engel curves, or equivalently, if all agents have identical homothetic preferences, consistent aggregation is possible…In sum, the conditions for consistent aggregation are so severe that we can safely say that they do not hold in reality. (p. 134-5).
In concluding his review of the literature on representative agents Kirman (1992) writes: Given the arguments presented here -that well-behaved individuals need not produce a well-behaved representative agent; that the reaction of a representative agent to change need not reflect how the individuals of the economy would respond to change; that the preferences of a representative agent over choices may be diametrically opposed to those of society as a whole -it is clear that the representative agent should have no future. (p. 134).
Franklin M. Fisher (1987) gives a broad review of all aspects of the aggregation problem; over consumers, over firms, and over commodities. He concludes: Such results show that the analytic use of such aggregates as 'capital', 'output', 'labour' or 'investment' as though the production side of the economy could be treated as a single firm is without sound foundation. This has not discouraged macroeconomists from continuing to work in such terms. (p. 55).
How is it possible that this devastating and largely unanimous criticism of the representative agent concept is essentially being ignored by the mainstream? Much of Harley's book is devoted to finding answers. Before turning to these, some preliminary comments: It should be pointed out first of all that the rejection of the representative agent is by no means a rejection of macroeconomics, in the sense of constructing and estimating models involving relationships between macro-variables. Relationships between macrovariables have long been observed and are fairly stable. 26 That relationships between micro-variables that generally go in the same direction at the individual level should be evident also in macro-data is highly plausible. For example, individuals consume on average roughly constant proportions of their income. Unless there are large shifts in the income distribution between households with different propensities to consume, a similar relationship will also be observed in the aggregate. In traditional macroeconomic modeling, such relationships are simply estimated from the data. In contrast to this, representative agent models impose prior restrictions on the estimate. These restrictions have two sources: one is the general maximization assumption. An example would be the Slutsky equation imposed on estimates of consumer demand functions. Other restrictions result from the specific parametric functions that are assumed in order to make the econometrics of these models tractable, for example Cobb-Douglas or CES production or utility functions. All of these restrictions are illegitimate, given that the representative agent model itself is flawed.
At this point I introduce a distinction and terminology used by Hartley. Marshall had founded partial equilibrium analysis with the idea that an industry may respond to a common disturbance with the same kind of response as the individual firms comprising it. In fact, he used this condition to define an industry. In this context he also referred to the response of the industry as being that of a representative firm. This use of the term is innocuous, but also superfluous. This type of analysis will be referred to as Marshallian and this term can be applied also to the associated concept of a representative firm.
The representative agent models used in contemporary economics are Walrasian in that they embody the central assumptions originally employed by Leon Walras. The basic substantial assumptions are that the economy is made up of rational agents, maximizing profits in the case of firms and utility in the case of consumers. The agents are always in a state of competitive equilibrium; when this is disturbed, they move instantaneously to the new equilibrium. From the methodological point of view, the emphasis is on the rigor of the mathematical analysis of the models; empirical relevance of the models is not seriously considered or tested.

The Intellectual Background
Immediately following The Second World War and during the following decades, socialism appeared to be the wave of the future. This was certainly the view of almost all intellectuals, of artists and of the rising elites of the newly liberated Third World countries. Concerted efforts were made at two American institutions to advance a counter ideology of unrestrained markets and limited democratic government. The earlier effort centered on the Rand Corporation, another was started about two decades later at the University of Chicago. Each of these efforts is described in a recent book; the Rand story is covered in Amadae (2003), that relating to Chicago in Leeson (2000). In this section I deal with the Chicago story because it is the logical continuation of the discussion of Keynesianism and early econometrics. First though I discuss the intellectual atmosphere that motivated the counterattacks organized at Rand and at Chicago. Amadae (2003) describes it as follows: …during the 1930s and 1940s there was a pervasive sense of dismay and defeat among the intellectuals of the West over what they took to be the inevitability of the triumph, both internal and external, of fascist or communist alternatives to democratic capitalism. Marxism -forged during the early days of mass production, mass warfare, and mass democracy -appeared to be the inevitable, if not yet victorious, structural principle that eventually would govern world affairs. (p. 2) The most prominent defenders of political and economic freedom in the Interwar Period were themselves pessimistic: Schumpeter's Capitalism, Socialism, and Democracy (1943), Hayek's The Road to Serfdom (1944), and Popper's The Open Society and Its Enemies (1945) convey the tenuous foothold of Western institutions of democracy, capitalism, and science before the rational choice revolution. These three widely read books, each expressing grave apprehension over the future viability of these sacred institutions, outlined the philosophical crises at the heart of Western civilization. (p. 16) Leeson (2000, pp. 14-15) has discussed the affinity of the early econometric movement to socialism: The econometrics movement was molded, to a large extent, by the desire to understand and tame the interwar business cycle… Keynesian and Marxian economics are modern versions of the 'endogenous instability of capitalism' thesis…The Great Depression gave an added dimension to these controversies. In the 1930s, many observers were concerned about the long run viability of capitalism and of the apparently infeasible combination of political liberty and economic freedom…Marschak initiated one of the earlier debates on the viability of socialism as an economic system, involving Pareto, Barone, von Mises, Schumpeter, von Hayek, Lange and Lerner. The 1940 Cowles Commission Report stated that unemployment was the primary economic problem to be tackled…Many economists lost their faith in the ability of markets to solve the problem of unemployment, and many embraced the new faith of economic planning. Tinbergen...for example, retrained as an economist under the influence of the onset of world depression; he regarded his econometric work, and in particular his estimation of parameters, as providing the tools to effect socialist intervention in the economy in order to minimize cyclical fluctuations and poverty… Lawrence Klein shared this approach to econometrics; later he would be persecuted because of his socialist convictions. Harold Hotelling also favored market socialism…Oscar Lange was a 'proclaimed socialist', and later a member of the Polish Communist government…Ragnar Frisch also had socialist leanings, according to Tinbergen…Frisch came to believe that uncovering the underlying structure of the economy -the structural parameters -would enable the business cycle to be tamed…Part of this optimism may reflect the initial training in physics which Tinbergen, Frisch, Koopmans and others had been exposed to…These ideological undercurrents were present in many of the business cycle research institutes which were established all over Europe and the United States in the 1920s.

Chicago Schools
Leeson who is the primary source for the present section focuses on the ideological campaign launched by Friedman and Stigler at the economics department of the University of Chicago during the 1960s. Both are however members of a tradition that dates further back. Reder (1987) dates the beginning of an identifiable 'Chicago School' to around 1930 with the arrival in the department of Frank H. Knight and Jacob Viner.
What Knight and Viner had in common was a continuing adherence to the main tenets of neoclassical price theory and resistance to the theoretical innovations of the www.economics-ejournal.org 1930s, Monopolistic Competition and Keynes's General Theory. This theoretical posture paralleled an antipathy to the interventionist aspects of the New Deal and the full employment Keynesianism of its later years. (p. 414).

Reder (1987) identifies a further important characteristic of the Chicago School:
For over half a century, the need to prepare for course and preliminary examinations, especially in price theory, has provided a disciplinary -cultural matrix for Chicago students…Although the specific content of examination questions has evolved with the development of the science, the basic paradigm remains substantially unchanged: economic phenomena are to be explained primarily as the outcome of decisions about quantities made by optimizing individuals who take market prices as data with the (quantity) decisions being coordinated through markets in which prices are determined so as to make aggregate quantities demanded equal to aggregate quantities supplied. (p.414).

Friedman and Stigler
The above characteristics of the Chicago School have remained in place; however with the advent of Friedman and Stigler the ideological battle assumed a new intensity and novel methods were employed in waging it. A further dramatic change occurred with the advent of Robert E. Lucas. I will therefore distinguish three Chicago Schools, shaped by Knight and Viner; Friedman and Stigler; finally, Lucas. After Lucas, the Chicago School merges with general trends in the economics profession that were however strongly influenced by the Chicago School. As a label for the political ideology of the Chicago School in all three of its manifestations I will use the current term 'neoliberalism' that has the same meaning as 'neoconservatism', the term often employed in the United States.
Both Friedman and Stigler viewed themselves as carrying forward the tradition of Frank Knight. The following quotation from Leeson (2000) illustrates both continuity and dramatic changes in emphasis: …Friedman and Stigler did not share Knight's fatalistic despair about the Chicago project for social and economic transformation… neither did they share his lack of interest in empirical or policy-oriented economics or his belief in the inherent contradiction between thought and action…. Both attended his seminars on the sociologist Max Weber; he was their dominant dinner- Leeson's description of this event as a 'conversion' is well chosen. The meeting resembles one of evangelists, culminating in a mass conversion. It is instructive to examine the 'Coase Theorem', as it was later called, that elicited such religious fervor in order to gain an insight into the nature of ideological beliefs. The theorem states that in the presence of externalities private parties can arrive at an efficient solution in the absence of government regulation. In the case of an external bad, those affected by it can pay the offender to desist. In the case of an external good, those who desire to have it can pay the producer to supply it. Prior to Coase, the standard position on externalities had been that of Pigou who advocated government intervention in order to restore efficiency.
As Coase well realized, his theorem requires a. the absence of transaction costs. Another obstacle b. that is prominent in the literature is the 'free rider' problem that arises whenever many individuals are subjected to an externality. I believe that there are several other serious obstacles. One c. typically occurs when a developer wishes to acquire a number of adjacent plots in order to realize a project. If his plan becomes public when he has acquired most of the plots, the remaining land owners can hold out for unreasonably high payments, since they are in a position to block the project. d. The compensation required may be larger than the means available to effected persons. In Germany there was a few years ago a legal case in which the owners of land and dwellings below a railroad bridge sued the national German railroad to force it to take action to prevent the flushing of toilets as the trains passed overhead. I don't know the outcome of the legal case, but I think that a payment to the railroad that would cause it to voluntarily desist, creating a precedent, would far exceed the cost of abandoning the land. e. Finally, as is suggested by the previous example, most people would feel a moral revulsion at paying an offender to desist. Such payments would also be an incentive for deliberate production of negative externalities in order to be paid for ceasing. The usual name given to such an action is extortion. For all of these reasons, private payments to deal with externalities are almost never observed. In reality the importance of the Coase solution is negligible relative to the Pigou solution. 27 How was it possible for a result that objectively appears quite modest to have been received with quasi-religious euphoria? I believe the answer is 'tunnel vision'. I borrow the term from Schaef (1987). The author is a psychologist who worked with Alcoholics Anonymous and applied insights obtained there to society at large. In AA it was observed that reformed alcoholics often acquired a new addiction. Sometimes it was another drug, such as smoking. However, the new addiction could also be to an idea; most often a fanatical desire to reform other alcoholics. Schaef found the phenomenon of addiction to be a characteristic feature of modern societies, attaching itself to a variety of activities, feelings and ideas. All consuming drives to consume, or to work ('workaholic') are obvious examples. Ideologies also produce fanatical adherents, willing to kill if they attain the means, as history abundantly testifies. To the intellectual, living in a more stable and peaceful society, the task becomes that of impaling the opponent intellectually with a stiletto-like argument.
I cannot in this space summarize Schaef's book on the psychology of addiction, but I will mention two characteristic traits: One is tunnel vision, which means that the addict (ideologist) ignores or denies those aspects of reality that contradict the picture of the world that he is attached to. The other is a diminishment of the moral sense. The alleged ultimate benefits promised by the ideology are taken to justify the twisting of truth, or even outright lies and other evil deeds in the present.
The religious fervor, biased argumentation, selective use of evidence and aggressive debating style that characterize the Chicago School as shaped by Friedman and Stigler can in my view be completely understood only by means of the psychology of addiction.

The Architecture of the Friedman/Stigler Chicago School
The logical structure created by Friedman and Stigler to advance their ideological agenda can be compared to a three-tiered building. The upper floor contains the neoliberal ideology. It is supported by the lower floors, each of which consists of two parts. The ideology rests directly on the second floor that contains the appropriate economic theory; the micro-part being the responsibility of Stigler, the macro-part of Friedman. The ground floor contains Friedman's methodology of positive economics, used to justify the theory constructed on the floor above. It also contains Stigler's sociology of economics, the basis for the effective advocacy of theory and ideology. In the writings of the two and of their followers the four building blocks tend to intermingle, with sometimes one, sometimes another assuming center stage.

Stigler's Sociology of Economics
Some preliminary remarks: a. Stigler often refers to economics and economists, but sometimes to science generally. I don't believe that he ever seriously studied the natural sciences. His writings in a sociological vein should be interpreted as being about economics, not science generally. b. He writes impressionistically, his claims are hardly ever backed by evidence; of course, that does not mean that they are false. c. Following Leeson, I also use the term 'sociology'. The sociology of science is a well established field of sociology, founded by Robert K. Merton. 28 Stigler's writings are not up to the standard established there; they can better be regarded as sociological impressions and as guides to persuasion.
A key feature of Stigler's sociology is elitism. I quote from Leeson's (2000) section entitled 'The elite and the masses'.
Stigler clearly distinguished between 'major scientific entrepreneurs' and the rest, some of whom could only employ 'an inferior mind', and some of whom were 'ersatz economists'. They entered the market as demanders, not suppliers, of ideas… Stigler concluded in his study of 'The Literature of Economics' that two-thirds of the articles surveyed were virtually worthless. There were commonly only about six really first-class scholars in any field; a small minority in the profession had 'superb instincts' with regard to the pursuit of ideas…for the 'mass' of scholars in any discipline, risk aversion, and a desire to preserve already acquired human capital, created a bias in favour of scientific conservatism. Given this structure of the sociology of economic knowledge, the 'most irresistible' of all the weapons of scholarship was 'infinite repetition' a 'form of the classical Chinese torture'. (pp. 51-52).
I quote Leeson (2000) on what in Stigler's view determines the long run acceptance of theories in economics.
Stigler noted that economists have a tendency to 'float on the tide of theory'. He reflected that empirical generalizations 'fail to achieve the continuity and the widespread influence of the formal theories' He was also aware that for theorists, statistically derived relationships could be 'frankensteins over which he has little or no control'… With respect to conclusions that contained policy implications, Stigler sought to elevate received theory over empirical analysis. The reason for this confidence in orthodoxy was that it was 'our most tested and reliable instrument for relating policies to effects'. Received theory, presumably, operated with a considerable advantage. The idea that a new theory 'is presumed innocent until shown guilty…is the exact opposite of the presumption I would use'. Not all theorists were to be trusted; unorthodox approaches were sometimes denied the label 'theory'. (p. 50).
Stigler's sociology of economics is a defense of orthodoxy. This defense is based on what I would call the 'Stigler twist'. Neither Stigler nor Friedman denies that only empirical testing should ultimately decide the fate of theories. Both stress however that testing is difficult and laborious and particular tests often indecisive. Stigler argues that orthodoxy has become that because of the long period of testing that it has successfully survived. We should therefore be biased in favour of orthodoxy. Under the Stigler twist, what begins as an empirical question ends up as a methodological prescription. Of course, that orthodoxy has become that because of a long history of successful empirical testing is precisely what its critics would deny. From Marx to current opponents of globalization, they would argue that the persistence of orthodoxy is due to the fact that it provides the argumentative foundation for policies that favour the rich and powerful, who in turn have the better means of influencing opinion.

Friedman's Methodology of Positive Economics
Leeson has very little to say about Friedman's methodology of positive economics. This is surprising given how influential this methodology has been, both generally and in advancing the Friedman/Stigler agenda. I believe that for most economists Friedman's methodology can be expressed in one sentence: 'Realism of the assumptions of a theory does not matter; all that matters is the theory's predictive success.' I will refer to this as the 'vulgar interpretation' of Friedman's methodology. I confess to having believed in this interpretation myself until rereading The Methodology of Positive Economics for the purpose of writing this section. What Friedman actually wrote is quite different and to my surprise I found myself in virtually complete agreement. 29 In the following I will review what Friedman actually wrote, followed by an evaluation and a discussion of the damage done by the vulgar interpretation.
The interpretation of Friedman's methodology hinges on the meaning one assigns to 'plausibility of assumptions'. Friedman makes it very clear that his purpose is to defend economic theory against those who attack its lack of 'descriptive realism'. The following passage defines the essence of his position and also points to the economic controversies that motivated his exposition: In so far as a theory can be said to have 'assumptions' at all, and in so far as their 'realism' can be judged independently of the validity of predictions, the relation between the significance of a theory and the 'realism' of its 'assumptions' is almost the opposite of that suggested by the view under criticism. Truly important and significant hypotheses will be found to have 'assumptions' that are wildly inaccurate descriptive representations of reality, and, in general, the more significant the theory, the more unrealistic the assumptions (in this sense). The reason is simple. A hypothesis is important if it 'explains' much by little, that is, if it abstracts the common and crucial elements from the mass of complex and detailed circumstances surrounding the phenomena to be explained and permits valid predictions on the basis of them alone… The theory of monopolistic and imperfect competition is one example of the neglect in economic theory of these propositions. The development of this analysis was explicitly motivated, and its wide acceptance and approval largely explained, by the belief that the assumptions of 'perfect competition' or 'perfect monopoly' said to underlie neoclassical economic theory are a false image of reality. And this belief was itself based almost entirely on the directly perceived descriptive inaccuracy of the assumptions rather than on any recognized contradiction of predictions derived from neoclassical economic theory. (pp. 14-15).
The most salient characteristic of the Chicago School is generally taken to be an unqualified belief in the competitiveness of existing markets. It is interesting to read what Friedman had to say about this: The confusion between descriptive accuracy and analytical relevance has led not only to criticisms of economic theory on largely irrelevant grounds but also to misunderstanding of economic theory and misdirection of efforts to repair supposed defects. 'Ideal types' in the abstract model developed by economic theorists have been regarded as strictly descriptive categories intended to correspond directly and fully to entities in the real world independently of the purpose for which the model is being used. The obvious discrepancies have led to necessarily unsuccessful attempts to construct theories on the basis of categories intended to be fully descriptive.
This tendency is perhaps most clearly illustrated by the interpretation given to the concepts of 'perfect competition' and 'monopoly' and the development of the theory of 'monopolistic' or 'imperfect competition'. Marshall, it is said, assumed 'perfect competition'; perhaps there once was such a thing. But clearly there is no longer, and we must therefore discard his theories. The reader will search long and hard-and I predict unsuccessfully-to find in Marshall any explicit assumption about perfect competition or any assertion that in a descriptive sense the world is composed of atomistic firms engaged in perfect competition. Rather, he will find Marshall saying: 'At one extreme are world markets in which competition acts directly from all parts of the globe; and at the other those secluded markets in which all direct competition from afar is shut out, though indirect and transmitted competition may make itself felt even in these; and about midway between these extremes lie the great majority of the markets which the economist and the business man have to study.' Marshall took the world as it is; he sought to construct an 'engine' to analyze it, not a photographic reproduction of it. (pp. 34-35).
Friedman gives an example involving the American cigarette industry during The Second World War. The firms appear to have been mainly concerned about maintaining market share and the predictions of the competitive model would have been falsified. The balanced position taken by Friedman in his 1953 essay may not be typical of his later views or of the Chicago School as an institution. The latter, according to Bhagwati, cited by Leeson , was …very Friedmanesque... The seminars seemed to oscillate between proving that elasticities were large with markets therefore stable, and formulating competitive hypotheses for apparently imperfectly-competitive industries and coming up with high enough 2 R s. Econometrics was the handmaiden of ideology: things looked imperfect to the naked eye, especially to that of Chamberlin and Joan Robinson, but they were 'really' not so and the world was 'as if' competitive...market imperfections were 'demonstrated' to be negligible and the imperfections rather of government intervention were the subject of active research. (p. 58).
My conclusion is that Friedman's essay on methodology is excellent and still very worthwhile reading for an economist. In essence it is a defense of the scientific method, with its emphasis on simplicity, against the claims of the adherents of descriptive realism. The limitation of the essay is that it very largely equates plausibility with descriptive realism. This is both false and the root of the vulgar interpretation that has done much harm. Plausibility is a much more general concept. The claims of critics notwithstanding, the basic assumptions of economic theory, that consumers try to get a maximum of satisfaction from their purchases and that firms attempt to maximize their expected long-run profits, are in fact highly plausible and constantly born out by experience. 30 Criticism is often aimed at a straw man, the homo oeconomicus, assumed to have complete information and perfect rationality. Economic theorists bear some fault for these misinterpretations because they seldom elaborate the concrete functioning of their concepts in empirical contexts.

Economic Theory and the Chicago School
I will deal only briefly with the microeconomic position of the Chicago School. There are two reasons: One is that this is not my own field of research and expertise. The other is that the microeconomic position of the Chicago School is not very original. Essentially it is an affirmation of Marshallian theory against the claims of the proponents of the theory of imperfect competition formulated by Edward H. Chamberlin and Joan Robinson. I believe that this position is correct; the theory of imperfect competition is simply not being used in the empirical analysis of economic conditions. Empirical analysis generally assumes competitive or oligopolistic industries, or monopolies. This is the position taken by Friedman in his essay on positive economics. Beyond this, I do not wish to deny a tendency on the part of members of the Chicago School, and more generally of adherents to the neoliberal ideology, to claim the existence competitive conditions where these may be objectively in doubt.
The macroeconomic position of the Chicago School, as developed by Friedman, is also simple to state. More complex is the history of its influence.
Friedman's macroeconomics stands Keynes on the head. Keynes believed the economy to be without a self-regulating mechanism that would lead to full employment. An actively managed fiscal policy was required. Monetary policy would be ineffective, due to various 'pathologies', particularly the 'liquidity trap'. Friedman takes the contrary position: The economy is inherently stable. Deviations from full employment are due to exogenous shocks, primarily due to an erratic monetary policy. The cure is to expand the money stock at a constant rate.
Friedman was instrumental in rehabilitating the quantity theory of money; much empirical work on it was done by members of the Chicago faculty and graduate students. An implication is that in order to have price stability, or a very low inflation rate, the money stock should grow at, or slightly above, the growth rate of real GDP.
Leeson very largely attributes the success of the Friedman/Stigler Chicago School to the efforts and persuasive skills of these two economists. I am more skeptical and a believer in the internal dynamics of ideologies. Almost by definition, ideologists have tunnel vision, ignoring or denying the aspects or reality uncongenial to them. Reality however has a way of catching up. The first half of the Twentieth Century witnessed the ascent of totalitarianism and socialism or its milder kin, social democracy, or American style liberalism. Even the Nazis called themselves Nationalsozialisten. Around 1970 the governments of the left were everywhere faltering and on the defense. They were experiencing rising unemployment, rising inflation, rising state deficits and ever larger public debts. The standard Keynesian remedies that they professed to be following no longer worked. Two related ideologies were thus on the defense; the broad political ideology of the left and Keynesianism as the set of macroeconomic beliefs shared by mainstream economists. The common element in these ideologies was the faith in active government intervention. The stage was thus set for a reversal of direction in both politics and economics. Who had how much effect in actually bringing the changes about? I cannot answer that question, nor does it interest me very much. Leeson stresses the influence of the Chicago School, specifically of Friedman and Stigler. Amadae, whose work is discussed in Section 6, emphasizes the role of the Rand Corporation and associated scholars. In politics, Margaret Thatcher and Ronald Reagan executed decisive turns to the right. Friedrich Hayek and Karl Popper are often mentioned as early and influential critics of leftist ideology.
A pillar of (post-Keynes) Keynesianism was the Phillips curve, an alleged negative association between the inflation rate and the rate of unemployment. Friedman challenged the Phillips curve in an influential presidential address to the American economics association in 1967. He argued that once the population had become used to a given rate of inflation they would tend to extrapolate it into the future. Prices and wages would accordingly rise in anticipation so that there would be no positive stimulus to real demand and hence to employment. He advanced the concept of a natural rate of unemployment to which the economy would tend. Leeson points out that a number of prominent economists had provided similar arguments, but without much impact.
But these (mostly scattered) judgments were not packaged in such a way as to convince the economics profession of the un-wisdom of believing in the long-run inverse trade-off. Only Friedman it seems was able to accomplish that. (p. 87).
Was it packaging or timing? Most likely both.

The Rand Corporation and the Cold War
This section is in large measure based on Amadae (2003). She describes efforts at the Rand Corporation in the post World War Two era to develop a counter ideology to communism and other forms of collectivism. These efforts, later supported by the Ford Foundation, transformed political science and much of the other social sciences as well. This is described in Part I of her book. She refers to the entire movement that resulted for these efforts as the 'rational choice movement', or as in the title of her book, 'rational choice liberalism'. The entire movement is divided into three parts, corresponding to newly established academic disciplines; each with its canonical text. These are: Social choice with the canonical text Kenneth Arrow (1951), Social Choice and Individual Values. Public choice with the canonical text James M. Buchanan and Gordon Tullock (1962), The Calculus of Consent. Positive political theory with the canonical text William H. Riker (1963), The Theory of Political Coalitions. Further canonical texts mentioned by Amadae are: John von Neumann and Oskar Morgenstern (1944), The Theory of Games and Economic Behavior;Anthony Downs (1957), An Economic Theory of Democracy; Mancur Olson (1965), The Logic of Collective Action. The canonical texts, with the exception of von Neumann and Morgenstern, all deal with aspects of social choice, however, the origin of the approach is in decision theory dealing with individual rational decision making under uncertainty and in the case of game theory also under strategic interaction. Under this aspect I would include Luce and Raiffa (1957), Games and Decisions, with the canonical texts.
In this section I will concentrate on the genesis, in the Cold War context, of the rational choice approach at the Rand Corporation, the Pentagon and the Ford Foundation. Arrow's work on social choice, specifically his 'impossibility theorem' will be the subject of the following section. Both topics allow an exemplary analysis of the role of ideology. Furthermore, for the subsequent developments in the political and social sciences none can approach the fame and influence attached to Arrow's work. First I will cite Amadae's assessment of the importance of the rational choice movement: The history of the development and establishment of policy science as an institutionalized and disciplinary norm has all the drama of a Hollywood screenplay involving the missile gap, the Sputniks, John F. Kennedy's presidential election, the overhaul of decision making procedures throughout Robert S. McNamara's Department of Defense, and the introduction of these same policy tools into domestic politics in Lyndon B. Johnson's Great Society program. This set of events gathered a momentum of its own but had crucial intersections with rational choice theory's development in the academy. Two vignettes illustrate the extent of these interconnections. Thomas Schelling, one of the key figures in establishing rational choice theory as mainstay approach to international relations, was one of RAND's alumni who formed McNamara's team of defense analysts in the Pentagon. In a pivotal series of U.S. Senate hearings, Schelling testified on the behalf of the administration about the development and use of rational policy tools throughout the U.S. Department of Defense in 1968. Subsequently he and other RAND alumni, including Howard Raiffa, helped to establish rational choice theory as part of the mainstream American intellectual endowment by virtue of their prominent academic posts at Harvard University's professional school of business.
It is no exaggeration to say that virtually all the roads to rational choice theory lead from RAND. (p. 10).

Rand and the Pentagon
The origins of the Rand Corporation reach back to rather humble applications of what today would be called operations research during World War II. After the war, in the context of the developing Cold War, Rand developed a much more ambitious vision: to make all aspects of the defense effort 'scientific', which in essence meant to make the required decisions subject to mathematical, mainly decision theoretic, analysis. This project was to have two grand subdivisions: a. The use of the mathematical theory of games, developed by von Neumann and Morgenstern, in order to gain an edge relative to the Soviet Union by means of an intellectually superior approach to the formulation of strategy. b. The cost/benefit analysis of military procurement decisions. Amadae discusses these in considerable detail. I discuss here only the two most important instances of the Rand approach. Their importance derives from the fact that they resulted in the transfer of vast sums to the defense effort and in the centralization of decision making away from the military to the systems analysts at the Pentagon.
The first episode concerns the famous (better infamous) 'missile gap'. Game theory was used to scare the public and the political elites into allocating unprecedented sums to the defense effort. These funds continued to flow even after the government had officially conceded the non-existence of the 'gap'. I cite Amadae's introduction to this subject (Amadae 2003): Despite this complete reversal of official U.S. government position and the resulting public outcry, every policy idea based on the belief in the (nonexistent) Soviet missile advantage was implemented over the next seven years. This chapter explores the processes of knowledge production and political interaction that manufactured the 'gap' in the public mind and public record, initiated a sea change in American national security policy whose rationale originated in the missile gap, and empowered a new policy elite whose authority was grounded in the supposed objectivity of rational policy analysis. (p. 29).
The next episode involves the planning-programming-budgeting-system (PPBS) developed at the RAND corporation and installed at the Pentagon by Secretary McNamara. As described by Amadae, the proponents of PPBS …were agreed on inverting the policy process: instead of fiscal appropriations being handed down from Congress to meet operational needs, defense planners would articulate their needs using presumptively objective and thus incontrovertible costeffectiveness studies. Instead of Congress' determining how much national security the nation could afford, national defense imperatives should determine defense allocations on the principle that ´there [be] no presumption that the defense budget is now, or should be, near any immovable upper limit`.
The authority for this bold restructuring came from the supposed scientific rigor promised by such a budgetary process. (p. 63).
The breakthrough for the establishment of PPBS at the pentagon came at the beginning of McNamara's tenure with a huge procurement project to supply the air force and the navy with new fighter planes. The military desired two different planes designed for the differing mission requirements of the two services.
McNamara prevailed over the strenuous objections of the generals and the contract for the combined plane was awarded to General Dynamics in Fort Worth. Extensive congressional hearings on this decision where subsequently held. They revealed that the sophisticated 'scientific' analysis on which the decision allegedly rested had never been performed. Amadae writes: Ultimately the investigating committee was led to conclude, as McNamara himself admitted, that the TFX decision rested on 'rough judgments'. In 1963 Congressmen could only suspect what history would confirm. The Navy later canceled its contract with General Dynamics in 1968, and the Air Force ultimately obtained only six hundred of the originally contracted twenty-four hundred planes, at a cost of $22 million each instead of the initially proposed $2.8 million. (p. 67).
It is not far from the truth that the claims of scientific objectivity that were used to propagate the Rand methodologies of game theory and PPBS were a hoax, used to provide a respectable cloak to what were actually naked power grabs.
Game theory was ultimately abandoned as a tool for the formulation of American foreign and military policy. Not so PPBS. In August of 1965, President Lyndon B. Johnson mandated that PPBS became standard operating procedure for all federal agencies.

Rand and the Neocons: A Postscript
The preceding deals with the military role of the Rand Corporation to the extent that it was discussed by Amadae. She has almost nothing on the Vietnam War, after all the most dramatic and consequential event in which McNamara was involved. Moreover, the Vietnam War was shaped by the same patterns of thought that characterized McNamara's actions at the Pentagon from the beginning: a faith in advanced military technology coupled with contempt for the views of the military services. Furthermore, there are similarities between the Vietnam War and the current Iraq War. The attitudes just attributed to McNamara are the same as those characterizing the secretary of defense managing the Iraq War, Donald Rumsfeld. The consequence in each case has been a disastrously mismanaged war, resulting ultimately in the ouster of the secretary of defense and the search for an exit strategy.
I did some more research in the internet and I chanced on the article by Husain (2003) which traces the role of key individual associated with Rand up to and including the current presidency of George W. Bush. The following owes much to that article from which the quotations are also taken.
Two influential Cold War strategists at Rand were Albert Wohlstetter and Andrew Marshall. Wohlstetter's principal contribution was the doctrine of nuclear deterrence that became the key concept of America's defense strategy vis-à-vis the Soviet Union.
The essence of the doctrine was that effective deterrence required an assured counterstrike capability after a first strike carried out by the Soviet Union.
…looking ahead to anticipate future threats was a central part of Wohlstetter's methodology. In an essay published in 1959 Wohlstetter argued for significantly boosting of America's nuclear umbrella, in number and in its capacity to resist a first strike, in anticipation of Soviet moves to deploy more missiles with greater accuracies in the next ten years. To the imperatives of nuclear deterrence, he went on to recommend a large conventional force capable of fighting a general war against the USSR alongside a full blown nuclear conflict… Wohlstetter also projected that in the 1960s, the American deterrent would have to deter not just the USSR, but China as well. Grimly pointing out that the Russians suffered 20 million dead in World War II and still emerged as a superpower; he wondered how much damage they would consider as 'unacceptable'…All this meant that there would be no rest for the military in the 1960s according to Wohlstetter (Husian 2003).
Wohlstetter left Rand in 1962 to enter academia and in 1964 took a professorship in the political science department at the University of Chicago.
It was here that he met a bright young student doing his dissertation in the Political Science department. His name was Paul Wolfowitz. Wolfowitz was drawn to Wohlstetter's intellect and temperament and began working under his supervision to carry his ideas further.
Wolfowitz became interested in strategic questions in the nuclear age and picked up where Wohlstetter left off. Where Wohlstetter had warned of preparing for a rearmed USSR and a nuclear China, Wolfowitz considered the third dimension along which nuclear strategy would evolve in the future: nuclear proliferation. (Husian 2003).
Wolfowitz took a job at the Pentagon and began to make alarmist projections of future Soviet nuclear capabilities. These were initially ignored. when George Bush Sr. became CIA director in 1976 he requested an assessment of Soviet nuclear capabilities. There resulted, with the active participation of Wolfowitz, the so called 'Team B reports'.
The contents of the Team B reports are alarming for the threats they saw in the Soviet nuclear build up, and startling for the methodologies they used. They projected that by 1984, the USSR would deploy about 500 Backfire strategic bombers, whereas the real figure in 1984 was less the half that. They claimed that the Soviet Union was working on an anti acoustic submarine, and failing to find any evidence of one, stated quite seriously that one may already be deployed since it appears to have evaded detection! … The Reagan defense buildup of the 1980s and the evil empire rhetoric of the second cold war built on the work of Team B members. The result was the largest defense budget increases in peacetime history. (Husian 2003).
I turn to the role of the other influential Rand strategist Andrew Marshall. He is described as a secretive individual about whose 23 years at Rand little is known.
His best known protégé is probably Donald Rumsfeld, whose association with Marshall is decades old, dating from Rumsfeld's early days in the Pentagon. Rumsfeld became an early proponent of ballistic missile defense, a Marshall idea and belonged to that clique of hawkish policy makers who were opposed to Kissinger's ideas of detente and engagement with China. (Husian 2003).
Marshall developed the concept of a 'Revolution in Military Affairs' (RMA) involving novel technologies, novel strategies and novel management methods.
When the Bush administration came to power, the RMA was put into practice. Rumsfeld was made the Secretary of Defense…and empowered Andrew Marshall to conduct a sweeping review of the military and make recommendations to make the military into a 21st century fighting force. The RMA was no longer part of the lunatic fringe from where it had originated. Its adherents were now in control, and were going to make their presence felt . (Husian 2003).
There exists a complex web consisting of the various threads of the Neoconservative movement and of movements that have allied themselves with it. Given the enormous effect that the combined movement has had on the politics of the United States, and hence of the world, it is surprising how little scholarly attention has been focused on it. Moreover, information on one aspect of the phenomenon is usually not connected with other aspects. For example, the Wikipedia article on 'Neoconservatism' http://en.wikipedia.org/wiki/Neoconservatism is focused on publicists, most prominently Irving Kristol, who is credited with inventing the term. A list of Neoconservative think-tanks does not include Rand.
The Neoconservative ideology rose to the apex of power under the administration of George W. Bush. A large part of the credit for bringing this about is usually given to the political operative and later presidential advisor Carl Rove. The various institutions and movements that Rove gathered under the umbrella of the Republican Party include. the military industrial complex; other highly politicized business sectors, particularly energy, construction and finance; Christian Evangelicals; large segments of the massmedia that either espoused the neoconservative cause, or else muted their criticism; a variety of neoconservative think-tanks.
It is no exaggeration to say that the ultimate goal of this alliance is the creation of an American empire that dominates the world. This goal has been rarely publicized. The fullest discussion of it has been given by Chalmers Johnson (2004). Much of the book is devoted to the extensive system of American military bases that gird the globe. The following quote focuses on American plans to dominate space.
Even prior to the Afghan war, a group of right-wing 'defense intellectuals' had started to advocate a comprehensive new strategy for global domination. Many had served in earlier Republican administrations and most of them were again given high appointive positions when George W. Bush became president. They focused on plans for the next decade or two in much the same way that Captain Alfred T. Mahan of the navy, Senator Henry Cabot Lodge, and Assistant Secretary of the Navy Theodore Roosevelt had emphasized sea power, Pacific bases, and a two-ocean navy at the end of the nineteenth century. Rarely taking the public into their confidence, the members of this new clique were masters of media manipulation, something they acknowledged they had 'learned' as a result of bitter experience during the Vietnam www.economics-ejournal.org War. The terrorist incidents of 2001, much like the sinking of the battleship Maine m 1898, gave a tremendous boost to their private agenda. It mobilized popular sentiment and patriotism behind military initiatives that might otherwise have elicited serious mainstream doubts and protests.
'The determination to militarize outer space and dominate the globe from orbiting battle stations armed with an array of weapons includes high-energy lasers that could be directed toward any target on earth or against other nations' satellites. The Space Command's policy statement, 'Vision for 2020,' argues that 'the globalization of the world economy, will continue, with a widening gulf between 'haves' and 'havenots,'' and that the Pentagon's mission is therefore to 'dominate the space dimension of military operations to protect U.S. interests and investments' in an increasingly dangerous and implicitly anti-American world. One crucial goal of policy should be 'denying other countries access to space.' (pp. 80-81).
The principal actors advancing this agenda, as named by Johnson, are se same ones mentioned earlier in this section.

The Origins
A crucial role in the institutionalization of rational choice first in the government and later in academia was played by H. Rowan Gaither Jr. He had managed the transformation of RAND from a division of the Air Force to a private foundation. Seeking philanthropic aid for RAND he had met Henry Ford II. This had momentous consequences as described by Amadae (2003): Meanwhile, Gaither had so impressed Ford at their meeting that the auto manufacturer asked him also to produce a policy statement for the Ford Foundation…The report describes a society managed by an educated elite outside the public arena and suggests that it is the duty of philanthropies to support this elite.
The report is unequivocal in suggesting that philanthropies and their beneficiaries manifest an objectivity that best entitles them to provide leadership in a democratic society. (pp. 35-36).
A third institution founded by Gaither with a grant from the Ford Foundation was the Center for the Advanced Study of the Behavioral Sciences (CASBS) at Stanford University. Amadae describes the Center's purpose and functioning: H. Rowan Gaither Jr.'s steadfast support of the social sciences as tools for social management and rational defense management had a two-fold impact on the emergence of the rational choice framework. Both at RAND and through the Ford Foundation's establishment of the CASBS, theorists had the freedom to generate a body of ideas. Furthermore, the empowerment of the defense rationalists helped to gain currency for their ideas of rational and objective policy analysis. As these theorists found their way back to academia after stints of service in Washington, they returned with the prestige helpful to making their idea set part of the mainstream intellectual endowment of American society. (p. 79).
These passages illustrate a profound contradiction of the rational choice movement. Ostensively it aims at providing an intellectual bulwark defending capitalist democracy against the temptations of totalitarian ideologies. But the architects of this bulwark were www.economics-ejournal.org in an elitist rather than a democratic tradition. The idea of social control by scientific elites hardly differs from the communist ideology that they so arduously combated. The anti-democratic tendency of the Western intellectual tradition, going back to Plato, was a theme of Popper's The Open Society and its Enemies. B. F. Skinner, the American psychologist and leader of the behaviorist school, advocated social control by a scientific elite, as illustrated by his utopian novel Walden Two.
Regarding the anti-democratic tendencies of elites generally, Amadae has the following comment: The history that emerges is relevant to broader discussions of the tension between the ideal of liberal democracy and the tendency of elites to develop means to control societal decision-making processes. Since its inception as a social form predating the French and American Revolutions, and going back to at least the British civil wars, the drama of democratization has in part been about conveying the appearance of inclusion while designing means to retain actual control over decisionmaking in the hands of a social elite. (p. 31).

Arrow and the Consequences
The subject matter of this section is one on which a huge, highly technical and mathematical literature exists. My own discussion will be entirely non-technical, it is however based on a rigorous paper (Hillinger, 2005b) to which the interested reader may turn.
Arrow's 'general possibility theorem', popularly known as his 'impossibility theorem', or as 'Arrow's paradox', along with its reception and interpretation, is surely one of the strangest episodes in the annals of science. I will try to show that it is this reception and interpretation that deserves to be called 'paradoxical', rather than the theorem itself.
In discussing Arrow's work I will first take at face value the interpretation given by him that has been accepted essentially unchanged by the huge subsequent literature. According to this standard interpretation, Arrow has proven that there is no way of going from a set of individual preferences over some alternatives to a social ordering, given that we require the aggregation to satisfy a few very reasonable conditions. Arrow's theorem is usually interpreted as applying to voting. Since voting is the most important formal method for reaching decisions in a democracy, the theorem is taken to imply the impossibility of democracy. Arrow has stressed that his theorem is even more general than that since it applies to any method for aggregating preferences. Specifically, the theorem applies also to market outcomes; these cannot be rationally justified as being particularly desirable.
There is only one other specific analytical result in economics, or for that matter in all of social science, that has achieved equivalent fame. It is the argument given by Adam Smith to show that a competitive market economy is efficient. The market achieves efficiency as if guided by an 'invisible hand'.
This brief analytical passage could easily have been overlooked amidst the many, discursive pages of the Wealth of Nations; happily it was not and became the foundation of modern economics. Though verbal, the argument matches exactly the most elegant mathematical derivation of what now is called 'the first theorem of welfare economics'; that a competitive economy is Pareto efficient. 31 A large literature dealing with the 'Arrow paradox' appeared, but it is fair to say, did not find a way around it. For example, it was shown that if preferences satisfy the condition of being single peaked, than the paradox becomes inoperative; however, the condition cannot generally be expected to hold. In the following I will discuss a series of quite different paradoxes involved in the reception and interpretation of Arrow's paradox.

First Paradox
The rational choice movement arose in the context of the Cold War with the purpose of countering totalitarian ideologies and providing the ideological foundation for capitalist democracy. For this purpose they felt that they had to both attack the foundations of totalitarian ideologies and provide an alternative. Both aims were accomplished most clearly by Arrow. As the foundation of totalitarianism he identified the idealistic philosophies of Rousseau, Kant, and Marx. Specifically, he attacked their idea of a 'general will', separate from and superior to individual preferences. 32 What is noteworthy in Arrow's discussion of these three philosophers is the manner in which his set-theoretic proof undermines their philosophical systems: he insists on a thorough-going individualism that is incompatible with any standard for collective social norms or self-legislation that may impinge on an individual's right to have any (transitive) set of desires, and he defines collective rationality in accordance with this priority granted to individual desire. Here Arrow's philosophical position clearly reflects the attempt to erect a basis for American economic and political liberalism that cannot be thwarted by authoritarianism. (p. 113-114).
Both Arrow and his interpreters, such as Amadae, seem to be blind to the obvious paradox that Arrow, on his own terms, proved the impossibility of capitalist democracy. Totalitarian systems are based on the belief that their leaders are already in the possession of the relevant truth; they see no need for the aggregation of individual preferences. If they allow voting at all, it is only to affirm the pronouncements and actions of the leader.

Second Paradox
The only formal result in all of social science that has attracted more attention than Arrow's theorem is 'the first theorem of welfare economics'; that a competitive economy is Pareto efficient. 33 The origin of this theorem is Adam Smith' argument regarding the 'invisible hand': But the annual revenue of every society is always precisely equal to the exchangeable value of the whole annual produce of its industry, or rather is precisely the same thing with that exchangeable value. As every individual, therefore, endeavors as much as he can both to employ his capital in the support of domestick industry, and so to direct that industry that its produce may be of the greatest value; every individual necessarily labours to render the annual revenue of the society as great as he can. He generally, indeed, neither intends to promote the publick interest, nor knows how much he is promoting it. By preferring the support of domestick to that of foreign industry, he intends only his own security; and by directing that industry in such a manner as its produce may be of the greatest value, he intends only his own gain, and he is in this, as in many other cases, led by an invisible hand to promote an end which was no part of his intention. (Smith, 1776(Smith, [1993, pp. 291-2).
There is an apparent contradiction between Smith's argument, in its modern form the first theorem of welfare economics, and Arrow's. For the general case of unrestricted individual preferences, Arrow was unwilling to grant any rationality to either voting or the market. However, efficiency, which is what the first theorem of welfare economics attests competitive markets, is certainly a strong rationality property. The welfare theorem would otherwise not have been assigned the importance that it has. The contradiction is not at a logical level, since Arrow has a broader concept of rationality that requires all possible states to be included in a ranking; The rationality of the market is more restrictive since possible initial distributions of wealth are not evaluated. For a given distribution, the market is rational in producing the 'best' outcome in the Pareto sense. This is more than Arrow is willing to grant.

Third Paradox
One would think that given Arrow's proof of the impossibility of rationally aggregating preferences that the subject would cease to receive the attention of scholars. On the occasion of Arrow's receiving the Nobel Prize, Paul Samuelson (1972) had written: What Kenneth Arrow proved once and for all is that there cannot possibly be found such an idea! voting scheme: The search of the great minds of recorded history for the perfect democracy; it turns out, is the search for a chimera, for a logical selfcontradiction. (p.71).
Why would anyone wish to chase the chimera that had defeated the great minds of the past? After all, given the proof that a square with the same area as a given circle cannot be constructed using compass and ruler alone, this problem ceased to occupy the attention of mathematicians. Similarly, given the impossibility of constructing a perpetuum mobile, engineers do not try to construct machines that run without energy. In the case of collective choice the experience has been the opposite; before Arrow the subject barely existed, after Arrow the literature veritably explodes. Dennis Mueller's text Public Choice III has 768 pages with 64 large and closely printed pages of references, almost all post-Arrow.
The theory of collective choice has undeniably produced much printed paper: What else has it produced? I am tempted to say very little. The mathematical theory of voting began about 300 years ago with the work of Borda and Condorcet who realized the defects of plurality voting and began looking for a superior alternative. Plurality voting is still universally used and theorists do not agree on the superiority of any other method. McLean and Urken (1995), in their review of social choice theory reach the following conclusion: ... modem social choice theorists develop very general models and are consequently reluctant to give advice. To the nonspecialist, the obvious question thrown up by social choice is, What is the best electoral system? This is an urgent practical question all over the world, never more so than since the collapse of communism began in 1989. But whereas theoretical molecular biology has started to play a large role in curing bodily disease, theoretical social choice has played almost no role in curing constitutional disease. Social choice theorists have usually regarded 'the earnest efforts of electoral reformers…with the same kind of amused contempt as mathematicians in the past reserved for claims by amateurs to have succeeded in squaring the circle' (Barry 1986, 1). Barry continues: A few years ago, at a conference on the theory of democracy, a group of five eminent social choice theorists were trying to decide which of several restaurants to dine at. Since each knew the preferences of the others and could immediately compute the outcome to be expected from any proposed procedure, it was impossible to find any agreement on a method of voting. (The impasse was in the end resolved by one of their number setting off in the direction of the restaurant he favoured; after he had gone about thirty yards the others fell in behind). (Barry 1986, 1-2) Since 1989, Czechoslovakia, Hungary, Poland, and the three ex-Soviet Baltic republics have all written new voting laws… Other new democracies, including those in the former Soviet Union, will shortly have to do the same. Even at the peak of the Enlightenment, only three countries (the United States in 1787; France in 1789, 1791, and 1793; and Poland in 1791) wrote constitutions containing voting laws. In two years we have had at least six; but as far as we know, not one of them consulted any social choice theorists. (In 1990 members of the Mongolian legislature asked the president of the Public Choice Society for advice on a new constitution; but we have reason to suspect that the advice he gave them may have been along the lines lampooned by Barry.) It adds up to a lamentable failure of social science. Too much research in social choice has been conducted in the spirit of the pure mathematician's prayer: 'May it never be of any use to anybody!' These negative evaluations are what Arrow's theorem leads one to expect. The motivation for producing this literature does appear paradoxical.

Fourth Paradox
The proof of the impossibility of democracy appears to be one of the most momentous discoveries of all time, its potential impact on the human race perhaps greater than the heliocentric hypothesis of Copernicus and Galileo, Newton's theory of universal gravitation, or Einstein's relativity theory. Children in school learn these laws of the natural world, but they and their elders remain ignorant of Arrow's law concerning the social world. Countries that have traditionally thought of themselves as democracies continue to think this way. The United States keeps pursuing its agenda of spreading democracy; sometimes with apparent success, as in post-WWII Europe, at other times with disastrous consequences as presently in Iraq. No one associates these developments with Arrow's theorem.

Fifth Paradox
Barely after the appearance of Arrow's impossibility theorem, Fleming (1952) and Harsanyi (1953Harsanyi ( , 1955 published what may be called possibility theorems on collective choice in the form of plausible and non-contradictory sets of axioms. Arrow assumed that preferences are to be expressed as weak orderings of the form meaning that a is not worse than b and b is not worse than c. The axioms proposed by Fleming and Harsanyi imply a cardinal representation of the form a b c ≥ ≥ ( ) ( ) ( ) 10, 2, 1 u a u b u c = = = which would suggest in the present example that the gain in going from b to a is greater than the gain in going from c to b. The results of these three authors show that there is an open door to social choice, based on a cardinal representation, and another one, based on the ordinal representation, that is closed. One would have expected choice theorists to pass through the open door; they chose instead to bang their heads against the closed one. In my paper on voting (Hillinger 2005b) I went through the open door.

On the Possibility of the Scientific Method in Economics:
The Case of the Quantity Theory

Background and History
For the purpose of this section I define science as a social enterprise that is able to secure agreement on aspects of reality on the basis of publicly available and publicly scrutinized evidence. This definition accords with that given by Ravetz, cited earlier: an immature science is one that has not agreed on criteria for determining factual truth. This definition corresponds to the social system of science that has been attained in the natural sciences and is the basis of the prestige that they enjoy; it is also the motivation for imitation-science that attempts to share this prestige. I do not believe that science by this narrow definition exhaust meaningful thought either in the social realm or elsewhere. This will be the subject of the next section. Nevertheless, science by the narrow definition is immensely powerful and also exerts its influence on other modes of thought. Also, immense efforts at being scientific, or at least appearing so, have been made in economics and other social 'sciences'; with little apparent success. There also exists a substantial literature on the methodology of economics that is concerned with this issue, but has been unable to reach a consensual verdict. For all these reasons, the question of the possibility of science in the social realm remains open and important.
I deal with the issue pragmatically by pointing out that there have been at least two fields that fully meet the narrow criterion for science. They are the quantity theory of money and the newer research on happiness. Examination of these fields also reveals the difficulties that scientific findings in economic and social thought have in spreading beyond a narrow circle of experts.
The first formulation of a quantitative macroeconomic relationship is due to Copernicus, better known for his advocacy of the heliocentric theory. In a statement to the Polish king, based on an address to the Prussian diet in 1522, he writes: 'Money usually depreciates when it becomes too abundant.' Regarding this earliest rough formulation of the quantity theory of money, the economic historian Spiegel (1971) wrote: Copernicus's tract was not published until the nineteenth century and may not have had much influence on the thought of his contemporaries. In any event, his discovery, whatever its range and effect may have been, is especially remarkable because chronologically it antedates the large-scale movement of precious metals from America to Europe. By the power of reasoning and by the ability to invent fruitful hypotheses, a great mind may discover relations that ordinary people can recognize only if driven by the stimulus of observation. (p. 88). 34 The subsequent histories of the heliocentric hypothesis and the quantity theory of money are highly instructive in illustrating the difference with regard to the establishment of factual truth between the natural and the social sciences. The heliocentric hypothesis was long forbidden by the Church, but as the prestige of science increased and that of the Church declined, the hypothesis became accepted by all educated people including the hierarchy of the Church. Once the hypothesis was broadly accepted, it was never again seriously brought into question.
The quantity theory of money had a very different fate. It did not face the same kind of opposition as the heliocentric hypothesis, instead it was for some centuries simply ignored. The incorporation of the hypothesis into classical economics was based on abstract economic reasoning rather than on systematic empirical evidence. During periods of high inflation, people tended to become aware of the relationship between money and prices, but when the memory of inflation faded so did this understanding. Keynesian economics emphasized the various 'pathologies' that might prevent a change in the quantity of money from having any effect.

Modern Work
Serious empirical work on the quantity theory began at Chicago as part of the monetarist attack on Keynesianism. 35 This was around 1950, some 400 years after the first formulation of the hypothesis by Copernicus! The Chicago studies provided strong evidence in favour of the quantity theory.
Largely inspired by Chicago, empirical studies of major inflations, which were at the same time test of the quantity theory, attracted a considerable number of scholars in following decades. Capie (1991) assembled 21 such studies. Together they provide the most impressive validation of the quantity theory that is available. In the following paragraph I will try to justify this statement.
I have defined 'science' as 'a social enterprise that is able to secure agreement on aspects of reality on the basis of publicly available and publicly scrutinized evidence'. Crucially important in this context is the greatest possible diversity of both investigators and evidence. If a hypothesis is confirmed only by one set of cooperating investigators working on a narrowly defined set of data, then the confidence in the validity of the hypothesis will not be great. The criteria of diversity for investigators and evidence are extremely well met by the Capie collection. The investigators come from a number of different countries and their articles were published over a span of four decades -the earliest being Bresciani-Turoni in 1937 and the last Capie in 1986. The inflations studied are highly diverse in time and location. Following the introductory Part I, Part II has 10 articles on inflations before 1900; Part III has 6 articles on inflations in the 1920s; Part IV has 5 articles on inflations in the 1940s. The most impressive example for the diversity of evidence is the article by Francis T. Lui 'Cagan's Hypothesis and the First Nationwide Inflation of Paper Money in World History'. It deals with the introduction of paper money and subsequent inflation in 12 th Century China.
Two more publications are relevant. One is Fischer, Sahay and Végh (2002). They do a cross-sectional study of the inflationary experience of 133 countries since 1957. From their conclusions I quote the statement that bears on the quantity theory: 'As expected, the long-run (cross-section) relationship between money growth and inflation is very strong'.
The final paper to be discussed, Dwyer and Hafer (1988), is my favourite empirical investigation of the quantity theory. It illustrates a common sense approach to doing empirical science, untouched by self-defeating econometric fashions. I quote their introductory paragraph because it illustrates the inability of the economic profession to come to an agreement regarding the evidence on an empirical phenomenon.
Many economists recently have been claiming that money has little or no effect on inflation and economic activity. For example, Lyle E. Gramley, past governor of the Federal Reserve Board, has been quoted as saying 'the relationship between growth of the economy and the growth of the money supply is just no longer there.' Meanwhile, even a noted monetarist such as Beryl W. Sprinkel, the current chairman of the Council of Economic Advisers, says: 'It's a problem. Nobody knows where we are going.' These recent statements are hardly novel, nor have they changed all that much over the years. In 1971, Federal Reserve Board Governor Andrew F. Brimmer noted that it has 'not [been] demonstrated convincingly that the relationship between the money supply and economic activity is especially close.' Two decades and innumerable empirical studies later, their statement applies just as well to current macroeconomics.
Dwyer and Hafer discuss the quantity theory in terms of the two equations where the dots indicate growth rates, M is the money stock, k a proportionality constant, Y nominal national income, p the price level, y real national income. For their investigation they use data on 62 countries for the period 1979-1984. Since the quantity theory has been generally interpreted as an equilibrium relationship they use 5 year averages for the main part of their study and do a cross-sectional analysis with these. In such a cross-section no systematic change I k can be expected.
The coefficient of M & is highly significant and almost exactly unity, as the quantity theory predicts for a constant k. The estimated constant term is of low significance and a graphic plot shows that a line through the origin fits the data very well.
The variation in inflation rates between the countries is equally well explained by the variation in the growth of the money stock: Finally, there is no long-run relationship between the growth rates of the money stock and real income: All in all an impressive validation of the quantity theory, using only elementary techniques.

Concluding Remarks on Evidence and Ideology in Relation to the Quantity Theory
My conclusion from this survey is that the validity of the quantity theory has been established beyond reasonable doubt, using the most rigid scientific criteria. What has been the consequence? My conjecture is that few contemporary economists are aware of the cited evidence and few would express any confidence in the quantity theory. I have no direct evidence for this statement, but I do have some indirect evidence.
The most recent advanced macroeconomics text in my library is Sørensen and Whitta-Jacobsen (2005). The subtitle of the book 'Growth and Business Cycles' already indicates a lack of interest in the money/inflation nexus. There is no explicit mention of the quantity theory, but the relationship does appear in their Figure 3.3: Money growth and inflation, Denmark and the US, 1870-2000. Strangely, this is in Chapter 3: 'Capital Accumulation and Growth' and in a section titled 'Money?'. The principal argument of this section is that money has no effect on output in the long run. The purpose of the figure is apparently to demonstrate that the effect of money growth is exhausted by inflation. 36 The section concludes: 'This should explain why you will hear no more of money in Book One of this text, while monetary policy will be at the heart of the analysis throughout Book Two'. Book Two is titled: The Short Run: Economic Fluctuations, Short-run Unemployment and Stabilization Policy. Much of it is devoted to the Phillips curve and related issues. The student of this text will never have heard of the quantity theory and the empirical relationship that it denotes will hardly have made an impression on him.
Further evidence that the quantity theory has faded from economists' awareness is furnished by the monetary policies pursued by central banks. They are universally pursuing active discretionary policies. Generally, the key consideration is some form of inflation targeting; monetary policy is related to an inflation forecast. Secondarily, an estimate of the stage of the business cycle also plays a role. For most industrialized countries, the growth rate of the money stock has been well above that suggested by the quantity theory as appropriate to secure the desired low inflation rate in the long run. Currently inflation is accelerating world wide and the central banks are stepping on the brake by raising their discount rates. Their focus on inflation targeting has evidently led them away from being guided by the quantity theory.
The natural sciences are characterized by the steady accretion of knowledge. Knowledge, once firmly established, may subsequently be further refined; rarely is it refuted or simply forgotten. Why is this progress not visible in the social sciences? In general terms I will deal with this question in the final section. Here I examine it in relation to the quantity theory.
The history of the quantity theory, as well as of macroeconomics generally, is one of the increasing dominance of ideology over empirical science. The early work done at Chicago to rehabilitate the quantity theory was pure empirical science. The motivation for this work was however partly ideological: to discredit Keynesianism with its interventionist bias and thus to promote the neoliberal ideology. To advance this agenda further, Chicago monetarism went beyond the quantity theory and asserted that shortrun fluctuations of output had their origin in shocks originating in the monetary sector. This aspect of monetarism steered empirical work away from the quantity theory towards the investigation of the short-run effects of monetary changes on output.
A related development was the discrediting of structural macro-econometric modeling, at least in academic economics. The large-scale macro-econometric models moved to governmental agencies or private research institutes. For this there were essentially three causes. The major one was the unsatisfactory nature of the large-scale models, both in their conceptualization and in their performance. A second element was the rising interest in time series analysis, strongly influenced by the work of C. W. J. Granger. A third factor was the aforementioned development of monetarism; the assumed unicausal influence from money to the macro-economy called for time series analysis rather than multi-equation structural modeling.
Granger's methodology emphasized statistical testing which was in line with the general tendency of econometrics. An important role in the present context was played by the 'causality test' developed by Granger. There was an explosion of papers testing for causality from money on macroeconomic variables. Ultimately, this research failed to identify any reliable relationships involving money and served to undermine the belief in the existence of any such relationship.
Another major reason for the eclipse of the quantity theory is what may be called the professional ideology of central banks. The implication of the quantity theory for monetary policy, strongly advocated by Friedman, is to let the money supply grow at a constant rate. This would eliminate discretionary monetary policy and drastically reduce the importance of central banks. I read a newspaper interview with Milton Friedman in which he was asked about monetary targeting, he replied that it was a make work program for central bank professional staffs.
This history is an interesting example of the interplay of reality and ideology. Friedman's proposal of a constant monetary growth rate was based both on his scientific work on the quantity theory of money and his ideological preference for governmental non-intervention. It was defeated by the professional, self-interest motivated, ideology of the central banks.

Social Science and the Ideologies of the Left
The preceding sections have concentrated on ideologies of the right and their impact on economics and on political science in so far as it has been absorbed into the economists' style of formal reasoning. This section deals with the ideologies of the left that have impacted sociology, psychology and adjacent fields, as well as the humanities. This is a huge field in regard to many aspects of which I feel unqualified. I therefore limit myself to a few selected topics where I feel that I have something to say that fits in with the subject of this paper. Fortunately, there exists a highly readable book, written by a sociologist, which covers large parts of this field. It is Stanislav Andreski (1972), Social Sciences as Sorcery.
The initial impetus for writing this section came from the suggestion of a referee that I consult the writings of Steve Fuller. He is a sociologist and philosopher of science and politically on the left. A good introduction to his work is his homepage which contains both some of his articles as well as reviews of some of his books. What interested me most is his discussion of Thomas Kuhn's theory of scientific revolutions, which I www.economics-ejournal.org discuss in the following section. Fuller's own philosophy of science may be said to be of the general leftist 'anything goes' variety. It is the subject of Section 9.2. In Section 9.3 I summarize the interaction of ideology and social science since Adam Smith. Thomas Kuhn's (1962) The Structure of Scientific Revolutions, often referred to simply as 'Structure' is one of the most influential books of the second half of the Twentieth Century. I make no attempt at summarizing it here. 38 Instead I focus on a few central aspects, particularly the ideological background described by Fuller (2000).

Kuhn's Theory of Scientific Revolutions 37
The story begins with James B. Conant, president of Harvard University from 1933 to 1953 and a key figure in the military-industrial-academic complex. A chemist by training, he played important roles in the development of the atomic bomb as well as the founding of the National Science Foundation. He was a believer in an elitist role of science and was worried that public support of science might be insufficient for the role that he felt American science had to play in the struggle with the Soviet Union. He became committed to the agenda of developing a science education for the general public that would stress, not the critical scrutiny of science on the part of the public, but the need to support the scientific elites.
Kuhn was a graduate student in physics when recruited by Conant to teach in the general education program. Conant interested Kuhn in the history of science and was closely involved with supervising his dissertation which ultimately became Structure. Kuhn has credited Conant with originating the principal ideas in Structure. The book paints an elitist picture of science. Normal science works within and elaborates world view or paradigm set up by one of the greatest scientists of the order of Newton or Einstein. With time, unsolved problems within a paradigm accumulate and ultimately a new genius emerges who proposes a new paradigm that is incommensurate with the old. Since not even an ordinary scientist can question the paradigm within which he works, it is clear that the ordinary public would be even less able to do this.
Kuhn held his theory to be applicable to all of science, with the focus being on the natural sciences. However, the impact of Structure has been almost entirely within the social sciences and humanities. Typical of the view of natural scientists is the remark made to me by a physicist that progress in natural science is measured by the steady accumulation of significant digits in the quantitative relations that describe the laws of nature. Laws of physics and geometry discovered by Archimedes, Euclid and Pythagoras in antiquity are as valid today as then.
Conant and Kuhn initiated the field of science studies, which Fuller also cultivates. The evolution of this field illustrates the fact that projects started with an ideological agenda may evolve quite differently then anticipated by their founders. Science studies found a home in the social sciences and humanities, associated itself with the political left. As Fuller himself has put it: '…science studies was connected with other broadly academic leftist movements, such as feminism, postmodernism, and multiculturalism.' Rather then instilling respect, if not awe, the movement became highly critical of elitist science.

Leftist Philosophy of Science
The social movement often referred to as the Sixty-eight Generation, after the student revolts of that year, alternatively also as The New Left made its mark throughout the social sciences and humanities. Many members of the movement made careers in academia, or became active in the media where they could influence public opinion and the general culture. The dominant trait of the movement was anti-authoritarianism and further traits such as multi-culturalism and relativism served to buttress their central aim. A term that has also been used to refer to this set of traits is postmodernism.
The most influential philosopher of science is Paul Feyerabend. most importantly through his 1975 book Against Method. He argued that there is no single valid philosophical or scientific method and that all forms of expression, including even mythology and astrology are equally valid. A noteworthy application of this doctrine occurred at a lecture of the distinguished sociologist and co-founder of the Frankfurter Schule Theodore W. Adorno. The female students in the auditorium formed a semicircle around the eloquent professor and silently bared their breasts.
The most prominent contemporary postmodernist philosopher and sociologist of science is Steve Fuller. In several books and articles beginning with his 1988 Social Epistemology he has pleaded for a radical democratization of science. I have sympathy for the view that in a democratic society the direction of science should ultimately also be determined democratically. How this is to be affected is an important and difficult issue. The contemporary drive to require schools to teach creationism as an equally valid alternative to evolution illustrates the dangers inherent in a naïve implementation of this program.
My view of postmodernism is that it is untenable for both logical and pragmatic reasons. The logical problem is that if all position s are relative, so is relativism itself and so it cannot be regarded as objective truth. The pragmatic reason is that that I feel it is impossible to have a meaningful argument about anything without the implicit assumption of the existence of an objective truth.

10
Novelty, Replication and the Growth of Knowledge

Novelty Without Validation
The material surveyed in this paper allows only one conclusion: social science has been increasingly dominated by varieties of both political ideologies and forms of scientism; contact with reality was increasingly lost; the most basic problems of society are not being identified, let alone adequately analyzed and workable solutions proposed. I do not mean to imply that a great deal of valuable material is not contained in the vast flood of social science publications. The problem is rather that the social sciences have not developed a mechanism for securing agreement on what is true, or relevant. Referring again to Ravetz' criterion of an immature science, the social sciences are unable to agree on what is factually the case. Among the criteria that the social sciences have adopted from the natural sciences that of novelty is prominent. It is reflected in the torrential and ever expanding flow of social science publications. In the absence of criteria of relevance, this flood of publications produces alternating fashions instead of a cumulative growth of knowledge.
The state of the social sciences can best be understood by looking at examples were agreement on factual truth, or on relevance, was actually secured. It is useful to distinguish two different kinds of knowledge that differ in how specific they are and consequently in the ease of verification.

Quantitative Knowledge and Replication
The type of knowledge that is typically associated with the natural sciences concerns quantitative empirical laws that are established and verified in controlled experiments. In this case it is evident how replication of the experiments leads to agreement. It is often claimed that the relative backwardness of the social sciences is due to their inability to conduct controlled experiments. This assertion is invalid for several reasons: The alternative to establishing empirical regularities in controlled experiments is to establish them by statistical means. This paper has several examples that demonstrate that such replication is possible: the regularities of business cycles; the quantity theory of money and happiness research. The problem is both that more is not done by way of replication and that the results of replication do not become widely known, since they are not considered to be 'novel'.
The Journal of Money, Credit and Banking started a project on replication in the early 1980s that was reviewed by Anderson and Dewald (1994). The idea was to have authors submit their data, as well as computer software used, along with their articles, so that readers could check if the reported results actually follow from the cited data and reported statistical methods. Apart from some practical difficulties, such as data revisions, the project generated little interest and few such replications were actually performed. The project reveals a lack of understanding of what replication in science means; namely, the demonstration of the same empirical regularity against different data sets by independent investigators. This is the only meaningful form of replication since a regularity that applies only to one data set is hardly of interest. 39 Moreover, the verification of an empirical regularity against different data sets obviates the need to check an investigator's arithmetic; if he made errors of this, or any other kind, his results will not be replicated on other data sets. The previous sections have demonstrated that genuine replication in this sense is also possible in the social sciences if the required effort is made.
There is also a sense in which replication in the social sciences is possible. Societies have differed enormously both in their institutions and cultures as well as in the challenges they were confronted with; differences associated both with geographic dispersion and with alternate locations along the axis of historical time. This heterogeneity produces many 'natural experiments' that can be exploited analogously to controlled experiments. This is an avenue that has been insufficiently explored. Thus, in the vast literatures on different aspects of history, comparatively little attention has been paid to the question of what types of institutions have tended to produce favourable outcomes. In particular, little effort has been devoted to the study of the conditions that are favourable to the emergence of democracy An example of the successful exploitation of natural experiments is Titmuss (1970) that has been regarded as something of social science classic. He made an international comparison of blood banks and found that those that operated on a voluntary basis worked well, those operated commercially worked badly. Titmuss' findings apparently had no impact on policy. Subsequent decades saw major scandals involving commercial blood banks, particularly in connection with HIV infected blood.
Still another unexploited source of experimentation is the legislative process. The consequences of new laws are rarely exactly those that were intended. Often, the discrepancy between intended and experienced consequences of a law is major. The consequence is a constant process of revision of existing laws. If the provisional and experimental character of laws were explicitly recognized, it would be possible to design them so as to minimize costs, particularly of unforeseen outcomes, and to maximize the learning effect. The prevailing political culture tends in the opposite direction: politicians feel that they must project an aura of confidence and pretend that they are making laws for all time.

Knowledge that Requires Judgment
Thus far I have discussed knowledge regarding issues that can be resolved by specific evidence. For example: Are the rich happier than the middle class? Is the quantity theory of money valid? There is another kind of knowledge that cannot be confirmed, or disconfirmed, in any simple, straightforward manner. Here there are many different kinds of evidence that may be difficult to evaluate as being either in favour of, or opposed to a hypothesis at issue. Nevertheless, regarding many such hypotheses a consensus was eventually reaches, sometimes after a lengthy period of doubt, or even rejection. Examples are most readily found in fields that are evolutionary in the widest sense of the word; in addition to biological evolution, also the evolution of our planet, the solar system, or the universe. Hypotheses in these fields cannot, by their very nature, verified experimentally.
The following may be cited as examples: The hypothesis of continental drift, proposed by Alfred Wegener in 1912. It remained controversial until the 1960 when it was incorporated into a new theory of plate tectonics.
Another example is Mendelian genetics. Mendel published his two laws of heredity in 1866. Regarding the reception and subsequent developments I quote from Wikipedia: http://en.wikipedia.org/wiki/Gregor_Mendel: Mendel's results were largely neglected. Though they were not completely unknown to biologists of the time, they were not seen as being important. Even Mendel himself did not see their ultimate applicability, and thought they only applied to certain categories of species. In 1900, however, the work was 're-discovered'… …the 'rediscovery' made Mendelism an important but controversial theory…The model of heredity was highly contested by other biologists because it implied that heredity was discontinuous, in opposition to the apparently continuous variation observable. Many biologists also dismissed the theory because they were not sure it would apply to all species, and there seemed to be very few true Mendelian characters in nature. However later work by biologists and statisticians such as R.A. Fisher showed that if multiple Mendelian factors were involved for individual traits, www.economics-ejournal.org they could produce the diverse amount of results observed in nature. Thomas Hunt Morgan and his assistants would later integrate the theoretical model of Mendel with the chromosome theory of inheritance, in which the chromosomes of cells were thought to hold the actual hereditary particles, and create what is now known as classical genetics, which was extremely successful and cemented Mendel's place in history.
The description shows that Mendel's laws could be firmly established only in the context of a complex, sophisticated theory of genetics that could explain a large variety of observations, some of them initially thought to contradict Mendel.
In the natural sciences there evidently exists a tendency, even if sometimes slow and halting, for opinion to converge on relevant and valid knowledge. Evidently this is not the case in the social sciences. The problem of voting, considered in some detail in the present paper offers a striking example. Voting theory, viewed as an essentially mathematical discipline began more than 200 years ago with the work of Borda and Condorcet, both mathematicians. Today it is a recognized discipline, with professional journals, associations, meetings and a voluminous record of publications. Yet, as argued in Section 2.4, no progress is visible in the sense of a convergence of opinion regarding a superior method of voting.
I have argued that the problem of voting is just an instance of the problem of aggregating judgments. Doing this is the business of the opinion research industry. Within this industry, the procedure to be followed is non-controversial. It is what I have called utilitarian voting: the judgments are expressed on a numerical scale and averaged. Recently I discovered that utilitarian voting is used by engineers in such field as automatic control and robotics. The typical problem is to obtain a 'best' value for some parameter that is observed by several independent sensors. The estimated is obtained by averaging the observations, a process that the engineering literature also refers to as 'utilitarian voting'.
The contrast between these fields is striking. The engineers and the opinion researchers are pragmatists; faced with real world problems, they espouse, without much debate, a fairly obvious solution. In academic social choice theory the task is to publish as many mathematical papers as possible in order to climb the academic career ladder. To admit a simple solution would undermine the raison d'être of the discipline.
What is true about voting theory is also true regarding the other major topics discussed in this paper: The stylized facts of business cycles were forgotten when they did not fit in with the prevailing ideologies. Economic measurement as a field of academic research has been largely abandoned without having arrived at acceptable solutions. The situation is similar throughout the various fields of social science. In any of these fields, if we ask what has been discovered in the form of significant confirmed knowledge, we find very little if any.
The inability to secure agreement on relevant and confirmed knowledge in the social sciences is the mirror image of a mistaken emphasis on novelty. The cumulative growth of knowledge in the natural sciences is intimately connected to observational instruments that penetrate realms unreachable by unaided human senses. In early science it was primarily the microscope and the telescope; in modern science it is the giant particle accelerators and observatories that inform us about the infinitely small and the infinitely large. In contrast to this, homo sapiens has been observing and thinking about his society since the beginning of his evolution. Today we have more refined methods for obtaining and organizing social data, the national income and product accounts are an example, that this could lead to a steady stream of startling new discoveries, as in quantum physics or cosmology, is hardly to be expected.
Society has not found a way to organize social thought so as to secure agreement on what is relevant. In one aspect this problem is less difficult than in the natural sciences, in another aspect more so. It is easier because fundamental aspects of society are more obvious. Adam Smith identified the two most basic socially relevant drives: selfishness and empathy. As a social animal man is predisposed to enhance both his own survival and that of his group. To extend empathy beyond the small groups in which homo sapiens has lived throughout most of his evolutionary history is a task for civilization that has so far been accomplished only very imperfectly.
The larger difficulty arises from a commitment to a democratic society. In the natural sciences it has generally been sufficient to secure a consensus within each scientific discipline. To the extent that scientific findings have commercial applications it is only needed to find an entrepreneur who will bring the resulting innovation to market; consumers can then decide to buy it, without any real understanding of the scientific principles involved. Questions of social organization, for example the design of a constitution, require not only a consensus among scholars, but also a wider social consensus. The constitution cannot be produced by an entrepreneur and sold on an individual basis. The fundamental problem of social organization of our time is that social science has not been able to accomplish the first of these steps, let alone the second.
Subsequent to Smith, social thought largely separated the two forces he identified and associated empathy with the political Left, selfishness with the political Right. In academic social science, sociology and psychology came to be associated with the Left, economics with the Right. Communism, the pure ideology of the left aimed at a cooperative society without selfishness or competition. Neoliberals, while a bit more ambivalent, certainly favour competition over cooperation. European social democracy and American liberalism have attempted to bring the two forces into balance, with limited success.

Ideology and Reality: Concluding Remarks
The dysfunctional state of the social sciences is in my view only one instance of the increasingly dysfunctional state of institutions, including universities, and beyond that of the general culture. With modifications, this statement applies not only to the traditionally democratic countries of the West, but also to the rest of the world. I would define the current culture as a field strewn with the ruins of ideologies that failed; no one having an idea of how a new and functional architecture can be constructed from the rubble. An important book that appeared recently in Germany, Grünewald (2006), illustrates the cultural situation. The author is a psychologist and psychotherapist who founded and directs an opinion research institute. Over the past two decades, psychologists at this institute conducted in depth interviews with almost all segments of the German population and on many topics that impact their lives. The interviews are not standardized; each lasts a minimum of two hours and attempts to reach deep layers of feeling and thinking. At the beginning of the interview, the subject might not be consciously aware of these deeper layers hat lie below the image that the she usually www.economics-ejournal.org projects. Similar research in other countries would undoubtedly yield somewhat different results, but I am convinced that the principal results are a reflection of the current state of what is by now in large parts a world culture.
The most important findings are: a. The prevailing attitude is one detachment. People are generally well informed, but non-judgmental. Their attitude towards the world is that of a spectator, as if watching a play in a theater. b. Emotionally they feel themselves and their society to be stressed, engaged in a rat race without visible progress. c. People lead to a considerable extent substitute, or virtual lives. This means that their attention and emotions are focused on stories and events presented in the media that are without any real connection to their lives.
Of these characteristics, the lack of a firm set of beliefs and values is in my view fundamental, while the other aspects are more in the nature of consequences. Grünewald explains the current mood of detachment as a reaction against the fervent hopes with which earlier generations embraced their ideologies. Nowhere was the alteration of ideologies more pronounced than in Germany. During the 19 th and early 20 th Centuries socialist movements struggled against autocratic regimes. The brief experiment in democracy of the Weimar Republic ended with the rise of Hitler. The collapse of the 'Thousand Year Reich' was followed in the West by social democracy promising ever rising living standards while all existential risks were taken care off by the state. In the East, communism was propagated as the ultimate form of social organization. German unification in 1991 provided once more a euphoric feeling of optimism. All of these hopes having been defeated, it is not surprising that people feel it is better to be noncommittal and emotionally detached.
There is I believe an additional cause of the turn to the worse of both the general culture and the social sciences. It is the advent of centralized, bureaucratized mass education following WWII. Of course, I share the belief that an educated public is a prerequisite for democracy to function. However, I believe equally that centrally directed mass 'education' is a self-contradiction. The essence of bureaucratic processes is standardization and the elimination of individual variation. This contradicts the essence of any meaningful education that must nurture the capability for critical, independent thought. Public education, like other social innovations, originated in the Eighteenth Century with the aim of teaching basic literacy to industrial workers. This was a limited, well defined task that public education was able to accomplish. Today's public education systems produce conformist individuals trained to uncritically repeat whatever opinion they think is expected of them. 40 The post-WWII educational system did produce the 68 generation, who viewed themselves as revolutionary, but were highly conformist within their own ideological framework. Many of those active in the student revolts studied either the traditional social sciences, or the newly developing applied fields such as education or social work. These fields, without any clearly established standards, underwent a huge expansion. Many of those entering these fields came from backgrounds without any intellectual tradition, were of a doctrinaire mindset and in the mass of limited ability. The preferred careers sought by the 68 generation were in education, government, politics and the media. The resulting desolate state of both social science and the general culture is not surprising.
In 1960 the sociologist Daniel Bell published The End of Ideology: On the Exhaustion of Political Ideas in the Fifties. Since the 68 generation and the subsequent rise of neoliberalism were still in the future, the claim at that time was premature. Today, the end of ideology is at least for the contemporary world a reality. I do not claim that various ideologies, be they scientism, neoconservatism, or other 'isms' no longer have any advocates, the difference is that they no longer carry conviction for large parts of the world population.
Analogous to the idea of an end to ideology is the idea of The End Of History (Fukuyama, 1992). Fukuyama, another Rand graduate, argued that American style neoliberalism was the ultimate form of social organization so that with its worldwide adaptation the end of historical evolution would be reached. Fukuyama's mentor Hegel had claimed that the Prussian state of his day exemplified the end of historical evolution. Such claims are actually good predictors, not of what they claim, but rather of the end of an era. Only a society that has lost its creative drive will try to assuage that loss with fantasies of ever lasting uniformity.
Ideology is a subject of enormous importance that has received little by way of objective scholarly study. Beginning with Marx, ideology has been a label that we use to characterize the views we oppose. But ideology is pervasive; it needs to be analyzed objectively and brought out into the open if the decline of reason in modern societies is to be reversed.
I do not know what the future holds; with its general loss of faith in the institutions of society I can envision two possible paths. One is an increase in authoritarian government to compensate the increasing dysfunction of institutions. Authoritarian governments, based on the associated elites, will not be able to deal with the increasing complexity of the modern world. This path is likely to end in dictatorship. The better, though perhaps less likely path is the emergence of genuine democracy in which the public regains control over the elites that it appoints to perform various tasks. 41 This path requires more commitment and intelligence than is currently visible. History though always comes up with surprises