III Simulation-Based Methods12.4 Maximum Simulated Likelihood Estimation 393 12.5 Moment-Based Simulation Estimation 398 13.3 Bayesian Analysis of Linear Regression 435 13.5 Markov Chain
Trang 3This book provides a comprehensive treatment of microeconometrics, the analysis ofindividual-level data on the economic behavior of individuals or firms using regres-sion methods applied to cross-section and panel data The book is oriented to the prac-titioner A good understanding of the linear regression model with matrix algebra isassumed The text can be used for Ph.D courses in microeconometrics, in appliedeconometrics, or in data-oriented microeconomics sub-disciplines; and as a referencework for graduate students and applied researchers who wish to fill in gaps in theirtool kit Distinguishing features include emphasis on nonlinear models and robustinference, as well as chapter-length treatments of GMM estimation, nonparametricregression, simulation-based estimation, bootstrap methods, Bayesian methods, strati-fied and clustered samples, treatment evaluation, measurement error, and missing data.The book makes frequent use of empirical illustrations, many based on seven large andrich data sets
A Colin Cameron is Professor of Economics at the University of California, Davis Hecurrently serves as Director of that university’s Center on Quantitative Social ScienceResearch He has also taught at The Ohio State University and has held short-termvisiting positions at Indiana University at Bloomington and at a number of Australianand European universities His research in microeconometrics has appeared in leading
econometrics and economics journals He is coauthor with Pravin Trivedi of sion Analysis of Count Data.
Regres-Pravin K Trivedi is John H Rudy Professor of Economics at Indiana University atBloomington He has also taught at The Australian National University and University
of Southampton and has held short-term visiting positions at a number of Europeanuniversities His research in microeconometrics has appeared in most leading econo-
metrics and health economics journals He coauthored Regression Analysis of Count Data with A Colin Cameron and is on the editorial boards of the Econometrics Journal and the Journal of Applied Econometrics.
i
Trang 5A Colin Cameron Pravin K Trivedi
University of California, Indiana University
Davis
i
Trang 6CAMBRIDGE UNIVERSITY PRESS
Cambridge, New York, Melbourne, Madrid, Cape Town, Singapore, São Paulo, Delhi Cambridge University Press
32 Avenue of the Americas, New York, NY 10013-2473, USA www.cambridge.org
Information on this title: www.cambridge.org/9780521848053
© A Colin Cameron and Pravin K Trivedi 2005 This publication is in copyright Subject to statutory exception and to the provisions of relevant collective licensing agreements,
no reproduction of any part may take place without the written permission of Cambridge University Press.
First published 2005 8th printing 2009 Printed in the United States of America
A catalog record for this publication is available from the British Library.
Library of Congress Cataloging in Publication Data
Cameron, Adrian Colin.
Microeconomics : methods and applications / A Colin Cameron, Pravin K Trivedi.
Cambridge University Press has no responsibility for the persistence or accuracy of URLs for external or third-party Internet Web sites referred to in this publication and does not guarantee that any content on such Web sites is,
or will remain, accurate or appropriate Information regarding prices, travel timetables, and other factual information given in this work are correct at the time of first printing, but Cambridge University Press does not guarantee the accuracy of such information thereafter.
Trang 7my mother and the memory of my father the memory of my parents
v
Trang 92.8 Causal Modeling and Estimation Strategies 35
vii
Trang 106.5 Nonlinear Instrumental Variables 192
6.6 Sequential Two-Step m-Estimation 200
Trang 118.4 Tests for Some Common Misspecifications 274
8.5 Discriminating between NonnestedModels
9.2 Nonparametric Example: Hourly Wage 295
Trang 12III Simulation-Based Methods
12.4 Maximum Simulated Likelihood Estimation 393
12.5 Moment-Based Simulation Estimation 398
13.3 Bayesian Analysis of Linear Regression 435
13.5 Markov Chain Monte Carlo Simulation 445
13.6 MCMC Example: Gibbs Sampler for SUR 452
IV Models for Cross-Section Data
14.2 Binary Outcome Example: Fishing Mode Choice 464
Trang 1315.9 Ordered, Sequential, and Ranked Outcomes 519
16.6 Selection Example: Health Expenditures 553
17.10 Discrete-Time Proportional Hazards 600
17.11 Duration Example: Unemployment Duration 603
xi
Trang 1417.12 Practical Considerations 608
18.2 Unobserved Heterogeneity and Dispersion 612
18.3 Identification in Mixture Models 618
18.4 Specification of the Heterogeneity Distribution 620
18.5 Discrete Heterogeneity and Latent Class Analysis 621
20.3 Count Example: Contacts with Medical Doctor 671
20.4 Parametric Count Regression Models 674
20.6 Multivariate Counts and Endogenous Regressors 685
20.7 Count Example: Further Analysis 690
V Models for Panel Data
21.2 Overview of Models and Estimators 698
21.3 Linear Panel Example: Hours and Wages 708
21.4 Fixed Effects versus Random Effects Models 715
Trang 1522.2 GMM Estimation of Linear Panel Models 744
22.3 Panel GMM Example: Hours and Wages 754
22.4 Random and Fixed Effects Panel GMM 756
22.6 Difference-in-Differences Estimator 768
22.7 Repeated Cross Sections and Pseudo Panels 770
23.3 Nonlinear Panel Example: Patents and R&D 762
24.7 Clustering Example: Vietnam Health Care Use 848
Trang 1625.3 Treatment Effects and Selection Bias 865
25.4 Matching and Propensity Score Estimators 871
25.5 Differences-in-Differences Estimators 878
25.6 Regression Discontinuity Design 879
25.8 Example: The Effect of Training on Earnings 889
26.4 Measurement Errors in Nonlinear Models 911
26.5 Attenuation Bias Simulation Examples 919
27.3 Handling Missing Data without Models 928
A.6 Multivariate Normal Limit Distributions 951
Trang 17List of Figures
4.1 Quantile regression estimates of slope coefficient 89
9.3 Nonparametric regression of log wage on education 297
9.4 Kernel density estimates using different kernels 300
9.7 Nonparametric estimate of derivative of y with respect to x 317
11.1 Bootstrap estimate of the density of t-test statistic 368
12.1 Halton sequence draws compared to pseudo-random draws 411
12.2 Inverse transformation method for unit exponential draws 413
13.1 Bayesian analysis for mean parameter of normal density 424
14.1 Charter boat fishing: probit and logit predictions 466
16.2 Inverse Mills ratio as censoring point c increases 540
17.1 Strike duration: Kaplan–Meier survival function 575
17.2 Weibull distribution: density, survivor, hazard, and cumulativehazard functions
585
17.3 Unemployment duration: Kaplan–Meier survival function 604
17.4 Unemployment duration: survival functions by unemployment insurance 605
17.5 Unemployment duration: Nelson–Aalen cumulated hazard function 606
17.6 Unemployment duration: cumulative hazard function byunemployment insurance
606
xv
Trang 1818.1 Length-biased sampling under stock sampling: example 627
18.2 Unemployment duration: exponential model generalized residuals 633
18.3 Unemployment duration: exponential-gamma model generalizedresiduals
633
18.4 Unemployment duration: Weibull model generalized residuals 635
18.5 Unemployment duration: Weibull-IG model generalized residuals 636
19.1 Unemployment duration: Cox CR baseline survival functions 661
19.2 Unemployment duration: Cox CR baseline cumulative hazards 662
21.3 Hours and wages: within (fixed effects) regression 713
23.1 Patents and R&D: pooled (overall) regression 793
25.2 RD design: treatment assignment in sharp and fuzzy designs 883
25.3 Training impact: earnings against propensity score by treatment 892
Trang 19List of Tables
4.1 Loss Functions and Corresponding Optimal Predictors 67
4.2 Least Squares Estimators and Their Asymptotic Variance 83
4.3 Least Squares: Example with Conditionally Heteroskedastic Errors 84
4.5 Returns to Schooling: Instrumental Variables Estimates 111
5.4 Linear Exponential Family Densities: Leading Examples 148
5.6 Nonlinear Least-Squares Estimators and Their Asymptotic Variance 156
5.7 Exponential Example: Least-Squares and ML Estimates 161
6.2 GMM Estimators in Linear IV Model and Their Asymptotic Variance 186
6.3 GMM Estimators in Nonlinear IV Model and Their Asymptotic Variance 195
7.2 Wald Test Size and Power for Probit Regression Example 253
8.1 Specification m-Tests for Poisson Regression Example 270
8.2 Nonnested Model Comparisons for Poisson Regression Example 284
10.2 Computational Difficulties: A Partial Checklist 350
xvii
Trang 2011.1 Bootstrap Statistical Inference on a Slope Coefficient: Example 367
12.1 Monte Carlo Integration: Example for x Standard Normal 392
12.2 Maximum Simulated Likelihood Estimation: Example 398
12.3 Method of Simulated Moments Estimation: Example 404
13.3 Gibbs Sampling: Seemingly Unrelated Regressions Example 454
14.2 Fishing Mode Choice: Logit and Probit Estimates 465
15.2 Fishing Mode Multinomial Choice: Logit Estimates 493
15.3 Fishing Mode Choice: Marginal Effects for Conditional Logit Model 493
16.1 Health Expenditure Data: Two-Part and Selection Models 554
17.2 Hazard Rate and Survivor Function Computation: Example 582
17.3 Strike Duration: Kaplan–Meier Survivor Function Estimates 583
17.4 Exponential and Weibull Distributions: pdf, cdf, Survivor Function,Hazard, Cumulative Hazard, Mean, and Variance
584
17.5 Standard Parametric Models and Their Hazard and Survivor Functions 585
17.6 Unemployment Duration: Description of Variables 603
17.7 Unemployment Duration: Kaplan–Meier Survival and Nelson–AalenCumulated Hazard Functions
19.2 Unemployment Duration: Competing and Independent RiskEstimates of Exponential Model with and without IG Frailty
659
19.3 Unemployment Duration: Competing and Independent RiskEstimates of Weibull Model with and without IG Frailty
660
20.1 Proportion of Zero Counts in Selected Empirical Studies 666
20.2 Summary of Data Sets Used in Recent Patent–R&D Studies 667
20.3 Contacts with Medical Doctor: Frequency Distribution 672
20.4 Contacts with Medical Doctor: Variable Descriptions 672
20.5 Contacts with Medical Doctor: Count Model Estimates 673
20.6 Contacts with Medical Doctor: Observed and Fitted Frequencies 674
Trang 21L I S T O F T A B L E S
21.1 Linear Panel Model: Common Estimators and Models 699
21.2 Hours and Wages: Standard Linear Panel Model Estimators 710
21.3 Hours and Wages: Autocorrelations of Pooled OLS Residuals 714
21.4 Hours and Wages: Autocorrelations of Within Regression Residuals 715
21.5 Pooled Least-Squares Estimators and Their Asymptotic Variances 721
21.6 Variances of Pooled OLS Estimator with Equicorrelated Errors 724
22.1 Panel Exogeneity Assumptions and Resulting Instruments 752
22.2 Hours and Wages: GMM-IV Linear Panel Model Estimators 755
23.1 Patents and R&D Spending: Nonlinear Panel Model Estimators 794
24.1 Stratification Schemes with Random Sampling within Strata 823
24.2 Properties of Estimators for Different Clustering Models 832
24.4 Vietnam Health Care Use: FE and RE Models for Positive Expenditure 851
24.5 Vietnam Health Care Use: Frequencies for Pharmacy Visits 852
24.6 Vietnam Health Care Use: RE and FE Models for Pharmacy Visits 852
25.3 Training Impact: Sample Means in Treated and Control Samples 890
25.4 Training Impact: Various Estimates of Treatment Effect 891
25.5 Training Impact: Distribution of Propensity Score for Treated andControl Units Using DW (1999) Specification
894
25.7 Training Evaluation: DW (2002) Estimates of ATET 896
26.1 Attenuation Bias in a Logit Regression with Measurement Error 919
26.2 Attenuation Bias in a Nonlinear Regression with AdditiveMeasurement Error
920
27.2 Missing Data Imputation: Linear Regression Estimates with 10%
Missing Data and High Correlation Using MCMC Algorithm
936
27.3 Missing Data Imputation: Linear Regression Estimates with 25%
Missing Data and High Correlation Using MCMC Algorithm
937
27.4 Missing Data Imputation: Linear Regression Estimates with 10%
Missing Data and Low Correlation Using MCMC Algorithm
937
27.5 Missing Data Imputation: Logistic Regression Estimates with 10%
Missing Data and High Correlation Using MCMC Algorithm
938
27.6 Missing Data Imputation: Logistic Regression Estimates with 25%
Missing Data and Low Correlation Using MCMC Algorithm
939
B.1 Continuous Random Variable Densities and Moments 957
B.3 Discrete Random Variable Probability Mass Functions and Moments 959
xix
Trang 23This book provides a detailed treatment of microeconometric analysis, the analysis ofindividual-level data on the economic behavior of individuals or firms This type ofanalysis usually entails applying regression methods to cross-section and panel data.The book aims at providing the practitioner with a comprehensive coverage of sta-tistical methods and their application in modern applied microeconometrics research.These methods include nonlinear modeling, inference under minimal distributionalassumptions, identifying and measuring causation rather than mere association, andcorrecting departures from simple random sampling Many of these features are ofrelevance to individual-level data analysis throughout the social sciences
The ambitious agenda has determined the characteristics of this book First, though oriented to the practitioner, the book is relatively advanced in places A cook-book approach is inadequate because when two or more complications occur simulta-neously – a common situation – the practitioner must know enough to be able to adaptavailable methods Second, the book provides considerable coverage of practical dataproblems (see especially the last three chapters) Third, the book includes substantialempirical examples in many chapters to illustrate some of the methods covered Fi-nally, the book is unusually long Despite this length we have been space-constrained
al-We had intended to include even more empirical examples, and abbreviated tations will at times fail to recognize the accomplishments of researchers who havemade substantive contributions
presen-The book assumes a good understanding of the linear regression model with matrixalgebra It is written at the mathematical level of the first-year economics Ph.D se-quence, comparable to Greene (2003) We have two types of readers in mind First, thebook can be used as a course text for a microeconometrics course, typically taught inthe second year of the Ph.D., or for data-oriented microeconomics field courses such
as labor economics, public economics, and industrial organization Second, the bookcan be used as a reference work for graduate students and applied researchers whodespite training in microeconometrics will inevitably have gaps that they wish to fill.For instructors using this book as an econometrics course text it is best to introducethe basic nonlinear cross-section and linear panel data models as early as possible,
xxi
Trang 24initially skipping many of the methods chapters The key methods chapter (Chapter 5)covers maximum-likelihood and nonlinear least-squares estimation Knowledge ofmaximum likelihood and nonlinear least-squares estimators provides adequate back-ground for the most commonly used nonlinear cross-section models (Chapters 14–17and 20), basic linear panel data models (Chapter 21), and treatment evaluation meth-ods (Chapter 25) Generalized method of moments estimation (Chapter 6) is neededespecially for advanced linear panel data methods (Chapter 22).
For readers using this book as a reference work, the chapters have been written to be
as self-contained as possible The notable exception is that some command of generalestimation results in Chapter 5, and occasionally Chapter 6, will be necessary Mostchapters on models are structured to begin with a discussion and example that is acces-sible to a wide audience
The Web site www.econ.ucdavis.edu/faculty/cameron provides all the data andcomputer programs used in this book and related materials useful for instructionalpurposes
This project has been long and arduous, and at times seemingly without an end Itscompletion has been greatly aided by our colleagues, friends, and graduate students
We thank especially the following for reading and commenting on specific chapters:Bijan Borah, Kurt Br¨ann¨as, Pian Chen, Tim Cogley, Partha Deb, Massimiliano DeSantis, David Drukker, Jeff Gill, Tue Gorgens, Shiferaw Gurmu, Lu Ji, Oscar Jorda,Roger Koenker, Chenghui Li, Tong Li, Doug Miller, Murat Munkin, Jim Prieger,Ahmed Rahmen, Sunil Sapra, Haruki Seitani, Yacheng Sun, Xiaoyong Zheng, andDavid Zimmer Pian Chen gave detailed comments on most of the book We thankRajeev Dehejia, Bronwyn Hall, Cathy Kling, Jeffrey Kling, Will Manning, BrianMcCall, and Jim Ziliak for making their data available for empirical illustrations Wethank our respective departments for facilitating our collaboration and for the produc-tion and distribution of the draft manuscript at various stages We benefited from thecomments of two anonymous reviewers Guidance, advice, and encouragement fromour Cambridge editor, Scott Parris, have been invaluable
Our interest in econometrics owes much to the training and environments we countered as students and in the initial stages of our academic careers The first authorthanks The Australian National University; Stanford University, especially TakeshiAmemiya and Tom MaCurdy; and The Ohio State University The second author thanksthe London School of Economics and The Australian National University
en-Our interest in writing a book oriented to the practitioner owes much to our exposure
to the research of graduate students and colleagues at our respective institutions, Davis and IU-Bloomington
UC-Finally, we thank our families for their patience and understanding without whichcompletion of this project would not have been possible
A Colin CameronDavis, CaliforniaPravin K TrivediBloomington, Indiana
Trang 25Chapters 2 and 3 set the scene for the remainder of the book by introducing thereader to key model and data concepts that shape the analyses of later chapters.
A key distinction in econometrics is between essentially descriptive models anddata summaries at various levels of statistical sophistication and models that go be-yond associations and attempt to estimate causal parameters The classic definitions
of causality in econometrics derive from the Cowles Commission simultaneous tions models that draw sharp distinctions between exogenous and endogenous vari-ables, and between structural and reduced form parameters Although reduced formmodels are very useful for some purposes, knowledge of structural or causal parame-ters is essential for policy analyses Identification of structural parameters within thesimultaneous equations framework poses numerous conceptual and practical difficul-ties An increasingly-used alternative approach based on the potential outcome model,also attempts to identify causal parameters but it does so by posing limited questionswithin a more manageable framework Chapter 2 attempts to provide an overview ofthe fundamental issues that arise in these and other alternative frameworks Readerswho initially find this material challenging should return to this chapter after gaininggreater familiarity with specific models covered later in the book
equa-The empirical researcher’s ability to identify causal parameters depends not only
on the statistical tools and models but also on the type of data available An mental framework provides a standard for establishing causal connections However,observational, not experimental, data form the basis of much of econometric inference.Chapter 3 surveys the pros and cons of three main types of data: observational data,data from social experiments, and data from natural experiments The strengths andweaknesses of conducting causal inference based on each type of data are reviewed
experi-1
Trang 27C H A P T E R 1Overview
1.1. IntroductionThis book provides a detailed treatment of microeconometric analysis, the analysis
of individual-level data on the economic behavior of individuals or firms A broaderdefinition would also include grouped data Usually regression methods are applied tocross-section or panel data
Analysis of individual data has a long history Ernst Engel (1857) was among theearliest quantitative investigators of household budgets Allen and Bowley (1935),Houthakker (1957), and Prais and Houthakker (1955) made important contributionsfollowing the same research and modeling tradition Other landmark studies that werealso influential in stimulating the development of microeconometrics, even thoughthey did not always use individual-level information, include those by Marschak andAndrews (1944) in production theory and by Wold and Jureen (1953), Stone (1953),and Tobin (1958) in consumer demand
As important as the above earlier cited work is on household budgets and demandanalysis, the material covered in this book has stronger connections with the work ondiscrete choice analysis and censored and truncated variable models that saw their firstserious econometric applications in the work of McFadden (1973, 1984) and Heckman(1974, 1979), respectively These works involved a major departure from the over-whelming reliance on linear models that characterized earlier work Subsequently, theyhave led to significant methodological innovations in econometrics Among the earliertextbook-level treatments of this material (and more) are the works of Maddala (1983)and Amemiya (1985) As emphasized by Heckman (2001), McFadden (2001), and oth-ers, many of the fundamental issues that dominated earlier work based on market dataremain important, especially concerning the conditions necessary for identifiability ofcausal economic relations Nonetheless, the style of microeconometrics is sufficientlydistinct to justify writing a text that is exclusively devoted to it
Modern microeconometrics based on individual-, household-, and level data owes a great deal to the greater availability of data from cross-sectionand longitudinal sample surveys and census data In the past two decades, with the
establishment-3
Trang 28expansion of electronic recording and collection of data at the individual level, datavolume has grown explosively So too has the available computing power for analyzinglarge and complex data sets In many cases event-level data are available; for example,marketing science often deals with purchase data collected by electronic scanners insupermarkets, and industrial organization literature contains econometric analyses ofairline travel data collected by online booking systems There are now new branches ofeconomics, such as social experimentation and experimental economics, that generate
“experimental” data These developments have created many new modeling nities that are absent when only aggregated market-level data are available Meanwhilethe explosive growth in the volume and types of data has also given rise to numerousmethodological issues Processing and econometric analysis of such large microdata-bases, with the objective of uncovering patterns of economic behavior, constitutes thecore of microeconometrics Econometric analysis of such data is the subject matter ofthis book
opportu-Key precursors of this book are the books by Maddala (1983) and Amemiya (1985).Like them it covers topics that are presented only briefly, or not at all, in undergraduateand first-year graduate econometrics courses Especially compared to Amemiya (1985)this book is more oriented to the practitioner The level of presentation is nonethelessadvanced in places, especially for applied researchers in disciplines that are less math-ematically oriented than economics
A relatively advanced presentation is needed for several reasons First, the data are
often discrete or censored, in which case nonlinear methods such as logit, probit,
and Tobit models are used This leads to statistical inference based on more difficultasymptotic theory
Second, distributional assumptions for such data become critically important One
response is to develop highly parametric models that are sufficiently detailed to capturethe complexities of data, but these models can be challenging to estimate A more com-mon response is to minimize parametric assumptions and perform statistical inferencebased on standard errors that are “robust” to complications such as heteroskedasticityand clustering In such cases considerable knowledge can be needed to ensure validstatistical inference even if a standard regression package is used
Third, economic studies often aim to determine causation rather than merely
mea-sure correlation, despite access to observational rather than experimental data Thisleads to methods to isolate causation such as instrumental variables, simultaneousequations, measurement error correction, selection bias correction, panel data fixedeffects, and differences-in-differences
Fourth, microeconomic data are typically collected using cross-section and panel
surveys, censuses, or social experiments Survey data collected using these methods
are subject to problems of complex survey methodology, departures from simple dom sampling assumptions, and problems of sample selection, measurement errors,and incomplete, and/or missing data Dealing with such issues in a way that can sup-port valid population inferences from the estimated econometric models populationrequires use of advanced methods
ran-Finally, it is not unusual that two or more complications occur simultaneously,
such as endogeneity in a logit model with panel data Then a cookbook approach
Trang 291 2 D I S T I N C T I V E A S P E C T S O F M I C R O E C O N O M E T R I C S
becomes very difficult to implement Instead, considerable understanding of the ory underlying the methods is needed, as the researcher may need to read econometricsjournal articles and adapt standard econometrics software
the-1.2. Distinctive Aspects of Microeconometrics
We now consider several advantages of microeconometrics that derive from its tive features
distinc-1.2.1 Discreteness and Nonlinearity
The first and most obvious point is that microeconometric data are usually at a lowlevel of aggregation This has a major consequence for the functional forms used toanalyze the variables of interest In many, if not most, cases linear functional formsturn out to be simply inappropriate More fundamentally, disaggregation brings to the
forefront heterogeneity of individuals, firms, and organizations that should be
prop-erly controlled (modeled) if one wants to make valid inferences about the undprop-erlyingrelationships We discuss these issues in greater detail in the following sections.Although aggregation is not entirely absent in microdata, as for example whenhousehold- or establishment-level data are collected, the level of aggregation is usu-ally orders of magnitude lower than is common in macro analyses In the latter case theprocess of aggregation leads to smoothing, with many of the movements in oppositedirections canceling in the course of summation The aggregated variables often showsmoother behavior than their components, and the relationships between the aggre-gates frequently show greater smoothness than the components For example, a rela-tion between two variables at a micro level may be piecewise linear with many nodes.After aggregation the relationship is likely to be well approximated by a smooth func-tion Hence an immediate consequence of disaggregation is the absence of features ofcontinuity and smoothness both of the variables themselves and of the relationshipsbetween them
Usually individual- and firm-level data cover a huge range of variation, both in thecross-section and time-series dimensions For example, average weekly consumption
of (say) beef is highly likely to be positive and smoothly varying, whereas that of an dividual household in a given week may be frequently zero and may also switch to pos-itive values from time to time The average number of hours worked by female workers
in-is unlikely to be zero, but many individual females have zero market hours of work(corner solutions), switching to positive values at other times in the course of their la-bor market history Average household expenditure on vacations is usually positive, butmany individual households may have zero expenditure on vacations in any given year.Average per capita consumption of tobacco products will usually be positive, but manyindividuals in the population have never consumed these products and never will, irre-spective of price and income considerations As Pudney (1989) has observed, micro-data exhibit “holes, kinks and corners.” The holes correspond to nonparticipation in theactivity of interest, kinks correspond to the switching behavior, and corners correspond
5
Trang 30to the incidence of nonconsumption or nonparticipation at specific points of time.That is, discreteness and nonlinearity of response are intrinsic to microeconometrics.
An important class of nonlinear models in microeconometrics deals with limited
dependent variables (Maddala, 1983) This class includes many models that provide
suitable frameworks for analyzing discrete responses and responses with limited range
of variation Such tools of analyses are of course also available for analyzing data, if required The point is that they are indispensable in microeconometrics andgive it its distinctive feature
macro-1.2.2 Greater Realism Macroeconometrics is sometimes based on strong assumptions; the representative
agent assumption is a leading example A frequent appeal is made to microeconomicreasoning to justify certain specifications and interpretations of empirical results How-ever, it is rarely possible to say explicitly how these are affected by aggregation overtime and micro units Alternatively, very extreme aggregation assumptions are made.For example, aggregates are said to reflect the behavior of a hypothetical representativeagent Such assumptions also are not credible
From the viewpoint of microeconomic theory, quantitative analysis founded onmicrodata may be regarded as more realistic than that based on aggregated data Thereare three justifications for this claim First, the measurement of the variables involved
in such hypotheses is often more direct (though not necessarily free from measurementerror) and has greater correspondence to the theory being tested Second, hypothesesabout economic behavior are usually developed from theories of individual behavior Ifthese hypotheses are tested using aggregated data, then many approximations and sim-plifying assumptions have to be made The simplifying assumption of a representativeagent causes a great loss of information and severely limits the scope of an empiricalinvestigation Because such assumptions can be avoided in microeconometrics, andusually are, in principle the microdata provide a more realistic framework for testingmicroeconomic hypotheses This is not a claim that the promise of microdata is nec-essarily achieved in empirical work Such a claim must be assessed on a case-by-casebasis Finally, a realistic portrayal of economic activity should accommodate a broadrange of outcomes and responses that are a consequence of individual heterogeneityand that are predicted by underlying theory In this sense microeconomic data sets cansupport more realistic models
Microeconometric data are often derived from household or firm surveys, typicallyencompassing a wide range of behavior, with many of the behavioral outcomes tak-ing the form of discrete or categorical responses Such data sets have many awkwardfeatures that call for special tools in the formulation and analysis that, although notentirely absent from macroeconometric work, nevertheless are less widely used
1.2.3 Greater Information Content
The potential advantages of microdata sets can be realized if such data are tive Because sample surveys often provide independent observations on thousands of
Trang 31informa-1 2 D I S T I N C T I V E A S P E C T S O F M I C R O E C O N O M E T R I C S
cross-sectional units, such data are thought to be more informative than the standard,usually highly serially correlated, macro time series typically consisting of at most afew hundred observations
As will be explained in the next chapter, in practice the situation is not so clear-cutbecause the microdata may be quite noisy At the individual level many (idiosyncratic)factors may play a large role in determining responses Often these cannot be observed,leading one to treat them under the heading of a random component, which can be avery large part of observed variation In this sense randomness plays a larger role inmicrodata than in macrodata Of course, this affects measures of goodness of fit of theregressions Students whose initial exposure to econometrics comes through aggregate
time-series analysis are often conditioned to see large R2 values When encounteringcross-section regressions for the first time, they express disappointment or even alarm
at the “low explanatory power” of the regression equation Nevertheless, there remains
a strong presumption that, at least in certain dimensions, large microdata sets are highlyinformative
Another qualification is that when one is dealing with purely cross-section data,very little can be said about the intertemporal aspects of relationships under study.This particular aspect of behavior can be studied using panel and transition data
In many cases one is interested in the behavioral responses of a specific group ofeconomic agents under some specified economic environment One example is theimpact of unemployment insurance on the job search behavior of young unemployedpersons Another example is the labor supply responses of low-income individualsreceiving income support Unless microdata are used such issues cannot be addresseddirectly in empirical work
1.2.4 Microeconomic Foundations
Econometric models vary in the explicit role given to economic theory At one end ofthe spectrum there are models in which the a priori theorizing may play a dominantrole in the specification of the model and in the choice of an estimation procedure Atthe other end of the spectrum are empirical investigations that make much less use ofeconomic theory
The goal of the analysis in the first case is to identify and estimate fundamentalparameters, sometimes called deep parameters, that characterize individual taste andpreferences and/or technological relationships As a shorthand designation, we call
this the structural approach Its hallmark is a heavy dependence on economic theory
and emphasis on causal inference Such models may require many assumptions, such
as the precise specification of a cost or production function or specification of thedistribution of error terms The empirical conclusions of such an exercise may not
be robust with respect to the departures from the assumptions In Section 2.4.4 weshall say more about this approach At the present stage we simply emphasize that ifthe structural approach is implemented with aggregated data, it will yield estimates
of the fundamental parameters only under very stringent (and possibly unrealistic)conditions Microdata sets provide a more promising environment for the structuralapproach, essentially because they permit greater flexibility in model specification
7
Trang 32The goal of the analysis in the second case is to model relationship(s) between sponse variables of interest conditionally on variables the researcher takes as given, or
re-exogenous More formal definitions of endogeneity and exogeneity are given in ter 2 As a shorthand designation, we call this a reduced form approach The essential
Chap-point is that reduced form analysis does not always take into account all causal
inter-dependencies A regression model in which the focus is on the prediction of y given
regressors x, and not on the causal interpretation of the regression parameters, is often
referred to as a reduced form regression As will be explained in Chapter 2, in generalthe parameters of the reduced form model are functions of structural parameters Theymay not be interpretable without some information about the structural parameters
1.2.5 Disaggregation and Heterogeneity
It is sometimes said that many problems and issues of macroeconometrics arise fromserial correlation of macro time series, and those of microeconometrics arise fromheteroskedasticity of individual-level data Although this is a useful characterization ofthe modeling effort in many microeconometric analyses, it needs amplification and issubject to important qualifications In a range of microeconometric models, modeling
of dynamic dependence may be an important issue
The benefits of disaggregation, which were emphasized earlier in this section, come
at a cost: As the data become more disaggregated the importance of controlling forinterindividual heterogeneity increases Heterogeneity, or more precisely unobservedheterogeneity, plays a very important role in microeconometrics Obviously, manyvariables that reflect interindividual heterogeneity, such as gender, race, educationalbackground, and social and demographic factors, are directly observed and hence can
be controlled for In contrast, differences in individual motivation, ability, intelligence,and so forth are either not observed or, at best, imperfectly observed
The simplest response is to ignore such heterogeneity, that is, to absorb it into theregression disturbance After all this is how one treats the myriad small unobservedfactors This step of course increases the unexplained part of the variation More seri-
ously, ignoring persistent interindividual differences leads to confounding with other
factors that are also sources of persistent interindividual differences Confounding issaid to occur when the individual contributions of different regressors (predictor vari-ables) to the variation in the variable of interest cannot be statistically separated Sup-
pose, for example, that the factor x1(schooling) is said to be the source of variation in
y (earnings), when another variable x2 (ability), which is another source of variation,does not appear in the model Then that part of total variation that is attributable tothe second variable may be incorrectly attributed to the first variable Intuitively, theirrelative importances are confounded A leading source of confounding bias is the in-correct omission of regressors from the model and the inclusion of other variables thatare proxies for the omitted variable
Consider, for example, the case in which a program participation (0/1 dummy)
variable D is included in the regression mean function with a vector of regressors x,
Trang 331 2 D I S T I N C T I V E A S P E C T S O F M I C R O E C O N O M E T R I C S
where u is an error term The term “treatment” is used in biological and experimental
sciences to refer to an administered regimen involving participants in some trial Ineconometrics it commonly refers to participation in some activity that may impact anoutcome of interest This activity may be randomly assigned to the participants or may
be self-selected by the participant Thus, although it is acknowledged that individualschoose their years of schooling, one still thinks of years of schooling as a “treatment”variable Suppose that program participation is taken to be a discrete variable Thecoefficientα of the “treatment variable” measures the average impact of the program participation (D= 1), conditional on covariates If one does not control for unob-served heterogeneity, then a potential ambiguity affects the interpretation of the results
If d is found to have a significant impact, then the following question arises: Is α nificantly different from zero because D is correlated with some unobserved variable that affects y or because there is a causal relationship between D and y? For example,
sig-if the program considered is university education, and the covariates do not include ameasure of ability, giving a fully causal interpretation becomes questionable Becausethe issue is important, more attention should be given to how to control for unobservedheterogeneity
In some cases where dynamic considerations are involved the type of data availablemay put restrictions on how one can control for heterogeneity Consider the example
of two households, identical in all relevant respects except that one exhibits a tematically higher preference for consuming good A One could control for this byallowing individual utility functions to include a heterogeneity parameter that reflectstheir different preferences Suppose now that there is a theory of consumer behaviorthat claims that consumers become addicted to good A, in the sense that the more theyconsume of it in one period, the greater is the probability that they will consume more
sys-of it in the future This theory provides another explanation sys-of persistent vidual differences in the consumption of good A By controlling for heterogeneouspreferences it becomes possible to test which source of persistence in consumption –preference heterogeneity or addiction – accounts for different consumption patterns.This type of problem arises whenever some dynamic element generates persistence
interindi-in the observed outcomes Several examples of this type of problem arise interindi-in variousplaces in the book
A variety of approaches for modeling heterogeneity coexist in microeconometrics
A brief mention of some of these follows, with details postponed until later
An extreme solution is to ignore all unobserved interindividual differences If served heterogeneity is uncorrelated with observed heterogeneity, and if the outcomebeing studied has no intertemporal dependence, then the aforementioned problems willnot arise Of course, these are strong assumptions and even with these assumptions notall econometric difficulties disappear
unob-One approach for handling heterogeneity is to treat it as a fixed effect and to
esti-mate it as a coefficient of an individual specific 0/1 dummy variable For example, in
a cross-section regression, each micro unit is allowed its own dummy variable cept) This leads to an extreme proliferation of parameters because when a new individ-ual is added to the sample, a new intercept parameter is also added Thus this approachwill not work if our data are cross sectional The availability of multiple observations
(inter-9
Trang 34per individual unit, most commonly in the form of panel data with T time-series servations for each of the N cross-section units, makes it possible to either estimate
ob-or eliminate the fixed effect, fob-or example by first differencing if the model is linearand the fixed effect is additive If the model is nonlinear, as is often the case, the fixedeffect will usually not be additive and other approaches will need to be considered
A second approach to modeling unobserved heterogeneity is through a random fects model There are a number of ways in which the random effects model can be
ef-formulated One popular formulation assumes that one or more regression parameters,often just the regression intercept, varies randomly across the cross section In anotherformulation the regression error is given a component structure, with an individualspecific random component The random effects model then attempts to estimate theparameters of the distribution from which the random component is drawn In somecases, such as demand analysis, the random term can be interpreted as random prefer-ence variation Random effects models can be estimated using either cross-section orpanel data
1.2.6 Dynamics
A very common assumption in cross-section analysis is the absence of ral dependence, that is, an absence of dynamics Thus, implicitly it is assumed thatthe observations correspond to a stochastic equilibrium, with the deviation from theequilibrium being represented by serially independent random disturbances Even inmicroeconometrics for some data situations such an assumption may be too strong.For example, it is inconsistent with the presence of serially correlated unobserved het-erogeneity Dependence on lagged dependent variables also violates this assumption.The foregoing discussion illustrates some of the potential limitations of a singlecross-section analysis Some limitations may be overcome if repeated cross sectionsare available However, if there is dynamic dependence, the least problematic approachmight well be to use panel data
intertempo-1.3. Book Outline
The book is split into six parts Part 1 presents the issues involved in microeconometricmodeling Parts 2 and 3 present general theory for estimation and statistical inferencefor nonlinear regression models Parts 4 and 5 specialize to the core models used inapplied microeconometrics for, respectively, cross-section and panel data Part 6 coversbroader topics that make considerable use of material presented in the earlier chapters.The book outline is summarized in Table 1.1 The remainder of this section detailseach part in turn
1.3.1 Part 1: Preliminaries Chapters 2 and 3 expand on the special features of the microeconometric approach
to modeling and microeconomic data structures within the more general statistical
Trang 35Table 1.1 Book Outline
Part and Chapter Backgrounda Example
1 Preliminaries
2 Causal and Noncausal Models – Simultaneous equations models
3 Microeconomic DataStructures
– Observational data
2 Core Methods
4 Linear Models – Ordinary least squares
5 Maximum Likelihood andNonlinear Least-SquaresEstimation
– m-estimation or extremum
estimation
6 Generalized Method ofMoments and SystemsEstimation
5,7 Conditional moment test
9 Semiparametric Methods – Kernel regression
10 Numerical Optimization 5 Newton–Raphson iterative method
3 Simulation-Based Methods
11 Bootstrap Methods 7 Percentile t-method
12 Simulation-Based Methods 5 Maximum simulated likelihood
13 Bayesian Methods – Markov chain Monte Carlo
4 Models for Cross-Section Data
14 Binary Outcome Models 5 Logit, probit for y = (0, 1)
15 Multinomial Models 5,14 Multinomial logit for
y = (1, , m)
16 Tobit and Selection Models 5,14 Tobit for y = max(y∗, 0)
17 Transition Data: SurvivalAnalysis
5 Cox proportional hazards for
y = min(y∗, c)
18 Mixture Models andUnobserved Heterogeneity
5,17 Unobserved heterogeneity
19 Models for Multiple Hazards 5,17 Multiple hazards
20 Models of Count Data 5 Poisson for y = 0, 1, 2,
5 Models for Panel Data
21 Linear Panel Models: Basics – Fixed and random effects
22 Linear Panel Models:
5 Data (y i j , x i j ) correlated over j
25 Treatment Evaluation 5,21 Regressor d= 1 if in program
26 Measurement Error Models 5 Logit model with measurement
Trang 36arena of regression analysis Many of the issues raised in these chapters are pursuedthroughout the book as the reader develops the necessary tools.
1.3.2 Part 2: Core Methods
Chapters 4–10 detail the main general methods used in classical estimation and tistical inference The results given in Chapter 5 in particular are extensively usedthroughout the book
sta-Chapter 4 presents some results for the linear regression model, emphasizing those
issues and methods that are most relevant for the rest of the book Analysis is relativelystraightforward as there is an explicit expression for linear model estimators such asordinary least squares
Chapters 5 and 6 present estimation theory that can be applied to nonlinear models
for which there is usually no explicit solution for the estimator Asymptotic theory
is used to obtain the distribution of estimators, with emphasis on obtaining robuststandard error estimates that rely on relatively weak distributional assumptions A quitegeneral treatment of estimation, along with specialization to nonlinear least-squaresand maximum likelihood estimation, is presented in Chapter 5 The more challenginggeneralized method of moments estimator and specialization to instrumental variablesestimation are given separate treatment in Chapter 6
Chapter 7 presents classical hypothesis testing when estimators are nonlinear and the hypothesis being tested is possibly nonlinear in parameters Specification tests in
addition to hypothesis tests are the subject of Chapter 8
Chapter 9 presents semiparametric estimation methods such as kernel regression.
The leading example is flexible modeling of the conditional mean For the patents
ex-ample, the nonparametric regression model is E[y |x] = g(x), where the function g(·)
is unspecified and is instead estimated Then estimation has an infinite-dimensional
component g(·) leading to a nonstandard asymptotic theory With additional sors some further structure is needed and the methods are called semiparametric orseminonparametric
regres-Chapter 10 presents the computational methods used to compute a parameter
esti-mate when the estimator is defined implicitly, usually as the solution to some first-orderconditions
1.3.3 Part 3: Simulation-Based Methods
Chapters 11–13 consider methods of estimation and inference that rely on simulation.These methods are generally more computationally intensive and, currently, less uti-lized than the methods presented in Part 2
Chapter 11 presents the bootstrap method for statistical inference This yields the
empirical distribution of an estimator by obtaining new samples by simulation, such
as by repeated resampling with replacement from the original sample The bootstrapcan provide a simple way to obtain standard errors when the formulas from asymp-totic theory are complex, as is the case for some two-step estimators Furthermore, if
Trang 371 3 B O O K O U T L I N E
implemented appropriately, the bootstrap can lead to better statistical inference insmall samples
Chapter 12 presents simulation-based estimation methods, developed for models
that involve an integral over a probability distribution for which there is no form solution Estimation is still possible by making multiple draws from the relevantdistribution and averaging
closed-Chapter 13 presents Bayesian methods, which combine a distribution for the
ob-served data with a specified prior distribution for parameters to obtain a posterior tribution of the parameters that is the basis for estimation Recent advances make com-putation possible even if there is no closed-form solution for the posterior distribution.Bayesian analysis can provide an approach to estimation and inference that is quite dif-ferent from the classical approach However, in many cases only the Bayesian tool kit
dis-is adopted to permit classical estimation and inference for problems that are otherwdis-iseintractable
1.3.4 Part 4: Models for Cross-Section Data Chapters 14–20 present the main nonlinear models for cross-section data This part is
the heart of the book and presents advanced topics such as models for limited dent variables and sample selection The classes of models are defined by the range ofvalues taken by the dependent variable
depen-Binary data models for dependent variable that can take only two possible values,
say y = 0 or y = 1, are presented in Chapter 14 In Chapter 15 an extension is made to
multinomial models, for dependent variable that takes several discrete values
Exam-ples include employment status (employed, unemployed, and out of the labor force)and mode of transportation to work (car, bus, or train) Linear models can be informa-tive but are not appropriate, as they can lead to predicted probabilities outside the unitinterval Instead logit, probit, and related models are used
Chapter 16 presents models with censoring, truncation, sample selection
Exam-ples include annual hours of work, conditional on choosing to work, and hospital penditures, conditional on being hospitalized In these cases the data are incompletely
ex-observed with a bunching of observations at y = 0 and with the remaining y > 0.
The model for the observed data can be shown to be nonlinear even if the underlyingprocess is linear, and linear regression on the observed data can be very misleading.Simple corrections for censoring, truncation, or sample selection such as the Tobitmodel exist, but these are very dependent on distributional assumptions
Models for duration data are presented in Chapters 17–19 An example is length
of unemployment spell Standard regression models include the exponential, Weibull,and Cox proportional hazards model Additionally, as in Chapter 16, the dependentvariable is often incompletely observed For example, the data may be on the length of
a current spell that is incomplete, rather than the length of a completed spell
Chapter 20 presents count data models Examples include various measures of
health utilization such as number of doctor visits and number of days hospitalized.Again the model is nonlinear, as counts and hence the conditional mean are nonnega-tive Leading parametric models include the Poisson and negative binomial
13
Trang 381.3.5 Part 5: Models for Panel Data Chapters 21–23 present methods for panel data Here the data are observed in several
time periods for each of the many individuals in the sample, so the dependent variableand regressors are indexed by both individual and time Any analysis needs to controlfor the likely positive correlation of error terms in different time periods for a given in-dividual Additionally, panel data can provide sufficient data to control for unobservedtime-invariant individual-specific effects, permitting identification of causation underweaker assumptions than those needed if only cross-section data are available.The basic linear panel data model is presented in Chapter 21, with emphasis on
fixed effects and random effects models Extensions of linear models to permit lagged
dependent variables and endogenous regressors are presented in Chapter 22 Panelmethods for the nonlinear models of Part 4 are presented in Chapter 23
The panel data methods are placed late in the book to permit a unified self-containedtreatment Chapter 21 could have been placed immediately after Chapter 4 and is writ-ten in an accessible manner that relies on little more than knowledge of least-squaresestimation
1.3.6 Part 6: Further Topics
This part considers important topics that can generally relate to any and all modelscovered in Parts 4 and 5 Chapter 24 deals with modeling of clustered data in sev-eral different models Chapter 25 discusses treatment evaluation Treatment evaluation
is a general term that can cover a wide variety of models in which the focus is onmeasuring the impact of some “treatment” that is either exogenously or randomly as-signed to an individual on some measure of interest, denoted an “outcome variable.”Chapter 26 deals with the consequences of measurement errors in outcome and/orregressor variables, with emphasis on some leading nonlinear models Chapter 27considers some methods of handling missing data in linear and nonlinear regressionmodels
1.4. How to Use This Book
The book assumes a basic understanding of the linear regression model with matrixalgebra It is written at the mathematical level of the first-year economics Ph.D se-quence, comparable to Greene (2003)
Although some of the material in this book is covered in a first-year sequence,most of it appears in second-year econometrics Ph.D courses or in data-oriented mi-croeconomics field courses such as labor economics, public economics, or industrialorganization This book is intended to be used as both an econometrics text and as anadjunct for such field courses More generally, the book is intended to be useful as areference work for applied researchers in economics, in related social sciences such associology and political science, and in epidemiology
For readers using this book as a reference work, the models chapters have beenwritten to be as self-contained as possible For the specific models presented in Parts 4
Trang 391 5 S O F T W A R E
Table 1.2 Outline of a 20-Lecture 10-Week Course
1–3 4, Appx A Review of linear models and asymptotic theory4–7 5 Estimation: m-estimation, ML, and NLS
8 10 Estimation: numerical optimization9–11 14, 15 Models: binary and multinomial12–14 16 Models: censored and truncated
For instructors using this book as a course text it is best to introduce the basic linear cross-section and linear panel data models as early as possible, skipping many
non-of the methods chapters The most commonly used nonlinear cross-section modelsare presented in Chapters 14–16; these require knowledge of maximum likelihoodand least-squares estimation, presented in Chapter 5 Chapter 21 on linear panel datamodels requires even less preparation, essentially just Chapter 4
Table 1.2 provides an outline for a one-quarter second-year graduate course taught
at the University of California, Davis, immediately following the required first-yearstatistics and econometrics sequence A quarter provides sufficient time to cover thebasic results given in the first half of the chapters in this outline With additional timeone can go into further detail or cover a subset of Chapters 11–13 on computation-ally intensive estimation methods (simulation-based estimation, the bootstrap, which
is also briefly presented in Chapter 7, and Bayesian methods); additional cross-sectionmodels (durations and counts) presented in Chapters 17–20; and additional panel datamodels (linear model extensions and nonlinear models) given in Chapters 22 and 23
At Indiana University, Bloomington, a 15-week semester-long field course in croeconometrics is based on material in most of Parts 4 and 5 The prerequisite coursesfor this course cover material similar to that in Part 2
mi-Some exercises are provided at the end of each chapter after the first three ductory chapters These exercises are usually learning-by-doing exercises; some arepurely methodological whereas others entail analysis of generated or actual data Thelevel of difficulty of the questions is mostly related to the level of difficulty of the topic
intro-1.5. Software
There are many software packages available for data analysis Popular packages withstrong microeconometric capabilities include LIMDEP, SAS, and STATA, all of which
15
Trang 40offer an impressive range of canned routines and additionally support user-defined cedures using a matrix programming language Other packages that are also widelyused include EVIEWS, PCGIVE, and TSP Despite their time-series orientation, thesecan support some cross-section data analysis Users who wish to do their own pro-gramming also have available a variety of options including GAUSS, MATLAB, OX,and SAS/IML The latest detailed information about these packages and many otherscan be efficiently located via an Internet browser and a search engine.
pro-1.6. Notation and Conventions
Vector and matrix algebra are used extensively
Vectors are defined as column vectors and represented using lowercase bold For
example, for linear regression the regressor vector x is a K × 1 column vector with jth entry x j and the parameter vector β is a K × 1 column vector with jth entry β j, so
Then the linear regression model y = β1x1+ β2x2 + · · · + β K x K + u is expressed as
y= xβ + u At times a subscript i is added to denote the typical ith observation The linear regression equation for the i th observation is then
i β + u i The sample is one of N observations, {(y i , x i), i = 1, , N} In this book observa- tions are usually assumed to be independent over i
Matrices are represented using uppercase bold In matrix notation the sample is
(y, X), where y is an N × 1 vector with ith entry y i and X is a matrix with i th row xi,so
where u is an N × 1 column vector with ith entry u i
Matrix notation is compact but at times it is clearer to write products of matrices
as summations of products of vectors For example, the OLS estimator can be lently written in either of the following ways:
equiva-β= (XX)−1Xy=
N