1. Trang chủ
  2. » Luận Văn - Báo Cáo

Lecture Undergraduate econometrics - Chapter 7: The multiple regression model

72 32 0

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 72
Dung lượng 226,86 KB

Các công cụ chuyển đổi và chỉnh sửa cho tài liệu này

Nội dung

In this chapter, students will be able to understand: Model specification and the data, estimating the parameters of the multiple regression model, sampling properties of the least squares estimator, interval estimation, hypothesis testing for a single coefficient, measuring goodness of fit.

Trang 1

Chapter 7

The Multiple Regression Model

• When we turn an economic model with more than one explanatory variable into its

corresponding statistical model, we refer to it as a multiple regression model

• Most of the results we developed for the simple regression model in Chapters 3–6 can

be extended naturally to this general case There are slight changes in the interpretation of the β parameters, the degrees of freedom for the t-distribution will change, and we will need to modify the assumption concerning the characteristics of

the explanatory (x) variables

Trang 2

7.1 Model Specification and the Data

When we turn an economic model with more than one explanatory variable into its corresponding econometric model, we refer to it as a multiple regression model Most of the results we developed for the simple regression model in Chapter 3–6 can be extended naturally to this general case There are slight changes in the interpretation of the β parameters, the degrees of freedom for the t-distribution will change, and we will need to

modify the assumption concerning the characteristics of the explanatory (x) variables

As an example for introducing and analyzing the multiple regression model, consider a model used to explain total revenue for a fast-food hamburger chain in the San Francisco Bay area We begin with an outline of this model and the questions that we hope it will answer

Trang 3

7.1.1 The Economic Model

• Each week the management of a Bay Area Rapid Food hamburger chain must decide how much money should be spent on advertising their products, and what specials (lower prices) should be introduced for that week

• How does total revenue change as the level of advertising expenditure changes? Does

an increase in advertising expenditure lead to an increase in total revenue? If so, is the increase in total revenue sufficient to justify the increased advertising expenditure?

• Management is also interested in pricing strategy Will reducing prices lead to an increase or decrease in total revenue? If a reduction in price leads only to a small increase in the quantity sold, total revenue will fall (demand is price inelastic); a price reduction that leads to a large increase in quantity sold will produce an increase in total revenue (demand is price elastic) This economic information is essential for effective management

Trang 4

• We initially hypothesize that total revenue, tr, is linearly related to price, p, and advertising expenditure, a Thus the economic model is:

tr = β1 + β2p + β3a (7.1.1)

where tr represents total revenue for a given week, p represents price in that week and

a is the level of advertising expenditure during that week Both tr and a are measured

in terms of thousands of dollars

• Let us assume that management has constructed a single weekly price series, p,

measured in dollars and cents, that describes overall prices

• The remaining items in Equation (7.1.1) are the unknown parameters β1, β2 and β3 that

describe the dependence of revenue (tr) on price (p) and advertising (a)

• In the multiple regression model the intercept parameter, β1, is the value of the

Trang 5

value zero In many cases this parameter has no clear economic interpretation, but it is almost always included in the regression model It helps in the overall estimation of the model and in prediction

• The other parameters in the model measure the change in the value of the dependent

variable given a unit change in an explanatory variable, all other variables held

constant For example, in Equation (7.1.1),

β2 = the change in tr ($1000) when p is increased by one unit ($1), and a is held

Trang 6

• The symbol ∂ stands for “partial differentiation.” It means that we calculate the

change in one variable, tr, when the variable p changes, all other factors, a, held

constant

• The sign of β2 could be positive or negative If an increase in price leads to an increase in revenue, then β2 > 0, and the demand for the chain’s products is price inelastic Conversely, a price elastic demand exists if an increase in price leads to a decline in revenue, in which case β2 < 0 Thus, knowledge of the sign of β2 provides

information on the price elasticity of demand The magnitude of β2 measures the amount of the change in revenue for a given price change

• The parameter β3 describes the response of revenue to a change in the level of advertising expenditure That is,

β3 = the change in tr ($1000) when a is increased by one unit ($1000), and p is

Trang 7

• The next step along the road to learning about β1, β2 and β3 is to convert the economic model into an econometric model

Trang 8

7.1.2 The Econometric Model

• The economic model in Equation (7.1.1) describes the expected behavior of many

individual franchises As such we should write it as E(tr) = β1 + β2p + β3a, where E(tr) is the “expected value” of total revenue

• Weekly data for total revenue, price and advertising will not follow an exact linear

relationship The Equation (7.1.1) describes, not a line as in Chapters 3-6, but a plane

• The plane intersects the vertical axis at β1 The parameters β2 and β3 measure the slope of the plane in the directions of the “price axis” and the “advertising axis,” respectively

• Table 7.1 shows representative weekly observations on total revenue, price and advertising expenditure for a hamburger franchise If we plot the data we obtain Figure 7.1 These data do not fall exactly on a plane, but instead resemble a “cloud.”

• To allow for a difference between observable total revenue and the expected value of

Trang 9

all the factors that cause weekly total revenue to differ from its expected value These factors might include the weather, the behavior of competitors, a new Surgeon General’s report on the deadly effects of fat intake, etc

• Denoting the t’th weekly observation by the subscript t, we have

tr t = E(tr) + e t = β1 + β2p t + β3a t + e t (7.1.2)

• The economic model in Equation (7.1.1) describes the average, systematic relationship

between the variables tr, p, and a The expected value E(tr) is the nonrandom, systematic component, to which we add the random error e to determine tr Thus, tr is

a random variable We do not know what the value of weekly total revenue will be until we observe it

• The introduction of the error term, and assumptions about its probability distribution, turn the economic model into the econometric model in Equation (7.1.2) The

Trang 10

econometric model provides a more realistic description of the relationship between the variables, as well as a framework for developing and assessing estimators of the unknown parameters

7.1.2a The General Model

• In a general multiple regression model a dependent variable y t is related to a number of

explanatory variables x t2 , x t3 ,…, x tK through a linear equation that can be written as

y t = β1 + β2x t2 + β3x t3 +…+ βK x tK (7.1.3)

• The coefficients β1, β2,…, βK are unknown parameters The parameter βK measures

the effect of a change in the variable x tK upon the expected value of y t , E(y t), all other variables held constant The parameter β1 is the intercept term The “variable” to which β is attached is x = 1

Trang 11

• The equation for total revenue can be viewed as a special case of Equation (7.1.3)

where K = 3, y t = tr t , x t1 = 1, x t2 = p t and x t3 = a t Thus we rewrite Equation (7.1.2) as

y t = β1 + β2x t2 + β3x t3 + e t (7.1.4)

7.1.2b The Assumptions of the Model

To make the statistical model in Equation (7.1.4) complete, assumptions about the

probability distribution of the random errors, e t, need to be made The assumptions that

we introduce for e t are similar to those introduced for the simple regression model in Chapter 3 They are

1 E[e t] = 0 Each random error has a probability distribution with zero mean Some errors will be positive, some will be negative; over a large number of observations they will average out to zero With this assumption we asset that the average of all the

Trang 12

omitted variables, and any other errors made when specifying the model, is zero Thus,

we are asserting that our model is on average correct

2 var(e t) = σ2 Each random error has a probability distribution with variance σ2 The variance σ2 is an unknown parameter and it measures the uncertainty in the statistical model It is the same for each observation, so that for no observations will the model uncertainty be more, or less, nor is it directly related to any economic variable Errors

with this property are said to be homoskedastic

3 cov(e t , e s) = 0 The covariance between the two random errors corresponding to any two different observations is zero The size of an error for one observation has no bearing on the likely size of an error for another observation Thus, any pair of errors

is uncorrelated

4 We will sometimes further assume that the random errors e t have normal probability

distributions That is, e t ~ N(0, σ2)

Trang 13

Because each observation on the dependent variable y t depends on the random error term

e t , each y t is also a random variable The statistical properties of y t follow from those of

e t These properties are

1 E(y t) = β1 + β2x t2 + β3x t3 The expected (average) value of y t depends on the values of the explanatory variables and the unknown parameters This assumption is equivalent

to E(e t ) = 0 It says that the average value of y t changes for each observation and is

given by the regression function E(y t) = β1 + β2x t2 + β3x t3

2 var(y t ) = var(e t) = σ2 The variance of the probability distribution of y t does not change

with each observation Some observations on y t are not more likely to be further from the regression function than others

3 cov(y t , y s ) = cov(e t , e s) = 0 Any two observations on the dependent variable are uncorrelated

Trang 14

4 We sometimes will assume the values of y t are normally distributed about their mean

That is, y t ~ N[(β1 + β2x t2 + β3x t3), σ2], which is equivalent to assuming that e t ~ N(0,

Trang 15

• The second assumption is that any one of the explanatory variables is not an exact linear function of any of the others This assumption is equivalent to assuming that no variable is redundant As we will see, if this assumption is violated, a condition called

“exact multicollinearity,” the least squares procedure fails

• To summarize then, let us construct a list of the assumptions for the general multiple regression model in Equation (7.1.3), much as we have done in the earlier chapters, to which we can refer as needed:

Assumptions of the Multiple Regression Model

MR1 y t = β1 + β2x t2 + … + βK x tK + e t , t = 1,…,T

MR2 E(y t) = β1 + β2x t2 + … + βK x tK ⇔ E(et) = 0

MR3 var(y t ) = var(e t) = σ2

MR4 cov(y t , y s ) = cov(e t , e s) = 0

Trang 16

MR5 The values of x tK are not random and are not exact linear functions of the

other explanatory variables

MR6 y t ~ N[(β1 + β2x t2 + β3x t3), σ2] ⇔ et ~ N(0, σ2)

Trang 17

7.2 Estimating the Parameters of the Multiple Regression Model

We consider the problem of using the least squares principle to estimate the unknown parameters of the multiple regression model We will discuss estimation in the context of the model in Equation (7.1.4), which is

y t = β1 + β2x t2 + β3x t3 + e t (7.2.1)

7.2.1 Least Squares Estimation Procedure

• With the least squares principle we minimize the sum of squared differences between

the observed values of y t and its expected value E[y t] = β1 + β2x t2 + β3x t3

• Mathematically, we minimize the sum of squares function S(β1, β2, β3), which is a function of the unknown parameters, given the data,

Trang 18

• Given the sample observations y t, minimizing the sum of squares function is a

straightforward exercise in calculus The solutions are the least squares estimates b1,

Trang 19

• The least squares estimates b1, b2 and b3 are:

Trang 20

1 2 2 3 3 1

T

t t t t

T

t t t t

Trang 21

1 2 2 3 3 2 1

2

2

2 1 2 2 2 3 2 3 1

T

t

t t t t t t

Trang 22

1 2 2 3 3 3 1

3

2

3 1 3 2 2 3 3 3 1

T

t

t t t t t t

Trang 25

• These formulas can be used to obtain least squares estimates in the model (7.2.1), whatever the data values are Looked at as a general way to use sample data, the formulas in (7.2.3) are referred to as estimation rules or procedures and are called the

least squares estimators of the unknown parameters

• Since their values are not known until the data are observed and the estimates

calculated, the least squares estimators are random variables

• When applied to a specific sample of data, the rules produce the least squares estimates, which are numeric values

7.2.2 Least Squares Estimates Using Hamburger Chain Data

Table 7.2 contains the output obtained when the EViews computer software is used to estimate β1, β2, and β3 for the hamburger revenue equation For the moment, we are concerned only with the least squares estimates, which, from the equation, are:

Trang 27

So in terms of the original economic variables found in Equation (7.1.1),

tr = − p + a (R7.3)

Based on these results, what can we say?

1 The negative coefficient of p t suggests that demand is price elastic and we estimate that

an increase in price of $1 will lead to a fall in weekly revenue of $6,642 Or, stated positively, a reduction in price of $1 will lead to an increase in revenue of $6,642

2 The coefficient of advertising is positive, and we estimate that an increase in advertising expenditure of $1,000 will lead to an increase in total revenue of $2,984

3 The estimated intercept implies that if both price and advertising expenditure were zero the total revenue earned would be $104,790 This is obviously not correct In this model, as in many others, the intercept is included in the model for mathematical completeness and to improve the model’s predictive ability

Trang 28

• The estimated equation can also be used for prediction Suppose management is interested in predicting total revenue for a price of $2 and an advertising expenditure

of $10,000 This prediction is given by

( ) ( )

ˆ 104.785 6.6419 2 2.9843 10121.34

Trang 29

Remark: A word of caution is in order about interpreting regression results

The negative sign attached to price implies that reducing the price will increase total revenue If taken literally, why should we not keep reducing the price to zero? Obviously that would not keep increasing total revenue This makes the following important point: estimated regression models describe the

relationship between the economic variables for values similar to those found

in the sample data Extrapolating the results to extreme values is generally not

a good idea In general, predicting the value of the dependent variable for values of the explanatory variables far from the sample values invites disaster

Trang 30

7.2.3 Estimation of the Error Variance σ2

To develop an estimation procedure for σ2 we use the least squares residuals, which represent the only sample information we have about the error term values For this parameter we follow the same steps that were outlined in Chapter 4.5

• The least squares residuals for the model in Equation (7.2.1) are:

Trang 31

• In the hamburger chain example we have K = 3 The estimate for our sample of data

Trang 32

7.3 Sampling Properties of the Least Squares Estimator

• In a general context, the least squares estimators (b1, b2, b3) in Equation (7.2.3) are random variables; they take on different values in different samples and their values are unknown until a sample is collected and their values computed

• The sampling properties of a least squares estimator tell us how the estimates vary from sample to sample They provide a basis for assessing the reliability of the estimates

• In Chapter 4 we found that the least squares estimator was unbiased and that there is

no other linear unbiased estimator that has a smaller variance, if the model

assumptions are correct This result remains true for the general multiple regression

model that we are considering in this chapter

Trang 33

The Gauss-Markov Theorem: For the multiple regression model, if

assumptions MR1-MR5 hold, then the least squares estimators are the Best

Linear Unbiased Estimators (BLUE) of the parameters in a multiple regression

model

• If we are able to assume that the errors are normally distributed, then y t will also be a normally distributed random variable

• The least squares estimators will also have normal probability distributions, since they

are linear functions of y t

• If the errors are not normally distributed, then the least squares estimators are

approximately normally distributed in large samples, in which T K is greater than,

perhaps, 50

• These facts are of great importance for the construction of interval estimates and the testing of hypotheses about the parameters of the regression model

Trang 34

7.3.1 The Variances and Covariances of the Least Squares Estimators

• The variances and covariances of the least squares estimates give us information about

the reliability of the estimators b1, b2, and b3

• Since the least squares estimators are unbiased, the smaller their variances the higher

is the probability that they will produce estimates “near” the true parameter values

• For K = 3 we can express the variances and covariances in an algebraic form that

provides useful insights into the behavior of the least squares estimator For example,

we can show that:

Trang 35

σcov( , )

Trang 36

• It is important to understand the factors affecting the variance of b2:

1 The larger σ2 the larger the variance of the least squares estimators This is to be expected since σ2 measures the overall uncertainty in the model specification If σ2 is

large, then data values may be widely spread about the regression function E[y t] = β1 +

β2x t2 + β3x t3 and there is less information in the data about the parameter values If σ2

is small, then data values are compactly spread about the regression function E[y t] = β1+ β2x t2 + β3x t3 and there is more information about what the parameter values might be

2 The larger is the sample size T the smaller the variances A larger value of T means a

larger value of the summation 2

2 2

(x tx )

∑ Since this term appears in the denominator

of Equation (7.3.1), when it large, var(b2) is small This outcome is also an intuitive one; more observations yield more precise parameter estimation

3 The larger the variation in an explanatory variable around its mean [measured in this

Ngày đăng: 02/03/2020, 14:05

TỪ KHÓA LIÊN QUAN

TÀI LIỆU CÙNG NGƯỜI DÙNG

TÀI LIỆU LIÊN QUAN