This project estimates the dynamic consumption function for the country Nigeria from 1969 to 2003. In the study, I focus on the linear dynamic specification in order to discuss the parameter estimates in carry out various tests. The first chapter focus on the consumption function theory and how various economy view various theory to show the relationship between income and consumption and how it can be specified in both linear and non linear form. The next chapter describe the data how it was obtained processed through Eviews and how it was interpreted to fit into the project.
The next chapter is hypothesis testing which is used to describe the significance of the speculation and structure the relationship that determines dependent variables in consumption. The t-test and F test is used to measure the proportion in the dependent variables. it also explain the significant testing where we reject or Do not reject null hypothesis. It further estimates the long run and short run marginal propensity to consume and how they affect consumption. it also explain some test such as Ramsey RESET test, Heteroskadasticity test, LM test, for serial correlation and test for their misspecification.
CHAPTER 2 : LITERATURE REVIEW
This chapter focus on the way the total disposable income and total consumption and savings are divided by a consumer.
The consumption function is a mathematical formula laid out by the famous john Maynard Keynes. The formula was designed shown the relationship between real disposal income and consumer spending. The function is used to calculate the amount of total consumption in an economy. An autonomous consumption is used to make up for consumption that is not influenced by current income and induced consumption that is influenced by the economy’s level of income.
The simple consumption function is shown as the linear function:
Where:
C = total consumption,
c0 = autonomous consumption (c0 > 0),
c1 is the marginal propensity to consume (ie the induced consumption) (0 < c1 < 1), and
Yd = disposable income (income after taxes and transfer payments, or W – T).
Autonomous consumption represents consumption when income is zero. In estimation, this is usually assumed to be positive. The marginal propensity to consume (MPC), on the other hand measures the rate at which consumption is changing when income is changing. In a geometric fashion, the MPC is actually the slope of the consumption function.
The MPC is assumed to be positive. Thus, as income increases, consumption increases. However, Keynes mentioned that the increases (for income and consumption) are not equal. According to him, “ as income increases, consumption increases but not by as much as the increase in income”.
There are four approaches to consumption function and they are:
Absolute income hypothesis (Keynes)
Relative income hypothesis (James Duesenberry 1949)
Permanent income hypothesis (Friedman)
Life Cycle hypothesis (Ando&Modigliani)
Absolute Income Hypothesis In economics was proposed by English economist John Maynard Keynes (1883-1946), and has been refined extensively during the 1960s and 1970s, notably by American economist James Tobin (1918-2002).
The theory examines the relationship between income and consumption, and asserts that the consumption level of a household depends not on its relative income but on its absolute level of income. As income rises, the theory asserts, consumption will also rise but not necessarily at the same rate.
Developed by James Stemble Duesenberry, the relative income hypothesis states that an individual’s attitude to consumption and saving is dictated more by his income in relation to others than by abstract standard of living. So an individual is less concerned with absolute level of consumption than by relative levels. The percentage of income consumed by an individual depends on his percentile position within the income distribution.
Secondly it hypothesizes that that present consumption is not influenced merely by present levels of absolute and relative income, but also by levels of consumption attained in previous period. It is difficult for a family to reduce a level of consumption once attained. The aggregate ratio of consumption to income is assumed to depend on the level of present income relative to past peak income.
CD = a = bYt
The current level of consumption is a straightforward function, driven by the current level of income. This implies that people adapt instantaneously to income changes.
– There is rapid adaptation to income changes
– The elasticity of consumption to current income changes
The elasticity with respect to current income in other theories will be less. They reduce the sensitivity to current income flows.
Permanent income hypothesis (PIH) This theory of consumption that was developed by the American economist Milton Friedman. In its simplest form, the hypothesis states that the choices made by consumers regarding their consumption patterns are determined not by current income but by their longer-term income expectations. The key conclusion of this theory is that transitory, short-term changes in income have little effect on consumer spending behavior.
Measured income and measured consumption contain a permanent (anticipated and planned) element and a transitory (windfall gain/unexpected) element. Friedman concluded that the individual will consume a constant proportion of his/her permanent income; and that low income earners have a higher propensity to consume; and high income earners have a higher transitory element to their income and a lower than average propensity to consume.
In Friedman’s permanent income hypothesis model, the key determinant of consumption is an individual’s real wealth, not his current real disposable income. Permanent income is determined by a consumer’s assets; both physical (shares, bonds, property) and human (education and experience). These influence the consumer’s ability to earn income. The consumer can then make an estimation of anticipated lifetime income. He also explains why there was no drop collapse in spending post WWI. Friedman argues that it would be more sensible for people to use current income, but also at the same time to form expectations about future levels of income and the relative amounts of risk.
Thus, they are forming an analysis of “ permanent income.”
Permanent Income = Past Income + Expected Future Income
Transitory Income – income that is earned in excess of, or perceived as an unexpected windfall. If you get income not equal to what you expected or to what you don’t expect to get again.
So, he argues that we tend to spend more out of permanent income than out of transitory.
In the Friedman analysis, he treats people as forming their level of expected future income based on their past incomes. This is known as adaptive expectations.
Adaptive Expectations – looking forward in time using past expectations. In this case, we use a distributed lag of past income.
YPt+1 ® E(Yt+1) = B0Yt + B1Yt-1 + B2Yt-2…
Where B0 > B1 > B2
It is also possible to add a constraint: B0 + B1 + B2 + B3 + … Bn = 1 this is expected income, the actual income can be thought of as: Yt-1 – Ypt+1 = Ytt+1 Using this, we can construct a new model of the consumption function: Ct = a = bYDt + cYtt
There are other factors that people can look at to think about future levels of income. For example, people can think about future interest rates and their effect on their income stream.
The Relative Income Hypothesis The Duesenberry approach says that people are not just concerned about absolute levels of possession. They are in fact concerned about their possessions relative to others. People are not necessarily happier if they have more money. They do however report higher happiness if they have more relative to others. The new utility function would be:
Current economists still support this idea. Ex: Robert Frank and Juliet Schor
Duesenberry argues that we have a greater tendency to resist spending decreases relative to falls in income than we do to increase expenditure relative to increases of income. The reason is that we don’t want to alter our standard of living downard.
CT = a +bYT + cYX
YX is the previous peak level of income (this keeps expenditure from falling in the face of income drops). It is also known as the Drag Effect.
A shift in expenditures relative to a previous level of income is known as the Ratchet Effect, and will be shown below.
Duesenberry argues that we will shift the curve up or move along the curve, but not we will resist shifts down. When WWII ended, a significant number of economists claimed that there would be a consumption decline and aggregate demand drop which did not occur. This provides supporting evidence.
A long-run consumption function can be drawn, assuming that there is a growth trend. If this is true, previous peak income would have been that of last year and thus would give a consumption function that looks like it depends on current income.
The Life Cycle Hypothesis
This is primarily attributed to Ando and Modigliani
The basic notion is that consumption spending will be smooth in the face of an erratic stream of income
Working Phase:
Maintain current consumption, pay off debt from youth years
Maintain current consumption, build up reserves
Age distribution now matters when we look at consumption, and in general, the propensity to consume. Debt and wealth are also taken into account when we look at the propensity to consume. The dependence structure of the population will affect or influence consumption patterns.
Lester Thurow (1976) – argued that this model doesn’t work because it doesn’t presume there is any motive for building wealth other than consumption. Thurow argues that their real motivation is status and power (both internal and external to the family). The permanent income hypothesis bears a resemblance to the life-cycle hypothesis in that in some sense, in both hypotheses, the individuals must behave as if they have some sense of the future.
CHAPTER 3: DATA
This chapter explain the data used in this project. The country used here is Nigeria and the data is downloaded from the IMF International Financial Statistics. The country table is chosen from Annual IFS series via beyond 20/20 WDS. From the Nigeria Annual IFS series the following variables were selected: 96F. CZF HOUSEH. CONS. EXPEND. INCL. NPISHS SA (Units: National Currency)
(Scale: Billions) = NCONS
64…ZF CPI: ALL ITEMS (2000= 100) (Units: Index Number) = CPI
99I. CZF GROSS NATIONAL DISPOSABLE INCOME SA (Units: National Currency)
(Scale: Billions) = NYD
99BIRZF GDP DEFLATOR (2000= 100) (Units: Index Number) = GDPdef ns) = NCONS
64…ZF CPI: ALL ITEMS (2000= 100) (Units: Index Number) = CPI
99I. CZF GROSS NATIONAL DISPOSABLE INCOME SA (Units: National Currency)
(Scale: Billions) = NYD
99BIRZF GDP DEFLATOR (2000= 100) (Units: Index Number) = GDPdef
Then click on show tables, from Download Icon, select Microsoft excel format.
From Excel, I did convert nominal series to real series by using a method called deflating by price Index.
Rcons, which is the Real Consumption expenditure series for USA, is obtain by dividing the NCONS (96F. CZF…): Household consumption expenditure by CPI (64…ZF)
Rcons
RYD is the Real Disposable Income for USA, which is obtain by deflating the Nominal Disposable Income (NYD) 99I. CZF by the GDP deflator (99BIRZF) as,
RYD=
Then the results are saved in a new spreadsheet and are then sent in EViews.
CHAPTER 4: ESTIMATED RESULT AND INTERPRETATION
This chapter reports the ordinary least squares result used for Nigeria between 1969 and 2003 specified both linear and log linear forms.
Interpretation of the intercept is explained as when the data series lies far from the origin of the X, Y plot of the variable, estimates of the will not have a meaningful economy theory interpretation.
Marginal propensity and Elasticities.
————- Linear specification
Estimated coefficient are marginal propensities, the proportion of a (N) income that will spent on consumption.
Where N is the unit of measurement
———— Log linear specification.
The estimated coefficient from a linear specification are not marginal propensities, they are elasticities because they measure the responsiveness of the independent variable changes in the independent variables. Thus, elasticities of 0. 70 will indicate that 70% of change in independent variable will increase the dependent by 70%.
Elasticity is very good because it tends to relate to range while the propensities tends to relate to the average.
Reporting of dynamic linear specification:
Dynamic linear specification: rcons c ryd rcons(-1)):
Dynamic linear equation:
Report of dynamic linear specification:
{4. 506} {0. 20} {0. 74} [Coefficient]
{5. 375} {0. 14} {0. 18} [Std. Error]
{0. 838} {1. 38} {4. 10} [t-Statistic]
{0. 408} {0. 18} {0. 00} [Prob]
Reporting of dynamic log linear specification:
Dynamic log linear specification: log(rcons) c log(ryd) log(rcons(-1))
Dynamic log linear equation:
Report of dynamic log linear specification
{0. 951} {0. 163} {0. 598} [Coefficient]
{0. 380} {0. 097} {0. 164} [Std. Error]
{2. 501} {1. 677} {3. 638} [t-Statistic]
{0. 018} {0. 104} {0. 001} [Prob]
Testing that the MPC of coefficients of Ryd and and Zero.
4. 506 + 0. 20 Ryd + 0. 74
{5. 375} {0. 14} {0. 18}
H0 : β= 0
H1 : β ≠0
tstatistics= == 1. 43
tcrit = tn-k, α/2
Where,
n= 34
k= 3,
= 5%
t= 34-3, 0. 05/2 = t 31, 0. 025
FIGURE 1. 5
0. 025
0. 025
We will therefore not reject the null hypothesis that MPC coefficient of Ryd () is equal to zero because 1. 432. 042 which also means that ordinary least squares of is not statistically significant to zero.
4. 506 + 0. 20 Ryd + 0. 74
{5. 375} {0. 14} {0. 18}
H0 : = 0
H1 : ≠0
T-statistics= == 4. 11
tcrit = tn-k, α/2
Where n= 34, k= 3,= 5%
t= 34-3, 0. 05/2= t 31, 0. 025
We will therefore reject the null hypothesis that MPC of the coefficient of Ryd () is equal to zero because 4. 112. 042 which also means that ordinary least squares of is statistically significant to zero.
Testing that the Elasticities of coefficients of Ryd and and Zero.
H0 : β= 0
H1 : β ≠0
t-statistics= == 1. 43
tcrit = tn-k, α/2
where n= 34, k= 3,= 5%
t= 34-3, 0. 05/2= t 30, 0. 025
0. 95
0. 025
0. 025
We will therefore not reject the null hypothesis that MPC coefficient of Ryd () is equal to zero because 1. 432. 042 which also means that ordinary least squares of is not statistically significant to zero.
4. 506 + 0. 20 Ryd + 0. 74
{5. 375} {0. 14} {0. 18}
H0 : = 0
H1 : ≠0
T-statistics= == 4. 11
tcrit = tn-k, α/2
Where n= 34, k= 3,= 5%
t= 34-3, 0. 05/2= t 30, 0. 025
CHAPTER 5: TESTING THE SIGNIFICANCE OF THE ESTIMATED
PARAMETER.
A statistical hypothesis test is a method of making statistical decisions using experimental data. In statistics, a result is called statistically significant if it is unlikely to have occurred by chance. The phrase “ test of significance” was coined by Ronald Fisher: “ Critical tests of this kind may be called tests of significance, and when such tests are available we may discover whether a second sample is or is not significantly different from the first.
Hypothesis testing is sometimes called confirmatory data analysis, in contrast to exploratory data analysis. In frequency probability, these decisions are almost always made using null-hypothesis tests; that is, ones that answer the question Assuming that the null hypothesis is true, what is the probability of observing a value for the test statistic that is at least as extreme as the value that was actually observed? One use of hypothesis testing is deciding whether experimental results contain enough information to cast doubt on conventional wisdom.
Statistical hypothesis testing is a key technique of frequents statistical inference, and is widely used, but also much criticized. The main direct alternative to statistical hypothesis testing is Bayesian inference. However, other approaches to reaching a decision based on data are available via decision theory and optimal decisions.
The critical region of a hypothesis test is the set of all outcomes which, if they occur, will lead us to decide that there is a difference. That is, cause the null hypothesis to be rejected in favor of the alternative hypothesis. The critical region is usually denoted by C.
Null Hypothesis is a phrase that was originally coined by English geneticist and statistician Ronald Fisher. In statistical hypothesis testing, the null hypothesis (H0) formally describes some aspect of the statistical “ behaviour” of a set of data. This description is assumed to be valid unless the actual behaviour of the data contradicts this assumption. Thus, the null hypothesis is contrasted against another or alternative hypothesis. Statistical hypothesis testing, which involves a number of steps, is used to decide whether the data contradicts the null hypothesis. This is called significance testing. A null hypothesis is never proven by such methods, as the absence of evidence against the null hypothesis does not establish its truth. In other words, one may either reject, or not reject the null hypothesis; one cannot accept it. This means that one cannot make decisions or draw conclusions that assume the truth of the null hypothesis. Just as failing to reject it does not “ prove” the null hypothesis, one does not conclude that the alternative hypothesis is disproven or rejected, even though this seems reasonable. One simply concludes that the null hypothesis is not rejected. Not rejecting the null hypothesis still allows for getting new data to test the alternative hypothesis again. On the other hand, rejecting the null hypothesis only means that the alternative hypothesis may be true, pending further testing.
Dynamic linear equation:
Test the overall significance of Dynamic linear regression model consumption function.
4. 506 + 0. 20 Ryd + 0. 74
{5. 375} {0. 14} {0. 18}
R2= 0. 70
N= 34
Ho: β = γ = 0
H1: β ≠γ ≠0
0. 05
0. 95
We will therefore reject the null hypothesis that MPC coefficients of Ryd () and () is equal to zero because Fs> Fc (36. 16 > 3. 32) which also means that ordinary least squares of is statistically significant, that is, it is not zero.
Intercept is not statistically significant at the 5% level in any case; in estimation the intercept has no economic meaning.
Interpretation — scale — factor.
The coefficient on RYD and the lagged dependent variable (LDV) statistically significant at 5%
The Regression on coefficient of the determination R2
R2 is a measure of the proportion of the variation in the dependent variable that is explained by the entire variable in the equation.
In statistics, the coefficient of determination, R2 is used in the context of statistical models whose main purpose is the prediction of future outcomes on the basis of other related information. It is the proportion of variability in a data set that is accounted for by the statistical model. It provides a measure of how well future outcomes are likely to be predicted by the model.
There are several different definitions of R2 which are only sometimes equivalent. One class of such cases includes that of linear regression. In this case, R2 is simply the square of the sample correlation coefficient between the outcomes and their predicted values, or in the case of simple linear regression, between the outcome and the values being used for prediction. In such cases, the values vary from 0 to 1. Important cases where the computational definition of R2 can yield negative values, depending on the definition used, arise where the predictions which are being compared to the corresponding outcome have not derived from a model-fitting procedure using those data.
R2 is expressed as follows:
R2 = ESS/TSS = 1 – RSS/TSS = 1 -∑ e2i /∑ (Yi – Y-) 2
Where; ESS= Explained sum of squares
TSS= Total sum of squares
RSS= Residuals sum of squares
Therefore R2 in dynamic linear regression model is explained as follows.
R2 is the coefficient of determination. It is also used to measure to measure the goodness of fit. It is also used to tell us how close the data point are to the fitted function that is 0 < R2 < 1. More precisely 0. 70 in dynamic linear regression model consumption function tell us the 70% of the variability in Rconst can be explained by the variability in the explanatory variables(Ryd and Rconst-1).
ESTIMATION OF LONG RUN AND SHORT RUN MARGINAL PROPENSITY TO CONSUME
The two different period of MPC. The first one is the Short Run Marginal consume and it shows us the marginal propensity to consume. This shows the marginal propensity to consume for one period of time.
This simple indicates that a 1 unit change in disposable income would have on consumption in the same period.
The second shows in the Long Run Marginal Propensity to Consume, it takes into consideration recent consumption behaviour, as well as disposable income when determining the level of consumption.
By taking into consideration the previous consumption, and the current income, it allows you to assess what effect the previous could probably have on consumption.
Long run =
Long Run is obtained through the use of steady state consumption. We can assume that
it allows dynamic equation to be written as
Taking to Left Hand Side we get
When C is factored out it is written as:
When solving for C we obtain
C =
So using the SSQ we get the Long Run where C is the consumption and is being determine by the Long run Marginal Propensity to Consume of the income variable.
CHAPTER 6: THE RAMSEY RESET TEST
Ramsey Regression Equation Specification Error Test (RESET) test (Ramsey, 1969) is a general specification test for the linear regression model. More specifically, it tests whether non-linear combinations of the estimated values help explain the endogenous variable. The intuition behind the test is that, if non-linear combinations of the explanatory variables have any power in explaining the endogenous variable, then the model is mis-specified. The RESET test is designed to detect omitted variables and incorrect functional form.
The Ramsey test then tests whether (β1x)2,(β2x)3…,(βk − 1x)k has any power in explaining y. This is executed by estimating the following and then testing, by a means of a F-test whether through are zero. If the null-hypothesis that all regression coefficients of the non-linear terms are zero is rejected, then the model suffers from mis-specification.
For a univariate x the test can also be performed by regressing on the truncated power series of the explanatory variable and using an F-Test for
Test rejection implies the same insight as the first version mentioned above.
The F-test compares regressions, the original one and the Ramsey’s auxiliary one, as done with the evaluation of linear restrictions. The original model is the restricted one opposed to the Ramsey’s unrestricted model.
F(k − 1, n − k), where:
n is the sample size;
k is number of parameters in the in the Ramsey’s model.
Furthermore, the linear model and the model with the non-linear power terms are subjected to the F-test, similarly as before:
~ F(k − 1, n − m − k) ,
where m + k is number of parameters in the Ramsey’s model, which are k − 1 variables in the Ramsey group (non-linear )
plus m + 1 the number of parameters in the original model
Test for misspecification.
Rejection of Ho implies the original model is inadequate and can be improved. A failure to reject H0 says the test has not been able to detect any misspecification.
Overall, the general philosophy of the test is: If we can significantly improve the model by artificially including powers of the predictions of the model, then the original model must have been inadequate.
€
Estimating this model, and then augmenting it with squares of the predictions, and squares and cubes of then predictions, yields the RESET test results The F-values are quite small and their corresponding p-values of 0. 93 and 0. 70 are well above the conventional significance level of 0. 05. There is no evidence from the RESET test to suggest the log-log model is inadequate.
CHAPTER 7: BREUSCH- GODFERYSERIAL CORRELATION LM TEST
The Breusch – Godfery Lagrange Multiplier (BGLM) Test. computes the Breusch (1978)-Godfrey (1978) Lagrange multiplier test for no independence in the error distribution. For a specified number of lags p, the test’s null of independent errors has alternatives. The test statistic, a T R^2 measure, is distributed Chi-squared (p) under the null hypothesis.
In statistics, the BG serial correlation LM test is a big test for autocorrelation in the residuals from a regression analysis and is considered more general than the standard Durbin-Watson statistic. The null hypothesis is that there is no serial correlation of any order up to p. un like the Durbin-Watson h statistic is only valid for nonstochastic regressors and first-order autoregressive schemes, the BG test has none of these restrictions, and is statistically more powerful than Durbin’s h statistic.
Characteristics of Breusch-Godfrey Test:
Allows for a relationship between ut and several of its lags
Estimate a regression and obtain residuals
Regress residuals on all regressors and lagged residuals
Obtain R2 from this regression
Letting T denote the number of observations.
The problem of time series is often a problem when the stochastic term in one period is not the independent in another.
Serial correlation does not affect the unbiasedness of the properties of ordinary least squared but it does affect the minimum variance hence, the estimated efficiency.
The LM test uses the form that in order to see if the estimated residual from the restricted forms is related to the lagged values of itself then
If serial correlation does not exist, the then the regresses of the unrestricted form and restricted form are the same.
There two forms of the test reported and they are:
(1) Chi-squared (£2): the test statistics is calculated as TR2, where T is the number of observations in the original regression and R2 is the R-squared in the auxiliary regression. This has a£2 (h) distribution for h restrictions (lags). So where h= 5.
If TR2 < (£2): (h) 0. 05 then we do not reject the null of no autocorrelation (at the 5% Significance level).
If TR2 > £2 (h) 0. 05 then we must reject the null of no autocorrelation
(At the 5% Significance level).
(2) F Test
The test statistic is calculated as Fcal = (T-k-1-h) R2/h (1-R2), where k is the number of regressors in the original equation (here k = 3). This has an F (h, T-k-1-h) distribution.
If Fcal < F tables, 0. 05 then we do not reject the null of no autocorrelation (at the 5%
Significance level).
If Fcal > F tables, 0. 05 then we must reject the null hypothesis of no autocorrelation (at the 5% significance level)
The Chi- squared is going to be used in carrying out the test below
CHAPTER 8: HETEROSKADASTICITY TEST.
.
HETEROSKEDASTICITY
One of the assumptions of the classical linear regressions (CLR) model is that the error term in the linear model is drawn from a distribution with a constant variance when this is the case, it is said that the errors are homoskedastic. In a situation where this assumption does not hold, then the problem of heteroskedasticity occurs.
Heteroskedasticity, as a violation of the assumptions of the CLR model, causes the OLS estimates to lose some of their nice model.
Heteroskadasticity is likely to take place in a cross- sectional model than time series model.
Heteroskedasticity causes OLS to underestimate the variances and standard errors of the estimated coefficients
This implies that the t-test and F-test are not reliable
The t-statistics tend to be higher leading us to reject a null hypothesis that should not be rejected
F-statistic follows an F distribution with k degrees of freedom in the numerator and (n – k -1) degrees of freedom in the denominator
Reject the null hypothesis that there exists no heteroskedasticity if the F-statistic is greater than the critical F-value at the selected level of significance.
If the null cannot be rejected, then there exists heteroskedasticity in the data and an alternative estimation method to OLS must be followed.
Testing for heteroskadasticity
When testing for heteroskadasticity, there different ways to go because heteroskadasticy takes a number of different forms and it exact manifestation in a given equation is almost known.
The main focus will be on the white test which is more generally used than the other tests.
The white test is used to detect the heteroskadasticity by running a regression with the squared residuals as the dependent variable. In the right hand side of the second equation all the original independent variables, the squared of all the original independent variables with each other and the cross product of the entire original test with each other. The white test run the advantage of not assuming any particular form of heteroskadasticity which make it popular