Skip to main content

The General Linear Model IV

  • Chapter
  • First Online:
Book cover Introductory Econometrics
  • 3470 Accesses

Abstract

In this chapter we take up the problems occasioned by the failure of the rank condition (for the matrix of explanatory variables). This problem arises as a matter of course in analysis of variance (or covariance) models where some of the variables are classificatory. In this case, we are led to the construction of “dummy” variables representing the classificatory schemes. Since all such classificatory schemes are exhaustive, it is not surprising that the “dummy” variables are linearly dependent and, thus, the rank condition for the data matrix fails.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 69.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 89.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 119.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    The term is actually a misnomer. Two linearly dependent variables are said to be collinear since they lie on the same line. Thus if α 1 x 1 + α 2 x 2 = 0 then we have x 2 =  − (α 1/α 2)x 1 for α 2 ≠ 0. Three or more linearly dependent variables, however, lie on the same plane and this case is more properly referred to as coplanarity.

  2. 2.

    For full comprehension of the discussion in the first part of this chapter the reader is well advised to master Sect. 3 of Mathematics for Econometrics.

  3. 3.

    A more satisfactory approach may be to consider the ratio between the largest and smallest characteristic root of the correlation matrix; if this ratio exceeds, say, 500 we may conclude that the coefficient estimates obtained by the regression are, numerically, unreliable.

  4. 4.

    Thank you to Professor David Hendry reminding the author of this point. Moreover, Professor Hendry reminds us that collinearity is a property of a specific parameterization of the model. Hendry adds a variable in his DHSY model, see Chap. 16 of Hendry and Doornik [183]. Variable selection can benefit by (deliberately) creating perfect collinearities (in special cases).

  5. 5.

    Let \( \widehat{\theta} \) be an estimator of a parameter vector θ. The MSE matrix of \( \widehat{\theta} \) is defined by

    $$ \mathrm{MSE}\left(\widehat{\theta}\right)=\mathrm{E}\left(\widehat{\theta}-\theta \right){\left(\widehat{\theta}-\theta \right)}^{\prime } $$

    If \( \mathrm{E}\left(\widehat{\theta}\right)=\overline{\theta} \) we note that \( \mathrm{MSE}\left(\widehat{\theta}\right)=\mathrm{Cov}\left(\widehat{\theta}\right)+\left(\overline{\theta}-\theta \right){\left(\overline{\theta}-\theta \right)}^{\prime } \), so that the MSE matrix is the sum of the covariance and bias matrices of an estimator.

  6. 6.

    In July 1983, The Lincoln Center campus of Fordham University hosted a symposium on ridge regression, emphasizing RR and multicollinearity. The papers were published as a special issue of Communications in Statistics 13 (1984). Guerard and Horton [149, 150] used ridge regression to address issues of multicollinearity in executive compensation; compensation being a function of sales, assets, and profits of companies.

  7. 7.

    It is necessary to impose the restriction in (4.46) only if we wish to reconcile the models in (4.44) and (4.45). On the other hand the model in (4.46) could be postulated ab initio, in which case such restriction need not be imposed.

  8. 8.

    Chan et al. [50, 51] used seemingly unrelated regression (SUR) to model CAPM monthly excess returns as functions of CAPM excess returns of the value-weighted or equal-weighted market index return; EP, BP, CP; size as measured by the natural logarithm of market capitalization (LS). Chan et al. [50, 51] define cash as the sum of earnings and depreciation without explicit correction for other noncash revenue or expenses. Betas were simultaneously estimated and cross-sectional correlations of residuals were addressed. When fractile portfolios were constructed by sorting on the EP ratio, the highest EP quintile portfolio outperformed the lowest EP quintile portfolio, and the EP effect was not statistically significant. The highest BP stocks outperformed the lowest BP stocks. The portfolios composed (sorted) of the highest BP and CP outperformed the portfolios composed of the lowest BP and CP stocks. In the authors’ multiple regressions, the size and book-to-market variables were positive and statistically significant. The EP coefficient was negative and statistically significant at the 10% level. Thus, no support was found for the Graham and Dodd low PE approach. In the monthly univariate SUR analysis, with each month variable being deflated by an annual (June) cross-sectional mean, Chan et al. [50, 51] found that the EP coefficient was negative (but not statistically significant), the size coefficient was negative (but not statistically significant), the book to market coefficient was positive and statistically significant, and the cash flow coefficient was positive and statistically significant. In their multiple regressions, Chan et al. [50, 51] report BP and CP variables were positive and statistically significant but EP was not significant. Applying an adaptation of the Fama-MacBeth [112] time series of portfolio cross sections to the Japanese market produced negative and statistically significant coefficients on EP and size but positive and statistically significant coefficients for the BP and CP variables. Chan et al. [50, 51] summarized their findings: “The performance of the book to market ratio is especially noteworthy; this variable is the most important of the four variables investigated.”

  9. 9.

    In 1991, Markowitz headed the Daiwa Securities Trust Global Portfolio Research Department (GPRD). Markowitz hired John Guerard, formerly of Drexel, Burnham, Lambert, to build expected returns models. The Markowitz team estimated stock selection models using Graham and Dodd [137] fundamental valuation variables, earnings, book value, cash flow and sales, relative variables, defined as the ratio of the absolute fundamental variable ratios divided by the 60-month averages of the fundamental variables. Bloch et al. [34] reported a set of some 200 simulations of United States and Japanese equity models. Bloch et al. [34] found that Markowitz [236] mean-variance efficient portfolios using the lower EP values in Japan underperformed the universe benchmark, whereas BP, CP, and SP (sales-to-price, or sales yield) variables outperformed the universe benchmark.

  10. 10.

    The use of non-financial stocks led to a customized index for the Markowitz Global Portfolio Research Group (GPRD) analysis. The Chan et al. and an initial Guerard presentation occurred in September 1991 at the Berkeley Program in Finance, Santa Barbara, on Fundamental Analysis. Bill Ziemba presented a very interesting study comparing U.S. and Japanese fundamental strategies at the same Berkeley Program meeting. Markowitz refers to this meeting in his Nobel Prize lecture (1991). Ziemba used capitalization-weighted regressions. The Chan et al., Guerard, and Ziemba studies found statistical significance with expectation and reported fundamental data.

  11. 11.

    In a thorough assessment of value versus growth in Japan and the United States, Lakonishok et al. [207] examined the intersection of Compustat and CRSP databases for annual portfolios for NYSE and AMEX common stocks, April 1963 to April 1990. Their value measures were three current value ratios: EP, BP and CP. Their growth measure was the five-year average annual sales growth (GS). They performed three types of tests: a univariate ranking into annual decile portfolios for each of the four variables, bivariate rankings on CP (value) and GS (growth, glamour), and finally a multivariate regression adaptation of the Fama and MacBeth [71] time series pooling of cross sectional regressions. Lakonishok et al. [139] used the Fama-MacBeth methodology to construct portfolios and pool (average over time) a time series of twenty-two one-year cross-sectional univariate regressions for each of the 22 years in their study period. The univariate regression coefficient for SG was significantly negative. The EP, BP, and CP coefficients were all significantly positive. When Lakonishok, Shleifer, and Vishny performed a multivariate regression using all four variables, they found significantly positive coefficients for BP and EP (but not CP) and significantly negative coefficients for SG. Overall, Lakonishok et al. [207] concluded that buying out-of-favor value stocks outperformed growth (glamour) over the April 1968 to April 1990 period, that future growth was difficult to predict from past growth alone and that the actual future growth of the glamour stocks was much lower than past growth relative to the growth of value stocks, and that the value strategies ex post were not significantly riskier than growth (glamour) strategies.

  12. 12.

    Haugen and Baker [177] extended their 1996 study in a recent volume to honor Harry Markowitz [158]. Haugen and Baker estimate their model using weighted least squares. In a given month we will simultaneously estimate the payoffs to a variety of firm and stock characteristics using a weighted least squares multiple regression procedure of the following form:

    $$ {\mathrm{r}}_{\mathrm{j},\kern0.5em \mathrm{t}}=\sum \limits_{\mathrm{i}=1}^{\mathrm{n}}{\mathrm{p}}_{\mathrm{i},\kern0.5em \mathrm{t}}\kern0.5em {\mathrm{F}}_{\mathrm{i},\mathrm{j},\mathrm{t} - 1}+{\upmu}_{\mathrm{j},\mathrm{t}} $$
    (4.55)

    where:

    r j,t = the total rate of return to stock j in month t.

    Pi,t = estimated weighted least squares regression coefficient (payoff) for factor i in month t.

    Fi,j,t-1 = normalized value for factor i for stock j at the end of month t-1.

    n = the number of factors in the expected return factor model.

    μj,t = component of total monthly return for stock j in month t unexplained by the set of

    factors.

    Haugen and Baker [177] estimated their equation (2) in each month in the period 1963 through 2007. In the manner of Fama and MacBeth [71], they then compute the average values for the monthly regression coefficients (payoffs) across the entire period. Dividing the mean payoffs by their standard errors we obtain t-statistics. The most significant factors are computed as follows:

    • Residual Return is last month’s residual stock return unexplained by the market.

    • Cash Flow-to-Price is the12-month trailing cash flow-per-share divided by the current price.

    • Earnings-to-Price is the 12-month trailing earnings-per-share divided by the current price.

    • Return on Assets is the12-month trailing total income divided by the most recently reported total assets.

    • Residual Risk is the trailing variance of residual stock return unexplained by market return.

    • 12-month Return is the total return for the stock over the trailing 12 months.

    • Return on Equity is the12-month trailing earnings-per-share divided by the most recently reported book equity.

    • Volatility is the 24-month trailing volatility of total stock return.

    • Book-to-Price is the most recently reported book value of equity divided by the current market price.

    • Profit Margin is 12-month trailing earnings before interest divided by 12-month trailing sales.

    • Three-month Return is the total return for the stock over the trailing 3 months.

    • Sales-to-Price is 12-month trailing sales-per-share divided by the market price.

    Haugen and Baker noted that the t-scores are large as compared to those obtained by Fama and MacBeth even though the length of the time periods covered by the studies is comparable. Last month’s residual return and the return over the preceding 3 months have negative predictive power relative to next month’s total return. This may be induced by the fact that the market tends to overreact to most information. The overreaction sets up a tendency for the market to reverse itself upon the receipt of the next piece of related information.

    The four measures of cheapness: cash-to-price, earnings-to-price, book-to-price, and sales-to-price, all have significant positive payoffs. Haugen and Baker [177] find statistically significant results for the four fundamental factors as did the previously studies we reviewed. Measures of cheapness have been frequently found in the past to be associated with relatively high stock returns, so it is not surprising that four measures of cheapness appear here as significant determinants of structure in the cross-section. Haugen and Baker [177] dismiss the problem of multicollinearity. Haugen and Baker present optimization analysis to support their stock selection modeling, and portfolio trading is controlled through a penalty function. When available, the optimizations are based on the largest 1000 stocks in the database. Estimates of portfolio volatility are based on the full covariance matrix of returns to the 1000 stocks in the previous 24 months. Two years of monthly return data, from 1963 through 1964, is used to construct the initial portfolios. Estimates of expected returns to the 1000 stocks are based on the factor model discussed above. The following constraints are applied to portfolio weights for each quarterly optimization:

    1. 1.

      The maximum weight in a portfolio that can be assigned to a single stock is limited to 5%. The minimum is 0% (Short selling is not permitted).

    2. 2.

      The maximum invested in any one stock in the portfolio is three times the market capitalization weight or 0.25%, whichever is greater, subject to the 5% limit.

    3. 3.

      The portfolio industry weight is restricted to be within 3% of the market capitalization weight of that industry. (Based on the two-digit SIC code.)

    4. 4.

      Turnover in the portfolio is penalized through a linear cost applied to the trading of each stock. As a simplification, all stocks are subject to the same linear turnover cost although in practice portfolio managers use differential trading costs in their optimizations.

    These constraints are designed to merely keep the portfolios diversified. Reasonable changes in the constraints do not materially affect the results. The portfolios are re-optimized quarterly.

    Trading costs are not reflected in the Haugen and Baker [177] optimization analysis; however, the Haugen and Baker [125] portfolios out-performed the benchmark by almost 5 percent with average annual turnover of 80 percent during the 1965–2007 period. Obviously, as Haugen and Baker conclude, transactions costs would have to be unrealistically extreme to significantly close the gap between the high and low expected return portfolios. Haugen and Baker [125] conclude with the following findings: (1) measures of current profitability and cheapness are overwhelmingly significant in determining the structure of the cross-section of stock returns; (2) the statistical significance of risk is also overwhelming, but the payoff to risk has the wrong sign period after period. The riskiest stocks over measures including market beta, total return volatility, and residual volatility tend to have the lowest returns; (3) one-year momentum pays off positively, and that last month’s residual return and last quarter’s total return pays off negatively; (4) strikingly, nearly all of the most significant factors over our total period are highly significant in our five sub-periods, and all have the same signs as they did in the total period; and (5) the ad hoc expected return factor model is very powerful in predicting the future relative returns on stocks.

  13. 13.

    Guerard [159] decomposed the MQ variable into (1) price momentum, (2) the consensus analysts’ forecasts efficiency variable, CIBF, which itself is composed of forecasted earnings yield, EP, revisions, EREV, and direction of revisions, EB, identified as breadth, Wheeler [325], and (3) the stock standard deviation, identified in Cragg and Malkiel [76] as a variable with predictive power regarding the stock price-earnings multiple. Guerard et al. [154] and Guerard and Mark [156] found that the consensus analysts’ forecast variable dominated analysts’ forecasted earnings yield, as measured by I/B/E/S one-year-ahead forecasted earnings yield, FEP, revisions, and breadth. Guerard reported domestic (U.S.) evidence that the predicted earnings yield is incorporated into the stock price through the earnings yield risk index. Moreover, CIBF dominates the historic low price-to-earnings effect, or high earnings-to-price, PE.

  14. 14.

    Cragg and Malkiel [76] created a database of five forecasters of long-term earnings forecasts for 185 companies in 1962 and 1963. These five forecast firms included two New York City banks (trust departments), an investment banker, a mutual fund manager, and the final firm was a broker and an investment advisor. The Cragg and Malkiel [50] forecasts were five-year average annual growth rates. The earnings forecasts were highly correlated with one another, the highest paired correlation was 0.889 (in 1962) and the lowest paired correlation was 0.450 (in 1963) with most correlations exceeding 0.7. They calculated used the Thiel Inequality Coefficient (1966) to measure the efficiency of the financial forecasts and found that the correlations of predicted and realized earnings growth were low, although most were statistically greater than zero. The TICs were large, according to Cragg and Malkiel [50], although they were less than 1.0 (showing better than no-change forecasting). The TICS were lower (better) within sectors; the forecasts in electronics and electric utility firms were best and foods and oils were the worst firms to forecast earnings growth. Elton and Gruber [105] built upon the Cragg and Malkiel study and found similar results. That is, a simple exponentially weighted moving average was a better forecasting model of annual earnings than additive or multiplicative exponential smoothing models with trend or regression models using time as an independent variable. Indeed, a very good model was a naïve model, which assumed a no-change in annual earnings per shares with the exceptional of the prior change had occurred in earnings. One can clearly see the random walk with drift concept of earnings in the Elton and Gruber [105]. Elton and Gruber compared the naïve and time series forecasts to three financial service firms, and found for their 180 firm sample that two of of the three firms were better forecasters than the naïve models. Elton et al. [106] build upon the Cragg and Malkiel [76] and Elton and Gruber [105] results and create an earnings forecasting database that evolves to include over 16,000 companies, the Institutional Brokerage Estimation Services, Inc. (I/B/E/S). Elton et al. [106] find than earnings revisions, more than the earnings forecasts, determine the securities that will out-perform the market. Guerard and Stone [219] found the I/B/E/S consensus forecasts were not statistically different than random walk with drift time series forecasts for 648 firms during the 1982–1985 period. Lim [145], using the I/B/E/S Detailed database from 1984 – December 1996, found forecast bias associated with small and more volatile stocks, experienced poor past stock returns, and had prior negative earnings surprises. Moreover, Lim [145] found that relative bias was negative associated with the size of the number of analysts in the brokerage firm. That is, smaller firms with fewer analysts, often with more stale data, produced more optimistic forecasts. Keane and Runkle [195] found during the 1983–1991 period that analysts’ forecasts were rational, once discretionary special charges are removed. The Keane and Runkle [195] study is one of the very few studies finding rationality of analysts’ forecasts; most find analysts are optimistic. Further work by Wheeler [325] will find that firms where analysts agree with the direction of earnings revisions, denoted breadth, will out-perform stocks with lesser agreement of earnings revisions. Guerard et al. [154] combined the work of Elton et al. [105] and Wheeler [325] to create a better earnings forecasting model, CTEF, which we use in this analysis. The CTEF variable continues to produce statistically significant excess return in backtest and in identifying real-time security mispricing, see Guerard [159].

  15. 15.

    The ICs on the analysts’ forecast variable, CTEF, and price momentum variable, PM, were lower than in the United States universe, reported in Guerard et al. [160].

  16. 16.

    A subset of ICs for the 1997–2011 and 2003–2011 time periods, for the GLER data, were reported in Guerard, Markowitz, and Xu (2013).

  17. 17.

    See Markowitz [232,233,234,235,236], Chap. 9.

  18. 18.

    Harry Markowitz reminds readers that he discussed the possibility of looking at security returns relative to index returns in Chap. 4, footnote 1, page 100, of Portfolio Selection (1959).

  19. 19.

    See Chap. 2 of Guerard (2010) for a history of multi-index and multi-factor risk models.

  20. 20.

    Blin and Bender estimated their APT factor model using principal components analysis, PCA. The Axioma Statistical Risk Model, World-Wide Equity Risk Factor Model, AX-WW2.1, estimates 15 principal components to measure risk. See Guerard, Markowitz, and Xu [162] for a comparison of Axioma Fundamental and statistically based risk models. Guerard et al. (2014) reported that the statistical model dominated the fundamental risk model in producing a higher set of returns for a given level of risk. Guerard (2012) reported statistically-based portfolio results using the Sungard APT risk model, which uses principal components in its estimation. The Sunguard Model was originally built by John Blin and Steve Bender. Blin, Bender and Guerard [33] used a 20-factor beta model of variance and covariances based on 3.5 years of weekly stock returns data. The Blin and Bender Arbitrage Pricing Theory (APT) model followed the Roll factor theory, but Blin and Bender estimated at least 20 orthogonal factors.

  21. 21.

    Note that when symbols are assigned their proper meaning the first equation below simply states ρ = ρ.

  22. 22.

    Although not perhaps in the Treynor spirit a more useful index might be

    δ = (rp(i) – r0) / (a + b rm*),

    the ratio of excess to market returns, in lieu of the original form

    TI = (rp(i) - r0) / b.

  23. 23.

    In Sharpe [222] it is somewhat incorrectly stated that the Sharpe ratio is equivalent to the Treynor ratio. As is seen from the discussion above noting in particular the modification in footnote 17, this is not true; the Treynor ratio as modified in that footnote is merely the ratio of excess returns divided by market returns, not the standard deviation.

  24. 24.

    The authors are indebted to Vishhu Anand, of Axioma, who ran the Axioma attribution analysis based on the Axioma Fundamental Risk Model.

  25. 25.

    Readers may question the use of a 12-year backtesting period. The USER model was tested with the Barra USE3 Model for the 1980–2009 period and asset selection of 449 basis points, annualized, is reported. The t-value on the USER variable is 4.60, which is highly statistically significant. Stone and Guerard (2010b) found good stock selection returns in the United States and Japan, 1980–2005, using similar models.

  26. 26.

    The Barra attribution model assumes the Bara Fundamental risk Model, USE3, determines risk, The Barra risk model was developed in Rosenberg [270], Rosenberg and Marathe [271] and thoroughly discussed in Rudd and Clasing [279] and Grinold and Kahn [148].

    The Axioma fundamentally-based risk model evolved from the MSCI Barra risk model. The first set is a fundamental risk model, such as the Axioma World-Wide Equity Risk Factor Model (WW21AxiomaMH), which seeks to forecast medium-horizon risk, or risk 3–6 months ahead. The Axioma Fundamental Risk Model uses nine style factors: exchange rate sensitivity, growth (historical earnings and sales growth), leverage (debt-to-assets), liquidity (1 month trading volume divided by market capitalization), medium-term momentum (cumulative returns of the past year, excluding the previous month), short-term momentum (last month return), size (natural logarithm of issuer market capitalization), value (book-to-price and earnings-to-price ratios), and volatility (3 months average of absolute returns divided by cross-sectional standard deviation).

    Statistically-based risk models developed in the works of Roll and Ross [273], Dhrymes et al. [83], Dhrymes et al. [84] and Gultekin et al. [116]. Guerard, Markowitz, Xu [162] tested the effectiveness of composite earnings forecasting model and the GLER model using APT and both Axioma Fundamental and Statistical Risk Models. Guerard et al. (2015) reported that the Axioma Statistical Risk Model produced higher Geometric Means, Sharpe Ratios, and Geometric Means that the Axioma Fundamentl Risk model for both CTEF and GLER Models.

  27. 27.

    Deng and Min [79] tested USER and GLER over identical time periods and simulation conditions, 12/31/1999 – 01/31/2011, and reported that GLER portfolios outperformed USER portfolios, primarily due to Momentum factor returns in global markets. Stock specific returns were very similar in the USER and GLER portfolios.

References

  1. APT Analytics Guide. (2011). SunGard APT, London www.sungard.com/apt/learnmore/

  2. Ashley, R., & Ye, H. (2012). On the granger causality between median inflation and price dispersion. Applied Economics, 44, 4221–4238.

    Article  Google Scholar 

  3. Beaton, A. E., & Tukey, J. W. (1974). The fitting of power series, meaning polynomials, illustrated on Bank-spectroscopic data. Technometrics, 16, 147–185.

    Article  Google Scholar 

  4. Benjamini, Y., & Hochberg, Y. (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society, Series B, 57, 289–300.

    Google Scholar 

  5. Blin, J. M., Bender, S., & Guerard, J. B. Jr. (1997). Earnings Forecasts, Revisions and Momentum in the Estimation of Efficient Market-Neutral Japanese and U.S. Portfolios. In A. Chen (Ed.), Research in Finance, 15.

    Google Scholar 

  6. Bloch, M., Guerard Jr., J. B., Markowitz, H. M., Todd, P., & Xu, G.-L. (1993). A comparison of some aspects of the U.S. and Japanese equity markets. Japan & the World Economy, 5, 3–26.

    Article  Google Scholar 

  7. Brush, J. S., & Boles, K. E. (1988). The predictive power in relative strength and CAPM. Journal of Portfolio Management, 20–23.

    Google Scholar 

  8. Brush, J. S. (2001). Price Momentum: A Twenty-Year Research Effort. Columbine Newsletter Special Issue.

    Google Scholar 

  9. Chan, L. K. C., Hamao, Y., & Lakonishok, J. (1991). Fundamentals and stock returns in Japan. Journal of Finance, 46, 1739–1764.

    Article  Google Scholar 

  10. Chan, L. K. C., Lakonishok, J., & Hamao, Y. (1993). Can Fundamentals Predict Japanese Stock Returns. Financial Analysts Journal.

    Google Scholar 

  11. Conner, G., & Korajczyk, R. A. (2010). Factor models in portfolio and asset pricing theory. In J. Guerard (Ed.), The handbook of portfolio construction: Contemporary applications of Markowitz techniques. New York: Springer.

    Google Scholar 

  12. Cragg, J. G., & Malkiel, B. G. (1968). The consensus and accuracy of some predictions of the growth of corporate earnings. Journal of Finance, 23, 67–84.

    Article  Google Scholar 

  13. Deng, S., & Min, X. (2013). Applied optimization in global efficient portfolio construction using earnings forecasts. Journal of Investing, 23, 104–114.

    Article  Google Scholar 

  14. Dhrymes, P. J., Friend, I., Gultekin, B., & Gultekin, M. (1984). A critical reexamination of the empirical evidence on the APT model. Journal of Finance, 39, 323–346.

    Article  Google Scholar 

  15. Dhrymes, P. J., Friend, I., Gultekin, B., & Gultekin, M. (1985). New tests of the APT and their implications. Journal of Finance, 40, 659–674.

    Article  Google Scholar 

  16. Dhrymes, P. J., & Guerard Jr., J. B. (2017). Returns, risk, portfolio selection, and evaluation. In J. Guerard (Ed.), Portfolio construction, Measurment, and efficiency: Essays in honor of Jack Tteynor. New York: Springer.

    Google Scholar 

  17. Efron, B. (1982). The jackknife, the bootstrap, and other resampling plans. Philadelphia: Society for Industrial and Applied Mathematics.

    Book  Google Scholar 

  18. Elton, E. J., & Gruber, M. J. (1972). Earnings estimates and the accuracy of Expectational data. Management Science, 18, B409–B424.

    Article  Google Scholar 

  19. Elton, E. J., Gruber, M. J., & Gultekin, M. (1981). Expectations and share prices. Management Science, 27, 975–987.

    Article  Google Scholar 

  20. Elton, E. J., Gruber, M. J., Brown, S. J., & Goetzman, W. N. (2007). Modern portfolio theory and investment analysis (7th ed.). New York: Wiley.

    Google Scholar 

  21. Fama, E. F., & MacBeth, J. D. (1973). Risk, return, and equilibrium: Empirical tests. Journal of Political Economy, 81, 607–636.

    Article  Google Scholar 

  22. Fama, E. F., & French, K. R. (2008). Dissecting anomalies. Journal of Finance, 63, 1653–1678.

    Article  Google Scholar 

  23. Farrell Jr., J. L. (1997). Portfolio management: Theory and applications. New York: McGraw-Hill/ Irwin.

    Google Scholar 

  24. Farrer, D. E., & Glauber, R. R. (1967). Multicollinearity in regression analysis: The problem revisited. The Review of Economics and Statistics, 49, 92–107.

    Article  Google Scholar 

  25. Ferson, W. E., & Harvey, C. R. (1991b). Sources of predictability in portfolio returns. Financial Analysts Journal, 47, 49–56.

    Article  Google Scholar 

  26. Ferson, W. E., & Harvey, C. R. (1995). Explaining the predictability of asset returns. In A. Chen (Ed.), Research in finance (Vol. 11). Westport, Ct: JAI Press.

    Google Scholar 

  27. Fisher, L., & Lorie, J. H. (1964). Rates of return on Investments in Common Stocks. Journal of Business, 7, 1–21.

    Google Scholar 

  28. Geary, R. C., & Leser, C. E. V. (1968). Significance tests in multiple regression. The American Statistician, 22, 20–21.

    Google Scholar 

  29. Graham, B., & Dodd, D. (1934). Security analysis: Principles and technique. New York: McGraw-Hill Book Company.

    Google Scholar 

  30. Granger, C. W. J. (1969). Investigating casual relations by economic models and cross-spectral methods. Econometrica, 37, 424–438.

    Article  Google Scholar 

  31. Granger, C. W. J. (1989a). Invited review: Combining forecasts-twenty years later. Journal of Forecasting, 8, 167–173.

    Article  Google Scholar 

  32. Grinhold, R., & Kahn, R. (1999). Active portfolio management. New York: McGraw-Hill/Irwin.

    Google Scholar 

  33. Guerard Jr., J. B., & Horton, R. L. (1984a). The management f executive compensation in large, dynamic firms: A ridge regression estimation. Communications is Statistics, 13, 183–190.

    Article  Google Scholar 

  34. Guerard Jr., J. B., & Horton, R. L. (1984b). The management f executive compensation in large, dynamic firms: A further look. Communications is Statistics, 13, 441–448.

    Google Scholar 

  35. Guerard Jr., J. B., Gultekin, M., & Stone, B. K. (1997). The role of fundamental data and analysts’ earnings breadth, forecasts, and revisions in the creation of efficient portfolios. In A. Chen (Ed.), Research in finance (Vol. 15). Greenwich, CT: JAI Press.

    Google Scholar 

  36. Guerard Jr., J. B., & Mark, A. (2003). The optimization of efficient portfolios: The case for a Quadratic R&D Term. In Research in Finance, 20, 213–247.

    Article  Google Scholar 

  37. Guerard Jr., J. B. (2010). The handbook of portfolio construction: Contemporary applications of Markowitz techniques. New York: Springer. Chapter 3.

    Book  Google Scholar 

  38. Guerard Jr., J. B. (2012). Global earnings forecast efficiency. In J. Kensinger (Ed.), Research in finance (Vol. 28). Dublin: Emerald.

    Chapter  Google Scholar 

  39. Guerard Jr., J. B., Gultekin, M. N., & Xu, G. (2012). Investing with momentum: The past, present, and future. Journal of Investing, 21, 68–80.

    Article  Google Scholar 

  40. Guerard Jr., J. B., Rachev, R. T., & Shao, B. (2013). Efficient global portfolios: Big data and investment universe. IBM Journal of Research and Development, 57(5), 11.

    Article  Google Scholar 

  41. Guerard Jr., J. B., Markowitz, H. M., & Xu, G. (2015). Earnings forecasting in a global stock selection model and efficient portfolio construction and management. International Journal of Forecasting, 31, 550–560.

    Article  Google Scholar 

  42. Hastie, T., Tibshirani, R., & Friedman, J. (2016). The elements of statistical learning: Data mining, inference, and prediction (2nd ed.). New York: Springer. 11th printing.

    Google Scholar 

  43. Haugen, R. A., & Baker, N. (1996). Communality in the determinants of expected results. Journal of Financial Economics, 41, 401–440.

    Article  Google Scholar 

  44. Haugen, R., & Baker, N. (2010). Case closed. In J. B. Guerard (Ed.), The handbook of portfolio construction: Contemporary applications of Markowitz techniques. New York: Springer.

    Google Scholar 

  45. Hawawini, G., & Keim, D. B. (1995). On the predictability of common stock returns: World-wide evidence. In R. A. Jarrow, V. Maksimovic, & W. T. Ziemba (Eds.), Handbooks in operations research and management science: Finance (Vol. 9, pp. 497–537). Amsterdam: Elsevier.

    Google Scholar 

  46. Hendry, D. F., & Doornik, J. A. (2014). Empirical model discovery and theory evaluation. Cambridge: MIT Press.

    Book  Google Scholar 

  47. Hoerl, A. E. (1962). Application of ridge analysis to regression problems. Chemical Engineering Progress, 58, 54–59.

    Google Scholar 

  48. Hoerl, A. E., & Kennard, R. W. (1970a). Ridge regression: Biased estimation of nonorthogonal problems. Technometrics, 12, 55–69.

    Article  Google Scholar 

  49. Keane, M. P., & Runkle, D. E. (1998). Are financial analysts’ forecasts of corporate profits rational? The Journal of Political Economy, 106, 768–805.

    Article  Google Scholar 

  50. King, R., Plosser, C., & Stock, J. (1991). Stochastic trends and economic fluctuations. American Economic Review, 81, 819–840.

    Google Scholar 

  51. Lakonishok, J., Shleifer, A., & Vishny, R. W. (1994). Contrarian investment, extrapolation and risk. Journal of Finance, 49, 1541–1578.

    Article  Google Scholar 

  52. Latane, H. A. (1959). Criteria for choice among risky ventures. Journal of Political Economy, 67, 144–155.

    Article  Google Scholar 

  53. Latane, H. A., Tuttle, D. L., & Jones, C. P. (1975). Security analysis and portfolio management (2nd ed.). New York: The Roland Press.

    Google Scholar 

  54. Leamer, E. E. (1972). A class of informative priors and distributed lag analysis. Econometrica, 40, 1059–1081.

    Article  Google Scholar 

  55. Leamer, E. E. (1973). Multicollinearity: A Bayesian interpretation. Review of Economics and Statistics, 55, 371–380.

    Article  Google Scholar 

  56. Leamer, E. E. (1978). Specification searches: Ad hoc inference with nonexperimental data. New York: Wiley.

    Google Scholar 

  57. Lim, T. (2001). Rationality and analysts’ forecast bias. Journal of Finance, 56, 369–385.

    Article  Google Scholar 

  58. Lintner, J. (1965b). Security prices, risk, and the maximum gain from diversification. Journal of Finance, 20, 587–615.

    Google Scholar 

  59. Markowitz, H. M. (1952). Portfolio selection. Journal of Finance, 7, 77–91.

    Google Scholar 

  60. Markowitz, H. M. (1956). The optimization of a quadratic function subject to linear constraints. Naval Research Logistics Quarterly, 3, 111–133.

    Article  Google Scholar 

  61. Markowitz, H. M. (1959). Portfolio selection: Efficient diversification of investment, Cowles foundation monograph no.16. New York: John Wiley & Sons.

    Google Scholar 

  62. Markowitz, H. M. (1976). Investment in the Long run: New evidence for an old rule. Journal of Finance, 31, 1273–1286.

    Article  Google Scholar 

  63. Markowitz, H. M. (1987). Mean-variance analysis in portfolio choice and capital markets. Oxford: Basil Blackwell.

    Google Scholar 

  64. Markowitz, H. M., & Xu, G. (1994). Data mining corrections. Journal of Portfolio Management, 21, 60–69.

    Article  Google Scholar 

  65. Mossin, J. (1966). Equilibrium in a capital asset market. Econometrica, 34, 768–783.

    Article  Google Scholar 

  66. Rosenberg, B. (1974). Extra-market components of covariance in security returns. Journal of Financial and Quantitative Analysis, 9, 263–274.

    Article  Google Scholar 

  67. Rosenberg, B., & Marathe, V. (1979). Tests of capital asset pricing hypotheses. In H. Levy (Ed.), Research in finance (Vol. 1). Westport, Ct: JAI Press.

    Google Scholar 

  68. Ross, S. A. (1976). The arbitrage theory of capital asset pricing. Journal of Economic Theory, 13, 341–360.

    Article  Google Scholar 

  69. Ross, S. A., & Roll, R. (1980). An empirical investigation of the arbitrage pricing theory. Journal of Finance, 35, 1071–1103.

    Article  Google Scholar 

  70. Rudd, A., & Rosenberg, B. (1979). Realistic Portfolio Optimization. In E. Elton & M. Gruer (Eds.), Portfolio theory, 25 years after. Amsterdam: North-Holland.

    Google Scholar 

  71. Rudd, A., & Clasing, H. K. (1982). Modern portfolio theory: The principles of investment management. Homewood, IL: Dow-Jones Irwin.

    Google Scholar 

  72. Sharpe, W. F. (1963). A simplified model for portfolio analysis. Management Science, 9, 277–293.

    Article  Google Scholar 

  73. Sharpe, W. F. (1964). Capital asset prices: A theory of market equilibrium under conditions of risk. Journal of Finance, 19, 425–442.

    Google Scholar 

  74. Sharpe, W. F. (1966). Mutual fund performance. Journal of Business: A Supplement, 1(2), 119–138.

    Article  Google Scholar 

  75. Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Association, Series B, 58, 267–288.

    Google Scholar 

  76. Treynor, J. L. (1965). How to rate Management of Investment Funds. Harvard Business Review, 43, 63–75.

    Google Scholar 

  77. Treynor, J. L., & Mazuy, K. K. (1966). Can mutual funds outguess the market. Harvard Business Review, 44, 131–136.

    Google Scholar 

  78. Wheeler, L. B. (1994). Changes in consensus earnings estimates and their impact on stock returns. In B. Bruce & C. B. Epstein (Eds.), The handbook of corporate earnings analysis. Chicago: Probus.

    Google Scholar 

  79. Williams, J. B. (1938). The theory of investment value. Cambridge: Harvard University Press.

    Google Scholar 

  80. Wonnacott, T. H., & Wonnacott, R. J. (1981). In E. Robert (Ed.), Regression: A second course in statistics. Malabar, FL: Krieger Publishing Company.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Appendix: Modern Regression Analysis, the Case of Least Angle Regression

Appendix: Modern Regression Analysis, the Case of Least Angle Regression

There are many modern approaches to identifying the best-subsets of variables for linear regression models. Weighted latent root regression model first applied in Bloch et al. [34], and later validated in Guerard, Rachev, and Shao [161] was introduced in Chap. 4. Tibshirayi [304] reported variable selection modeling using the LASSO technique. Hastie, Johnstone, and Tibshirayi [174] introduced the Least Angle Regression (LAR) technique for variable selection. LASSO and LAR techniques are widely used and will make reasonable and relevant benchmarks for analyzing composite model building techniques. Hastie, Johnstone, and Tibshirayi [174] introduce LAR to the reader by discussing automatic model-building algorithms, including forward selection, all subsets, and back elimination. One can measure the goodness of fit in terms of predictive accuracy, but we will use a different manner, how well model subsets perform in terms of portfolio geometric means using out-of-sample variable weights. LAR is a variation of forward selection; that is the technique selects the variable with the largest absolute correlation, xj1, with the response variable, y and performs simple linear regression of y on xj1. The regression produces a residual vector orthogonal to xj1, now considered to be the response, or dependent variable. One projects the other predictor variables orthogonally to xj1, and repeat the selection process. The application of K steps produces a set of predictor variables, xj1, xj2, xj3, …, xjk to construct a K-parameter linear model. Hastie et al. [174] state that forward selection is an aggressive fitting technique that can be overly greedy, eliminating at a second step useful prediction correlated with xj1.

Forward Stagewise regression, is a cautious version of forward selection and is geometrically related to LASSO. All variables are standardized, to be zero mean and unit variance, or

$$ {\sum}_{i=1}^n{y}_i=0,{\sum}_{i=1}^n{x}_{ij}=0,{\sum}_{i=1}^n{x}_{ij}^z=1, $$
(A.1)

A prediction vector, m, can be written as an Xnxm matrix.

$$ \widehat{\mu}={\sum}_{j=1}^m{x}_j{\widehat{\beta}}_j=X\widehat{\beta}, $$
(A.2)

The total squared error is

$$ S\left(\widehat{\beta}\right)={\left\Vert y-\widehat{\mu}\right\Vert}^2={\sum}_{i=1}^n{\left({y}_1-{\mu}_i\right)}^2 $$
(A.3)

If \( \mathrm{T}\left(\widehat{\beta}\right) \) is the absoluter norm of \( \widehat{\beta} \), then

$$ \mathrm{T}\left(\widehat{\beta}\right)={\sum}_{j=1}^{nm}\left|{\widehat{\beta}}_j\right| $$
(A.4)

The LASSO chooses \( \widehat{\beta} \) by minimizing \( \mathrm{S}\left(\widehat{\beta}\right) \) subject to a bound on \( \mathrm{T}\left(\widehat{\beta}\right) \) The LASSO tends to shrink the OLS coefficients to zero, the shrinkage improves predictive accuracy, and LASSO is parsimonious, with only a subset of covariates have non-zero values of \( {\widehat{\beta}}_i \) \( {\widehat{\beta}}_j \). Stagewise, or forward Stagewise linear regression, is an iterative process that begins with \( \widehat{\mu}=0 \) and builds the regression function in successive small steps. If \( c\left(\widehat{\mu}\right) \) is the vector of current correlations,

$$ \widehat{c}=c\left(\widehat{\mu}\right)={x}^1\left(y-\widehat{\mu}\right) $$
(A.5)

Then the next step of the Stagewise algorithm is taken in the direction of the greatest current correlation

$$ \widehat{J}=\arg \max \mid {\widehat{c}}_j\mid \mathrm{and}\;\widehat{\mu}\to \widehat{\mu}+\upvarepsilon \operatorname {sign}\;\left({\widehat{c}}_j\right)\cdot {x}_j. $$
(A.6)

The Stagewise procedure does not lead to the “big” choice \( \upvarepsilon =\mid {\widehat{c}}_j\mid \) the Classic Forward Selection technique, but rather to a “small” choice, which is not overly greedy.

LAR is a version of the Stagewise procedure to accelerate calculations. All coefficients start as equal to zero. LAR is similar to Forward Selection in that it finds the predictor most correlated with the response.

LAR next proceeds to find the predictor with as much correlation with the current residual. LAR finds in an equiangular direction between the first two predictors and a third predictor that is equally-correlated with the “most correlated” set, along the “least angle direction.” LAR creates a regression model, one covariate at a step, such that after K steps, only K of the \( {\widehat{\beta}}_j \)s are non-zero.

$$ {\widehat{\mu}}_1={\widehat{\mu}}_0+{\widehat{\gamma}}_1{x}_1 $$
(A.7)

LAR uses \( {\widehat{\gamma}}_1 \) \( {\widehat{\gamma}}_1 \) such that \( {\overline{y}}_2-\widehat{\mu} \) is equally correlated with x 1 x 1 and x 2 x 2. Efron et al. (2004) demonstrate that \( {\overline{y}}_2-{\widehat{\mu}}_1 \) \( {\overline{y}}_2- \)bisects the angle between x 1 x 1 and x 2 such that \( {c}_1\left({\widehat{\mu}}_1\right)={c}_2\left({\widehat{\mu}}_1\right) \).

$$ {\widehat{\mu}}_2={\widehat{\mu}}_1+{\widehat{\gamma}}_2{u}_2, $$
(A.8)

where u 2 u 2 is the unit vector lying along the bisector.The LAR subsequent steps are taken along equiangular vectors. The covariate vectors are linearly independent.

LAR begins at \( {\widehat{\mu}}_0=0 \) with the Stagewise procedure. Each step builds \( \widehat{\mu} \). If the current \( {\widehat{\mu}}_A \) is \( \widehat{c}={x}^1\left(y-{\widehat{\mu}}_A\right) \) and is the current correlations vector. The active set A is the set of indices corresponding to

$$ \widehat{c}=\max \left\{|{\widehat{c}}_j|\right\}\mathrm{and}\;\mathrm{A}=\left\{j|{\widehat{c}}_j|=\widehat{c}\right\} $$
(A.9)

Let \( {s}_j=\operatorname{sign}\left\{{\widehat{c}}_j\right\} \) for j ϵ A.The inner product vector, a, is:

$$ \mathrm{a}\equiv {x}^1{u}_A $$

The next LAR step update is:

$$ {\widehat{\mu}}_{A+}={\widehat{\mu}}_A+{\widehat{\gamma}}_{\mu A} $$

where

$$ \widehat{\gamma}={\min}_{\mathrm{j}\in \mathrm{A}}+\left\{\frac{c-{\widehat{c}}_j}{A_A-{a}_j},\frac{c-{\widehat{c}}_j}{A_A-{a}_j}\right\} $$
(A.10)
$$ {\displaystyle \begin{array}{lll}\upmu \left(\upgamma \right)& =& {\widehat{\mu}}_A+{\gamma}_1{\mu}_A\\ {}{c}_j\left(\gamma \right)& =& {x}_j\left(y-\mu \left(\gamma \right)\right)={\widehat{c}}_j-{\widehat{\gamma}}_{aj}\\ {}\mid {c}_j\left(\gamma \right)\mid & =& \widehat{c}-\gamma {A}_A\end{array}} $$
(A.11)

Thus, \( \widehat{\gamma} \) is the smallest positive value of γ such that the new index j joins the active set, A+. The new maximum absolute correlation is \( {c}_{+}=\widehat{c}-\widehat{\gamma}{A}_A \).

$$ {\displaystyle \begin{array}{l}{\overline{y}}_k={\widehat{\mu}}_{k-1}+{Y}_k{J}_k^{-1}{X}_k^1\left(y-{\widehat{\mu}}_{k-1}\right)+\frac{{\widehat{c}}_k}{A_k}{U}_k\\ {}{X}_k^1\left(y-{\mu}_{k-1}^a\right)={\widehat{C}}_k{I}_A\\ {}{\widehat{\gamma}}_k=\frac{{\widehat{c}}_k}{A_k}\\ {}{\widehat{\mu}}_k-{\widehat{\mu}}_{k-1}=\frac{{\widehat{\gamma}}_k}{{\overline{\gamma}}_k}\left({\overline{y}}_k-{\widehat{\mu}}_{k-1}\right)\end{array}} $$

Because \( {\widehat{\gamma}}_k \) is less than \( {\overline{\gamma}}_k \), \( {\overline{\mu}}_k \) lies closer to \( {\overline{y}}_k \) to \( {\widehat{\mu}}_{k-1} \). Successive LAR steps approach but do not reach the OLS estimate \( {\overline{y}}_k \). LAR is computationally thrifty.

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Dhrymes, P. (2017). The General Linear Model IV. In: Introductory Econometrics. Springer, Cham. https://doi.org/10.1007/978-3-319-65916-9_4

Download citation

Publish with us

Policies and ethics