Post on 27-May-2018
Department of Business Administration- School of Economics and Management – Lund University
Market Value Approximation Using MultiplesAn Investigation of American Large-, Mid-, and Small-Cap Stocks
Thesis Supervisor: Master Students:
Naciye Sekerci Berglund, Oscar 900115
Zisiadou, Argyro 920706
May, 2015
Master’s Programme in Corporate & Financial Management (EAGCM)
1
Berglund Oscar Zisiadou Argyro
Abstract
The aim of this paper is to provide answers to questions related to valuation accuracy and error determinants by investigating the US market over the last 15 years (2000-2014). The first questions are related to the market efficiency in its weak form, while the rest of the paper is focused on valuation accuracy using the multiples approach and the error determinant variables that can influence the valuation accuracy. The main results for the market efficiency are that only the small capitalization index (S&P 600) is efficient in its weak form. Regarding valuation accuracy, it was discovered that the 12-month forward-looking multiples are more accurate at approximating market values than trailing multiples. In addition, the authors were able to conclude that equity multiples performed better than entity multiples. Lastly, through estimations, it is proven that valuation errors are influenced by specific error determinants both from the current period and the past period observations, while the valuation error of the previous year appears to have a significant influence on the new valuation error.
Keywords: Valuation Error, Valuation Accuracy, Error Determinants, Multiples, Market Efficiency, US market, Capitalization
Department of Business Administration- School of Economics and Management – Lund University
Acknowledgements
We would like to acknowledge our supervisor, Nacyie Sekerci. We would also like to acknowledge and thank Julienne Stewart-Sandgren. Without their help and support we would not have been able to write this thesis. Thank you!
3
Berglund Oscar Zisiadou Argyro
Table of Contents
1. Introduction 81.1 Background 81.2 Aim and Objectives 91.3 Research Purpose101.4 Research Limitations 111.5 Outline of the Thesis 11
2. Literature and Theoretical Review 132.1 Market Efficiency 132.2 Discounted Cash Flow Model 142.3 Multiples152.4 Mispricing and Error Determinants 19
3. Methodology 213.1 Research Approach 213.2 Research Design 223.3 Data Collection 243.4 Data Analysis 24
3.4.1 Market Efficiency – Time Series Approach 243.4.2 Multiples 303.4.3 Valuation Accuracy – Statistical Approach 333.4.4 Error Determinants – Panel Data Approach 35
4 Analysis and Discussion 464.1 Market Efficiency 46
4.1.1 Descriptive Statistics 464.1.2 Model Approach 474.1.3 Residual Diagnostics 484.1.4 Forecasting 504.1.4.1 Forecasting Large Capitalization 504.1.4.2 Forecasting Mid Capitalization 514.1.4.3Forecasting Small Capitalization 52
4.2 Valuation Accuracy 534.3 Valuations Error Determinants 57
5. Conclusion 645.1 Practical Implications 665.2 Future Research 66
References 67
Appendix A 69
Appendix B 70
Appendix C 94
Department of Business Administration- School of Economics and Management – Lund University
Table of Figures
4.1.1 DIFFERENCE IN VALUATION ERROR FOR DIFFERENT SIC-LEVELS 56 4.1.2 PERCENTAGE OF SIGNIFICANCE OF ERROR DETERMINANTS (TIME T) 604.1.3 PERCENTAGE OF SIGNIFICANCE OF ERROR DETERMINANTS (TIME T-1) 61 4.1.4 GOODNESS OF FIT 62
List of Tables
4.1.1 DESCRIPTIVE STATISTICS 464.1.2 BEST MODEL APPROACH 484.1.3 LARGE CAPITALIZATION FORECASTING RESULTS514.1.4 MID CAPITALIZATION FORECASTING RESULTS 514.1.5 SMALL CAPITALIZATION FORECASTING RESULTS524.1.6 AVERAGE VALUATION ERRORS – TOTAL SAMPLE 534.1.7 P/E MEDIAN ERROR DETERMINANTS AND DIAGNOSTIC TESTS 58
5
Berglund Oscar Zisiadou Argyro
Abbreviations
Abbreviation MeaningAbs. AbsoluteAIC Akaike Information CriterionAR AutoregressiveARCH AutoRegressive Conditional HeteroskedasticityARIMA AutoRegressive Integrated Moving AverageARMA AutoRegressive Moving AverageBDS Brock-Dechert-ScheinkmanBPG Breusch-Pagan-GodfreyCap. CapitalizationCAPM Capital Asset Pricing ModelCorr CorrelationCov CovarianceCSU Cross-section UnitsCV Continuing ValueDCF Discounted Cash FlowDDM Dividend Discount ModelDW Durbin - WatsonE-GARCH Exponential Generalized AutoRegressive Conditional HeteroskedasticityEBIT Earnings Before Interests and TaxesEBITDA Earnings Before Interests, Taxes, Depreciation and AmortizationEV Enterprise ValueEV/EBITDA (1) 12-month Forward EV/EBITDAFE Fixed EffectsFFIG Fama and French Industry GroupingsGARCH Generalized AutoRegressive Conditional HeteroskedasticityGICS Global Industry Classification StandardICB Industry Classification BenchmarkIPO Initial Public OfferingIT Information TechnologyIV Instrument VariableJB Jarque - BeraLM Langrange MultiplierM/B Market to Book ratioMA Moving AverageNAICS North America Industry Classification SystemOLS Ordinary Least SquareP/E Price to EarningsP/E (1) 12-month Forward P/E
Department of Business Administration- School of Economics and Management – Lund University
P/EBITDA (1) 12-month Forward P/EBITDAP/S Price to SalesR&D Research and DevelopmentRE Random EffectsRHS Right Hand SideRIV Residual Income ValuationRW Random WalkSE Standard ErrorsSIC Standard industrial ClassificationT-GARCH Threshold Generalized AutoRegressive Conditional HeteroskedasticityTA Total AssetsWACC Weighted Average Cost of Capital
1. Introduction
1.1 Background
Students of finance, private investors, and finance professionals are all concerned with
valuation of corporate entities. As aid in these efforts, they have different models to choose
from, where the most prominent and commonly used are the DCF-method, the RIV-model,
7
Berglund Oscar Zisiadou Argyro
the DDM-model, and the Multiples approach (Lie & Lie, 2002). Although, the DCF-method
is widely accepted by economists, a vast majority of finance professionals mainly use the
multiples approach in their daily work (Demirakos, Strong & Walker, 2004; Kaplan &
Ruback, 1995). Reasons for this are many. For example, the multiples approach tends to be
less time consuming to complete, the method is easy to grasp for individuals not well versed
in finance, and ratios are very accessible since they are usually quoted in financial
publications and on trading platforms (Schreiner, 2007). Surprisingly, although a vast
majority of finance professionals, around 65%, favor using the multiples approach, very little
prior research has been conducted on the subject of corporate valuation using multiples
(Schreiner, 2007).
In addition, the research to date, which is scant, seem to only focus on multiple valuation of
large cap companies, mainly in the United States. Furthermore, to the best of our knowledge,
there is only one published article looking at European large cap stocks. Filling the gap, we
believe that our contribution of valuing US mid- and small-cap stocks through the multiples
approach could prove useful both for professional finance practitioners and private investors.
1.2 Aim and Objectives
The main objective of this thesis is to investigate which corporate valuation multiple(s)
provides the closest approximation to market value. In this effort, we will use the valuation
approach outlined in the papers by Lie & Lie (2002), Schreiner (2007), Kaplan & Ruback
(1995), and Ek & Lillhage (2012). The approach is described in detail in the Methodology
chapter of this thesis. A secondary objective, a natural extension, will be to see how the results
of previous studies support our sample of mid- and small-cap stocks. Essentially, we aim to
explore if the previously deemed best multiples for valuing large cap stocks prove to be the
Department of Business Administration- School of Economics and Management – Lund University
best for our sample as well. Based on our results, we also aim to be able to draw conclusions
on as to what variables may distort valuation results when using the multiples approach.
Furthermore, we will devote some time to the famous market efficiency hypothesis first
mentioned by Eugene Fama in 1970. In addition, we aim to investigate whether any model
can predict the price movements of the market indices. As a result, we will be able to
conclude whether the market efficiency hypothesis holds or not.
In order to reach these aims, we endeavor to address eight main research questions, which are
divided into three categories: Market Efficiency, Valuation Accuracy and Error Determinants:
A. Market Efficiency:
1. Is the S&P large-, mid-, and small-cap indices efficient in its weak form during year
2000 through 2014?
B. Valuation Accuracy:
2. What is the average valuation error for each multiple?
3. Which multiple gives the best, that is, closest approximation to market value?
4. On average, are there any notable differences in valuation errors for the different
indices?
5. Do equity value multiples outperform entity value multiples in terms of valuation
accuracy?
6. Do forward looking multiples outperform trailing multiples in terms of valuation
accuracy?
C. Error Determinants:
7. Is there any ‘error variable’ that can significantly influence the valuation error?
9
Berglund Oscar Zisiadou Argyro
8. Is there significant correlation between the valuation error of the present period (t) and
error variable observation from the previous period (t-1).
The study uses a hypothesis method to address each question, except questions 2 and 3.
Questions 2 and 3 are kept in their current form as it is not possible to rephrase them into
hypotheses.
1.3 Research Purpose
The purpose of our research is to provide answers to the questions listed above. The results
will further be compared to the results from previous papers written in the same field of
interest, that is, the papers by Kaplan & Ruback (1995), Lie & Lie (2002), Schreiner (2007),
and Ek & Lillhage (2012). From this, we will be able to provide an analysis based on
empirical observations on whether mid- and small-cap markets need to be valued differently
or not, as well as whether the U.S market can be approached in the same way as the Nordic
and the European market or not. In addition, we intend to investigate if the market efficiency
hypothesis holds for the markets and period we examine; that is, US large-, mid-, and small-
cap markets during the time span 2000 through 2014. Based on the results we will be able to
connect mispricing to market (in)efficiency.
1.4 Research Limitations
We would have liked to extend the research further by exploring two separate periods; that is,
before and after the economic crisis. Further, using more forward multiples would have given
us the opportunity to compare our results with previous research, including two year
multiples. Although, these are areas for future research, we feel that this could have been done
if time permitted. Instead, we have chosen to keep the methodology mainly intact compared to
Department of Business Administration- School of Economics and Management – Lund University
the studies we draw on, and have been able to focus more on data collection and increasing
the sample. The final sample includes three-hundred (300) companies with annual
observations for each company over fifteen (15) years, that is, year 2000 through year 2014.
For each year and company we have collected twenty two (22) different variables (see
Appendix C for a full list of all the variables).
1.5 Outline of the Thesis
This thesis is separated into five (5) different chapters. Chapter 1 provides a general
introduction to the main topic of research, which is valuation accuracy using the multiples
approach. Moreover, the main questions we intend to answer will be presented in this chapter.
The literature review and the theoretical background is accounted for in Chapter 2. Chapter 3
contains the research approaches and the data analysis for all the prospects examined in this
thesis including the market efficiency, the valuation accuracy, the multiples and the error
determinants. Furthermore, this chapter, contains all the empirical approaches; that is,
statistical and econometric estimations and diagnostic tests. Chapter 4 presents the results of
our analysis, with an accompanying discussion of our findings. Finally, Chapter 5 includes all
the conclusions of our research along with answers to our initial hypothesis, followed by
recommendations for further research.
11
Berglund Oscar Zisiadou Argyro
2 Literature and Theoretical Review
There are a great many variables that play into the valuation of a corporate entity. Depending
on how thorough the analyst wants to be and how detailed the valuation needs to be, there
may be a need to adjust every single line item of a company’s financial statements. However,
for this research paper we have decided to limit ourselves to mainly investigating one
approach in particular; that is, the multiples approach. This being the case, our literature and
theoretical review will mainly delve into what others have discovered on the subject of
corporate valuation using multiples as well as the advantages and disadvantages of some of
the competing models mentioned in Section 1.1. Naturally, market efficiency, as it holds an
important place in this thesis, will be discussed.
2.1 Market Efficiency
Eugene Fama in his paper from 1970, tried to explain the market efficiency and created three
different forms. The first form is the weak form efficiency in which, future prices cannot be
predicted by analyzing the prices from the past. The second form is the semi-strong form
efficiency in which stock prices adjust to publicly available information rapidly and in an
unbiased fashion, so that no excess returns can be earned by trading on that information. The
third and last form is the strong form efficiency in which stock prices reflect all information,
both private and public, and no one can earn excess returns. Since then, many researchers
tried to investigate the weak form efficiency hypothesis in different periods and markets. For
Department of Business Administration- School of Economics and Management – Lund University
instance, Gu (2004) used daily observation of the NASDAQ composite index from 1971 to
2001 and proved that there is no weak form efficiency. On the other hand, Chan, Gup and Pan
(1992), examined the United States and the major Asian Markets and came to the conclusion
that they are efficient in the weak form.
2.2 Discounted Cash Flow Model
The DCF-model, when used correctly, can estimate the entity value of a firm; that is, market
value of equity plus net debt and cash (e.g. Koller, Goedhart & Wessels, 2010; Damodaran,
2006). According to Koller et al. (2010) description of the model, it does so by discounting all
the future projected free cash flows to firm at an appropriate discount rate. They further point
out that the discount rate used is usually the firm’s weighted average cost of capital, or
WACC for short. Moreover, the WACC has two parts to it, the cost of equity and the cost of
debt. Depending on the firm’s capital mix, different weights are allocated to the different costs
to determine the WACC. At the end of the manually forecasted period a continuing value, CV
for short, is calculated by using an appropriate perpetuity formula. Once again, this value is
then discounted back to present time using the WACC. All the discounted cash flows are then
summed to arrive at entity or enterprise value. Then, according to Koller et al. (2010), once
the entity value is calculated it is then possible to subtract net debt to arrive at equity values.
In total, it is probably the most extensive and complete model in terms of flexibility and
ability to capture specific claims on or of the firm (Koller et al, 2010).
As such, it is very useful when, for example, valuing unorthodox transactions such as
management buyouts. Indeed, the DCF-model is what Kaplan & Ruback (1995) used in the
their research. In their study, Kaplan & Ruback (1995) looked at 51 high-leveraged
transactions spanning from 1983 to 1989. More specifically, they calculated the entity values
13
Berglund Oscar Zisiadou Argyro
of the firms in their sample by discounting the projected cash flows with the WACC, and then
calculating a CV-value for the year after projections ended. They used the information
provided by respective management for their cash flows and WACC projections. Doing this,
they found that the DCF-model performed very well when comparing the calculated entity
values to the market transactions, with a mean valuation error of 8% (median only 6%)
(Kaplan & Ruback, 1995).
Not surprising to anyone familiar with the model, Kaplan & Ruback (1995) also found that
the valuation errors changed drastically when they altered variables in the CAPM-formula
used to calculate the WACC. Specifically, they changed betas and market-risk premium. This
fact led the authors to raise one of the main drawbacks of the DCF-model, namely that it is
very dependent on the assumptions the user makes regarding growth rates, WACC, CV-year,
among many other things. Kaplan and Ruback (1995), however, maintain that their results are
based solely “on a number of ad hoc assumptions” (p.1060), that both professionals and
academics should be able to improve, which in turn should lead to more accurate valuations.
Regardless if this is true or not, there is still another major drawback with the DCF-method in
that it tends to be very time consuming to complete. When done fundamentally, every line
item of the financial statements should be forecasted, at least for a period of a few years
(Koller et al., 2010). As such, the model has an obvious disadvantage to other, less time-
consuming approaches such as using multiples when computing corporate value.
2.3 Multiples
At its core, a multiple is a ratio of some variables pertaining to corporate value, for example
stock price to earnings (P/E). Depending on how the market values other, comparable firms
one can quickly compute an approximate value of the target firm (Schreiner, 2007). The
Department of Business Administration- School of Economics and Management – Lund University
procedure of calculating a firm multiple has four steps to it. Step one is to decide on which
numerator and denominator to use. For example, price to earnings and enterprise value to
EBITDA are two popular ones. In those instances the numerator would be price and enterprise
value respectively and the denominator or the value driver would be earnings and EBITDA
respectively. The second step is to decide on what peers to use (Schreiner, 2007). Essentially
one can do this through any design, however the consensus among the authors published in
the field is to use firms with similar industry codes (Lie & Lie, 2002; Schreiner, 2007). The
most commonly used industry classification systems are SIC, NAICS, FFIG, GICS and ICB
(Schreiner, 2007). Most systems rank a company with a four-digit code. For example code
6021 under the SIC-system is National Commercial Banks, code 602 is Commercial Banks,
and code 60-67 is Finance, Insurance, and Real Estate companies. Therefore, the more
numbers that match, the closer a firm is to the target company. The norm in the papers we
have examined, seem to use peers with at least a three-digit industry code match (Ek &
Lillhage, 2012; Lie & Lie, 2002; Schreiner, 2007). Evidently, at least three digits are used
because Alford (1992) found that multiple valuation accuracy improves when the industry
code is narrowed from zero up to three digits. Using four digits had no apparent advantage in
valuation accuracy versus only using three (Alford, 1992). Another issue that comes up when
selecting a peer sample, is the number of peers needed in order to form an adequate
comparable. Lie & Lie (2002) recommend using a peer sample of at least five firms. If there
are less than five firms for a given industry code, the code is then relaxed one level (Lie &
Lie, 2002). Thus, in our previous example, code 602 would be relaxed to 60 if there were less
than five comparable firms in the 602 industry. Schreiner (2007) on the other hand, maintains
that the optimal peer group consists of four to eight peers. Should there be less than four firms
for any given level-three industry code Schreiner (2007) suggests simply making do with the
smaller sample as relaxing the industry code level decreases comparability between the peers.
15
Berglund Oscar Zisiadou Argyro
The only time when Schreiner (2007) recommends relaxing the criteria or using a different
valuation method altogether, is when there are less than two peers.
Step three is concerned with combining each peer multiple into a single number. There are
several ways of doing this, for example using the arithmetic mean, geometric mean, harmonic
mean, or simply taking the median. Ek & Lillhage (2012) and Lie & Lie (2002) use the
arithmetic mean and the median in their papers. Schreiner (2007) however, argues that the
arithmetic mean is inappropriate to use when aggregating peer multiples as it relies too
heavily on outliers. He instead recommends using the median or harmonic mean (Schreiner,
2007).
Step four is the actual valuation and is very straight forward. The computed peer multiple
from step three is simply multiplied with the value driver, for example earnings, of the target
firm. It is plausible that it is the simplicity of this process that is attributable to the multiples
approach widespread practice. In fact, according to Dr. Klaus Spremann of the Swiss Institute
of Banking and Finance: “[a]ccounting based market-multiples are the most common
technique in equity valuation” (Schreiner, 2007, p.VII). Dr. Spremann further acknowledges
that multiples “are used in research reports…stock recommendations…in fairness opinions…
pitch books of investment bankers… [and even]…road shows of firms seeking an IPO”
(Schreiner, 2007, p. VII). With such widespread use, it is rather surprising to learn that there
is not much research on the subject. Out of the scarce sample of relevant research papers we
found, the most significant on the subject of valuation using multiples were written by Kaplan
& Ruback (1995), Lie & Lie (2002), Schreiner (2007), and Ek & Lillhage (2012). These
authors all raise important benefits of using multiples. For instance, the fact that they are
relatively quick to compute when compared to other more labor intensive models, such as the
DCF- or RIV-model. According Lie & Lie,
Department of Business Administration- School of Economics and Management – Lund University
[t]he theoretical emphasis is usually on the discounted cash flow valuation (DCF)
technique, but it is cumbersome to use and sensitive to a host of assumptions.
Consequently, investment bankers and appraisers regularly use valuation by multiples
(Lie & Lie, 2002, p. 1).
Another attractive feature with multiples is the ease of use and simplicity, which makes the
concept very graspable to people not well versed in corporate finance (DeAngelo, 1990, cited
in Schreiner, 2007). In addition, they are also very useful when initially screening stocks for
investment purposes as relevant multiples are usually quoted in financial newspapers and on
trading platforms (Schreiner, 2007). In sum, multiples make for a very powerful and useful
valuation tool. Given this, a very relevant question tends to come up: which is in fact the best
multiple to use? Naturally, the answer depends on the context surrounding the question.
However, in general, Lie & Lie found that asset based multiples yielded more accurate and
less biased results than sales and earnings estimates did. Secondly, they found that adjusting
entity values for cash had little to no effect on valuation accuracy, while using forward
looking value drivers did. They also found EBITDA to be a better value driver than EBIT in
terms valuation accuracy for entity multiples. Lastly, they discovered that the overall accuracy
and bias of the valuation were greatly influenced by company size, profitability, and amount
of intangibles on the balance sheet (Lie & Lie, 2002). Schreiner (2007) on the other hand
found that earnings based multiples performed best, especially the two-year forward looking
price to earnings multiple, which had the overall best performance. Ek & Lillhage’s (2012)
findings support Schreiner’s in this regard. Although Ek & Lillhage (2012) did not look at the
two-year forward looking P/E multiple specifically, they found that the regular P/E multiple
performed the best in their sample. Schreiner (2007) also had another noteworthy discovery,
namely that equity multiples outperformed entity multiples in every regard. In addition, he
found, somewhat in line with the findings of Lie & Lie (2002), that knowledge related
17
Berglund Oscar Zisiadou Argyro
multiples, such as ratios based on R&D, outperform traditional multiples in science-based
industries (Schreiner, 2007).
However, one of the drawbacks of the multiples approach is that it combines all the
complexities of the business and provides one value for a given point in time (Schreiner,
2007). As such, it doesn’t require any fundamental analysis of the underlying business, which
arguably is very important from an investor’s perspective. In addition, since the multiples are
based on market values, there are instances where values will be somewhat inflated due to the
market being too hot (Schreiner, 2007).
2.4 Mispricing & Error Determinants
As described in the previous section, Lie & Lie (2002) found that company size and amount
of intangibles on the balance sheet affected the accuracy of a firm valuation. Especially
companies with a large part of their asset base consisting of intangible assets were severely
mispriced using traditional multiples, such as price to book or price to sales. Price to book
gave a median valuation error of -71%, while price to sales gave a median valuation error of -
151%. Based on their results Lie & Lie suggest that multiples are not suited well for valuing
tech companies. It should however, be noted that Lie & Lie identified tech companies by
searching for companies with a .com in their name. Their paper is also from 2002 essentially
right after the IT-crash, which may have also played into the valuation of especially IT- and
tech companies. Schreiner (2007) on the other hand found that multiples performed rather
well in valuing science-based companies, the trick is just to use the right multiple. He
proposed using knowledge-related multiples such as P/(EBIT+RD) or P/(E+R&D) for which
Department of Business Administration- School of Economics and Management – Lund University
he received a better valuation result with a median valuation error of 29.8% and 28.07%
respectively. The results are not near as accurate as for valuing non-science based companies
with traditional multiples, but the improvement from the sample of Lie & Lie (2002) still
suggest that there may be some multiple combination out there that can capture the value of
science-based companies well. Nevertheless, it could be questioned why is it difficult to value
science-based companies. For one, many of these science-based firms are still to post profits
for the first time, which obviously pose a problem since most valuation models work by
discounting cash flows. Consequently, when valuing many of these firms it essentially
becomes a guessing game without any historic earnings to base future cash-flow predictions
on. The science-based companies that do post profits and expense intangible assets and
research, R&D for example, may experience severely decreased earnings in the short run.
However, according to Eberhart, Maxwell & Siddique (2004) a majority of firms that
suddenly increase their R&D expenses, later enjoy longer periods of above average returns,
something that markets are slow to incorporate - in essence another source of mispricing. In
support of the accounting treatment’s role in mispricing Lev, Sarath & Sougiannis (2005)
found consistent undervaluation of firms that expense R&D and consistent overvaluation of
firms that capitalize R&D, suggesting that markets have trouble valuing investments in R&D.
3 Methodology
In this section, we initially describe the Time Series Analysis, used in this thesis, in order to
examine the Market Efficiency in the three (3) different capitalization indices. The second
19
Berglund Oscar Zisiadou Argyro
part contains the main approach for our thesis, which is the Panel Data Analysis that will be
used to estimate the main models for the valuation accuracy and the error determinants.
3.1 Research Approach
As many researchers propose, the most appropriate approach for our investigative topic is a
quantitative method. That means that we are going to base our analysis on data gathered and
connect it to the topic we are trying to investigate. The fact that we are analyzing three
different capitalization indices and the first 100 firms of each index for a period of 15 years,
lead us to use the Panel Data Approach. Moreover, because we are examining market
efficiency for three indices in their weak form over a period of 15 years with daily
observations, the most appropriate method of analysis is the Time Series Approach. This
approach was proposed by Fama (1970) when he mentioned the market efficiency hypothesis.
This research project has also chosen the hypothesis approach in order to provide clear and
indisputable answers to our research questions.
3.2 Research Design
With this thesis we attempted to answer eight (8) questions divided into three main categories.
The first part aims to answer questions related to the Market Efficiency Hypothesis, the
second part is related to Valuation Accuracy, and the third part pertains to Error Determinants.
Regarding market efficiency, we attempted to answer if markets are efficient in its weak form
for all three (3) capitalization indices in US Market (large-cap, mid-cap, small-cap) during our
specified time period, that is, year 2000 through 2014.
Department of Business Administration- School of Economics and Management – Lund University
Pertaining to valuation accuracy and mispricing, we attempted to shed light on whether or not
the different indices need to be valued differently as well as which multiple provides the best
approximation of market value. In addition, we tried to determine which of our specified
variables were connected to mispricing.
A. Market Efficiency
1. Is the S&P large-, mid-, and small-cap indices efficient in its weak form during year
2000 through 2014?
B. Valuation Accuracy
2. What is the average valuation error for each multiple?
3. Which multiple gives the best, that is, closest approximation to market value?
4. On average, are there any notable differences in valuation errors for the different
indices?
5. Do equity value multiples outperform entity value multiples in terms of valuation
accuracy?
6. Do forward looking multiples outperform trailing multiples in terms of valuation
accuracy?
C. Error Determinants
7. Is there any ‘error variable’ that can significantly influence the valuation error?
8. Is there significant correlation between the valuation error of the present period (t) and
error variable observation from the previous period (t-1).
In order to address these questions, we developed seven (7) hypotheses. The first three
hypotheses are connected to the market efficiency whereas the remaining hypotheses are
connected to the valuation accuracy and mispricing.
21
Berglund Oscar Zisiadou Argyro
H1: The large capitalization index (S&P500) is efficient in its weak form over the past 15
years.
H2: The mid capitalization index (S&P400) is efficient in its weak form over the past 15
years.
H3: The small capitalization index (S&P600) is efficient in its weak form over the past 15
years.
H4: Equity multiples outperform entity multiples in terms of valuation accuracy.
H5: There is no difference across the capitalizations by using the synthetic peer multiples.
H6: Forward-looking multiples outperform trailing multiples in terms of valuation accuracy.
H7: There is no connection between the valuation error and past observations of the error
determinants.
3.3 Data Collection Method
Our main sources for datasets are Thomson Reuters - Datastream Professional 5.1 as well as
the NASDAQ website (NASDAQ, 2015). Specifically, the list with firms under each
Capitalization Category (S&P500, S&P400, S&P600) were downloaded from the NASDAQ
(NASDAQ, 2015) webpage. The same source was used in order to identify the SIC Code for
each firm separately. The rest of our variables were gathered in Thomson Reuters -
Datastream Professional 5.1 (see Appendix C, Table C1)
Department of Business Administration- School of Economics and Management – Lund University
Regarding data frequency, we use annual observations for all of the variables pertaining to the
panel data analysis, as well as the multiples calculations. For the market efficiency time series
analysis, the observations are daily. All variables have been gathered from Thomson Reuters -
Datastream Professional 5.1 (see Appendix C, Table C1).
3.4 Data Analysis
Having as purpose to analyze different aspects of valuation accuracy and the error
determinants, it is prudent to separate the analysis into subsections so that it will be easier for
the reader to follow the procedure.
3.4.1 Market Efficiency - Time Series Approach
The term Time Series Analysis is used to describe the different econometric approaches that
can be used in order to make estimations and forecasts on time-series data. Conceivable from
the name itself, time series data is a dataset that contains observations for a specific variable
of interest over a specific period of time. For instance, a time-series dataset can be daily
observations of a stock index over the last ten years.
Halkos (2011a) explains that there are both parametric and non-parametric methods that can
be used while working on time series analysis datasets. In the parametric approaches, the
underlying stochastic process has a specific configuration that includes an certain number of
parameters such as Autoregressive model (AR) or Moving Average model (MA) or even a
combination of both like the ARMA or ARIMA approach. Moreover, Halkos (2011a) also
mentions that there can be a separation between linear and non-linear approaches as well.
23
Berglund Oscar Zisiadou Argyro
In the case of Market Efficiency, we are going to use the time-series analysis because the
dataset for each index examined will contain daily observations over the last 15 years (2000-
2014). Moreover, the estimations and forecasts for each index will be done separately and
without any combined estimations. Our goal is to model the movements of each index
autonomously and not in regard to the other indices.
As in any statistical or econometric procedure, before the estimation and the forecast, a
descriptive analysis is required. A descriptive analysis will provide an opportunity to the
researchers or analysts to get a general idea of the observations they are going to use in their
estimations. The most important indicators that require our attention are the mean value and
the standard deviation. Moreover, skewness and kurtosis give a general idea whether the
observations are normally distributed or not. However, we should not jump into conclusions
only based on the descriptive statistics. We also need to test the datasets as a number of
different problems may occur.
Problems may occur in Time-series datasets
1. Autocorrelation
One of the assumptions of an OLS estimation is that the error terms are assumed to be
independent over time. Autocorrelated error terms are the most important consideration for a
practicing forecaster because when the problem of the autocorrelation occurs, the estimation
and the forecast will be biased. (Halkos, 2011a; Brooks, 2008)
In order to test for autocorrelation we can use a number of different tests, such as Durbin-
Watson statistics (DW).
H0= independent error terms over time, the error terms are completely random and serially
uncorrelated.
Department of Business Administration- School of Economics and Management – Lund University
If there is a serial correlation issue, autocorrelation may occur because the residuals are not
serially uncorrelated.
2. Heteroskedasticity
One of the assumptions of an OLS estimation is that the variance of the error terms are
assumed to be constant (homoscedastic). However, if the variance of the residuals is not
constant but depend upon the value of x, then the residuals are heteroskedastic. (Halkos,
2011a)
In order to test for Heteroskedasticity, we can use the Goldfeld-Quandt test (Halkos, 2011a),
Breusch – Pagan – Godfrey test (Halkos, 2011a), or the White test (Halkos, 2011a)
H0= Error term variance is constant (homoscedastic)
3. Multicollinearity
With the term Multicollinearity we identify the issue when two or more independent variables
(x) are excessively linearly correlated with each other then autonomous influence cannot be
determined (Halkos, 2011a; Brooks, 2008). What is more, the problem of Multicollinearity
does not occur due to the estimation but due to data and variable specification. (Brooks,
2008).
In order to test for Multicollinearity will can check the R2. A high R2 with low values for t-
statistics of the coefficients indicates Multicollinearity issues. Moreover, the case of (Near)
Multicollinearity can be tested through the correlation matrix, where if: Corr(x i,xj)>=0.80,
then those two variables are highly correlated.
25
Berglund Oscar Zisiadou Argyro
However, in our case, Multicollinearity will not occur due to the fact that the independent
variables will be AR and MA terms and not specific variables that can be linearly correlated
(explain the same thing).
4. Non-normality
The assumption for normality is that the residuals follow the normal distribution. In order to
test for normality we use the skewness, the kurtosis and the Jarque-Bera test (Brooks, 2008;
Halkos, 2011a)
In order for a series to be normally distributed the value for skewness should be equal to zero
(0) and the value for kurtosis should be equal to three (3). The Jarque-Bera test assumes to
follow the Chi-squared distribution.
H0= Residuals are normally distributed
5. Non-stationarity
A common assumption in many time series techniques is that the data is stationary, which
means that the mean value, the variance and the autocorrelation structure do not change over
time.
In order to test for stationarity we can use either the correlogram or the Augmented Dickey-
Fuller test and the Phillips-Perron test.
H0= Mean value, variance and autocorrelation structure are constant over time
6. Non-linearity
The issue of non-linearity describes the situation where the relationship between the
dependent and the independent variables is not a linear one (Brooks, 2008; Halkos, 2011a).
Department of Business Administration- School of Economics and Management – Lund University
That is not a serious problem, because the most obvious solution is to try and estimate the
model with a non-linear approach, such as the exponential one.
In order to test for non-linearity we can the BDS test (Halkos, 2011a).
H0= Residuals are independent and identically distributed (i.i.d)
The ARIMA Model
The ARIMA model (AutoRegressive Integrated Moving Average) is a well-known and useful
estimation, which is widely used in Time Series Analysis. This approach is a generalization of
an autoregressive moving average (ARMA) model. Their effectiveness derives from the fact
that these models can help the researchers or the analysts to understand the available data or
even make predictions (forecasts) based on those series.
The general reference to the model is ARIMA(p,d,q) where p,d and q are the parameters refer
to the order of autoregressive, integrated and moving average terms respectively. This model
constitutes one of the most important parts of the Box-Jenkins approach, which was described
in the Time Series Analysis: Forecasting and Control (1971) book by George Box and
Gwilym Jenkins.
As a rule of thumb, it is widely accepted that the first 80% of the observations are used for the
estimation of the model while the remaining 20% are used for the forecasting procedure.
These are the percentages we are going to use for our estimations as well. The procedure of
estimation, diagnostic testing and forecast will be described in the next subsection.
The ARIMA procedure
The procedure is as follows. We aim to start with descriptive statistics of the market returns so
as to get a sense of the variables we are using. Afterwards we make ARIMA estimations
27
Berglund Oscar Zisiadou Argyro
starting with an ARIMA (10,1,10), meaning we will include AR terms (autocorrelation) from
1 to 10 and MA terms (moving average) from 1 to 10 and 1 integration term because we are
using the returns of the prices generated as the first difference of the logarithm of the prices.
That means:
ARIMA (10,1,10) = maxAR(10), maxIntegration (1), maxMA(10)
and in those estimations we are going to exclude AR and MA terms which are insignificant
with the purpose of reaching the most optimal model by using the AIC and the Loglikelihood
criterion. Note that the lower the AIC, the higher the Loglikelihood, the better the estimation.
Thus, the estimations with the lowest AIC and highest Loglikelihood will be the most
appropriate.The same procedure will be followed both for the GARCH/T-GARCH and E-
GARCH. The residuals will be tested on all these different tests already mentioned. What is
important to mention is that we are going to use 80% of our sample (01/03/2000-12/30/2011)
for the estimations and the last 20% (01/02/2012-12/31/2014) for the forecast.
The results of the forecast will be qualified based on four (4) values: the Root Mean Squared
Error, the Mean Absolute Error, the Mean Abs. Percent Error and the Theil Inequality
Coefficient[1]. Note that the forecast model that gives the lowest values on those criteria or
most of them is the best possible forecast of all.
If the Random Walk (RW) (see Appendix A, eq. 1)appears to be the best forecast estimation
we can assume that the market is efficient in its weak form because there is randomness in the
returns and there is no model that can predict the market movements. Otherwise, we can
assume that the market is inefficient in its weak form.
In this thesis, we are going to investigate three different American capitalization indices
(S&P500, S&P400, S&P600) for a period of 15 years (2000-2014). In addition, we are going
Department of Business Administration- School of Economics and Management – Lund University
to use daily observation for the last 15 years (01/03/2000-12/31/2014). By using different
estimations such as ARIMA, GARCH/T-GARCH, E-GARCH and Random Walk for all three
indices, we aim to find the best forecast method and prove market (in)efficiency. Lazar and
Ureche (2007) investigated the market efficiency of the emerging markets and they proved
that most of the markets in their sample are not efficient in their weak form. The tests used in
that paper were, BDS, LM test, Q-statistics and Runs test. In our case we are going to test the
residuals of the ARIMA models with the following tests: Q-statistics (for serial correlation),
Jarque-Bera (for normality), ARCH test (for heteroskedasticity), LM test (for linear
dependence), BDS (for non-linearity).
3.4.2 Multiples
As touched upon in section 2.3, there are four basic steps to any multiple valuation:
Step 1 consists of selecting a target firm, that is, the corporate entity one wish to value and
selecting the numerator and the denominator of the multiple. The most common selection for
the numerator is usually price or enterprise value. The choice of denominator/value driver is
usually more flexible. Some of the more commonly quoted and used ones are earnings, sales,
assets, EBIT, and EBITDA (see Appendix C, Table C2 for full list of multiples). It should be
noted that Schreiner (2007) bring up the matching principle, that is, value drivers that capture
claims of both debt and equity holders such as EBITDA should be matched with the
corresponding numerator, in this example enterprise value. We have chosen to ignore that
principle in this thesis, as one of our main goals is to find which multiple best approximates
market value. We therefore find it pertinent to examine all possible combinations we have
time for, even if they may violate the matching principle.
29
Berglund Oscar Zisiadou Argyro
Step 2 consists of selecting an appropriate peer group. For this thesis we started by taking the
one hundred (100) largest firms seen to market capitalization of the S&P large-, mid-, and
small-cap indices. We chose these particular indices because of the ease of access to data in
Datastream 5.1. The reason we only included the one hundred largest firms from each index is
simply because time limitations did not allow us to collect and examine any more data. With
our sample at hand, we then assigned each firm their respective SIC code as listed on Nasdaq.
We chose to use the SIC-system solely because its availability. Schreiner (2007) argued that
the GICS or ICB system yield a better estimation of peers than does the SIC system.
However, both the GICS and ICB system are professionally managed and as such there is a
fee to use it.
Step 3 consists of compiling each target firm’s peers’ multiples into a single number. As noted
in section 2.3 there are several ways of doing this, for example using the arithmetic mean,
harmonic mean, or simply taking the median. Ek & Lillhage (2012) and Lie & Lie (2002) use
the arithmetic mean and the median in their papers. Schreiner however, argued that the
arithmetic mean is inappropriate to use when aggregating peer multiples as it relies too
heavily on outliers. He instead recommended using the median or harmonic mean (Schreiner,
2007). We decided to use all three ways in this thesis. Once again, one of our main objectives
is to find which multiple provides the best approximation of market value. We therefore think
it is prudent to employ all of the previously cited methods so nothing is left unchecked. Due to
our sample size we sometimes found that we did not have enough peers for a given SIC-code
to compute a peer multiple. Lie & Lie recommend using a peer sample of at least five firms. If
there are less than five firms for a given industry code, the code is then relaxed one level (Lie
& Lie, 2002). So for example, code 602 would be relaxed to 60 if there were less than five
comparable firms in the 602 industry. Schreiner on the other hand, maintains that the optimal
peer group consists of four to eight peers. Should there be less than four firms for any given
Department of Business Administration- School of Economics and Management – Lund University
level-three industry code he suggests simply making do with the smaller sample as relaxing
the industry code level decreases comparability between the peers (Schreiner, 2007). We
decided to do a hybrid version where we relaxed the SIC-code one level if there were less than
three peers for a given SIC-code. We decided on three because our limited sample had a lot of
instances where there only were three or four companies for a given SIC-code. According to
Alford (1992) the valuation accuracy decreases substantially when moving from an industry-
code level of at least three down to two. So in order to maintain the highest valuation accuracy
possible, we opted for going with Schreiner’s (2007) recommendations instead and to
decrease the number of peers first hand and then relax the SIC-code one level if absolutely
necessary. Just like Lie & Lie (2002) we use one year estimations when compiling the peer
multiples. So peer values at time t are compiled into one number, which is then used in step 4
to calculate the estimated value for the target firm at time t.
Step 4 consists of the actual valuation and is, as noted in Section 2.4, very straight forward.
The calculated peer/SIC-multiple from step 3 is simply multiplied with the target firm’s value
driver in order to arrive at theoretical market price or enterprise value. For example, if the
calculated peer/SIC P/E-multiple is 10 and the target firm’s earnings are 2, we get a
theoretical market price of 20. Note that we get market price and not enterprise value in this
example as the numerator in the multiple, that is, P is price and not enterprise value.
3.4.3 Valuation Accuracy - Statistical Approach
In order to have good comparability with previous papers on multiples valuation, we opted to
use the same method in calculating the valuation error. Kaplan & Ruback (1995), Lie & Lie
(2002), Schreiner (2007), and Ek & Lillhage (2012) all calculate the valuation error by taking
the natural logarithm of the ratio of the estimated value and the observed/market value.
31
Berglund Oscar Zisiadou Argyro
Valuation Error=ln EV∗¿EV
¿
Using this equation we were able to calculate a mean valuation error for each multiple for our
different indices, which arguably provides some insight into which multiple works the best.
However, before assigning the title of best multiple to one of the ratios it is also interesting to
look at what percentage of the sample has a valuation error of less than a certain percentage.
Kaplan & Ruback (1995), Lie & Lie (2002), Schreiner (2007), and Ek & Lillhage (2012)
crown the best multiple by looking at what multiple had the largest part of the sample with a
valuation error of less than 15%. For comparability, we did this as well.
In order to obtain these values, descriptive statistics by classification will be used. As
classification we are going to use the period dimension. Moreover, by using the Simple
Hypothesis Test, the number observations which are placed out of the range of (+/-) 0.15 will
allow the researchers to calculate the percentage of the observations included in the range
over the total number of observations.
3.4.4 Error Determinants - Panel Data Approach
The term Panel Data Analysis is used to describe the econometric approach that can be used
to combine two dimensions in the same estimation. In other words, when in an analysis the
researchers include both cross sectional units and period dimension, which means
observations for more than one year for each firm, they are using the Panel Data Analysis.
This analysis is preferable in many cases as it can provide a solution to more complex
problems than Time Series Analysis or Cross Section Analysis when used separately.
Moreover, by combining both dimensions, the analysts are increasing the degrees of freedom
and as a result the significance of the tests (Brooks, 2008).
Department of Business Administration- School of Economics and Management – Lund University
A panel data can be either balanced or unbalanced. The balanced panel data has no missing
values in periods in any cross-section unit whereas the unbalanced panel dataset has fewer
observations in some cross-section units (Brooks, 2008).
Once again, before the final estimation is specified both the regressions and their residuals
should be tested for a number of different problems that may occur.
Problems may occur in Panel Data Analysis
1. Multicollinearity
With the term Multicollinearity refers to the issue when two or more independent variables (x)
are excessively linearly correlated with each other and an autonomous influence cannot be
determined (Halkos, 2011a; Brooks, 2008). What is more, the problem of Multicollinearity
does not occur due to the estimation but due to data and variable specification. (Brooks,
2008).
In order to test for Multicollinearity we can check the R2. A high R2 with low values for t-
statistics of the coefficients indicates Multicollinearity issues. Moreover, the case of Near
Multicollinearity can be tested through the correlation matrix, where if: Corr(xi,xj)>=0.80,
then those two variables are highly correlated (Brooks, 2008; Halkos, 2011a).
2. Heterogeneity
One of the main problems that appear when using Panel Datasets is Heterogeneity. With the
term Heterogeneity, econometric theory refers to the difference across the cross section units
being examined that can occur when using Panel Dataset. (Brooks, 2008)
H0= no difference across the cross section units (homogeneity)
33
Berglund Oscar Zisiadou Argyro
3. Endogeneity
Another problem that needs to be taken into consideration is the Endogeneity issue. The four
assumptions of the OLS is that there should be no relationship between the error term and the
corresponding x variable or mathematically expressed Cov (ui,t,xi,t)=0. If that assumption does
not hold, then there is a correlation between the error term and the corresponding variable. So
in that case the independent variable is endogenous. (Angrist, Pischke, 2009)
H0= no correlation between the independent variable and the error term (exogenous variable)
There are four different factors that can cause endogeneity to an estimation. Briefly, these
factors are: 1. Omitted Variables, 2. Simultaneity, 3. Measurement Error, 4. Selection Bias.
(Angrist, Pischke, 2009)
4. Heteroskedasticity
One of the assumptions of an OLS estimation is that the variance of the error terms are
assumed to be constant (homoscedastic). However, if the variance of the residuals are not
constant but depend upon the value of x, then the residuals are heteroskedastic. (Halkos,
2011a; Brooks, 2008)
H0= Error term variance is constant (homoscedastic)
5. Autocorrelation
One of the assumptions of an OLS estimation is that the error terms are assumed to be
independent over time. Autocorrelated error terms are the most important consideration for a
Department of Business Administration- School of Economics and Management – Lund University
practicing forecaster because when the problem of the autocorrelation occurs, the estimation
and the forecast will be biased. (Halkos, 2011a; Brooks, 2008)
H0= independent error terms over time,
the error terms are completely random and serially uncorrelated
6. Non-normality
The assumption of normality is that the residuals follow the normal distribution. In order to
test for normality we use the skewness, the kurtosis and the Jarque-Bera test (Brooks, 2008;
Halkos, 2011a)
In order for a series to be normally distributed the value of skewness should be equal to zero
(0) and the value of kurtosis should be equal to three (3). The Jarque-Bera test assumes to
follow the Chi-squared distribution.
H0= Residuals are normally distributed
What are the consequences of those problems?
1. Multicollinearity
In case that two or more independent variables are highly correlated, the regression becomes
very sensitive to small changes in its specification (Brooks, 2008)
2. Heterogeneity
When the sample is heterogeneous and the analysts ignore that problem, then the result will be
a pooled regression. That can lead to another potential problem the poolability. Of course, a
35
Berglund Oscar Zisiadou Argyro
pooled regression will not be an appropriate estimation since the estimated coefficients will be
biased and unreliable (Brooks, 2008).
3. Endogeneity
By ignoring the endogeneity issue, the estimation will be biased. That means that the
outcomes will not be reliable and future expectations can not be based on them (Angrist,
Pischke, 2009).
4. Heteroskedasticity
The non-constant variance on the residuals can lead to biased estimations. That means that the
estimated values do not represent reality, so forecasts should not be based on those
estimations (Halkos, 2011a; Brooks, 2008).
5. Autocorrelation
The estimators that derived from the regression are unbiased but ineffective. Once again, that
means that the estimated values do represent the reality, so forecasts should not be based on
those estimations. Even in large samples, the coefficients are not efficient which means that
they do not have the minimum variance (Brooks, 2008; Halkos, 2011a)
6. Non-normality
There are no serious consequences deriving from the non-normality problem.
How to test for those problems?
1. Multicollinearity
In order to test for Multicollinearity, a correlation matrix is required. Through that matrix it is
easy to identify if two corresponding variables are highly correlated. If two corresponding
Department of Business Administration- School of Economics and Management – Lund University
variables have a correlation value equal to or over 0.80, then near Multicollinearity occurs and
those two variables explain the same thing.
2. Heterogeneity
The formal test for Heterogeneity is the Redundant Fixed Effects – Likelihood Ratio. When
using fixed effects specification in both dimensions, the Redundant Fixed Effects -
Likelihood Ratio test gives two different values, both the F-test and Chi-square test for both
dimensions. If the probability of these values is lower than 0.05 (5% level of significance),
then the null hypothesis is rejected for the specific dimension and there is heterogeneity. It is
possible to have heterogeneity in both dimension or only in one (Brooks, 2008).
3. Endogeneity
The test for endogeneity is one that needs to be done manually because the software used, E-
Views 8, does not have an endogeneity test. The manual test is as follows. Based on theory
the researcher should decide which of the independent variables can be endogenous. This
variable should be used in a new estimation as dependent variables where all the other
independent variables will follow on the RHS. The residuals of the regression should be
saved. These residuals will afterwards be included in the first estimation specification. If the
residuals appear to be significant, which mean that the p-value will be lower than 0.05 and the
t-statistics are higher than 2 in absolute value (based on the Rule of thumb), then the null
hypothesis is rejected and the examined independent variable is endogenous (Angrist,
Pischke, 2009).
4. Heteroskedasticity
In order to test for Heteroskedasticity, we can use the Goldfeld-Quandt test (Halkos, 2011a),
Breusch – Pagan – Godfrey test (Halkos, 2011a), White test (Halkos, 2011a)
37
Berglund Oscar Zisiadou Argyro
However, these tests are not applicable on Panel Data, thus the estimation needs to be tested
manually for Heteroskedasticity. This manual test is as follows. From the estimated
regression, that included the possible heterogeneity and endogeneity specifications, the
residuals should be saved and squared. Those residuals will be used afterwards as dependent
variables followed by all the independent variables in the RHS of the regression but without
the heterogeneity and endogeneity effects. If the probability of the F-statistics from that
estimation are lower than 5%, then the null hypothesis is rejected and the estimation is has
Heteroskedasticity issues.
5. Autocorrelation
In order to test for autocorrelation we can use a number of different tests, such as Durbin-
Watson statistics (DW). A DW value equal to 2 indicates that there is no autocorrelation. On
the other hand, DW value lower than 2 indicates negative autocorrelation and DW value
greater than 2 indicates positive autocorrelation (Halkos, 2011a; Brooks, 2008)
6. Non-normality
The formal test of Normality is the Jarque – Bera test. If the probability of Jarque – Bera is
lower than 0.05, then the null hypothesis is rejected and the residuals are not normally
distributed.
Moreover, when the residuals are normally distributed the value of skewness is equal to zero
(0) and the value of kurtosis is equal to three (3). If these values are not reached, it can be
Department of Business Administration- School of Economics and Management – Lund University
concluded that the residuals do not follow the normal distribution (Halkos, 2011a, Halkos,
2011b, Brooks,2008).
How can the problems be solved?
1. Multicollinearity
The solution to Multicollinearity is a straight forward procedure. Because the two highly
correlated independent variables explain the same thing, it is up to the researchers to decide
which one of the two corresponding highly correlated variables will be excluded (Brooks,
2008).
2. Heterogeneity
If the Redundant Fixed Effects – Likelihood Ratio indicates that heterogeneity occurs in the
estimation, there is a number of different solutions.
The first and most preferable solution is to introduce Random Effects (RE) in the model
because it is the specification that gives more degrees of freedom compared to the Fixed
Effects specification. However, in order to use that specification, the RE need to be well-
specified. To test the specification of RE, the Correlated Random Effects – Hausman Test is
used. The null hypothesis for the test is that the RE are well specified. If the null hypothesis is
rejected, the Random Effects are not a suitable solution for heterogeneity. In that case the
Fixed Effects should be used instead (Brooks,2008).
Last but not least, there is a possibility that combined effects may be required. That means
that one of the dimensions may have well specified RE but the other one does not. However,
if the software cannot run a combined effect regression, the Within Transformation is the
solution to the problem. The same solution applies to when RE are needed in both dimensions
of an unbalanced panel data. Once again, the software cannot estimate a regression with RE
39
Berglund Oscar Zisiadou Argyro
in both dimensions when the panel dataset is unbalanced. Thus, one of the dimensions should
be transformed by using the Within Transformation (Brooks, 2008).
3. Endogeneity
When at least one of the independent variables is endogenous, the Valid Instrument Variables
need to be used in order to mitigate the problem of endogeneity. With the term Valid
Instrument Variable (IV), we actually mean a variable which is partially correlated with the
endogenous independent variable and not correlated with the error term. However, because
the error term is not observable we cannot test the correlation between the IV and the error
term. Instead, the correlation between the dependent variable and the IV has to be tested.
Thus, in that case, it is important that the dependent variable and the error term are
uncorrelated (Angrist, Pischke, 2009).
4. Heteroskedasticity
The solution to the Heteroskedasticity issue is either to transform the variables which will
reduce their size or to make use of the White “Robust” standard errors (Brooks,2008).
5. Autocorrelation
One of the methods that can be used to solve the autocorrelation issue is to include both the
dependent and independent variables in the estimation lags of the variables (Halkos, 2011a).
The term lag is indicating the use of the value from the previous period of the specified
variable.
6. Non-normality
The issue of non-normality can be solved by either transforming the variables through the log
of the variables or by excluding a few outliers. However, in the case that the sample is big
enough that problem can be ignored (Brooks,2008).
Department of Business Administration- School of Economics and Management – Lund University
The Estimation Procedure
The variables we use in this part will be the valuation errors that have been calculated based
on description in Section 3.4.3 as dependent variable and as error explanatory variables the
researchers decided to use some firm-based information that can affect the investors decision
and lead to mis-valuation. Based on Ek & Lillhage’s (2012) findings, some explanatory
variables are the natural logarithm of the total assets, the Market/Book ratio, and the annual
volatility and the R&D intensity (see Appendix A, eq. 2) of each firm. In our case some other
explanatory variables will be included. One of the variables will be a dummy variable
indicating if the credit rating of the firm is investment grade or not, by assigning the value one
(1) if so, and zero (0) otherwise (Moody’s, 2015). The decision to include the credit rating
variable was based on the notion that mis-valuation may be caused by investor bias and how
investors perceive the firm in question. Because the credit rating of each firm in our sample is
easily accessible to investor, it may be a source of information that can influence the
valuation. Another explanatory variable will be the Tobin’s Q (see Appendix A, eq.3), which
is a performance measure for a firm. Specifically, it is total market value of the firm divided
by total asset value of the firm. We included this variable as we believe investors may be
influenced and show some bias due to the historic performance of a firm. Finally, another
variable we believe may lead to mis-valuation is the profit margin. Much like in the case of
Tobin’s Q, we believe that miv-valuation may be caused by investor bias stemming from too
high of a reliance on a firm’s historic performance. Thus, the analysis consists of determining
which of these error variables are significant and in which cases. The analysis is as follows.
The first step is to present the descriptive statistics of the sample to gain a general sense the
variables we are using, which then will be followed by the correlation matrix. This matrix will
give us the opportunity to check for highly correlated variables before we start the
estimations. The next step is to specify the regression. We will use the valuation errors
41
Berglund Oscar Zisiadou Argyro
calculated in Section 3.4.3 as the dependent variables and the previously listed terms, such as
Tobin’s Q as explanatory variables. Through this procedure we will try to identify which of
these explanatory variables can significantly influence the valuation errors. The estimation has
the following expression:
Valuation errori,t= α+β1*X1,i,t+β2*X2,i,t+β3*X3,1i,t+β4*X4,i,t+β5*X5,i,t +β6*X6,i,t+β7*X7,i,t+ui,t
The independent variables are as follows:
X1,i,t: ln (Total Assets)
X2,i,t: Market/Book (M/B)
X3,i,t: Volatility
X4,i,t: R&D intensity (see Appendix A, eq.2)
X5,i,t: Credit Rating
X6,i,t: Tobin’s Q (see Appendix A, eq.3)
X7,i,t: Profit Margin
With the expression at hand, a pooled regression is used in order to attain a benchmark result.
Obviously the benchmark will not be used because that means that the sample is a cross
sectional sample and has no period dimension. The next step is to run the Redundant Fixed
Effects – Likelihood Ratio which will indicate the heterogeneity issue and will specify the
variables that face that problem. Afterwards, Random Effects will be introduced in the
regression because, as explained above, they are more preferable. However, these RE need to
be well-specified which leads us to the Correlated Random Effects – Hausman Test. That test
Department of Business Administration- School of Economics and Management – Lund University
will give us the solution to the heterogeneity issue by indicating the specification needed,
either the Random Effects or the Fixed Effects or perhaps a combined specification.
The next step is the endogeneity testing procedure, which will provide a solution if that
appears in our estimations. Since we do not know which of the seven (7) explanatory
variables that can be endogenous, all of the variables will be tested consecutively and if any
variable appears to be endogenous, then the procedure described in this Section will be used
in order to mitigate endogeneity.
After solving the two main issues of the regression, the estimation will be tested for
heterogeneity and White Robust SE’s will be used if needed. After mitigating any potential
heterogeneity issue, the estimation will be tested for autocorrelation by using the DW ratio
given in the statistics following the estimation. If the DW ratio indicates autocorrelation,
either positive or negative, the lags of both the dependent or independent variables will be
used till a DW close to two (2) is achieved, while at the same time providing the best AIC.
Lastly, the residuals of the final estimation will be tested for normality. In the case that the
residuals are not normally distributed, we can ignore that issue because of the sufficient size
of our sample.
By repeating the above described procedure, the researchers will have the ability to identify
which error variables significantly influence the valuation errors depending on the way the
valuation errors were calculated each time. That means, answers will be given for all different
cases. Specifically, the valuation errors, based on Section 3.4.2, will be calculated for all
different multiples (11 in total), three different approaches (median, average and harmonic)
and all different sample categorization (total sample and each capitalization separately). That
43
Berglund Oscar Zisiadou Argyro
gives a final number of 132 estimations and outcomes that will be presented and discussed in
Section 4.3.
4 Analysis and Discussion
In this chapter we present the results of our different analysis. The chapter is divided into
three main Sections; Market Efficiency, Valuation Accuracy, and Error Determinants.
4.1 Market Efficiency
4.1.1 Descriptive Statistics
Department of Business Administration- School of Economics and Management – Lund University
The results of the descriptive statistics of the market returns are as follow:
Table 4.1.1 Descriptive Statistics of market returns
Indices S&P500 S&P400 S&P600
Mean 0.075913 0.016055 0.086034
Std. Dev 0.248609 0.237141 0.261584
Skewness -0.638427 -0.834145 -0.697049
Kurtosis 3.310569 3.222872 3.290829
JB p-value 0.604320 0.437687 0.553482
Normality yes yes yes
From the table above (Table 4.1.1) we can conclude that the lowest returns are observed in the
mid-cap index with a value equal to 0.016055 or 1.6055% while the highest returns are
observed in Small Cap with a value equal to 0.086034 or 8.6034%. The market risk, measured
by using the standard deviation, indicates that the riskiest of all indices is the Small Cap and at
the same time the least risky is the Mid Cap. The negative skewness indicates that the lower
deviations from the mean are larger than the upper deviations, suggesting a greater probability
of large decreases than rises. What is more, Kurtosis, in all cases, has coefficients greater than
3 (but really close to 3) and Skewness has values close to 0, so we can assume that the returns
seem to follow the asymptotic chi-squared distribution, as required. Normality can also be
proved with the use of Jarque-Bera test and its probability. As we can see, the probabilities, in
45
Berglund Oscar Zisiadou Argyro
all cases, are greater than 0.05, so we cannot reject the null hypothesis in a 5% level of
significance.
4.1.2 Model Approach
The next step is to estimate the most appropriate model based on AIC and the Loglikelihood
Criterion. After repeating the procedure for all different combinations we came to the
conclusion that there is a specification for each model that gives the best estimation. The table
below (Table 4.1.2) presents the results with the best specifications and the terms that have
been excluded:
Table 4.1.2 Best Model Approach
Indices ARIMA GARCH/T-GARCH E-GARCH
S&P500 ARIMA (10,1,10) GARCH (2,2)
With ARIMA (10,1,10)
E-GARCH (2,1)
With ARIMA (10,1,10)
- excluded No No AR(4)
AIC -5.781425 -6.237441 -6.277221
S&P400 ARIMA (10,1,10) GARCH (2,1)
With ARIMA (10,1,10)
E-GARCH (1,2)
With ARIMA (10,1,10)
- excluded AR(4), MA(1), MA(4) AR(3), MA(3) AR(1), AR(8), MA(8)
Department of Business Administration- School of Economics and Management – Lund University
AIC -5.4498221 -5.991598 -6.007784
S&P600 ARIMA (10,1,10) GARCH (2,2)
With ARIMA (10,1,10)
E-GARCH (1,1)
With ARIMA (10,1,10)
- excluded AR(3), AR(5), MA(5) AR(8), MA(8) AR(4), AR(7), AR(8)
AIC -5.567518 -5.789050 -5.809214
4.1.3 Residual Diagnostics
The residuals for each test:
Jarque-Bera: H0: Skewness being equal to 0 and kurtosis being equal to 3
Q-statistics: H0: Residuals have no serial correlation
LM Test: H0: Residual have no autocorrelation
ARCH Heteroskedasticity Test: Ho:Residuals are homoskedastic
BDS Test: H0: Residuals have linearity
The residuals of all these estimations were tested and the results are the following: [1] None
of the residuals meets the normality requirements because their skewness in most cases is far
from 0 and their kurtosis is far form 3. At the same time, the probability is below 5%, which
means that we can reject the null hypothesis of having a skewness equal to 0 and kurtosis
equal to 3; following a normal distribution, so all residuals suffer from non-normality. [2]
Based on Q-statistics we can support the belief that apart from ARIMA estimations and
GARCH (2,2) for S&P500, none of the estimations have a serial correlation problem using up
to 36 lags because the probabilities of Q-statistics are greater than 5% in all cases, with the
exception of ARIMA and S&P500 GARCH (2,2) estimations where the probabilities are
below 5%, so only in these cases the null hypothesis can be rejected and it can be assumed
that there is serial correlation. [3] Regarding the LM test, it should be noted that this test is
47
Berglund Oscar Zisiadou Argyro
applicable only in an ARIMA approach and not in any GARCH approach. For that reason we
have tested the residuals only for the ARIMA estimations. Our conclusions is that all the
ARIMA based LM tests have insignificant residuals, which mean that their t-statistics are
lower than 1.96. Note that when t-statistics is lower than 1.96, there is no significance on 55
level of significance. As an alternative, we can use the Rule of Thumb: significance |t-
statistics|>2 suggests significance. That means that we can reject the null hypothesis of the test
and assume that there is no autocorrelation. [4] The next test is the Heteroskedasticity ARCH
test. With a low probability of F-statistics in that test for ARIMA estimations and E-GARCH
(1,1) for the S&P600, we can reject the null hypothesis of the test and suggest that there is
heteroskedasticity instead. However, we cannot reject the null hypothesis for any other case.
[5] Last test is the BDS test for linearity issues. For all indices and for all different
approaches, there is significance for the test. That means that the probability of the z-statistics
in all dimensions are lower than 0.05 and the z-statistics are greater than 3 in all cases. That
leads us to reject the null hypothesis of the test and come to the conclusion that none of our
estimations have linearity in its residuals.
4.1.4 Forecasting
After presenting the results of the residual diagnostic tests, it is time to present the results of
the forecasts which will give us the answer to the main question of this Section: Is the market
is efficient in its weak form or not? If the market is efficient in its weak form then a Random
Walk should be followed by the price movements and there should be no correlation between
the past and the future price movements, otherwise the market will be inefficient and a
number of investors will be able to gain abnormal returns. As mentioned above, for the
forecast the last 20% of our observations will be used.
Department of Business Administration- School of Economics and Management – Lund University
4.1.4.1 Large CapitalizationThe table below (Table 4.1.3) presents the results from the forecast for all four estimations.
Four different critical values are presented. Based on these values we can conclude if the
market if efficient in its weak form. The model that reaches the lowest values for each critical
value, or at least most of them, is the model that gives the best estimation. For the large cap
(S&P500), we can suggest that because Random Walk has only one of the lowest values of
the four, while GARCH has two and E-GARCH has one, it is not the best model to use order
to describe the market prices. That means that, for the S&P500 index the stock price
movements do not follow the RW, which indicates that there is no market efficiency in its
weak form for the large capitalization.
Table 4.1.3 Large Capitalization Forecasting Results
ARIMA GARCH (2,2)
E-GARCH (2,1)
Random Walk
Root Mean Squared Error
0.007316 0.007274 0.007285 0.007290
Mean Absolute Error 0.005374 0.005324 0.005375 0.005329
Mean Abs. Percent Error 114.6901 145.7314 180.5384 99.20992
Theil Inequality Coefficient
0.936366 0.936503 0.917324 0.995350
49
Berglund Oscar Zisiadou Argyro
4.1.4.2 Mid CapitalizationThe table below (Table 4.1.4) presents the results from the forecast for all four estimations.
Four different critical values are presented and based on those we can conclude if the market
is efficient in its weak form. The model that reaches the lowest values for each critical value,
or at least most of them, is the model that gives the best estimation. For the mid-cap index
(S&P400), we can suggest that because Random Walk has only one of the lowest values out
of the four, while GARCH has three, it is not the best model to describe the market prices.
That means that, for the S&P400 index the stock price movements do not follow the RW,
which indicates that there is no market efficiency in its weak form for the mid capitalization.
Table 4.1.4 Mid Capitalization Forecasting Results
ARIMA GARCH (2,1)
E-GARCH (1,2)
Random Walk
Root Mean Squared Error
0.008669 0.008559 0.008622 0.008566
Mean Absolute Error 0.006528 0.006379 0.006487 0.006398
Mean Abs. Percent Error 120.3124 112.5558 120.9649 98.92613
Theil Inequality Coefficient
0.883868 0.262480 0.908428 0.971812
4.1.4.3 Small CapitalizationThe table below (Table 4.1.5) presents the results from the forecast for all four estimations.
Four different critical values are presented and based on those we can conclude if the market
is efficient in its weak form. The model that reaches the lowest values for each critical value,
or at least most of them, it is the model that gives the best estimation. For the small-cap index
(S&P600), we can suggest that because Random Walk has three out of the four lowest critical
values, while GARCH has only one, is the best model to describe the market prices. That
Department of Business Administration- School of Economics and Management – Lund University
means that, for the S&P600 index the stock price movements do follow the RW, which
indicates that there is market efficiency in its weak form for the small capitalization.
Table 4.1.5 Small Capitalization Forecasting Results
ARIMA GARCH (2,2)
E-GARCH (1,1)
Random Walk
Root Mean Squared Error
0.009616 0.009498 0.009583 0.009450
Mean Absolute Error 0.007376 0.007284 0.007369 0.007232
Mean Abs. Percent Error 118.3928 109.4687 106.9880 98.30536
Theil Inequality Coefficient
0.905076 0.902495 0.915511 0.972271
4.2 Valuation Accuracy
1. What is the average valuation error for each multiple?
In the table below we present the average multiple valuation error for the total sample. All the
valuation errors for the different indices can be found in Appendix B (See Appendix B9-B12).
51
Berglund Oscar Zisiadou Argyro
2. Which multiple gives the best, that is, closest approximation to market value?
For large cap the P/EBITDA 1 calculated with the median proved to be the best multiple
with an average valuation error of only 0.0879% or 8.70 bp, and a standard deviation of
0.393. However, P/E 1 calculated with the median proved to be the best multiple in terms of
Table 4.1.6:Average Valuation Errors - Total Sample
Department of Business Administration- School of Economics and Management – Lund University
percentage of the sample falling within the fraction of 0.15. 57.16% percent of the sample had
a valuation error less than 15% for P/E (1).
For mid cap P/TA calculated with the median proved the have the lowest valuation error on
average of only 0.1199% or 11.99 bp, and a standard deviation of 0.659. In terms of
percentage of the sample falling within the fraction of 0.15, EV/EBITDA 1 proved to be the
best multiple with 49.26% of the sample with a valuation error less than 15%.
For small cap EV/EBITDA 1 calculated with the median was the best multiple with an
average valuation error of only 0.21% or 21 bp, and a standard deviation of 0.5075. In terms
of percentage of the sample falling within the fraction of 0.15, P/E 1 calculated with the
median proved to be the best one with 49.31% of the sample having a valuation error of less
than 15%.
3. On average, are there any notable differences in valuation errors for the different
indices?
Forward-looking earnings based multiples perform the best for all indices, in terms of
percentage of the sample having a valuation error less than 15%. P/S and P/TA perform the
worst in the same regard, which is notably since assets based multiples performed very well in
the study by Lie & Lie (2002). In addition, there’s a slight difference in absolute value of the
valuation errors for the different caps, where large cap has the smallest absolute valuation
error for the best performing multiple, mid cap somewhat larger and small cap somewhat
larger than mid cap. The conclusion we draw from this is that it is harder to value smaller
companies, albeit the difference is very small. Drawing on other knowledge however, it
makes sense. Arguably there is less media coverage and insight into smaller companies than
53
Berglund Oscar Zisiadou Argyro
there generally is for larger companies, which we believe could explain some of this
mispricing.
4. Do equity value multiples outperform entity value multiples in terms of valuation
accuracy?
In terms of accounting for the largest part of the sample within the valuation fraction of 0.15,
equity based multiples performed better than entity based multiples for all indices except mid-
cap where EV/EBITDA 1 performed the best. The difference for the mid-cap index was
however, very small with EV/EBITDA 1 accounting for 49.26% of the sample and P/E 1
accounting for 47.93% of the sample. Both multiples were calculated with the median.
5. Do forward-looking multiples outperform trailing multiples in terms of valuation
accuracy?
Yes, it would seem so. All of our best performing multiples were indeed forward-looking
multiples.
All of our samples, that, is the total sample, large-cap, mid-cap, and small-cap showed
undervaluation for the SIC-multiple compiled by using the median and harmonic mean, while
it showed overvaluation for the average. From an investors perspective it therefore seems
better to use either the median or harmonic mean when compiling the peer multiple, as it has a
tendency to undervalue the stock, essentially resulting in an extra cushion or margin of safety
for said investor. Another nice feature is the fact that the peer multiples calculated with the
median performed the best in all of our tests. From a practical viewpoint this is nice, as the
method is very easy to understand and quick to calculate. In accordance with Schreiner’s
(2007) findings, we also discovered that forward-looking multiples outperform trailing ones.
Department of Business Administration- School of Economics and Management – Lund University
Once again, just like Schreiner we found that the 12-month forward-looking price to earnings
multiple performed especially well.
Total Sample Large Cap Mid Cap Small Cap0.00%
10.00%
20.00%
30.00%
40.00%
50.00%
60.00%
70.00%
All SIC LevelsSIC Level (3 & 4)
Figure 4.1.1: Difference in Valuation Error for Different SIC-levels
Another interesting discovery is that valuation accuracy improved when we excluded the SIC-
multiples calculated at a SIC-level less than three. This is also agreement with Schreiner’s
(2007) argument that the industry code level should only be relaxed when absolutely
necessary, as the similarity of peers is paramount to getting an accurate valuation.
Interestingly, P/TA performed the worst for many of our samples, which is the opposite of
what Lie & Lie (2002) found in their study, where asset based multiples performed the best.
4.3 Valuation Error Determinants
The main questions of this Section are the following:
1. Is there any error variable that can significantly influence the valuation error?55
Berglund Oscar Zisiadou Argyro
2. Is there significant correlation between the valuation error of the present period (t) and
error variable observation from the previous period (t-1).
As already mentioned in Section 3.4.4, the researchers used seven (7) explanatory variables
with the purpose to determine the valuation errors. The initial step was the descriptive
statistics for all valuation errors and explanatory variables. Through this the authors got a
general idea of the sample they have to use, specifically, from looking at the mean values and
the standard deviation. Moving forward, the correlation matrix indicated that there is no
Multicollinearity between the valuation errors and the error determinants. That means there is
no reason to exclude any of the explanatory variables.
The next part was the estimation procedure that started with a pooled (OLS) regression as a
benchmark, which will not be used in the results. The researchers repeated the procedure for
every valuation error calculated in Section 3.4.3 and all the results, both the tests and the final
estimated coefficients can be found in Appendix B (Tables B13 –B45). One of the results,
specifically the estimation of P/E median is presented in the table below (Table 4.1.7)
The table is separated into two parts, where on the left side of the table the reader can observe
all the final estimations of the coefficients and the regression in total, while on the right side
of the table the reader can obtain information about all the tests the researchers used and the p-
values which will lead either to acceptance or rejection of the null hypotheses briefly
described in Section 3.4.4. What is important to mention is that the values of
Heteroskedasticity and Autocorrelation presented on the right side of the table with all the
tests are those initially estimated, before any correction was made, while the values presented
on the left side are the final values after taking into consideration any potential problem and
Department of Business Administration- School of Economics and Management – Lund University
correcting the errors. All the hypotheses were checked in 5% level of significance, which
means that if the p-value is lower than 0.05, the null hypothesis was rejected.
In the table 4.1.7 the level of significance is presented with the use of (*). Specifically,
*indicates significance at the 0.1level, **indicates significance at the 0.05 level and ***indicates
significance at the 0.01 level.
Focusing the discussion on table 4.1.7, it is easily recognizable that the greatest fit is given in
the Small Cap index, which has a fit of 51.0834%. That means that almost 51.1% of the error
variables can explain the valuation error of P/E median.
Moreover, in none of the four (4) categories Total Assets were found to be significant in terms
of explaining the valuation error. The M/B ratio appears to be significant only in the Large
capitalization index and seems to have a negative impact on valuation error. That means that
57
Table 4.1.7: P/E Median Error Determinants & Diagnostic Tests
Berglund Oscar Zisiadou Argyro
the higher the M/B, the lower the valuation error and vice versa. What appears to be
significant in all cases, or almost all cases, is the R&D intensity variable and Tobin’s Q. Both
of them have a negative effect on the valuation error just like M/B. Moving now to the lags, it
is obvious that the Credit Rating from the previous period can cause a significant valuation
error both in the Large Cap and Small Cap indices. Due to the fact that lags were used only in
order to solve Autocorrelation, some of the lags were not included in the estimation because
they were either exaggerating the problem or they were giving a higher AIC. Compared to the
current period, there is still significance on Tobin’s Q, although only in the total sample
estimation. R&D intensity, however, showed no significance compared to the current period.
Not surprising, the Profit margin from the previous period is significant, which lead us to the
conclusion that the investor decisions can be influenced by the previous period profits. The
difference between the two periods is that the significant explanatory variables from time t
appear to reduce the valuation error; whereas the significant explanatory variables from time
t-1 appears to worsen the valuation error. What is essential to mention is that the current
valuation error can be influenced by the previous periods valuation error.
Pertaining to the overall results, initially, all the estimations were suffering from
autocorrelation and a high percentage of the estimations were facing Heteroskedasticity
issues. None of the estimations had endogenous variables, which means that the problem of
endogeneity did not appear and the researchers did not have to deal with that issue. Finally,
all the residuals from the final estimation are not normally distributed, but due to the large
sample size that is an issue that can be ignored.
The graph below (Graph 4.1.2) present the percentage of significance of all coefficients
estimated for time t.
Department of Business Administration- School of Economics and Management – Lund University
Constat_term
ln(TA) M/B Volatility R&D int dummy CR
Tobin's Q Profit Margin
0.00%
10.00%
20.00%
30.00%
40.00%
50.00%
60.00%
70.00%
80.00%
Total SampleLarge Cap.Mid Cap.Small Cap.
Figure 4.1.2: Percentage of significance of Error Determinants (time t)
As we can see, in general, the highest significance can be found in Tobin’s Q and the Profit
Margin with the Total sample and Large Cap index reaching the highest values, where the
credit rating from that period is the one that has the lowest percentages.
These percentages were calculated based on the number of coefficients that were significant at
5% level of significance over the total number of coefficients (33 in time t).
The next graph (Graph 4.1.3) presents the significance of the coefficient of the previous
period.
In comparison with the previous graph, it is recognizable that the most significant of all
coefficients appear to be previous year’s valuation error, which actually reaches 100% for the
total sample, and almost 40% for the Small Cap, the lowest one of all four. The Total Assets
59
Berglund Oscar Zisiadou Argyro
come in second place with more than 63% significant coefficients on the Large Cap index,
which is followed by Volatility and Tobin’s Q. However, what is important to mention is that
both volatility and R&D intensity do not have any significance in the Small Cap index for
time t-1 while at the same time, Tobin’s Q has the lowest value of all existing percentages in
the Small Cap index. In time t-1, the percentage was calculated based on the number of
significant coefficients divided by the total number of coefficients, which differs since some
of the lags were excluded.
Lastly, we have a graph presenting the Goodness of Fit (R2). In this graph (Graph 4.1.4), the
researchers are trying to present the percentage of Goodness of Fit that goes under each of the
three categories. The three categories are as follows:
0%=<R2< 40% Low fit
40%=<R2<80% Medium fit
ln(TA)(-1) M/B(-1) Volatility(-1)
R&D int (-1)
dummy CR (-1)
Tobin's Q (-1)
Profit Margin (-
1)
Valuation Error(-1)
0.00%
10.00%
20.00%
30.00%
40.00%
50.00%
60.00%
70.00%
80.00%
90.00%
100.00%
Total SampleLarge Cap.Mid Cap.Small Cap.
Department of Business Administration- School of Economics and Management – Lund University
80%=<R2=<100% High fit
Total Large Mid Small0%
10%
20%
30%
40%
50%
60%
70%
80%
90%
100%
80%=<R^2=<100%40%=<R^2<80%0%=<R^2<40%
Figure 4.1.4: Goodness of Fit
As we can see from the graph, in the Total Sample 50% of the estimations have a medium
goodness of fit, while only 15.63% of estimations are reaching a high R2. In the Large Cap
index the percentage of the medium fit increases at 54.55% while at the same time the
percentage of the high fit increases at 24.24% which leads to a decrease on the low fit. The
highest percentage of the low fit is observed on the Mid Cap index with a value of 63.64%,
which at the same time we see really low percentage of the high fit, only 9.09%. Finally, the
Small Cap index has a high percentage on the high R2, leading to equal percentages to the
other two categories (low and medium) with a value of 24.24%.
Once again, in order to calculate the percentages the researchers used the number of observed
values under each category over the total number of observations, which equals 33. However,
61
Berglund Oscar Zisiadou Argyro
for the Total Sample, the total number of observation is not 33 but 32. The reason for that is
the fact that one of the estimations in the category has an R2 equal to 100%, which is the P/TA
harmonic estimation (see Appendix B, table B30). Such a high Goodness of Fit is not
reasonable, because it actually means that the explanatory variables can completely explain
the valuation error and there are no standard errors in that estimation. That derives from the
formula of R2 (see Appendix A, eq. 4). In order to have a reasonable estimation on the
percentages presented above and more realistic results we decided that it would be fair to
exclude such an irrational value, which is probably not reflecting reality.
With this said, it is respectable to suggest that there are error variables that can explain the
valuation error. That means that the valuation errors are not random but they are influenced by
other factors. These error variables can be values from the current period but also values from
the previous period. Moreover, the valuation error by itself from the previous period can have
an impact on the current valuation error. Based on the Credit Rating dummy, we can suggest
that the investors’ decisions can be influenced by investment credit ratings and cause an
increase in the valuation error.
5 Conclusion
Our analysis suggest that earnings-based multiples are the best multiples to use when valuing
firms through the multiples approach, that is, earnings-based multiples yield the closest
approximation of market value. We found that this is true both for equity and entity multiples,
where P/E 1 performed very well for equity multiples and EV/EBITDA 1 performed very well
for entity multiples. Interestingly, one of our hybrid multiples performed very well for the
Department of Business Administration- School of Economics and Management – Lund University
large-cap index, namely P/EBITDA 1, which is rather surprising since both Schreiner (2007)
and Ek & Lillhage (2012) are adamant about the importance of matching theory. Overall,
equity-based multiples, that is, multiples with price or capitalization in the numerator
outperformed enterprise-based multiples and forward-looking multiples, that is, multiples that
incorporate next year’s expected earnings, outperformed trailing ones. Both of these findings
are in accordance with what Schreiner (2007) discovered. Schreiner’s (2007) best performing
multiple was P/E 2, and ours was P/E 1. We expect that P/E 2 would have performed even
better than P/E 1 in our sample as well, unfortunately earnings estimations for the next two
years were not available in Datastream 5.1.
Regarding the different indices, we discovered that the average valuation error for our best
performing multiples increased as we moved from the large-cap index down through and
towards the small-cap index, suggesting that valuation accuracy becomes worse for smaller
firms. The error, however, was marginal and the best multiples were still forward-looking,
earnings-based equity multiples suggesting that this is still the best way to value firms through
the multiples approach, regardless of company size.
We can also conclude that using the median when compiling the peer multiples is the best
approach (see section 3.4.2). All of our best performing multiples were computed with the
median. The median also tended to have negatively biased results, which is a nice feature for
investors on the buy side. Lastly, in accordance with Schreiner’s (2007) argument, we found
that the valuation errors decreased when we excluded the multiples computed with a SIC-level
of 2 or 1, confirming that similarity of peers is highly important in order to achieve high
valuation accuracy.
Regarding the market efficiency theory, we can conclude that our econometric analysis
yielded results in line with previous research findings. Out of the three indices we analyzed;
63
Berglund Oscar Zisiadou Argyro
that is, the S&P large-, mid-, and small-cap, only the small-cap index is efficient in its weak
form. Our results suggest that for the other two; that is, the large- and mid-cap index, there are
approaches which can explain price movements and let the investors exploit the market
inefficiency.
For the valuation error determinants, we can conclude that there is a significant connection of
the valuation error with both current and past period observations. Moreover, we found that
the Large Cap index is has better Goodness of fit compared to the other two capitalizations.
Contradictory to Ek and Lillhage’s (2012) findings, we can prove that R&D intensity has
significance in some of the estimations; whereas in their study R&D intensity has no
significance. That leads us to conclude that investors active in the US markets are influenced
by the R&D intensity; whereas the Scandinavian investors seem to not be influenced by the
R&D intensity.
5.1 Practical Implications
We believe our findings can be of interest for investors, both institutional and private, in their
endeavors to invest successfully. After all, price paid for any given financial instrument is one
of the most important variables of a successful investment: in order to make sure it is a fair
and appropriate price, a proper valuation is needed. We also encourage students and
academics to continue the research within this field.
Department of Business Administration- School of Economics and Management – Lund University
5.2 Future Research
Including this thesis, there have now been studies looking at multiples valuation for American
and European large cap stocks. To the best of our knowledge, this is the first study to look at
mid-cap and small-cap firms. However, we believe that there is still room for more research.
For example, it would be interesting to see a more comprehensive study, incorporating more
firms - for example the whole S&P small cap. In addition, there is to our knowledge still no
study looking at Asian companies or emerging markets. It would also be interesting to see a
study that focused more on the difference in valuation accuracy between various industries.
Regarding the error determinants, we believe that there is still room to improve on the model
and add on more error determinants. It will also be interesting to see a study with more,
unconventional multiples included, more specially, one will emerge that can reduce the
valuation error for science-based firms.
65
Berglund Oscar Zisiadou Argyro
References
Alford, A. W. (1992). The Effect of the Set of Comparable Firms on the Accuracy of the Price-Earnings Valuation Method, Journal of Accounting Research, vol. 30, no.1, pp. 94-108, Available online: http://www.jstor.org/stable/2491093?seq=1#page_scan_tab_contents [Accessed 3 May 2015]
Angrist, J.D., & Pischke, J.S. (2009). Mostly Harmless Econometrics, An Empiricist’s Companion. NY: Princeton University Press
Chan, C.K., & Gup, B.E, & Pan, M.S. (1992). An empirical analysis of the stock prices in major Asian markets and the U.S., Financial Review, vol. 17, no. 2, pp. 289-307 Available online: http://onlinelibrary.wiley.com/doi/10.1111/j.1540-6288.1992.tb01319.x/pdf [Accessed 15 March 2015]
Damodaran, A. (2006), Damodaran on Valuation: Security Analysis for Investment and Corporate Finance, New York, USA, Wiley Finance
Demirakos, E.G, & Strong, N.C., & Walker, M. (2004). What Valuation Models Do Analysts Use?, Accounting Horizons, vol. 18, no. 4, pp. 221-240 Available online: http://eds.b.ebscohost.com.ludwig.lub.lu.se/eds/pdfviewer/pdfviewer?sid=859cc27e-6556-452e-aff0-08d4ec049f90%40sessionmgr110&vid=1&hid=126 [Accessed 22 March 2015]
Eberhart, A. C., Maxwell, W. F., & Siddique, A. R. (2004). An examination of long‐term abnormal stock returns and operating performance following R&D increases, The Journal of Finance, vol 59, no. 2, pp. 623-650
Ek, D., & Lillhage, J.(2012). Valuation using Multiples - Accuracy and Error Determinants, Unpublished Master’s thesis, Business Management Project, Lund University, School of Economics and Management Library
Fama, E. (1970).Efficient Capital Markets: A Review of Theory and Empirical Work The Journal of Finance, vol. 25, no. 2, pp. 383-417
Gu, A.Y (2004). Increasing market efficiency: Evidence from the NASDAQ, American Business Review, vol. 22, no. 2, pp. 51-64
Halkos, G. (2011a). Econometric Theory and Practice: Instructions in using E-Views, Minitab, SPSS and Excel (ed.) Athens, Greece: Gutenberg
Halkos, G. (2011b). Statistical Theory, Applications and Use of Statistical Programs in PCs, (ed.) Athens, Greece: Tipothito,
Koller, T., Goedhart, M., Wessels, D. (2010). Valuation: Measuring and Managing the Value of Companies, McKinsey & Company, New Jersey, USA, John Wiley & Sons, Inc
Department of Business Administration- School of Economics and Management – Lund University
Lazar, D., & Ureche, S. (2007). Testing for Efficiency of the Stock Market in Emerging Economies, The Journal of the Faculty of Economics and Business Administration, Babes-Bolyai University, Available online: http://steconomice.uoradea.ro/anale/volume/2007/v2-statistics-and-economic-informatics/24.pdf [Accessed 3 March 2015]
Lev, B., Sarath, B., & Sougiannis, T. (2005). R&D Reporting Biases and Their Consequences, Contemporary Accounting Research, vol. 22, no. 4, pp. 977-1026, Available online: http://onlinelibrary.wiley.com/doi/10.1506/7XMH-QQ74-L6GG-CJRX/abstract [Accessed 29 April 2015]
Lie, E.G., & Lie, H.J., (2002). Multiples Used to Estimate Corporate Value, Financial Analyst Journal, vol. 58, no.2, pp. 44-54, Available online: https://www.biz.uiowa.edu/faculty/elie/ValByMult.pdf [15 March 2015]
Moody’s (2015). Credit Ratings, Research, Tools and Analysis, Available online: https://www.moodys.com [Accessed 27 March 2015]
NASDAQ (2015). American Stock Exchange Available online:http://www.nasdaq.com
Schreiner, A. (2007). Equity Valuation Using Multiples: An empirical investigation, Germany: Deutscher Universitats-Verlag,
Yoo, Y.K. (2006). The Valuation Accuracy of Equity Valuation Using a Combination of Multiples, Review of Accounting and Finance, vol. 5, no. 2, pp.108-123, Available online: http://www.emeraldinsight.com/doi/full/10.1108/14757700610668958 [Accessed 15 March 2015]
67
Berglund Oscar Zisiadou Argyro
Appendix AExpressions
RandomWalk :StockReturns=c+Stock Returns−1+u i (1)
R∧Dintensity= R∧DSales (2)
Tobi n' sQ=Total Market V alueTotal Assets (3)
R2=1− SSESST (4)
Department of Business Administration- School of Economics and Management – Lund University
Appendix B
69
B1 – B8: Valuation Errors - Descriptive Statistics
Department of Business Administration- School of Economics and Management – Lund University
B10: Average Valuation Errors - Large Cap.
73
Department of Business Administration- School of Economics and Management – Lund University
B12: Average Valuation Errors – Small Cap.
75
Berglund Oscar Zisiadou Argyro
B13: P/E Median Error Determinants & Diagnostic Tests
B14: P/E Average Error Determinants & Diagnostic Tests
Department of Business Administration- School of Economics and Management – Lund University
77
B15: P/E Harmonic Error Determinants & Diagnostic Tests
B16: P/S Median Error Determinants & Diagnostic Tests
Berglund Oscar Zisiadou Argyro
B17: P/S Average Error Determinants & Diagnostic Tests
B18: P/S Harmonic Error Determinants & Diagnostic Tests
Department of Business Administration- School of Economics and Management – Lund University
79
B19: P/EBITDA Median Error Determinants & Diagnostic Tests
B20: P/EBITDA Average Error Determinants & Diagnostic Tests
Berglund Oscar Zisiadou Argyro
B21: P/EBITDA Harmonic Error Determinants & Diagnostic Tests
B22: P/EBIT Median Error Determinants & Diagnostic Tests
B23: P/EBIT Average Error Determinants & Diagnostic Tests
B24: P/EBIT Harmonic Error Determinants & Diagnostic Tests
B25: P/(EBIT+RD) Median Error Determinants & Diagnostic Tests
B27: P/(EBIT+RD) Harmonic Error Determinants & Diagnostic Tests
B29: P/TA Average Error Determinants & Diagnostic Tests
Department of Business Administration- School of Economics and Management – Lund University
81
B31: EV/EBITDA Median Error Determinants & Diagnostic Tests
B32: EV/EBITDA Average Error Determinants & Diagnostic Tests
B33: EV/EBITDA Harmonic Error Determinants & Diagnostic Tests
B34: PEV/EBIT Median Error Determinants & Diagnostic Tests
B35: EV/EBIT Average Error Determinants & Diagnostic Tests
B36: EV/EBIT Harmonic Error Determinants & Diagnostic Tests
B37: P/e (1) Median Error Determinants & Diagnostic Tests
Berglund Oscar Zisiadou Argyro
B39: P/E (1) Harmonic Error Determinants & Diagnostic Tests
B40: P/EBITDA (1) Median Error Determinants & Diagnostic Tests
B41 : P/EBITDA (1) Average Error Determinants & Diagnostic Tests
B42: P/EBITDA (1) Harmonic Error Determinants & Diagnostic Tests
B43: EV/EBITDA (1) Median Error Determinants & Diagnostic Tests
Department of Business Administration- School of Economics and Management – Lund University
83
B45 : EV/EBITDA (1) Harmonic Error Determinants & Diagnostic Tests
Berglund Oscar Zisiadou Argyro
Appendix C
C1: Variable Codes
Variable Code Variable Code
Assets WC02999 Market/Book PTBV
Volatility WC08806 R&D WC01201
Enterprise Value WC18100 EBIT WC18191
EBITDA WC18198 Sales DWSL
Earnings WC07250 Stock Price P
P/S FTSPS Profit Margin WC08316
S&P400 S&PMID P/E PE
10-year gov index U.S BMUS10Y Capitalization MV
12-month forward Enterprise Value DIEV S&P500 $SPX
12-month forward P/E PEFD12 S&P600 S&P600I
12-month forward EBITDA EBD1FD12 10-year treas. bond U.S S00311
C2: List of Multiples
List of MultiplesP/EP/E (1)P/SP/EBITP/(EBIT+RD)P/EBITDAP/EBITDA (1)P/TAEV/EBITEV/EBITDAEV/EBITDA (1)