Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Authors D av i d G e l t n e r a n d D av i d C . L i n g
Abstract This paper surveys some of the major technical issues in the
design and construction of real estate research indices, both
appraisal-based and transactions-based. The paper considers
property sampling issues, differences between transaction prices,
market values, and appraised values, the trade-off between
random measurement error and temporal lag bias, optimal
reporting and property revaluation frequencies, and the uses and
limitations of modern statistical techniques. Although one of the
conclusions of the analysis is that most research questions are
best addressed with transactions-based, rather than appraisal-
based, indices in the United States, the paper suggests how
appraisal-based indices can still be useful for some research
purposes.
The past decade has seen the development of technological and information
advances that hold great potential for moving investment real estate performance
measurement to a new level of accuracy and usefulness for the industry. In
particular, the development of large-scale electronic databases of commercial
property transaction prices, and the advance of econometric techniques honed by
the academic real estate community offers a tremendous new opportunity to
advance the level of information and knowledge about the commercial real estate
asset class.
Geltner and Ling (2001, 2006) conclude that the real estate investment industrys
needs for performance measurement, research, and decision support are too diverse
to be optimally met by a single index product or a single type of index. They
present arguments for creating two separate families of index products: one
focused on the asset class research support role, the other focused on the
evaluation benchmarking and performance attribution support role.
Although Geltner and Ling (2001, 2006) discuss the general characteristics of
benchmark and research indices, the present paper focuses more narrowly, but
more deeply, on some basic technical considerations associated with the design
and construction of commercial real estate return indices for asset class research
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 1 2 G e l t n e r a n d L i n g
D e n s i t y o f Tr a n s a c t i o n s v e r s u s R e p o r t i n g F r e q u e n c y
For a given index reporting frequency, statistical quality improves the greater the
frequency of property revaluation, the greater the number of properties in the
population, and the more statistically efficient and effective is the index
construction methodology employed. Holding these characteristics constant, the
greater the index return reporting frequency, the lower the dynamic statistical
quality per reporting period (as the underlying sample density, n, shrinks
accordingly). Thus, broadly speaking, there is a trade-off between statistical
quality per period and the frequency of index reporting, holding constant the
overall quantity and quality of raw valuation data and index construction
methodology. This trade-off is depicted in Exhibit 1a.
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 1 4 G e l t n e r a n d L i n g
E x h i b i t 1a The General Trade-off between Index Statistical Quality Per Period, and the Index Reporting
Increase Statistical Quality Frequency
Per Period
On the other hand, the usefulness of an index for research purposes clearly
increases the greater the frequency of reporting, holding statistical quality constant
(per period). Thus, the usefulness, or utility, of a research index is two-
dimensional. It is likely that there is diminishing marginal substitutability between
these two dimensions: statistical quality per period, and the frequency of periods.
Thus, utility isoquants are convex, as indicated in Exhibit 1b.
The result is that the optimal index reporting frequency, in general, may be either
ambiguous or not unique, as suggested in Exhibit 1c. Multiple indices optimally
designed for different reporting frequencies may be useful for representing and
studying the same underlying population of properties. For example, it might be
useful to have both a quarterly index, optimally designed as such, and a separate
annual index, optimally designed for that frequency, that is not simply a temporal
aggregate (such as compounded growth) of the quarterly returns from the quarterly
index.
The frequency of valuation observations per property is generally beyond the
control of the index producer in a transactions-based index. This is because
C o n s i d e r a t i o n s i n t h e D e s i g n a n d C o n s t r u c t i o n 4 1 5
E x h i b i t 1b Index Utility Isoquants: Index Utility Increases With Both Statistical Quality Per Period and
Reporting Frequency: U0 U1 U2; With Diminishing Marginal Substitutability (convex isoquants)
Increase Statistical Quality
Per Period
U2
U1
U0
properties are sold at the discretion of their owners, not for the sake of providing
input data for an index. As a result, in a transaction-based index, only three of
the four quality-determining attributes can be manipulated: return reporting
frequency, scope of the subject property population, and index construction
methodology. On the other hand, the frequency and timing of reappraisals might
be determined to some extent by policy set by the index producer in the case of
an appraisal-based index that tracks a population of properties that is regularly
marked to market. However, appraisals are costly (as is the collection of
transaction price data), so there are limits to how much the index producer can
control valuation observation frequency, even in the case of an appraisal-based
index.
In the remainder of this paper, it is generally assumed that the index reporting
frequency (m), and the valuation observation frequency (), are determined
exogenously. The focus will be on the other two dimensions of index design: the
optimal index construction technology, and the optimal definition of property
population scope.
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 1 6 G e l t n e r a n d L i n g
U2
U1
U0
S o m e B a s i c C o n c e p t s : Tr a n s a c t i o n P r i c e , A p p r a i s e d
Va l u e , a n d M a r k e t Va l u e
Transaction prices and appraised valuations (or appraisals) are empirically
observable values, but occur and exist only when a property transacts or is
appraised. In contrast, market values are conceptual constructs, also referred to as
true value. Market value exists for each property at each point in time, although
any given propertys (or a portfolio of properties) market value generally changes
continuously through time because information arrives continually that is relevant
to property values.
The market value of a property is frequently defined in real estate as the most
likely (or the expected) transaction price of the property, as of a given point in
time. It may therefore be thought of as the mean of the ex ante transaction price
probability distribution as of the stated date. Market value is therefore the
opportunity cost of holding onto the property rather than selling.
Market value should also closely approximate the actual transaction price in a
highly liquid, dense market where homogeneous assets are frequently bought and
C o n s i d e r a t i o n s i n t h e D e s i g n a n d C o n s t r u c t i o n 4 1 7
sold by numerous buyers and sellers, such as the trading of equity shares in the
stock market. In such an environment, market values represent market clearing
prices at which the number of buyers equals the number of sellers for
homogeneous, divisible assets.
In private real estate markets, market values are not empirically observable, unlike
transaction prices or appraised values. This is because whole assets must be bought
and sold, and these lumpy assets are unique and infrequently traded, and
exchanged in a private transaction between two parties. The inability to observe
market values, however, does not imply they do not exist.
The classical model (whose roots go back at least to Plato) is that observable
transaction prices are individual draws from underlying probability distributions
that are centered around the unobservable true market values of the properties
being transacted. Transaction prices at any given point in time therefore exhibit
cross-sectional dispersion around the underlying (unobservable) true market values
as of that point in time. This difference between the observable transaction price
and the unobservable true value is often referred to as transaction price noise,
or transaction price error. Cross-sectional dispersion occurs because market
participants cannot observe the true market value of the property being traded, so
neither side in the negotiation knows exactly at what price the property should
trade. In a given transaction, one side or the other will typically end up getting
the better deal, although it will usually be impossible to know which side it is.
Either the buyer or seller may have better information, be under a little less
pressure to close the deal, or simply be more skillful or lucky in the negotiation.
Appraised values are also dispersed cross-sectionally around true market values,
as of any given point in time. If two appraisers are employed to value a property,
and they are prevented from communicating with each other, they will almost
certainly not arrive at the same estimate of market value. At least one of them
must be wrong in the sense that their valuation differs from the true market
value of the subject property. In fact, both appraisers are probably wrong in
this sense. The difference between a given appraised value and the (unobservable
true) market value is called appraisal error, although there is no implication that
the appraiser has exhibited any incompetence, negligence, or impropriety.
Although appraised values are dispersed around the underlying true values, unlike
transaction price dispersion, the appraisal value dispersion is not necessarily
centered on the true value. In other words, the ex ante mean appraised value may
not equal the true market value as of the same point in time. Such bias may result
from very rational behavior on the part of the appraiser given the nature of the
empirical information available in the real estate market. A major bias that is
likely to exist is that appraised values lag true contemporaneous market values.
This is referred to as temporal lag bias.2 Normative and behavioral models
consistent with the temporal lag bias hypothesis of appraisal have been put forth
in the academic literature (e.g., Quan and Quigley 1991; and Chinloy, Cho, and
Megbolugbe 1997).
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 1 8 G e l t n e r a n d L i n g
A S i m p l e E x a m p l e o f Te m p o r a l L a g B i a s
While there may be more than one source or type of temporal lag bias in
appraisals, the following illustrates what is probably the major hypothesis for
commercial property indices. Suppose you hire an appraiser who faces the
following choice. The appraiser can give you a value estimate based on Method
A, which will be unbiased, but this value estimate has only a 50% chance of
being within 10% of the true market value. Alternatively, the appraiser can provide
a value estimate based on Method B, which is biased. In particular, the expected
value of the Method B appraisal equals the true market value from six months
ago, rather than the true value of today. Thus, if true market values were 1% lower
six months ago, Method B has a 1% bias on the low side. However, there is a
75% chance that the Method B appraisal will be within, say, 5% of the true market
value of the property. Furthermore, Method B will provide you with more solid
historical evidence explicitly documenting the estimated value (e.g., more
comps). The appraiser says he will charge you the same price for either method.
Which would you prefer?
The most typical answer is Method B, because it will probably give you a more
useful estimate of value, provided you are concerned about documenting the value
of the individual property that is the subject of the appraisal. Said differently, you
are probably willing to accept a little bit of temporal lag bias in order to get a
more precise and well-documented estimate of market value. Of course, if you
own the property and are in the process of negotiating a sale price, you would
probably hesitate to sell for a price equal to the appraised value because you
would probably be aware, in this example, that market values have been increasing
over the past half year and, furthermore, that the appraisal likely contains some
error. You would try to sell for as high a price as you could, above the appraisal
if possible.
However, if you are not concerned about accurately estimating the value of a
single property, but rather primarily interested in obtaining an up-to-date estimate
of the aggregate value of a large portfolio of properties, then you should prefer
the result of Method A for the individual property in question. This is because
the purely random valuation error in Method A will be diversified away at the
portfolio level, thanks to what is known in basic statistics as the Square Root of
N Rule. In contrast, the temporal lag bias in Method B is systematic, and will
not be diversified away at the portfolio level. But the standard practices and
procedures of the appraisal profession in the United States, and the typical method
of employing and using appraisers in the institutional real estate investment
industry in the country, probably results in appraisal methods more similar to
Method B than A.
C o n s i d e r a t i o n s i n t h e D e s i g n a n d C o n s t r u c t i o n 4 1 9
A S i m p l e M o d e l o f P r o p e r t y Va l u e E s t i m a t i o n f o r A s s e t
Class Research
A simple stylized model of the property value estimation process as viewed from
the perspective of asset class research follows. The objective of this process is to
estimate the market value (as defined in the previous section) of a specified
population of properties (the market or asset class of interest) or,
equivalently, the market value of a representative property within that population,3
as of each period of time, t. True market value is defined as Vt. Suppose that
within each period of time, n properties are sold with observable transaction
prices. These transaction prices are labeled Pi,t, for i 1, 2, . . . , n.4 Each
transacting property is an equally valid representative of the type or class of
properties in the population of interest. As described in the preceding section,
these n observable transaction prices are dispersed randomly around the true
market value:
where i,t is the random error or noise term, which is iid normal with zero mean
and standard deviation .5
Now consider the arithmetic average across N of the transaction prices starting
with transactions that are among the n that occur within period t. This estimator
is V t[N]:
P .
N
1
V t[N] i,t (2)
N i1
Basic statistics relates that V t[N] will have a purely random standard deviation (or
standard error) of / N. That is, the purely random error (or noise)
component in the market value estimate is reduced according to the Square Root
of N Rule. Furthermore, since each transacting property is an equally valid
representative of the population, if the N transaction observations all occur within
period t (which is only possible if N n), then V t[N] will also have a mean of
Vt. In this case, V t[N] will be an unbiased estimator of the current (period t) market
value of the population.6
Now consider what happens if the sample size N is greater than n, the number of
transaction observations in any one period t. The random error magnitude of the
estimator will still equal / N, but the mean will no longer be Vt. Some of the
N transaction price observations will have to be drawn from one or more earlier
periods in time; for example, from t1, or perhaps even from t2, and t3, and
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 2 0 G e l t n e r a n d L i n g
so on. This will cause the mean of V t[N] to be some average of Vt and previous
market values of the property population: Vt1, and perhaps Vt2, Vt3, and so on.
This will induce temporal lag bias into the V t[N] estimator.
Given a population transaction density of n per period, the policy question
becomes: how shall N be selected so as to develop the estimate of Vt each period?
Clearly there is a trade-off between the two types of error that will exist in the
estimates. The larger the N, the smaller will be the purely random error whose
magnitude is: / N. However, increasing N beyond n will add temporal lag bias.
The ratio N/n determines how far back in time to reach for the estimation sample.
If N 2n, the transaction observations are drawn from two periods of time, t and
t1. To simplify the analysis of the policy question without loss of generality,
assume a sample size would never be chosen that is less than a whole multiple
of n. That is, if the decision is to go back to a given period of time in order to
increase the transaction sample size, all transactions within that period will be
included. Thus, the ratio N/n will be an integer, and the maximum lag in the
estimation sample, L, will be one less than the N/n ratio: L N/n1. The general
expectation of V t[N] can now be specified as:
V
N / n1 L
1 1
E[V t[N]] Vts . (3)
N/n s0 L1 s0
ts
This implies that there will be L/2, or (N/n1)/2, periods of lag bias in the
estimator.
To use this model to decide the optimal sample size (and hence, the optimal lag
bias, if any), an objective function needs to be specified. A logical one often used
in statistics is to minimize the mean squared error (MSE) in the estimate, where
the error is the total error, including both the random noise and the lag bias. That
is, N is selected (or equivalently L) so as to minimize the expectation:
E (P V ) (P V )
1
N
n
i1
i,t t
2n
in1
i,t1 t
. . . (P V )
N 2
i,tL t (4a)
iNn1
N
n (V V )
N N / n1
1
2
2
i
2
ts t
2
i1 s1
(n(L1))
n (V V ) .
n(L1) L
1
2
2
i
2
ts t
2
i1 s1
C o n s i d e r a t i o n s i n t h e D e s i g n a n d C o n s t r u c t i o n 4 2 1
If, for illustrative purposes, the true market value follows a random walk: Vt
Vt1 rt, where rt is iid with zero mean,7 and the true market volatility is r,8
this objective function simplifies to:
VAR[i] n2VAR[rt] N / n1 2 2 n2 r2
N / n1
MSE s s2
N N2 s1 N N2 s1
r
2 2 L
s2. (4b)
n(L 1) (L 1)2 s1
The MSE thus consists of two terms. The first term on the RHS of Equation 4b
is the purely random error effect, which consists of the cross-sectional price
dispersion variance 2 times a noise factor, 1/(n(L1)), that is diminishing in
the number of lags. The second term on the RHS of Equation 4b is the lag bias
error effect. It is the true market return volatility (squared) times a lag factor,
which is the sum of the squared lags divided by (L1)2.
The objective is to find the value of L that minimizes the MSE. Clearly the noise
factor in the overall error diminishes with every increment in L (and more so the
larger is n, the transaction density per period). If purely random error were the
only type of error, as large a lag as possible could be selected (i.e., use all of the
historical price data just to estimate the current value of Vt). However, the lag
factor component in the overall error term increases with each increment in L,
thus presenting a trade-off. As lags are added, the noise factor is reduced by a
diminishing increment, while the lag factor in the MSE is increased by an
increasing increment.9 Thus, if the MSE is not already minimized at L 0, then
it will be minimized at some finite value of L as we step up through the integers:
L 1, 2, . . . n.
The optimal lag can be characterized by examining the finite differential of the
MSE as a function of L:
s s
L1 L
2 2
MSE 1
s1 s1
r2 2. (5)
L L2 L1 n(L 1)(L 2)
Setting this incremental error equal to zero, generates following optimal lag
criterion:10
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 2 2 G e l t n e r a n d L i n g
MSE
0,
L
s s
L1 L
2 2
1
s1
s1
r2 2, (6)
L2 L1 n(L 1)(L 2)
L1 L
s2 s 2
2
s1
s1
(L 1)(L 2) .
L2 L1 n 2r
Exhibit 2 shows the optimal maximum lag, L, measured in the number of index
return reporting periods (t), as a function of the transaction density per period, n,
and the ratio of the cross-sectional to longitudinal standard deviation: ( / r).
Thus, for example, if there are two transaction observations per period, and the
cross-sectional standard error in the transaction price observations is four times
Dispersion Ratio ( / r)
n1 6 3 1 0
n2 4 2 0 0
n 10 1 0 0 0
n 35 0 0 0 0
the longitudinal volatility per period in the true market returns, then the optimal
lag is four periods. That is, the MSE of the value estimate each period will be
minimized with a sample size of N 10, which will include two observations
from the current period t, plus eight additional observations, two from each period
going back through t4. Such an index estimation methodology will result in
L/2 4/2 2 periods of temporal lag bias (E[V t[N]] Vt2). The purely random
component of the total error in the index (log) value levels will have a standard
deviation of / N / 10 /3.16.
Note that the optimal lag diminishes rapidly as the transaction density increases
and as the dispersion ratio decreases. The optimal lag is zero (i.e., the population
value estimation uses contemporaneous transaction observations only) for
transaction densities as low as two transaction observations per period with the
dispersion ratio as high as unity. With this in mind, it is important to note that
the dispersion ratio decreases with the length of the index reporting periods (i.e.,
with the inverse of the index reporting frequency). For serially independent market
returns, the dispersion ratio is exactly proportional to the inverse of the square
root of the length of time in each index reporting period. Thus, for example, the
dispersion ratio in a quarterly index would be twice that of an annual index of
the same market.
On the other hand, the transaction density is a direct linear function of the length
of time in each reporting period, such that n for a quarterly index is one-fourth
that for an annual index of the same property population. The combination of
these two relationships keeps the optimal lag measured in calendar time
approximately the same, for a given market (that is, for a fixed population of
properties characterized by a given transaction density per unit of calendar time).
There is considerable empirical evidence that typical values of and r range
from 5% to 10% per year.12 Thus a dispersion ratio of about unity, with a range
of between 1/2 and 2, is probably realistic for annual frequency periods, implying
a dispersion ratio between 1 and 4 would be plausible for quarterly periods.
Assuming 10%, Exhibit 3 shows the root MSE (RMSE),13 as well as the
optimal lag (in number of periods) and the optimal sample size (N). The RMSEs
are based on the optimal lag methodology, under three different periodic volatility
assumptions that correspond to three different index reporting frequencies (annual,
quarterly, and monthly). It is assumed that r 10% per year.14
Exhibit 3 reveals the types of trade-offs a research index designer faces. In
principle, the index designer can control the reporting frequency and the
estimation sample size for each periodic value estimate. However, the designer
cannot control the underlying transaction density per unit of calendar time, or the
underlying market price dispersion (), or the volatility per unit of calendar time
(r). Thus, if the underlying transaction density per month is one, the index will
face a transaction density per index reporting period of either n 1, n 3, or n
12, depending on whether the reporting frequency is set at the monthly,
quarterly, or annual frequency. Similarly, if the underlying market volatility is 10%
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 2 4 G e l t n e r a n d L i n g
E x h i b i t 3 RMSE at Optimal Lag (L, & Optimal Sample Size N) as a Function of Transaction Density and
Market Volatility (both per period)
per year, then the index will face an underlying true volatility per period of either
r 2.9%, r 5%, or r 10%, depending on whether the index is to be
monthly, quarterly, or annual, respectively.
In this case, the index designer faces the possibilities indicated in the cells on the
diagonal in the table in Exhibit 3. The designer can improve the accuracy of the
index as measured by its RMSE by reducing the index reporting frequency
(increasing the period length), moving from a RMSE of 5.42% for a monthly
index to as little as 2.89% error in an annual index. However, the usefulness of
an index for research purposes is not purely a function of its RMSE. For many
important research questions, a higher frequency index would be more useful, and
for some questions the random noise component of the RMSE would not matter
as much as the lag bias component.
Conceivably, the index designer might attempt an alternative strategy to improve
the accuracy of the index. Instead of reducing the reporting frequency (aggregating
across time), the index designer might attempt to broaden the underlying
population of properties covered by the index, thereby increasing the underlying
transaction density per unit of calendar time. This would increase the transaction
density n for any given period length (i.e., without reducing index reporting
frequency). With this type of strategy (aggregating across space or across property
market segments), the movement down within the columns of Exhibit 3 reveal the
kind of improvements in valuation RMSE that can be obtained, first from tripling
the size of the covered population of properties, then by quadrupling it again. For
example, viewing all properties as representative of the target market, the top
row might refer to an index covering 1,000 properties, the middle row to an index
covering 3,000 properties, and the bottom row to an index covering 12,000
properties. However, the trade-off here may be that as one expands the cross-
sectional scope of the covered population of properties, one loses precision in the
definition of the actual real estate market that is represented by the index.
C o n s i d e r a t i o n s i n t h e D e s i g n a n d C o n s t r u c t i o n 4 2 5
For example, instead of an index of Boston office properties, one might be dealing
with an index of all office properties in the Northeast, or in the entire U.S.
I m p l i c a t i o n s f o r A p p r a i s a l - B a s e d a n d Tr a n s a c t i o n s -
Based Indices
In this section, the simple property value estimation model is used to examine
whether it would be better for a real estate research index to be based on traditional
appraisals or on direct transaction prices. First consider Exhibit 4a, which presents
a qualitative picture of the general noise-versus-lag trade-off and property value
estimation methodology optimization modeled in the previous section.
In Exhibit 4a, the vertical axis quantifies how up-to-date the value estimate is;
points farther up the axis represent less temporal lag bias in the value estimate
(and hence in any index based on such estimates). The horizontal axis quantifies
how precise is the value estimate; points farther to the right on the axis have less
purely random error (noise). The dashed curves represent user utility isoquants,
reflecting the utility (usefulness) of the value estimates for those using them. Such
utility increases from southwest to northeast: U0 U1 U2. The isoquants
E x h i b i t 4a The Noise vs. Lag Trade-off Frontier and Optimal Property Value Estimation Methodology
Reduced temporal lag bias
A
LA /2
U2
U1
T U0
2 /NA
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 2 6 G e l t n e r a n d L i n g
T
LB /2 B
U2
U1
U0
2 /NB
temporal lag bias. The isoquants are now shallower, tilted more horizontally to
skew the trade-off preference toward minimizing temporal lag bias. The result is
a different optimal point on the trade-off frontier, point B, that specifies a
smaller optimal estimation sample size: NB NA, resulting in less average lag:
LB /2 (NB /n1)/2 (NA /n1)/2, but with more purely random error.
The minimization of the MSE performed in the previous section may be viewed
as a particular utility function, which reflects a particular weighting of lag bias
error and purely random error. In particular, it weights these two forms of error
equally. Minimization of the MSE (or, equivalently, of the RMSE) is not
necessarily the most appropriate objective function for all uses of a research index,
but it is a widely used objective function in statistics. Exhibit 5 quantifies a
particular optimization based on the minimum-MSE objective function and the
stylized model of the previous section. Recall from Exhibit 2 that with one
transaction price observation available per period (n 1) and a cross-sectional-
to-longitudinal dispersion ratio of ( / r) 4, the optimal maximum lag in the
sample is LA 6 (first row, first column in Exhibit 2). This, in turn, implies an
optimal transaction sample size of NA n(LA1) 1(61) 7. In this situation,
if it assumed that the random error standard deviation is 10%, then the
RMSE (or standard total error percent) is 5.09% based on Equation 4b.16
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 2 8 G e l t n e r a n d L i n g
0
Lag Bias in Periods
Optimum @
Tange nt Pt. A.
3
A
U = MIN[RMSE]
= 5.09%
1 1/2 1/7
Now consider the difference between an index based on the stylized appraisal-
based value estimation technology versus one based on the stylized
transactions-based technology. Traditional appraisals of commercial properties
consider only a few comparable sales transactions (comps), spanning perhaps
a year or more in time. In contrast, transactions-based index regressions typically
employ sample sizes of at least a couple dozen transaction price observations
within each index reporting period. By taking as the objective the valuation of a
given subject property, the appraiser limits the relevant transaction sample to only
those properties that are very similar in structural and locational characteristics to
the subject property. The result is a sample density, labeled n (the number of
available relevant transaction observations per unit of time) that is inevitably very
small. In contrast, by taking as the objective the valuation of an aggregate subject
population of properties (the market or asset class that the index is meant to
track), the regression (mass appraisal) methodology is able to expand the
relevant transaction sample density many-fold, to a much larger value of n.
For example, a traditional appraisal of a specific Class A office building in
downtown Boston might be able to find, say, one comp per calendar quarter. Thus,
C o n s i d e r a t i o n s i n t h e D e s i g n a n d C o n s t r u c t i o n 4 2 9
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 3 0 G e l t n e r a n d L i n g
0 Corner Solution
Lag Bias in Periods
3 UC: RMSE=1.69%
A B
UB: RMSE=3.47%
UA: RMSE=5.09%
1 1/7 1/10
random error dimension. This is because in the present model, the individual
appraisals have less noise than individual transaction prices.19 In a population
where the properties are reappraised more frequently than they tend to transact,
another reason for such a result would be the larger cross-sectional sample size
of appraisals compared to transactions during each index reporting period.
Although the advantage of an appraisal-based index in this regard should not be
ignored, it is important to realize that many research questions are more sensitive
to temporal lag bias than to purely random error, particularly when the random
error is already low, as would be the case when the underlying property population
is sufficiently large to provide a large transaction sample each period.
Exhibit 6 reveals that the valuation procedure for index construction can be
conceptualized as an expansion of the noise-versus-lag trade-off frontier, from the
traditional appraisal-based frontier oriented toward disaggregate valuation, labeled
TDis, to the transactions-based regression procedure frontier aimed at aggregate
valuation labeled TAgg. Indeed, this frontier is not only expanded, but its shape
is changed. For estimating aggregate market value, the feasible noise-versus-lag
trade-off frontier becomes much steeper and straighter as it is pushed out to the
right toward the 2 0 boundary in the noise dimension (as it is impossible to
have a variance less than zero). Although this frontier is still concave in shape, it
is effectively kinked at the lag 0 boundary in the vertical dimension. The
result is that, even with the MSE-minimization utility function (which does not
value lag bias error minimization any more than random error minimization), the
optimum in aggregate valuation will often be a corner solution at zero lag. This
important general result is shown in Exhibit 8.20
S u m m a r i z i n g t h e M o d e l s I m p l i c a t i o n s
The stylized model described in the preceding two sections is clearly a
simplification of reality. For example, appraisers do not mechanically estimate
values by taking a simple average of the prices of the comps they use, and they
do not ignore changes in the market since the times when the comps transacted.
Nevertheless, appraisers tend to be conservative (and usually rightfully so) when
deviating from the direct market evidence presented by sales of comparable
properties. The model captures the essence of the distinction between basing a
real estate performance index on the aggregation of traditional individual property
appraisals versus constructing the index directly from transaction prices in a
regression-based procedure. With this in mind, the key policy implications from
the preceding analysis can be summarized as follows.
Consider a population of properties that are regularly appraised (marked to
market) and among which there are transactions in every period. However, the
transactions are less frequent than the appraisals; for example, the properties might
be reappraised annually, but transact on average only every 510 years. For a
research index meant to track the investment performance of such a population
over time, the following general conclusions apply:
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 3 2 G e l t n e r a n d L i n g
Lag Bias
in Periods
Corner Solution
Optimum @ Zero
Lag.
0 C
T Agg
T Dis
LA /2
A B
U1
T Agg
T Dis
U0
NA Disaggregate (Appraisal) Optimal Sample Size (# comps) nA(LA1), nA comps density / period;
NC Aggregate (Transactions Based Regression) Optimal Sample Size (obs per period);
Q Number of Properties (or market segments) in the Index Population.
For research that is highly sensitive to temporal lag bias, but less sensitive
to random error in the index returns, transactions-based indices are
preferable to appraisal-based indices, because of the temporal lag bias
inherent in appraisal-based returns.
For research that is highly sensitive to random error, but not very
sensitive to temporal lag bias, an appraisal-based index may be preferable
to a transactions-based index because of the greater frequency of the
appraisals in the population and (possibly) because the appraisals may
be less noisy than the transaction prices.
For research that is equally sensitive to lag bias error and random error
(e.g., as represented by the MSE-minimization objective for the index),
transactions-based indices are preferable to appraisal-based indices.
Except for indices tracking small populations of properties where
transaction density is less than two or three dozen observations per index
reporting period, transactions-based indices minimizing the MSE
criterion can be produced with no temporal lag bias.
C o n s i d e r a t i o n s i n t h e D e s i g n a n d C o n s t r u c t i o n 4 3 3
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 3 4 G e l t n e r a n d L i n g
The problem of stale appraisals may not be solved by simply requiring all data-
contributors to fully reappraise their properties each period. In practice, such brute-
force techniques may work fairly well at the annual frequency, for example, in
Britain with the IPD Annual Index. But there is some evidence that the brute-
force technique does not work well at quarterly or greater frequencies. For
example, the IPD Monthly Index in Great Britain shows signs of temporal lag
bias of the type expected from a stale appraisal effect. The brute-force technique
has trouble working at quarterly or monthly frequencies for two reasons: expense,
and lack of sufficient current real estate market information.
Data contributors to appraisal-based indices often try to mitigate the stale appraisal
problem by updating their valuation reports between major annual reappraisals.
But in order for an appraisal update to be temporally independent, it must be
completely independent of the previous value report (or any prior value report)
for the subject property. This is not merely an independent valuation report in
the legalistic sense that an external fee appraiser signs off on the update. Temporal
independence for the purpose of avoiding the stale appraisal effect requires that
the appraiser estimating the updated value be entirely unaware of any previous
value estimates for the subject property. Superficial updating of the value reports
(so-called desktop appraisals) between the annual full (or clean-slate)
reappraisals will be inadequate, because they will inevitably rely heavily on prior
valuation reports.
Fundamentally, there is often too little purely contemporaneous valuation
information, within the current quarter or month, for the appraiser to get a
sufficiently precise indication of the subject propertys value. Unless the property
is a very common or homogeneous type that trades frequently, there may be little
or no transaction price evidence available purely from the current period. This
makes it extremely difficult for the appraiser to avoid the use of stale information.
This is a bigger problem the shorter is the time interval between valuation updates.
C o n s i d e r a t i o n s i n t h e D e s i g n a n d C o n s t r u c t i o n 4 3 5
I n d e x e s t o Tr a c k D i f f e r e n t Ty p e s o f P r o p e r t y P o p u l a t i o n s
The above analysis is oriented to the type of index design problem where the
index is meant to track a population of properties that is owned by a particular
class of investors who require that the properties be regularly marked to market.
This is true of the NCREIF Index in the U.S. and the IPD Index in Great Britain.
In such populations, appraisals are available for index construction, and they are
more frequent than transaction prices. However, if we take the population of all
commercial properties of a given type and/or geographic location; appraisals will
often not be available for most properties in the population, and/or appraisals will
often be less frequent than transactions. For indices designed to track such
populations or markets, transactions-based indices have even more advantages.
Indeed, in such circumstances, transactions-based indices may be the only
solution, and the emphasis should be on using statistically rigorous techniques
such as hedonic or repeat-sales regressions rather than simple median price indices
that do not control for quality differences in the properties that transact each
period.
Another type of aggregate property value estimation process is represented by the
estimation of net asset value (NAV) for real estate investment trust (REIT)
properties. The process of regular estimation of REIT NAV by stock market
analysts who follow the REIT industry is akin to the process involved in
transactions-based index estimation in that NAV estimates are aggregate (or mass
appraisal) valuations. In other words, NAV estimation is oriented toward
optimizing the value estimate of a portfolio of properties, namely all those owned
by a given REIT. However, the actual NAV estimation process is less formal and
less based on econometric techniques than transactions-based index estimation.
An intriguing possibility in NAV estimation is the role played by the subject
REITs share price on the stock exchange. There is considerable evidence that
REIT share prices lead the property market in time.24 The NAV analysts attempt
to apply cap rate information from relevant property markets to NOI information
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 3 6 G e l t n e r a n d L i n g
from the REITs own reported accounts. However, in the numerous judgments that
must be made in the NAV valuation process, it is likely that analysts are influenced
by the availability of a prominent, very appropriate comp, i.e., the REITs
current stock price. The influence of the REIT share price, as a leading indicator
of the actual relevant property market value, would act to make the NAV estimate
less lagged than would be a typical property appraisal. This is combined with the
fact that, as noted in the previous analysis, optimal mass appraisals (which is what
NAV estimates essentially are) have less lag than traditional individual property
appraisal. The implied result is that NAV estimates, though informal and therefore
difficult to rigorously compare to econometrically-based valuation methods, may
provide a rather accurate and up-to-date estimate of their subject property portfolio
values. Clayton (2004) provides some evidence that this is indeed the case.
P r o p e r t y S a m p l i n g I s s u e s i n a Tr a n s a c t i o n s - B a s e d
Research Index
Indices of property populations that are regularly marked to market can be quite
useful for research purposes if optimally designed for research, as described above.
Nevertheless, it is also desirable to develop indices of broader populations of
properties, and for specific commercial property markets. Therefore, this section
presents some basic considerations in the design of such indices. In particular,
what is the optimal population size and scope for a transactions-based research
index.
Statistics theory puts forth that the power and usefulness of a sample drawn from
a large population is a function of three characteristics:
1. How much random dispersion exists in the underlying population;
2. The absolute size of the sample (not the percentage of the population
included in the sample); and
3. The representativeness of the sample (i.e., the lack of bias in its
collection).
As noted previously, the effect of sample size on statistical power is a function
of the square root of the number of independent observations in the sample.
Because the square root is a concave function, declining benefits to scale exist for
sample size increments, other things being equal. However, there may also be
economies of scale in collecting data when tracking commercial property markets,
at least up to a point. Thus, the benefit/cost ratio of expanding the sample size
may be more favorable than implied by the Square Root of N Rule.
Because of declining returns to scale, it may be wasteful to attempt to include all
or even a large fraction of a large population in a statistical sample to be used for
research purposes. Rather, if the population is diverse and complex, it may be
more efficient to stratify the sample. Sample stratification refers to the
identification of strata or cells consisting of sub-populations or segments of
C o n s i d e r a t i o n s i n t h e D e s i g n a n d C o n s t r u c t i o n 4 3 7
the overall population that are relatively homogeneous. For example, one cell or
strata could be office properties in Florida. It is necessary to sample from each
cell, but within each cell the Square Root of N Rule applies.
How big a property sample would be necessary for a transactions-based research
index of U.S. commercial property investment performance? The answer depends
on a number of considerations. At a general level, the most important include:
1. What types of research questions will the index be optimized to address?
2. How much random dispersion exists in the underlying population
regarding the variables that are important for addressing the target
research questions?
3. How much standard error (e.g., in terms of a 95% confidence interval
around the best estimate) is acceptable in the index reports?
As suggested previously, there is evidence that cell sample sizes as low as two
dozen transaction observations per index period and sometimes even less are
sufficient to produce useful return indices in a commercial property market when
regression analysis is combined with noise-filtering techniques such as the
Bayesian use of ridge regression (see, for example, Geltner 1997a; Gatzlaff and
Geltner 1998; and Geltner and Goetzmann, 2000).
To investigate how to determine the minimum necessary cell sample size, consider
a simple stylized example, building on the previous model. Assume as before that
the random noise in transaction prices () has a standard deviation of 10%. Now
assume that the true market return volatility (r) is 5% per index reporting period.
Assume also that the research index periodic return will have a standard error
equal to no more than one-half the true periodic return volatility. Thus, the sample
needs to be designed so that the index returns will have a standard error of 2.5%
each.25
As before, the standard error of the index value level is / NC, where NC is the
transaction sample size within each cell.26 Because returns are first differences of
value levels, the index return standard error is equal to the square root of twice
the square of the index value level standard error. Thus, the index return standard
error is 2( / NC)2. Setting this equal to 2.5% and solving for NC, the equation
becomes:
2 (10%)2
NC 1 64, (7)
(r) /2 (2 5%)2 /2
2
where is the fraction of the true return volatility the index return standard error
should fall within. Said differently, the index requires a cell sample size of 64
transactions per period per cell. This result holds regardless of how many
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 3 8 G e l t n e r a n d L i n g
properties are in the underlying population within the cell and would apply with
classical regression technology.27 The Bayesian noise-filtering technique noted
previously can often reduce this sample size requirement by half or more.28
This type of analysis can be used to determine how many cells, or property market
segments, on which the index can report; that is, how finely the overall
commercial property asset class can be carved up for research purposes. For
example, suppose the total population of U.S commercial properties consists of
1,000,000 properties, which transact, on average, once every five years. This
provides an average of 200,000 transactions per year, or 50,000 per quarter. If a
minimum cell size is NC 50 (which is a bit conservative), this implies that the
overall commercial real estate market (or asset class) in the U.S. can be broken
into 1,000 cells, or market segments, that can be included in a quarterly index.
For example, assume the overall market can be usefully classified into, say, five
property sectors (office, retail, etc), with an average of three size or quality
categories per sector (e.g., Class A, B, etc. or, alternatively, below $5 million; $5
50 million; and greater than $50 million). Assume also that there are 35
metropolitan areas or geographic regions of interest. This would imply a total of
525 cells (5 3 35), well within the 1,000 cell limit imposed by the accuracy
requirements for the index. However, some gerrymandering of cell definitions
would normally be necessary to achieve sufficient sample size in all cells (e.g.,
requiring the combination of some smaller market segments into composite cells).
Moving up the strata to define, for example, national property sector indices, the
relevant sample size increases, leading to greater statistical power and less need
to gerrymander the cells. In the previous example with 525 cells of 50 observations
each, the resulting 26,250 (525 50) overall transaction sample per quarter would
include an average of 5,250 (26,250/5) observations per quarter, per property type
sector. This far exceeds the number of observations necessary to provide a highly
accurate return index. In order to aggregate properly across cells, however, the
relative sizes of each market segment must be known as a proportion of the total
population. Thus, scientific sample stratification requires not only a determination
of the minimum sample size in each cell, but the relative magnitude of the
population within each cell. For example, we would like to know the relative
market value of each MSA, each property type within each MSA, and each sub-
type or size category within each type. In general, it is considered acceptable to
estimate the relative cell sizes without recourse to an entire population census.
Nevertheless, some method of surveying or otherwise estimating the relative cell
magnitudes would need to be applied.
Such a stratified index can be designed to be far more useful than an index with
a larger overall sample size but where the observations are drawn randomly from
the population, with over-representation from some cells and no representation at
all from others.
C o n s i d e r a t i o n s i n t h e D e s i g n a n d C o n s t r u c t i o n 4 3 9
Conclusion
Since the founding of NCREIF almost three decades ago, statistical methodologies
useful for investment performance index construction have advanced dramatically,
notably including developments such as the repeated-measures regression (RMR)
and related noise-filtering techniques. In recent years, electronic databases of
commercial property prices have been developed that far exceed the quality and
coverage of those databases available only a few years ago. Together, these two
developments offer capabilities for transactions-based indices, mass appraisal, and
other tools that could be very useful for improving real estate research indices.
Real estate indices developed or improved going forward should strive to
incorporate these modern standard statistical methodologies and databases.29
This paper has surveyed some of the major technical issues in the design and
construction of such modern real estate research indices, both appraisal-based and
transactions-based. It has discussed property sampling issues, differences between
transaction prices, market values, and appraised values, the trade-off between
random measurement error and temporal lag bias, optimal reporting and property
revaluation frequencies, and the uses and limitations of modern statistical theory.
Although one of the conclusions suggested by the analysis is that most research
questions are best addressed with transactions-based, rather than appraisal-based,
indices in the U.S. (largely because of the temporal lag bias problem in appraisal-
based indices), this paper discusses how appraisal-based indices can still be useful
for some research purposes, especially if they are optimized for such purposes.
Clearly, great benefit can be obtained from expanding the property universe
coverage in developing new research-oriented indices going forward. In this
regard, the paper discusses how stratified statistical sampling techniques can be
used to facilitate the most efficient and effective expansion of asset class coverage.
Finally, as the U.S. real estate investment industry continues to improve the
quantity and quality of indexing /benchmarking information products, it should
carefully guard against the use of, or reliance on, black boxes. To increase
the credibility of the private real estate asset class, and to ensure continued
improvement of the index products, it is vitally important that proprietary
production techniques and methodologies used in the development and production
of index products be fully available to public and academic scrutiny and criticism
(while nevertheless protecting the confidentiality of proprietary data).
Endnotes
1
In an appraisal-based index, revaluation frequency refers to how often the properties
composing the index are reappraised. In practice, reappraisal has multiple definitions,
ranging from superficial desk-top appraisal updates to full independent fee
appraisals. For the purpose of determining or understanding the important dynamic
characteristics of an index, superficial reappraisals do not have much effect. As a result,
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 4 0 G e l t n e r a n d L i n g
model, had declining appraiser weights on comps with the age of the comp, and also
relaxed the random walk assumption of the underlying true real estate returns. [Also,
Geltner (1997b) considered the use of REIT returns to improve appraisal accuracy.]
12
For the evidence on , see: Goetzmann (1993), Diaz and Wolverton (1998), Geltner,
Young, and Graff (1994), Geltner (1998), and Crosby, Lavers, and Murdoch (1998). For
the evidence on r, see: Fisher, Geltner, and Webb (1994), Fisher, Gatzlaff, Geltner, and
Haurin (2003), and Fisher, Geltner, and Pollakowski (2007).
13
Recall that the model is using log values, so the value error units equate to percentage
errors in straight levels.
14
Note that return standard errors are greater than valuation standard errors. The square
of the noise variance component doubles in the returns compared to the log value
estimates, as the returns are the first-differences of the log value levels across time.
15
The covariance between the real estate returns and the stock market, or the beta of
real estate, will be unaffected by purely random error. However, the correlation
coefficient between the two returns series, and the precision in the estimation of the
beta, will be reduced by the presence of noise in the real estate returns, while the real
estate volatility will be increased by purely random error.
16
With ( / r) 4, and 10%, r 2.5%. Plugging these values (and N7) into
Equation 4b, gives:
5.09% .102
7
.025
7
2
s.
6
s1
2
17
Empirical evidence from the NCREIF Index suggests that this is, roughly, a typical
amount of lag and of error in such an appraisal (see Fisher and Geltner, 2000; and
Geltner and Goetzmann, 2000).
18
Substituting r 2.5% and LA 6 into the lag bias error component of Equation 4b,
the root of the lag bias error component becomes:
2 6
.025
s2 (.00357)2(91) 0.0341.
7 s1
19
In the real world, there is some question whether this is in fact the case. For example,
the studies of appraisal accuracy cited in the earlier endnote suggest typical random
appraisal error magnitudes at least in the 5% to 10% range, which is not much, if any,
below the transaction dispersion magnitude found in repeat-sales regression models of
housing prices (see, e.g., Goetzmann, 1993).
20
Giaccotto and Clapp (1992) found a similar result. However, they characterized the
different appraisal objectives as that between estimating returns versus estimating
values, whereas the focus here is on the difference between estimating values (or,
hence, returns) for individual properties versus aggregate portfolios or indexes of many
properties, or in effect, using transactions-based versus traditional appraisal-based
indices for tracking changes in values.
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 4 2 G e l t n e r a n d L i n g
21
Perhaps this could be accomplished by a rotation of appraisers.
22
Ling, Naranjo, and Nimalendran (2000) propose the use of a latent variable model for
the construction of total return indices that makes use of both appraisal-based and
transaction-based returns.
23
The tendency of the appraised values themselves to exhibit temporal lag bias was
addressed in the previous model.
24
See, for example, Gyourko and Keim (1992), Barkham and Geltner (1995), Giliberto
and Mengden (1996), and Geltner and Goetzmann (2000).
25
With normally distributed errors, this would mean that each periodic return reported by
the index would have about a two-thirds chance of being within 250 basis points of the
true return that period.
26
Note that according to the stylized model discussed previously: NC nC, the observation
sample size used to construct the index will equal the cell populations available
transaction density per period, as transactions-based indices generally have an optimal
lag of zero. Exceptions to this rule might occur, for example, with a very small
underlying property population size, or with a very short index reporting period, or a
research purpose that places extreme importance on reducing random error while not
caring much about temporal lag bias error.
27
This assumes the cell population is large. If the cell population is very small, then the
necessary 2.5% standard error might be achieved with a smaller sample size.
28
Other recently developed econometric techniques may also be useful in this regard, such
as spatial autocorrelation models.
29
A recent example is the transactions-based version of the NCREIF Index developed and
published by the MIT Center for Real Estate. [See: Fisher, Geltner, and Pollakowski
(2007).]
References
Bailey, M., R. Muth, and H. Nourse. A Regression Method for Real Estate Price Index
Construction. Journal of the American Statistical Association, 1963, 58, 93342.
Barkham, R. and D. Geltner. Price Discovery in American and British Property Markets.
Real Estate Economics, 1995, 23:1, 2144.
Blundell, G.F. and C.W.R. Ward. Property Portfolio Allocation: a Multi-Factor Model. Land
Development Studies, 1987, 4:2, 14556.
Brown, G. The Information Content of Property Valuations. Journal of Valuation, 1985, 3,
35062.
Chaplin, R. Unsmoothing Valuation-based Indices Using Multiple Regimes. Journal of
Property Research, 1997, 14:3, 189210.
Chinloy, P., M. Cho, and I.F. Megbolugbe. Appraisal, Transaction Incentives, and
Smoothing. Journal of Real Estate Finance and Economics, 1997, 14, 89112.
Clayton, J. Tracking Value in the Private Real Estate Market: A Comparison of Green
Street NAV and NCREIF Value Indices. Working Paper, University of Cincinnati, 2004.
Crosby, N., A. Lavers, and J. Murdoch. Property Valuation Variation and the Margin of
Error in the UK. Journal of Property Research, 1998, 15:4, 30530.
Diaz, J. III and M.L. Wolverton. A Longitudinal Examination of the Appraisal Smoothing
Process. Real Estate Economics, 1998, 26, 34958.
C o n s i d e r a t i o n s i n t h e D e s i g n a n d C o n s t r u c t i o n 4 4 3
Fisher, J. and D. Geltner. De-Lagging the NCREIF Index: Transaction Prices and Reverse
Engineering. Real Estate Finance, 2000, 17, 722.
Fisher, J., D. Gatzlaff, D. Geltner, and D. Haurin. Controlling for the Impact of Variable
Liquidity in Commercial Real Estate Price Indices. Real Estate Economics, 2003, 31:2:
269303.
Fisher, J., D. Geltner, and H. Pollakowski. A Quarterly Transactions-Based Index of
Institutional Real Estate Investment Performance and Movements in Supply and Demand.
Journal of Real Estate Finance & Economics, 2007.
Fisher, J., D. Geltner, and R.B. Webb. Value Indices of Commercial Real Estate: A
Comparison of Index Construction Methods. Journal of Real Estate Finance & Economics,
1994, 9:2, 13764.
Fu, Y. Estimating the Lagging Error in Real Estate Price Indices, Real Estate Economics,
2003, 31:1, 7598.
Gatzlaff, D. and D. Geltner. A Transaction-Based Index of Commercial Property and its
Comparison to the NCREIF Index. Real Estate Finance, 1998, 15, 722.
Geltner, D. Bias & Precision of Estimates of Housing Investment Risk Based on Repeat-
Sales Indexes: A Simulation Analysis. Journal of Real Estate Finance & Economics, 1997a,
14:1/2, 15572.
Geltner, D. The Use of Appraisals in Portfolio Valuation and Index Construction. Journal
of Property Valuation & Investment, 1997b, 15:5, 42348.
. How Accurate is the NCREIF Index as a Benchmark, and Who Cares? Real Estate
Finance, 1998, 14:4, 2538.
Geltner, D. and W. Goetzmann. Two Decades of Commercial Property Returns: A
Repeated-Measures Regression-Based Version of the NCREIF Index. Journal of Real Estate
Finance and Economics, 2000, 21, 522.
Geltner, D. and D.C. Ling. Ideal Research and Benchmark Indexes in Private Real Estate:
Some Conclusions from the RERI/PREA Technical Report. Real Estate Finance, 2001,
17:4, 112.
. Benchmarking and Return Performance in the Private Commercial Real Estate
Market, International Real Estate Review, forthcoming 2006.
Geltner, D., M. Young, and R. Graff. Random Disaggregate Appraisal Error in Commercial
Property: Evidence from the Russell-NCREIF Database. Journal of Real Estate Research,
1994, 9:4, 40319.
Giaccotto, C. and J. Clapp. Appraisal-Based Real Estate Returns under Alternative Market
Regimes. Journal of the American Real Estate and Urban Economics Association, 1992,
20:1, 124.
Giliberto, S.M. and A. Mengden. REITs and Real Estate: Two Markets Reexamined. Real
Estate Finance, 1996, 13:1, 5660.
Goetzmann, W. The Single Family Home in the Investment Portfolio. Journal of Real
Estate Finance & Economics, 1993, 6:3, 20122.
Gyourko, J. and D. Keim. What Does the Stock Market Tell Us About Real Estate Returns.
Real Estate Economics, 1992, 20:3, 45786.
Hamilton, S.W. and J. Clayton. Smoothing in Commercial Property Valuations: Evidence
from the Trenches. Real Estate Finance, 1999, 16, 1626.
Lai, T-Y. and K. Wang. Appraisal Smoothing: The Other Side of the Story. Real Estate
Economics, 26, 51136.
J R E R Vo l . 2 8 N o . 4 2 0 0 6
4 4 4 G e l t n e r a n d L i n g
Ling, D., A. Naranjo, and M. Nimalendran. Estimating Returns on Commercial Real Estate:
A New Methodology Using Latent Variable Regression. Real Estate Economics, 2000,
28(2). 20531.
Quan, D.C. and John M. Quigley. Price Formation and the Appraisal Function in Real
Estate Markets. Journal of Real Estate Finance and Economics, 1991, 4, 12746.
The authors thank the Pension Real Estate Association and the Real Estate Research
Institute for providing partial funding for this research. However, all statements and
assertions are the opinions of the authors. They also thank Ko Wang, the editor, and
several anonymous referees for helpful comments and suggestions.