Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Approach
Most of the widely used statistical models are based on “average behavior”; that is, they are built
around explaining and predicting the outcome of the mean of available data. And, in most
analyses, observations that are too “far” from the mean, i.e. outliers, are discarded. But, an
important question needs to be addressed: do outliers (also called extreme values) contain
information that can be useful in practice and are there any statistical techniques that provide for
their modeling? The goal of this paper is twofold. First, to offer a short exposition of the
fundamental principles governing the analysis of extreme values and, second, to offer a traffic
example of how the use of extreme value analysis can improve on the results of more commonly
used statistical techniques. The results of the case study suggest that both the sign and the
magnitude of an estimated trend term vary significantly between extreme value and regression
models, a finding with very important policy implications.
keywords: traffic data analysis, extreme values, outliers, extreme order statistics
INTRODUCTION
Transportation and traffic considerations most frequently revolve around the analysis of
available data; data on accident counts, traffic injuries, freight volumes, passengers, traffic
flows, highway speeds, etc. Appropriate models are fit to these data to “explain” and predict the
behavior of corresponding variables. As is well known, most of the commonly used models are
based on “average behavior”; that is, most models are built around explaining and predicting the
And, while this approach is very satisfactory in modeling most of the commonly
encountered problems there are, many times, transportation problems that deal with boundary
conditions. These conditions can be, for example, congested traffic, speed limit violators,
extreme terminal delays, etc. From a methodological standpoint all these problems have been
dealt with using the commonly available statistical analysis tools (linear, non-linear and Poisson
regression, time-series analyses, etc.). But, an important question needs to be addressed: are the
1
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
commonly used statistical methods still unbiased, efficient and asymptotically consistent when
dealing with boundary or extreme conditions? And, furthermore, are there any other statistical
techniques that can provide for improved modeling of such phenomena and also allow for useful
inferences to be made?
In most statistical analyses the usual course of action is to discard observations that are
“too far” from the mean; these observations are commonly termed outliers. But, as is widely
recognized, outliers frequently contain information that can shed much light into the phenomena
being studied. Furthermore, in many cases, these outliers can help predict the occurrence of
other such outliers much better than by studying the remainder of the values. For example, it is
quite possible that by examining the occurrence of, say, very high traffic occupancy values
(occupancy is frequently recognized as the “best” indicator of congestion), the future occurrence
of such values can be better predicted. That is, by concentrating the study on congestion itself,
rather than on the entire series of occupancy values, it may be easier and more effective to
Interestingly, the study of phenomena far from the mean, phenomena that are infrequent,
or, in mathematical terms, phenomena that occur in the tail of probability distributions have, for
a long time, been the focus of many scientific disciplines. Seismologists and geotechnical
engineers have been very interested in the occurrence of earthquakes of extreme magnitudes;
oceanographers and hydraulic engineers have been interested in the occurrence of the T-year
flood level; meteorologists and structural engineers have been interested in the occurrence of
strong hurricanes. All these phenomena, powerful hurricanes and earthquakes, T-year floods, etc.
have been described as extreme events and, as such, their analysis is widely known as “Analysis
of Extreme Values” or “Extreme Order Statistics” (as a point of scientific interest it should be
noted that extreme value analyses have been used to study a very wide variety of subjects
ranging from the atmospheric ozone levels to earthquakes, to hurricanes, to stock market returns,
to population aging and sea levels. Of-course, a discussion of these studies is beyond the scope
2
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
of this paper; interested readers should refer to (1), (2) and (3) for both an in-depth discussion of
The spirit and the goal of the analysis of extreme values is to develop models that can
capture the essential information contained in them, i.e. in the tails of distributions, and to
improve on the projection of re-occurrence of such events. The potential for applying these
occurrence, trends in traffic flows, extreme highway speeds, high accident areas, etc. can all be
studied using the tools provided by extreme value analysis. The goal of this paper is twofold.
First, to offer a short exposition of the fundamental principles governing extreme value analysis
and, second, to offer a traffic example of how the use of extreme value analysis can improve on
the results of more commonly used statistical techniques. The remainder of this paper is
organized as follows; the next section reviews some of the fundamental mathematical principles
upon which extreme value analysis is built. The third section describes and offers some results
from applying extreme value analysis to a traffic database and, finally, the fourth section offers
Fundamental Concepts
The issues related to the analysis of Extreme Order Statistics (Extreme Values) may be
described, in broad terms, as an effort to make statistical inferences about the extreme values in a
population or a random process. The traditional approach (4) is based on the extreme value limit
distributions first identified by Fisher and Tippett (5) and applied to a large number of cases as
reviewed in the introductory section. A typical application consists of the generalized extreme
value distribution fitted to annual (or of any predefined time period) maxima (that is, the
maximum value of a given series for a given time period) and can be defined as
3
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
1κ
y − µ
H ( y; µ ,ψ ,κ ) = exp − 1 − κ (1)
ψ
over the range of y that κ ( y − µ ) < ψ , where ψ > 0 and κ , µ are arbitrary. The case κ = 0 is
y − µ
H ( y; µ ,ψ ,0 ) = exp − exp − (2)
ψ
the type I family of Gumbel distribution(4). These distributions are justified as the limiting
stable distributions of extreme value theory. The numerical fitting procedures were developed by
Prescott and Walden (6, 7) and a Newton-Raphson algorithm was developed by Hosking (8) and
amended by Macleod (9). For the asymptotics of Maximum Likelihood, see Smith (10); it should
be noted though that there is evidence that Maximum Likelihood estimation does not work well
in small samples (8, 9, 10), and other methods have been proposed, the most well known of
which is the “probability weighted moments” approach proposed by Hosking et al. (11). It
should also be noted that an extension to the usual time period maxima is to use several of the
Threshold methods, the second most important area of interest in Extreme Order
Statistics after the Maxima method, have been extensively developed by hydrologists over the
last 30 years, often under the acronym of the POT (peaks over threshold) method. These
methods are based on fitting a stochastic model to either the exceedances or the peaks over a
threshold for a given series. Such methods have been used for quite some time, but the first
systematic treatment was by Todorovic and Zelenhasic (15) and Todorovic and Rousselle (16).
The approach most commonly employed in practice is based on the generalized Pareto
distribution, which has an interpretation similar to that which motivates Eq. (1). The remainder
of this Section deals with POT methods since this class of models shows most promise in
transportation and traffic related problems. It studies the occurrence of phenomena over a
4
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
“tolerance” value (acceptable threshold), such as highway speeds over the speed limit, traffic
flows in the unstable region, etc. These are typical cases where the analyst may wish to identify
and analyze classes of events that could potentially lead to severe consequences in a particular
context. In this case, in which the potential extremist events have not yet been identified or
crucial.
process of exceedance times (when, for example, congestion occurs during a time period)
combined with independent excesses over the threshold (the magnitude of congestion). Consider
a threshold u, and let N denote the number of exceedances of u during a period of, say, m
months. Suppose the excesses, or the differences between the observations over the threshold
and the thresholf itself, are independent with common distribution function G. The theory upon
which POT models is built suggests that G is the generalized Pareto distribution given by
1k
y
G ( y;σ , κ ) = 1 − 1 − κ (3)
σ
where σ > 0 and κ is arbitrary; the range of y is 0 < y < ∞ if κ ≤ 0 , o < y < σ κ if κ > 0 .
When κ = 0 then Eq. (3) corresponds to the exponential distribution with mean σ . There are
several ways of motivating Eq. (3). Pickards (17) showed that this equation arises as a limiting
distribution for excesses over thresholds if and only if the parent distribution is in the domain of
attraction of one of the extreme value distributions. Another motivation suggests that if Y is
generalized Pareto and u>0, then the conditional distribution of Y-u given Y>u is also
generalized Pareto. Another property suggests that if N has a Poisson distribution and,
5
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
variables, then max (Y1 ,..., YN ) has the generalized extreme value distribution. Thus, a Poisson
process of exceedance times with generalized Pareto excesses implies the classical extreme
value distributions. These properties suggest that the generalized Pareto distribution will be a
practical family for statistical estimation, provided that the threshold is taken sufficiently high.
Of-course, the question that naturally arises is how high to take the threshold. Theoretical results
are available that guide toward threshold selection (see, for example, Smith (18)), but in this
Methods for estimating the generalized Pareto distribution parameters have been
thoroughly reviewed by Hosking and Wallis (20). Maximum Likelihood estimators exist in large
1
samples provided that κ < 1 and are asymptotically normal and efficient if κ ≤ . In general, it
2
∑ log (1 − κ Y σ )
N
− n log σ + (1 κ − 1) i
1
Writing σ = κ τ and differentiating with respect to τ and κ , the maximum with respect to κ
is achieved when
∑ log (1 − τ Y )
N
κ = κ (τ ) = − N −1 i
1
Thus, the maximum likelihood estimation reduces to solving the single equation
∑Y
N
N τ = {1 κ (τ ) − 1} i (1 − τ Yi )
1
6
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
The most frequently encountered “competitor” for maximum likelihood estimation is the
probability weighted moments method of Hosking and Wallis (20). The rth probability weighted
moment is defined as
σ
α r = E Y {1 − G ( Y )} =
r
( r + 1)( r + 1 + k )
where G is the generalized Pareto distribution function. The method essentially consists of
equating the theoretical and sample-based values of the probability weighted moments for r=0,1.
POT Regression
The most widely used application of statistics is the description of systematic variation in a
explanatory or independent variables). In the context of POT models, regression assumes that κ
parameters κ i = κ ( zTi γ ) and σ i = σ ( xTi β ) , where xi and zi are known vectors of covariates of
respective lengths p+1 and q+1. Suppose in addition that the Yi may be upper truncated at
random, and say that δ i = 0 if yi is the observed value of the random variable Yi , but that δ i = 1
if yi is a lower bound for the true but unobserved Yi . The contribution to the overall log-
likelihood L = ∑l i made by Yi is
− yi σ i − log σ i + δ i log σ i , κi = 0
li ( yi ;σ i ,κ i ) = 1 yi yi (4)
κ − 1 log 1 − κ i σ − log σ i + δ i log σ i + log 1 − κ i σ , κ i ≠ 0
i i
i
7
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
∂L
∑ z κ ∂l ( y∂;κσ ,κ )
N
= i i i i
∂γ
i i
i =1 i
(5)
∂l ( y ;σ ,κ )
= ∑xσ
∂L N
i i i i
∂β ∂σ
i i
i =1 i
Athens’ main traffic corridors has increased over the last 4 years; furthermore, with the 2004
Olympic Games fast approaching, it is imperative to determine whether key road axes that will
serve as main gateways for the Olympic Family are approaching congestion or whether policy
measures instituted by the authorities have begun to show an effect in tapering congestion.
For information purposes, it is noted that the urban area of Athens, the capital of Greece,
has an area of 60 km2 and a population of approximately 3.8 million people. During the last
decade the population in the greater Athens area has increased by about 10% while at the same
time car ownership has increased considerably, approaching 310 automobiles per 1000
inhabitants. This has led to an increase in travel time by 26% in the last 12 years which has
resulted in the deterioration of traffic conditions in the capital. Overall, planning authorities had
to deal with a 3.5% annual increase in traffic for the last ten years. An increasing proportion of
the signalized intersection approaches in the center of the city is highly congested (levels of
service E-F). Obviously, travel times in such a congested network are quite long, and the
8
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
The Data
Background
A dynamic traffic map and ATIS for the central Athens area has been operating since 1996. As a
first attempt to meet driver information needs, travel time information is given over the internet
updated every 15 minutes. To do this, 6 origin points have been defined at the major city
entrances, from which a total of 17 possible (commonly used) routes have been considered.
Speeds and travel times along these 17 routes are being estimated using flow and occupancy data
collected directly from the 140 controllers installed in various points of the down-town urban
network. These are single (inductive) loops measuring traffic volume and queue presence at an
Raw data covering traffic volumes and occupancies are selected by sensors of the Urban
Traffic Control Center (UTCC) and arrive to the special processing facilities of the National
Technical University of Athens (NTUA) every 90 seconds. All data are batch processed, with a
combination of specific tools and algorithms. Data integrity is ensured by the controllers
themselves as malfunctioning loops default shortly to the value of 255 and they are
not taken into account in the analysis period or thereafter. Data quality control is provided
automatically by the data management software that performs screening and data repair
functions. The processed data files are then transmitted using the File Transfer Protocol (FTP)
procedure from the batch processing enviroment to the UNIX based Web Server of the NTUA
School of Civil Engineering (FCE) via the University’s telematics (ISDN) network. Combining
all the information contained in the data files, ten Graphic Interchange Format (GIF) images are
generated. These are included in Web Pages of the NTUA Faculty of Civil Engineering Web
Server giving access to information on the traffic situation in Athens to Internet users.
The first two GIF images contain volume and congestion data for the entire central
Athens urban area. The remaining eight GIFs (four for volume and four for congestion) are
blown-up parts of the network, generated along with the main image every quarter of an hour via
9
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
a UNIX script. Users can have access to those blown-up parts by clicking on a part of the main
image in the Web Page (The final result is available at the address: www/transport.ntua.gr/map)
The Database
Data on traffic flow and occupancies are collected, as already mentioned, at 90 sec intervals at
140 locations around Athens, continuously since 1996. All this data has been stored and is
available for analysis. For the purposes of this analysis further manual checking was performed.
Detail police records were used to exclude periods were major or minor events (e.g.
demonstrations) or incidents external to normal traffic operations (e.g. roadworks) had been
reported. In an effort to reduce the size of the data set to be processed without losing much
essential information, a new data set was created containing occupancy measurements (at 90
second intervals) for the first two months (January and February) of the years 1998 – 2001. This
yielded more than 15,000 measurements for each of the 17 cross-sections along a part of the
eastern section of the Olympic Ring that is examined in this study (Figure 1). To make more
practical sense of the results, and after preliminary testing indicated this to be a statistically
acceptable approach, data were further aggregated into directions: occupancy toward the city
center (direction NE to SW) and toward the Olympic Stadium (the Olympic Stadium is in the
opposite direction from the city center; in this paper only the results for the direction toward the
Olympic Stadium are presented (direction SW to NE) – the results for the other direction are
available from the authors upon request). All data refer to 4 measuring locations along the
aforementioned corridor (shown by arrows in Figure 1). The distance between the first and the
last location is 4100m, with intermediate distances between the first and the second, the second
and the third etc, equal to 1900m, 1200m and 1000m respectively.
10
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
The Methodological Approach
As previously mentioned, the main goal of this study is to demonstrate the potential applications
of Extreme Order Statistics to transportation problems. The case study selected for this
one of Athens’ major roadway corridors that serves not only a large part of the daily commuting
activity, but is also part of the Olympic Ring and will serve as a major artery for the Olympic
visitors in the 2004 Games. The goal is to examine whether congestion has increased since 1998,
or whether measures instituted by the traffic authorities have been effective; to achieve this, the
main parameter considered is that of a trend. That is, interest in this study focuses around
determining the existence of a trend in congestion over the past four years. In light of the
discussion in the previous Section, the approach taken here is to fit the model:
yi = α + bti + ε i , 1 ≤ i ≤ n (6)
where yi is the exceedance magnitude for the ith observation (occupancy magnitude above the
92% threshold), α is the constant, ti is the time trend for the ith observation (ti = 1,…, n, where n
= total number of observations), and the ith residual ε i is assumed to come from a distribution in
Estimation Results
Three basic models were fitted to the data: the first was a least squares model fitted to the entire
data set; the second was again a least squares model fitted to the exceedances data (as previously
noted, exceedance is considered any occupancy value over 92% and clearly indicates the onset
of congestion); the third was the model shown in Eq. (6) fitted again to the exceedances data
using maximum likelihood estimation. The results for all three models are presented in Table 1.
It should be noted here that the Generalized Gamma Model (GGM) was initially estimated as the
11
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
distribution of the error terms in the third model. Then, since most of the other widely used
parametric forms of the Generalized Pareto distribution (such as the Exponential, the Beta, the
Weibull, the Standard Gamma, the Reverse Weibull, etc.) are nested within the GGM model,
they were evaluated using the likelihood ratio test. In most cases, as can be seen in Table 1, the
Reverse Weibull model could not be rejected and fitted the data very well and better than all
other models (this result was also reached using graphical diagnostics not shown in this paper).
Interestingly, Heckert et al. (21) found the Reverse Weibull as the “best” fitting distribution to
hurricane speeds since its upper tail is finite, lending itself well to modeling phenomena bounded
from above (in the occupancy case examined here, 100% occupancy can be considered as the
upper bound).
The results are very interesting. First, it appears that both the sign and the magnitude of
the trend term vary between the models estimated. Second, the least squares estimated trend
varies significantly between the full and the extreme values/exceedances (congestion) data set as
shown in the second and third columns of Table 1. Third, the trend term demonstrates marked
differences for some loops within given daily time periods. For example, congestion has
increased at loop 67 at similar magnitudes for all time periods; the estimated, via Maximum
Likelihood, slope coefficient is between 0.57 and 0.99. Congestion has remained unchanged at
loop 70, while it has increased significantly for most periods (excluding the period 13:30-17:00)
at loop 93. Interestingly, traffic at loop 87 has decreased for the morning peak period (06:30-
10:00) and for the early afternoon (13:30-17:00), but shows a large increase for the late morning
It is also worth noting that had the trend term been estimated via simple least squares the
results would have been statistically both biased and inconsistent for both data sets. This is a
typical example of estimation errors that occur when either an inappropriate technique is used to
analyze a given data set, or when a wrong functional form is specified during estimation (for a
more elaborate discussion see Washington et al. (22)). Of-course, since this study focuses on
12
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
extreme loadings that are indicative of congestion trends, the extreme values (exceedances) data
set is more appropriate. A least squares estimation on the congestion data yields similar results
as those of Eq. (6) for loop 67, but largely contradictory results for all other loops. Least squares
estimation for loop 70, for example, would suggest a decreasing trend in congestion rather than
no trend (estimated from Eq. (6)), and would also suggest no trend (or slightly decreasing) for
CONCLUSIONS
As is well known, most of the commonly used statistical models are based on explaining and
predicting “average behavior”; that is, most models are built around explaining and predicting
the outcome of the mean of the available data. And, while this approach is very satisfactory in
modeling most of the commonly encountered problems there are, many times, transportation
problems that deal with boundary conditions. These conditions can be, for example, congested
traffic, speed limit violators, extreme terminal delays, etc. In particular, severe traffic conditions
that may lead to a general network breakdown can be seen and treated as a physical calamity.
From a methodological standpoint all these problems have been dealt either by using the
altogether. But, an important question needs to be addressed: do outliers (also called extreme
values) contain information that can be useful in practice and are there any statistical techniques
that provide for their modeling? As such, the first part of this paper offered a short exposition of
the fundamental principles governing the analysis of extreme values and, in second offered a
traffic example of how the use of extreme value analysis can improve on the results of more
The analysis of Extreme Order Statistics (Extreme Values) may be described, in broad
terms, as an effort to make statistical inferences about the extreme values in a population or a
random process. This analysis is based on two fundamental principles: first, the generalized
13
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
extreme value distribution can be fitted to minima and/or maxima (that is, the minimum and/or
maximum value of a given series for a given time period) and, second, the generalized Pareto
distribution can be fitted to exceedances of a given measure over a threshold value. Using these
two principles inferences and predictions can be made regarding extreme values (outliers). To
demonstrate the principles of extreme value analysis, this paper presented a case study of traffic
The basic problem considered in this paper is whether traffic flows and congestion in
one of Athens’ main traffic arteries has increased over the last 4 years; further, with the 2004
Olympic Games fast approaching, it is imperative to determine whether key road axes that will
serve as main gateways for the Olympic Family are approaching endemic congestion conditions
or whether policy measures instituted by the authorities have begun to show an effect in tapering
congestion. To estimate this trend in congestion, two basic models were estimated: first, a least
squares models was fit, and then an extreme values trend model was estimated (via maximum
likelihood estimation). The results were very poignant. It appears, rather conclusively, that both
the sign and the magnitude of the trend term vary between the models estimated. Had the
commonly used least squares estimates been used to determine the trend the results would have
pointed toward either no increase in congestion or slight decrease. Extreme value techniques
point toward an increase in congestion, a finding with tremendous policy implications. Although
this case study is only a first step in demonstrating the usefulness of extreme value analyses in
transportation problems, it is hoped that more researchers will make use of the available
statistical tools and gain a deeper understanding of the occurrence of extreme events.
REFERENCES
1. Galambos, J. (1978) The Asymptotic Theory of Extreme Order Statistics. Wiley, New York.
2. Hipel, K.W. (ed.) (1994) Stochastic and Statistical Methods in Hydrology and
Environmental Engineering: Extreme Values. Kluwer Academic Publishers, New York.
14
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
3. Reiss, R.D. and Thomas, M. (2001) Statistical Analysis of Extreme Values 2nd Ed.
Birkhauser, Germany.
4. Gumbel, E. J. (1958) Statistics of Extremes. Columbina University Press, New York.
5. Fisher, R.A. and Tippett, L.H.C. (1928) Limiting forms of the frequency distributions of the
largest or smallest numbers of a sample. Procceedings of the Cambridge Philosophical Society,
24, 180-190.
6. Prescott, P. and Walden, A.T. (1980) Maximum likelihood estimation of the parameters of
the generalized extreme-value distribution. Biometrika, 67, 723-724.
7. Prescott, P. (1983) Maximum likelihood estimation of the parameters of the three parameter
generalized extreme-value distribution from censored samples. Journal of Statistical Computing
and Simulation, 16, 241-250.
8. Hosking, J.R.M. (1985) Algorithm AS 215: Maximum likelihood estimation of the
parameters of the generalized extreme-value distribution. Applied Statistics, 34, 301-210.
9. Macleod, A.J. (1989) AS R76 – A remark on algorithm AS 215: Maximum likelihood
estimation of the parameters of the generalized extreme-value distribution. Applied Statistics, 38,
198-199.
10. Smith, R.L. (1985) Maximum likelihood estimation in a class of nonregular cases.
Biometrika, 72, 67-92.
11. Hoskinh, J.R.M., Wallis, J.R. and Wood, E.F. (1985) Estimation of the generalized extreme-
value distribution by the method of probability-weighted moments. Technometrics, 27, 251-261.
12. Leadbetter, M.R., Lindgren, G. and Rootgen, H. (1983) Extremes and Related Properties of
Random Sequences and Series. Springer, New York.
13. Smith, R.L. (1986) Extreme value theory based on the r largest annual events, Journal of
Hydrology, 86, 27-43.
14. Tawn, J.A. (1988) An extreme-value theory model for dependent observations. Journal of
Hydrology, 101, 227-250.
15. Todorovic, P., and Zelenhasic, E. (1970). A stochastic model of flood analysis, Water
Resources Research, 6(6), 1641-1648.
16. Todorovic P. and Rouselle J (1977). Analyse stochastique des crues et son effet sur les
plaines inondables, Colloque canadien d'hydrologie: inondations, pp. 279-288, Edmonton,
Alberta.
17. Pickards, J. (1975) Statistical inference using extreme order statistics. Annals of Statistics, 3,
119-131.
15
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
18. Smith, R.L. (1987) Estimating tails of probability distributions. Annals of Statistics, 15,
1174-1207.
19. Bensalah, Y. (2000) Steps in applying extreme value theory to finance: A review. Working
paper 2000-20, Bank of Canada.
20. Hosking, J.R.M. and Wallis, J.R. (1987) Parameter and quantile estimation for the
generalized Pareto distribution, Technometrics, 29, 339-349.
21. Heckert, N.A. and Simiu, E. (1998) Estimates of hurricane wind speeds by “Peaks over
Threshold” method. ASCE Journal of Structural Engineering, 124(4), 445-449.
22. Washington, S., Karlaftis, M.G. Mannering, F.L. (2003) Statistical and Econometric
techniques for transportation data analysis, Chapman & Hall/CRC Press, Boca Raton, Fl.
16
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
Table 1. Parameter Estimates for Various Trend Models
All Data Extreme Values (congestion Data)
GGM3 RW4
LSE1 LSE1 ML2 scale shape scale
5
Loop 93
06:30-20:30 4.2 -0.02 2.51 0.02 2.6
06:30-10:00 7.6 ns 1.72 0.04 -4.6
10:00-13:30 3.7 ns 4.66 0.051
13:30-17:00 1.1 ns ns
17:00-20:30 0.9 -0.6 4.73 0.048
Loop 87
06:30-20:30 -1.3 ns 1.1 0.037
06:30-10:00 0.7 ns -1.94 0.03
10:00-13:30 -0.3 0.8 4.23 0.011
13:30-17:00 -0.7 ns -0.8 0.042
17:00-20:30 -2.5 0.7 2.94 0.027
Loop 67
06:30-20:30 0.97 1 0.84 0.03 1.85
06:30-10:00 1.23 0.8 0.87 0.048
10:00-13:30 0.61 0.9 0.94 0.051
13:30-17:00 NS 0.7 0.57 0.053
17:00-20:30 1.92 1.01 0.99 0.043
Loop 70
06:30-20:30 -1.92 -1.6 ns
06:30-10:00 1.1 ns ns
10:00-13:30 -4.7 -3.4 ns
13:30-17:00 0.7 ns ns
17:00-20:30 -0.7 -3.9 ns
1
Least Squares Estimate for the trend term
2
Maximum Likelihood Estimates for the trend term
3
Generalized Gamma Model
4
Reverse (Negative) Weibull
5
ns denotes no statistical significance (90% significance level)
17
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.
Figure 1: Loop detector location sequence along the SW to NE corridor
18
TRB 2003 Annual Meeting CD-ROM Paper revised from original submittal.