Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
www.elsevier.com/locate/spa
Urn schemes and reinforced random walks
P. Muliere
,S
t(x, ,)
for all x S and c E.
Theorem 2.3. For all n0 and all nite sequences (s
0
, . . . , s
n
) of elements of S,
P[X
0
= s
0
, . . . , X
n
= s
n
] = 0 if (s
0
, . . . , s
n
) is not admissible; otherwise
P[X
0
= s
0
, . . . , X
n
= s
n
] =
xS
_
cE
!
x
(c)1
i=0
(n
x
(c) + i)
t(x)1
)=0
() +
cE
n
x
(c))
_
, (2.4)
where the convention is made that
1
0
=1.
62 P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978
Proof. Let o=(s
0
, . . . , s
n
) be a nite sequence of elements of S. If o is not admissible,
it is evident that P[X
0
= s
0
, . . . , X
n
= s
n
] = 0. If o is admissible, then
P[X
0
= s
0
, . . . , X
n
= s
n
] =
n1
i=0
_
n
s
i
(c(s
i
, s
i+1
)) + m
s
i
(c(s
i
, s
i+1
))
cE
(n
s
i
(c) + m
s
i
(c))
_
, (2.5)
where m
s
0
0 and
m
s
i
(c) =
i1
r=0
I [s
r
= s
i
, c(s
r
, s
r+1
) = c]
for c E, and i = 1, . . . , n 1.
Given an element x of the sequence (s
0
, . . . , s
n1
), let (,
1
, . . . , ,
t(x)
) be the ordered
sequence of elements which follow x in o. Then the group of factors in (2.5) relative
to x can be rewritten as
n
x
(c(x, ,
1
))
cE
n
x
(c)
n
x
(c(x, ,
2
)) + I [,
1
= ,
2
]
1 +
cE
n
x
(c)
n
x
(c(x, ,
t(x)
)) +
t(x)1
i=1
I [,
i
= ,
t(x)1
]
t(x) 1 +
cE
n
x
(c)
=
cE
!
x
(c)1
i=0
(n
x
(c) + i)
t(x)1
)=0
() +
cE
n
x
(c))
assuming that
0
1
=0. From this, (2.4) follows.
If o = (s
0
, . . . , s
n
) and [ = (,
0
, . . . , ,
n
) are two equivalent sequences there is a
one-to-one correspondence between the transitions in o and those in [. Therefore,
either o and [ are both admissible or they are both not admissible. In the former case,
P[X
0
=s
0
, . . . , X
n
=s
n
] depends only on the transition counts t(x, ,), x, , S, and hence
it is the same as P[X
0
= ,
0
, . . . , Y
n
= ,
n
]. If, on the contrary, o and [ are both not
admissible, then P[X
0
= s
0
, . . . , X
n
= s
n
] = P[X
0
= ,
0
, . . . , X
n
= ,
n
] = 0. The argument
above proves the following result.
Corollary 2.6. The process {X
n
} is partially exchangeable.
For all x S, set
R
x
={, S: n
x
(c(x, ,)) 0}.
Then dene R
(0)
={x
0
} and, for all n1, set
R
(n)
=
_
xR
(n1)
R
x
.
The states in R
(n)
are those that the RUP process {X
n
} with initial state x
0
reaches
with positive probability in n steps. Let R =
n=0
R
(n)
. Then it follows from Theorem
2.3 that, for all n0,
P[X
0
R, . . . , X
n
R] = 1.
Remark 2.7. By a suitable denition of the law of motion q and the urn composition
function U one can have R = S or RS.
P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978 63
Set t
0
= 0 and, for all n1, dene
t
n
= inf {) t
n1
: X
)
= x
0
}. (2.8)
The process {X
n
} is recurrent if
P[X
n
= x
0
for innitely many n] = 1
or, equivalently, if
P
_
n=0
{t
n
}
_
= 1. (2.9)
When (2.9) holds, Theorem 7 of Diaconis and Freedman (1980) implies that {X
n
}
is a mixture of Markov chains. Since with probability one the states visited by the
process {X
n
} are elements of R, this means that there is a probability distribution j
on the set P of stochastic matrices on R R such that, for all n1 and all nite
sequences (x
1
, . . . , x
n
) of elements of R,
P[X
0
= x
0
, . . . , X
n
= x
n
] =
_
P
n1
)=0
(x
)
, x
)+1
) j(d). (2.10)
Let H be a random element of P with probability distribution j. The next theorem
describes the measure j by showing that the rows of H are independent Dirichlet
processes.
For all x R, let H(x) be the xth row of H and :(x) be the measure on R which
assigns mass n
x
(c) to q(x, c) for each c E such that n
x
(c) 0 and mass 0 to all
other elements of R; nally set
(x) = P[X
n
= x for innitely many n].
Lemma 2.11. Assume that {X
n
} is recurrent and let x
1
, . . . , x
m
, m1, be distinct el-
ements of R such that (x
i
) = 1 for i = 1, . . . , m. Then H(x
1
), . . . , H(x
m
) are mutually
independent random distributions on R and, for i = 1, . . . , m, the law of H(x
i
) is that
of a Dirichlet process with parameter :(x
i
).
Proof. Let x R be such that (x) =1. We begin by showing that H(x) is a Dirichlet
process of parameter :(x). For all n1, let Y
n
be the color of the ball picked from
the urn associated with x the nth time it is sampled. Then {Y
n
} is a P olya sequence of
parameter U(x). Let us order the elements of E by writing E ={c
1
, . . . , c
k
}. Then, as
a special case of the theorem of Blackwell and MacQueen (1973), there is a random
vector O(x) such that, with probability one,
lim
n
_
1
n
n
i=1
I [Y
i
= c
1
], . . . ,
1
n
n
i
I [Y
i
= c
k
]
_
= O(x) (2.12)
and O(x) has Dirichlet distribution with parameter (n
x
(c
1
), . . . , n
x
(c
k
)). Furthermore,
conditionally on O(x), the random variables of the sequence {Y
n
} are i.i.d. with distri-
bution O(x). We stress the fact that we are here referring to the Dirichlet distribution
in the general sense of Ferguson (1973) who allows some of the variables to be
degenerate at zero.
64 P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978
For all m1, let H
(m)
be the doubly innite matrix indexed by elements of R whose
(s, ,) entry H
(m)
(s, ,) is the number of transitions from state s R to state , R
divided by the number of transitions from state s made by the process {X
n
} up to time
m. Diaconis and Freedman (1980) showed that, when {X
n
} is recurrent, H
(m)
converges
almost surely in the topology of coordinate-wise convergence to a stochastic matrix H
whose probability distribution is the measure j appearing in (2.10). However, with
probability one,
lim
m
H
(m)
(x, q(x, c)) = lim
n
1
n
n
i=1
I [Y
i
= c]
for all c E, whereas lim
m
H
(m)
(x, ,) = 0 for all , R such that q(x, c) = ,
for all c E. This fact and (2.12) show that H(x) is a Dirichlet process on R with
parameter :(x). Notice that if x
1
, . . . , x
m
are dierent elements of R such that (x
i
) =1
for i = 1, . . . , m, then H(x
1
), . . . , H(x
m
) are independent since the sequences of colors
produced by the urns associated with (x
1
, . . . , x
m
) are independent.
The next lemma shows that every state in R is visited innitely often with probability
one by the process {X
n
} when it is recurrent.
Lemma 2.13. Assume that {X
n
} is recurrent. Then, for every x R, (x) = 1.
Proof. We prove by induction on n that for all n0 and x R
(n)
(x) = 1. In fact
(x
0
)=1 by assumption. Assume that (,)=1 for every state in R
(n)
and let x R
(n+1)
.
Then there is a , R
(n)
such that n
,
(c(,, x)) 0. Since (,) = 1,
P[X
n
= , for innitely many n|H = ] = 1 (2.14)
for all belonging to a set of j-measure one; furthermore, Lemma 2.11 implies that
H(,) is a Dirichlet process with parameter :(,) and thus H(,, x) has Beta distribution
with parameter (n
,
(c(,, x)),
cE
n
,
(c) n
,
(c(,, x))). Therefore there is a measurable
subset A
,
of P with j(A
,
) = 1 such that for all A
,
, (2.14) holds and
(,, x) 0. (2.15)
Eq. (2.14) says that , is a recurrent state for a Markov chain on R with transition
matrix A
,
. Moreover, since the class of recurrent states of a Markov chain is
closed, (2.15) implies that x is also recurrent for the same Markov chain, that is
P[X
n
= x for innitely many n|H = ] = 1
if A
,
. Therefore
[(x) =
_
A
,
P[X
n
= x for innitely many n|H = ]j(d) = 1.
Lemmas 2.11 and 2.13 immediately imply the following theorem which characterizes
the measure j appearing in (2.10).
Theorem 2.16. If {X
n
} is recurrent, the rows of H are mutually independent random
probability distributions on R and, for all x R, the law of H(x) is that of a Dirichlet
process with parameter :(x).
P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978 65
Remark 2.17. For all n1 and (x
1
, . . . , x
n
) R
n
, the law of the vector of probability
measures
(H(x
1
), . . . , H(x
n
))
is a product of Dirichlet processes. This kind of random probability measure has been
considered in a dierent context by Cifarelli and Regazzini (1978).
Remark 2.18. As a by-product of the previous theorem we get a dierent description
of the process {X
n
} when it is recurrent. With every x S, associate a random gen-
erator which produces an innite exchangeable sequence of colors {Y
n
(x)}, Y
n
(x)
E = {c
1
, . . . , c
k
} for all n. We assume that the sequences {Y
n
(x)}, x S, are inde-
pendent and that, given a random vector O(x) with Dirichlet distribution with pa-
rameter (n
x
(c
1
), . . . , n
x
(c
k
)), the random variables Y
n
(x) are conditionally i.i.d. with
distribution O(x). Set X
0
= x
0
. For all n1, if X
0
= x
0
, . . . , X
n1
= x
n1
S, let
X
n
= q(x
n1
, Y
t(x
n1
)+1
(x
n1
)) where t(x
n1
) counts the number of transitions from
state x
n1
made in the nite sequence (x
0
, . . . , x
n1
). When this process is recurrent it
as the same law as a recurrent RUP(S, E, U, q) process with initial state x
0
. In fact, this
would be the natural construction of a RUP process if one were to adopt the original
denition of partial exchangeability due to De Finetti (1938, 1959).
This fact stimulates some ideas for possible generalizations. To begin with, we may
try to drop the assumption that the de Finettis measure for the sequences {Y
n
(x)}
is Dirichlet. The process {X
n
} thus generated is still partially exchangeable. Under
what circumstances is it recurrent? In Section 4 we will present an example of such a
generalization by considering the case where {Y
n
(x)} are exchangeable sequences of
dichotomous random variables with a de Finettis measure which need not be a Beta.
Another possible generalization comes from replacing the assumption of independence
for the sequences {Y
n
(x)}, x S, with that of conditional independence; this opens the
doors to the analysis of nonparametric hierarchical models by means of mixtures of
reinforced urn processes.
Let (s
0
, . . . , s
n
) be an admissible sequence. Given that we observed X
0
=s
0
, . . . , X
n
=s
n
,
for all x S the urn associated with x contains
n
x
(c) + !
x
(c) (2.19)
balls of color c, for each c E; set
U(x) to be the k-uple which describes this
composition. The following result is almost obvious but has important implications
when one uses the process {X
n
} for predictive purposes.
Theorem 2.20. Assume that {X
n
} is recurrent. Let (s
0
, . . . , s
n
) be an admissible
sequence such that P[X
0
= s
0
, . . . , X
n
= s
n
] 0. Then, given X
0
= s
0
, . . . , X
n
= s
n
, the
process (X
n
, X
n+1
, . . .) RUP(S, E,
U, q) with initial state s
n
and is recurrent.
The probability distribution j has a nice conjugacy property, as follows from
Theorem 2.16 and the previous result.
66 P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978
Corollary 2.21. Assume that {X
n
} is recurrent. Let (s
0
, . . . , s
n
) be an admissible
sequence such that P[X
0
=s
0
, . . . , X
n
=s
n
] 0. Then, given X
0
=s
0
, . . . , X
n
=s
n
, the rows
of H are mutually independent random probability distributions and, for all x R,
the law of H(x) is that of a Dirichlet process on R with parameter
:(x) +
cE
!
x
(c)o(q(x, c)),
where o(,) is the measure which gives mass 1 to , and 0 to all other elements of R,
for each , R.
Following Diaconis and Freedman (1980), we dene an x
0
-block for the process
{X
n
} to be a nite sequence of states which begins by x
0
and contains no further
x
0
. Endow the countable space S
, B
2
S
, . . ., be the sequence
of the successive x
0
-blocks in {X
n
}: the fact that {B
n
} is exchangeable was proved by
Diaconis and Freedman (1980). This obvioulsy implies that if is a function which
maps measurably S
(c
0
) 0 and n
k
)=0
(1 J
)
) where the {J
)
} are mutually independent [0, 1] random variables.
The beta-Stacy process arises by taking the J
)
to be beta variables. These processes are
P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978 67
widely used in Bayesian nonparametric survival studies. In this section we use RUPs
to characterize the beta-Stacy process.
Let E ={w, b} where w is for white and b for black. Let x
0
=0 and assume that
n
0
(w) = 0 whereas, for all )1, n
)
(w) 0. Finally, dene the law of motion q by
setting
q(x, b) = x + 1 and q(x, w) = 0
for all x S. Let {X
n
} RUP(S, E, U, q) with initial state x
0
= 0.
Therefore, the process {X
n
} begins at 0 and, given that at stage n0 it is at state
x S, it moves to state x + 1 if the ball sampled from the urn associated with x is
black and to state 0 if it is white. Urn compositions are updated in the usual way. If
n
)
(b) =0 for some )1, let N =min{)1: n
)
(b) =0} and note that, with probability
one, the process {X
n
} visits only the states {0, . . . , N}.
By applying (2.4) we get for all admissible nite sequences (x
0
, . . . , x
n
) of elements
of S
P[X
0
= x
0
, . . . , X
n
= x
n
] =
)=1
B(n
)
(w) + t(), 0), n
)
(b) + t(), ) + 1))
B(n
)
(w), n
)
(b))
,
where, for all a, b 0, B(a, b) is the usual beta integral.
For all n1, let 1
n
=X
t
n
1
where the sequence of stopping times {t
n
} was dened
in (2.8). When {X
n
} is recurrent, 1
n
is the last state of the nth 0-block, that is
1
n
= (B
n
).
Equivalently, 1
n
measures the length of the sequence of states strictly between the nth
zero and the (n + 1)th zero in the sequence {X
n
}.
Lemma 3.23. The process {X
n
} is recurrent if and only if
lim
n
n
i=0
n
i
(b)
n
i
(b) + n
i
(w)
= 0. (3.24)
Proof. For all n1,
P[t
1
n] =
n1
i=0
n
i
(b)
n
i
(b) + n
i
(w)
.
Hence, P[t
1
] = 1 if (3.24) holds.
By induction on n, assume that for an n1, P[
n
i=1
{t
i
}] = 1. Then
P[t
n+1
] =
_
n
i=1
{t
i
}
P[t
n+1
| 1
1
, . . . , 1
n
] dP
=
_
n
i=1
{t
i
}
_
1 lim
k
P[t
n+1
k | 1
1
, . . . , 1
n
]
_
dP.
However, for k max(1
1
, . . . , 1
n
) + 1 = !
P[t
n+1
k | 1
1
, . . . , 1
n
]6
k1
)=!
n
)
(b)
n
)
(b) + n
)
(w)
68 P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978
and the last quantity goes to 0 for k because of (3.24). Hence P[t
n+1
] =1
and this proves that (3.24) is a sucient condition for the recurrence of {X
n
}.
In order to prove that the condition is also necessary, notice that P[t
1
] = 1 if
{X
n
} is recurrent. Therefore
lim
n
n
i=0
n
i
(b)
n
i
(b) + n
i
(w)
= lim
n
P[t
1
n + 1] = 0.
Therefore, whenever (3.24) holds, the sequence {1
n
} is exchangeable and, by de
Finettis Representation Theorem, there exists a random distribution function F such
that, given F, the random variables of the sequence {1
n
} are i.i.d. with distribution F.
With the next theorem we prove that F is a beta-Stacy process on S with parameters
{(n
)
(w), n
)
(b))} (Walker and Muliere, 1997); this means that F is a neutral to the
right process such that F(0) =0 with probability one and, for )1, [F()) F() 1)]
has the same distribution as
P
)
= W
)
)1
i=1
(1 W
i
), (3.25)
where {W
)
} is a sequence of independent random variables such that, for all )1, W
)
has distribution Beta(n
)
(w), n
)
(b)). Note that in (3.25) we assumed that
0
i=1
=1 and
that Beta(a, 0) is the point mass at 1 for all a 0.
Theorem 3.26. F is a beta-Stacy process on S with parameters
{n
)
(w), n
)
(b)}.
Proof. Fix n1 and integers 0 = k
0
k
1
6 6k
n
. Then,
P[1
1
=k
1
, . . . , 1
n
= k
n
]
=P[X
1
= 1, . . . , X
k
1
= k
1
, X
k
1
+1
= 0, . . . , X
n1
)=1
k
)
+n1
= 0, . . . , X
n
)=1
k
)
+n1
= k
n
, X
n
)=1
k
)
+n
= 0]
=
)=1
B(n
)
(w) + t(), 0), n
)
(b) + t(), ) + 1))
B(n
)
(w), n
)
(b))
=E
_
_
)=1
W
t( ), 0)
)
(1 W
)
)
t( ), )+1)
_
_
,
where {W
)
} is a sequence of independent random variables such that, for all )1,
W
)
has distribution Beta(n
)
(w), n
)
(b)). However
t(), 0) =
_
_
n
i=1
I [k
i
= )] if ) {k
1
, . . . , k
n
},
0 otherwise.
Also, because the k
i
s are ordered, for all )1,
t(), ) + 1) =
n1
i=0
(n i) I [k
i
6)6k
i+1
1]
P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978 69
assuming that the indicator of an empty set is 0. Therefore
P[1
1
=k
1
, . . . , 1
n
= k
n
]
=E
_
_
n
)=1
W
k
)
k
1
1
)=0
(1 W
)
)
n
k
2
1
)=k
1
(1 W
)
)
n1
k
n
1
)=k
n1
(1 W
)
)
_
_
=E
_
_
W
k
1
k
1
1
)=0
(1 W
)
) W
k
n
k
n
1
)=0
(1 W
)
)
_
_
=E
_
_
n
)=1
P
k
)
_
_
where, for all k1, we dened
P
k
= W
k
k1
i=1
(1 W
i
)
and we used the convention that
)1
)
=1. Since {1
n
} is exchangeable, this shows
that, for all n1 and all sequences (k
1
, . . . , k
n
) of strictly positive elements of S,
P[1
1
= k
1
, . . . , 1
n
= k
n
] = E
_
_
n
)=1
P
k
)
_
_
. (3.27)
If one or more elements of (k
1
, . . . , k
n
) are 0, then trivially
P[1
1
= k
1
, . . . , 1
n
= k
n
] = 0. (3.28)
Eqs. (3.27) and (3.28) are sucient to prove that the de Finettis measure of the
sequence {1
n
} is that for the beta-Stacy process on S with parameters {(n
)
(w), n
)
(b))}.
Note that F is a random distribution function since we assumed the recurrence condition
(3.24) holds.
Remark 3.29. One could prove the result by applying Theorem 2.16. We preferred a
more direct proof based only on the expression for the nite dimensional distributions
of the process {X
n
} since it suggests a possible generalization that will be considered
in Section 4.
Remark 3.30. For all k1,
P[1
1
= k] =
n
k
(w)
n
k
(b) + n
k
(w)
k1
)=0
n
)
(b)
n
)
(b) + n
)
(w)
.
Also, for all n1 and k1,
P[1
n+1
=k|1
1
, 1
2
, . . . , 1
n
] =
n
k
(w)+n
k
n
k
(b)+n
k
(w)+n
k
+m
k
k1
)=1
n
)
(b)+m
)
n
)
(b)+n
)
(w)+n
)
+ m
)
(3.31)
on the set
i=1
{t
i
} if, for all 06)6k 1, we dene n
)
=
n
i=1
I [1
i
= )] and
m
)
=
n
i=1
I [1
i
)]. Therefore, if (3.24) holds, (3.31) is true with probability one.
70 P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978
Remark 3.32. By specifying dierent sequences of parameters
{(n
)
(w), n
)
(b))},
one may obtain dierent laws for the random distribution F which are commonly
used in Bayesian nonparametrics; for example, the Dirichlet process or the simple
homogeneous process. For more details on this issue we refer to the paper of Walker
and Muliere (1997).
4. Neutral to the right processes via reinforced random walks
As in the previous section, let S = {0, 1, 2, . . .} and let F represent a neutral to
the right process on S (Doksum, 1974). For all k1, set F
k
to be the random mass
assigned by F to {1, . . . , k}. Then, by construction,
F
k
= 1
k
)=1
(1 J
)
), (4.33)
where {J
)
} is a sequence of mutually independent random variables dened on (0, 1).
Let E(1 J
)
) = j
)
for ) = 1, 2, . . . . We state without proof:
Lemma 4.34. The random measure F on S dened by (4.33) is a.s. a random
probability distribution if, and only if,
)=1
j
)
= 0. (4.35)
The aim of this section is to describe a scheme for generating a partially exchange-
able and recurrent process {Z
n
} based on F; this process generalizes the RUP process
{X
n
} introduced in Section 3 along the lines suggested in Remark 2.18.
Let the state space S, the set of colors E and the law of motion q be the same as
those dened in Section 3. Set {Y
n
(0)} to be an innite sequence of black colors. For
all )1, let {Y
n
())} be an innite exchangeable sequence of random elements of E
such that, given J
)
, the random variables Y
n
()) are i.i.d. and
P[Y
n
()) = w|J
)
] = 1 P[Y
n
()) = b|J
)
] = J
)
.
Set Z
0
= 0. For all n1, given Z
0
= 0, . . . , Z
n1
= :
n1
, let
Z
n
= q(:
n1
, Y
t(:
n1
)+1
(:
n1
)).
Recall that t(:
n1
) counts the number of transitions from state :
n1
made in the nite
admissible sequence (0, . . . , :
n1
). This is equivalent to say that, for all n1 and for
all nite sequences (0, . . . , :
n1
) of elements of S,
P[Z
n
= 0|Z
0
= 0, . . . , Z
n
= :
n1
] =
E(J
t(:
n1
, 0)+1
:
n1
(1 J
:
n1
)
t (:
n1
, :
n1
+1)
)
E(J
t(:
n1
, 0)
:
n1
(1 J
:
n1
)
t (:
n1
, :
n1
+1)
)
= 1 P[Z
n
= :
n
+ 1|Z
0
= 0, . . . , Z
n1
= :
n1
]
P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978 71
if :
n1
1, whereas
P[Z
n
= 1|Z
0
= 0, . . . , Z
n1
= :
n1
] = 1
if :
n1
= 0. For all )1 and n, m0, let us dene
z
)
(n, m) = E[J
n
)
(1 J
)
)
m
].
Then one may check that, for all admissible sequences (0, . . . , :
n
) of elements of S,
P[Z
0
= 0, . . . , Z
n
= :
n
] =
)=1
z
)
(t(), 0), t(), ) + 1)). (4.36)
The process {Z
n
} is partially exchangeable. With an argument analogous to that of
Lemma 3.23 it is not dicult to show that {Z
n
} is recurrent when (4.35) holds and
so {Z
n
} is a mixture of Markov chains. Notice that the particular scheme described in
Section 3 arises when the random variables J
)
have distribution Beta(n
)
(w), n
)
(b)).
For all n1, let 1
n
S be the last coordinate of the nth 0-block of the process {Z
n
}
when it is recurrent. Then {1
n
} is exchangeable. The next theorem describes the de
Finettis measure of the sequence {1
n
}: it can also be considered as a characterization
of a neutral to the right process via a mixture of Markov chains.
Theorem 4.37. The de Finettis measure for the exchangeable sequence {1
n
} is the
neutral to the right process prior dened by (4.33).
Proof. As in the proof of Theorem 3.26, we aim to show that, for all n1 and all
nite sequences (k
1
, . . . , k
n
) of strictly positive elements of S,
P[1
1
= k
1
, . . . , 1
n
= k
n
] = E
_
_
n
)=1
P
k
)
_
_
,
where, for all )1, P
)
is the random variable allocating the random mass to ) S
dened by P
)
= F
)
F
)1
= J
)
)1
i=1
(1 J
i
).
Eq. (4.36) implies that
P[1
1
= k
1
, . . . , 1
n
= k
n
] =
)=1
z(t(), 0), t(), ) + 1)), (4.38)
where, for all x, , S, t(x, ,) counts the transitions from x to , made in the admissible
sequence
(0, 1, . . . , k
1
, 0, 1, . . . , k
2
, . . . , 0, 1, . . . , k
n
, 0).
However,
)=1
z
)
(t(), 0), t(), ) + 1)) =
)=1
E[J
t ( ), 0)
)
(1 J
)
)
t ( ), )+1)
].
From here the proof proceeds analogously to that of Theorem 3.26.
72 P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978
Remark 4.39. For all n1 and all k1,
P[1
n+1
= k|1
1
, . . . , 1
n
] =
z
k
(n
k
+ 1, m
k
)
z
k
(n
k
, m
k
)
k1
)=1
z
)
(n
)
, m
)
+ 1)
z
)
(n
)
, m
)
)
,
where for all 06)6k 1, we dene n
)
=
n
i=1
I [1
i
=)] and m
)
=
n
i=1
I [1
i
)]. This
form for the predictive distribution characterizes neutral to the right processes (Walker
and Muliere, 1999).
5. An RUP for a multiple state process
Some nonparametric models for survival analysis arise naturally by specifying the
dening elements of a RUP process. An example where a disease process is under
observation is the competing risk model for cancer patients of Lagakos (1976). Three
disease states are introduced: at state 0 the presence of cancer for the patient is ac-
knowledged, state 1 indicates a progressive state of the cancer and state 2 is death.
A patient in state 0 is in a competing risk situation in that death due to cancer is in
competition with death due to other causes. Another example is the survival model of
Walker (1998) where a surrogate endpoint or intermediate response variable related to
survival time is available. During follow up, the surrogate endpoint may occur earlier
than the survival time and give potentially important informations for the prediction of
the patients ultimate survival time.
Here we put the three state models of Lagakos and Walker into our RUP framework.
This is readily extended to a higher number of states. For modelling situations like those
of Lagakos and Walker we dene a process
{X
n
} RUP(S, E, U, q).
Let S={(!, x): !, x=0, 1, . . .} where x might indicate the time since we started observing
a patient and ! is a level: ! could measure the level of a certain disease or could be a
ag whose value 1 or 0 indicates that a surrogate endpoint has been observed or not.
Let E ={b, w, r} where w is for white, b for black and r for red. We dene the
law of motion q for the process {X
n
} by setting, for all (!, x) S,
q((!, x), w) = (0, 0), q((!, x), b) = (!, x + 1) and q((!, x), r) = (! + 1, x + 1).
Hence, whenever a white ball is sampled the process {X
n
} is set back to (0, 0): this
models death of the patient under observation. If a black ball is sampled the patient
proceeds to the next time period at the same disease level, whereas the level is increased
by one unit whenever a red ball is picked from the urn.
We will assume that n
(0, 0)
(w) = n
(0, 0)
(r) = 0. If for a given !0, n
(!, x)
(r) = 0 for
all x = 0, 1, . . . , then with probability one the process {X
n
} will visit states of level at
most !. In the rest of the section we will restrict the analysis to the two-level case
where n
(1, x)
(r) = 0 for all x = 0, 1, . . . ; extensions to the general nite-level case are
straightforward. Set M =inf {x0: n
(0, x)
(r) 0} and N =inf {x0: n
(0, x)
(b) =0} with
the usual convention that inf () =. We will assume that M is nite and M6N.
When the last assumption does not hold, the process {X
n
} is not dierent from that
studied in Section 3. In fact, we can always recover the process of that section by
P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978 73
setting n
(0, x)
(r) =0 for all x =0, 1, . . . . Note that when n
(0, x)
(w) =0 for all x =0, 1, . . .
there is no competing risk and a patient whose disease is at level 0 can die only after
the disease has stepped one level higher.
Dene the projections L : S {0, 1, . . .} and D : S {0, 1, . . .} by setting, for all
(!, x) S,
L((!, x)) = ! and D((!, x)) = x.
As before, we consider the sequence of stopping times {t
n
} dened by setting t
0
= 0
and, for all n1,
t
n
= inf {) t
n1
: X
)
= (0, 0)}.
We then introduce a new sequence of stopping times {o
n
} where
o
n
= inf {) t
n1
: L(X
)
) = 1}
for n = 1, 2, . . . . We write t
n
o
n
for min(t
n
, o
n
): that is t
n
o
n
is the time when
{X
n
} visits (0, 0) for the nth time or when it rst reaches a state of level 1 after the
(n 1)th visit to state (0, 0), whichever occurs rst.
Lemma 5.40. The process {X
n
} is recurrent if and only if
lim
n
n
i=1
n
(!, i)
(b)
cE
n
(!, i)
(c)
= 0 (5.41)
for ! = 0, 1.
Proof. For all n1,
P[t
1
o
1
n] =
n1
i=1
n
(0, i)
(b)
cE
n
(0, i)
(c)
.
Hence, P[t
1
o
1
] = 1 if (5.41) holds. Thus,
P[t
1
] =
_
{t
1
o
1
}
P[t
1
|X
t
1
o
1
] dP
=
_
{o
1
t
1
,t
1
o
1
}
P[t
1
|X
o
1
] dP+
_
{o
1
t
1
,t
1
o
1
}
P[t
1
|X
t
1
] dP
= 1
_
{o
1
t
1
,t
1
o
1
}
lim
n
P[t
1
n|X
o
1
] dP
= 1
_
{o
1
t
1
,t
1
o
1
}
lim
n
n1
i=D(X
o
1
)
n
(1, i)
(b)
cE
n
(1, i)
(c)
dP
= 1.
where the last equality follows from (5.41).
Now prove by induction on n that if P[
n
i=1
{t
i
}]=1, then P[t
n+1
o
n+1
]=
1 and P[t
n+1
] = 1; the argument is analogous to that in the proof Lemma 3.23
but uses conditional probabilities given
X
t
1
o
1
, X
t
1
, . . . , X
t
n
o
n
, X
t
n
.
74 P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978
Therefore, (5.41) is sucient to prove that {X
n
} is recurrent. It is also necessary. In
fact, if P[t
1
] = 1,
lim
n
n
i=1
n
(0, i)
(b)
cE
n
(0, i)
(c)
= lim
n
P[t
1
o
1
n + 1] = 0.
Furthermore, since M is nite and M6N, P[o
1
= M + 1] 0 and, for all nM + 1,
n
i=M+1
n
(1, i)
(b)
cE
n
(1, i)
(c)
= P[t
1
n + 1|o
1
= M + 1]6
P[t
1
n + 1]
P[o
1
= M + 1]
.
Hence,
lim
n
n
i=1
n
(1, i)
(b)
cE
n
(1, i)
(c)
6 lim
n
n
i=M+1
n
(1, i)
(b)
cE
n
(1, i)
(c)
= 0.
On the space S
,
((s
0
, . . . , s
n
)) =
_
n
i=0
I [L(s
i
) = 0] 1,
n
i=0
I [L(s
i
) = 1]
_
.
When {X
n
} is recurrent, the sequence {B
n
} of its (0, 0)-blocks is exchangeable and
this implies that { (B
n
)} is exchangeable. For all n1, set (U
n
, J
n
) = (B
n
). Thus
U
n
is equal to the right coordinate of the last state at level 0 in the (0, 0)-block B
n
:
that is, U
n
measures the time spent by patient n with disease at level 0. If we indicate
with 1
n
the length of B
n
minus one, that is if 1
n
is equal to the total survival time of
patient n, then J
n
= 1
n
U
n
measures the length of the time spent by patient n with
disease at level 1 including the day when the disease made the transition from level 0
to level 1.
Using Theorem 2.3 or Theorem 2.16 one can nd an expression for the nite-
dimensional distributions of the exchangeable process {(U
n
, J
n
)}. We will just touch
upon this by computing the distribution of (U
1
, J
1
): this takes on interest for predictive
purposes. In fact, when the expression of the distribution of (U
1
, J
1
) is known, it
follows from Theorem 2.20 and Corollary 2.21 that the expression of the conditional
distribution of (U
n+1
, J
n+1
) given ((U
1
, J
1
), . . . , (U
n
, J
n
)) is immediately available: this
is true even when data are censored.
Note that P[U
1
=0, J
1
=k] =0 for k0. For h, k 0 it follows from Theorem 2.16
that
P[U
1
= h, J
1
= 0] =
_
P
_
h1
x=1
((0, x), (0, x + 1))
_
((0, h), (0, 0))j(d), (5.42)
P[U
1
= h, J
1
= k] =
_
P
_
h1
x=1
((0, x), (0, x + 1))
_
((0, h), (1, h + 1))
_
h+k1
x=h+1
((1, x), (1, x + 1))
_
((1, k), (0, 0)) j(d). (5.43)
P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978 75
Furthermore Theorem 2.16 implies that, if H is a random element of P with probability
distribution j then, for all x = 1, 2 . . . , the random vector
(H((0, x), (0, 0)), H((0, x), (0, x + 1)), H((0, x), (1, x + 1)))
=(O
(0)
x
(w), O
(0)
x
(b), O
(0)
x
(r))
has Dirichlet distribution with parameter
(n
(0, x)
(w), n
(0, x)
(b), n
(0, x)
(r)),
whereas
H((1, x), (0, 0)) = 1 H((1, x), (1, x + 1)) = W
(1)
x
has distribution Beta(n
(1, x)
(w), n
(1, x)
(b)). Therefore, Eqs. (5.42) and (5.43) become
respectively
P[U
1
= h, J
1
= 0] =E
_
O
(0)
h
(w)
h1
x=1
O
(0)
x
(b)
_
=
n
(0, h)
(w)
cE
n
(0, h)
(c)
h1
x=1
n
(0, x)
(b)
cE
n
(0, x)
(c)
, (5.44)
P[U
1
= h, J
1
= k] =E
__
h1
x=1
O
(0)
x
(b)
_
O
(0)
h
(r)
_
h+k1
x=h+1
(1 W
(1)
x
)
_
W
(1)
h+k
_
=
n
(0, h)
(r)
cE
n
(0, h)
(c)
_
h1
x=1
n
(0, x)
(b)
cE
n
(0, x)
(c)
_
n
(1, h+k)
(w)
cE
n
(1, h+k)
(c)
_
h+k1
x=h+1
n
(1, x)
(b)
cE
n
(1, x)
(c)
_
. (5.45)
Example 5.46. When ! is a ag whose value indicates if a surrogate endpoint has
been observed or not, let j
n
=o
n
t
n1
1 be the surrogate response time for the nth
patient under observation.
Using (5.44) and (5.45) we may easily compute the joint distribution of (1
1
, j
1
). In
fact, for all t 0,
P[1
1
= t, j
1
t] =P[U
1
= t, J
1
= 0]
=
n
(0, t)
(w)
cE
n
(0, t)
(c)
t1
x=1
n
(0, x)
(b)
cE
n
(0, x)
(c)
whereas, if u t,
P[1
1
= t, j
1
= u] = P[U
1
= u, J
1
= t u]
=
n
(0, u)
(r)
cE
n
(0, u)
(c)
_
u1
x=1
n
(0, x)
(b)
cE
n
(0, x)
(c)
_
n
(1, t)
(w)
cE
n
(1, t)
(c)
_
t1
x=u+1
n
(1, x)
(b)
cE
n
(1, x)
(c)
_
.
76 P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978
These expressions, together with Theorem 2.20 and Corollary 2.21, justify a nonpara-
metric estimator for the distribution of (1, j) recently proposed by Walker (1998).
Example 5.47. Assume that n
(0, x)
(w)=0 for x=1, 2, . . . . This corresponds to a situation
where there is no competing risk and a patient dies only when the disease level is 1.
Butler and Huzurbar (1997) report that a model with the same dynamics has been
described by Gross et al. (1971) for kidney patients. With more than one level for
the disease process, similar models have also been considered for representing the
progressive states of AIDS for a patient as described by Longini et al. (1989).
In this situation, Theorem 2.16 implies that for all n1 and sequences ((h
1
, k
1
), . . . ,
(h
n
, k
n
)) with h
i
, k
i
0
P[(U
1
, J
1
) = (h
1
, k
1
), . . . , (U
n
, J
n
) = (h
n
, k
n
)] = E
_
n
i=1
P
(0)
h
i
P
(1)
k
i
h
i
)=1
(1 W
(1)
)
)
_
(5.48)
where, for x = 1, 2, . . . , and ! = 0, 1,
P
(!)
x
= W
(!)
x
x1
)=1
(1 W
(!)
)
)
and the random variables W
(!)
x
are mutually independent and such that W
(0)
x
has distri-
bution Beta(n
(0, x)
(r), n
(0, x)
(b)) whereas W
(1)
x
has distribution Beta(n
(1, x)
(w), n
(1, x)
(b)).
From (5.48) one can get a description of the bivariate process which acts as de
Finettis measure for the exchangeable sequence {(U
n
, J
n
)}. Furthermore (5.48) implies
that the de Finettis measure for the exchangeable sequence {U
n
} is a beta-Stacy
process with parameter {(n
(0, ))
(r), n
(0, ))
(b))}. The next theorem will show that this
result holds true in a more general situation.
Theorem 5.49. When {X
n
} is recurrent, the de Finettis measure for the exchangeable
process {U
n
} is a beta-Stacy process.
Proof. Instead of using a direct argument based on the nite-dimensional distributions
of the exchangeable process {(U
n
, J
n
)}, we will recur to an urn argument which seems
more appropriate in the present context.
Coupled with the process {X
n
}, we consider a process
{
X
n
} RUP(
S,
E,
U, q)
of the type studied in Section 3 and dened by setting
S ={0, 1, . . .},
E ={, b} where
is for pink and b for black and, for all x
S,
q(x, ) = 0 and q(x, b) = x + 1.
Finally, for x = 0, 1, . . . , we assume that
n
x
() = n
(0, x)
(w) + n
(0, x)
(r) and n
x
(b) = n
(0, x)
(b).
The process {
X
n
} is partially exchangeable and it is recurrent when condition (5.41)
holds because of Lemma 3.23. Hence, if
1
n
represents the last coordinate of the
nth 0-block of {
X
n
}, then {
1
n
} is exchangeable and Theorem 3.26 implies that its
P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978 77
de Finettis measure is a beta-Stacy process with parameter {( n
)
(), n
)
(b))}. However,
for all n1 and all nite sequences (h
1
, . . . , h
n
) of nonnegative integers
P[U
1
= h
1
, . . . , U
n
= h
n
] = P[
1
1
= h
1
, . . . ,
1
n
= h
n
]. (5.50)
This fact and the unicity of the de Finettis measure prove that the de Finettis
measure for the sequence {U
n
} is a beta-Stacy process with parameter {(n
(0, ))
(w) +
n
(0, ))
(r), n
(0, ))
(b))}.
6. Concluding remarks
A few obvious possibilities for generalizations come to mind when one accepts the
perspective which generated the idea of reinforced urn processes. The most immediate
one is perhaps that of extending the analysis to continuous time processes; in this
direction the work of Freedman (1996) on mixtures of continuos-time Markov chains
may be of help.
In the terminology of Hill et al. (1987) the urn function which was used for gen-
erating RUPs is the identity; when x S is the current state of the process a ball of
color c is added to urn U(x) with probability equal to the proportion of the number of
balls of color c contained in the urn. We may conceive reinforced random walks on a
state space of urns generated by urn functions dierent from the identity. The question
is: for what urn functions are these processes partially exchangeable?
Another natural candidate for generalizations is the updating mechanism for urn
compositions; it might be of interest to consider situations where the extraction of a
color from an urn need not increase its probability of being extracted again from the
same urn as is the case for the Bernard Friedmans urn, Friedman (1949).
At this point in time it is not clear what applications the above ideas might have.
However, one possibility for the continuous time version of RUPs would be Bayesian
nonparametric hierarchical analysis where the Markov chains are left unobserved at a
lower stage in the model.
References
Blackwell, D., MacQueen, J.B., 1973. Ferguson distributions via P olya-urn schemes. Ann. Statist. 1, 353355.
Butler, R.W., Huzurbar, A.V., 1997. Stochastic network models for survival analysis. J. Amer. Statist. Assoc.
92, 246257.
Cifarelli, D.M., Regazzini, E., 1978. Problemi statistisci non parametrici in condizioni di scambiabilit a
parziale. Impiego di medie associative. In: Istituto di Matematica Finanziaria dellUniversit a di Torino,
Serie III, No. 12.
Coppersmith, D., Diaconis, P., 1986. Random walk with reinforcement. Unpublished.
De Finetti, B., 1938. Sur la condition d equivalence partielle, In: VI Colloque Geneve, Act. Sci. Ind. No.
739. Herman, Paris.
De Finetti, B., 1959. La probabilit a e la statistica nei rapporti con linduzione, secondo i diversi punti di
vista, in: Atti Corso CIME su Induzione e Statistica. Varenna, pp. 1115.
Diaconis, P., 1988. Recent progress on de Finettis notions of exchangeability. In: Bernardo, J.M., DeGroot,
M.H., Lindley, D.V., Smith, A.F.M (Eds.), Bayesian Statistics 3. Oxford University Press, Oxford.
Diaconis, P., Freedman, D., 1980. De Finettis Theorem for Markov chains. Ann. Probab. 8, 115130.
Doksum, K.A., 1974. Tailfree and neutral random probabilities and their posterior distributions. Ann. Probab.
2, 183201.
78 P. Muliere et al. / Stochastic Processes and their Applications 88 (2000) 5978
Ferguson, T.S., 1973. A Bayesian analysis of some nonparametric problems. Ann. Statist. 1, 209230.
Friedman, B., 1949. A simple urn model. Comm. Pure Appl. Math. 2, 5970.
Freedman, D.A., 1996. De Finettis theorem in continuous time. Technical Report No. 456, Statistics
Department, University of California, Berkeley.
Gross, A.J., Clark, V.A., Liu, V., 1971. Estimation of survival parameters where one of two organs must
function for survival. Biometrics 27, 369377.
Hill, B.M., Lane, D., Sudderth, W.D., 1987. Exchangeable urn processes. Ann. Probab. 15, 15921596.
Lagakos, S., 1976. A stochastic model for censored survival time in the presence of an auxiliary variable.
Biometrics 32, 551559.
Longini, I., Clark, W., Byers, R., Ward, J., Darrow, W., Lemp, G., Hethcote, H., 1989. Statistical analysis
of the stages of HIV infection using a Markov model. Statist. Med. 8, 831843.
Mauldin, R.D., Sudderth, W.D., Williams, S.C., 1992. P olya trees and random distributions. Ann. Statist. 20,
12031221.
Pemantle, R., 1988. Phase transition in reinforced random walk and RWRE on trees. Ann. Probab. 16,
12291241.
Walker, S.G., 1998. A nonparametric approach to a survival study with surrogate endpoints. Biometrics 54,
662672.
Walker, S.G., Muliere, P., 1997. Beta-Stacy processes and a generalization of the P olya-urn scheme. Ann.
Statist. 25, 17621780.
Walker, S.G., Muliere, P., 1999. A characterization of a neutral to the right prior via an extension of
Johnsons sucientness postulate. Ann. Statist. 27, 589599.