Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Xuzi Zhou
Dept. of Computer Science
University of Kentucky
United States
xiwei@netlab.uky.edu
Germany
eosten@ix.urz.uni-heidelberg.de
United States
xuzizhou@netlab.uky.edu
Hui Lin
Jinze Liu
University of Kentucky
University of Kentucky
United States
United States
hui.lin@uky.edu
liuj@netlab.uky.edu
compare, and buy specific items or groups of items without
the necessity of personal presence. To sell their products
better, most online shopping websites provide
recommendation information to the customers 1 who have
visited their website before. Recommender systems are
mechanisms that can be used to help users to make purchase
decisions.
A recommender system is actually a program that
utilizes algorithms to predict customers purchase interests
by profiling their shopping patterns. There are many
research publications about recommender systems since the
mid-1990s [1]. Different approaches and models have been
proposed and applied to real world industrial applications.
The most popular recommendation technique is the
Collaborative Filtering (CF) model [2, 3]. In CF, previous
transactions are analyzed in order to establish connections
between users and products. When recommending items to
a user, the CF-based recommender systems try to find
information related to the current user to compute ratings for
every possible item. Items with the highest rating scores will
be presented to the user.
In the research area of recommender systems, most
models work with rating datasets, such as Netflix movie
rating data [21]. The rating information is very important for
obtaining good prediction accuracy because it precisely
indicates users preferences and the degree of their interest
on certain items. However, the rating information is not
always available. Some websites do not have a rating
mechanism and thus their users cannot leave any rating
feedback on the products. This situation requires evaluating
implicit information which results in a lower prediction
accuracy of the recommender systems. The datasets from
the retargeting company2 are such a kind of data that contain
I.
INTRODUCTION
1
The term product and item, customer and user will be used
interchangeably.
2
We are not allowed to disclose the companys name.
410
III.
A. Notational Conventions
In this paper, we use a matrix to store the clicking
relationships between users and items, called the user-item
clicking matrix, denoted by R = [rij ] i 1,..., m where m is the
j 1,..., n
RELATED WORKS
hu
411
ii)
where xik'
Ju
2
ij
(1 - J ) u
j S (i;u )
npi
N
(1)
and so R P QT .
If we use ru to denote the u-th row of the rating matrix
R, then the user factor vector pu can be obtained via
(6)
pu ru Q
From (3) and (6), we have
rui ru Q qiT
(7)
Essentially, the matrix Q is the only one in decomposed
matrices that will be used in prediction.
The first tier is the similarity score and the second tier is the
popularity score. S(i;u) is the set of items that were viewed by
user u and are similar to item i. ij is the Pearson correlation
coefficient [10] between item i and item j. The popularity score is
the ratio between the view count of item i (denoted by npi) and the
global maximum view count (denoted by N). We use [0, 1] to
control the weight of each part.
U ij
k 1
d
k 1
'
( xik
xi )( x 'jk x j )
'
( xik
xi ) 2
d
k 1
1
) is a variation of xik (nck is
1 log nck
rui
xik (1
(2)
( x 'jk x j ) 2
412
TABLE I.
Item set
t1
ti
tj
rk
User set
u1
uk
um
(8)
| t j ))
k 1
rki
n
rkj
j 1
where P(t i | u k )
rkj
m
rkj
k 1
P(u k | t j )
( p
ij
Tu (t j ))
(9)
j 1
Tu (t j )
(10)
rki
n
rkj
j 1
IV.
P(u k | t j )
rkj
D
m r
kj
k 1
# of users
# of items
# of clicks
3699
20,471
499
134,982
5202
148,409
1,004
300,757
8631
112,738
94
1,559,529
9093
70,049
2,303
120,836
B. Evaluation Strategy
As stated before, our datasets are different from Netflix
datasets; therefore, we do not evaluate by the prediction
error (e.g. RMSE, MAE) but by the hit rate (also referred to
as recall rate, precision).
To compare the prediction accuracies, we apply the
models described above on four datasets to get top-N
recommendations with N = 10 for users in test user set.
We call the recommended item set the predicted set. As
evaluation criterion we adopted the hit rate of
recommendations (the higher, the better), i.e.
rui
Site ID
(P(t | u ) P(u
P(ti | t j )
STATISTICS ON DATASETS
tn
HitRate
The hit rates of all models are also tested for different
Ns on these datasets.
(11)
EXPERIMENT STUDY
A. Description of Datasets
The data that we used is gathered by a retargeting
company for research use. It consists of the browsing
history from 139 online shopping sites in one week
(08/08/2010 08/14/2010). In this dataset, each row
represents a transaction, which has four attributes, namely
product ID, website ID, user ID, and date.
We select four sub datasets from the whole data as our
testing datasets. Statistics are shown in Table I.
413
TABLE II.
Hit Rate
0.5
20.0%
21.2%
14.2%
Within the item popularity-based model, an itempopularity list is constructed by collecting statistics on the
browsing history. The filtering step is applied on this list to
obtain the final recommended items.
In the item similarity-based model, the parameter is
tweaked to get the best ratio of similarity-based score and
popularity-based score. is chosen from the interval of [0,
1] with step size 0.1.
Considering the filtered item similarity-based model,
we perform the filtering step of the item popularity-based
model on the ordered top-(2N) recommendation list
generated by the item similarity-based model to produce a
new top-N list. The same step will be applied to the other
models.
With bipartite graph model, we first build a probability
transition matrix with (8) and (11). The prediction is
computed with the Markov chain in the matrix.
Hit Rate
Model
414
Hit Rate
Hit Rate
Model
Model
Nevertheless, the filtering step has no effect on SVDbased model. It can be inferred that the latent factors in
SVD not only capture the users clicking count information
but the clicking patterns, i.e. the browsing habit. So no
filtering step is needed.
For website with site ID 5202, the results shown in
Fig.4 are quite different from those on 3699. The filtered
models (except f-SVD) perform better than the others. IS
and BG, f-IS and f-BG have a very similar hit rate,
respectively. IP got the worst prediction accuracy once
again. However, SVD (with 70 factors), the champion of
the previous experiment, only got a hit rate of 18.6%. The
latent factors seem not to capture the correlation between
users and items very well. Thus, latent factors do not work
for all cases. The f-SVD again has no improvement on
SVD.
Fig. 5 shows the results on the website with site ID
9093. The filtered bipartite graph model performs best.
SVD (with 100 factors) has a similar hit rate on site 5202.
The results show that in some datasets, the local
relationship among items (obtained by BG and IS-like
models) plays a more crucial role in predicting the next
item while in some other datasets, capturing the global
effects (obtained by SVD-like latent factor models) is more
important.
Hit Rate
TABLE III.
PERFORMANCE LIST
Performance Rank
Model
415
Dataset
3699
IP
6
IS
5
f-IS
3
BG
4
f-BG
2
SVD
5202
8631
9093
1
6
1
4
TOTAL
19 (6)
18 (5)
11 (2)
15 (4)
8 (1)
13 (3)
1
5
Hit Rate
Hit Rate
Hit Rate
Hit Rate
V.
6
Since SVD and f-SVD have same hit rate, we use SVD to stand for both
SVD and f-SVD in this context.
416
[5]
[6]
[7]
[8]
[9]
REFERENCES
[1]
[2]
[3]
[21] J. Bennet, and S. Lanning, The Netflix Prize, KDD Cup and
Workshop, 2007. www.netflixprize.com
[4]
417