Professional Documents
Culture Documents
+ MODEL
ARTICLE IN PRESS
Siemens AG, Corporate Research and Technology, Otto-Hahn-Ring 6, Geb. 31, D-81730 Mnchen, Germany
b
University of Maryland, College Park, A.V. Williams Building, College Park, MD 20740, USA
Received 6 September 2005; received in revised form 29 October 2006; accepted 3 November 2006
Abstract
Online communities that allow their users to express their personal preferences, such as the members they trust and the products
they appreciate, are becoming increasingly popular. Exploiting these communities as playgrounds for sociological research we
present two frameworks for analyzing the correlation between interpersonal trust and interest similarity. We obtain empirical results
from applying the two frameworks on two real, operational communities, that suggest there is a strong correlation between both
trust and interest similarity. We believe our findings particularly relevant for ongoing research in recommender systems and
collaborative filtering, where people are suggested products based on their similarity with other customers, and propose ways in
which trust models can be integrated into these systems.
2006 Elsevier B.V. All rights reserved.
Keywords: Trust; Similarity; Collaborative filtering; e-commerce; Distributed systems
In today's networked worlds, uncertainty and anonymity are important factors that bear strong implications
on decision-making. Several researchers have therefore
proposed to incorporate the concept of interpersonal trust
into electronic interactions [32], and proposed several
computational trust models [37,43,41]. These models
intend to predict which people are deemed trustworthy,
based on the people one given agent trusts, the people the
trusted agents trust, and so forth. Trust can be seen as an
adequate decision-support tool that dissects relevant and
reliable information sources from unreliable [28,26].
Corresponding author.
E-mail addresses: cai.ziegler@siemens.com (C.-N. Ziegler),
golbeck@cs.umd.edu (J. Golbeck).
1
Recommender systems, or collaborative filtering systems, make
product recommendations based upon the opinions of people bearing
high similarity with the user for whom to make recommendations.
1. Introduction
0167-9236/$ - see front matter 2006 Elsevier B.V. All rights reserved.
doi:10.1016/j.dss.2006.11.003
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
2
prevail. Complementing the evaluation framework, Section 5 presents an experimental study performed on
FilmTrust, a community catering to movie lovers. Suggestions for the exploitation of correlation between trust
and similarity are provided in Section 6, while Section 7
mentions open questions and possible future work.
2. Trust and recommender systems
Online recommender systems [49,33] intend to provide people with recommendations of products they
might appreciate, taking into account their past ratings
profile, purchase history, or interest. There are three main
types of filtering systems [19]: collaborative, contentbased, and economic filtering. While content-based filtering takes into account properties attributed to the
nature of products themselves, collaborative filtering
(CF) relies upon building neighborhoods of like-minded
customers [53] whose rating history may then serve to
generate new recommendations. Economic filtering,
which filters items based on cost factors, has seen little
practical application and exerts marginal impact only.
Sinha and Swearingen [57] have found that people
prefer receiving recommendations from people they
know and trust, i.e., friends and family-members, rather
than from recommender systems. As a result, some
researchers have started to focus on computational trust
models as appropriate means to supplement or replace
current collaborative filtering approaches.
Kautz et al. [29] mine social network structures in
order to render expertise information exchange and
collaboration feasible. Olsson [47] proposes an architecture combining trust, collaborative filtering and content-based filtering in one single framework, giving only
vague information and insight, though. Another agentbased approach has been presented by Montaner et al.
[42], who introduce opinion-based filtering. Montaner
claims that trust should be derived from user similarity,
implying that friends are exactly those people that resemble our very nature. However, Montaner's model
only extends to the agent world and does not reflect
evidence acquired from real-world social studies concerning the formation of trust. Similar agent-based systems have been devised by Kinateder and Rothermel
[31], Kinateder and Pearson [30], and Chen and Singh
[11].
Apart from research in multi-agent systems, online
communities also make use of trust networks. The wellknown reviewers' community Epinions (http://www.
epinions.com) provides information filtering facilities
based upon personalized webs of trust [20]. Guha, chief
architect of Epinions, states that the trust-based filtering
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
C.-N. Ziegler, J. Golbeck / Decision Support Systems xx (2006) xxxxxx
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
4
We intend to establish a formal framework for investigating interactions between trust and similarity,
believing that given an application domain, such as, for
instance, the book-reading domain, people's trusted
peers are on average considerably more similar to their
sources of trust than arbitrary peers. More formally, let A
denote the set of all community members, trust(ai) the set
of all users trusted by ai, and sim: A A [ 1,+1] some
similarity function:
P
P
X aj atrustai simai ; aj X aj aAqfai g simai ; aj
NN
jtrustai j
jAj 1
a aA
a aA
i
1
For instance, given that agent ai is interested in
Science-Fiction and Artificial Intelligence, the chances
that aj, trusted by ai, also likes these two topics are
much higher than for peer ae not explicitly trusted by ai.
Various social processes are involved, such as participation in those social groups that best reflect our own
interests and desires.
4.1. Model and data acquisition
In order to verify or refute our hypothesis for specific
domains, we need to define an information model, determine metrics and methods for evaluation, and apply
our framework to real-world data.
4.1.1. Information model
In our information model proposed, we assume
relationships of binary preference, i.e., for every ai A,
we divide the set of agents A into trusted and non-trusted
ones. Moreover, ratings are assumed implicit and nonquantifiable with respect to the degree of appreciation.
Though making the data more imprecise, the model
becomes more general and bindings with real-world
data, such as the All Consuming dataset, are facilitated.
(a) Set of agents A = {a1, a2,, an}. A contains all
agents part of the community, identified uniquely
through URLs etc.
(b) Set of products B = {b1, b2,, bm}. All domainrelevant products are stored in set B. Unique
identifiers either refer to proprietary product codes
from an online store, such as Amazon.com's
ASINs, or represent globally accepted standard
codes, like ISBNs.
(c) Trust rating function trust: A 2A . For every
ai A, trust(ai) returns the set of agents Ai A
that ai trusts.
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
C.-N. Ziegler, J. Golbeck / Decision Support Systems xx (2006) xxxxxx
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
6
jDj
X
vi;k s
k1
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
C.-N. Ziegler, J. Golbeck / Decision Support Systems xx (2006) xxxxxx
q
X
sco pm scdke
scdke
;
gq
where
g0 : 1; g1 : 1
m0
sco pm1
sib pm1 1
1
;
sib pq 1
and n {2,,q}
gn : gn1 gn1 gn2 d
1
sib pqn1 1
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
8
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
C.-N. Ziegler, J. Golbeck / Decision Support Systems xx (2006) xxxxxx
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
10
4.3.2.1. Histogram curves. The bell-shaped distribution of s, depicted in Fig. 3(B), looks more condensed
with respect to s V and has its peak slightly below the
latter plot's curve. The differences between z and z V are
even more pronounced, though, e.g., the shape of z's
histogram looks more regular than z V's pendant. The
approximation of z's distribution, applying polynomial
regression of degree 5, strongly resembles the Erlang-k
distribution, supposing k = 2. For similarity degrees
above 0.35, peaks of z's histogram are considerably
less explicit than for z or have effectively disappeared,
as is the case for degrees above 0.6.
4.3.2.2. Matching ziW and siW. Fig. 4(B) gives the
scatter plot of our lower bound analysis. The strong bias
towards the upper region has become less articulate,
though still clearly visible. Interestingly, the increase of
ratio ziW: siW for siW N 0.15 still persists.
4.4. Statistical significance
Fig. 4. Scatter plot for the upper bound (A) and lower bound
(B) analysis.
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
C.-N. Ziegler, J. Golbeck / Decision Support Systems xx (2006) xxxxxx
Table 1
KruskalWallis ANOVA test results for the upper bound experiment
zV
sV
Rank sum
Mean rank
261
261
73702.0
60719.0
284.56
234.44
11
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
12
information. These disconnected nodes are not considered in our analysis. That leaves 431 remaining members. Because relationships are directional, that is, a
person can list someone as a friend without that person
listing them back, there are different in-and outdegrees
for each node. The average outdegree (people the user
has listed as a friend), is 3.72, while the average indegree (people who listed a user as a friend) is 3.01.
Thus, looking at all edges, the average degree is 6.73.
The average trust rating is 6.8 on a scale from 1 to 10,
with a standard deviation of 2.23. The values are normally distributed with the exception of a spike for trust
values of 10, see Fig. 6(A).
The other relevant feature of the Web site is a movie
rating and review system. Users can choose any film and
rate it on a scale of one half star to four stars. They can
also write free-text reviews about movies. The user's
Movies page displays data for every movie that he has
rated or reviewed.
In FilmTrust, 1486 unique movies have been rated or
reviewed by members. There are a total of 4833 reviews
for 387 different movies. There are 13,997 ratings for
1479 different movies. This averages to 9.46 ratings per
film. However, this distribution is skewed. 1437 movies
have fewer than 50 ratings with about half (790) given
only one rating. The average number of ratings given to
this group of 1437 films is 2.66. On the other hand, fifty
films have over 80 ratings, with an average of 202
ratings for each film in that group. This distribution is
due to the fact that during the registration process, users
are presented with and asked to rate the top fifty films
from the AFI Top 100 Movies of All Time list. Thus, the
movies on this list are much more likely to be rated. Of
the 672 members, 563 have rated or reviewed films. On
average, each person has rated or reviewed 24.9 films.
These explicit ratings of trust values and movies
facilitate an analysis of the correlation between trust and
user similarity.
5.2. Profile similarity computation
When users join the FilmTrust network, they are
presented with a list of the top 50 films on the American
Film Institute's top 100 movies list6 and asked to assign
a star rating to any movies they have seen. Thus, there is
a core set of films with ratings from many users that
permit a better and more facile analysis.
Let ti (aj) represent the trust rating that user ai A
has assigned to user aj and rj (bk) represent the rating
6
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
C.-N. Ziegler, J. Golbeck / Decision Support Systems xx (2006) xxxxxx
Table 2
KruskalWallis ANOVA test results for FilmTrust data
Trust value
Rank sum
Mean rank
1
2
3
4
5
6
7
8
9
10
158
203
195
296
1235
1043
1217
1142
636
898
636,734.5
820,833.5
752,781.0
1,129,921.0
4,681,499.5
3,874,654.5
4,233,673.5
3,771,554.0
2,049,333.5
2,713,791.0
4029.97
4043.51
3860.42
3817.30
3790.69
3714.91
3478.78
3302.59
3222.22
3022.04
13
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
14
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
C.-N. Ziegler, J. Golbeck / Decision Support Systems xx (2006) xxxxxx
[5] Thomas Beth, Malte Borcherding, Birgit Klein. Valuation of trust in
open networks, Proceedings of the 1994 European Symposium on
Research in Computer Security, 1994, pp. 318, Brighton, UK.
[6] John Breese, David Heckerman, Carl Kadie, Empirical analysis
of predictive algorithms for collaborative filtering, Proceedings
of the Fourteenth Annual Conference on Uncertainty in Artificial
Intelligence, Morgan Kaufmann, Madison, WI, USA, July 1998,
pp. 4352.
[7] Alexander Budanitsky, Graeme Hirst, Semantic distance in
WordNet: an experimental, application-oriented evaluation of
five measures, Proceedings of the Workshop on WordNet and
Other Lexical Resources, Pittsburgh, PA, USA, June 2000.
[8] Ernest Burgess, Paul Wallin, Homogamy in social characteristics,
American Journal of Sociology 2 (49) (1943) 109124.
[9] Donn Byrne, Interpersonal attraction and attitude similarity,
Journal of Abnormal and Social Psychology 62 (1961) 713715.
[10] Donn Byrne, The Attraction Paradigm, Academic Press, New
York, NY, USA, 1971.
[11] Mao Chen, Jaswinder Singh, Computing and using reputations
for internet ratings, Proceedings of the 3rd ACM Conference
on Electronic Commerce, ACM Press, Tampa, FL, USA, 2001,
pp. 154162.
[12] Edd Dumbill, Finding friends with XML and RDF, IBM's XML
Watch, June 2002.
[13] Sabine Einwiller, The significance of reputation and brand in
creating trust between an online vendor and its customers, in:
Otto Petrovic, Markus Fallenbck, Christian Kittl (Eds.), Trust in
the Network Economy, Springer-Verlag, Heidelberg, Germany,
ISBN: 3-211-06853-8, 2003, pp. 113127.
[14] Andrew Fernandes, Risking trust in a public key infrastructure:
old techniques of managing risk applied to new technology,
Decision Support Systems 31 (3) (2001) 303322.
[15] Barbara Given, Teaching to the brain's natural learning systems,
Association for Supervision and Curriculum Development,
Alexandria, VA, USA, May 2002.
[16] Jennifer Golbeck, Computing and Applying Trust in Web-based
Social Networks, PhD thesis, University of Maryland, College
Park, MD, USA, April 2005.
[17] Jennifer Golbeck, James Hendler, Accuracy of metrics for
inferring trust and reputation in Semantic Web-based social
networks, Proceedings of the 14th International Conference on
Knowledge Engineering and Knowledge Management, Northamptonshire, UK, October 2004.
[18] Jennifer Golbeck, Bijan Parsia, James Hendler, Trust networks
on the Semantic Web, Proceedings of Cooperative Intelligent
Agents, Helsinki, Finland, August 2003.
[19] David Goldberg, David Nichols, Brian Oki, Douglas Terry, Using
collaborative filtering to weave an information tapestry, Communications of the ACM, ISSN: 0001-0782 35 (12) (1992) 6170.
[20] Ramanathan Guha, Open rating systems, Technical Report,
Stanford Knowledge Systems Laboratory, Stanford, CA, USA,
2003.
[21] Ramanthan Guha, Ravi Kumar, Prabhakar Raghavan, Andrew
Tomkins, Propagation of trust and distrust, Proceedings of the
Thirteenth International World Wide Web Conference, ACM
Press, New York, NY, USA, May 2004.
[22] Fritz Heider, The Psychology of Interpersonal Relations, Wiley,
New York, NY, USA, 1958.
[23] Jonathan Herlocker, Joseph Konstan, Al Borchers, John Riedl,
An algorithmic framework for performing collaborative filtering,
Proceedings of the 22nd Annual International ACM SIGIR
Conference on Research and Development in Information
[24]
[25]
[26]
[27]
[28]
[29]
[30]
[31]
[32]
[33]
[34]
[35]
[36]
[37]
[38]
[39]
[40]
15
Retrieval, ACM Press, Berkeley, CA, USA, ISBN: 1-58113096-1, 1999, pp. 230237.
Ted Huston, George Levinger, Interpersonal attraction and
relationships, Annual Review of Psychology 29 (1978) 115156.
Carlos Jensen, John Davis, Shelly Farnham, Finding others
online: reputation systems for social online spaces, Proceedings
of the SIGCHI Conference on Human Factors in Computing
Systems, ACM Press, Minneapolis, MN, USA, ISBN: 1-58113453-3, 2002, pp. 447454.
Andrew Jones, On the concept of trust, Decision Support
Systems 33 (3) (2002) 225232.
Edward Jones, Linda Bell, Elliot Aronson, The reciprocation of
attraction from similar and dissimilar others, in: Charles
McClintock (Ed.), Experimental Social Psychology, Holt,
Rinehart, and Winston, New York, NY, USA, 1972.
Audun Jsang, Rolan Ismail, Colin Boyd, Survey of trust and
reputation systems for online service provision, Decision Support
Systems (2005) (To appear).
Henry Kautz, Bart Selman, Mehul Shah, Referral Web:
combining social networks and collaborative filtering, Communications of the ACM, ISSN: 0001-0782 40 (3) (March 1997)
6365.
Michael Kinateder, Siani Pearson, A privacy-enhanced peer-topeer reputation system, Proceedings of the 4th International
Conference on Electronic Commerce and Web Technologies
LNCS, vol. 2378, Springer-Verlag, Prague, Czech Republic,
September 2003.
Michael Kinateder, Kurt Rothermel, Architecture and algorithms
for a distributed reputation system, in: Paddy Nixon, Sotirios
Terzis (Eds.), Proceedings of the First International Conference
on Trust Management LNCS, vol. 2692, Springer-Verlag, Crete,
Greece, April 2003, pp. 116.
Mathias Klang, Who do you trust? Beyond encryption, secure ebusiness, Decision Support Systems 31 (3) (2001) 293301.
Joseph Konstan, Introduction to recommender systems: algorithms and evaluation, ACM Transactions on Information
Systems, ISSN: 1046-8188 22 (1) (2004) 14.
Raph Levien, Attack-resistant Trust Metrics, PhD thesis,
University of California at Berkeley, Berkeley, CA, USA,
2004. To appear.
Raph Levien, Alexander Aiken, Attack-resistant trust metrics for
public key certification, Proceedings of the 7th USENIX Security
Symposium, San Antonio, TX, USA, January 1998.
Stephen Marsh, Optimism and pessimism in trust, in: Jose Ramirez
(Ed.), Proceedings of the IberoAmerican Conference on Artificial
Intelligence, McGraw-Hill, Caracas, Venezuela, 1994.
Stephen Marsh, Formalising Trust as a Computational Concept,
PhD thesis, Department of Mathematics and Computer Science,
University of Stirling, Stirling, UK, 1994.
Paolo Massa, Paolo Avesani, Trust-aware collaborative filtering
for recommender systems, in: Robert Meersman, Zahir Tari
(Eds.), Proceedings of the DOA/CoopIS/ODBASE Confederated
International Conferences (1), LNCS, vol. 3290, SpringerVerlag, Larnaca, Cyprus, ISBN: 3-540-23663-5, October 2004,
pp. 492508.
Paolo Massa, Bobby Bhattacharjee, Using trust in recommender systems: an experimental analysis, in: Christian Jensen,
Stefan Poslad, Theodosis Dimitrakos (Eds.), Proceedings of the
2nd International Conference on Trust Management, LNCS,
vol. 2995, Springer-Verlag, Oxford, UK, March 2004.
Ueli Maurer, Modelling a public key infrastructure, in: Elisa
Bertino (Ed.), Proceedings of the 1996 European Symposium on
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003
ARTICLE IN PRESS
16
[41]
[42]
[43]
[44]
[45]
[46]
[47]
[48]
[49]
[50]
[51]
[52]
[53]
[54]
[55]
[56]
[57]
Cai-Nicolas Ziegler is consultant and researcher in the domain of text mining and
natural language processing (NLP) at Siemens AG, Corporate Research and Technology (CT) in Munich, Germany. Before
joining Siemens AG, he has been a postdoctoral researcher at DBIS, the Databases
and Information Systems group of the University of Freiburg, Germany. Cai-Nicolas
received his PhD in Computer Science from
the University of Freiburg in 2005, and his Diploma (equivalent to
MSc) from the University of Passau.
Please cite this article as: Cai-Nicolas Ziegler, Jennifer Golbeck, Investigating interactions of trust and interest similarity, Decision Support
Systems (2006), doi:10.1016/j.dss.2006.11.003