Serendipitous Recommendations via Innovators Noriaki Kawamae SIGIR ’10 Summarized and presented by Sang-il Song Page 1 Contents • • Introduction Related Work – Novelty Metric – Novelty vs. Serendipity • Key Idea – New serendipity measure based on time – Innovative consumers – User Flow • Detail Model – Personal Innovator Probability (PIP) – User Flow Probability (UFP) – Combination of PIP and UFP • Experiments – Datasets – Comparison of top-N precision – Comparison of other metrics • Discussion • Conclusion Page 2 INTRODUCTION • Traditional Collaborative Filtering – Focuses on optimizing accuracy measurements such as precision and recall – Neglects other factors such as novelty and serendipity • Popularity Bias – The most popular items in any log collection are the ones that a given user will recognize with high probability, or be broadly familiar with. – Highly accurate recommendations appear far too obvious and of little help to users, and seem to fail to enhance user satisfaction – Recommendation systems that appropriately discount popularity lead to an increase in the total sales • Novel and serendipitous recommendations – Broaden user’s view in the recommendation flows – Focus on the surprise of each user with the recommendation Page 3 Related Work • Novelty and serendipity metrics – Measuring the Degree of “non-obviousness” of recommendations • Interest-based Recommendation (Yang et al., 2005) – Define novelty in terms of user knowledge and his/her degree of interest model • Topic diversification approach (Ziegler et al., 2005) – Topical diversification approach to balance and diversify personalized recommendat ion lists • Random Walk based Techniques (Focuss et al. 2007) – Computing similarity of nodes solving sparsity problem of data • Item centric evaluation for recommendation (Celma et al. 2008) – User-centric evaluation aims to measure the perceived quality of the recommendati ons • These works don’t consider “dynamics” – The changes in user preferences over time – New items are different from old items in terms of their serendipity Page 4 Novelty vs. Serendipity • Distinction between novelty and serendipity • Example – There is a user whose favorite director “James Cameron” – If the system recommends his latest movie “Avatar (2009)” • Certainly novel but probably not serendipitous – If the systems shows “Princess Mononoke (1997)”, which has them es similar to “Avatar” • Non-obvious recommendation • Providing serendipity to the user • Take much longer for find it by themselves in the absence of any help Page 5 Key Idea 1 - How to quantify serendipity? • “Time” can be used to quantify serendipity? – The difference between the time at which target user would purchase ite ms and the time at which user these items are recommended to user – In the below figure • | tˆa3 - t0 | indicates a longer acquisition delay than | tˆa2 - t0 | • i3 seems to offer more serendipity to ua than i2 Page 6 Key Idea 2 - “Innovative” consumers • “Innovative” Consumers – The early consumers who focus on undiscovered or unreleased items are called “innovative” – Innovators become aware of items well before their release – The purchase logs of innovators include more serendipitous items that oth er like-minded consumers would like to acquire but have not yet become aware of • Innovators are able to discover even poorly marketed items – These items, once “discovered”, are rapidly acquired by like-minded follow ers • Innovative consumer is captured into recommendation model u sing “Personal Innovator Probability (PIP)” Page 7 Key Idea 3 - User Flow • Traditional algorithms based on item-item similarity compute the simil arity scores among items using the number of consumers who purcha sed both items – Assumption: “Those who agreed in terms of past behavior will tend to agree in the future” • In the below figure – Each item of set i1, i2, i3 has been purchased by same users in common with ia – Each item is treaded equally according to the similarity defined for ia in traditional system – There needs to differentiate these items by focusing on the order of item purchase time – i3 is more appropriate item for a user who has purchased ia than i1 or i2 • This property is modeled as User Flow Probability (UFP) Page 8 Personal Innovator Degree • Personal Innovator Degree (PID) – Weight user logs appropriately given a target user and to identify logs tha t match the precedence preference of the target user – Two time factors and one item factor PID(ua , ub ) = å r(i;b, a) ´ w(i;a)´ v(i) iÎCab Cab r(i;b, a) w(i;a) v(i) Innovator Recency Discriminativity The set of common items that both ua and ub ha ve purchased The degree of ub as a personal innovator to ua for common item i How recently ua purchased common item i How important common item i is as a discrimina tor of like-minded users Page 9 Personal Innovator Degree Detail • Degree of innovator – A – Smaller tb,i and bigger ta,i make Degree of innovator bigger • Recency – ea,i w(i;a) = exp(- ) ea Degree of Innovator B가 A보다 아이템을 먼저 구매했을 경우 ➜ 커진다 B보다 A가 먼저 아이템을 구매했을 경우 ➜ 0 – Smaller elapsed time passed since ua purchased j makes it bigger • Degree of discriminator Recency 아이템이 나오자마자 구매한 경우 ➜ 1 아이템이 나온지 한참후에 구매한 경우 ➜ 0에 가까움 1 log(1+Ui ) – v(i) = – Unpopular item is important discriminator Degree of discriminator Unpopular한 아이템 ➜ 유저의 특성 파악에 유용한 정보 ➜ 큰 값 Popular한 아이템 ➜ 유저의 특성 파악에 의미 없는 정보 ➜ 0에 가까움 Page 10 Personal Innovator Probability • Personal Innovator Probability – Improved version of Personal innovator Degree – Solve the sparsity problem of data • The ratio of users who share common items decreases rapidly in inverse propo rtion to the number of times • Innovator probability – Using Markov Chain (Random Walk) – Nodes are corresponding to user and edge is p(b | a) = PID(ub, ua ) å PID(u, ua ) u – Normalized PID over all users – eeT P = a P +(1- a ) U – Ṗ is ergodic Markov chain – P = e- b ((b P) + 1 (b P)2 + 2! 모두 유저간 edge에 대하여 (1-a)/|U|의 weight를 더해줌으 로써 Markov Chain Process가 수렴함을 보장 Ergodic Markov Chain 어떤 state에서 다른 어떠한 state로 이동이 가능 Random Walk 시 수렴 보장 + 1 (b P)N-1 + ) (N -1)! N – Weighted Hitting Time (weight: b ) N! Page 11 Continuous Time Markov Chain • Markov Chain – random process endowed with the Markov property – the next state depends only on the current state and not on the past • Continuous-Time Markov Chain – Continuous parameter space (cf. discrete parameter space) – Characterized by transition rate qij between state i and j • How quickly that i➝j transition happens Pr(i ® j) = qij , i¹ j – Jump Rate: The rate that stochastic process leaves item i qi = å qij j¹i Page 12 Distribution of the number of users who rated a DVD at a specific age of the DVD X-axis: the number of users Y-axis: the age of movies Distribution of the number of users who rated a DVD at a specific age of the DVD in Netflix • The fitting lines is exponential functions of time • The number of user who evaluated an item decreases exponentially as the item ages Page 13 User Flow Probability • The transition probability – The probability that transition from item i to item j occurs pij = å qi exp(-qi t u,ij ) t=0 일 때, pij = qi t=∞ 일 때, pij = 0 t=0~∞사이에서 transition이 일어날 확률: 1 User들이 아이템 i를 선택한지 얼마 안되서 j를 선택한 경우, pij가 커진다 u • User Flow Probability – Representing relationships of items – Continuous-time Markov Chain – UFP is computed using Taylor approximation U(t f ) = tf n (tQ) ò ån=0 n! dt 0 ¥ Continuous-time Markov Chain 모델 사용 (PIP와 유사) 단, hitting time이 시간에 대한 함수 특정 시점 tf에서의 trend는 0~tf까지 hitting 의 적분값 UFP(ib,ia ) = lim u(ib | ia , t f ) t f ®¥ UFP는 특정 시점이 아닌 모든 시간의 평균적인 trend 를 의미 Page 14 Recommendation based on both PIP and UFP • Personal Innovator Probability – the relationship between users – The probability of innovator of user • User Flow Probability – Presenting relationship between items • Combination Model p(ib | ua ) µ åUFP(ib ,ia )d (ia | u j )PIP(u j , ua ) j – δ(i,u) represents the evidence that u has purchased item i • 1 if I is in the purchased history • 0 otherwise • Computational Cost User kNN의 관점에서 보면 1. Similarity measure로 PIP 사용 2. Item rating의 aggregation에서 weight를 UFP 줌 – PIP, UFP can be pre-computed on all user-user pair and item-item pair – The recommendation algorithm is computed in the linear time Page 15 Experiment - Dataset • Dataset – purchase log data of on-line music download services • 2005. 04. 01 ~ 2006. 07. 31 • #user: 84620 • #item: 44527 – Purchase log data of on-line video download service • 2005. 09. 01 ~ 2006. 2. 28 • #user: 7537 • #item: 4064 – Netflix • • • • • 1999. 11. 11 ~ 2005. 12 .31 #user: 480,189 #item: 17,770 #rating: 100,480,507 Binary rating (1 if purchase) – Query log of search engine • 2006. 04. 01 ~ 2006. 05. 31 • 33,325,842 records • <user id, query keyword id, timestamp> Page 16 Comparison of top-N precision Top-10 Precision 0.5 0.45 Popular 0.4 Cosine 0.35 Pearson Item 0.3 bPLSA 0.25 0.2 MEA 0.15 EABIF PID 0.1 PIP 0.05 PIP+UFP 0 Video Music NetFlix - Popular - Cosine User-based KNN with Cosine Similarity - Pearson User-based KNN with Pearson coefficient - Item Item-based KNN (2001) - bPLSA Probabilistic Latent Semantic Analysis wit h Bernoulli Distribution (2003) - MEA Maximum Entropy Approach (2002) - EABIF Early-Adoption-Based information Flow m ethod (2006) - PID Personal Innovator Degree (2009) - PIP+UFP Query • PIP+UFP and PID shows higher precision than other recommendation algorithms – – – – Video: 0.4612 and 0.4602 Music: 0.3987 and 0.3922 Netflix: 0.1804 and 0.1793 Query: 0.1523 and 0.1588 Page 17 Comparison with Other Metric (1) • Item Coverage (IC) – The ratio of the number of unique items appearing in the top-N recomme ndation list from the total number of items – Measure of the item domain in the system from which the system can rec ommend • Average Elapsed Time (AE) – How much time has passed from each item’s release day to its day of pur chase by each user – A short elapsed time indicates that the item is novel • Average Difference Time (AD) – The Difference observed time when each user u purchased item I in the te st data without recommendation and the time of test Page 18 Comparison with Other Metric (2) • Gini Coefficient – A measure of the statistical dispersion of the distribution of users over items – defined as the ratio of the areas on the Lorenz curve diagram – Its value is between 0 and 1 • • • • High Gini coefficient indicates the distribution is extremely biased 0 corresponds to perfect equality 1 corresponds to perfect inequality High Gini coefficient means that a few particular items tend to be ranked highly by most u sers and thus recommendations are not strongly personalized Page 19 Comparison with Other Metric (3) Item Coverage Average Elapsed Time 0.4 Popular Cosine 0.3 50 Popular Cosine 40 Pearson 30 Item 0.2 bPLSA MEA 0.1 Pearson EABIF Item bPLSA 20 MEA EABIF 10 PID 0 Music Video NetFlix Query PID 0 PIP Music Average Distance Time 20 10 5 NetFlix Query 1 Popular Cosine 0.8 Cosine Pearson 0.6 Item Pearson bPLSA 0.4 bPLSA MEA MEA Item 0.2 EABIF 0 Music Video NetFlix Query PIP Gini Coefficient Popular 15 Video PID EABIF 0 Music Video NetFlix Query PID Page 20 Discussion • Effect of Innovator – PIP improves AD – Support that innovators know more serendipitous items than others • PID vs PIP – PID considers only pairwise comparison – PIP uses an Markov chain to model how innovators are followed through multiple steps – PIP will recommend surprisingly interesting items that might not otherwis e be discovered • Novel Item Recommendation – PIP shows shortest elapsed time – It is best at identifying novel items Page 21 Conclusion • Contribution – Impact of innovator of user (PIP) – Impact of trends on the transition probability of items (UFP) – New Metric for measuring serendipitousness • Estimated time offset between recommendation time and purchase ti me • Future Work – Incorporating user speciality and level of expertise in a particular domain – Extending model to personalized information retrieval more generally • It considers binary rating only (purchase or not) • There needs comparison to state-of-the-art recommendation algorith m Page 22