Closed ndey96 closed 6 years ago
Chapter 8: Evaluating Recommender Systems 3 types of recommender systems evaluations. Ratings, Usage and Ranking
Rating Prediction: Predict what user will rate. Metrics: RMSE, MSE, weighted MSE, etc. All follow some sort of cost function that is like Sum(predicted_rating-actual_rating)
Usage Prediction Assumption that if it has not been consumed, that then if recommended, would not be consumed. Take a set of actions, hide a portion, and predict if user would interact with it. This creates the Precision-Recall tradeoff and ROC curves to analyze. Use F-measure to balance precision and recall.
Ranking Prediction Predict the order to put recommendations for the user. 2 methods for evaluating Reference Based: Establish reference list of ranking and compute ‘accuracy’ using things like Spearman’s or Kendall (look em up if you want). Establishing reference list is tricky since it must be inferred. Can do funky set theory things. Utility Based: Assign some power law weighting of recommendations and maximize utility function. Easier to understand.
Chapter 13: Music Recommendations
Unique Because:
Chapter 26: Novelty and Diversity in Recommendations
Metrics for Diversity
Proposed Definition of Diversity and Novelty:
Novelty: The unexpectedness of a single content piece (ex: Global Tail Novelty is evaluating novelty) Diversity: The intradifferences and interdifferences between sets of recommended content (ex: Average Intra List Diversity and Interrecommendation diversity are evaluating diversity).
Closing since @MatthewMcLeod read this
Diversity and Novelty Enhancements
Approaches can be summarized by two approaches.
There is usually trade-off between diversity and accuracy. One can do reorder where only items evaluated as the same "chance" will be reordered to improve diversity. This preserves accuracy but is not as flexible for increasing diversity.
Clustering Method Cluster user actions in a number of clusters, and then rather than make recommendations off of entire user history, perform recommendation on each cluster.
Fusion Based Methods Aggregate different recommendation systems outputs. Aggregation is hopefully more diverse.
In user studies, can trick the users into thinking recommendations are more diverse by reordering the set that they see.
Read the relevant sections of the recommender systems handbook