Methods and Metrics for Cold-start Recommendations
A. Schein, A. Popescul, L. Ungar, und D. Pennock. Proceedings of the 25th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Seite 253--260. New York, NY, USA, ACM, (2002)
We have developed a method for recommending items that combines content and collaborative data under a single probabilistic framework. We benchmark our algorithm against a naïve Bayes classifier on the cold-start problem, where we wish to recommend items that no one in the community has yet rated. We systematically explore three testing methodologies using a publicly available data set, and explain how these methods apply to specific real-world applications. We advocate heuristic recommenders when benchmarking to give competent baseline performance. We introduce a new performance metric, the CROC curve, and demonstrate empirically that the various components of our testing strategy combine to obtain deeper understanding of the performance characteristics of recommender systems. Though the emphasis of our testing is on cold-start recommending, our methods for recommending and evaluation are general.
Methods and metrics for cold-start recommendations