%0 Conference Paper %1 wu2009learning %A Wu, Lei %A Yang, Linjun %A Yu, Nenghai %A Hua, Xian S. %B WWW '09: Proceedings of the 18th international conference on World wide web %C New York, NY, USA %D 2009 %I ACM %K learning ol_web2.0 tagging widely_related %P 361--370 %T Learning to tag %U http://dx.doi.org/10.1145/1526709.1526758 %X Social tagging provides valuable and crucial information for large-scale web image retrieval. It is ontology-free and easy to obtain; however, irrelevant tags frequently appear, and users typically will not tag all semantic objects in the image, which is also called semantic loss. To avoid noises and compensate for the semantic loss, tag recommendation is proposed in literature. However, current recommendation simply ranks the related tags based on the single modality of tag co-occurrence on the whole dataset, which ignores other modalities, such as visual correlation. This paper proposes a multi-modality recommendation based on both tag and visual correlation, and formulates the tag recommendation as a learning problem. Each modality is used to generate a ranking feature, and Rankboost algorithm is applied to learn an optimal combination of these ranking features from different modalities. Experiments on Flickr data demonstrate the effectiveness of this learning-based multi-modality recommendation strategy.