TY - JOUR AU - Zesch, Torsten AU - Gurevych, Iryna T1 - Wisdom of crowds versus wisdom of linguists - measuring the semantic relatedness of words. JO - Natural Language Engineering PY - 2010/ VL - 16 IS - 1 SP - 25 EP - 59 UR - http://dblp.uni-trier.de/db/journals/nle/nle16.html#ZeschG10 M3 - KW - datasets KW - kallimachos KW - measure KW - posts KW - relatedness KW - semantic L1 - SN - N1 - N1 - AB - ER - TY - CONF AU - Du, Lan AU - Buntine, Wray Lindsay AU - Jin, Huidong A2 - Webb, Geoffrey I. A2 - 0001, Bing Liu A2 - Zhang, Chengqi A2 - Gunopulos, Dimitrios A2 - Wu, Xindong T1 - Sequential Latent Dirichlet Allocation: Discover Underlying Topic Structures within a Document. T2 - ICDM PB - IEEE Computer Society CY - PY - 2010/ M2 - VL - IS - SP - 148 EP - 157 UR - http://dblp.uni-trier.de/db/conf/icdm/icdm2010.html#DuBJ10 M3 - KW - genre KW - kallimachos KW - plot KW - toread L1 - SN - 978-0-7695-4256-0 N1 - N1 - AB - ER - TY - CONF AU - Levy, Omer AU - Goldberg, Yoav A2 - Morante, Roser A2 - tau Yih, Wen T1 - Linguistic Regularities in Sparse and Explicit Word Representations. T2 - CoNLL PB - ACL CY - PY - 2014/ M2 - VL - IS - SP - 171 EP - 180 UR - http://dblp.uni-trier.de/db/conf/conll/conll2014.html#LevyG14 M3 - KW - kallimachos KW - posts KW - representation KW - similarity KW - toread KW - word L1 - SN - 978-1-941643-02-0 N1 - N1 - AB - ER - TY - GEN AU - Yu, Hsiang-Fu AU - Jain, Prateek AU - Kar, Purushottam AU - Dhillon, Inderjit S. A2 - T1 - Large-scale Multi-label Learning with Missing Labels JO - PB - AD - PY - 2013/ VL - IS - SP - EP - UR - http://arxiv.org/abs/1307.5101 M3 - KW - classification KW - kallimachos KW - label KW - large KW - learning KW - multi L1 - N1 - Large-scale Multi-label Learning with Missing Labels N1 - AB - The multi-label classification problem has generated significant interest in

recent years. However, existing approaches do not adequately address two key

challenges: (a) the ability to tackle problems with a large number (say

millions) of labels, and (b) the ability to handle data with missing labels. In

this paper, we directly address both these problems by studying the multi-label

problem in a generic empirical risk minimization (ERM) framework. Our

framework, despite being simple, is surprisingly able to encompass several

recent label-compression based methods which can be derived as special cases of

our method. To optimize the ERM problem, we develop techniques that exploit the

structure of specific loss functions - such as the squared loss function - to

offer efficient algorithms. We further show that our learning framework admits

formal excess risk bounds even in the presence of missing labels. Our risk

bounds are tight and demonstrate better generalization performance for low-rank

promoting trace-norm regularization when compared to (rank insensitive)

Frobenius norm regularization. Finally, we present extensive empirical results

on a variety of benchmark datasets and show that our methods perform

significantly better than existing label compression based methods and can

scale up to very large datasets such as the Wikipedia dataset. ER - TY - CONF AU - Mirowski, Piotr AU - Ranzato, Marc'Aurelio AU - LeCun, Yann A2 - of the NIPS 2010 Workshop on Deep Learning, Proceedings T1 - Dynamic Auto-Encoders for Semantic Indexing T2 - PB - CY - PY - 2010/ M2 - VL - IS - SP - EP - UR - http://yann.lecun.com/exdb/publis/pdf/mirowski-nipsdl-10.pdf M3 - KW - deep KW - kallimachos KW - lda KW - learning KW - model KW - toread L1 - SN - N1 - Neuer Tab N1 - AB - ER - TY - GEN AU - Karampatziakis, Nikos AU - Mineiro, Paul A2 - T1 - Discriminative Features via Generalized Eigenvectors JO - PB - AD - PY - 2013/ VL - IS - SP - EP - UR - http://arxiv.org/abs/1310.1934 M3 - KW - analysis KW - eigenvector KW - feature KW - kallimachos L1 - N1 - Discriminative Features via Generalized Eigenvectors N1 - AB - Representing examples in a way that is compatible with the underlying

classifier can greatly enhance the performance of a learning system. In this

paper we investigate scalable techniques for inducing discriminative features

by taking advantage of simple second order structure in the data. We focus on

multiclass classification and show that features extracted from the generalized

eigenvectors of the class conditional second moments lead to classifiers with

excellent empirical performance. Moreover, these features have attractive

theoretical properties, such as inducing representations that are invariant to

linear transformations of the input. We evaluate classifiers built from these

features on three different tasks, obtaining state of the art results. ER - TY - CONF AU - Kohlschütter, Christian AU - Fankhauser, Peter AU - Nejdl, Wolfgang A2 - T1 - Boilerplate Detection using Shallow Text Features T2 - Proc. of 3rd ACM International Conference on Web Search and Data Mining New York City, NY USA (WSDM 2010). PB - CY - PY - 2010/ M2 - VL - IS - SP - EP - UR - M3 - KW - features KW - kallimachos KW - text KW - toread L1 - SN - N1 - N1 - AB - ER -