Luo, C.; Li, Y. & Chung, S. M. (2009),
'Text document clustering based on neighbors', Data & Knowledge Engineering
68
(11)
, 1271 - 1288
.
[BibTeX]
[Endnote]
Clustering is a very powerful data mining technique for topic discovery from text documents. The partitional clustering algorithms, such as the family of k-means, are reported performing well on document clustering. They treat the clustering problem as an optimization process of grouping documents into k clusters so that a particular criterion function is minimized or maximized. Usually, the cosine function is used to measure the similarity between two documents in the criterion function, but it may not work well when the clusters are not well separated. To solve this problem, we applied the concepts of neighbors and link, introduced in [S. Guha, R. Rastogi, K. Shim, ROCK: a robust clustering algorithm for categorical attributes, Information Systems 25 (5) (2000) 345-366], to document clustering. If two documents are similar enough, they are considered as neighbors of each other. And the link between two documents represents the number of their common neighbors. Instead of just considering the pairwise similarity, the neighbors and link involve the global information into the measurement of the closeness of two documents. In this paper, we propose to use the neighbors and link for the family of k-means algorithms in three aspects: a new method to select initial cluster centroids based on the ranks of candidate documents; a new similarity measure which uses a combination of the cosine and link functions; and a new heuristic function for selecting a cluster to split based on the neighbors of the cluster centroids. Our experimental results on real-life data sets demonstrated that our proposed methods can significantly improve the performance of document clustering in terms of accuracy without increasing the execution time much.
Hu, J.; Fang, L.; Cao, Y.; Zeng, H.-J.; Li, H.; Yang, Q. & Chen, Z. (2008),
Enhancing text clustering by leveraging Wikipedia semantics., in
Sung-Hyon Myaeng; Douglas W. Oard; Fabrizio Sebastiani; Tat-Seng Chua & Mun-Kew Leong, ed.,
'SIGIR'
, ACM,
, pp. 179-186
.
[BibTeX]
[Endnote]
Kostoff, R. N. (2007),
'Literature-related discovery (LRD): Potential treatments for cataracts', Technological Forecasting and Social Change
In Press, Corrected Proof
, --
.
[BibTeX]
[Endnote]
Literature-related discovery (LRD) is the linking of two or more literature concepts that have heretofore not been linked (i.e., disjoint), in order to produce novel, interesting, plausible, and intelligible knowledge (i.e., potential discovery). The open discovery systems (ODS) component of LRD starts with a problem to be solved, and generates solutions to that problem through potential discovery. We have been using ODS LRD to identify potential treatments or preventative actions for challenging medical problems, among myriad other applications. This paper describes the second medical problem we addressed (cataract) using ODS LRD; the first problem addressed was Raynaud's Phenomenon (RP), and was described in the third paper of this Special Issue. Cataract was selected because it is ubiquitous globally, appears intractable to all forms of treatment other than surgical removal of cataracts, and is a major cause of blindness in many developing countries. The ODS LRD study had three objectives: a) identify non-drug non-surgical treatments that would 1) help prevent cataracts, or 2) reduce the progression rate of cataracts, or 3) stop the progression of cataracts, or 4) maybe even reverse the progression of cataracts; b) demonstrate that we could solve an ODS LRD problem with no prior knowledge of any results or prior work (unlike the case with the RP problem); c) determine whether large time savings in the discovery process were possible relative to the time required for conducting the RP study. To that end, we used the MeSH taxonomy of MEDLINE to restrict potential discoveries to selected semantic classes, as a substitute for the manually-intensive process used in the RP study to restrict potential discoveries to selected semantic classes. We also used additional semantic filtering to identify potential discovery within the selected semantic classes. All these goals were achieved. As will be shown, we generated large amounts of potential discovery in more than an order of magnitude less time than required for the RP study. We identified many non-drug non-surgical treatments that may be able to reduce or even stop the progression rate of cataracts. Time, and much testing, will determine whether this is possible. Finally, the methodology has been developed to the point where ODS LRD problems can be solved with no results or knowledge of any prior work.
Bloehdorn, S.; Cimiano, P. & Hotho, A. (2006),
Learning Ontologies to Improve Text Clustering and Classification
'From Data and Information Analysis to Knowledge Engineering'
, Springer Berlin Heidelberg,
, pp. 334--341
.
[BibTeX]
[Endnote]
Recent work has shown improvements in text clustering and classification tasks by integrating conceptual features extracted from ontologies. In this paper we present text mining experiments in the medical domain in which the ontological structures used are acquired automatically in an unsupervised learning process from the text corpus in question. We compare results obtained using the automatically learned ontologies with those obtained using manually engineered ones. Our results show that both types of ontologies improve results on text clustering and classification tasks, whereby the automatically acquired ontologies yield a improvement competitive with the manually engineered ones.
ER -
Cimiano, P.; Hotho, A. & Staab, S. (2004),
Clustering Ontologies from Text, in
'Proceedings of the Conference on Languages Resources and Evaluation (LREC)'
, ELRA - European Language Ressources Association, Lisbon, Portugal
.
[BibTeX]
[Endnote]
Hotho, A.; Staab, S. & Stumme, G. (2003),
Wordnet improves text document clustering, in
'Proc. SIGIR Semantic Web Workshop'
.
[BibTeX]
[Endnote]
Hotho, A.; Staab, S. & Stumme, G. (2003),
Wordnet improves text document clustering, in
'Proc. SIGIR Semantic Web Workshop'
.
[BibTeX]
[Endnote]
Hotho, A.; Staab, S. & Stumme, G. (2003),
Explaining Text Clustering Results using Semantic Structures, in
'Proc. of the 7th European Conference on Principles and Practice of Knowledge Discovery in Databases, PKDD'
, pp. 217-228
.
[BibTeX]
[Endnote]
Hotho, A.; Staab, S. & Stumme, G. (2003),
Explaining Text Clustering Results using Semantic Structures, in
Nada Lavrač; Dragan Gamberger & Hendrik BlockeelLjupco Todorovski, ed.,
'Knowledge Discovery in Databases: PKDD 2003, 7th European Conference on Principles and Practice of Knowledge Discovery in Databases'
, Springer, Heidelberg
, pp. 217-228
.
[BibTeX]
[Endnote]
Common text clustering techniques offer rather poor capabilities
for explaining to their users why a particular result has been
achieved. They have the disadvantage that they do not relate
semantically nearby terms and that they cannot explain how
resulting clusters are related to each other.
In this paper, we discuss a way of integrating a large thesaurus
and the computation of lattices of resulting clusters into common text clustering
in order to overcome these two problems.
As its major result, our approach achieves an explanation using an
appropriate level of granularity at the concept level as well as
an appropriate size and complexity of the explaining lattice of
resulting clusters.
Hotho, A.; Staab, S. & Stumme, G. (2003),
Ontologies improve text document clustering, in
'Proceedings of the 2003 IEEE International Conference on Data Mining'
, IEEE Computer Society, Melbourne, Florida
, pp. 541-544 (Poster
.
[BibTeX]
[Endnote]
Hotho, A.; Staab, S. & Stumme, G. (2003),
Ontologies improve text document clustering, in
'Proceedings of the 2003 IEEE International Conference on Data Mining'
, IEEE Computer Society, Melbourne, Florida
, pp. 541-544 (Poster
.
[BibTeX]
[Endnote]
Hotho, A.; Staab, S. & Stumme, G. (2003),
'Text Clustering Based on Background Knowledge'
, Technical report, University of Karlsruhe, Institute AIFB
.
[BibTeX]
[Endnote]
Text document clustering plays an important role in providing intuitive
navigation and browsing mechanisms by organizing large amounts of information
into a small number of meaningful clusters. Standard partitional or agglomerative
clustering methods efficiently compute results to this end.
However, the bag of words representation used for these clustering methods is often
unsatisfactory as it ignores relationships between important terms that do not
co-occur literally. Also, it is mostly left to the user to find out why a particular partitioning
has been achieved, because it is only specified extensionally. In order to
deal with the two problems, we integrate background knowledge into the process of
clustering text documents.
First, we preprocess the texts, enriching their representations by background knowledge
provided in a core ontology — in our application Wordnet. Then, we cluster
the documents by a partitional algorithm. Our experimental evaluation on Reuters
newsfeeds compares clustering results with pre-categorizations of news. In the experiments,
improvements of results by background knowledge compared to the baseline
can be shown for many interesting tasks.
Second, the clustering partitions the large number of documents to a relatively small
number of clusters, which may then be analyzed by conceptual clustering. In our approach,
we applied Formal Concept Analysis. Conceptual clustering techniques are
known to be too slow for directly clustering several hundreds of documents, but they
give an intensional account of cluster results. They allow for a concise description
of commonalities and distinctions of different clusters. With background knowledge
they even find abstractions like “food” (vs. specializations like “beef” or “corn”).
Thus, in our approach, partitional clustering reduces first the size of the problem
such that it becomes tractable for conceptual clustering, which then facilitates the
understanding of the results.
Hotho, A.; Staab, S. & Stumme, G. (2003),
'Text Clustering Based on Background Knowledge'
, Technical report, University of Karlsruhe, Institute AIFB
.
[BibTeX]
[Endnote]
Text document clustering plays an important role in providing intuitive
navigation and browsing mechanisms by organizing large amounts of information
into a small number of meaningful clusters. Standard partitional or agglomerative
clustering methods efficiently compute results to this end.
However, the bag of words representation used for these clustering methods is often
unsatisfactory as it ignores relationships between important terms that do not
co-occur literally. Also, it is mostly left to the user to find out why a particular partitioning
has been achieved, because it is only specified extensionally. In order to
deal with the two problems, we integrate background knowledge into the process of
clustering text documents.
First, we preprocess the texts, enriching their representations by background knowledge
provided in a core ontology — in our application Wordnet. Then, we cluster
the documents by a partitional algorithm. Our experimental evaluation on Reuters
newsfeeds compares clustering results with pre-categorizations of news. In the experiments,
improvements of results by background knowledge compared to the baseline
can be shown for many interesting tasks.
Second, the clustering partitions the large number of documents to a relatively small
number of clusters, which may then be analyzed by conceptual clustering. In our approach,
we applied Formal Concept Analysis. Conceptual clustering techniques are
known to be too slow for directly clustering several hundreds of documents, but they
give an intensional account of cluster results. They allow for a concise description
of commonalities and distinctions of different clusters. With background knowledge
they even find abstractions like “food” (vs. specializations like “beef” or “corn”).
Thus, in our approach, partitional clustering reduces first the size of the problem
such that it becomes tractable for conceptual clustering, which then facilitates the
understanding of the results.
Staab, S. & Hotho, A. (2003),
Ontology-based Text Document Clustering., in
'Intelligent Information Processing and Web Mining, Proceedings of the International IIS: IIPWM'03 Conference held in Zakopane'
, pp. 451-452
.
[BibTeX]
[Endnote]
Hotho, A. & Stumme, G. (2002),
Conceptual Clustering of Text Clusters, in
G. Kókai & J. Zeidler, ed.,
'Proc. Fachgruppentreffen Maschinelles Lernen (FGML 2002)'
, pp. 37-45
.
[BibTeX]
[Endnote]
Hotho, A. & Stumme, G. (2002),
Conceptual Clustering of Text Clusters, in
'Proceedings of FGML Workshop'
, Special Interest Group of German Informatics Society (FGML --- Fachgruppe Maschinelles Lernen der GI e.V.),
, pp. 37-45
.
[BibTeX]
[Endnote]
Hotho, A.; Maedche, A. & Staab, S. (2001),
Text Clustering Based on Good Aggregations, in
'ICDM '01: Proceedings of the 2001 IEEE International Conference on Data Mining'
, IEEE Computer Society, Washington, DC, USA
, pp. 607--608
.
[BibTeX]
[Endnote]
Toivonen, J.; Visa, A.; Vesanen, T.; Back, B. & Vanharanta, H. (2001),
Validation of Text Clustering Based on Document Contents., in
Petra Perner, ed.,
'MLDM'
, Springer,
, pp. 184-195
.
[BibTeX]
[Endnote]
Sanderson, M. & Croft, W. B. (1999),
Deriving concept hierarchies from text, in
'Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR'99'
, pp. 206--213
.
[BibTeX]
[Endnote]
Baker, L. D. & McCallum, A. K. (1998),
Distributional clustering of words for text classification, in
W. Bruce Croft; Alistair Moffat; Cornelis J. van Rijsbergen; Ross Wilkinson & Justin Zobel, ed.,
'Proceedings of SIGIR-98, 21st ACM International Conference on Research and Development in Information Retrieval'
, ACM Press, New York, US, Melbourne, AU
, pp. 96--103
.
[BibTeX]
[Endnote]