QuickSearch:   Number of matching entries: 0.

Search Settings

    AuthorTitleYearJournal/ProceedingsReftypeDOI/URL
    Haridas, M. & Caragea, D. Exploring Wikipedia and DMoz as Knowledge Bases for Engineering a User Interests Hierarchy for Social Network Applications 2009
    Vol. 5871On the Move to Meaningful Internet Systems: OTM 2009, pp. 1238-1245 
    incollection DOI URL 
    Abstract: The outgrowth of social networks in the recent years has resulted in opportunities for interesting data mining problems, such as interest or friendship recommendations. A global ontology over the interests specified by the users of a social network is essential for accurate recommendations. We propose, evaluate and compare three approaches to engineering a hierarchical ontology over user interests. The proposed approaches make use of two popular knowledge bases, Wikipedia and Directory Mozilla, to extract interest definitions and/or relationships between interests. More precisely, the first approach uses Wikipedia to find interest definitions, the latent semantic analysis technique to measure the similarity between interests based on their definitions, and an agglomerative clustering algorithm to group similar interests into higher level concepts. The second approach uses the Wikipedia Category Graph to extract relationships between interests, while the third approach uses Directory Mozilla to extract relationships between interests. Our results show that the third approach, although the simplest, is the most effective for building a hierarchy over user interests.
    BibTeX:
    @incollection{haridas2009exploring,
      author = {Haridas, Mandar and Caragea, Doina},
      title = {Exploring Wikipedia and DMoz as Knowledge Bases for Engineering a User Interests Hierarchy for Social Network Applications},
      booktitle = {On the Move to Meaningful Internet Systems: OTM 2009},
      publisher = {Springer},
      year = {2009},
      volume = {5871},
      pages = {1238-1245},
      url = {http://dx.doi.org/10.1007/978-3-642-05151-7_35},
      doi = {http://dx.doi.org/10.1007/978-3-642-05151-7_35}
    }
    
    Silva, L.D. & Jayaratne, L. Semi-automatic extraction and modeling of ontologies using Wikipedia XML Corpus 2009 Applications of Digital Information and Web Technologies, 2009. ICADIWT '09. Second International Conference on the, pp. 446-451  inproceedings DOI URL 
    Abstract: This paper introduces WikiOnto: a system that assists in the extraction and modeling of topic ontologies in a semi-automatic manner using a preprocessed document corpus derived from Wikipedia. Based on the Wikipedia XML Corpus, we present a three-tiered framework for extracting topic ontologies in quick time and a modeling environment to refine these ontologies. Using natural language processing (NLP) and other machine learning (ML) techniques along with a very rich document corpus, this system proposes a solution to a task that is generally considered extremely cumbersome. The initial results of the prototype suggest strong potential of the system to become highly successful in ontology extraction and modeling and also inspire further research on extracting ontologies from other semi-structured document corpora as well.
    BibTeX:
    @inproceedings{silva2009semiautomatic,
      author = {Silva, L. De and Jayaratne, L.},
      title = {Semi-automatic extraction and modeling of ontologies using Wikipedia XML Corpus},
      booktitle = {Applications of Digital Information and Web Technologies, 2009. ICADIWT '09. Second International Conference on the},
      year = {2009},
      pages = {446-451},
      url = {http://ieeexplore.ieee.org/xpls/abs_all.jsp?isnumber=5273826&arnumber=5273871&count=156&index=116},
      doi = {http://dx.doi.org/10.1109/ICADIWT.2009.5273871}
    }
    
    Grineva, M., Grinev, M., Turdakov, D. & Velikhov, P. Harnessing Wikipedia for Smart Tags Clustering 2008 Proceedings of the International Workshop on Knowledge Acquisition from the Social Web (KASW2008)  inproceedings  
    Abstract: The quality of the current tagging services can be greatly improved if the service is able to cluster tags by their meaning. Tag clouds clustered by higher level topics enable the users to explore their tag space, which is especially needed when tag clouds become large. We demonstrate TagCluster - a tool for automated tag clustering that harnesses knowledge from Wikipedia about semantic relatedness between tags and names of categories to achieve smart clustering. Our approach shows much better quality of clusters compared to the existing techniques that rely on tag co-occurrence analysis in the tagging service.
    BibTeX:
    @inproceedings{grineva2008harnessing,
      author = {Grineva, Maria and Grinev, Maxim and Turdakov, Denis and Velikhov, Pavel},
      title = {Harnessing Wikipedia for Smart Tags Clustering},
      booktitle = {Proceedings of the International Workshop on Knowledge Acquisition from the Social Web (KASW2008)},
      year = {2008}
    }
    
    Medelyan, O. & Legg, C. Integrating Cyc and Wikipedia: Folksonomy meets rigorously defined common-sense 2008
    Vol. 8Proceedings of the WIKI-AI: Wikipedia and AI Workshop at the AAAI 
    inproceedings URL 
    Abstract: Integration of ontologies begins with establishing mappings between their concept entries. We map categories from the largest manually-built ontology, Cyc, onto Wikipedia articles describing corresponding concepts. Our method draws both on Wikipedia’s rich but chaotic hyperlink structure and Cyc’s carefully defined taxonomic and common-sense knowledge. On 9,333 manual alignments by one person, we achieve an F-measure of 90%; on 100 alignments by six human subjects the average agreement of the method with the subject is close to their agreement with each other. We cover 62.8% of Cyc categories relating to common-sense knowledge and discuss what further information might be added to Cyc given this substantial new alignment.
    BibTeX:
    @inproceedings{medelyan2008integrating,
      author = {Medelyan, O. and Legg, C.},
      title = {Integrating Cyc and Wikipedia: Folksonomy meets rigorously defined common-sense},
      booktitle = {Proceedings of the WIKI-AI: Wikipedia and AI Workshop at the AAAI},
      year = {2008},
      volume = {8},
      url = {http://scholar.google.de/scholar.bib?q=info:hgFpsjJR__4J:scholar.google.com/&output=citation&hl=de&as_sdt=2000&ct=citation&cd=58}
    }
    
    Nazir, F. & Takeda, H. Extraction and analysis of tripartite relationships from Wikipedia 2008 IEEE International Symposium on Technology and Society, pp. 1-13  inproceedings DOI URL 
    Abstract: Social aspects are critical in the decision making process for social actors (human beings). Social aspects can be categorized into social interaction, social communities, social groups or any kind of behavior that emerges from interlinking, overlapping or similarities between interests of a society. These social aspects are dynamic and emergent. Therefore, interlinking them in a social structure, based on bipartite affiliation network, may result in isolated graphs. The major reason is that as these correspondences are dynamic and emergent, they should be coupled with more than a single affiliation in order to sustain the interconnections during interest evolutions. In this paper we propose to interlink actors using multiple tripartite graphs rather than a bipartite graph which was the focus of most of the previous social network building techniques. The utmost benefit of using tripartite graphs is that we can have multiple and hierarchical links between social actors. Therefore in this paper we discuss the extraction, plotting and analysis methods of tripartite relations between authors, articles and categories from Wikipedia. Furthermore, we also discuss the advantages of tripartite relationships over bipartite relationships. As a conclusion of this study we argue based on our results that to build useful, robust and dynamic social networks, actors should be interlinked in one or more tripartite networks.
    BibTeX:
    @inproceedings{nazir2008extraction,
      author = {Nazir, F. and Takeda, H.},
      title = {Extraction and analysis of tripartite relationships from Wikipedia},
      booktitle = {IEEE International Symposium on Technology and Society},
      year = {2008},
      pages = {1--13},
      url = {http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=4559785},
      doi = {http://dx.doi.org/10.1109/ISTAS.2008.4559785}
    }
    
    Auer, Sö. & Lehmann, J. What Have Innsbruck and Leipzig in Common? Extracting Semantics from Wiki Content 2007 ESWC, pp. 503-517  inproceedings URL 
    Abstract: Wikis are established means for the collaborative authoring, versioning and publishing of textual articles. The Wikipedia project, for example, succeeded in creating the by far largest encyclopedia just on the basis of a wiki. Recently, several approaches have been proposed on how to extend wikis to allow the creation of structured and semantically enriched content. However, the means for creating semantically enriched structured content are already available and are, although unconsciously, even used by Wikipedia authors. In this article, we present a method for revealing this structured content by extracting information from template instances. We suggest ways to efficiently query the vast amount of extracted information (e.g. more than 8 million RDF statements for the English Wikipedia version alone), leading to astonishing query answering possibilities (such as for the title question). We analyze the quality of the extracted content, and propose strategies for quality improvements with just minor modifications of the wiki systems being currently used.
    BibTeX:
    @inproceedings{auer2007what,
      author = {Auer, Sören and Lehmann, Jens},
      title = {What Have Innsbruck and Leipzig in Common? Extracting Semantics from Wiki Content},
      booktitle = {ESWC},
      year = {2007},
      pages = {503-517},
      url = {http://www.springerlink.com/content/3131t21p634191n2/}
    }
    
    Ponzetto, S.P. & Strube, M. Deriving a Large-Scale Taxonomy from Wikipedia. 2007 AAAI, pp. 1440-1445  inproceedings URL 
    Abstract: We take the category system inWikipedia as a conceptual network. We label the semantic relations between categories using methods based on connectivity in the network and lexicosyntactic matching. As a result we are able to derive a large scale taxonomy containing a large amount of subsumption, i.e. isa, relations. We evaluate the quality of the created resource by comparing it with ResearchCyc, one of the largest manually annotated ontologies, as well as computing semantic similarity between words in benchmarking datasets.
    BibTeX:
    @inproceedings{ponzetto2007deriving,
      author = {Ponzetto, Simone Paolo and Strube, Michael},
      title = {Deriving a Large-Scale Taxonomy from Wikipedia.},
      booktitle = {AAAI},
      publisher = {AAAI Press},
      year = {2007},
      pages = {1440-1445},
      url = {http://dblp.uni-trier.de/db/conf/aaai/aaai2007.html#PonzettoS07}
    }
    
    Strube, M. & Ponzetto, S.P. WikiRelate! Computing Semantic Relatedness Using Wikipedia. 2006 AAAI  inproceedings URL 
    Abstract: Wikipedia provides a knowledge base for computing word relatedness in a more structured fashion than a search engine and with more coverage than WordNet. In this work we present experiments on using Wikipedia for computing semantic relatedness and compare it to WordNet on various benchmarking datasets. Existing relatedness measures perform better using Wikipedia than a baseline given by Google counts, and we show that Wikipedia outperforms WordNet when applied to the largest available dataset designed for that purpose. The best results on this dataset are obtained by integrating Google, WordNet and Wikipedia based measures. We also show that including Wikipedia improves the performance of an NLP application processing naturally occurring texts.
    BibTeX:
    @inproceedings{strube2006wikirelate,
      author = {Strube, Michael and Ponzetto, Simone Paolo},
      title = {WikiRelate! Computing Semantic Relatedness Using Wikipedia.},
      booktitle = {AAAI},
      publisher = {AAAI Press},
      year = {2006},
      url = {http://www.dit.unitn.it/~p2p/RelatedWork/Matching/aaai06.pdf}
    }
    
    Ruiz-Casado, M., Alfonseca, E. & Castells, P. Automatic Extraction of Semantic Relationships for WordNet by Means of Pattern Learning from Wikipedia 2005
    Vol. 3513Natural Language Processing and Information Systems, pp. 233-242 
    incollection DOI URL 
    Abstract: This paper describes an automatic approach to identify lexical patterns which represent semantic relationships between concepts, from an on-line encyclopedia. Next, these patterns can be applied to extend existing ontologies or semantic networks with new relations. The experiments have been performed with the Simple English Wikipedia and WordNet 1.7. A new algorithm has been devised for automatically generalising the lexical patterns found in the encyclopedia entries. We have found general patterns for the hyperonymy, hyponymy, holonymy and meronymy relations and, using them, we have extracted more than 1200 new relationships that did not appear in WordNet originally. The precision of these relationships ranges between 0.61 and 0.69, depending on the relation.
    BibTeX:
    @incollection{ruizcasado2005automatic,
      author = {Ruiz-Casado, Maria and Alfonseca, Enrique and Castells, Pablo},
      title = {Automatic Extraction of Semantic Relationships for WordNet by Means of Pattern Learning from Wikipedia},
      booktitle = {Natural Language Processing and Information Systems},
      publisher = {Springer},
      year = {2005},
      volume = {3513},
      pages = {233-242},
      url = {http://dx.doi.org/10.1007/11428817_7},
      doi = {http://dx.doi.org/10.1007/11428817_7}
    }
    

    Created by JabRef on 30/04/2024.