@incollection{becker2000conceptual, abstract = {Conceptual Information Systems are based on a formalization of the concept of ‘concept’ as it is discussed in traditional philosophical logic. This formalization supports a human-centered approach to the development of Information Systems. We discuss this approach by means of an implemented Conceptual Information System for supporting IT security management in companies and organizations.}, address = {Berlin/Heidelberg}, affiliation = {Entrust Technologies (Switzerland) Ltd liab. Co Glatt Tower CH-8301 Glattzentrum Switzerland}, author = {Becker, Klaus and Stumme, Gerd and Wille, Rudolf and Wille, Uta and Zickwolff, Monika}, booktitle = {Knowledge Engineering and Knowledge Management Methods, Models, and Tools}, doi = {10.1007/3-540-39967-4_27}, editor = {Dieng, Rose and Corby, Olivier}, interhash = {dacb08013d9496d41d4f9f39bce7ecd1}, intrahash = {283f8a780ac47746cc3031ad47bfdf9c}, isbn = {978-3-540-41119-2}, keyword = {Computer Science}, pages = {352--365}, publisher = {Springer}, series = {Lecture Notes in Computer Science}, title = {Conceptual Information Systems Discussed through an IT-Security Tool}, url = {http://dx.doi.org/10.1007/3-540-39967-4_27}, volume = 1937, year = 2000 } @article{poelmans2012semiautomated, abstract = {We propose an iterative and human-centred knowledge discovery methodology based on formal concept analysis. The proposed approach recognizes the important role of the domain expert in mining real-world enterprise applications and makes use of specific domain knowledge, including human intelligence and domain-specific constraints. Our approach was empirically validated at the Amsterdam-Amstelland police to identify suspects and victims of human trafficking in 266,157 suspicious activity reports. Based on guidelines of the Attorney Generals of the Netherlands, we first defined multiple early warning indicators that were used to index the police reports. Using concept lattices, we revealed numerous unknown human trafficking and loverboy suspects. In-depth investigation by the police resulted in a confirmation of their involvement in illegal activities resulting in actual arrestments been made. Our human-centred approach was embedded into operational policing practice and is now successfully used on a daily basis to cope with the vastly growing amount of unstructured information.}, author = {Poelmans, Jonas and Elzinga, Paul and Ignatov, Dmitry I. and Kuznetsov, Sergei O.}, doi = {10.1080/03081079.2012.721662}, eprint = {http://www.tandfonline.com/doi/pdf/10.1080/03081079.2012.721662}, interhash = {18d6f6312af57cc72d7e26de4903dc9f}, intrahash = {9bb41c50dd5333f94a807482489c0732}, journal = {International Journal of General Systems}, number = 8, pages = {774--804}, title = {Semi-automated knowledge discovery: identifying and profiling human trafficking}, url = {http://www.tandfonline.com/doi/abs/10.1080/03081079.2012.721662}, volume = 41, year = 2012 } @article{ley2009lessons, abstract = {The DBLP Computer Science Bibliography evolved from an early small experimental Web server to a popular service for the computer science community. Many design decisions and details of the public XML-records behind DBLP never were documented. This paper is a review of the evolution of DBLP. The main perspective is data modeling. In DBLP persons play a central role, our discussion of person names may be applicable to many other data bases. All DBLP data are available for your own experiments. You may either download the complete set, or use a simple XML-based API described in an online appendix.}, acmid = {1687577}, author = {Ley, Michael}, interhash = {a75ae2987d55512b7d0731c7a11a1722}, intrahash = {bb968ff4ba9ae93bc80ba05d16a98ff4}, issn = {2150-8097}, issue_date = {August 2009}, journal = {Proceedings of the VLDB Endowment}, month = aug, number = 2, numpages = {8}, pages = {1493--1500}, publisher = {VLDB Endowment}, title = {DBLP: some lessons learned}, url = {http://dl.acm.org/citation.cfm?id=1687553.1687577}, volume = 2, year = 2009 } @book{koester2006fooca, abstract = {This book deals with Formal Concept Analysis (FCA) and its application to Web Information Retrieval. It explains how Web search results retrieved by major Web search engines such as Google or Yahoo can be conceptualized leading to a human-oriented form of representation. A generalization of Web search results is conducted, leading to an FCA-based introduction of FooCA. FooCA is an application in the field of Conceptual Knowledge Processing and supports the idea of a holistic representation of Web Information Retrieval.}, address = {Mühltal}, author = {Koester, Bjoern}, interhash = {fe53b2b1fa6be34259647954fca36bf8}, intrahash = {5571d950ada3ee1892e5c043ac438271}, publisher = {Verlag Allgemeine Wissenschaft}, series = {Beiträge zur begrifflichen Wissensverarbeitung}, title = {FooCA: web information retrieval with formal concept analysis}, url = {http://www.bjoern-koester.de/fooca/web_information_retrieval_with_formal_concept_analysis.html}, year = 2006 } @inproceedings{baader2007completing, abstract = {We propose an approach for extending both the terminological and the assertional part of a Description Logic knowledge base by using information provided by the knowledge base and by a domain expert. The use of techniques from Formal Concept Analysis ensures that, on the one hand, the interaction with the expert is kept to a minimum, and, on the other hand, we can show that the extended knowledge base is complete in a certain, well-defined sense.}, acmid = {1625311}, address = {San Francisco, CA, USA}, author = {Baader, Franz and Ganter, Bernhard and Sertkaya, Baris and Sattler, Ulrike}, booktitle = {Proceedings of the 20th international joint conference on Artifical intelligence}, interhash = {8ab382f3aa141674412ba7ad33316a9b}, intrahash = {87f98ae486014ba78690ffa314b67da8}, location = {Hyderabad, India}, numpages = {6}, pages = {230--235}, publisher = {Morgan Kaufmann Publishers Inc.}, title = {Completing description logic knowledge bases using formal concept analysis}, url = {http://dl.acm.org/citation.cfm?id=1625275.1625311}, year = 2007 } @article{birkholz2012scalable, abstract = {Studies on social networks have proved that endogenous and exogenous factors influence dynamics. Two streams of modeling exist on explaining the dynamics of social networks: 1) models predicting links through network properties, and 2) models considering the effects of social attributes. In this interdisciplinary study we work to overcome a number of computational limitations within these current models. We employ a mean-field model which allows for the construction of a population-specific socially informed model for predicting links from both network and social properties in large social networks. The model is tested on a population of conference coauthorship behavior, considering a number of parameters from available Web data. We address how large social networks can be modeled preserving both network and social parameters. We prove that the mean-field model, using a data-aware approach, allows us to overcome computational burdens and thus scalability issues in modeling large social networks in terms of both network and social parameters. Additionally, we confirm that large social networks evolve through both network and social-selection decisions; asserting that the dynamics of networks cannot singly be studied from a single perspective but must consider effects of social parameters. }, author = {Birkholz, Julie M. and Bakhshi, Rena and Harige, Ravindra and van Steen, Maarten and Groenewegen, Peter}, interhash = {a8ef0aac2eab74fc8eb3f9d3dc8a32dd}, intrahash = {aefcc2aa922b048bec85d5070494ed81}, journal = {CoRR}, month = sep, title = {Scalable Analysis of Socially Informed Network Models: the data-aware mean-field approach }, url = {http://arxiv.org/abs/1209.6615}, volume = {abs/1209.6615}, year = 2012 } @inproceedings{pfaltz2012entropy, abstract = {We introduce the concepts of closed sets and closure operators as mathematical tools for the study of social networks. Dynamic networks are represented by transformations. It is shown that under continuous change/transformation, all networks tend to "break down" and become less complex. It is a kind of entropy. The product of this theoretical decomposition is an abundance of triadically closed clusters which sociologists have observed in practice. This gives credence to the relevance of this kind of mathematical analysis in the sociological context. }, author = {Pfaltz, John L.}, booktitle = {Proceedings of the SOCINFO}, interhash = {753f13a5ffaa0946220164c2b05c230f}, intrahash = {044d0b1f6e737bede270a40bbddb0b06}, title = {Entropy in Social Networks}, year = 2012 } @article{obiedkov2009building, abstract = {The use of lattice-based access control models has been somewhat restricted by their complexity. We argue that attribute exploration from formal concept analysis can help create lattice models of manageable size, while making it possible for the system designer to better understand dependencies between different security categories in the domain and, thus, providing certain guarantees for the relevance of the constructed model to a particular application. In this paper, we introduce the method through an example.}, author = {Obiedkov, Sergei and Kourie, Derrick G. and Eloff, J.H.P.}, doi = {10.1016/j.cose.2008.07.011}, interhash = {367ceb95cd5e3964aa2d7d00ad21da09}, intrahash = {7be2b4bf0987c4d18adf7243eae690c0}, issn = {0167-4048}, journal = {Computers and Security}, number = {1–2}, pages = {2--7}, title = {Building access control models with attribute exploration}, url = {http://www.sciencedirect.com/science/article/pii/S0167404808000497}, volume = 28, year = 2009 } @article{kleinberg2013analysis, abstract = {The growth of the Web has required us to think about the design of information systems in which large-scale computational and social feedback effects are simultaneously at work. At the same time, the data generated by Web-scale systems—recording the ways in which millions of participants create content, link information, form groups and communicate with one another—have made it possible to evaluate long-standing theories of social interaction, and to formulate new theories based on what we observe. These developments have created a new level of interaction between computing and the social sciences, enriching the perspectives of both of these disciplines. We discuss some of the observations, theories and conclusions that have grown from the study of Web-scale social interaction, focusing on issues including the mechanisms by which people join groups, the ways in which different groups are linked together in social networks and the interplay of positive and negative interactions in these networks.}, author = {Kleinberg, Jon}, doi = {10.1098/rsta.2012.0378}, eprint = {http://rsta.royalsocietypublishing.org/content/371/1987/20120378.full.pdf+html}, interhash = {b4686f01da53c975f342dbb40bdd1a90}, intrahash = {e3898cfb7206a7fee8eb3a5419aa030f}, journal = {Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences}, month = mar, number = 1987, title = {Analysis of large-scale social and information networks}, url = {http://rsta.royalsocietypublishing.org/content/371/1987/20120378.abstract}, volume = 371, year = 2013 } @article{barabsi2013network, abstract = {Professor Barabási's talk described how the tools of network science can help understand the Web's structure, development and weaknesses. The Web is an information network, in which the nodes are documents (at the time of writing over one trillion of them), connected by links. Other well-known network structures include the Internet, a physical network where the nodes are routers and the links are physical connections, and organizations, where the nodes are people and the links represent communications.}, author = {Barabási, Albert-László}, doi = {10.1098/rsta.2012.0375}, eprint = {http://rsta.royalsocietypublishing.org/content/371/1987/20120375.full.pdf+html}, interhash = {e2cfdd2e3c7c68581e3ab691909ed28b}, intrahash = {208c1f9d6d8eff67cee07ebdf3cd0fc1}, journal = {Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences}, number = 1987, title = {Network science}, url = {http://rsta.royalsocietypublishing.org/content/371/1987/20120375.abstract}, volume = 371, year = 2013 } @inproceedings{ollivier2007finding, abstract = {We introduce a new method for finding nodes semantically related to a given node in a hyperlinked graph: the Green method, based on a classical Markov chain tool. It is generic, adjustment-free and easy to implement. We test it in the case of the hyperlink structure of the English version of Wikipedia, the on-line encyclopedia. We present an extensive comparative study of the performance of our method versus several other classical methods in the case of Wikipedia. The Green method is found to have both the best average results and the best robustness.}, acmid = {1619874}, author = {Ollivier, Yann and Senellart, Pierre}, booktitle = {Proceedings of the 22nd national conference on Artificial intelligence}, interhash = {a291b1b4e195dd09a11c8ffe329fc0e5}, intrahash = {76e219fe6e8a257b30c6665af8b273da}, isbn = {978-1-57735-323-2}, location = {Vancouver, British Columbia, Canada}, numpages = {7}, pages = {1427--1433}, publisher = {AAAI Press}, title = {Finding related pages using Green measures: an illustration with Wikipedia}, url = {http://dl.acm.org/citation.cfm?id=1619797.1619874}, volume = 2, year = 2007 } @inproceedings{takahashi2011evaluating, abstract = {We propose a method to evaluate the significance of historical entities (people, events, and so on.). Here, the significance of a historical entity means how it affected other historical entities. Our proposed method first calculates the tempo-spacial impact of historical entities. The impact of a historical entity varies according to time and location. Historical entities are collected from Wikipedia. We assume that a Wikipedia link between historical entities represents an impact propagation. That is, when an entity has a link to another entity, we regard the former is influenced by the latter. Historical entities in Wikipedia usually have the date and location of their occurrence. Our proposed iteration algorithm propagates such initial tempo-spacial information through links in the similar manner as PageRank, so the tempo-spacial impact scores of all the historical entities can be calculated. We assume that a historical entity is significant if it influences many other entities that are far from it temporally or geographically. We demonstrate a prototype system and show the results of experiments that prove the effectiveness of our method.}, acmid = {1995980}, address = {New York, NY, USA}, author = {Takahashi, Yuku and Ohshima, Hiroaki and Yamamoto, Mitsuo and Iwasaki, Hirotoshi and Oyama, Satoshi and Tanaka, Katsumi}, booktitle = {Proceedings of the 22nd ACM conference on Hypertext and hypermedia}, doi = {10.1145/1995966.1995980}, interhash = {6665836546bedb1ee5d56a4d16a0848e}, intrahash = {e4769d86e71c9e7ba77d5d4af6f21e0c}, isbn = {978-1-4503-0256-2}, location = {Eindhoven, The Netherlands}, numpages = {10}, pages = {83--92}, publisher = {ACM}, title = {Evaluating significance of historical entities based on tempo-spatial impacts analysis using Wikipedia link structure}, url = {http://doi.acm.org/10.1145/1995966.1995980}, year = 2011 } @inproceedings{zesch2007analysis, abstract = {In this paper, we discuss two graphs in Wikipedia (i) the article graph, and (ii) the category graph. We perform a graph-theoretic analysis of the category graph, and show that it is a scale-free, small world graph like other well-known lexical semantic networks. We substantiate our findings by transferring semantic relatedness algorithms defined on WordNet to the Wikipedia category graph. To assess the usefulness of the category graph as an NLP resource, we analyze its coverage and the performance of the transferred semantic relatedness algorithms. }, address = {Rochester}, author = {Zesch, Torsten and Gurevych, Iryna}, booktitle = {Proceedings of the TextGraphs-2 Workshop (NAACL-HLT)}, interhash = {0401e62edb9bfa85dd498cb40301c0cb}, intrahash = {332ed720a72bf069275f93485432314b}, month = apr, pages = {1--8}, publisher = {Association for Computational Linguistics}, title = {Analysis of the Wikipedia Category Graph for NLP Applications}, url = {http://acl.ldc.upenn.edu/W/W07/W07-02.pdf#page=11}, year = 2007 } @article{batagelj2011algorithms, abstract = {The structure of a large network (graph) can often be revealed by partitioning it into smaller and possibly more dense sub-networks that are easier to handle. One of such decompositions is based on “}, author = {Batagelj, Vladimir and Zaveršnik, Matjaž}, doi = {10.1007/s11634-010-0079-y}, interhash = {a0bd7331f81bb4da72ce115d5943d6e4}, intrahash = {cd0d5266688af6bb98bde7f99e3a54c1}, issn = {1862-5347}, journal = {Advances in Data Analysis and Classification}, language = {English}, number = 2, pages = {129--145}, publisher = {Springer}, title = {Fast algorithms for determining (generalized) core groups in social networks}, url = {http://dx.doi.org/10.1007/s11634-010-0079-y}, volume = 5, year = 2011 } @article{larowe2009scholarly, abstract = {The Scholarly Database aims to serve researchers and practitioners interested in the analysis, modelling, and visualization of large-scale data sets. A specific focus of this database is to support macro-evolutionary studies of science and to communicate findings via knowledge-domain visualizations. Currently, the database provides access to about 18 million publications, patents, and grants. About 90% of the publications are available in full text. Except for some datasets with restricted access conditions, the data can be retrieved in raw or pre-processed formats using either a web-based or a relational database client. This paper motivates the need for the database from the perspective of bibliometric/scientometric research. It explains the database design, setup, etc., and reports the temporal, geographical, and topic coverage of data sets currently served via the database. Planned work and the potential for this database to become a global testbed for information science research are discussed at the end of the paper.}, author = {La Rowe, Gavin and Ambre, Sumeet and Burgoon, John and Ke, Weimao and Börner, Katy}, doi = {10.1007/s11192-009-0414-2}, interhash = {1819f263b0ea1b99ec15d0c22b38207e}, intrahash = {c24611ec1f2efbdcf7f5b26d49af320e}, issn = {0138-9130}, journal = {Scientometrics}, language = {English}, number = 2, pages = {219--234}, publisher = {Springer Netherlands}, title = {The Scholarly Database and its utility for scientometrics research}, url = {http://dx.doi.org/10.1007/s11192-009-0414-2}, volume = 79, year = 2009 } @inproceedings{yan2012better, abstract = {Usually scientists breed research ideas inspired by previous publications, but they are unlikely to follow all publications in the unbounded literature collection. The volume of literature keeps on expanding extremely fast, whilst not all papers contribute equal impact to the academic society. Being aware of potentially influential literature would put one in an advanced position in choosing important research references. Hence, estimation of potential influence is of great significance. We study a challenging problem of identifying potentially influential literature. We examine a set of hypotheses on what are the fundamental characteristics for highly cited papers and find some interesting patterns. Based on these observations, we learn to identify potentially influential literature via Future Influence Prediction (FIP), which aims to estimate the future influence of literature. The system takes a series of features of a particular publication as input and produces as output the estimated citation counts of that article after a given time period. We consider several regression models to formulate the learning process and evaluate their performance based on the coefficient of determination (R2). Experimental results on a real-large data set show a mean average predictive performance of 83.6% measured in R^2. We apply the learned model to the application of bibliography recommendation and obtain prominent performance improvement in terms of Mean Average Precision (MAP).}, acmid = {2232831}, address = {New York, NY, USA}, author = {Yan, Rui and Huang, Congrui and Tang, Jie and Zhang, Yan and Li, Xiaoming}, booktitle = {Proceedings of the 12th ACM/IEEE-CS joint conference on Digital Libraries}, doi = {10.1145/2232817.2232831}, interhash = {85d10c6d37bcbfa057c51acc325a8116}, intrahash = {9269d2dd9bf4bc8c0e7c668011fcfc1b}, isbn = {978-1-4503-1154-0}, location = {Washington, DC, USA}, numpages = {10}, pages = {51--60}, publisher = {ACM}, series = {JCDL '12}, title = {To better stand on the shoulder of giants}, url = {http://doi.acm.org/10.1145/2232817.2232831}, year = 2012 } @article{liu2012fulltext, author = {Liu, Xiaozhong and Zhang, Jinsong and Guo, Chun}, interhash = {011df26355ad51a88947017fd2791a98}, intrahash = {f9c6133bf4503003822f99860f864698}, journal = {Journal of the American Society for Information Science and Technology}, title = {Full-Text Citation Analysis: A New Method to Enhance Scholarly Network}, url = {http://discern.uits.iu.edu:8790/publication/Full%20text%20citation.pdf}, year = 2012 } @article{nanba2000classification, abstract = {We are investigating automatic generation of a review (or survey) article in a specific subject domain. In a research paper, there are passages where the author describes the essence of a cited paper and the differences between the current paper and the cited paper (we call them citing areas). These passages can be considered as a kind of summary of the cited paper from the current author's viewpoint. We can know the state of the art in a specific subject domain from the collection of citing areas. FUrther, if these citing areas are properly classified and organized, they can act 8.', a kind of a review article. In our previous research, we proposed the automatic extraction of citing areas. Then, with the information in the citing areas, we automatically identified the types of citation relationships that indicate the reasons for citation (we call them citation types). Citation types offer a useful clue for organizing citing areas. In addition, to support writing a review article, it is necessary to take account of the contents of the papers together with the citation links and citation types. In this paper, we propose several methods for classifying papers automatically. We found that our proposed methods BCCT-C, the bibliographic coupling considering only type C citations, which pointed out the problems or gaps in related works, are more effective than others. We also implemented a prototype system to support writing a review article, which is based on our proposed method.}, author = {Nanba, H. and Kando, N. and Okumura, M.}, interhash = {a8fbc36d3ee8de28f65ef2486bb18cd2}, intrahash = {7a99ee2d1444ae569beb7bee04137e4b}, journal = {11th ASIS SIG/CR Classification Research Workshop}, misc = {10.7152/acro.v11i1.12774}, pages = {117--134}, title = {Classification of research papers using citation links and citation types: Towards automatic review article generation}, url = {http://journals.lib.washington.edu/index.php/acro/article/download/12774/11255}, year = 2000 } @inproceedings{angelova2008characterizing, abstract = {Social networks and collaborative tagging systems are rapidly gaining popularity as a primary means for storing and sharing data among friends, family, colleagues, or perfect strangers as long as they have common interests. del.icio.us is a social network where people store and share their personal bookmarks. Most importantly, users tag their bookmarks for ease of information dissemination and later look up. However, it is the friendship links, that make delicious a social network. They exist independently of the set of bookmarks that belong to the users and have no relation to the tags typically assigned to the bookmarks. To study the interaction among users, the strength of the existing links and their hidden meaning, we introduce implicit links in the network. These links connect only highly "similar" users. Here, similarity can reflect different aspects of the user’s profile that makes her similar to any other user, such as number of shared bookmarks, or similarity of their tags clouds. We investigate the question whether friends have common interests, we gain additional insights on the strategies that users use to assign tags to their bookmarks, and we demonstrate that the graphs formed by implicit links have unique properties differing from binomial random graphs or random graphs with an expected power-law degree distribution. }, author = {Angelova, Ralitsa and Lipczak, Marek and Milios, Evangelos and Prałat, Paweł}, booktitle = {Proceedings of the Mining Social Data Workshop (MSoDa)}, interhash = {f74d27a66d2754f3d5892d68c4abee4c}, intrahash = {02d6739886a13180dd92fbb7243ab58b}, month = jul, organization = {ECAI 2008}, pages = {21--25}, title = {Characterizing a social bookmarking and tagging network}, url = {http://www.math.ryerson.ca/~pralat/papers/2008_delicious.pdf}, year = 2008 } @inproceedings{baur2007generating, abstract = {The modeling of realistic networks is of great importance for modern complex systems research. Previous procedures typically model the natural growth of networks by means of iteratively adding nodes, geometric positioning information, a definition of link connectivity based on the preference for nearest neighbors or already highly connected nodes, or combine several of these approaches. Our novel model is based on the well-know concept of k-cores, originally introduced in social network analysis. Recent studies exposed the significant k-core structure of several real world systems, e.g. the AS network of the Internet. We present a simple and efficient method for generating networks which strictly adhere to the characteristics of a given k-core structure, called core fingerprint. We show-case our algorithm in a comparative evaluation with two well-known AS network generators. }, author = {Baur, Michael and Gaertler, Marco and Görke, Robert and Krug, Marcus and Wagner, Dorothea}, booktitle = {Proceedings of the European Conference of Complex Systems}, interhash = {387eebb80bbfaafab5ac201c88ebd263}, intrahash = {e2fef8dce15087afbcc3489f2029d2c6}, month = oct, title = {Generating Graphs with Predefined k-Core Structure}, url = {http://i11www.ira.uka.de/extra/publications/bggkw-ggpcs-07.pdf}, year = 2007 }