@inproceedings{decke2024structured, abstract = {This article investigates the application of computer vision and graph-based models in solving mesh-based partial differential equations within high-performance computing environments. Focusing on structured, graded structured, and unstructured meshes, the study compares the performance and computational efficiency of three computer vision-based models against three graph-based models across three datasets. The research aims to identify the most suitable models for different mesh topographies, particularly highlighting the exploration of graded meshes, a less studied area. Results demonstrate that computer vision-based models, notably U-Net, outperform the graph models in prediction performance and efficiency in two (structured and graded) out of three mesh topographies. The study also reveals the unexpected effectiveness of computer vision-based models in handling unstructured meshes, suggesting a potential shift in methodological approaches for data-driven partial differential equation learning. The article underscores deep learning as a viable and potentially sustainable way to enhance traditional high-performance computing methods, advocating for informed model selection based on the topography of the mesh.}, author = {Decke, Jens and Wünsch, Olaf and Sick, Bernhard and Gruhl, Christian}, booktitle = {International Conference on Architecture of Computing Systems (ARCS)}, interhash = {ed6bd9576bb4b4cfbbce236485b92f99}, intrahash = {bf3baf1e39a1493c9cf4fcb18c6a9334}, note = {(accepted)}, publisher = {Springer}, title = {From Structured to Unstructured: A Comparative Analysis of Computer Vision and Graph Models in solving Mesh-based PDEs}, year = 2024 } @article{mueller-2014b, abstract = {The combination of ubiquitous and social computing is an emerging research area which integrates different but complementary methods, techniques and tools. In this paper, we focus on the Ubicon platform, its applications, and a large spectrum of analysis results. Ubicon provides an extensible framework for building and hosting applications targeting both ubiquitous and social environments. We summarize the architecture and exemplify its implementation using four real-world applications built on top of Ubicon. In addition, we discuss several scientific experiments in the context of these applications in order to give a better picture of the potential of the framework, and discuss analysis results using several real-world data sets collected utilizing Ubicon.}, author = {Atzmueller, Martin and Becker, Martin and Kibanov, Mark and Scholz, Christoph and Doerfel, Stephan and Hotho, Andreas and Macek, Bjoern-Elmar and Mitzlaff, Folke and Mueller, Juergen and Stumme, Gerd}, doi = {10.1080/13614568.2013.873488}, interhash = {6364e034fa868644b30618dc887c0270}, intrahash = {d38f1e01e735253b4cad2c98c1027659}, issn = {1361-4568}, journal = {New Review of Hypermedia and Multimedia}, month = {#mar#}, number = 20, pages = {53--77}, title = {Ubicon and its Applications for Ubiquitous Social Computing}, url = {http://dx.doi.org/10.1080/13614568.2013.873488}, volume = 1, year = 2014 } @article{atzmueller2014ubicon, abstract = {The combination of ubiquitous and social computing is an emerging research area which integrates different but complementary methods, techniques and tools. In this paper, we focus on the Ubicon platform, its applications, and a large spectrum of analysis results. Ubicon provides an extensible framework for building and hosting applications targeting both ubiquitous and social environments. We summarize the architecture and exemplify its implementation using four real-world applications built on top of Ubicon. In addition, we discuss several scientific experiments in the context of these applications in order to give a better picture of the potential of the framework, and discuss analysis results using several real-world data sets collected utilizing Ubicon.}, author = {Atzmueller, Martin and Becker, Martin and Kibanov, Mark and Scholz, Christoph and Doerfel, Stephan and Hotho, Andreas and Macek, Bjoern-Elmar and Mitzlaff, Folke and Mueller, Juergen and Stumme, Gerd}, doi = {10.1080/13614568.2013.873488}, eprint = {http://www.tandfonline.com/doi/pdf/10.1080/13614568.2013.873488}, interhash = {6364e034fa868644b30618dc887c0270}, intrahash = {5d1ed63c337f8473d2b5b3b6c02a5f20}, journal = {New Review of Hypermedia and Multimedia}, number = 1, pages = {53-77}, title = {Ubicon and its applications for ubiquitous social computing}, url = {http://www.tandfonline.com/doi/abs/10.1080/13614568.2013.873488}, volume = 20, year = 2014 } @inproceedings{jaschke2013attribute, abstract = {We propose an approach for supporting attribute exploration by web information retrieval, in particular by posing appropriate queries to search engines, crowd sourcing systems, and the linked open data cloud. We discuss underlying general assumptions for this to work and the degree to which these can be taken for granted.}, author = {Jäschke, Robert and Rudolph, Sebastian}, booktitle = {Contributions to the 11th International Conference on Formal Concept Analysis}, editor = {Cellier, Peggy and Distel, Felix and Ganter, Bernhard}, interhash = {000ab7b0ae3ecd1d7d6ceb39de5c11d4}, intrahash = {45e900e280661d775d8da949baee3747}, month = may, organization = {Technische Universität Dresden}, pages = {19--34}, title = {Attribute Exploration on the Web}, url = {http://nbn-resolving.de/urn:nbn:de:bsz:14-qucosa-113133}, urn = {urn:nbn:de:bsz:14-qucosa-113133}, year = 2013 } @article{robertson2013programming, abstract = {The aim of ‘programming the global computer’ was identified by Milner and others as one of the grand challenges of computing research. At the time this phrase was coined, it was natural to assume that this objective might be achieved primarily through extending programming and specification languages. The Internet, however, has brought with it a different style of computation that (although harnessing variants of traditional programming languages) operates in a style different to those with which we are familiar. The ‘computer’ on which we are running these computations is a social computer in the sense that many of the elementary functions of the computations it runs are performed by humans, and successful execution of a program often depends on properties of the human society over which the program operates. These sorts of programs are not programmed in a traditional way and may have to be understood in a way that is different from the traditional view of programming. This shift in perspective raises new challenges for the science of the Web and for computing in general.}, author = {Robertson, David and Giunchiglia, Fausto}, doi = {10.1098/rsta.2012.0379}, eprint = {http://rsta.royalsocietypublishing.org/content/371/1987/20120379.full.pdf+html}, interhash = {c671d953e4eb09fc3fe67f93ccd2024c}, intrahash = {a802922683b23455f903551ee2b24b42}, journal = {Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences}, month = mar, number = 1987, title = {Programming the social computer}, url = {http://rsta.royalsocietypublishing.org/content/371/1987/20120379.abstract}, volume = 371, year = 2013 } @article{mone2013beyond, abstract = {The leading open source system for processing big data continues to evolve, but new approaches with added features are on the rise.}, acmid = {2398364}, address = {New York, NY, USA}, author = {Mone, Gregory}, doi = {10.1145/2398356.2398364}, interhash = {b685e2b6666b7e319b3f6730a8843819}, intrahash = {f9b6c86ba3a498ddea95ca0185c49197}, issn = {0001-0782}, issue_date = {January 2013}, journal = {Communications of the ACM}, month = jan, number = 1, numpages = {3}, pages = {22--24}, publisher = {ACM}, title = {Beyond Hadoop}, url = {http://doi.acm.org/10.1145/2398356.2398364}, volume = 56, year = 2013 } @article{wilson2012practices, abstract = {Scientists spend an increasing amount of time building and using software. However, most scientists are never taught how to do this efficiently. As a result, many are unaware of tools and practices that would allow them to write more reliable and maintainable code with less effort. We describe a set of best practices for scientific software development that have solid foundations in research and experience, and that improve scientists' productivity and the reliability of their software.}, author = {Wilson, Greg and Aruliah, D. A. and Brown, C. Titus and Hong, Neil P. Chue and Davis, Matt and Guy, Richard T. and Haddock, Steven H. D. and Huff, Katy and Mitchell, Ian M. and Plumbley, Mark and Waugh, Ben and White, Ethan P. and Wilson, Paul}, interhash = {78f98610c430aa34dc2e161bb8069401}, intrahash = {e28ce8ccadfa439cce3bcdcb5289b499}, journal = {CoRR}, month = oct, title = {Best Practices for Scientific Computing}, url = {http://arxiv.org/abs/1210.0530}, volume = {abs/1210.0530}, year = 2012 } @inproceedings{derose2008building, abstract = {The rapid growth of Web communities has motivated many solutions for building community data portals. These solutions follow roughly two approaches. The first approach (e.g., Libra, Citeseer, Cimple) employs semi-automatic methods to extract and integrate data from a multitude of data sources. The second approach (e.g., Wikipedia, Intellipedia) deploys an initial portal in wiki format, then invites community members to revise and add material. In this paper we consider combining the above two approaches to building community portals. The new hybrid machine-human approach brings significant benefits. It can achieve broader and deeper coverage, provide more incentives for users to contribute, and keep the portal more up-to-date with less user effort. In a sense, it enables building "community wikipedias", backed by an underlying structured database that is continuously updated using automatic techniques. We outline our ideas for the new approach, describe its challenges and opportunities, and provide initial solutions. Finally, we describe a real-world implementation and preliminary experiments that demonstrate the utility of the new approach.}, author = {DeRose, P. and Chai, Xiaoyong and Gao, B.J. and Shen, W. and Doan, An Hai and Bohannon, P. and Zhu, Xiaojin}, booktitle = {24th International Conference on Data Engineering}, doi = {10.1109/ICDE.2008.4497473}, interhash = {00f45357225b1e75ed93bddb8d456fb7}, intrahash = {38a2e84d3dfd845d9c260d5f15161c6f}, month = apr, pages = {646--655}, publisher = {IEEE}, title = {Building Community Wikipedias: A Machine-Human Partnership Approach}, url = {http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=4497473&tag=1}, year = 2008 } @inproceedings{jeffery2008payasyougo, abstract = {A primary challenge to large-scale data integration is creating semantic equivalences between elements from different data sources that correspond to the same real-world entity or concept. Dataspaces propose a pay-as-you-go approach: automated mechanisms such as schema matching and reference reconciliation provide initial correspondences, termed candidate matches, and then user feedback is used to incrementally confirm these matches. The key to this approach is to determine in what order to solicit user feedback for confirming candidate matches.

In this paper, we develop a decision-theoretic framework for ordering candidate matches for user confirmation using the concept of the value of perfect information (VPI). At the core of this concept is a utility function that quantifies the desirability of a given state; thus, we devise a utility function for dataspaces based on query result quality. We show in practice how to efficiently apply VPI in concert with this utility function to order user confirmations. A detailed experimental evaluation on both real and synthetic datasets shows that the ordering of user feedback produced by this VPI-based approach yields a dataspace with a significantly higher utility than a wide range of other ordering strategies. Finally, we outline the design of Roomba, a system that utilizes this decision-theoretic framework to guide a dataspace in soliciting user feedback in a pay-as-you-go manner.}, acmid = {1376701}, address = {New York, NY, USA}, author = {Jeffery, Shawn R. and Franklin, Michael J. and Halevy, Alon Y.}, booktitle = {Proceedings of the 2008 ACM SIGMOD international conference on Management of data}, doi = {10.1145/1376616.1376701}, interhash = {3ceaf563712b776c1ed97a8cb061f63b}, intrahash = {3bff24fb9eb1e39fa97a524aabb8dee9}, isbn = {978-1-60558-102-6}, location = {Vancouver, Canada}, numpages = {14}, pages = {847--860}, publisher = {ACM}, title = {Pay-as-you-go user feedback for dataspace systems}, url = {http://doi.acm.org/10.1145/1376616.1376701}, year = 2008 } @mastersthesis{olson2012cloud, abstract = {My thesis describes the design and implementation of systems that empower individuals to help their communities respond to critical situations and to participate in research that helps them understand and improve their environments. People want to help their communities respond to threats such as earthquakes, wildfires, mudslides and hurricanes, and they want to participate in research that helps them understand and improve their environment. “Citizen Science” projects that facilitate this interaction include projects that monitor climate change, water quality and animal habitats. My thesis explores the design and analysis of community-based sense and response systems that enable individuals to participate in critical community activities and scientific research that monitors their environments.}, author = {Olson, Michael J.}, interhash = {a9cdee464e76cd5210c13d7f66981e83}, intrahash = {d9e22a1a5e9404a805aee5cb0fd406c4}, school = {California Institute of Technology}, title = {Cloud computing for citizen science}, type = {Master's thesis}, url = {http://resolver.caltech.edu/CaltechTHESIS:08232011-122341638}, year = 2012 } @inproceedings{ls_leimeister, address = {Cologne, Germany}, author = {Gierczak, Michael and Söllner, Matthias and Leimeister, Jan Marco}, booktitle = {ConLife 2012 Academic Conference}, interhash = {55d8152a9bc0f44d511c03baadb60426}, intrahash = {36de6094774092f9602aa515dc783599}, note = 339, title = {Untersuchung bestehender Geschäftsmodelle etablierter Cloud Anbieter}, url = {http://pubs.wi-kassel.de/wp-content/uploads/2013/03/JML_360.pdf}, year = 2012 } @article{raykar2010learning, abstract = {For many supervised learning tasks it may be infeasible (or very expensive) to obtain objective and reliable labels. Instead, we can collect subjective (possibly noisy) labels from multiple experts or annotators. In practice, there is a substantial amount of disagreement among the annotators, and hence it is of great practical interest to address conventional supervised learning problems in this scenario. In this paper we describe a probabilistic approach for supervised learning when we have multiple annotators providing (possibly noisy) labels but no absolute gold standard. The proposed algorithm evaluates the different experts and also gives an estimate of the actual hidden labels. Experimental results indicate that the proposed method is superior to the commonly used majority voting baseline.}, acmid = {1859894}, author = {Raykar, Vikas C. and Yu, Shipeng and Zhao, Linda H. and Valadez, Gerardo Hermosillo and Florin, Charles and Bogoni, Luca and Moy, Linda}, interhash = {8113daf47997fddf48e4c6c79f2eba56}, intrahash = {14220abe8babfab01c0cdd5ebd5e4b7c}, issn = {1532-4435}, issue_date = {3/1/2010}, journal = {Journal of Machine Learning Research}, month = aug, numpages = {26}, pages = {1297--1322}, publisher = {JMLR.org}, title = {Learning From Crowds}, url = {http://dl.acm.org/citation.cfm?id=1756006.1859894}, volume = 11, year = 2010 } @incollection{li2011incorporating, abstract = {In scientific cooperation network, ambiguous author names may occur due to the existence of multiple authors with the same name. Users of these networks usually want to know the exact author of a paper, whereas we do not have any unique identifier to distinguish them. In this paper, we focus ourselves on such problem, we propose a new method that incorporates user feedback into the model for name disambiguation of scientific cooperation network. Perceptron is used as the classifier. Two features and a constraint drawn from user feedback are incorporated into the perceptron to enhance the performance of name disambiguation. Specifically, we construct user feedback as a training stream, and refine the perceptron continuously. Experimental results show that the proposed algorithm can learn continuously and significantly outperforms the previous methods without introducing user interactions.}, address = {Berlin/Heidelberg}, affiliation = {Intelligent and Distributed Computing Lab, School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, 430074 P.R. China}, author = {Li, Yuhua and Wen, Aiming and Lin, Quan and Li, Ruixuan and Lu, Zhengding}, booktitle = {Web-Age Information Management}, doi = {10.1007/978-3-642-23535-1_39}, editor = {Wang, Haixun and Li, Shijun and Oyama, Satoshi and Hu, Xiaohua and Qian, Tieyun}, interhash = {3baace12cb4481dcceb53c2d47f413b5}, intrahash = {96f2ae8551126527c2dfe69c8fa22f6c}, isbn = {978-3-642-23534-4}, keyword = {Computer Science}, pages = {454--466}, publisher = {Springer}, series = {Lecture Notes in Computer Science}, title = {Incorporating User Feedback into Name Disambiguation of Scientific Cooperation Network}, url = {http://dx.doi.org/10.1007/978-3-642-23535-1_39}, volume = 6897, year = 2011 } @article{lofi2012information, abstract = {Recent years brought tremendous advancements in the area of automated information extraction. But still, problem scenarios remain where even state-of-the-art algorithms do not provide a satisfying solution. In these cases, another aspiring recent trend can be exploited to achieve the required extraction quality: explicit crowdsourcing of human intelligence tasks. In this paper, we discuss the synergies between information extraction and crowdsourcing. In particular, we methodically identify and classify the challenges and fallacies that arise when combining both approaches. Furthermore, we argue that for harnessing the full potential of either approach, true hybrid techniques must be considered. To demonstrate this point, we showcase such a hybrid technique, which tightly interweaves information extraction with crowdsourcing and machine learning to vastly surpass the abilities of either technique.}, address = {Berlin/Heidelberg}, affiliation = {Institut für Informationssysteme, Technische Universität Braunschweig, Braunschweig, Germany}, author = {Lofi, Christoph and Selke, Joachim and Balke, Wolf-Tilo}, doi = {10.1007/s13222-012-0092-8}, interhash = {941feeaa7bb134e0a5f8b5c0225756b8}, intrahash = {37cc8f1d19105a073544d6594fbbc033}, issn = {1618-2162}, journal = {Datenbank-Spektrum}, keyword = {Computer Science}, number = 2, pages = {109--120}, publisher = {Springer}, title = {Information Extraction Meets Crowdsourcing: A Promising Couple}, url = {http://dx.doi.org/10.1007/s13222-012-0092-8}, volume = 12, year = 2012 } @inproceedings{paton2011feedback, abstract = {User feedback is gaining momentum as a means of addressing the difficulties underlying information integration tasks. It can be used to assist users in building information integration systems and to improve the quality of existing systems, e.g., in dataspaces. Existing proposals in the area are confined to specific integration sub-problems considering a specific kind of feedback sought, in most cases, from a single user. We argue in this paper that, in order to maximize the benefits that can be drawn from user feedback, it should be considered and managed as a first class citizen. Accordingly, we present generic operations that underpin the management of feedback within information integration systems, and that are applicable to feedback of different kinds, potentially supplied by multiple users with different expectations. We present preliminary solutions that can be adopted for realizing such operations, and sketch a research agenda for the information integration community.}, author = {Paton, Norman W. and Fernandes, Alvaro A. A. and Hedeler, Cornelia and Embury, Suzanne M.}, booktitle = {Proceedings of the Conference on Innovative Data Systems Research (CIDR)}, interhash = {1874e5c09919244808457021d2d884d1}, intrahash = {cd75210156615616e4f25c91143040c4}, pages = {175--183}, title = {User Feedback as a First Class Citizen in Information Integration Systems}, url = {http://www.cidrdb.org/cidr2011/Papers/CIDR11_Paper21.pdf}, year = 2011 } @article{dean2008mapreduce, abstract = {MapReduce is a programming model and an associated implementation for processing and generating large datasets that is amenable to a broad variety of real-world tasks. Users specify the computation in terms of a map and a reduce function, and the underlying runtime system automatically parallelizes the computation across large-scale clusters of machines, handles machine failures, and schedules inter-machine communication to make efficient use of the network and disks. Programmers find the system easy to use: more than ten thousand distinct MapReduce programs have been implemented internally at Google over the past four years, and an average of one hundred thousand MapReduce jobs are executed on Google's clusters every day, processing a total of more than twenty petabytes of data per day.}, acmid = {1327492}, address = {New York, NY, USA}, author = {Dean, Jeffrey and Ghemawat, Sanjay}, doi = {10.1145/1327452.1327492}, interhash = {b8a00982bf087c8543855897b7362a04}, intrahash = {bff539224836d703c2d21141985fa1a3}, issn = {0001-0782}, issue_date = {January 2008}, journal = {Communications of the ACM}, month = jan, number = 1, numpages = {7}, pages = {107--113}, publisher = {ACM}, title = {MapReduce: simplified data processing on large clusters}, url = {http://doi.acm.org/10.1145/1327452.1327492}, volume = 51, year = 2008 } @inproceedings{chai2009efficiently, abstract = {Many applications increasingly employ information extraction and integration (IE/II) programs to infer structures from unstructured data. Automatic IE/II are inherently imprecise. Hence such programs often make many IE/II mistakes, and thus can significantly benefit from user feedback. Today, however, there is no good way to automatically provide and process such feedback. When finding an IE/II mistake, users often must alert the developer team (e.g., via email or Web form) about the mistake, and then wait for the team to manually examine the program internals to locate and fix the mistake, a slow, error-prone, and frustrating process.

In this paper we propose a solution for users to directly provide feedback and for IE/II programs to automatically process such feedback. In our solution a developer U uses hlog, a declarative IE/II language, to write an IE/II program P. Next, U writes declarative user feedback rules that specify which parts of P's data (e.g., input, intermediate, or output data) users can edit, and via which user interfaces. Next, the so-augmented program P is executed, then enters a loop of waiting for and incorporating user feedback. Given user feedback F on a data portion of P, we show how to automatically propagate F to the rest of P, and to seamlessly combine F with prior user feedback. We describe the syntax and semantics of hlog, a baseline execution strategy, and then various optimization techniques. Finally, we describe experiments with real-world data that demonstrate the promise of our solution.}, acmid = {1559857}, address = {New York, NY, USA}, author = {Chai, Xiaoyong and Vuong, Ba-Quy and Doan, AnHai and Naughton, Jeffrey F.}, booktitle = {Proceedings of the 35th SIGMOD international conference on Management of data}, doi = {10.1145/1559845.1559857}, interhash = {5860215447e374b059597c0e3864e388}, intrahash = {d6c9fbf442a935dc0618107f8fb54d44}, isbn = {978-1-60558-551-2}, location = {Providence, Rhode Island, USA}, numpages = {14}, pages = {87--100}, publisher = {ACM}, title = {Efficiently incorporating user feedback into information extraction and integration programs}, url = {http://doi.acm.org/10.1145/1559845.1559857}, year = 2009 } @inproceedings{marcus2011crowdsourced, abstract = {Amazon's Mechanical Turk (\MTurk") service allows users to post short tasks (\HITs") that other users can receive a small amount of money for completing. Common tasks on the system include labelling a collection of images, com- bining two sets of images to identify people which appear in both, or extracting sentiment from a corpus of text snippets. Designing a work ow of various kinds of HITs for ltering, aggregating, sorting, and joining data sources together is common, and comes with a set of challenges in optimizing the cost per HIT, the overall time to task completion, and the accuracy of MTurk results. We propose Qurk, a novel query system for managing these work ows, allowing crowd- powered processing of relational databases. We describe a number of query execution and optimization challenges, and discuss some potential solutions.}, author = {Marcus, Adam and Wu, Eugene and Madden, Samuel and Miller, Robert C.}, booktitle = {Proceedings of the 5th Biennial Conference on Innovative Data Systems Research}, doi = {1721.1/62827}, interhash = {b6b7d67c3c09259fb2d5df3f52e24c9d}, intrahash = {29723ba38aa6039091769cd2f69a1514}, month = jan, pages = {211--214}, publisher = {CIDR}, title = {Crowdsourced Databases: Query Processing with People}, url = {http://dspace.mit.edu/handle/1721.1/62827}, year = 2011 } @inproceedings{horowitz2010anatomy, abstract = {We present Aardvark, a social search engine. With Aardvark, users ask a question, either by instant message, email, web input, text message, or voice. Aardvark then routes the question to the person in the user's extended social network most likely to be able to answer that question. As compared to a traditional web search engine, where the challenge lies in finding the right document to satisfy a user's information need, the challenge in a social search engine like Aardvark lies in finding the right person to satisfy a user's information need. Further, while trust in a traditional search engine is based on authority, in a social search engine like Aardvark, trust is based on intimacy. We describe how these considerations inform the architecture, algorithms, and user interface of Aardvark, and how they are reflected in the behavior of Aardvark users.}, acmid = {1772735}, address = {New York, NY, USA}, author = {Horowitz, Damon and Kamvar, Sepandar D.}, booktitle = {Proceedings of the 19th international conference on World wide web}, doi = {10.1145/1772690.1772735}, interhash = {418d79b49ede3a8d15ef5eb8453094f0}, intrahash = {787ecbd5796ada03f15bdda85497e1fd}, isbn = {978-1-60558-799-8}, location = {Raleigh, North Carolina, USA}, numpages = {10}, pages = {431--440}, publisher = {ACM}, title = {The anatomy of a large-scale social search engine}, url = {http://doi.acm.org/10.1145/1772690.1772735}, year = 2010 } @inproceedings{chan2009mathematical, abstract = {Human computation is a technique that makes use of human abilities for computation to solve problems. Social games use the power of the Internet game players to solve human computation problems. In previous works, many social games were proposed and were quite successful, but no formal framework exists for designing social games in general. A formal framework is important because it lists out the design elements of a social game, the characteristics of a human computation problem, and their relationships. With a formal framework, it simplifies the way to design a social game for a specific problem. In this paper, our contributions are: (1) formulate a formal model on social games, (2) analyze the framework and derive some interesting properties based on model's interactions, (3) illustrate how some current social games can be realized with the proposed formal model, and (4) describe how to design a social game for solving a specific problem with the use of the proposed formal model. This paper presents a set of design guidelines derived from the formal model and demonstrates that the model can help to design a social game for solving a specific problem in a formal and structural way.}, author = {Chan, Kam Tong and King, I. and Yuen, Man-Ching}, booktitle = {Proceedings of the International Conference on Computational Science and Engineering, CSE '09}, doi = {10.1109/CSE.2009.166}, interhash = {a54732b662bcb0d763139a38f6525b56}, intrahash = {216d582316e970eb498423ee8448edbe}, month = aug, pages = {1205--1210}, title = {Mathematical Modeling of Social Games}, url = {http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=5283086&tag=1}, volume = 4, year = 2009 }