2008

From WikiPapers
Jump to: navigation, search
<< 2005 - 2006 - 2007 - 2008 - 2009 - 2010 - 2011 >>

This is a list of 3 events celebrated and 765 publications published in 2008.

Events

Name City Country DateThis property is a special property in this wiki.
RecentChangesCamp 2008 Palo Alto United States 9 May 2008
WikiSym 2008 Porto Portugal 8 September 2008
Wikimania 2008 Alexandria Egypt 17 July 2008


Publications

Title Author(s) Keyword(s) Published in Language Abstract R C
"Who is this" quiz dialogue system and users' evaluation Sawaki M.
Minami Y.
Ryuichiro Higashinaka
Kohji Dohsaka
Maeda E.
Interactive systems
User interfaces
2008 IEEE Workshop on Spoken Language Technology, SLT 2008 - Proceedings English In order to design a dialogue system that users enjoy and want to be near for a long time, it is important to know the effect of the system's action on users. This paper describes "Who is this" quiz dialogue system and its users' evaluation. Its quiz-style information presentation has been found effective for educational tasks. In our ongoing effort to make it closer to a conversational partner, we implemented the system as a stuffed-toy (or CG equivalent). Quizzes are automatically generated from Wikipedia articles, rather than from hand-crafted sets of biographical facts. Network mining is utilized to prepare adaptive system responses. Experiments showed the effectiveness of person network and the relationship of user attribute and interest level. 0 0
'Wikivism': From communicative capitalism to organized networks Paul Stacey Politics
Internet
Wiki
Hypertextual
Organization
Postrepresentative
Cultural Poltics English This article examines two different approaches to the political significance of networked technologies like the Internet. It considers Richard Kahn and Douglas Kellner’s “critical/reconstructive” methodology and Jodi Dean’s account of “communicative capitalism,” and shows how the respective approaches are insufficient to elucidate the genuinely radical possibilities we may harbor for the Internet. The case study of “hypertextual databases” or “wikis” is used, both to contextualize the limitations of the above arguments and to present a more radical overture for thinking about network politics. I also utilize Ned Rossiter’s concept of “organized networks” and show how these social-technical forms can provide a more radical proposition for thinking about the political possibilities of wikis. I proceed to translate wikis as specific kinds of organized networks that take us beyond a purely perfunctory language – whether as “information-rich data banks” or else animating the “fantasy of abundance” – and allow us to see them in a decidedly “political” way, as necessarily “incomplete” and thus eminently “rewritable” formations. This essay then concludes by examining the wider implications this “political” reading has for the way in which we understand the multiple situations of nascent forms of democratic politics. 2 1
(Wiki + ResTechs) = (Fresh documentation + Organic knowledge management + Training materials + Good, cheap technical writers) Niki Serakiotou
Aime Diaw
Thomas Bui
Richard Roberts
Carolynne White
Knowledge management
Wiki technology
Proceedings ACM SIGUCCS User Services Conference English Most Information Technology Departments in academia have their historical roots in a culture best described by words such as "geeks," "beta-testers," "troubleshooters," "debuggers," "early adopters," and so on. This culture was partly created by the fact that the mission of academia is indeed to keep looking forward to new developments and cutting-edge technologies. It is the role of industry to adopt the outcome of academia's frenetic efforts to move forward and to then produce commodity-like products. At the same time, in most academic institutions, Information Technology Departments have also evolved into providers of services that use mature but still evolving technology. The consumers (customers) have come to treat these technologies as commodity products or utilities, and they expect to be able to easily learn how to use them, and to be able to easily figure out what are the services offered and how they can be configured. In such an environment, "deployment cycles" and "personnel training" are often seen as a complete waste of time by the academics, as they themselves are pushing for the next generation of technology to be adopted by IT as fast as possible. For these reasons, the creation of detailed documentation by technical writers is usually not deemed worth funding, and with the advent of internet searches and FAQ blogging is seen often as completely unnecessary. This is a true statement for many areas of technology, where there exists ambient literature and FAQs (such as any MS Office product, for example). It is disastrous for customer satisfaction though, if it is applied on issues of configuration that are specific to the institution. Copyright 2008 ACM. 0 0
2Lip: The step towards the web3D Jacek Jankowski
Kruk S.R.
Design
Web3D
Wikipedia
WPF
Proceeding of the 17th International Conference on World Wide Web 2008, WWW'08 English The World Wide Web allows users to create and publish a variety of resources, including multimedia ones. Most of the contemporary best practices for designing web interfaces, however, do not take into account the 3D techniques. In this paper we present a novel approach for designing interactive web applications-2-Layer Interface Paradigm (2LIP). The background layer of the 2LIP-type user interface is a 3D scene, which a user cannot directly interact with. The foreground layer is HTML content. Only taking an action on this content (e.g. pressing a hyperlink, scrolling a page) can affect the 3D scene. We introduce a reference implementation of 2LIP: Copernicus - The Virtual 3D Encyclopedia, which shows one of the potential paths of the evolution of Wikipedia towards Web 3.0. Based on the evaluation of Copernicus we prove that designing web interfaces according to 2LIP provides users a better browsing experience, without harming the interaction. 0 0
A Conceptual Model of Wiki Technology Diffusion A. J. Hester
J. E. Scott
English Wiki technology is an emerging trend making way in organizational environments. Although numerous benefits of using a Wiki in applications of collaborative knowledge creation and sharing have been reported, little research on the adoption and diffusion of Wiki technology has been published. The contribution of this paper is the theoretically informed emphasis on the need to consider a variety of contextual factors influencing Wiki technology diffusion. Implementations of Wiki technology should include careful consideration of organizational culture, as well as user perceptions of Wiki organizational compatibility, relative advantage, and complexity. Efforts should also be made to achieve and maintain a critical mass of Wiki users. 0 0
A Geospatial Wiki for m-Learning Christian Safran
Bilal Zaka
E-Learning
M-learning
Geotagging
Geowiki
CSSE English 0 0
A Method for Measuring Co-authorship Relationships in MediaWiki Libby Veng-Sam Tang
Robert P. Biuk-Aghai
Simon Fong
Analysis
Co-authorship
Wiki
WikiSym English Collaborative writing through wikis has become increasingly popular in recent years. When users contribute to a wiki article they implicitly establish a co-authorship relationship. Discovering these relationships can be of value, for example in finding experts on a given topic. However, it is not trivial to determine the main co-authors for a given author among the potentially thousands who have contributed to a given author’s edit history. We have developed a method and algorithm for calculating a co-authorship degree for a given pair of authors. We have implemented this method as an extension for the MediaWiki system and demonstrate its performance which is satisfactory in the majority of cases. This paper also presents a method of determining an expertise group for a chosen topic. 5 2
A New Internet A New Year M. Castelluccio Strategic Finance A wiki, according to the guy who invented them, is the simplest online database that could possibly work. Ward Cunningham launched his first wiki in 1995, and the format has been widely adopted since by academics, artists, hackers, and business professionals. The most famous wiki is Wikipedia, the online encyclopedia. Like other wikis, Wikipedia has an open editing system where the readers are the contributing editors and proofreaders. The readers write the articles. One of the problems with defining wikis is that the word, which actually means quick" in Hawaiian can refer to the software the community or the database. The community can be seen and operated as an intranet or a common workspace for collaborators. The reality is a little amorphous so why not go to the wiki {(Wikipedia)} for their take on it -- they should know." 0 0
A Search Engine for Browsing the Wikipedia Thesaurus Kotaro Nakayama
Takahiro Hara
Sojiro Nishio
Data mining
Association Thesaurus
Link Structure Analysis
XML Web Services
13th International Conference on Database Systems for Advanced Applications, Demo session (DASFAA) Wikipedia has become a huge phenomenon on the WWW. As a corpus for knowledge extraction, it has various impressive characteristics such as a huge amount of articles, live updates, a dense link structure, brief link texts and URL identification for concepts. In our previous work, we proposed link structure mining algorithms to extract a huge scale and accurate association thesaurus from Wikipedia. The association thesaurus covers almost 1.3 million concepts and the significant accuracy is proved in detailed experiments. To prove its practicality, we implemented three features on the association thesaurus; a search engine for browsing Wikipedia Thesaurus, an XML Web service for the thesaurus and a Semantic Web support feature. We show these features in this demonstration. 0 0
A Self-Adaptive Explicit Semantic Analysis Method for Computing Semantic Relatedness Using Wikipedia Weiping Wang
Peng Chen
Bowen Liu
Semantic relatedness
Wikipedia
Explicit semantic analysis
Self-adaptive
FITME English 0 0
A Wiki Service for Investment Knowledge Management C. F. So
K. S. Wong
K. F. Fung
S. Y. Ho
C. L. Lai
Raymond Y. K. Lau
Web Services
Pervasive Computing
Knowledge management
Financial Investment
RSS Feeding
Wiki Service
ICEBE English 0 0
A Wiki for Collaboration and Publication in Research Christoph von Hinten
Andreas Hense
Matthias Razum
Wiki
Digital object repository
Persistent identifier
Collaboration
List of publications
Fedora
ESCIENCE English 0 0
A Wiki for classroom writing Reading Teacher English 0 1
A Wiki that knows where it is being used: insights from potential users Maria Plummer
Linda Plotnick
Starr R. Hiltz
Quentin Jones
SIGMIS Database English With the advent of extensive wireless networks that blanket physically compact urban enclaves such as office complexes, shopping centers, or university campuses, it is possible to create software applications that provide location-based mobile/online services. One such application is CampusWiki, which integrates location information into a wiki structure. In the design science research reported in this paper, we employed a form of "action research" in which we engaged users as participants in an iterative process of designing and evaluating CampusWiki. Two qualitative studies were undertaken early in the design process in which semi-structured interviews were used to assess potential users' reactions to CampusWiki. Through this research, the designers were able to assess whether their intentions matched the mental models of potential users of the application. The results showed that although many of the perceived benefits were as designed by the developers, a misunderstanding of the location-aware feature led users to unanticipated concerns and expectations. These findings are important in guiding designers and implementers on the desirable and possibly undesirable features of such systems. 0 0
A Wikipedia-Based Multilingual Retrieval Model Martin Potthast
Benno Stein
Maik Anderka
30th European Conference on IR Research (ECIR 08) English This paper introduces CL-ESA, a new multilingual retrieval model for the analysis of cross-language similarity. The retrieval model exploits the multilingual alignment of Wikipedia: given a document d written in language L we construct a concept vector d for d, where each dimension i in d quantifies the similarity of d with respect to a document chosen from the “L-subset” of Wikipedia. Likewise, for a second document d′ written in language L′, , we construct a concept vector d′, using from the L′-subset of the Wikipedia the topic-aligned counterparts of our previously chosen documents. Since the two concept vectors d and d′ are collection-relative representations of d and d′ they are language-independent. I. e., their similarity can directly be computed with the cosine similarity measure, for instance. We present results of an extensive analysis that demonstrates the power of this new retrieval model: for a query document d the topically most similar documents from a corpus in another language are properly ranked. Salient property of the new retrieval model is its robustness with respect to both the size and the quality of the index document collection. 0 0
A bilingual dictionary extracted from the Wikipedia link structure Maike Erdmann
Kotaro Nakayama
Takahiro Hara
Shojiro Nishio
DASFAA English 0 0
A bliki model to support political discourse formation Lourenco R.P. Bliki
Blogs
Deliberative democracy
Discourse formation
Public participation
Wiki
WikiSym 2008 - The 4th International Symposium on Wikis, Proceedings English This paper presents a bliki model which combines the potential of a blog and a wiki in order to support the creation of a constellation of political discourses in the public sphere and promote critical reflection among them. The model is inspired on traditional local public participation events that usually take the form of a face-to-face meeting where participants discuss some issue or problem and, in the end, produce a document summarizing the highlights of the event. The model embraces the collaborative paradigm fostered by the Internet and tries to support the creation of a set of documents (using a wiki), each one representing a particular discourse, whose content is generated from the discussion held in a blog. It is based on the identification of relevant ideas expressed in the blog, which are then sorted according to the point of view (discourse) they support, and are incorporated into the wiki document representing that discourse by those participants supporting it. 0 0
A bush encroachment decision support system's metamorphosis Winschiers-Theophilus H.
Fendler J.
Stanley C.
Joubert D.
Zimmermann I.
Mukumbira S.
Community value design
Cross-cultural design
Cross-cultural usability
Cultural appropriation
Decision support system
Namibia
Recommendation support system
Southern Africa
Wiki
Proceedings of the 20th Australasian Conference on Computer-Human Interaction: Designing for Habitus and Habitat, OZCHI'08 English Since the inception of our bush-encroachment decision support system, we have gone through many cycles of adaptations while striving towards what we believed to be a usable system. A fundamental difference between community based users and individualistic users necessitates a change in the design and evaluation methods as well as a community agreement of concepts and values guiding the design. In this paper we share the lessons learned along the story depicting the metamorphosis of a bush encroachment decision support system in Southern African rangelands. Above and beyond community members participating in the design and evaluation of the system, they establish the community grounded values determining the system's quality concepts such as usability. 0 0
A collaborative multilingual database project on aymara implemented in peru and bolivia Beck H.
Legg S.
Hardman M.J.
Lord G.
Llanque-Chana J.
Lowe E.
Collaboration
Linguistics database
Web development
Wiki
American Society of Agricultural and Biological Engineers Annual International Meeting 2008, ASABE 2008 English A web-based collaborative environment including on-line authoring tools that is managed by a central database was developed in collaboration with several countries including Peru, Bolivia, and the United States. The application involved developing a linguistics database and eLearning environment for documenting, preserving, and promoting language training for Aymara, a language indigenous to Peru and Bolivia. The database, an ontology management system called Lyra, incorporates all elements of the language (dialogues, phrase patterns, phrases, words, and morphemes) as well as cultural multimedia resources (images and sound recordings). The organization of the database enables a high level of integration among language elements and cultural resources. Authoring tools are used by experts in the Aymara language to build the linguistic database. These tools are accessible on-line as part of the collaborative environment using standard web browsers incorporating the Java plug-in. The eLearning student interface is a web-based program written in Flash. The Flash program automatically interprets and formats data objects retrieved from the database in XML format. The student interface is presented in Spanish and English. A web service architecture is used to publish the database on-line so that it can be accessed and utilized by other application programs in a variety of formats. 0 0
A comparison of World Wide Web resources for identifying medical information P.T. Johnson
J.K. Chen
J. Eng
M.A. Makary
E.K. Fishman
Academic Radiology The objective is to compare the utility of a search engine, Google, with other medical and non-medical, web-based resources for identifying specific medical {information.This} institutional review board-approved case cross-over study randomly assigned 89 medical student volunteers to use either Google or any other web-based resource (excluding Google) to research 10 advanced medical questions in a multiple choice exam. Primary outcome measures were resource efficiency (inversely related to number of links used to identify the correct answer for each question) and correctness (number of correct answers/total number of questions answered). For Google searches, the sites providing the information in question were also {evaluated.The} most frequently selected {non-Google} resources were Yahoo (n = 531), Ask.com (n = 110), and the interactive encyclopedia Wikipedia.com (n = 74). Google was more efficient than all other resources (1.50 vs. 1.94 mean links, P .0001), with no significant difference in correctness (97\% 756/780 vs. 96\% 747/780, P = .16). After a Google search, the four most common categories of sites that provided the correct answer were dictionary/encyclopedia sites, medical websites, National Library of Medicine resources, or journal websites. Yahoo was less efficient than Google (1.90 vs. 1.54 mean links, P .0001). However, {non-Google} search engines were more efficient than web sites (eg, Wikipedia, medical websites) and {PubMed} (1.87 vs. 2.54 mean links, P = {.0004).Google} is an efficient web resource for identifying specific medical information, by guiding users to an array of medical {resources.All} rights reserved Elsevier. 0 1
A conversation with Jimmy Wales G Morse HARVARD BUSINESS REVIEW The founder of Wikipedia analyzes why wikis are becoming popular tools for sharing knowledge in the workplace. He encourages managers to provide institutional support for these highly practical forums but to be judicious about direct participation. 0 0
A gene wiki for community annotation of gene function Jon W. Huss
Camilo Orozco
James Goodale
Chunlei Wu
Serge Batalov
Tim J. Vickers
Faramarz Valafar
Andrew I. Su
English This manuscript describes the creation of comprehensive gene wiki, seeded with data from public domain sources, which will enable and encourage community annotation of gene function. 0 5
A generic method for multi word extraction from wikipedia Bekavac B.
Tadic M.
Croatian
Multi word expressions
Multi word extraction
Wikipedia
Proceedings of the International Conference on Information Technology Interfaces, ITI English This paper presents the generic method for multiword expression extraction from Wikipedia. The method is using the properties of this specific encyclopedic genre in its HTML format and it relies on the intention of the authors of articles to link to other articles. The relevant links were processed by applying local regular grammars within the NooJ development environment. We tested the method on a Croatian version of Wikipedia and we present the results obtained. 0 0
A grammar for standardized Wiki markup Martin Junghans
Dirk Riehle
Rama Gurram
Matthias Kaiser
Mário Lopes
Umit Yalcinalp
WikiCreole
Wiki engine performance
Wiki markup
Wiki markup grammar
Wiki markup parser
Wiki markup standard
Wiki markup test cases
Wiki
WikiSym 2008 - The 4th International Symposium on Wikis, Proceedings English Today's wiki engines are not interoperable. The rendering engine is tied to the processing tools which are tied to the wiki editors. This is an unfortunate consequence of the lack of rigorously specified standards. This paper discusses an EBNF-based grammar for Wiki Creole 1.0, a community standard for wiki markup, and demonstrates its benefits. Wiki Creole is being specified using prose, so our grammar revealed several categories of ambiguities, showing the value of a more formal approach to wiki markup specification. The formalization of Wiki Creole using a grammar shows performance problems that today's regular-expression-based wiki parsers might face when scaling up. We present an implementation of a wiki markup parser and demonstrate our test cases for validating Wiki Creole parsers. We view the work presented in this paper as an important step towards decoupling wiki rendering engines from processing tools and from editing tools by means of a precise and complete wiki markup specification. This decoupling layer will then allow innovation on these different parts to proceed independently and as is expected at a faster pace than before. 0 0
A language-independent approach to keyphrase extraction and evaluation Paukkeri M.-S.
Nieminen I.T.
Polla M.
Timo Honkela
Coling 2008 - 22nd International Conference on Computational Linguistics, Proceedings of the Conference English We present Likey, a language-independent keyphrase extraction method based on statistical analysis and the use of a reference corpus. Likey has a very light-weight preprocessing phase and no parameters to be tuned. Thus, it is not restricted to any single language or language family. We test Likey having exactly the same configuration with 11 European languages. Furthermore, we present an automatic evaluation method based on Wikipedia intra-linking. © 2008. Licensed under the Creative Commons. 0 0
A lexical approach for Spanish question answering Tellez A.
Juarez A.
Hernandez G.
Denicia C.
Villatoro E.
Montes M.
Villasenor L.
Lecture Notes in Computer Science English This paper discusses our system's results at the Spanish Question Answering task of CLEF 2007. Our system is centered in a full data-driven approach that combines information retrieval and machine learning techniques. It mainly relies on the use of lexical information and avoids any complex language processing procedure. Evaluation results indicate that this approach is very effective for answering definition questions from Wikipedia. In contrast, they also reveal that it is very difficult to respond factoid questions from this resource solely based on the use of lexical overlaps and redundancy. 0 0
A model for Ranking entities and its application to Wikipedia Gianluca Demartini
Firan C.S.
Tereza Iofciu
Ralf Krestel
Wolfgang Nejdl
Proceedings of the Latin American Web Conference, LA-WEB 2008 English Entity Ranking (ER) is a recently emerging search task in Information Retrieval, where the goal is not finding documents matching the query words, but instead finding entities which match types and attributes mentioned in the query. In this paper we propose a formal model to define entities as well as a complete ER system, providing examples of its application to enterprise, Web, and Wikipedia scenarios. Since searching for entities on Web scale repositories is an open challenge as the effectiveness of ranking is usually not satisfactory, we present a set of algorithms based on our model and evaluate their retrieval effectiveness. The results show that combining simple Link Analysis, Natural Language Processing, and Named Entity Recognition methods improves retrieval performance of entity search by over 53% for P@ 10 and 35% for MAP. 0 0
A new, dynamic resource for both students and teachers of family medicine: the family medicine research wiki at FMDRL Caroline R. Richardson Ann Fam Med English 0 0
A nonlinear representation of page history in P2P wiki system IFIP International Federation for Information Processing English 0 0
A pervasive wiki application based on VoiceXML Constantinos Kolias
Vassilis Kolias
Ioannis Anagnostopoulos
Georgios Kambourakis
Eleftherios Kayafas
VoiceXML
Audio information system
Information system
Wiki
PETRA English 0 0
A pitfall of wiki solution for biological databases Masanori Arita English Not a few biologists tend to consider wiki as a solution to manage and reorganize data by a community. However, in its basic functionality, wiki lacks a measure to check data consistency and is not suitable for a database. To circumvent this pitfall, installation of page dependency through in-line page searches is necessary. We also introduce two existing approaches that support in-line queries. 0 0
A qualitative analysis on collaborative learning experience of student journalists using Wiki Ma W.W.K.
Yuen A.H.K.
Journalistic writing processes
Revision
Social interaction
Wiki
Lecture Notes in Computer Science English Education in journalism emphasizes internships, apprenticeships, and other opportunities to learn journalism by doing journalism; however, most computer-mediated communication tools do not have such a provision. The fully open structure of Wiki matches the principles of learning journalism while, from a technical point of view, Wiki provides a very easy way for users to report, write and edit. In a case study, a group of undergraduate journalism students were exposed to a student-written Wiki to jointly compose news reporting. Analysis of student journalists' responses to the open-ended questions revealed revision as the core processing capability of Wiki. The motivational factors to revision include accuracy (fact checking), story enrichment, and personal interest toward the news topic. In addition, learners are also affected by the social interactions among the community users within Wiki. The qualitative data shows students both value the process and face challenges in managing the complexity of shared editing. 0 0
A resource-poor approach for linking ontology classes to Wikipedia articles Nils Reiter
Matthias Hartung
Anette Frank
STEP English 0 0
A search engine for browsing the Wikipedia thesaurus Kotaro Nakayama
Takahiro Hara
Shojiro Nishio
DASFAA English 0 0
A semantic wiki on cooperation in public administration Krabina B. Intercommunal cooperation
Public administration
Semantic wiki
CEUR Workshop Proceedings English Authorities cooperate in various ways. The Web portal www. verwaltungskooperation.at aims to share knowledge on collaboration projects. A semantic wiki approach was used to facilitate best practice documentation with Semantic Web and Web 2.0 technology. 0 0
A study of ontology convergence in a Semantic Wiki Chrysovalanto Kousetti
Millard D.E.
Yvonne Howard
Emergent ontologies
Semantic web
Semantic wiki
WikiSym 2008 - The 4th International Symposium on Wikis, Proceedings English Semantic Wikis propose a combination of both easy collaboration and semantic expressivity; characteristics of the WikiWikiWeb and the Semantic Web respectively. In this paper we look to define and analyse the Semantic Wiki method, in order to explore the effect of different Semantic Wiki characteristics on the quality of the semantic networks authored within them. We look at a number of different Semantic Wiki implementations, including their semantic expressivity and usability. We focus on support for ontology creation, and perform an evaluation on the effect of type suggestion tools on ontology convergence within a seeded and unseeded Wiki (using Semantic MediaWiki and our own MOCA extension). We find that seeding a Wiki with typed pages and links has a strong effect on the quality of the emerging structure and that convergence tools have the potential to replicate that effect with an unseeded Wiki, but that they have limited impact on the reuse of elements from the evolving ontology. 0 0
A study on multi-agent spatial database update mechanism based on Wiki idea Proceedings of SPIE - The International Society for Optical Engineering English 0 0
A systemic and cognitive view on collaborative knowledge building with wikis Ulrike Cress
Joachim Kimmerle
International Journal of Computer-Supported Collaborative Learning 0 0
A visual-analytic toolkit for dynamic interaction graphs Yang X.
Asur S.
Parthasarathy S.
Mehta S.
Dynamic interaction networks
Graph visualization
Visual analytics
Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining English In this article we describe a visual-analytic tool for the interrogation of evolving interaction network data such as those found in social, bibliometric, WWW and biological applications. The tool we have developed incorporates common visualization paradigms such as zooming, coarsening and filtering while naturally integrating information extracted by a previously described event-driven framework for characterizing the evolution of such networks. The visual front-end provides features that are specifically useful in the analysis of interaction networks, capturing the dynamic nature of both individual entities as well as interactions among them. The tool provides the user with the option of selecting multiple views, designed to capture different aspects of the evolving graph from the perspective of a node, a community or a subset of nodes of interest. Standard visual templates and cues are used to highlight critical changes that have occurred during the evolution of the network. A key challenge we address in this work is that of scalability - handling large graphs both in terms of the efficiency of the back-end, and in terms of the efficiency of the visual layout and rendering. Two case studies based on bibliometric and Wikipedia data are presented to demonstrate the utility of the toolkit for visual knowledge discovery. 0 0
A web 2.0 based computer knowledge learning platform Yuanxin O.
Chao L.
Huan L.
Pingan Z.
Xiaodan Zhang
Blogs
Learning platform
RSS
Web 2.0
Wiki
Proceedings - International Conference on Computer Science and Software Engineering, CSSE 2008 English Traditional web-based online learning systems usually focus on the dispatch of knowledge, and lack of ways for students to get involved. Introduction to Computer Basics (ICB) is one of the first professional courses for freshmen majored in computer science, as well as information technology. To make the learning platform of ICB more helpful, a Web 2.0 based computer knowledge learning platform is presented, which changes the focus from course content to the students participation. Web2.0 elements including personal and group spaces, wiki cyclopedia, interest mining and personalized recommendation, and RSS resource subscription are integrated. The platform has been put into use already, and got satisfaction from both teachers and students. 0 0
A web 2.0 case tool supporting packaged software implementation Wu H.
Arnaout G.
Case-based reasoning
Packaged Software Implementation
Web 2.0
Wiki
ICEIS 2008 - Proceedings of the 10th International Conference on Enterprise Information Systems English Companies increasingly rely on implementing packaged software instead of developing custom solutions. Packaged software implementation (PSI) is the process to solve business problems by customizing and integrating an off-the-shelf software package. However, there has been a lack of CASE (Computer Aided Software Engineering) tools to support PSI. This paper presents a Web 2.0 based tool that supports casebased reasoning in PSI. The tool helps users explore past design cases, find a similar case, and reuse the design for that case in new problem situations. Our belief is that by utilizing the social power of a large group of users, better designs can be achieved at lower risks and lower costs. 0 0
A wiki for the life sciences where authorship matters Robert Hoffmann English WikiGenes is the first wiki system to combine the collaborative and largely altruistic possibilities of wikis with explicit authorship. In view of the extraordinary success of Wikipedia there remains no doubt about the potential of collaborative publishing, yet its adoption in science has been limited. Here I discuss a dynamic collaborative knowledge base for the life sciences that provides authors with due credit and that can evolve via continual revision and traditional peer review into a rigorous scientific tool. 0 3
A wiki instance in the enterprise: Opportunities, concerns and reality Catalina Danis
David Singer
Enterprise
Organization
Wiki
Work practice
English We describe the design and deployment of a wiki-based application that supports yearly planning work by members of a globally distributed, 900-member research organization. The goal of the ResearchWiki is to enable greater transparency in the work of the organization and to provide a basis for broader collaboration among researchers. We motivate the design based on results from 40 interviews and discuss two cycles of usage of the ResearchWiki based on analyses of logs and an additional 20 interviews. We conclude with a discussion of the interplay of technology, work practice and organizational change and with design recommendations. Copyright 2008 ACM. 0 0
A wiki way of working Internet Reference Services Quarterly English 0 0
A wiki-based approach to method tailoring Fredrik Karlsson Dynamic method tailoring
Evolutionary method engineering
Method components
Method rationale
Wiki
ICPW English 0 0
AMYPdb: A database dedicated to amyloid precursor proteins Pawlicki S.
Le Bechec A.
Delamarche C.
BMC Bioinformatics English Background: Misfolding and aggregation of proteins into ordered fibrillar structures is associated with a number of severe pathologies, including Alzheimer's disease, prion diseases, and type II diabetes. The rapid accumulation of knowledge about the sequences and structures of these proteins allows using of in silico methods to investigate the molecular mechanisms of their abnormal conformational changes and assembly. However, such an approach requires the collection of accurate data, which are inconveniently dispersed among several generalist databases. Results: We therefore created a free online knowledge database (AMYPdb) dedicated to amyloid precursor proteins and we have performed large scale sequence analysis of the included data. Currently, AMYPdb integrates data on 31 families, including 1,705 proteins from nearly 600 organisms. It displays links to more than 2,300 bibliographic references and 1,200 3D-structures. A Wiki system is available to insert data into the database, providing a sharing and collaboration environment. We generated and analyzed 3,621 amino acid sequence patterns, reporting highly specific patterns for each amyloid family, along with patterns likely to be involved in protein misfolding and aggregation. Conclusion: AMYPdb is a comprehensive online database aiming at the centralization of bioinformatic data regarding all amyloid proteins and their precursors. Our sequence pattern discovery and analysis approach unveiled protein regions of significant interest. AMYPdb is freely accessible [1]. 0 0
Accuracy estimate and optimization techniques for simrank computation Dmitry Lizorkin
Maxim Grinev
Pavel Velikhov
Denis Turdakov
Proceedings of the VLDB Endowment English The measure of similarity between objects is a very useful tool in many areas of computer science, including informa-tion retrieval. SimRank is a simple and intuitive measure of this kind, based on graph-theoretic model. SimRank is typ-ically computed iteratively, in the spirit of PageRank. How-ever, existing work on SimRank lacks accuracy estimation of iterative computation and has discouraging time complexity. In this paper we present a technique to estimate the ac-curacy of computing SimRank iteratively. This technique provides a way to find out the number of iterations required to achieve a desired accuracy when computing SimRank. We also present optimization techniques that improve the com-putational complexity of the iterative algorithm from O(n 4) to O(n 3) in the worst case. We also introduce a threshold sieving heuristic and its accuracy estimation that further improves the efficiency of the method. As a practical illustration of our techniques we computed SimRank scores on a subset of English Wikipedia corpus, consisting of the complete set of articles and category links. 0 0
AceWiki: Collaborative ontology management in controlled natural language Kuhn T. CEUR Workshop Proceedings English AceWiki is a prototype that shows how a semantic wiki using controlled natural language - Attempto Controlled English (ACE) in our case - can make ontology management easy for everybody. Sentences in ACE can automatically be translated into first-order logic, OWL, or SWRL. AceWiki integrates the OWL reasoner Pellet and ensures that the ontology is always consistent. Previous results have shown that people with no background in logic are able to add formal knowledge to AceWiki without being instructed or trained in advance. 0 0
Acquiring a Taxonomy from the German Wikipedia Laura Kassner
Vivi Nastase
Michael Strube
English 0 0
Action research as a congruent methodology for understanding wikis: the case of Wikiversity C. Lawler Journal of Interactive Media in Education 0 0
Active knowledge generation by university students through cooperative learning Arribillaga E. Collaborative tools
Educational Technology
Human Capital Development.
2008 ITI 6th International Conference on Information and Communications Technology, ICICT 2008 English Social and cultural transformations caused by the globalisation have fostered changes in current universities, institutions which, doing an intensive and responsible use of technologies,have to create a continuous improvement-based pedagogical model consisting on communities.To this end, we propose here the adoption of the so-called hacker ethic, which highlights the importance of collaborative, passionate, creative as well as socially-valuable work. Applying this ethic to higher education, current universities may become Net-Academy-based universities.Therefore, these institutions require a new digital culture that allow the transmission of hacker ethic's values and, in turn, a Net-Academy-based learning model that enable students transform into knowledge generators. In this way, wikitechnology-based systems may help universities to achieve the transformation they need. We present here an experiment to check whether these kind of resources transmit to the students the values of the hacker ethic allowing them to become active knowledge generators. This experiment revealed the problems of such technologies with the limits of the scope of the community created and the non-so-active knowledge-generator role of the students. Against these shortcomings, we address here a Wikipedia-based methodology and discuss the possibilities of this alternative to help current universities upgrade into Net-Academy-based universities. 0 0
Ad-hoc knowledge engineering with semantic knowledge wikis Jochen Reutelshoefer
Joachim Baumeister
Frank Puppe
CEUR Workshop Proceedings English A couple of semantic wikis have been proposed to serve as collaborative knowledge engineering environments - however, the knowledge of almost all systems is currently based on the expressiveness of OWL (Lite/DL). In this paper we present the concept of a semantic knowledge wiki that extends the capabilities of semantic wikis by strong problem-solving methods. We show how problem-solving knowledge is connected with standard ontological knowledge using an upper ontology. Furthermore, we discuss different possibilities to formalize problemsolving knowledge, for example by semantic knowledge annotation, structured text and explicit markups. 0 0
Adaptive indexing for content-based search in P2P systems Aoying Zhou
Rong Zhang
Weining Qian
Quang Hieu Vu
Tianming Hu
Data and Knowledge Engineering One of the major challenges in {Peer-to-Peer} {(P2P)} file sharing systems is to support content-based search. Although there have been some proposals to address this challenge, they share the same weakness of using either servers or super-peers to keep global knowledge, which is required to identify importance of terms to avoid popular terms in query processing. As a result, they are not scalable and are prone to the bottleneck problem, which is caused by the high visiting load at the global knowledge maintainers. To that end, in this paper, we propose a novel adaptive indexing approach for content-based search in {P2P} systems, which can identify importance of terms without keeping global knowledge. Our method is based on an adaptive indexing structure that combines a Chord ring and a balanced tree. The tree is used to aggregate and classify terms adaptively, while the Chord ring is used to index terms of nodes in the tree. Specifically, at each node of the tree, the system classifies terms as either important or unimportant. Important terms, which can distinguish the node from its neighbor nodes, are indexed in the Chord ring. On the other hand, unimportant terms, which are either popular or rare terms, are aggregated to higher level nodes. Such classification enables the system to process queries on the fly without the need for global knowledge. Besides, compared to the methods that index terms separately, term aggregation reduces the indexing cost significantly. Taking advantage of the tree structure, we also develop an efficient search algorithm to tackle the bottleneck problem near the root. Finally, our extensive experiments on both benchmark and Wikipedia datasets validated the effectiveness and efficiency of the proposed method. 2008. 0 0
Agent simulation of collaborative knowledge processing in Wikipedia Jinsheng Xu
Levent Yilmaz
Jinghua Zhang
Wikipedia
Agent simulation
Collaborative knowledge processing
User innovation community
SpringSim English 0 1
Aggregate documents: Making sense of a patchwork of topical documents Shilman M. Aggregation
Analysis
Information retrieval
Summarization
DocEng'08 - Proceedings of the 8th ACM Symposium on Document Engineering English With the dramatic increase in quantity and diversity of online content, particularly in the form of user generated content, we now have access to unprecedented amounts of information. Whether you are researching the purchase of a new cell phone, planning a vacation, or trying to assess a political candidate, there are now countless resources at your fingertips. However, finding and making sense of all this information is laborious and it is difficult to assess high-level trends in what is said. Web sites like Wikipedia and Digg democratize the process of organizing the information from countless document into a single source where it is somewhat easier to understand what is important and interesting. In this talk, I describe a complementary set of automated alternatives to these approaches, demonstrate these approaches with a working example, the commercial web site Wize.com, and derive some basic principles for aggregating a diverse set of documents into a coherent and useful summary. 0 0
Agile documentation with uScrum Joaquim Baptista Agile documentation
Agile management
Complexity measures
Orders of ignorance
Risk
Uncertainty
UScrum
Wiki
Writing
SIGDOC 2008 - Proceedings of the 26th ACM International Conference on Design of Communication English uScrum (uncertainty Scrum) is an agile process developed by a small team at Altitude Software to manage the process of writing user documentation. uScrum manages uncertainty and the unknown, allowing writers to quickly react to changing conditions. uScrum uses orders of ignorance to understand the difficulty of tasks, allowing the team to effectively prioritize regular work together with difficult creative work. uScrum overbooks writers on iterative cycles called sprints, then lets the writers micro-manage their tasks to overcome obstacles. After each sprint the team decides what to publish and whether to proceed with unfinished work. Copyright 2008 ACM. 0 0
Agile learning & teaching with wikis: Building a pattern Marija Cubric Agile development
Blended learning
Wiki
WikiSym 2008 - The 4th International Symposium on Wikis, Proceedings English In this paper, we describe a blended learning process that is based on the agile development principles [1] and techniques [2] and supported by wikis. The process presented here is applicable to any learning and teaching scenario where the module1 objectives require development of non-cognitive skills such as: teamwork, giving and receiving feedback, collaborative writing, incremental and iterative development etc We believe that understanding agile principles and being able to apply agile techniques to any development activity are highly relevant skills that are important for the employability of our students. 0 0
Altruism, selfishness, and destructiveness on the social web John Riedl Lecture Notes in Computer Science English Many online communities are emerging that, like Wikipedia, bring people together to build community-maintained artifacts of lasting value (CALVs). What is the nature of people's participation in building these repositories? What are their motives? In what ways is their behavior destructive instead of constructive? Motivating people to contribute is a key problem because the quantity and quality of contributions ultimately determine a CALV's value. We pose three related research questions: 1) How does intelligent task routing-matching people with work-affect the quantity of contributions? 2) How does reviewing contributions before accepting them affect the quality of contributions? 3) How do recommender systems affect the evolution of a shared tagging vocabulary among the contributors? We will explore these questions in the context of existing CALVs, including Wikipedia, Facebook, and MovieLens. 0 0
An Analysis of Topical Coverage of Wikipedia Alexander Halavais
Derek Lackaff
English Many have questioned the reliability and accuracy of Wikipedia. Here a different issue, but one closely related: how broad is the coverage of Wikipedia? Differences in the interests and attention of Wikipedia’s editors mean that some areas, in the traditional sciences, for example, are better covered than others. Two approaches to measuring this coverage are presented. The first maps the distribution of topics on Wikipedia to the distribution of books published. The second compares the distribution of topics in three established, field-specific academic encyclopedias to the articles found in Wikipedia. Unlike the top-down construction of traditional encyclopedias, Wikipedia’s topical coverage is driven by the interests of its users, and as a result, the reliability and completeness of Wikipedia is likely to be different depending on the subject-area of the article. 0 8
An Analysis of Wikipedia M. Rahman JITTA : Journal of Information Technology Theory and Application Wikipedia is defined by its founders as the free encyclopedia that anyone can edit." This property we argue makes Wikipedia a public good and hence subject to under-provision. A puzzling feature of Wikipedia however is its enormous size at roughly seven times that of its commercial counterparts. What is driving this growth? And how can we assess the reliability of this giant encyclopedia arising solely from free-editing? We model contribution to Wikipedia and its reliability. We demonstrate that Wikipedia is indeed subject to free-riding and offer a novel explanation for the mitigation of under-provision under such circumstances. We also find that the public-good feature of Wikipedia and free-riding introduce a lower-bound in the quality of Wikipedia. This finding is consistent with a previous empirical study that established Wikipedia's surprisingly high level of quality. We identify Wikipedia as part of a general Internet phenomenon that we call the Collaborative Net and that includes features such as citizen journalism and online reviews. {PUBLICATION} {ABSTRACT}" 0 1
An Empirical Research on Extracting Relations from Wikipedia Text Jin-Xia Huang
Pum-Mo Ryu
Key-Sun Choi
Information extraction
Feature-based
Relatedness information
Relation classification
IDEAL English A feature based relation classification approach is presented, in which probabilistic and semantic relatedness features between patterns and relation types are employed with other linguistic information. The importance of each feature set is evaluated with Chi-square estimator, and the experiments show that, the relatedness features have big impact on the relation classification performance. A series experiments are also performed to evaluate the different machine learning approaches on relation classification, among which Bayesian outperformed other approaches including Support Vector Machine (SVM). 0 0
An Unsupervised Approach to Biography Production using Wikipedia Fadi Biadsy
Julia Hirschberg
Elena Filatova
ACL-08: HLT, pp. 807-815 We describe an unsupervised approach to multi-document sentence-extraction based summarization for the task of producing biographies. We utilize Wikipedia to automatically construct a corpus of biographical sentences and TDT4 to construct a corpus of non-biographical sentences. We build a biographical-sentence classifier from these corpora and an SVM regression model for sentence ordering from the Wikipedia corpus. We evaluate our work on the DUC2004 evaluation data and with human judges. Overall, our system significantly outperforms all systems that participated in DUC2004, according to the ROUGE-L metric, and is preferred by human subjects. 0 0
An activity theoretic model for information quality change Stvilia
B.
Gasser
L.
Wikipedia
Activity Theory
Information quality
First Monday, 13(4) To manage information quality (IQ) effectively, one needs to know how IQ changes over time, what causes it to change, and whether the changes can be predicted. In this paper we analyze the structure of IQ change in Wikipedia, an open, collaborative general encyclopedia. We found several patterns in Wikipedia’s IQ process trajectories and linked them to article types. Drawing on the results of our analysis, we develop a general model of IQ change that can be used for reasoning about IQ dynamics in many different settings, including traditional databases and information repositories. 0 1
An analysis of topical coverage of Wikipedia A. Halavais
D. Lackaff
Journal of Computer Mediated Communication Many have questioned the reliability and accuracy of Wikipedia. This article looks at a different but closely related one in the following: How broad is the coverage of Wikipedia? Differences in the interests and attention of Wikipedia's editors mean that some areas, in the traditional sciences, for example, are better covered than others. Two approaches to measuring this coverage are presented. The first maps the distribution of topics on Wikipedia to the distribution of books published. The second compares the distribution of topics in three established, field-specific academic encyclopedias to the articles found in Wikipedia. Unlike the top-down construction of traditional encyclopedias, Wikipedia's topical coverage is driven by the interests of its users, and as a result, the reliability and completeness of Wikipedia is likely to be different depending on the subject area of the article. 0 8
An application of wikis for mediated collaborative learning to Spanish L2 Carmen Martinez-Carrillo
Kostas Pentikousis
Wiki
Spanish L2
Web-based Education
Collaborative E-Learning
Sociocultural Theories of Learning
EuroIMSA English Information and Communication Technologies facilitate the emergence of new contexts and practices of learning that educational institutions have to adapt to their pedagogical discourse. This paper presents an application of wikis for mediated collaborative learning of Spanish as a second language (“Spanish L2”). We employed the off the-shelf wiki environment provided by wikispaces in this pilot project. Students were asked to create a set of pages in Spanish about topics of their interest, with the aim of creating a complete cross-linked wiki. At the end of the project, the students answered 25 open questions related with their experiences in improving their language skills, technical use, and collaborative writing with wikis; working collaboratively in small groups; and searching and reading available information in Spanish on the Internet. Our results indicate that students can quickly appropriate wikis as a tool for learning Spanish L2. Moreover, the wiki supported the shift from individual to collaborative writing effectively as indicated by a change from (a) personal to collective responsibility; (b) fragmentary and individual comprehension to general and collective knowledge; and (c) monocontextuality to polycontextuality. Notably, the wiki was a catalyst in language skill development and proved a good tool for cultural knowledge acquisition. 0 0
An approach for extracting bilingual terminology from Wikipedia Maike Erdmann
Kotaro Nakayama
Takahiro Hara
Shojiro Nishio
DASFAA English 0 1
An approach to deep web crawling by sampling Lu J.
Yafang Wang
Liang J.
Jilin Chen
Liu J.
Proceedings - 2008 IEEE/WIC/ACM International Conference on Web Intelligence, WI 2008 English Crawling deep web is the process of collecting data from search interfaces by issuing queries. With wide availability of programmable interface encoded in web services, deep web crawling has received a large variety of applications. One of the major challenges crawling deep web is the selection of the queries so that most of the data can be retrieved at a low cost. We propose a general method in this regard. In order to minimize the duplicates retrieved, we reduced the problem of selecting an optimal set of queries from a sample of the data source into the well-known set-covering problem and adopt a classical algorithm to resolve it. To verify that the queries selected from a sample also produce a good result for the entire data source, we carried out a set of experiments on large corpora including Wikipedia and Reuters. We show that our sampling-based method is effective by empirically proving that 1) The queries selected from samples can harvest most of the data in the original database; 2) The queries with low overlapping rate in samples will also result in a low overlapping rate in the original database; and 3) The size of the sample and the size of the terms from where to select the queries do not need to be very large. 0 0
An audio wiki supporting mobile collaboration Lei Wang
Paul Roe
Binh Pham
Dian Tjondronegoro
SAC English Wikis have proved to be very effective collaboration and knowledge management tools in large variety of fields thanks to their simplicity and flexible nature. Another important development for the internet is the emergence of powerful mobile devices supported by fast and reliable wireless networks. The combination of these developments begs the question of how to extend wikis on mobile devices and how to leverage mobile devices' rich modalities to supplement current wikis. Realizing that composing and consuming through auditory channel is the most natural and efficient way for mobile device user, this paper explores the use of audio as the medium of wiki. Our work, as the first step towards this direction, creates a framework called Mobile Audio Wiki which facilitates asynchronous audio-mediated collaboration on the move. In this paper, we present the design of Mobile Audio Wiki. As a part of such design, we propose an innovative approach for a light-weight audio content annotation system for enabling group editing, versioning and cross-linking among audio clips. To elucidate the novel collaboration model introduced by Mobile Audio Wiki, its four usage modes are identified and presented in storyboard format. Finally, we describe the initial design for presentation and navigation of Mobile Audio Wiki. 0 0
An effective, low-cost measure of semantic relatedness obtained from Wikipedia links. Milne
David
Witten
Ian .H.
Proceedings of the first AAAI Workshop on Wikipedia and Artificial Intelligence (WIKIAI'08), Chicago, I.L. This paper describes a new technique for obtaining measures of semantic relatedness. Like other recent approaches, it uses Wikipedia to provide structured world knowledge about the terms of interest. Our approach is unique in that it does so using the hyperlink structure of Wikipedia rather than its category hierarchy or textual content. Evaluation with manually defined measures of semantic relatedness reveals this to be an effective compromise between the ease of computation of the former approach and the accuracy of the latter. 0 0
An effective, low-cost measure of semantic relatedness obtained from wikipedia links Milne D.
Witten I.H.
AAAI Workshop - Technical Report English This paper describes a new technique for obtaining measures of semantic relatedness. Like other recent approaches, it uses Wikipedia to provide structured world knowledge about the terms of interest. Our approach is unique in that it does so using the hyperlink structure of Wikipedia rather than its category hierarchy or textual content. Evaluation with manually defined measures of semantic relatedness reveals this to be an effective compromise between the ease of computation of the former approach and the accuracy of the latter. Copyright 0 1
An empirical research on extracting relations from Wikipedia text Huang J.-X.
Ryu P.-M.
Choi K.-S.
Feature-based
Information extraction
Relatedness information
Relation classification
Lecture Notes in Computer Science English A feature based relation classification approach is presented, in which probabilistic and semantic relatedness features between patterns and relation types are employed with other linguistic information. The importance of each feature set is evaluated with Chi-square estimator, and the experiments show that, the relatedness features have big impact on the relation classification performance. A series experiments are also performed to evaluate the different machine learning approaches on relation classification, among which Bayesian outperformed other approaches including Support Vector Machine (SVM). 0 0
An empirical study of the effects of NLP components on Geographic IR performance N Stokes
Y Li
A Moffat
JW Rong
INTERNATIONAL JOURNAL OF GEOGRAPHICAL INFORMATION SCIENCE Natural language processing {(NLP)} techniques, such as toponym detection and resolution, are an integral part of most geographic information retrieval {(GIR)} architectures. Without these components, synonym detection, ambiguity resolution and accurate toponym expansion would not be possible. However, there are many important factors affecting the success of an {NLP} approach to {GIR,} including toponym detection errors, toponym resolution errors and query overloading. The aim of this paper is to determine how severe these errors are in state-of-the-art systems, and to what extent they affect {GIR} performance. We show that a careful choice of weighting schemes in the {IR} engine can minimize the negative impact of these errors on {GIR} accuracy. We provide empirical evidence from the {GeoCLEF} 2005 and 2006 datasets to support our observations. 0 0
An integrated probabilistic and logic approach to encyclopedia relation extraction with multiple features Yu X.
Lam W.
Coling 2008 - 22nd International Conference on Computational Linguistics, Proceedings of the Conference English We propose a new integrated approach based on Markov logic networks (MLNs), an effective combination of probabilistic graphical models and first-order logic for statistical relational learning, to extracting relations between entities in encyclopedic articles from Wikipedia. The MLNs model entity relations in a unified undirected graph collectively using multiple features, including contextual, morphological, syntactic, semantic as well as Wikipedia characteristic features which can capture the essential characteristics of relation extraction task. This model makes simultaneous statistical judgments about the relations for a set of related entities. More importantly, implicit relations can also be identified easily. Our experimental results showed that, this integrated probabilistic and logic model significantly outperforms the current state-of-the-art probabilistic model, Conditional Random Fields (CRFs), for relation extraction from encyclopedic articles. © 2008. Licensed under the Creative Commons. 0 0
An n-gram and initial description based approach for entity ranking track Murugeshan M.S.
Saswati Mukherjee
Entity Ranking
List Completion
N-gram checking
Lecture Notes in Computer Science English The most important work that takes the center stage in the Entity Ranking track of INEX is proper query formation. Both the subtasks, namely Entity Ranking and List Completion, would immensely benefit if the given query can be expanded with more relevant terms, thereby improving the efficiency of the search engine. This paper stresses on the correct identification of "Meaningful n-grams" from the given title and proper selection of the "Prominent n-grams" among them as the utmost important task that improves query formation and hence improves the efficiencies of the overall Entity Ranking tasks. We also exploit the Initial Descriptions (IDES) of the Wikipedia articles for ranking the retrieved answers based on their similarities with the given topic. List completion task is further aided by the related Wikipedia articles that boosted the score of retrieved answers. 0 0
An unsupervised approach to biography production using Wikipedia Biadsy F.
Hirschberg J.
Elena Filatova
ACL-08: HLT - 46th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference English We describe an unsupervised approach to multi-document sentence-extraction based summarization for the task of producing biographies. We utilize Wikipedia to automatically construct a corpus of biographical sentences and TDT4 to construct a corpus of non-biographical sentences. We build a biographical-sentence classifier from these corpora and an SVM regression model for sentence ordering from the Wikipedia corpus. We evaluate our work on the DUC2004 evaluation data and with human judges. Overall, our system significantly outperforms all systems that participated in DUC2004, according to the ROUGE-L metric, and is preferred by human subjects. 0 0
Analyzing Wiki-based Networks to Improve Knowledge Processes in Organizations Müller
C.
Meuthrath
B.
Baumgraß
A.
Collaboration network
Knowledge work
Network analysis
Social software
Wiki
Journal of Universal Computer Science, 14(4) Increasingly wikis are used to support existing corporate knowledge exchange processes. They are an appropriate software solution to support knowledge processes. However, it is not yet proven whether wikis are an adequate knowledge management tool or not. This paper presents a new approach to analyze existing knowledge exchange processes in wikis based on network analysis. Because of their dynamic characteristics four perspectives on wiki networks are introduced to investigate the interrelationships between people, information, and events in a wiki information space. As an analysis method the Social Network Analysis (SNA) is applied to uncover existing structures and temporal changes. A scenario data set of an analysis conducted with a corporate wiki is presented. The outcomes of this analysis were utilized to improve the existing corporate knowledge processes. 0 0
Analyzing wiki-based networks to improve knowledge processes in organizations Journal of Universal Computer Science English 0 0
And the anthropological challenge to designing interhuman relations Reframing semiotic telematic knowledge spaces R. Stettler Technoetic Arts: A Journal of Speculative Research Number 2 0 0
Annotating documents by Wikipedia concepts Schonhofen P. Proceedings - 2008 IEEE/WIC/ACM International Conference on Web Intelligence, WI 2008 English We present a technique which is able to reliably label words or phrases of an arbitrary document with Wikipedia articles (concepts) best describing their meaning. First it scans the document content, and when it finds a word sequence matching the title of a Wikipedia article, it attaches the article to the constituent word(s). The collected articles are then scored based on three factors: (1) how many other detected articles they semantically relate to, according to the Wikipedia link structure; (2) how specific is the concept they represent; and (3) how similar is the title by which they were detected to their "official" title. If a text location refers to multiple Wikipedia articles, only the one with the highest score is retained. Experiments on 24,000 randomly selected Wikipedia article bodies showed that 81% of phrases annotated by article authors were correctly identified. Moreover, out of the 5 concepts deemed as the most important by our algorithm during a final ranking, in average 72% was indeed marked in the original text. 0 0
Application of social network analysis in knowledge processes Mueller C.
Gronau N.
Lembcke R.
Knowledge process
Modeling method
Social network analysis
Wiki
16th European Conference on Information Systems, ECIS 2008 English Social network analysis (SNA) is increasingly applied to analyze existing knowledge exchange processes in companies. In the past, these processes were mainly analyzed by traditional business process modeling methods. Same of these methods are working with a specific knowledge perspective, e.g. KMDL (Knowledge Modeling and Description Language). The knowledge needs and demands are modeled along knowledge-intensive business processes. The SNA however has no direct reference to processes; here persons with specific relationships (e.g. knowledge exchange) are mainly the object of interest. Therefore a combination of both views - the SNA and knowledge-intensive business process modeling - is valuable. In the following contribution two approaches are introduced to combine social network analysis with the modeling of knowledge intensive business processes based on KMDL. In addition to a theoretical introduction to SNA and KMDL there are two examples introduced, which illustrate the practical implementation of these approaches. In this work in progress traditional techniques like interviews are combined with electronically gathered data. Also by using different perspectives a better understanding of all events in knowledge processes can be extracted. 0 0
Application wikis to mashup makers to next generation mashups: keynote Stewart Nickolas WikiSym English 0 0
Applications of Semantic Web Methodologies and Techniques to Social Networks and Social Websites Sheila Kinsella
John G. Breslin
Alexandre Passant
Stefan Decker
Reasoning Web One of the most visible trends on the Web is the emergence of {Social} Web" sites which facilitate the creation and gathering of knowledge through the simplification of user contributions via blogs tagging and folksonomies wikis podcasts and the deployment of online social networks. The Social Web has enabled community-based knowledge acquisition with efforts like the Wikipedia demonstrating the "wisdom of the crowds" in creating the world's largest online encyclopaedia. Although it is difficult to define the exact boundaries of what structures or abstractions belong to the Social Web a common property of such sites is that they facilitate collaboration and sharing between users with low technical barriers although usually on single sites. As more social websites form around the connections between people and their objects of interest and as these "object-centred networks" grow bigger and more diverse more intuitive methods are needed for representing and navigating the content items in these sites: both within and across social websites. Also to better enable user access to multiple sites interoperability among social websites is required in terms of both the content objects and the person-to-person networks expressed on each site. This requires representation mechanisms to interconnect people and objects on the Social Web in an interoperable and extensible way. The Semantic Web provides such representation mechanisms: it can be used to link people and objects by representing the heterogeneous ties that bind us all to each other (either directly or indirectly). In this paper we will describe methods that build on agreed-upon Semantic Web formats to describe people content objects and the connections that bind them together explicitly or implicitly enabling social websites to interoperate by appealing to some common semantics. We will also focus on how developers can use the Semantic Web to augment the ways in which they createreuse and link content on social networking sites and social websites." 0 0
Applying Web 2.0 design principles in the design of cooperative applications Pinkwart N. Cooperative applications
CSCL
Computer-Supported Cooperative Work
Social software
Lecture Notes in Computer Science English "Web 2.0" is a term frequently mentioned in media - apparently, applications such as Wikipedia, Social Network Services, Online Shops with integrated recommender systems, or Sharing Services like flickr, all of which rely on user's activities, contributions, and interactions as a central factor, are fascinating for the general public. This leads to a success of these systems that seemingly exceeds the impact of most "traditional" groupware applications that have emerged from CSCW research. This paper discusses differences and similarities between novel Web 2.0 tools and more traditional CSCW application in terms of technologies, system design and success factors. Based on this analysis, the design of the cooperative learning application LARGO is presented to illustrate how Web 2.0 success factors can be considered for the design of cooperative environments. 0 0
Applying the logic of multiple-valued argumentation to social web: SNS and wikipedia Shusuke Kuribara
Safia Abbas
Hajime Sawamura
Argumentation
Logic of multiple-valued argumentation
SNS
Social web
Wikipedia
Lecture Notes in Computer Science English The Logic of Multiple-Valued Argumentation (LMA) is an argumentation framework that allows for argument-based reasoning about uncertain issues under uncertain knowledge. In this paper, we describe its applications to Social Web: SNS and Wikipedia. They are said to be the most influential social Web applications to the present and future information society. For SNS, we present an agent that judges the registration approval for Mymixi in mixi in terms of LMA. For Wikipedia, we focus on the deletion problem of Wikipedia and present agents that argue about the issue on whether contributed articles should be deleted or not, analyzing arguments proposed for deletion in terms of LMA. These attempts reveal that LMA can deal with not only potential applications but also practical ones such as extensive and contemporary applications. 0 0
Article Recommendation Based on a Topic Model for Wikipedia Selection for Schools Choochart Haruechaiyasak
Chaianun Damrongrat
Latent Dirichlet Allocation (LDA)
Wikipedia
Content-based filtering
Educational Web contents
Recommender system
English 0 0
Article recommendation based on a topic model for Wikipedia Selection for Schools Choochart Haruechaiyasak
Chaianun Damrongrat
Content-based filtering
Educational Web contents
Latent Dirichlet Allocation (LDA)
Recommender system
Wikipedia
Lecture Notes in Computer Science English The 2007 Wikipedia Selection for Schools is a collection of 4,625 selected articles from Wikipedia as educational for children. Users can currently access articles within the collection via two different methods: (1) by browsing on either a subject index or a title index sorted alphabetically, and (2) by following hyperlinks embedded within article pages. These two retrieval methods are considered static and subjected to human editors. In this paper, we apply the Latent Dirichlet Allocation (LDA) algorithm to generate a topic model from articles in the collection. Each article can be expressed by a probability distribution on the topic model. We can recommend related articles by calculating the similarity measures among the articles' topic distribution profiles. Our initial experimental results showed that the proposed approach could generate many highly relevant articles, some of which are not covered by the hyperlinks in a given article. 0 0
Articulations of wikiwork: Uncovering valued work in wikipedia through barnstars Travis Kriplean
Ivan Beschastnikh
David W. McDonald
Articulation work
Barnstars
Commons-based peer production
Online community
Wikipedia
English Successful online communities have complex cooperative arrangements, articulations of work, and integration practices. They require technical infrastructure to support a broad division of labor. Yet the research literature lacks empirical studies that detail which types of work are valued by participants in an online community. A content analysis of Wikipedia barnstars - personalized tokens of appreciation given to participants - reveals a wide range of valued work extending far beyond simple editing to include social support, administrative actions, and types of articulation work. Our analysis develops a theoretical lens for understanding how wiki software supports the creation of articulations of work. We give implications of our results for communities engaged in large-scale collaborations. Copyright 2008 ACM. 0 1
Assessing students' WIKI contributions ASEE Annual Conference and Exposition, Conference Proceedings English 0 0
Assigning Trust to Wikipedia Content B. Thomas Adler
Krishnendu Chatterjee
Luca de Alfaro
Marco Faella
Ian Pye
Vishwanath Raman
WikiSym English The Wikipedia is a collaborative encyclopedia: anyone can contribute to its articles simply by clicking on an "edit" button. The open nature of the Wikipedia has been key to its success, but has also created a challenge: how can readers develop an informed opinion on its reliability? We propose a system that computes quantitative values of trust for the text in Wikipedia articles; these trust values provide an indication of text reliability. The system uses as input the revision history of each article, as well as information about the reputation of the contributing authors, as provided by a reputation system. The trust of a word in an article is computed on the basis of the reputation of the original author of the word, as well as the reputation of all authors who edited text near the word. The algorithm computes word trust values that vary smoothly across the text; the trust values can be visualized using varying text-background colors. The algorithm ensures that all changes to an article's text are reflected in the trust values, preventing surreptitious content changes. We have implemented the proposed system, and we have used it to compute and display the trust of the text of thousands of articles of the English Wikipedia. To validate our trust-computation algorithms, we show that text labeled as low-trust has a significantly higher probability of being edited in the future than text labeled as high-trust. 0 7
Association thesaurus construction methods based on link co-occurrence analysis for Wikipedia Masahiro Ito
Kotaro Nakayama
Takahiro Hara
Shojiro Nishio
English Wikipedia, a huge scale Web based encyclopedia, attracts great attention as an invaluable corpus for knowledge extraction because it has various impressive characteristics such as a huge number of articles, live updates, a dense link structure, brief anchor texts and URL identification for concepts. We have already proved that we can use Wikipedia to construct a huge scale accurate association thesaurus. The association thesaurus we constructed covers almost 1.3 million concepts and its accuracy is proved in detailed experiments. However, we still need scalable methods to analyze the huge number of Web pages and hyperlinks among articles in the Web based encyclopedia. 0 0
Augmented Social Cognition Chi E.H.
Peter Pirolli
Bongwon Suh
Aniket Kittur
Pendleton B.
Mytkowicz T.
AAAI Spring Symposium - Technical Report English Research in Augmented Social Cognition is aimed at enhancing the ability of a group of people to remember, think, and reason; to augment their speed and capacity to acquire, produce, communicate, and use knowledge; and to advance collective and individual intelligence in socially mediated information environments. In this paper, we describe the emergence of this research endeavor, and summarize some results from the research. In particular, we have found that (1) analyses of conflicts and coordination in Wikipedia have shown us the scientific need to understand social sensemaking environments; and (2) information theoretic analyses of social tagging behavior in del.icio.us shows the need to understand human vocabulary systems. 0 0
Augmenting Domain-Specific Thesauri With Knowledge from Wikipedia Olena Medelyan
David N. Milne
English 0 1
Augmenting domain-specific thesau with knowledge from wikipedia Olena Medelyan
Milne D.
Thesauri
Wikipedia
Word sense disambiguation
New Zealand Computer Science Research Student Conference, NZCSRSC 2008 - Proceedings English We propose a new method for extending a domain-specific thesaurus with valuable information from Wikipedia. The main obstacle is to disambiguate thesaurus concepts to correct Wikipedia articles. Given the concept name, we first identify candidate mappings by analyzing article titles, their redirects and disambiguation pages. Then, for each candidate, we compute a link-based similarity score to all mappings of context terms related to this concept. The article with the highest score is then used to augment the thesaurus concept. It is the source for the extended gloss, explaining the concept's meaning, synonymous expressions that can be used as additional non-descriptors in the thesaurus, translations of the concept into other languages, and new domain-relevant concepts. Copyright is held by the author/owner(s). 0 0
Augmenting domain-specific thesauri with knowledge from Wikipedia Medelyan
O.
Milne
D.
Proceedings of the NZ Computer Science Research Student Conference (NZCSRSC 2008), Christchurch, New Zealand. 0 1
Augmenting wikipedia-extraction with results from the web Fei Wu
Raphael Hoffmann
Weld D.S.
AAAI Workshop - Technical Report English Not only is Wikipedia a comprehensive source of quality information, it has several kinds of internal structure (e.g., relational summaries known as infoboxes), which enable self-supervised information extraction. While previous efforts at extraction from Wikipedia achieve high precision and recall on well-populated classes of articles, they fail in a larger number of cases, largely because incomplete articles and infrequent use of infoboxes lead to insufficient training data. This paper explains and evaluates a method for improving recall by extracting from the broader Web. There are two key advances necessary to make Web supplementation effective: 1) a method to filter promising sentences from Web pages, and 2) a novel retraining technique to broaden extractor recall. Experiments show that, used in concert with shrinkage, our techniques increase recall by a factor of up to 8 while maintaining or increasing precision. Copyright 0 0
Automated Building of Error Corpora of Polish Marcin Milkowski Error corpora
Normativity
Revision history
Corpora building
Corpus Linguistics, Computer Tools, and Applications – State of the Art. PALC 2007, Peter Lang. Internationaler Verlag der Wissenschaften 2008, 631-639 The paper shows how to automatically develop error corpora out of revision history of documents. The idea is based on a hypothesis that minor edits in documents represent correction of typos, slips of the tongue, grammar, usage and style mistakes. This hypothesis has been confirmed by frequency analysis of revision history of articles in the Polish Wikipedia. Resources such as revision history in Wikipedia, Wikia, and other collaborative editing systems, can be turned into corpora of errors, just by extracting the minor edits. The most theoretically interesting aspect is that the corrections will represent the average speaker's intuitions about usage, and this seems to be a promising way of researching normativity in claims about proper or improper Polish. By processing the revision history, one can gain pairs of segments in the corpus: first representing the error, and the other representing the correction. Moreover, it is relatively easy to tag parts of speech, compare subsequent versions, and prepare a text file containing the resulting corpus. 0 0
Automated construction Chinese domain ontology from Wikipedia Li Lian
Jun Ma
JingSheng Lei
Ling Song
LeBo Liu
Proceedings - 4th International Conference on Natural Computation, ICNC 2008 English Wikipedia (Wiki) is a collaborative on-line encyclopedia, where web users are able to share their knowledge about a certain topic. How to make use of the rich knowledge in the Wiki is a big challenge. In this paper we propose a method to construct domain ontology from the Chinese Wiki automatically. The main Idea in this paper is based on the entry segmenting and Feature Text (FT) extracting, where we segment the name of entries and establish the concept hierarchy firstly. Secondly, we extract the FTs from the descriptions of entries to eliminate the redundant information. Finally we calculate the similarity between pairs of FTs to revise the concept hierarchy and gain non-taxonomy relations between concepts. The primary experiment indicates that our method is useful for Chinese domain ontology construction. 0 0
Automated object shape modelling by clustering of web images Scardino G.
Infantino I.
Gaglio S.
Image annotation
Images clustering
Visual image search
VISAPP 2008 - 3rd International Conference on Computer Vision Theory and Applications, Proceedings English The paper deals with the description of a framework to create shape models of an object using images fromthe web. Results obtained from different image search engines using simple keywords are filtered, and it is possible to select images viewing a single object owning a well-defined contour. In order to have a large set of valid images, the implemented system uses lexical web databases (e.g. WordNet) or free web encyclopedias (e.g. Wikipedia), to get more keywords correlated to the given object. The shapes extracted from selected images are represented by Fourier descriptors, and are grouped by K-means algorithm. Finally, the more representative shapes of main clusters are considered as prototypical contours of the object. Preliminary experimental results are illustrated to show the effectiveness of the proposed approach. 0 0
Automatic Edit-War Detection in Wikipedia Dennis Hoppe Bauhaus-University Weimar German 0 0
Automatic Vandalism Detection in Wikipedia Robert Gerling Bauhaus-University Weimar German We present results of a new approach to detect destructive article revisions, so-called vandalism, in Wikipedia. Vandalism detection is a one-class classification problem, where vandalism edits are the target to be identified among all revisions. Interestingly, vandalism detection has not been addressed in the Information Retrieval literature by now. In this paper we discuss the characteristics of vandalism as humans recognize it and develop features to render vandalism detection as a machine learning task. We compiled a large number of vandalism edits in a corpus, which allows for the comparison of existing and new detection approaches. Using logistic regression we achieve 83% precision at 77% recall with our model. Compared to the rule-based methods that are currently applied in Wikipedia, our approach increases the F-Measure performance by 49% while being faster at the same time. 0 4
Automatic Wikibook prototyping Chou J.-L.
Wu S.-H.
Table of content generation
Wikibook
Wikipedia
Proceedings - ICCE 2008: 16th International Conference on Computers in Education English Wikipedia is the world's largest collaboratively edited source of encyclopedic knowledge. Wikibook is a sub-project of Wikipedia. The purpose of Wikibook is to enable a free textbook to be edited by various contributors, in the same way that Wikipedia is composed and edited. However, editing a book requires more effort than editing separate articles. Therefore, how to help users cooperatively edit a book is a new research issue. In this paper, we investigate how to automatically extract content from Wikipedia and generate a prototype of a Wikibook. Applying search technology, our system can retrieve relevant articles from Wikipedia. A table of contents is built automatically based on link analysis and. Our experiment shows that given a topic, our system can generate a table of contents, which can be treated as a prototype of a Wikibook. 0 0
Automatic extraction of useful facet hierarchies from text databases Dakka W.
Ipeirotis P.G.
Proceedings - International Conference on Data Engineering English Databases of text and text-annotated data constitute a significant fraction of the information available in electronic form. Searching and browsing are the typical ways that users locate items of interest in such databases. Faceted interfaces represent a new powerful paradigm that proved to be a successful complement to keyword searching. Thus far, the identification of the facets was either a manual procedure, or relied on apriori knowledge of the facets that can potentially appear in the underlying collection. In this paper, we present an unsupervised technique for automatic extraction of facets useful for browsing text databases. In particular, we observe, through a pilot study, that facet terms rarely appear in text documents, showing that we need external resources to identify useful facet terms. For this, we first identify important phrases in each document. Then, we expand each phrase with "context" phrases using external resources, such as WordNet and Wikipedia, causing facet terms to appear in the expanded database. Finally, we compare the term distributions in the original database and the expanded database to identify the terms that can be used to construct browsing facets. Our extensive user studies, using the Amazon Mechanical Turk service, show that our techniques produce facets with high precision and recall that are superior to existing approaches and help users locate interesting items faster. 0 0
Automatic thesaurus generation using co-occurrence Brussee R.
Wartena C.
Belgian/Netherlands Artificial Intelligence Conference English This paper proposes a characterization of useful thesaurus terms by the informativity of co-occurrence with that term. Given a corpus of documents, informativity is formalized as the information gain of the weighted average term distribution of all documents containing that term. While the resulting algorithm for thesaurus generation is unsupervised, we find that high informativity terms correspond to large and coherent subsets of documents. We evaluate our method on a set of DutchWikipedia articles by comparing high informativity terms with keywords for the Wikipedia category of the articles. 0 0
Automatic vandalism detection in wikipedia: Towards a machine learning approach Smets K.
Goethals B.
Verdonk B.
AAAI Workshop - Technical Report English Since the end of 2006 several autonomous bots are, or have been, running on Wikipedia to keep the encyclopedia free from vandalism and other damaging edits. These expert systems, however, are far from optimal and should be improved to relieve the human editors from the burden of manually reverting such edits. We investigate the possibility of using machine learning techniques to build an autonomous system capable to distinguish vandalism from legitimate edits. We highlight the results of a small but important step in this direction by applying commonly known machine learning algorithms using a straightforward feature representation. Despite the promising results, this study reveals that elementary features, which are also used by the current approaches to fight vandalism, are not sufficient to build such a system. They will need to be accompanied by additional information which, among other things, incorporates the semantics of a revision. Copyright 0 3
Automatically Refining the Wikipedia Infobox Ontology Fei Wu
Daniel S. Weld
Semantic web
Ontology
Wikipedia
Markov Logic Networks
17th International World Wide Web Conference (www-08) The combined efforts of human volunteers have recently extracted numerous facts fromWikipedia, storing them asmachine-harvestable object-attribute-value triples inWikipedia infoboxes. Machine learning systems, such as Kylin, use these infoboxes as training data, accurately extracting even more semantic knowledge from natural language text. But in order to realize the full power of this information, it must be situated in a cleanly-structured ontology. This paper introduces KOG, an autonomous system for refining Wikipedia’s infobox-class ontology towards this end. We cast the problem of ontology refinement as a machine learning problem and solve it using both SVMs and a more powerful joint-inference approach expressed in Markov Logic Networks. We present experiments demonstrating the superiority of the joint-inference approach and evaluating other aspects of our system. Using these techniques, we build a rich ontology, integratingWikipedia’s infobox-class schemata with WordNet. We demonstrate how the resulting ontology may be used to enhance Wikipedia with improved query processing and other features. 0 0
Automatically refining the Wikipedia infobox ontology Fei Wu
Daniel S. Weld
English The combined efforts of human volunteers have recently extracted numerous facts from Wikipedia, storing them as machine-harvestable object-attribute-value triples in Wikipedia infoboxes. Machine learning systems, such as Kylin, use these infoboxes as training data, accurately extracting even more semantic knowledge from natural language text. But in order to realize the full power of this information, it must be situated in a cleanly-structured ontology. This paper introduces KOG, an autonomous system for refining Wikipedia's infobox-class ontology towards this end. We cast the problem of ontology refinement as a machine learning problem and solve it using both SVMs and a more powerful joint-inference approach expressed in Markov Logic Networks. We present experiments demonstrating the superiority of the joint-inference approach and evaluating other aspects of our system. Using these techniques, we build a rich ontology, integrating Wikipedia's infobox-class schemata with WordNet. We demonstrate how the resulting ontology may be used to enhance Wikipedia with improved query processing and other features. 0 0
Automatischer Aufbau eines multilingualen Thesaurus durch Extraktion semantischer und lexikalischer Relationen aus der Wikipedia Daniel Kinzler Universtität Leipzig German 0 0
Autopoiesis in virtual organizations M. Pamkowska Informatica Economica Virtual organizations continuously gain popularity because of the benefits created by them. Generally, they are defined as temporal adhocracies, project oriented, knowledge-based network organizations. The goal of this paper is to present the hypothesis that knowledge system developed by virtual organization is an autopoietic system. The term autopoiesis" was introduced by Maturana for self-productive systems. In this paper Wikipedia is described as an example of an autopoietic system. The first part of the paper covers discussion on virtual organizations. Next autopoiesis' interpretations are delivered and the value of autopoiesis for governance of virtual organizations is presented. The last parts of the work comprise short presentation of Wikipedia its principles and conclusions of Wikipedia as an autopoietic system." 0 0
Babel Wiki workshop: Cross-language collaboration WikiSym English 0 0
Back to basics: Electronic pedagogy from the (virtual) ground up Ehrlich J. Blogs
Concordance
Curriculum
Digital divide
EEBO
Enduring
Imagery
Mash-up
McTighe
Pedagogy
Performance
Podcast
Sampling
UbD
Understanding
Webcast
Wiggins
Wiki
YouTube
Shakespeare English Educators today have no agreement on what "electronic Shakespeares" in the classroom might mean, or even what they should mean. In this essay, I will survey a variety of electronic resources that have been used in classrooms effectively, and examine a variety of ways that students can use them to learn. In doing so, I hope to suggest ways in which these tools can be used effectively in the classroom today, regardless of which tools ultimately become part of a teacher's arsenal in 20 years. Jerome McGann argues that "humanities education ... will not take the use of digital technology seriously until one demonstrates how its tools improve the ways we explore and explain aesthetic works ...". It is my hope that this exploration can help to provide that kind of explanation. By looking in turn at wikis, blogs, online texts, concordances and a wide range of "performance media", I hope to identify those areas where good pedagogy can be enhanced by the strengths of the Internet. In doing so, I argue that using the Internet in the classroom can be a humanizing tool and a promotion of the kinds of close reading the Internet may otherwise discourage. Nonetheless, electronic Shakespeares look a lot like electronic Joyces and electronic bibles: like Shakespeare's cultural ubiquity, his exceptionality is a social construction, not apparent in the application of electronic methods. 0 0
Becoming a Part of a "Community" Online in Order to Acquire Language Skills L. Thelmadatter MEXTESOL Journal Number 1 It is a commons belief among foreign language students that one needs to go to the country where the language is spoken to really master it. This belief has been shown to have validity in concepts such as sociocultural theory, which basically states that one needs to become part of a community in order to master the skills associated with it. Given that most foreign language learners do not have the luxury of living in a place where the language is spoken by native speakers, alternatives, such as online resources show promise as a substitute. This article outlines several types of online interactions or "virtual communities" such as chat rooms, role-playing games and Wikipedia analyzing how students work with them and their feedback given in surveys and reflective essays as part of coursework. 0 0
Bell Labs Technical Journal: Overview Hofmann M.
Tomsu M.
Bell Labs Technical Journal English A review of the research papers included in the Bell Labs Technical Journal, addressing various important aspects of modern content networks and investigating the implications of the paradigm shift on technology and business models, is presented. The research papers highlight Internet and mobile Internet as the most effective tool for empowering users in the information age. Wikipedia encyclopedia and user-generated video content and the ability to create new business models allow an effective information dissemination. Continuously expanding bandwidth provides access to content for more than a billion users, while broadband connectivity and data transport quality address the requirements of today's business cases. The research papers also highlight that a service-oriented infrastructure provide infrastructure virtualization and describe the development of content delivery solutions such as IPTV and video-on-demand (VoD). 0 0
Big data: Wikiomics Mitch Waldrop Nature English Pioneering biologists are trying to use wiki-type web pages to manage and interpret data, reports Mitch Waldrop. But will the wider research community go along with the experiment? 0 1
Big data: open-source format needed to aid wiki collaboration Tin-Lap Lee English 0 0
Blind relevance feedback with wikipedia: Enterprise track Peng Y.
Mao M.
NIST Special Publication English In this year's Enterprise track experiment, we focused on testing Blind Relevance Feedback, especially using online Wikipedia as query expansion collection. We demonstrated that using Wikipedia as query expansion collection returns better infNDCG than not using it. 0 0
Building a textual entailment system for the RTE3 competition. Application to a QA system Adrian Iftene Proceedings of the 2008 10th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing, SYNASC 2008 English Textual entailment recognition (RTE) is the task of deciding, when given two text fragments, whether the meaning of one text is entailed (can be inferred) from the other text. Last year, we built our first Textual Entailment (TE) system, with which we participated in the RTE31 competition. The main idea of this system is to transform the hypothesis making use of extensive semantic knowledge from sources like DIRT, WordNet, Wikipedia and acronyms database. Additionally, the system applies complex grammar rules for rephrasing in English and uses the results of a module we built to acquire the extra background knowledge needed. In the first part, we presented the system architecture and the results, whose best run ranked 3rd in RTE3 among 45 participating runs of 26 groups. The second part of the paper presents the manner in which we adapted the TE system in order to include it in a Question Answering (QA) system. The aim of using the TE system as a module in the general architecture of a QA system is to improve the ranking between possible answers for questions in which the answer type is Measure, Person, Location, Date and Organization. 0 0
Building an SVG interface to MediaWiki: Creating a web-based map of interpersonal relations in the sphere of culture Edwards P.
Goodden N.
SVG Open 2008 - 6th International Conference on Scalable Vector Graphics English In 2006, e-2 started a project with artist Nat Goodden which maps key cultural figures by using information about their acquaintanceship with others to link them together. The name of the project is Who knew Whom, and the first working prototype of the SVG interface can be found at http:// culturalcartography.net/ (the SVG "view" can be activated from within each of the biographical pages). The ideal tool for the artist to input entries was a wiki, so we used MediaWiki, the software behind WikiPedia. The capability to engage large numbers of contributors in the future was one of the criteria for this, as was the ability of the software to map links between pages. The artist wanted an interface similar to ThinkMap's visual thesaurus (http://www.visualthesaurus.com/), which shows how an imaginative understanding of both visual and conceptual semantics can draw the viewer in to an understanding of highly complex interrelationships. In our case, individual names are animated amongst a field of other connected names, and can be clicked to move them to the foreground, and potentially change most of the interface elements. We chose SVG for the interface, as we had previous experience of using SVG to animate text and had found that it out-performed all alternatives. This paper focuses on the way in which we developed custom solutions for font handling, animation and the use of AJAX in realising the project. Font embedding in SVG viewers is badly supported, and we needed to include a cursive font in the interface and have it display correctly in SVG viewers which did not support SVG fonts. We therefore embedded the font definition as a set of paths, themselves derived from the TrueType font. Kerning information was derived in the same way and made available to the application as a JSON object, and all text operations were carried out using scripts. Animation of the interface was to be handled entirely using scripts, as declarative animation suffered from a similar lack of support. AJAX was needed to constantly update the interface, and to retrieve more in-depth information about each subject. The project illustrates the efficacy with which SVG can lend itself to interface design, and it also indicates that the fledgling support for SVG in modern browsers such as Safari and Firefox is very good, and can be harnessed in ways which can sidestep the lack of certain features in their implementations. 0 0
Building community wikipedias: A machine-human partnership approach DeRose P.
Chai X.
Gao B.J.
Shen W.
Doan A.
Bohannon P.
Zhu X.
Proceedings - International Conference on Data Engineering English The rapid growth of Web communities has motivated many solutions for building community data portals. These solutions follow roughly two approaches. The first, approach (e.g., Libra, Citeseer, Cimple) employs semi-automatic methods to extract and integrate data from a multitude of data sources. The second approach (e.g., Wikipedia, Intellipedia) deploys an initial portal in wild format, then invites community members to revise and add material. In this paper we consider combining the above two approaches to building community portals. The new hybrid machine-human approach brings significant benefits. It can achieve broader and deeper coverage, provide more incentives for users to contribute, and keep the portal more up-to-date with less user effort. In a sense, it enables building "community wikipedias", backed by an underlying structured database that is continuously updated using automatic techniques. We outline our ideas for the new approach, describe its challenges and opportunities, and provide initial solutions. Finally, we describe a real-world implementation and preliminary experiments that demonstrate the utility of the new approach. 0 0
Building knowledge online: master students' use and evaluation of wiki and forum Carmen Lebherz
Ulf-Dietrich Reips
Klaus Jonas
ICLS English 0 0
Building semantic kernels for text classification using Wikipedia Pu Wang
Carlotta Domeniconi
English Document classification presents difficult challenges due to the sparsity and the high dimensionality of text data, and to the complex semantics of the natural language. The traditional document representation is a word-based vector (Bag of Words, or BOW), where each dimension is associated with a term of the dictionary containing all the words that appear in the corpus. Although simple and commonly used, this representation has several limitations. It is essential to embed semantic information and conceptual patterns in order to enhance the prediction capabilities of classification algorithms. In this paper, we overcome the shortages of the BOW approach by embedding background knowledge derived from Wikipedia into a semantic kernel, which is then used to enrich the representation of documents. Our empirical evaluation with real data sets demonstrates that our approach successfully achieves improved classification accuracy with respect to the BOW technique, and to other recently developed methods. 0 1
Business 2.0: A novel model for delivery of business services Yap P.-H.
Ong K.-L.
Xiaolong Wang
B2B
Business analytics
Data analysis
Web 2.0
5th International Conference Service Systems and Service Management - Exploring Service Dynamics with Science and Innovative Technology, ICSSSM'08 English Web 2.0, regardless of the exact definition, has proven to bring about significant changes to the way the Internet was used. Evident by key innovations such as Wikipedia, FaceBook, YouTube, and Blog sites, these community-based Website in which contents are generated and consumed by the same group of users are changing the way businesses operate. Advertisements are no longer 'forced' upon the viewers but are instead 'intelligently' targeted based on the contents of interest. In this paper, we investigate the concept of Web 2.0 in the context of business entities. We asked if Web 2.0 concepts could potentially lead to a change of paradigm or the way businesses operate today. We conclude with a discussion of a Web 2.0 application we recently developed that we think is an indication that businesses will ultimately be affected by these community-based technologies; thus bringing about Business 2.0 - a paradigm for businesses to cooperate with one another to deliver improved products and services to their own customers. 0 0
CAWS: An Awareness Based Wiki System to Improve Team Collaboration Ilaria Liccardi
Hugh C. Davis
Su White
Collaboration
Awareness
Co-authoring
Wiki
CAWS
ICALT English Effective collaborative authoring techniques require tools that consider the social aspects of collaboration in addition to the technical aspects. Collaborative authoring is fundamentally different to individual writing because of the communications that must inevitably take place between team members. Despite the fact that collaborative authoring has greatly increased in popularity in recent years, most collaborative authoring efforts are performed using tools that are primarily designed for individual authors. The lack of regard for the collaborative process leads to a number of common problems.This paper presents research into the use of a prototype wiki-based system (CAWS) to more effectively support the collaborative process. The results of two field studies into the use of the system are examined, in order to investigate the effectiveness of the techniques employed by the tool. 0 2
CE2 - Towards a large scale hybrid search engine with integrated ranking support Haofen Wang
Thanh Tran
Che-Hung Liu
Annotations
Hybrid search
Ranking
Scalable storage
International Conference on Information and Knowledge Management, Proceedings English The Web contains a large amount of documents and increasingly, also semantic data in the form of RDF triples. Many of these triples are annotations that are associated with documents. While structured query is the principal mean to retrieve semantic data, keyword queries are typically used for document retrieval. Clearly, a form of hybrid search that seamlessly integrates these formalisms to query both documents and semantic data can address more complex information needs. In this paper, we present CE2, an integrated solution that leverages mature database and information retrieval technologies to tackle challenges in hybrid search on the large scale. For scalable storage, CE2 integrates database with inverted indices. Hybrid query processing is supported in CE2 through novel algorithms and data structures, which allow for advanced ranking schemes to be integrated more tightly into the process. Experiments conducted on Dbpedia and Wikipedia show that CE2 can provide good performance in terms of both effectiveness and effciency. 0 0
Can you ever trust a wiki? Impacting perceived trustworthiness in wikipedia Aniket Kittur
Bongwon Suh
Chi E.H.
Collaboration
Social computing
Stability
Trust
Visualisation
Wiki
Wikipedia
English Wikipedia has become one of the most important information resources on the Web by promoting peer collaboration and enabling virtually anyone to edit anything. However, this mutability also leads many to distrust it as a reliable source of information. Although there have been many attempts at developing metrics to help users judge the trustworthiness of content, it is unknown how much impact such measures can have on a system that is perceived as inherently unstable. Here we examine whether a visualization that exposes hidden article information can impact readers' perceptions of trustworthiness in a wiki environment. Our results suggest that surfacing information relevant to the stability of the article and the patterns of editor behavior can have a significant impact on users' trust across a variety of page types. Copyright 2008 ACM. 0 0
Case study-using a wiki to teach information decision science Olson T. Collaboration
Innovation
Textbook
Web 2.0
Wiki
Proceedings of the 7th European Conference on e-Learning, ECEL 2008 English The Carlson School of Management at the University of Minnesota, Information and Decision Science Department (IDSc), has developed a wiki textbook for our undergraduate core curriculum course. Starting in the Spring Semester of 2008, all seven sections, (approximately 400 students), are using the course wiki textbook taught by four different instructors. The web based wiki contains over 200 articles; ranging from classics discussing databases to recent articles describing internet security. The wiki textbook was originally developed to deal with the rapid changes in the information technology field and the challenges of keeping a traditional textbook current. In addition to providing timely content for students, there are a number of other benefits in using a wiki textbook. In the current semester no issues have been reported, and the students have identified several benefits: Students only need to have access to an internet browser Students are able to download and save posted readings There is no text book to purchase and a savings of $180 for each student The total saving for first semester for all students exceeded $70,000. No text book for students to carry Students print-out and carry only the materials required for each The wiki site has fourteen topics or chapters, and each topic is broken down into four sections. Section One has learning objects, topic terminology, key concepts and selective readings, Section Two is primarily case studies or analytic articles on the topic. Section Three (Additional Material) is a place holder for several articles if students want to study the topic further. Section Four (Student Sandbox) is for student interaction and active learning assignments. This paper focuses on the following: Basic setup, Faculty site, Student site, Security (access and monitoring), Content, Searching, Copyrights, Editing, Multiple instructors using the same site, Maintenance In discussion with several book publishers and authorities on using technology in the classroom, the University of Minnesota feels this topic has tremendous potential both for students and educators in the field of technology. 0 0
Categorizing queries by topic directory He M.
Cutler M.
Wu K.
Proceedings - The 9th International Conference on Web-Age Information Management, WAIM 2008 English The categorization of a web user query by topic or category can be used to select useful web sources that contain the required information. In pursuit of this goal, we explore methods for mapping user queries to category hierarchies under which deep web resources are also assumed to be classified. Our sources for these category hierarchies, or directories, are Yahoo! Directory and Wikipedia. Forwarding an unrefined query (in our case a typical fact finding query sent to a question answering system) directly to these directory resources usually returns no directories or incorrect ones. Instead, we develop techniques to generate more specific directory finding queries from an unrefined query and use these to retrieve better directories. Despite these engineered queries, our two resources often return multiple directories that include many incorrect results, i.e., directories whose categories are not related to the query, and thus web resources for these categories are unlikely to contain the required information. We develop methods for selecting the most useful ones. We consider a directory to be useful if web sources for any of its narrow categories are likely to contain the searched for information. We evaluate our mapping system on a set of 250 TREC questions and obtain precision and recall in the 0.8 to 1.0 range. 0 0
Catriple: Extracting triples from wikipedia categories Qiaoling Liu
Kaifeng Xu
Lei Zhang
Haofen Wang
Yiqin Yu
Yue Pan
Lecture Notes in Computer Science English As an important step towards bootstrapping the Semantic Web, many efforts have been made to extract triples from Wikipedia because of its wide coverage, good organization and rich knowledge. One kind of important triples is about Wikipedia articles and their non-isa properties, e.g. (Beijing, country, China). Previous work has tried to extract such triples from Wikipedia infoboxes, article text and categories. The infobox-based and text-based extraction methods depend on the infoboxes and suffer from a low article coverage. In contrast, the category-based extraction methods exploit the widespread categories. However, they rely on predefined properties, which is too effort-consuming and explores only very limited knowledge in the categories. This paper automatically extracts properties and triples from the less explored Wikipedia categories so as to achieve a wider article coverage with less manual effort. We manage to realize this goal by utilizing the syntax and semantics brought by super-sub category pairs in Wikipedia. Our prototype implementation outputs about 10M triples with a 12-level confidence ranging from 47.0% to 96.4%, which cover 78.2% of Wikipedia articles. Among them, 1.27M triples have confidence of 96.4%. Applications can on demand use the triples with suitable confidence. 0 0
Citing Wikipedia. Donna Shaw American Journalism Review The article presents several several scenarios where the online encyclopedia Wikipedia was cited in a newspaper story. The author highlights the editors, the stories, and the editors' explanations as to why Wikipedia was cited as a source. John Leach of the {Arizona} Republic" states that a citation of Wikipedia led to the creation of rules governing its use. Lois Wilson of the {"Star-Gazette"} in Elmira New York states that she decided whether Wikipedia can be used on a case by case basis and that generally reporters use it as a part of their research." 0 0
ClueBot and Vandalism in Wikipedia Jacobi Carter English 0 1
Clustering Documents with Active Learning Using Wikipedia Anna Huang
David N. Milne
Eibe Frank
Ian H. Witten
Data Mining, IEEE International Conference on English Wikipedia has been applied as a background knowledge base to various text mining problems, but very few attempts have been made to utilize it for document clustering. In this paper we propose to exploit the semantic knowledge in Wikipedia for clustering, enabling the automatic grouping of documents with similar themes. Although clustering is intrinsically unsupervised, recent research has shown that incorporating supervision improves clustering performance, even when limited supervision is provided. The approach presented in this paper applies supervision using active learning. We first utilize Wikipedia to create a concept-based representation of a text document, with each concept associated to a Wikipedia article. We then exploit the semantic relatedness between Wikipedia concepts to find pair-wise instance-level constraints for supervised clustering, guiding clustering towards the direction indicated by the constraints. We test our approach on three standard text document datasets. Empirical results show that our basic document representation strategy yields comparable performance to previous attempts; and adding constraints improves clustering performance further by up to 20%. 0 0
Clustering XML documents using closed frequent subtrees: A structural similarity approach Kutty S.
Thanh Tran
Nayak R.
Yanyan Li
Clustering
Frequent Mining
Frequent subtrees
INEX
Structural mining
XML document mining
Lecture Notes in Computer Science English This paper presents the experimental study conducted over the INEX 2007 Document Mining Challenge corpus employing a frequent subtree-based incremental clustering approach. Using the structural information of the XML documents, the closed frequent subtrees are generated. A matrix is then developed representing the closed frequent subtree distribution in documents. This matrix is used to progressively cluster the XML documents. In spite of the large number of documents in INEX 2007 Wikipedia dataset, the proposed frequent subtree-based incremental clustering approach was successful in clustering the documents. 0 0
Co-occurrence network of reuters news Ozgur A.
Cetin B.
Bingol H.
Importance of people
Ranking
Social network
Wikipedia
International Journal of Modern Physics C English Networks describe various complex natural systems including social systems. We investigate the social network of co-occurrence in Reuters-21578 corpus, which consists of news articles that appeared in the Reuters newswire in 1987. People are represented as vertices and two persons are connected if they co-occur in the same article. The network has small-world features with power-law degree distribution. The network is disconnected and the component size distribution has power-law characteristics. Community detection on a degree-reduced network provides meaningful communities. An edge-reduced network, which contains only the strong ties has a star topology. "Importance" of persons are investigated. The network is the situation in 1987. After 20 years, a better judgment on the importance of the people can be done. A number of ranking algorithms, including Citation count and PageRank, are used to assign ranks to vertices. The ranks given by the algorithms are compared against how well a person is represented in Wikipedia. We find up to medium level Spearman's rank correlations. A noteworthy finding is that PageRank consistently performed worse than the other algorithms. We analyze this further and find reasons. 0 0
CoScripter: Automating & sharing how-to knowledge in the enterprise Leshed G.
Haber E.M.
Matthews T.
Lau T.
Automation
Knowledge sharing
Procedural knowledge
Programming-by-demonstration
Scripting
User study
Wiki
Conference on Human Factors in Computing Systems - Proceedings English Modern enterprises are replete with numerous online processes. Many must be performed frequently and are tedious, while others are done less frequently yet are complex or hard to remember. We present interviews with knowledge workers that reveal a need for mechanisms to automate the execution of and to share knowledge about these processes. In response, we have developed the CoScripter system (formerly Koala [11]), a collaborative scripting environment for recording, automating, and sharing web-based processes. We have deployed CoScripter within a large corporation for more than 10 months. Through usage log analysis and interviews with users, we show that CoScripter has addressed many user automation and sharing needs, to the extent that more than 50 employees have voluntarily incorporated it into their work practice. We also present ways people have used CoScripter and general issues for tools that support automation and sharing of how-to knowledge. Copyright 2008 ACM. 0 0
Collaboration in context: Comparing article evolution among subject disciplines in Wikipedia Katherine Ehmann
Andrew Large
And Jamshid Beheshti
Wikipedia
Open source
Encyclopedia
Reference materials
Information sources
Article quality
Article development
This exploratory study examines the relationships between article and Talk page contributions and their effect on article quality in Wikipedia. The sample consisted of three articles each from the hard sciences, soft sciences, and humanities, whose talk page and article edit histories were observed over a five–month period and coded for contribution types. Richness and neutrality criteria were then used to assess article quality and results were compared within and among subject disciplines. This study reveals variability in article quality across subject disciplines and a relationship between Talk page discussion and article editing activity. Overall, results indicate the initial article creator’s critical role in providing a framework for future editing as well as a remarkable stability in article content over time. 0 2
Collaborative e-learning environments enhanced by wiki technologies Ioannis Giannoukos
Ioanna Lykourentzou
Giorgos Mpardis
Vassilis Nikolopoulos
Vassili Loumos
Eleftherios Kayafas
English E-learning environments have met rapid technological advancements in the previous years. Nevertheless, current e-learning techniques do not adequately support student interaction and collaboration, resulting in decreased student progress and motivation. In this paper, a blended technique combining collaborative forums and wiki technologies is proposed. Through collaborative forums, students discuss course related topics assigned by the tutors to produce new educational material. This material is then stored in the wiki platform for further use. The proposed technique was applied on an e-learning course provided by the National Technical University of Athens and its effectiveness was evaluated using student activity data and questionnaire analysis. Results showed that the technique adequately supported teamwork, increasing student motivation and progress while simultaneously producing satisfactory level educational material. 0 0
Collaborative editing for improved usefulness and usability of transcript-enhanced webcasts Munteanu C.
Baecker R.
Penn G.
Automatic speech recognition
Field study
Navigational tools
Text transcripts
Webcasting
Wiki
Conference on Human Factors in Computing Systems - Proceedings English One challenge in facilitating skimming or browsing through archives of on-line recordings of webcast lectures is the lack of text transcripts of the recorded lecture. Ideally, transcripts would be obtainable through Automatic Speech Recognition (ASR). However, current ASR systems can only deliver, in realistic lecture conditions, a Word Error Rate of around 45% - above the accepted threshold of 25%. In this paper, we present the iterative design of a webcast extension that engages users to collaborate in a wiki-like manner on editing the ASR-produced imperfect transcripts, and show that this is a feasible solution for improving the quality of lecture transcripts. We also present the findings of a field study carried out in a real lecture environment investigating how students use and edit the transcripts. Copyright 2008 ACM. 0 0
Collaborative end-user development on handheld devices Ahmadi N.
Repenning A.
Ioannidou A.
Proceedings - 2008 IEEE Symposium on Visual Languages and Human-Centric Computing, VL/HCC 2008 English Web 2.0 has enabled end users to collaborate through their own developed artifacts, moving on from text (e.g., Wikipedia, Blogs) to images (e.g., Flickr) and movies (e.g., YouTube), changing end-user's role from consumer to producer. But still there is no support for collaboration through interactive end-user developed artifacts, especially for emerging handheld devices, which are the next collaborative platform. Featuring fast always-on networks, Web browsers that are as powerful as their desktop counterparts, and innovative user interfaces, the newest generation of handheld devices can run highly interactive content as Web applications. We have created Ristretto Mobile, a Web-compliant framework for running end-user developed applications on handheld devices. The Webbased Ristretto Mobile includes compiler and runtime components to turn end-user applications into Web applications that can run on compatible handheld devices, including the Apple iPhone and Nokia N800. Our paper reports on the technological and cognitive challenges in creating interactive content that runs efficiently and is user accessible on handheld devices. 0 0
Collaborative enterprise integrated modelling Chiara Ghidini
Marco Rospocher
Luciano Serafini
Faatz A.
Barbara Kump
Tobias Ley
Viktoria Pammer
Stefanie Lindstaedt
CEUR Workshop Proceedings English Enterprise modelling focuses on the construction of a structured description, the so-called enterprise model, which represents (a subset of) the aspects relevant to the activity of an enterprise. Nowadays, knowledge engineering provides sophisticated methodologies and tools to support enterprise modelling and production of formal enterprise models. Recently, it has become clearer that enterprise modelling is a collaborative activity; however, most of the modelling techniques and tools devised so far do not consider this collaborative dimension properly. Therefore, we envisage the necessity to expand, extend, and generalize current methodologies in order to exploit this new dimension. In this line, we propose a flexible framework for enterprise modelling that supports an agile collaboration between the actors involved in the modelling activities, without sticking to a pre-defined protocol of interaction. This methodology produces an integrated enterprise model model that, not only contains a structural (formal) description, but it also includes more informal and semi-formal information. We also present a tool, based on a semantic wiki, that supports the proposed methodology. 0 0
Collaborative knowledge semantic graph image search Shieh J.-R.
Yeh Y.-T.
Lin C.-H.
Lin C.-Y.
Wu J.-L.
Keyword Expansion
Re-ranking
Social Network
Proceeding of the 17th International Conference on World Wide Web 2008, WWW'08 English In this paper, we propose a Collaborative Knowledge Semantic Graphs Image Search (CKSGIS) system. It provides a novel way to conduct image search by utilizing the collaborative nature in Wikipedia and by performing network analysis to form semantic graphs for search-term expansion. The collaborative article editing process used by Wikipedia's contributors is formalized as bipartite graphs that are folded into networks between terms. When a user types in a search term, CKSGIS automatically retrieves an interactive semantic graph of related terms that allow users to easily find related images not limited to a specific search term. Interactive semantic graph then serve as an interface to retrieve images through existing commercial search engines. This method significantly saves users' time by avoiding multiple search keywords that are usually required in generic search engines. It benefits both naive users who do not possess a large vocabulary and professionals who look for images on a regular basis. In our experiments, 85% of the participants favored CKSGIS system rather than commercial search engines. 0 0
Collaborative spaces and user-generated content in a lifelong learning program Koracak B.V. Blogs
Collaboration
Lifelong learning
Social
Social software
Web 2.0
Wiki
MCCSIS'08 - IADIS Multi Conference on Computer Science and Information Systems; Proceedings of e-Learning 2008 English Within the CARNet E-learning Academy - a mixed-mode educational program on e-learning consisting of three specialised certificates - a set of resources has been developed in a wiki, in addition to the regular LMS learning environment. The initial plan was to develop a collaborative space in which the Academy participants could explore various e-learning tools and build, use, develop and expand this common testing laboratory. However, over four offerings of the programs the resources proved to be underused. This paper will focus on the resources, tools and activities within the E-learning Toolkit, analyse the results of their use and suggest conclusions and new developments that should be undertaken. 0 0
Combining multiple resources to build reliable wordnets Fiser D.
Sagot B.
Lecture Notes in Computer Science English This paper compares automatically generated sets of synonyms in French and Slovene wordnets with respect to the resources used in the construction process. Polysemous words were disambiguated via a five-language word-alignment of the SEERA.NET parallel corpus, a subcorpus of the JRC Acquis. The extracted multilingual lexicon was disambiguated with the existing wordnets for these languages. On the other hand, a bilingual approach sufficed to acquire equivalents for monosemous words. Bilingual lexicons were extracted from different resources, including Wikipedia, Wiktionary and EUROVOC thesaurus. A representative sample of the generated synsets was evaluated against the goldstandards. 0 0
Combining semantic wikis and controlled natural language Kuhn T. AceWiki
Attempto Controlled English (ACE)
Controlled Natural Language
Ontology
Semantic web
Semantic wiki
CEUR Workshop Proceedings English We demonstrate AceWiki that is a semantic wiki using the controlled natural language Attempto Controlled English (ACE). The goal is to enable easy creation and modification of ontologies through the web. Texts in ACE can automatically be translated into first-order logic and other languages, for example OWL. Previous evaluation showed that ordinary people are able to use AceWiki without being instructed. 0 0
Combining structure and semantics for ontology-based corporate wikis Alexandre Passant
Laublet P.
Architecture of participation
Collective intelligence
Ontology
Semantic web
Web 2.0
Wiki
Lecture Notes in Business Information Processing English While wikis offer new means to collaboratively build, organize and share knowledge in organizations, such knowledge cannot be easily understood by computers in a query answering or reusability process. This paper details the features and architecture of a wiki-farm system that combines structure and semantics in order to collaboratively produce ontology-based data and immediately reuse it in wiki pages to enrich browsing and querying capabilities of the system. 0 0
Combining wikipedia and newswire texts for question answering in spanish De Pablo-Sanchez C.
Martinez-Fernandez J.L.
Gonzalez-Ledesma A.
Samy D.
Martinez P.
Moreno-Sandoval A.
Al-Jumaily H.
Lecture Notes in Computer Science English This paper describes the adaptations of the MIRACLE group QA system in order to participate in the Spanish monolingual question answering task at QA@CLEF 2007. A system, initially developed for the EFE collection, was reused for Wikipedia. Answers from both collections were combined using temporal information extracted from questions and collections. Reusing the EFE subsystem has proven not feasible, and questions with answers only in Wikipedia have obtained low accuracy. Besides, a co-reference module based on heuristics was introduced for processing topic-related questions. This module achieves good coverage in different situations but it is hindered by the moderate accuracy of the base system and the chaining of incorrect answers. 0 0
Communication key as Mapa gets a wiki: "non-techie" staff unfazed by IT innovation Human Resource Management International Digest English 0 0
Comparative studies on enterprise knowledge sharing and wiki 2008 IEEE International Symposium on Knowledge Acquisition and Modeling Workshop Proceedings, KAM 2008 English 0 0
Comparison of Wikipedia and other encyclopedias for accuracy, breadth, and depth in historical articles Lucy Holman Rector Encyclopedia
Reference services
Reference Services Review English Purpose – This paper seeks to provide reference librarians and faculty with evidence regarding the comprehensiveness and accuracy of Wikipedia articles compared with respected reference resources.

Design/methodology/approach – This content analysis evaluated nine Wikipedia articles against comparable articles in Encyclopaedia Britannica, The Dictionary of American History and American National Biography Online in order to compare Wikipedia's comprehensiveness and accuracy. The researcher used a modification of a stratified random sampling and a purposive sampling to identify a variety of historical entries and compared each text in terms of depth, accuracy, and detail.

Findings – The study did reveal inaccuracies in eight of the nine entries and exposed major flaws in at least two of the nine Wikipedia articles. Overall, Wikipedia's accuracy rate was 80 percent compared with 95-96 percent accuracy within the other sources. This study does support the claim that Wikipedia is less reliable than other reference resources. Furthermore, the research found at least five unattributed direct quotations and verbatim text from other sources with no citations.

Research limitations/implications – More research must be undertaken to analyze Wikipedia entries in other disciplines in order to judge the source's accuracy and overall quality. This paper also shows the need for analysis of Wikipedia articles' histories and editing process.

Practical implications – This research provides a methodology for further content analysis of Wikipedia articles.

Originality/value – Although generalizations cannot be made from this paper alone, the paper provides empirical data to support concerns regarding the accuracy and authoritativeness of Wikipedia.
0 4
Concept-based feature generation and selection for information retrieval Egozi O.
Evgeniy Gabrilovich
Shaul Markovitch
Proceedings of the National Conference on Artificial Intelligence English Traditional information retrieval systems use query words to identify relevant documents. In difficult retrieval tasks, however, one needs access to a wealth of background knowledge. We present a method that uses Wikipedia-based feature generation to improve retrieval performance. Intuitively, we expect that using extensive world knowledge is likely to improve recall but may adversely affect precision. High quality feature selection is necessary to maintain high precision, but here we do not have the labeled training data for evaluating features, that we have in supervised learning. We present a new feature selection method that is inspired by pseudo-relevance feedback. We use the top-ranked and bottom-ranked documents retrieved by the bag-of-words method as representative sets of relevant and non-relevant documents. The generated features are then evaluated and filtered on the basis of these sets. Experiments on TREC data confirm the superior performance of our method compared to the previous state of the art. Copyright © 2008, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. 0 0
Concurrency awareness in a P2P Wiki system 2008 International Symposium on Collaborative Technologies and Systems, CTS'08 English 0 0
Conflictual consensus in the chinese version of Wikipedia Liao H.-T. International Symposium on Technology and Society, Proceedings English The paper examines how the recent development of the Chinese version of Wikipedia (CW) has developed to accommodate the diverse regional differences of its contributors. Although contributors are all users of the Chinese language, the orthographic, linguistic, regional and political differences among them do exist. Thus, CW has to attend to the different needs of users from four regions of origin (Mainland, Hong Kong/Macau, Taiwan, and Singapore/Malaysia). The paper shows how a technological polity is built, with an aim to accommodate regional diversity, by importing Wikipedia governance principles, implementing user-generated character conversion, and establishing the "Avoid Region-Centric Policy". It has been observed that although the orthographic and lexical differences have been preserved and respected, the offline political and ideological differences seem to threaten its potential growth, especially when compared to its self-censored copycat Baidu Baike. This paper then suggests it is neither the internal conflicts nor the external competition that matters most to CW, but rather the evolution of its polity 0 0
Constructing a Global Ontology by Concept Mapping using Wikipedia Thesaurus Minghua Pei
Kotaro Nakayama
Takahiro Hara
Sojiro Nishio
Data mining
Association Thesaurus
Ontology Mapping
Global Ontology
International Symposium on Mining And Web (IEEE MAW) conjunction with IEEE AINA 0 0
Constructing a global ontology by concept mapping using Wikipedia thesaurus Minghua Pei
Kotaro Nakayama
Takahiro Hara
Shojiro Nishio
Proceedings - International Conference on Advanced Information Networking and Applications, AINA English Recently, the importance of semantics on the WWW is widely recognized and a lot of semantic information (RDF, OWL etc.) is being built/published on the WWW. However, the lack of ontology mappings becomes a serious problem for the Semantic Web since it needs well defined relations to retrieve information correctly by inferring the meaning of information. One to one mapping is not an efficient method due to the nature of distributed environment. Therefore, it would be a considerable method to map the concepts by using a large-scale intermediate ontology. On the other hand, Wikipedia is a large-scale of concept network covering almost all concepts in the real world. In this paper, we propose an intermediate ontology construction method using Wikipedia Thesaurus, an association thesaurus extracted from Wikipedia. Since Wikipedia Thesaurus provides associated concepts without explicit relation type, we propose an approach of concept mapping using two sub methods; "name mapping" and "logic-based mapping". 0 0
Context and keyword extraction in plain text using a graph representation Chahine C.A.
Nathalie Chaignaud
Kotowicz J.P.
Pecuchet J.P.
SITIS 2008 - Proceedings of the 4th International Conference on Signal Image Technology and Internet Based Systems English Document indexation is an essential task achieved by archivists or automatic indexing tools. To retrieve relevant documents to a query, keywords describing this document have to be carefully chosen. Archivists have to find out the right topic of a document before starting to extract the keywords. For an archivist indexing specialized documents, experience plays an important role. But indexing documents on different topics is much harder. This article proposes an innovative method for an indexing support system. This system takes as input an ontology and a plain text document and provides as output contextualized keywords of the document. The method has been evaluated by exploiting Wikipedia's category links as a termino-ontological resources. 0 0
Contextualized Knowledge Acquisition in a Personal Semantic Wiki Ludger Elst
Malte Kiesel
Sven Schwarz
Georg Buscher
Andreas Lauer
Andreas Dengel
EKAW English 0 0
Copernicus: 3D Wikipedia Jacek Jankowski Design
Web3D
Wikipedia
SIGGRAPH English 0 0
Coreference resolution for questions and answer merging by validation Sven Hartrumpf
Glockner I.
Johannes Leveling
Lecture Notes in Computer Science English For its fourth participation at QA@CLEF, the German question answering (QA) system InSicht was improved for CLEF 2007 in the following main areas: questions containing pronominal or nominal anaphors are treated by a coreference resolver; the shallow QA methods are improved; and a specialized module is added for answer merging. Results showed a performance drop compared to last year mainly due to problems in handling the newly added Wikipedia corpus. However, dialog treatment by coreference resolution delivered very accurate results so that follow-up questions can be handled similarly to isolated questions. 0 0
Corpus Co-Occurrence, Dictionary and Wikipedia Entries as Resources for Semantic Relatedness Information Sabine Schulte im Walde Michael Roth LREC'08 Distributional, corpus-based descriptions have frequently been applied to model aspects of word meaning. However, distributional models that use corpus data as their basis have one well-known disadvantage: even though the distributional features based on corpus co-occurrence were often successful in capturing meaning aspects of the words to be described, they generally fail to capture those meaning aspects that refer to world knowledge, because coherent texts tend not to provide redundant information that is presumably available knowledge. The question we ask in this paper is whether dictionary and encyclopaedic resources might complement the distributional information in corpus data, and provide world knowledge that is missing in corpora. As test case for meaning aspects, we rely on a collection of semantic associates to German verbs and nouns. Our results indicate that a combination of the knowledge resources should be helpful in work on distributional descriptions. 0 0
Creating and consuming web 2.0 in art education Buffington M.L. Art education
Blogs
Creating Web 2.0
Digital technologies
Podcasts
Social bookmarking
Technology in art education
Web 2.0
Wiki
Computers in the Schools English Web 2.0 has the potential to dramatically affect educational practices in a variety of ways. Within the field of art education, students and teachers can utilize Web 2.0 to further student understanding and art-making. The potential for students to create the Web is an aspect of Web 2.0 that is particularly well suited for an art classroom. Through examples from teachers and from the author's own practice, this article details potential uses of delicious, flickr, blogs, podcasts, and wiki. Through these technologies, students may collaboratively build knowledge, develop a deeper understanding of their own artworks and those of other artists, and interact with artworks in new ways. © 2008 by The Haworth Press. All rights reserved. 0 0
Creating critical mass in collaboration systems: Insights from wikipedia Pattarawan Prasarnphanich
Christian Wagner
Collaborative knowledge creation
Collective action
Critical mass
Digital ecosystem
2008 2nd IEEE International Conference on Digital Ecosystems and Technologies, IEEE-DEST 2008 English Digital ecosystems that rely on peer production, where users are consumers as well as producers of information and knowledge, are becoming increasingly popular and viable. Supported by Web 2.0 technologies such as wikis, these systems have the potential to replace existing knowledge management systems which generally rely on a small group of experts. The fundamental question for all such systems is under which conditions, the collective acts of knowledge contribution are started and become self-sustaining? Our article addresses this question, using Wikipedia as an exemplary system. Through a collective action framework, we apply critical mass theory to explain emergence and sustainability of the peer production approach. 0 1
Creating wiki communities in blended learning environment and the creation of the Moodle New Wiki Proceedings of the 2008 International Conference on Frontiers in Education: Computer Science and Computer Engineering, FECS 2008 English 0 0
Critical assessment of using wikis in legal education E. Hoorn
D. van Hoorn
JILT-Journal of Information Law & Technology Wikis serve to support collaborative writing on the Web. The best known example of a wiki is Wikipedia, an open encyclopedia on the web. This paper is meant to explore possible uses of a wiki-environment in legal education. Firstly, it describes the actual use of a closed wiki environment in a class on Cybercrime in the Netherlands. Secondly, the paper explores the possibilities for international collaboration of students without face to face contact. Innovative use of wikis in learning situations demands insights in educational design and best practices of educators. We will show that for students as well as educators the use of a wiki is an easy and effective way of using technology in order to get engaged in new forms of learning. The paper is intended for legal educators who share an interest in innovative approaches to legal education. 0 0
Critical literacy and web 2.0: Exercising and negotiating power McLeod J.
Vasinda S.
Classroom examples
Critical literacy
Definitions
Dynamic
Multiple perspectives
Power
Web 2.0
Wiki
Computers in the Schools English In this article, the authors describe Web 2.0 as tools that have increased the urgency for students' and teachers' critical literacy skills and have also participated in the implementation of critical literacy. The authors define and position both Web 2.0 and critical literacy. Further, students' and teachers' power dynamics within both critical literacy and Web 2.0 are explored. Examples of combining Web 2.0 and critical literacy illuminate the critical literacy principles in the context of schools. © 2008 by The Haworth Press. All rights reserved. 0 0
Cross-language retrieval with wikipedia Schonhofen P.
Benczur A.
Biro I.
Csalogany K.
Lecture Notes in Computer Science English We demonstrate a twofold use of Wikipedia for cross-lingual information retrieval. As our main contribution, we exploit Wikipedia hyperlinkage for query term disambiguation. We also use bilingual Wikipedia articles for dictionary extension. Our method is based on translation disambiguation; we combine the Wikipedia based technique with a method based on bigram statistics of pairs formed by translations of different source language terms. 0 0
Cross-lingual blog analysis based on multilingual blog distillation from multilingual wikipedia entries Mariko Kawaba
Hiroyuki Nakasaki
Takehito Utsuro
Tomohiro Fukuhara
ICWSM 2008 - Proceedings of the 2nd International Conference on Weblogs and Social Media English The goal of this paper is to cross-lingually analyze multilingual blogs collected with a topic keyword. The framework of collecting multilingual blogs with a topic keyword is designed as the blog distillation (feed search) procedure. Mulitlingual queries for retrieving blog feeds are created from Wikipedia entries. Finally, we cross-lingually and cross-culturally compare less well known facts and opinions that are closely related to a given topic. Preliminary evaluation results support the effectiveness of the proposed framework. Copyright © 2008, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. 0 0
Cross-lingual blog analysis by cross-lingual comparison of characteristic terms and blog posts Hiroyuki Nakasaki
Mariko Kawaba
Takehito Utsuro
Tomohiro Fukuhara
Hiroshi Nakagawa
Kando N.
Proceedings of the 2nd International Symposium on Universal Communication, ISUC 2008 English The goal of this paper is to cross-lingually analyze multilingual blogs collected with a topic keyword. The framework of collecting multilingual blogs with a topic keyword is designed as the blog feed retrieval procedure. Multilingual queries for retrieving blog feeds are created from Wikipedia entries. Finally, we cross-lingually and crossculturally compare less well known facts and opinions that are closely related to a given topic. Preliminary evaluation results support the effectiveness of the proposed framework. 0 0
Crowdsourcing user studies with Mechanical Turk Aniket Kittur
Chi E.H.
Bongwon Suh
Mechanical Turk
Micro task
Remote user study
Wikipedia
Conference on Human Factors in Computing Systems - Proceedings English User studies are important for many aspects of the design process and involve techniques ranging from informal surveys to rigorous laboratory studies. However, the costs involved in engaging users often requires practitioners to trade off between sample size, time requirements, and monetary costs. Micro-task markets, such as Amazon's Mechanical Turk, offer a potential paradigm for engaging a large number of users for low time and monetary costs. Here we investigate the utility of a micro-task market for collecting user measurements, and discuss design considerations for developing remote micro user evaluation tasks. Although micro-task markets have great potential for rapidly collecting user measurements at low costs, we found that special care is needed in formulating tasks in order to harness the capabilities of the approach. Copyright 2008 ACM. 0 0
Dealing with conflicting modifications in a Wiki WEBIST 2008 - 4th International Conference on Web Information Systems and Technologies, Proceedings English 0 0
Decoding Wikipedia Categories for Knowledge Acquisition Vivi Nastase
Michael Strube
English 0 0
Decoding Wikipedia categories for knowledge acquisition Vivi Nastase
Michael Strube
Proceedings of the National Conference on Artificial Intelligence English This paper presents an approach to acquire knowledge from Wikipedia categories and the category network. Many Wikipedia categories have complex names which reflect human classification and organizing instances, and thus encode knowledge about class attributes, taxonomic and other semantic relations. We decode the names and refer back to the network to induce relations between concepts in Wikipedia represented through pages or categories. The category structure allows us to propagate a relation detected between constituents of a category name to numerous concept links. The results of the process are evaluated against ResearchCyc and a subset also by human judges. The results support the idea that Wikipedia category names are a rich source of useful and accurate knowledge. Copyright © 2008, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. 0 0
Delta-encoding for document revision control system of Wikipedia Khairulmizam Samsudin Wan Muhammad Salehuddin Wan Hassan Revision control
Document revision
Delta-encoding
Sixth IEEE Student Conference on Research and Development (SCOReD 2008) A revision control system keep track of changes for multiple versions of the same unit of information. It is often used in engineering and software development to manage storing, retrieval, logging, identification and merging of source files and electronic documents. Changes to these documents are noted by incrementing an associated number or letter code and associated historically with the person making the change. Revision control system is an important component of collaborative software platform that allows several member of a development team to work concurrently on an electronic document. Wikipedia, a free content encyclopedia is an example of a successful application of collaborative technology. A poorly implemented document revision control system will affect the performance and cause difficulty in managing Wikipedia huge amount of electronic data. In this work, efficiency of the current revision control system of Wikipedia will be evaluated. Feasibility of delta-encoding to address the current limitation of Wikipedia document revision control system will be presented. 0 0
Design and code reviews in the age of the internet Meyer B. Communications of the ACM English Eiffel Software has applied code reviews while developing EiffelStudio, a large integrated development environment (IDE), which supports Web-based code review. EiffelStudio applies the seamless development principle, treating specification, design, and analysis as a continuum rather than as a sequence of separate steps. The tools needed to support code distributed reviews include voice communication similar to a conference call, written communication, Google Docs for shared documents, and WebEx sharing tool for sharing screens. The standard-review page structure includes choice of abstractions, other aspects of API design and architecture, contracts, programming style, global comments, and adherence to official coding practices. The EiffelStudio team ensures that the review results are made available as pat of the EiffelStudio open-source development site based on WiKi pages. 0 0
Design and implementation of a secure mobile wiki system Costantinos Kolias
Stefanos Demertzis
Georgios Kambourakis
SSL
XML
Mobile systems
Security
Wiki
WBE English 0 0
Developing a Pain Management Resource Wiki for Cancer Patients and Their Caregivers Linda Hasman
Deborah Chiarella
Journal of Consumer Health On the Internet English Utilizing Web 2.0 technologies such as blogs, wikis, and social networking sites can provide innovative and creative means to communicate with library patrons. They can also serve as a purposeful information repository for individuals looking for information on a particular topic. A pain management resource for cancer patients and their caregivers was created to provide a concise one-stop location for this community to access the latest and most reliable consumer health information regarding this topic. The process of planning, organizing, building, and maintaining the wiki is chronicled in this article. Patient education material in wiki format can serve as a valuable information tool for others interested in pursuing such technology in other libraries. Patient education material in wiki format can serve as a valuable information tool for others interested in pursuing such technology in other libraries. 0 0
Developing virtual simulation game for authentic learning: Realizing partnership between university and industry Starcic A.I. Business simulation
Constructivism
Hephaistos simulation software
Higher education
Lifelong learning
Virtual community
Virtual simulation game
Wiki
WSEAS Transactions on Communications English The paper describes the development of virtual simulation-game-based learning environment. The key objective is to offer students and lifelong learners simulation-game environment that preserves real life authenticity. Presented is a simulation-game tailored for diversified student groups in graduate courses and employees in the tourism sector, aiming at giving them competences for international and global business management. The constructivism theory provides the framework of the simulation-game pedagogical approaches: learning by doing, learning with live experiences, experiential learning, problem-based learning, and learning in safety. The paper presents the partnership between university and industry in the design process of simulation-game and virtual community for assisting collaboration among users: entrepreneurs, students, teachers and trainers. Focus groups were formed in the design process of simulation-game, engaging all relevant stakeholders, and needs assessment was performed at the beginning of the project. 0 0
Development of mechatronics educational material applying "Wiki" 2008 International Conference on Control, Automation and Systems, ICCAS 2008 English 0 0
Die Schöne und das Tier: Semantic Web und Wikis Thomas Tunsch EVA 2008 Berlin German Although to a large extent the Semantic Web specifies fundamentals and future potentials of the WWW, it is associated with current projects as well. With standards like the CIDOC Conceptual Reference Model for the domain of cultural heritage, main principles for the semantic network are available already today.

In contrast, Wikis seem to exemplify the danger of subjectivity and absence of verification for many experts in museums, especially due to general participation beyond traditional areas of expertise.

Semantic MediaWiki is an effective tool for converting seeming contradictions into a prolific challenge for forward-looking international collaboration.
0 0
Disconnected in a connected world Karpinski J.L. Blogs
Medical libraries
Podcasts
RSS feeds
Social tagging
User generated content
Web 2.0
Wiki
Medical Reference Services Quarterly English This article outlines five Web 2.0 resources and looks at the use of these tools among medical and nursing professionals and students at the Hospital, Medical School, and Nursing School of the University of Pennsylvania. Questionnaires showed that a majority of the individuals surveyed were unfamiliar with Web 2.0 resources. Additional respondents recognized the tools but did not use them in a medical or nursing context, with a minimal number using any tools to expand their medical or nursing knowledge. A lack of time to set up and use the resources, difficulty of set-up and use, skepticism about the quality of user-generated medical content, and a lack of perceived need for Web 2.0 resources contributed substantially to non-use. The University of Pennsylvania Biomedical Library is responding by increasing the availability of basic, quick, and easy-to-use instructional materials for selected Web 2.0 resources. 0 0
Discovering behavioral patterns in collective authorship of place-based information Darren Hardy Geotagging
Peer production
Wikipedia
Robot
Internet Research 9.0: Rethinking Community, Rethinking Place While current GIS research has focused on technological issues of visualization and data organization, the emergence of new forms of collective authorship suggest we need new information frameworks and behaviors. How do individuals contribute place-based information to a digital commons? What are the authorship dynamics of such collective effort? For my research, I will use spatial data mining methods to characterize authorship behavior on a corpus of 1 million geotagged articles across 20 languages from Wikipedia. 0 0
Discovering semantics in multimedia content using Wikipedia Angela Fogarolli
Marco Ronchetti
Content retrieval and filtering: Search over semi-structural Web sources
E-Learning
Multimedia
Wikipedia
Lecture Notes in Business Information Processing English Semantic-based information retrieval is an area of ongoing work. In this paper we present a solution for giving semantic support to multimedia content information retrieval in an e-Learning environment where very often a large number of multimedia objects and information sources are used in combination. Semantic support is given through intelligent use of Wikipedia in combination with statistical Information Extraction techniques. 0 0
Distinguishing Between Instances and Classes in the Wikipedia Taxonomy Cäcilia Zirn
Vivi Nastase
Michael Strube
English 0 0
Document Summarization using Wikipedia Krishnan Ramanathan
Yogesh Sankarasubramaniam
Nidhi Mathur
Ajay Gupta
Document summarization
Wikipedia
ROUGE
First IEEE international conference on Human computer interaction (IHCI) Although most of the developing world is likely to first access the Internet through mobile phones, mobile devices are constrained by screen space, bandwidth and limited attention span. Single document summarization techniques have the potential to simplify information consumption on mobile phones by presenting only the most relevant information contained in the document. In this paper we present a language independent single-document summarization method. We map document sentences to semantic concepts in Wikipedia and select sentences for the summary based on the frequency of the mapped-to concepts. Our evaluation on English documents using the ROUGE package indicates our summarization method is competitive with the state of the art in single document summarization. 0 0
Document and query expansion models for blog distillation Arguello J.
Elsas J.L.
Yoo C.
Callan J.
Carbonell J.G.
NIST Special Publication English This paper presents the CMU submission to the 2008 TREC blog distillation track. Similar to last year's experiments, we evaluate different retrieval models and apply a query expansion method that leverages the link structure in Wikipedia. We also explore using a corpus that combines several different representations of the documents, using both the feed XML and permalink HTML, and apply initial experiments with spam filtering. 0 0
Document clustering using incremental and pairwise approaches Thanh Tran
Nayak R.
Bruza P.
Clustering
Content
INEX 2007
Structure
XML
Lecture Notes in Computer Science English This paper presents the experiments and results of a clustering approach for clustering of the large Wikipedia dataset in the INEX 2007 Document Mining Challenge. The clustering approach employed makes use of an incremental clustering method and a pairwise clustering method. The approach enables us to perform the clustering task on a large dataset by first reducing the dimension of the dataset to an undefined number of clusters using the incremental method. The lower-dimension dataset is then clustered to a required number of clusters using the pairwise method. In this way, clustering of the large number of documents is performed successfully and the accuracy of the clustering solution is achieved. 0 0
Documenting service-oriented architectures with ontobrowse semantic wiki Happel H.-J.
Seedorf S.
CEUR Workshop Proceedings English Documenting and maintaining an enterprise-wide service-oriented architecture (SOA) causes a substantial management effort which should be addressed by intelligent, scalable solutions. A particular challenge is that business experts, developers and software architects take different perspectives on a SOA, which leads towards a scattering of architectural information throughout various formats and sources. Ontobrowse Semantic Wiki specifically addresses these issues by providing an ontology-based integration and documentation platform for architectural information. In this paper, we first identify the main issues arising from documentation and maintenance by introducing the case of an enterprise SOA in an insurance company. Then we give a detailed description of the Ontobrowse approach and its implementation. At last, we explain how ontologies, artifact mappings and architectural rules are created to support the Enterprise SOA case. 0 0
Does WIKIPEDIA provide evidence-based health care information? A content analysis Ingrid Muhlhauser
Friederike Oser
Fortbildung Und Qualitat Im Gesundheitswesen Zeitschrift Fuhr Evidenz Patients and consumers are increasingly searching the Internet for medical and healthcare information. Using the criteria of evidence-based medicine the present study analyses the websites of Wikipedia and two major German statutory health insurances for content and presentation of patient information. 22 senior students of health sciences and education evaluated one topic each. In a first step, they identified the evidence for their specific question. Afterwards they used their results as reference for the evaluation of the three websites. Using a check list each student and a second researcher independently rated content and presentation of the information offered. All these websites failed to meet relevant criteria, and key information such as the presentation of probabilities of success on patient-relevant outcomes, probabilities of unwanted effects, and unbiased risk communication was missing. On average items related to the objectives of interventions, the natural course of disease and treatment options were only rated as partially fulfilled". Overall there were only minor differences between the three providers except for items related to the specific nature of the websites such as disclosure of authorship conflict of interest and support offers. In addition the Wikipedia information tended to achieve lower comprehensibility. In conclusion the quality of the healthcare information provided by Wikipedia and two major German statutory health insurances is comparable. They do not meet important criteria of evidence-based patient and consumer information though." 0 0
Does Wikipedia information help Netflix predictions? John Lees-Miller
Fraser Anderson
Bret Hoehn
Russell Greiner
Proceedings - 7th International Conference on Machine Learning and Applications, ICMLA 2008 English We explore several ways to estimate movie similarity from the free encyclopedia Wikipedia with the goal of improving our predictions for the Netflix Prize. Our system first uses the content and hyperlink structure of Wikipedia articles to identify similarities between movies. We then predict a user's unknown ratings by using these similarities in conjunction with the user's known ratings to initialize matrix factorization and k-Nearest Neighbours algorithms. We blend these results with existing ratings-based predictors. Finally, we discuss our empirical results, which suggest that external Wikipedia data does not significantly improve the overall prediction accuracy. 0 0
Does Wikipedia provide evidence based health care information? A content analysis Ingrid Muhlhauser
Friederike Oser
Zeitschrift für Evidenz, Fortbildung und Qualität im Gesundheitswesen English Patients and consumers are increasingly searching the Internet for medical and healthcare information. Using the criteria of evidence-based medicine the present study analyses the websites of Wikipedia and two major German statutory health insurances for content and presentation of patient information. 22 senior students of health sciences and education evaluated one topic each. In a first step, they identified the evidence for their specific question. Afterwards they used their results as reference for the evaluation of the three websites. Using a check list each student and a second researcher independently rated content and presentation of the information offered. All these websites failed to meet relevant criteria, and key information such as the presentation of probabilities of success on patient-relevant outcomes, probabilities of unwanted effects, and unbiased risk communication was missing. On average items related to the objectives of interventions, the natural course of disease and treatment options were only rated as “partially fulfilled.” Overall, there were only minor differences between the three providers, except for items related to the specific nature of the web sites such as disclosure of authorship, conflict of interest and support offers. In addition, the Wikipedia information tended to achieve lower comprehensibility. In conclusion, the quality of the healthcare information provided by Wikipedia and two major German statutory health insurances is comparable. They do not meet important criteria of evidence-based patient and consumer information though. 0 1
Don't look now, but we've created a bureaucracy: the nature and roles of policies and rules in Wikipedia Brian Butler
Elisabeth Joyce
Jacqueline Pike
Collaboration
Community
Dynamics
Policy
Rules
Wikipedia
Wiki
Conference on Human Factors in Computing Systems English Wikis are sites that support the development of emergent, collective infrastructures that are highly flexible and open, suggesting that the systems that use them will be egalitarian, free, and unstructured. Yet it is apparent that the flexible infrastructure of wikis allows the development and deployment of a wide range of structures. However, we find that the policies in Wikipedia and the systems and mechanisms that operate around them are multi-faceted. In this descriptive study, we draw on prior work on rules and policies in organizations to propose and apply a conceptual framework for understanding the natures and roles of policies in wikis. We conclude that wikis are capable of supporting a broader range of structures and activities than other collaborative platforms. Wikis allow for and, in fact, facilitate the creation of policies that serve a wide variety of functions. 11 5
Dublin City University at CLEF 2007: Cross-Language Speech Retrieval Experiments Ying Zhang
Gareth J. Jones
Ke Zhang
Advances in Multilingual and Multimodal Information Retrieval The Dublin City University participation in the {CLEF} 2007 {CL-SR} English task concentrated primarily on issues of topic translation. Our retrieval system used the {BM25F} model and pseudo relevance feedback. Topics were translated into English using the Yahoo! {BabelFish} free online service combined with domain-specific translation lexicons gathered automatically from Wikipedia. We explored alternative topic translation methods using these resources. Our results indicate that extending machine translation tools using automatically generated domain-specific translation lexicons can provide improved {CLIR} effectiveness for this task. 0 0
Dublin City University at CLEF 2007: Cross-language speech retrieval experiments YanChun Zhang
Jones G.J.F.
Zhang K.
Lecture Notes in Computer Science English The Dublin City University participation in the CLEF 2007 CL-SR English task concentrated primarily on issues of topic translation. Our retrieval system used the BM25F model and pseudo relevance feedback. Topics were translated into English using the Yahoo! BabelFish free online service combined with domain-specific translation lexicons gathered automatically from Wikipedia. We explored alternative topic translation methods using these resources. Our results indicate that extending machine translation tools using automatically generated domain-specific translation lexicons can provide improved CLIR effectiveness for this task. 0 0
Dynamic element retrieval in the wikipedia collection Crouch C.J.
Crouch D.B.
Kamat N.
Malik V.
Mone A.
Lecture Notes in Computer Science English This paper describes the successful adaptation of our methodology for the dynamic retrieval of XML elements to a semi-structured environment. Working with text that contains both tagged and untagged elements presents particular challenges in this context. Our system is based on the Vector Space Model; basic functions are performed using the Smart experimental retrieval system. Dynamic element retrieval requires only a single indexing of the document collection at the level of the basic indexing node (i.e., the paragraph). It returns a rank-ordered list of elements identical to that produced by the same query against an all-element index of the collection. Experimental results are reported for both the 2006 and 2007 Ad-hoc tasks. 0 0
Dynamic perspectives on social characteristics and sustainability in online community networks P. Otto
M. Simon
System Dynamics Review Online community networks can help organizations improve collaboration. However, in spite of their potential value, there has been little empirical research into two important network factors that determine their success: social characteristics of users and changes in operations that result from network evolution. Our research addresses these deficiencies by using a cultural framework. Derived from anthropology, it extends previous system dynamics research on online community networks. The framework acts as a lens, enabling a better understanding of the effects that changes in these factors bring to online community networks. Using data collected from Wikipedia for model calibration, our findings suggest that, contrary to conventional wisdom, removing policies that focus on building group commitment does not lower performance. The results also show that online networks need structural control, otherwise their attractiveness, credibility and, subsequently, content value might all decrease. To ensure sustainability the network must be monitored, especially during the early stages of its evolution, so that rules and regulations that ensure value and validity can be selectively employed. Copyright 2008 John Wiley Sons, Ltd. 0 1
Dynamics of platform-based markets F. Zhu Harvard University, Massachusetts English Platform-based markets are prevalent in today's economy. Understanding the driver of platform success is of critical importance for platform providers. In this dissertation, I first develop a dynamic model to characterize conditions under which different factors drive the success of a platform, and then use the theoretical framework to analyze market-level data from the video game industry. I find that game players' marginal utility decreases rapidly with additional games after the number of games reaches a certain point, and quality is more influential than indirect network effects in driving the success of video game consoles. I also use individual-level data from Chinese Wikipedia to examine contributors' incentives to contribute. I take advantage of China's block of Chinese Wikipedia in mainland China in 2005 as a natural experiment to establish the causal relationship between contributors' incentives to contribute and the number of the beneficiaries of their contributions. I find that while on average contributors' incentives to contribute drop significantly after the block, the contribution levels of those contributors with small collaboration networks do not decrease after the block. In addition, these contributors join Wikipedia significantly earlier than the average contributor. The results suggest that other market factors such as altruism could be more influential than indirect network effects in encouraging user participation in the early stage of Chinese Wikipedia. The overall research casts doubt on the popular belief that indirect network effects are the primary force driving platform success and suggests that in many cases, other market forces could be dominant. Late movers could therefore take over market leaderships by exploiting these market forces. 0 0
EBay, Wikipedia, and the future of the footnote Margaret M. Knapp Theatre History Studies 0 0
EFS: Expert finding system based on wikipedia link pattern analysis Yang K.-H.
Chen C.-Y.
Lee H.-M.
Ho J.-M.
Automatic term recognition
Expert finding
Wikipedia
Conference Proceedings - IEEE International Conference on Systems, Man and Cybernetics English Building an expert finding system is very important for many applications especially in the academic environment. Previous work uses e-mails or web pages as corpus to analyze the expertise for each expert. In this paper, we present an Expert Finding System, abbreviated as EFS to build experts' profiles by using their journal publications. For a given proposal, the EFS first looks up the Wikipedia web site to get relative link information, and then list and rank all associated experts by using those information. In our experiments, we use a real-world dataset which comprises of 882 people and 13,654 papers, and are categorized into 9 expertise domains. Our experimental results show that the EFS works well on several expertise domains like "Artificial Intelligence" and "Image & Pattern Recognition" etc. 0 0
EGovernment 2.0 - How can government benefit from web 2.0? Ostergaard S.D.
Hvass M.
Autonomous IT
Blogging
Citizen self service
EGovernment
Social computing
Social inclusion
Web 2.0
Wiki
WMSCI 2008 - The 12th World Multi-Conference on Systemics, Cybernetics and Informatics, Jointly with the 14th International Conference on Information Systems Analysis and Synthesis, ISAS 2008 - Proc. English The purpose of this paper is to identify the major global trends in social computing/collaborative computing as they seem to evolve under the 'web 2.0' heading, evaluate possible future trends seen from a technology point of view as well as from a socio-economic perspective and compare these trends to the current state and need for eGovernment. Finally, some general guidelines, ideas and projections for eGovernment will be proposed and discussed. 0 0
Early response to false claims in Wikipedia P.D. Magnus First Monday A number of studies have assessed the reliability of entries in Wikipedia at specific times. One important difference between Wikipedia and traditional media, however, is the dynamic nature of its entries. An entry assessed today might be substantially extended or reworked tomorrow. This study paper assesses the frequency with which small, inaccurate changes are quickly corrected. 0 3
Editor's Introduction: Writing and New Media Special Issue. Christina Haas Written Communication The article discusses various reports published within the article, including one by Jeff Bezemer, and Gunther Kress on social semiotics of writing, and another by John Jones on revision patterns in Wikipedia articles. 0 0
Edutech Wiki - an all-in-one solution to support whole scholarship? Daniel K. Schneider World Conference on Educational Multimedia, Hypermedia and Telecommunications English One of the greatest challenges to educational technologists is to apply our research to ourselves, e.g. use technology to enhance our own intellectual development, as well as the way we teach and interact with others. There are many explanations why technology is underused. One is the uncomfortable relationship between research and teaching. Boyer distinguishes between scholarship of discovery, integration, application and teaching. Only the first one (research) is relevant in terms of career planning. It is a bad strategy to invest in time-consuming technology-enhanced pedagogies, in collective knowledge building or in application activities. However, most academics still might agree that these would be “nice to have”. Edutech Wiki (http://edutechwiki.unige.ch/en/) is an attempt to lower the cost of engagement in Boyer’s “lesser” scholarships though using the same medium for multiple purposes. We will report on some outcomes, some of which can be considered good practice. 0 0
Eduwiki: An effective teachers' knowledge management platform Yiping Zhou
Chaohua Gong
Computer-Supported Cooperative Work
Eduwiki
Knowledge management
Teachers' knowledge management
Wiki
2008 International Conference on Wireless Communications, Networking and Mobile Computing, WiCOM 2008 English Wikis' simplicity and flexible nature have begun drawing the attention of teachers and educators, who realize that Wikis facilitate collaborative finding, buildng, and sharing of knowledge. Knowledge management can enable schools to evolve from bureaucracies forged during an industrial era to educational knowledge ecologies that are prepared to compete in a networked information-driven global society. Referred to the advantages of Wikis, the paper describes the needs analysis and architecture of Eduwiki, which focus on Wiki basic service, Personal space, Group space, Resource center and other mechanisms of Eduwiki. These mechanisms are effective when it is used as tools of monitoring and recording processes of teachers' work The paper also gives some practical evidences and they suggest that Eduwiki can be used as an effective tool for collaborative lessong-preparing. Also, the teachers' evaluations are the important reference standards for further improving availability of Eduwiki. 0 0
Effects of short-term memory and content representation type on mobile language learning Chen
Nian-Shing
Sheng Hsieh-Wen Kinshuk
Language, Learning \& Technology 0 0
Efficient Ranking and Computation of Semantic Relatedness and its Application to Word Sense Disambiguation Maxim Grinev
Dmitry Lizorkin
Denis Turdakov
Pavel Velikhov
English 0 0
El enfoque de colaboración de Wikipedia y el proyecto Wikiprofesional. Pedro Urra Gonzalez ACIMED En este artículo se considera el proyecto de colaboración científica {WikiProteínas,} enfocado a recolectar e integrar los conocimientos sobre proteínas y su importancia en la biología y la medicina. Dicha iniciativa se realiza en el marco proyectual de las tecnologías Wiki y {WikiProfesional} {(Http://www.wikiprofessional.org/conceptweb). 0 0
Employing a domain specific ontology to perform semantic search Morneau M.
Mineau G.W.
Lecture Notes in Computer Science English Increasing the relevancy of Web search results has been a major concern in research over the last years. Boolean search, metadata, natural language based processing and various other techniques have been applied to improve the quality of search results sent to a user. Ontology-based methods were proposed to refine the information extraction process but they have not yet achieved wide adoption by search engines. This is mainly due to the fact that the ontology building process is time consuming. An all inclusive ontology for the entire World Wide Web might be difficult if not impossible to construct, but a specific domain ontology can be automatically built using statistical and machine learning techniques, as done with our tool: SeseiOnto. In this paper, we describe how we adapted the SeseiOnto software to perform Web search on the Wikipedia page on climate change. SeseiOnto, by using conceptual graphs to represent natural language and an ontology to extract links between concepts, manages to properly answer natural language queries about climate change. Our tests show that SeseiOnto has the potential to be used in domain specific Web search as well as in corporate intranets. 0 0
End-user programming with application wikis Ludovic Dubost
Stewart Nickolas
Peter Thoeny
Dirk Riehle
WikiSym English 0 0
Enhancing blended learning - developing a community based methopedia Ryberg T.
Brenstein E.
Pilt L.
Moczadlo R.
Niemczik C.
Dirckinck-Holmfeld L.
Blended learning methods
Community based wiki
Learning activities
Learning designs
Methopedia
Sharing best practices
Proceedings of the 7th European Conference on e-Learning, ECEL 2008 English This paper will present some theoretical, methodological and technical discussions of the community based Methopedia wiki, which is under development, and will launch in its first incarnation September 2008. Methopedia is one activity within the EU-funded multilateral ICT project "Community of Integrated Blended Learning in Europe" (COMBLE), which is composed of partners from business, university and education. Methopedia is envisioned to be a community based wiki-system, combined with social networking features, where trainers and teachers from various sectors (vocational training, enterprises, universities, and public bodies) can consult, discuss, describe and share blended learning methods and methodologies. The aim of the COMBLE project and Methopedia is to contribute to leveraging the quality of education and training in Europe by engaging practitioners and researchers in developing, sharing and promoting innovative blended learning concepts and methods. To enable practitioners to share methods and learning designs an important task is to construct a shared, flexible pedagogical model/template for describing these. The model, which we outline in this paper, through discussing and synthesising literature within the field of learning design, needs to be further developed through theoretical and conceptual discussions. However, based on existing research, we argue that it is necessary that the model is developed through engaging the partner-network of trainers from business and education, and position them as co-creators of models and frameworks. We conclude the paper by discussing and outlining requirements and functionalities, some of which follow from the discussion of learning designs and blended learning, while others are related to the technical and socio-technical aspects of Methopedia. 0 0
Enhancing text clustering by leveraging Wikipedia semantics Jian Hu
Lujun Fang
Yang Cao
Hua-Jun Zeng
Hua Li
Qiang Yang
Zheng Chen
English Most traditional text clustering methods are based on "bag of words" (BOW) representation based on frequency statistics in a set of documents. BOW, however, ignores the important information on the semantic relationships between key terms. To overcome this problem, several methods have been proposed to enrich text representation with external resource in the past, such as WordNet. However, many of these approaches suffer from some limitations: 1) WordNet has limited coverage and has a lack of effective word-sense disambiguation ability; 2) Most of the text representation enrichment strategies, which append or replace document terms with their hypernym and synonym, are overly simple. In this paper, to overcome these deficiencies, we first propose a way to build a concept thesaurus based on the semantic relations (synonym, hypernym, and associative relation) extracted from Wikipedia. Then, we develop a unified framework to leverage these semantic relations in order to enhance traditional content similarity measure for text clustering. The experimental results on Reuters and OHSUMED datasets show that with the help of Wikipedia thesaurus, the clustering performance of our method is improved as compared to previous methods. In addition, with the optimized weights for hypernym, synonym, and associative concepts that are tuned with the help of a few labeled data users provided, the clustering performance can be further improved. 0 0
Enhancing the student experience using web 2.0 technologies (Wikis, Blogs and Webcam Recordings) to encourage student engagement and to develop collaborative learning: A case study Clarke J.
Hunter J.
Wells M.
Blogs
Collaborative learning
Webcams
Wiki
Proceedings of the 7th European Conference on e-Learning, ECEL 2008 English The use of Web 2.0 technologies on the Digital Economy module first began as part of a TQEF funded project during 05/06 delivery. Part of the coursework assessment requires the use of the technically demanding software package Dreamweaver. Support material in the form of Blackboard quizzes; FAQs; and screen capture demonstrations together with additional seminar material relating to the module was prepared and feedback from students was very positive and encouraged a greater level of student engagement and interaction. Following a post delivery review, the module team therefore reworked the group coursework element and revised the presentation of seminar and workshop materials to ensure that students accessed all material at an appropriate time and were encouraged to engage more actively with the module. For the 06/07 delivery each group was provided with voluntary access to wiki pages as a secure collaborative environment to record their group's assessment ideas and to work on them accordingly. Not all groups used the facility but all groups were required to post up their final website and presentation slides and comments to their wiki pages by the date of the assessment presentation itself. This case study will examine in detail how students and staff engaged with the wikis, blogs, podcasts and webcam recordings in the 07/08 module delivery. We will then discuss how Web 2.0 technologies have impacted on the students' engagement with the module; the impact, if any, on the quality of their work; the impact on the feedback process; and how, exactly, the student experience has been enhanced. 0 0
Enriching Multilingual Language Resources by Discovering Missing Cross-Language Links in Wikipedia Jong-Hoon Oh
Daisuke Kawahara
Kiyotaka Uchimoto
Jun'ichi Kazama
Kentaro Torisawa
Web mining
Wikipedia
Cross-Language Links
Language resources
WI-IAT English 0 1
Enriching multilingual language resources by discovering missing cross-language links in Wikipedia Oh J.-H.
Daisuke Kawahara
Kiyotaka Uchimoto
Jun'ichi Kazama
Kentaro Torisawa
Proceedings - 2008 IEEE/WIC/ACM International Conference on Web Intelligence, WI 2008 English We present a novel method for discovering missing crosslanguage links between English and Japanese Wikipedia articles. We collect candidates of missing cross-language links - a pair of English and Japanese Wikipedia articles, which could be connected by cross-language links. Then we select the correct cross-language links among the candidates by using a classifier trained with various types of features. Our method has three desirable characteristics for discovering missing links. First, our method can discover cross-language links with high accuracy (92% precision with 78% recall rates). Second, the features used in a classifier are language-independent. Third, without relying on any external knowledge, we generate the features based on resources automatically obtained from Wikipedia. In this work, we discover approximately 105 missing crosslanguage links from Wikipedia, which are almost two-thirds as many as the existing cross-language links in Wikipedia. 0 1
Enriching the crosslingual link structure of wikipedia - A classification-based approach Sorg P.
Philipp Cimiano
AAAI Workshop - Technical Report English The crosslingual link structure of Wikipedia represents a valuable resource which can be exploited for crosslingual natural language processing applications. However, this requires that it has a reasonable coverage and is furthermore accurate. For the specific language pair German/English that we consider in our experiments, we show that roughly 50% of the articles are linked from German to English and only 14% from English to German. These figures clearly corroborate the need for an approach to automatically induce new cross-language links, especially in the light of such a dynamically growing resource such as Wikipedia. In this paper we present a classification-based approach with the goal of inferring new cross-language links. Our experiments show that this approach has a recall of 70% with a precision of 94% for the task of learning cross-language links on a test dataset. 0 0
Entity ranking based on category expansion Jamsen J.
Nappila T.
Arvola P.
Lecture Notes in Computer Science English This paper introduces category and link expansion strategies for the XML Entity Ranking track at INEX 2007. Category expansion is a coefficient propagation method for the Wikipedia category hierarchy based on given categories or categories derived from sample entities. Link expansion utilizes links between Wikipedia articles. The strategies are evaluated within the entity ranking and list completion tasks. 0 0
Entity ranking from annotated text collections using multitype topic models Shiozaki H.
Eguchi K.
Lecture Notes in Computer Science English Very recently, topic model-based retrieval methods have produced good results using Latent Dirichlet Allocation (LDA) model or its variants in language modeling framework. However, for the task of retrieving annotated documents when using the LDA-based methods, some post-processing is required outside the model in order to make use of multiple word types that are specified by the annotations. In this paper, we explore new retrieval methods using a 'multitype topic model' that can directly handle multiple word types, such as annotated entities, category labels and other words that are typically used in Wikipedia. We investigate how to effectively apply the multitype topic model to retrieve documents from an annotated collection, and show the effectiveness of our methods through experiments on entity ranking using a Wikipedia collection. 0 0
Entity ranking in Wikipedia Anne M. Vercoustre
James A. Thom
Jovan Pehcevski
English The traditional entity extraction problem lies in the ability of extracting named entities from plain text using natural language processing techniques and intensive training from large document collections. Examples of named entities include organisations, people, locations, or dates. There are many research activities involving named entities; we are interested in entity ranking in the field of information retrieval. In this paper, we describe our approach to identifying and ranking entities from the INEX Wikipedia document collection. Wikipedia offers a number of interesting features for entity identification and ranking that we first introduce. We then describe the principles and the architecture of our entity ranking system, and introduce our methodology for evaluation. Our preliminary results show that the use of categories and the link structure of Wikipedia, together with entity examples, can significantly improve retrieval effectiveness. 0 0
Entity-based query reformulation using Wikipedia Yang Xu
Fan Ding
Bin Wang
CIKM English Many real world applications increasingly involve both structured data and text, and entity based retrieval is an important problem in this realm. In this paper, we present an automatic query reformulation approach based on entities detected in each query. The aim is to utilize semantics associated with entities for enhancing document retrieval. This is done by expanding a query with terms/phrases related to entities in the query. We exploit Wikipedia as a large repository of entity information. Our reformulated approach consists of three major steps : (1) detect representative entity in a query; (2) expand the query with entity related terms/phrases; and (3) facilitate term dependency features. We evaluate our approach in ad-hoc retrieval task on four TREC collections, including two large web collections. Experiments results show that significant improvement is possible by utilizing entity corresponding information. 0 0
Environment for statistical computing Jaromir Antoch Computer Science Review 0 0
Experiences of wiki use in Finnish companies MindTrek - 12th International MindTrek Conference: Entertainment and Media in the Ubiquitous Era English 0 0
Experiment management system-A way towards a transparent Tokamak Kramer-Flecken A.
Landgraf B.
Krom J.G.
Client/server approach
Database
Logbook
Metadata
Remote data access
Remote participation
Web technology
Wiki
Fusion Engineering and Design English At TEXTOR extensive collaborations with foreign institutes have been established. For a successful collaboration, the participating scientists must be able to plan the experiment in advance. Therefore they need tools to submit and track their experiment proposals and pulse plans. During and after the experiment access to raw and evaluated data of a TEXTOR discharge is needed to analyze TEXTOR data remotely from the home labs. A goal-oriented data analysis needs in addition a various items of data not recorded with conventional data logging. This discharge specific comments and settings are stored in a logbook. At TEXTOR a logbook with different views is implemented for the scientific experiment and the machine operation, respectively. In a further development towards a TEXTOR experiment management system the information of auxiliary heating devices is stored. The electronic logbook is realized by a client/server approach which makes the access independent from the hard- and software on the client side and allows to access the data and data supplements (metadata) by any common web browsers. All information from TEXTOR is accessible for everybody via a uniform resource locator (URL). The concept of the electronic logbook, the first experience and its possibilities for data analysis will be discussed in this paper. © 2007 Elsevier B.V. All rights reserved. 0 0
Exploiting Locality of Wikipedia Links in Entity Ranking Jovan Pehcevski
Anne M. Vercoustre
James A. Thom
Advances in Information Retrieval English Information retrieval from web and XML document collections is ever more focused on returning entities instead of web pages or XML elements. There are many research fields involving named entities; one such field is known as entity ranking, where one goal is to rank entities in response to a query supported with a short list of entity examples. In this paper, we describe our approach to ranking entities from the Wikipedia XML document collection. Our approach utilises the known categories and the link structure of Wikipedia, and more importantly, exploits link co-occurrences to improve the effectiveness of entity ranking. Using the broad context of a full Wikipedia page as a baseline, we evaluate two different algorithms for identifying narrow contexts around the entity examples: one that uses predefined types of elements such as paragraphs, lists and tables; and another that dynamically identifies the contexts by utilising the underlying XML document structure. Our experiments demonstrate that the locality of Wikipedia links can be exploited to significantly improve the effectiveness of entity ranking. 0 0
Exploiting Wikipedia for directional inferential text similarity Leong C. Wee
S. Hassan
Directional
Inference
Semantics
Similarity
Wikipedia
Proceedings - International Conference on Information Technology: New Generations, ITNG 2008 English In natural languages, variability of semantic expression refers to the situation where the same meaning can be inferred from different words or texts. Given that many natural language processing tasks nowadays (e.g. question answering, information retrieval, document summarization) often model this variability by requiring a specific target meaning to be inferred from different text variants, it is helpful to capture text similarity in a directional manner to serve such inference needs. In this paper, we show how Wikipedia can be used as a semantic resource to build a directional inferential similarity metric between words, and subsequently, texts. Through experiments, we show that our Wikipedia-based metric performs significantly better when applied to a standard evaluation dataset, with a reduction in error rate of 16.1% over the random metric baseline. 0 0
Exploiting XML structure to improve information retrieval in peer-to-peer systems Winter J. Content-based XML-retrieval
Distributed search
Information retrieval
Peer-to-Peer
XML information retrieval
ACM SIGIR 2008 - 31st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Proceedings English With the advent of XML as a standard for representation and exchange of structured documents, a growing amount of XML-documents are being stored in Peer-to-Peer (P2P) networks. Current research on P2P search engines proposes the use of Information Retrieval (IR) techniques to perform content-based search, but does not take into account structural features of documents. P2P systems typically have no central index, thus avoiding single-points-of-failures, but distribute all information among participating peers. Accordingly, a querying peer has only limited access to the index information and should select carefully which peers can help answering a given query by contributing resources such as local index information or CPU time for ranking computations. Bandwidth consumption is a major issue. To guarantee scalability, P2P systems have to reduce the number of peers involved in the retrieval process. As a result, the retrieval quality in terms of recall and precision may suffer substantially. In the proposed thesis, document structure is considered as an extra source of information to improve the retrieval quality of XML-documents in a P2P environment. The thesis centres on the following questions: how can structural information help to improve the retrieval of XML-documents in terms of result quality such as precision, recall, and specificity? Can XML structure support the routing of queries in distributed environments, especially the selection of promising peers? How can XML IR techniques be used in a P2P network while minimizing bandwidth consumption and considering performance aspects? To answer these questions and to analyze possible achievements, a search engine is proposed that exploits structural hints expressed explicitly by the user or implicitly by the self-describing structure of XML-documents. Additionally, more focused and specific results are obtained by providing ranked retrieval units that can be either XML-documents as a whole or the most relevant passages of theses documents. XML information retrieval techniques are applied in two ways: to select those peers participating in the retrieval process, and to compute the relevance of documents. The indexing approach includes both content and structural information of documents. To support efficient execution of multi term queries, index keys consist of rare combinations of (content, structure)-tuples. Performance is increased by using only fixed-sized posting lists: frequent index keys are combined with each other iteratively until the new combination is rare, with a posting list size under a pre-set threshold. All posting lists are sorted by taking into account classical IR measures such as term frequency and inverted term frequency as well as weights for potential retrieval units of a document, with a slight bias towards documents on peers with good collections regarding the current index key and with good peer characteristics such as online times, available bandwidth, and latency. When extracting the posting list for a specific query, a re-ordering on the posting list is performed that takes into account the structural similarity between key and query. According to this preranking, peers are selected that are expected to hold information about potentially relevant documents and retrieval units The final ranking is computed in parallel on those selected peers. The computation is based on an extension of the vector space model and distinguishes between weights for different structures of the same content. This allows weighting XML elements with respect to their discriminative power, e.g. a title will be weighted much higher than a footnote. Additionally, relevance is computed as a mixture of content relevance and structural similarity between a given query and a potential retrieval unit. Currently, a first prototype for P2P Information Retrieval of XML-documents called SPIRIX is being implemented. Experiments to evaluate the proposed techniques and use of structural hints will be performed on a distributed version of the INEX Wikipedia Collection. 0 0
Exploiting semantics in collaborative software development tasks Panagiotou D.
Mentzas G.
Collaborative
Semantic web
Semantic Wikis
Software Development
Frontiers in Artificial Intelligence and Applications English Modern software development is highly knowledge intensive; it requires that software developers create and share new knowledge during their daily work. However, current software development environments are 'syntantic', i.e. they do not facilitate understanding the semantics of software artifacts and hence cannot fully support the knowledge-driven activities of developers. In this paper we present SoWiSE, a Semantic Wiki environment which focuses on the software development domain and strives to address these problems. SoWiSE aims at providing software developers such a tool to ease their daily work and facilitate cooperation, context-aware navigation, ontology-based browsing and searching, concept-based source code documentation and related problem solving. © 2008 The authors and IOS Press. All rights reserved. 0 0
Exploiting the Collective Intelligence Contained in Wikipedia to Automatically Describe the Content of a Document Anuradha Jambunathan
Marco Ronchetti
The Semantic Web: a view on data integration, reasoning, human factors, collective intelligence and technology adoption English The Wikipedia phenomenon very interesting from the point of view

of the collective, social effort to produce a large, strongly interlinked body of knowledge. It also offers, for the first time in history, a general source of information coded in electronic form and freely available to anyone. As such, it can be used as a reference for tools aiming at mining semantic meaning from generic documents. In this paper, we propose a clustering-based method that exploits some of the implicit knowledge built into Wikipedia to refine and

ameliorate existing approaches.
0 0
Exploiting the collective intelligence contained in Wikipedia to automatically describe the content of a document Marco Ronchetti Anuradha Jambunathan Semantic relatedness
Semantic Analysis
The Wikipedia phenomenon is very interesting from the point of view of the collective, social effort to produce a large, strongly interlinked body of knowledge. It also offers, for the first time in history, a general source of information coded in electronic form and freely available to anyone. As such, it can be used as a reference for tools aiming at mining semantic meaning from generic documents. In this paper, we propose a clustering-based method that exploits some of the implicit knowledge built into Wikipedia to refine and ameliorate existing approaches. 0 0
Exploring Words with Semantic Relations from Chinese Wikipedia Y Li
K Y Huang
F J Ren
Y X Zhong
INFORMATION-AN INTERNATIONAL INTERDISCIPLINARY JOURNAL This paper introduces a way of exploring words with semantic relations from Chinese Wikipedia documents. A corpus with structured documents is generated from Chinese Wikipedia pages. Then considering of the hyperlinks, text overlaps and word frequencies, word pairs with semantic relations are explored. Words can be self clustered into groups with tight semantic relations. We roughly measure the semantic relatedness with different document based algorithms and analyze the reliability of our measures in comparing experiment. 0 0
Exploring motivations for contributing to open source initiatives: The roles of contribution context and personal values Shaul Oreg
Oded Nov
Personal values
Motivation
Open source
Wikipedia
Computers in Human Behavior English We explore contextual and dispositional correlates of the motivation to contribute to open source initiatives. We examine how the context of the open source project, and the personal values of contributors, are related to the types of motivations for contributing. A web-based survey was administered to 300 contributors in two prominent open source contexts: software and content. As hypothesized, software contributors placed a greater emphasis on reputation-gaining and self-development motivations, compared with content contributors, who placed a greater emphasis on altruistic motives. Furthermore, the hypothesized relationships were found between contributors' personal values and their motivations for contributing. 0 1
Exploring motivations for contributing to open source initiatives: The roles of contribution context and personal values. Shaul Oreg and Oded Nov Personal values; Motivations; Open source; Wikipedia Computers in Human Behavior, , issue 5, 2055-2073 We explore contextual and dispositional correlates of the motivation to contribute to open source initiatives. We examine how the context of the open source project, and the personal values of contributors, are related to the types of motivations for contributing. A web-based survey was administered to 300 contributors in two prominent open source contexts: software and content. As hypothesized, software contributors placed a greater emphasis on reputation-gaining and self-development motivations, compared with content contributors, who placed a greater emphasis on altruistic motives. Furthermore, the hypothesized relationships were found between contributors' personal values and their motivations for contributing. 0 0
Exploring the knowledge in semi structured data sets with rich queries Umbrich J.
Sebastian Blohm
CEUR Workshop Proceedings English Semantics can be integrated in to search processing during both document analysis and querying stages. We describe a system that incorporates both, semantic annotations of Wikipedia articles into the search process and allows for rich annotation search, enabling users to formulate queries based on their knowledge about how entities relate to one another while simultaneously retaining the freedom of free text search where appropriate. The outcome of this work is an application consisting of semantic annotators, an extended search engine and an interactive user interface. 0 0
Exploring the relationship between anxiety and usability evaluation - An online study of internet and wiki anxiety MCCSIS'08 - IADIS Multi Conference on Computer Science and Information Systems; Proceedings of Interfaces and Human Computer Interaction 2008 English 0 0
Exploring words with semantic correlations from Chinese Wikipedia Yanyan Li
Huang K.
Tsuchiya S.
Ren F.
Zhong Y.
IFIP International Federation for Information Processing English In this paper, we work on semantic correlation between Chinese words based on Wikipedia documents. A corpus with about 50,000 structured documents is generated from Wikipedia pages. Then considering of hyper-links, text overlaps and word frequency, about 300,000 word pairs with semantic correlations are explored from these documents. We roughly measure the degree of semantic correlations and find groups with tight semantic correlations by self clustering. 0 0
Extending the makna semanticwiki to support workflows Dello K.
Nixon L.
Robert Tolksdorf
CEUR Workshop Proceedings English Semantic wikis combine the advantages introduced by the wiki principle with the potential of SemanticWeb technologies. However, there is still a very limited support for coordination, collaboration and integration in current semantic wikis. In this paper, we present a solution for this through the integration of our Makna semantic wiki with a workflow system. The resulting implementation is presented and an example given how this integration leads to better coordination, collaboration and integration support. 0 0
External query expansion in the blogosphere Weerkamp W.
Maarten de Rijke
NIST Special Publication English We describe the participation of the University of Amsterdam's ILPS group in the blog track at TREC 2008. We mainly explored different ways of using external corpora to expand the original query. In the blog post retrieval task we did not succeed in improving over a simple baseline (equal weights for both the expanded and original query). Obtaining optimal weights for the original and the expanded query remains a subject of investigation. In the blog distillation task we tried to improve over our (strong) baseline using external expansion, but due to differences in the run setup, comparing these runs is hard. Compared to a simpler baseline, we see an improvement for the run using external expansion on the combination of news, Wikipedia and blog posts. 0 0
Extracting Lexical Semantic Knowledge from Wikipedia and Wiktionary Torsten Zesch
Christof Muller
Iryna Gurevych
Wikipedia API
Ubiquitous Knowledge Processing Lab#Wiktionary_API
LREC'08 Recently, collaboratively constructed resources such as Wikipedia and Wiktionary have been discovered as valuable lexical semantic knowledge bases with a high potential in diverse Natural Language Processing (NLP) tasks. Collaborative knowledge bases however significantly differ from traditional linguistic knowledge bases in various respects, and this constitutes both an asset and an impediment for research in NLP. This paper addresses one such major impediment, namely the lack of suitable programmatic access mechanisms to the knowledge stored in these large semantic knowledge bases. We present two application programming interfaces for Wikipedia and Wiktionary which are especially designed for mining the rich lexical semantic information dispersed in the knowledge bases, and provide efficient and structured access to the available knowledge. As we believe them to be of general interest to the NLP community, we have made them freely available for research purposes. 0 1
Extracting bilingual word pairs from Wikipedia Tyers
F.
Pienaar
J.
Under-resourced languages
Machine translation
Language resources
Bilingual terminology
Interwiki links
SALTMIL workshop at Language Resources and Evaluation Conference (LREC) 2008 A bilingual dictionary or word list is an important resource for many purposes, among them, machine translation. For many language pairs these are either non-existent, or very often unavailable owing to licensing restrictions. We describe a simple, fast and computationally inexpensive method for extracting bilingual dictionary entries from Wikipedia (using the interwiki link system) and assess the performance of this method with respect to four language pairs. Precision was found to be in the 69-92% region, but open to improvement. 0 1
Extracting communities from complex networks by the k-dense method K. Saito
T. Yamada
K. Kazama
Communications and Computer Sciences IEICE Transactions on Fundamentals of Electronics To understand the structural and functional properties of large-scale complex networks, it is crucial to efficiently extract a set of cohesive subnetworks as communities. There have been proposed several such community extraction methods in the literature, including the classical k-core decomposition method and, more recently, the k-clique based community extraction method. The k-core method, although computationally efficient, is often not powerful enough for uncovering a detailed community structure and it produces only coarse-grained and loosely connected communities. The k-clique method, on the other hand, can extract fine-grained and tightly connected communities but requires a substantial amount of computational load for large-scale complex networks. In this paper, we present a new notion of a subnetwork called k-dense, and propose an efficient algorithm for extracting k-dense communities. We applied our method to the three different types of networks assembled from real data, namely, from blog trackbacks, word associations and Wikipedia references, and demonstrated that the k-dense method could extract communities almost as efficiently as the k-core method, while the qualities of the extracted communities are comparable to those obtained by the k-clique method. 0 0
Extracting concept hierarchy knowledge from the Web based on Property Inheritance and Aggregation Hattori S.
Katsumi Tanaka
Proceedings - 2008 IEEE/WIC/ACM International Conference on Web Intelligence, WI 2008 English Concept hierarchy knowledge, such as hyponymy and meronymy, is very important for various natural language processing systems. While WordNet and Wikipedia are being manually constructed and maintained as lexical ontologies, many researchers have tackled how to extract concept hierarchies from very large corpora of text documents such as the Web not manually but automatically. However, their methods are mostly based on lexico-syntactic patterns as not necessary but sufficient conditions of hyponymy and meronymy, so they can achieve high precision but low recall when using stricter patterns or they can achieve high recall but low precision when using looser patterns. Therefore, we need necessary conditions of hyponymy and meronymy to achieve high recall and not low precision. In this paper, not only "Property Inheritance "from a target concept to its hyponyms but also "Property Aggregation" from its hyponyms to the target concept is assumed to be necessary and sufficient conditions of hyponymy, and we propose a method to extract concept hierarchy knowledge from the Web based on property inheritance and property aggregation. 0 0
Extracting structured knowledge for Semantic Web by mining Wikipedia Kotaro Nakayama CEUR Workshop Proceedings English Since Wikipedia has become a huge scale database storing wide-range of human knowledge, it is a promising corpus for knowledge extraction. A considerable number of researches on Wikipedia mining have been conducted and the fact that Wikipedia is an invaluable corpus has been confirmed. Wikipedia's impressive characteristics are not limited to the scale, but also include the dense link structure, URI for word sense disambiguation, well structured Infoboxes, and the category tree. One of the popular approaches in Wikipedia Mining is to use Wikipedia's category tree as an ontology and a number of researchers proved that Wikipedia's categories are promising resources for ontology construction by showing significant results. In this work, we try to prove the capability of Wikipedia as a corpus for knowledge extraction and how it works in the Semantic Web environment. We show two achievements; Wikipedia Thesaurus, a huge scale association thesaurus by mining the Wikipedia's link structure, and Wikipedia Ontology, a Web ontology extracted by mining Wikipedia articles. 0 0
Extraction and analysis of tripartite relationships from Wikipedia Nazir F.
Hideaki Takeda
International Symposium on Technology and Society, Proceedings English Social aspects are critical in the decision making process for social actors (human beings). Social aspects can be categorized into social interaction, social communities, social groups or any kind of behavior that emerges from interlinking, overlapping or similarities between interests of a society. These social aspects are dynamic and emergent. Therefore, interlinking them in a social structure, based on bipartite affiliation network, may result in isolated graphs. The major reason is that as these correspondences are dynamic and emergent, they should be coupled with more than a single affiliation in order to sustain the interconnections during interest evolutions. In this paper we propose to interlink actors using multiple tripartite graphs rather than a bipartite graph which was the focus of most of the previous social network building techniques. The utmost benefit of using tripartite graphs is that we can have multiple and hierarchical links between social actors. Therefore in this paper we discuss the extraction, plotting and analysis methods of tripartite relations between authors, articles and categories from Wikipedia. Furthermore, we also discuss the advantages of tripartite relationships over bipartite relationships. As a conclusion of this study we argue based on our results that to build useful, robust and dynamic social networks, actors should be interlinked in one or more tripartite networks. 0 1
Feminist invitational collaboration in a digital age: Looking over disciplinary and national borders Z. Wei Gender, Communication, and Technology 0 0
Fernanda Bertini Viegas and Martin Wattenberg. Anonymous Issues in Science & Technology The article provides information on the history flow designed by Martin Wattenberg and Fernanda Bertini Viegas, that presents visualization of the flow of editing that takes place on all Wikipedia entries. An example is presented that show the history of the popular entry for chocolate in 2003 that each color corresponds to a different contributor. The revision lines corresponds to the beginning of changed or updated text, and a line's length shows the length of the text. Furthermore, the visualization therefore briefly takes the level of debate and controversy surrounding a topic. 0 0
Fine-grained addressability to support large-scale collaborative document development Viral Gupta Computer science
Large-scale documents
English The Web has made it possible for large, distributed collaborations to develop sophisticated document bases. As these collaborations increase in size, there is a need to support reference, navigation and search through the document bases by individuals who are not computer pro-fessionals. Although collaborative tools such as wikis have been developed that address some of the requirements of these communities, the tools lack support for simple and convenient fine-grained addressability to parts of the documents. Such addressability is essential for formal documents, such as standards and legal documents. In this thesis we develop a solution to the problem of fine-grained addressability that is based on MediaWiki, a popular and powerful col- laboration tool that is the software infrastructure for Wikipedia. Although some collaboration tools that support fine-grained addressability already exist, they have not addressed some of the open research issues of fine-grained addressability, such as dealing with transclusion, semantic annotation and hyperscope support. It also deals with the research problems that were raised by this goal. An architecture and reference implementation was developed to provide a proof of concept and to test the viability of the proposed solutions to the research problems. The thesis also discusses the various design decisions that were made in the course of solving the research problems and developing the reference implementation. 0 0
Flexible concurrency control for real-time collaborative editors Imine A. Collaborative editors
Consistency
Operational transformation
Optimistic replication
Real-time collaboration
Proceedings - International Conference on Distributed Computing Systems English Real-time Collaborative Editors (RCE) provide computer support for modifying simultaneously shared documents, such as articles, wiki pages and programming source code, by dispersed users. Due to data replication, Operational Transformation (OT) is considered as the efficient and safe method for consistency maintenance in the literature of collaborative editors. Indeed, it is aimed at ensuring copies convergence even though the users's updates are executed in any order on different copies. Unfortunately, existing OT algorithms often fail to achieve this objective. Moreover, these algorithms have limited scalability with the number of users as they use vector timestamps to enforce causality dependency. In this paper, we present a novel framework for managing collaborative editing work in a scalable and decentralized fashion. It may be deployed easily on P2P networks as it supports dynamic groups where users can leave and join at any time. 0 0
Flood little, cache more: Effective result-reuse in P2P IR systems Carl Zimmer
Bedathur S.
Gerhard Weikum
Lecture Notes in Computer Science English State-of-the-art Peer-to-Peer Information Retrieval (P2P IR) systems suffer from their lack of response time guarantee especially with scale. To address this issue, a number of techniques for caching of multi-term inverted list intersections and query results have been proposed recently. Although these enable speedy query evaluations with low network overheads, they fail to consider the potential impact of caching on result quality improvements. In this paper, we propose the use of a cache-aware query routing scheme, that not only reduces the response delays for a query, but also presents an opportunity to improve the result quality while keeping the network usage low. In this regard, we make three-fold contributions in this paper. First of all, we develop a cache-aware, multi-round query routing strategy that balances between query efficiency and result-quality. Next, we propose to aggressively reuse the cached results of even subsets of a query towards an approximate caching technique that can drastically reduce the bandwidth overheads, and study the conditions under which such a scheme can retain good result-quality. Finally, we empirically evaluate these techniques over a fully functional P2P IR system, using a large-scale Wikipedia benchmark, and using both synthetic and real-world query workloads. Our results show that our proposal to combine result caching with multi-round, cache-aware query routing can reduce network traffic by more than half while doubling the result quality. 0 0
Flyspeck in a semantic wiki collaborating on a large scale formalization of the kepler conjecture Christoph Lange
McLaughlin S.
Rabe F.
CEUR Workshop Proceedings English Semantic wikis have been successfully applied to many problems in knowledge management and collaborative authoring. They are particularly appropriate for scientific and mathematical collaboration. In previous work we described an ontology for mathematical knowledge based on the semantic markup language OMDoc and a semantic wiki using both. We are now evaluating these technologies in concrete application scenarios. In this paper we evaluate the applicability of our infrastructure to mathematical knowledge management by focusing on the Flyspeck project, a formalization of Thomas Hales' proof of the Kepler Conjecture. After describing the Flyspeck project and its requirements in detail, we evaluate the applicability of two wiki prototypes to Flyspeck, one based on Semantic MediaWiki and another on our mathematics-specific semantic wiki SWiM. 0 0
FolksoViz: A subsumption-based folksonomy visualization using the wikipedia Kangpyo Lee
Hyunwoo Kim
Chungsu Jang
Hyoung Kim-Joo
Journal of KISS: Computing Practices Folksonomy, which is created through the collaborative tagging from many users, is one of the driving factors of Web 2.0. Tags are said to be the web metadata describing a web document. If we are able to find the semantic subsumption relationships between tags created through the collaborative tagging, it can help users understand the metadata more intuitively. In this paper, targeting del.icio.us tag data, we propose a method named {FolksoViz} for deriving subsumption relationships between tags by using Wikipedia texts. For this purpose, we propose a statistical model for deriving subsumption relationships based on the frequency of each tag on the Wikipedia texts, and {TSD} {(Tag} Sense Disambiguation) method for mapping each tag to a corresponding Wikipedia text. The derived subsumption pairs are visualized effectively on the screen. The experiment shows that our proposed algorithm managed to find the correct subsumption pairs with high accuracy. 0 0
FolksoViz: A subsumption-based folksonomy visualization using wikipedia texts Kangpyo L.
Hyunwoo K.
Chungsu J.
Kim H.-J.
Collaborative tagging
Folksonomy
Subsumption
Visualisation
Web 2.0
Wikipedia
Proceeding of the 17th International Conference on World Wide Web 2008, WWW'08 English In this paper, targeting del.icio.us tag data, we propose a method, FolksoViz, for deriving subsumption relationships between tags by using Wikipedia texts, and visualizing a folksonomy. To fulfill this method, we propose a statistical model for deriving subsumption relationships based on the frequency of each tag on the Wikipedia texts, as well as the TSD (Tag Sense Disambiguation) method for mapping each tag to a corresponding Wikipedia text. The derived subsumption pairs are visualized effectively on the screen. The experiment shows that the FolksoViz manages to find the correct subsumption pairs with high accuracy. 0 0
Folksoviz: a subsumption-based folksonomy visualization using wikipedia texts Kangpyo Lee
Hyunwoo Kim
Chungsu Jang
Hyoung-Joo Kim
Collaborative tagging
Folksonomy
Subsumption
Visualisation
Web 2.0
Wikipedia
World Wide Web English 0 0
Forming teams within wiki MCCSIS'08 - IADIS Multi Conference on Computer Science and Information Systems; Proceedings of Intelligent Systems and Agents 2008 English 0 0
Fostering collaboration amongst off-campus students McIntosh P.C.
Weaver D.
Online collaboration
Open learning
Wiki
ASCILITE 2008 - The Australasian Society for Computers in Learning in Tertiary Education English While the importance of developing and giving feedback on generic skills is widely recognised for on-campus students, this is still largely ignored for off-campus students, primarily due to the practical difficulties involved. This paper reports on a pilot project that introduced a compulsory group work project into an off-campus unit, delivered through Open Universities Australia. Wikis were used as the technical tool to facilitate the project, which was mostly successful for students, albeit stressful and time-consuming for staff. While all participating students successfully completed group projects, only one group demonstrated high levels of collaboration during the process. © 2008 P. Craig McIntosh and Debbi Weaver. 0 0
From Friday to Sunday: The hacker ethic and shifting notions of labour, leisure and intellectual property Jr.
James Brown
Leisure Studies Leisure studies scholars have theorised how the Web is changing leisure experiences, and this essay continues that work by discussing the Web and shifting notions of leisure, labour and intellectual property. Much online activity is described under the umbrella term of 'piracy'. By discussing online cultural production in terms of what Pekka Himanen calls the hacker ethic, we can rethink rhetorics of piracy and better understand the positive and negative aspects of online activities. Rather than thinking of online activity as derivative, we can reframe Web texts as doing what all cultural texts do - build upon the past. The ethic of the Web is built on a hacker approach to work, play, collaboration, intellectual property. Facebook applications and Wikipedia entries are just two examples of Web users' embrace of the hacker ethic. But is this labour or leisure? Is Wikipedia, a text edited and maintained by volunteers, the result of work or play? Himanen provides a new way to view online activities that sit in between the categories of labour and leisure. Further, the hacker ethic allows us to understand the contested terms of labour and leisure alongside a third contested term: intellectual property. This paper provides a framework to help us better understand the new immaterial aspects of leisure activity happening on the Web. A discussion of these activities in terms of the hacker ethic allows scholars to explore shifting notions of labour, leisure and intellectual property without resorting to rhetorics of piracy. 0 0
… further results