Quality

From WikiPapers
(Redirected from Wikipedia Quality)
Jump to: navigation, search

quality is included as keyword or extra keyword in 2 datasets, 1 tools and 35 publications.

Datasets

Dataset Size Language Description
EPIC/Oxford Wikipedia quality assessment English EPIC/Oxford Wikipedia quality assessment This dataset comprises the full, anonymized set of responses from the blind assessment of a sample of Wikipedia articles across languages and disciplines by academic experts. The study was conducted in 2012 by EPIC and the University of Oxford and sponsored by the Wikimedia Foundation.
PAN Wikipedia quality flaw corpus 2012 324 MB English PAN Wikipedia quality flaw corpus 2012 is an evaluation corpus for the "Quality Flaw Prediction in Wikipedia" task of the PAN 2012 Lab, held in conjunction with the CLEF 2012 conference.

Tools

Tool Operating System(s) Language(s) Programming language(s) License Description Image
WikiTrust English New BSD License
GPL
WikiTrust is an open-source, on-line reputation system for Wikipedia authors and content.


Publications

Title Author(s) Published in Language DateThis property is a special property in this wiki. Abstract R C
Factors That Influence the Quality of Crowdsourcing Al Sohibani M.
Al Osaimi N.
Al Ehaidib R.
Al Muhanna S.
Dahanayake A.
Advances in Intelligent Systems and Computing 2015 Crowdsourcing is a technique that aims to obtain data, ideas, and funds, conduct tasks, or even solve problems with the aid of a group of people. It's a useful technique to save money and time. The quality of data is an issue that confronts crowdsourcing websites; as the data is obtained from the crowd, and how they control the quality of data. In some of the crowdsourcing websites they have implemented mechanisms in order to manage the data quality; such as, rating, reporting, or using specific tools. In this paper, five crowdsourcing websites: Wikipedia, Amazon Mechanical Turk, YouTube, Rally Fighter, and Kickstarter are studied as cases in order to identify the possible quality assurance methods or techniques that are useful to represent crowdsourcing data. A survey is conducted to gather general opinion about the range of reliability of crowdsourcing sites, their passion and contribution to improve the contents of these sites. Combining those to the available knowledge in the crowdsourcing research, the paper highlights the factors that influence the data quality in crowdsourcing. © Springer International Publishing Switzerland 2015. 0 0
On Measuring Malayalam Wikipedia Vasudevan T V International Journal of Emerging Engineering Research and Technology English September 2014 Wikipedia is a popular, multilingual, free internet encyclopedia. Anyone can edit articles in it. This paper presents an overview of research in the Malayalam edition of Wikipedia. History of Malayalam Wikipedia

is explained first. Different research lines related with Wikipedia are explored next. This is followed by an analysis of Malayalam Wikipedia’s fundamental components such as Articles, Authors and Edits along with

Growth and Quality. General trends are measured comparing with Wikipedias in other languages.
0 0
Assessing quality score of wikipedia articles using mutual evaluation of editors and texts Yu Suzuki
Masatoshi Yoshikawa
International Conference on Information and Knowledge Management, Proceedings English 2013 In this paper, we propose a method for assessing quality scores of Wikipedia articles by mutually evaluating editors and texts. Survival ratio based approach is a major approach to assessing article quality. In this approach, when a text survives beyond multiple edits, the text is assessed as good quality, because poor quality texts have a high probability of being deleted by editors. However, many vandals, low quality editors, delete good quality texts frequently, which improperly decreases the survival ratios of good quality texts. As a result, many good quality texts are unfairly assessed as poor quality. In our method, we consider editor quality score for calculating text quality score, and decrease the impact on text quality by vandals. Using this improvement, the accuracy of the text quality score should be improved. However, an inherent problem with this idea is that the editor quality scores are calculated by the text quality scores. To solve this problem, we mutually calculate the editor and text quality scores until they converge. In this paper, we prove that the text quality score converges. We did our experimental evaluation, and confirmed that our proposed method could accurately assess the text quality scores. Copyright is held by the owner/author(s). 0 0
Effects of implicit positive ratings for quality assessment of Wikipedia articles Yu Suzuki Journal of Information Processing English 2013 In this paper, we propose a method to identify high-quality Wikipedia articles by using implicit positive ratings. One of the major approaches for assessing Wikipedia articles is a text survival ratio based approach. In this approach, when a text survives beyond multiple edits, the text is assessed as high quality. However, the problem is that many low quality articles are misjudged as high quality, because every editor does not always read the whole article. If there is a low quality text at the bottom of a long article, and the text has not seen by the other editors, then the text survives beyond many edits, and the text is assessed as high quality. To solve this problem, we use a section and a paragraph as a unit instead of a whole page. In our method, if an editor edits an article, the system considers that the editor gives positive ratings to the section or the paragraph that the editor edits. From experimental evaluation, we confirmed that the proposed method could improve the accuracy of quality values for articles. 0 0
Network analysis of user generated content quality in Wikipedia Myshkin Ingawale
Amitava Dutta
Rahul Roy
Priya Seetharaman
Online Information Review English 2013 Purpose - Social media platforms allow near-unfettered creation and exchange of user generated content (UGC). Drawing from network science, the purpose of this paper is to examine whether high and low quality UGC differ in their connectivity structures in Wikipedia (which consists of interconnected user generated articles). Design/methodology/approach - Using Featured Articles as a proxy for high quality, a network analysis was undertaken of the revision history of six different language Wikipedias, to offer a network-centric explanation for the emergence of quality in UGC. Findings - The network structure of interactions between articles and contributors plays an important role in the emergence of quality. Specifically the analysis reveals that high-quality articles cluster in hubs that span structural holes. Research limitations/implications - The analysis does not capture the strength of interactions between articles and contributors. The implication of this limitation is that quality is viewed as a binary variable. Extensions to this research will relate strength of interactions to different levels of quality in UGC. Practical implications - The findings help harness the "wisdom of the crowds" effectively. Organisations should nurture users and articles at the structural hubs from an early stage. This can be done through appropriate design of collaborative knowledge systems and development of organisational policies to empower hubs. Originality/value - The network centric perspective on quality in UGC and the use of a dynamic modelling tool are novel. The paper is of value to researchers in the area of social computing and to practitioners implementing and maintaining such platforms in organisations. Copyright 0 0
The category structure in wikipedia: To analyze and know its quality using k-core decomposition Wang Q.
Xiaolong Wang
Zheng Chen
Lecture Notes in Computer Science English 2013 Wikipedia is a famous and free encyclopedia. A network based on its category structure is built and then analyzed from various aspects, such as the connectivity distribution, evolution of the overall topology. As an innovative point of our paper, the model that is on the base of the k-core decomposition is used to analyze evolution of the overall topology and test the quality (that is, the error and attack tolerance) of the structure when nodes are removed. The model based on removal of edges is compared. Our results offer useful insights for the growth and the quality of the category structure, and the methods how to better organize the category structure. 0 0
Assessing the accuracy and quality of Wikipedia entries compared to popular online encyclopaedias Imogen Casebourne
Chris Davies
Michelle Fernandes
Naomi Norman
English 2 August 2012 8 0
Classifying Wikipedia Articles Using Network Motif Counts and Ratios Guangyu Wu
Martin Harrigan
Pádraig Cuningham
WikiSym English August 2012 Because the production of Wikipedia articles is a collaborative process, the edit network around a article can tell us something about the quality of that article. Articles that have received little attention will have sparse networks; at the other end of the spectrum, articles that are Wikipedia battle grounds will have very crowded networks. In this paper we evaluate the idea of characterizing edit networks as a vector of motif counts that can be used in clustering and classification. Our objective is not immediately to develop a powerful classifier but to assess what is the signal in network motifs. We show that this motif count vector representation is effective for classifying articles on the Wikipedia quality scale. We further show that ratios of motif counts can effectively overcome normalization problems when comparing networks of radically different sizes. 0 0
Mutual Evaluation of Editors and Texts for Assessing Quality of Wikipedia Articles Yu Suzuki
Masatoshi Yoshikawa
WikiSym English August 2012 In this paper, we propose a method to identify good quality Wikipedia articles by mutually evaluating editors and texts. A major approach for assessing article quality is a text survival ratio based approach. In this approach, when a text survives beyond multiple edits, the text is assessed as good quality. This approach assumes that poor quality texts are deleted by editors with high possibility. However, many vandals delete good quality texts frequently, then the survival ratios of good quality texts are improperly decreased by vandals. As a result, many good quality texts are unfairly assessed as poor quality. In our method, we consider editor quality for calculating text quality, and decrease the impacts on text qualities by the vandals who has low quality. Using this improvement, the accuracy of the text quality should be improved. However, an inherent problem of this idea is that the editor qualities are calculated by the text qualities. To solve this problem, we mutually calculate the editor and text qualities until they converge. We did our experimental evaluation, and we confirmed that the proposed method could accurately assess the text qualities. 0 0
A Breakdown of Quality Flaws in Wikipedia Maik Anderka
Benno Stein
2nd Joint WICOW/AIRWeb Workshop on Web Quality (WebQuality 12) English 2012 The online encyclopedia Wikipedia is a successful example of the increasing popularity of user generated content on the Web. Despite its success, Wikipedia is often criticized for containing low-quality information, which is mainly attributed to its core policy of being open for editing by everyone. The identification of low-quality information is an important task since Wikipedia has become the primary source of knowledge for a huge number of people around the world. Previous research on quality assessment in Wikipedia either investigates only small samples of articles, or else focuses on single quality aspects, like accuracy or formality. This paper targets the investigation of quality flaws, and presents the first complete breakdown of Wikipedia's quality flaw structure. We conduct an extensive exploratory analysis, which reveals (1) the quality flaws that actually exist, (2) the distribution of flaws in Wikipedia, and (3) the extent of flawed content. An important finding is that more than one in four English Wikipedia articles contains at least one quality flaw, 70% of which concern article verifiability. 0 0
Assessing quality values of Wikipedia articles using implicit positive and negative ratings Yu Suzuki Lecture Notes in Computer Science English 2012 In this paper, we propose a method to identify high-quality Wikipedia articles by mutually evaluating editors and text using implicit positive and negative ratings. One of major approaches for assessing Wikipedia articles is a text survival ratio based approach. However, the problem of this approach is that many low quality articles are misjudged as high quality, because of two issues. This is because, every editor does not always read the whole articles. Therefore, if there is a low quality text at the bottom of a long article, and the text have not seen by the other editors, then the text survives beyond many edits, and the survival ratio of the text is high. To solve this problem, we use a section or a paragraph as a unit of remaining instead of a whole page. This means that if an editor edits an article, the system treats that the editor gives positive ratings to the section or the paragraph that the editor edits. This is because, we believe that if editors edit articles, the editors may not read the whole page, but the editors should read the whole sections or paragraphs, and delete low-quality texts. From experimental evaluation, we confirmed that the proposed method could improve the accuracy of quality values for articles. 0 0
Classifying Wikipedia articles using network motif counts and ratios Guangyu Wu
Martin Harrigan
Pádraig Cunningham
WikiSym 2012 English 2012 Because the production of Wikipedia articles is a collaborative process, the edit network around a article can tell us something about the quality of that article. Articles that have received little attention will have sparse networks; at the other end of the spectrum, articles that are Wikipedia battle grounds will have very crowded networks. In this paper we evaluate the idea of characterizing edit networks as a vector of motif counts that can be used in clustering and classification. Our objective is not immediately to develop a powerful classifier but to assess what is the signal in network motifs. We show that this motif count vector representation is effective for classifying articles on the Wikipedia quality scale. We further show that ratios of motif counts can effectively overcome normalization problems when comparing networks of radically different sizes. 0 0
Codification and collaboration: Information quality in social media Kane G.C.
Ransbotham S.
International Conference on Information Systems, ICIS 2012 English 2012 This paper argues that social media combines the codification and collaboration features of earlier generations of knowledge management systems. This combination potentially changes the way knowledge is created, potentially requiring new theories and methods for understanding these processes. We forward the specialized social network method of two-mode networks as one such approach. We examine the information quality of 16,244 articles built through 2,677,397 revisions by 147,362 distinct contributors to Wikipedia's Medicine Wikiproject. We find that the structure of the contributor-artifact network is associated with information quality in these networks. Our findings have implications for managers seeking to cultivate effective knowledge creation environments using social media and to identify valuable knowledge created external to the firm. 0 0
Doctors use, but don’t rely totally on, Wikipedia English 2012 0 0
First results from an investigation into the validity of developer reputation derived from wiki articles and source code Prause C.R.
Eisenhauer M.
2012 5th International Workshop on Co-operative and Human Aspects of Software Engineering, CHASE 2012 - Proceedings English 2012 The internal quality of software is often neglected by developers for various reasons like time pressure or a general dislike for certain activities. Yet internal quality is important to speed up development and to keep software maintainable. We present a way to use reputation systems to improve the internal quality of software by putting artifacts like wiki articles and source code under their control. Specifically, we show that reputation scores derived from such artifacts reflect actual reputation in the developer community using data from a work group wiki and an open source project. 0 0
FlawFinder: A Modular System for Predicting Quality Flaws in Wikipedia Oliver Ferschke
Iryna Gurevych
Marc Rittberger
PAN English 2012 With over 23 million articles in 285 languages, Wikipedia is the largest free knowledge base on the web. Due to its open nature, everybody is allowed to access and edit the contents of this huge encyclopedia. As a downside of this open access policy, quality assessment of the content becomes a critical issue and is hardly manageable without computational assistance. In this paper, we present FlawFinder, a modular system for automatically predicting quality flaws in unseen Wikipedia articles. It competed in the inaugural edition of the Quality Flaw Prediction Task at the PAN Challenge 2012 and achieved the best precision of all systems and the second place in terms of recall and F1-score. 0 1
Mutual evaluation of editors and texts for assessing quality of Wikipedia articles Yu Suzuki
Masatoshi Yoshikawa
WikiSym 2012 English 2012 In this paper, we propose a method to identify good quality Wikipedia articles by mutually evaluating editors and texts. A major approach for assessing article quality is a text survival ratio based approach. In this approach, when a text survives beyond multiple edits, the text is assessed as good quality. This approach assumes that poor quality texts are deleted by editors with high possibility. However, many vandals delete good quality texts frequently, then the survival ratios of good quality texts are improperly decreased by vandals. As a result, many good quality texts are unfairly assessed as poor quality. In our method, we consider editor quality for calculating text quality, and decrease the impacts on text qualities by the vandals who has low quality. Using this improvement, the accuracy of the text quality should be improved. However, an inherent problem of this idea is that the editor qualities are calculated by the text qualities. To solve this problem, we mutually calculate the editor and text qualities until they converge. We did our experimental evaluation, and we confirmed that the proposed method could accurately assess the text qualities. 0 0
Network Analysis of User Generated Content Quality in Wikipedia Myshkin Ingawale
Amitava Dutta
Rahul Roy
Priya Seetharaman
Online Information Review 2012 Social media platforms allow near-unfettered creation and exchange of User Generated Content (UGC). We use Wikipedia, which consists of interconnected user generated articles. Drawing from network science, we examine whether high and low quality UGC in Wikipedia differ in their connectivity structures. Using featured articles as a proxy for high quality, we undertake a network analysis of the revision history of six different language Wikipedias to offer a network-centric explanation for the emergence of quality in UGC. The network structure of interactions between articles and contributors plays an important role in the emergence of quality. Specifically, the analysis reveals that high quality articles cluster in hubs that span structural holes. The analysis does not capture the strength of interactions between articles and contributors. The implication of this limitation is that quality is viewed as a binary variable. Extensions to this research will relate strength of interactions to different levels of quality in user generated content. Practical implications Our findings help harness the ‘wisdom of the crowds’ effectively. Organizations should nurture users and articles at the structural hubs, from an early stage. This can be done through appropriate design of collaborative knowledge systems and development of organizational policies to empower hubs. Originality The network centric perspective on quality in UGC and the use of a dynamic modeling tool are novel. The paper is of value to researchers in the area of social computing and to practitioners implementing and maintaining such platforms in organizations. 0 0
On the Use of PU Learning for Quality Flaw Prediction in Wikipedia Edgardo Ferretti
Donato Hernández Fusilier
Rafael Guzmán Cabrera
Manuel Montes y Gómez
Marcelo Errecalde
Paolo Rosso
PAN English 2012 In this article we describe a new approach to assess Quality Flaw Prediction in Wikipedia. The partially supervised method studied, called PU Learning, has been successfully applied in classifications tasks with traditional corpora like Reuters-21578 or 20-Newsgroups. To the best of our knowledge, this is the first time that it is applied in this domain. Throughout this paper, we describe how the original PU Learning approach was evaluated for assessing quality flaws and the modifications introduced to get a quality flaws predictor which obtained the best F1 scores in the task “Quality Flaw Prediction in Wikipedia” of the PAN challenge. 0 1
Quality of information sources about mental disorders: A comparison of Wikipedia with centrally controlled web and printed sources Reavley N.J.
MacKinnon A.J.
Morgan A.J.
Alvarez-Jimenez M.
Hetrick S.E.
Killackey E.
Nelson B.
Purcell R.
Yap M.B.H.
Jorm A.F.
Psychological Medicine English 2012 Background Although mental health information on the internet is often of poor quality, relatively little is known about the quality of websites, such as Wikipedia, that involve participatory information sharing. The aim of this paper was to explore the quality of user-contributed mental health-related information on Wikipedia and compare this with centrally controlled information sources. Method Content on 10 mental health-related topics was extracted from 14 frequently accessed websites (including Wikipedia) providing information about depression and schizophrenia, Encyclopaedia Britannica, and a psychiatry textbook. The content was rated by experts according to the following criteria: accuracy, up-to-dateness, breadth of coverage, referencing and readability. Results Ratings varied significantly between resources according to topic. Across all topics, Wikipedia was the most highly rated in all domains except readability. Conclusions The quality of information on depression and schizophrenia on Wikipedia is generally as good as, or better than, that provided by centrally controlled websites, Encyclopaedia Britannica and a psychiatry textbook. 0 2
QualityRank: Assessing quality of wikipedia articles by mutually evaluating editors and texts Yu Suzuki
Masatoshi Yoshikawa
HT'12 - Proceedings of 23rd ACM Conference on Hypertext and Social Media English 2012 In this paper, we propose a method to identify high-quality Wikipedia articles by mutually evaluating editors and texts. A major approach for assessing articles using edit history is a text survival ratio based approach. However, the problem is that many high-quality articles are identified as low quality, because many vandals delete high-quality texts, then the survival ratios of high-quality texts are decreased by vandals. Our approach's strongest point is its resistance to vandalism. Using our method, if we calculate text quality values using editor quality values, vandals do not affect any quality values of the other editors, then the accuracy of text quality values should improve. However, the problem is that editor quality values are calculated by text quality values, and text quality values are calculated by editor quality values. To solve this problem, we mutually calculate editor and text quality values until they converge. Using this method, we can calculate a quality value of a text that takes into consideration that of its editors. From experimental evaluation, we confirmed that the proposed method can improve the accuracy of quality values for articles. Copyright 2012 ACM. 0 0
Don't bite the newbies: how reverts affect the quantity and quality of Wikipedia work Aaron Halfaker
Aniket Kittur
John Riedl
WikiSym English 2011 Reverts are important to maintaining the quality of Wikipedia. They fix mistakes, repair vandalism, and help enforce policy. However, reverts can also be damaging, especially to the aspiring editor whose work they destroy. In this research we analyze 400,000 Wikipedia revisions to understand the effect that reverts had on editors. We seek to understand the extent to which they demotivate users, reducing the workforce of contributors, versus the extent to which they help users improve as encyclopedia editors. Overall we find that reverts are powerfully demotivating, but that their net influence is that more quality work is done in Wikipedia as a result of reverts than is lost by chasing editors away. However, we identify key conditions – most specifically new editors being reverted by much more experienced editors – under which reverts are particularly damaging. We propose that reducing the damage from reverts might be one effective path for Wikipedia to solve the newcomer retention problem. 0 2
Measuring article quality in Wikipedia: Lexical clue model Xu Y.
Luo T.
IEEE Symposium on Web Society English 2011 Wikipedia is the most entry-abundant on-line encyclopedia. Some studies published by Nature proved that the scientific entries in Wikipedia are of good quality comparable to those in the Encyclopedia Britannica which are mainly maintained by experts. But the manual partition of the articles in Wikipedia from a WikiProject implies that high-quality articles are usually reached grade by grade via being repeatedly revised. So many work address to automatically measuring the article quality in Wikipedia based on some assumption of the relationship between the article quality and contributors' reputations, view behaviors, article status, inter-article link, or so on. In this paper, a lexical clue based measuring method is proposed to assess article quality in Wikipedia. The method is inspired the idea that the good articles have more regular statistic features on lexical usage than the primary ones due to the more revise by more people. We select 8 lexical features derived from the statistic on word usages in articles as the factors that can reflect article quality in Wikipedia. A decision tree is trained based on the lexical clue model. Using the decision tree, our experiments on a well-labeled collection of 200 Wikipedia articles shows that our method has more than 83% precise and recall. 0 0
Quantifying the trustworthiness of social media content Moturu S.T.
Hongyan Liu
Distributed and Parallel Databases English 2011 The growing popularity of social media in recent years has resulted in the creation of an enormous amount of user-generated content. A significant portion of this information is useful and has proven to be a great source of knowledge. However, since much of this information has been contributed by strangers with little or no apparent reputation to speak of, there is no easy way to detect whether the content is trustworthy. Search engines are the gateways to knowledge but search relevance cannot guarantee that the content in the search results is trustworthy. A casual observer might not be able to differentiate between trustworthy and untrustworthy content. This work is focused on the problem of quantifying the value of such shared content with respect to its trustworthiness. In particular, the focus is on shared health content as the negative impact of acting on untrustworthy content is high in this domain. Health content from two social media applications, Wikipedia and Daily Strength, is used for this study. Sociological notions of trust are used to motivate the search for a solution. A two-step unsupervised, feature-driven approach is proposed for this purpose: a feature identification step in which relevant information categories are specified and suitable features are identified, and a quantification step for which various unsupervised scoring models are proposed. Results indicate that this approach is effective and can be adapted to disparate social media applications with ease. 0 0
The effects of wikis on foreign language students writing performance Alshumaimeri Y. Procedia - Social and Behavioral Sciences English 2011 This study investigated the use of wikis in improving writing skills among 42 male students at the Preparatory Year (PY) in King Saud University in Saudi Arabia. Research questions investigated writing accuracy and quality. Performance results on pre- and post-tests revealed that both groups improved significantly overtime in both accuracy and quality. However, the experimental group significantly outperformed the control group in both accuracy and quality of writing in the post-test. The implications of the results are that wikis can benefit teachers and students by improving their writing skills in accuracy and quality in a collaborative environment. 0 0
Answer reliability on Q&A sites Pnina Shachaf 16th Americas Conference on Information Systems 2010, AMCIS 2010 English 2010 Similar to other Web 2.0 platforms, user-created content on question answering (Q&A) sites raises concerns about information quality. However, it is possible that some of these sites provide accurate information while others do not. This paper evaluates and compares answer reliability on four Q&A sites. Content analysis of 1,522 transactions from Yahoo! Answers, Wiki Answers, Askville, and the Wikipedia Reference Desk, reveals significant differences in answer quality among these sites. The most popular Q&A site (that attracts the largest numbers of users, questions, and answers) provides the least accurate, complete, and verifiable information. 0 0
Determinants of Wikipedia Quality: the Roles of Global and Local Contribution Inequality Oded Nov Ofer Arazy Computer Supported Cooperative Work (CSCW) 2010 The success of Wikipedia and the relative high quality of its articles seem to contradict conventional wisdom. Recent studies have begun shedding light on the processes contributing to Wikipedia’s success, highlighting the role of coordination and contribution inequality. In this study, we expand on these works in two ways. First, we make a distinction between global (Wikipedia-wide) and local (article-specific) inequality and investigate both constructs. Second, we explore both direct and indirect effects of these inequalities, exposing the intricate relationships between global inequality, local inequality, coordination, and article quality. We tested our hypotheses on a sample of a Wikipedia articles using structural equation modeling and found that global inequality exerts significant positive impact on article quality, while the effect of local inequality is indirect and is mediated by coordination. 0 1
A jury of your peers: Quality, experience and ownership in Wikipedia Aaron Halfaker
Aniket Kittur
Kraut R.
John Riedl
WikiSym English 2009 Wikipedia is a highly successful example of what mass collaboration in an informal peer review system can accomplish. In this paper, we examine the role that the quality of the contributions, the experience of the contributors and the ownership of the content play in the decisions over which contributions become part of Wikipedia and which ones are rejected by the community. We introduce and justify a versatile metric for automatically measuring the quality of a contribution. We find little evidence that experience helps contributors avoid rejection. In fact, as they gain experience, contributors are even more likely to have their work rejected. We also find strong evidence of ownership behaviors in practice despite the fact that ownership of content is discouraged within Wikipedia. Copyright 0 6
Issues of cross-contextual information quality evaluation—The case of Arabic, English, and Korean Wikipedias Besiki Stvilia
Abdullah Al-Faraj
& Yong Jeong Yi.
Library & Information Science Research, 31(4), 232-239 2009 Objective: An initial exploration into the issue of information quality evaluation across different cultural and community contexts based on data collected from the Arabic, English, and Korean Wikipedias showed that different Wikipedia communities may have different understandings of and models for quality. It also showed the feasibility of using some article edit-based metrics for automated quality measurement across different Wikipedia contexts. A model for measuring context similarity was developed and used to evaluate the relationship between similarities in sociocultural factors and the understanding of information quality by the three Wikipedia communities. 0 0
Wikibugs: Using template messages in open content collections Loris Gaio
Den Besten M.
Alessandro Rossi
Dalle J.-M.
WikiSym English 2009 In the paper we investigate an organizational practice meant to increase the quality of commons-based peer production: the use of template messages in wiki-collections to highlight editorial bugs and call for intervention. In the context of SimpleWiki, an online encyclopedia of the Wikipedia family, we focus on {complex}, a template which is used to flag articles disregarding the overall goals of simplicity and readability. We characterize how this template is placed on and removed from articles and we use survival analysis to study the emergence and successful treatment of these bugs in the collection. Copyright 0 0
Comparison of Wikipedia and other encyclopedias for accuracy, breadth, and depth in historical articles Lucy Holman Rector Reference Services Review English 2008 Purpose – This paper seeks to provide reference librarians and faculty with evidence regarding the comprehensiveness and accuracy of Wikipedia articles compared with respected reference resources.

Design/methodology/approach – This content analysis evaluated nine Wikipedia articles against comparable articles in Encyclopaedia Britannica, The Dictionary of American History and American National Biography Online in order to compare Wikipedia's comprehensiveness and accuracy. The researcher used a modification of a stratified random sampling and a purposive sampling to identify a variety of historical entries and compared each text in terms of depth, accuracy, and detail.

Findings – The study did reveal inaccuracies in eight of the nine entries and exposed major flaws in at least two of the nine Wikipedia articles. Overall, Wikipedia's accuracy rate was 80 percent compared with 95-96 percent accuracy within the other sources. This study does support the claim that Wikipedia is less reliable than other reference resources. Furthermore, the research found at least five unattributed direct quotations and verbatim text from other sources with no citations.

Research limitations/implications – More research must be undertaken to analyze Wikipedia entries in other disciplines in order to judge the source's accuracy and overall quality. This paper also shows the need for analysis of Wikipedia articles' histories and editing process.

Practical implications – This research provides a methodology for further content analysis of Wikipedia articles.

Originality/value – Although generalizations cannot be made from this paper alone, the paper provides empirical data to support concerns regarding the accuracy and authoritativeness of Wikipedia.
0 4
The quality and trust of wiki content in a learning community Peacock T.
Fellows G.
Eustace K.
ASCILITE 2007 - The Australasian Society for Computers in Learning in Tertiary Education English 2007 User generated content is having an ever-increasing influence and presence on the Internet. Wiki communities, in particular Wikipedia, have gained wide spread attention and criticism. This research explores criticisms and strengths of wiki communities, and methods to reconcile the two. This research tests wiki software in an educational setting to determine indicators of article quality. The results give insight into the use of wiki systems in educational settings, suggest possible methods of improving the validity of content created within wiki communities, and provide groundwork for further research in the area. 0 0
Wikipedia vs The Old Guard PC Pro English 2007 0 0
An empirical exploration of Wikipedia's credibility Thomas Chesney First Monday English 6 November 2006 Wikipedia is an free, online encyclopaedia; anyone can add content or edit existing content. The idea behind Wikipedia is that members of the general public can add their own personal knowledge, anonymously if they wish. Wikipedia then evolves over time into a comprehensive knowledge base on all things. Its popularity has never been questioned, although some have speculated about its authority. By its own admission, Wikipedia contains errors. A number of people have tested Wikipedia’s accuracy using destructive methods, i.e. deliberately inserting errors. This has been criticised by Wikipedia. This short study examines Wikipedia’s credibility by asking 258 research staff with a response rate of 21 percent, to read an article and assess its credibility, the credibility of its author and the credibility of Wikipedia as a whole. Staff were either given an article in their own expert domain or a random article. No difference was found between the two group in terms of their perceived credibility of Wikipedia or of the articles’ authors, but a difference was found in the credibility of the articles — the experts found Wikipedia’s articles to be more credible than the non–experts. This suggests that the accuracy of Wikipedia is high. However, the results should not be seen as support for Wikipedia as a totally reliable resource as, according to the experts, 13 percent of the articles contain mistakes. 0 0
Internet encyclopaedias go head to head Jim Giles Nature English 14 December 2005 Jimmy Wales' Wikipedia comes close to Britannica in terms of the accuracy of its science entries, a Nature investigation finds. 0 50