List of master's theses
This is a list of all the master's theses available in WikiPapers. Currently, there are 26 master's theses.
To create a new "master's thesis" go to Form:Publication.
|Title||Author(s)||Keyword(s)||Published in||Language||DateThis property is a special property in this wiki.||Abstract||R||C|
|Wikimédia France : un modèle de la contribution ?||Céline Rabaud||French||June 2015||4||0|
|Fidarsi di Wikipedia||Simone Dezaiacomo||Wikipedia
Teoria delle decisioni e processi cognitivi
|Italian||15 July 2014||Lo scopo dello studio è comprendere i fenomeni alla base della fiducia degli utenti verso l'enciclopedia online Wikipedia. Per farlo è necessario prima di tutto comprendere e modellizzare l'organizzazione della struttura dei processi socio-produttivi sottostanti alla produzione del contenuto di Wikipedia, procedendo quindi nel verificare empiricamente e descrivere le capacità di autocorrezione della stessa. Oltre a quelli utilizzati in questo studio, saranno anche descritti gli approcci e i risultati trattati in letteratura, riportando i principali studi che nel corso degli anni hanno affrontato questi argomenti, sebbene mantenendoli indipendenti.
Per comprendere la struttura della community degli editor di Wikipedia, si è ipotizzata l'esistenza di un modello di tipo Core-Periphery. Per studiare il modello sono state eseguite delle analisi su dati derivanti da un campione di pagine della versione italiana di Wikipedia. I risultati ottenuti dall'analisi di queste informazioni rappresentano le basi utilizzate per la selezione delle pagine oggetto dell'iniezione degli errori, costituendo un metodo per stimare le diverse probabilità di autocorrezione per ciascuna pagina. Per quanto riguarda le capacità di resilienza di Wikipedia, i risultati sono ricavati utilizzando un approccio empirico. Questo consiste nell'inserimento di errori all'interno del campione di pagine sotto specifici vincoli metodologici per poi valutare in quanto tempo e con quali modalità questi errori vengono corretti.
E' stata effettuata un'analisi specifica per la scelta delle tipologie di errore e delle variabili da considerare nell'inserimento di questi.Questa analisi ha portato alla definizione di 2 esperimenti tra loro distinti, i cui risultati portano ad interessanti conclusioni sia visti separatamente che combinati tra loro. Sulla base dei risultati di questi esperimenti è stato possibile discutere sulle capacità di autocorrezione del sistema, elemento chiave nello studio delle dinamiche della fiducia verso Wikipedia.
|Secure Wiki System: A plugin-based solution to wiki security||Kasper Lindberg||English||March 2012||Wiki systems have become an important tool for knowledge sharing among people. From the small wikis for knowledge sharing in organizations to the larger project-related wikis on the Internet. In addition, Wikipedia, which is in a class of its own when it comes to size, has managed to collect an impressive amount of information based solely on the cooperation between strangers from around the world. Any open wiki, with a user-community so large that members of the community have a certain degree of anonymity, suffer from the effects of directed and random vandalism. This vandalism is a problem because it reduces the trustworthiness of the content provided by the wiki system. The secure wiki model is an integrity model that has been proposed to help prevent vandalism and improve the trustworthiness of articles in wiki system. This model is based on both static and dynamic document access controls, which enforce a simple integrity based security policy. This thesis improves this model by proposing a new policy for use with the model. The proposed policy is evaluated and compared to the original policy. The evaluation shows that the new policy is highly configurable and can be configured in such a way that it requires significantly less reviewers than the original policy, which can benefit small systems with a low number of users. An implementation of a base wiki system have been created, which on its own equals any other wiki in terms of its vulnerability to vandalism. In addition to this, an implementation of the secure wiki model has also been created. The implementation is made as a plugin to the base wiki system and adds an integrity model to the existing soft-security model that is used by the base system and other wiki implementations. The integrity model provides harder security guarantees and limits the ability of attackers to compromise the integrity of wiki articles, without compromising the all can edit policy of open wiki systems.||6||0|
|Valorisation du bénévolat sur Wikipédia||Vincent Juhel||French||February 2012||Wikipédia a un fonctionnement atypique dont les recherches s’attardent majoritairement autour de la qualité des articles potentiellement rédigés par n’importe qui. J’ai cherché par cette thèse professionnelle à présenter un regard quantitatif et qualitatif de la véritable valeur que ce projet apporte aux lecteurs, rédacteurs, donateurs mais également ce qu’il aurait représenté s’il avait été une entreprise classique. Le premier objectif était d’évaluer la valeur du travail de ces bénévoles, qui, en dépit sa gratuité, apporte une véritable richesse. Mieux définir cette richesse, c’est aussi mieux convaincre les donateurs et avoir plus de poids vis à vis des partenaires. Le deuxième objectif a été de définir les contours d’une stratégie cherchant à maximiser la valeur produite par une communauté de bénévoles en grande partie autogérée. Mieux maîtriser la valeur produite pour mieux orienter et motiver le travail des contributeurs.||0||1|
|Avoimen suomenkielisen morfologian liittäminen Wikimedian hakujärjestelmään||Niklas Laxström||University of Helsinki||Finnish||1 January 2012||In my thesis I investigated the feasibility of using a Finnish morphology implementation with the Lucene search system. With the same Lucene-search package that is used by the Wikimedia Foundation I built two search indexes: one with the existing Porter stemming algorithm and the other one with morphological analysis. The corpus I used was the current text dump of Finnish Wikipedia. [...] See http://laxstrom.name/blag/2012/02/13/exploring-the-states-of-open-source-search-stack-supporting-finnish/||9||0|
|Källkritik||J. Kolmodin||Lärarutbildning, Mälardalens högskola||Swedish||2012||0||0|
|Wikipedia i undervisning, Wikipedia and education||P. Cöster-Ahl||Lärarutbildning, Malmö Högskola||Swedish||2012||0||0|
|Aufbau eines wissenschaftlichen Textcorpus auf der Basis der Daten der englischsprachigen Wikipedia||Markus Fuchs||University of Regensburg||German||2009||With the growth in popularity over the last eight years, Wikipedia has become a very promising resource in academic studies. Some of its properties make it attractive for a wide range of research fields (information retrieval, information extraction, natural language processing, ...), e.g. free availability and up to date content. However, efficient and structured access to this information is not easy, as most of Wikipedia's contents are encoded in its own markup language (wikitext). And, unfortunately, there is no formal definition of wikitext, which makes parsing very difficult and burdensome. In this thesis, we present a system that lets the researcher automatically build a richly annotated corpus containing the information most commonly used in research projects. To this end, we built our own wikitext parser based on the original converter used by Wikipedia itself to convert wikitext into HTML. The system stores all data in a relational database, which allows for efficient access and extensive retrieval functionality.||0||0|
|Conflicts in collaboration: a study of tensions in the process of collective writing on Web 2.0||Aline de Campos||Federal University of Rio Grande do Sul - UFRGS||Portuguese||2009||From the context of collaboration as a process of the collective intelligence (LÉVY, 2003) and wisdom of crowds (SUROWIECKI, 2006) this work aims to study the conflict as an important factor of these collective processes. The imbalances are part of human history, however, recurrently we find a purpose to see this issues as negative and of annihilation of relations, leaving aside the aspect of the potential impulsion to beneficial reconfiguration of the process in which it operates. An interesting practice in the study of conflicts in collaboration is the online collective writing. The independence of space and time and the multiplicity of voices that can focus on a project textual, opens space for negotiations, debates and tensions of various kinds. In addition, sometimes there is too optimistic view, which relieves the issues of structural, dynamics and behavior in the production of meaning and the conflict that may arise from it. In this sense, is questioned: what is the influence of conflicts in the process of online collective writing? This project, through theoretical and empirical research, seeks to answer this question in a communicational perspective, which takes into account the relationships and interactions beyond the harmony permanently assigned to these processes for various areas of knowledge. For the empirical verification are presented two collaborative projects of textual production: the Wikipedia, the free encyclopedia, wide used and with popularity and the project of the Laboratory of Computer-mediated Interaction of the Federal University of Rio Grande do Sul, the Co-dex, the social dictionary, environment created for concepts, reviews and biographies of the area of communication and information science. In both observations were made for systematic verification of the tensions from the interaction and collective production so that, together with the developed theoretical contribution, the guiding question of this work could be investigated. It is concluded that the conflicts that permeate the processes of collaborative production of meaning relate to aspects of a textual and relational order and that they are related and influence each other. It is believed that the tensions have strong relevance in the development of relations between the collaborators and building content to introduce imbalances needed for a "majorant reequilibration" (Piaget, 1977).||0||0|
|Mining Structured Information from Unstructured Texts using Wikipedia-based Semantic Analysis||Amir Hossein Jadidinejad||Islamic Azad University||Persian||2009||0||0|
|Motivational voluntary knowledge sharing among users in the open source encyclopedia; case study (the Arabic Wikipedia)||Mohammed N M Abu-Shaaban||Edinburgh Napier University||English||2009||This research aims to examine the motives behind participation in the Arabic version of Wikipedia. Building on previous research, it argues that there are four main motivational factors: egoism, altruism, principalism and collectivism. Data is collected to examine the importance of each factor. Data collection occurs in two stages; first, Wikipedia content is researched and examined for statements that denote motivational behaviour and second, interviews are carried out with a sample of Arabic Wikipedians to discuss the results of the first stage and reach more definitive conclusions. The research also finds a correlation between the various contents of the articles (religious, political or country-related) and the motivational factor behind contributing users in a given article.||0||0|
|Um Método Automático para Estimativa da Qualidade de Enciclopédias Colaborativas On-Line: Um Estudo de Caso Sobre a Wikipédia||Daniel Hasan Dalip||Universidade Federal de Minas Gerais||Portuguese||2009||The old dream of a universal repository containing all the human knowledge and culture is becoming possible through the Internet and the Web. Moreover, this is happening with the direct collaborative, participation of people. Wikipedia is a great example. It is an enormous repository of information with free access and edition, created by the community in a collaborative manner. However, this large amount of information, made available democratically and virtually without any control, raises questions about its relative quality. In this work we explore a significant number of quality indicators, some of them proposed by us and used here for the first time, and study their capability to assess the quality of Wikipedia articles. Furthermore, we explore machine learning techniques to combine these quality indicators into one single assessment judgment. Through experiments, we show that the most important quality indicators are the easiest ones to extract on a open digital library, namely, textual features related to length, structure and style. We were also able to determine which indicators did not contribute significantly to the quality assessment. These were, coincidentally, the most complex features, such as those based on link analysis. Finally, we compare our combination method with state-of-the-art solutions and show significant improvements in terms of effective quality prediction.||0||0|
|Vandalism Detection in Wikipedia: a Bag-of-Words Classifier Approach||A. Belani||Cornell University||English||2009||A bag-of-words based probabilistic classifier is trained using regularized logistic regression to detect vandalism in the English Wikipedia. Isotonic regression is used to calibrate the class membership probabilities. Learning curve, reliability, ROC, and cost analysis are performed.||0||0|
|Wiki Culture: En analyse af organisatorisk samarbejde på Wikipedia||Niels Møller Christensen||IT University of Copenhagen||Danish||2009||This thesis investigates the organisational structures behind the collaborative work on Wikipedia. The objectives are to identify formal an informal group dynamics, the relation between the individual writer and the community and furthermore how culture and norms develop in relation to the collaborative article writing. In addition the thesis discusses the power structures of Wikipedia, and whether or not informal hierarchical structures emerge despite the open and relatively flat structure of Wikipedia. This study is made through a discourse analysis of a Wikipedia article and the underlying communication between the writers of the article. The analysis is supplemented visually by discourse mapping and quantitatively by the use of statistical material obtained by data gathering and processing. The analysis is based on both sociological and media theories – especially concerning social media. The primary sociological theories are Jean Lave and Etienne Wenger's theory of community of practice and Michel Maffesoli's theories concerning neotribalism. The primary media theories are Adrian Mackenzie's theory of pre-individualism and Pierre Levy's theories of anthropological spaces and collective intelligence. The analysis shows that informal and hierarchical group structures emerge through the collaboration of Wikipedia articles despite the open and flat principles of the encyclopedia. Furthermore the analysis indicates that these group structures are upheld through both practical work and development of mutual norms, culture and language and that an interplay between individual and collective identity of the writers is necessary for the collaboration on articles.||0||0|
|WikiNet: Construction d'une ressource lexico-sémantique multilingue à partir de Wikipedia||Samuel Reese||Institut Supérieur de l'Aéronautique et de l'Espace||French||2009||0||0|
|Fine-grained addressability to support large-scale collaborative document development||Viral Gupta||Computer science
|English||January 2008||The Web has made it possible for large, distributed collaborations to develop sophisticated document bases. As these collaborations increase in size, there is a need to support reference, navigation and search through the document bases by individuals who are not computer pro-fessionals. Although collaborative tools such as wikis have been developed that address some of the requirements of these communities, the tools lack support for simple and convenient fine-grained addressability to parts of the documents. Such addressability is essential for formal documents, such as standards and legal documents. In this thesis we develop a solution to the problem of fine-grained addressability that is based on MediaWiki, a popular and powerful col- laboration tool that is the software infrastructure for Wikipedia. Although some collaboration tools that support fine-grained addressability already exist, they have not addressed some of the open research issues of fine-grained addressability, such as dealing with transclusion, semantic annotation and hyperscope support. It also deals with the research problems that were raised by this goal. An architecture and reference implementation was developed to provide a proof of concept and to test the viability of the proposed solutions to the research problems. The thesis also discusses the various design decisions that were made in the course of solving the research problems and developing the reference implementation.||0||0|
|Methoden zur sprachübergreifenden Plagiaterkennung||Maik Anderka||University of Paderborn||German||2007||0||0|
|The transformation of the encyclopedia : a textual analysis and comparison of the Encyclopædia Britannica and Wikipedia||Mark W. Bell||Ball State University||2007||0||0|
|Wer schreibt die Wikipedia? Die Online-Enzyklopädie in der Vorstellungs- und Lebenswelt ihrer aktivsten Autoren.||Andreas Möllenkamp||University of Leipzig||German||2007||0||0|
|Wikipedia as Collective Action: Personal incentives and enabling structures||Benjamin Keith Johnson||Michigan State University||2007||Wikipedia is an online encyclopedia created by volunteers, and is an example of how developments in software platforms and the low cost of sharing and coordinating on the Internet are leading to a new paradigm of creative collaboration on a massive scale. This thesis addresses the questions of why individuals choose to give away their time and effort and how the challenges associated with collective action are addressed by Wikipedia’s technologies, organization, and community. Interviews with editors of the encyclopedia were used to identify what personal gains and other motivations compel contributors, what challenges to collaboration exist, and what technological and social structures aid their ability to create a freely available repository of human knowledge.||0||1|
|Wikipédia: le Refus du Pouvoir||Sylvain Firer-Blaess||Institut d'études politiques de Lyon||French||2007||0||0|
|Wolnosc i Wiedza: Aksjonormatywny Wymiar Wikipedii (Freedom and Knowledge: The Axiomatic Dimension of Wikipedia)||Julian Madej||Warsaw Univeristy||2007||0||0|
|A Study of Phenomena of Knowledge Sharing in Wikipedia||Chun-yu Huang||National Central University, Taiwan||Chinese||2006||Wikipedia is an encyclopedia on the Internet. It provides a lot of knowledge for the user. The first Wikipedia appeared in 2001 and was only in English. After six year of development, there are now various versions in more than 250 languages. Contents in Wikipedia were contributed and edited not by authorities, but by users of Wikipedia. As long as one wants, one can contribute to the contents of Wikipedia. Many users spent their time and energy to devote themselves to Wikipedia. Wikipedia gives no monetary reward to its contributor, but there are more and more users sharing their knowledge to Wikipedia. Does this reveal a massive pro-social phenomenon? This study thus attempts to look into factors that effect knowledge sharing of these sharing individuals. A web based questionnaire was designed, and known Wikipedia users were invited as informants. 156 valid samples were tallied out of a total of 181 returns. Empirical results reveal that reputation and altruism have positive effects on attitude of knowledge sharing, while expected reward has significant but negative effect on attitude of knowledge sharing. External control and community identification have moderating effect on the relationship between attitude of knowledge sharing and behavior of knowledge sharing. However, we failed to find evidence that support the effect of attitude of knowledge sharing on behavior of knowledge sharing. This is an issue that calls for more studies.||0||0|
|Automatic Ontology Extraction for Document Classification||Natalia Kozlova||Saarland University||2006||The amount of information in the world is enormous. Millions of documents in electronic libraries, thousands of them on each personal computer waiting for the expert to organize this information, to be assigned to appropriate categories. Automatic classification can help. However, synonymy, polysemy and word usage patterns problems usually arise. Modern knowledge representation mechanisms such as ontologies can be used as a solution to these issues. Ontology-driven classification is a powerful technique which combines the advantages of modern classification methods with semantic specificity of the ontologies. One of the key issues here is the cost and difficulty of the ontology building process, especially if we do not want to stick to any specific field. Creating a generally applicable but simple ontology is a challenging task. Even manually compiled thesauri such as WordNet can be overcrowded and noisy. We propose a flexible framework for efficient ontology extraction in document classification purposes. In this work we developed a set of ontology extraction rules. Our framework was tested on the manually created corpus of Wikipedia, the free encyclopedia. We present a software tool, developed with regard to the claimed principles. Its architecture is open for embedding new features in. The ontology-driven document classification experiments were performed on the Reuters collection. We study the behavior of different classifiers on different ontologies, varying our experimental setup. Experiments show that the performance of our system is better, in comparison to other approaches. In this work we observe and state the potential of automatic ontology extraction techniques and highlight directions for the further investigation.||0||0|
|D'Amour et de Neutralité||Martin Erpicum||French||2005||0||0|