=Paper=
{{Paper
|id=Vol-1177/CLEF2011wn-PAN-PalkovskiiEt2011
|storemode=property
|title=Using WordNet-based Semantic Similarity Measurement in External Plagiarism Detection - Notebook for PAN at CLEF 2011
|pdfUrl=https://ceur-ws.org/Vol-1177/CLEF2011wn-PAN-PalkovskiiEt2011.pdf
|volume=Vol-1177
}}
==Using WordNet-based Semantic Similarity Measurement in External Plagiarism Detection - Notebook for PAN at CLEF 2011==
Using WordNet-based semantic similarity measurement in External Plagiarism Detection Notebook for PAN at CLEF 2011 Yurii Palkovskii, Alexei Belov, Iryna Muzyka Zhytomyr State University, SkyLine. Inc, "MARS" p.e. palkovskiy@yandex.ru 1 Introduction Continuing our previous work started at PAN 2009 and PAN 2010 [7] we considered further research options based on the achieved baseline of the best performing algorithms. The research done by Potthast et al. [4] presented a sliced view of the presented approaches showing their performance on specific corpus metrics - external\intrinsic, obfuscation strategies (none, artificial high\low, simulated, translated), topic match, case length and document length thus defining the baseline for further studies. A brief analysis of the above named results [1,3] shows that there exists a direct correlation between the obfuscation degree (method) and the achieved performance. Table 1. Detection Efficiency (DE% Recall::Precision) in relation to "no obf." as a base value: 1 result 2 result obfuscation type: P R DE% P R DE% no obfuscation 94 96 100% 78 86 100% low 93 92 98% 81 85 103% high 93 75 98% 76 76 97% translation 92 70 97% 58 47 74% simulated 33 18 35% 19 22 24% I the above table the complexity of obfuscation rises from "no obfuscation" to "simulated" type. An important note should be made - we rearranged "translated" plagiarism with "simulated" treating the latter as the most advanced form of obfuscation due to the fact that the exact "translation" mechanism used for cases generation, followed the "bag of words" pattern, instead of Mechanical Turk human translation. The achieved score proves the above idea and opens a new vector for further developing the corpus including human made translations. Additionally, some minor criticism has been put forward in relation to the PAN plagiarism detection "definition of plagiarism encompasses much more than just character sequence matching" [2]. This particular opinion and the performance baseline difference mentioned above led us to the idea of switching from the quantum based statistical n-gram probing (n-gram fingerprinting and TFIDF measurement) to semantic similarity measurement (SSM). Our primary hypothesis stated that SSM used as a main comparer must be more resilient to simulated plagiarism and translated plagiarism as it relies on sense quantum instead of character quantum. The success of the SSM heavily depends on the exact method of plagiarism construction. Artificial plagiarism, starting from low to high obfuscation levels and including translated plagiarism, constructed by the PAN Random Plagiarist follows "bag-of-words" approach [4] and lacks many features that differentiate it from "human generated rewrite" thus giving the ground for the above mentioned criticism. Among these missing features we can name correct word order, correct grammar, writing style, word choice, etc. and more importantly - the sense structure behind the text that conveys sense itself. In an effort to overcome these limitations PAN 2010 has launched "simulated" plagiarism sections into the corpus. Generated by Mechanical Turk with multiple quality validation followed, these cases have become our primary target for SSM experiments as they represented the real world plagiarism cases generated by human, that convey all the previously missing features mentioned above. 2 External Plagiarism Detection Due to serious time constraints we decided to use the previously developed system for candidate document retrieval and focus our research on the SSM comparer development. Selecting the SSM as our primary document similarity comparison we were aware that most probably we will not be able to achieve better or even comparable results in relation to the existing PAN baseline due the following factors - the majority of plagiarism cases are not simulated but artificial plagiarism, extremely heavy performance load on the comparer that resulted in excessive processing time requirements. Still we strongly believe that SSM is the future of plagiarism detection so we decided to pursue this particular method to discover its possible benefits. Initially we planned to build our SSM comparer from scratch building it up around the idea of measuring the distance via WordNet synsets but when we discovered the Troy Simpson's project on SSM [5] we decided to use this project as a foundation of your own prototype. Our SSM text comparer comprises several open-source projects, namely: • WordNet 2.1 • WordNet.Net an open-source .NET Framework library for WordNet • C# Porter Stemmer implementation • C# Brill Tagger implementation • SSM Words Comparer by Troy Simpson and Thanh Dao • SSM Sentence Comparer by Troy Simpson and Thanh Dao The strategy to capture semantic similarity between two sentences. Given two sentences X and Y, we denote m to be length of X, n to be length of Y. The major steps can be described as follows: 1. Tokenization. 2. Perform word stemming. 3. Perform part of speech tagging. 4. Word sense disambiguation. 5. Building a semantic similarity relative matrix R[m, n] of each pair of word senses, where R[i, j] is the semantic similarity between the most appropriate sense of word at position i of X and the most appropriate sense of word at position j of Y. Thus, R[i,j] is also the weight of the edge connecting from i to j. 6. We formulate the problem of capturing semantic similarity between sentences as the problem of computing a maximum total matching weight of a bipartite graph, where X and Y are two sets of disjoint nodes 7. The match results from the previous step are combined into a single similarity value for two sentences. There following strategy was used: According to Dhanh Tao approach [5,6], the path length-based similarity measurement example the length between car and auto is 1, car and truck is 3, car and bicycle is 4, car and fork is 12. Figure 1. Hyponym taxonomy in WordNet used for path length similarity measurement [5,6]: At text level sliding window approach is used to utilize the SSM comparer. This approach was inspired by the style changing function measurement approach used in PAN 2010 intrinsic plagiarism detectors [3]. 3 Evaluation For detailed evaluation of the progress we developed an application to prepare a pre-selected corpora from the test corpus of PAN 2010 using the meatacriteria of the corpus itself. As we focused on the simulated plagiarism mainly - we trained our SSM comparer on this specific corpus. Taking into the consideration large amount of processing to be done and the lack of code optimization we need to further investigate the differences between our intermediate results that were achieved on our sub-corpus and the test corpus of the 2011. We suspect that PAN 2011 corpus has some specifics that presupposed the baseline change from 79% to 50% that will be announced at the workshop. 4 Conclusion Concluding our research, we would like to note that using SSM as plagiarism detection mechanism has to be much improved to achieve competitive results. Our research is a small step towards the better understanding of semantic similarity usage at large scale data processing. Migration to a more productive languages and or infrastructure with possible cauterization may yield better results to face the conditions of the PAN competition and we strongly believe that we will be able to launch full semantic search for the next PAN Competition. Among further research vectors we can name the following: 1. Better WSD methods. 2. Overall speed\performance optimization. 3. Further research on the problem of Semantic Hashing and Search. 4. Further research on the problem of semantic normalization. 5. Better definition of meta parameters used via machine learning. We suspect that SSM will do best on human translated plagiarism but this needs to be proved on a larger corpus yet. Bibliography [1] Cristian Grozea and Marius Popescu. Encoplot—Performance in the Second International Plagiarism Detection Challenge: Lab Report for PAN at CLEF 2010. In Braschler et al. ISBN 978-88-904810-0-0 [2] Debora Weber-Wulff, "Plagiarism Detection Competition" copy-shake-paste.blogspot.com. 2009. 21 June.2011. [3] Markus Muhr, Roman Kern, Mario Zechner, and Michael Granitzer. External and Intrinsic Plagiarism Detection using a Cross-Lingual Retrieval and Segmentation System: Lab Report for PAN at CLEF 2010. In Braschler et al. [2]. ISBN 978-88- 904810-0-0. [4] Martin Potthast, Benno Stein, Andreas Eiselt, Alberto Barrón-Cedeño, and Paolo Rosso. Overview of the 1st International Competition on Plagiarism Detection. In Benno Stein, Paolo Rosso, Efstathios Stamatatos, Moshe Koppel, and Eneko Agirre, editors, SEPLN 2009 Workshop on Uncovering Plagiarism, Authorship, and Social Software Misuse (PAN 09), pages 1–9. CEUR-WS.org, September 2009. URL http://ceur-ws.org/Vol-502. [5] Thanh Dao. "An improvement on capturing similarity between strings" www.codeproject.com. 2005. 29 Jul. 2011. http://www.codeproject.com/KB/recipes/improvestringsimilarity.aspx [6] Troy Simpson, Thanh Dao. "WordNet-based semantic similarity measurement" www.codeproject.com. 2005. 1 Oct. 2011. http://www.codeproject.com/KB/string/semanticsimilaritywordnet.aspx [7] Yurii Palkovskii, Alexei Belov, and Irina Muzika. Exploring Fingerprinting as External Plagiarism Detection Method: Lab Report for PAN at CLEF 2010. In Braschler et al. [2]. ISBN 978-88-904810-0-0.