scispace - formally typeset
Search or ask a question
Topic

Plagiarism detection

About: Plagiarism detection is a research topic. Over the lifetime, 1790 publications have been published within this topic receiving 24740 citations.


Papers
More filters
Journal Article
TL;DR: The experiments revealed that the best retrieval performance is obtained after removal of in-code comments and applying a combined weighting scheme based on terms frequencies, normalized term frequencies, and a cosine-based document normalization.
Abstract: Latent Semantic Analysis (LSA) is an intelligent information retrieval technique that uses mathematical algorithms for analyzing large corpora of text and revealing the underlying semantic information of documents. LSA is a highly parameterized statistical method, and its effectiveness is driven by the setting of its parameters which are adjusted based on the task to which it is applied. This paper discusses and evaluates the importance of parameterization for LSA based similarity detection of source-code documents, and the applicability of LSA as a technique for source-code plagiarism detection when its parameters are appropriately tuned. The parameters involve preprocessing techniques, weighting approaches; and parameter tweaking inherent to LSA processing – in particular, the choice of dimensions for the step of reducing the original post-SVD matrix. The experiments revealed that the best retrieval performance is obtained after removal of in-code comments (Java comment blocks) and applying a combined weighting scheme based on term frequencies, normalized term frequencies, and a cosine-based document normalization. Furthermore, the use of similarity thresholds (instead of mere rankings) requires the use of a higher number of dimensions. Povzetek: Prispevek analizira metodo LSA posebej glede plagiarizma izvirne kode.

23 citations

Proceedings ArticleDOI
18 Oct 2007
TL;DR: It is found that plagiarism that cannot be detected by the traditional methods can be identified by this new approach, and application of natural language processing can help to resolve this kind of problem.
Abstract: The problem of plagiarism has existed for a long time but with the advance of information technology the problem becomes worse. It is because there are many electronic versions of published materials available to everyone. The Web is an important and common source for plagiarism. Some plagiarism detection programs (such as Turnitin) were developed to attempt to deal with this problem. To determine whether an article is copied from the Web or other electronic sources, the plagiarism detection program should calculate the similarity between two articles. However, it is often difficult to detect plagiarism accurately after modification of the copied contents. For example, it is possible to simply replace a word with its synonym (e.g. "program" -- "software ") and change the entire sentence structure. Most plagiarism detection programs can only compare whether two words are the same lexically and count how many matched words are there in a paper. Thus, if the copied materials are modified deliberately, it becomes difficult to detect plagiarism.Application of natural language processing can help to resolve this kind of problem. The underlying syntactic structure and semantic meaning of two sentences can be compared to reveal their similarity. There are several steps in the matching procedure. First, the thesaurus (or the lexical hierarchical structure) is referenced to find out the synonyms, broader terms and narrower terms used in the paper being checked. Then, the paper will be compared with the documents in the database. Wordnet is a typical example of the thesaurus that can be used for this purpose. If it is suspected that the paper contains some contents from the database, the sentences of the paper may be parsed to construct their parsing trees and semantic representations for further detailed comparison. The context free grammar and the case grammar are used to represent the syntactic structure and semantic meaning of sentences in the system. It is found that plagiarism that cannot be detected by the traditional methods can be identified by this new approach.

23 citations

Proceedings Article
01 Jan 2007
TL;DR: In this work it is shown how a natural language parser can be used to fight against basic plagiarism hiding methods.
Abstract: The problem of plagiarism detection system design is a subject of numerous works of the last decades. Various advanced file-file comparison techniques were developed. However, most existing systems, aimed at natural language texts, do not perform any significant preprocessing of the input documents. So in many cases it is possible to hide the presence of plagiarism by utilizing some simple techniques. In this work we show how a natural language parser can be used to fight against basic plagiarism hiding methods.

23 citations

Proceedings ArticleDOI
29 May 2012
TL;DR: Evaluation results using real student program assignments show the effectiveness of the proposed method to automatically detect plagiarisms, i.e. illegal copies, among a set of programs submitted by students in elementary programming courses.
Abstract: This paper proposes a method to automatically detect plagiarisms, i.e. illegal copies, among a set of programs submitted by students in elementary programming courses. In such courses, programming assignments are so simple that submitted programs are very short and similar to each other. Existing plagiarism detection methods, therefore, may yield many false positive results. The proposed method solves the problem by using three types of similarity: code, comment, and inconsistence. The inconsistence similarity, a unique feature of the method, improves the precision and recall ratios, and helps to find evidences of plagiarisms. Evaluation results using real student program assignments show the effectiveness of the proposed method.

23 citations


Network Information
Related Topics (5)
Active learning
42.3K papers, 1.1M citations
78% related
The Internet
213.2K papers, 3.8M citations
77% related
Software development
73.8K papers, 1.4M citations
77% related
Graph (abstract data type)
69.9K papers, 1.2M citations
76% related
Deep learning
79.8K papers, 2.1M citations
76% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
202359
2022126
202183
2020118
2019130
2018125