2017
DOI: 10.1016/j.procs.2017.06.038
|View full text |Cite
|
Sign up to set email alerts
|

Plagiarism detection using document similarity based on distributed representation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(6 citation statements)
references
References 5 publications
0
6
0
Order By: Relevance
“…Longest common subsequence (LCS) method: consists of finding the longest subsequence common to all sequences in a set of sequences. The longest common subsequence problem is a classic computer science problem, the basis of data comparison programs such as the diff utility and has applications in computational linguistics and bioinformatics [24]. Word Mover's Distance (WMD): uses word embeddings to calculate the similarities, and precisely, it uses normalized Bag-of-words and word Embeddings to calculate the distance between documents [25].…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Longest common subsequence (LCS) method: consists of finding the longest subsequence common to all sequences in a set of sequences. The longest common subsequence problem is a classic computer science problem, the basis of data comparison programs such as the diff utility and has applications in computational linguistics and bioinformatics [24]. Word Mover's Distance (WMD): uses word embeddings to calculate the similarities, and precisely, it uses normalized Bag-of-words and word Embeddings to calculate the distance between documents [25].…”
Section: Methodsmentioning
confidence: 99%
“…The similarity between vectors was computed by using cosine similarity. [24] The aim of this approach is evaluating the validity of using the distributed representation to define the word similarity. They introduce three methods based on the following three document similarities: for two documents: The length of the longest common subsequence (LCS) divided by the length of the shorter document, the local maximal value of the length of LCS, and the local maximal value of the weighted length of LCS.…”
Section: Related Workmentioning
confidence: 99%
“…The LCS algorithm has several benefits like [3] such more scalable, reduce computing power as well as checking the grammar. In the realm of image plagiarism detection [4] numerous methodologies have been proposed, including the scale-invariant feature transform (SIFT) algorithm. This algorithm aims to analyze and extract the primary elements of two images to ascertain their similarity.…”
Section: Literature Surveymentioning
confidence: 99%
“…In December 2016 [15], Plagiarism detection method is being proposed by the author, which lies on the principle of local maximal value of the longest common subsequence (LCS) by its length and weight. They introduce three methods based on the following three document similarities: for two documents, • the length of LCS divided by the length of the shorter document, • the local maximal value of the length of LCS, and • the local maximal value of the weighted length of LCS.…”
Section: Related Workmentioning
confidence: 99%
“…Here in table 1 below is the complete comparison of available work for the plagiarism detection taken from different references such as [15], [16], [17], [18], [19],20], [21]. Table 1.…”
Section: Related Workmentioning
confidence: 99%