2022
DOI: 10.14778/3565816.3565836
|View full text |Cite
|
Sign up to set email alerts
|

PromptEM

Abstract: Entity Matching (EM), which aims to identify whether two entity records from two relational tables refer to the same real-world entity, is one of the fundamental problems in data management. Traditional EM assumes that two tables are homogeneous with the aligned schema, while it is common that entity records of different formats (e.g., relational, semi-structured, or textual types) involve in practical scenarios. It is not practical to unify their schemas due to the different formats. To support EM on format-d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
0
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 14 publications
(2 citation statements)
references
References 27 publications
0
0
0
Order By: Relevance
“…We study the scheduling and coordination of the individual ER algorithms, in order to resolve the multiple datasets, and show the scalability of our approach….." The conflict between efficiency and effectiveness: The entity resolution models based on PLMs can be divided into two categories regarding representation learning: independent or interdependent representation [9]. Interdependent representation models [2,3,10,11] have a deep interaction between pairs of records through attention mechanisms, resulting in better matching quality. Despite being effective, interdependent representation models come with a poor scalability for the quadratic searching space of record pairs, thus need additional blocking steps.…”
Section: Joint Entity Resolution On Multiple Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…We study the scheduling and coordination of the individual ER algorithms, in order to resolve the multiple datasets, and show the scalability of our approach….." The conflict between efficiency and effectiveness: The entity resolution models based on PLMs can be divided into two categories regarding representation learning: independent or interdependent representation [9]. Interdependent representation models [2,3,10,11] have a deep interaction between pairs of records through attention mechanisms, resulting in better matching quality. Despite being effective, interdependent representation models come with a poor scalability for the quadratic searching space of record pairs, thus need additional blocking steps.…”
Section: Joint Entity Resolution On Multiple Datasetsmentioning
confidence: 99%
“…By constructing prompt templates, downstream tasks can be transformed into fill-in-the-blank forms of upstream tasks, which can more effectively utilize the original network structure of PLM and the prior knowledge obtained from pretraining. PromptEM [10] is the first work that applies prompt tuning for entity resolution tasks and performs well under low-resource and sufficient resource settings. Recently, some work also adopted the prompt-based method to eliminate embedding bias in PLMs [24], which can also be used to improve the embedding quality of records in entity resolution tasks.…”
Section: Related Workmentioning
confidence: 99%