TY - GEN
T1 - CopycHats
T2 - 2024 Workshop on Human-In-the-Loop Data Analytics, HILDA 2024, Co-located with SIGMOD 2024
AU - Solomon, Matan
AU - Genossar, Bar
AU - Gal, Avigdor
N1 - Publisher Copyright:
© 2024 Copyright held by the owner/author(s).
PY - 2024/6/14
Y1 - 2024/6/14
N2 - Schema Matching, the task of finding correspondences among attributes of different schemata, plays an important role in data integration. The task has been extensively researched, leading to the development of multiple algorithmic approaches, many of which incorporate humans to some extent, by performing the matching, validating algorithmic solutions, or generating reliable ground truth for algorithms to be trained against. Human matching is a temporal process, in which previous decisions and ongoing cognitive processes influence future behavior. Therefore, planning a human matching task necessitates careful consideration, for example, the ordering of questions posed to the matcher. Various strategies exist for optimally choosing the sequence of matching questions and evaluating them may be limited by notable inherent human limitations. In this work, we propose to leverage Large Language Models (LLMs) to create artificial agents that emulate human agents in order to evaluate question sequencing strategies. We offer an alternative to traditional human-based evaluations, which overcome those limitations. We test our suggested evaluation framework and discuss the similarities and differences between artificial and human agents in the context of schema matching evaluation.
AB - Schema Matching, the task of finding correspondences among attributes of different schemata, plays an important role in data integration. The task has been extensively researched, leading to the development of multiple algorithmic approaches, many of which incorporate humans to some extent, by performing the matching, validating algorithmic solutions, or generating reliable ground truth for algorithms to be trained against. Human matching is a temporal process, in which previous decisions and ongoing cognitive processes influence future behavior. Therefore, planning a human matching task necessitates careful consideration, for example, the ordering of questions posed to the matcher. Various strategies exist for optimally choosing the sequence of matching questions and evaluating them may be limited by notable inherent human limitations. In this work, we propose to leverage Large Language Models (LLMs) to create artificial agents that emulate human agents in order to evaluate question sequencing strategies. We offer an alternative to traditional human-based evaluations, which overcome those limitations. We test our suggested evaluation framework and discuss the similarities and differences between artificial and human agents in the context of schema matching evaluation.
UR - http://www.scopus.com/inward/record.url?scp=85198224672&partnerID=8YFLogxK
U2 - 10.1145/3665939.3665963
DO - 10.1145/3665939.3665963
M3 - ???researchoutput.researchoutputtypes.contributiontobookanthology.conference???
AN - SCOPUS:85198224672
T3 - HILDA 2024 - Workshop on Human-In-the-Loop Data Analytics Co-located with SIGMOD 2024
BT - HILDA 2024 - Workshop on Human-In-the-Loop Data Analytics Co-located with SIGMOD 2024
Y2 - 14 June 2024
ER -