Intermediate training of BERT for product matching


Peeters, Ralph ; Bizer, Christian ; Glavaš, Goran



URL: http://ceur-ws.org/Vol-2726/paper1.pdf
Additional URL: http://ceur-ws.org/Vol-2726/
Document Type: Conference or workshop publication
Year of publication: 2020
Book title: DI2KG 2020 : Proceedings of the 2nd International Workshop on Challenges and Experiences from Data Integration to Knowledge Graphs co-located with 46th International Conference on Very Large Data Bases (VLDB 2020), Tokyo, Japan, August 31, 2020
The title of a journal, publication series: CEUR Workshop Proceedings
Volume: 2726
Page range: 1-2
Conference title: DI2KG 2020
Location of the conference venue: Online
Date of the conference: 31.08.2020
Publisher: Piai, Federico
Place of publication: Aachen, Germany
Publishing house: RWTH Aachen
ISSN: 1613-0073
Publication language: English
Institution: School of Business Informatics and Mathematics > Wirtschaftsinformatik V (Bizer)
School of Business Informatics and Mathematics > Text Analytics for Interdisciplinary Research (Juniorprofessur) (Glavaš 2017-2021)
Subject: 004 Computer science, internet
Keywords (English): e-commerce , product matching , deep learning
Abstract: Transformer-based models like BERT have pushed the state-of the-art for a wide range of tasks in natural language processing. General-purpose pre-training on large corpora allows Transformers to yield good performance even with small amounts of training data for task-specific fine-tuning. In this work, we apply BERT to the task of product matching in e-commerce and show that BERT is much more training data efficient than other state-of-the-art methods. Moreover, we show that we can further boost its effectiveness through an intermediate training step, exploiting large collections of product offers. Our intermediate training leads to strong performance (>90% F1) on new, unseen products without any product-specific fine-tuning. Further fine-tuning yields additional gains, resulting in improvements of up to 12% F1 for small training sets. Adding the masked language modeling objective in the intermediate training step in order to further adapt the language model to the application domain leads to an additional increase of up to 3% F1.

Dieser Eintrag ist Teil der Universitätsbibliographie.




Metadata export


Citation


+ Search Authors in

+ Page Views

Hits per month over past year

Detailed information



You have found an error? Please let us know about your desired correction here: E-Mail


Actions (login required)

Show item Show item