Improving hierarchical product classification using domain-specific language modelling


Brinkmann, Alexander ; Bizer, Christian



URL: http://sites.computer.org/debull/A21june/p14.pdf
Additional URL: http://sites.computer.org/debull/A21june/issue1.ht...
Document Type: Article
Year of publication: 2021
The title of a journal, publication series: Bulletin of the Technical Committee on Data Engineering / IEEE Computer Society
Volume: 44
Issue number: 2
Page range: 14-25
Place of publication: Washington, DC
Publishing house: IEEE Computer Soc.
Publication language: English
Institution: School of Business Informatics and Mathematics > Information Systems V: Web-based Systems (Bizer 2012-)
Subject: 004 Computer science, internet
Keywords (English): e-commerce , product categorization , deep learning , language modelling
Abstract: In order to deliver a coherent user experience, product aggregators such as market places or price portals integrate product offers from many web shops into a single product categorization hierarchy. Recently, transformer models have shown remarkable performance on various NLP tasks. These models are pre-trained on huge cross-domain text corpora using self-supervised learning and fine-tuned afterwards for specific downstream tasks. Research from other application domains indicates that additional selfsupervised pre-training using domain-specific text corpora can further increase downstream performance without requiring additional task-specific training data. In this paper, we first show that transformers outperform a more traditional fastText-based classification technique on the task of assigning product offers from different web shops into a product hierarchy. Afterwards, we investigate whether it is possible to improve the performance of the transformer models by performing additional self-supervised pretraining using different corpora of product offers, which were extracted from the Common Crawl. Our experiments show that by using large numbers of related product offers for masked language modelling, it is possible to increase the performance of the transformer models by 1.22% in wF1 and 1.36% in hF1 reaching a performance of nearly 89% wF1.




Dieser Eintrag ist Teil der Universitätsbibliographie.




Metadata export


Citation


+ Search Authors in

+ Page Views

Hits per month over past year

Detailed information



You have found an error? Please let us know about your desired correction here: E-Mail


Actions (login required)

Show item Show item