HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Intermediate Training of BERT for Product Matching

{Goran Glavas Christian Bizer Ralph Peeters}

Intermediate Training of BERT for Product Matching

Abstract

Transformer-based models like BERT have pushed the state-of the-art for a wide range of tasks in natural language processing. General-purpose pre-training on large corpora allows Transformers to yield good performance even with small amounts of training data for task-specific fine-tuning. In this work, we apply BERT to the task of product matching in e-commerce and show that BERTis much more training data efficient than other state-of-the-art methods. Moreover, we show that we can further boost its effectiveness through an intermediate training step, exploiting large collections of product offers. Our intermediate training leads to strong performance (>90% F1) on new, unseen products without any product-specific fine-tuning. Further fine-tuning yields additional gains, resulting in improvements of up to 12% F1 for small training sets. Adding the masked language modeling objective in the intermediate training step in order to further adapt the language model to the application domain leads to an additional increase of up to 3% F1.

Benchmarks

BenchmarkMethodologyMetrics
entity-resolution-on-wdc-computers-smallBERT
F1 (%): 96.53
entity-resolution-on-wdc-computers-xlargeBERT
F1 (%): 97.37

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Intermediate Training of BERT for Product Matching | Papers | HyperAI