HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

Baseline Needs More Love: On Simple Word-Embedding-Based Models and Associated Pooling Mechanisms

Dinghan Shen; Guoyin Wang; Wenlin Wang; Martin Renqiang Min; Qinliang Su; Yizhe Zhang; Chunyuan Li; Ricardo Henao; Lawrence Carin

Baseline Needs More Love: On Simple Word-Embedding-Based Models and Associated Pooling Mechanisms

Abstract

Many deep learning architectures have been proposed to model the compositionality in text sequences, requiring a substantial number of parameters and expensive computations. However, there has not been a rigorous evaluation regarding the added value of sophisticated compositional functions. In this paper, we conduct a point-by-point comparative study between Simple Word-Embedding-based Models (SWEMs), consisting of parameter-free pooling operations, relative to word-embedding-based RNN/CNN models. Surprisingly, SWEMs exhibit comparable or even superior performance in the majority of cases considered. Based upon this understanding, we propose two additional pooling strategies over learned word embeddings: (i) a max-pooling operation for improved interpretability; and (ii) a hierarchical pooling operation, which preserves spatial (n-gram) information within text sequences. We present experiments on 17 datasets encompassing three tasks: (i) (long) document classification; (ii) text sequence matching; and (iii) short text tasks, including classification and tagging. The source code and datasets can be obtained from https:// github.com/dinghanshen/SWEM.

Code Repositories

nyk510/scdv-python
Mentioned in GitHub
dinghanshen/SWEM
Official
tf
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
named-entity-recognition-ner-on-conll-2003SWEM-CRF
F1: 86.28
named-entity-recognition-on-conll-2000SWEM-CRF
F1: 90.34
natural-language-inference-on-multinliSWEM-max
Matched: 68.2
Mismatched: 67.7
natural-language-inference-on-snliSWEM-max
% Test Accuracy: 83.8
paraphrase-identification-on-msrpSWEM-concat
Accuracy: 71.5
F1: 81.3
question-answering-on-quora-question-pairsSWEM-concat
Accuracy: 83.03%
question-answering-on-wikiqaSWEM-concat
MAP: 0.6788
MRR: 0.6908
sentiment-analysis-on-mrSWEM-concat
Accuracy: 78.2
sentiment-analysis-on-sst-2-binarySWEM-concat
Accuracy: 84.3
sentiment-analysis-on-sst-5-fine-grainedSWEM-concat
Accuracy: 46.1
sentiment-analysis-on-yelp-binarySWEM-hier
Error: 4.19
sentiment-analysis-on-yelp-fine-grainedSWEM-hier
Error: 36.21
subjectivity-analysis-on-subjSWEM-concat
Accuracy: 93
text-classification-on-ag-newsSWEM-concat
Error: 7.34
text-classification-on-dbpediaSWEM-concat
Error: 1.43
text-classification-on-trec-6SWEM-aver
Error: 7.8
text-classification-on-yahoo-answersSWEM-concat
Accuracy: 73.53

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp