Command Palette
Search for a command to run...
Sarkar Snigdha Sarathi Das Arzoo Katiyar Rebecca J. Passonneau Rui Zhang

Abstract
Named Entity Recognition (NER) in Few-Shot setting is imperative for entity tagging in low resource domains. Existing approaches only learn class-specific semantic features and intermediate representations from source domains. This affects generalizability to unseen target domains, resulting in suboptimal performances. To this end, we present CONTaiNER, a novel contrastive learning technique that optimizes the inter-token distribution distance for Few-Shot NER. Instead of optimizing class-specific attributes, CONTaiNER optimizes a generalized objective of differentiating between token categories based on their Gaussian-distributed embeddings. This effectively alleviates overfitting issues originating from training domains. Our experiments in several traditional test domains (OntoNotes, CoNLL'03, WNUT '17, GUM) and a new large scale Few-Shot NER dataset (Few-NERD) demonstrate that on average, CONTaiNER outperforms previous methods by 3%-13% absolute F1 points while showing consistent performance trends, even in challenging scenarios where previous approaches could not achieve appreciable performance.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| few-shot-ner-on-few-nerd-inter | CONTaiNER | 10 way 1~2 shot: 48.35 10 way 5~10 shot: 57.12 5 way 1~2 shot: 55.95 5 way 5~10 shot: 61.83 |
| few-shot-ner-on-few-nerd-intra | CONTaiNER | 10 way 1~2 shot: 33.84 10 way 5~10 shot: 47.49 5 way 1~2 shot: 40.43 5 way 5~10 shot: 53.70 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.