Command Palette
Search for a command to run...
Sheng Zhang; Xutai Ma; Kevin Duh; Benjamin Van Durme

Abstract
We unify different broad-coverage semantic parsing tasks under a transduction paradigm, and propose an attention-based neural framework that incrementally builds a meaning representation via a sequence of semantic relations. By leveraging multiple attention mechanisms, the transducer can be effectively trained without relying on a pre-trained aligner. Experiments conducted on three separate broad-coverage semantic parsing tasks -- AMR, SDP and UCCA -- demonstrate that our attention-based neural transducer improves the state of the art on both AMR and UCCA, and is competitive with the state of the art on SDP.
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| amr-parsing-on-ldc2014t12-1 | Broad-Coverage Semantic Parsing as Transduction | F1 Full: 71.3 |
| amr-parsing-on-ldc2017t10 | Zhang et al. | Smatch: 77.0 |
| ucca-parsing-on-semeval-2019-task-1 | Neural Transducer | English-Wiki (open) F1: 76.6 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.