Command Palette
Search for a command to run...
Tom Hosking Hao Tang Mirella Lapata

Abstract
We propose a generative model of paraphrase generation, that encourages syntactic diversity by conditioning on an explicit syntactic sketch. We introduce Hierarchical Refinement Quantized Variational Autoencoders (HRQ-VAE), a method for learning decompositions of dense encodings as a sequence of discrete latent variables that make iterative refinements of increasing granularity. This hierarchy of codes is learned through end-to-end training, and represents fine-to-coarse grained information about the input. We use HRQ-VAE to encode the syntactic form of an input sentence as a path through the hierarchy, allowing us to more easily predict syntactic sketches at test time. Extensive experiments, including a human evaluation, confirm that HRQ-VAE learns a hierarchical representation of the input space, and generates paraphrases of higher quality than previous systems.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| paraphrase-generation-on-mscoco | HRQ-VAE | BLEU: 27.90 iBLEU: 19.04 |
| paraphrase-generation-on-paralex | HRQ-VAE | BLEU: 39.49 iBLEU: 24.93 |
| paraphrase-generation-on-quora-question-pairs-1 | HRQ-VAE | BLEU: 33.11 iBLEU: 18.42 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.