Command Palette
Search for a command to run...
Sai Surya; Abhijit Mishra; Anirban Laha; Parag Jain; Karthik Sankaranarayanan

Abstract
The paper presents a first attempt towards unsupervised neural text simplification that relies only on unlabeled text corpora. The core framework is composed of a shared encoder and a pair of attentional-decoders and gains knowledge of simplification through discrimination based-losses and denoising. The framework is trained using unlabeled text collected from en-Wikipedia dump. Our analysis (both quantitative and qualitative involving human evaluators) on a public test data shows that the proposed model can perform text-simplification at both lexical and syntactic levels, competitive to existing supervised methods. Addition of a few labelled pairs also improves the performance further.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| text-simplification-on-asset | UNTS (Unsupervised) | BLEU: 76.14* SARI (EASSEu003e=0.2.1): 35.19 |
| text-simplification-on-turkcorpus | UNMT (Unsupervised) | BLEU: 74.02 SARI (EASSEu003e=0.2.1): 37.20 |
| text-simplification-on-turkcorpus | UNTS-10k (Weakly supervised) | SARI (EASSEu003e=0.2.1): 37.15 |
| text-simplification-on-turkcorpus | UNTS (Unsupervised) | SARI (EASSEu003e=0.2.1): 36.29 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.