HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

TASK AWARE MULTI-TASK LEARNING FOR SPEECH TO TEXT TASKS

{Inchul Hwang Chanwoo Kim Sangha Kim Seokchan Ahn Hyojung Han Beomseok Lee Nikhil Kumar Lakumarapu Mohd Abbas Zaidi Sathish Indurthi}

Abstract

In general, the direct Speech-to-text translation (ST) is jointly trained with Automatic Speech Recognition (ASR), and Machine Translation (MT) tasks. However, the issues with the current joint learning strategies inhibit the knowledge transfer across these tasks. We propose a task modulation network which allows the model to learn task specific features, while learning the shared features simultaneously. This proposed approach removes the need for separate finetuning step resulting in a single model which performs all these tasks. This single model achieves a performance of 28.64 BLEU score on ST MuST-C English-German, WER of 11.61% on ASR TEDLium v3, 23.35 BLEU score on MT WMT’15 English-German task. This sets a new state-of-the-art performance (SOTA) on the ST task while outperforming the existing end-to-end ASR systems.

Benchmarks

BenchmarkMethodologyMetrics
speech-to-text-translation-on-must-c-en-deTask Modulation + Multitask Learning(ASR/MT) + Data Augmentation
Case-sensitive sacreBLEU: 28.88

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
TASK AWARE MULTI-TASK LEARNING FOR SPEECH TO TEXT TASKS | Papers | HyperAI