Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
Modality Confidence Aware Training for Robust End-to-End Spoken Language Understanding
by
Seltzer, Michael L
, Lin, Ju
, Kalinli, Ozlem
, Shrivastava, Akshat
, Le, Duc
, Kim, Suyoun
in
Errors
/ Hypotheses
/ Representations
/ Speech recognition
2023
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
Modality Confidence Aware Training for Robust End-to-End Spoken Language Understanding
by
Seltzer, Michael L
, Lin, Ju
, Kalinli, Ozlem
, Shrivastava, Akshat
, Le, Duc
, Kim, Suyoun
in
Errors
/ Hypotheses
/ Representations
/ Speech recognition
2023
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
Modality Confidence Aware Training for Robust End-to-End Spoken Language Understanding
Paper
Modality Confidence Aware Training for Robust End-to-End Spoken Language Understanding
2023
Request Book From Autostore
and Choose the Collection Method
Overview
End-to-end (E2E) spoken language understanding (SLU) systems that generate a semantic parse from speech have become more promising recently. This approach uses a single model that utilizes audio and text representations from pre-trained speech recognition models (ASR), and outperforms traditional pipeline SLU systems in on-device streaming scenarios. However, E2E SLU systems still show weakness when text representation quality is low due to ASR transcription errors. To overcome this issue, we propose a novel E2E SLU system that enhances robustness to ASR errors by fusing audio and text representations based on the estimated modality confidence of ASR hypotheses. We introduce two novel techniques: 1) an effective method to encode the quality of ASR hypotheses and 2) an effective approach to integrate them into E2E SLU models. We show accuracy improvements on STOP dataset and share the analysis to demonstrate the effectiveness of our approach.
Publisher
Cornell University Library, arXiv.org
Subject
This website uses cookies to ensure you get the best experience on our website.