Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
Low-Resourced Alphabet-Level Pivot-Based Neural Machine Translation for Translating Korean Dialects
by
Park, Seong-Bae
, Park, Junho
in
alpha-level tokenization
/ Alphabets
/ Analysis
/ Chinese languages
/ dialect translation
/ Dialects
/ Experiments
/ Foreign languages
/ Interpreters
/ Japanese language
/ Korean language
/ lack of parallel corpus
/ Language
/ Large language models
/ Machine translation
/ Morphology
/ neural machine translation
/ Normalization
/ Parallel corpora
/ Phonetics
/ Sentences
/ sequence-to-sequence model
/ Sequences
/ Speech
/ Standard dialects
/ Translating and interpreting
/ Translation
/ Translation methods and strategies
2025
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Low-Resourced Alphabet-Level Pivot-Based Neural Machine Translation for Translating Korean Dialects
by
Park, Seong-Bae
, Park, Junho
in
alpha-level tokenization
/ Alphabets
/ Analysis
/ Chinese languages
/ dialect translation
/ Dialects
/ Experiments
/ Foreign languages
/ Interpreters
/ Japanese language
/ Korean language
/ lack of parallel corpus
/ Language
/ Large language models
/ Machine translation
/ Morphology
/ neural machine translation
/ Normalization
/ Parallel corpora
/ Phonetics
/ Sentences
/ sequence-to-sequence model
/ Sequences
/ Speech
/ Standard dialects
/ Translating and interpreting
/ Translation
/ Translation methods and strategies
2025
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
Low-Resourced Alphabet-Level Pivot-Based Neural Machine Translation for Translating Korean Dialects
by
Park, Seong-Bae
, Park, Junho
in
alpha-level tokenization
/ Alphabets
/ Analysis
/ Chinese languages
/ dialect translation
/ Dialects
/ Experiments
/ Foreign languages
/ Interpreters
/ Japanese language
/ Korean language
/ lack of parallel corpus
/ Language
/ Large language models
/ Machine translation
/ Morphology
/ neural machine translation
/ Normalization
/ Parallel corpora
/ Phonetics
/ Sentences
/ sequence-to-sequence model
/ Sequences
/ Speech
/ Standard dialects
/ Translating and interpreting
/ Translation
/ Translation methods and strategies
2025
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
Low-Resourced Alphabet-Level Pivot-Based Neural Machine Translation for Translating Korean Dialects
Journal Article
Low-Resourced Alphabet-Level Pivot-Based Neural Machine Translation for Translating Korean Dialects
2025
Request Book From Autostore
and Choose the Collection Method
Overview
Developing a machine translator from a Korean dialect to a foreign language presents significant challenges due to a lack of a parallel corpus for direct dialect translation. To solve this issue, this paper proposes a pivot-based machine translation model that consists of two sub-translators. The first sub-translator is a sequence-to-sequence model with minGRU as an encoder and GRU as a decoder. It normalizes a dialect sentence into a standard sentence, and it employs alphabet-level tokenization. The other type of sub-translator is a legacy translator, such as off-the-shelf neural machine translators or LLMs, which translates the normalized standard sentence to a foreign sentence. The effectiveness of the alphabet-level tokenization and the minGRU encoder for the normalization model is demonstrated through empirical analysis. Alphabet-level tokenization is proven to be more effective for Korean dialect normalization than other widely used sub-word tokenizations. The minGRU encoder exhibits comparable performance to GRU as an encoder, and it is faster and more effective in managing longer token sequences. The pivot-based translation method is also validated through a broad range of experiments, and its effectiveness in translating Korean dialects to English, Chinese, and Japanese is demonstrated empirically.
Publisher
MDPI AG
Subject
MBRLCatalogueRelatedBooks
Related Items
Related Items
This website uses cookies to ensure you get the best experience on our website.