Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
Image caption generation using Visual Attention Prediction and Contextual Spatial Relation Extraction
by
Kumaraswamy, Suresh
, Sasibhooshan, Reshmi
, Sasidharan, Santhoshkumar
in
Artificial neural networks
/ Attention
/ Big Data
/ Cider
/ Coders
/ Communication
/ Datasets
/ Encoders-Decoders
/ Engineering
/ Experiments
/ Extraction
/ Feature extraction
/ Feature maps
/ Image retrieval
/ Natural language
/ Networks
/ Neural networks
/ Predictions
/ Semantics
/ Short term memory
/ Spatial attention
/ Visual attention
/ Wavelet transforms
/ Work
2023
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Image caption generation using Visual Attention Prediction and Contextual Spatial Relation Extraction
by
Kumaraswamy, Suresh
, Sasibhooshan, Reshmi
, Sasidharan, Santhoshkumar
in
Artificial neural networks
/ Attention
/ Big Data
/ Cider
/ Coders
/ Communication
/ Datasets
/ Encoders-Decoders
/ Engineering
/ Experiments
/ Extraction
/ Feature extraction
/ Feature maps
/ Image retrieval
/ Natural language
/ Networks
/ Neural networks
/ Predictions
/ Semantics
/ Short term memory
/ Spatial attention
/ Visual attention
/ Wavelet transforms
/ Work
2023
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
Image caption generation using Visual Attention Prediction and Contextual Spatial Relation Extraction
by
Kumaraswamy, Suresh
, Sasibhooshan, Reshmi
, Sasidharan, Santhoshkumar
in
Artificial neural networks
/ Attention
/ Big Data
/ Cider
/ Coders
/ Communication
/ Datasets
/ Encoders-Decoders
/ Engineering
/ Experiments
/ Extraction
/ Feature extraction
/ Feature maps
/ Image retrieval
/ Natural language
/ Networks
/ Neural networks
/ Predictions
/ Semantics
/ Short term memory
/ Spatial attention
/ Visual attention
/ Wavelet transforms
/ Work
2023
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
Image caption generation using Visual Attention Prediction and Contextual Spatial Relation Extraction
Journal Article
Image caption generation using Visual Attention Prediction and Contextual Spatial Relation Extraction
2023
Request Book From Autostore
and Choose the Collection Method
Overview
Automatic caption generation with attention mechanisms aims at generating more descriptive captions containing coarser to finer semantic contents in the image. In this work, we use an encoder-decoder framework employing Wavelet transform based Convolutional Neural Network (WCNN) with two level discrete wavelet decomposition for extracting the visual feature maps highlighting the spatial, spectral and semantic details from the image. The Visual Attention Prediction Network (VAPN) computes both channel and spatial attention for obtaining visually attentive features. In addition to these, local features are also taken into account by considering the contextual spatial relationship between the different objects. The probability of the appropriate word prediction is achieved by combining the aforementioned architecture with Long Short Term Memory (LSTM) decoder network. Experiments are conducted on three benchmark datasets—Flickr8K, Flickr30K and MSCOCO datasets and the evaluation results prove the improved performance of the proposed model with CIDEr score of 124.2.
This website uses cookies to ensure you get the best experience on our website.