Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
EAN: Event Adaptive Network for Enhanced Action Recognition
by
Guo, Guodong
, Tian, Yuan
, Yan, Yichao
, Gao, Zhiyong
, Zhai, Guangtao
in
Activity recognition
/ Design
/ Interaction models
/ Modelling
/ Modules
/ Sensors
/ Video
2022
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
EAN: Event Adaptive Network for Enhanced Action Recognition
by
Guo, Guodong
, Tian, Yuan
, Yan, Yichao
, Gao, Zhiyong
, Zhai, Guangtao
in
Activity recognition
/ Design
/ Interaction models
/ Modelling
/ Modules
/ Sensors
/ Video
2022
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
EAN: Event Adaptive Network for Enhanced Action Recognition
Journal Article
EAN: Event Adaptive Network for Enhanced Action Recognition
2022
Request Book From Autostore
and Choose the Collection Method
Overview
Efficiently modeling spatial–temporal information in videos is crucial for action recognition. To achieve this goal, state-of-the-art methods typically employ the convolution operator and the dense interaction modules such as non-local blocks. However, these methods cannot accurately fit the diverse events in videos. On the one hand, the adopted convolutions are with fixed scales, thus struggling with events of various scales. On the other hand, the dense interaction modeling paradigm only achieves sub-optimal performance as action-irrelevant parts bring additional noises for the final prediction. In this paper, we propose a unified action recognition framework to investigate the dynamic nature of video content by introducing the following designs. First, when extracting local cues, we generate the spatial–temporal kernels of dynamic-scale to adaptively fit the diverse events. Second, to accurately aggregate these cues into a global video representation, we propose to mine the interactions only among a few selected foreground objects by a Transformer, which yields a sparse paradigm. We call the proposed framework as Event Adaptive Network because both key designs are adaptive to the input video content. To exploit the short-term motions within local segments, we propose a novel and efficient Latent Motion Code module, further improving the performance of the framework. Extensive experiments on several large-scale video datasets, e.g., Something-to-Something V1 &V2, Kinetics, and Diving48, verify that our models achieve state-of-the-art or competitive performances at low FLOPs. Codes are available at: https://github.com/tianyuan168326/EAN-Pytorch.
Publisher
Springer Nature B.V
Subject
This website uses cookies to ensure you get the best experience on our website.