Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
Unified Odd‐Descent Regularization for Input Optimization
by
Lai, Zhao‐Rong
, Zhou, Zheng‐Sen
2025
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Do you wish to request the book?
Unified Odd‐Descent Regularization for Input Optimization
by
Lai, Zhao‐Rong
, Zhou, Zheng‐Sen
2025
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
Journal Article
Unified Odd‐Descent Regularization for Input Optimization
2025
Request Book From Autostore
and Choose the Collection Method
Overview
Activation‐descent regularization is a crucial approach in input optimization for ReLU networks, but traditional methods face challenges. Converting discrete activation patterns into differentiable forms introduces half‐space division, high computational complexity, and instability. We propose a novel local descent regularization method based on a network of arbitrary odd functions, which unifies half‐space processing, simplifies expression, reduces computational complexity, and enriches the expression of the activation descent regularization term. Furthermore, by selecting an arbitrary differentiable odd function, we can derive an exact gradient descent direction, solving the non‐differentiability problem caused by the non‐smooth nature of ReLU, thus improving optimization efficiency and convergence stability. Experiments demonstrate the competitive performance of our approach, particularly in adversarial learning applications. This work contributes to both theory and practice of regularization for input optimization.
MBRLCatalogueRelatedBooks
Related Items
Related Items
We currently cannot retrieve any items related to this title. Kindly check back at a later time.
This website uses cookies to ensure you get the best experience on our website.