Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
Adversarial Attacks to Latent Representations of Distributed Neural Networks in Split Computing
by
Abdi, Mohammad
, Restuccia, Francesco
, Ashdown, Jonathan
, Zhang, Milin
in
Artificial neural networks
/ Edge computing
/ Information theory
/ Neural networks
/ Robustness
2025
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
Adversarial Attacks to Latent Representations of Distributed Neural Networks in Split Computing
by
Abdi, Mohammad
, Restuccia, Francesco
, Ashdown, Jonathan
, Zhang, Milin
in
Artificial neural networks
/ Edge computing
/ Information theory
/ Neural networks
/ Robustness
2025
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
Adversarial Attacks to Latent Representations of Distributed Neural Networks in Split Computing
Paper
Adversarial Attacks to Latent Representations of Distributed Neural Networks in Split Computing
2025
Request Book From Autostore
and Choose the Collection Method
Overview
Distributed deep neural networks (DNNs) have been shown to reduce the computational burden of mobile devices and decrease the end-to-end inference latency in edge computing scenarios. While distributed DNNs have been studied, to the best of our knowledge, the resilience of distributed DNNs to adversarial action remains an open problem. In this paper, we fill the existing research gap by rigorously analyzing the robustness of distributed DNNs against adversarial action. We cast this problem in the context of information theory and rigorously proved that (i) the compressed latent dimension improves the robustness but also affect task-oriented performance; and (ii) the deeper splitting point enhances the robustness but also increases the computational burden. These two trade-offs provide a novel perspective to design robust distributed DNN. To test our theoretical findings, we perform extensive experimental analysis by considering 6 different DNN architectures, 6 different approaches for distributed DNN and 10 different adversarial attacks using the ImageNet-1K dataset.
Publisher
Cornell University Library, arXiv.org
This website uses cookies to ensure you get the best experience on our website.