Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
A Unified Generalization Framework for Model Merging: Trade-offs, Non-Linearity, and Scaling Laws
by
Tang, Anke
, Zhang, Miao
, Yin, Quanjun
, Li, Qinglun
, Wang, Mengzhu
, Shen, Li
in
Algorithms
/ Optimization
/ Visual tasks
2026
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
A Unified Generalization Framework for Model Merging: Trade-offs, Non-Linearity, and Scaling Laws
by
Tang, Anke
, Zhang, Miao
, Yin, Quanjun
, Li, Qinglun
, Wang, Mengzhu
, Shen, Li
in
Algorithms
/ Optimization
/ Visual tasks
2026
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
A Unified Generalization Framework for Model Merging: Trade-offs, Non-Linearity, and Scaling Laws
Paper
A Unified Generalization Framework for Model Merging: Trade-offs, Non-Linearity, and Scaling Laws
2026
Request Book From Autostore
and Choose the Collection Method
Overview
Model merging efficiently aggregates capabilities from multiple fine-tuned models into a single one, operating purely in parameter space without original data or expensive re-computation. Despite empirical successes, a unified theory for its effectiveness under heterogeneous finetuning hyperparameters (e.g., varying learning rates, batch sizes) remains missing. Existing federated learning theories focus purely on optimization, which fails to explain model merging and inherently leads to theoretical paradoxes. To address this challenge, we pioneer the integration of \\(L_2\\)-Stability theory into heterogeneous environments to rigorously decouple the excess risk of the merged model \\(\\boldsymbol{x}_{avg}\\) into optimization and generalization errors. This comprehensive analysis yields three main contributions: (i) We mathematically establish the fundamental \\textit{Optimization-Generalization Trade-off}, explicitly resolving the paradox of why over-trained experts lead to catastrophic merging collapse. (ii) \\textit{A unified theoretical framework} is provided to explain not only linear merging algorithms (e.g., TA, AdaMerging) but also state-of-the-art \\textit{non-linear} merging algorithms (e.g., TIES, DARE), proving how sparsification operators strictly tighten the generalization bound by suppressing task heterogeneity. (iii) Rather than heuristic guidelines, we derive \\textit{Quantitative Scaling Laws} that theoretically predict the precise impact of hyperparameter choices, enabling practitioners to strategically construct ``merge-friendly'' experts. Extensive experiments on the ResNet and ViT architectures across 20 visual classification tasks, involving thousands of finetuning models, robustly confirm that our theoretical scaling laws accurately predict the empirical generalization behaviors of \\(\\boldsymbol{x}_{avg}\\).
Publisher
Cornell University Library, arXiv.org
Subject
MBRLCatalogueRelatedBooks
Related Items
Related Items
This website uses cookies to ensure you get the best experience on our website.