Information-geometric perspectives on merging variational foundation models

Jamoussi, Nour; Serra, Giuseppe; Stavrou, Photios A.; Kountouris, Marios
NeurIPS 2025, 39th Conference on Neural Information Processing Systems, CCFM Workshop on Continual and Compatible Foundation Model Updates, 6 December 2025, San Diego, CA, USA

We propose an information-geometric framework for merging variational foundation models that preserves global robustness while integrating domain-specific knowledge in a principled manner. Assuming that the foundation models have been pretrained or fine-tuned using the Improved Variational Online Newton (IVON) optimizer, matching Adam’s computational cost while providing Bayesian advantages, we formulate the merging problem between the pretrained and fine-tuned models as an information-geometric projection. Under mild assumptions, this reduces to computing a barycenter in the variational parameter space, yielding a computationally efficient and theoretically grounded merging rule. The framework naturally extends to multi-model barycentric merging, minimizing the average discrepancy among fine-tuned models.


Type:
Conférence
City:
San Diego
Date:
2025-12-06
Department:
Systèmes de Communication
Eurecom Ref:
8503
Copyright:
© NIST. Personal use of this material is permitted. The definitive version of this paper was published in NeurIPS 2025, 39th Conference on Neural Information Processing Systems, CCFM Workshop on Continual and Compatible Foundation Model Updates, 6 December 2025, San Diego, CA, USA and is available at :

PERMALINK : https://www.eurecom.fr/publication/8503