Universal Model Adapter Fusion (UMAF): A Framework for Cross-Architecture Capability Transfer and Fusion
Abstract
The Universal Model Adapter Fusion (UMAF) framework enables the transfer and fusion of capabilities across language models of diverse architectures and scales. By leveraging a universal latent space for capability representation, a size interpolator for scaling, a fusion module for dynamic combination, and an adapter generator for parameter adjustments, UMAF produces lightweight, architecture-agnostic adapters. This paper provides a robust theoretical foundation, detailed methodological clarifications, and expanded experimental validation. UMAF demonstrates significant potential for creating bespoke, efficient AI models with enhanced interpretability and modularity.
Large language models (LLMs) have achieved remarkable success in specialized tasks, yet transferring and combining their capabilities across models with differing architectures or scales remains a sig