CALM: Consensus-Aware Localized Merging for Multi-Task Learning

  • Kunda Yan
  • , Min Zhang
  • , Sen Cui*
  • , Zikun Qu
  • , Bo Jiang
  • , Feng Liu
  • , Changshui Zhang*
  • *Corresponding author for this work

Research output: Contribution to journalConference articlepeer-review

Abstract

Model merging aims to integrate the strengths of multiple fine-tuned models into a unified model while preserving task-specific capabilities. Existing methods, represented by task arithmetic, are typically classified into global-and localaware methods. However, global-aware methods inevitably cause parameter interference, while local-aware methods struggle to maintain the effectiveness of task-specific details in the merged model. To address these limitations, we propose a Consensus-Aware Localized Merging (CALM) method which incorporates localized information aligned with global task consensus, ensuring its effectiveness post-merging. CALM consists of three key components: (1) class-balanced entropy minimization sampling, providing a more flexible and reliable way to leverage unsupervised data; (2) an efficient-aware framework, selecting a small set of tasks for sequential merging with high scalability; (3) a consensusaware mask optimization, aligning localized binary masks with global task consensus and merging them conflict-free. Experiments demonstrate the superiority and robustness of our CALM, significantly outperforming existing methods and achieving performance close to traditional MTL.

Original languageEnglish
Pages (from-to)70309-70329
Number of pages21
JournalProceedings of Machine Learning Research
Volume267
StatePublished - 2025
Event42nd International Conference on Machine Learning, ICML 2025 - Vancouver, Canada
Duration: 13 Jul 202519 Jul 2025

Fingerprint

Dive into the research topics of 'CALM: Consensus-Aware Localized Merging for Multi-Task Learning'. Together they form a unique fingerprint.

Cite this