Model-Dowser: Data-Free Importance Probing to Mitigate Catastrophic Forgetting in Multimodal Large Language Models

Hyeontaek Hwang*, Nguyen Dinh Son*, Daeyoung Kim
School of Computing, KAIST
*Equal contribution    Corresponding author
ICML 2026

Abstract

Fine-tuning Multimodal Large Language Models (MLLMs) on task-specific data is an effective way to improve performance on downstream applications. However, such adaptation often leads to a degradation in generalization on pretrained tasks, a phenomenon known as Catastrophic Forgetting. Existing methods that aim to mitigate this issue either become ineffective when fine-tuning deeper layers of the language decoder or scale poorly with increasing model size.

To address these limitations, we propose Model-Dowser, a novel sparse fine-tuning approach for MLLMs. Model-Dowser measures a principled importance score for each model parameter with respect to pretrained generalization (prior to downstream adaptation) by jointly considering weight magnitudes, input activations, and output sensitivities. During fine-tuning, Model-Dowser selectively preserves high-importance parameters and updates the remaining.

Comprehensive experiments on two representative MLLMs, LLaVA and NVILA, demonstrate that Model-Dowser effectively mitigates catastrophic forgetting and consistently outperforms prior methods, while remaining resource-efficient and scalable to multi-billion-parameter models.

Motivation

Recent work on catastrophic forgetting in MLLMs is predominantly evaluated under shallow fine-tuning settings, where only the last few layers of the language decoder are updated. However, earlier layers play a critical role in multimodal understanding, suggesting that shallow fine-tuning may not fully exploit the model's adaptation capacity.

We examine forgetting under deeper fine-tuning regimes and find that post-merging methods degrade rapidly once fine-tuning extends to earlier decoder layers. Sparse fine-tuning methods are more stable, but their memory overhead limits scalability. Model-Dowser remains robust across all fine-tuning depths while matching the memory complexity of standard fine-tuning.

Radar chart comparing Model-Dowser against baselines on LLaVA and NVILA fine-tuned on ImageNet-R.

Radar charts comparing upstream and downstream task balance across methods.

H-score stability across varying fine-tuning depths for Model-Dowser and baselines.

H-score stability across varying fine-tuning depths. Model-Dowser (red) consistently outperforms prior methods.

Method Overview

Model-Dowser consists of a three-stage pipeline that identifies and preserves functionally critical parameters before downstream fine-tuning:

  1. Probing — Sample Jacobian matrices and input activations using synthetically generated prompts. This data-free step leverages the MLLM's own generative capability to probe its functional response without requiring access to the original pretraining data.
  2. Compute Score — Assign a sensitivity-based importance score to each parameter: Sij = ‖Ji2 · |Wij| · |hj−1|, jointly capturing output sensitivity, connection strength, and input activity.
  3. Sparse Fine-tuning — Freeze the top-importance parameters and update only the least important ρ% during downstream adaptation, preserving pretrained knowledge while enabling task-specific learning.
Overall architecture of Model-Dowser showing the three-stage pipeline: Probing, Compute Score, and Sparse Fine-tuning.

Overall architecture of Model-Dowser. Parameters highlighted in yellow are updated during sparse fine-tuning.

Key Contributions

  • Forgetting diagnosis across fine-tuning depth. We reveal that MLLMs experience severe catastrophic forgetting as fine-tuning extends to earlier decoder layers, and that existing approaches are either ineffective or exhibit inconsistent behavior under this regime.
  • A scalable sparse fine-tuning method. Model-Dowser introduces a data-free importance score derived from input activations and output sensitivity, selectively freezing critical parameters to enable effective downstream learning without loss of pretrained knowledge.
  • Theoretical justification. We provide a theoretical analysis showing that our importance score captures the sensitivity of model outputs to individual parameter perturbations, explaining why preserving high-score parameters retains pretrained generalization.
  • State-of-the-art results with practical efficiency. Experiments on LLaVA-1.5-7B and NVILA-Lite-2B across diverse downstream tasks show that Model-Dowser consistently outperforms prior methods while requiring no additional memory beyond standard fine-tuning.

Key Findings

1. Robust Balance Between Upstream and Downstream Performance

Across all benchmarks, different methods often achieve comparable downstream performance, but their ability to preserve pretrained knowledge varies significantly. Model-Dowser achieves the highest H-scores among all evaluated methods by preserving functionally sensitive parameters during sparse fine-tuning. For example, on NVILA-Lite-2B, Model-Dowser achieves H-scores of 85.7 and 71.2 on COCO-Caption and ImageNet-R respectively, outperforming the second-best (SPIDER) by 7.4 and 10.7 points. On LLaVA-1.5-7B, it achieves H-scores of 79.9 and 69.7 on the same tasks.

Performance comparison on NVILA-Lite-2B and LLaVA-1.5-7B across all downstream tasks

Radar charts for ImageNet-R on LLaVA and NVILA

ImageNet-R

Radar charts for COCO-Caption

COCO-Caption

Radar charts for Flickr30k

Flickr30k

Radar charts for IconQA

IconQA

2. Increased Vulnerability of Early Decoder Layers

Layer-wise analysis reveals that catastrophic forgetting is most severe when fine-tuning extends to early decoder layers. Post-merging methods (Grafting, DARE, Tailor) maintain stability across the last 4–16 layers but collapse when updates reach earlier layers, because post-hoc weight patching cannot recover from extensive disruption. SPIDER is more stable but still underperforms Model-Dowser across all 32 layers on LLaVA and 28 layers on NVILA. Model-Dowser preserves sensitive functional anchors across all depths, sustaining stability throughout.

Performance comparison across fine-tuning depths on COCO and ImageNet-R.

Performance across fine-tuning depths on COCO-Caption and ImageNet-R. The x-axis shows the number of fine-tuned layers, counted from the output layer toward the input layer.

Performance comparison across fine-tuning depths on Flickr30k and IconQA.

Performance across fine-tuning depths on Flickr30k and IconQA.

3. Robustness to Update Ratios

Model-Dowser maintains a wide operational window for the update ratio ρ. Average upstream performance remains stable for mask ratios up to ρ = 0.25, and consistently outperforms Full-FT (ρ = 1.0) across all settings. This demonstrates that the functional importance identified by sensitivity scoring is highly concentrated: as long as the most critical parameters are protected, the model remains robust to significant task-specific updates elsewhere.

Performance across mask ratios on COCO-Caption and ImageNet-R.

Upstream/downstream performance across update ratios (ρ) on COCO-Caption and ImageNet-R for NVILA-Lite-2B (a–b) and LLaVA-1.5-7B (c–d).

Performance across mask ratios on Flickr30k and IconQA.

Same analysis on Flickr30k and IconQA.

4. Stability of Data-Free Importance Estimation

The data-free importance estimation yields a stable, robust ranking of parameter sensitivity. With R = 8 Rademacher vectors and N = 64 Monte Carlo samples, Model-Dowser achieves a Hamming distance of 0.050 ± 0.002 and Spearman correlation of 0.891 ± 0.003 relative to real-data scoring, while random selection shows zero correlation.

Significance analysis of data-free importance estimation.

Significance analysis of the data-free importance estimation.

Related Links

LLaVA and NVILA are the two representative MLLM architectures used in our experiments.

BibTeX

@misc{hwang2026modeldowserdatafreeimportanceprobing,
      title={Model-Dowser: Data-Free Importance Probing to Mitigate Catastrophic Forgetting in Multimodal Large Language Models},
      author={Hyeontaek Hwang and Nguyen Dinh Son and Daeyoung Kim},
      year={2026},
      eprint={2602.04509},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2602.04509},
}