Scaling Laws for Multimodal Foundation Models: A Unified Framework
Abstract
Foundation models trained on multiple data modalities — text, images, and audio — have demonstrated capabilities that exceed the sum of their unimodal components. In this work, we present a unified empirical framework for characterizing scaling laws in multimodal foundation models. Through controlled experiments training over 200 model configurations on curated text-image-audio datasets totaling 4.2T tokens, we derive modality-specific and cross-modal scaling exponents. We introduce the Cross-Modal Alignment Tax (CMAT) and the Unified Scaling Exponent (USE) framework, enabling principled compute allocation decisions in multimodal training.
1. Introduction
The scaling laws established by Kaplan et al. (2020) [1] and refined by Hoffmann et al. (2022) [2] have provided the machine learning community with predictive tools for optimizing the allocation of compute, data, and parameters in language model training. These laws, however, were derived exclusively for unimodal text models. As the field moves toward multimodal foundation models that jointly process text, images, and audio [3, 4], a critical question emerges: do the same scaling relationships hold, and if not, how must they be modified?
This question has practical urgency. Training a multimodal model at frontier scale costs tens of millions of dollars. Without reliable scaling predictions, organizations risk misallocating compute — overinvesting in one modality at the expense of others, or training models that are either parameter-starved or data-starved relative to their compute budget.
We address this challenge with three contributions:
- Empirical scaling laws for multimodal models across three modalities (text, image, audio), derived from over 200 training runs.
- The Cross-Modal Alignment Tax (CMAT), a quantitative measure of the additional compute overhead required for modality alignment.
- The Unified Scaling Exponent (USE) framework, a tensor-based formulation that extends power-law scaling to heterogeneous data regimes.
2. Related Work
Kaplan et al. (2020) [1] established that language model loss scales as a power law in model parameters , dataset size , and compute :
Hoffmann et al. (2022) [2] refined these results, showing that the Kaplan et al. recommendations were undertrained and that optimal performance requires scaling and in roughly equal proportion (the "Chinchilla" law). Clark et al. (2022) [5] extended scaling analysis to sparse mixture-of-expert models.
For multimodal models, scaling behavior has been studied in narrower settings. Cherti et al. (2023) [6] examined scaling in CLIP-style contrastive models, finding that image-text alignment benefits from data scale more than parameter scale. Aghajanyan et al. (2023) [7] studied scaling in text-code-math mixtures but did not extend to non-textual modalities.
Our work is the first to provide a unified scaling framework spanning text, image, and audio modalities with a single mathematical formalism.
3. Methodology
3.1 Training Setup
We train decoder-only transformer models using a modality-agnostic tokenization scheme:
- Text: BPE tokenizer (64K vocabulary)
- Images: ViT-based patch tokenizer producing 256 tokens per image
- Audio: Mel-spectrogram encoder producing 128 tokens per second of audio
Modality-specific projection layers map each tokenizer's output to a shared embedding dimension . We train 216 configurations spanning:
| Parameter | Range |
|---|---|
| Model size () | 125M – 34B |
| Dataset size () | 10B – 4.2T tokens |
| Modality ratios | Text: 40–80%, Image: 10–40%, Audio: 5–20% |
| Compute () | – FLOPs |
All models are trained with AdamW (, ), cosine learning rate schedule with warmup, and bf16 mixed precision on clusters of H100 GPUs.
3.2 The Cross-Modal Alignment Tax
We define the Cross-Modal Alignment Tax as the additional compute required for a multimodal model to match the per-modality loss of a unimodal model at the same parameter count:
)}{C_{\text{uni}}(L_{m_i}^)} - 1
where is a target loss on modality , is the compute to achieve that loss in the multimodal setting, and is the compute for a unimodal model.
Intuitively, CMAT captures the cost of learning shared representations. If CMAT = 0, multimodal training is "free" in terms of per-modality performance. If CMAT > 0, there is a tax for joint training.
3.3 Unified Scaling Exponent Framework
We model the loss of a multimodal model as:
The first term is the standard parameter scaling. The second captures per-modality data scaling with modality-specific exponents . The third — the cross-modal interaction term — is our key innovation: it captures how jointly scaling two modalities produces effects not predicted by their independent scaling curves.
The cross-modal interaction coefficients and exponents form a symmetric tensor which we term the modality interaction tensor. This tensor is estimated via nonlinear least squares fitting across all training runs.
4. Results and Discussion
4.1 Modality-Specific Scaling Exponents
We find that each modality exhibits distinct scaling behavior:
| Modality | (data exponent) | (irreducible offset) |
|---|---|---|
| Text | ||
| Image | ||
| Audio |
Text scales most efficiently with data (), consistent with the observation that language has higher token-level information density. The parameter scaling exponent is shared across modalities within confidence intervals, suggesting that model capacity benefits all modalities roughly equally.
4.2 Cross-Modal Alignment Tax
The empirically measured CMAT values are:
- CMAT(text, image) = (23% compute overhead)
- CMAT(text, audio) = (31% overhead)
- CMAT(image, audio) = (18% overhead)
Notably, CMAT decreases with model scale. At 125M parameters, CMAT(text, image) ; at 34B, it drops to . This suggests that larger models amortize the alignment cost more efficiently, providing an additional incentive for scale in multimodal training.
4.3 Cross-Modal Interaction Effects
The modality interaction tensor reveals significant superlinear benefits:
The negative indicates that jointly scaling text and image data produces lower loss than predicted by independent scaling — a synergistic effect. This manifests most strongly on cross-modal benchmarks: for instance, visual question answering accuracy scales as in the multimodal setting versus for a text-only model fine-tuned on VQA.
The audio interactions are weaker (, ), suggesting that audio contributes less cross-modal synergy at current data scales.
4.4 Predictive Accuracy
We evaluate the USE framework's predictive power by fitting on 80% of training runs and predicting held-out losses. The mean absolute percentage error (MAPE) across all held-out configurations is:
This demonstrates that accounting for cross-modal interactions is essential for accurate loss prediction in multimodal settings.
4.5 Optimal Compute Allocation
Using the USE framework, we derive the optimal modality data ratio as a function of total compute :
This shows that as compute increases, the optimal allocation shifts toward modalities with stronger cross-modal interactions, not simply toward modalities with steeper individual scaling curves.
5. Conclusion
We have presented the first unified scaling law framework for multimodal foundation models spanning text, image, and audio. Our key findings are:
- Modalities exhibit distinct data scaling exponents, with text scaling most efficiently.
- The Cross-Modal Alignment Tax quantifies the cost of joint training, and this tax decreases with model scale.
- Cross-modal interactions produce superlinear scaling benefits, particularly between text and images.
- The Unified Scaling Exponent framework predicts multimodal losses within 3.2% MAPE, enabling principled compute allocation.
These results have immediate practical implications: organizations training multimodal models can use our framework to determine optimal parameter counts, dataset compositions, and compute budgets before committing to expensive training runs. We release our full dataset of 216 training run logs and the USE fitting toolkit to support future research.
References
[1] J. Kaplan et al., "Scaling Laws for Neural Language Models," arXiv:2001.08361, 2020.
[2] J. Hoffmann et al., "Training Compute-Optimal Large Language Models," NeurIPS, 2022.
[3] J. Alayrac et al., "Flamingo: a Visual Language Model for Few-Shot Learning," NeurIPS, 2022.
[4] OpenAI, "GPT-4 Technical Report," arXiv:2303.08774, 2023.
[5] A. Clark et al., "Unified Scaling Laws for Routed Language Models," ICML, 2022.
[6] M. Cherti et al., "Reproducible Scaling Laws for Contrastive Language-Image Learning," CVPR, 2023.
[7] A. Aghajanyan et al., "Scaling Data-Constrained Language Models," NeurIPS, 2023.


