Scaling Laws For Generative Mixed-Modal Language Models

  • Unimodal scaling laws
Modalityαβ
Code0.370.32
Image-Text0.120.11
Image0.130.13
Molecules0.370.26
Speech-Text0.320.24
Speech0.310.24
Text0.180.22
  • Unimodal parameters differ widely—Code and Molecules scale “easiest,” Images “hardest”
    • Across 7 modalities (Text, Image, Image-Text, Speech, Speech-Text, Code, Molecules) the fitted α and β exponents range from ~0.11 (Images) to ~0.37 (Code/Molecules).

  • Bi-modal scaling law

    • The authors extend Hoffmann-style (a.k.a. “Chinchilla”) scaling to the mixed-modal case by writing the loss for two modalities as the average of the two unimodal scaling laws minus an interaction term Ci,j (benefit) plus small approximation/optimization penalties.
    • = model size, = dataset on modality
    • A good approximation
  • = Maximum synergy: the gap between the best possible joint loss and the average of two separate models

  • = how much more model capacity helps with the loss (functional-approx. competition)

  • = how much more data helps with the loss (optimization competition)

  • Modalities compete in the low-compute regime and turn synergistic only after a “competition barrier” is crossed
     • When the interaction term is smaller than the approximation/optimization penalties, each modality drags the other down; the resulting model under-performs two separate unimodal baselines.
     • Once model capacity (N) and data (|D|) are large enough that Ci,j dominates, the same training run becomes strictly better than training two separate models. The paper derives the exact inequality and calls the boundary the competition barrier

    • For Speech + Text, the law forecast a break-even point around a 28 B-parameter model trained on ~45 B tokens.
       • A 30 B parameter run with 50 B tokens indeed beat the two corresponding unimodal baselines, validating the theory.
  • Four emergent training phenomena fall directly out of the interaction parameters
     1. Coordinate-ascent behaviour: perplexity for one modality plateaus while another drops, then roles swap—exactly what you expect when modalities “take turns” using limited capacity.
     1.  2. The frequency of these plateaus shrinks with larger models and with smaller αi,j, linking emergence to the functional-approximation term.
     3. Optimal batch size for a mixed run correlates with βi,j; you can estimate a good batch by summing the unimodal optima and adjusting with β.
     - What the paper found – When the authors swept four global batch sizes (0.5 M → 8 M tokens) for every bi-modal pair, the batch that minimised validation loss almost always satisfied  -  - so higher βi,j ⇒ a proportionally larger batch is worth paying for .

      - _A larger β means the optimiser needs many fresh examples before each parameter update to overcome that noise. A big batch gives you those examples “in parallel” instead of by adding more steps
      
    

 4. Gradient-spike instability rises roughly with log N / αi,j, letting the scaling law warn when training will turn unstable. - Large models (N) or small interaction exponents (αi,j) both push the ratio up and training gets touchier — more spikes, more restarts, lower final quality . - Why it makes sense – αi,j governs the functional-approximation penalty - When α is small, that penalty falls only slowly as you scale parameters, so the optimiser is forced to chase steep, modality-specific minima. The bigger the log-width of the model (log N) compared with how fast that penalty shrinks (α), the more often gradients “blow up”.