Multi-objective optimisation
2021-07-13 — 2026-04-12
Wherein the difficulty of tuning machine learning hyperparameters is attributed to Pareto optimality, and the use of Lagrange multipliers is presented as a classical remedy.
Optimising for an objective defined as a weighted sum of multiple objectives with unknown weights can be difficult. Useful in multi-task learning, for example, or in weighting regularisation in regression including neural nets.
HT Cheng Soon Ong for pointing out Jonas Degrave and Ira Korshunova’s illustrated explanation of a tricky thing, Why machine learning algorithms are hard to tune (and the fix). Their summary:
“Machine learning hyperparameters are hard to tune. One way to think of why it is hard, is because it is a Pareto front of multiple objectives. One way to solve that problem is to look at Lagrange multipliers, as proposed by a paper in 1988 (Platt and Barr 1987).”
A follow-up post describes how we can make machine learning algorithms tunable.
1 Chebyshev scalarization
I just heard about this from Austin Tripp’s blog, and it looks cool. TBC
