mlprep
mlprep/ML Breadthhard12 min

You're ranking content and need to optimize for both engagement (clicks) and satisfaction (survey scores). How does multi-task learning help, and what are the failure modes?

formulate your answer, then —

tldr

Multi-task learning: shared representation trained jointly on multiple objectives captures more signal than separate models, especially when one task has sparse labels. Hard parameter sharing: shared layers + task-specific heads, single forward pass. Seesaw problem: conflicting objectives cause training instability. MMOE (Multi-gate Mixture of Experts) mitigates seesaw by letting tasks weight different expert sub-networks. At serving: combine task scores via weighted product or linear blend tuned by business policy.

follow-up

  • How does MMOE differ from a standard mixture-of-experts model, and what does the per-task gating add?
  • When two tasks conflict (clickbait is high-engagement but low-satisfaction), what should your model actually learn to do?
  • How would you detect that multi-task learning is hurting one of your tasks compared to a single-task baseline?