You're ranking content and need to optimize for both engagement (clicks) and satisfaction (survey scores). How does multi-task learning help, and what are the failure modes?
formulate your answer, then —
tldr
Multi-task learning: shared representation trained jointly on multiple objectives captures more signal than separate models, especially when one task has sparse labels. Hard parameter sharing: shared layers + task-specific heads, single forward pass. Seesaw problem: conflicting objectives cause training instability. MMOE (Multi-gate Mixture of Experts) mitigates seesaw by letting tasks weight different expert sub-networks. At serving: combine task scores via weighted product or linear blend tuned by business policy.
follow-up
- How does MMOE differ from a standard mixture-of-experts model, and what does the per-task gating add?
- When two tasks conflict (clickbait is high-engagement but low-satisfaction), what should your model actually learn to do?
- How would you detect that multi-task learning is hurting one of your tasks compared to a single-task baseline?