Learning to solve auxiliary tasks concurrently with a principal task of interest can improve performance when data is scarce or the principal task is complex. This idea is inspired by the improved generalization capability induced by solving multiple tasks simultaneously, leading to a robust shared representation. However, selecting optimal auxiliary tasks typically requires manual design or costly meta-learning approaches. We propose Detaux, a framework that discovers an unrelated auxiliary classification task via weakly supervised disentanglement at the representation level. Isolating variations relevant to the principal task in one subspace while generating orthogonal subspaces with high separability allows us to discover auxiliary labels by clustering in these subspaces, allowing a transition from Single-Task Learning (STL) to Multi-Task Learning (MTL). In particular, the original labels associated with the principal task and the newly discovered ones can be fed into any MTL framework. Experiments and ablation studies highlight the effectiveness of Detaux and reveal an unexplored link between disentangled representations and MTL. The source code is available at https://github.com/intelligolabs/Detaux.
Skenderi, G., Capogrosso, L., Toaiari, A., Denitto, M., Fummi, F., Melzi, S. (2026). Disentangled Latent Spaces Facilitate Data-Driven Auxiliary Learning. In Image Analysis and Processing – ICIAP 2025 23rd International Conference, Rome, Italy, September 15–19, 2025, Proceedings, Part I (pp.143-155). Springer Science and Business Media Deutschland GmbH [10.1007/978-3-032-10185-3_12].
Disentangled Latent Spaces Facilitate Data-Driven Auxiliary Learning
Melzi S.
2026
Abstract
Learning to solve auxiliary tasks concurrently with a principal task of interest can improve performance when data is scarce or the principal task is complex. This idea is inspired by the improved generalization capability induced by solving multiple tasks simultaneously, leading to a robust shared representation. However, selecting optimal auxiliary tasks typically requires manual design or costly meta-learning approaches. We propose Detaux, a framework that discovers an unrelated auxiliary classification task via weakly supervised disentanglement at the representation level. Isolating variations relevant to the principal task in one subspace while generating orthogonal subspaces with high separability allows us to discover auxiliary labels by clustering in these subspaces, allowing a transition from Single-Task Learning (STL) to Multi-Task Learning (MTL). In particular, the original labels associated with the principal task and the newly discovered ones can be fed into any MTL framework. Experiments and ablation studies highlight the effectiveness of Detaux and reveal an unexplored link between disentangled representations and MTL. The source code is available at https://github.com/intelligolabs/Detaux.| File | Dimensione | Formato | |
|---|---|---|---|
|
Skenderi-2026-arXiv-preprint.pdf
accesso aperto
Tipologia di allegato:
Submitted Version (Pre-print)
Licenza:
Creative Commons
Dimensione
6.13 MB
Formato
Adobe PDF
|
6.13 MB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


