Multi-task learning (MTL) has shown effectiveness in exploiting shared information across tasks to improve generalization. MTL assumes tasks share similarities that can improve performance. In addition, boosting algorithms have demonstrated exceptional performance across diverse learning problems, primarily due to their ability to focus on hard-to-learn instances and iteratively reduce residual errors. This makes them a promising approach for learning multi-task problems. However, real-world MTL scenarios often involve tasks that are not well-aligned (known as outlier or adversarial tasks), which do not share beneficial similarities with others and can, in fact, deteriorate the performance of the overall model. To overcome this challenge, we propose Robust-Multi-Task Gradient Boosting (R-MTGB), a novel boosting framework that explicitly models and adapts to task heterogeneity during training. R-MTGB structures the learning process into three sequential blocks: (1) learning shared patterns, (2) partitioning tasks into outliers and non-outliers with regularized parameters, and (3) fine-tuning task-specific predictors. This architecture enables R-MTGB to automatically detect and penalize outlier tasks while promoting effective knowledge transfer among related tasks. Our method integrates these mechanisms seamlessly within gradient boosting, allowing robust handling of noisy or adversarial tasks without sacrificing accuracy. Extensive experiments on both synthetic benchmarks and real-world datasets demonstrate that our approach successfully isolates outliers, transfers knowledge, and consistently reduces prediction errors for each task individually, and achieves overall performance gains across all tasks. These results highlight robustness, adaptability, and reliable convergence of R-MTGB in challenging MTL environments.
翻译:多任务学习(MTL)已被证明能有效利用任务间的共享信息以提升泛化性能。MTL假设任务间存在可提升性能的相似性。此外,提升算法因其专注于难学习实例并迭代减少残差误差的能力,已在多种学习问题上展现出卓越性能,这使其成为解决多任务学习问题的有前景方法。然而,现实世界的MTL场景常包含未良好对齐的任务(称为离群任务或对抗性任务),这些任务与其他任务缺乏有益的相似性,实际上可能损害整体模型性能。为克服这一挑战,我们提出鲁棒多任务梯度提升(R-MTGB),这是一种新颖的提升框架,能在训练过程中显式建模并适应任务异质性。R-MTGB将学习过程构建为三个顺序模块:(1)学习共享模式,(2)通过正则化参数将任务划分为离群任务与非离群任务,(3)微调任务特定预测器。该架构使R-MTGB能自动检测并惩罚离群任务,同时促进相关任务间的有效知识迁移。我们的方法将这些机制无缝集成于梯度提升中,允许在不牺牲准确性的前提下鲁棒处理噪声或对抗性任务。在合成基准和真实数据集上的大量实验表明,我们的方法能成功隔离离群任务、迁移知识,并持续降低每个任务的预测误差,同时实现所有任务的整体性能提升。这些结果凸显了R-MTGB在挑战性MTL环境中的鲁棒性、适应性和可靠收敛性。