论文标题

利用收敛行为来平衡多任务学习中的冲突任务

Leveraging convergence behavior to balance conflicting tasks in multi-task learning

论文作者

Nakamura, Angelica Tiemi Mizuno, Wolf, Denis Fernando, Grassi Jr, Valdir

论文摘要

多任务学习是一种学习范式,它使用关联的任务来改善性能概括。学习多个任务的一种常见方法是通过硬参数共享方法,其中单个体系结构用于共享相同的参数子集,从而在训练过程中产生归纳偏差。由于其简单性,提高概括并降低计算成本的潜力,它引起了科学和工业社区的关注。但是,任务通常相互冲突,这使得定义应如何将多个任务的梯度组合在一起以同时学习的挑战。为了解决这个问题,我们使用多目标优化的概念提出了一种考虑梯度的时间行为的方法,以创建动态偏见,以调整反向传播过程中每个任务的重要性。该方法的结果是更多地关注在最后一次迭代期间发散或没有受益的任务,以确保同时学习正向所有任务的绩效最大化。结果,我们从经验上表明,所提出的方法的表现优于学习矛盾的任务的最先进方法。与采用的基线不同,我们的方法确保所有任务都达到良好的概括性能。

Multi-Task Learning is a learning paradigm that uses correlated tasks to improve performance generalization. A common way to learn multiple tasks is through the hard parameter sharing approach, in which a single architecture is used to share the same subset of parameters, creating an inductive bias between them during the training process. Due to its simplicity, potential to improve generalization, and reduce computational cost, it has gained the attention of the scientific and industrial communities. However, tasks often conflict with each other, which makes it challenging to define how the gradients of multiple tasks should be combined to allow simultaneous learning. To address this problem, we use the idea of multi-objective optimization to propose a method that takes into account temporal behaviour of the gradients to create a dynamic bias that adjust the importance of each task during the backpropagation. The result of this method is to give more attention to the tasks that are diverging or that are not being benefited during the last iterations, allowing to ensure that the simultaneous learning is heading to the performance maximization of all tasks. As a result, we empirically show that the proposed method outperforms the state-of-art approaches on learning conflicting tasks. Unlike the adopted baselines, our method ensures that all tasks reach good generalization performances.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源