论文标题

通过指数移动平均减肥策略来减轻多任务学习中的负转移

Mitigating Negative Transfer in Multi-Task Learning with Exponential Moving Average Loss Weighting Strategies

论文作者

Lakkapragada, Anish, Sleiman, Essam, Surabhi, Saimourya, Wall, Dennis P.

论文摘要

多任务学习(MTL)与使用一组常规的单任务模型相比,它可以在多个任务上更有效地训练模型在多个任务上训练模型,因此在深度学习中越来越感兴趣。但是,MTL可能是不切实际的,因为某些任务可以主导培训和伤害其他任务,因此与多任务相比,在单个任务模型中,某些任务的表现更好。这些问题被广泛归类为负转移,并且已经提出了许多先前的方法来减轻这些问题。当前一种减轻负转移的方法是加重每个损失,以使它们处于相同的范围。尽管当前的损失平衡方法依赖于优化或复杂的数值分析,但没有直接根据其观察到的幅度直接扩展损失。我们建议通过指数移动平均值根据缩放缩放的损失平衡进行多种技术,并根据三个已建立的数据集对当前表现最佳的方法进行基准测试。在这些数据集上,与当前表现最佳的方法相比,它们的性能可比性(即使不是更高)。

Multi-Task Learning (MTL) is a growing subject of interest in deep learning, due to its ability to train models more efficiently on multiple tasks compared to using a group of conventional single-task models. However, MTL can be impractical as certain tasks can dominate training and hurt performance in others, thus making some tasks perform better in a single-task model compared to a multi-task one. Such problems are broadly classified as negative transfer, and many prior approaches in the literature have been made to mitigate these issues. One such current approach to alleviate negative transfer is to weight each of the losses so that they are on the same scale. Whereas current loss balancing approaches rely on either optimization or complex numerical analysis, none directly scale the losses based on their observed magnitudes. We propose multiple techniques for loss balancing based on scaling by the exponential moving average and benchmark them against current best-performing methods on three established datasets. On these datasets, they achieve comparable, if not higher, performance compared to current best-performing methods.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源