论文标题
小组发出:通过小组一对多任务进行快速的DETR培训
Group DETR: Fast DETR Training with Group-Wise One-to-Many Assignment
论文作者
论文摘要
检测变压器(DETR)依赖于一对一的分配,将一个地面真实对象分配给一个预测,用于端到端检测,而无需NMS后处理。众所周知,将一个地面对象分配给多个预测的一对多分配成功地用于检测方法,例如更快的R-CNN和FCO。虽然天真的一对多任务对DETR不起作用,但在DETR培训中应用一对多的任务仍然具有挑战性。在本文中,我们介绍了Group Detr,这是一种简单而有效的DETR培训方法,它介绍了一种一对多分配的团体方式。这种方法涉及使用多个对象查询,在每个组中进行一对一分配,并分别执行解码器自我注意。它类似于与自动学习对象查询增强的数据增强。这也等同于同时训练相同体系结构的参数共享网络,从而引入更多的监督,从而改善了DETR培训。推理过程与经过正常训练的DETR相同,并且只需要一组无需修改的查询即可。组DETR具有通用性,适用于各种DETR变体。该实验表明,组DETR显着加快了训练收敛的速度,并提高了各种基于DITR的模型的性能。代码将在\ url {https://github.com/atten4vis/groupdetr}上找到。
Detection transformer (DETR) relies on one-to-one assignment, assigning one ground-truth object to one prediction, for end-to-end detection without NMS post-processing. It is known that one-to-many assignment, assigning one ground-truth object to multiple predictions, succeeds in detection methods such as Faster R-CNN and FCOS. While the naive one-to-many assignment does not work for DETR, and it remains challenging to apply one-to-many assignment for DETR training. In this paper, we introduce Group DETR, a simple yet efficient DETR training approach that introduces a group-wise way for one-to-many assignment. This approach involves using multiple groups of object queries, conducting one-to-one assignment within each group, and performing decoder self-attention separately. It resembles data augmentation with automatically-learned object query augmentation. It is also equivalent to simultaneously training parameter-sharing networks of the same architecture, introducing more supervision and thus improving DETR training. The inference process is the same as DETR trained normally and only needs one group of queries without any architecture modification. Group DETR is versatile and is applicable to various DETR variants. The experiments show that Group DETR significantly speeds up the training convergence and improves the performance of various DETR-based models. Code will be available at \url{https://github.com/Atten4Vis/GroupDETR}.