PP-YoLoE | PP-YoLov2全面升级Anchor-Free,速度精度完美超越YoLoX和YoLov5

技术

picture.image

picture.image

PP-YOLOE是基于PP-YOLOv2的卓越的单阶段Anchor-free模型,超越了多种流行的yolo模型。PP-YOLOE有一系列的模型,即s/m/l/x,可以通过width multiplier和depth multiplier配置。PP-YOLOE避免使用诸如deformable convolution或者matrix nms之类的特殊算子,以使其能轻松地部署在多种多样的硬件上。

picture.image

picture.image

1

模型架构

picture.image

PP-YOLOE由以下方法组成:

  • 可扩展的backbone和neck
  • Task Alignment Learning
  • Efficient Task-aligned head with DFL和VFL
  • SiLU激活函数

1.1、Backbone

picture.image

PP-YOLOE的Backbone主要是使用RepVGG模块以及CSP的模型思想对ResNet及逆行的改进,同时也使用了SiLU激活函数、Effitive SE Attention等模块,下面我们一一道来。

1、RepVGG

RepVGG,这个网络就是在VGG的基础上面进行改进,主要的思路包括:

  1. 在VGG网络的Block块中加入了Identity和残差分支,相当于把ResNet网络中的精华应用 到VGG网络中;
  2. 模型推理阶段,通过Op融合策略将所有的网络层都转换为3×3卷积,便于网络的部署和加速。

picture.image

picture.image

上图展示了模型推理阶段的重参数化过程,其实就是一个OP融合和OP替换的过程。图A从结构化的角度展示了整个重参数化流程, 图B从模型参数的角度展示了整个重参数化流程。整个重参数化步骤如下所示:

步骤1 :首先通过式3将残差块中的卷积层和BN层进行融合,该操作在很多深度学习框架的推理阶段都会执行。图中的蓝色框中执行3×3卷积+BN层的融合,图中的黑色矩形框中执行1×1卷积+BN层的融合,图中的黄色矩形框中执行3×3卷积(卷积核设置为全1)+BN层的融合。其中表示转换前的卷积层参数,表示BN层的均值,表示BN层的方差,和分别表示BN层的尺度因子和偏移因子,W’和b’分别表示融合之后的卷积的权重和偏置。

picture.image

步骤2 :将融合后的卷积层转换为3×3卷积,即将具体不同卷积核的卷积均转换为具有3×3大小的卷积核的卷积。由于整个残差块中可能包含1×1卷积分支和Identity两种分支,如图中的黑框和黄框所示。对于1×1卷积分支而言,整个转换过程就是利用3×3卷积核替换1×1卷积核,具体的细节如图中的紫框所示,即将1×1卷积核中的数值移动到3×3卷积核的中心点即可;对于Identity分支而言,该分支并没有改变输入的特征映射的数值,那么可以设置一个3×3的卷积核,将所有的9个位置处的权重值都设置为1,那么它与输入的特征映射相乘之后,保持了原来的数值,具体的细节如图中的褐色框所示。

步骤3 :合并残差分支中的3×3卷积。即将所有分支的权重W和偏置B叠加起来,从而获得一个融合之后的3×3卷积层。

为什么要用VGG式模型?

除了相信简单就是美以外,VGG式极简模型至少还有5大现实的优势:

  1. 3×3卷积非常快。在GPU上,3×3卷积的计算密度(理论运算量除以所用时间)可达1×1和5×5卷积的4倍。
  2. 单路架构非常快,因为并行度高。同样的计算量,“大而整”的运算效率远超“小而碎”的运算。
  3. 单路架构省内存。例如,ResNet的shortcut虽然不占计算量,却增加了一倍的显存占用。
  4. 单路架构灵活性更好,容易改变各层的宽度(如剪枝)。
  5. RepVGG主体部分只有一种算子:3×3卷积接ReLU。在设计专用芯片时,给定芯片尺寸或造价可以集成海量的3×3卷积+ReLU计算单元来达到很高的效率。

picture.image

下图表示RepVGG推理融合后的ONNX输出,可以看出简化了很多。

picture.image


        
          
class RepVggBlock(nn.Layer):  
    def \_\_init\_\_(self, ch\_in, ch\_out, act='relu'):  
        super(RepVggBlock, self).__init__()  
        self.ch_in = ch_in  
        self.ch_out = ch_out  
        self.conv1 = ConvBNLayer(ch_in, ch_out, 3, stride=1, padding=1, act=None)  
        self.conv2 = ConvBNLayer(ch_in, ch_out, 1, stride=1, padding=0, act=None)  
        self.act = get_act_fn(act) if act is None or isinstance(act, (str, dict)) else act  
  
    def forward(self, x):  
        if hasattr(self, 'conv'):  
            y = self.conv(x)  
        else:  
            y = self.conv1(x) + self.conv2(x)  
        y = self.act(y)  
        return y  
  
    def convert\_to\_deploy(self):  
        if not hasattr(self, 'conv'):  
            self.conv = nn.Conv2D(in_channels=self.ch_in, out_channels=self.ch_out, kernel_size=3, stride=1, padding=1, groups=1)  
        kernel, bias = self.get_equivalent_kernel_bias()  
        self.conv.weight.set_value(kernel)  
        self.conv.bias.set_value(bias)  
  
    def get\_equivalent\_kernel\_bias(self):  
        # 融合推理  
        kernel3x3, bias3x3 = self._fuse_bn_tensor(self.conv1)  
        kernel1x1, bias1x1 = self._fuse_bn_tensor(self.conv2)  
        return kernel3x3 + self._pad_1x1_to_3x3_tensor(kernel1x1), bias3x3 + bias1x1  
  
    def \_pad\_1x1\_to\_3x3\_tensor(self, kernel1x1):  
        if kernel1x1 is None:  
            return 0  
        else:  
            return nn.functional.pad(kernel1x1, [1, 1, 1, 1])  
  
    def \_fuse\_bn\_tensor(self, branch):  
        if branch is None:  
            return 0, 0  
        kernel = branch.conv.weight  
        running_mean = branch.bn._mean  
        running_var = branch.bn._variance  
        gamma = branch.bn.weight  
        beta = branch.bn.bias  
        eps = branch.bn._epsilon  
        std = (running_var + eps).sqrt()  
        t = (gamma / std).reshape((-1, 1, 1, 1))  
        return kernel * t, beta - running_mean * gamma / std  

      

2、Swish激活函数

从代码和公式来看,Swish包含了SiLU,换句话说SiLU是Swish的一种特例。

picture.image

所以画图基本上都使用了SiLU代替Swish,因为YOLOE中的Swish的,也就是SiLU激活函数。

picture.image

β是个常数或可训练的参数。Swish 具备无上界有下界、平滑、非单调的特性。Swish 在深层模型上的效果优于 ReLU。

例如,仅仅使用 Swish 单元替换 ReLU 就能把 Mobile NASNetA 在 ImageNet 上的 top-1 分类准确率提高 0.9%,Inception-ResNet-v 的分类准确率提高 0.6%。

picture.image

picture.image


        
          
class ConvBNLayer(nn.Layer):  
    def \_\_init\_\_(self, ch\_in, ch\_out, filter\_size=3, stride=1, groups=1, padding=0, act=None):  
        super(ConvBNLayer, self).__init__()  
        self.conv = nn.Conv2D(in_channels=ch_in, out_channels=ch_out, kernel_size=filter_size, stride=stride, padding=padding, groups=groups, bias_attr=False)  
        self.bn = nn.BatchNorm2D(ch_out, weight_attr=ParamAttr(regularizer=L2Decay(0.0)), bias_attr=ParamAttr(regularizer=L2Decay(0.0)))  
        self.act = get_act_fn(act) if act is None or isinstance(act, (str, dict)) else act  
  
    def forward(self, x):  
        x = self.conv(x)  
        x = self.bn(x)  
        x = self.act(x)  
  
        return x  

      

3、Effective SE Attention

该模块主要是来自于《CenterMask:Real-Time Anchor-Free Instance Segmentation》中的eSE模块;

picture.image

在输出的内部添加了一个channel上的attention模块eSE。原始的SE模块中使用2个FC去进行channel权重映射,但是为了减少计算量通常会将FC中的channel给剪裁一些(小于输入的channel),这就引入了一些信息的损失,为此文章直接将2个FC替换为了1个FC。

picture.image


        
          
class EffectiveSELayer(nn.Layer):  
    def \_\_init\_\_(self, channels, act='hardsigmoid'):  
        super(EffectiveSELayer, self).__init__()  
        self.fc = nn.Conv2D(channels, channels, kernel_size=1, padding=0)  
        self.act = get_act_fn(act) if act is None or isinstance(act, (str, dict)) else act  
  
    def forward(self, x):  
        x_se = x.mean((2, 3), keepdim=True)  
        x_se = self.fc(x_se)  
        return x * self.act(x_se)  

      

4、CSPNet结构

picture.image

CSPNet的主要思想还是Partial Dense Block,设计Partial Dense Block的目的是:

  • 增加梯度路径:通过分裂合并策略,可以使梯度路径的数目翻倍。由于采用了跨阶段的策略,可以减轻使用显式特征映射复制进行连接的缺点;
  • 平衡各层的计算:通常情况下,DenseNet底层的信道数远远大于增长率。由于部分dense block中的dense layer操作所涉及的底层信道只占原始信道的一半,因此可以有效地解决近一半的计算瓶颈;
  • 减少内存流量:假设dense block在DenseNet中的基本特征映射大小为,增长率为d,并且共m层。然后,该dense block的CIO为,部分dense block的CIO为。虽然m和d通常比c小得多,但部分dense block最多可以节省网络内存流量的一半。

picture.image


        
          
class CSPResStage(nn.Layer):  
    def \_\_init\_\_(self, block\_fn, ch\_in, ch\_out, n, stride, act='relu', attn='eca'):  
        super(CSPResStage, self).__init__()  
        ch_mid = (ch_in + ch_out) // 2  
        if stride == 2:  
            self.conv_down = ConvBNLayer(ch_in, ch_mid, 3, stride=2, padding=1, act=act)  
        else:  
            self.conv_down = None  
        self.conv1 = ConvBNLayer(ch_mid, ch_mid // 2, 1, act=act)  
        self.conv2 = ConvBNLayer(ch_mid, ch_mid // 2, 1, act=act)  
        self.blocks = nn.Sequential(* [block_fn(ch_mid // 2, ch_mid // 2, act=act, shortcut=True) for i in range(n)])  
        if attn:  
            self.attn = EffectiveSELayer(ch_mid, act='hardsigmoid')  
        else:  
            self.attn = None  
  
        self.conv3 = ConvBNLayer(ch_mid, ch_out, 1, act=act)  
  
    def forward(self, x):  
        if self.conv_down is not None:  
            x = self.conv_down(x)  
        y1 = self.conv1(x)  
        y2 = self.blocks(self.conv2(x))  
        y = paddle.concat([y1, y2], axis=1)  
        if self.attn is not None:  
            y = self.attn(y)  
        y = self.conv3(y)  
        return y  

      

5、SPP结构

SPP-Net全名为Spatial Pyramid Pooling(空间金字塔池化结构),2015年由微软研究院的何恺明提出,主要解决2个问题:

  1. 有效避免了R-CNN算法对图像区域剪裁、缩放操作导致的图像物体剪裁不全以及形状扭曲等问题。
  2. 解决了卷积神经网络对图像重复特征提取的问题,大大提高了产生候选框的速度,且节省了计算成本。

picture.image

SPP 显著特点

  1. 不管输入尺寸是怎样,SPP 可以产生固定大小的输出
  2. 使用多个窗口(pooling window)
  3. SPP 可以使用同一图像不同尺寸(scale)作为输入, 得到同样长度的池化特征。

其它特点

  1. 由于对输入图像的不同纵横比和不同尺寸,SPP同样可以处理,所以提高了图像的尺度不变(scale-invariance)和降低了过拟合(over-fitting)
  2. 实验表明训练图像尺寸的多样性比单一尺寸的训练图像更容易使得网络收敛(convergence)
  3. SPP 对于特定的CNN网络设计和结构是独立的。(也就是说,只要把SPP放在最后一层卷积层后面,对网络的结构是没有影响的, 它只是替换了原来的pooling层)
  4. 不仅可以用于图像分类而且可以用来目标检测

picture.image

通过spp模块实现局部特征和全局特征(所以空间金字塔池化结构的最大的池化核要尽可能的接近等于需要池化的featherMap的大小)的featherMap级别的融合,丰富最终特征图的表达能力,从而提高MAP。


        
          
class SPP(nn.Layer):  
    def \_\_init\_\_(self, ch\_in, ch\_out, k, pool\_size, act='swish', data\_format='NCHW'):  
        super(SPP, self).__init__()  
        self.pool = []  
        self.data_format = data_format  
        for i, size in enumerate(pool_size):  
            pool = self.add_sublayer('pool{}'.format(i), nn.MaxPool2D(kernel_size=size, stride=1, padding=size // 2, data_format=data_format, ceil_mode=False))  
            self.pool.append(pool)  
        self.conv = ConvBNLayer(ch_in, ch_out, k, padding=k // 2, act=act)  
  
    def forward(self, x):  
        outs = [x]  
        for pool in self.pool:  
            outs.append(pool(x))  
        if self.data_format == 'NCHW':  
            y = paddle.concat(outs, axis=1)  
        else:  
            y = paddle.concat(outs, axis=-1)  
  
        y = self.conv(y)  
        return y  
  
  
class CSPStage(nn.Layer):  
    def \_\_init\_\_(self, block\_fn, ch\_in, ch\_out, n, act='swish', spp=False):  
        super(CSPStage, self).__init__()  
  
        ch_mid = int(ch_out // 2)  
        self.conv1 = ConvBNLayer(ch_in, ch_mid, 1, act=act)  
        self.conv2 = ConvBNLayer(ch_in, ch_mid, 1, act=act)  
        self.convs = nn.Sequential()  
        next_ch_in = ch_mid  
        for i in range(n):  
            self.convs.add_sublayer(str(i), eval(block_fn)(next_ch_in, ch_mid, act=act, shortcut=False))  
            if i == (n - 1) // 2 and spp:  
                self.convs.add_sublayer('spp', SPP(ch_mid * 4, ch_mid, 1, [5, 9, 13], act=act))  
            next_ch_in = ch_mid  
        self.conv3 = ConvBNLayer(ch_mid * 2, ch_out, 1, act=act)  
  
    def forward(self, x):  
        y1 = self.conv1(x)  
        y2 = self.conv2(x)  
        y2 = self.convs(y2)  
        y = paddle.concat([y1, y2], axis=1)  
        y = self.conv3(y)  
        return y  

      

1.2、Neck

yoloe的neck结构采用的依旧是FPN+PAN结构模式,将Neck部分用立体图画出来,更直观的看下两部分之间是如何通过FPN结构融合的。

picture.image

如图所示,FPN是自顶向下的,将高层特征通过上采样和低层特征做融合得到进行预测的特征图。

picture.image

和FPN层不同,yoloe在FPN层的后面还添加了一个自底向上的特征金字塔。FPN是自顶向下,将高层的强语义特征传递下来,对整个金字塔进行增强,不过只增强了语义信息,对定位信息没有传递,而本文就是针对这一点,在FPN的后面添加一个自底向上的金字塔。这样的操作是对FPN的补充,将低层的强定位特征传递上去。

picture.image


        
          
class CustomCSPPAN(nn.Layer):  
    __shared__ = ['norm\_type', 'data\_format', 'width\_mult', 'depth\_mult', 'trt']  
  
    def \_\_init\_\_(self, in\_channels=[256, 512, 1024], out\_channels=[1024, 512, 256], norm\_type='bn', act='leaky',  
                 stage\_fn='CSPStage', block\_fn='BasicBlock', stage\_num=1, block\_num=3, drop\_block=False,  
                 block\_size=3, keep\_prob=0.9, spp=False, data\_format='NCHW', width\_mult=1.0,  
                 depth\_mult=1.0, trt=False):  
  
        super(CustomCSPPAN, self).__init__()  
        out_channels = [max(round(c * width_mult), 1) for c in out_channels]  
        block_num = max(round(block_num * depth_mult), 1)  
        act = get_act_fn(act, trt=trt) if act is None or isinstance(act, (str, dict)) else act  
        self.num_blocks = len(in_channels)  
        self.data_format = data_format  
        self._out_channels = out_channels  
        in_channels = in_channels[::-1]  
        fpn_stages = []  
        fpn_routes = []  
        for i, (ch_in, ch_out) in enumerate(zip(in_channels, out_channels)):  
            if i > 0:  
                ch_in += ch_pre // 2  
  
            stage = nn.Sequential()  
            for j in range(stage_num):  
                stage.add_sublayer(str(j), eval(stage_fn)(block_fn, ch_in if j == 0 else ch_out, ch_out, block_num, act=act, spp=(spp and i == 0)))  
  
            if drop_block:  
                stage.add_sublayer('drop', DropBlock(block_size, keep_prob))  
  
            fpn_stages.append(stage)  
  
            if i < self.num_blocks - 1:  
                fpn_routes.append(ConvBNLayer(ch_in=ch_out, ch_out=ch_out // 2, filter_size=1, stride=1, padding=0, act=act))  
  
            ch_pre = ch_out  
  
        self.fpn_stages = nn.LayerList(fpn_stages)  
        self.fpn_routes = nn.LayerList(fpn_routes)  
  
        pan_stages = []  
        pan_routes = []  
        for i in reversed(range(self.num_blocks - 1)):  
            pan_routes.append(ConvBNLayer(ch_in=out_channels[i + 1], ch_out=out_channels[i + 1], filter_size=3, stride=2, padding=1, act=act))  
            ch_in = out_channels[i] + out_channels[i + 1]  
            ch_out = out_channels[i]  
            stage = nn.Sequential()  
            for j in range(stage_num):  
                stage.add_sublayer(str(j), eval(stage_fn)(block_fn, ch_in if j == 0 else ch_out, ch_out, block_num, act=act, spp=False))  
            if drop_block:  
                stage.add_sublayer('drop', DropBlock(block_size, keep_prob))  
  
            pan_stages.append(stage)  
  
        self.pan_stages = nn.LayerList(pan_stages[::-1])  
        self.pan_routes = nn.LayerList(pan_routes[::-1])  
  
    def forward(self, blocks, for\_mot=False):  
        blocks = blocks[::-1]  
        fpn_feats = []  
  
        for i, block in enumerate(blocks):  
            if i > 0:  
                block = paddle.concat([route, block], axis=1)  
            route = self.fpn_stages[i](block)  
            fpn_feats.append(route)  
  
            if i < self.num_blocks - 1:  
                route = self.fpn_routes[i](route)  
                route = F.interpolate(route, scale_factor=2., data_format=self.data_format)  
  
        pan_feats = [fpn_feats[-1], ]  
        route = fpn_feats[-1]  
        for i in reversed(range(self.num_blocks - 1)):  
            block = fpn_feats[i]  
            route = self.pan_routes[i](route)  
            block = paddle.concat([route, block], axis=1)  
            route = self.pan_stages[i](block)  
            pan_feats.append(route)  
  
        return pan_feats[::-1]  

      

1.3、Head

picture.image

对于PP-YOLOE的head部分,其依旧是TOOD的head,也就是T-Head,主要是包括了Cls Head和Loc Head。具体来说,T-head首先在FPN特征基础上进行分类与定位预测;然后TAL基于所提任务对齐测度计算任务对齐信息;最后T-head根据从TAL传回的信息自动调整分类概率与定位预测。

picture.image

由于2个任务的预测都是基于这个交互特征来完成的,但是2个任务对于特征的需求肯定是不一样的,因为作者设计了一个layer attention来为每个任务单独的调整一下特征,这个部分的结构也很简单,可以理解为是一个channel-wise的注意力机制。这样的话就得到了对于每个任务单独的特征,然后再利用这些特征生成所需要的类别或者定位的特征图。

picture.image

picture.image


        
          
class PPYOLOEHead(nn.Layer):  
    __shared__ = ['num\_classes', 'trt', 'exclude\_nms']  
    __inject__ = ['static\_assigner', 'assigner', 'nms']  
  
    def \_\_init\_\_(self,  
                 in\_channels=[1024, 512, 256],  
                 num\_classes=80,  
                 act='swish',  
                 fpn\_strides=(32, 16, 8),  
                 grid\_cell\_scale=5.0,  
                 grid\_cell\_offset=0.5,  
                 reg\_max=16,  
                 static\_assigner\_epoch=4,  
                 use\_varifocal\_loss=True,  
                 static\_assigner='ATSSAssigner',  
                 assigner='TaskAlignedAssigner',  
                 nms='MultiClassNMS',  
                 eval\_input\_size=[],  
                 loss\_weight={'class': 1.0, 'iou': 2.5, 'dfl': 0.5,},  
                 trt=False,  
                 exclude\_nms=False):  
        super(PPYOLOEHead, self).__init__()  
        assert len(in_channels) > 0, "len(in\_channels) should > 0"  
        self.in_channels = in_channels  
        self.num_classes = num_classes  
        self.fpn_strides = fpn_strides  
        self.grid_cell_scale = grid_cell_scale  
        self.grid_cell_offset = grid_cell_offset  
        self.reg_max = reg_max  
        self.iou_loss = GIoULoss()  
        self.loss_weight = loss_weight  
        self.use_varifocal_loss = use_varifocal_loss  
        self.eval_input_size = eval_input_size  
  
        self.static_assigner_epoch = static_assigner_epoch  
        self.static_assigner = static_assigner  
        self.assigner = assigner  
        self.nms = nms  
        self.exclude_nms = exclude_nms  
        # stem  
        self.stem_cls = nn.LayerList()  
        self.stem_reg = nn.LayerList()  
        act = get_act_fn(act, trt=trt) if act is None or isinstance(act, (str, dict)) else act  
        for in_c in self.in_channels:  
            self.stem_cls.append(ESEAttn(in_c, act=act))  
            self.stem_reg.append(ESEAttn(in_c, act=act))  
        # pred head  
        self.pred_cls = nn.LayerList()  
        self.pred_reg = nn.LayerList()  
        for in_c in self.in_channels:  
            self.pred_cls.append(nn.Conv2D(in_c, self.num_classes, 3, padding=1))  
            self.pred_reg.append(nn.Conv2D(in_c, 4 * (self.reg_max + 1), 3, padding=1))  
        # projection conv  
        self.proj_conv = nn.Conv2D(self.reg_max + 1, 1, 1, bias_attr=False)  
        self._init_weights()  
  
    @classmethod  
    def from\_config(cls, cfg, input\_shape):  
        return {'in\_channels': [i.channels for i in input_shape], }  
  
    def forward\_train(self, feats, targets):  
        anchors, anchor_points, num_anchors_list, stride_tensor = generate_anchors_for_grid_cell(feats, self.fpn_strides, self.grid_cell_scale, self.grid_cell_offset)  
  
        cls_score_list, reg_distri_list = [], []  
        for i, feat in enumerate(feats):  
            avg_feat = F.adaptive_avg_pool2d(feat, (1, 1))  
            cls_logit = self.pred_cls[i](self.stem_cls[i](feat, avg_feat) + feat)  
            reg_distri = self.pred_reg[i](self.stem_reg[i](feat, avg_feat))  
            # cls and reg  
            cls_score = F.sigmoid(cls_logit)  
            cls_score_list.append(cls_score.flatten(2).transpose([0, 2, 1]))  
            reg_distri_list.append(reg_distri.flatten(2).transpose([0, 2, 1]))  
        cls_score_list = paddle.concat(cls_score_list, axis=1)  
        reg_distri_list = paddle.concat(reg_distri_list, axis=1)  
  
        return self.get_loss([cls_score_list, reg_distri_list, anchors, anchor_points, num_anchors_list, stride_tensor], targets)  
  
    def forward\_eval(self, feats):  
        if self.eval_input_size:  
            anchor_points, stride_tensor = self.anchor_points, self.stride_tensor  
        else:  
            anchor_points, stride_tensor = self._generate_anchors(feats)  
        cls_score_list, reg_dist_list = [], []  
        for i, feat in enumerate(feats):  
            b, _, h, w = feat.shape  
            l = h * w  
            avg_feat = F.adaptive_avg_pool2d(feat, (1, 1))  
            cls_logit = self.pred_cls[i](self.stem_cls[i](feat, avg_feat) + feat)  
            reg_dist = self.pred_reg[i](self.stem_reg[i](feat, avg_feat))  
            reg_dist = reg_dist.reshape([-1, 4, self.reg_max + 1, l]).transpose([0, 2, 1, 3])  
            reg_dist = self.proj_conv(F.softmax(reg_dist, axis=1))  
            # cls and reg  
            cls_score = F.sigmoid(cls_logit)  
            cls_score_list.append(cls_score.reshape([b, self.num_classes, l]))  
            reg_dist_list.append(reg_dist.reshape([b, 4, l]))  
  
        cls_score_list = paddle.concat(cls_score_list, axis=-1)  
        reg_dist_list = paddle.concat(reg_dist_list, axis=-1)  
  
        return cls_score_list, reg_dist_list, anchor_points, stride_tensor  
  
    def forward(self, feats, targets=None):  
        assert len(feats) == len(self.fpn_strides), "The size of feats is not equal to size of fpn\_strides"  
  
        if self.training:  
            return self.forward_train(feats, targets)  
        else:  
            return self.forward_eval(feats)  

      

2

样本匹配

2.1 ATSS Assigner思想

ATSS论文指出One-Stage Anchor-Based和Center-Based Anchor-Free检测算法间的差异主要来自于正负样本的选择,基于此提出ATSS(Adaptive Training Sample Selection)方法,该方法能够自动根据GT的相关统计特征选择合适的Anchor Box作为正样本,在不带来额外计算量和参数的情况下,能够大幅提升模型的性能。

ATSS选取正样本的方法如下:其简要流程为:

  1. 计算每个 gt bbox 和多尺度输出层的所有 anchor 之间的 IoU
  2. 计算每个 gt bbox 中心坐标和多尺度输出层的所有 anchor 中心坐标的 l2 距离
  3. 遍历每个输出层,遍历每个 gt bbox,找出当前层中 topk (超参,默认是 9 )个最小 l2 距离的 anchor 。假设一共有 l 个输出层,那么对于任何一个 gt bbox,都会挑选出 topk×l 个候选位置
  4. 对于每个 gt bbox,计算所有候选位置 IoU 的均值和标准差,两者相加得到该 gt bbox 的自适应阈值
  5. 遍历每个 gt bbox,选择出候选位置中 IoU 大于阈值的位置,该位置认为是正样本,负责预测该 gt bbox
  6. 如果 topk 参数设置过大,可能会导致某些正样本位置不在 gt bbox 内部,故需要过滤掉这部分正样本,设置为背景样本

1、ATSS主要2大特性:

  1. 保证了所有的正样本Anchor都是在Ground Truth的周围。
  2. 最主要是根据不同层的特性对不同层的正样本的阈值进行了微调。

2、ATSS的贡献

  • 指出Anchor-Base检测器和Anchor-Free检测器之间的本质区别实际上是如何定义正训练样本和负训练样本;
  • 提出自适应训练样本选择,以根据目标的统计特征自动选择正负样本;
  • 证明了在图像上的每个位置上平铺多个Anchor来提升检测的性能是没效果的;

        
          
class ATSSAssigner(nn.Layer):  
    """Bridging the Gap Between Anchor-based and Anchor-free Detection  
     via Adaptive Training Sample Selection  
    """  
    __shared__ = ['num\_classes']  
  
    def \_\_init\_\_(self, topk=9, num\_classes=80, force\_gt\_matching=False, eps=1e-9):  
        super(ATSSAssigner, self).__init__()  
        self.topk = topk  
        self.num_classes = num_classes  
        self.force_gt_matching = force_gt_matching  
        self.eps = eps  
  
    def \_gather\_topk\_pyramid(self, gt2anchor\_distances, num\_anchors\_list, pad\_gt\_mask):  
        pad_gt_mask = pad_gt_mask.tile([1, 1, self.topk]).astype(paddle.bool)  
        gt2anchor_distances_list = paddle.split(gt2anchor_distances, num_anchors_list, axis=-1)  
        num_anchors_index = np.cumsum(num_anchors_list).tolist()  
        num_anchors_index = [0, ] + num_anchors_index[:-1]  
        is_in_topk_list = []  
        topk_idxs_list = []  
        for distances, anchors_index in zip(gt2anchor_distances_list, num_anchors_index):  
            num_anchors = distances.shape[-1]  
            topk_metrics, topk_idxs = paddle.topk(distances, self.topk, axis=-1, largest=False)  
            topk_idxs_list.append(topk_idxs + anchors_index)  
            topk_idxs = paddle.where(pad_gt_mask, topk_idxs, paddle.zeros_like(topk_idxs))  
            is_in_topk = F.one_hot(topk_idxs, num_anchors).sum(axis=-2)  
            is_in_topk = paddle.where(is_in_topk > 1, paddle.zeros_like(is_in_topk), is_in_topk)  
            is_in_topk_list.append(is_in_topk.astype(gt2anchor_distances.dtype))  
        is_in_topk_list = paddle.concat(is_in_topk_list, axis=-1)  
        topk_idxs_list = paddle.concat(topk_idxs_list, axis=-1)  
        return is_in_topk_list, topk_idxs_list  
  
    @paddle.no\_grad()  
    def forward(self, anchor\_bboxes, num\_anchors\_list, gt\_labels, gt\_bboxes, pad\_gt\_mask, bg\_index, gt\_scores=None, pred\_bboxes=None):  
        """  
        ATSS匹配步骤如下:  
        1. 计算所有预测bbox与GT之间的IoU  
        2. 计算所有预测bbox与GT之间的距离  
        3. 在每个pyramid level上,对于每个gt,选择k个中心距离gt中心最近的bbox,总共选择k*l个bbox作为每个gt的候选框  
        4. 获取这些候选框对应的iou,计算mean和std,设 mean + std为 iou 阈值  
        5. 选择iou大于或等于阈值的样本为正样本  
        6. 将正样本的中心限制在gt内  
        7. 如果Anchor框被分配到多个gts,则选择具有最高的IoU的那个。  
        Args:  
            anchor\_bboxes (Tensor, float32): pre-defined anchors, shape(L, 4),  
                    "xmin, xmax, ymin, ymax" format  
            num\_anchors\_list (List): num of anchors in each level  
            gt\_labels (Tensor, int64|int32): Label of gt\_bboxes, shape(B, n, 1)  
            gt\_bboxes (Tensor, float32): Ground truth bboxes, shape(B, n, 4)  
            pad\_gt\_mask (Tensor, float32): 1 means bbox, 0 means no bbox, shape(B, n, 1)  
            bg\_index (int): background index  
            gt\_scores (Tensor|None, float32) Score of gt\_bboxes,  
                    shape(B, n, 1), if None, then it will initialize with one\_hot label  
            pred\_bboxes (Tensor, float32, optional): predicted bounding boxes, shape(B, L, 4)  
        Returns:  
            assigned\_labels (Tensor): (B, L)  
            assigned\_bboxes (Tensor): (B, L, 4)  
            assigned\_scores (Tensor): (B, L, C), if pred\_bboxes is not None, then output ious  
        """  
        assert gt_labels.ndim == gt_bboxes.ndim and gt_bboxes.ndim == 3  
  
        num_anchors, _ = anchor_bboxes.shape  
        batch_size, num_max_boxes, _ = gt_bboxes.shape  
  
        # 1. 计算所有预测bbox与GT之间的IoU, [B, n, L]  
        ious = iou_similarity(gt_bboxes.reshape([-1, 4]), anchor_bboxes)  
        ious = ious.reshape([batch_size, -1, num_anchors])  
  
        # 2. 计算所有预测bbox与GT之间的距离, [B, n, L]  
        gt_centers = bbox_center(gt_bboxes.reshape([-1, 4])).unsqueeze(1)  
        anchor_centers = bbox_center(anchor_bboxes)  
        gt2anchor_distances = (gt_centers - anchor_centers.unsqueeze(0)).norm(2, axis=-1).reshape([batch_size, -1, num_anchors])  
  
        # 3. 在每个pyramid level上,对于每个gt,选择k个中心距离gt中心最近的bbox,总共选择k*l个bbox作为每个gt的候选框  
        # based on the center distance, [B, n, L]  
        is_in_topk, topk_idxs = self._gather_topk_pyramid(gt2anchor_distances, num_anchors_list, pad_gt_mask)  
  
        # 4. 获取这些候选框对应的iou,计算mean和std,设 mean + std为 iou 阈值  
        iou_candidates = ious * is_in_topk  
        iou_threshold = paddle.index_sample(iou_candidates.flatten(stop_axis=-2), topk_idxs.flatten(stop_axis=-2))  
        iou_threshold = iou_threshold.reshape([batch_size, num_max_boxes, -1])  
        iou_threshold = iou_threshold.mean(axis=-1, keepdim=True) + iou_threshold.std(axis=-1, keepdim=True)  
        is_in_topk = paddle.where(iou_candidates > iou_threshold.tile([1, 1, num_anchors]), is_in_topk, paddle.zeros_like(is_in_topk))  
  
        # 6. 将正样本的中心限制在gt内, [B, n, L]  
        is_in_gts = check_points_inside_bboxes(anchor_centers, gt_bboxes)  
  
        # 选择正样本, [B, n, L]  
        mask_positive = is_in_topk * is_in_gts * pad_gt_mask  
  
        # 7. 如果Anchor框被分配到多个gts,则选择具有最高的IoU的那个。  
        mask_positive_sum = mask_positive.sum(axis=-2)  
        if mask_positive_sum.max() > 1:  
            mask_multiple_gts = (mask_positive_sum.unsqueeze(1) > 1).tile([1, num_max_boxes, 1])  
            is_max_iou = compute_max_iou_anchor(ious)  
            mask_positive = paddle.where(mask_multiple_gts, is_max_iou, mask_positive)  
            mask_positive_sum = mask_positive.sum(axis=-2)  
        # 8. 确认每个gt\_bbox 都匹配到了 anchor  
        if self.force_gt_matching:  
            is_max_iou = compute_max_iou_gt(ious) * pad_gt_mask  
            mask_max_iou = (is_max_iou.sum(-2, keepdim=True) == 1).tile([1, num_max_boxes, 1])  
            mask_positive = paddle.where(mask_max_iou, is_max_iou, mask_positive)  
            mask_positive_sum = mask_positive.sum(axis=-2)  
        assigned_gt_index = mask_positive.argmax(axis=-2)  
  
        # 匹配目标  
        batch_ind = paddle.arange(end=batch_size, dtype=gt_labels.dtype).unsqueeze(-1)  
        assigned_gt_index = assigned_gt_index + batch_ind * num_max_boxes  
        assigned_labels = paddle.gather(gt_labels.flatten(), assigned_gt_index.flatten(), axis=0)  
        assigned_labels = assigned_labels.reshape([batch_size, num_anchors])  
        assigned_labels = paddle.where(mask_positive_sum > 0, assigned_labels, paddle.full_like(assigned_labels, bg_index))  
  
        assigned_bboxes = paddle.gather(gt_bboxes.reshape([-1, 4]), assigned_gt_index.flatten(), axis=0)  
        assigned_bboxes = assigned_bboxes.reshape([batch_size, num_anchors, 4])  
  
        assigned_scores = F.one_hot(assigned_labels, self.num_classes)  
        if pred_bboxes is not None:  
            # assigned iou  
            ious = batch_iou_similarity(gt_bboxes, pred_bboxes) * mask_positive  
            ious = ious.max(axis=-2).unsqueeze(-1)  
            assigned_scores *= ious  
        elif gt_scores is not None:  
            gather_scores = paddle.gather(gt_scores.flatten(), assigned_gt_index.flatten(), axis=0)  
            gather_scores = gather_scores.reshape([batch_size, num_anchors])  
            gather_scores = paddle.where(mask_positive_sum > 0, gather_scores, paddle.zeros_like(gather_scores))  
            assigned_scores *= gather_scores.unsqueeze(-1)  
  
        return assigned_labels, assigned_bboxes, assigned_scores  

      

2.2、Task-aligned Assigner思想(TOOD)

TOOD提出了Task Alignment Learning (TAL) 来显式的把2个任务的最优Anchor拉近。这是通过设计一个样本分配策略和任务对齐loss来实现的。样本分配计算每个Anchor的任务对齐度,同时任务对齐loss可以逐步将分类和定位的最佳Anchor统一起来。

picture.image

类似于近期提出的One-Stage检测器,所提TOOD采用了类似的架构:Backbone-FPN-Head。考虑到效率与简单性,类似ATSS, TOOD在每个位置放置一个Anchor。

正如所讨论的,由于分类与定位任务的发散性,现有One-Stage检测器存在任务不对齐(Task Mis-Alignment)约束问题。本文提出通过显式方式采用T-head+TAL对2个任务进行对齐,见上图。T-headTAL通过协同工作方式改善2个任务的对齐问题;

TOOD选取样本的方法具体来说:

  • 首先,T-head在FPN特征基础上进行分类与定位预测;
  • 然后,TAL基于所提任务对齐测度计算任务对齐信息;
  • 最后,T-head根据从TAL传回的信息自动调整分类概率与定位预测。

1、Task-Aligned Head

本文所提T-Head见下图,它具有非常简单的结构:特征提取器+TAP

为增强分类与定位之间的相互作用,作者通过特征提取器学习任务交互(Task-Interactive)特征,如中蓝色框部分。这种设计不仅有助于任务交互,同时可以为2个任务提供多级多尺度特征。

picture.image

假设X表示FPN特征,特征提取器采用N个连续卷积计算任务交互特征:

picture.image

因此,通过特征提取器可以得到丰富的多尺度特征并用于送入到后续2个TAP模块中进行分类与定位对齐。

2、Task-Aligned Sample Assignment

为与NMS搭配,训练样例的Anchor分配需要满足以下规则:

  • 正常对齐的Anchor应当可以预测高分类得分,同时具有精确定位;
  • 不对齐的Anchor应当具有低分类得分,并在NMS阶段被抑制。

基于上述两个规则,作者设计了一种新的Anchor对齐度量以显式度量Anchor层面的对齐度 。该对齐度量将集成到样本分配与损失函数中以动态提炼每个Anchor的预测。

Anchor Alignment metric 考虑到分类得分与IoU表征了预测质量,我们采用2者的高阶组合度量任务对齐度,公式定义如下:

其中,s与u分别表示分类得分与IoU值,而用于控制两者的影响。因此,t在联合优化中起着非常重要的作用,它激励网络动态的聚焦于高质量的Anchor上

Training sample assignment 正如已有研究表明,训练样例分配对于检测器的训练非常重要

为提升两个任务的对齐性,TOOD聚焦于任务对齐Anchor,采用一种简单的分配规则选择训练样本:对每个实例,选择m个具有最大t值的Anchor作为正样例,选择其余的Anchor作为负样例。然后,通过新的损失函数(针对分类与定位的对齐而设计的损失函数)任务进行训练。


        
          
class TaskAlignedAssigner(nn.Layer):  
    def \_\_init\_\_(self, topk=13, alpha=1.0, beta=6.0, eps=1e-9):  
        super(TaskAlignedAssigner, self).__init__()  
        self.topk = topk  
        self.alpha = alpha  
        self.beta = beta  
        self.eps = eps  
  
    @paddle.no\_grad()  
    def forward(self, pred\_scores, pred\_bboxes, anchor\_points, num\_anchors\_list, gt\_labels, gt\_bboxes, pad\_gt\_mask, bg\_index, gt\_scores=None):  
        """  
        Task-Aligned Assigner计算步骤如下:  
        1. 计算所有 bbox与 gt 之间的对齐度  
        2. 选择 top-k bbox 作为每个 gt 的候选项  
        3. 将正样品的中心限制在 gt 内(因为Anchor-Free检测器只能预测大于0的距离)  
        4. 如果一个Anchor被分配给多个gt,将选择IoU最高的那个。  
        Args:  
            pred\_scores (Tensor, float32): predicted class probability, shape(B, L, C)  
            pred\_bboxes (Tensor, float32): predicted bounding boxes, shape(B, L, 4)  
            anchor\_points (Tensor, float32): pre-defined anchors, shape(L, 2), "cxcy" format  
            num\_anchors\_list (List): num of anchors in each level, shape(L)  
            gt\_labels (Tensor, int64|int32): Label of gt\_bboxes, shape(B, n, 1)  
            gt\_bboxes (Tensor, float32): Ground truth bboxes, shape(B, n, 4)  
            pad\_gt\_mask (Tensor, float32): 1 means bbox, 0 means no bbox, shape(B, n, 1)  
            bg\_index (int): background index  
            gt\_scores (Tensor|None, float32) Score of gt\_bboxes, shape(B, n, 1)  
        Returns:  
            assigned\_labels (Tensor): (B, L)  
            assigned\_bboxes (Tensor): (B, L, 4)  
            assigned\_scores (Tensor): (B, L, C)  
        """  
        assert pred_scores.ndim == pred_bboxes.ndim  
        assert gt_labels.ndim == gt_bboxes.ndim and gt_bboxes.ndim == 3  
  
        batch_size, num_anchors, num_classes = pred_scores.shape  
        _, num_max_boxes, _ = gt_bboxes.shape  
  
        # 计算GT与预测box之间的iou, [B, n, L]  
        ious = iou_similarity(gt_bboxes, pred_bboxes)  
        # 获取预测bboxes class score  
        pred_scores = pred_scores.transpose([0, 2, 1])  
        batch_ind = paddle.arange(end=batch_size, dtype=gt_labels.dtype).unsqueeze(-1)  
        gt_labels_ind = paddle.stack([batch_ind.tile([1, num_max_boxes]), gt_labels.squeeze(-1)], axis=-1)  
        bbox_cls_scores = paddle.gather_nd(pred_scores, gt_labels_ind)  
        # 计算bbox与 gt 之间的对齐度, [B, n, L]  
        alignment_metrics = bbox_cls_scores.pow(self.alpha) * ious.pow(self.beta)  
  
        # check the positive sample's center in gt, [B, n, L]  
        is_in_gts = check_points_inside_bboxes(anchor_points, gt_bboxes)  
  
        # 选择 top-k 预测 bbox 作为每个 gt 的候选项  
        is_in_topk = gather_topk_anchors(alignment_metrics * is_in_gts, self.topk, topk_mask=pad_gt_mask.tile([1, 1, self.topk]).astype(paddle.bool))  
  
        # select positive sample, [B, n, L]  
        mask_positive = is_in_topk * is_in_gts * pad_gt_mask  
  
        # 如果一个Anchor被分配给多个gt,将选择IoU最高的那个, [B, n, L]  
        mask_positive_sum = mask_positive.sum(axis=-2)  
        if mask_positive_sum.max() > 1:  
            mask_multiple_gts = (mask_positive_sum.unsqueeze(1) > 1).tile([1, num_max_boxes, 1])  
            is_max_iou = compute_max_iou_anchor(ious)  
            mask_positive = paddle.where(mask_multiple_gts, is_max_iou, mask_positive)  
            mask_positive_sum = mask_positive.sum(axis=-2)  
        assigned_gt_index = mask_positive.argmax(axis=-2)  
  
        # assigned target  
        assigned_gt_index = assigned_gt_index + batch_ind * num_max_boxes  
        assigned_labels = paddle.gather(gt_labels.flatten(), assigned_gt_index.flatten(), axis=0)  
        assigned_labels = assigned_labels.reshape([batch_size, num_anchors])  
        assigned_labels = paddle.where(mask_positive_sum > 0, assigned_labels, paddle.full_like(assigned_labels, bg_index))  
  
        assigned_bboxes = paddle.gather(gt_bboxes.reshape([-1, 4]), assigned_gt_index.flatten(), axis=0)  
        assigned_bboxes = assigned_bboxes.reshape([batch_size, num_anchors, 4])  
  
        assigned_scores = F.one_hot(assigned_labels, num_classes)  
        # rescale alignment metrics  
        alignment_metrics *= mask_positive  
        max_metrics_per_instance = alignment_metrics.max(axis=-1, keepdim=True)  
        max_ious_per_instance = (ious * mask_positive).max(axis=-1, keepdim=True)  
        alignment_metrics = alignment_metrics / (max_metrics_per_instance + self.eps) * max_ious_per_instance  
        alignment_metrics = alignment_metrics.max(-2).unsqueeze(-1)  
        assigned_scores = assigned_scores * alignment_metrics  
  
        return assigned_labels, assigned_bboxes, assigned_scores  

      

3

损失函数

3.1、分类损失varifocal loss

Focal loss定义:

picture.image

其中a是前景背景的损失权重,p的y次是不同样本的权重,难分样本的损失权重会增大。当训练一个密集的物体检测器使连续的IACS回归时,本文从focal loss中借鉴了样本加权思想来解决类不平衡问题。但是,与focal loss同等对待正负样本的损失不同,而varifocal loss选择不对称地对待它们。varifocal loss定义如下:

picture.image

其中p是预测的IACS得分,q是目标IoU分数。对于训练中的正样本,将q设置为生成的bbox和gt box之间的IoU(gt IoU),而对于训练中的负样本,所有类别的训练目标q均为0。

备注:Varifocal Loss会预测Iou-aware Cls_score(IACS)与分类两个得分,通过p的y次来有效降低负样本损失的权重,正样本选择不降低权重。此外,通过q(Iou感知得分)来对Iou高的正样本损失加大权重,相当于将训练重点放在高质量的样本上面。


        
          
@staticmethod  
def \_varifocal\_loss(pred\_score, gt\_score, label, alpha=0.75, gamma=2.0):  
    weight = alpha * pred_score.pow(gamma) * (1 - label) + gt_score * label  
    loss = F.binary_cross_entropy(pred_score, gt_score, weight=weight, reduction='sum')  
    return loss  

      

3.2、回归损失

1、GIoULoss

GIOU的计算很简单,对于两个bounding box A,B。我们可以算出其最小凸集(包围A、B的最小包围框)C。有了最小凸集,就可以计算GIOU:

picture.image

计算方法很简单,从公式可以看出,GIOU有几个特点:

  • GIOU是IOU的下界,且取值范围为(-1, 1]。当两个框不重合时,IOU始终为0,不论A、B相隔多远,但是对于GIOU来说,A,B不重合度越高(离的越远),GIOU越趋近于-1。关于这点,下面再详细解释一下。
  • GIOU其实就是在IOU的基础上减掉了一个东西,这个减掉的东西,让我门避免了两个bbox不重合时Loss为0的情况。至于减掉的东西怎么去直观的理解,似乎不好找到一个很好的解释?
  • 可导:这一点需要强调下,由于max,min,分段函数(比如ReLU)这些都是可导的,所以用1-GIOU作为Loss是可导的。

当IOU=0时:

picture.image

显然, A∪B值不变,最大化GIOU就是要最小化C,最小化C就会促成2个框不断靠近,而不是像最小化IOU那样loss为0。

YOLO V3涨了2个点,Faster RCNN,MaskRCNN这种涨点少了些。主要原因在于Faster RCNN,MaskRCNN本身的Anchor很多,出现完全无重合的情况比较少,这样GIOU和IOU Loss就无明显差别。所以提升不是太明显。

在TOOD中,bbox(Bouding box)通过对齐的anchor(具有更大的分类得分、更精确的定位)预测得到,这样的bbox通常经过NMS后仍可以得以保留。此外,t可以在训练阶段通过对损失加权选择高质量的bbox。因此,采用t度量bbox的质量,同时结合GIoU Loss定义了TOOD的Reg Loss如下:

picture.image


        
          
@register  
@serializable  
class GIoULoss(object):  
    """  
    Generalized Intersection over Union, see https://arxiv.org/abs/1902.09630  
    Args:  
        loss\_weight (float): giou loss weight, default as 1  
        eps (float): epsilon to avoid divide by zero, default as 1e-10  
        reduction (string): Options are "none", "mean" and "sum". default as none  
    """  
  
    def \_\_init\_\_(self, loss\_weight=1., eps=1e-10, reduction='none'):  
        self.loss_weight = loss_weight  
        self.eps = eps  
        assert reduction in ('none', 'mean', 'sum')  
        self.reduction = reduction  
  
    def bbox\_overlap(self, box1, box2, eps=1e-10):  
        """calculate the iou of box1 and box2  
        Args:  
            box1 (Tensor): box1 with the shape (..., 4)  
            box2 (Tensor): box1 with the shape (..., 4)  
            eps (float): epsilon to avoid divide by zero  
        Return:  
            iou (Tensor): iou of box1 and box2  
            overlap (Tensor): overlap of box1 and box2  
            union (Tensor): union of box1 and box2  
        """  
        x1, y1, x2, y2 = box1  
        x1g, y1g, x2g, y2g = box2  
  
        xkis1 = paddle.maximum(x1, x1g)  
        ykis1 = paddle.maximum(y1, y1g)  
        xkis2 = paddle.minimum(x2, x2g)  
        ykis2 = paddle.minimum(y2, y2g)  
        w_inter = (xkis2 - xkis1).clip(0)  
        h_inter = (ykis2 - ykis1).clip(0)  
        overlap = w_inter * h_inter  
  
        area1 = (x2 - x1) * (y2 - y1)  
        area2 = (x2g - x1g) * (y2g - y1g)  
        union = area1 + area2 - overlap + eps  
        iou = overlap / union  
  
        return iou, overlap, union  
  
    def \_\_call\_\_(self, pbox, gbox, iou\_weight=1., loc\_reweight=None):  
        x1, y1, x2, y2 = paddle.split(pbox, num_or_sections=4, axis=-1)  
        x1g, y1g, x2g, y2g = paddle.split(gbox, num_or_sections=4, axis=-1)  
        box1 = [x1, y1, x2, y2]  
        box2 = [x1g, y1g, x2g, y2g]  
        iou, overlap, union = self.bbox_overlap(box1, box2, self.eps)  
        xc1 = paddle.minimum(x1, x1g)  
        yc1 = paddle.minimum(y1, y1g)  
        xc2 = paddle.maximum(x2, x2g)  
        yc2 = paddle.maximum(y2, y2g)  
  
        area_c = (xc2 - xc1) * (yc2 - yc1) + self.eps  
        miou = iou - ((area_c - union) / area_c)  
        if loc_reweight is not None:  
            loc_reweight = paddle.reshape(loc_reweight, shape=(-1, 1))  
            loc_thresh = 0.9  
            giou = 1 - (1 - loc_thresh) * miou - loc_thresh * miou * loc_reweight  
        else:  
            giou = 1 - miou  
        if self.reduction == 'none':  
            loss = giou  
        elif self.reduction == 'sum':  
            loss = paddle.sum(giou * iou_weight)  
        else:  
            loss = paddle.mean(giou * iou_weight)  
        return loss * self.loss_weight  

      

2、L1 loss

均绝对误差(Mean Absolute Error,MAE) 是指模型预测值f(x)和真实值y之间距离的均值,其公式如下:

picture.image

忽略下标i ,设n=1,以f(x)−y为横轴,MAE的值为纵轴,得到函数的图形如下:

picture.image

MAE曲线连续,但是在y−f(x)=0处不可导。而且 MAE 大部分情况下梯度都是相等的,这意味着即使对于小的损失值,其梯度也是大的。这不利于函数的收敛和模型的学习。但是,无论对于什么样的输入值,都有着稳定的梯度,不会导致梯度爆炸问题,具有较为稳健性的解。

相比于MSE,MAE有个优点就是,对于离群点不那么敏感。因为MAE计算的是误差y−f(x)的绝对值,对于任意大小的差值,其惩罚都是固定的。


        
          
loss_l1 = F.l1_loss(pred_bboxes_pos, assigned_bboxes_pos)  

      

3、DF Loss

对于任意分布来建模框的表示,它可以用积分形式嵌入到任意已有的和框回归相关的损失函数上,例如最近比较流行的GIoU Loss。这个实际上也就够了,不过涨点不是很明显,我们又仔细分析了一下,发现如果分布过于任意,网络学习的效率可能会不高,原因是一个积分目标可能对应了无穷多种分布模式。如下图所示:

picture.image

考虑到真实的分布通常不会距离标注的位置太远,所以我们又额外加了个loss,希望网络能够快速地聚焦到标注位置附近的数值,使得他们概率尽可能大。基于此,我们取了个名字叫Distribution Focal Loss (DFL):

picture.image

其形式上与QFL的右半部分很类似,含义是以类似交叉熵的形式去优化与标签y最接近的一左一右两个位置的概率,从而让网络快速地聚焦到目标位置的邻近区域的分布中去。

picture.image

QFL和DFL的作用是正交的,他们的增益互不影响


        
          
def \_df\_loss(self, pred\_dist, target):  
    target_left = paddle.cast(target, 'int64')  
    target_right = target_left + 1  
    weight_left = target_right.astype('float32') - target  
    weight_right = 1 - weight_left  
    loss_left = F.cross_entropy(pred_dist, target_left, reduction='none') * weight_left  
    loss_right = F.cross_entropy(pred_dist, target_right, reduction='none') * weight_right  
    return (loss_left + loss_right).mean(-1, keepdim=True)  

      

3.3、总损失

其中,a表示分类损失的权重系数,b表示回归损失的权重系数,c表示DFL损失的权重系数。


        
          
    def get\_loss(self, head\_outs, gt\_meta):  
        pred_scores, pred_distri, anchors, anchor_points, num_anchors_list, stride_tensor = head_outs  
  
        anchor_points_s = anchor_points / stride_tensor  
        pred_bboxes = self._bbox_decode(anchor_points_s, pred_distri)  
  
        gt_labels = gt_meta['gt\_class']  
        gt_bboxes = gt_meta['gt\_bbox']  
        pad_gt_mask = gt_meta['pad\_gt\_mask']  
        # Epoch小于100使用ATSS匹配  
        if gt_meta['epoch\_id'] < self.static_assigner_epoch:  
            assigned_labels, assigned_bboxes, assigned_scores = \  
                self.static_assigner(  
                    anchors,  
                    num_anchors_list,  
                    gt_labels,  
                    gt_bboxes,  
                    pad_gt_mask,  
                    bg_index=self.num_classes,  
                    pred_bboxes=pred_bboxes.detach() * stride_tensor)  
            alpha_l = 0.25  
        else:  
             # Epoch大于100使用TAL匹配  
            assigned_labels, assigned_bboxes, assigned_scores = \  
                self.assigner(  
                pred_scores.detach(),  
                pred_bboxes.detach() * stride_tensor,  
                anchor_points,  
                num_anchors_list,  
                gt_labels,  
                gt_bboxes,  
                pad_gt_mask,  
                bg_index=self.num_classes)  
            alpha_l = -1  
        # rescale bbox  
        assigned_bboxes /= stride_tensor  
        # cls loss  
        if self.use_varifocal_loss:  
            one_hot_label = F.one_hot(assigned_labels, self.num_classes)  
            loss_cls = self._varifocal_loss(pred_scores, assigned_scores, one_hot_label)  
        else:  
            loss_cls = self._focal_loss(pred_scores, assigned_scores, alpha_l)  
  
        assigned_scores_sum = assigned_scores.sum()  
        if paddle_distributed_is_initialized():  
            paddle.distributed.all_reduce(assigned_scores_sum)  
            assigned_scores_sum = paddle.clip(assigned_scores_sum / paddle.distributed.get_world_size(), min=1)  
        loss_cls /= assigned_scores_sum  
  
        loss_l1, loss_iou, loss_dfl = self._bbox_loss(pred_distri, pred_bboxes, anchor_points_s, assigned_labels, assigned_bboxes, assigned_scores, assigned_scores_sum)  
        loss = self.loss_weight['class'] * loss_cls + self.loss_weight['iou'] * loss_iou + self.loss_weight['dfl'] * loss_dfl  
        out_dict = {  
            'loss': loss,  
            'loss\_cls': loss_cls,  
            'loss\_iou': loss_iou,  
            'loss\_dfl': loss_dfl,  
            'loss\_l1': loss_l1,  
        }  
        return out_dict  

      

4

模型推理与部署

4.1、模型推理


        
          
# inference single image  
CUDA_VISIBLE_DEVICES=0 python tools/infer.py -c configs/ppyoloe/ppyoloe_crn_l_300e_coco.yml -o weights=https://paddledet.bj.bcebos.com/models/ppyoloe_crn_l_300e_coco.pdparams --infer_img=demo/000000014439_640x640.jpg  
  
# inference all images in the directory  
CUDA_VISIBLE_DEVICES=0 python tools/infer.py -c configs/ppyoloe/ppyoloe_crn_l_300e_coco.yml -o weights=https://paddledet.bj.bcebos.com/models/ppyoloe_crn_l_300e_coco.pdparams --infer_dir=demo  

      

picture.image

4.2、导出ONNX


        
          
# export inference model  
python tools/export_model.py configs/ppyoloe/ppyoloe_crn_l_300e_coco.yml --output_dir=output_inference -o weights=https://paddledet.bj.bcebos.com/models/ppyoloe_crn_l_300e_coco.pdparams  
  
# install paddle2onnx  
pip install paddle2onnx  
  
# convert to onnx  
paddle2onnx --model_dir output_inference/ppyoloe_crn_l_300e_coco --model_filename model.pdmodel --params_filename model.pdiparams --opset_version 11 --save_file ppyoloe_crn_l_300e_coco.onnx  
  

      

4。3、导出TensorRT Engine


        
          
python tools/export_model.py configs/ppyoloe/ppyoloe_crn_l_300e_coco.yml -o weights=https://paddledet.bj.bcebos.com/models/ppyoloe_crn_l_300e_coco.pdparams -o trt=True  

      

5

参考

[1].https://github.com/PaddlePaddle/PaddleDetection

6

推荐阅读

CVPR2022 | Anchor-Free之Label Assignment全新范式,目标检测经典之作!!!

YOLOv5-Face | 原理超精细讲解、训练步骤还原、C++边缘部署(就要这么学!!!)

YOLOv5-Lite 详解教程 | 嚼碎所有原理、训练自己数据集、TensorRT部署落地应有尽有

长按扫描下方二维码添加小助手并加入交流群, 群里博士大佬云集, 每日讨论话题有目标检测、语义分 割、 超分辨率、模型部署、数学基础知识、算法面试题分享的等等内容,当然也少不了搬砖人的扯犊子

长按扫描下方二维码添加小助手。

可以一起讨论遇到的问题

picture.image

声明:转载请说明出处

扫描下方二维码关注【 集智书童 】公众号,获取更多实践项目源码和论文解读,非常期待你我的相遇,让我们以梦为马,砥砺前行!

picture.image

picture.image

0
0
0
0
关于作者
关于作者

文章

0

获赞

0

收藏

0

相关资源
云原生数仓如何构建高性能向量检索技术
火山引擎ByteHouse团队基于社区 ClickHouse 进行技术演进,提出了全新的向量检索功能设计思路,满足业务对向量检索稳定性与性能方面的需求。
相关产品
评论
未登录
看完啦,登录分享一下感受吧~
暂无评论