YOLOv11改进 | Conv/卷积篇 | 利用2024最新YOLOv9的GELAN模块替换C3k2结构(附轻量化版本 + 高效涨点版本 + 结构图

一、本文介绍

本文给大家带来的改进机制是利用2024/02/21号发布的 YOLOv9其中提出的GELAN模块来改进YOLOv11中的C3k2 ,GELAN融合了CSPNet和ELAN机制同时其中利用到了RepConv在获取更多有效特征的同时在推理时专用单 分支结构 从而不影响推理速度,同时本文的内容提供了两种版本一种是参数量更低涨点效果略微弱一些的版本,另一种是参数量稍多一些但是效果要不参数量低的效果要好一些(均为我个人整理),提供两种版本是为了适配不同需求的读者,具体选择那种大家可以根据自身的需求来选择即可,文章中我都均已提供, 同时本文的结构存在大量的二次创新机会,后面我也会提供。


目录

一、本文介绍

二、GELAN的原理

2.1 Generalized ELAN

2.2 Generalized ELAN结构图

三、核心代码

四、手把手教你添加GELAN机制

4.1 修改一

4.2 修改二

4.3 修改三

4.4 修改四

五、GELAN的yaml文件和运行记录

5.1 GELAN低参数量版本的yaml文件

5.2  GELAN高参数量版本的yaml文件

5.3 训练代码

5.3 GELAN的训练过程截图

5.3.1 低参数量版本

5.3.2 高参数量版本

五、本文总结


二、GELAN的原理

2.1 Generalized ELAN

在本节中,我们描述了提出的新 网络架构 - GELAN。通过结合两种 神经网络 架构CSPNet和ELAN,这两种架构都是以梯度 路径规划 设计的,我们设计了考虑了轻量级、推理速度和准确性的广义高效层聚合网络(GELAN)。其整体架构如图4所示。我们推广了ELAN的能力,ELAN原本只使用卷积层的堆叠,到一个新的架构,可以使用任何计算块。

这张图(图4)展示了广义高效层聚合网络(GELAN)的架构,以及它是如何从CSPNet和ELAN这两种神经网络架构演变而来的。这两种架构都设计有梯度路径规划。

a) CSPNet: 在CSPNet的架构中,输入通过一个转换层被分割为两部分,然后分别通过任意的计算块。之后,这些分支被重新合并(通过concatenation),并再次通过转换层。

b) ELAN: 与CSPNet相比,ELAN采用了堆叠的卷积层,其中每一层的输出都会与下一层的输入相结合,再经过卷积处理。

c) GELAN: 结合了CSPNet和ELAN的设计,提出了GELAN。它采用了CSPNet的分割和重组的概念,并在每一部分引入了ELAN的层级卷积处理方式。不同之处在于GELAN不仅使用卷积层,还可以使用任何计算块,使得网络更加灵活,能够根据不同的应用需求定制。

GELAN的设计考虑到了轻量化、推理速度和精确度,以此来提高模型的整体性能。图中显示的模块和分区的可选性进一步增加了网络的适应性和可定制性。GELAN的这种结构允许它支持多种类型的计算块,这使得它可以更好地适应各种不同的计算需求和 硬件 约束。

总的来说,GELAN的架构是为了提供一个更加通用和高效的网络,可以适应从轻量级到复杂的 深度学习 任务,同时保持或增强计算效率和性能。通过这种方式,GELAN旨在解决现有架构的限制,提供一个可扩展的解决方案,以适应未来深度学习的发展。

大家看图片一眼就能看出来它融合了什么,就是将CSPHet的anyBlock模块堆叠的方式和ELAN融合到了一起。


2.2 Generalized ELAN结构图

YOLOv9最主要的创新目前能够得到的就是其中的 GELAN结构 我也是分析其代码根据论文将其结构图绘画出来。

下面的文件为YOLOv9的yaml文件。可以看到的是其提出了一种结构名字RepNCSPELAN4,其中的结构图concat后的通道数我没有画是因为它有计算中间的参数的变量是根据个人设置来的。

其代码和结构图如下所示!

  1. class RepNCSPELAN4(nn.Module):
  2. # csp-elan
  3. def __init__(self, c1, c2, c5=1): # c5 = repeat
  4. super().__init__()
  5. c3 = int(c2 / 2)
  6. c4 = int(c3 / 2)
  7. self.c = c3 // 2
  8. self.cv1 = Conv(c1, c3, 1, 1)
  9. self.cv2 = nn.Sequential(RepNCSP(c3 // 2, c4, c5), Conv(c4, c4, 3, 1))
  10. self.cv3 = nn.Sequential(RepNCSP(c4, c4, c5), Conv(c4, c4, 3, 1))
  11. self.cv4 = Conv(c3 + (2 * c4), c2, 1, 1)
  12. def forward(self, x):
  13. y = list(self.cv1(x).chunk(2, 1))
  14. y.extend((m(y[-1])) for m in [self.cv2, self.cv3])
  15. return self.cv4(torch.cat(y, 1))
  16. def forward_split(self, x):
  17. y = list(self.cv1(x).split((self.c, self.c), 1))
  18. y.extend(m(y[-1]) for m in [self.cv2, self.cv3])
  19. return self.cv4(torch.cat(y, 1))


三、核心代码

核心代码的使用方式看章节四!

  1. import torch
  2. import torch.nn as nn
  3. import numpy as np
  4. __all__ = ['RepNCSPELAN4_low', 'RepNCSPELAN4_high']
  5. class RepConvN(nn.Module):
  6. """RepConv is a basic rep-style block, including training and deploy status
  7. This code is based on https://github.com/DingXiaoH/RepVGG/blob/main/repvgg.py
  8. """
  9. default_act = nn.SiLU() # default activation
  10. def __init__(self, c1, c2, k=3, s=1, p=1, g=1, d=1, act=True, bn=False, deploy=False):
  11. super().__init__()
  12. assert k == 3 and p == 1
  13. self.g = g
  14. self.c1 = c1
  15. self.c2 = c2
  16. self.act = self.default_act if act is True else act if isinstance(act, nn.Module) else nn.Identity()
  17. self.bn = None
  18. self.conv1 = Conv(c1, c2, k, s, p=p, g=g, act=False)
  19. self.conv2 = Conv(c1, c2, 1, s, p=(p - k // 2), g=g, act=False)
  20. def forward_fuse(self, x):
  21. """Forward process"""
  22. return self.act(self.conv(x))
  23. def forward(self, x):
  24. """Forward process"""
  25. id_out = 0 if self.bn is None else self.bn(x)
  26. return self.act(self.conv1(x) + self.conv2(x) + id_out)
  27. def get_equivalent_kernel_bias(self):
  28. kernel3x3, bias3x3 = self._fuse_bn_tensor(self.conv1)
  29. kernel1x1, bias1x1 = self._fuse_bn_tensor(self.conv2)
  30. kernelid, biasid = self._fuse_bn_tensor(self.bn)
  31. return kernel3x3 + self._pad_1x1_to_3x3_tensor(kernel1x1) + kernelid, bias3x3 + bias1x1 + biasid
  32. def _avg_to_3x3_tensor(self, avgp):
  33. channels = self.c1
  34. groups = self.g
  35. kernel_size = avgp.kernel_size
  36. input_dim = channels // groups
  37. k = torch.zeros((channels, input_dim, kernel_size, kernel_size))
  38. k[np.arange(channels), np.tile(np.arange(input_dim), groups), :, :] = 1.0 / kernel_size ** 2
  39. return k
  40. def _pad_1x1_to_3x3_tensor(self, kernel1x1):
  41. if kernel1x1 is None:
  42. return 0
  43. else:
  44. return torch.nn.functional.pad(kernel1x1, [1, 1, 1, 1])
  45. def _fuse_bn_tensor(self, branch):
  46. if branch is None:
  47. return 0, 0
  48. if isinstance(branch, Conv):
  49. kernel = branch.conv.weight
  50. running_mean = branch.bn.running_mean
  51. running_var = branch.bn.running_var
  52. gamma = branch.bn.weight
  53. beta = branch.bn.bias
  54. eps = branch.bn.eps
  55. elif isinstance(branch, nn.BatchNorm2d):
  56. if not hasattr(self, 'id_tensor'):
  57. input_dim = self.c1 // self.g
  58. kernel_value = np.zeros((self.c1, input_dim, 3, 3), dtype=np.float32)
  59. for i in range(self.c1):
  60. kernel_value[i, i % input_dim, 1, 1] = 1
  61. self.id_tensor = torch.from_numpy(kernel_value).to(branch.weight.device)
  62. kernel = self.id_tensor
  63. running_mean = branch.running_mean
  64. running_var = branch.running_var
  65. gamma = branch.weight
  66. beta = branch.bias
  67. eps = branch.eps
  68. std = (running_var + eps).sqrt()
  69. t = (gamma / std).reshape(-1, 1, 1, 1)
  70. return kernel * t, beta - running_mean * gamma / std
  71. def fuse_convs(self):
  72. if hasattr(self, 'conv'):
  73. return
  74. kernel, bias = self.get_equivalent_kernel_bias()
  75. self.conv = nn.Conv2d(in_channels=self.conv1.conv.in_channels,
  76. out_channels=self.conv1.conv.out_channels,
  77. kernel_size=self.conv1.conv.kernel_size,
  78. stride=self.conv1.conv.stride,
  79. padding=self.conv1.conv.padding,
  80. dilation=self.conv1.conv.dilation,
  81. groups=self.conv1.conv.groups,
  82. bias=True).requires_grad_(False)
  83. self.conv.weight.data = kernel
  84. self.conv.bias.data = bias
  85. for para in self.parameters():
  86. para.detach_()
  87. self.__delattr__('conv1')
  88. self.__delattr__('conv2')
  89. if hasattr(self, 'nm'):
  90. self.__delattr__('nm')
  91. if hasattr(self, 'bn'):
  92. self.__delattr__('bn')
  93. if hasattr(self, 'id_tensor'):
  94. self.__delattr__('id_tensor')
  95. class RepNBottleneck(nn.Module):
  96. # Standard bottleneck
  97. def __init__(self, c1, c2, shortcut=True, g=1, k=(3, 3), e=0.5): # ch_in, ch_out, shortcut, kernels, groups, expand
  98. super().__init__()
  99. c_ = int(c2 * e) # hidden channels
  100. self.cv1 = RepConvN(c1, c_, k[0], 1)
  101. self.cv2 = Conv(c_, c2, k[1], 1, g=g)
  102. self.add = shortcut and c1 == c2
  103. def forward(self, x):
  104. return x + self.cv2(self.cv1(x)) if self.add else self.cv2(self.cv1(x))
  105. class RepNCSP(nn.Module):
  106. # CSP Bottleneck with 3 convolutions
  107. def __init__(self, c1, c2, n=1, shortcut=True, g=1, e=0.5): # ch_in, ch_out, number, shortcut, groups, expansion
  108. super().__init__()
  109. c_ = int(c2 * e) # hidden channels
  110. self.cv1 = Conv(c1, c_, 1, 1)
  111. self.cv2 = Conv(c1, c_, 1, 1)
  112. self.cv3 = Conv(2 * c_, c2, 1) # optional act=FReLU(c2)
  113. self.m = nn.Sequential(*(RepNBottleneck(c_, c_, shortcut, g, e=1.0) for _ in range(n)))
  114. def forward(self, x):
  115. return self.cv3(torch.cat((self.m(self.cv1(x)), self.cv2(x)), 1))
  116. def autopad(k, p=None, d=1): # kernel, padding, dilation
  117. # Pad to 'same' shape outputs
  118. if d > 1:
  119. k = d * (k - 1) + 1 if isinstance(k, int) else [d * (x - 1) + 1 for x in k] # actual kernel-size
  120. if p is None:
  121. p = k // 2 if isinstance(k, int) else [x // 2 for x in k] # auto-pad
  122. return p
  123. class Conv(nn.Module):
  124. # Standard convolution with args(ch_in, ch_out, kernel, stride, padding, groups, dilation, activation)
  125. default_act = nn.SiLU() # default activation
  126. def __init__(self, c1, c2, k=1, s=1, p=None, g=1, d=1, act=True):
  127. super().__init__()
  128. self.conv = nn.Conv2d(c1, c2, k, s, autopad(k, p, d), groups=g, dilation=d, bias=False)
  129. self.bn = nn.BatchNorm2d(c2)
  130. self.act = self.default_act if act is True else act if isinstance(act, nn.Module) else nn.Identity()
  131. def forward(self, x):
  132. return self.act(self.bn(self.conv(x)))
  133. def forward_fuse(self, x):
  134. return self.act(self.conv(x))
  135. class RepNCSPELAN4_low(nn.Module):
  136. # csp-elan
  137. def __init__(self, c1, c2, c5=1): # c5 = repeat
  138. super().__init__()
  139. c3 = int(c2 / 2)
  140. c4 = int(c3 / 2)
  141. self.c = c3 // 2
  142. self.cv1 = Conv(c1, c3, 1, 1)
  143. self.cv3 = nn.Sequential(RepNCSP(c3, c3, c5))
  144. self.cv4 = Conv(c3 + (2 * c4), c2, 1, 1)
  145. def forward(self, x):
  146. temp = self.cv1(x)
  147. temp3 = self.cv3(temp)
  148. y = list(temp.chunk(2, 1))
  149. y.append(temp3)
  150. temp2 = torch.cat(y, 1)
  151. return self.cv4(temp2)
  152. def forward_split(self, x):
  153. y = list(self.cv1(x).split((self.c, self.c), 1))
  154. y.extend(m(y[-1]) for m in [self.cv2, self.cv3])
  155. return self.cv4(torch.cat(y, 1))
  156. class RepNCSPELAN4_high(nn.Module):
  157. # csp-elan
  158. def __init__(self, c1, c2, c5=1): # c5 = repeat
  159. super().__init__()
  160. c3 = c2
  161. c4 = int(c3 / 2)
  162. self.c = c3 // 2
  163. self.cv1 = Conv(c1, c3, 1, 1)
  164. self.cv2 = nn.Sequential(RepNCSP(c3 // 2, c4, c5), Conv(c4, c4, 3, 1))
  165. self.cv3 = nn.Sequential(RepNCSP(c4, c4, c5), Conv(c4, c4, 3, 1))
  166. self.cv4 = Conv(c3 + (2 * c4), c2, 1, 1)
  167. def forward(self, x):
  168. y = list(self.cv1(x).chunk(2, 1))
  169. y.extend((m(y[-1])) for m in [self.cv2, self.cv3])
  170. return self.cv4(torch.cat(y, 1))
  171. def forward_split(self, x):
  172. y = list(self.cv1(x).split((self.c, self.c), 1))
  173. y.extend(m(y[-1]) for m in [self.cv2, self.cv3])
  174. return self.cv4(torch.cat(y, 1))
  175. if __name__ == "__main__":
  176. # Generating Sample image
  177. image_size = (1, 24, 224, 224)
  178. image = torch.rand(*image_size)
  179. # Model
  180. mobilenet_v1 = RepNCSPELAN4_low(24, 24)
  181. out = mobilenet_v1(image)
  182. print(out.size())


四、手把手教你添加GELAN机制

4.1 修改一

第一还是建立文件,我们找到如下 ultralytics /nn/modules文件夹下建立一个目录名字呢就是'Addmodules'文件夹 !然后在其内部建立一个新的py文件将核心代码复制粘贴进去即可。


4.2 修改二

第二步我们在该目录下创建一个新的py文件名字为'__init__.py' ,然后在其内部导入我们的检测头如下图所示。


4.3 修改三

第三步我门中到如下文件'ultralytics/nn/tasks.py'进行导入和注册我们的模块 !


4.4 修改四

按照我的添加在parse_model里添加即可。

到此就修改完成了,大家可以复制下面的yaml文件运行。


五、GELAN的yaml文件和运行记录

5.1 GELAN低参数量版本的yaml文件

此版本训练信息:YOLO11-RepGELAN-low summary: 403 layers, 2,218,027 parameters, 2,218,011 gradients, 6.3 GFLOPs

  1. # Ultralytics YOLO 🚀, AGPL-3.0 license
  2. # YOLO11 object detection model with P3-P5 outputs. For Usage examples see https://docs.ultralytics.com/tasks/detect
  3. # Parameters
  4. nc: 80 # number of classes
  5. scales: # model compound scaling constants, i.e. 'model=yolo11n.yaml' will call yolo11.yaml with scale 'n'
  6. # [depth, width, max_channels]
  7. n: [0.50, 0.25, 1024] # summary: 319 layers, 2624080 parameters, 2624064 gradients, 6.6 GFLOPs
  8. s: [0.50, 0.50, 1024] # summary: 319 layers, 9458752 parameters, 9458736 gradients, 21.7 GFLOPs
  9. m: [0.50, 1.00, 512] # summary: 409 layers, 20114688 parameters, 20114672 gradients, 68.5 GFLOPs
  10. l: [1.00, 1.00, 512] # summary: 631 layers, 25372160 parameters, 25372144 gradients, 87.6 GFLOPs
  11. x: [1.00, 1.50, 512] # summary: 631 layers, 56966176 parameters, 56966160 gradients, 196.0 GFLOPs
  12. # YOLO11n backbone
  13. backbone:
  14. # [from, repeats, module, args]
  15. - [-1, 1, Conv, [64, 3, 2]] # 0-P1/2
  16. - [-1, 1, Conv, [128, 3, 2]] # 1-P2/4
  17. - [-1, 2, RepNCSPELAN4_low, [256]]
  18. - [-1, 1, Conv, [256, 3, 2]] # 3-P3/8
  19. - [-1, 2, RepNCSPELAN4_low, [512]]
  20. - [-1, 1, Conv, [512, 3, 2]] # 5-P4/16
  21. - [-1, 2, RepNCSPELAN4_low, [512]]
  22. - [-1, 1, Conv, [1024, 3, 2]] # 7-P5/32
  23. - [-1, 2, RepNCSPELAN4_low, [1024]]
  24. - [-1, 1, SPPF, [1024, 5]] # 9
  25. - [-1, 2, C2PSA, [1024]] # 10
  26. # YOLO11n head
  27. head:
  28. - [-1, 1, nn.Upsample, [None, 2, "nearest"]]
  29. - [[-1, 6], 1, Concat, [1]] # cat backbone P4
  30. - [-1, 2, RepNCSPELAN4_low, [512]] # 13
  31. - [-1, 1, nn.Upsample, [None, 2, "nearest"]]
  32. - [[-1, 4], 1, Concat, [1]] # cat backbone P3
  33. - [-1, 2, RepNCSPELAN4_low, [256]] # 16 (P3/8-small)
  34. - [-1, 1, Conv, [256, 3, 2]]
  35. - [[-1, 13], 1, Concat, [1]] # cat head P4
  36. - [-1, 2, RepNCSPELAN4_low, [512]] # 19 (P4/16-medium)
  37. - [-1, 1, Conv, [512, 3, 2]]
  38. - [[-1, 10], 1, Concat, [1]] # cat head P5
  39. - [-1, 2, RepNCSPELAN4_low, [1024]] # 22 (P5/32-large)
  40. - [[16, 19, 22], 1, Detect, [nc]] # Detect(P3, P4, P5)


5.2  GELAN高参数量版本的yaml文件

此版本训练信息:YOLO11-RepGELAN-high summary: 651 layers, 3,837,803 parameters, 3,837,787 gradients, 12.1 GFLOPs

  1. # Ultralytics YOLO 🚀, AGPL-3.0 license
  2. # YOLO11 object detection model with P3-P5 outputs. For Usage examples see https://docs.ultralytics.com/tasks/detect
  3. # Parameters
  4. nc: 80 # number of classes
  5. scales: # model compound scaling constants, i.e. 'model=yolo11n.yaml' will call yolo11.yaml with scale 'n'
  6. # [depth, width, max_channels]
  7. n: [0.50, 0.25, 1024] # summary: 319 layers, 2624080 parameters, 2624064 gradients, 6.6 GFLOPs
  8. s: [0.50, 0.50, 1024] # summary: 319 layers, 9458752 parameters, 9458736 gradients, 21.7 GFLOPs
  9. m: [0.50, 1.00, 512] # summary: 409 layers, 20114688 parameters, 20114672 gradients, 68.5 GFLOPs
  10. l: [1.00, 1.00, 512] # summary: 631 layers, 25372160 parameters, 25372144 gradients, 87.6 GFLOPs
  11. x: [1.00, 1.50, 512] # summary: 631 layers, 56966176 parameters, 56966160 gradients, 196.0 GFLOPs
  12. # YOLO11n backbone
  13. backbone:
  14. # [from, repeats, module, args]
  15. - [-1, 1, Conv, [64, 3, 2]] # 0-P1/2
  16. - [-1, 1, Conv, [128, 3, 2]] # 1-P2/4
  17. - [-1, 2, RepNCSPELAN4_high, [256]]
  18. - [-1, 1, Conv, [256, 3, 2]] # 3-P3/8
  19. - [-1, 2, RepNCSPELAN4_high, [512]]
  20. - [-1, 1, Conv, [512, 3, 2]] # 5-P4/16
  21. - [-1, 2, RepNCSPELAN4_high, [512]]
  22. - [-1, 1, Conv, [1024, 3, 2]] # 7-P5/32
  23. - [-1, 2, RepNCSPELAN4_high, [1024]]
  24. - [-1, 1, SPPF, [1024, 5]] # 9
  25. - [-1, 2, C2PSA, [1024]] # 10
  26. # YOLO11n head
  27. head:
  28. - [-1, 1, nn.Upsample, [None, 2, "nearest"]]
  29. - [[-1, 6], 1, Concat, [1]] # cat backbone P4
  30. - [-1, 2, RepNCSPELAN4_high, [512]] # 13
  31. - [-1, 1, nn.Upsample, [None, 2, "nearest"]]
  32. - [[-1, 4], 1, Concat, [1]] # cat backbone P3
  33. - [-1, 2, RepNCSPELAN4_high, [256]] # 16 (P3/8-small)
  34. - [-1, 1, Conv, [256, 3, 2]]
  35. - [[-1, 13], 1, Concat, [1]] # cat head P4
  36. - [-1, 2, RepNCSPELAN4_high, [512]] # 19 (P4/16-medium)
  37. - [-1, 1, Conv, [512, 3, 2]]
  38. - [[-1, 10], 1, Concat, [1]] # cat head P5
  39. - [-1, 2, RepNCSPELAN4_high, [1024]] # 22 (P5/32-large)
  40. - [[16, 19, 22], 1, Detect, [nc]] # Detect(P3, P4, P5)

5.3 训练代码

大家可以创建一个py文件将我给的代码复制粘贴进去,配置好自己的文件路径即可运行。

  1. import warnings
  2. warnings.filterwarnings('ignore')
  3. from ultralytics import YOLO
  4. if __name__ == '__main__':
  5. model = YOLO('ultralytics/cfg/models/v8/yolov8-C2f-FasterBlock.yaml')
  6. # model.load('yolov8n.pt') # loading pretrain weights
  7. model.train(data=r'替换数据集yaml文件地址',
  8. # 如果大家任务是其它的'ultralytics/cfg/default.yaml'找到这里修改task可以改成detect, segment, classify, pose
  9. cache=False,
  10. imgsz=640,
  11. epochs=150,
  12. single_cls=False, # 是否是单类别检测
  13. batch=4,
  14. close_mosaic=10,
  15. workers=0,
  16. device='0',
  17. optimizer='SGD', # using SGD
  18. # resume='', # 如过想续训就设置last.pt的地址
  19. amp=False, # 如果出现训练损失为Nan可以关闭amp
  20. project='runs/train',
  21. name='exp',
  22. )


5.3 GELAN的训练过程截图

5.3.1 低参数量版本


5.3.2 高参数量版本


五、本文总结

到此本文的正式分享内容就结束了,在这里给大家推荐我的YOLOv11改进有效涨点专栏,本专栏目前为新开的平均质量分98分,后期我会根据各种最新的前沿顶会进行论文复现,也会对一些老的改进机制进行补充,如果大家觉得本文帮助到你了,订阅本专栏,关注后续更多的更新~