RepLKNet

  • impressively manages to scale the kernel size to 31×31 with improved performance
  • the performance starts to saturate as the kernel size continues growing, compared to the scaling trend of ad- vanced ViTs such as Swin Transformer
  • explore the possibility of training extreme convolutions larger than 31×31 and test whether the performance gap can be eliminated by strategically enlarging convolutions.