Woongkyu Park
A latency-efficient integration of channel attention for ConvNets.
Park, Woongkyu; Choi, Yeongyu; Mekala, Mahammad Shareef; SangChoi, Gyu; Yoo, Kook-Yeol; Jung, Ho-youl
Authors
Abstract
Designing fast and accurate neural networks is becoming essential in various vision tasks. Recently, the use of attention mechanisms has increased, aimed at enhancing the vision task performance by selectively focusing on relevant parts of the input. In this paper, we concentrate on squeeze-and-excitation (SE)-based channel attention, considering the trade-off between latency and accuracy. We propose a variation of the SE module, called squeeze-and-excitation with layer normalization (SELN), in which layer normalization (LN) replaces the sigmoid activation function. This approach reduces the vanishing gradient problem while enhancing feature diversity and discriminability of channel attention. In addition, we propose a latency-efficient model named SELNeXt, where the LN typically used in the ConvNext block is replaced by SELN to minimize additional latency-impacting operations. Through classification simulations on ImageNet-1k, we show that the top-1 accuracy of the proposed SELNeXt outperforms other ConvNeXtbased models in terms of latency efficiency. SELNeXt also achieves better object detection and instance segmentation performance on COCO than Swin Transformer and ConvNeXt for small-sized models. Our results indicate that LN could be a considerable candidate for replacing the activation function in attention mechanisms. In addition, SELNeXt achieves a better accuracy-latency trade-off, making it favorable for real-time applications and edge computing.
Citation
PARK, W., CHOI, Y., MEKALA, M.S., CHOI, G.S., YOO, K.-Y. and JUNG, H.-Y. 2025. A latency-efficient integration of channel attention for ConvNets. Computers, maerials and continua [online], 82(3), pages 3965-3981. Available from: https://doi.org/10.32604/cmc.2025.059966
Journal Article Type | Article |
---|---|
Acceptance Date | Jan 15, 2025 |
Online Publication Date | Mar 6, 2025 |
Publication Date | Mar 31, 2025 |
Deposit Date | Mar 27, 2025 |
Publicly Available Date | Mar 27, 2025 |
Journal | Computers, materials and continua |
Print ISSN | 1546-2218 |
Electronic ISSN | 1546-2226 |
Publisher | Tech Science Press |
Peer Reviewed | Peer Reviewed |
Volume | 82 |
Issue | 3 |
Pages | 3965-3981 |
DOI | https://doi.org/10.32604/cmc.2025.059966 |
Keywords | Attention mechanism; Convolutional neural networks; Image classification; Object detection; Semantic segmentation |
Public URL | https://rgu-repository.worktribe.com/output/2761834 |
Files
PARK 2025 A latency-efficient integration (VOR)
(1.1 Mb)
PDF
Publisher Licence URL
https://creativecommons.org/licenses/by/4.0/
Copyright Statement
© 2025 The Authors. Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
You might also like
A multimodel-based screening framework for C-19 using deep learning-inspired data fusion.
(2024)
Journal Article
Object-aware multi-criteria decision-making approach using the heuristic data-driven theory for intelligent transportation systems.
(2023)
Presentation / Conference Contribution
ASXC2 approach: a service-X cost optimization strategy based on edge orchestration for IIoT.
(2023)
Journal Article
Downloadable Citations
About OpenAIR@RGU
Administrator e-mail: publications@rgu.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2025
Advanced Search