TY - JOUR
T1 - UPANets: Learning from the Universal Pixel Attention Neworks
AU - Tseng, Ching-Hsun
AU - Lee, Shin-Jye
AU - Feng, Jianan
AU - Mao, Shengzhong
AU - Wu, Yuping
AU - Shang, Jiayu
AU - Zeng, Xiaojun
N1 - Funding Information:
This research was funded by Taiwan Ministry of Science and Technology grant number MOST 111-2410-H-A49-019.
Publisher Copyright:
© 2022 by the authors.
PY - 2022/9/4
Y1 - 2022/9/4
N2 - With the successful development in computer vision, building a deep convolutional neural network (CNNs) has been mainstream, considering the character of shared parameters in a convolutional layer. Stacking convolutional layers into a deep structure improves performance, but over-stacking also ramps up the needed resources for GPUs. Seeing another surge of Transformers in computer vision, the issue has aroused severely. A resource-hungry model is hardly implemented for limited hardware or single-customers-based GPU. Therefore, this work focuses on these concerns and proposes an efficient but robust backbone, which equips with channel and spatial direction attentions, so the attentions help to expand receptive fields in shallow convolutional layers and pass the information to every layer. An attention-boosted network based on already efficient CNNs, Universal Pixel Attention Networks (UPANets), is proposed. Through a series of experiments, UPANets fulfil the purposes of learning global information with less needed resources and outshine many existing SOTAs in CIFAR-{10, 100}.
AB - With the successful development in computer vision, building a deep convolutional neural network (CNNs) has been mainstream, considering the character of shared parameters in a convolutional layer. Stacking convolutional layers into a deep structure improves performance, but over-stacking also ramps up the needed resources for GPUs. Seeing another surge of Transformers in computer vision, the issue has aroused severely. A resource-hungry model is hardly implemented for limited hardware or single-customers-based GPU. Therefore, this work focuses on these concerns and proposes an efficient but robust backbone, which equips with channel and spatial direction attentions, so the attentions help to expand receptive fields in shallow convolutional layers and pass the information to every layer. An attention-boosted network based on already efficient CNNs, Universal Pixel Attention Networks (UPANets), is proposed. Through a series of experiments, UPANets fulfil the purposes of learning global information with less needed resources and outshine many existing SOTAs in CIFAR-{10, 100}.
KW - CNN
KW - attention
KW - computer vision
KW - image classification
UR - http://dx.doi.org/10.3390/e24091243
UR - https://www.scopus.com/pages/publications/85138545436
U2 - 10.3390/e24091243
DO - 10.3390/e24091243
M3 - Article
SN - 1099-4300
VL - 24
SP - 1243
JO - Entropy
JF - Entropy
IS - 9
ER -