TY - GEN
T1 - Low Light Video Enhancement Based on Temporal-Spatial Complementary Feature
AU - Zhang, Gengchen
AU - Zeng, Yuhang
AU - Fu, Ying
N1 - Publisher Copyright:
© 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.
PY - 2022
Y1 - 2022
N2 - Under low light conditions, the quality of video data is heavily affected by noise, artifacts, and weak contrast, leading to low signal-to-noise ratio. Therefore, enhancing low light video to obtain high-quality information expression is a challenging problem. Deep learning based methods have achieved good performance on low light enhancement tasks and a majority of them are based on Unet. However, the widely used Unet architecture may generate pseudo-detail textures, as the simple skip connections of Unet introduce feature inconsistency between encoding and decoding stages. To overcome these shortcomings, we propose a novel network 3D Swin Skip Unet (3DS 2 Unet) in this paper. Specifically, we design a novel feature extraction and reconstruction module based on Swin Transformer and a temporal-channel attention module. Temporal-spatial complementary feature is generated by two modules and then fed into the decoder. The experimental results show that our model can well restore the texture of objects in the video, and performs better in removing noise and maintaining object boundaries between frames under low light conditions.
AB - Under low light conditions, the quality of video data is heavily affected by noise, artifacts, and weak contrast, leading to low signal-to-noise ratio. Therefore, enhancing low light video to obtain high-quality information expression is a challenging problem. Deep learning based methods have achieved good performance on low light enhancement tasks and a majority of them are based on Unet. However, the widely used Unet architecture may generate pseudo-detail textures, as the simple skip connections of Unet introduce feature inconsistency between encoding and decoding stages. To overcome these shortcomings, we propose a novel network 3D Swin Skip Unet (3DS 2 Unet) in this paper. Specifically, we design a novel feature extraction and reconstruction module based on Swin Transformer and a temporal-channel attention module. Temporal-spatial complementary feature is generated by two modules and then fed into the decoder. The experimental results show that our model can well restore the texture of objects in the video, and performs better in removing noise and maintaining object boundaries between frames under low light conditions.
KW - Attention mechanism
KW - Low light video enhancement
KW - Swin-Transformer
UR - http://www.scopus.com/inward/record.url?scp=85145006474&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-20497-5_30
DO - 10.1007/978-3-031-20497-5_30
M3 - Conference contribution
AN - SCOPUS:85145006474
SN - 9783031204968
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 368
EP - 379
BT - Artificial Intelligence - Second CAAI International Conference, CICAI 2022, Revised Selected Papers
A2 - Fang, Lu
A2 - Povey, Daniel
A2 - Zhai, Guangtao
A2 - Mei, Tao
A2 - Wang, Ruiping
PB - Springer Science and Business Media Deutschland GmbH
T2 - 2nd CAAI International Conference on Artificial Intelligence, CAAI 2022
Y2 - 27 August 2022 through 28 August 2022
ER -