TY - JOUR
T1 - Enhanced ADHD detection
T2 - Frequency information embedded in a visual-language framework
AU - Hu, Runze
AU - Zhu, Kaishi
AU - Hou, Zhenzhe
AU - Wang, Ruideng
AU - Liu, Feifei
N1 - Publisher Copyright:
© 2024
PY - 2024/7
Y1 - 2024/7
N2 - This paper presents the Frequency-Integrated Visual-Language Network (FIVLNet), a deep learning (DL) framework tailored to improve the diagnostic accuracy for Attention Deficit Hyperactivity Disorder (ADHD) using magnetic resonance imaging (MRI) scans. Traditional DL approaches in ADHD diagnosis often overlook the sequential dependencies of MRI images or fail to adequately capture their complex structural details, resulting in suboptimal classification accuracy. To address this, the proposed FIVLNet synergistically integrates both high and low-frequency data from MRI images, based on the Convolutional Neural Network (CNN) and the cross-attention mechanism, subsequently achieving more comprehensive representations of the MRI images. Furthermore, in order to enrich the model's learning capacity, textual embeddings from Contrastive Language-Image Pre-training (CLIP) are introduced to provide additional modalities of information. FIVLNet also preserves a lightweight architecture, which necessitates a smaller number of learnable parameters compared to existing models.
AB - This paper presents the Frequency-Integrated Visual-Language Network (FIVLNet), a deep learning (DL) framework tailored to improve the diagnostic accuracy for Attention Deficit Hyperactivity Disorder (ADHD) using magnetic resonance imaging (MRI) scans. Traditional DL approaches in ADHD diagnosis often overlook the sequential dependencies of MRI images or fail to adequately capture their complex structural details, resulting in suboptimal classification accuracy. To address this, the proposed FIVLNet synergistically integrates both high and low-frequency data from MRI images, based on the Convolutional Neural Network (CNN) and the cross-attention mechanism, subsequently achieving more comprehensive representations of the MRI images. Furthermore, in order to enrich the model's learning capacity, textual embeddings from Contrastive Language-Image Pre-training (CLIP) are introduced to provide additional modalities of information. FIVLNet also preserves a lightweight architecture, which necessitates a smaller number of learnable parameters compared to existing models.
KW - Deep learning
KW - MRI image processing
KW - Multimodal learning
UR - http://www.scopus.com/inward/record.url?scp=85190164172&partnerID=8YFLogxK
U2 - 10.1016/j.displa.2024.102712
DO - 10.1016/j.displa.2024.102712
M3 - Article
AN - SCOPUS:85190164172
SN - 0141-9382
VL - 83
JO - Displays
JF - Displays
M1 - 102712
ER -