TY - GEN
T1 - Feature Segmentation-based Recognition Technology for Explosive Devices
AU - Li, Xin
AU - Feng, Lingyun
AU - Zeng, Bingru
AU - Hu, Zenghao
AU - Liu, Chunmei
AU - Wang, Yabin
N1 - Publisher Copyright:
© 2025 SPIE.
PY - 2025/7/21
Y1 - 2025/7/21
N2 - A novel feature segmentation-based recognition system for explosive devices is proposed, leveraging advanced deep learning techniques to enhance detection accuracy and robustness in complex environments. Effective management of explosive devices is critical for global security, yet existing methods struggle with accuracy and real-time performance. We hope to apply this system to the detection and recognition of conventional explosives or non-conventional explosives, especially highly camouflaged non-conventional explosives. The proposed system integrates Transformer-based Grounding DINO (Grounding by Distillation with Implicit Neural Optimization) with SAM (Segment Anything Model), combining human inputs (e.g., category names or pointers) with CNN (Convolutional Neural Net) and ViTs (Vision Transformers) to improve accuracy, real-time processing, and adaptability. Experiments using visible light and X-ray image datasets of simulated explosive devices demonstrate superior performance of the deep learning model over traditional algorithms in both image segmentation and object recognition tasks. The proposed explosive device identification method significantly enhances accuracy and real-time performance, with great potential for public safety and emergency response applications.
AB - A novel feature segmentation-based recognition system for explosive devices is proposed, leveraging advanced deep learning techniques to enhance detection accuracy and robustness in complex environments. Effective management of explosive devices is critical for global security, yet existing methods struggle with accuracy and real-time performance. We hope to apply this system to the detection and recognition of conventional explosives or non-conventional explosives, especially highly camouflaged non-conventional explosives. The proposed system integrates Transformer-based Grounding DINO (Grounding by Distillation with Implicit Neural Optimization) with SAM (Segment Anything Model), combining human inputs (e.g., category names or pointers) with CNN (Convolutional Neural Net) and ViTs (Vision Transformers) to improve accuracy, real-time processing, and adaptability. Experiments using visible light and X-ray image datasets of simulated explosive devices demonstrate superior performance of the deep learning model over traditional algorithms in both image segmentation and object recognition tasks. The proposed explosive device identification method significantly enhances accuracy and real-time performance, with great potential for public safety and emergency response applications.
KW - Explosive Device Recognition
KW - Feature Segmentation
KW - Grounding DINO
KW - SAM
UR - https://www.scopus.com/pages/publications/105022928086
U2 - 10.1117/12.3076606
DO - 10.1117/12.3076606
M3 - Conference contribution
AN - SCOPUS:105022928086
T3 - Proceedings of SPIE - The International Society for Optical Engineering
BT - Third Asia Conference on Computer Vision, Image Processing, and Pattern Recognition, CVIPPR 2025
A2 - Chen, Lei
PB - SPIE
T2 - 3rd Asia Conference on Computer Vision, Image Processing, and Pattern Recognition, CVIPPR 2025
Y2 - 23 May 2025 through 25 May 2025
ER -