Accurate recognition of auditory brainstem response (ABR) wave latencies is essential for clinical practice but remains a subjective and time-consuming process. Existing AI approaches face challenges in generalization, complexity, and semantic sparsity due to single sampling-point analysis. This study introduces the Derivative-Guided Patch Dual-Attention Transformer (Patch-DAT), a novel, lightweight, and generalizable deep learning (DL) model for the automated recognition of latencies for waves I, III, and V. Patch-DAT divides the ABR time series into overlapping patches to aggregate semantic information, better capturing local temporal patterns. Meanwhile, leveraging the fact that ABR waves occur at the zero crossing of the first derivative, Patch-DAT incorporates a first derivative-guided dual-attention mechanism to model global dependencies. Trained and validated on large-scale, diverse datasets from two hospitals, Patch-DAT (with a size of 0.36 MB) achieves accuracies of 92.29% and 98.07% at 0.1 ms and 0.2 ms error scales, respectively, on a held-out test set. It also performs well on an independent dataset with accuracies of 88.50% and 95.14%, demonstrating strong generalization across clinical settings. Ablation studies highlight the contributions of the patching strategy and dual-attention mechanisms. Compared to previous state-of-the-art DL models, Patch-DAT shows superior accuracy and reduced complexity, making it a promising solution for object recognition of ABR latencies. Additionally, we systematically investigate how sample size and data heterogeneity affect model generalization, indicating the importance of large, diverse datasets in training robust DL models. Future work will focus on expanding dataset diversity and improving model interpretability to further improve clinical relevance.
基金:
National Key Research and Development Program of China [2022YFA1004100]; National Natural Science Foundation of China [62301096, 62371084]; Natural Science Foundation of Chongqing, China [CSTB2023NSCQ-MSX0659]; Science and Technology Research Program of Chongqing Municipal Education Commission [KJQN202400632]
第一作者机构:[1]Chongqing Univ Posts & Telecommun, Sch Commun & Informat Engn, Chongqing 400065, Peoples R China[2]Chongqing Univ Posts & Telecommun, Inst Adv Sci, Chongqing 400065, Peoples R China
通讯作者:
推荐引用方式(GB/T 7714):
Liu Yin,Sun Huanghong,Li Qiang,et al.Derivative-Guided Dual-Attention Mechanisms in Patch Transformer for Efficient Automated Recognition of Auditory Brainstem Response Latency[J].IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING.2025,33:1865-1877.doi:10.1109/TNSRE.2025.3558730.
APA:
Liu, Yin,Sun, Huanghong,Li, Qiang,Li, Kangkang,Fu, Xinxing...&Gao, Chenqiang.(2025).Derivative-Guided Dual-Attention Mechanisms in Patch Transformer for Efficient Automated Recognition of Auditory Brainstem Response Latency.IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING,33,
MLA:
Liu, Yin,et al."Derivative-Guided Dual-Attention Mechanisms in Patch Transformer for Efficient Automated Recognition of Auditory Brainstem Response Latency".IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING 33.(2025):1865-1877