资源类型:
期刊
WOS体系:
Review
Pubmed体系:
Journal Article;Review
收录情况:
◇ SCIE
文章类型:
论著
机构:
[1]School of Computer Science and Software Engineering, Southwest Petroleum University, Sichuan, China
[2]Operation and Development Department, Sichuan Clinical Research Center for Cancer, Sichuan Cancer Hospital & Institute, Chengdu, China
四川省肿瘤医院
[3]Department of Biomedical Engineering, Michigan Technological University, Houghton, MI 49931, USA
[4]Department of Medical Physics, University of Wisconsin, Madison, WI 53705, USA
ISSN:
0140-0118
关键词:
Knowledge distillation
Ultrasound elastography
Unsupervised learning
Motion tracking
摘要:
Accurate and efficient motion estimation is a crucial component of real-time ultrasound elastography (USE). However, obtaining radiofrequency ultrasound (RF) data in clinical practice can be challenging. In contrast, although B-mode (BM) data is readily available, elastographic data derived from BM data results in sub-optimal elastographic images. Furthermore, existing conventional ultrasound devices (e.g., portable devices) cannot provide elastography modes, which has become a significant obstacle to the widespread use of traditional ultrasound devices. To address the challenges above, we developed a teacher-student guided knowledge distillation for an unsupervised convolutional neural network (TSGUPWC-Net) to improve the accuracy of BM motion estimation by employing a well-established convolutional neural network (CNN) named modified pyramid warping and cost volume network (MPWC-Net). A pre-trained teacher model based on RF is utilized to guide the training of a student model using BM data. Innovations outlined below include employing spatial attention transfer at intermediate layers to enhance the guidance effect of the model. The loss function consists of smoothness of the displacement field, knowledge distillation loss, and intermediate layer loss. We evaluated our method on simulated data, phantoms, and in vivo ultrasound data. The results indicate that our method has higher signal-to-noise ratio (SNR) and contrast-to-noise ratio (CNR) values in axial strain estimation than the model trained on BM. The model is unsupervised and requires no ground truth labels during training, making it highly promising for motion estimation applications.© 2024. International Federation for Medical and Biological Engineering.
基金:
Sichuan Provincial Natural Science Foundation of China under Grant 2022NSFSC0833 and in part by research grants from the Science and Technology Council of Nanchong (SXQHJH046, SXHZ019).
被引次数:
2
WOS:
WOS:001203369900002
PubmedID:
38627356
中科院(CAS)分区:
最新[2023]版:
大类
|
4 区
医学
小类
|
2 区
数学与计算生物学
4 区
计算机:跨学科应用
4 区
工程:生物医学
4 区
医学:信息
JCR分区:
最新[2023]版:
Q2
COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS
Q2
MATHEMATICAL & COMPUTATIONAL BIOLOGY
Q3
ENGINEERING, BIOMEDICAL
Q3
MEDICAL INFORMATICS
影响因子:
2.6
最新[2023版]
2.7
最新五年平均
2.6
出版当年[2023版]
0
出版当年五年平均
2.6
出版前一年[2023版]
第一作者:
Xiang Tianqiang
第一作者机构:
[1]School of Computer Science and Software Engineering, Southwest Petroleum University, Sichuan, China
通讯作者:
Peng Bo;Jiang Jingfeng
通讯机构:
[3]Department of Biomedical Engineering, Michigan Technological University, Houghton, MI 49931, USA
[4]Department of Medical Physics, University of Wisconsin, Madison, WI 53705, USA
推荐引用方式(GB/T 7714):
Xiang Tianqiang,Li Yan,Deng Hui,et al.Teacher-student guided knowledge distillation for unsupervised convolutional neural network-based speckle tracking in ultrasound strain elastography[J].Medical & Biological Engineering & Computing.2024,62(8):2265-2279.doi:10.1007/s11517-024-03078-z.
APA:
Xiang Tianqiang,Li Yan,Deng Hui,Tian Chao,Peng Bo&Jiang Jingfeng.(2024).Teacher-student guided knowledge distillation for unsupervised convolutional neural network-based speckle tracking in ultrasound strain elastography.Medical & Biological Engineering & Computing,62,(8)
MLA:
Xiang Tianqiang,et al."Teacher-student guided knowledge distillation for unsupervised convolutional neural network-based speckle tracking in ultrasound strain elastography".Medical & Biological Engineering & Computing 62..8(2024):2265-2279