1
|
Wu YC, Chang CY, Huang YT, Chen SY, Chen CH, Kao HK. Artificial Intelligence Image Recognition System for Preventing Wrong-Site Upper Limb Surgery. Diagnostics (Basel) 2023; 13:3667. [PMID: 38132251 PMCID: PMC10743305 DOI: 10.3390/diagnostics13243667] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2023] [Revised: 11/30/2023] [Accepted: 12/02/2023] [Indexed: 12/23/2023] Open
Abstract
Our image recognition system employs a deep learning model to differentiate between the left and right upper limbs in images, allowing doctors to determine the correct surgical position. From the experimental results, it was found that the precision rate and the recall rate of the intelligent image recognition system for preventing wrong-site upper limb surgery proposed in this paper could reach 98% and 93%, respectively. The results proved that our Artificial Intelligence Image Recognition System (AIIRS) could indeed assist orthopedic surgeons in preventing the occurrence of wrong-site left and right upper limb surgery. At the same time, in future, we will apply for an IRB based on our prototype experimental results and we will conduct the second phase of human trials. The results of this research paper are of great benefit and research value to upper limb orthopedic surgery.
Collapse
Affiliation(s)
- Yi-Chao Wu
- Department of Electronic Engineering, National Yunlin University of Science and Technology, Yunlin 950359, Taiwan;
| | - Chao-Yun Chang
- Interdisciplinary Program of Green and Information Technology, National Taitung University, Taitung 950359, Taiwan; (C.-Y.C.); (Y.-T.H.); (S.-Y.C.)
| | - Yu-Tse Huang
- Interdisciplinary Program of Green and Information Technology, National Taitung University, Taitung 950359, Taiwan; (C.-Y.C.); (Y.-T.H.); (S.-Y.C.)
| | - Sung-Yuan Chen
- Interdisciplinary Program of Green and Information Technology, National Taitung University, Taitung 950359, Taiwan; (C.-Y.C.); (Y.-T.H.); (S.-Y.C.)
| | - Cheng-Hsuan Chen
- Department of Electrical Engineering, National Central University, Taoyuan 320317, Taiwan;
- Department of Electrical Engineering, Fu Jen Catholic University, New Taipei City 242062, Taiwan
| | - Hsuan-Kai Kao
- Department of Orthopedic Surgery, Chang Gung Memorial Hospital at Linkou, Taoyuan 333423, Taiwan
- Bone and Joint Research Center, Chang Gung Memorial Hospital at Linkou, Taoyuan 333423, Taiwan
- College of Medicine, Chang Gung University, Taoyuan 333423, Taiwan
| |
Collapse
|
2
|
Paridar R, Asl BM. Frame rate improvement in ultrafast coherent plane wave compounding. ULTRASONICS 2023; 135:107136. [PMID: 37647702 DOI: 10.1016/j.ultras.2023.107136] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/24/2023] [Revised: 07/18/2023] [Accepted: 08/10/2023] [Indexed: 09/01/2023]
Abstract
Coherent plane wave compounding (CPWC), as an ultrafast ultrasound imaging technique, makes a significant breakthrough in frame rate enhancement. However, there exists a compromise between the quality of the final image and the frame rate in CPWC. In this paper, we propose an efficient method to minimize the number of required emissions, and consequently, improve the frame rate, while maintaining the image quality. To this end, we down-sample the angle interval using two specific sampling factors. More precisely, we construct two different subsets, each of which consists of a few numbers of emissions. The optimal values of the angle intervals are achieved based on the beampattern that corresponds to the reference case (that is, the case where all plane waves are used). Finally, in order to keep the image quality comparable with the reference case, we apply some modifications to the image reconstruction procedure. In the proposed algorithm, the Delay-and-Sum beamformed images of two considered subsets are convolved to achieve the final reconstructed image. The obtained results confirm the efficiency of the proposed method in terms of frame rate improvement compared to the reference case. In particular, by using the proposed method, the required emissions in PICMUS data reduce to 16, which is 4.6 times smaller compared to the reference case. Also, the gCNR values of the proposed method and the reference case are obtained as 0.98 and 0.97, respectively, for in-vivo dataset. This demonstrates that the proposed method successfully preserves the quality of the reconstructed image by using much fewer emissions.
Collapse
Affiliation(s)
- Roya Paridar
- Department of Biomedical Engineering, Tarbiat Modares University, Tehran, Iran
| | | |
Collapse
|
3
|
Liang Z, Chen K, Luo T, Jiang W, Wen J, Zhao L, Song W. HTC-Net: Hashimoto's thyroiditis ultrasound image classification model based on residual network reinforced by channel attention mechanism. Health Inf Sci Syst 2023; 11:24. [PMID: 37234207 PMCID: PMC10205956 DOI: 10.1007/s13755-023-00225-y] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2023] [Accepted: 04/23/2023] [Indexed: 05/27/2023] Open
Abstract
Convolutional neural network (CNN) is efficient in extracting and aggregating local features in the spatial dimension of the images. However, obtaining the inapparent texture information of the low-echo area in the ultrasound images is not easy, and it is especially challenging for the early lesion recognition in Hashimoto's thyroiditis (HT) ultrasound images. In this paper, a HT ultrasound image classification model HTC-Net based on residual network reinforced by channel attention mechanism is proposed. HTC-Net strengthens the features of the important channels by reinforced channel attention mechanism through which the high-level semantic information is enchanced and the low-level semantic information is suppressed. Residual network assists HTC-Net focus on the key local areas of the ultrasound images while pay attention to the global semantic information. Furthermore, in order to solve the problem of uneven distribution caused by large amount of difficult-to-classify samples in the data sets, a new feature loss function TanCELoss with weight factor dynamically adjusting is constructed. TanCELoss function can better assist HTC-Net to transform difficult-to-classify samples into easy-to-classify samples gradually, and improve the balancing distribution of the samples. The experiments are implemented based on data sets collected by the Endocrinology Department of four branches from Guangdong Provincial Hospital of Chinese Medicine. Both quantitative testing and visualization results show that HTC-Net obtains STOA performance for early lesions recognition in HT ultrasound images. HTC-Net has great application value especially under the condition of owning only small data samples.
Collapse
Affiliation(s)
- Zhipeng Liang
- School of Computer Science and Technology, Guangdong University of Technology, Guangzhou, 510006 China
| | - Kang Chen
- School of Computer Science and Technology, Guangdong University of Technology, Guangzhou, 510006 China
| | - Tianchun Luo
- School of Computer Science and Technology, Guangdong University of Technology, Guangzhou, 510006 China
| | - Wenchao Jiang
- School of Computer Science and Technology, Guangdong University of Technology, Guangzhou, 510006 China
| | - Jianxuan Wen
- Department of Endocrinology, Guangdong Provincial Hospital of Chinese Medicine, Guangzhou, 510120 China
| | - Ling Zhao
- Department of Endocrinology, Guangdong Provincial Hospital of Chinese Medicine, Guangzhou, 510120 China
| | - Wei Song
- Department of Endocrinology, Guangdong Provincial Hospital of Chinese Medicine, Guangzhou, 510120 China
| |
Collapse
|
4
|
Afrakhteh S, Iacca G, Demi L. A two-dimensional angular interpolation based on radial basis functions for high frame rate ultrafast imaging. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2023; 154:3454-3465. [PMID: 38015029 DOI: 10.1121/10.0022515] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/19/2023] [Accepted: 11/06/2023] [Indexed: 11/29/2023]
Abstract
To solve the problem of reduced image quality in plane wave imaging (PWI), coherent plane wave compounding (CPWC) has been introduced, based on a combination of plane wave images from several directions (i.e., with different angles). However, the number of angles needed to reach a reasonable image quality affects the maximum achievable frame rate in CPWC. In this study, we suggest reducing the tradeoff between the image quality and the frame rate in CPWC by employing two-dimensional (2D) interpolation based on radial basis functions. More specifically, we propose constructing a three-dimensional spatio-angular structure to integrate both spatial and angular information into the reconstruction prior to 2D interpolation. The rationale behind our proposal is to reduce the number of transmissions and then apply the 2D interpolation along the angle dimension to reconstruct the missing information corresponding to the angles not selected for CPWC imaging. To evaluate the proposed technique, we applied it to the PWI challenges in the medical ultrasound database. Results show that we can achieve 3× to 4× improvement in frame rate while maintaining acceptable image quality compared to the case of using all the angles.
Collapse
Affiliation(s)
- Sajjad Afrakhteh
- Department of Information Engineering and Computer Science, University of Trento, Trento, Italy
| | - Giovanni Iacca
- Department of Information Engineering and Computer Science, University of Trento, Trento, Italy
| | - Libertario Demi
- Department of Information Engineering and Computer Science, University of Trento, Trento, Italy
| |
Collapse
|
5
|
Nguon LS, Park S. Extended aperture image reconstruction for plane-wave imaging. ULTRASONICS 2023; 134:107096. [PMID: 37392616 DOI: 10.1016/j.ultras.2023.107096] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/24/2022] [Revised: 05/05/2023] [Accepted: 06/26/2023] [Indexed: 07/03/2023]
Abstract
B-mode images undergo degradation in the boundary region because of the limited number of elements in the ultrasound probe. Herein, a deep learning-based extended aperture image reconstruction method is proposed to reconstruct a B-mode image with an enhanced boundary region. The proposed network can reconstruct an image using pre-beamformed raw data received from the half-aperture of the probe. To generate a high-quality training target without degradation in the boundary region, the target data were acquired using the full-aperture. Training data were acquired from an experimental study using a tissue-mimicking phantom, vascular phantom, and simulation of random point scatterers. Compared with plane-wave images from delay and sum beamforming, the proposed extended aperture image reconstruction method achieves improvement at the boundary region in terms of the multi-scale structure of similarity and peak signal-to-noise ratio by 8% and 4.10 dB in resolution evaluation phantom, 7% and 3.15 dB in contrast speckle phantom, and 5% and 3 dB in in vivo study of carotid artery imaging. The findings in this study prove the feasibility of a deep learning-based extended aperture image reconstruction method for boundary region improvement.
Collapse
Affiliation(s)
- Leang Sim Nguon
- Department of Electronic and Electrical Engineering, Ewha Womans University, Seoul 03760, Korea
| | - Suhyun Park
- Department of Electronic and Electrical Engineering, Ewha Womans University, Seoul 03760, Korea.
| |
Collapse
|
6
|
Molinier N, Painchaud-April G, Le Duff A, Toews M, Bélanger P. Ultrasonic imaging using conditional generative adversarial networks. ULTRASONICS 2023; 133:107015. [PMID: 37269681 DOI: 10.1016/j.ultras.2023.107015] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/10/2023] [Revised: 03/17/2023] [Accepted: 04/11/2023] [Indexed: 06/05/2023]
Abstract
The Full Matrix Capture (FMC) and Total Focusing Method (TFM) combination is often considered the gold standard in ultrasonic nondestructive testing, however it may be impractical due to the amount of time required to gather and process the FMC, particularly for high cadence inspections. This study proposes replacing conventional FMC acquisition and TFM processing with a single zero-degree plane wave (PW) insonification and a conditional Generative Adversarial Network (cGAN) trained to produce TFM-like images. Three models with different cGAN architectures and loss formulations were tested in different scenarios. Their performances were compared with conventional TFM computed from FMC. The proposed cGANs were able to recreate TFM-like images with the same resolution while improving the contrast in more than 94% of the reconstructions in comparison with conventional TFM reconstructions. Indeed, thanks to the use of a bias in the cGANs' training, the contrast was systematically increased through a reduction of the background noise level and the elimination of some artifacts. Finally, the proposed method led to a reduction of the computation time and file size by a factor of 120 and 75, respectively.
Collapse
Affiliation(s)
- Nathan Molinier
- PULÉTS, École de Technologie Supérieure (ÉTS), Montréal, H3C 1K3, QC, Canada.
| | | | - Alain Le Duff
- Evident Industrial (formerly Olympus IMS), Québec, G1P 0B3, QC, Canada.
| | - Matthew Toews
- Department of Systems Engineering, École de Technologie Supérieure, Université du Québec, Montréal, H3C 1K3, QC, Canada.
| | - Pierre Bélanger
- PULÉTS, École de Technologie Supérieure (ÉTS), Montréal, H3C 1K3, QC, Canada; Department of Mechanical Engineering, École de Technologie Supérieure, Université du Québec, Montréal, H3C 1K3, QC, Canada.
| |
Collapse
|
7
|
Fouad M, Ghany MAAE, Schmitz G. A Single-Shot Harmonic Imaging Approach Utilizing Deep Learning for Medical Ultrasound. IEEE TRANSACTIONS ON ULTRASONICS, FERROELECTRICS, AND FREQUENCY CONTROL 2023; 70:237-252. [PMID: 37018250 DOI: 10.1109/tuffc.2023.3234230] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Tissue harmonic imaging (THI) is an invaluable tool in clinical ultrasound due to its enhanced contrast resolution and reduced reverberation clutter in comparison with fundamental mode imaging. However, harmonic content separation based on high-pass filtering suffers from potential contrast degradation or lower axial resolution due to spectral leakage, whereas nonlinear multipulse harmonic imaging schemes, such as amplitude modulation and pulse inversion, suffer from a reduced frame rate and comparatively higher motion artifacts due to the necessity of at least two pulse echo acquisitions. To address this problem, we propose a deep-learning-based single-shot harmonic imaging technique capable of generating comparable image quality to pulse amplitude modulation methods, yet at a higher frame rate and with fewer motion artifacts. Specifically, an asymmetric convolutional encoder-decoder structure is designed to estimate the combination of the echoes resulting from the half-amplitude transmissions using the echo produced from the full amplitude transmission as input. The echoes were acquired with the checkerboard amplitude modulation technique for training. The model was evaluated across various targets and samples to illustrate generalizability as well as the possibility and impact of transfer learning. Furthermore, for possible interpretability of the network, we investigate if the latent space of the encoder holds information on the nonlinearity parameter of the medium. We demonstrate the ability of the proposed approach to generate harmonic images with a single firing that are comparable to those from a multipulse acquisition.
Collapse
|
8
|
Luijten B, Chennakeshava N, Eldar YC, Mischi M, van Sloun RJG. Ultrasound Signal Processing: From Models to Deep Learning. ULTRASOUND IN MEDICINE & BIOLOGY 2023; 49:677-698. [PMID: 36635192 DOI: 10.1016/j.ultrasmedbio.2022.11.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/10/2022] [Revised: 11/02/2022] [Accepted: 11/05/2022] [Indexed: 06/17/2023]
Abstract
Medical ultrasound imaging relies heavily on high-quality signal processing to provide reliable and interpretable image reconstructions. Conventionally, reconstruction algorithms have been derived from physical principles. These algorithms rely on assumptions and approximations of the underlying measurement model, limiting image quality in settings where these assumptions break down. Conversely, more sophisticated solutions based on statistical modeling or careful parameter tuning or derived from increased model complexity can be sensitive to different environments. Recently, deep learning-based methods, which are optimized in a data-driven fashion, have gained popularity. These model-agnostic techniques often rely on generic model structures and require vast training data to converge to a robust solution. A relatively new paradigm combines the power of the two: leveraging data-driven deep learning and exploiting domain knowledge. These model-based solutions yield high robustness and require fewer parameters and training data than conventional neural networks. In this work we provide an overview of these techniques from the recent literature and discuss a wide variety of ultrasound applications. We aim to inspire the reader to perform further research in this area and to address the opportunities within the field of ultrasound signal processing. We conclude with a future perspective on model-based deep learning techniques for medical ultrasound.
Collapse
Affiliation(s)
- Ben Luijten
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands.
| | - Nishith Chennakeshava
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands
| | - Yonina C Eldar
- Faculty of Math and Computer Science, Weizmann Institute of Science, Rehovot, Israel
| | - Massimo Mischi
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands
| | - Ruud J G van Sloun
- Department of Electrical Engineering, Eindhoven University of Technology, Eindhoven, The Netherlands; Philips Research, Eindhoven, The Netherlands
| |
Collapse
|
9
|
Gao J, Xu L, Zou Q, Zhang B, Wang D, Wan M. A progressively dual reconstruction network for plane wave beamforming with both paired and unpaired training data. ULTRASONICS 2023; 127:106833. [PMID: 36070635 DOI: 10.1016/j.ultras.2022.106833] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/12/2022] [Revised: 08/12/2022] [Accepted: 08/16/2022] [Indexed: 06/15/2023]
Abstract
High-frame-rate plane wave (PW) imaging suffers from unsatisfactory image quality due to the absence of focus in transmission. Although coherent compounding from tens of PWs can improve PW image quality, it in turn results in a decreased frame rate, which is limited for tracking fast moving tissues. To overcome the trade-off between frame rate and image quality, we propose a progressively dual reconstruction network (PDRN) to achieve adaptive beamforming and enhance the image quality via both supervised and transfer learning in the condition of single or a few PWs transmission. Specifically, the proposed model contains a progressive network and a dual network to form a close loop and provide collaborative supervision for model optimization. The progressive network takes the channel delay of each spatial point as input and progressively learns coherent compounding beamformed data with increased numbers of steered PWs step by step. The dual network learns the downsampling process and reconstructs the beamformed data with decreased numbers of steered PWs, which reduces the space of the possible learning functions and improves the model's discriminative ability. In addition, the dual network is leveraged to perform transfer learning for the training data without sufficient steered PWs. The simulated, in vivo, vocal cords (VCs), and public available CUBDL dataset are collected for model evaluation.
Collapse
Affiliation(s)
- Junling Gao
- The Key Laboratory of Biomedical Information Engineering of Ministry of Education, School of Life Science and Technology, Xi'an Jiaotong University, Xi'an 710049, PR China
| | - Lei Xu
- The Key Laboratory of Biomedical Information Engineering of Ministry of Education, School of Life Science and Technology, Xi'an Jiaotong University, Xi'an 710049, PR China; Xi'an Hospital of Traditional Chinese Medicine, Xi'an 710021, PR China
| | - Qin Zou
- The Key Laboratory of Biomedical Information Engineering of Ministry of Education, School of Life Science and Technology, Xi'an Jiaotong University, Xi'an 710049, PR China
| | - Bo Zhang
- The Key Laboratory of Biomedical Information Engineering of Ministry of Education, School of Life Science and Technology, Xi'an Jiaotong University, Xi'an 710049, PR China
| | - Diya Wang
- The Key Laboratory of Biomedical Information Engineering of Ministry of Education, School of Life Science and Technology, Xi'an Jiaotong University, Xi'an 710049, PR China.
| | - Mingxi Wan
- The Key Laboratory of Biomedical Information Engineering of Ministry of Education, School of Life Science and Technology, Xi'an Jiaotong University, Xi'an 710049, PR China.
| |
Collapse
|
10
|
Goudarzi S, Rivaz H. Deep reconstruction of high-quality ultrasound images from raw plane-wave data: A simulation and in vivo study. ULTRASONICS 2022; 125:106778. [PMID: 35728310 DOI: 10.1016/j.ultras.2022.106778] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/05/2021] [Revised: 05/27/2022] [Accepted: 05/28/2022] [Indexed: 06/15/2023]
Abstract
This paper presents a novel beamforming approach based on deep learning to get closer to the ideal Point Spread Function (PSF) in Plane-Wave Imaging (PWI). The proposed approach is designed to reconstruct a high-quality version of Tissue Reflectivity Function (TRF) from echo traces acquired by transducer elements using only a single plane-wave transmission. In this approach, first, a model for the TRF is introduced by setting the imaging PSF as an isotropic (i.e., circularly symmetric) 2D Gaussian kernel convolved with a cosine function. Then, a mapping function between the pre-beamformed Radio-Frequency (RF) channel data and the proposed output is constructed using deep learning. Network architecture contains multi-resolution decomposition and reconstruction using wavelet transform for effective recovery of high-frequency content of the desired output. We exploit step by step training from coarse (mean square error) to fine (ℓ0.2) loss functions. The proposed method is trained on 1174 simulation ultrasound data with the ground-truth echogenicity map extracted from real photographic images. The performance of the trained network is evaluated on the publicly available simulation and in vivo test data without any further fine-tuning. Simulation test results show an improvement of 37.5% and 65.8% in terms of axial and lateral resolution as compared to Delay-And-Sum (DAS) results, respectively. The contrast is also improved by 33.7% in comparison to DAS. Furthermore, the reconstructed in vivo images confirm that the trained mapping function does not need any fine-tuning in the new domain. Therefore, the proposed approach maintains high resolution, contrast, and framerate simultaneously.
Collapse
Affiliation(s)
- Sobhan Goudarzi
- Department of Electrical and Computer Engineering, Concordia University, Montreal, QC, Canada.
| | - Hassan Rivaz
- Department of Electrical and Computer Engineering, Concordia University, Montreal, QC, Canada
| |
Collapse
|
11
|
Hyun D, Wiacek A, Goudarzi S, Rothlubbers S, Asif A, Eickel K, Eldar YC, Huang J, Mischi M, Rivaz H, Sinden D, van Sloun RJG, Strohm H, Bell MAL. Deep Learning for Ultrasound Image Formation: CUBDL Evaluation Framework and Open Datasets. IEEE TRANSACTIONS ON ULTRASONICS, FERROELECTRICS, AND FREQUENCY CONTROL 2021; 68:3466-3483. [PMID: 34224351 PMCID: PMC8818124 DOI: 10.1109/tuffc.2021.3094849] [Citation(s) in RCA: 25] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
Deep learning for ultrasound image formation is rapidly garnering research support and attention, quickly rising as the latest frontier in ultrasound image formation, with much promise to balance both image quality and display speed. Despite this promise, one challenge with identifying optimal solutions is the absence of unified evaluation methods and datasets that are not specific to a single research group. This article introduces the largest known international database of ultrasound channel data and describes the associated evaluation methods that were initially developed for the challenge on ultrasound beamforming with deep learning (CUBDL), which was offered as a component of the 2020 IEEE International Ultrasonics Symposium. We summarize the challenge results and present qualitative and quantitative assessments using both the initially closed CUBDL evaluation test dataset (which was crowd-sourced from multiple groups around the world) and additional in vivo breast ultrasound data contributed after the challenge was completed. As an example quantitative assessment, single plane wave images from the CUBDL Task 1 dataset produced a mean generalized contrast-to-noise ratio (gCNR) of 0.67 and a mean lateral resolution of 0.42 mm when formed with delay-and-sum beamforming, compared with a mean gCNR as high as 0.81 and a mean lateral resolution as low as 0.32 mm when formed with networks submitted by the challenge winners. We also describe contributed CUBDL data that may be used for training of future networks. The compiled database includes a total of 576 image acquisition sequences. We additionally introduce a neural-network-based global sound speed estimator implementation that was necessary to fairly evaluate the results obtained with this international database. The integration of CUBDL evaluation methods, evaluation code, network weights from the challenge winners, and all datasets described herein are publicly available (visit https://cubdl.jhu.edu for details).
Collapse
|
12
|
Qu X, Yan G, Zheng D, Fan S, Rao Q, Jiang J. A Deep Learning-Based Automatic First-Arrival Picking Method for Ultrasound Sound-Speed Tomography. IEEE TRANSACTIONS ON ULTRASONICS, FERROELECTRICS, AND FREQUENCY CONTROL 2021; 68:2675-2686. [PMID: 33886467 DOI: 10.1109/tuffc.2021.3074983] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Ultrasound sound-speed tomography (USST) has shown great prospects for breast cancer diagnosis due to its advantages of nonradiation, low cost, 3-D breast images, and quantitative indicators. However, the reconstruction quality of USST is highly dependent on the first-arrival picking of the transmission wave. Traditional first-arrival picking methods have low accuracy and noise robustness. To improve the accuracy and robustness, we introduced a self-attention mechanism into the bidirectional long short-term memory (BLSTM) network and proposed the self-attention BLSTM (SAT-BLSTM) network. The proposed method predicts the probability of the first-arrival time and selects the time with maximum probability. A numerical simulation and prototype experiment were conducted. In the numerical simulation, the proposed SAT-BLSTM showed the best results. For signal-to-noise ratios (SNRs) of 50, 30, and 15 dB, the mean absolute errors (MAEs) were 48, 49, and 76 ns, respectively. The BLSTM had the second-best results, with MAEs of 55, 56, and 85 ns, respectively. The MAEs of the Akaike information criterion (AIC) method were 57, 296, and 489 ns, respectively. In the prototype experiment, the MAEs of the SAT-BLSTM, the BLSTM, and the AIC were 94, 111, and 410 ns, respectively.
Collapse
|