Abstract
Successful ultrasound-guided supraclavicular block (SCB) requires the understanding of sonoanatomy and identification of the optimal view. Segmentation using a convolutional neural network (CNN) is limited in clearly determining the optimal view. The present study describes the development of a computer-aided diagnosis (CADx) system using a CNN that can determine the optimal view for complete SCB in real time. The aim of this study was the development of computer-aided diagnosis system that aid non-expert to determine the optimal view for complete supraclavicular block in real time. Ultrasound videos were retrospectively collected from 881 patients to develop the CADx system (600 to the training and validation set and 281 to the test set). The CADx system included classification and segmentation approaches, with Residual neural network (ResNet) and U-Net, respectively, applied as backbone networks. In the classification approach, an ablation study was performed to determine the optimal architecture and improve the performance of the model. In the segmentation approach, a cascade structure, in which U-Net is connected to ResNet, was implemented. The performance of the two approaches was evaluated based on a confusion matrix. Using the classification approach, ResNet34 and gated recurrent units with augmentation showed the highest performance, with average accuracy 0.901, precision 0.613, recall 0.757, f1-score 0.677 and AUROC 0.936. Using the segmentation approach, U-Net combined with ResNet34 and augmentation showed poorer performance than the classification approach. The CADx system described in this study showed high performance in determining the optimal view for SCB. This system could be expanded to include many anatomical regions and may have potential to aid clinicians in real-time settings.
Trial registration The protocol was registered with the Clinical Trial Registry of Korea (KCT0005822, https://cris.nih.go.kr).
Similar content being viewed by others
Introduction
Supraclavicular block (SCB) is useful for both surgical anesthesia and perioperative analgesia in patients undergoing upper limb surger1. Ultrasound-guided SCB was shown to be safer than landmark or nerve stimulator techniques, as it alleviates concerns about the proximity of the brachial plexus (BP) to the pleura and reduces inadvertent vascular punctures2,3. Ultrasound can easily visualize the BP, lying close (postero-lateral) to the pulsatile subclavian artery (SA) above the hyperechoic first rib at the supraclavicular fossa4. In this position, all the components of the BP are surrounded by a sheath at a shallow depth just above the clavicle. SCB, also called the ‘spinal of the arm’, can therefore anesthetize almost the entire upper extremity1.
However, simply using ultrasound alone does not guarantee the success of SCB2 or avoid complications such as pneumothorax5. Optimising the ultrasound view for SCB requires the probe should be angled or tilted until a lower trunk is visualized laterally to the SA and above the clavicle6. This so-called corner pocket (CP) approach7 is a common and safe procedure with a low risk of having pneumothorax8; however, few studies have focused on classifying the optimal view for SCB.
Ultrasound-guided regional anesthesia (UGRA) has been a useful technique for perioperative anesthesia as it can prevent inadvertent injuries caused by needles1. Recently, AI-assisted UGRA has become a novel technology that aids the process of local anesthetic injection to peripheral nerves. The greatest advantage of a deep learning-based computer-aided diagnosis (CADx) system is that it is suitable for the increasing demand for training UGRA for residents9,10,11. Feedback by the CADx system can be helpful for learners who do not have much access to guided training12,13. Bowness, James et al. developed a U-Net based ScanNav that performs segmentation on seven different anatomical structures, including the supraclavicular region, using 253 scan videos obtained from 244 healthy individuals14. Notably, ScanNav visualizes a color overlay on real-time ultrasound to highlight key anatomical structures, specifically, brachial plexus (BP) region. It may support non-experts in training and clinical practice, and experts in teaching UGRA15. However, the issue of clinicians having to perform time-consuming labelling work remains a challenge for conducting the segmentation task.
Tyagi et al. developed a segmentation model with dice coefficient performances of 0.84 and 0.94 in the Supraclavicular and interscalene areas, respectively, using approximately 35,000 frames of ultrasound images obtained from 196 patients16. However, while showing high performance on images from a single ultrasound device, the model has limitations in its generality, as it performs poorly on lower quality images from different ultrasound devices.
To handle current limitations on AI-assisted UGRA for SCB, we suggest a deep learning-based CADx system that can distinguish between optimal and non-optimal views for the CP approach. The CADx systems were developed with a classification approach that trains all ultrasound images, distinguishes between optimal and non-optimal images, and a segmentation approach that trains images by designating anatomical structures, as in previous studies. Our CADx system showed considerable accuracy and robustness across three ultrasound machines. Furthermore, it can visualize the optimality score and semantic segmentation results in a real-time environment.
Methods
Study design
The CADx systems were developed using a convolutional neural network (CNN) to detect the optimal view for ultrasound-guided SCB. The developed systems utilized a classification approach and a segmentation approach (Fig. 1). This single-centre, retrospective study was approved by the institutional review board of Chungnam National University Hospital (CNUH, Daejeon, Korea, CNUH IRB 2022-05-071, Chairperson Prof. Jeong Lan Kim) on 10 June 2022, with the modified protocol approved on 7 November 2022. Due to the retrospective nature of the study, CNUH waived the need of obtaining informed consent. The protocol was registered with the Clinical Trial Registry of Korea (KCT0007482, https://cris.nih.go.kr). All experiments were performed in accordance with relevant guidelines and regulations.
Ultrasound image acquisition and curation
All patients at CNUH underwent routine sequential ultrasound imaging technique (SUIT) before any BP block6, as SUIT has been shown useful to identify individual elements of the BP and vascular structures above the clavicle17. Patients were maintained in a supine position with the head turned to the contralateral side and the ipsilateral shoulder slightly elevated with a pillow (Supplementary Fig. 1a). The probe was applied to the upper part of the interscalene groove and slid downward from the fifth cervical nerve root (C5) to the first thoracic nerve root (T1) until the complete BP was evaluated on the first rib at the supraclavicular fossa; the probe was subsequently slid upward in the reverse direction to the interscalene level6.
The optimal image for each SCB was defined as an image that enabled visualization of the corner pocket by the first rib, the SA and the neural component (Supplementary Fig. 1b)7. Optimal and non-optimal images were distinguished by several ultrasound anatomic characteristics. In optimal images, the SA presented large and round hypoechoic appearance with clear hyperechoic rim. In addition, while sliding the probe from the cephalad to caudad direction, the first rib was visible from lateral to medial under the BP. The BP appeared as a honeycomb mass or small hypoechoic clustered structure and was located on the inferolateral side of the SA and completely above the first rib. Both the nerve itself and the relationship between each BP and its adjacent structures were important in determining the optimal view6.
Patient data at CNUH were collected using a video capture device (SurgBox, MTEG Co. Ltd. Korea) and stored on a video archiving and communication system (VACS; MTEG Co. Ltd.)18. The ultrasound videos were reviewed by two regional anaesthesiologists, who distinguished between optimal and non-optimal imaging datasets and selected well-focused, high-quality videos of appropriate brightness and depth. All patient-identifying information on ultrasound images was pseudonymized.
Dataset
Ultrasound videos were collected retrospectively from 881 patients scheduled to undergo elective surgery at CNUH from January 2019 to August 2022. Videos of image quality too poor to detect BP were excluded. The images were sampled from the videos to construct datasets for training, validation, and test steps. The sampling rate was a half second which is lower than the minimal unit of the labeled classification interval.
The method proposed in this study consists of classification and segmentation approaches. For the training and validation datasets, the former used all frames as labels, while the latter used masks from a single frame of the optimal view as labels. Therefore, in the classification approach, the labels consist of frames from both optimal and non-optimal views in the videos, and in the segmentation approach, masks from one section of the optimal view are used as labels, which corresponds to the number of ultrasound videos.
The training and validation dataset consisted of 600 ultrasound videos obtained with a high-resolution ultrasound system (X-Porte, FUJIFILM SonoSite, Inc., Bothell, USA) and HFL50xp 15–6 MHz probe (X-Porte). The training set consisted of 3060 optimal images and 28,085 non-optimal images from 506 videos. The validation set consisted of 618 optimal images and 5799 non-optimal images from 94 videos (Supplementary Fig. 2).
Three test datasets were prepared. For generalizability, two datasets were constructed from images taken by ultrasound machines different from the training dataset. The test set 1 consisted of 1030 optimal images and 4532 non-optimal images from 100 videos obtained from an additional 100 patients using the X-Porte ultrasound system. The test set 2 consisted of 836 optimal images and 7718 non-optimal images from 100 videos obtained from 100 patients with a Venue Go ultrasound machine (GE Healthcare, Florida, USA) and a 12–4 MHz linear probe. The test set 3 consisted of 755 optimal images and 5032 non-optimal images obtained from 81 videos of 81 patients with a TE 7 ultrasound machine (Mindray, Shenzhen, China) and a 11–3 MHz linear probe. The compositions of these datasets, including the demographic characteristics of included patients, are shown in Table 1.
Preprocessing
The images from the datasets were processed before being fed into the CADx system. For the CNN training, the acquired ultrasound images were resized to 224 × 224 pixels. Adapted from the work by Pi et al.19, we designed an automated background removal algorithm that can be deployed in a real-time setting (Supplementary Fig. 6). To improve performance during the training phase, augmentation techniques, such as random cropping, noise addition, blurring, affine transformation, non-linear spatial transformation, and adjusting brightness and contrast were randomly applied. The software used for image preprocessing was OpenCV (version 4.5.5.64) and an image augmentation library called Albumentations20. Examples of the augmentation results are shown in Supplementary Fig. 3.
Development of CNN for determining optimal views: classification approach
Residual neural network (ResNet) was selected for the backbone network of the CADx system for its excellence in classification tasks despite of its relatively small model size. The performances of the ResNet models were compared by training various models from shallow to deep layers.
The ResNet model has a CNN encoder part and a fully connected (FC) layer. An optimality score for SCB is then computed from the output of the FC layer normalized by a sigmoid function21. As proposed by Van Boxtel et al.22, ResNet (18, 34, 50, 101, and 152 layers) models, pre-trained with the ImageNet dataset, were trained with a binary cross entropy loss. As the number of optimal view images was far less than the non-optimal view, the weight for the optimal view class was multiplied by the ratio of the number of optimal to non-optimal view examples. Among the ResNet models, the one that performed best according to our evaluation method was selected for further modifications to improve its classification performance.
Recurrent neural network
In another model, it was devised that the encoder branch of the ResNet was connected to a recurrent neural network (RNN) as shown in Fig. 1c. The features of sequential images extracted by the encoder were taken as serial inputs to RNN to capture time-dependent features23. For the classification task, only the output of the last RNN unit was used to classify its label (many-to-one). Similar to the proposed structure by Chen et al.24 where CNN was jointly combined with a long short-term memory (LSTM) network25, we adopted the gated recurrent units which are known to require less memory than LSTM but can deal with the vanishing gradient problem of a vanilla RNNs26. The extracted features were embedded into a latent space of 128 dimensions. The hidden dimension and the number of nodes of RNN were set to 256 and 8. Finally, an FC layer was joined to the output of the last node for the optimal view classification.
Supervised contrastive learning
One of our proposed networks utilized supervised contrastive learning27 before getting trained for the optimal view classification task as shown in Fig. 1b. We utilized the labels of data to better segregate the ultrasound images in a feature space similar in essence to self-supervised learning that learns the representations of unlabeled data28. Using the method in Khosla et al.27, the encoder network was pre-trained with the supervised contrastive learning loss, and then the linear classifier was jointly trained with the encoder.
Development of CNN for determining optimal views: segmentation approach
The segmentation approach was a CADx system that predicts the regions of the SA, first rib and BP to determine whether the view is optimal or not. Because the optimal view could not be directly determined from the predicted result of the segmentation model, the classification model was applied as a cascaded structure after application of the semantic segmentation model. U-Net29, as a semantic segmentation network was combined with the ResNet encoder (Fig. 1d). This approach is similar to Van Boxtel et al.22, but basically different in that the segmentation network is followed by the classification network. The predicted segmentation maps, along with the original ultrasound images, were reused as input to the classification model, as pseudo-labels.
Implementation details
All the models except for the segmentation network were trained maximum of 30 epochs, with an initial learning rate of 0.0001 with a decay of 0.1 for every 10 epochs. The segmentation network was trained maximum of 100 epochs, with an initial learning rate of 0.001 with a decay of 0.1 for every 25 epochs. In all cases, a batch size of 64 and an Adam optimizer were used30. The models were developed on an NVIDIA V100 GPU and were implanted by Python (version 3.9.; Python Software Foundation, Beaverton, OR) and PyTorch (version 1.11.0) software. All source codes of this study are uploaded on https://github.com/nistring/Ultrasound-Optimal-View-Detection.
Evaluation metrics
On test datasets, the labeled ground truth was compared with the model’s prediction. Receiver operating characteristic (ROC) curves and precision-recall (PR) curves were plotted with the areas under the ROC (AUROC) and PR curves (AUPRC). Since AUPRC is more appropriate than AUROC when there’s a highly imbalanced dataset31, AUPRC was mainly used as a metric for comparing the models’ performances. Subsequently, accuracy, precision, recall, and f1-score were calculated on an optimality threshold value where the f1-score had the highest value on ROC curves. The Inference speed was measured by a unit of frames per second (fps).
In the inference phase, gradient-weighted class activation mapping (Grad-CAM)32 was applied as described in Hassanien et al.33. The Grad-CAM was calculated by the weights of the last convolutional layer of CNN to identify the region within each ultrasound image that influenced the CNN's prediction (Fig. 2b).
Results
Comparative performances among classification approaches
In this work, we sought to discover the best backbone layer for predicting the optimal view for SCB among the ResNet with different number of layers. Comparing performances of ResNet with different number of layers (18, 34, 50, 101, and 152 layers), AUROC and AUPRC didn't significantly improve and began to level off for the networks deeper than ResNet34 (Supplementary Fig. 4).
Thus, ResNet34 was selected as the backbone model and several approaches was evaluated. (Table 2). The predicted probability on test-set lower than the threshold value was classified as non-optimal view. The threshold value was 0.81 obtained when the f1-score on validation set was highest. It is evident that the image augmentation made the model more generalizable to images from other US machines (Table 2; Fig. 3). ResNet34 with GRU and augmentation showed the highest performance, with a mean accuracy of 0.901, a mean precision of 0.613, a mean recall of 0.757, a mean f1-score of 0.677, and a mean AUROC of 0.936. This method, however, showed the lowest performance at 153.2 fps (Table 2; Fig. 3b).
By comparison, ResNet34 with supervised contrasting learning (SCL) and augmentation showed a mean accuracy of 0.899, a mean precision of 0.606, a mean recall of 0.74, a mean f1-score of 0.666 and a mean AUROC of 0.932, with no significant differences compared with ResNet34 with GRU and augmentation, and a highest inference speed of 910.8 fps (Table 2; Fig. 3b).
Comparative performances of the classification and segmentation approaches
In the segmentation approach, it was evident that ResNet with deeper layer (50, 101, and 152) was never better than ResNet34, and this is consistent with results obtained in the classification approach. U-Net with ResNet34 as the final model was evaluated the performance based on the highest probability value on the test set by determining an f1-score threshold of 0.88 in the validation set.
The cascaded model with augmentation showed a mean accuracy of 0.903, a mean precision of 0.67, a mean recall of 0.606, a mean f1-score of 0.635, a mean AUROC of 0.939 and a mean 500.6 fps (Table 2; Fig. 3). This segmentation approach showed poorer performance than the classification approach of ResNet34 with augmentation methods.
Figure 2 shows the qualitative results of optimal view determination of ultrasound-guided SCB block using both the classification and segmentation approaches. Application of the trained CNN model during the inference phase enabled visualization of the gradient-weighted class activation mapping (Grad-CAM)34 results by overlapping the original ultrasound images. Red colour in the heatmap is indicative of higher chance of finding the optimal view around that area (Fig. 2b). Examples of qualitative results of the CADx system with the ultrasound equipment used are shown in Supplementary Videos 1–6. Figure 4 shows examples of evaluation of the performance of each ultrasound video, consisting of comparisons between the CNN prediction of the optimal view section and the ground truth section. CNN with higher overlapping area between the predicted score and the binary ground truth value is proved to demonstrate a better performance. The threshold values obtained from the validation set for the classification and segmentation approaches were 0.81 and 0.88, respectively. When the threshold was exceeded, the model predicted that test sets 1–3 would provide optimal views.
Discussion
SCB is a common procedure usually with the CP approach, which has a low risk of causing pneumothorax8. Despite this, there have been few studies focused on identifying the optimal view for SCB. To address the current limitations of AI-assisted UGRA for SCB, we proposed a CADx system utilizing deep learning technology. The CADx is likely to be a proper model capable of distinguishing between optimal and non-optimal views for the CP approach with a high AUPRC and a decent inference speed. We also confirmed that the CADx system guarantees considerable robustness across three different ultrasound machines.
The CADx systems have been developed in a variety of fields to assist the optimal view determination such as cardiac23, fetal24, breast19 and thyroid35 ultrasound. More specifically, some practical studies about AI-assisted UGRA detect BP at the interscalene level15,22 and visualize the relevant anatomic structures in real time14. However, to our best knowledge, no studies have described methods for optimal view detection for SCB and our study broadens the applicability of the CADx for SCB.
During labelling in videos, large intra-individual variations even within the optimal views were observed and many optimal images did not show large inter-individual variations when compared with non-optimal views. Nevertheless, the most important ultrasound image characteristics distinguishing optimal from non-optimal views was the finding that BP was located lateral or postero-lateral to the SCA on top of the first rib in optimal views. One concern with the classification approach was the lack of transparency of the process used to determine the outcome, as the CNN model was unable to provide an explanation for the outcome36. The Grad-CAM result of the classification approach, however, showed that the heatmap was consistently activated in the area of the first rib. Although it was not intended by clinicians during SCB, the first rib was located in the middle of optimal images. This indirect determination of the method used by the classification model was somewhat consistent with clinical inferences.
The deep learning-based classification and segmentation approaches developed in this study both used ResNet34, enabling the application of a light-weight model for real-time processing in a clinical environment. In general, the number of parameters and the performance of a model tend to be proportional; however, this proportionality was not observed in the ImageNet dataset and other domains such as the chest X-ray dataset37. This trend is qualitatively similar to Guo et al.35 where even ResNet18 was shown to be highly effective enough to recognize a target image. This indicates that there might be no need for large-sized models in the ultrasound classification task, and also it showed similar results (Supplementary Fig. 4). A simple and cost-effective model without additional information presented here was sufficient for the ultrasound image classification task. This result demonstrates the feasibility of the CADx system in a real-time clinical environment.
In this study, we confirmed that the proposed classification approach shows near real-time performance, and for future application in a real clinic environment, two approaches can be considered. The first is embedding the CAD system into the ultrasound device being used, for which a high-performance GPU system would also need to be installed. Another approach is to apply the CAD system to a screen that captures the output from a split channel of the ultrasound monitor. This has the advantage of allowing for simultaneous comparison of the original ultrasound image and the image to which the CAD system has been applied.
In addition, the CADx system showed similar high performance when evaluating test sets acquired from two other ultrasound imaging devices (Venue Go, TE7) and a test set acquired with the same device as the training set (X-Porte). Thus, the CADx system may be applicable to images acquired with many types of ultrasound machines.
In addition to the cascaded architecture proposed in this study, a model that effectively predicts optimal and non-optimal view segments could potentially be developed using a single architecture through multi-task learning (MTL). However, this approach would require labels for both classification and segmentation, making the labeling process in the preparation of ultrasound video training data burdensome. And also, the optimal view for SCB in this study was defined as where relevant anatomical structures (SA, first rib and BP) are simultaneously observed, they are regarded as quintessential and contain useful information on the optimal view. Therefore, the output of the segmentation network was concatenated with the original ultrasound images to be reused as input to the classification network in cascaded structure.
This study had several limitations. First, since this study is a single centre study and the data were analyzed retrospectively, there may have been selection bias. Multi-center, prospective studies are needed to evaluate the generalizability of the developed model. Second, although the developed CADx systems showed high quality performance, their clinical efficacy has not yet been determined. Although these models have been evaluated using quantitative measures, such as accuracy, f1-score, and fps, it is unclear whether these metrics are associated with clinical efficacy. Thus, real-time clinical application of the CNN model is required to determine whether it improves performance outcomes. Finally, the segmentation approach did not measure the dice coefficient for the test set, as the objective of this study was to distinguish between optimal and non-optimal views for ultrasound-guided SCB, so only the classification performance was evaluated quantitatively.
In conclusion, this study described the development of CADx systems, using both classification and segmentation approaches, which could optimally detect corner pocket images for complete SCB. Both approaches showed high performance in detecting optimal views and functioned well in real-time settings. This proposed method may be applicable to various anatomical structures14 and to systems of tracking nerves along their courses and selective trunk identification38,39.
Data availability
The data used during the current study are available from the corresponding author on reasonable request.
Change history
10 November 2023
A Correction to this paper has been published: https://doi.org/10.1038/s41598-023-46973-5
References
Jones, M. R. et al. Upper extremity regional anesthesia techniques: A comprehensive review for clinical anesthesiologists. Best Pract. Res. Clin. Anaesthesiol. 34, e13–e29 (2020).
Perlas, A. et al. Ultrasound-guided supraclavicular block: Outcome of 510 consecutive cases. Reg. Anaesth. Pain Med. 34, 171–176 (2009).
Abrahams, M., Aziz, M., Fu, R. & Horn, J.-L. Ultrasound guidance compared with electrical neurostimulation for peripheral nerve block: A systematic review and meta-analysis of randomized controlled trials. Br. J. Anaesth. 102, 408–417 (2009).
Karmakar, M. Atlas of Sonoanatomy for Regional Anesthesia and Pain Medicine. (McGraw Hill Professional, 2017).
Barrington, M. J. & Uda, Y. Did ultrasound fulfill the promise of safety in regional anesthesia?. Current Opinion in Anesthesiology 31, 649–655 (2018).
Karmakar, M. K., Pakpirom, J., Songthamwat, B. & Areeruk, P. High definition ultrasound imaging of the individual elements of the brachial plexus above the clavicle. Reg. Anesth. Pain Med. 45, 344–350 (2020).
Soares, L. G., Brull, R., Lai, J. & Chan, V. W. Eight ball, corner pocket: the optimal needle position for ultrasound-guided supraclavicular block. Reg. Anesth. Pain Med. 32, 94 (2007).
Gauss, A. et al. Incidence of clinically symptomatic pneumothorax in ultrasound-guided infraclavicular and supraclavicular brachial plexus block. Anesthesia 69, 327–336 (2014).
Kim, T. E. & Tsui, B. C. Simulation-based ultrasound-guided regional anesthesia curriculum for anesthesiology residents. Korean J. Anesthesiol. 72, 13–23 (2019).
Niazi, A. U., Haldipur, N., Prasad, A. G. & Chan, V. W. Ultrasound-guided regional anesthesia performance in the early learning period: Effect of simulation training. Reg. Anesth. Pain Med. 37, 51–54 (2012).
Mendiratta-Lala, M., Williams, T., de Quadros, N., Bonnett, J. & Mendiratta, V. The use of a simulation center to improve resident proficiency in performing ultrasound-guided procedures. Academic radiology 17, 535–540 (2010).
Ramlogan, R. et al. Challenges and training tools associated with the practice of ultrasound-guided regional anesthesia: A survey of the American society of regional anesthesia and pain medicine. Reg. Anesth. Pain Med. 35, 224–226 (2010).
Nix, C. M. et al. A scoping review of the evidence for teaching ultrasound-guided regional anesthesia. Reg. Anesth. Pain Med. 38, 471–480 (2013).
Bowness, J., Varsou, O., Turbitt, L. & Burkett-St Laurent, D. Identifying anatomical structures on ultrasound: assistive artificial intelligence in ultrasound-guided regional anesthesia. Clin. Anat. 34, 802–809 (2021).
Bowness, J. S. et al. Exploring the utility of assistive artificial intelligence for ultrasound scanning in regional anesthesia. Reg. Anesth. Pain Med. 47, 375–379 (2022).
Tyagi, A. et al. Automated Real Time Delineation of Supraclavicular Brachial Plexus in Neck Ultrasonography Videos: A Deep Learning Approach. arXiv preprint arXiv:2308.03717 (2023).
Songthamwat, B., Pakpirom, J., Pangthipampai, P., Vorapaluk, P. & Karmakar, M. K. Reliability of a sequential ultrasound imaging technique (SUIT) to identify the individual elements of the brachial plexus above the clavicle. Reg. Anesth. Pain Med. 46, 1107–1109 (2021).
Kim, D., Hwang, W., Bae, J., Park, H. & Kim, K. G. Video archiving and communication system (VACS): A progressive approach, design, implementation, and benefits for surgical videos. Healthc. Inform. Res. 27, 162–167 (2021).
Pi, Y., Li, Q., Qi, X., Deng, D. & Yi, Z. Automated assessment of BI-RADS categories for ultrasound images using multi-scale neural networks with an order-constrained loss function. Appl. Intell. 52, 12943–12956 (2022).
Buslaev, A. et al. Albumentations: Fast and flexible image augmentations. Information 11, 125 (2020).
He, K., Zhang, X., Ren, S. & Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 770–778 (2016).
Van Boxtel, J., Vousten, V., Pluim, J. & Rad, N. M. Hybrid Deep neural network for brachial plexus nerve segmentation in ultrasound images. In 29th European Signal Processing Conference (EUSIPCO) 1246–1250 (2021).
Howard, J. P. et al. Improving ultrasound video classification: an evaluation of novel deep learning methods in echocardiography. J. Med. Artif. Intell. 3, 66 (2020).
Chen, H. et al. Ultrasound standard plane detection using a composite neural network framework. IEEE Trans. Cybern. 47, 1576–1586 (2017).
Hochreiter, S. & Schmidhuber, J. Long short-term memory. Neural Comput. 9, 1735–1780 (1997).
Cho, K. et al. Learning phrase representations using RNN encoder–decoder for statistical machine translation. arXiv preprint arXiv:1406.1078 (2014).
Khosla, P. et al. Supervised contrastive learning. Adv. Neural Inf. Process. Syst. 33, 18661–18673 (2020).
Jaiswal, A., Babu, A. R., Zadeh, M. Z., Banerjee, D. & Makedon, F. A survey on contrastive self-supervised learning. Technologies 9, 2 (2020).
Ronneberger, O., Fischer, P. & Brox, T. Medical Image Computing and Computer-Assisted Intervention–MICCAI 18th International Conference, Munich, Germany, October 5–9, 2015, Proceedings, Part III, vol. 18 234–241 (2015).
Kingma, D. P. & Ba, J. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).
Qi, Q., Luo, Y., Xu, Z., Ji, S. & Yang, T. Stochastic optimization of areas under precision-recall curves with provable convergence. Adv. Neural Inf. Process. Syst. 34, 1752–1765 (2021).
Gildenblat, J. et al. PyTorch library for CAM methods. GitHub https://github.com/jacobgil/pytorch-grad-cam (2021).
Hassanien, M. A., Singh, V. K., Puig, D. & Abdel-Nasser, M. Predicting breast tumor malignancy using deep ConvNeXt radiomics and quality-based score pooling in ultrasound sequences. Diagnostics 12, 1053 (2022).
Selvaraju, Ramprasaath R., et al. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision 618–626 (2017).
Guo, M. et al. Recognition of thyroid ultrasound standard plane images based on residual network. Comput. Intell. Neurosci. 2021, 1–11 (2021).
Chen, J. & See, K. C. Artificial intelligence for COVID-19: rapid review. J. Med. Internet Res. 22, e21476 (2020).
Ke, A., Ellsworth, W., Banerjee, O., Ng, A. Y. & Rajpurkar, P. CheXtransfer: Performance and parameter efficiency of ImageNet models for chest X-ray interpretation. In Proceedings of the Conference on Health, Inference, and Learning 116–124 (2021).
Jo, Y. et al. Comparison of the ulnar nerve blockade between intertruncal and corner pocket approaches for supraclavicular block: a randomized controlled trial. Korean J. Anesthesiol. 74, 522–530 (2021).
Karmakar, M. K., Areeruk, P., Mok, L. Y. & Sivakumar, R. K. Ultrasound-guided selective trunk block to produce surgical anesthesia of the whole upper extremity: a case report. A&A Pract. 14, e01274 (2020).
Funding
This work received support from the Chungnam National University Hospital Research Fund (2020-CF-005), the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (NRF-2022M3J6A1084843) and was the result of a study on the “HPC Support” Project, supported by the ‘Ministry of Science and ICT’ and ‘National IT Industry Promotion Agency (NIPA)’.
Author information
Authors and Affiliations
Contributions
Study concept/design: Y.J., B.H., Y.S.S. Data collection: Y.J., B.H., J.J. Data processing: D.L., D.B., B.K.C., N.A. Data analysis and interpretation: D.L., D.B., B.K.C., N.A. Manuscript preparation: all authors. Manuscript editing: Y.S.S., Y.J., B.H., D.L., D.B. Final approval of this version to be submitted: all authors.
Corresponding authors
Ethics declarations
Competing interests
Bo Kyung Choi, Nisan Aryal are staff members of commercial organization (MTEG Co. Ltd.) and involved in a AI Voucher support project as co-authors supported from ‘National IT Industry Promotion Agency (NIPA).’ But, they have no financial or personal conflict of interest relevant to this article. All the others has no competing interest.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The original online version of this Article was revised: The Funding section in the original version of this Article was incomplete. It now reads: “This work received support from the Chungnam National University Hospital Research Fund (2020-CF-005), the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (NRF-2022M3J6A1084843) and was the result of a study on the “HPC Support” Project, supported by the ‘Ministry of Science and ICT’ and ‘National IT Industry Promotion Agency (NIPA)’.”
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Jo, Y., Lee, D., Baek, D. et al. Optimal view detection for ultrasound-guided supraclavicular block using deep learning approaches. Sci Rep 13, 17209 (2023). https://doi.org/10.1038/s41598-023-44170-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-023-44170-y