Skip to main content

Convolutional neural network for brachial plexus segmentation at the interscalene level

Abstract

Background

Regional anesthesia with ultrasound-guided brachial plexus block is widely used for patients undergoing shoulder and upper limb surgery, but needle misplacement can result in complications. The purpose of this study was to develop and validate a convolutional neural network (CNN) model for segmentation of the brachial plexus at the interscalene level.

Methods

This prospective study included patients who underwent ultrasound-guided brachial plexus block in the Anesthesiology Department of Beijing Jishuitan Hospital between October 2019 and June 2022. A Unet semantic segmentation model was developed to train the CNN to identify the brachial plexus features in the ultrasound images. The degree of overlap between the predicted segmentation and ground truth segmentation (manually drawn by experienced clinicians) was evaluated by calculation of the Dice index and Jaccard index.

Results

The final analysis included 502 images from 127 patients aged 41 ± 14 years-old (72 men, 56.7%). The mean Dice index was 0.748 ± 0.190, which was extremely close to the threshold level of 0.75 for good overlap between the predicted and ground truth segregations. The Jaccard index was 0.630 ± 0.213, which exceeded the threshold value of 0.5 for a good overlap.

Conclusion

The CNN performed well at segregating the brachial plexus at the interscalene level. Further development could allow the CNN to be used to facilitate real-time identification of the brachial plexus during interscalene block administration.

Clinical trial registration

The trial was registered prior to patient enrollment at the Chinese Clinical Trial Registry (ChiCTR2200055591), the site url is https://www.chictr.org.cn/. The date of trial registration and patient enrollment is 14/01/2022.

Peer Review reports

Background

The brachial plexus is a network of nerves that provides sensory and motor innervation to the upper chest, shoulder, arm, forearm and hand [1]. The brachial plexus is formed by the ventral rami of cervical nerves C5–C8 and thoracic nerve T1 and extends from the spine to pass above the first rib and through the posterior triangle of the neck between the anterior and middle scalene muscles to reach the axilla [2]. The brachial plexus emerges from the spine as five roots (C5–C8 and T1) that merge to form three trunks: superior (C5 and C6), middle (C7) and inferior (C8 and T1).

Detailed knowledge of the brachial plexus anatomy is essential not only to surgeons managing brachial plexus injuries [3] but also to anesthetists performing brachial plexus block [4]. Regional anesthesia with brachial plexus block is widely used as an adjunct or alternative to general anesthesia during upper limb surgery [5,6,7,8]. However, most peripheral nerve blocks are performed by a limited number of specialists, which reflects the challenges and complexities of these procedures [9]. The success rate of ultrasound-guided brachial plexus block is 77–94% [9,10,11], and the procedure is associated with a risk of complications such as neurological symptoms, systemic toxicity due to unintended blood vessel puncture, and pneumothorax [12]. Widespread acceptance of ultrasound-guided brachial plexus block is limited by the difficulty of learning skills related to the acquisition and interpretation of images using ultrasound and needle guidance under ultrasound guidance.

The use of artificial intelligence (AI) in medicine is increasing, and research is ongoing into the potential clinical applications of AI in anesthesiology [13] and medical imaging [14]. Since ultrasound-guided peripheral nerve block relies on imaging, one possible use of AI would be to improve image optimization and interpretation in real-time, thereby helping physicians to recognize the target nerve(s) and avoid complications due to puncture of other nerves, arteries, veins or pleura.

Convolutional neural networks (CNNs) can help identify target structures of interest by learning from a large number of labeled ultrasound images. Since ultrasound scans are two-dimensional grayscale moving images, they are not easy for novices to identify. But AI can provide color landmarks as a reference, which may lead to a wider range of applications for ultrasound-guided nerves. [Applying artificial intelligence to the use of ultrasound as an educational tool: A focus on Ultrasound-guided regional anesthesia bowness BJA] And AI advances may improve skills in ultrasound image interpretation.

Therefore, the aim of this study was to develop and validate a CNN model for image segmentation that would be capable of recognizing the brachial plexus at the interscalene level. A Unet semantic segmentation model was developed to train the CNN to identify the features of the brachial plexus in ultrasound images so that it learned to automatically detect and segregate the brachial plexus.

Methods

Study design and patients

This prospective study included patients who underwent ultrasound-guided brachial plexus block in the Anesthesiology Department of Beijing Jishuitan Hospital (Beijing, China) between October 2019 and June 2022. The inclusion criteria were as follows: (1) aged 18–65 years-old; (2) scheduled to receive ultrasound-guided brachial plexus block as anesthesia for orthopedic surgery, anesthesia for closed fracture reduction, or pain therapy; (3) American Society of Anesthesiologists (ASA) physical status classification stage I–III; and (4) volunteered to participate in this study. The exclusion criteria were: (1) incapable of cooperating with the study protocol or maintaining the required position for imaging; (2) brachial plexus dysfunction; (3) previous surgery to the nerve segment to be scanned; (4) wound or infection of the neck or at the interscalene level; and (5) allergic to the ultrasonic coupling agent. The study was approved by the Ethics Committee of Beijing Jishuitan Hospital (approval no. 202104-14), and all patients provided informed written consent.

Baseline demographic and clinical characteristics

The following baseline demographic and clinical characteristics were obtained: age, sex, height and body weight.

Procedure

The patient was placed in the supine position with their head slightly turned to the contralateral side. A high-frequency probe was used, and the medial side was displayed on the right-hand side of the image. After pre-scanning of the image acquisition area to optimize the ultrasound parameters, continuous scanning was performed from the supraclavicular area (start point) to the level of the transverse process of C5 (end point). The probe was fine-tuned to optimize the image before it was acquired, and the probe was maintained in a stable position for 1 s before and after the button was pressed for image acquisition. A cross-sectional image was acquired at the supraclavicular region, at the region between the supraclavicular region and the level of the transverse process of C7, and at the level of the transverse process of C7 (Fig. 1).

Fig. 1
figure 1

Ultrasound images acquired at various levels. (A) Images acquired at the supraclavicular region. The area containing nerves (and including as many nerves as possible) but excluding the middle scalene muscle was marked (N, right-hand images). The subclavian artery was also marked (SA). (B) Images acquired at a region between the supraclavicular area and the level of the transverse process of C7. The area containing all the nerves (C5–C7 as well as C8 if it was clearly visible on the surface of the first rib) was marked (N). (C) Images acquired at the level of the transverse process of C7. The area containing the nerves (C5 and C6 plus C7 if the surface of the transverse process of C7 was clearly displayed) was marked (N)

Quality control of the images

For quality testing, 10% of the images were randomly selected after acquisition or marking. The images were reviewed by three physicians, each of which had at least 5 years of experience in clinical practice. Images with the essential structures displayed and marked relatively precisely were considered to be of high quality; the criteria used to define a poor quality image are shown in Supplementary Table 1. An image was defined as being of high quality if two of the three reviewers classified it as high quality (i.e., not poor quality).

Establishment of the CNN

The principle used for brachial plexus extraction mimicked the process of brachial plexus recognition by human physicians. The process consisted of multiple convolution and pooling layers to detect local features followed by fully connected layers, with the softmax function used in the final layer prior to output (Supplementary Figure S1). A large number of high-quality images were input into the pool in different batches for computer learning and recognition.And a Unet model was used for segregation. The results of the recognition process were then validated.

Validation of the CNN

This study evaluates the accuracy of image separation by calculating the S ø rensen Dice coefficient (Dice index) and Jaccard index, both of which are statistics that estimate the similarity between two samples and are used in image segmentation to compare algorithm output against a reference mask.

The Dice index and Jaccard index were determined according to the degree of overlap between the segmentation predicted by the CNN and the ground truth segmentation manually drawn by the physicians. The Dice index was calculated as (2 × area of the overlap between the predicted segmentation and the ground truth segmentation) / (area of the predicted segmentation + area of the ground truth segmentation) (Fig. 2A). The Dice index had a value of 0–1, and a value ≥ 0.75 was taken to indicate good overlap between the predicted and ground truth segmentations, i.e., good segregation by the CNN. The Jaccard index was calculated as (area of the overlap between the predicted segmentation and ground truth segmentation) / (area of the region formed by merger of the predicted segmentation and ground truth segmentation) (Fig. 2B). A Jaccard value > 0.5 was taken to indicate a good overlap.

Fig. 2
figure 2

Metrics used to validate the convolutional neural network. (A) Calculation of the Dice index. (B) Calculation of the Jaccard index

Statistical analysis

SPSS Statistics 27.0 software (IBM Corp., Armonk, NY, USA) was used for the analysis. Normally distributed quantitative data are described as the mean ± standard deviation. Non-normally distributed quantitative data are presented as the median (range). Qualitative data are described as frequency (constituent ratio or percentage).

Results

Baseline characteristics of the study participants

The final analysis included 502 images from 127 patients aged 41 ± 14 years-old (72 men, 56.7%). The mean height and mean body weight of the participants was 168.10 ± 8.29 cm and 67.65 ± 13.13 kg, respectively (Table 1).

Table 1 Baseline characteristics of the study participants

Establishment of CNN

The neural and vascular structures in ultrasound images can be recognized after training using the Unet semantic segmentation model (Fig. 1). In this model, we collected ultrasound images of the intermuscular sulcus area, and it can be seen that the brachial plexus nerves in different positions of the intermuscular sulcus are well recognized (Fig. 1S).

Dice index and Jaccard index

The mean Dice index of the CNN for segregation of the brachial plexus at the interscalene level was 0.748 ± 0.190, which was extremely close to the threshold level of 0.75 for good overlap between the predicted and ground truth segregations (Table 2). Furthermore, the Jaccard index was 0.630 ± 0.213, which exceeded the threshold value of 0.5 for a good overlap (Table 2). The Dice index was mainly distributed around a value of 0.8, while the Jaccard index was mainly distributed around a value of 0.6 (Fig. 3).

Table 2 Dice index and Jaccard index
Fig. 3
figure 3

Distribution of the Dice index and Jaccard index. Each circle is color-coded according to the number of number of segmentations with a particular Dice Index and Jaccard index. The color represents the frequency of occurrence and the darker the color, the more IOU and DICE appear. IOU: Jaccard index (also known as the Intersection-Over-Union).

Discussion

The main finding of this study was that the developed CNN was able to segregate the brachial plexus in ultrasound images obtained at the interscalene level with a mean Dice index of 0.748 and a mean Jaccard index of 0.630. These results indicate that the CNN established in this study performed well in the segregation of the brachial plexus at the interscalene level. We anticipate that, after further development, the CNN could be used to train ultrasonographers and anesthetists to recognize the brachial plexus and to facilitate the identification of the brachial plexus of a patient in real-time during ultrasound-guided interscalene block.

An interscalene block is generally performed on patients undergoing surgery of the shoulder, upper arm or elbow but is not recommended for hand operations because it tends to spare the inferior trunk [15]. However, incorrect needle placement can result in complications such as pneumothorax, nerve injury, epidural or intrathecal injection of local anesthetic, spinal cord trauma, shortness of breath due to unintended phrenic nerve block, and Horner syndrome due to unintended cervical plexus block [16]. The present study developed a SegNet semantic segmentation model to train the CNN to identify the brachial plexus features in ultrasound images, so that it learned to automatically segregate the brachial plexus in ultrasound images obtained at the interscalene level. Notably, the CNN algorithm performed well in the segregation of the brachial plexus at the interscalene level, as evidenced by a Dice index of 0.748 (very close to the threshold level of 0.75 for a good overlap with the ground truth segmentation) and a Jaccard index of 0.630 (which exceeded the threshold of 0.5 for a good overlap). The performance of our CNN compares well with that of algorithms for brachial plexus segregation reported by other studies. Liu et al. developed a deep adversarial network comprising a segmentation network and discriminator network, and the Jaccard index for brachial plexus segmentation was 0.693 without elastic transformation and 0.733 with elastic transformation [17]. Moreover, the authors reported that the Jaccard values for their deep adversarial network were numerically higher than those reported previously for a support vector machine and conditional random fields model (0.454) [18], a fully convolutional network (0.572) [19], a CNN incorporating a conditional random field (0.693) [20], piecewise training of conditional random fields with CNN pairwise potentials (0.716) [21], and a recurrent convolutional neural network (0.721) [22]. Wang et al. used preprocessing and a modified U-Net architecture to generate a model that segregated the brachial plexus with a Dice index of 0.709 [23]. Additionally, a deep learning model described by Yang et al. was able to locate the interscalene brachial plexus in ultrasound images more accurately than nonexperts, as evidenced by the distance between the lateral midpoints of the nerve sheath contours of the model predictions and the ground truth (0.8 mm for the model and 3.4 mm for non-experts, P < 0.001) [24].

This study has some limitations. First, the CNN was designed for brachial plexus segmentation at the interscalene level, so it remains to be determined whether the algorithm would also be capable of brachial plexus segmentation at the supraclavicular, infraclavicular and axillary levels. Second, only the brachial plexus was targeted, hence it was not established whether the CNN could be used for segmentation of other peripheral nerves. Third, patients with brachial plexus dysfunction were excluded, so further research is required to evaluate whether the algorithm would be suitable for use in patients with brachial plexus dysfunction or anatomic abnormalities. Fourth, the study only included a training set and did not utilize a validation set. Additional studies are merited to further develop and validate the CNN described in this study.

Conclusion

In conclusion, this study describes an efficient CNN that exhibited good performance in the segmentation of the brachial plexus at the interscalene level, with a Dice index of 0.748 and a Jaccard index of 0.630. We anticipate that this CNN could be developed into a useful tool to train clinicians to recognize the brachial plexus in ultrasound images and to assist anesthetists performing brachial plexus block.

Data availability

All data generated or analyzed during this study are included in this article and supplementary information files.

Abbreviations

CNN:

Convolutional neural network

AI:

Artificial intelligence

ASA:

American Society of Anesthesiologists

References

  1. Polcaro L, Charlick M, Daly DT, Anatomy. Head and Neck, Brachial Plexus. StatPearls. Treasure Island (FL): StatPearls Publishing Copyright © 2022. StatPearls Publishing LLC.; 2022.

  2. Gilcrease-Garcia BM, Deshmukh SD, Parsons MS. Anatomy, imaging, and pathologic conditions of the Brachial Plexus. Radiographics. 2020;40:1686–714. https://doi.org/10.1148/rg.2020200012.

    Article  PubMed  Google Scholar 

  3. Noland SS, Bishop AT, Spinner RJ, Shin AY. Adult traumatic brachial plexus injuries. J Am Acad Orthop Surg. 2019;27:705–16. https://doi.org/10.5435/jaaos-d-18-00433.

    Article  PubMed  Google Scholar 

  4. Feigl GC, Litz RJ, Marhofer P. Anatomy of the brachial plexus and its implications for daily clinical practice: regional anesthesia is applied anatomy. Reg Anesth Pain Med. 2020;45:620–7. https://doi.org/10.1136/rapm-2020-101435.

    Article  PubMed  Google Scholar 

  5. Iwashita S, Hashiguchi H, Okubo A, Yoneda M, Takai S. Nerve Block for Pain Relief during Arthroscopic Rotator Cuff Repair. J Nippon Med Sch. 2020;87:87–91. https://doi.org/10.1272/jnms.JNMS.2020_87-206.

    Article  PubMed  CAS  Google Scholar 

  6. Bojaxhi E, Lumermann LA, Mazer LS, Howe BL, Ortiguera CJ, Clendenen SR. Interscalene brachial plexus catheter versus single-shot interscalene block with periarticular local infiltration analgesia for shoulder arthroplasty. Minerva Anestesiol. 2019;85:840–5. https://doi.org/10.23736/s0375-9393.19.13387-1.

    Article  PubMed  Google Scholar 

  7. Dhir S, Brown B, Mack P, Bureau Y, Yu J, Ross D. Infraclavicular and supraclavicular approaches to brachial plexus for ambulatory elbow surgery: a randomized controlled observer-blinded trial. J Clin Anesth. 2018;48:67–72. https://doi.org/10.1016/j.jclinane.2018.05.005.

    Article  PubMed  Google Scholar 

  8. Ertikin A, Argun G, Mısırlıoğlu M, Aydın M, Arıkan M, Kadıoğulları N. Comparison between the two-injection technique and the four-injection technique in Axillary Brachial Plexus Block with Articaine. Turk J Anaesthesiol Reanim. 2017;45:289–96. https://doi.org/10.5152/tjar.2017.16023.

    Article  PubMed  PubMed Central  CAS  Google Scholar 

  9. Turbitt LR, Mariano ER, El-Boghdadly K. Future directions in regional anaesthesia: not just for the cognoscenti. Anaesthesia. 2020;75:293–7. https://doi.org/10.1111/anae.14768.

    Article  PubMed  CAS  Google Scholar 

  10. Qin Q, Yang D, Xie H, Zhang L, Wang C. Ultrasound guidance improves the success rate of axillary plexus block: a meta-analysis. Braz J Anesthesiol. 2016;66:115–9. https://doi.org/10.1016/j.bjane.2015.01.002.

    Article  PubMed  Google Scholar 

  11. Schwemmer U, Papenfuss T, Greim C, Brederlau J, Roewer N. Ultrasound-guided interscalene brachial plexus anaesthesia: differences in success between patients of normal and excessive weight. Ultraschall Med. 2006;27:245–50. https://doi.org/10.1055/s-2006-926591.

    Article  PubMed  CAS  Google Scholar 

  12. Abhinaya RJ, Venkatraman R, Matheswaran P, Sivarajan G. A randomised comparative evaluation of supraclavicular and infraclavicular approaches to brachial plexus block for upper limb surgeries using both ultrasound and nerve stimulator. Indian J Anaesth. 2017;61:581–6. https://doi.org/10.4103/ija.IJA_402_16.

    Article  PubMed  PubMed Central  Google Scholar 

  13. Hashimoto DA, Witkowski E, Gao L, Meireles O, Rosman G. Artificial Intelligence in Anesthesiology: current techniques, clinical applications, and limitations. Anesthesiology. 2020;132:379–94. https://doi.org/10.1097/aln.0000000000002960.

    Article  PubMed  Google Scholar 

  14. Giger ML. Machine learning in Medical Imaging. J Am Coll Radiol. 2018;15:512–20. https://doi.org/10.1016/j.jacr.2017.12.028.

    Article  PubMed  Google Scholar 

  15. Yamashita R, Nishio M, Do RKG, Togashi K. Convolutional neural networks: an overview and application in radiology. Insights Imaging. 2018;9:611–29. https://doi.org/10.1007/s13244-018-0639-9.

    Article  PubMed  PubMed Central  Google Scholar 

  16. Zisquit J, Nedeff N. Interscalene Block. StatPearls [Internet] Treasure Island (FL): StatPearls Publishing. 2022.

  17. Liu C, Liu F, Wang L, Ma L, Lu ZM. Segmentation of nerve on ultrasound images using deep adversarial network. Int J Innov Comput Inf Control. 2018;14:53–64. https://doi.org/10.24507/ijicic.14.01.53.

    Article  Google Scholar 

  18. Lucchi A, Li Y, Smith K, Fua P. Structured image segmentation using Kernelized features. Fitzgibbon, Lazebnik A, Perona S, Sato P, Schmid Y, C, editors Computer vision – ECCV 2012 ECCV 2012 lecture notes in Computer Science, vol 7573 Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-33709-3_29.

  19. Shelhamer E, Long J, Darrell T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans Pattern Anal Mach Intell. 2017;39:640–51. https://doi.org/10.1109/tpami.2016.2572683.

    Article  PubMed  Google Scholar 

  20. Arnab A, Jayasumana S, Zheng S, Torr PHS. Higher Order conditional Random fields in deep neural networks. Leibe B, Matas J, Sebe N, Welling M, editors Computer vision – ECCV 2016 ECCV 2016 Lecture Notes in Computer Science, vol 9906 Springer, Cham. https://doi.org/10.1007/978-3-319-46475-6_33.

  21. Lin G, Shen C, van den Hengel A, Reid I. Exploring context with deep structured models for semantic segmentation. IEEE Trans Pattern Anal Mach Intell. 2018;40:1352–66. https://doi.org/10.1109/tpami.2017.2708714.

    Article  PubMed  Google Scholar 

  22. Pinheiro P, Collobert R. Recurrent convolutional neural networks for scene labeling. Proc Mach Learn Res. 2014;32:82–90.

    Google Scholar 

  23. Wang R, Shen H, Zhou M. Ultrasound nerve segmentation of brachial plexus based on optimized ResU-Net. 2019 IEEE International Conference on Imaging Systems and Techniques. 2019.

  24. Yang XY, Wang LT, Li GD, Yu ZK, Li DL, Guan QL, et al. Artificial intelligence using deep neural network learning for automatic location of the interscalene brachial plexus in ultrasound images. Eur J Anaesthesiol. 2022;39:758–65. https://doi.org/10.1097/eja.0000000000001720.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

We are grateful to all the patients who agreed to participate in this study.

Funding

None.

Author information

Authors and Affiliations

Authors

Contributions

YX and GW carried out the studies, participated in collecting data, and drafted the manuscript. YZ and YX performed the statistical analysis and participated in its design. FZ, HC and YY participated in acquisition, analysis, or interpretation of data and draft the manuscript. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Geng Wang.

Ethics declarations

Ethics approval and consent to participate

This work has been carried out in accordance with the Declaration of Helsinki (2000) of the World Medical Association. The study was approved by the Ethics Committee of Beijing Jishuitan Hospital (approval no. 202104-14), and all patients provided informed written consent.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Electronic supplementary material

Below is the link to the electronic supplementary material.

Supplementary Material 1

Supplementary Material 2

Supplementary Material 3

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Xi, Y., Chong, H., Zhou, Y. et al. Convolutional neural network for brachial plexus segmentation at the interscalene level. BMC Anesthesiol 24, 17 (2024). https://doi.org/10.1186/s12871-024-02402-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12871-024-02402-2

Keywords