This abstract is published in the Journal of the American Society of Echocardiography (volume 33, number 6, year 2020), as part of the proceedings of the American Society Echocardiography Conference 2020.

Title: Automated View Detection of Contrast and Non-Contrast Cardiac Views in Echocardiography; A Multi-Centre, Multi-Vendor Study

Authors: Gao, Shan and Stojanovski, David and Parker, Andrew and Marques, Patricia and Heitner, Stephen and Yadava, Mrinal and Upton, Ross and Woodward, Gary and Lamata, Pablo and Beqiri, Arian and others

BACKGROUND: Correct identification of views acquired in a 2D echocardiographic examination is paramount to post-processing and quantification steps performed as part of most clinical workflows. In many exams, microbubble contrast is used which greatly affects the appearance of the cardiac views. Here we present a fully automated convolutional neural network (CNN) which identifies apical 2, 3, and 4 chamber, and short axis (SAX) views with and without contrast in a completely independent, external dataset.

METHODS: Data was taken from 1014 subjects from a prospective multisite, multi-vendor, UK trial with the number of frames in each view greater than 17,500. Prior to view classification model training, images were processed using standard techniques to ensure homogenous and normalised image inputs. A CNN was built using the minimum number of convolutional layers required with batch normalisation, and dropout for reducing overfitting. Before processing, the data was split into 90% for model training (211,958 frames), and 10% used as a validation dataset (23,946 frames). Image frames from different subjects were separated out entirely amongst the different datasets. Further, a separate trial dataset of 240 studies acquired in the USA was used as an independent test dataset (39,401 frames).

Figure 1: Confusion matrices for the validation dataset (left) and the independent test dataset (right)

RESULTS: Figure 1 shows the confusion matrix for both validation data (left) and independent test data (right), with an overall accuracy of 96% and 95% for the validation and test
datasets respectively. The accuracy for the non-contrast data of >99% exceeds that seen in other works such as Østvik et al. US Med. Biol. 45, 374–384. The combined datasets included images acquired across ultrasound manufacturers and models from 12 clinical sites.

CONCLUSION: We have developed a CNN capable of accurately identifying all relevant cardiac views used in “real world” echo exams, including views acquired with contrast, that could improve efficiency of quantification steps performed after image acquisition in routine clinical practice. This was tested on an independent dataset acquired in a different country to that used to train the model and was found to perform similarly.

Share on

Share on facebook
Share on twitter
Share on linkedin

Further reading

$10m funding accelerates innovation and growth for EchoGo

Artificial intelligence (AI) health-tech company Ultromics announced today it has  closed a $10 million financing round to accelerate its next stage of business growth. The capital has been raised …

Off-cart analysis and diagnosis can enhance workflow efficiencies and reduce exposure to COVID

Healthcare systems are accustomed to adapting to change, but the current COVID-19 crisis is unique. Whilst many past changes to healthcare delivery have allowed for technology adoption over time …

Ultromics collaborates with Mayo Clinic to optimize the impact of AI tools on care pathways

Ultromics, a health-tech company which applies AI to echocardiography, to earlier detect and better diagnose heart disease, has entered a know-how and clinical research agreement with Mayo Clinic. Underpinned …