Rheumatic Heart Disease (RHD) remains a leading cause of cardiovascular morbidity in low-resource settings. Our work at DSAIL focuses on building deployable AI systems that support early diagnosis of RHD using echocardiographic data—bridging clinical needs with scalable machine learning.
Our initial work tackled the challenge of identifying the parasternal long axis (PLAX) view, which is critical for RHD screening. We developed a binary classifier using logistic regression and later CNNs to distinguish PLAX from non-PLAX frames. To support scalable annotation, we built a web-based tool (Echo Label) deployed on Google Cloud Platform, enabling cardiologists to tag echo frames with WHF-guided metadata.
Highlights
To reduce expert annotation burden, we explored unsupervised clustering of echo videos using PCA and agglomerative hierarchical clustering. These methods grouped videos by view (PLAX, PSAX, A4C) without labels, revealing strong potential for semi-automated annotation workflows.
Technical Methods:
We extended our pipeline to multi-task classification of echocardiographic views, RHD conditions, and severity using self-supervised learning. Two models were compared:
Key Components:
Both were pre-trained on 38,000+ unlabeled frames and fine-tuned on a curated set of 2,655 labeled images. DINOv2 achieved up to 99% accuracy on severity classification, outperforming SimCLR on most tasks.
Tasks:
We are now fine-tuning MedGemma, a multimodal vision-language model, for RHD detection in echocardiographic images. This involves adapting MedGemma to:
Technical Direction:
Our long-term goal is to deliver interpretable, deployable AI tools that support frontline clinicians in diagnosing RHD early and accurately—especially in underserved regions. This project exemplifies how multimodal learning, clinical collaboration, and thoughtful deployment can be leveraged to address global health challenges.