Home

Awesome

EchoCLIP: A Multimodal Foundation Model For Echocardiography

EchoCLIP is a multimodal foundation model for echocardiography. It is finetuned from CLIP weights on a dataset of >1M pairs of echocardiogram images and their associated expert interpretation text. It can be used for semantic search amongst echo videos as well as zero-shot prediction on a wide range of clinically relevant tasks. For more details, see our paper:

(link will be added once preprint is released)

<!-- [Multimodal Foundation Models For Echocardiogram Interpretation](https://arxiv.org/abs/) -->

Quickstart

This repo contains example code for loading and using EchoCLIP and its long-context variant, EchoCLIP-R. To get started, clone this repo and navigate into it. Then, create a new conda environment and install the required packages:

git clone https://github.com/echonet/echo_CLIP
cd echo_CLIP
conda env create -n echo-clip
conda activate echo-clip
python -m pip install -r requirements.txt

You should now be able to run embedding_example.py and zero_shot_example.py.

Repo contents