Skip to content

cardiAc ultrasound Segmentation & Color-dopplEr dealiasiNg Toolbox (ASCENT)

License

Notifications You must be signed in to change notification settings

creatis-myriad/ASCENT

Repository files navigation

ASCENT

Welcome to the code repository for cardiAc ultrasound Segmentation & Color-dopplEr dealiasiNg Toolbox (ASCENT).

python PyTorch Lightning Config: Hydra Template

Imports: isort Code style: black pre-commit code-quality

license

Publications

Journal Paper

Conference Paper

Description

ASCENT is a toolbox to segment cardiac structures (left ventricle, right ventricle, etc.) on ultrasound images and perform dealiasing on color Doppler echocardiography. It combines one of the best segmentation framework, nnUNet with Lightning, Hydra, and monai. The main reasons of doing so are to take advantage of each library:

  • nnUNet's heuristic rules for hyperparameters determination and training scheme give excellent segmentation results.
  • Lightning reduces boilerplate and provides better PyTorch code organization.
  • Hydra offers pluggable architectures, dynamic configurations, and easy configuration overriding through command lines.
  • Monai simplifies the data loading and pre-processing.

For now, ASCENT provides only nnUNet 2D and 3D_fullres architectures (similar to monai's DynUNet). You can easily plug your own models in ASCENT pipeline.

nnUNet implemented in ASCENT is the V1.

Table of Contents

How to run

ASCENT has been tested on Linux (Ubuntu 20, Red Hat 7.6), macOS and Windows 10/11.

Automatic Mixed Precision (AMP) is buggy on Windows devices, e.g. Nan in loss computation. For Windows users, it is recommended to disable it during the run by adding trainer.precision=32 to the train/evaluate/predict command to avoid errors.

Install

  1. Download the repository:
    # clone project
    git clone https://github.com/creatis-myriad/ASCENT
    cd ASCENT
  2. Create a virtual environment (Conda is strongly recommended):
    # create conda environment
    conda create -n ascent python=3.10
    conda activate ascent
  3. Install PyTorch according to instructions. Grab the one with GPU for faster training:
    # example for linux or Windows
    conda install pytorch torchvision torchaudio pytorch-cuda=11.8 -c pytorch -c nvidia
  4. Install the project in editable mode and its dependencies:
    pip install -e .

Several new commands will be added to the virtual environment once the installation is completed. These commands all start with ascent_.

Data

Before doing any preprocessing and training, you must first reformat the dataset to the appropriate format and place the converted dataset in the data/ folder. Here is an example of the converted CAMUS dataset using this conversion script.

Refer to here if you want to have a different data/ folder location.

The reformatted dataset should look like this:

data/
├── CAMUS_challenge
│   ├──raw/
│   │  ├──imagesTr/
│   │  │  ├──patient0001_2CH_ED_0000.nii.gz
│   │  │  ├──patient0001_2CH_ES_0000.nii.gz
│   │  │  ├──patient0001_4CH_ED_0000.nii.gz
│   │  │  └── ...
│   │  │
│   │  ├──labelsTr/
│   │  │  ├──patient0001_2CH_ED.nii.gz
│   │  │  ├──patient0001_2CH_ES.nii.gz
│   │  │  ├──patient0001_4CH_ED.nii.gz
│   │  │  └── ...
│   │  │
│   │  ├──imagesTs/
│   │  ├──labelsTs/
│   │  ├──dataset.json

More details can be found in nnUNet's dataset conversion instructions.

Important note

ASCENT uses Hydra to handle the configurations and runs. To know more about Hydra's CLI, refer to its documentation.

Preprocess

ASCENT preprocesses and determines the optimized hyperparameters according to nnUNet's heuristic rules. After placing the dataset in the correct data/ folder, you can run the preprocessing and planning with the following command:

ascent_preprocess_and_plan dataset=XXX

XXX refers to the dataset name, e.g. CAMUS_challenge.

It is possible to preprocess multiple dataset at once using the --multirun flag of Hydra:

ascent_preprocess_and_plan --multirun dataset=XXX,YYY,ZZZ

By default, ASCENT creates a 2D and a 3D experiment configuration files that includes the batch size, U-Net architecture, patch size, etc. You can choose to plan only 2D or 3D experiment by overriding pl2d or pl3d like follows:

ascent_preprocess_and_plan dataset=XXX pl2d=False

Once ascent_preprocess_and_plan is completed, the cropped and preprocessed data will be located respectively at data/XXX/cropped and data/XXX/preprocessed. New config files are also generated in ascent/configs/experiment/, ascent/configs/datamodule/, and ascent/configs/model/. These configs files are named as XXX_2d.yaml or XXX_3d.yaml, depending on the requested planner(s).

You may override other configurations as long as they are listed in ascent/configs/preprocess_and_plan.yaml. You can also run the following command to display all available configurations:

ascent_preprocess_and_plan -h

Model training

With the preprocessing being done, you can now train the model. For all experiments, ASCENT automatically detects the presence of GPU and utilize the GPU if it is available. ASCENTS creates 10-Fold cross validations with train/validation/test splits with 0.8/0.1/0.1 ratio. You can disable the test splits by overriding datamodule.test_splits=False to create 5-Fold train/validation splits with 0.8/0.2 ratio.

Below is an example to train a 2D model on CAMUS dataset with the pre-determined hyperparameters:

ascent_train experiment=camus_challenge_2d logger=tensorboard

# train on cpu
ascent_train experiment=camus_challenge_2d trainer.accelerator=cpu logger=tensorboard

You can override any parameter from command line like this:

ascent_train experiment=camus_challenge_2d trainer.max_epochs=20 datamodule.batch_size=8 logger=tensorboard

More advanced Hydra overriding is explained here.

If you want to check if all the configurations are correct without running the experiment, simply run:

ascent_train experiment=camus_challenge_2d --cfg job --resolve

Hydra creates new output directory for every executed run. Default ASCENT's nnUNet logging structure is as follows:

├── logs
│   ├── dataset_name
│   │   ├── nnUNet
│   │   │   ├── 2D                          # nnUNet variant (2D or 3D)
│   │   │   │   ├── Fold_X                      # Fold to train on
│   │   │   │   │   ├── runs                        # Logs generated by single runs
│   │   │   │   │   │   ├── YYYY-MM-DD_HH-MM-SS       # Datetime of the run
│   │   │   │   │   │   │   ├── .hydra                  # Hydra logs
│   │   │   │   │   │   │   ├── wandb                   # Weights&Biases logs
│   │   │   │   │   │   │   ├── checkpoints             # Training checkpoints
│   │   │   │   │   │   │   └── ...                     # Any other thing saved during training
│   │   │   │   │   │   └── ...
│   │   │   │   │   │
│   │   │   │   │   └── multiruns                   # Logs generated by multiruns
│   │   │   │   │       ├── YYYY-MM-DD_HH-MM-SS       # Datetime of the multirun
│   │   │   │   │       │   ├──1                        # Multirun job number
│   │   │   │   │       │   ├──2
│   │   │   │   │       │   └── ...
│   │   │   │   │       └── ...

At the end of the training, the prediction on validation or test data will be executed and saved in the output folder, named as validation_raw or testing_raw. To disable this, override test=False:

ascent_train experiment=camus_challenge_2d test=False logger=tensorboard

Refer to here if you want to have a different logs/ folder location.

Model evaluation

If you skipped the evaluation on validation or test data during, you may evaluate your model afterward by specifying the fold and the ckpt_path using ascent_evaluate:

ascent_evaluate experiment=camus_challenge_2d  fold=0 ckpt_path="/path/to/ckpt" logger=tensorboard

This will create a new output directory containing the prediction folder.

Run inference

To run inference on unseen data, you may use the ascent_predict:

ascent_predict dataset=CAMUS_challenge model=camus_challenge_2d ckpt_path=/path/to/checkpoint input_folder=/path/to/input/folder/ output_folder=/path/to/output/folder

By default, ASCENT applies test time augmentation during inference. To disable this, override tta=False. If you wish to save the predicted softmax probabilities as well, activate the save_npz=True flag.

Experiment tracking

ASCENTS supports all the logging frameworks proposed by PyTorch Lightning: Weights&Biases, Neptune, Comet, MLFlow, Tensorboard.

For nnUNet experiments, Weights&Biases logger is used by default. This requires you to create an account. After signing up, rename the .env.example file to .env and specify your WANDB API Key as follows:

### API keys ###
WANDB_API_KEY=<your-wandb-api-key>

The environment variables in the .env file is automatically loaded by pyrootutils before each run.

You can simply override logger to use your logger of preference:

# to use the default tensorboard logger of PyTorch Lightning
ascent_train experiment=camus_challenge_2d logger=tensorboard

Define custom data and logs path

In some cases, you may want to specify your own data and logs paths instead of using the default data/ and logs/. You can do this by setting them in environments variables after renaming the .env.example file to .env. In the .env, simply override:

# custom data path
ASCENT_DATA_PATH="path/to/data"

# custom logs path
ASCENT_LOGS_PATH="paths/to/logs"

After that, you must override paths=custom in all your commands, e.g.:

# to use custom data and logs paths
ascent_train experiment=camus_challenge_2d paths=custom

Advanced Hydra overriding

If you wish to perform some advanced Hydra overriding, kindly refer to the these documentations.

Example in Jupyter Notebook

An example of how to use ASCENT on ACDC dataset in a Jupyter Notebook can be found here.

Resources

This project was inspired by:

References

If you find this repository useful, please consider citing the paper implemented in this repository relevant to you from the list below:

@article{ling_dealiasing_2023,
   title = {Phase {Unwrapping} of {Color} {Doppler} {Echocardiography} using {Deep} {Learning}},
   doi = {10.1109/TUFFC.2023.3289621},
   journal = {IEEE Transactions on Ultrasonics, Ferroelectrics, and Frequency Control},
   author = {Ling, Hang Jung and Bernard, Olivier and Ducros, Nicolas and Garcia, Damien},
   month = aug,
   year = {2023},
   pages = {810--820},
}

@inproceedings{ling_temporal_2023,
   title = {Extraction of {Volumetric} {Indices} from {Echocardiography}: {Which} {Deep} {Learning} {Solution} for {Clinical} {Use}?},
   doi = {10.1007/978-3-031-35302-4_25},
   series = {Lecture {Notes} in {Computer} {Science}},
   booktitle = {Functional {Imaging} and {Modeling} of the {Heart}},
   publisher = {Springer Nature Switzerland},
   author = {Ling, Hang Jung and Painchaud, Nathan and Courand, Pierre-Yves and Jodoin, Pierre-Marc and Garcia, Damien and Bernard, Olivier},
   month = june,
   year = {2023},
   pages = {245--254},
}

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •