A Unified Model for Compressed Sensing MRI Across Undersampling Patterns

Caltech
In submission

*Indicates Equal Advising

Abstract

Compressed Sensing MRI reconstructs images of the body’s internal anatomy from undersampled measurements, thereby reducing the scan time—the time subjects need to remain still. Recently, deep neural networks have shown great potential for reconstructing high-fidelity images from highly undersampled measurements in the frequency space. However, one needs to train multiple models for different un- dersampling patterns and desired output image resolutions, since most networks operate on a fixed discretization. Such approaches are highly impractical in clinical settings, where undersampling patterns and image resolutions are frequently changed to accommodate different real-time imaging and diagnostic requirements.

We propose a unified model robust to different measure- ment undersampling patterns and image resolutions in com- pressed sensing MRI. Our model is based on neural opera- tors, a discretization-agnostic architecture. Neural operators are employed in both image and measurement space, which capture local and global image features for MRI reconstruc- tion. Empirically, we achieve consistent performance across different undersampling rates and patterns, with an average 11% SSIM and 4 dB PSNR improvement over a state-of-the- art CNN, End-to-End VarNet. For efficiency, our inference speed is also 1,400x faster than diffusion methods. The resolution-agnostic design also enhances zero-shot super- resolution and extended field of view in reconstructed images. Our unified model offers a versatile solution for MRI, adapt- ing seamlessly to various measurement undersampling and imaging resolutions, making it highly effective for flexible and reliable clinical imaging.

BibTeX

@article{wang2024ultrasound,
  title={Ultrasound Lung Aeration Map via Physics-Aware Neural Operators},
  author={Wang, Jiayun and Ostras, Oleksii and Sode, Masashi and Tolooshams, Bahareh and Li, Zongyi and Azizzadenesheli, Kamyar and Pinton, Giammarco and Anandkumar, Anima},
  year={2024}
}