Prior image-constrained reconstruction using style-based generative models - Tensorflow implementation
Varun A. Kelkar, Mark A. Anastasio
University of Illinois at Urbana-Champaign, Urbana, IL - 61801, USA
Contact: [email protected], [email protected]
Abstract: Obtaining a useful estimate of an object from highly incomplete imaging measurements remains a holy grail of imaging science. Deep learning methods have shown promise in learning object priors or constraints to improve the conditioning of an ill-posed imaging inverse problem. In this study, a framework for estimating an object of interest that is semantically related to a known prior image, is proposed. An optimization problem is formulated in the disentangled latent space of a style-based generative model, and semantically meaningful constraints are imposed using the disentangled latent representation of the prior image. Stable recovery from incomplete measurements with the help of a prior image is theoretically analyzed. Numerical experiments demonstrating the superior performance of our approach as compared to related methods are presented
- Linux/Unix-based systems recommended. The code hasn't been tested on Windows.
- 64 bit Python 3.6+. The code has been tested with Python 3.7.4 installed via Anaconda
- Tensorflow 1.14/1.15. The code has been tested with Tensorflow 1.14. Tensorflow 2+ is not supported.
- imageio - Install via
pip
orconda
. - scikit-image
Additional dependencies that are required for the various reconstruction methods are as follows:
- Prox-TV
Can be installed viapip install prox_tv
. In our experience, this works on Linux and Mac, but not on Windows.
- Cuda toolkit 10.0 (higher versions may work but haven't been tested)
- GCC 7.2+. The code has been tested with GCC 7.2.0
- TF-Wavelets: Relevant portions included with this code.
The directory scripts
contains the top level scripts:
get_network_weights.sh
: Script for downloading our pretrained StyleGAN2 network weights and storing them instylegan2/nets/
.run_projector.sh
: Script for projecting an image onto the range of a given StyleGAN2.recon_*.sh
: Scripts for running the various reconstruction algorithms.
The directory stylegan2
contains the original StyleGAN2 code, with the following modifications:
- Addition of regularization based on Gaussianized disentangled latent-space for projecting images onto the range of the StyleGAN2 (contained in
stylegan2/projector.py
with regularization and optimization parameters controlled instylegan2/run_projector.py
). stylegan2/model_oop.py
- A small object-oriented wrapper around StyleGAN2.- Various scripts for related to the theoretical results presented in our paper.
stylegan2/nets/
is the location for saving the trained network .pkl files.
The directory pic_recon
contains the following sub-directories:
src
Stores all python codes for image reconstruction.masks_mri
stores the MRI undersampling masksground_truths
- images of ground truthsprior_images
- Images and latent-projections of prior images.results
- stores the results of the reconstruction.
StyleGAN2 network weights, trained on brain MR images and FFHQ dataset images can be found here : https://databank.illinois.edu/datasets/IDB-4499850. They can be downloaded using scripts/get_network_weights.sh
. More information can be found in TRAINED_WEIGHTS_INFO.md
.
- Make sure
cuda-toolkit/10
andgcc/7.2+
are loaded. - Download StyleGAN2 weights, for example
bash scripts/get_network_weights.sh FastMRIT1T2
. - Run
bash scripts/run_projector.sh
. If you wish, you may setnetwork
insiderun_projector.sh
to the path to the StyleGAN2 network.pkl
you want. The image size is 256x256x1 for brain images, and 128x128x3 for face images. The projected images along with their latent representations will be stored instylegan2/projected_images/
The simplest way to get a recon algorithm alg
is as follows:
- Download the appropriate StyleGAN2 weights. For brain images, use
bash scripts/get_network_weights.sh CompMRIT1T2
.
For face images, use
bash scripts/get_network_weights.sh FFHQ
.
The weights are stored instylegan2/nets/
- Specify the correct network pkl path as
network_path
inscripts/recon_${alg}.sh
. The correct one for brain images is already specified. (applies only to CSGM and PICGM). - Run
bash scripts/recon_${alg}.sh
for an algorithmalg
, wherealg
can beplstv
,csgm
,piccs
orpicgm
.
The ground truth (GT) images can be found in pic_recon/ground_truths/
. The prior images (PIs) can be found in pic_recon/prior_images/
. The GTs and the PIs are organized according to data_type
, which can be either faces
or brain
.
For the brain images, we provide two example GT-PI pairs. Additional GT-PI pairs can be downloaded from The Cancer Imaging Archive (TCIA) Brain-Tumor-Progression dataset. Links to the data use policy can be found in pic_recon/ground_truths/brain/README.md
.
The Shutterstock Collection containing the GT-PI pairs for the face images can be found here: https://www.shutterstock.com/collections/298591136-e81133c6. A Shutterstock license is needed to use these images. The preprocessing steps used are described in our paper, and include manual cropping and resizing to 128x128x3. You may also use the automatic face alignment procedure developed by V. Kazemi and J. Sullivan for preprocessing, similar to the original FFHQ dataset. This is likely to improve the reconstruction performance of CSGM and PICGM further.
All the recon scripts contain the following common arguments:
process
: This is the index of the image to be reconstructed.data_type
: Can be eitherfaces
orbrain
.mask_type
: More generally refers to the forward model. For random Gaussian sensing, this argument takes the formgaussian_${m_by_n}
. For simulated MRI with random cartesian undersampling, this argument takes the formmask_rand_${n_by_m}x
. The Gaussian sensing matrices are generated in the code using a fixed seed. The undersampling masks for MRI are located insidepic_recon/masks_mri/
.snr
: Measurement SNR.gt_filename
: Path to the ground truth.pi_filename
: (only for PICCS and PICGM) Path to the prior image.- Various regularization parameters.
The results will be stored in pic_recon/results/${data_type}/${process}/${algorithm}_${data_type}_${mask_type}_SNR${snr}/
.
@article{kelkar2021prior,
title={Prior Image-Constrained Reconstruction using Style-Based Generative Models},
author={Kelkar, Varun A and Anastasio, Mark A},
journal={arXiv preprint arXiv:2102.12525},
year={2021}
}
- FastMRI initiative database: Zbontar, Jure, et al. "fastMRI: An open dataset and benchmarks for accelerated MRI." arXiv preprint arXiv:1811.08839 (2018).
- The Cancer Imaging Archive Clark, K., Vendt, B., Smith, K., Freymann, J., Kirby, J., Koppel, P., Moore, S., Phillips, S., Maffitt, D., Pringle, M., Tarbox, L., & Prior, F. (2013). The Cancer Imaging Archive (TCIA): Maintaining and Operating a Public Information Repository. Journal of Digital Imaging, 26(6), 1045–1057. https://doi.org/10.1007/s10278-013-9622-7
- TCIA Brain-Tumor-Progression Schmainda KM, Prah M (2018). Data from Brain-Tumor-Progression. The Cancer Imaging Archive. https://doi.org/10.7937/K9/TCIA.2018.15quzvnb
- FFHQ dataset A Style-Based Generator Architecture for Generative Adversarial Networks Tero Karras (NVIDIA), Samuli Laine (NVIDIA), Timo Aila (NVIDIA) https://arxiv.org/abs/1812.04948, https://github.com/NVlabs/ffhq-dataset