3D Latent Diffusion Models (LDM) for generating large CT and MRI images with corresponding segmentation masks. Supports variable volume size and voxel spacing with precise control of organ/tumor size.
![]() |
![]() |
| Generated MR T2w prostate and T1w brain image | Generated CT image/mask pair |
NV-Generate-CTMR generates high-resolution synthetic 3D medical volumes using latent diffusion models built on the MAISI (Medical AI for Synthetic Imaging) framework. It produces CT images with paired segmentation masks and MRI volumes across multiple contrasts — enabling synthetic training data generation, data augmentation for rare pathologies, and privacy-preserving data sharing.
Key capabilities:
- CT generation with paired 132-class segmentation masks, supporting volumes up to 512x512x768 voxels with controllable organ and tumor size
- MRI generation across T1, T2, FLAIR, and additional contrasts for brain, abdomen, breast, and prostate anatomy
- Brain MRI synthesis with cross-sequence ControlNet for generating matched multi-contrast brain volumes (T1w, T2w, FLAIR, SWI)
- Variable resolution with configurable volume size and voxel spacing for each generation
Live Demo (no GPU required)
- 🎆 March 2026 🎇 — Released NV-Generate-MR-Brain models
rflow-mr-brainfor fast high-resolution 3D MR brain image generation, which covers both whole brain and skull-stripped brain generation for T1w, T2w, FLAIR, SWI images. - [October 2025] — Released rectified flow models
rflow-mrfor fast high-resolution 3D MR image generation. Upgraded previous MAISI repo to this NV-Generate-CTMR repo. - [March 2025] — Released rectified flow models
rflow-ctfor fast high-resolution 3D CT image generation and paired CT image/mask synthesis.rflow-ctis 33x faster thanddpm-ctand generates better quality images for the head region and small output volumes. - [August 2024] — Initial release
ddpm-ctsupporting 3D latent diffusion (DDPM) for CT image generation and paired CT image/mask synthesis.
- Overview
- News
- 1. Model Variants
- 2. Quick Start
- 3. Documentation: details of data preparation, training, and inference tutorials
- 4. Performance: accuracy, speed, and GPU memory usage
- 5. License
- 6. Citation
- 7. Resources
- 8. Acknowledgements
This repository provides four model variants for medical image generation: rflow-mr-brain, rflow-mr, rflow-ct, and ddpm-ct.
rflow-mr-brain |
rflow-mr |
rflow-ct |
ddpm-ct |
|
|---|---|---|---|---|
| Modality | MRI (Brain) | MRI | CT | CT |
| Model Weights | NV-Generate-MR-Brain | NV-Generate-MR | NV-Generate-CT | NV-Generate-CT |
| Architecture | MAISI-v2 (Rectified Flow) | MAISI-v2 (Rectified Flow) | MAISI-v2 (Rectified Flow) | MAISI-v1 (DDPM) |
| Paper | MAISI-v2 | MAISI-v2 | MAISI-v2 | MAISI-v1 |
| Inference Steps | 30 | 30 | 30 | 1000 |
| Max Volume | 512x512x256 | 512x512x128 | 512x512x768 | 512x512x768 |
| Use Case | MR Brain multi-contrast synthesis | MR image-only generation | CT image/mask pair generation | CT image/mask pair generation |
| License | NVIDIA Open Model | NVIDIA Non-Commercial | NVIDIA Open Model | NVIDIA Open Model |
Summary: Use rflow-ct for CT (whole-body inference). Use rflow-mr-brain for brain MRI (multi-contrast). Use rflow-mr for other MRI anatomies (fine-tune on your own data).
rflow-mr-brain |
rflow-mr |
rflow-ct |
ddpm-ct |
|
|---|---|---|---|---|
| Modality | MRI (brain) | MRI | CT | CT |
| Release Date | March 2026 | October 2025 | March 2025 | August 2024 |
| Model Weights | NV-Generate-MR-Brain | NV-Generate-MR | NV-Generate-CT | NV-Generate-CT |
| Quick Start | 2.2 MR Brain Image Generation | 2.5 MR Image Generation | 2.3 CT Paired Image/Mask, 2.4 CT Image | 2.3 CT Paired Image/Mask |
| Architecture | MAISI-v2 (Rectified Flow) | MAISI-v2 (Rectified Flow) | MAISI-v2 (Rectified Flow) | MAISI-v1 (DDPM) |
| Paper | MAISI-v2 | MAISI-v2 | MAISI-v2 | MAISI-v1 |
| Network Detail | config_network_rflow.json | config_network_rflow.json | config_network_rflow.json | config_network_ddpm.json |
| Inference Steps | 30 | 30 | 30 (33× faster than ddpm-ct) |
1000 |
| Max Volume | 512×512×256 | 512×512×128 | 512×512×768 | 512×512×768 |
| Use Case | MR image-only generation for brain (T1w, T2w, FLAIR, SWI; whole brain and skull-stripped) | MR image-only generation with user specified contrast | CT image-only generation; CT image/mask pair generation | CT image-only generation; CT image/mask pair generation |
| Model: Foundation VAE | same VAE with ddpm-ct |
trained on CT and MR (with additional abdomen MRI) | same VAE with ddpm-ct |
trained on CT and MR |
| Model: Foundation Diffusion Model | does not take body region as input, takes modality as input (brain-focused) | does not take body region as input, takes modality as input. We recommend finetuning with users' own MRI data. | does not take body region as input, has API for modality input (always set as 'ct' but expandable) | takes body region as input, no API for modality input |
| Model: ControlNet | Coming soon | N/A | generate image/mask pairs, with contrastive loss | generate image/mask pairs, no contrastive loss |
pip install -r requirements.txtPlease refer to inference_diff_unet_tutorial.ipynb for the inference tutorial that generates CT or MR image without mask.
You can also run it in command line to generate MR image without mask. Please change "modality" in configs/config_maisi_diff_model_rflow-mr-brain.json according to configs/modality_mapping.json to control the output MR contrast. Currently we support both whole brain and skull-stripped brain generation for T1w, T2w, FLAIR, SWI images.
"mri":8, # MRI without specifying contrast or skull condition, can be any of them
"mri_t1":9, # T1w whole-brain MRI
"mri_t2":10, # T2w whole-brain MRI
"mri_flair":11, # FLAIR whole-brain MRI
"mri_swi":20, # SWI whole-brain MRI
"mri_t1_skull_stripped":29, # T1w skull-stripped brain MRI
"mri_t2_skull_stripped":30, # T2w skull-stripped brain MRI
"mri_flair_skull_stripped":31, # FLAIR skull-stripped brain MRI
"mri_swi_skull_stripped":32, # SWI skull-stripped brain MRInetwork="rflow"
generate_version="rflow-mr-brain"
python -m scripts.download_model_data --version ${generate_version} --root_dir "./" --model_only
python -m scripts.diff_model_infer -t ./configs/config_network_${network}.json -e ./configs/environment_maisi_diff_model_${generate_version}.json -c ./configs/config_maisi_diff_model_${generate_version}.jsonexport MONAI_DATA_DIRECTORY="./temp_work_dir"
network="rflow"
generate_version="rflow-ct" # can change to "ddpm-ct"
python -m scripts.inference -t ./configs/config_network_${network}.json -i ./configs/config_infer.json -e ./configs/environment_${generate_version}.json --random-seed 0 --version ${generate_version}See also: inference_tutorial.ipynb
network="rflow"
generate_version="rflow-ct" # can change to "ddpm-ct"
python -m scripts.download_model_data --version ${generate_version} --root_dir "./" --model_only
python -m scripts.diff_model_infer -t ./configs/config_network_${network}.json -e ./configs/environment_maisi_diff_model_${generate_version}.json -c ./configs/config_maisi_diff_model_${generate_version}.jsonChange "modality" in configs/config_maisi_diff_model_rflow-mr.json according to configs/modality_mapping.json to control the output MR contrast. Supported contrasts: T1/T2 brain, FLAIR skull-stripped brain, T2 prostate, T1 breast, T1/T2 abdomen. But if you are going to synthesize brain images, we recommend using rflow-mr-brain model instead. Please see 2.2 MR Brain Image Generation.
network="rflow"
generate_version="rflow-mr"
python -m scripts.download_model_data --version ${generate_version} --root_dir "./" --model_only
python -m scripts.diff_model_infer -t ./configs/config_network_${network}.json -e ./configs/environment_maisi_diff_model_${generate_version}.json -c ./configs/config_maisi_diff_model_${generate_version}.jsonA reference implementation for MR-to-CT synthesis based on NV-Generate-CTMR (rflow-ct) is available here: https://github.com/brudfors/maisi-mr-to-ct.
If you've adapted NV-Generate-CTMR for other imaging tasks or applications and would like to share your work, please feel free to open an issue or contact the maintainers — we'd love to link to your repo.
| Guide | Description |
|---|---|
| Setup | Full installation guide, dependencies, model weight download |
| Inference | Detailed inference parameters, spacing tables |
| Training | VAE, Diffusion Model, and ControlNet training guides |
| Data Preparation | Dataset formats and preparation steps |
| Evaluation | FID evaluation tool and benchmark results |
| Troubleshooting | Common issues and solutions |
| Applications | Community adaptations (MR-to-CT synthesis) |
| Inference Tutorial | Quick start CT paired generation (notebook) |
| Diffusion Inference | CT/MR image-only generation (notebook) |
| Training Tutorials | VAE, diffusion, and ControlNet training |
Training, inference, data preparation, and evaluation details are covered in the guides linked above.
On the unseen autoPET 2023 benchmark:
| Model | FID Score | Inference Steps | Speed vs ddpm-ct |
|---|---|---|---|
rflow-ct |
5.124 | 30 | 33x faster |
ddpm-ct |
6.083 | 1000 | baseline |
For inference parameters, see Documentation. For GPU memory and timing, see Performance.
| Component | License |
|---|---|
| Source code | Apache 2.0 |
| NV-Generate-CT weights | NVIDIA Open Model |
| NV-Generate-MR weights | NVIDIA Non-Commercial |
| NV-Generate-MR-Brain weights | NVIDIA Open Model |
This project will download and install additional third-party open source software projects. Review the license terms of these open source projects before use.
@article{zhao2026maisi,
title={MAISI-v2: Accelerated 3D high-resolution medical image synthesis with rectified flow and region-specific contrastive loss},
author={Zhao, Can and Guo, Pengfei and Yang, Dong and Tang, Yucheng and He, Yufan and Simon, Benjamin and Belue, Mason and Harmon, Stephanie and Turkbey, Baris and Xu, Daguang},
journal={Proceedings of the 40th AAAI Conference on Artificial Intelligence (AAAI 2026)},
year={2026}
}@inproceedings{guo2025maisi,
title={MAISI: Medical AI for synthetic imaging},
author={Guo, Pengfei and Zhao, Can and Yang, Dong and Xu, Ziyue and Nath, Vishwesh and Tang, Yucheng and Simon, Benjamin and Belue, Mason and Harmon, Stephanie and Turkbey, Baris and others},
booktitle={2025 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)},
pages={4430--4441},
year={2025},
organization={IEEE}
}- NV-Generate-CT on HuggingFace -- CT model weights and model card
- NV-Generate-MR on HuggingFace -- MR model weights and model card
- NV-Generate-MR-Brain on HuggingFace -- Brain MRI model weights and model card
- MAISI Live Demo -- Try online without GPU
- MAISI-v1 Paper (WACV 2025)
- MAISI-v2 Paper (AAAI 2026)
- Built with MONAI -- Medical Open Network for AI
This project was conducted by NVIDIA in collaboration with the University of Zurich, Istanbul Medipol University, and Forithmus.
We would like to thank the following people for their contributions to the development of the NV-Generate-MR-Brain models: Bjoern Menze, Ibrahim Ethem Hamamci, Sezgin Er, Suprosanna Shit, Utku Türkbey, etc.


