Direct
Deep learning framework for MRI reconstruction
Install / Use
/learn @NKI-AI/DirectREADME
.. raw:: html
<div align="center"> <img src="https://github.com/NKI-AI/direct/assets/71031687/14ce8234-7ef1-4e32-84c6-966dc393e7ca" width="400"/> <br> <figcaption margin-top:10px; font-size:24px !important; font-weight:bold !important;">DIRECT: Deep Image REConstruction Toolkit</figcaption> </div>.. raw:: html
<div align="center"> <br /> <a href="https://doi.org/10.21105/joss.04278"> <img src="https://joss.theoj.org/papers/10.21105/joss.04278/status.svg" alt="JOSS"></a> <a href="https://github.com/NKI-AI/direct/actions/workflows/tox.yml"> <img src="https://github.com/NKI-AI/direct/actions/workflows/tox.yml/badge.svg" alt="TOX"></a> <a href="https://github.com/NKI-AI/direct/actions/workflows/pylint.yml"> <img src="https://github.com/NKI-AI/direct/actions/workflows/pylint.yml/badge.svg" alt="Pylint"></a> <a href="https://github.com/NKI-AI/direct/actions/workflows/black.yml"> <img src="https://github.com/NKI-AI/direct/actions/workflows/black.yml/badge.svg" alt="Black"></a> <a href="https://app.codacy.com/gh/NKI-AI/direct?utm_source=github.com&utm_medium=referral&utm_content=NKI-AI/direct&utm_campaign=Badge_Grade_Settings"> <img src="https://api.codacy.com/project/badge/Grade/1c55d497dead4df69d6f256da51c98b7" alt="Codacy"></a> <a href="https://codecov.io/gh/NKI-AI/direct"> <img src="https://codecov.io/gh/NKI-AI/direct/branch/main/graph/badge.svg?token=STYAUFCKJY" alt="Codecov"></a> </div> <p align="center"> <a href="https://docs.aiforoncology.nl/direct/installation.html">Installation</a> • <a href="https://docs.aiforoncology.nl/direct/getting_started.html">Quick Start</a> • <a href="https://docs.aiforoncology.nl/direct/index.html">Documentation</a> • <a href="https://docs.aiforoncology.nl/direct/model_zoo.html">Model Zoo</a> <br> </p> <br />DIRECT is a Python, end-to-end pipeline for solving Inverse Problems emerging in Imaging Processing.
It is built with PyTorch and stores state-of-the-art Deep Learning imaging inverse problem solvers such as denoising, dealiasing and reconstruction.
By defining a base forward linear or non-linear operator, DIRECT can be used for training models for recovering images such as MRIs from partially observed or noisy input data.
DIRECT stores inverse problem solvers such as the vSHARP, Learned Primal Dual algorithm, Recurrent Inference Machine and Recurrent Variational Network, which were part of the winning solutions in Facebook & NYUs FastMRI challenge in 2019, the Calgary-Campinas MRI reconstruction challenge at MIDL 2020 and the CMRxRecon challenge 2023.
For a full list of the baselines currently implemented in DIRECT see here <#baselines-and-trained-models>_.
.. raw:: html
<div align="center"> <img src=".github/direct.png"/> <figcaption>Zero-filled reconstruction, Compressed-Sensing (CS) reconstruction using the BART toolbox, Reconstruction using a RIM model trained with DIRECT</figcaption> </div>Projects
In the projects <https://github.com/NKI-AI/direct/tree/main/projects>_ folder baseline model configurations are provided for each project.
Baselines and trained models
We provide a set of baseline results and trained models in the DIRECT Model Zoo <https://docs.aiforoncology.nl/direct/model_zoo.html>. Baselines and trained models include the vSHARP <https://arxiv.org/abs/2309.09954>, Recurrent Variational Network (RecurrentVarNet) <https://arxiv.org/abs/2111.09639>, the Recurrent Inference Machine (RIM) <https://www.sciencedirect.com/science/article/abs/pii/S1361841518306078>, the End-to-end Variational Network (VarNet) <https://arxiv.org/pdf/2004.06688.pdf>, the Learned Primal Dual Network (LDPNet) <https://arxiv.org/abs/1707.06474>, the X-Primal Dual Network (XPDNet) <https://arxiv.org/abs/2010.07290>, the KIKI-Net <https://pubmed.ncbi.nlm.nih.gov/29624729/>, the U-Net <https://arxiv.org/abs/1811.08839>, the Joint-ICNet <https://openaccess.thecvf.com/content/CVPR2021/papers/Jun_Joint_Deep_Model-Based_MR_Image_and_Coil_Sensitivity_Reconstruction_Network_CVPR_2021_paper.pdf>, and the AIRS Medical fastmri model (MultiDomainNet) <https://arxiv.org/pdf/2012.06318.pdf>_.
License and usage
DIRECT is not intended for clinical use. DIRECT is released under the Apache 2.0 License <LICENSE>_.
Citing DIRECT
If you use DIRECT in your own research, or want to refer to baseline results published in the DIRECT Model Zoo <model_zoo.rst>_\ , please use the following BiBTeX entry:
.. code-block:: text
@article{DIRECTTOOLKIT,
doi = {10.21105/joss.04278},
url = {https://doi.org/10.21105/joss.04278},
year = {2022},
publisher = {The Open Journal},
volume = {7},
number = {73},
pages = {4278},
author = {George Yiasemis and Nikita Moriakov and Dimitrios Karkalousos and Matthan Caan and Jonas Teuwen},
title = {DIRECT: Deep Image REConstruction Toolkit},
journal = {Journal of Open Source Software}
}
Related Skills
YC-Killer
2.7kA library of enterprise-grade AI agents designed to democratize artificial intelligence and provide free, open-source alternatives to overvalued Y Combinator startups. If you are excited about democratizing AI access & AI agents, please star ⭐️ this repository and use the link in the readme to join our open source AI research team.
openclaw-plugin-loom
Loom Learning Graph Skill This skill guides agents on how to use the Loom plugin to build and expand a learning graph over time. Purpose - Help users navigate learning paths (e.g., Nix, German)
best-practices-researcher
The most comprehensive Claude Code skills registry | Web Search: https://skills-registry-web.vercel.app
research_rules
Research & Verification Rules Quote Verification Protocol Primary Task "Make sure that the quote is relevant to the chapter and so you we want to make sure that we want to have it identifie
