We develop Open Foundation Models from numerous public datasets using their heterogeneous expert annotations.
A Fully Open AI Foundation Model Applied to Chest Radiography
DongAo Ma1, Jiaxuan Pang1, Michael B. Gotway2, Jianming Liang1
1 Arizona State University, 2 Mayo Clinic
Nature (2025)
Paper (PDF, Supplementary, Peer Review) | Poster | Github | CodeOcean for Reprodicible Run
Ark+: Supervised training a single high-performance AI foundation model from many differently labeled datasets—no label consolidation required
DongAo Ma1, Jiaxuan Pang1, Shivasakthi Senthil Velan1, Michael B. Gotway2, Jianming Liang1
1 Arizona State University, 2 Mayo Clinic
Medical Image Analysis (2025)
This article is the culmination of a technological investigation into
methodology for fully supervised learning from heterogeneous labels
associated with numerous (big or small and public or private) datasets.
Paper (Free access before 1/18/2026)
Foundation Ark: Accruing and Reusing Knowledge for Superior and Robust Performance
DongAo Ma1, Jiaxuan Pang1, Michael B. Gotway2, Jianming Liang1
1 Arizona State University, 2 Mayo Clinic
International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI 2023) (Oral + Poster)
🏆 Best Paper Award Runner-up
★ MICCAI 2023 STAR Awards
Paper (PDF, Arxiv) | Code | Poster | Oral Presentation (YouTube, BiliBili)
You can request the pretrained Ark+ models in our paper throught this Google Form or wjx.cn.
An example of initializing the model and loading the pretrained weights can be found at: Zeroshot Transfer
Create Swin Transformer Base/Large model from the official model or from timm (v0.5.4).
Below is an example of how to load pre-trained weights into the Swin Transformer model:
import torch
from timm.models.swin_transformer import SwinTransformer
# Initialize the model
model = SwinTransformer(
num_classes=args.num_class,
img_size=768,
patch_size=4,
window_size=12,
embed_dim=192,
depths=(2, 2, 18, 2),
num_heads=(6, 12, 24, 48)
)
# Load the checkpoint
checkpoint = torch.load('<PATH_TO_MODEL>/Ark6_swinLarge768_ep50.pth.tar', map_location="cpu")
state_dict = checkpoint['teacher']
# Remove "module." prefix if present
state_dict = {k.replace("module.", ""): v for k, v in state_dict.items()}
# Identify and delete unnecessary keys
k_del = [k for k in state_dict.keys() if "attn_mask" in k] + ['head.weight', 'head.bias']
print(f"Removing key(s) {k_del} from pretrained checkpoint for scaled input size")
# Delete identified keys
for k in k_del:
if k in state_dict: # Ensure the key exists
del state_dict[k]
# Load the model weights
msg = model.load_state_dict(state_dict, strict=False)
print('Loaded with msg:', msg)If you encounter a size mismatch error when loading a pretrained model, please verify the version of the timm package, as later versions have updated the Swin Transformer architectures.
If you use this code or use our pre-trained weights for your research, please cite our paper:
@article{ma2025fully,
title={A fully open {AI} foundation model applied to chest radiography},
author={Ma, DongAo and Pang, Jiaxuan and Gotway, Michael B and Liang, Jianming},
journal={Nature},
pages={1--11},
year={2025},
publisher={Nature Publishing Group}
}
@article{ma2026ark+,
author = {DongAo Ma and Jiaxuan Pang and Shivasakthi {Senthil Velan} and Michael B. Gotway and Jianming Liang},
title = {Ark+: Supervised training a single high-performance AI foundation model from many differently labeled datasets—no label consolidation required},
journal = {Medical Image Analysis},
volume = {108},
pages = {103828},
year = {2026},
issn = {1361-8415},
doi = {https://doi.org/10.1016/j.media.2025.103828},
url = {https://www.sciencedirect.com/science/article/pii/S1361841525003743}
}
@InProceedings{ma2023foundation,
author="Ma, DongAo and Pang, Jiaxuan and Gotway, Michael B. and Liang, Jianming",
title="Foundation Ark: Accruing and Reusing Knowledge for Superior and Robust Performance",
booktitle="Medical Image Computing and Computer Assisted Intervention -- MICCAI 2023",
year="2023",
publisher="Springer Nature Switzerland",
address="Cham",
pages="651--662",
isbn="978-3-031-43907-0"
}
This research has been supported in part by ASU and Mayo Clinic through a Seed Grant and an Innovation Grant, and in part by the NIH under Award Number R01HL128785. The content is solely the responsibility of the authors and does not necessarily represent the official views of the NIH. This work has utilized the GPUs provided in part by the ASU Research Computing and in part by the Bridges-2 at Pittsburgh Supercomputing Center through allocation BCS190015 and the Anvil at Purdue University through allocation MED220025 from the Advanced Cyberinfrastructure Coordination Ecosystem: Services & Support (ACCESS) program, which is supported by National Science Foundation grants #2138259, #2138286, #2138307, #2137603, and #2138296. We also acknowledge Google for granting us access to CXR Foundation API, which enabled us to generate the embeddings for the target datasets. The content of this paper is covered by patents pending.
Released under the ASU GitHub Project License.