liu.seSearch for publications in DiVA
Change search
Link to record
Permanent link

Direct link
Akbar, Muhammad Usman
Publications (5 of 5) Show all publications
Akbar, M. U., Wang, W. & Eklund, A. (2025). Beware of diffusion models for synthesizing medical images - A comparison with GANs in terms of memorizing brain MRI and chest x-ray images. Machine Learning: Science and Technology
Open this publication in new window or tab >>Beware of diffusion models for synthesizing medical images - A comparison with GANs in terms of memorizing brain MRI and chest x-ray images
2025 (English)In: Machine Learning: Science and Technology, E-ISSN 2632-2153Article in journal (Refereed) Published
Abstract [en]

Diffusion models were initially developed for text-to-image generation and are now being utilized to generate high quality synthetic images. Preceded by GANs, diffusion models have shown impressive results using various evaluation metrics. However, commonly used metrics such as FID and IS are not suitable for determining whether diffusion models are simply reproducing the training images. Here we train StyleGAN and a diffusion model, using  BRATS20, BRATS21 and a chest x-ray pneumonia dataset, to synthesize brain MRI and chest x-ray images, and measure the correlation between the synthetic images and all training images. Our results show that diffusion models are more likely to memorize the training images, compared to StyleGAN, especially for small datasets and when using 2D slices from 3D volumes. Researchers should be careful when using diffusion models (and to some extent GANs) for medical imaging, if the final goal is to share the synthetic images. 

Keywords
Synthetic images, GANs, diffusion models, memorization
National Category
Radiology, Nuclear Medicine and Medical Imaging Probability Theory and Statistics
Identifiers
urn:nbn:se:liu:diva-210499 (URN)10.1088/2632-2153/ad9a3a (DOI)001408876900001 ()2-s2.0-85217039477 (Scopus ID)
Funder
Åke Wiberg Foundation, M22-0088Vinnova, 2021-01954
Note

Funding Agencies|ITEA/VINNOVA project ASSIST [2021-01954]; LiU Cancer and the akeWiberg foundation

Available from: 2024-12-16 Created: 2024-12-16 Last updated: 2025-03-03
Batool, H., Mukhtar, A., Khawaja, S. G., Alghamdi, N. S., Khan, A. M., Qayyum, A., . . . Eklund, A. (2025). Knowledge Distillation and Transformer Based Framework for Automatic Spine CT Report Generation. IEEE Access, 1-1
Open this publication in new window or tab >>Knowledge Distillation and Transformer Based Framework for Automatic Spine CT Report Generation
Show others...
2025 (English)In: IEEE Access, E-ISSN 2169-3536, p. 1-1Article in journal (Refereed) Published
Abstract [en]

Spine Computed Tomography (SCT) is essential for identifying fractures, tumors and degenerative spine diseases, assisting medical practitioners in formulating an accurate diagnosis and treatment. One of the core element of SCT is reporting. The effectiveness of spine reporting is often limited by challenges such as an inadequate infrastructure and lack of experts. Automated SCT analysis has the potential to revolutionize spinal healthcare and improve patient outcomes. To achieve this objective, we proposed a framework for spine report generation that utilizes transformer architecture, trained on textual reports alongside the visual features extracted from the sagittal slices of the SCT volume. A foundation model is used to perform Knowledge Distillation (KD) alongside an encoder to ensure an optimal performance. The proposed framework is evaluated on the public dataset (VerSe20). The incorporation of KD results improved both the BERT and BLEU1 score on the dataset, from 0.7486 to 0.7522 and 0.6361 to 0.7291. Additionally, the proposed framework is evaluated using four different types of reports: original radiologist reports, reports without spine-level annotations, rephrased reports, and reports generated by ChatGPT-4o (ChatGPT). The evaluation without spine-level annotations demonstrates superior performance across most metrics, achieving the highest BLEU-1 and ROUGE-L scores, with a BLEU-1 of 0.9293 and a ROUGE-L score of 0.9297. In contrast, the other techniques achieved moderate scores across all metrics. Finally, experienced radiologists assessed the spine report and have given high rating to the original reports across all three criteria (completeness, conciseness and correctness), in comparison to the generated reports. This study’s findings suggest that omitting spine-level annotations can improve the quality of text generation.

Keywords
Spine Report Generation, Knowledge Distillation, Foundation Model, ChatGPT
National Category
Radiology and Medical Imaging
Identifiers
urn:nbn:se:liu:diva-211967 (URN)10.1109/access.2025.3546131 (DOI)001446493800034 ()2-s2.0-105001061916 (Scopus ID)
Note

Funding Agencies|Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia [PNURSP2025R40]

Available from: 2025-03-01 Created: 2025-03-01 Last updated: 2025-04-08
Akbar, M. U., Larsson, M., Blystad, I. & Eklund, A. (2024). Brain tumor segmentation using synthetic MR images - A comparison of GANs and diffusion models. Scientific Data, 11(1), Article ID 259.
Open this publication in new window or tab >>Brain tumor segmentation using synthetic MR images - A comparison of GANs and diffusion models
2024 (English)In: Scientific Data, E-ISSN 2052-4463, Vol. 11, no 1, article id 259Article in journal (Refereed) Published
Abstract [en]

Large annotated datasets are required for training deep learning models, but in medical imaging data sharing is often complicated due to ethics, anonymization and data protection legislation. Generative AI models, such as generative adversarial networks (GANs) and diffusion models, can today produce very realistic synthetic images, and can potentially facilitate data sharing. However, in order to share synthetic medical images it must first be demonstrated that they can be used for training different networks with acceptable performance. Here, we therefore comprehensively evaluate four GANs (progressive GAN, StyleGAN 1–3) and a diffusion model for the task of brain tumor segmentation (using two segmentation networks, U-Net and a Swin transformer). Our results show that segmentation networks trained on synthetic images reach Dice scores that are 80%–90% of Dice scores when training with real images, but that memorization of the training images can be a problem for diffusion models if the original dataset is too small. Our conclusion is that sharing synthetic medical images is a viable option to sharing real images, but that further work is required. The trained generative models and the generated synthetic images are shared on AIDA data hub.

Place, publisher, year, edition, pages
Nature Publishing Group, 2024
Keywords
Deep learning, brain tumor, magnetic resonance imaging, synthetic images, generative adversarial networks, diffusion models
National Category
Radiology, Nuclear Medicine and Medical Imaging Medical Imaging
Identifiers
urn:nbn:se:liu:diva-201435 (URN)10.1038/s41597-024-03073-x (DOI)001177063000006 ()38424097 (PubMedID)2-s2.0-85186294143 (Scopus ID)
Funder
Vinnova, 2021-01954Vinnova, 2021-01420Åke Wiberg Foundation, M22-0088
Note

Funding Agencies|ITEA/VINNOVA project ASSIST [2021-01420]; LiU Cancer; VINNOVA AIDA [M22-0088]; Ake Wiberg foundation; Wallenberg Center for Molecular Medicine as an associated clinical fellow;  [2021-01954]

Available from: 2024-03-09 Created: 2024-03-09 Last updated: 2025-02-09Bibliographically approved
Gustafsson, C. J., Löfstedt, T., Åkesson, M., Rogowski, V., Akbar, M. U., Hellander, A., . . . Eklund, A. (2024). Federated training of segmentation models for radiation therapy treatment planning. Paper presented at ESTRO. Radiotherapy and Oncology, 194, S4819-S4822
Open this publication in new window or tab >>Federated training of segmentation models for radiation therapy treatment planning
Show others...
2024 (English)In: Radiotherapy and Oncology, ISSN 0167-8140, E-ISSN 1879-0887, Vol. 194, p. S4819-S4822Article in journal, Meeting abstract (Refereed) Published
Abstract [en]

Radiotherapy treatment planning takes substantial time, several hours per patient, as it involves manual segmentation of tumor and risk organs. Segmentation networks can be trained to automatically perform the segmentations, but typically require large annotated datasets for training. Sharing of sensitive data between hospitals, to create a larger dataset, is often difficult due to ethics and GDPR. Here we therefore demonstrate that federated learning is a solution to this problem, as then only the segmentation model is sent between each hospital and a global server. We export and preprocess brain tumor images from the oncology departments in Linköping and Lund, and use federated learning to train a global segmentation model using two different frameworks.

Place, publisher, year, edition, pages
Elsevier, 2024
Keywords
Radiotherapy, deep learning, federated learning
National Category
Medical Imaging Cancer and Oncology
Identifiers
urn:nbn:se:liu:diva-207369 (URN)10.1016/s0167-8140(24)01903-0 (DOI)
Conference
ESTRO
Funder
Vinnova, 2021-01954
Available from: 2024-09-07 Created: 2024-09-07 Last updated: 2025-08-30
Ericsson, L., Hjalmarsson, A., Akbar, M. U., Ferdian, E., Bonini, M., Hardy, B., . . . Marlevi, D. (2024). Generalized super-resolution 4D Flow MRI - using ensemble learning to extend across the cardiovascular system. IEEE journal of biomedical and health informatics, 28(12), 7239-7250
Open this publication in new window or tab >>Generalized super-resolution 4D Flow MRI - using ensemble learning to extend across the cardiovascular system
Show others...
2024 (English)In: IEEE journal of biomedical and health informatics, ISSN 2168-2194, E-ISSN 2168-2208, Vol. 28, no 12, p. 7239-7250Article in journal (Refereed) Published
Abstract [en]

4D Flow Magnetic Resonance Imaging (4D Flow MRI) is a non-invasive measurement technique capable of quantifying blood flow across the cardiovascular system. While practical use is limited by spatial resolution and image noise, incorporation of trained super-resolution (SR) networks has potential to enhance image quality post-scan. However, these efforts have predominantly been restricted to narrowly defined cardiovascular domains, with limited exploration of how SR performance extends across the cardiovascular system; a task aggravated by contrasting hemodynamic conditions apparent across the cardiovasculature. The aim of our study was therefore to explore the generalizability of SR 4D Flow MRI using a combination of existing super-resolution base models, novel heterogeneous training sets, and dedicated ensemble learning techniques; the latter-most being effectively used for improved domain adaption in other domains or modalities, however, with no previous exploration in the setting of 4D Flow MRI. With synthetic training data generated across three disparate domains (cardiac, aortic, cerebrovascular), varying convolutional base and ensemble learners were evaluated as a function of domain and architecture, quantifying performance on both in-silico and acquired in-vivo data from the same three domains. Results show that both bagging and stacking ensembling enhance SR performance across domains, accurately predicting high-resolution velocities from low-resolution input data in-silico. Likewise, optimized networks successfully recover native resolution velocities from downsampled in-vivo data, as well as show qualitative potential in generating denoised SR-images from clinicallevel input data. In conclusion, our work presents a viable approach for generalized SR 4D Flow MRI, with the novel use of ensemble learning in the setting of advanced fullfield flow imaging extending utility across various clinical areas of interest.

Place, publisher, year, edition, pages
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC, 2024
Keywords
Superresolution, Magnetic resonance imaging, Data models, Training, Ensemble learning, Biomedical imaging, Hemodynamics
National Category
Cardiology and Cardiovascular Disease Medical Imaging Radiology, Nuclear Medicine and Medical Imaging
Identifiers
urn:nbn:se:liu:diva-205978 (URN)10.1109/jbhi.2024.3429291 (DOI)001373825400019 ()39012742 (PubMedID)2-s2.0-85198708713 (Scopus ID)
Note

Funding Agencies|European Union ERC, MultiPRESS [101075494]; NIH [R01HL170059]

Available from: 2024-07-19 Created: 2024-07-19 Last updated: 2025-02-10
Organisations

Search in DiVA

Show all publications