A Generative AI-Based Decision Support Framework for Early Prediction and Management of Diabetes Mellitus

Keywords

Generative Artificial Intelligence
Decision Support System
Diabetes Mellitus
Deep Learning
Reinforcement Learning

How to Cite

Lalwani , S., & Gupta , H. (2025). A Generative AI-Based Decision Support Framework for Early Prediction and Management of Diabetes Mellitus. PromptAI Academy Journal, 5, e093. https://doi.org/10.37497/PromptAI.5.2026.93
Bookmark and Share

Abstract

Purpose: This paper presents a Generative Artificial Intelligence (GenAI)-based Decision Support System (DSS) framework for the early prediction and clinical management of diabetes mellitus. The study aims to integrate generative modeling, counterfactual simulation, and reinforcement learning into a unified architecture to enhance diagnostic precision, explainability, and personalized treatment recommendations.

Design/Methodology/Approach: The proposed Generative Decision Support Model (GDSM) employs multimodal generative pre-training (GMP) using longitudinal Electronic Health Records (EHRs), combining structured and unstructured data such as laboratory investigations, prescriptions, and clinical narratives. A transformer-based generative backbone synthesizes patient trajectories through cross-modal representation learning, temporal sparse attention, and contrastive alignment. A world model simulator enables counterfactual reasoning for evaluating hypothetical clinical interventions, while the policy optimization layer applies conservative offline reinforcement learning to generate safe and actionable recommendations.

Findings: Experimental results demonstrate that the GDSM significantly improves early diabetes prediction accuracy and clinical interpretability compared to conventional rule-based or discriminative models. By simulating patient responses and optimizing treatment strategies, the framework enhances data efficiency, calibration, and reliability of predictive outcomes. The integration of SHAP-based explainability and Human-in-the-Loop (HITL) feedback mechanisms increases model transparency and clinician trust.

Practical Implications: The framework enables personalized diabetes care through predictive simulation and ethically grounded AI decision support. It also establishes a scalable pathway for generative AI adoption in clinical informatics and chronic disease management.

Originality/Value: This study introduces a next-generation generative DSS architecture that transitions from predictive analytics to proactive, interpretable, and patient-centered clinical decision-making.

https://doi.org/10.37497/PromptAI.5.2026.93

References

Achiam, J., Held, D., Tamar, A., & Abbeel, P. (2017). Constrained policy optimization. In Proceedings of the 34th International Conference on Machine Learning (ICML).

Beltagy, I., Peters, M. E., & Cohan, A. (2020). Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150.

Brown, T. B., Mann, B., Ryder, N., et al. (2020). Language models are few-shot learners. In Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS).

Che, S.-H., Purushotham, C., Cho, K., Sontag, D., & Liu, Y. (2018). Recurrent neural networks for multivariate time series with missing values (GRU-D). Scientific Reports, 8, 6085.

Choi, E., Bahadori, M. T., Schuetz, A., Stewart, W. F., & Sun, J. (2016). RETAIN: An interpretable predictive model for healthcare using reverse time attention. In Proceedings of the 30th Conference on Neural Information Processing Systems (NeurIPS).

Choi, E., Biswal, S., Malin, B., Duke, J., Stewart, W. F., & Sun, J. (2017). Generating multi-label discrete patient records using generative adversarial networks (medGAN). In Proceedings of the Machine Learning for Healthcare Conference (MLHC).

Child, R., Gray, S., Radford, A., & Sutskever, I. (2019). Generating long sequences with sparse transformers. arXiv preprint arXiv:1904.10509.

Cohen, J. P., Peng, E., & Solomon, A. (2019). ClinicalBERT: Modeling clinical notes and predicting hospital readmission. arXiv preprint arXiv:1904.05342.

Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL-HLT).

Dosovitskiy, A., Beyer, L., Kolesnikov, A., et al. (2021). An image is worth 16×16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations (ICLR).

Dudík, M., Langford, J., & Li, L. (2011). Doubly robust policy evaluation and learning. Machine Learning, 76(1), 33–51.

Esteban, C., Hyland, S. L., & Rätsch, G. (2017). Real-valued (medical) time series generation with recurrent conditional GANs. arXiv preprint arXiv:1706.02633.

Gottesman, O., Johansson, F., Meier, J., et al. (2019). Guidelines for reinforcement learning in healthcare. Nature Medicine, 25(1), 16–18.

Gu, Y., Tinn, R., Cheng, H., et al. (2021). Domain-specific language model pretraining for biomedical NLP (PubMedBERT). In Findings of the Association for Computational Linguistics (ACL Findings).

Ha, D., & Schmidhuber, J. (2018). World models. arXiv preprint arXiv:1803.10122.

Hafner, D., Lillicrap, T., Fischer, I., et al. (2019). Learning latent dynamics for planning from pixels (PlaNet). In Proceedings of the International Conference on Machine Learning (ICML).

Jaegle, A., Borgeaud, S., Alayrac, J., et al. (2021). Perceiver IO: A general architecture for structured inputs & outputs. In Proceedings of the 38th International Conference on Machine Learning (ICML).

Jiang, N., & Li, L. (2016). Doubly robust off-policy value evaluation for reinforcement learning. In Proceedings of the 33rd International Conference on Machine Learning (ICML).

Kitaev, N., Kaiser, Ł., & Levskaya, A. (2020). Reformer: The efficient transformer. In Proceedings of the International Conference on Learning Representations (ICLR).

Komorowski, M., Celi, A., Badawi, O., Gordon, A., & Faisal, A. (2018). The artificial intelligence clinician learns optimal treatment strategies for sepsis. Nature Medicine, 24(11), 1716–1720.

Kumar, A., Zhou, A., Tucker, G., & Levine, S. (2020). Conservative Q-learning for offline reinforcement learning. In Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS).

Lee, J., Yoon, W., Kim, S., et al. (2020). BioBERT: A pre-trained biomedical language representation model for biomedical text mining. Bioinformatics, 36(4), 1234–1240.

Levine, S., Kumar, A., Tucker, G., & Fu, J. (2020). Offline reinforcement learning: Tutorial, review, and perspectives. arXiv preprint arXiv:2005.01643.

Li, Y., Rao, S., Solares, J., et al. (2020). BEHRT: Transformer for electronic health records. Scientific Reports, 10, 7155.

Lundberg, S. M., & Lee, S.-I. (2017). A unified approach to interpreting model predictions. In Proceedings of the 31st Conference on Neural Information Processing Systems (NeurIPS).

Radford, A., Kim, J. W., Hallacy, C., et al. (2021). Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning (ICML).

Rasmy, M., Xiang, Y., Xie, Z., et al. (2021). Med-BERT: Pretrained contextualized embeddings on large-scale EHR for disease prediction. NPJ Digital Medicine, 4(86), 1–13.

Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD).

Song, H., Rajan, D., Thiagarajan, J., & Spanias, A. (2018). Attend and diagnose: Clinical time series analysis using attention models. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence (AAAI).

Thomas, P., Theocharous, G., & Ghavamzadeh, M. (2015). High-confidence off-policy evaluation. In Proceedings of the 29th AAAI Conference on Artificial Intelligence (AAAI).

Vaswani, A., Shazeer, N., Parmar, N., et al. (2017). Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NeurIPS).

Wang, E., Sun, Z., Wang, H., et al. (2022). MedCLIP: Contrastive learning from paired text and medical images. arXiv preprint arXiv:2210.10163.