Document Type

Article

Publication Title

Proceedings of the 14th International Conference on Data Science, Technology and Applications (DATA 2025)

Publication Date

6-23-2025

Meeting Name

14th International Conference on Data Science, Technology and Applications (DATA 2025)

Meeting Date

June 10-12, 2025

Meeting Location

Bilbao, Spain

Abstract/ Summary

This paper introduces a tamper-resistant framework for large language models (LLMs) in medical applications, utilizing quantum gradient descent (QGD) to detect malicious parameter modifications in real time. Integrated into a LLaMA-based model, QGD monitors weight amplitude distributions, identifying adversarial fine-tuning anomalies. Tests on the MIMIC and eICU datasets show minimal performance impact (accuracy: 89.1 to 88.3 on MIMIC) while robustly detecting tampering. PubMedQA evaluations confirm preserved biomedical question-answering capabilities. Compared to baselines like selective unlearning and cryptographic fingerprinting, QGD offers superior sensitivity to subtle weight changes. This quantum-inspired approach ensures secure, reliable medical AI, extensible to other high-stakes domains.

Share

COinS