Show simple item record

AuthorUllah, Imdad
AuthorHassan, Najm
AuthorGill, Sukhpal Singh
AuthorSuleiman, Basem
AuthorAhanger, Tariq Ahamed
AuthorShah, Zawar
AuthorQadir, Junaid
AuthorKanhere, Salil S.
Available date2025-07-08T03:58:11Z
Publication Date2024
Publication NameIET Blockchain
ResourceScopus
Identifierhttp://dx.doi.org/10.1049/blc2.12091
ISSN26341573
URIhttp://hdl.handle.net/10576/66085
AbstractThe generative Artificial Intelligence (AI) tools based on Large Language Models (LLMs) use billions of parameters to extensively analyse large datasets and extract critical information such as context, specific details, identifying information, use this information in the training process, and generate responses for the requested queries. The extracted data also contain sensitive information, seriously threatening user privacy and reluctance to use such tools. This article proposes the conceptual model called PrivChatGPT, a privacy-preserving model for LLMs consisting of two main components, that is, preserving user privacy during the data curation/pre-processing and preserving private context and the private training process for large-scale data. To demonstrate the applicability of PrivChatGPT, it is shown how a private mechanism could be integrated into the existing model for training LLMs to protect user privacy; specifically, differential privacy and private training using Reinforcement Learning (RL) were employed. The privacy level probabilities are associated with the document contents, including the private contextual information, and with metadata, which is used to evaluate the disclosure probability loss for an individual's private information. The privacy loss is measured and the measure of uncertainty or randomness is evaluated using entropy once differential privacy is applied. It recursively evaluates the level of privacy guarantees and the uncertainty of public databases and resources during each update when new information is added for training purposes. To critically evaluate the use of differential privacy for private LLMs, other mechanisms were hypothetically compared such as Blockchain, private information retrieval, randomisation, obfuscation, anonymisation, and the use of Tor for various performance measures such as the model performance and accuracy, computational complexity, privacy vs. utility, training latency, vulnerability to attacks, and resource consumption. It is concluded that differential privacy, randomisation, and obfuscation can impact the training models' utility and performance; conversely, using Tor, Blockchain, and Private Information Retrieval (PIR) may introduce additional computational complexity and high training latency. It is believed that the proposed model could be used as a benchmark for privacy-preserving LLMs for generative AI tools.
Languageen
PublisherJohn Wiley and Sons Inc
Subjectartificial intelligence
blockchain applications and digital technology
blockchain platforms
blockchain standards
data protection
information security
models and analysis
security of data
TitlePrivacy preserving large language models: ChatGPT case study based vision and framework
TypeArticle
Pagination706-724
Issue NumberS1
Volume Number4
dc.accessType Open Access


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record