عرض بسيط للتسجيلة

المؤلفUllah, Imdad
المؤلفHassan, Najm
المؤلفGill, Sukhpal Singh
المؤلفSuleiman, Basem
المؤلفAhanger, Tariq Ahamed
المؤلفShah, Zawar
المؤلفQadir, Junaid
المؤلفKanhere, Salil S.
تاريخ الإتاحة2025-07-08T03:58:11Z
تاريخ النشر2024
اسم المنشورIET Blockchain
المصدرScopus
المعرّفhttp://dx.doi.org/10.1049/blc2.12091
الرقم المعياري الدولي للكتاب26341573
معرّف المصادر الموحدhttp://hdl.handle.net/10576/66085
الملخصThe generative Artificial Intelligence (AI) tools based on Large Language Models (LLMs) use billions of parameters to extensively analyse large datasets and extract critical information such as context, specific details, identifying information, use this information in the training process, and generate responses for the requested queries. The extracted data also contain sensitive information, seriously threatening user privacy and reluctance to use such tools. This article proposes the conceptual model called PrivChatGPT, a privacy-preserving model for LLMs consisting of two main components, that is, preserving user privacy during the data curation/pre-processing and preserving private context and the private training process for large-scale data. To demonstrate the applicability of PrivChatGPT, it is shown how a private mechanism could be integrated into the existing model for training LLMs to protect user privacy; specifically, differential privacy and private training using Reinforcement Learning (RL) were employed. The privacy level probabilities are associated with the document contents, including the private contextual information, and with metadata, which is used to evaluate the disclosure probability loss for an individual's private information. The privacy loss is measured and the measure of uncertainty or randomness is evaluated using entropy once differential privacy is applied. It recursively evaluates the level of privacy guarantees and the uncertainty of public databases and resources during each update when new information is added for training purposes. To critically evaluate the use of differential privacy for private LLMs, other mechanisms were hypothetically compared such as Blockchain, private information retrieval, randomisation, obfuscation, anonymisation, and the use of Tor for various performance measures such as the model performance and accuracy, computational complexity, privacy vs. utility, training latency, vulnerability to attacks, and resource consumption. It is concluded that differential privacy, randomisation, and obfuscation can impact the training models' utility and performance; conversely, using Tor, Blockchain, and Private Information Retrieval (PIR) may introduce additional computational complexity and high training latency. It is believed that the proposed model could be used as a benchmark for privacy-preserving LLMs for generative AI tools.
اللغةen
الناشرJohn Wiley and Sons Inc
الموضوعartificial intelligence
blockchain applications and digital technology
blockchain platforms
blockchain standards
data protection
information security
models and analysis
security of data
العنوانPrivacy preserving large language models: ChatGPT case study based vision and framework
النوعArticle
الصفحات706-724
رقم العددS1
رقم المجلد4
dc.accessType Open Access


الملفات في هذه التسجيلة

Thumbnail

هذه التسجيلة تظهر في المجموعات التالية

عرض بسيط للتسجيلة