عرض بسيط للتسجيلة

المؤلفHu, Minghui
المؤلفGao, Ruobin
المؤلفSuganthan, Ponnuthurai Nagaratnam
تاريخ الإتاحة2025-01-20T05:12:04Z
تاريخ النشر2023
اسم المنشورIEEE Transactions on Neural Networks and Learning Systems
المصدرScopus
المعرّفhttp://dx.doi.org/10.1109/TNNLS.2023.3292063
الرقم المعياري الدولي للكتاب2162237X
معرّف المصادر الموحدhttp://hdl.handle.net/10576/62283
الملخصKnowledge distillation (KD) is a conventional method in the field of deep learning that enables the transfer of dark knowledge from a teacher model to a student model, consequently improving the performance of the student model. In randomized neural networks, due to the simple topology of network architecture and the insignificant relationship between model performance and model size, KD is not able to improve model performance. In this work, we propose a self-distillation pipeline for randomized neural networks: the predictions of the network itself are regarded as the additional target, which are mixed with the weighted original target as a distillation target containing dark knowledge to supervise the training of the model. All the predictions during multi-generation self-distillation process can be integrated by a multi-teacher method. By induction, we have additionally arrived at the methods for infinite self-distillation (ISD) of randomized neural networks. We then provide relevant theoretical analysis about the self-distillation method for randomized neural networks. Furthermore, we demonstrated the effectiveness of the proposed method in practical applications on several benchmark datasets.
اللغةen
الناشرInstitute of Electrical and Electronics Engineers Inc.
الموضوعBiological neural networks
Closed-form solutions
Knowledge distillation (KD)
Knowledge engineering
Neurons
Pipelines
Predictive models
random vector functional link (RVFL)
randomized neural network
self-distillation
Training
العنوانSelf-Distillation for Randomized Neural Networks
النوعArticle
الصفحات1-10
dc.accessType Open Access


الملفات في هذه التسجيلة

Thumbnail

هذه التسجيلة تظهر في المجموعات التالية

عرض بسيط للتسجيلة