[1]ZHANG Xiongtao,CHEN Tianyu,ZHAO Kang,et al.TSK fuzzy classifier based on multi-teacher adaptive knowledge distillation[J].CAAI Transactions on Intelligent Systems,2025,20(5):1136-1147.[doi:10.11992/tis.202410028]
Copy
CAAI Transactions on Intelligent Systems[ISSN 1673-4785/CN 23-1538/TP] Volume:
20
Number of periods:
2025 5
Page number:
1136-1147
Column:
学术论文—机器感知与模式识别
Public date:
2025-09-05
- Title:
-
TSK fuzzy classifier based on multi-teacher adaptive knowledge distillation
- Author(s):
-
ZHANG Xiongtao1; 2; CHEN Tianyu1; 2; ZHAO Kang1; 2; LI Shuimiao2; 3; SHEN Qing1; 2
-
1. School of Information Engineering, Huzhou University, Huzhou 313000, China;
2. Zhejiang Province Key Laboratory of Smart Management and Application of Modern Agricultural Resources, Huzhou 313000, China;
3. Information Technology Center, Huzhou University, Huzhou 313000, China
-
- Keywords:
-
TSK fuzzy classifier; knowledge distillation; multiple teacher networks; adaptive allocation of weights; dark knowledge; fuzzy system; different perspectives; deep learning
- CLC:
-
TP181
- DOI:
-
10.11992/tis.202410028
- Abstract:
-
Currently, hierarchical and deep fuzzy systems demonstrate excellent performance, but they often suffer from high model complexity. Lightweight Takagi-Sugeno-Kang (TSK) fuzzy classifiers based on distillation learning typically rely on single-teacher knowledge distillation. However, if the teacher model underperforms, then the distillation effect and the overall model performance can be compromised. Furthermore, traditional multiteacher distillation approaches often assign weights to teacher model outputs using label-free strategies, which may allow low-quality teachers to mislead the student model. Aiming to address these issues, this paper introduces a TSK fuzzy classifier based on multiteacher adaptive knowledge distillation (TSK-MTAKD). The method employs multiple deep neural networks, each with different neural expression capabilities, as teacher models. The proposed distillation framework extracts dark knowledge from these models and transfers it to a TSK fuzzy system, leveraging its strong capability to handle uncertainty. Additionally, an adaptive weight allocator is introduced, which performs cross-entropy calculations between the output of the teacher model and the true label. Outputs that are closer to the true label are assigned higher weights, thereby improving model robustness and the quality of dark knowledge. Experimental results on 13 UCI benchmark datasets validate the advantages of the TSK-MTAKD approach.