Application of a Multi-Teacher Distillation Regression Model Based on Clustering Integration and Adaptive Weighting in Dam Deformation Prediction

被引:0
作者
Guo, Fawang [1 ]
Yuan, Jiafan [2 ]
Li, Danyang [2 ]
Qin, Xue [2 ]
机构
[1] POWERCHINA Guiyang Engineering Corporation Limited, Guiyang
[2] College of Big Data and Information Engineering, Guizhou University, Guiyang
关键词
cluster ensemble; dam deformation prediction; deep learning; information entropy; knowledge distillation;
D O I
10.3390/w17070988
中图分类号
学科分类号
摘要
Deformation is a key physical quantity that reflects the safety status of dams. Dam deformation is influenced by multiple factors and has seasonal and periodic patterns. Due to the challenges in accurately predicting dam deformation with traditional linear models, deep learning methods have been increasingly applied in recent years. In response to the problems such as an excessively long training time, too-high model complexity, and the limited generalization ability of a large number of complex hybrid models in the current research field, we propose an improved multi-teacher distillation network for regression tasks to improve the performance of the model. The multi-teacher network is constructed using a Transformer that considers global dependencies, while the student network is constructed using Temporal Convolutional Network (TCN). To improve distillation efficiency, we draw on the concept of clustering integration to reduce the number of teacher networks and propose a loss function for regression tasks. We incorporate an adaptive weight module into the loss function and assign more weight to the teachers with more accurate prediction results. Finally, knowledge information is formed based on the differences between the teacher networks and the student network. The model is applied to a concrete-faced rockfill dam located in Guizhou province, China, and the results demonstrate that, compared to other knowledge distillation methods, this approach exhibits higher accuracy and practicality. © 2025 by the authors.
引用
收藏
相关论文
共 47 条
[41]  
Zhang J., Badcleaner: Defending backdoor attacks in federated learning via attention-based multi-teacher distillation, IEEE Trans. Dependable Secur. Comput, 21, pp. 4559-4573, (2024)
[42]  
Bai L., An information-theoretical framework for cluster ensemble, IEEE Trans. Knowl. Data Eng, 31, pp. 1464-1477, (2018)
[43]  
Lee K., Fast and accurate facial expression image classification and regression method based on knowledge distillation, Appl. Sci, 13, (2023)
[44]  
Bai S., An empirical evaluation of generic convolutional and recurrent networks for sequence modeling, arXiv, (2018)
[45]  
Saputra M.R.U., Distilling knowledge from a deep pose regressor network, Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 263-272
[46]  
Chen G., Choi W., Yu X., Han T., Chandraker M., Learning efficient object detection models with knowledge distillation, Adv. Neural Inf. Process. Syst, 30, pp. 742-751, (2017)
[47]  
Takamoto M., An efficient method of training small models for regression problems with knowledge distillation, Proceedings of the 2020 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), pp. 67-72