Longitudinal crack prediction method of continuous cast slab based on random forest and clustering
ZHANG He1,2, DUAN Hai-yang1,2, WANG Xu-dong1,2, YAO Man1,2
1. School of Materials Science and Engineering, Dalian University of Technology, Dalian 116024, Liaoning, China; 2. Key Laboratory of Solidification Control and Digital Preparation Technology (Liaoning Province), Dalian 116024, Liaoning, China
Abstract:Longitudinal crack is a common surface defect of casting slab. Accurate prediction of longitudinal cracking on slab surface is of great significance for improving the quality of casting slab. Aiming at the temporal and spatial variation trends of the temperature of the mold thermocouple during the formation and propagation of longitudinal cracks, this paper captures and extracts the typical variation features of the thermocouple temperature in time series, and the Random Forest(RF) algorithm is used to reduce the dimension of the captured features, and the features closely related to longitudinal cracks are extracted. On this basis, a longitudinal crack detection model based on K-means(K Means) clustering was established. The results show that the proposed longitudinal crack prediction model based on temperature-time series features and clustering algorithm can correctly distinguish and identify samples with longitudinal cracks from samples under normal conditions, which provides feasible way for introducing machine learning methods into abnormal monitoring of continuous casting process.
Nakato H, Ozawa M, Kinoshita K, et al. Factors affecting the formation of shell and longitudinal cracks in mold during high speed continuous casting of slabs[J]. Transactions of the Iron, 1984, 24(11): 957.
Lieftucht I D, Arzberger M, Reifferscheid M, et al. Online prediction of longitudinal facial cracks in thin slab casting using principal component analysis[J]. Journal of Iron Steel Research, International,2008, 15(S1): 255.
Breiman L J M L. Random forests[J]. Machine Learning,2001, 45(1): 5.
[17]
Quinlan J R. Induction of decision trees[J]. Machine Learning,1986, 1(1): 81.
[18]
Degenhardt F, Seifert S, Szymczak S. Evaluation of variable selection methods for random forests and omics data sets[J]. Briefings in Bioinformatics, 2019, 20(2): 492.
[19]
Nainggolan R,Perangin-Angin R,Simarmata E, et al. Improved the performance of the K-Means cluster using the sum of squared error (SSE) optimized by using the elbow method[J]. Journal of Physics: Conference Series, 2019, 1361(1):012015.
[20]
Macqueen J. Some methods for classification and analysis of multivariate observations[C]//Proceedings of the Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability. Oakland, CA, USA,1967.
[21]
Arthur D,Vassilvitskii S. K-Means++: The advantages of careful seeding[C]//Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms.New Orleans, Louisiana, USA, 2007.