Prediction of Mechanical Properties of Cold-Rolled Steel Based on Improved Graph Attention Network
Abstract
:1. Introduction
2. Theoretical Foundation
2.1. Granger Causality
2.2. Graph Attention Networks
3. NGC–EGAT Method
3.1. NGC–EGAT Overall Framework
3.2. Neural Granger Causal Graph Structure
3.3. Embedding Graph Attention Network
4. Experiment and Analysis
4.1. Description of Experimental Data
4.2. Model Performance Evaluation Index and Parameter Setting
4.3. Structural Analysis of Neural Causality Graph
4.4. Model Performance Comparison
- GATv2 [18] is an improvement of GAT model. The key hyperparameters of GATv2 are set as the hidden size is 128, the number of heads is 7, and the learning rate is 0.01;
- GCN [19] is a semi-supervised deep learning model designed for graph-structured data. The key hyperparameters of GCN are set to the hidden size of 256 and the learning rate of 0.01;
- DeepGCNs [20], a variant of traditional GCN, defines a differentiable generalized aggregation function to unify different message aggregation operations, adopts a deeper structure, and solves the problem of information disappearance in graph-structured data. The key hyperparameters of DeepGCNs are set as the size of hidden is 128, the number of model layers is 4, and the learning rate is 0.01;
- GraphUNet [21] is a U-Net model based on graph-structured data. It realizes the feature learning of graph nodes and hierarchical representation of graph data through the graph convolution operation of hierarchical structure. The key hyperparameters for GraphUNet are set to a hidden size of 256, a U-Net depth of 4, and a learning rate of 0.01;
- CNN–LSTM [22] is a deep neural network that integrates CNN and LSTM models. The hidden layer size of CNN–LSTM is 64, the time step is 2, the model has 2 layers, and the learning rate is set to 0.01.
4.5. Ablation Experiment
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Diao, Y.; Yan, L.; Gao, K. A strategy assisted machine learning to process multi-objective optimization for improving mechanical properties of carbon steels. J. Mater. Sci. Technol. 2022, 109, 86–93. [Google Scholar] [CrossRef]
- Jacobs, L.J.M.; Atzema, E.H.; Moerman, J.; de Rooij, M.B. Quantification of the in fluence of Anisotropic Plastic Yielding on Cold Rolling Force. J. Mater. Process. Technol. 2023, 319, 118055. [Google Scholar] [CrossRef]
- Sheng, H.; Wang, P.; Tang, C. Predicting mechanical properties of cold-rolled steel strips using micro-magnetic ndt technologies. Materials 2022, 15, 2151. [Google Scholar] [CrossRef] [PubMed]
- Kano, M.; Nakagawa, Y. Data-based process monitoring, process control, and quality improvement: Recent developments and applications in steel industry. Comput. Chem. Eng. 2008, 32, 12–24. [Google Scholar] [CrossRef]
- Li, W.; Gu, J.; Deng, Y.; Mu, W.; Li, J. New comprehension on the microstructure, texture and deformation behaviors of UNS S32101 duplex stainless steel fabricated by direct cold rolling process. Mater. Sci. Eng. A 2022, 845, 143150. [Google Scholar] [CrossRef]
- Li, F.; He, A.; Song, Y.; Wang, Z.; Xu, X.; Zhang, S. Deep learning for predictive mechanical properties of hot-rolled strip in complex manufacturing systems. Int. J. Miner. Metall. Mater. 2022, 30, 1093–1103. [Google Scholar] [CrossRef]
- Yan, Y.F.; Lü, Z.M. Multi-objective quality control method for cold-rolled products oriented to customized requirements. Int. J. Miner. Metall. Mater. 2021, 28, 1332–1342. [Google Scholar] [CrossRef]
- Liu, X.; Cong, Z.; Peng, K.; Dong, J.; Li, L. DA-CBGRU-Seq2Seq based soft sensor for mechanical properties of hot rolling process. IEEE Sens. J. 2023, 23, 14234–14244. [Google Scholar] [CrossRef]
- Xu, Z.W.; Liu, X.M.; Zhang, K. Mechanical properties prediction for hot rolled alloy steel using convolutional neural network. IEEE Access 2019, 7, 47068–47078. [Google Scholar] [CrossRef]
- Chen, H.; Jiang, Y.; Zhang, X.; Zhou, Y.; Wang, L.; Wei, J. Spatio-Temporal Graph Attention Network for Sintering Temperature Long-Range Forecasting in Rotary Kilns. IEEE Trans. Ind. Inform. 2022, 19, 1923–1932. [Google Scholar] [CrossRef]
- Sun, B.; Lv, M.; Zhou, C.; Li, Y. A multimode structured prediction model based on dynamic attribution graph attention network for complex industrial processes. Inf. Sci. 2023, 640, 119001. [Google Scholar] [CrossRef]
- Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2021, arXiv:1710.10903. [Google Scholar]
- Tank, A.; Covert, I.; Foti, N.; Shojaie, A.; Fox, E.B. Neural granger causality. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4267–4279. [Google Scholar] [CrossRef] [PubMed]
- Yang, Y.; Wang, X.; Song, M.; Yuan, J.; Tao, D. Spagan: Shortest path graph attention network. arXiv 2021, arXiv:2101.03464. [Google Scholar]
- Grover, A.; Leskovec, J. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 855–864. [Google Scholar]
- Choudhury, A. Prediction and analysis of mechanical properties of low carbon steels using machine learning. J. Inst. Eng. (India) Ser. D 2022, 103, 303–310. [Google Scholar] [CrossRef]
- Li, X.; Zheng, M.; Yang, X.; Chen, P.; Ding, W. A property-oriented design strategy of high-strength ductile RAFM steels based on machine learning. Mater. Sci. Eng. A 2022, 840, 142891. [Google Scholar] [CrossRef]
- Brody, S.; Alon, U.; Yahav, E. How attentive are graph attention networks? arXiv 2021, arXiv:2105.14491. [Google Scholar]
- Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
- Li, G.; Xiong, C.; Thabet, A.; Ghanem, B. Deepergcn: All you need to train deeper gcns. arXiv 2020, arXiv:2006.07739. [Google Scholar]
- Gao, H.; Ji, S. Graph u-nets. In International Conference on Machine Learning; PMLR: Long Beach, CA, USA, 9–15 June 2019; pp. 2083–2092. [Google Scholar]
- Kim, T.Y.; Cho, S.B. Predicting residential energy consumption using CNN-LSTM neural networks. Energy 2019, 182, 72–81. [Google Scholar] [CrossRef]
Parameters | Values | Parameters | Values |
---|---|---|---|
hid_dim | 512 | num_heads | 14 |
emb_dim | 128 | walk_len | 10 |
cont_size | 10 | walks_pn | 10 |
Methods | Target Variables | RMSE | R2 | MAE |
---|---|---|---|---|
NGC | YS | 2.926 | 0.930 | 1.859 |
TS | 2.501 | 0.947 | 1.518 | |
EL | 0.596 | 0.970 | 0.356 | |
Pearson | YS | 2.947 | 0.926 | 1.842 |
TS | 2.538 | 0.945 | 1.573 | |
EL | 0.692 | 0.960 | 0.481 |
Target Variables | Metrics | Model | |||||
---|---|---|---|---|---|---|---|
GATv2 | GCN | DeepGCNs | GraphUNet | CNN–LSTM | NGC–EGAT | ||
YS | RMSE | 3.803 | 3.903 | 3.035 | 3.486 | 4.764 | 2.926 |
0.878 | 0.871 | 0.922 | 0.897 | 0.831 | 0.930 | ||
MAE | 2.704 | 2.674 | 1.921 | 2.407 | 2.920 | 1.859 | |
TS | RMSE | 3.135 | 3.086 | 2.658 | 3.122 | 4.673 | 2.501 |
0.916 | 0.919 | 0.940 | 0.917 | 0.839 | 0.947 | ||
MAE | 1.969 | 2.066 | 1.637 | 2.117 | 3.554 | 1.518 | |
EL | RMSE | 0.712 | 0.699 | 0.642 | 0.712 | 1.221 | 0.596 |
0.957 | 0.959 | 0.965 | 0.957 | 0.884 | 0.970 | ||
MAE | 0.482 | 0.480 | 0.432 | 0.497 | 0.930 | 0.356 |
Methods | Target Variables | RMSE | MAE | |
---|---|---|---|---|
NGC–GCN | YS | 3.903 | 0.871 | 2.674 |
TS | 3.086 | 0.919 | 2.066 | |
EL | 0.699 | 0.959 | 0.480 | |
NGC–GAT | YS | 3.827 | 0.876 | 2.466 |
TS | 3.029 | 0.922 | 1.911 | |
EL | 0.685 | 0.960 | 0.430 | |
NGC–EGAT | YS | 2.926 | 0.930 | 1.859 |
TS | 2.501 | 0.947 | 1.518 | |
EL | 0.596 | 0.970 | 0.356 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Luo, X.; Guo, R.; Zhang, Q.; Tang, X. Prediction of Mechanical Properties of Cold-Rolled Steel Based on Improved Graph Attention Network. Symmetry 2024, 16, 188. https://doi.org/10.3390/sym16020188
Luo X, Guo R, Zhang Q, Tang X. Prediction of Mechanical Properties of Cold-Rolled Steel Based on Improved Graph Attention Network. Symmetry. 2024; 16(2):188. https://doi.org/10.3390/sym16020188
Chicago/Turabian StyleLuo, Xiaoyang, Rongping Guo, Qiwen Zhang, and Xingchang Tang. 2024. "Prediction of Mechanical Properties of Cold-Rolled Steel Based on Improved Graph Attention Network" Symmetry 16, no. 2: 188. https://doi.org/10.3390/sym16020188
APA StyleLuo, X., Guo, R., Zhang, Q., & Tang, X. (2024). Prediction of Mechanical Properties of Cold-Rolled Steel Based on Improved Graph Attention Network. Symmetry, 16(2), 188. https://doi.org/10.3390/sym16020188