1. Introduction
Steel-reinforced concrete is one of the most important building materials used in the construction industry [
1]. The carbonization of concrete and corrosion of reinforcing steel due to air exposure results in a reduction in the strength and durability of the structure [
2]. Thus, non-corrosive materials have become more significant during the past three decades as precast and lightweight concrete constructions with great durability have been a priority [
3]. High-performance fibers implanted in the shape of a textile mesh are combined with conventional concrete to create textile-reinforced concrete (TRC) [
4,
5]. It is intended to improve the concrete’s mechanical properties, including tensile strength, ductility, and impact resistance, which are typically low in conventional concrete [
6,
7,
8,
9,
10,
11].
In TRC, the textile reinforcement serves as a replacement for traditional steel reinforcement, offering advantages like corrosion resistance, ease of handling and installation, and improved crack control [
12,
13]. The process of manufacturing textile-reinforced concrete involves applying a thin layer of cementitious mortar to the textile reinforcement, which is then impregnated with a high-performance cementitious matrix. The fibers’ strong bond to the concrete matrix is ensured by this impregnation, allowing for efficient load transfer between the two materials [
14]. The resulting TRC panels or elements can be used in various applications, including architectural facades, cladding, thin-shell structures, and precast elements. TRC offers design flexibility, allowing for thinner and lighter elements compared to traditional concrete, while maintaining or improving structural performance [
15,
16]. Recently, there has been an increase in interest in the usage of textile-reinforced concrete due to its potential for reducing material consumption, energy consumption during manufacturing, and carbon footprint [
17,
18]. Additionally, TRC exhibits good durability and fire resistance properties, making it an attractive choice for sustainable construction practices [
19,
20,
21,
22]. Based on the kind of fibers utilized for reinforcement and the arrangement of the textile reinforcement, there are various types of textile-reinforced concrete (TRC) [
23,
24,
25,
26,
27]. Fibers generally come in two types: natural and artificial. Animals, plants, and natural minerals are the primary sources of natural textile fibers; in contrast, man-made materials include ceramics and synthetic materials made with mineral fibers [
28]. Jute, flax, bamboo, cotton, sisal, and coir are a few examples of common natural fibers.
Table 1 displays man-made fabrics and their characteristics.
Additionally, textile reinforcements can be categorized based on their configurations, which include woven fabrics, non-woven fabrics, and grid structures. Woven fabrics consist of interlaced fibers, similar to traditional textiles, while non-woven fabrics have fibers bonded together mechanically or chemically. Grid structures involve the arrangement of fibers in a grid pattern, creating a mesh-like reinforcement. The specific type of TRC chosen depends on the requirements, such as the desired mechanical properties, environmental conditions, and cost considerations. Corrosion of the embedded metals and reinforcing steel is the primary cause of concrete deterioration. As a result, non-corrosive metals are now utilized, such as textile reinforcement. In previous studies, only textile reinforcement, excluding steel reinforcement, was used, and it did not give satisfactory flexural strength compared to conventional beams. In this study, carbon fabric has been chosen as a textile reinforcement to reinforce the concrete partially along with a steel reinforcement to improve the tensile strength of concrete and reduce concrete deterioration. As the reinforcement is placed over the fabric, the foreign ions will first reach the fabric, thus reducing the concrete deterioration. Structural health monitoring (SHM) is an important aspect of evaluating the safety, integrity, and functionality of structures [
29]. SHM involves the continuous or periodic monitoring of structural parameters to detect changes, damage, or potential failures in real-time [
30,
31]. Here are some commonly employed SHM techniques:
Strain Gauges—Strain gauges are frequently employed in structural health monitoring to measure the deformation or strain in structural components such as bridges, buildings, and other infrastructure. Strain gauges are embedded in the surface of the structural component where deformation or strain needs to be monitored. They are often attached to critical locations, such as areas prone to high stress or where damage is more likely to occur [
32,
33].
Fiber Optic Sensors—Because of their unique properties, fiber optic sensors have become increasingly common in structural health monitoring, including high sensitivity, resistance to electromagnetic interference, and their ability to cover large areas with distributed sensing. Fiber Bragg grating sensors and distributed fiber optic sensors are most commonly used [
34,
35].
Acoustic emission sensors are strategically placed on or within the structure being monitored. These sensors are sensitive to the high-frequency stress waves produced by structural changes [
36,
37,
38].
Electromagnetic techniques are employed in structural health monitoring to assess the condition of structures by utilizing electromagnetic waves and their interactions with the materials. These techniques are non-destructive and can provide valuable information about the integrity and potential damage of structures [
39,
40,
41].
Ultrasonic techniques are widely used in structural health monitoring to assess the condition of materials and structures by utilizing ultrasonic waves. Ultrasonic methods are non-destructive and can provide valuable information about the integrity, thickness, and potential defects within structures [
42,
43,
44].
The choice of monitoring technique depends on factors such as the specific application, accessibility of the structure, required sensitivity, and desired level of monitoring coverage. Developing a comprehensive monitoring strategy tailored to the structure’s characteristics and anticipated failure modes to ensure an effective and reliable assessment of the structure’s health is important. Generally, sensors are bonded to the surface of the structure or embedded within the concrete during construction. In this study, strain gauges were used to monitor the stress–strain behavior of the carbon fabric-reinforced concrete beam.
6. Prediction Model to Determine the Ultimate Load of TRC Beams by ANN
Artificial Neural Networks (ANNs) are extensively used in predictive analysis due to their ability to learn complex patterns and relationships from data. ANNs are versatile and can be applied to various predictive analysis tasks, including regression, classification, time-series forecasting, and pattern recognition. Collecting relevant data is the first step in predictive analysis. Raw data often need preprocessing before being fed into an ANN. Identifying and selecting relevant features are crucial for the performance of an ANN. Too many irrelevant or redundant features may lead to overfitting, while too few features may result in poor predictive performance. Designing the architecture of the neural network involves finding the number of layers, the number of neurons in each layer, the activation functions, and the overall structure of the network. During the training phase, the ANN learns from the historical data to identify patterns and relationships. After training, the ANN’s performance is assessed using validation and testing datasets. Once the ANN demonstrates satisfactory performance, it can be deployed for making predictions on new, real-world data [
46,
47]. In this research, an ANN is used to predict the load-carrying capacity of partially replaced carbon fabric-reinforced concrete beams by taking the steel reinforcement percentage and number of fabric layers as variables. From the experimental investigations carried out, datasets were taken to carry out the predictive analysis.
6.1. Random Forest Regression Algorithm
A versatile machine learning method for predicting numerical values is called random forest regression. The random forest regressor prediction class implements a random forest regression model using the sklearn random forest regressor. This ensemble learning technique works by building a large number of decision trees during training. The mean or average prediction made by each tree is returned for regression tasks. Many decision trees are trained on random subsets of the training data using a technique called bagging. This introduces randomness into the model and reduces overfitting. When splitting nodes during tree construction, only a random subset of features is considered. This further decorrelates the individual trees to improve performance. Combining multiple decision trees produces a more robust and accurate model compared to a single decision tree [
48].
6.1.1. Training of Random Forest Regression Algorithm
Training in random forest regression refers to the process of building a random forest model to predict continuous numeric values (regression) based on input features. Random forest algorithms have three main hyperparameters which need to be set before training. These include node size, the number of trees, and the number of features sampled. From there, the random forest classifier can be used to solve regression. The train_data_points() method generates some dummy input feature data X and target values y to train the model. The RandomForestRegressor.fit() method is used to train the regressor on these data.
6.1.2. Prediction of Random Forest Regression Algorithm
Once the model is trained and validated, it can be used to make predictions on new, unseen data by inputting the features of the new data into the trained model, which will then output the predicted target values. A random forest regression model combines multiple decision trees to create a single model. Each tree in the forest builds from a different subset of the data and makes its own independent prediction. The final prediction for the input is based on the average or weighted average of all the individual trees’ predictions. For making predictions on new data, get_predic_input_value() reads in a CSV file uploaded and returns the input feature data. predict_output() feeds these data into the trained random forest regressor model to generate predictions. The forest’s predictions are averaged to produce the final predicted regression value.
6.1.3. Visualization of Random Forest Regression Algorithm
Visualization involves various techniques used to gain insights into the trained model, understand its behavior, and interpret its results. Visualization techniques are essential for understanding and interpreting the behavior of a random forest regression model, identifying potential issues such as overfitting or underfitting, and gaining insights into the relationships between features and the target variable. create_dataframe() packages the predictions with the original input data into a Pandas DataFrame. Graph() generates a scatter plot of the training data and adds the new prediction point. This allows for visualization of how the prediction relates to the original training data distribution. A scatter plot of random forest regression prediction is shown in
Figure 24.
6.2. Support Vector Machine (SVM) Regression
The SVM prediction class implements a support vector regression model using Sklearn’s SVR.VM regression aims to find the hyperplane (defined by support vectors) that best fits the training data while maximizing the flatness of the mapping. This makes the method robust and less prone to overfitting. SVR maps inputs to a high-dimensional feature space, where a linear regression is carried out, using a kernel trick. Common kernels include radial basis function and polynomial kernels. A tube with radius e is added around the fitted hyperplane to tolerate some prediction errors and prevent massive overfitting. This tube constraint controls model flexibility. Only support vectors that define the hyperplane margins influence the prediction. Data points inside the tube do not.
6.2.1. Training of SVM Regression
Training in support vector machine (SVM) regression involves the process of fitting a model to a dataset in order to learn the relationship between input features and continuous target variables. Training in SVM regression involves finding the optimal hyperplane that best fits the training data while maximizing the margin and minimizing prediction errors. train_data_points() generates dummy input feature data X and target values y. SVR.fit() fits the regression model on this training data. Key hyperparameters like C, epsilon, and kernel parameters control model behavior.
6.2.2. Prediction of SVM Regression
Prediction in SVM regression involves applying the learned model to new data points to estimate the continuous target variable and assessing the performance of these predictions to understand the model’s accuracy and reliability. To make predictions, the values of the features for the new data point are input into the trained SVM regression model. The model then applies the learned decision function to these feature values to estimate the continuous target variable. get_predic_input_value() obtains new input data from the uploaded CSV. predict_output() feeds these data into the trained SVR model to generate predicted values. The position of the new data relative to the tube and support vectors determines the prediction.
6.2.3. Visualization of SVM Regression
Visualization in support vector machine (SVM) regression can be challenging compared to classification tasks, as the output is a continuous variable rather than a discrete class. While visualizing SVM regression models may not be as straightforward as in classification tasks, these techniques can still provide valuable insights into the model’s behavior, performance, and the relationships between input features and the target variable. A plot shows the training data distribution and where the new prediction point falls relative to it. This provides insight into the prediction logic for the given inputs. The gradio interface allows for the interactive testing of different input data CSVs to see the prediction performance. The scatter plot of support vector machine regression is shown in
Figure 25.
6.3. Multilayer Perceptron (MLP) Regression
The MLP Regressor Algorithm implements a multilayer perceptron neural network for regression using sklearn’s MLP Regressor. This feed-forward ANN model maps inputs to appropriate outputs. It is made up of various node layers, such as an input layer, one or more hidden layers, and an output layer. Each node in one layer connects with a certain weight wij to every node in the following layer.
6.3.1. Training of MLP Regression
Training an MLP regression model involves iteratively adjusting the weights and biases of the network to minimize the difference between the predicted and actual target values, ultimately resulting in a model that can accurately predict continuous variables. The fit() method trains the MLP Regressor on the provided input feature data X and target values y. Backpropagation is used to iteratively adjust weights and biases to minimize the loss between the predicted and actual target values.
6.3.2. Prediction of MLP Regression
Prediction involves using a trained neural network model to estimate the continuous target variable (or response variable) for new, unseen data points based on their input features. To make predictions on new data, get_predic_input_value() obtains new input data from the uploaded CSV. The trained MLP Regressor generates predicted values by passing new data through the network architecture. Outputs are generated by calculating activations through the network layer by layer.
6.3.3. Visualization of MLP Regression
Visualizing the relationship between each input feature and the target variable can provide insights into their correlation. A scatter plot shows true targets vs. predictions to visually assess performance. The gradio interface allows for interactive testing to gauge prediction accuracy on varying input data.
Figure 26 shows the scatter plot of multilayer perceptron regression.
6.4. XGBoost Algorithm
The XGBoost class implements the XGBoost library for gradient-boosted decision tree regression. This ensemble technique combines multiple decision tree models. Trees are added sequentially, with each new tree correcting errors from the existing sequence. This boosting process reduces bias and variance for improved performance. Regularization helps prevent overfitting as more trees are added.
6.4.1. Training of XGBoost Algorithm
Training in the XGBoost algorithm involves iteratively building decision trees and optimizing them to minimize a specified objective function. Through gradient boosting and regularization techniques, XGBoost produces an ensemble of decision trees that can make accurate predictions for regression tasks. The fit() method trains an XGB Regressor on the provided X input features and the y target variable. Key hyperparameters like n_estimators control ensemble size and performance.
6.4.2. Prediction of XGBoost Algorithm
For prediction, the values of the features for the new data point are inputted into the trained XGBoost model. The model then uses the ensemble of decision trees to generate a prediction for the target variable based on the input features. get_predict_input_values() obtains new input data from the uploaded CSV. The trained model uses the learned tree sequence to generate predictions for this new data. Predictions are made by the weighted summation of outputs from individual trees.
6.4.3. Visualization of XGBoost Algorithm
Visualization in the XGBoost algorithm primarily revolves around understanding the model’s structure, feature importance, and decision-making process. A scatter plot can visually assess prediction accuracy and can compare actual vs. predicted values. The gradio interface allows for interactively testing predictions on new input CSVs.
Figure 27 shows the XGBoost prediction scatter plot.
6.5. Output
In the prediction model created based on the regression analysis carried out, steel reinforcement percentage and the number of fabric layers are taken as inputs to predict the ultimate load of the TRC beams.
Figure 28 shows the prediction model created by ANN.