ArticlePDF Available

Computational Complexity and Its Influence on Predictive Capabilities of Machine Learning Models for Concrete Mix Design

MDPI
Materials
Authors:

Abstract and Figures

The design of concrete mixtures is crucial in concrete technology, aiming to produce concrete that meets specific quality and performance criteria. Modern standards require not only strength but also eco-friendliness and production efficiency. Based on the Three Equation Method, conventional mix design methods involve analytical and laboratory procedures but are insufficient for contemporary concrete technology, leading to overengineering and difficulty predicting concrete properties. Machine learning-based methods offer a solution, as they have proven effective in predicting concrete compressive strength for concrete mix design. This paper scrutinises the association between the computational complexity of machine learning models and their proficiency in predicting the compressive strength of concrete. This study evaluates five deep neural network models of varying computational complexity in three series. Each model is trained and tested in three series with a vast database of concrete mix recipes and associated destructive tests. The findings suggest a positive correlation between increased computational complexity and the model’s predictive ability. This correlation is evidenced by an increment in the coefficient of determination (R²) and a decrease in error metrics (mean squared error, Minkowski error, normalized squared error, root mean squared error, and sum squared error) as the complexity of the model increases. The research findings provide valuable insights for increasing the performance of concrete technical feature prediction models while acknowledging this study’s limitations and suggesting potential future research directions. This research paves the way for further refinement of AI-driven methods in concrete mix design, enhancing the efficiency and precision of the concrete mix design process.
This content is subject to copyright.
Citation: Ziolkowski, P.
Computational Complexity and Its
Influence on Predictive Capabilities
of Machine Learning Models for
Concrete Mix Design. Materials 2023,
16, 5956. https://doi.org/10.3390/
ma16175956
Academic Editors: Krzysztof
Schabowicz and Eddie Koenders
Received: 18 July 2023
Revised: 24 August 2023
Accepted: 25 August 2023
Published: 30 August 2023
Copyright: © 2023 by the author.
Licensee MDPI, Basel, Switzerland.
This article is an open access article
distributed under the terms and
conditions of the Creative Commons
Attribution (CC BY) license (https://
creativecommons.org/licenses/by/
4.0/).
materials
Article
Computational Complexity and Its Influence on Predictive
Capabilities of Machine Learning Models for Concrete
Mix Design
Patryk Ziolkowski
Faculty of Civil and Environmental Engineering, Gdansk University of Technology, Gabriela Narutowicza 11/12,
80-233 Gdansk, Poland; patziolk@pg.edu.pl; Tel.: +48-58-347-2385
Abstract:
The design of concrete mixtures is crucial in concrete technology, aiming to produce
concrete that meets specific quality and performance criteria. Modern standards require not only
strength but also eco-friendliness and production efficiency. Based on the Three Equation Method,
conventional mix design methods involve analytical and laboratory procedures but are insufficient
for contemporary concrete technology, leading to overengineering and difficulty predicting concrete
properties. Machine learning-based methods offer a solution, as they have proven effective in pre-
dicting concrete compressive strength for concrete mix design. This paper scrutinises the association
between the computational complexity of machine learning models and their proficiency in predict-
ing the compressive strength of concrete. This study evaluates five deep neural network models of
varying computational complexity in three series. Each model is trained and tested in three series
with a vast database of concrete mix recipes and associated destructive tests. The findings suggest
a positive correlation between increased computational complexity and the model’s predictive ability.
This correlation is evidenced by an increment in the coefficient of determination (R
2
) and a decrease
in error metrics (mean squared error, Minkowski error, normalized squared error, root mean squared
error, and sum squared error) as the complexity of the model increases. The research findings provide
valuable insights for increasing the performance of concrete technical feature prediction models
while acknowledging this study’s limitations and suggesting potential future research directions.
This research paves the way for further refinement of AI-driven methods in concrete mix design,
enhancing the efficiency and precision of the concrete mix design process.
Keywords:
applied machine learning; buildings; cement; concrete mix design; concrete strength
prediction; concrete; construction industry; data mining; green building; innovation; sustainability;
sustainable development; sustainable
1. Introduction
The field of concrete mix design has seen significant advancements in recent years with
the integration of machine learning techniques. The ability of these models to predict the
compressive strength of concrete has the potential to revolutionise the construction industry
by providing more efficient and accurate methods for mix design. However, the prediction
accuracy of these models depends on many factors, including the computational complexity
of the model itself. This study aims to investigate the influence of computational complexity
on the prediction accuracy of concrete compressive strength in machine learning models for
concrete mix design. The findings of this research will provide insights into the trade-off
between model accuracy and computational efficiency and will guide the development of
more effective machine learning models for concrete mix design in the future. The concrete
mix composition consists of cement, water, a combination of fine and coarse aggregates,
and supplementary materials referred to as additives and admixtures, where additives are
incorporated during the cement manufacturing stage, whereas admixtures are introduced
Materials 2023,16, 5956. https://doi.org/10.3390/ma16175956 https://www.mdpi.com/journal/materials
Materials 2023,16, 5956 2 of 36
during concrete mix preparation. These substances are formulated to enhance the chemical
properties and performance of concrete, specifically with regard to compressive strength,
durability, and workability. There are various types of additives and admixtures [
1
4
],
including accelerators, substances that improve fresh concrete properties [
5
], materials
that enhance durability, fibres that reinforce concrete [
6
], set-retarding admixtures, and
water-reducing agents.
Properly designing a concrete mixture is a crucial aspect of the construction process,
with multiple factors to consider. It should be designed with economy in mind, ensuring
that the desired properties can be achieved using the most cost-effective raw materials. The
mixture must also be optimised for the specific technology used in the construction process,
considering elements such as workability and setting speed. Environmental conditions,
such as temperature [
7
,
8
], precipitation, distance away from the construction area, as well
as the volume of traffic, must also be considered when designing the concrete mix. The
final composition of the mix is determined by construction specifications, such as the
desired compressive strength or resistance to environmental elements such as chloride
ingression and, increasingly, ecological considerations, such as low emissivity. To address
environmental concerns, various solutions exist to reduce concrete carbonation, including
admixtures of graphene nanoparticles. Designing a concrete mix involves selecting the
proper proportions of primary and secondary components to achieve the desired properties.
Once the mix is prepared, it is conveyed to the construction location and poured into the
formwork, where it undergoes the progression of hardening and increasing strength. The
hardening process is initiated by the cement’s hydration, which involves a heat-releasing
chemical reaction that occurs upon contact between cement and water [
9
]. The reaction
initiates the formation of various components, such as tobermorite gel [
10
], hydroxide, as
well as additional components, which improve the bonding between coarse and fine aggre-
gates. During this procedure, the hydration products steadily accumulate on the cement
grains and replace the water in the mixture. The ultimate hydration stage occurs when all
the water molecules are fully integrated, or no unreacted cement is left in the mixture. After
the hydration process begins, hardened concrete acquires some of its compressive strength
within a few days, and most of its compressive strength is attained after roughly 28 days
(although some types of concrete may take longer to reach their full strength). The quan-
tity of water required to hydrate the cement completely ranges from 20% to 25% of its
weight, excluding water trapped in pores. However, specific models suggest that 42% of
the cement’s weight is needed for proper hydration [
11
]. The design methods for concrete
mixtures currently employed in engineering practice have been derived from solutions
developed more than a decade ago and rely on estimating the bending strength of the con-
crete mortar. Implementing these techniques in practice might be a tedious and inefficient
process that does not consider the intricate chemical composition and variability of modern
concrete mixes. The current challenges in the field at hand necessitate novel technological
solutions. Machine learning-based methods could offer a promising avenue, as they have
demonstrated varying degrees of success in predicting concrete compressive strength.
Machine learning, a prominent subfield of artificial intelligence, has garnered signif-
icant attention in recent years due to its vast applications and transformative potential
across various domains. The fundamental concept behind machine learning involves em-
powering computers to acquire knowledge from data, recognize patterns, and arrive at
informed choices while minimizing the need for extensive human involvement. By utilis-
ing algorithms and statistical models, machine learning systems can adapt and improve
their performance over time, making them valuable tools for a multitude of tasks, rang-
ing from natural language processing [
12
] and image recognition [
13
] to real estate value
forecasting
[
14
17
] and medical purposes [
18
]. The foundation of machine learning lies in
its ability to extract knowledge from data, which is achieved by employing various learning
paradigms. These paradigms include supervised, unsupervised, and reinforcement learn-
ing, each catering to different problem domains. Supervised learning, the most common
approach, involves training a model using labelled data with known outcomes, while
Materials 2023,16, 5956 3 of 36
unsupervised learning deals with discovering hidden structures in unlabelled data [
19
].
Reinforcement learning, on the other hand, focuses on learning through trial and error,
with a model receiving feedback in the form of rewards or penalties [
20
,
21
]. A plethora of
algorithms have been developed for each learning paradigm, and the choice of the algo-
rithm largely depends on the specific problem and the available data. Popular algorithms
include linear regression, decision trees, support vector machines, neural networks, and
clustering algorithms. These algorithms often involve tuning various parameters, known
as hyperparameters, to optimise the model’s performance. Machine learning has proven
particularly effective in addressing complex problems with high-dimensional data. This has
been facilitated by the advent of deep learning, a subset of machine learning that relies on
artificial neural networks with multiple layers. These networks, drawing inspiration from
the organization and operation of the human brain, excel in acquiring complex patterns
and representations from extensive datasets, making them suitable for a diverse array of
uses, including natural language processing and computer vision. The success of machine
learning in diverse fields has prompted researchers to investigate its potential for predicting
and optimising properties of materials, such as concrete.
In this paper, the influence of computational complexity on the performance of ma-
chine learning models used for predicting the compressive strength of concrete is explored.
This research assessed three sets of five deep neural network models (MLM1, MLM2,
MLM3, MLM4, MLM5), each with differing levels of computational complexity. Through
an examination of various machine learning models, the goal is to identify the trade-offs
between accuracy and computational efficiency. This could provide valuable insights into
the development of robust and cost-effective models for concrete mix design.
2. Concrete Mix Design and Machine Learning
2.1. Prediction of Concrete Technical Properties in Concrete Mix Design
Designing an optimal concrete mix is a multifaceted challenge requiring a compre-
hensive understanding of concrete technology and significant practical experience. The
primary objective of the design process is to determine suitable material compositions
to achieve the desired properties in fresh concrete during transportation and placement
and in hardened concrete. Distinct mechanical properties are anticipated at each stage of
the concrete fabrication process. Various characteristics influence concrete performance,
including plasticity, durability, compressive strength, and modulus of elasticity. The signifi-
cance of these properties may vary at different stages, for instance, adequate compressive
strength is crucial for the designed ultimate limit state, whereas sufficient durability is
critical in aggressive environments. Designing a mix with inappropriate specifications can
result in severe consequences. Therefore, fearing noncompliance with the necessary criteria,
concrete mix manufacturers often intentionally exceed the designed parameters [22].
Global corporate engineering practices related to concrete technology exhibit con-
siderable variation and notable commonalities. Within the European Union, the primary
standard governing concrete technology issues is EN 206 Concrete: Specification, Per-
formance, Production, and Conformity [
23
], while EN 1992-1-1: Eurocode 2: Design of
Concrete Structures [
24
] provides guidelines for the design of concrete structures. Both stan-
dards have national equivalents and appendices, such as DIN EN 206 [
25
] in Germany and
PN-EN 206 + A1: 2016-12 in Poland. Member states of the European Union employ diverse
methods for designing concrete mixes. In Poland, the Bukowski, Eyman, Klaus, Kopycinski,
and Paszkowski methods are predominantly used alongside the double-coating method.
Conversely, the Bolomey, Fuller, and 0.45 power gradation chart methods are more preva-
lent in the United States. Most of these approaches are derived from the “three equations
method” representing a combined experimental-analytical strategy for concrete mix design.
The approach that combines experimental and analytical methods involves determin-
ing the required quantity of ingredients through analytical calculations and confirming its
accuracy through destructive lab testing. This technique enables the researchers to establish
the proportions of cement, water, and aggregate by weight for a given volume, employing
Materials 2023,16, 5956 4 of 36
three formulas related to consistency (1), strength (2), and impermeability (3). The con-
sistency formula (1) is integrated into the water requirement equation, which assists in
identifying the optimal consistency. The demand for water in cement and aggregate relies
on factors such as particle size, form, texture, ratio within the mixture, and the desired
consistency of the concrete blend. To account for the water needs of concrete additives and
admixtures, they are added to the aggregate or cement based on the particle dimensions.
Factors such as grain size, shape, surface roughness, composition proportion, and the con-
crete mix’s required consistency influence the cement–water and aggregate–water demand
indices created by Bolomey and Stern [26].
W=C·wc+K·wk, (1)
fcm =A C
W+pa, (2)
fcm =A1,2C
W±a, (3)
C
ρc
+K
ρk
+W=1000, (4)
In Equation (1),
W
denotes the volume of water (in litres) present in one cubic meter
of concrete. The weight of cement in one cubic meter of concrete is represented by C
(in kilograms). The cement–water demand index
wc
, indicates the volume of water (in
litres per kilogram) required to be combined with one kilogram of a specific cement class
in one decimetre cubed. K represents aggregate weight in one cubic meter of concrete,
measured in kilograms. Lastly, the aggregate–water demand index
wk
, signifies the volume
of water (in litres per kilogram) needed to be added to one kilogram of a certain dry
aggregate fraction to achieve the desired consistency. The subsequent formula, called
the concrete compressive strength formula, exists in two variations: Bolomey and Feret.
This formula illustrates the connection between concrete compressive strength and factors
such as water–cement ratio, cement grade, and aggregate grade. The Feret and Bolomey
versions of the formula are represented by Equations (2) and (3), respectively. In these
equations,
fcm
denotes the average concrete compressive strength in MPa, while
A
,
A1
,
and
A2
are coefficients contingent upon the aggregate’s type and strength class and the
cement’s strength class.
A1
is utilised when the cement-to-water ratio (C/W) is less than
2.5, and
A2
is employed when the ratio exceeds 2.5. In the equations, C signifies the cement
weight in 1 cubic meter of concrete (measured in kg), Wrepresents the water quantity in
1 cubic meter of concrete (measured in L), pis the air content in one cubic meter of concrete
(measured in dm
3
), and
a
is a numeric value that depends on cement and aggregate quality,
typically considered constant equal to 0.5. The
a
value is positive for water–cement ratios
greater than or equal to 2.5, while it is negative for ratios below 2.5. The Feret equation
is applicable when the aggregate strength is inferior to the grout strength, specifically
in the case of porous concrete. The water-tightness equation, designated as Equation (4),
asserts that the total volume of individual components in a concrete mix is equivalent
to the mix’s overall volume. In this equation,
W
signifies the water content in litres per
cubic meter (m
3
) of concrete, while C stands for the cement weight in kilograms per cubic
meter (m
3
) of concrete. Additionally,
ρc
represents the density of cement in kg/dm
3
, K
denotes the weight of cement in kg per cubic meter (m
3
) of concrete, and
ρk
refers to the
aggregate density in kg/dm
3
. Using the equations mentioned above, one can determine
the quantitative composition of a concrete mix, which consists of the quantities of cement,
water, and aggregate in a cubic meter (m
3
) of the mixture. The three equations method
does have specific boundary constraints, for instance, the porosity of the concrete mix must
not surpass 0.002 of the mix volume without air-entraining additives or 0.008 of the mix
volume when incorporating air-entraining additives.
Materials 2023,16, 5956 5 of 36
The process of creating a concrete mix design incorporates several steps such as formu-
lating initial suppositions, defining the necessary characteristics of both the fresh and cured
concrete, selecting and assessing the components of the mix, crafting the blend, testing
its properties in a lab setting, and finally devising a workable formula. In formulating
initial suppositions, key considerations include the concrete’s intended application and the
specific traits of the new structure. These factors include location, degree of reinforcement,
and the structural cross-section’s geometric properties. The primary attributes of interest
for fresh concrete are its bulk density, consistency, and air content. As for the hardened
concrete, we look at its frost and fire resistance and the grade of its compressive strength. It
is essential to scrutinise the technological process and evaluate the conditions under which
the concrete matures and the method used for compacting the fresh mix. The concrete expo-
sure class is a significant parameter, defining the level and nature of environmental stress
the material can withstand. Further specifications such as the concrete’s impermeability
need to be established. Parameters such as maximum aggregate size and mix workability
also need determination. The components of the concrete mix, including the appropriate
type of cement, water, and aggregate quality, should be selected and appraised as per
relevant standards. Following the mix design and lab tests, the final stage involves creating
a functional formula for one cubic meter of the concrete mix. One should also anticipate po-
tential recipe modifications due to the aggregate’s moisture content and adjust it according
to specific circumstances, such as the transportation vehicle’s capacity [27,28].
2.2. Machine Learning in Prediction of Concrete Technical Properties
Machine learning has permeated numerous scientific domains, showcasing its versatil-
ity and potency. It has a particularly large number of applications in the field of structural
and material engineering. Machine learning has found use within this branch of science
in areas such as structural health monitoring, crack detection, life-cycle cost assessment,
or prediction of diffusivity [
29
]. Particular interest has been shown in predicting technical
properties [30], most of which were devoted to concrete compressive strength prediction.
The complexities of predicting concrete strength through machine learning were first
articulated by Yeh et al. [
31
] in 1998. They experimented with seven input variables using
artificial neural networks (ANN) and linear regression to forecast the strength of high-
strength concrete. While their model was trained on a vast array of concrete samples, these
were not scrutinised for content. Their study considered concrete samples in the maturing
phase, including those as young as three days old, which may have led to skewed results.
In 2003, this subject was further refined by Seung-Chang Lee [
32
], who employed
a unique, modular network architecture consisting of five ANNs. Each ANN represented
concrete at various stages of maturation up to its maximum strength. Lee used the parame-
ter condensation technique to ascertain the number of neurons in the input layer. Despite
claiming his condensation and weighing techniques as beneficial for optimal network
performance, the practicality of his ANN model, which illustrates the maturation process
from pouring to full strength, is questionable. From an engineering perspective, attention
should be devoted to concrete that has achieved full or near-full strength.
In 2005, Hola, J. and Schabowicz, K. [
33
,
34
] introduced a novel nondestructive ap-
proach for assessing concrete strength. Rather than relying on the concrete mix composition,
they trained their ANN model on data obtained from nondestructive concrete testing tools.
Their database incorporated ultrasonic wave velocity, reflection number, hardness, pull-out
strength, concrete age, and bulk density. To evaluate their lab results, they experimented on
concrete compressive strength samples with a 28-day strength ranging from 24 to 105 MPa.
Using the Levenberg–Marquardt training method, they developed the ANN with eight hid-
den neurons within one layer. The authors posited that the average compressive strength
comparison between the ANN and nondestructive tests was comparable.
In 2006, a neural-expert system for predicting the compressive strength of high-
performance concrete was proposed by Gupta et al. [
35
]. They focused on training the
algorithm through example inferences, employing a multilayer ANN trained with gen-
Materials 2023,16, 5956 6 of 36
eralised backpropagation for interval training patterns. They also used input variables
from unrelated metrics, such as curing time, that were unrelated to the recipe. However,
these strategies may lead to algorithm training based on insignificant patterns and unclear
results. The use of a neural-expert system for concrete compressive strength prediction was
also explored by Dac-Khuong Bui et al. [
36
] with a focus on the practical application of
this method.
The advent of deep learning in this field was introduced by Fangming Deng et al. [
37
]
in 2018. They prepared a database of recycled concrete samples for algorithm training.
They chose not to train the algorithm on the concrete mix composition with direct amounts
of individual components, but they called deep features on several ratios. This approach
was emulated in the current study with the inclusion of feature scaling. Deng and his team
used Softmax regression to identify a suitable prediction model. They claimed that deep
learning compared to ANN, provided better generalisation capabilities, superior efficiency,
and greater precision. However, these claims were not definitive and warranted further
research. Given that convolution neural networks are computationally costly, the authors
used a limited database of 74 samples compared to 741 in the current study. This limited
sample size might lead to underfitting, implying a model that does not fully capture the
modelled phenomenon. A comparable level of accuracy between artificial neural networks
and deep neural networks was reported by Hosein Naderpour et al. [38] in 2018.
Ziolkowski, P. et al. [
39
] introduced an algorithm in 2019 that assists in designing
a concrete mix by predicting the strength based on the mix composition. While this al-
gorithm accurately predicted the strength of the concrete mix, it underperformed in the
high-strength spectrum of 40 MPa and beyond and was insufficient in predicting the prop-
erties of mixtures with additives and admixtures. Furthermore, it neglected other essential
parameters that contribute to concrete’s performance, such as durability, which is essential
for maintaining structural service quality over time [40].
In a publication from 2020, Nunez, I. and his team [
41
] shared insights on using
machine learning to accurately forecast the compressive strength of recycled aggregate
concrete, consequently refining the concrete mix design process. The researchers recognised
the critical importance of an effective optimisation method for concrete mix design in light of
the inherent variability and the absence of reliable compressive strength prediction formulae
for recycled aggregate concrete. Three innovative machine learning models were developed
in their study, specifically the Gaussian process model, the recurrent neural network
model, and the gradient-boosted regression tree model. Based on their findings, they
reported superior predictive outcomes, particularly with the gradient-boosted regression
trees model.
Another noteworthy contribution from the same year is a study by Marani, A. and
his colleagues [
42
], who explored the use of machine learning to forecast the compressive
strength of ultra-high-performance concrete. Their algorithm was trained on a comprehen-
sive dataset of 810 samples from freely accessible resources, encompassing 15 variables as
input data. Rather uniquely, they capitalised on their dataset to generate 6513 records, a sub-
stantial number of synthetic data samples, using tabular generative adversarial networks.
The wealth of data facilitated a more robust training of their machine learning model.
Upon evaluation, the model trained with synthetic data yielded exceptional predictive
performance when assessed with the primary dataset.
In 2021 Ziolkowski, P. et al. [
43
] introduced a new adaptive machine learning method
that more precisely estimates the compressive strength of concrete based on the composition
of its primary ingredients. Unlike previous models that had mixed success in forecasting
concrete strength and struggled to encapsulate the variability inherent in current concrete
mixes, this method incorporated two observations for each concrete batch in the model. The
authors built this machine learning model using a deep neural network architecture and
trained it on a comprehensive database of concrete recipes before translating it into a math-
ematical formula. The algorithm was tested on four concrete mix recipes calculated using
Materials 2023,16, 5956 7 of 36
contemporary design methods such as Bolomey and Fuller, with the findings revealing that
this new algorithm outperformed nonadaptive models trained on the same dataset.
Adil, M. et al. [
44
] investigated the effect of the number of neurons and layers in
ANN for generalised concrete mix design. They developed an ANN with 17 inputs and
five outputs related to the concrete mix’s composition and properties. The authors proposed
optimising the network with one or two hidden layers. It represented a significant departure
from previous work, where concrete’s technical parameters were predicted based on the
composition ratio.
Feng, W., in their paper [
45
] explores the mechanical characteristics of rubber-modified
recycled aggregate concrete (RRAC). The authors utilised machine learning (ML) models to
predict these thermomechanical properties of RRAC, employing a unique algorithm called
the beetle antennae search (BAS) to tune the hyperparameters of these models. Four ML
models were tested: random forest, logistic regression, multiple linear regression, and back-
propagation neural network (BPNN). Among them, BPNN yielded the most accurate and
reliable UCS and peak strain predictions, suggesting that ML models, particularly BPNN,
can serve as robust tools for predicting the properties of sustainable construction materials
such as RRAC. This study highlights the potential of RRAC in sustainable construction and
the effective use of ML models in predicting its performance.
Tavares, C. et al., in their two-part study [
46
,
47
], proposed an innovative method that
utilises machine learning (ML) for the optimised mixture design of ultra-high-performance
concrete (UHPC). This methodology presents an attractive alternative to resource-consuming
experimental runs by employing orthogonal arrays for data collection, which could enable
ML design optimisation. The researchers used an ensemble of ML techniques, specifi-
cally random forest and k-nearest neighbours, to create performance density diagrams
(PDDs). These diagrams serve as an intuitive tool to demonstrate the trade-offs between
mix proportions and mechanical performance of UHPC, providing practical assistance to
designers in the construction industry. Their research has shown promising results, where
the PDDs effectively predicted the behaviour of most mixtures in the test set. This method
facilitated the design of a UHPC mixture averaging 155 MPa at age 56 days, maintaining
the fine-aggregate-to-cementitious-material ratio above the unit. It represents a substantial
advancement in developing mix design tools for UHPC, leading to cost and eco-efficiency
improvements. Notably, this methodology was further extended in the second part of
their study to allow simultaneous evaluation of performance, cost, and carbon footprint.
This approach lays a foundation for the broader adoption of ML techniques in sustainable
construction and the development of mix designs for UHPC.
Endzhievskaya, I.G. et al. [
48
] presented a study on road concrete’s physical and
mechanical characteristics. The authors employed machine learning techniques, specifically
a random forest and decision trees. These methods were advantageous due to their ease
of use, minimum hyperparameters for tuning, and the ability to predict with low errors.
Their findings indicate that components such as air-entraining additives and specific sizes
of crushed stone contribute significantly to improving compressive and bending strengths.
Machine learning’s predictive accuracy demonstrated its potential in optimising road
concrete mixtures, enhancing road surfaces’ quality and service life.
The study presented by Taffese, W.Z. and Espinosa-Leal, L. [
49
] stands out in concrete
properties prediction. Their research leverages machine learning, specifically decision tree-
based ensemble methods, to develop multitarget models predicting compressive strength
and nonsteady-state chloride migration coefficients (D
nssm
) of concrete. This work’s novelty
lies in developing a single model that simultaneously predicts two crucial concrete prop-
erties, compressive strength and D
nssm
. The gradient boosting model demonstrated the
most impressive prediction accuracy, yielding the mean absolute error (MAE) of 6.683 and
1.363, mean squared error (MSE) of 83.369 and 3.712, and root mean squared error (RMSE)
of 9.131 and 1.927 for compressive strength and D
nssm
, respectively. The authors stress the
necessity to expand the model with comprehensive datasets encompassing a wider range
of concrete properties to improve its versatility.
Materials 2023,16, 5956 8 of 36
3. Materials and Methods
3.1. Essentials
Machine learning models can be used to predict concrete’s technical properties based
on the mix composition. This study tries to determine the impact of increasing the computa-
tional complexity of the artificial neural network on the model’s performance. The quantity
of layers is one of the deep neural network (DNN)’s [
50
,
51
] features that represents a criti-
cal determinant of model complexity and its inherent capability to discern and replicate
complex patterns embedded in the data. This aspect forms the basis of the term “deep”
within deep learning, where an increased number of layers, representing greater depth,
facilitates the modelling of progressively intricate and abstract features. Each layer within
a DNN can be conceptualised as performing successive transformations of the raw input
into higher-level, abstract features. For instance, in applying convolutional neural networks
(CNNs) [
52
54
], commonly used for image recognition tasks, initial layers may decipher
basic, low-level features such as edges and colours. As the depth of the network increases,
subsequent layers amalgamate these rudimentary features to detect more abstract patterns,
encompassing shapes and, ultimately, entire objects or scenes.
However, while an increased depth can enable a model to learn more complex rep-
resentations, it poses new challenges. The augmentation in the number of layers directly
expands the model’s parameters, thereby escalating the risk of overfitting. Overfitting
manifests when a model excessively adapts to the training data, compromising its ability to
generalise to unseen data effectively [
55
57
]. This becomes particularly problematic when
the volume of available training data is limited compared to the complexity of the model.
Furthermore, training highly deep networks introduces additional technical difficulties.
One notable issue is vanishing or exploding gradients, which can decelerate training or
result in suboptimal solutions [
58
,
59
]. Techniques such as parameter initialisation, batch
normalisation, and incorporating residual connections have been suggested to alleviate
these concerns. Consequently, the selection of an optimal layer quantity necessitates a deli-
cate balancing act, taking into account the trade-off between a network’s capacity to model
intricate patterns (which tends to increase with depth) and the potential challenges associ-
ated with overfitting and training difficulties. Strategies such as early stopping, dropout,
and using validation sets can help to manage the risks inherent to increased depth [
60
62
].
Despite these challenges, the capacity to train deep networks remains a pivotal factor
propelling recent advancements in artificial intelligence and machine learning.
The analysis adopted a classical approach involving the construction of a model that
will estimate the concrete compressive strength determined by the quantitative concrete
mix composition. The prepared analysis used a database from previous studies [
39
,
43
],
which contains several hundred records of concrete recipes, along with corresponding
destructive compressive strength tests on normalised samples in the laboratory. The number
of records has been increased by using a dedicated AI model to generate reliable synthetic
data [
63
65
]. The database used contains recipes for the concrete mix composition, which
were designed as intended for incorporation into various concrete elements. At the same
time, differentiation of dimensions, functions, and purpose is assumed here. Some recipes
contained admixtures for various purposes, such as workability improvers, plasticisers or
setting retarders. It is taken for granted that the concrete production process employed
met the necessary quality standards. However, due to varying design requirements and
the use of different admixtures, some differences between formulations may be difficult
to quantify. Therefore, the procedure of removing univariate outliers as multiples of the
standard deviation was used, described in detail later in the paper. Individual components
of concrete mixes and the water–cement ratio were assigned input variables, while the
concrete compressive strength was treated as the output variable. The presented study,
along with many studies in the literature, focuses on predicting one of the main technical
properties of concrete, which is concrete compressive strength. However, it should be noted
that many other technical properties affect the final behaviour of concrete, especially at
various stages of the technological production process. The quality of this process is also
Materials 2023,16, 5956 9 of 36
essential, which is influenced by factors such as the curing process [
66
,
67
] or the concrete
pouring temperature [
68
]. Figure 1shows a flowchart illustrating the research procedures
outlined in this investigation.
Materials2023,14,xFORPEERREVIEW9of38
recipescontainedadmixturesforvariouspurposes,suchasworkabilityimprovers,plas-
ticisersorseingretarders.Itistakenforgrantedthattheconcreteproductionprocess
employedmetthenecessaryqualitystandards.However,duetovaryingdesignrequire-
mentsandtheuseofdierentadmixtures,somedierencesbetweenformulationsmay
bediculttoquantify.Therefore,theprocedureofremovingunivariateoutliersasmul-
tiplesofthestandarddeviationwasused,describedindetaillaterinthepaper.Individual
componentsofconcretemixesandthewatercementratiowereassignedinputvariables,
whiletheconcretecompressivestrengthwastreatedastheoutputvariable.Thepresented
study,alongwithmanystudiesintheliterature,focusesonpredictingoneofthemain
technicalpropertiesofconcrete,whichisconcretecompressivestrength.However,it
shouldbenotedthatmanyothertechnicalpropertiesaectthenalbehaviourofconcrete,
especiallyatvariousstagesofthetechnologicalproductionprocess.Thequalityofthis
processisalsoessential,whichisinuencedbyfactorssuchasthecuringprocess[66,67]
ortheconcretepouringtemperature[68].Figure1showsaowchartillustratingthere-
searchproceduresoutlinedinthisinvestigation.
Figure1.Flowchartshowingtheproceduresdescribedinthisstudy.
3.2.DataProcessing
Thedatabaseusedinthisresearchcontains6187records,generatedusingadedicated
AImodelfromtheoriginaldatabasethatcontained741recordsofconcreterecipes[39,43],
alongwithcorrespondingcompressivestrengthtestsconductedunderlaboratorycondi-
tionsonstandardsamplesaccordingtoPN-EN:206[23].Thissethassixvariables,asfol-
lows:fck—concretecompressivestrength(MPa),C—cement(kg/m3),W—water–cement
ratio(-),FAneaggregate(kg/m3),CA—coarseaggregate(kg/m3).Thesesyntheticdata
donotcreatenewknowledgebuthelpstoachievebeerrobustnessoftheAImodel.The
parametersutilizedhavebeenshowcasedinTable1.Afundamentalstatisticalanalysis
waspreparedforeachvariableintheanalyseddataset.Table2showseachvariablesmax-
imal,minimal,mean,median,anddominantvalues.
Figure 1. Flowchart showing the procedures described in this study.
3.2. Data Processing
The database used in this research contains 6187 records, generated using a dedicated
AI model from the original database that contained 741 records of concrete recipes
[39,43]
,
along with corresponding compressive strength tests conducted under laboratory con-
ditions on standard samples according to PN-EN: 206 [
23
]. This set has six variables, as
follows: f
ck
—concrete compressive strength (MPa), C—cement (kg/m
3
), W—water–cement
ratio (-), FA—fine aggregate (kg/m
3
), CA—coarse aggregate (kg/m
3
). These synthetic data
do not create new knowledge but helps to achieve better robustness of the AI model. The
parameters utilized have been showcased in Table 1. A fundamental statistical analysis was
prepared for each variable in the analysed dataset. Table 2shows each variable’s maximal,
minimal, mean, median, and dominant values.
Table 1. The parameters utilized within the dataset.
Parameter
Concrete
Compressive
Strength
Cement Water–Cement
Ratio Fine Aggregate Coarse Aggregate
Type Target Input Input Input Input
Description
The 28-day
compressive
strength of concrete
that is considered to
have most of its
strength (MPa).
Content of cement
added to the
mixture (kg/m3).
Water-to-cement
ratio (-).
Content of fine
aggregate added to
the mixture
(kg/m3).
Content of coarse
aggregate added to
the mixture
(kg/m3).
Materials 2023,16, 5956 10 of 36
Table 2. Ranges of input variables for the database.
Input Variable Minimum Maximum Mean Median Dominant
Cement 87.00 kg/m3540.00 kg/m3322.15 kg/m3312.45 kg/m3380.00 kg/m3
Water–cement ratio 0.30 0.80 0.58 0.58 0.58
Fine aggregate 472.00 kg/m3995.60 kg/m3767.96 kg/m3774.00 kg/m3594.00 kg/m3
Coarse aggregate 687.80 kg/m31198.00 kg/m3969.92 kg/m3963.00 kg/m3932.00 kg/m3
The tabular long short-term memory (TLSTM) model [
69
,
70
] was used to generate
credible synthetic data. The experiments were carried out to validate the excellence of
the produced data, specifically a principal components analysis (PCA) [
71
,
72
] and Jensen–
Shannon divergence (JSD) [
73
76
]. In order to verify the statistical integrity of deeper,
multifield distributions and correlations, a comparative analysis was performed using PCA
computed first on the original data, then on the synthetic data. The basis of PCA lies in
capturing the essential shape of all the features in a few key features, referred to as the
principal components. Consider a dataset with just two columns, as graphed below for
illustrative purposes. PCA can be visualised as an exercise in fitting an ellipsoid to the
data, where the axis of the ellipsoid, signifying the directions of maximum variability in the
data, represents the principal components. In a more complex multidimensional scenario,
the objective of PCA becomes analogous to rotating an object in hand to achieve a view of
maximal width, which is then determined to be the first principal component. Subsequent
rotations, while maintaining horizontal steadiness, aim to achieve the view of maximal
height, thus determining the second principal component. The approach is structured
around identifying the axis with maximum variability, always maintaining perpendicularity
to the previously chosen axis. Consequently, the newly created dimensions encapsulate the
essence of the fundamental shape of the data. The quality of synthetic data can be assessed
by evaluating the distributional distance between the principal components in the original
data and those in the synthetic data. The proximity of the principal components directly
influences the quality of synthetic data, with closer principal components yielding better
quality. Given the ubiquity of PCA in machine learning for dimensionality reduction and
visualisation, this score provides an immediate assessment of the utility of the obtained
synthetic data for machine learning applications. The approach hence aims to measure
the statistical integrity of the synthetic data by evaluating its conformance to the structure
encapsulated in the principal components of the original data. The results of PCA are
presented in Figure 2.
Materials2023,14,xFORPEERREVIEW11of38
(A)(B)
Figure2.Principalcomponentanalysis:trainingdata(A),syntheticdata(B).Trainingdataarethe
datausedtogeneratesyntheticdata.
JSDrepresentsthedegreeofresemblancebetweentheelddistributionsoftheorig-
inalandsyntheticdata.Itisamethodcommonlyappliedforcomparingtwodistributions.
TheaverageJensen–Shannondivergencevalueacrossalleldsinverselycorrelateswith
thedataquality,withlowervalueindicatinghigherquality.Avisualcomparisonofthe
originalandsyntheticelddistributionsisfacilitatedbypresentingbarchartsinFigure3.
(A)(B)
(C)(D)
Figure 2.
Principal component analysis: training data (
A
), synthetic data (
B
). Training data are the
data used to generate synthetic data.
JSD represents the degree of resemblance between the field distributions of the original
and synthetic data. It is a method commonly applied for comparing two distributions. The
average Jensen–Shannon divergence value across all fields inversely correlates with the
Materials 2023,16, 5956 11 of 36
data quality, with lower value indicating higher quality. A visual comparison of the original
and synthetic field distributions is facilitated by presenting bar charts in Figure 3.
Materials2023,14,xFORPEERREVIEW11of38
(A)(B)
Figure2.Principalcomponentanalysis:trainingdata(A),syntheticdata(B).Trainingdataarethe
datausedtogeneratesyntheticdata.
JSDrepresentsthedegreeofresemblancebetweentheelddistributionsoftheorig-
inalandsyntheticdata.Itisamethodcommonlyappliedforcomparingtwodistributions.
TheaverageJensen–Shannondivergencevalueacrossalleldsinverselycorrelateswith
thedataquality,withlowervalueindicatinghigherquality.Avisualcomparisonofthe
originalandsyntheticelddistributionsisfacilitatedbypresentingbarchartsinFigure3.
(A)(B)
(C)(D)
Materials2023,14,xFORPEERREVIEW12of38
(E)
Figure3.Fielddistributioncomparisons:concretecompressivestrength(A),cement(B),waterce-
mentratio(C),neaggregate(D),coarseaggregate(E).Purplebarscorrespondtotrainingdata.
Greenbarscorrespondtosyntheticdata.Verticalaxesarepercentages.
Thecompressivestrengthofconcretewasassessedinalaboratoryseingutilising
standardisedsamplesundertheEN-206-01standard[77].Thesamplesinquestion,stand-
ardisedasperthespecications,werecylinderswithadiameterof15cmandahightof
30cmandcubeswithasidelengthof15cm.Theresultswerepresentedasthecompres-
sivestrengthofcylindricalsamples,whiletheresultsfromthecubicsampleswerecon-
vertedaccordingtothestandardmentionedabove[77]torepresentthestrengthsobtain-
ablefromcylindricalsamples.Thesampleswerefabricatedusingordinaryportlandce-
ment,andthesandemployedwasfreefromclaycontamination.AspertheEN-206-01
standard[77],thestrengthwasexaminedafter28days,generallywhentheconcreteat-
tainedfullstrength.Itshouldbeemphasisedthatthetimerequiredforconcretetoachieve
fullstrengthmainlydependsonthetypeofcementused,andforsometypesofcement,
thisperiodmaybelongerorshorter.However,itwasassumedthatthecementusedin
themixturedidnotresultinanyreductionorextensionofthetimetoachievestrength.
Tostandardisetheinvestigation,sampleswithoutfullstrengthwereexcludedfromthe
dataset.Thequantityofrecordsmentionedaboveisdevoidofsuchsamples.Duetothe
factthatallANNhasbeentrainedonspecicdatasets,itisadvisabletooperatewithin
themaximumvaluesofthemodelsinputparametersandpreferablyavoiddeviantareas.
Insomecases,inputparametersthatfalloutsidethemaximumvaluesofthemodelcan
leadtounreliableorinaccurateresults.
Figure4displaysscaerplotsthatelucidatetheproportionalrelationshipbetween
targetedandinputvariables.Theutilityofscaerplotsasacomprehensivemethodfor
scrutinisingtheinterdependenceofvariablesiswell-established[78].Itisavisuallystrik-
ingdemonstrationofthenexusbetweenthesetwovariablecategories.Owingtothemany
possiblecombinations,onlyaselectnumberofexamplesinvolvingtargetvariableswere
furnishedforillustrativepurposes.Thesaidplotsspecicallycatertotheobjectivevaria-
blecorrespondingtotheconcretecompressivestrength.
Figure 3.
Field distribution comparisons: concrete compressive strength (
A
), cement (
B
), water–cement
ratio (
C
), fine aggregate (
D
), coarse aggregate (
E
). Purple bars correspond to training data. Green bars
correspond to synthetic data. Vertical axes are percentages.
The compressive strength of concrete was assessed in a laboratory setting utilising
standardised samples under the EN-206-01 standard [
77
]. The samples in question, stan-
dardised as per the specifications, were cylinders with a diameter of 15 cm and a hight of
30 cm and cubes with a side length of 15 cm. The results were presented as the compressive
strength of cylindrical samples, while the results from the cubic samples were converted
according to the standard mentioned above [
77
] to represent the strengths obtainable from
cylindrical samples. The samples were fabricated using ordinary portland cement, and the
sand employed was free from clay contamination. As per the EN-206-01 standard [
77
], the
strength was examined after 28 days, generally when the concrete attained full strength. It
should be emphasised that the time required for concrete to achieve full strength mainly
depends on the type of cement used, and for some types of cement, this period may be
Materials 2023,16, 5956 12 of 36
longer or shorter. However, it was assumed that the cement used in the mixture did not
result in any reduction or extension of the time to achieve strength. To standardise the
investigation, samples without full strength were excluded from the dataset. The quantity
of records mentioned above is devoid of such samples. Due to the fact that all ANN has
been trained on specific datasets, it is advisable to operate within the maximum values of
the model’s input parameters and preferably avoid deviant areas. In some cases, input
parameters that fall outside the maximum values of the model can lead to unreliable or
inaccurate results.
Figure 4displays scatter plots that elucidate the proportional relationship between
targeted and input variables. The utility of scatter plots as a comprehensive method for
scrutinising the interdependence of variables is well-established [
78
]. It is a visually striking
demonstration of the nexus between these two variable categories. Owing to the many
possible combinations, only a select number of examples involving target variables were
furnished for illustrative purposes. The said plots specifically cater to the objective variable
corresponding to the concrete compressive strength.
Materials2023,14,xFORPEERREVIEW13of38
(A)(B)
(C)(D)
Figure4.Thescaerplotsdepicttherelationshipbetweentheinputvariablesandthetargetvariable.
TheverticalaxisrepresentsthecompressivestrengthofconcreteinMPa,whilethehorizontalaxis
representsinputvariables:(A)cement(kg/m3),(B)watercementratio(-),(C)neaggregate(kg/m3),
(D)coarseaggregate(kg/m3).
3.3.Training,Testing,andModelSelection
Theabovedatabasewasusedtotrainaseriesofdeeparticialneuralnetworkmod-
els.Thegoaloftheanalysisistoinvestigatetheinuenceofthecomputationalcomplexity
ofDNNsontheaccuracyofpredictingtechnicalparametersofconcrete.Createdmodels,
giventhequantitativecomposition,canestimatethecompressivestrengthofconcrete.In
theanalysisbelow,veneuralnetworkmodelsofvaryingcomputationalcomplexity,dif-
ferentiatedbythenumberofhiddenlayers(MLM1,MLM2,MLM3,MLM4,MLM5),were
compared,repeatingtheentireprocessinthreeseries(I,II,III)forvalidationpurposes.
Fromtheleastcomplexmodel,MLM1hastwohiddenlayers,toMLM5,whichhassix
hiddenlayers.Thenumberofneuronsinatypicalhiddenlayerisfour.Eachmodelhas
veparameters,fourinputvariables,andoneoutputvariable.Foreectivetrainingof
deepneuralnetworks,thedatasetistypicallydividedintothreeindependentparts:train-
ing,validation(orselection),andtesting.Thisisastandardprocedureappliedindeep
learning[79,80].Thetrainingsetisusedtooptimisetheparametersoftheneuralnetwork.
Thevalidationsetallowsitseectivenesstobeassessedduringthelearningprocessand
tochoosethebestmodel,whilethetestsetisusedtoevaluatethenalperformanceofthe
model.Anoutliereliminationprocedurewasimplementedinthedatasetsunderinvesti-
gationwhereanydatapointexceedingthreetimesthemedianvalueofeachvariable,cal-
culatedfromthedataset’scentre,wasexcluded.Theseexclusioncriteria,formulatedto
targetunivariateoutliers,wereemployedtosafeguardtheprecisionanddependabilityof
thesubsequentstatisticalanalyses[81,82].Duetotheinuentialeectofunivariateoutli-
ers,thepotentialdistortionofresultscouldyieldamisinterpretationofthedataset’sactual
Figure 4.
The scatter plots depict the relationship between the input variables and the target variable.
The vertical axis represents the compressive strength of concrete in MPa, while the horizontal axis
represents input variables: (
A
) cement (kg/m
3
), (
B
) water–cement ratio (-), (
C
) fine aggregate (kg/m
3
),
(D) coarse aggregate (kg/m3).
3.3. Training, Testing, and Model Selection
The above database was used to train a series of deep artificial neural network models.
The goal of the analysis is to investigate the influence of the computational complexity of
DNNs on the accuracy of predicting technical parameters of concrete. Created models,
given the quantitative composition, can estimate the compressive strength of concrete.
In the analysis below, five neural network models of varying computational complexity,
differentiated by the number of hidden layers (MLM1, MLM2, MLM3, MLM4, MLM5),
were compared, repeating the entire process in three series (I, II, III) for validation pur-
poses. From the least complex model, MLM1 has two hidden layers, to MLM5, which has
Materials 2023,16, 5956 13 of 36
six hidden layers. The number of neurons in a typical hidden layer is four. Each model
has five parameters, four input variables, and one output variable. For effective training
of deep neural networks, the dataset is typically divided into three independent parts:
training, validation (or selection), and testing. This is a standard procedure applied in deep
learning
[
79
,
80
]. The training set is used to optimise the parameters of the neural network.
The validation set allows its effectiveness to be assessed during the learning process and to
choose the best model, while the test set is used to evaluate the final performance of the
model. An outlier elimination procedure was implemented in the datasets under inves-
tigation where any data point exceeding three times the median value of each variable,
calculated from the dataset’s centre, was excluded. These exclusion criteria, formulated to
target univariate outliers, were employed to safeguard the precision and dependability of
the subsequent statistical analyses [
81
,
82
]. Due to the influential effect of univariate outliers,
the potential distortion of results could yield a misinterpretation of the dataset’s actual char-
acteristics. Thus, outlier removal enhances the sample’s representativeness, contributing to
more robust and reliable results. It is essential to acknowledge that this procedure, despite
potentially impacting the sample size, is essential in affirming this study’s validity. The
applied process ensures the statistical integrity of the analyses, reinforcing the reliability of
this study’s conclusions. The used dataset was allocated as follows: 59.6% of the records
(3689) were assigned to the training set, 19.9% (1229 records) to the selection set, 19.9%
(1231 records) to the test set, and 0.6% (38 records) were unused. The number of input,
target, and unused variables in the final models and the division of subsets are presented
in Figure 5.
Materials2023,14,xFORPEERREVIEW14of38
characteristics.Thus,outlierremovalenhancesthesamplesrepresentativeness,contrib-
utingtomorerobustandreliableresults.Itisessentialtoacknowledgethatthisprocedure,
despitepotentiallyimpactingthesamplesize,isessentialinarmingthisstudy’svalidity.
Theappliedprocessensuresthestatisticalintegrityoftheanalyses,reinforcingtherelia-
bilityofthisstudy’sconclusions.Theuseddatasetwasallocatedasfollows:59.6%ofthe
records(3689)wereassignedtothetrainingset,19.9%(1229records)totheselectionset,
19.9%(1231records)tothetestset,and0.6%(38records)wereunused.Thenumberof
input,target,andunusedvariablesinthenalmodelsandthedivisionofsubsetsare
presentedinFigure5.
(A)(B)
Figure5.Numberofinput,target,andunusedvariablesinnalmodelsanddivisionofsubsets.The
diagramincludesavariablebarchart(A)andasamplepiechart(B).
Thearchitectureoftherstmodel,MLM1,consistsof20neurons,includingfourin-
putneurons,nineneuronsspreadovertwohiddenlayers,withfourneuronsinonescal-
inglayer,oneneuroninthedescalinglayer,oneneuroninthebondinglayer,andone
neuronintheoutputlayer.Thearchitectureofthesecondmodel,MLM2,consistsof28
neurons,includingfourinputneurons,17neuronsspreadoverthreehiddenlayers,with
fourneuronsinonescalinglayer,oneneuroninthedescalinglayer,oneneuroninthe
bondinglayer,andoneneuronintheoutputlayer.Thearchitectureofthethirdmodel,
MLM3,consistsof36neurons,includingfourinputneurons,25neuronsspreadoverfour
hiddenlayers,withfourneuronsinonescalinglayer,oneneuroninthedescalinglayer,
oneneuroninthebondinglayer,andoneneuronintheoutputlayer.Thearchitectureof
thefourthmodel,MLM4,consistsof44neurons,includingfourinputneurons,33neurons
spreadovervehiddenlayers,withfourneuronsinonescalinglayer,oneneuroninthe
descalinglayer,oneneuroninthebondinglayer,andoneneuronintheoutputlayer.The
architectureofthefthmodel,MLM5,consistsof52neurons,includingfourinputneu-
rons,41neuronsspreadoversixhiddenlayers,withfourneuronsinonescalinglayer,
oneneuroninthedescalinglayer,oneneuroninthebondinglayer,andoneneuroninthe
outputlayer.Figure6showsthearchitecturesofthemodelsanalysedinthisresearch.
(A)(B)
Figure 5.
Number of input, target, and unused variables in final models and division of subsets. The
diagram includes a variable bar chart (A) and a sample pie chart (B).
The architecture of the first model, MLM1, consists of 20 neurons, including four
input neurons, nine neurons spread over two hidden layers, with four neurons in one
scaling layer, one neuron in the descaling layer, one neuron in the bonding layer, and
one neuron in the output layer. The architecture of the second model, MLM2, consists of
28 neurons, including four input neurons, 17 neurons spread over three hidden layers,
with four neurons in one scaling layer, one neuron in the descaling layer, one neuron in
the bonding layer, and one neuron in the output layer. The architecture of the third model,
MLM3, consists of 36 neurons, including four input neurons, 25 neurons spread over four
hidden layers, with four neurons in one scaling layer, one neuron in the descaling layer,
one neuron in the bonding layer, and one neuron in the output layer. The architecture of
the fourth model, MLM4, consists of 44 neurons, including four input neurons, 33 neurons
spread over five hidden layers, with four neurons in one scaling layer, one neuron in the
descaling layer, one neuron in the bonding layer, and one neuron in the output layer. The
architecture of the fifth model, MLM5, consists of 52 neurons, including four input neurons,
41 neurons spread over six hidden layers, with four neurons in one scaling layer, one
neuron in the descaling layer, one neuron in the bonding layer, and one neuron in the
output layer. Figure 6shows the architectures of the models analysed in this research.
Materials 2023,16, 5956 14 of 36
Materials2023,14,xFORPEERREVIEW14of38
characteristics.Thus,outlierremovalenhancesthesamplesrepresentativeness,contrib-
utingtomorerobustandreliableresults.Itisessentialtoacknowledgethatthisprocedure,
despitepotentiallyimpactingthesamplesize,isessentialinarmingthisstudy’svalidity.
Theappliedprocessensuresthestatisticalintegrityoftheanalyses,reinforcingtherelia-
bilityofthisstudy’sconclusions.Theuseddatasetwasallocatedasfollows:59.6%ofthe
records(3689)wereassignedtothetrainingset,19.9%(1229records)totheselectionset,
19.9%(1231records)tothetestset,and0.6%(38records)wereunused.Thenumberof
input,target,andunusedvariablesinthenalmodelsandthedivisionofsubsetsare
presentedinFigure5.
(A)(B)
Figure5.Numberofinput,target,andunusedvariablesinnalmodelsanddivisionofsubsets.The
diagramincludesavariablebarchart(A)andasamplepiechart(B).
Thearchitectureoftherstmodel,MLM1,consistsof20neurons,includingfourin-
putneurons,nineneuronsspreadovertwohiddenlayers,withfourneuronsinonescal-
inglayer,oneneuroninthedescalinglayer,oneneuroninthebondinglayer,andone
neuronintheoutputlayer.Thearchitectureofthesecondmodel,MLM2,consistsof28
neurons,includingfourinputneurons,17neuronsspreadoverthreehiddenlayers,with
fourneuronsinonescalinglayer,oneneuroninthedescalinglayer,oneneuroninthe
bondinglayer,andoneneuronintheoutputlayer.Thearchitectureofthethirdmodel,
MLM3,consistsof36neurons,includingfourinputneurons,25neuronsspreadoverfour
hiddenlayers,withfourneuronsinonescalinglayer,oneneuroninthedescalinglayer,
oneneuroninthebondinglayer,andoneneuronintheoutputlayer.Thearchitectureof
thefourthmodel,MLM4,consistsof44neurons,includingfourinputneurons,33neurons
spreadovervehiddenlayers,withfourneuronsinonescalinglayer,oneneuroninthe
descalinglayer,oneneuroninthebondinglayer,andoneneuronintheoutputlayer.The
architectureofthefthmodel,MLM5,consistsof52neurons,includingfourinputneu-
rons,41neuronsspreadoversixhiddenlayers,withfourneuronsinonescalinglayer,
oneneuroninthedescalinglayer,oneneuroninthebondinglayer,andoneneuroninthe
outputlayer.Figure6showsthearchitecturesofthemodelsanalysedinthisresearch.
(A)(B)
Figure 6.
The topology of the deep neural network (DNN) for each model is presented as follows:
MLM1 (
A
)
,
MLM2 (
B
), MLM3 (
C
), MLM4 (
D
), and MLM5 (
E
). The figure shows the DNN architecture,
which includes input neurons (green), scaling neurons (orange), hidden neurons (yellow), descaling
neurons (red), bonding neurons (purple), and output neurons (blue).
The data features, represented as input variables, are assigned to the input neurons of
the neural network structure, while the output neuron is connected to the target variables.
To enhance the model’s effectiveness, a method known as feature scaling was implemented
across all models. This process entails converting the numerical attributes of the data into
a specific scale [
83
]. This scaling and the subsequent descaling were carried out using the
mean standard deviation (MSD) as a scaler [
84
]. The models maintain consistent activation
functions, with the hyperbolic tangent [
85
87
] being used for the hidden layers and the
linear tangent [
87
] for the output layer. A bonding layer was also incorporated into the
models. The constructed models were meticulously calibrated to minimise the associated
loss function. The quantification of the model’s error in computing the index loss was
executed utilising the normalized squared error (NSE). It is to be noted that lower NSE
values are indicative of a model’s superior predictive capabilities. Conversely, values
tending towards one highlight the model’s weaker predictive potential, while values close
to zero signify a commendable predictive performance. To further enhance the performance
of our model and inhibit the potential for overfitting or underfitting, the adoption of
regularisation strategies was necessitated. Specifically, the L2 method [
88
90
] was instituted
as our chosen regularisation function. This regularisation phase is instrumental in tuning
the model by minimising the adjusted loss function. It contributes towards mitigating
biases and, consequently, facilitates more precise predictions. It is pertinent to underscore
the importance of the regularisation step in the model-building process, as it aids in
ensuring that the model is absorbing the intrinsic patterns and relationships within the
data instead of merely reproducing the training data. In successfully preventing overfitting
Materials 2023,16, 5956 15 of 36
and underfitting, a model is deemed accurate and generalisable, rendering it suitable for
application to novel, unseen data. In this context, the efficacy of the L2 regularisation
method [
88
90
] is particularly pronounced. It appends a penalty term to the loss function
proportional to the square of the magnitude of the network parameters’ weights. As a
result, the weights are driven towards zero, facilitating the generation of smaller, simpler
models less susceptible to overfitting.
The quasi-Newton method [
91
,
92
] was employed as the optimisation algorithm in the
present study. The quasi-Newton method is a popular choice for optimisation algorithms
due to its efficiency and effectiveness in tackling large-scale optimisation problems [
92
]. It
does so by using first-order derivative information to build up an approximation of the
Hessian matrix [
93
], which represents the second-order partial derivatives of the objective
function [
91
]. This method has shown substantial success in solving nonlinear optimisa-
tion problems that arise in diverse applications, thanks to its robustness and ability to
converge to the solution more quickly than traditional gradient descent algorithms. The
quasi-Newton method’s effectiveness is enhanced by its ability to handle functions that
are not necessarily smooth, making it a versatile tool in optimisation [
94
]. Employing
this algorithm in the current study was an integral part of the process, allowing for the
efficient optimisation of the model’s parameters. The resulting loss history for the model
and series are presented in Figure 7A–C (for series I, II, and III, respectively), illustrating
the model’s learning progression throughout the training process. The loss history repre-
sents how the loss function of a machine learning model changes over the course of its
training and selection process. The loss function quantifies the discrepancy between the
predicted outputs of the model and the actual target values, and the goal of training is to
minimize this loss. The adopted training strategy proved to be highly effective in optimis-
ing the model’s performance, providing the desired level of accuracy while minimising
computational resources.
Materials2023,14,xFORPEERREVIEW16of38
Thequasi-Newtonmethod[91,92]wasemployedastheoptimisationalgorithmin
thepresentstudy.Thequasi-Newtonmethodisapopularchoiceforoptimisationalgo-
rithmsduetoitseciencyandeectivenessintacklinglarge-scaleoptimisationproblems
[92].Itdoessobyusingrst-orderderivativeinformationtobuildupanapproximation
oftheHessianmatrix[93],whichrepresentsthesecond-orderpartialderivativesofthe
objectivefunction[91].Thismethodhasshownsubstantialsuccessinsolvingnonlinear
optimisationproblemsthatariseindiverseapplications,thankstoitsrobustnessandabil-
itytoconvergetothesolutionmorequicklythantraditionalgradientdescentalgorithms.
Thequasi-Newtonmethod’seectivenessisenhancedbyitsabilitytohandlefunctions
thatarenotnecessarilysmooth,makingitaversatiletoolinoptimisation[94].Employing
thisalgorithminthecurrentstudywasanintegralpartoftheprocess,allowingforthe
ecientoptimisationofthemodelsparameters.Theresultinglosshistoryforthemodel
andseriesarepresentedinFigure7A–C(forseriesI,II,andIII,respectively),illustrating
themodelslearningprogressionthroughoutthetrainingprocess.Thelosshistoryrepre-
sentshowthelossfunctionofamachinelearningmodelchangesoverthecourseofits
trainingandselectionprocess.Thelossfunctionquantiesthediscrepancybetweenthe
predictedoutputsofthemodelandtheactualtargetvalues,andthegoaloftrainingisto
minimizethisloss.Theadoptedtrainingstrategyprovedtobehighlyeectiveinoptimis-
ingthemodelsperformance,providingthedesiredlevelofaccuracywhileminimising
computationalresources.
(a)(b)
(c)(d)
Figure 7. Cont.
Materials 2023,16, 5956 16 of 36
Materials2023,14,xFORPEERREVIEW17of38
(e)
(A)
(a)(b)
(c)(d)
(e)
(B)
Figure 7. Cont.
Materials 2023,16, 5956 17 of 36
Materials2023,14,xFORPEERREVIEW18of38
(a)(b)
(c)(d)
(e)
(C)
Figure7.(A)LosshistorydiagramforspecicmodelsinseriesI:MLM1(epochfrom1to171)(a),
MLM2(epochfrom1to285)(b),MLM3(epochfrom1to332)(c),MLM4(epochfrom1to295)(d),
MLM5(epochfrom1to321)(e).(B)LosshistorydiagramforspecicmodelsinseriesII:MLM1
(epochfrom1to147)(a),MLM2(epochfrom1to248)(b),MLM3(epochfrom1to294)(c),MLM4
(epochfrom1to320)(d),MLM5(epochfrom1to298)(e).(C)Losshistorydiagramforspecic
modelsinseriesIII:MLM1(epochfrom1to128)(a),MLM2(epochfrom1to278)(b),MLM3(epoch
from1to257)(c),MLM4(epochfrom1to296)(d),MLM5(epochfrom1to280)(e).
Itshouldbenotedthatthemodelsweretrainedonaspecicsetofdata.Therefore,
whenusingthemodels,oneshouldoperatewithinthevaluesindicatedinTable2asmin-
imalandmaximal.Thefollowingresearchdoesnotconsidertheimpactofusingadditives
andadmixturesonconcrete.Theusablerangeforthewater–cementratioextendsfrom
approximately0.3tobeyond0.8.Aproportionof0.3resultsinhighlyrigidconsistency
(unlesssuperplasticisersareemployed),whilearatioof0.8yieldsconcretethatisdamp
Figure 7.
(
A
) Loss history diagram for specific models in series I: MLM1 (epoch from 1 to 171) (
a
),
MLM2 (epoch from 1 to 285) (
b
), MLM3 (epoch from 1 to 332) (
c
), MLM4 (epoch from 1 to 295) (
d
),
MLM5 (epoch from 1 to 321) (
e
). (
B
) Loss history diagram for specific models in series II: MLM1
(epoch from 1 to 147) (
a
), MLM2 (epoch from 1 to 248) (
b
), MLM3 (epoch from 1 to 294) (
c
), MLM4
(epoch from 1 to 320) (
d
), MLM5 (epoch from 1 to 298) (
e
). (
C
) Loss history diagram for specific
models in series III: MLM1 (epoch from 1 to 128) (
a
), MLM2 (epoch from 1 to 278) (
b
), MLM3 (epoch
from 1 to 257) (c), MLM4 (epoch from 1 to 296) (d), MLM5 (epoch from 1 to 280) (e).
It should be noted that the models were trained on a specific set of data. Therefore,
when using the models, one should operate within the values indicated in Table 2as
minimal and maximal. The following research does not consider the impact of using
additives and admixtures on concrete. The usable range for the water–cement ratio extends
from approximately 0.3 to beyond 0.8. A proportion of 0.3 results in highly rigid consistency
(unless superplasticisers are employed), while a ratio of 0.8 yields concrete that is damp
and lacking in strength [
95
,
96
]. All records outside the range water–cement 0.3–0.8 have
been removed from the dataset.
Materials 2023,16, 5956 18 of 36
3.4. Results and Discussion
In the following study, five models of deep artificial neural networks with varying
degrees of computational complexity were analysed, named consecutively MLM1, MLM2,
MLM3, MLM4, and MLM5, starting from the least complex network to the most complex,
in three series I, II, III. First, a feature correlation analysis was prepared to determine precise
relationships between individual variables. The result of the analysis is a feature correlation
heatmap visible in Figure 8.
Materials2023,14,xFORPEERREVIEW19of38
andlackinginstrength[95,96].Allrecordsoutsidetherangewatercement0.3–0.8have
beenremovedfromthedataset.
3.4.ResultsandDiscussion
Inthefollowingstudy,vemodelsofdeeparticialneuralnetworkswithvarying
degreesofcomputationalcomplexitywereanalysed,namedconsecutivelyMLM1,
MLM2,MLM3,MLM4,andMLM5,startingfromtheleastcomplexnetworktothemost
complex,inthreeseriesI,II,III.First,afeaturecorrelationanalysiswaspreparedtodeter-
minepreciserelationshipsbetweenindividualvariables.Theresultoftheanalysisisa
featurecorrelationheatmapvisibleinFigure8.
Figure8.Featurecorrelationheatmap.
Thefeaturecorrelationheatmaprevealstherelationshipsofindividualvariables,
whereavaluecloserto1.0indicatesastrongercorrelation[97].Onecannoticethatthe
inputvariablesrelatedtothewatercementratioandthecementcontenthavethestrong-
estassociationwiththeoutputvariable.Therelationshipsofotherinputvariableswith
theoutputvariablearemuchweaker,withtheamountofneaggregatehavingamore
signicantimpactthanwatercontentandwatercontenthavingagreaterimpactthanthe
amountofcoarseaggregate.Itcanbeobservedthattheimpactofthewatercementratio
andthequantityofcementonthestrengthofconcreteisevident.Severalresearchpapers
intheliteraturehavearmedthecrucialroleofthesefactorsindeterminingthecompres-
sivestrengthofconcrete[98].
Eachofthemodelswassubjectedtoagoodness-of-ttest[99–103].Thisassessment
providesameanstoquantifythediscrepancybetweenobservedvaluesandthosepre-
dictedbythemodel.Astandardmetricusedtogaugethegoodness-of-tinscienticin-
vestigationsisthecoecientofdetermination,denotedasR
2
[104,105].Thisparameteris
utilizedtomeasuretheextentofdisparitybetweenobservedvaluesandprojectedpredic-
tions.Morespecically,R
2
determinesthefractionofthisdiscrepancythatcanbeac-
countedforbythemodel.Inascenariowherethemodeltisideal,resultinginoutput
valuesperfectlymatchingthetargetvalues,theR
2
coecientwouldequatetoone.Figure
9A–Cprovideadetailedvisualrepresentationofagoodness-of-tanalysisforseriesI,II,
andIII,utilisingthestatisticalmeasureknownasthecoecientofdetermination(R
2
).
Figure 8. Feature correlation heatmap.
The feature correlation heatmap reveals the relationships of individual variables,
where a value closer to 1.0 indicates a stronger correlation [
97
]. One can notice that the
input variables related to the water–cement ratio and the cement content have the strongest
association with the output variable. The relationships of other input variables with the
output variable are much weaker, with the amount of fine aggregate having a more signifi-
cant impact than water content and water content having a greater impact than the amount
of coarse aggregate. It can be observed that the impact of the water–cement ratio and the
quantity of cement on the strength of concrete is evident. Several research papers in the
literature have affirmed the crucial role of these factors in determining the compressive
strength of concrete [98].
Each of the models was subjected to a goodness-of-fit test [
99
103
]. This assessment
provides a means to quantify the discrepancy between observed values and those predicted
by the model. A standard metric used to gauge the goodness-of-fit in scientific investiga-
tions is the coefficient of determination, denoted as R
2
[
104
,
105
]. This parameter is utilized
to measure the extent of disparity between observed values and projected predictions.
More specifically, R
2
determines the fraction of this discrepancy that can be accounted for
by the model. In a scenario where the model fit is ideal, resulting in output values perfectly
matching the target values, the R
2
coefficient would equate to one. Figure 9A–C provide
a detailed visual representation of a goodness-of-fit analysis for series I, II, and III, utilising
the statistical measure known as the coefficient of determination (R2).
In the considered issue, a series of R
2
values were calculated for the target variable f
ck
in the individual models and series. In series I, it was 0.5691 for MLM1, 0.6268 for MLM2,
0.6053 for MLM3, 0.6438 for MLM4, and 0.6453 for MLM5. In series II, it was 0.5467 for
MLM1, 0.6017 for MLM2, 0.6227 for MLM3, 0.6285 for MLM4, and 0.6514 for MLM5. In
series III, it was 0.5272 for MLM1, 0.5959 for MLM2, 0.6136 for MLM3, 0.6337 for MLM4,
and 0.6571 for MLM5. The values of the coefficient of determination (R
2
) for each model
are shown in Figure 10.
Materials 2023,16, 5956 19 of 36
Materials2023,14,xFORPEERREVIEW20of38
(a)(b)
(c)(d)
(e)
(A)
Materials2023,14,xFORPEERREVIEW21of38
(a)(b)
(c)(d)
(e)
(B)
(a)(b)
Figure 9. Cont.
Materials 2023,16, 5956 20 of 36
Materials2023,14,xFORPEERREVIEW21of38
(a)(b)
(c)(d)
(e)
(B)
(a)(b)
Materials2023,14,xFORPEERREVIEW22of38
(c)(d)
(e)
(C)
Figure9.(A)Goodness-of-tchartforMLM1(a),MLM2(b),MLM3(c),MLM4(d),MLM5(e)in
seriesI.Thechartshowspredictedvalueoftargetvariableversusrealone.(B)Goodness-of-tchart
forMLM1(a),MLM2(b),MLM3(c),MLM4(d),MLM5(e)inseriesII.Thechartshowspredicted
valueoftargetvariableversusrealone.(C)Goodness-of-tchartforMLM1(a),MLM2(b),MLM3
(c),MLM4(d),MLM5(e)inseriesIII.Thechartshowspredictedvalueoftargetvariableversusreal
one.
Intheconsideredissue,aseriesofR2valueswerecalculatedforthetargetvariablefck
intheindividualmodelsandseries.InseriesI,itwas0.5691forMLM1,0.6268forMLM2,
0.6053forMLM3,0.6438forMLM4,and0.6453forMLM5.InseriesII,itwas0.5467for
MLM1,0.6017forMLM2,0.6227forMLM3,0.6285forMLM4,and0.6514forMLM5.In
seriesIII,itwas0.5272forMLM1,0.5959forMLM2,0.6136forMLM3,0.6337forMLM4,
and0.6571forMLM5.Thevaluesofthecoecientofdetermination(R2)foreachmodel
areshowninFigure10.
Figure 9. Cont.
Materials 2023,16, 5956 21 of 36
Materials2023,14,xFORPEERREVIEW22of38
(c)(d)
(e)
(C)
Figure9.(A)Goodness-of-tchartforMLM1(a),MLM2(b),MLM3(c),MLM4(d),MLM5(e)in
seriesI.Thechartshowspredictedvalueoftargetvariableversusrealone.(B)Goodness-of-tchart
forMLM1(a),MLM2(b),MLM3(c),MLM4(d),MLM5(e)inseriesII.Thechartshowspredicted
valueoftargetvariableversusrealone.(C)Goodness-of-tchartforMLM1(a),MLM2(b),MLM3
(c),MLM4(d),MLM5(e)inseriesIII.Thechartshowspredictedvalueoftargetvariableversusreal
one.
Intheconsideredissue,aseriesofR2valueswerecalculatedforthetargetvariablefck
intheindividualmodelsandseries.InseriesI,itwas0.5691forMLM1,0.6268forMLM2,
0.6053forMLM3,0.6438forMLM4,and0.6453forMLM5.InseriesII,itwas0.5467for
MLM1,0.6017forMLM2,0.6227forMLM3,0.6285forMLM4,and0.6514forMLM5.In
seriesIII,itwas0.5272forMLM1,0.5959forMLM2,0.6136forMLM3,0.6337forMLM4,
and0.6571forMLM5.Thevaluesofthecoecientofdetermination(R2)foreachmodel
areshowninFigure10.
Figure 9.
(
A
) Goodness-of-fit chart for MLM1 (
a
), MLM2 (
b
), MLM3 (
c
), MLM4 (
d
), MLM5 (
e
) in
series I. The chart shows predicted value of target variable versus real one. (
B
) Goodness-of-fit chart
for MLM1 (
a
), MLM2 (
b
), MLM3 (
c
), MLM4 (
d
), MLM5 (
e
) in series II. The chart shows predicted
value of target variable versus real one. (
C
) Goodness-of-fit chart for MLM1 (
a
), MLM2 (
b
), MLM3 (
c
),
MLM4 (
d
), MLM5 (
e
) in series III. The chart shows predicted value of target variable versus real one.
Materials2023,14,xFORPEERREVIEW22of38
(e)
(C)
Figure9.(A)Goodness-of-tchartforMLM1(a),MLM2(b),MLM3(c),MLM4(d),MLM5(e)in
seriesI.Thechartshowspredictedvalueoftargetvariableversusrealone.(B)Goodness-of-tchart
forMLM1(a),MLM2(b),MLM3(c),MLM4(d),MLM5(e)inseriesII.Thechartshowspredicted
valueoftargetvariableversusrealone.(C)Goodness-of-tchartforMLM1(a),MLM2(b),MLM3
(c),MLM4(d),MLM5(e)inseriesIII.Thechartshowspredictedvalueoftargetvariableversusreal
one.
Intheconsideredissue,aseriesofR2valueswerecalculatedforthetargetvariablefck
intheindividualmodelsandseries.InseriesI,itwas0.5691forMLM1,0.6268forMLM2,
0.6053forMLM3,0.6438forMLM4,and0.6453forMLM5.InseriesII,itwas0.5467for
MLM1,0.6017forMLM2,0.6227forMLM3,0.6285forMLM4,and0.6514forMLM5.In
seriesIII,itwas0.5272forMLM1,0.5959forMLM2,0.6136forMLM3,0.6337forMLM4,
and0.6571forMLM5.Thevaluesofthecoecientofdetermination(R2)foreachmodel
areshowninFigure10.
Figure10.Thecoecientofdetermination(R2)valuesforMLM1,MLM2,MLM3,MLM4,andMLM5
inseriesI,II,andIII.
Itcanbeobservedthatareasonablygoodperformanceofthecreatedmodelswas
achieved.Simultaneously,withtheincreasedcomputationalcomplexity,ahigherR2value
Figure 10.
The coefficient of determination (R
2
) values for MLM1, MLM2, MLM3, MLM4, and MLM5
in series I, II, and III.
It can be observed that a reasonably good performance of the created models was
achieved. Simultaneously, with the increased computational complexity, a higher R
2
value
can be noticed in more complex models, suggesting that these models exhibit better predic-
tive capabilities than less complex models. This observation is evident in three series, except
for series I, where the MLM2 model has a higher R
2
coefficient value than the MLM3 model.
This deviation was not observed in the remaining series and may result from various
reasons. An exhaustive analysis of the model’s errors was performed by computing a range
of error metrics across each series. This analysis included metrics such as mean squared
error, Minkowski error, normalized squared error, root mean squared error, sum squared
error [
106
109
]. Furthermore, a detailed report outlining the minimum and maximum
values and the mean and standard deviation for absolute, relative, and percentage errors
of the model concerning the test data was provided. Histograms were constructed for
Materials 2023,16, 5956 22 of 36
the test subset to obtain a more tangible understanding of the distribution of errors in the
model. The outcomes of this analysis provide a rigorous evaluation of the model’s accuracy
and precision, illuminating potential avenues for further refinement. Table 3and Figure 11
present individual error metrics for each model, divided into respective subsets, for series I,
II, and III.
Table 3. Error metrics for various data subsets.
Series I
MLM1
Training Selection Testing
Sum squared error 550.201 342.198 317.183
Mean squared error 0.149 0.278 0.258
Root mean squared error 0.386 0.528 0.508
Normalised squared error
0.416 0.479 0.432
Minkowski error 1893.670 975.049 911.882
MLM2
Training Selection Testing
Sum squared error 489.493 310.386 295.271
Mean squared error 0.133 0.253 0.240
Root mean squared error 0.364 0.503 0.490
Normalised squared error
0.330 0.394 0.374
Minkowski error 1680.550 876.432 842.034
MLM3
Training Selection Testing
Sum squared error 458.547 313.610 305.511
Mean squared error 0.124 0.255 0.248
Root mean squared error 0.353 0.505 0.498
Normalised squared error
0.289 0.402 0.400
Minkowski error 1561.140 865.786 860.635
MLM4
Training Selection Testing
Sum squared error 447.218 307.594 290.030
Mean squared error 0.121 0.250 0.236
Root mean squared error 0.348 0.500 0.485
Normalised squared error
0.275 0.387 0.361
Minkowski error 1510.870 842.743 807.330
MLM5
Training Selection Testing
Sum squared error 416.643 296.432 291.111
Mean squared error 0.113 0.241 0.236
Root mean squared error 0.336 0.491 0.486
Normalised squared error
0.239 0.359 0.364
Minkowski error 1403.290 810.031 798.990
Series II
MLM1
Training Selection Testing
Sum squared error 559.448 340.780 325.536
Mean squared error 0.152 0.277 0.264
Root mean squared error 0.389 0.527 0.514
Normalised squared error
0.431 0.475 0.455
Minkowski error 1944.850 974.611 947.078
Materials 2023,16, 5956 23 of 36
Table 3. Cont.
MLM2
Training Selection Testing
Sum squared error 493.298 322.494 305.668
Mean squared error 0.134 0.262 0.248
Root mean squared error 0.366 0.512 0.498
Normalised squared error
0.335 0.425 0.401
Minkowski error 1687.480 905.087 867.447
MLM3
Training Selection Testing
Sum squared error 454.568 310.576 297.960
Mean squared error 0.123 0.253 0.242
Root mean squared error 0.351 0.503 0.492
Normalised squared error
0.284 0.394 0.381
Minkowski error 1539.040 852.571 837.393
MLM4
Training Selection Testing
Sum squared error 423.191 298.929 297.827
Mean squared error 0.115 0.243 0.242
Root mean squared error 0.339 0.493 0.492
Normalised squared error
0.246 0.365 0.381
Minkowski error 1431.560 823.304 820.941
MLM5
Training Selection Testing
Sum squared error 435.508 304.092 286.750
Mean squared error 0.118 0.247 0.233
Root mean squared error 0.344 0.497 0.483
Normalised squared error
0.261 0.378 0.353
Minkowski error 1458.890 834.351 799.642
Series III
MLM1
Training Selection Testing
Sum squared error 575.272 347.520 332.199
Mean squared error 0.156 0.283 0.270
Root mean squared error 0.395 0.532 0.519
Normalised squared error
0.455 0.494 0.473
Minkowski error 2007.800 1004.980 969.083
MLM2
Training Selection Testing
Sum squared error 500.105 321.972 307.902
Mean squared error 0.136 0.262 0.250
Root mean squared error 0.368 0.512 0.500
Normalised squared error
0.344 0.424 0.407
Minkowski error 1706.130 905.095 879.370
MLM3
Training Selection Testing
Sum squared error 473.081 318.669 301.983
Mean squared error 0.128 0.259 0.245
Root mean squared error 0.358 0.509 0.495
Normalised squared error
0.308 0.415 0.391
Minkowski error 1608.550 889.350 844.676
Materials 2023,16, 5956 24 of 36
Table 3. Cont.
MLM4
Training Selection Testing
Sum squared error 419.356 315.651 295.121
Mean squared error 0.114 0.257 0.240
Root mean squared error 0.337 0.507 0.490
Normalised squared error
0.242 0.407 0.374
Minkowski error 1407.230 862.350 814.506
MLM5
Training Selection Testing
Sum squared error 455.535 317.058 283.630
Mean squared error 0.123 0.258 0.230
Root mean squared error 0.351 0.508 0.480
Normalised squared error
0.286 0.411 0.345
Minkowski error 1529.650 866.807 798.284
In Figure 11, a clear error downward trend can be noticed for the NSE metric in the
training, selection, and testing sections and for the ME metric in the training section in
all three series, with the increase in computational complexity. Simultaneously, a milder
error downward trend can be observed in the remaining metrics, along with the increase
in computational complexity. Statistical calculations were conducted for individual target
variables, presented in Table 4. Table 4provides minimums, maximums, averages, and
standard deviations of absolute and percentage errors of the model for the test data.
Figure 12 shows the values of mean error (absolute error, relative error, percentage error)
for every model in series I, II, and III.
Materials2023,14,xFORPEERREVIEW25of38
Rootmeansquarederror0.3680.5120.500
Normalisedsquarederror0.3440.4240.407
Minkowskierror1706.130905.095879.370
MLM3
TrainingSelectionTesting
Sumsquarederror473.081318.669301.983
Meansquarederror0.1280.2590.245
Rootmeansquarederror0.3580.5090.495
Normalisedsquarederror0.3080.4150.391
Minkowskierror1608.550889.350844.676
MLM4
TrainingSelectionTesting
Sumsquarederror419.356315.651295.121
Meansquarederror0.1140.2570.240
Rootmeansquarederror0.3370.5070.490
Normalisedsquarederror0.2420.4070.374
Minkowskierror1407.230862.350814.506
MLM5
TrainingSelectionTesting
Sumsquarederror455.535317.058283.630
Meansquarederror0.1230.2580.230
Rootmeansquarederror0.3510.5080.480
Normalisedsquarederror0.2860.4110.345
Minkowskierror1529.650866.807798.284
(A)
Figure 11. Cont.
Materials 2023,16, 5956 25 of 36
Materials2023,14,xFORPEERREVIEW26of38
(B)
Figure11.ErrormetricsforMLM1,MLM2,MLM3,MLM4,MLM5inseriesI,II,andIII.Errormet-
ricsused:meansquarederror(MSE)(B),Minkowskierror(ME)(A),normalizedsquarederror
(NSE)(B),rootmeansquarederror(RMSE)(B),sumsquarederror(SSE)(A).
InFigure11,aclearerrordownwardtrendcanbenoticedfortheNSEmetricinthe
training,selection,andtestingsectionsandfortheMEmetricinthetrainingsectioninall
threeseries,withtheincreaseincomputationalcomplexity.Simultaneously,amilderer-
rordownwardtrendcanbeobservedintheremainingmetrics,alongwiththeincreasein
computationalcomplexity.Statisticalcalculationswereconductedforindividualtarget
variables,presentedinTable4.Table4providesminimums,maximums,averages,and
standarddeviationsofabsoluteandpercentageerrorsofthemodelforthetestdata.Fig-
ure12showsthevaluesofmeanerror(absoluteerror,relativeerror,percentageerror)for
everymodelinseriesI,II,andIII.
Tab l e 4.Errorstatisticsforcalculationoftargetvaluefck.
SeriesI
MinimumMaximumMeanDeviation
Absoluteerror
MLM10.000001192090.06953680.009366550.00818558
MLM20.000001043080.01208250.001453320.00118896
MLM30.0000001192090.08793880.002732780.0041617
MLM40.000001192090.1486370.003314780.00663424
MLM50.000001192090.1486370.003314780.00663424
Relativeerror
MLM10.00000114130.06657430.008967490.00783684
MLM20.000008510710.09858320.0118580.009701
MLM30.0000002963040.2185790.006792550.0103442
MLM40.000002132660.2659130.005930160.0118687
MLM50.000001192090.1486370.003314780.00663424
Percentageerror
MLM10.000114136.657430.8967490.783684
MLM20.0008510719.858321.18580.9701
Figure 11.
Error metrics for MLM1, MLM2, MLM3, MLM4, MLM5 in series I, II, and III. Error metrics
used: mean squared error (MSE) (
B
), Minkowski error (ME) (
A
), normalized squared error (NSE) (
B
),
root mean squared error (RMSE) (B), sum squared error (SSE) (A).
Table 4. Error statistics for calculation of target value fck.
Series I
Minimum Maximum Mean Deviation
Absolute error
MLM1 0.00000119209 0.0695368 0.00936655 0.00818558
MLM2 0.00000104308 0.0120825 0.00145332 0.00118896
MLM3 0.000000119209 0.0879388 0.00273278 0.0041617
MLM4 0.00000119209 0.148637 0.00331478 0.00663424
MLM5 0.00000119209 0.148637 0.00331478 0.00663424
Relative error
MLM1 0.0000011413 0.0665743 0.00896749 0.00783684
MLM2 0.00000851071 0.0985832 0.011858 0.009701
MLM3 0.000000296304 0.218579 0.00679255 0.0103442
MLM4 0.00000213266 0.265913 0.00593016 0.0118687
MLM5 0.00000119209 0.148637 0.00331478 0.00663424
Percentage error
MLM1 0.00011413 6.65743 0.896749 0.783684
MLM2 0.000851071 9.85832 1.1858 0.9701
MLM3 0.0000296304 21.8579 0.679255 1.03442
MLM4 0.000213266 26.5913 0.593016 1.18687
MLM5 0.00000119209 0.148637 0.00331478 0.00663424
Series II
Minimum Maximum Mean Deviation
Absolute error
MLM1 0.00000119209 0.0695368 0.00936655 0.00818558
MLM2 0.00000104308 0.0120825 0.00145332 0.00118896
MLM3 0.000000119209 0.0879388 0.00273278 0.0041617
Materials 2023,16, 5956 26 of 36
Table 4. Cont.
MLM4 0.00000119209 0.148637 0.00331478 0.00663424
MLM5 0.00000119209 0.148637 0.00331478 0.00663424
Relative error
MLM1 0.0000011413 0.0665743 0.00896749 0.00783684
MLM2 0.00000851071 0.0985832 0.011858 0.009701
MLM3 0.000000296304 0.218579 0.00679255 0.0103442
MLM4 0.00000213266 0.265913 0.00593016 0.0118687
MLM5 0.00000119209 0.148637 0.00331478 0.00663424
Percentage error
MLM1 0.00011413 6.65743 0.896749 0.783684
MLM2 0.000851071 9.85832 1.1858 0.9701
MLM3 0.0000296304 21.8579 0.679255 1.03442
MLM4 0.000213266 26.5913 0.593016 1.18687
MLM5 0.00000119209 0.148637 0.00331478 0.00663424
Series III
Minimum Maximum Mean Deviation
Absolute error
MLM1 0.00000119209 0.0695368 0.00936655 0.00818558
MLM2 0.00000104308 0.0120825 0.00145332 0.00118896
MLM3 0.000000119209 0.0879388 0.00273278 0.0041617
MLM4 0.00000119209 0.148637 0.00331478 0.00663424
MLM5 0.00000119209 0.148637 0.00331478 0.00663424
Relative error
MLM1 0.0000011413 0.0665743 0.00896749 0.00783684
MLM2 0.00000851071 0.0985832 0.011858 0.009701
MLM3 0.000000296304 0.218579 0.00679255 0.0103442
MLM4 0.00000213266 0.265913 0.00593016 0.0118687
MLM5 0.00000119209 0.148637 0.00331478 0.00663424
Percentage error
MLM1 0.00011413 6.65743 0.896749 0.783684
MLM2 0.000851071 9.85832 1.1858 0.9701
MLM3 0.0000296304 21.8579 0.679255 1.03442
MLM4 0.000213266 26.5913 0.593016 1.18687
MLM5 0.00000119209 0.148637 0.00331478 0.00663424
Materials2023,14,xFORPEERREVIEW28of38
(A)
(B)
(C)
Figure12.Thevaluesofmeanerror(absoluteerror(A),relativeerror(B),percentageerror(C))for
MLM1,MLM2,MLM3,MLM4,MLM5inseriesI,II,and
III.
Figure 12. Cont.
Materials 2023,16, 5956 27 of 36
Materials2023,14,xFORPEERREVIEW28of38
(A)
(B)
(C)
Figure12.Thevaluesofmeanerror(absoluteerror(A),relativeerror(B),percentageerror(C))for
MLM1,MLM2,MLM3,MLM4,MLM5inseriesI,II,and
III.
Figure 12.
The values of mean error (absolute error (
A
), relative error (
B
), percentage error (
C
)) for
MLM1
Materials2023,14,xFORPEERREVIEW28of38
(A)
(B)
(C)
Figure12.Thevaluesofmeanerror(absoluteerror(A),relativeerror(B),percentageerror(C))for
MLM1,MLM2,MLM3,MLM4,MLM5inseriesI,II,and
III.
, MLM2
Materials2023,14,xFORPEERREVIEW28of38
(A)
(B)
(C)
Figure12.Thevaluesofmeanerror(absoluteerror(A),relativeerror(B),percentageerror(C))for
MLM1,MLM2,MLM3,MLM4,MLM5inseriesI,II,and
III.
, MLM3
Materials2023,14,xFORPEERREVIEW28of38
(A)
(B)
(C)
Figure12.Thevaluesofmeanerror(absoluteerror(A),relativeerror(B),percentageerror(C))for
MLM1,MLM2,MLM3,MLM4,MLM5inseriesI,II,and
III.
, MLM4
Materials2023,14,xFORPEERREVIEW28of38
(A)
(B)
(C)
Figure12.Thevaluesofmeanerror(absoluteerror(A),relativeerror(B),percentageerror(C))for
MLM1,MLM2,MLM3,MLM4,MLM5inseriesI,II,and
III.
, MLM5
Materials2023,14,xFORPEERREVIEW28of38
(A)
(B)
(C)
Figure12.Thevaluesofmeanerror(absoluteerror(A),relativeerror(B),percentageerror(C))for
MLM1,MLM2,MLM3,MLM4,MLM5inseriesI,II,and
III.
in series I, II, and III.
Analysing Table 4and Figure 12A–C, one can observe that the average error decreases
with the increase in the model’s computational complexity for absolute error, relative error,
and percentage error. Figures 1315 represent the distribution of the relative error for the
target variable f
ck
. The abscissa represents the centres of the containers, and the ordinate
represents their corresponding frequencies. Error histograms show the distribution of
errors from the model for the test subset. A normal distribution for the target variable is
expected. For Figure 13A, the maximum frequency is 36.75%, which corresponds to the bin
centred at 0%. The minimum frequency is 0.2%, which corresponds to the bins with centres
at
48.277%. For Figure 13B, the maximum frequency is 39%, corresponding to the bin
centred at 0%. The minimum frequency is 0.12%, corresponding to the bins with centres at
47.614%. For Figure 13C, the maximum frequency is 38.76%, which corresponds to the
bin centred at 0%. The minimum frequency is 0.28%, which corresponds to the bins with
centres at 47.656%. For Figure 13D, the maximum frequency is 42.11%, corresponding to
the bin centred at 0%. The minimum frequency is 0.04%, corresponding to the bins with
centres at
53.646%. For Figure 13E, the maximum frequency is 42.37%, which corresponds
to the bin centred at 0%. The minimum frequency is 0.12%, which corresponds to the bins
with centres at 52.676%.
Materials 2023,16, 5956 28 of 36
Materials2023,14,xFORPEERREVIEW29of38
AnalysingTab l e 4andFigure12A–C,onecanobservethattheaverageerrorde-
creaseswiththeincreaseinthemodelscomputationalcomplexityforabsoluteerror,rel-
ativeerror,andpercentageerror.Figures13–15representthedistributionoftherelative
errorforthetargetvariablefck.Theabscissarepresentsthecentresofthecontainers,and
theordinaterepresentstheircorrespondingfrequencies.Errorhistogramsshowthedis-
tributionoferrorsfromthemodelforthetestsubset.Anormaldistributionforthetarget
variableisexpected.ForFigure13A,themaximumfrequencyis36.75%,whichcorre-
spondstothebincentredat0%.Theminimumfrequencyis0.2%,whichcorrespondsto
thebinswithcentresat−48.277%.ForFigure13B,themaximumfrequencyis39%,corre-
spondingtothebincentredat0%.Theminimumfrequencyis0.12%,correspondingto
thebinswithcentresat−47.614%.ForFigure13C,themaximumfrequencyis38.76%,
whichcorrespondstothebincentredat0%.Theminimumfrequencyis0.28%,whichcor-
respondstothebinswithcentresat47.656%.ForFigure13D,themaximumfrequencyis
42.11%,correspondingtothebincentredat0%.Theminimumfrequencyis0.04%,corre-
spondingtothebinswithcentresat−53.646%.ForFigure13E,themaximumfrequencyis
42.37%,whichcorrespondstothebincentredat0%.Theminimumfrequencyis0.12%,
whichcorrespondstothebinswithcentresat−52.676%.
(A)(B)
(C)(D)
Materials2023,14,xFORPEERREVIEW30of38
(E)
Figure13.ErrorhistogramfordevelopedmodelsinseriesI:MLM1(A),MLM2(B),MLM3(C),
MLM4(D),MLM5(E).
(A)(B)
(C)(D)
(E)
Figure 13.
Error histogram for developed models in series I: MLM1 (
A
), MLM2 (
B
), MLM3 (
C
),
MLM4 (D), MLM5 (E).
Materials2023,14,xFORPEERREVIEW30of38
(E)
Figure13.ErrorhistogramfordevelopedmodelsinseriesI:MLM1(A),MLM2(B),MLM3(C),
MLM4(D),MLM5(E).
(A)(B)
(C)(D)
(E)
Figure 14. Cont.
Materials 2023,16, 5956 29 of 36
Materials2023,14,xFORPEERREVIEW30of38
(E)
Figure13.ErrorhistogramfordevelopedmodelsinseriesI:MLM1(A),MLM2(B),MLM3(C),
MLM4(D),MLM5(E).
(A)(B)
(C)(D)
(E)
Figure 14.
Error histogram for developed models in series II: MLM1 (
A
), MLM2 (
B
), MLM3 (
C
),
MLM4 (D), MLM5 (E).
Materials2023,14,xFORPEERREVIEW31of38
Figure14.ErrorhistogramfordevelopedmodelsinseriesII:MLM1(A),MLM2(B),MLM3(C),
MLM4(D),MLM5(E).
(A)(B)
(C)(D)
(E)
Figure15.ErrorhistogramfordevelopedmodelsinseriesIII:MLM1(A),MLM2(B),MLM3(C),
MLM4(D),MLM5(E).
ForFigure14A,themaximumfrequencyis36.75%,whichcorrespondstothebin
centredat0%.Theminimumfrequencyis0.08%,whichcorrespondstothebinswithcen-
tresat−49.165%.ForFigure14B,themaximumfrequencyis41.3%,correspondingtothe
bincentredat0%.Theminimumfrequencyis0.04%,correspondingtothebinswithcen-
tresat−54.49%.ForFigure14C,themaximumfrequencyis38.93%,whichcorrespondsto
thebincentredat0%.Theminimumfrequencyis0.33%,whichcorrespondstothebins
withcentresat−44.438%.ForFigure14D,themaximumfrequencyis42.6%,correspond-
ingtothebincentredat0%.Theminimumfrequencyis0.12%,correspondingtothebins
Figure 15. Cont.
Materials 2023,16, 5956 30 of 36
Materials2023,14,xFORPEERREVIEW31of38
Figure14.ErrorhistogramfordevelopedmodelsinseriesII:MLM1(A),MLM2(B),MLM3(C),
MLM4(D),MLM5(E).
(A)(B)
(C)(D)
(E)
Figure15.ErrorhistogramfordevelopedmodelsinseriesIII:MLM1(A),MLM2(B),MLM3(C),
MLM4(D),MLM5(E).
ForFigure14A,themaximumfrequencyis36.75%,whichcorrespondstothebin
centredat0%.Theminimumfrequencyis0.08%,whichcorrespondstothebinswithcen-
tresat−49.165%.ForFigure14B,themaximumfrequencyis41.3%,correspondingtothe
bincentredat0%.Theminimumfrequencyis0.04%,correspondingtothebinswithcen-
tresat−54.49%.ForFigure14C,themaximumfrequencyis38.93%,whichcorrespondsto
thebincentredat0%.Theminimumfrequencyis0.33%,whichcorrespondstothebins
withcentresat−44.438%.ForFigure14D,themaximumfrequencyis42.6%,correspond-
ingtothebincentredat0%.Theminimumfrequencyis0.12%,correspondingtothebins
Figure 15.
Error histogram for developed models in series III: MLM1 (
A
), MLM2 (
B
), MLM3 (
C
),
MLM4 (D), MLM5 (E).
For Figure 14A, the maximum frequency is 36.75%, which corresponds to the bin
centred at 0%. The minimum frequency is 0.08%, which corresponds to the bins with
centres at
49.165%. For Figure 14B, the maximum frequency is 41.3%, corresponding to
the bin centred at 0%. The minimum frequency is 0.04%, corresponding to the bins with
centres at
54.49%. For Figure 14C, the maximum frequency is 38.93%, which corresponds
to the bin centred at 0%. The minimum frequency is 0.33%, which corresponds to the bins
with centres at
44.438%. For Figure 14D, the maximum frequency is 42.6%, corresponding
to the bin centred at 0%. The minimum frequency is 0.12%, corresponding to the bins with
centres at
54.26%. For Figure 14E, the maximum frequency is 39.65%, which corresponds
to the bin centred at 0%. The minimum frequency is 0.24%, which corresponds to the bins
with centres at 45.284%.
For Figure 15A, the maximum frequency is 35.93%, which corresponds to the bin
centred at 0%. The minimum frequency is 0.04%, which corresponds to the bins with
centres at
48.049%. For Figure 15B, the maximum frequency is 37.62%, corresponding
to the bin centred at 0%. The minimum frequency is 0.24%, corresponding to the bins
with centres at
45.596%. For Figure 15C, the maximum frequency is 41.64%, which
corresponds to the bin centred at 0%. The minimum frequency is 0.12%, which corresponds
to the bins with centres at
54.004%. For Figure 15D, the maximum frequency is 42.29%,
corresponding to the bin centred at 0%. The minimum frequency is 0.2%, corresponding to
the bins with centres at
50.63%. For Figure 15E, the maximum frequency is 36.95%, which
corresponds to the bin centred at 0%. The minimum frequency is 0.24%, which corresponds
to the bins with centres at 40.473%.
It should be noted that all error histograms presented in Figures 1315 have a bell
shape, which indicates a correctly obtained normal distribution [108,109]. The models are
designed to estimate concrete compressive strength as determined by the composition of
the concrete mixture. It is important to bear in mind that numerous additional elements,
primarily linked to the technological process of concrete production and environmental
conditions, exert influence over the concrete’s strength. The initial critical aspect involves
the proper maintenance of concrete after the setting process. Incorrect handling of concrete
can lead to a substantial deterioration its attributes, particularly its durability. Taking
environmental aspects into account, it is essential to consider the effect of environmental
aggression [
110
] that detrimentally affect concrete quality, such as frost action or exposure
large amounts of alkalis. One of pivotal concern is pouring concrete under unfavourable
weather conditions, particularly subjecting it to excessive shrinkage due to rapid drying
under high temperatures or freezing early in the setting process. The arrangement and
size distribution of the aggregates dictate the requirements for an appropriately workable
concrete paste as well as the source, shape, and texture of aggregates. It impact especially
the workability and durability of the concrete [
111
,
112
]. Admixtures and additives are also
important in whole process, especially those that profoundly affect the chemical properties
Materials 2023,16, 5956 31 of 36
of the mixture. This assessment has omitted several factors, including environmental
elements, technological processes, and properties of raw materials. It is assumed that the
quality of the produced concrete samples met acceptable standards. The source code of all
AI models presented in this research is available in an open repository [113].
4. Summary and Conclusions
In concrete structure production technology, a key challenge lies in ensuring pre-
dictable characteristics in the raw concrete mix and the hardened end product. Concrete
mix manufacturers are responsible for guaranteeing that the concrete they deliver to
construction sites meets the desired standards. However, achieving these standards consis-
tently throughout the manufacturing process can prove daunting. Reliable prediction of
concrete technical parameters is intricate, and most current solutions in the engineering
field are estimations, which have increasingly become obsolete due to rapid advancements
in material engineering. Predictive analytics dedicated to forecasting various phenomena,
attributes, and patterns based on machine learning holds the potential to enhance the
methodology behind concrete mix design substantially. The paper analyses the impact of
computational complexity on the effectiveness of predicting the compressive strength of
concrete using machine learning models. This study focuses on the growing interest in
applying machine learning algorithms in material engineering, specifically in predicting the
compressive strength of concrete, a key indicator of its quality. Computational complexity
in this research refers to the number of hidden layers in the neural network architecture.
In the context of machine learning models, computational complexity is essential, as it
can influence the speed and effectiveness of the model’s training and its ability to gener-
alise to new data. This study evaluated five deep neural network models (MLM1, MLM2,
MLM3, MLM4, MLM5) of varying computational complexity in three series. Each of the
MLM1-MLM5 models underwent training, selection, and testing in each three series. The
crux of this research was to establish an ideal deep neural network structure and train
it using a vast database of concrete mix recipes and their associated laboratory-based
destructive tests.
Presented machine learning models predicts the compressive strength of the concrete
mix based on its unique composition. Based on the obtained results, the following conclu-
sions can be formulated. There is a relationship between the computational complexity
of deep neural network models and their ability to predict the compressive strength of
concrete. The conducted analyses showed that as the computational complexity of the
model increases, so does its predictive ability. This means that the more complex the neural
network architecture, the more effective it is in predicting the compressive strength of
concrete within the conducted analyses. Several parameters point to the above conclusion.
In all three series, the coefficient of determination (R
2
) increase was observed along with the
increase in the model’s computational complexity. The smallest value of R
2
was captured
for model MLM1 and the largest for MLM5. It can be observed that errors in the five
analysed metrics, namely mean squared error (MSE), Minkowski error (ME), normalized
squared error (NSE), root mean squared error (RMSE), sum squared error (SSE), in training,
selection, and testing decrease with the increase in the model’s computational complexity,
with the greatest error decrease observable for the NSE metric in training, selection, and
testing and the ME metric in training. Furthermore, it is worth examining the mean values
of absolute error, relative error, and percentage error, which tend to decrease as the compu-
tational complexity of the model increases. The error histograms for all analysed models
in all series have a normal distribution. While the proposed method offers a promising
solution, it possesses certain limitations and does not comprehensively encapsulate all
the interactions between concrete mix components and their resulting properties. This is
a research gap that necessitates further exploration. Nonetheless, the outcomes of this paper
inspire optimism for this method’s expanded application in practical engineering settings.
Future investigations should aim to broaden this method’s utility in the concrete
mix design process by predicting additional fresh and hardened concrete properties such
Materials 2023,16, 5956 32 of 36
as durability, workability, air entrainment, and reliability. A more holistic strategy for
optimising concrete mix design is also essential to development. The findings from this
research lay a robust groundwork for further refinement and application of the proposed
AI-driven method in concrete mix design. AI-driven methods automate the traditionally
labour-intensive tasks, granting civil and structural engineers the liberty to tackle more
intricate and inventive challenges. It is anticipated that this method will improve both the
efficiency and precision of the design process, particularly in an era where advancements
in hardware are accelerating the computational capabilities of AI.
Funding:
The funding for these studies was obtained from the Faculty of Civil and Environmental
Engineering at Gdansk University of Technology (Gdansk Tech) through the Grants for Young
Scientists program.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: Not applicable.
Conflicts of Interest: The author declares no conflict of interest.
References
1.
Suchorzewski, J.; Prieto, M.; Mueller, U. An Experimental Study of Self-Sensing Concrete Enhanced with Multi-Wall Carbon
Nanotubes in Wedge Splitting Test and DIC. Constr. Build. Mater. 2020,262, 120871. [CrossRef]
2.
Nowek, A.; Kaszubski, P.; Abdelgader, H.S.; Górski, J. Effect of Admixtures on Fresh Grout and Two-Stage (Pre-Placed Aggregate)
Concrete. Struct. Concr. 2007,8, 17–23. [CrossRef]
3.
Kujawa, W.; Olewnik-Kruszkowska, E.; Nowaczyk, J. Concrete Strengthening by Introducing Polymer-Based Additives into the
Cement Matrix-a Mini Review. Materials 2021,14, 6071. [CrossRef]
4.
Suchorzewski, J.; Chitvoranund, N.; Srivastava, S.; Prieto, M.; Malaga, K. Recycling Potential of Cellular Lightweight Concrete
Insulation as Supplementary Cementitious Material. In Proceedings of the RILEM Bookseries; Springer: Berlin/Heidelberg, Germany,
2023; Volume 44, pp. 133–141.
5.
Liu, G.; Cheng, W.; Chen, L.; Pan, G.; Liu, Z. Rheological Properties of Fresh Concrete and Its Application on Shotcrete. Constr.
Build. Mater. 2020,243, 118180. [CrossRef]
6.
McNamee, R.; Sjöström, J.; Boström, L. Reduction of Fire Spalling of Concrete with Small Doses of Polypropylene Fibres. Fire
Mater. 2021,45, 943–951. [CrossRef]
7.
Cos-Gayón López, F.; Benlloch Marco, J.; Calvet Rodríguez, V. Influence of High Temperatures on the Bond between Carbon
Fibre-Reinforced Polymer Bars and Concrete. Constr. Build. Mater. 2021,309, 124967. [CrossRef]
8.
Gupta, S.; Kua, H.W.; Pang, S.D. Effect of Biochar on Mechanical and Permeability Properties of Concrete Exposed to Elevated
Temperature. Constr. Build. Mater. 2020,234, 117338. [CrossRef]
9. Marchon, D.; Flatt, R.J. Mechanisms of Cement Hydration. Sci. Technol. Concr. Admix. 2016,41, 129–145. [CrossRef]
10.
Liu, Y.; Kumar, D.; Lim, K.H.; Lai, Y.L.; Hu, Z.; Ambikakumari Sanalkumar, K.U.; Yang, E.H. Efficient Utilization of Municipal
Solid Waste Incinerator Bottom Ash for Autoclaved Aerated Concrete Formulation. J. Build. Eng. 2023,71, 106463. [CrossRef]
11.
Kocaba, V.; Gallucci, E.; Scrivener, K.L. Methods for Determination of Degree of Reaction of Slag in Blended Cement Pastes. Cem.
Concr. Res. 2012,42, 511–525. [CrossRef]
12.
Boinski, T.; Chojnowski, A. Towards Facts Extraction from Text in Polish Language. In Proceedings of the 2017 IEEE International
Conference on INnovations in Intelligent SysTems and Applications, INISTA 2017, Gdynia, Poland, 3–5 July 2017; IEEE:
Piscataway, NJ, USA, 2017; pp. 13–17.
13.
Pawlicki, M.; Marchewka, A.; Chora´s, M.; Kozik, R. Gated Recurrent Units for Intrusion Detection. In Proceedings of the Advances
in Intelligent Systems and Computing; Springer: Cham, Switzerland, 2020; Volume 1062, pp. 142–148.
14.
Renigier-Biłozor, M.; Janowski, A.; d’Amato, M. Automated Valuation Model Based on Fuzzy and Rough Set Theory for Real
Estate Market with Insufficient Source Data. Land Use Policy 2019,87, 104021. [CrossRef]
15.
Renigier-Biłozor, M.; Chmielewska, A.; Walacik, M.; Janowski, A.; Lepkova, N. Genetic Algorithm Application for Real Estate
Market Analysis in the Uncertainty Conditions. J. Hous. Built Environ. 2021,36, 1629–1670. [CrossRef]
16.
Renigier-Biłozor, M.; Janowski, A.; Walacik, M.; Chmielewska, A. Modern Challenges of Property Market Analysis-Homogeneous
Areas Determination. Land Use Policy 2022,119, 106209. [CrossRef]
17.
Chmielewska, A.; Renigier-Biłozor, M.; Janowski, A. Representative Residential Property Model—Soft Computing Solution. Int. J.
Environ. Res. Public Health 2022,19, 15114. [CrossRef]
18.
De Prado, R.P.; García-Galán, S.; Muñoz-Expósito, J.E.; Marchewka, A. Acceleration of Genome Sequencing with Intelligent
Cloud Brokers. In Proceedings of the Advances in Intelligent Systems and Computing; Springer: Cham, Switzerland, 2018; Volume 681,
pp. 133–140.
Materials 2023,16, 5956 33 of 36
19. Van Engelen, J.E.; Hoos, H.H. A Survey on Semi-Supervised Learning. Mach. Learn. 2020,109, 373–440.
20. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018; ISBN 0262352702.
21. Li, Y. Deep Reinforcement Learning: An Overview. arXiv 2017, arXiv:1701.07274.
22.
Ambroziak, A.; Ziolkowski, P. Concrete Compressive Strength under Changing Environmental Conditions during Placement
Processes. Materials 2020,13, 4577. [CrossRef] [PubMed]
23.
Tam, C.T.; Babu, D.S.; Li, W. EN 206 Conformity Testing for Concrete Strength in Compression. Procedia Eng.
2017
,171, 227–237.
[CrossRef]
24. EN 1992-1-1: 2004; Eurocode 2: Design of Concrete Structures. British Standards Institution: London, UK, 2004.
25.
DIN EN 206-1:2001-07; Beton–Teil 1: Festlegung, Eigenschaften, Herstellung Und Konformität; Deutsche Fassung EN 206-1:2000.
German Institute for Standardisation: Berlin, Germany, 2001.
26.
Abdelgader, H.S.; El-Baden, A.S.; Shilstone, J.M. Bolomeya Model for Normal Concrete Mix Design. J. Concr. Plant Int.
2012
,
2, 68–74.
27.
Zhang, C.; Nerella, V.N.; Krishna, A.; Wang, S.; Zhang, Y.; Mechtcherine, V.; Banthia, N. Mix Design Concepts for 3D Printable
Concrete: A Review. Cem. Concr. Compos. 2021,122, 104155. [CrossRef]
28.
Li, N.; Shi, C.; Zhang, Z.; Wang, H.; Liu, Y. A Review on Mixture Design Methods for Geopolymer Concrete. Compos. Part B Eng.
2019,178, 107490. [CrossRef]
29.
Liu, Q.F.; Iqbal, M.F.; Yang, J.; Lu, X.Y.; Zhang, P.; Rauf, M. Prediction of Chloride Diffusivity in Concrete Using Artificial Neural
Network: Modelling and Performance Evaluation. Constr. Build. Mater. 2021,268, 121082. [CrossRef]
30.
Iqbal, M.F.; Liu, Q.F.; Azim, I.; Zhu, X.; Yang, J.; Javed, M.F.; Rauf, M. Prediction of Mechanical Properties of Green Concrete
Incorporating Waste Foundry Sand Based on Gene Expression Programming. J. Hazard. Mater.
2020
,384, 121322. [CrossRef]
[PubMed]
31.
Yeh, I.C. Modeling of Strength of High-Performance Concrete Using Artificial Neural Networks. Cem. Concr. Res.
1998
,
28, 1797–1808. [CrossRef]
32. Lee, S.C. Prediction of Concrete Strength Using Artificial Neural Networks. Eng. Struct. 2003,25, 849–857. [CrossRef]
33.
Hola, J.; Schabowicz, K. Application of Artificial Neural Networks to Determine Concrete Compressive Strength Based on
Non-Destructive Tests. J. Civ. Eng. Manag. 2005,11, 23–32. [CrossRef]
34.
Hola, J.; Schabowicz, K. New Technique of Nondestructive Assessment of Concrete Strength Using Artificial Intelligence. NDT E
Int. 2005,38, 251–259. [CrossRef]
35.
Gupta, R.; Kewalramani, M.A.; Goel, A. Prediction of Concrete Strength Using Neural-Expert System. J. Mater. Civ. Eng.
2006
,
18, 462–466. [CrossRef]
36.
Bui, D.K.; Nguyen, T.; Chou, J.S.; Nguyen-Xuan, H.; Ngo, T.D. A Modified Firefly Algorithm-Artificial Neural Network Expert
System for Predicting Compressive and Tensile Strength of High-Performance Concrete. Constr. Build. Mater.
2018
,180, 320–333.
[CrossRef]
37.
Deng, F.; He, Y.; Zhou, S.; Yu, Y.; Cheng, H.; Wu, X. Compressive Strength Prediction of Recycled Concrete Based on Deep
Learning. Constr. Build. Mater. 2018,175, 562–569. [CrossRef]
38.
Naderpour, H.; Rafiean, A.H.; Fakharian, P. Compressive Strength Prediction of Environmentally Friendly Concrete Using
Artificial Neural Networks. J. Build. Eng. 2018,16, 213–219. [CrossRef]
39.
Ziolkowski, P.; Niedostatkiewicz, M. Machine Learning Techniques in Concrete Mix Design. Materials
2019
,12, 1256. [CrossRef]
[PubMed]
40. McCormac, J.C.; Brown, R.H. Design of Reinforced Concrete; John Wiley & Sons: Hoboken, NJ, USA, 2015; ISBN 1118879104.
41.
Nunez, I.; Marani, A.; Nehdi, M.L. Mixture Optimization of Recycled Aggregate Concrete Using Hybrid Machine Learning
Model. Materials 2020,13, 4331. [CrossRef] [PubMed]
42.
Marani, A.; Jamali, A.; Nehdi, M.L. Predicting Ultra-High-Performance Concrete Compressive Strength Using Tabular Generative
Adversarial Networks. Materials 2020,13, 4757. [CrossRef] [PubMed]
43.
Ziolkowski, P.; Niedostatkiewicz, M.; Kang, S.B. Model-Based Adaptive Machine Learning Approach in Concrete Mix Design.
Materials 2021,14, 1661. [CrossRef]
44.
Adil, M.; Ullah, R.; Noor, S.; Gohar, N. Effect of Number of Neurons and Layers in an Artificial Neural Network for Generalized
Concrete Mix Design. Neural Comput. Appl. 2022,34, 8355–8363. [CrossRef]
45.
Feng, W.; Wang, Y.; Sun, J.; Tang, Y.; Wu, D.; Jiang, Z.; Wang, J.; Wang, X. Prediction of Thermo-Mechanical Properties of
Rubber-Modified Recycled Aggregate Concrete. Constr. Build. Mater. 2022,318, 125970. [CrossRef]
46.
Tavares, C.; Wang, X.; Saha, S.; Grasley, Z. Machine Learning-Based Mix Design Tools to Minimize Carbon Footprint and Cost of
UHPC. Part 1: Efficient Data Collection and Modeling. Clean. Mater. 2022,4, 100082. [CrossRef]
47.
Tavares, C.; Grasley, Z. Machine Learning-Based Mix Design Tools to Minimize Carbon Footprint and Cost of UHPC. Part 2: Cost
and Eco-Efficiency Density Diagrams. Clean. Mater. 2022,4, 100094. [CrossRef]
48.
Endzhievskaya, I.G.; Endzhievskiy, A.S.; Galkin, M.A.; Molokeev, M.S. Machine Learning Methods in Assessing the Effect of
Mixture Composition on the Physical and Mechanical Characteristics of Road Concrete. J. Build. Eng.
2023
,76, 107248. [CrossRef]
49.
Taffese, W.Z.; Espinosa-Leal, L. Multitarget Regression Models for Predicting Compressive Strength and Chloride Resistance of
Concrete. J. Build. Eng. 2023,72, 106523. [CrossRef]
Materials 2023,16, 5956 34 of 36
50.
Gulli, A.; Pal, S. Deep Learning with Keras: Beginners Guide to Deep Learning with Keras; Packt Publishing Ltd.: Birmingham, UK,
2017; ISBN 9781787128422.
51. Cichy, R.M.; Kaiser, D. Deep Neural Networks as Scientific Models. Trends Cogn. Sci. 2019,23, 305–317. [CrossRef]
52. Saxena, A. An Introduction to Convolutional Neural Networks. Int. J. Res. Appl. Sci. Eng. Technol. 2022,10, 943–947. [CrossRef]
53.
Li, Z.; Liu, F.; Yang, W.; Peng, S.; Zhou, J. A Survey of Convolutional Neural Networks: Analysis, Applications, and Prospects.
IEEE Trans. Neural Netw. Learn. Syst. 2022,33, 6999–7019. [CrossRef]
54.
Zhang, Q.; Zhang, M.; Chen, T.; Sun, Z.; Ma, Y.; Yu, B. Recent Advances in Convolutional Neural Network Acceleration.
Neurocomputing 2019,323, 37–51. [CrossRef]
55.
Li, H.; Li, J.; Guan, X.; Liang, B.; Lai, Y.; Luo, X. Research on Overfitting of Deep Learning. In Proceedings of the 2019 15th
International Conference on Computational Intelligence and Security, CIS 2019, Macau, China, 13–16 December 2019; IEEE:
Piscataway, NJ, USA, 2019; pp. 78–81.
56. Salman, S.; Liu, X. Overfitting Mechanism and Avoidance in Deep Neural Networks. arXiv 2019, arXiv:1901.06566.
57.
Bejani, M.M.; Ghatee, M. A Systematic Review on Overfitting Control in Shallow and Deep Neural Networks. Artif. Intell. Rev.
2021,54, 6391–6438. [CrossRef]
58.
Liu, M.; Chen, L.; Du, X.; Jin, L.; Shang, M. Activated Gradients for Deep Neural Networks. IEEE Trans. Neural Netw. Learn. Syst.
2023,34, 2156–2168. [CrossRef]
59.
Rehmer, A.; Kroll, A. On the Vanishing and Exploding Gradient Problem in Gated Recurrent Units. IFAC-PapersOnLine
2020
,
53, 1243–1248. [CrossRef]
60.
Garbin, C.; Zhu, X.; Marques, O. Dropout vs. Batch Normalization: An Empirical Study of Their Impact to Deep Learning.
Multimed. Tools Appl. 2020,79, 12777–12815. [CrossRef]
61.
Salehinejad, H.; Valaee, S. Ising-Dropout: A Regularization Method for Training and Compression of Deep Neural Networks. In
Proceedings of the ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing-Proceedings, Brighton,
UK, 12–17 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 3602–3606.
62.
Piotrowski, A.P.; Napiorkowski, J.J.; Piotrowska, A.E. Impact of Deep Learning-Based Dropout on Shallow Neural Networks
Applied to Stream Temperature Modelling. Earth-Sci. Rev. 2020,201, 103076. [CrossRef]
63.
Tjoa, E.; Cuntai, G. Quantifying Explainability of Saliency Methods in Deep Neural Networks With a Synthetic Dataset. IEEE
Trans. Artif. Intell. 2022,4, 858–870. [CrossRef]
64.
Hernandez, M.; Epelde, G.; Alberdi, A.; Cilla, R.; Rankin, D. Synthetic Data Generation for Tabular Health Records: A Systematic
Review. Neurocomputing 2022,493, 28–45. [CrossRef]
65.
Juneja, T.; Bajaj, S.B.; Sethi, N. Synthetic Time Series Data Generation Using Time GAN with Synthetic and Real-Time Data
Analysis. In Proceedings of the Lecture Notes in Electrical Engineering; Springer: Berlin/Heidelberg, Germany, 2023; Volume 1011
LNEE, pp. 657–667.
66.
Ravikumar, D.; Zhang, D.; Keoleian, G.; Miller, S.; Sick, V.; Li, V. Carbon Dioxide Utilization in Concrete Curing or Mixing Might
Not Produce a Net Climate Benefit. Nat. Commun. 2021,12, 855. [CrossRef]
67.
Shi, J.; Liu, B.; Wu, X.; Qin, J.; Jiang, J.; He, Z. Evolution of Mechanical Properties and Permeability of Concrete during Steam
Curing Process. J. Build. Eng. 2020,32, 101796. [CrossRef]
68.
Li, Y.; Nie, L.; Wang, B. A Numerical Simulation of the Temperature Cracking Propagation Process When Pouring Mass Concrete.
Autom. Constr. 2014,37, 203–210. [CrossRef]
69.
Patel, S.K.; Parmar, J.; Katkar, V. Graphene-Based Multilayer Metasurface Solar Absorber with Parameter Optimization and
Behavior Prediction Using Long Short-Term Memory Model. Renew. Energy 2022,191, 47–58. [CrossRef]
70.
Zhou, X.; Lin, W.; Kumar, R.; Cui, P.; Ma, Z. A Data-Driven Strategy Using Long Short Term Memory Models and Reinforcement
Learning to Predict Building Electricity Consumption. Appl. Energy 2022,306, 118078. [CrossRef]
71.
Wold, S. Principal Component Analysis Why Principal Component Analysis? In IEEE Signal Processing Letters; Elsevier: Amster-
dam, The Netherlands, 2002; Volume 9, pp. 40–42; ISBN 0-387-95442-2.
72.
Vidal, R.; Ma, Y.; Sastry, S. Principal Component Analysis. In Interdisciplinary Applied Mathematics; Springer: New York, NY, USA,
2016; Volume 40, pp. 25–62. [CrossRef]
73.
Thiyagalingam, J.; Shankar, M.; Fox, G.; Hey, T. Scientific Machine Learning Benchmarks. Nat. Rev. Phys.
2022
,4, 413–420.
[CrossRef]
74.
Menéndez, M.L.; Pardo, J.A.; Pardo, L.; Pardo, M.C. The Jensen-Shannon Divergence. J. Frankl. Inst.
1997
,334, 307–318. [CrossRef]
75.
Fuglede, B.; Topsoe, F. Jensen-Shannon Divergence and Hilbert Space Embedding. In Proceedings of the International symposium
on Information theory, 2004. ISIT 2004. Proceedings, Chicago, IL, USA, 27 June–2 July 2004; IEEE: Piscataway, NJ, USA, 2004;
p. 31.
76.
Nielsen, F. On a Generalization of the Jensen-Shannon Divergence and the Jensen-Shannon Centroid. Entropy
2020
,22, 221.
[CrossRef] [PubMed]
77.
Toniolo, G.; Di Prisco, M. Reinforced Concrete Design to Eurocode 2; Springer Tracts in Civil Engineering; Springer International
Publishing: Berlin/Heidelberg, Germany, 2017; pp. 1–836.
78. Keim, D.A.; Hao, M.C.; Dayal, U.; Janetzko, H.; Bak, P. Generalized Scatter Plots. Inf. Vis. 2010,9, 301–311. [CrossRef]
79.
Larochelle, H.; Bengio, Y.; Louradour, J.; Lamblin, P. Exploring Strategies for Training Deep Neural Networks. J. Mach. Learn. Res.
2009,10, 1–40.
Materials 2023,16, 5956 35 of 36
80.
Zhu, H.; Akrout, M.; Zheng, B.; Pelegris, A.; Jayarajan, A.; Phanishayee, A.; Schroeder, B.; Pekhimenko, G. Benchmarking and An-
alyzing Deep Neural Network Training. In Proceedings of the 2018 IEEE International Symposium on Workload Characterization,
IISWC 2018, Raleigh, NC, USA, 30 September–2 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 88–100.
81.
Banerjee, S.; Iglewicz, B. A Simple Univariate Outlier Identification Procedure Designed for Large Samples. Commun. Stat. Simul.
Comput. 2007,36, 249–263. [CrossRef]
82.
Seo, S.; Gary, M.; Marsh, P.D. A Review and Comparison of Methods for Detecting Outliersin Univariate Data Sets. Dep. Biostat.
Grad. Sch. Public Health 2006, 1–53.
83.
Wan, X. Influence of Feature Scaling on Convergence of Gradient Iterative Algorithm. In Proceedings of the Journal of Physics:
Conference Series; IOP Publishing: Bristol, UK, 2019; Volume 1213, p. 32021.
84.
Zheng, A.; Casari, A. Feature Engineering for Machine Learning: Principles and Techniques for Data Scientists; O’Reilly Media, Inc.:
Sevastopol, CA, USA, 2018; ISBN 978-1491953242.
85.
Namin, A.H.; Leboeuf, K.; Muscedere, R.; Wu, H.; Ahmadi, M. Efficient Hardware Implementation of the Hyperbolic Tangent
Sigmoid Function. In Proceedings of the IEEE International Symposium on Circuits and Systems, Taipei, Taiwan, 24–27 May 2009;
IEEE: Piscataway, NJ, USA, 2009; pp. 2117–2120.
86.
Zamanlooy, B.; Mirhassani, M. Efficient VLSI Implementation of Neural Networks with Hyperbolic Tangent Activation Function.
IEEE Trans. Very Large Scale Integr. Syst. 2013,22, 39–48. [CrossRef]
87. Rasamoelina, A.D.; Adjailia, F.; Sincak, P. A Review of Activation Function for Artificial Neural Network. In Proceedings of the
SAMI 2020-IEEE 18th World Symposium on Applied Machine Intelligence and Informatics, Herl’any, Slovakia, 23–25 January
2020; IEEE: Piscataway, NJ, USA, 2020; pp. 281–286.
88. Van Laarhoven, T. L2 Regularization versus Batch and Weight Normalization. arXiv 2017, arXiv:1706.05350.
89.
Cortes, C.; Mohri, M.; Rostamizadeh, A. L2 Regularization for Learning Kernels. In Proceedings of the 25th Conference on
Uncertainty in Artificial Intelligence UAI 2009, Montreal, QC, Canada, 18–21 June 2009; pp. 109–116.
90.
Moore, R.C.; DeNero, J. L1 and L2 Regularization for Multiclass Hinge Loss Models. In Proceedings of the Symposium on
Machine Learning in Speech and Natural Language Processing, Bellevue, WA, USA, 27 June 2011.
91.
Goldfarb, D.; Ren, Y.; Bahamou, A. Practical Quasi-Newton Methods for Training Deep Neural Networks. Adv. Neural Inf. Process.
Syst. 2020,33, 2386–2396.
92.
Byrd, R.H.; Hansen, S.L.; Nocedal, J.; Singer, Y. A Stochastic Quasi-Newton Method for Large-Scale Optimization. SIAM J. Optim.
2016,26, 1008–1031. [CrossRef]
93.
Nilsen, G.K.; Munthe-Kaas, A.Z.; Skaug, H.J.; Brun, M. Efficient Computation of Hessian Matrices in TensorFlow. arXiv
2019
,
arXiv:1905.05559.
94.
Keskar, N.; Wächter, A. A Limited-Memory Quasi-Newton Algorithm for Bound-Constrained Non-Smooth Optimization. Optim.
Methods Softw. 2019,34, 150–171. [CrossRef]
95.
Ayanlere, S.A.; Ajamu, S.O.; Odeyemi, S.O.; Ajayi, O.E.; Kareem, M.A. Effects of Water-Cement Ratio on Bond Strength of
Concrete. Mater. Today Proc. 2023,86, 134–139. [CrossRef]
96.
Almusallam, A.A.; Maslehuddin, M.; Waris, M.A.; Al-Amoudi, O.S.B.; Al-Gahtani, A.S. Plastic Shrinkage Cracking of Concrete in
Hot-Arid Environments. Arab. J. Sci. Eng. 1998,23, 57–71.
97. Wilkinson, L.; Friendly, M. History Corner the History of the Cluster Heat Map. Am. Stat. 2009,63, 179–184. [CrossRef]
98.
Argiz, C.; Menéndez, E.; Sanjuán, M.A. Efecto de La Adición de Mezclas de Ceniza Volante y Ceniza de Fondo Procedentes Del
Carbón En La Resistencia Mecánica y Porosidad de Cementos Portland. Mater. Constr. 2013,63, 49–64. [CrossRef]
99.
Berk, R.H.; Jones, D.H. Goodness-of-Fit Test Statistics That Dominate the Kolmogorov Statistics. Z. Wahrscheinlichkeitstheorie
Verwandte Geb. 1979,47, 47–59. [CrossRef]
100.
Larntz, K.; Read, T.R.C.; Cressie, N.A.C. Goodness-of-Fit Statistics for Discrete Multivariate Data; Springer Science & Business Media:
Berlin/Heidelberg, Germany, 1989; Volume 84; ISBN 1461245788.
101. Mckinley, R.L.; Mills, C.N. A Comparison of Several Goodness-of-Fit Statistics. Appl. Psychol. Meas. 1985,9, 49–57. [CrossRef]
102.
Lospinoso, J.; Snijders, T.A.B. Goodness of Fit for Stochastic Actor-Oriented Models. Methodol. Innov.
2019
,12, 2059799119884282.
[CrossRef]
103.
Baum, J.; Kanagawa, H.; Gretton, A. A Kernel Stein Test of Goodness of Fit for Sequential Models. In Proceedings of the
International Conference on Machine Learning, PMLR, Honolulu, HI, USA, 23–29 July 2023; pp. 1936–1953.
104.
Piepho, H.P. A Coefficient of Determination (R2) for Generalized Linear Mixed Models. Biom. J.
2019
,61, 860–872. [CrossRef]
[PubMed]
105.
Camirand Lemyre, F.; Chalifoux, K.; Desharnais, B.; Mireault, P. Squaring Things Up with R2: What It Is and What It Can (and
Cannot) Tell You. J. Anal. Toxicol. 2022,46, 443–448. [CrossRef] [PubMed]
106.
Botchkarev, A. A New Typology Design of Performance Metrics to Measure Errors in Machine Learning Regression Algorithms.
Interdiscip. J. Inf. Knowl. Manag. 2019,14, 45–76. [CrossRef]
107.
Naser, M.Z.; Alavi, A.H. Error Metrics and Performance Fitness Indicators for Artificial Intelligence and Machine Learning in
Engineering and Sciences. Archit. Struct. Constr. 2021, 1–19. [CrossRef]
108.
Michael, W.; Berry, A.M. Supervised and Unsupervised Learning for Data Science; Springer: Berlin/Heidelberg, Germany, 2020; ISBN
978-3-030-22474-5.
109. Kubat, M. An Introduction to Machine Learning; Springer: Berlin/Heidelberg, Germany, 2021; ISBN 9783030819354.
Materials 2023,16, 5956 36 of 36
110.
Sanjuán, M.Á.; Estévez, E.; Argiz, C.; del Barrio, D. Effect of Curing Time on Granulated Blast-Furnace Slag Cement Mortars
Carbonation. Cem. Concr. Compos. 2018,90, 257–265. [CrossRef]
111.
Poloju, K.K.; Anil, V.; Manchiryal, R.K. Properties of Concrete as Influenced by Shape and Texture of Fine Aggregate. Am. J. Appl.
Sci. Res. 2017,3, 28–36.
112.
Chinchillas-Chinchillas, M.J.; Corral-Higuera, R.; Gómez-Soberón, J.M.; Arredondo-Rea, S.P.; Alamaral-Sánchez, J.L.; Acuña-
Aguero, O.H.; Rosas-Casarez, C.A. Influence of the Shape of the Natural Aggregates, Recycled and Silica Fume on the Mechanical
Properties of Pervious Concrete. Int. J. Adv. Comput. Sci. Its Appl. 2014,4, 216–220.
113.
Ziolkowski, P. Source Code AI Models (MLM1-5-series I–III-QNM opt); Gdansk University of Technology: Gda ´nsk, Poland, 2023.
[CrossRef]
Disclaimer/Publisher’s Note:
The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.
... Both studies demonstrated that by integrating optimization methods with ML, it is possible to design more cost-effective and durable concrete mixtures. Ziolkowski P.'s 2023 study [54] explores the relationship between computational complexity and predictive accuracy in ML models applied to concrete mix design. The research evaluates five deep neural network models of varying complexity, trained on a dataset comprising concrete mix compositions and the corresponding compressive strength test results. ...
... ML models have demonstrated considerable effectiveness in predicting the technical properties of concrete, particularly compressive strength, based on the composition of the mix. This study aims to build on previous research [54] by not only examining the impact of computational complexity on the predictive performance of ML models but also investigating the influence of various optimization algorithms. Specifically, this research evaluates the performance of the Adaptive Moment Estimation (ADAM) algorithm and Stochastic Gradient Descent (SGD), alongside the previously utilized Quasi-Newton Method (QNM) [63][64][65]. ...
... The analysis aimed to understand how these factors interact to influence the predictive accuracy of compressive strength. The data used in this study were drawn from previous work [45,48,54] and include a comprehensive database of concrete mix recipes, along with results from corresponding compressive strength tests. This dataset, initially composed of several hundred records, was further expanded using a dedicated AI model to generate reliable synthetic data [70][71][72]. ...
Article
Full-text available
The proper design of concrete mixtures is a critical task in concrete technology, where optimal strength, eco-friendliness, and production efficiency are increasingly demanded. While traditional analytical methods, such as the Three Equations Method, offer foundational approaches to mix design, they often fall short in handling the complexity of modern concrete technology. Machine learning-based models have demonstrated notable efficacy in predicting concrete compressive strength, addressing the limitations of conventional methods. This study builds on previous research by investigating not only the impact of computational complexity on the predictive performance of machine learning models but also the influence of different optimization algorithms. The study evaluates the effectiveness of three optimization techniques: the Quasi-Newton Method (QNM), the Adaptive Moment Estimation (ADAM) algorithm, and Stochastic Gradient Descent (SGD). A total of forty-five deep neural network models of varying computational complexity were trained and tested using a comprehensive database of concrete mix designs and their corresponding compressive strength test results. The findings reveal a significant interaction between optimization algorithms and model complexity in enhancing prediction accuracy. Models utilizing the QNM algorithm outperformed those using the ADAM and SGD in terms of error reduction (SSE, MSE, RMSE, NSE, and ME) and increased coefficient of determination (R²). These insights contribute to the development of more accurate and efficient AI-driven methods in concrete mix design, promoting the advancement of concrete technology and the potential for future research in this domain.
... This method utilized deep neural networks and was trained on a comprehensive database of concrete mix recipes, allowing for more precise strength prediction compared to earlier models. In 2023 he studied how computational complexity of deep neural networks affect the predictive capability of the models 11 . Chen et al. 56 proposed a convolution-based deep learning approach for predicting the compressive strength of fiber-reinforced concrete (FRC) exposed to elevated temperatures. ...
... The chosen architecture must be capable of capturing the complex relationships between temperature data and concrete strength. Based on experience from previous research [9][10][11] and analyses, it was decided that a deep neural network (DNN) would be best for this task. This network should consist of an input layer representing the extracted features, one or more hidden layers utilizing nonlinear activation functions to capture complex relationships within the data, and an output layer designed to provide continuous predictions of compressive strength. ...
Article
Full-text available
The continuous evolution of construction technologies, particularly in reinforced concrete production, demands advanced, reliable, and efficient methodologies for real-time monitoring and prediction of concrete compressive strength. Traditional laboratory methods for assessing compressive strength are time-intensive and can introduce delays in construction workflows. This study introduces a comprehensive framework for a system designed to predict early-age compressive strength of concrete through continuous monitoring of the cement hydration process using a custom artificial intelligence (AI) model. The system integrates a network of temperature sensors, communication modules, and a centralized database server to collect, transmit, and analyze real-time data during the concrete curing process. The AI model, a deep neural network leverages this data to generate accurate strength predictions. The system architecture emphasizes scalability, robustness, and integration with existing construction management systems. Empirical results indicate that the proposed system achieves high predictive accuracy, with an R² value of 0.996 and RMSE of 0.143 MPa, offering a robust tool for real-time decision-making in construction. This study also critically evaluates the system's performance, identifying key strengths such as predictive accuracy and real-time processing capabilities, and addresses challenges related to wireless communication reliability and sensor power supply. Recommendations are provided for enhancing system precision, improving communication technologies, optimizing power management, and ensuring scalability across diverse construction contexts. The developed system, which is part of the "CONCRESENSE" project and protected under European patent number 245107 (2024), represents a significant advancement in construction technology, with substantial implications for enhancing the safety, efficiency, and quality of reinforced concrete structures.
... While these approaches are driven mainly by human intuition, the margin of embodied carbon reduction is only affected by concrete volume reduction while the material composition remains the same. Considering that embodied carbon is determined by the carbon factor of the concrete ingredient, optimizing the mix composition to reduce embodied carbon that is devoid of substantial errors and bias will enhance significant reduction in embodied carbon with the use machine learning and artificial intelligence [7]. Since the effectiveness of prediction using machine learning is predicated on the viability of dataset [8-10], the use of experimental dataset as against empirical portends better reliability [11]. ...
Article
Full-text available
The demand for sustainable concrete in meeting the net zero carbon target places a burden in optimizing concrete response to structural strength that satisfy acceptable embodied carbon. In most cases, a low carbon concrete is deficient in structural requirement and vice versa. This dilemma informs the need for a tool that can predict compressive strength as well as embodied carbon using the same input data. Since the use of alternative materials as cement replacement to enhance sustainability is emerging in the quest for a sustainable concrete, an optimal material that satisfy both conditions of structural integrity and sustainability is still lacking. Paucity of data in the emerging lightweight low carbon concrete using alternative materials, portends an upheave in the bias for prediction of the behaviour of lightweight low carbon concrete. This study therefore uses concrete data of lightweight low carbon concrete from laboratory experiment for the prediction of compressive strength and embodied carbon with their performance evaluated using eight machine leaning regression models. The results obtained indicates that the XG boost regression model exhibited excellent performance with a low Mean Squared Error (MSE) of 50.15, Mean absolute error(MAE) = 5.26, Mean absolute percentage error(MAPE) = 11.76 %, Explained variance score = 0.97, Root mean square error(RMSE) = 7.08 and a high R squared value of 0.96. The tool predicted compressive strength and embodied carbon for lightweight carbon concrete using multiple output regression such that the output can be limited to the yearly structural embodied carbon threshold to achieving 2050 net zero target. The developed tool when compared with concrete of similar mix ingredients performed more than 95 % in predicting concrete compressive strength and associated embodied carbon. In line with the inclusion of embodied carbon in carbon regulations of buildings in the UK as suggested by the professionals in the construction industry, the developed learner model and prediction tool has integrated concrete strength and embodied carbon to initiate a holistic approach to design and construction, balancing performance, cost, and environmental impact.
... Taking advantage of the importance of visual data associated with item recommendation, Hiriyannaiah et al. [46] developed a deep learning-based visual recommendation system using a convolutional autoencoder neural network for classification. However, the higher computational complexities of deep learning algorithms [47] make them difficult to implement. To improve tourism recommendation, Malika et al. [48] proposed a long short-term memory-based method for point of interest recommendation. ...
Article
Full-text available
Object ratings in recommendation algorithms are used to represent the extent to which a user likes an object. Most existing recommender systems use these ratings to recommend the top-K objects to a target user. To improve the accuracy and diversity of recommender systems, we proposed a neighbourhood-based diffusion recommendation algorithm (NBD) that distributes the resources among objects using the rating scores of the objects based on the likings of the target user neighbours. Specifically, the Adamic–Adar similarity index is used to calculate the similarity between the target user and other users to select the top K similar neighbours to begin the diffusion process. In this approach, greater significance is put on common neighbours with fewer neighbour nodes. This is to reduce the effect of popular objects. At the end of the diffusion process, a modified redistribution algorithm using the sigmoid function is explored to finally redistribute the resources to the objects. This is to ensure that the objects recommended are personalized to target users. The evaluation has been conducted through experiments using four real-world datasets (Friendfeed, Epinions, MovieLens-100 K, and Netflix). The experiment results show that the performance of our proposed NBD algorithm is better in terms of accuracy when compared with the state-of-the-art algorithms.
... In the field of concrete performance, most of the existing studies conducted using ML and DL algorithms are limited to predicting mechanical and durable performances based on specific aged days and mix design [22][23][24]28]. However, concrete is a timedependent building material whose performance improves with aged days [34]. ...
Article
Full-text available
In this study, accelerated chloride diffusion tests are performed on ordinary Portland cement (OPC), ground granulated blast furnace slag (GGBFS), and fly ash (FA) concretes aged 4–6 years. Passed charge is evaluated according to ASTM-C-1202 for 12 mixtures, considering water–binder (W/B) ratios (0.37, 0.42, and 0.47), GGBFS replacement rates (0%, 30%, 50%), and FA replacement rates (0% and 30%). The effects of aged days on passed charge reduction behavior are quantified through repetitive regression analysis. Among existing machine learning (ML) models, linear, lasso, and ridge models are used to analyze the correlation of aged days and mix properties with passed charge. Passed charge analysis considering long-term age shows a significant variability decrease of passed charge by W/B ratio with increasing age and added admixtures (GGBFS and FA). Furthermore, the higher the water–binder ratio in GGBFS and FA concretes, the greater the decrease in passed charge due to aged days. The ML model-based regression analysis shows high correlation when compressive strength and independent variables are considered together. Future work includes a correlational analysis between mixture properties and chloride ingress durability performance using deep learning models based on the time series properties of evaluation data.
... What makes the proposed "a20index" valuable is its meaningful interpretation in the field of engineering, as it quantifies the percentage of samples for which the predicted values closely align with the experimental values, with a tolerance of ±20 % [15,32]. Moreover, the important analysis of input variables on the output is a critical step in predictive modeling, as it provides valuable insights into which factors exert the most significant influence on the target variable [58]. This understanding is instrumental in various applications, including optimizing processes, enhancing model interpretability, and guiding decision-making. ...
Article
Full-text available
This study addresses a critical gap in concrete strength prediction by conducting a comparative analysis of three deep learning algorithms: convolutional neural networks (CNNs), gated recurrent units (GRUs), and long short-term memory (LSTM) networks. Unlike previous studies that employed various machine learning algorithms on diverse concrete types, our study focuses on mixed-design concrete and fine-tuned deep learning algorithms. The objective is to identify the optimal deep learning (DL) algorithm for predicting concrete uniaxial compressive strength, a crucial parameter in construction and structural engineering. The dataset comprises experimental records for mixed-design concrete, and models were developed and optimized for predictive accuracy. The results show that the CNN model consistently outperformed GRU and LSTM. Hyperparameter tuning and regularization techniques further improved model performance. This research offers practical solutions for material property prediction in the construction industry, potentially reducing resource burdens and enhancing efficiency and construction quality.
Article
Full-text available
This work develops single multitarget regression models that predict compressive strength and non-steady-state chloride migration coefficients (Dnssm) of concrete simultaneously using machine learning algorithms. The data for this study are obtained from research projects and internationally published articles. Following data preprocessing, the compression strength ranged from 21 to 80 MPa, while the Dnssm ranged from 1.57 to 31.30 x10-12 m 2 /s. The algorithms used are five decision tree-based ensemble methods: bagging, random forest, AdaBoost, Gradient boosting, and XGBoost. In the development of the models, two scenarios are considered. Scenario 1 employs the default hyperparameter settings, while Scenario 2 employs hyperparameters chosen from among those identified through training single-target models. The performance evaluation results confirm that Gradient boosting is the best performing algorithm and Scenario 2 is the most appropriate modeling strategy for the considered dataset. It predicts compression strength with (MAE = 6.683, MSE = 83.369, and RMSE = 9.131) and Dnssm with (MAE = 1.363, MSE = 3.712, and RMSE = 1.927). The potential of the developed multitarget model to design concrete with the intended strength and Dnssm is supported by its remarkable generalization ability. However, in order to ensure the model's versatility, it is necessary to improve it by incorporating comprehensive datasets that include a broad range of concrete properties.
Article
Full-text available
Post-hoc analysis is a popular category in eXplainable artificial intelligence (XAI) study. In particular, methods that generate heatmaps have been used to explain the deep neural network (DNN), a black-box model. Heatmaps can be appealing due to the intuitive and visual ways to understand them but assessing their qualities might not be straightforward. Different ways to assess heatmaps' quality have their own merits and shortcomings. This paper introduces a synthetic dataset that can be generated adhoc along with the ground-truth heatmaps for more objective quantitative assessment. Each sample data is an image of a cell with easily recognized features that are distinguished from localization ground-truth mask, hence facilitating a more transparent assessment of different XAI methods. Comparison and recommendations are made, shortcomings are clarified along with suggestions for future research directions to handle the finer details of select post-hoc analysis methods. Furthermore, mabCAM is introduced as the heatmap generation method compatible with our ground-truth heatmaps. The framework is easily generalizable and uses only standard deep learning components.
Article
Full-text available
Residential properties are a major component of the environment and economy and a key element for the quality of human life. Faced with disruptive ideological and technological changes in the world, real estate analysis has also become a key research problem for many academic centers and private institutions. Due to the complex nature of properties, they are one of the most difficult and troublesome subjects of analysis. Given the rapid advancements in competitive automated analytical models, the problem of data representative sample selection may prove to be a very wide-reaching subject. The aim of this paper was the assessment of the particular soft computing methods’ (e.g., Self-Organizing Maps, Rough Set Theory) usefulness for selecting a representative property model. The obtained results confirm that the use of these methods leads to the creation of a model that enables a more reality-based view of the uncertainty and imprecise residential environment.
Chapter
Buildings are responsible for 40% of the total energy consumption annually in Europe, along with the respective greenhouse gas emissions. To mitigate these impacts, intensive research is ongoing in the sector of Nearly Zero-Energy Buildings (NZEBs). Within EU-funded project iClimaBuilt, RISE is developing and improving the formulation of cellular lightweight concrete insulation (CLC) as a sustainable and recyclable alternative for organic insulation materials (ex. Expanded polystyrene (EPS)). Compared to EPS, this material is non-combustible and achieves similar performance as existing insulation materials (thermal conductivity of 40 mW/(mK)) while economically more attractive (EPS 60 EUR/m3, CLC 50 EUR/m3). In addition, the material is easily recyclable since it is 100% mineral based. It is also similar or lighter in weight than traditional EPS (around 80 kg/m3). At these ultra-low densities as for concrete, the material integrity was ensured by adding polypropylene micro-fibres. To further improve the thermal conductivity of CLC, granulated silica aerogels were implemented in the mix. The CLC was implemented in a lightweight sandwich-panel in form of a prefabricated block similar to currently used EPS-sheets. The recyclability of the material as an SCM for new cement production is being evaluated by a reactivity study of grinded and reheated (recalcined) material. The panel prototypes will be tested in a laboratory and validated in the living labs – real-life test and experimentation buildings in different climate zones in Europe.KeywordsConcreteCellular-Lightweight ConcreteInsulationReactivitySCM
Chapter
Synthetic time series data generation is a wide area to research, and lot of attention has drawn recently. The assumption for generating multivariate time sequenced data is well proportioned and continuous without missing values. A productive time series data model should retain momentum, such that the new sequence maintains the actual relationship between different variations occurrence over time. Different existing approaches that put forward generative adversarial networks into the sequence system that prohibits care for temporary interactions that differ from sequential time data. Simultaneously, monitored sequence prediction models—allowing for fine control over network dynamics—are naturally determined. Time series data generation has problems like informative missing data values leading to untraceable challenge and long sequences with variable length. These problems are biggest challenges in making of a powerful generative algorithm. Herein, we are using an innovative structure to produce real-time sequence which associates the adaptability of different unattended prototype through controls provided during the supervised model training. Privacy data analysis safeguards data privacy and data sharing. Generation of accurate private synthetic data is a NP hard problem from any considered scenario. This paper discusses about privacy parameter and data analysis of real and synthetic data. The synthetic data is enclosed in the boundaries of real data and have similar behavior.KeywordsData privacyDynamic bayesian networkSynthetic dataTime series
Article
Contrary to the previous research on the general utilization of municipal solid waste (MSW) incineration bottom ash (IBA) in autoclaved aerated concrete (AAC), the given study strategically utilizes IBA for AAC formulation to improve IBA utilization efficiency and to enhance the hardened properties of the resulting IBA-AACs. A systematic classification and characterization approach was used to determine the suitable IBA fractions for each ingredient. The classified IBA with particle size less than 0.3 mm (IBA fines), IBA glass fraction (IBA-G), and coarse non-ferrous IBA with particle size greater than 1.18 mm (coarse IBA-ONF) were used as calcium, silica, and metallic aluminum (as an aerating agent) resource, respectively. The resulting IBA-AACs showed much higher specific strengths (about 40% enhancement) than the Control AAC due to the formation of relatively pure tobermorite gel. The leaching test results confirmed the safe utilization of AAC during and after the service life. The utilization of classified IBA optimized the amount of IBA that can be incorporated in AAC.
Article
The field of machine learning has taken a dramatic twist in re- cent times, with the rise of the Artificial Neural Network (ANN). These biologically inspired computational models can far exceed the per- performance of previous forms of artificial intelligence in common machine-learning tasks. One of the most impressive forms of ANN architecture is that of the Convolutional Neural Network (CNN). CNNs are primarily used to solve difficult image-driven pattern recognition tasks and with their precise yet simple architecture, offer a simplified method of getting started with ANNs. This document briefly introduces CNNs, discussing recently published papers and newly formed techniques in developing this bill- leniently fantastic image recognition models. This introduction assumes you are familiar with ANNs and machine learning fundamentals.
Article
Effective comprehension of highly complex and spatially heterogenous property market requires its’ appropriate recognition. One of the most critical steps in property analyses and valuation procedures is the identification of the sub-markets as the fundamental comparable units. The biggest challenge in this case is to define the criteria of the basis indicating the similarity (homogeneity) of property markets area. The objective of the study was to propose methodology (called “HO-MAR”) that enables objective identification of the homogenous (similar/comparable) areas (zones) that could indicate location of similar groups of property transactions (representative properties) used either in individual valuations or AVMs. The authors propose utilization of automated solutions based on robust geo-estimation that enables high efficacy of property submarkets identification. Robust geo-estimation is guaranteed by merging semi-automated data mining methods (e.g. entropy theory, rough set theory, fuzzy logic) and geoprocessing activities (Gauss filter, geocoding and reverse geocoding, tessellation model with mutual spatial overlapping) concerning spatial relation database application.