Applsci 13 01395
Applsci 13 01395
sciences
Article
Multi-Attribute Machine Learning Model for Electrical Motors
Performance Prediction
Maria Raluca Raia 1,2, * , Sebastian Ciceo 1,3,4 , Fabien Chauvicourt 3 and Claudia Martis 1
Abstract: Designing an electrical motor is a complex process that needs to deal with the non-linearity
phenomena caused by the saturation of the iron at high magnetic field strength, the multi-physical
nature of the investigated system and with requirements that may come into conflict. This paper
proposes to use geometric parametric models to evaluate the multi-physical performances of electrical
machines and build a machine learning model that is able to predict multi-physical characteristics
of electrical machines from input geometrical parameters. The focus of this work is to accurately
estimate the electromagnetic characteristics, motor losses and stator natural frequencies, using the
developed machine learning model, at the early-design stage of the electrical motor, when the
information about the housing is not available and to include the model in optimisation loops, to
speed-up the computational time. Three individual machine learning models are built for each
physics analysed, a model for the torque and back electromotive force harmonic orders, one model
for motor losses and another one for natural frequencies of the mode-shapes. The necessary data
is obtained by varying the geometrical parameters of 2D electromagnetic and 3D structural motor
parametric models. The accuracy of different machine learning regression algorithms are compared
to obtain the best model for each physics involved. Ultimately, the developed multi-attribute model
Citation: Raia, M.R.; Ciceo, S.;
is integrated in an optimisation routine, to compare the computational time with the classical finite
Chauvicourt, F.; Martis, C.
Multi-Attribute Machine Learning
element analysis (FEA) optimisation approach.
Model for Electrical Motors
Performance Prediction. Appl. Sci. Keywords: machine learning model; multi-physical analysis; electrical machine structural model;
2023, 13, 1395. https://doi.org/ motor parametric model
10.3390/app13031395
for electromagnetic and losses analysis and 3D FEA for structural characteristics. Hence,
the overall time is increased by using finite element (FE) analysis, that is time consuming
and memory intensive, even when parallelization techniques are used.
Traditionally, the process used to design a high-performance electrical machine is
multi-objective machine optimisation [1,2]. The optimal design is obtained by automatically
varying the geometric parameters within predefined limits, for imposed objectives and
constraints. The design space exploration is conducted using optimisation algorithms [3],
the designer having the freedom to choose the objectives, constraints and parameters
discretisation, selected based on manufacturing capabilities.
Various studies from the state of the art are focused on finding the optimal design of
electrical machines. Due to economical reasons (e.g., high cost of rare-earth materials) and
the need for high power densities, the cost optimisation procedure gained popularity in
the latter years, the machine being optimised to met the requirements at the lowest cost [4].
Different motor topologies, including permanent magnet and synchronous reluctance
machines, are analysed in order to select the best design at the lowest cost in [5], while in [6],
a permanent magnet synchronous machine (PMSM) is optimised to met the performance
and cost demands, with focus on high-volume mass production and its constraints. Another
optimisation objective is presented in [7], where the torque ripple is optimised to obtain a
high and smooth torque. Nevertheless, the computational cost of an optimisation loop can
drastically increase when a large number of machine designs are analysed. This is caused by
the FE-based simulations conducted to evaluate the performances of the machine designs.
Despite their well-known accuracy, the simulations based on FE may limit the optimisation
process, due to their high computational cost (simulation times may vary from several
minutes to several hours or even days [8]). To overcome the discussed issues, fast models
can be developed using machine learning models [9], reducing the computational burden
in the design stage, as most of the computations are carried out in the model building
phase. At the same time, several processes can be brought earlier in the design cycle. This
way, the system’s performances and sensitivities can be identified in the concept stage and
the designer can decide if the desired targets are met.
Machine learning models used for design, optimisation, fault detection and robustness
evaluation have been among the main research interests in electrical machines field over
the years. Some works have already focused on generating machine learning models
that allow replacement of the time consuming FEA and reduce the computational time.
In [10], a statistical analysis that uses multiple correlation coefficients has been used to
generate a fast model that is able to replace the FE model and reduce the computational
effort, whereas in [11], the same objective is accomplished by using an artificial neural
network. Another approach that uses online learning and dynamic weighting methods
is presented in [12]. Moreover, in [3] the focus is on analyzing the effectiveness of using
electrical machine machine learning models that incorporate tolerance or sensitivity aspects
in a multi-objective optimization run. Further works focusing on machine-learning-assisted
multi-objective optimisation are presented in [13], and in [14]. A recent work presents a
data-driven structural modelling for electrical machine airborne vibration that is intended to
be used in both design stage for optimisation purposes and in system-level simulations [15].
In [16], a multi-physics simulation workflow, based on reduce-order models, used to
predict the vibro-acoustic behaviour of electrical and decrease the computation time is
presented. The influence of the production mass tolerances modeled at system level and
the interaction between the uncertainties and the drive’s components, together with the
fitted machine learning can be found in [1]. Fast prediction of electromagnetic torque and
flux linkages used in system-level simulations are accomplished using a machine learning
electromagnetic model based on artificial neural networks in [17]. Machine learning models
employed to predict sound pressure levels are developed in [18,19], where it is proven that
the developed models can be considered as replacements of the FEA for future design and
optimization problems of the same motor.
Appl. Sci. 2023, 13, 1395 3 of 22
Input parameters
(geometry)
Step 1
Impose input
parameters and
analysis conditions
Design Space Exploration
with LHS
Step 4
Torque and Back Emf
Modes and Frequencies Predict motor
Harmonic Losses ML Model
ML Model performances using 3
ML Model
machine learning models
Ra
YT
TWS
ng
A
SO
TGD
/2
SO R
Figure 2. Stator cross-section parameterization where DOFs marked with black are kept fixed and
the DOFs marked with red are varied to construct the machine learning model [15].
Appl. Sci. 2023, 13, 1395 5 of 22
The parameterised model allows the generation of a set of feasible designs by perform-
ing a design of experiments using specific sampling techniques. The most important design
parameters, TWS, YT, SOAng, TGD, SO, are varied within imposed boundaries, while the
stator length (Lstk ) and the stator inner and outer radii are kept constant. Each parameter is
individually varied within imposed limits and the design space is filled with the help of
the Latin Hypercube Sampling Technique (LHS). The variation interval, with its lower (LB)
and upper (UB) boundaries, of the considered DOFs is presented in Table 1.
Table 1. Stator geometry DOFs considered for variation and their variation interval.
The variable DOFs are chosen based on their impact on the structural characteris-
tics and electromagnetic performances of the machine. The yoke thickness is the main
contributor to the stator vibration response due to its direct influence on the stator stiff-
ness. By increasing the yoke thickness, the stiffness value enlarges, shifting the natural
frequencies corresponding to each mode-shape to a higher value. In particular, mode 0
(breathing mode) is strongly influenced by the variation of this parameter. At the same
time, the yoke thickness, together with the tooth width have a massive impact on the
electromagnetic characteristics by influencing the saturation levels of the electrical machine.
The tooth width can also influence the structural behavior of the motor by creating local
tooth bending modes. On the other hand, the variation of tooth tip height, slot opening
and tooth tip, influence especially the electromagnetic quantities. The electromagnetic flux
density harmonics are affected by this variation and, indirectly the structural vibration
characteristics, by influencing the source of vibrations for the electrical motor.
3. Data Generation
Multi-physical FEA simulations are performed on the electromagnetic and structural
parametric models in order to extract the electromagnetic quantities (i.e., electromagnetic
torque, back electromotive force and motor losses) and the modal characteristics (i.e., mode-
shapes and their corresponding natural frequencies). Motor valid designs are obtained by
a variation of the stator geometrical parameters (TWS, YT, SOAng, TGD and SO) using
the LHS method, while the stator length (Lstk ), the stator inner and outer radii, together
with material properties remain invariant. For the electromagnetic analysis, the rotor and
magnet properties and geometry are also kept constant. Given these assumptions, 3D
FE modal analysis simulations are performed on the stator core, extracting the structural
mode-shapes and natural frequencies, while 2D FE electromagnetic analysis are performed
to obtain the electromagnetic characteristics and motor losses of the generated feasible
designs. At the end of the data generation process, the obtained information is harnessed
using different machine learning algorithms to obtain the most suitable machine learning
model to predict the electrical motors performances from input geometrical parameters.
Parameter Value
Rated Power 800 W
Base Speed 1650 rpm
Maximum Speed 4000 rpm
Rated Current 110 A (rms)
DC Link Voltage 12 V
Minimum Airgap 0.5 mm
Stack Length 58 mm
FEA FEA
reconstructed reconstructed
Back EMF [V]
Torque [Nm]
Figure 4. Electromagnetic torque and the back emf waveforms extracted from an electromagnetic
analysis (blue line) for JT = 4.39935 mm, TWS = 6.9565 mm, SO = 1.8815 mm, SOAng = 116.65 deg.,
Back EMF [V]
Torque [Nm]
TGD = 0.98125 mm and their reconstructed waveforms from the most important harmonics (red line).
Appl. Sci. 2023, 13, 1395 7 of 22
The generated data is intended to be used for training a machine learning model that
is capable to predict the performances of new electrical machine designs. Therefore, a
machine learning model suitable for waveforms should be chosen. However, this model
type includes the time as one of the dimensions, hence, it is more computationally expensive
in the training phase [22]. The problem can be simplified by using a harmonic model for
the torque and back emf values, instead of using their time varying waveforms. This
way, the time dimension is eliminated and the machine learning algorithm will be applied
on discrete datasets, reducing the complexity and the training time, compared with the
waveform machine learning model.
For that, the torque and back electromotive force waveforms are post processed by
applying the discrete Fourier decomposition, obtaining their corresponding harmonic order
amplitudes and phases. The harmonic orders obtained for electromagnetic torque and
back emf waveforms extracted from the electromagnetic analysis for JT = 4.39935 mm,
TWS = 6.9565 mm, SO = 1.8815 mm, SOAng = 116.65 deg., TGD = 0.98125 mm can be
identified in Figure 5, where with blue it is identified the DC component for electromagnetic
torque and the 1st harmonic order for back emf, and with red the higher harmonic orders,
having lower amplitudes. For the training process, only the first three most influential
harmonic orders are taking into account, for the electromagnetic torque (DC, 6th and
12th harmonic orders) and for the back emf (1st, 3rd and 11th harmonic orders). The
reconstructed torque and back emf waveforms using the most important harmonic orders
amplitude and phases are represented in Figure 4 with red lines. The torque waveform
was reconstructed using the 0 deg. phase for the DC component, −5.83 deg. for the 6th
harmonic and 18.94 deg. for the 11th harmonic order. The phases of the back emf harmonics
do not vary from one design to another, keeping a value of 91.2 deg. for the 1st harmonic,
−86 deg. for the 3rd one and −74.5 deg. for the 11th harmonic.
Back EMF [V]
Torque [Nm]
modifying the transmission path and shifting the stator structure eigen-frequencies, or by
influencing the excitation source, in this case, the air-gap forces. However, in some cases,
because the electric machine operates with variable frequency, it is impossible to eliminate
the resonance phenomena caused by variable frequency forces that excite the structure.
The modal analysis is based on the general motion equation, for a system with N
Degrees-Of-Freedom (DOFs):
where [ M] represents the system mass matrix, [ B] and [K ] represents the system damping
matrix and the stiffness matrix, respectively, of size N × N. The terms x (t) and f (t) are the
displacement and force vectors of size N × 1.
Modal analysis solves Equation (1) for free vibrations, i.e., without taking into considera-
tion the damping and the force, resulting in the following expression of the motion equation:
[ M] ẍ (t) + [K ] x (t) = 0 (2)
For this work, modal analysis is conducted by FEA, where the 3D structural parametric
model of the stator is used. A proper mesh type is chosen for the structural model to fulfill
the requirements regarding computational time and model accuracy. The structural mesh
is built from a 1 mm mesh size on the stator face, resulting in a 14.568 3D eight-noded
hexahedral solid (CHEXA(8) in Simcenter 3D terminology) and in 19.521 nodes. Once the
mesh is set, the material for the stator core is chosen. An isotropic material is used for the
stator steel, while free–free conditions are imposed for the modal analysis. The 3D structural
model is presented in Figure 6. The structural modal analysis is performed for a frequency
interval set to cover the audible frequency range, from 20 Hz to 20 kHz. The computational
time of the full process per discretized 3D structural mesh, on a workstation having an
Intel Core i7-9850H CPU running at 2.6 GHz, with 32 GB of RAM takes approximately 2
min (1 min and 38 s).
Figure 6. Parametric stator mesh having 19.521 nodes and the output nodes (with orange).
The first six mode-shapes of the base machine, identified by the automated process,
with and without longitudinal deflection, are presented in Table 3. Here, the mode-shapes
are identified as a combination of (m,n) where m is the order of the circumferential deflec-
tion and n is the order of the longitudinal deflection.
Appl. Sci. 2023, 13, 1395 9 of 22
FE Results
(0, 0) (0, 1)
(2, 0) (2, 1)
(3, 0) (3, 1)
(4, 0) (4, 1)
(5, 0) (5, 1)
(6, 0) (6, 1)
The process of running massive modal analysis simulations and identifying the modes
and natural frequencies without human intervention is allowed by developing a set of
scripts in Python and NX Open with the help of Simcenter 3D journaling capability. The
automated process allows the opening of a new Simcenter 3D session and control of the
simulation workflow from Python according to the user’s specifications. In the main Python
script, the user sets the number of DOEs, some analysis conditions (e.g., frequency range,
mesh size, material), and then the full process, starting from imposing the desired geometric
parameters, updating the mesh for each design, imposing simulation conditions, solving
the problem, exporting the data and identification of mode-shapes and their corresponding
natural frequency is carried out automatically. The mode-shapes are identified by taking
into consideration the displacement in radial, tangential and axial direction of stator
output nodes, highlighted in Figure 6 with orange. A developed script analyses the node’s
displacement values and identifies a deformation pattern of all the nodes belonging to the
Appl. Sci. 2023, 13, 1395 10 of 22
outer face of the stator. Moreover, the script has the capacity to distinguish between global
and local modes, selecting for this work only the first six global modes of the stator.
process is repeated for a imposed number of iterations in order to minimize the error and
obtain a better prediction.
One of the main drawbacks is that the GBR minimizes all errors, including the ones
given by the outliers, that can cause overfitting. However, to address the overfitting issues,
different methods, as regularization, setting a maximum depth and early stopping can be
chosen [38,39]. Another disadvantage is that this method is almost impossible to scale up,
because each decision tree is trained based on the previous one and it is hard to parallelize
the process. For processes that need to be scaled up, a scalable end-to-end tree boosting
system called XGBoost is widely used by data scientists. The XGBoost reduces the run time
and scales to billions of examples in distributed or memory-limited settings [40].
Considering that the datasets under test are discrete, with a non-linear relationship
between data and features, the GBR algorithm is a suitable method to train a machine
learning model. For the four datasets under test, the overfitting issue was avoided by
choosing a maximum depth of three and an early stopping coefficient to stop the training
process when validation score is not improving after 20 iterations. The loss function imposed
for the training process is the squared error.
Considering ŷi as the predicted value of the i − th sample, yi as the corresponding true
value for a number of n fitted points with the mean value ȳi , R2 coefficient is defined as:
By analysing the values, it can be seen that all tested methods show an ascendant
trend for the R2 score and a reduction of the MSE error when the number of samples
is increased. In terms of machine learning models, SVR model performs much better
than the GBR model for all types of samples, the GBR presenting a higher MSE and a
lower R2 than the SVR. At the same time, it can be observed that the GPR model fits the
targets more accurately, for each training sets. Starting with 750 samples, the GPR presents
excellent results, its R2 score being higher than 0.93 and MSE lower than 7%. In particular,
Appl. Sci. 2023, 13, 1395 14 of 22
for 750 samples, GPR succeeds to improve the scores for the 12th TH, R2 from 0.86 (for
250 samples) to 0.93 and the MSE from 13% (for 250 samples) to 7%. The same situation
applies for the 11th BmfH, the R2 value increases from 0.89 to 0.94 and the MSE reduces
from 11% to 1%. However, even if the process of generating 1000 samples from FEA is
more computationally expensive, GPR performs at its best for 1000 samples, enhancing,
compared with the 750 samples case, the fitting capabilities, especially for the third most
influential torque and back emf harmonics. For this case, the obtained R2 score is 0.95 and
MSE is 5% for 12th TH , and the R2 score is 0.99 and the MSE is 1% for the 11th BmfH.
Figure 7 shows the most influential torque and back emf harmonics over their actual target
(original) values obtained from the GPR 1000 samples machine learning model.
DC 6 th TH 12 th TH
Train Train Train
Test Test Test
Prediction
Prediction
Prediction
Original Original Original
1 st Bmf H 3 rd Bmf H 11 th Bmf H
Train Train Train
Test Test Test
Prediction
Prediction
Prediction
Original Original Original
Figure 7. GPR 1000 samples accuracy. Predicted vs. original values for the DC, 6th and 12th torque
and 1st, 3rd and 11th back emf harmonics orders.
The capability of the tested regression methods to predict the modes and their corre-
sponding natural frequencies are presented in Table 5, where the R2 and the MSE scores
are presented for each individual regression method, SVR, GBR, GPR, sequentially trained
using 250, 500, 750 and 1000 samples.
As can be observed, the machine learning models under test are able to predict very
well even when they are trained using 250 samples. By analysing the presented values, it
can be seen that except the GPR method, that keeps its scores constant, the SVR and the
GBR show an improvement of the R2 score and a reduction of the MSE error when the
number of samples is increased. Moreover, their performance starts to saturate beyond
750 samples. This behaviour is emphasized when the data size is increased from 750 to
1000 samples. Comparing the results, it is clear that the R2 and the MSE are the same in
both cases, and an expansion of data size beyond 750 does not influence the accuracy of
the developed models. Actually, the models developed for 750 samples perform well, their
R2 values being 0.99 for the SVR and GBR models and 1 for GPR model. Regarding the
MSE, its values are between 0% and 1% for SVR and GBR, while for GBR it is 0%. The
GPR model fits the natural frequencies more accurately than SVR and GBR. Even starting
with 250 samples, the model is able to be the best performer. The low number of training
samples necessary for a good performance is due to the fact that the global modes and
their natural frequency are highly affected by the yoke and tooth thickness and are not
as sensitive as the electromagnetic targets when the tooth tip angle, tooth tip height and
the slot opening are adjusted within the set range. Therefore, less designs are needed
to build the data-driven model and the structural characteristics obtained by imposing
250 input geometrical parameters are sufficient to obtain a good generalisation capacity
that allows the characterisation of a new design. Figure 8 shows the model capability to
Appl. Sci. 2023, 13, 1395 15 of 22
approximate the motor structural characteristics. Specifically, the frequency at which the
first six mode-shapes appear is estimated with a high accuracy by the GPR 250 samples
machine learning model.
Prediction
Prediction
1.4
3.5
1.3
1.35
1.2
3
1.3 1.1
1.3 1.35 1.4 1.45 1.1 1.2 1.3 1.4 1.5 3 3.5 4
Original 104 Original 103 Original 103
103 Mode4 103 Mode5 104 Mode6
Train Train Train
6.5 Test 9 Test 1 Test
Prediction
Prediction
Prediction
6
8 0.9
5.5
0.8
7
5
Figure 8. GPR 250 samples accuracy. Predicted vs. original values for stator modes and frequencies.
The fitting ability to predict the losses targets can be identified in Table 6, where the
scores for all types of motor losses are presented (i.e., total—Tot, winding losses—Wind,
iron—Iron, stator back iron—SBI, rotor back iron—RBI and magnet—Mag). The results
for R2 and MSE show that SVR performs better than the GBR for all dataset sizes. The
performances of all three models are improved by increasing the size of the training data,
the worst prediction being obtained at 250 samples for stator tooth losses. Both SVR and
GBR manage to increase the R2 to 1 and 0.99 and MSE to a value under 2 for 1000 samples
case, but the model that is more suitable to accurately fit the losses data is, also in this
case, GPR. This algorithm manages to predict, even for 250 samples, the losses with a
maximum MSE of 9%, minimising the MSE value to 1% for 500 samples and 750 cases and
to 0% for 1000 samples case. GPR perfectly fits the target data for 1000 samples case, where
the accuracy indicators show perfect results, R2 taking unity value and MSE being zero.
Figure 9 displays the predicted motor losses (total, winding, iron, stator back iron, stator
teeth, rotor back iron and the magnets) over their actual target values (original) obtained
from the GPR 1000 samples machine learning model. The obtained values are specific to
Appl. Sci. 2023, 13, 1395 16 of 22
the described number of stator DOFs. For less DOFs, a reduced number of feasible designs
is needed to build the data-driven model. Correspondingly, when the number of DOFs is
increased, the number of analysed designs must be enlarged to keep the same accuracy.
Prediction
Prediction
Prediction
Original Original
Prediction
Original Original
Figure 9. GPR 1000 samples accuracy. Predicted vs. original values for total motor, winding, iron,
stator back iron, stator teeth, rotor back iron and magnets losses.
system-level analysis and a fast optimisation routine. Therefore, an evaluation of the devel-
oped multi-attribute machine learning model capability to predict the optimal solution and
its computational cost is performed. The multi-attribute machine learning model under test
is defined by the combination of the best performance regressors for each physics involved,
GPR 1000 for electromagnetic targets, GPR 250 for natural frequencies of modes and GPR
1000 for losses, and from now on denoted with ML1000.
To quantify the advantage of using the ML1000 machine learning model over FEA,
the model is included within a geometric optimization process that targets to maximise
a function (e.g., natural frequency of a mode) with respect to imposed multi-physical
constraints. In parallel, a FEA based optimisation procedure, having identical objectives
and constraints to the ML1000 optimisation process, is performed on the structural and
electromagnetic FE motor model. After that, the optimisation results of the two presented
methods are compared from the point of view of accuracy and computational costs.
The natural frequency of Mode 2, average torque, torque ripple and total losses belong
to the constraint vector C ([ x ]).
Table 7. ML1000 vs. FEA Optimisation results expressed as percentage difference from the nomi-
nal design.
The computational cost of both methods, ML1000 and FEA based optimisations, is
identified in Figure 10. Here, the computation time is identified in function of the number of
machine geometric designs taken into consideration during the optimisation process (NO ).
120
FEA
110 ML1000
100
90
80
Time [h]
70
60
50
40
30
20
10
250 500 750 1000 1250 1500 1750 2000
No
The computational time is identified with blue for the FEA case and with red for the
ML1000 model. For the machine learning model, the computation time consists of both
time spent building the machine learning model and the time required for optimisation. As
it can be observed, for the machine learning model, the most time consuming process is the
extraction of data necessary for model identification. The cost of building the ML1000 model,
based on FEA simulations, is 60 h, considering that the structural and electromagnetic FEA
analysis are carried out in parallel. Once the data is available, the training time is negligible
(it usually takes couple of seconds for 1000 samples), while the run time is low and
considered negligible too. In the same figure, it can be noticed that the computational cost
of the ML1000 model is higher than the FEA for the cases where the number of simulation
used to build the model are higher than the number of optimisation designs. Starting with
1000 optimisation designs, the ML1000 model starts to be more time efficient than the FEA
approach. This region is marked with a grey area in Figure 10. After this threshold, it
can be noticed that the FEA cost time increases considerable with each additional design,
Appl. Sci. 2023, 13, 1395 19 of 22
compared with the machine learning model approach, for which the computational cost of
a new design is insignificant.
In order to quantify the time reduction, a computational efficiency factor that gives the
relative reduction in computational time, is introduced [47]. The computational efficiency
ratio (kcr ) is defined as the ratio between the number of designs used to build the machine
learning model (ND ) and the number of designs used in the machine geometric optimisation
process (NO ):
N
kcr = D (5)
NO
The saving of computational time for the proposed machine learning model is shown
in Figure 11 with red. For the cases where kcr is less than 1, marked with grey area, the
machine learning model is more time efficient than the FEA. The developed model presents
computational feasibility over a wide range. For example, if 2000 designs are considered
in the geometric optimisation process, the computational time reduction is 50%. The
computational efficiency factor values for other machine learning models, trained with
250, 500 and 750 samples are identified with black, blue and green lines. For these cases, it
can be observed that they start to be feasible starting with fewer optimisation designs, but
the drawback is that they are not as accurate as the ML1000 model is, as presented in the
previous section.
3
ML250
ML500
2.5 ML750
ML1000
2
k ctr
1.5
0.5
0
400 600 800 1000 1200 1400 1600 1800 2000
No
Figure 11. Machine learning model vs. FEA computational efficiency factor.
6. Conclusions
This paper presents a multi-physical machine learning approach for multi-attribute
performance prediction of electrical motors. For that, the full process of designing the
electrical motor is brought in the early-design stage, where all the characteristics (elec-
tromagnetic, motor losses, modal analysis mode-shapes and natural frequency) can be
evaluated and predicted by the designer. The multi-attribute model is constructed based
on the data achieved by conducting a series of 2D electromagnetic and 3D structural FE
analysis on a set of motor models generated by modifying the DOFs consisting in the
geometrical parameters of a base parametric motor model. Three independent machine
learning models are developed for each analysed physics, one for electromagnetic torque
and back emf harmonics, one for motor losses and one for the structural targets (modes
and natural frequencies). The tested methods are support vector regression (SVR), gradient
Appl. Sci. 2023, 13, 1395 20 of 22
boosting regressor (GBR) and Gaussian process regressor (GPR) and were selected due
to their capacity to predict discrete datasets. Moreover, their accuracy is tested for four
training datasets, 250, 500, 750 and 1000 samples. The fitting capabilities of the regression
models are individually analysed. Two types of key performance indicator (KPI) are used
to measure the fitting capabilities: the coefficient of determination (R2 ) and the mean
squared error (MSE). The results show that the GPR algorithm applied on 1000 samples is
the most suitable machine learning model to predict the electromagnetic torque and back
electromotive force harmonics orders, its R2 scores being higher than 0.95, while the MSE is
less than 5%. For structural targets, the regressor that is able to fit the data with the best
scores is the GPR model with 250 samples, R2 score being 1 and MSE 0%. Regarding the
ability of the machine learning models to predict the motor losses, the GPR 1000 samples
machine learning model presents the best results, with R2 scores of 1 and MSE values of
9%. It was demonstrated that the multi-attribute machine learning model developed by
combining the best models for each physics involved can provide accurate and significant
reduction in computational time, compared with the classical FEA approach.
Author Contributions: Conceptualization, M.R.R., S.C. and F.C.; writing—original draft preparation,
M.R.R.; writing—review and editing, S.C., F.C. and C.M.; supervision, S.C., F.C. and C.M. All authors
have read and agreed to the published version of the manuscript.
Funding: This paper was financially supported by the Project “Entrepreneurial competences and
excellence research in doctoral and postdoctoral programs—ANTREDOC”, project co-funded by
the European Social Fund financing agreement No. 56437/24.07.2019. This research was partially
supported by the project 38 PFE in the frame of the programme PDI-PFE-CDI 2021.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: Not applicable.
Acknowledgments: The authors would like to thank Davide Gorgoretti from Siemens Industry
Software NV, Leuven, Belgium, for the help with building the machine learning model.
Conflicts of Interest: The authors declare no conflict of interest.
12. Sykulski, J.K.; Al-Khoury, A.H.; Goddard, K.F. Minimal function calls approach with on-line learning and dynamic weighting for
computationally intensive design optimization. IEEE Trans. Magn. 2001, 37, 3423–3426. [CrossRef]
13. Silva, R.C.P.; Rahman, M.; Li, T.; Lowther, D.A. Surrogate-Based MOEA/D for Electric Motor Design With Scarce Function
Evaluations. IEEE Trans. Magn. 2017, 53, 7400704. [CrossRef]
14. Lim, D.-K.; Woo, D.-K.; Yeo, H.-K.; Jung, S.-Y.; Jung, H.-K. A new robust surrogate-assisted multi-objective optimization algorithm
for an IPMSM design. In Proceedings of the 2016 IEEE Conference on Electromagnetic Field Computation (CEFC), Miami, FL,
USA, 13–16 November 2016; p. 1.
15. Ciceo, S.; Chauvicourt, F.; Gyselinck, J.; Martis, C. Data-Driven Electrical Machines Structural Model Using the Vibration Synthesis
Method. IEEE Trans. Transp. Electrif. 2022, 69, 3771–3781. [CrossRef]
16. Pinto, D.E.; Pop, A.-C.; Myrria, G.; Kempkes, J.; Gyselinck, J.J.C. Vibration Analysis of Delta-Connected PMSMs Using Lookup
Table-Based Models—Influence of the 0-Sequence Component. IEEE Trans. Ind. Electron. 2022, 69, 6561–6571. [CrossRef]
17. Raia, M.R.; Ruba, M.; Nemes, R.O.; Martis, C. Artificial Neural Network and Data Dimensionality Reduction Based on Machine
Learning Methods for PMSM Model Order Reduction. IEEE Access 2021, 9, 102345–102354. [CrossRef]
18. Mohammadi, M.H.; Rahman, T.; Silva, R.C.P.; Wang, B.; Chang K.; Lowther, D.A. Effect of acoustic noise on optimal SynRM
design regions. IEEE Trans. Magn. 2018, 54, 8101304. [CrossRef]
19. Ibrahim, I.; Silva, R.; Mohammadi, M.H.; Ghorbanian, V.; Lowther, D.A. Surrogate-Based Acoustic Noise Prediction of Electric
Motors. IEEE Trans. Magn. 2020, 56, 7509304. [CrossRef]
20. Parekh, V.; Flore, D.; Schöps, S. Deep Learning-Based Prediction of Key Performance Indicators for Electrical Machines. IEEE
Access 2021, 9, 21786–21797. [CrossRef]
21. Hannon, B.; Sergeant, P.; Dupré, L.; Pfister, P.-D. Two-Dimensional Fourier-Based Modeling of Electric Machines—An Overview.
IEEE Trans. Magn. 2019, 55, 8107217. [CrossRef]
22. Towards Data Science. Available online: https://towardsdatascience.com/how-not-to-use-machine-learning-for-time-series-
forecasting-avoiding-the-pitfalls-19f9d7adf424 (accessed on 28 November 2022).
23. Wheeler, R.; Dorfi, H.; Keum, B. Vibration Modes of Radial Tires: Measurement, Prediction, and Categorization Under Different
Boundary and Operating Conditions. J. Passeng. Car Mech. Syst. J. 2005, 114, 2823–2837.
24. Castagnaro, E.; Bacco, G.; Bianchi, N. Impact of Geometry on the Rotor Iron Losses in Synchronous Reluctance Motors. IEEE
Trans. Ind. Appl. 2019, 55, 5865–5872. [CrossRef]
25. Bianchi, N.; Barcaro, M.; Bolognani, S. Electromagnetic and Thermal Analysis of Permanent Magnet Synchronous Machines. In
Finite Element Analysis—From Biomedical Applications to Industrial Developments; IntechOpen: London, UK, 2012. Available online:
https://www.intechopen.com/chapters/34521 (accessed on 28 November 2022).
26. Chiodetto, N.; Bianchi, N.; Alberti, L.; Improved Analytical Estimation of Rotor Losses in High-Speed Surface-Mounted PM
Synchronous Machines. IEEE Trans. Ind. Appl. 2017 53, 3548–3556. [CrossRef]
27. Popescu, M.; Staton, D.; Dorrell, D.; Marignetti, F.; Hawkins, D. Study of the thermal aspects in brushless permanent magnet
machines performance. In Proceedings of the 2013 IEEE Workshop on Electrical Machines Design, Control and Diagnosis
(WEMDCD), Paris, France, 11–12 March 2013; pp. 60–69.
28. Machine Learning in Python. Available online: https://scikit-learn.org/stable/ (accessed on 28 November 2022).
29. Towards Data Science. Available online: https://towardsdatascience.com/an-introduction-to-support-vector-regression-svr-a3
ebc1672c2 (accessed on 28 November 2022).
30. Towards Data Science. Available online: https://towardsdatascience.com/when-do-support-vector-machines-fail-3f23295ebef2
(accessed on 9 January 2023).
31. Machine Learning in Python. Available online: https://scikit-learn.org/stable/modules/generated/sklearn.svm.SVR.html
(accessed on 9 January 2023).
32. Lee, J.A.; Verleysen, M. Nonlinear Dimensionality Reduction; Springer Science & Business Media: New York, NY, USA, 2007.
33. Machine Learning in Python. Available online: https://scikit-learn.org/stable/modules/feature_selection.html (accessed on 9
January 2023).
34. Towards Data Science. Available online: https://towardsdatascience.com/all-you-need-to-know-about-gradient-boosting-
algorithm-part-1-regression-2520a34a502 (accessed on 28 November 2022).
35. KDnuggets. Available online: https://www.kdnuggets.com/2018/08/make-machine-learning-models-robust-outliers.html
(accessed on 9 January 2023).
36. Analytics India Magazine. Available online: https://analyticsindiamag.com/adaboost-vs-gradient-boosting-a-comparison-of-
leading-boosting-algorithms/ (accessed on 9 January 2023).
37. Towards Data Science. Available online: https://towardsdatascience.com/understanding-gradient-boosting-machines-9be756
fe76ab (accessed on 28 November 2022).
38. Data Science. Available online: https://datascience.eu/machine-learning/regularization-in-machine-learning/ (accessed on 9
January 2023).
39. KDnuggets. Available online: https://www.kdnuggets.com/2022/08/avoid-overfitting.html (accessed on 9 January 2023).
40. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International
Conference on Knowledge Discovery and Data Mining (KDD ’16), San Francisco, CA, USA, 13–17 August 2016; pp. 785–794.
Appl. Sci. 2023, 13, 1395 22 of 22
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.