8.4. Applications of the Growing Neural Network
Frameworks
8.4.1. Parallel Framework
For the
parallel approach, 100 runs of the growing neural network algorithm were performed
simultaneously for each analyte. The ranking of the variables after the first
step of the frameworks is shown in figure 55. In
contrast to the genetic algorithm framework, the variables are ranked separately
for each analyte, as for each analyte separate networks were grown. The final
topology resulted in fully connected networks for R22 with 5 input neurons and
6 hidden neurons and for R134a with 6 input neurons and 7 hidden neurons. The
input variables, which were used by these networks, are labeled in figure
55. These networks were subsequently trained with the complete calibration
data set and then used for the prediction of the validation data whereby the
errors are shown in table 4. The predictions
of the validation data not used during the network building process are the
best out of all methods used up to now with 2.04% for R22 and 2.61% for R134a.
The small size of the networks with only 43 respectively 57 adjustable parameters
is rewarded by an excellent generalization ability. The true-predicted plots
look pretty much like the true-predicted plots of the genetic algorithm framework
(figure 47) with low standard deviations
and no bias present and are not shown here. For both analytes, all variables
used by the networks were recorded within the first 16 seconds of exposure to
analyte and within the first 30 seconds after the end of exposure to analyte
(60 seconds to 90 seconds). The selected time points are located within the
same time intervals as the time points selected by the genetic algorithm framework
(except of the time-point 125 s). Yet, the parallel growing neural network framework
achieves better predictions using fewer variables and smaller networks. Additionally,
the variable selection of the parallel framework suggests to reduce the exposure
time to 20 seconds and to record the signal of the sensor for 90 seconds. The
reduced time of exposure to analyte would also reduce the time needed for the
recovery of the sensor signal resulting in a significantly shorter repetition
times.
figure 55:
Ranking of the time-points represented as frequency of being present in the
grown neural networks of the first step of the parallel framework.
The same
test for chance correlation and reproducibility was performed as for the genetic
algorithm framework before. The parallel framework was used for the increased
data set with 40 additional autoscaled random variables the same way as described
for the original data set. In the ranking of the variables after the first step,
no random variable was found in the top 34 variables for R22 and in the top
25 variables for R134a. This means that the growing neural network algorithm
and especially the parallel framework are very robust to selecting randomly
correlated variables. In figure 56, it is shown
that for R22 the first 8 variables and for R134a the first 7 variables in the
ranking were exactly the same as in the previous run of the parallel algorithm
resulting in exactly the same neural network topologies and predictions. Thus,
the parallel framework shows a high reproducibility not sensitive to the partitioning
of the calibration data set.
figure 56: Ranking of the time-points
and of the random variables after the first step of the parallel framework for
the increased data set.