Abstract
Detecting fall compensatory behaviour from large EEG datasets poses a difficult problem in big data which can be alleviated by evolutionary computation-based machine learning strategies. In this article, hyperheuristic optimisation solutions via evolutionary optimisation of deep neural network topologies and genetic programming of machine learning pipelines will be investigated. Wavelet extractions from signals recorded during physical activities present a binary problem for detecting fall compensation. The earlier results show that a Gaussian process model achieves an accuracy of 86.48%. Following this, artificial neural networks are evolved through evolutionary algorithms and score similarly to most standard models; the hyperparameters chosen are well outside the bounds of batch or manual searches. Five iterations of genetic programming scored higher than all other approaches, at a mean 90.52% accuracy. The best pipeline extracted polynomial features and performed Principal Components Analysis, before machine learning through a randomised set of decision trees, and passing the class prediction probabilities to a 72-nearest-neighbour algorithm. The best genetic solution could infer data in 0.02Â s, whereas the second best genetic programming solution (89.79%) could infer data in only 0.3 ms.
Graphical abstract

1 Introduction
Fall detection systems are deployed to automatically detect when an individual has fallen, which can then alert a caregiver or the emergency services. Detecting falls is possible through various means, such as wearable devices, smart home sensors, or computer vision. Fall detection systems analyse the data from these devices to automatically determine if the individual has suffered a fall [1]. Basic fall detection systems require the service user to pull an emergency cord, which may not necessarily be within reach. Hence, the need for automatic detection of falls, a subset of Human Activity Recognition (HAR). There are a growing number of solutions related to smart homes to improve the possibilities of independent living, including the automatic detection of falls. In most HAR applications. Machine Learning (ML) tools are employed to predict user behaviour. Oftentimes, these solutions are used in tandem with one another since ML predictions are imperfect [2].
Populations are ageing due to increasing life expectancy, and elderly healthcare service users are more prone to falls [3]. For example, the UK registered more deaths than births in 2021 [4]. With this demographic shift comes changes to those who provide care and who require care [5]. We can see here that changes are required for healthcare systems to continue to operate normally and effectively. One of the many solutions enabled by modern technology is the application of ML techniques for independent assisted living [6, 7].
Autonomous fall detection can be performed through Computer Vision [8], physically via accelerometers and gyroscopes (most commonly in pendants [9]), or by monitoring bioelectrical data from parts of the human body such as muscles and the brain [10,11,12]. Given the imperfections in the technology, it is ideal for exploring multiple approaches for later fusion, as well as the optimisation of those approaches.
In this paper, frontal lobe electroencephalogram (EEG) data incorporating ML models are used to predict compensatory fall-related behaviour.Footnote 1 Since costs must be considered when dealing with strained healthcare systems, the use of a low-cost single-electrode EEG sensor that can be embedded as a wearable is proposed. The EEG sensor is used to record data from the frontal lobe, and thus the goal is to detect compensation for falls, rather than balance-related signals directly from the cerebellum, as it is impossible to reach unless clinical and invasive approaches are used. Another issue that arises in biological signal processing is related to big data, and strategies are needed to identify which information is useful against the large amount of data which does not contribute to machine learning but will significantly increase computational complexity and cost.
The scientific contributions of this work are as follows: Genetic programming of learning pipelines provides the strongest models for fall detection from EEG; the algorithm is executed for five individual iterations, leading to the five best overall results. Two methods of hyperheuristic optimisation are explored; (a) evolutionary optimisation of neural network hyperparameters (neuroevolution), and (b) genetic programming of ML pipelines. The results show that when neuroevolution is executed for fifty generations over five iterations, deep learning finds difficulty for the available fall detection data. The best neuroevolved artificial neural network achieves 73.41% mean accuracy, which is worse than the ten other algorithms explored and better than 15 other algorithms. Exploration of the pareto frontiers of accuracy versus training time and accuracy versus inference time shows that the best model scores 90.52% mean accuracy and can infer brain activity in 0.019Â s per prediction. The second-best model scores a slightly lower 89.79% mean accuracy but can infer data objects in only 0.3 ms. All results are made open source, with Python code provided that is compatible with Scikit Learn.
The remainder of this article is as follows; Sect. 2 provides a background and review of the literature relevant to this study. Section 3 then describes the methodology followed by each of the experiments included in this work. Section 4 presents the results of all experiments, with Sects. 4.4 and 4.5 presenting the results for neural network neuroevolution and genetic programming of the ML pipelines, respectively. Finally, Sect. 5 concludes the findings of this study and presents suggestions for future work based on them.
2 Background and related work
Falls are most commonly caused by gait instability, confusion and agitation, urinary incontinence and frequency, and the use of prescription sedative and hypnotic drugs, according to a review by Oliver et al. [14]. Ageing, in general, leads to a decrease in balance [15]. Many injuries that occur commonly for people over 65 years of age are more severe and often preventable [16]. According to [17], 33% of adults considered to be elderly will experience a fall on average once per year. This risk is noted to rise later in life, with 50% of over 80’s suffering one or more falls per year. In the United States, there were more than 2.6 million fall-related injuries in year 2000 [18], of which 10,300 were fatal. Most of the time, serious physical injury or death does not occur but does lead to a loss of confidence, social withdrawal, and a feeling of lost independence [19]. This feeling of lost independence may lead to a higher incidence of depression [20]. Fall detection is the use of technology to automatically recognise when someone has fallen, which can then lead to healthcare providers or family members being alerted without any human intervention required. Automatic fall detection alleviates problems after a fall related to situations in which an emergency call button or cord cannot be reached [21]. Studies have shown that a fall event can be detected through several proposed methods that include, but are not limited to, wireless networks [22], computer vision [23], thermal image processing [24], acoustic classification [25], and HAR through wearable sensors [26].
Adkin et al. [27] report that compensatory balance reactions are recognisable within the recorded EEG data. There exists a great overlap in the functions of each of the brain’s lobes, but it is currently understood that much of the coordination involved in balance takes place in the cerebellum, since damage to this area of the brain can negatively affect balance and posture [28, 29]. Vice versa, the volume of the cerebellum was found to be larger within a subject group of high-speed ice skaters [30]. In [31], researchers found that there were significant levels of brain activation during falls within the frontal lobe, specifically the prefrontal cortex, the dorsolateral prefrontal cortex, and the frontal eye field. Given that the cerebellum is found deep within the brain, partly obscured by the cortex, the frontal lobe provides much easier access by non-invasive EEG. Consumer-level technology is the goal of the conducted research in this article, therefore the frontal lobe is selected as the most promising candidate for detecting fall compensatory behaviour. All of the subjects in this study were both healthy and conscious, and thus exhibited normal frontal lobe activity.
In Annese et al. [32], the authors proposed multimodality learning from both EEG and Electromyography (EMG) signals towards machine learning-based fall risk prediction within the design of a specialised digital processor. In this study, EEG focused on the motor cortex and EMG electrodes were placed on the leg muscles. Findings showed that a fall event could be detected 500ms prior to its occurrence because of the brain’s ability to anticipate and compensate for such events. The results in the dataset were almost perfect. The authors note the computational expense of the approach, and it is also worth noting that placing EMG on the legs and EEG with a cap is inconvenient and, therefore, not suitable for everyday use.
A more consumer-ready solution was presented in [11]. Their study explored the use of a helmet with embedded EEG electrodes for the classification of fall events. The dataset collected by the authors was classified at around 98% by a random forest ensemble. The authors note the complexity of having such an exhaustive EEG array and that, in the future, there may be methods of increasing prediction efficiency from an array of fewer electrodes. In addition to the financial costs involved with the trade-off between clinical and consumer-level sensors, errors arising from signal noise (which are more common when operating cheaper sensors [33]) can also be a source of problems for activity recognition. Machine learning-based approaches have been shown to be promising in the removal of signal artefacts [34, 35]. Low-cost sensors have been found to be prone to a variety of problems, and LaRocco et al. [36] argued for the need of algorithmic optimisation.
In the context of artificial intelligence, neuroevolution is a process in which evolutionary algorithms are implemented to generate the hyperparameters of an artificial neural network, given that their selection is a problem of combinatorial optimisation [37]. These hyperparameters can include the topology of the network, that is, how wide the hidden layers are and how deep the network structure is, alongside parameters such as their activation functions, learning rates, and momentum among others. It is, therefore, a form of Automated Machine Learning (AutoML), wherein complex sets of parameters present as the search space [38], with fitness of the solution derived from the ability of the neural network (e.g. from backpropagation on data).
The use of Neuroevolution has recently gained popularity in biological signal processing due to its promising ability to engineer appropriate models. In  [39], researchers proposed the use of neuroevolution for the classification of surface electromyography signals towards recognising hand gestures. Through the application of NeuroEvolution of Augmenting Topologies (NEAT) [40], results noted a mean classification accuracy of 88.76% on signal windows of 150ms. Similarly in the EEG domain, neuroevolution was proposed for the selection of channels prior to learning [41]; in this study, 64 channels of signals posed a problem prior to machine learning, and were gathered from four trans-humeral (upper-arm) amputees. Results showed that a particle swarm optimisation algorithm outperformed other heuristics.
Inspired by the findings of literature review, and given the noted research gaps, the goals of the experiments in this study are to employ meta-heuristics to explore hyperparameter optimisation through neuroevolution of deep neural networks and the genetic programming of machine learning pipelines (including individual hyperparameter sets and ensembles). The algorithms discovered by these approaches provide additional approaches to fall detection, and are made open source with the provision of Python source code in Appendix A for genetic programming pipelines, and Appendix B for neuroevolutionary neural network hyperparameters.
The NeuroSky EEG headset shown in Fig. 1 has a single electrode placed in the FP1 position within the 10–20 EEG electrode placement system. The NeuroSky is most often worn in the position that can be seen in Fig. 2. Although many of the commercial applications of the device are based on concentration classification [43], the NeuroSky has proposed applications in fatigue detection [44], blink detection [45], and fall detection [42].
The NeuroSky MindWave headset which was used to collect EEG fall data in [42]
3 Method
This section describes the methodology followed by the experiments carried out in this article. Firstly, the dataset and data preprocessing are detailed before explaining the hyperparameter optimisation and learning approaches. The aim of these studies is to explore hyperheuristic techniques to improve the detection of falls via biological signals classification.
Figure 3 shows the general approach used by the final outputs of this study. EEG signals are recorded in real-time from the sensor placement as detailed previously in Fig. 2. Following feature extraction, the model classifies whether or not fall compensation behaviour is occurring.
The diagram in Fig. 4 shows the optimisation of feature and model spaces, via genetic programming exercises that treat classification ability as fitness metrics. The goal of this algorithm is to improve the ability of fall compensation detection via EEG, a problem that persists due to the low quality of consumer-level EEG compared to clinical approaches. The neuroevolution approach takes place in the model space only, with hyperparameters of topology, activation, and loss function optimised.
3.1 Dataset and pre-processing
The initial dataset used for this study is the Preliminar Fall-UP Dataset [42]. This dataset, collected in 2019, comprises 11 physical activities performed by 4 subjects for three trials each. Of these activities, a fall event occurred in five and did not occur in six. Since fall events tend to happen quicker than non-fall events such as walking or standing, the dataset is imbalanced when considering binary classification. The activities can be observed in Table 1 alongside the binary class label applied for fall compensation detection. Only the Neurosky EEG brainwave data is used from this dataset.
Feature extraction from the data is required since waves are temporal, i.e. information is presented over time rather than from one singular data object. Time-windowing is a suitable method to extract descriptive information on a per-data-object basis. Feature extraction is the process of extracting these statistical descriptions for classification, and the usefulness is noted in several studies [46,47,48,49,50]. Furthermore, wavelet characteristics have been identified as particularly good features to inform the description of EEG signals [51, 52]. The feature extraction process for this work is as follows; the signals are initially divided into 0.5 s windows, and seven sets of features are extracted, leading to 39 individual features. The spectral entropies of the signals are computed via Fourier transform, which is given as \(F(X) = \sum _{n=0}^{N-1} f[n] e^{-j 2\pi nk/N}\), where \(e^{-j 2\pi nk/N}\) is the complex exponential with frequency index k, n is the sample index, and f[n] is the input sequence. Shannon entropy \(H(X)=- \sum _n^i=1 P(x_{i}) log P(x_{i})\) is also extracted, where P is the power spectrum and probability distribution of the input signal. For each wavelet scale up to 8, several features are extracted following a continuous wavelet transform. They are the absolute mean, energy, entropy, standard deviation, and variance. All features are normalised via min-max scaling on the scale \(0-1\). We first explore the relative entropy \(IG(T, a) = E(T) - E(T | a)\) via entropy \(E(s)= -\sum _j p_{j} log( p_{j} )\) to observe how much information each of the attributes carry for prediction, as well as which features are particularly useful, if any. The application of the aforementioned algorithms to the data leads to single data objects that describe a temporal sequence. For example, a single reading from a signal would give one value at a given timestep. This reading holds no useful information to classify the signal, since this information is derived over time given the behaviour of the wave. Given that most machine learning algorithms do not take into account temporal sequences, the methodology of this work instead generates mathematical descriptions of time windows to provide input data. To conclude this, one row of processed data describes half a second of EEG data and can be used as input to any machine learning approach, rather than being limited to those that are temporal.
Following preprocessing, it was observed that the classes were imbalanced at an approximate 5:1 ratio for non-falling and falling, respectively. Falling data was represented by 1102 data objects whereas the non-fall class had 5032, leading to a mean class label of 0.18 (where 0 is for non-fall and 1 is for fall) at a standard deviation of 0.384. For balancing, a simple random (seed = 1) undersample of 1102 non-fall instances are taken to provide a balanced dataset of 2204 data objects in total.
3.2 Hyperparameter optimisation and learning
The hyperparameter optimisation processes for the three sets of experiments is explained in this section. Firstly, classical linear and batch searches of statistical model parameters. Secondly, the optimisation of deep neural network topologies, activation functions, and loss functions. Finally, the third set of experiments is described in which a tree-based genetic search is used to optimise statistical model pipelines (that is, those detailed in Fig. 4).
Hyperparameters for the K-Nearest Neighbour [53] (KNN) and Random Forest [54] models are initially explored through a simple linear search \(k = \{10,20,..,90,100\}\) to discern whether hyperparameter tuning has a noticeable effect on predictive ability. KNN is a clusetering algorithm which classifies an unknown data object by its Euclidean distance to labelled points in n-dimensional space, where n is the number of attributes. Random Forests are an ensemble of Random Decision Trees (RDTs) voting on prediction, and RDTs classify data based on splitting to reduce entropy. Various ML algorithms are selected with a range of different statistical methods to provide a general overview of the classification ability using multiple methods (see Sect. 4.6 for more details). Following this, further tuning is performed via Adaptive Boosting [55] on all the selected models that are compatible with the algorithm.
The second set of hyperparameter optimisation experiments involves the evolutionary optimisation of neural network parameters, and the genetic programming of machine learning pipelines. The simulations are executed five times, with a population size of 30 for 50 generations. Each initialisation of the two search algorithms is given random seeds equal to their iteration, one through five.
The controlled hyperparameter limits for the evolutionary search experiments can be found in Table 2. Larger ranges from the original studies were attempted manually prior to experimentation, revealing severely low classification metrics. Each of the networks are given 300 epochs to train at a batch size of 200.
Finally, a Genetic Programming (GP) approach is then explored using a tree-based algorithm, as detailed in [56]. The GP tree is given access to all of the algorithms included with scikit-learn alongside the Extreme Gradient Boosting library [57]. The GP algorithm runs for a total of 50 generations, with a population size of 20. Mutation rate is set to 0.9, and there is a crossover rate of 0.1.
All algorithms in this work are trained by 10-fold cross-validation with a seed set to 1 for randomisation and are therefore directly comparable. For all heuristic searches, the population size was 25 simulated for 50 generations. The probability of crossover was selected as 0.8 and mutation at 0.1. The evolution strategy was \((\mu + \lambda )\). All algorithms were trained on an overclocked Intel Core i7-8700K CPU (4.3GHz) with scikit-learn [58], DEAP [59], and TPOT [56].
4 Results
This section presented the results of all planned experiments. First, we will explore the usefulness of the features extracted from the signals in conjunction with related observations. Following this, hyperparameter optimisation of models is explored through the selected methods of linear searching, neuroevolution, and genetic programming. The source code for the main experiments is given in Appendices A and B.
4.1 Data preprocessing
Following the preprocessing and subsampling (for class balance) strategies described in Sect. 3.1, the dataset comprised of 2204 samples. The measurements of relative entropy by 10-fold cross-validation are presented in Table 3. It can be observed that three features, in particular, carry more information relative to the rest; those were the absolute mean on the eighth wavelet scale, the variance of the third wavelet scale, and the variance of the fourth wavelet scale. An example as to why class balancing is used can be observed in Table 4. When the dataset is unbalanced, there is a higher frequency of EEG signals linked to activities related to not falling, and thus they are much easier to classify on average. Due to this, misleading results can be achieved; for example, the class balanced approach has a lower classification accuracy (83.3% vs. 92.21%), the ability to recognise the falling behaviour is improved from 885 correct instances to 980. The baseline (random guess or application of the most common label) for the balanced dataset is 50% while it is 82.03% for the unbalanced dataset. Therefore, balancing in this preliminary example provides a 33.3% advantage over the baseline, whereas leaving the dataset unbalanced provides only a 10.18% advantage.
When normalising, each of the values of attributes then shares a common scale, without distortion of ranges or information loss. A preliminary experiment is performed on the normalised and non-normalised data in Table 5. It is observed that the classification metrics increase slightly after normalisation is used as a preprocessing technique. Due to these examples and discussion, the normalised and equally balanced dataset is chosen for the remainder of the experiments presented in this work.
4.2 Hyperparameter tuning
A comprehensive investigation was conducted to find out the linear hyperparameter search for estimators in the Random Forest. The overall best approach of this search was a random forest of 80 trees, which had a mean accuracy of 84.94%. The model also had a precision of 0.81, a recall 0.915, and an F-Score of 0.856. These were the highest observed metrics within the linear search except for mean precision, where a Random Forest of 50 trees scored 0.81.
Following the same approach, a linear search of hyperparameter k in K-Nearest Neighbours can be observed. The strongest approach revealed during the search was \(k=40\), which had a mean accuracy of 73.37%, a precision of 0.793, a recall of 0.634, and an F score of 0.704.
4.3 Adaptive boosting
Models that had the ability to predict probabilities, and thus are compatible with the adaptive boosting algorithm [60], were adaptively boosted. The boosting results can be observed in Table 6 with a comparison between models then presented in Fig. 5. Adaptive boost leads to lower results on more than one occasion. Random Forests and Naive Bayes models lead to a lower mean classification accuracy. On the other hand, Logistic Regression and Stochastic Gradient Descent models can be improved with boosting. It must be noted that boosting is computationally expensive compared to many of the approaches explored in this work.
4.4 Neuroevolution of network topology
Following five evolutionary topology searches, solutions were presented with varying sizes of neural networks with different hyperparameters. Figure 6 shows how the mean accuracy of best solutions evolved over generations. Since many of the solutions presented close results, the search often stagnated relatively early, especially Iteration 2. The best solution found was by that of Iteration 3, which scored a mean 73.41% accuracy. The hyperparameters selected for this neural network were three hidden layers of 29, 10, and 9 neurons with a hyperbolic tangent activation function. The entire source code for the neural network hyperparameters can be found in Appendix B. Note that extremely fine values of parameters such as alpha and beta values, learning rates, and momentum were selected, to an extent that would not be tested manually or by batch search. The final results of all evolutionary neural network searches can be observed in Table 7, and the trade-off between accuracy versus training and inference times can be found in Table 8. It is worth noting here that although the models take much longer to train compared to the other algorithms explored in this work, there is no payoff in terms of gaining accuracy.
4.5 Genetic programming
As previously described, the genetic programming approach explored 50 generations with 30 solutions as a population size. The learning process for five iterations of the GP algorithm can be observed in Fig. 7, and the best final solutions are further detailed in Table 9. Although starting at the highest fitness, Iteration 1 had the lowest final score of 88.79%, with Iteration 2 (which started at the lowest fitness) scoring slightly more by the end of the simulation at 88.79%. The best solution found was that by Iteration 3, which scored 89.34%. Due to their complexity, the solutions are presented by their iteration ID in this work—the source code for all three machine learning pipelines can be found in Appendix A. Although wavelet features are extracted manually, it can be observed that there was further feature engineering through Principal Component Analysis (PCA) and Polynomial combinations, which are often also suggested in the literature [61,62,63].
Table 10 shows the tradeoff between model complexity as training and inference times compared with the average ability of the model. It can be noted that, although some algorithms were more complex and required considerably more resources, there was a diminishing return on ability. In fact, these models were outperformed by algorithms that could train in under one second. GP1-5 denote the best solutions after five individual Genetic Programming searches. Although the solution found by GP1 had the highest accuracy at a mean value of 90.52%, it took considerably longer to train than the other solutions at around 4.4Â s. The second-best solution, GP2, took the least training time of around 0.87Â s and achieved a mean accuracy of 89.79%. The second best solution also had the smallest inference time, at only 0.0003Â s per prediction. The best model extracted polynomial features and performed PCA on them prior to the prediction probabilities of an extra trees classifier were presented as input for a KNN of 72.
4.6 Comparison of all models
A final comparison of all models is provided in Table 11. For readability purposes of the table, a key to the abbreviations is as follows: RF—Random Forest, AB—Adaptive Boosting, KNN—K-Nearest Neighbours, LDA—Linear Discriminant Analysis, LR—Logistic Regression, L SVM—Linear Support Vector Machine, RBF SVM—Radial Basis Function Support Vector Machine, SGD—Stochastic Gradient Descent, QDA—Quadratic Discriminant Analysis, NB—Gaussian Naive Bayes. As can be observed, the best models were those that were explored through genetic programming. Interestingly, the adaptive boost of the Naive Bayes model was worse than random guessing, and this was the only instance of such an occurrence. The Receiver Operating Characteristic (ROC) and Precision-Recall curves are useful metrics for visualising model performance and ability. These can be found in in Figs. 8 and 9, respectively. The ROC curve shows a high true positive rate and low false positive rate, while the Precision-Recall curve demonstrates a high precision-recall trade-off. Furthermore, the area under the ROC curve (AUC-ROC) and the area under the Precision-Recall curve (AUC-PR) were found to be 0.943 and 0.939, respectively, providing valuable insights into the overall performance of the model.
4.7 Computational complexity versus accuracy tradeoff
The trade-off between model ability and complexity can be observed in Table 12. Additionally, Figs. 10 and 11 show a visual representations of accuracy versus training time, and accuracy versus inference time, respectively. The best single model bar genetic programming, the Gaussian Process, scored a mean accuracy of 86.48% but required around 359 s to train, which was by far the highest computational requirement of all models. Although the training time was high, inference took only 0.33 ms per data object. As can be observed, the genetic programming solutions have training times similar to most other models while attaining the highest classification accuracy scores on average. Inference time is an important aspect when it comes to the real-world application of the approach. The time taken to infer a data object is one and the same with the time taken to detect when someone has fallen. Thus, it is important to consider the accuracy/inference tradeoff when choosing a model given that it will affect the response time of a fall detection system. Therefore, even though the second solution is slightly worse at an average accuracy decrease of 0.73%, it is likely a more appropriate choice for use due to the inference time of 0.33 ms, as opposed to the 18.92 ms inference of the best performing model.
Finally, the best model, GP1, is validated by training based on three subjects and testing the remaining subjects, with the results presented in Table 13. Through leave-one-subject-out cross-validation, we observe a mean accuracy of 75.04%. The highest ability was found to be when Subject 1 was the test, scoring 82.95% accuracy. It was relatively more difficult to generalise to Subject 4, where the model reached only 67.02%. These results show that the dataset is not diverse enough for generalisation between individuals, and experiments to benchmark this are limited by the fact that only four subjects are present. This suggests that future experiments should concern a larger group of subjects, with the aim for further generalisation.
5 Conclusion and future work
To finally conclude, this work has explored how techniques such as genetic programming and neuroevolution may provide solutions to learning from low-cost and low quality EEG electrodes for consumer use. Fall detection from this electrode was observed to be a difficult task, but hyperheuristic solutions found machine learning pipelines and hyperparameter sets to enable this possibility. Although the problem was difficult, due in part to activities such as laying down being present in the category of not falling, genetic programming developed a machine learning pipeline that could detect falls with an average accuracy of 90.52% in only 0.019Â s per data object. The second best solution, also a genetic programming pipeline, achieved a slightly lower 89.79% mean accuracy but could classify data objects in only 0.3 ms.
The results presented in this work provide a good basis for further experiments in the future, given that some approaches were particularly worse than the more impressive set of results. Most interesting of these future works concerns the leave-one-subject-out validation, which found that some subjects could be generalised to with high accuracy, and some faltered compared to the overall metrics observed through 10-fold cross-validation. In future, and with a larger dataset, work could concern to which extent this activity can generalise more when observing a greater number of subjects. Given a larger number of subjects, generalisation can be aimed for as a goal of the Genetic Search via deriving fitness scores from leave-one-subject-out cross validation. Due to the nature of the algorithm, there is an experimental limitation within the search for neural net hyperparameters, since layers two and three were both a problem space of \(1-128\), the solution space was thus x, n in some cases provided that layer 2 was 0 and layer 3 was \(>0\). Therefore, a small number of solutions were identical but treated as unique individuals. In addition to running algorithms for a longer period of time, future work could also concern testing other hyperparameters for the evolutionary strategies, although this would add a further layer of complexity to the approach and would require considerably more computational resources. These include the population size, crossover and mutation rates, as well as the overall mutation strategy. In addition to the models explored, future work could involve the multimodal classification of falls by including information collected by other sensors e.g. those which are wearable and ambient sensors around the home environment. Finally, deep learning and data augmentation could be explored towards methods that can be tuned in the future as more data becomes available, since neural networks may then become a more viable option.
Availability of data and materials
All data is available from [42] (The FALL-UP dataset can be downloaded from: https://sites.google.com/up.edu.mx/har-up/).
Code availability
All code for implementation is presented in the appendices of this mansucript.
Notes
This paper is an extended version of the work reported by authors earlier in [13]. The studies involved in this work are an extension of the work which discovered that Genetic Programming was a viable solution to hyperparameter optimisation for EEG-based fall detection.
References
R.W. Broadley, How to Build Better Fall Detection Technology: A Search for Characteristics Unique to Falls and Methods to Robustly Evaluate Performance. PhD thesis, (University of Salford United Kingdom, 2020)
V.-R. Xefteris, A. Tsanousa, G. Meditskos, S. Vrochidis, I. Kompatsiaris, Performance, challenges, and limitations in multimodal fall detection systems: a review. IEEE Sens. J. 21(17), 18398–18409 (2021)
F. Prince, H. Corriveau, R. Hébert, D.A. Winter, Gait in the elderly. Gait Posture 5(2), 128–135 (1997)
Office for National Statistics: Dataset: Vital statistics in the UK: births, deaths and marriages. Office for National Statistics (2021). https://www.ons.gov.uk/peoplepopulationandcommunity/populationandmigration/populationestimates/datasets/vitalstatisticspopulationandhealthreferencetables
J. Braithwaite, R. Mannion, Y. Matsuyama, P.G. Shekelle, S. Whittaker, S. Al-Adawi, K. Ludlow, W. James, H.P. Ting, J. Herkes et al., The future of health systems to 2030: a roadmap for global progress and sustainability. Int. J. Quality Health Care 30(10), 823–831 (2018)
P. Rashidi, A. Mihailidis, A survey on ambient-assisted living tools for older adults. IEEE J. Biomed. Health Inform. 17(3), 579–590 (2012)
A.H. Sapci, H.A. Sapci, Innovative assisted living tools, remote monitoring technologies, artificial intelligence-driven solutions, and robotic systems for aging societies: systematic review. JMIR Aging 2(2), 15429 (2019)
N. Lu, Y. Wu, L. Feng, J. Song, Deep learning for fall detection: three-dimensional CNN combined with LSTM on video kinematic data. IEEE J. Biomed. Health Inform. 23(1), 314–323 (2018)
J. Santiago, E. Cotto, L.G. Jaimes, I. Vergara-Laurens, Fall detection system for the elderly, in 2017 IEEE 7th Annual Computing and Communication Workshop and Conference (CCWC). IEEE, pp. 1–4 (2017)
A. Leone, G. Rescio, A. Caroppo, P. Siciliano, A wearable EMG-based system pre-fall detector. Procedia Eng. 120, 455–458 (2015)
S.R. Dhole, A. Kashyap, A.N. Dangwal, R. Mohan, A novel helmet design and implementation for drowsiness and fall detection of workers on-site using EEG and random-forest classifier. Procedia Comput. Sci. 151, 947–952 (2019)
W. Siwadamrongpong , J. Chinrungrueng , S. Hasegawa , E. Nantajeewarawat, Fall detection and prediction based on IMU and EMG sensors for elders, in 2022 19th International Joint Conference on Computer Science and Software Engineering (JCSSE). IEEE, pp. 1–6 (2022)
J.J. Bird, Eeg wavelet classification for fall detection with genetic programming, in Proceedings of the 15th International Conference on PErvasive Technologies Related to Assistive Environments, pp. 376–382 (2022)
D. Oliver, F. Daly, F.C. Martin, M.E. McMurdo, Risk factors and risk assessment tools for falls in hospital in-patients: a systematic review. Age Ageing 33(2), 122–130 (2004)
D. Eibling, Balance disorders in older adults. Clin. Geriatr. Med 34(2), 175–181 (2018)
J.M. Rothschild, D.W. Bates, L.L. Leape, Preventable medical injuries in older patients. Arch. Intern. Med. 160(18), 2717–2728 (2000)
M.E. Tinetti, M. Speechley, S.F. Ginter, Risk factors for falls among elderly persons living in the community. New Engl. J. Med. 319(26), 1701–1707 (1988)
J.A. Stevens, P.S. Corso, E.A. Finkelstein, T.R. Miller, The costs of fatal and non-fatal falls among older adults. Injury Prev. 12(5), 290–295 (2006)
National Health Service: Falls. NHS (2021). https://www.nhs.uk/conditions/falls/
F.M. Alpass, S. Neville, Loneliness, health and depression in older males. Aging Mental Health 7(3), 212–216 (2003)
M. Mubashir, L. Shao, L. Seed, A survey on fall detection: principles and approaches. Neurocomputing 100, 144–152 (2013)
Y. Wang, K. Wu, L.M. Ni, Wifall: device-free fall detection by wireless networks. IEEE Trans. Mobile Comput. 16(2), 581–594 (2016)
C. Rougier, J. Meunier, A. St-Arnaud, J. Rousseau, Fall detection from human shape and motion history using video surveillance, in 21st International Conference on Advanced Information Networking and Applications Workshops (AINAW’07), (vol. 2, 2007). IEEE, pp. 875–880
A. Naser, A. Lotfi, J. Zhong, Multiple thermal sensor array fusion towards enabling privacy-preserving human monitoring applications. IEEE Internet Things J. 9(17), 16677–16688 (2022)
Y. Li, K. Ho, M. Popescu, A microphone array system for automatic fall detection. IEEE Trans. Biomed. Eng. 59(5), 1291–1301 (2012)
J. Chen, K. Kwong, D. Chang, J. Luk, R. Bajcsy, Wearable sensors for reliable fall detection, in 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference. IEEE, pp. 3551–3554 (2006)
A.L. Adkin, S. Quant, B.E. Maki, W.E. McIlroy, Cortical responses associated with predictable and unpredictable compensatory balance reactions. Exp. Brain Res. 172(1), 85–93 (2006)
E.V. Sullivan, M.J. Rosenbloom, A. Deshmukh, J.E. Desmond, A. Pfefferbaum, Alcohol and the cerebellum: effects on balance, motor coordination, and cognition. Alcohol Health Res. World 19(2), 138 (1995)
S.M. Morton, A.J. Bastian, Cerebellar control of balance and locomotion. The Neuroscientist 10(3), 247–259 (2004)
I.S. Park, N.J. Lee, T.-Y. Kim, J.-H. Park, Y.-M. Won, Y.-J. Jung, J.-H. Yoon, I.J. Rhyu, Volumetric analysis of cerebellum in short-track speed skating players. The Cerebellum 11(4), 925–930 (2012)
M. Mihara, I. Miyai, M. Hatakenaka, K. Kubota, S. Sakoda, Role of the prefrontal cortex in human balance control. Neuroimage 43(2), 329–336 (2008)
V.F. Annese, M. Crepaldi, D. Demarchi, D. De Venuto, A digital processor architecture for combined eeg/emg falling risk prediction, in 2016 Design, Automation & Test in Europe Conference & Exhibition (DATE). IEEE, pp. 714–719 (2016)
M.-H.J. Lin, S.N. Cross, W.J. Jones, T.L. Childers, Applying EEG in consumer neuroscience. Eur. J. Market. 52(1/2), 66–91 (2018)
S. Stalin, V. Roy, P.K. Shukla, A. Zaguia, M.M. Khan, P.K. Shukla, A. Jain, A machine learning-based big EEG data artifact detection and wavelet-based removal: an empirical approach. Math. Probl. Eng. 2021, 1–11 (2021)
M. Fabietti, M. Mahmud, A. Lotfi, Channel-independent recreation of artefactual signals in chronically recorded local field potentials using machine learning. Brain Inform. 9(1), 1–17 (2022)
J. LaRocco, M.D. Le, D.-G. Paeng, A systemic review of available low-cost EEG headsets used for drowsiness detection. Front. Neuroinform. 42, (2020)
K.O. Stanley, J. Clune, J. Lehman, R. Miikkulainen, Designing neural networks through neuroevolution. Nature Mach. Intell. 1(1), 24–35 (2019)
D. Floreano, P. Dürr, C. Mattiussi, Neuroevolution: from architectures to learning. Evolut. Intell. 1, 47–62 (2008)
Y. Dweiri, Y. Hajjar, O. Hatahet, A novel neuroevolution model for EMG-based hand gesture classification. Neural Comput. Appl. 35, 10621–10635 (2023)
K.O. Stanley, R. Miikkulainen, Evolving neural networks through augmenting topologies. Evolut. Comput. 10(2), 99–127 (2002)
O.P. Idowu, O. Adelopo, A.E. Ilesanmi, X. Li, O.W. Samuel, P. Fang, G. Li, Neuro-evolutionary approach for optimal selection of EEG channels in motor imagery based BCI application. Biomed. Signal Process. Control 68, 102621 (2021)
L. MartÃnez-Villaseñor, H. Ponce, J. Brieva, E. Moya-Albor, J. Núñez-MartÃnez, C. Peñafort-Asturiano, Up-fall detection dataset: a multimodal approach. Sensors 19(9), 1988 (2019)
B. Zhang, J. Wang, T. Fuhlbrigge, A review of the commercial brain-computer interface technology from perspective of industrial robotics, in 2010 IEEE International Conference on Automation and Logistics. IEEE, pp. 379–384 (2010)
D.H. Joshi, U. Jaliya, D. Thakore, Raw EEG-based fatigue and drowsiness detection: a review. Int. Inst. Technol. Res. Develop. 1(1) (2015)
M. Sahu, P. Shukla, A. Chandel, S. Jain, S. Verma, Eye blinking classification through neurosky mindwave headset using eegid tool, in International Conference on Innovative Computing and Communications, (Springer, 2021), pp. 789–799
F. Lotte, M. Congedo, EEG feature extraction. Brain–Computer Interfaces 1: Foundations and Methods, 127–143 (2016)
S. Krishnan, Y. Athavale, Trends in biomedical signal feature extraction. Biomed. Signal Process. Control 43, 41–63 (2018)
L. Hu, Z. Zhang, EEG Signal Processing and Feature Extraction, 1st edn. (Springer, Singapore, 2019)
J.J. Bird, D.R. Faria, L.J. Manso, A. Ekárt, C.D. Buckingham, A deep evolutionary approach to bioinspired classifier optimisation for brain-machine interaction. Complexity 2019, (2019)
J.J. Bird, M. Pritchard, A. Fratini, A. Ekárt, D.R. Faria, Synthetic biological signals machine-generated by GPT-2 improve the classification of EEG and EMG through data augmentation. IEEE Robot. Autom. Lett. 6(2), 3498–3504 (2021)
A. Subasi, EEG signal classification using wavelet feature extraction and a mixture of expert model. Expert Syst. Appl. 32(4), 1084–1093 (2007)
E.D. Übeyli, Combined neural network model employing wavelet coefficients for EEG signals classification. Digit. Signal Process. 19(2), 297–308 (2009)
E. Fix, J.L. Hodges, Discriminatory analysis. Nonparametric discrimination: Consistency properties. International Statistical Review/Revue Internationale de Statistique 57(3), 238–247 (1989)
L. Breiman, Random forests. Mach. Learn. 45(1), 5–32 (2001)
R.E. Schapire, Explaining adaboost, in Empirical Inference, (Springer, Berlin, Heidelberg, 2013), pp. 37–52
R.S. Olson, N. Bartley, R.J. Urbanowicz, J.H. Moore, Evaluation of a tree-based pipeline optimization tool for automating data science, in Proceedings of the Genetic and Evolutionary Computation Conference 2016, pp. 485–492 (2016)
T. Chen, T. He, M. Benesty, V. Khotilovich, Y. Tang, H. Cho, K. Chen et al., Xgboost: extreme gradient boosting. R package version 0.4-2 1(4), 1–4 (2015)
F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, E. Duchesnay, Scikit-learn: machine learning in Python. J. Mach. Learn. Res. 12, 2825–2830 (2011)
F.-A. Fortin, F.-M. De Rainville, M.-A. Gardner, M. Parizeau, C. Gagné, DEAP: evolutionary algorithms made easy. J. Mach. Learn. Res. 13, 2171–2175 (2012)
Y. Freund, R.E. Schapire, A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 55(1), 119–139 (1997)
A. Subasi, M.I. Gursoy, EEG signal classification using PCA, ICA, LDA and support vector machines. Expert Syst. Appl. 37(12), 8659–8666 (2010)
K. Mahajan, M. Vargantwar, S.M. Rajput, Classification of EEG using PCA, ICA and neural network. Int. J. Eng. Adv. Technol. 1(1), 80–83 (2011)
L.C.D. Nkengfack, D. Tchiotsop, R. Atangana, V. Louis-Door, D. Wolf, EEG signals analysis for epileptic seizures detection using polynomial transforms, linear discriminant analysis and support vector machines. Biomed. Signal Process. Control 62, 102141 (2020)
Funding
Not applicable, no funding was involved with this research and manuscript.
Author information
Authors and Affiliations
Contributions
J.B. and A.L. wrote the main text of the manuscript. All authors reviewed the manuscript.
Corresponding author
Ethics declarations
Conflict of interest
The authors confirm that there is no conflict of interest.
Ethical approval
Not applicable, the experiments in this study used publicly available data.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Area Editor: Sebastian Risi.
Appendices
Appendix A: Python source code for the genetic programming solutions
This appendix provides the source code for the final solutions found by the five iterations of genetic programming. Python 3.x code is presented and is compatible with the scikit-learn library.
1.1 A.1 Iteration 1

1.2 A.2 Iteration 2

1.3 A.3 Iteration 3

1.4 A.4 Iteration 4

1.5 A.5 Iteration 5

Appendix B: Python source code for the evolutionary MLP solutions
This appendix provides the source code for the final solutions found by the five iterations of evolutionary search for neural network hyperparameters. Python 3.x code is presented and is compatible with the scikit-learn library.
1.1 B.1 Iteration 1

1.2 B.2 Iteration 2

1.3 B.3 Iteration 3

1.4 B.4 Iteration 4

1.5 B.5 Iteration 5

Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Bird, J.J., Lotfi, A. Fall compensation detection from EEG using neuroevolution and genetic hyperparameter optimisation. Genet Program Evolvable Mach 24, 6 (2023). https://doi.org/10.1007/s10710-023-09453-3
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10710-023-09453-3
Keywords
- Evolutionary optimisation
- Genetic programming
- Fall detection
- EEG
- Hyperheuristics
- Signal classification