Fall detection studies using machine learning require the definition of a problem depending on the desired outcome. In this study, fall detection is a binary classification problem that involves making decisions between daily activities and fall events. Such classification problems can be addressed using supervised machine learning methods. In this study, supervised classification algorithms were preferred because the data used is labeled, and artificial intelligence models developed with labeled data tend to yield more successful results.
When training an artificial intelligence model, the following steps are typically followed: problem identification, data acquisition, data analysis, data preprocessing, creation of a training dataset, training of the dataset, testing the model with a test dataset, and evaluating its performance using performance criteria. The training process of the artificial intelligence model is completed by selecting the most successful model (Fig. 4).
3.1 Datasets
The fall detection system used in this study is based on data obtained through two different brand IMU and volunteers. The first dataset is obtained from the MTw commercial IMU and is available as an open source [21]. The second IMU dataset is obtained from an Activity Tracking Device (ATD) designed to detect falls and daily activities. Both IMUs are collected from different volunteers with different physical characteristics.
3.1.1 MTw Dataset
The dataset consists of data obtained by placing the Xsens sensor on the lower back, capturing daily life activities (DLAs), and fall events. The proposed experimental protocol was adopted for DLAs and fall events [28]. The experiments were conducted by 14 volunteers, consisting of 7 females and 7 males. Female volunteers are aged between 19–23, with weights ranging from 47–70 kg and heights ranging from 165–174 cm. Male volunteers are aged between 21–25, with weights ranging from 55–80 kg and heights ranging from 160–184 cm (Table 5).
Table 5
Anthropometric and physical information of volunteers for MTw.
Gender | Volunteer Code | Height (cm) | Weight (kg) | Age |
| 101 | 170 | 75 | 21 |
| 102 | 174 | 81 | 21 |
| 103 | 180 | 78 | 23 |
Male | 104 | 176 | 67 | 27 |
| 106 | 160 | 54 | 22 |
| 107 | 175 | 72 | 21 |
| 108 | 184 | 68 | 21 |
| 203 | 170 | 51 | 21 |
| 204 | 157 | 47 | 21 |
| 205 | 169 | 51 | 20 |
Female | 206 | 166 | 47 | 19 |
| 207 | 165 | 60 | 20 |
| 208 | 163 | 55 | 24 |
| 209 | 182 | 70 | 22 |
| | All Volunteer | |
Avarage | 170.79 | 62.57 | 21.64 |
Standard Deviation | 8.17 | 11.82 | 1.98 |
The MTw dataset consists of 2520 records (14 volunteers x 5 repetitions x (20 Falls + 16 DLAs)). These records are obtained from 20 fall activities and 16 DLAs of the 14 volunteers, each repeated 5 times.
3.1.2. ATD Dataset
The accelerations and fall events in the ATD dataset were determined based on the activity types collected with MTw. The experiments were conducted with 30 volunteers, consisting of 15 females and 15 males. The age, weight, and height ranges of the female volunteers were measured as 18–41 years, 45–73 kg, and 156–175 cm, respectively. The male volunteers were within the age range of 18–50, weight range of 62–100 kg, and height range of 168–191 cm (Table 6).
Table 6
Anthropometric and physical information of volunteers for ATD.
Gender | Volunteer Code | Height (cm) | Weight (kg) | Age |
Male | 101 | 179 | 100 | 39 |
102 | 174 | 96 | 37 |
103 | 191 | 69 | 18 |
104 | 178 | 86 | 20 |
105 | 175 | 89 | 18 |
106 | 168 | 93 | 50 |
107 | 170 | 62 | 21 |
108 | 168 | 90 | 37 |
109 | 182 | 75 | 20 |
110 | 192 | 60 | 18 |
111 | 178 | 76 | 34 |
112 | 176 | 71 | 21 |
113 | 170 | 87 | 41 |
114 | 174 | 81 | 37 |
115 | 180 | 67 | 22 |
Female | 201 | 163 | 50 | 34 |
202 | 169 | 69 | 19 |
203 | 166 | 53 | 21 |
204 | 162 | 62 | 19 |
205 | 156 | 73 | 41 |
206 | 172 | 70 | 18 |
207 | 163 | 45 | 19 |
208 | 165 | 53 | 19 |
209 | 162 | 59 | 23 |
210 | 169 | 73 | 21 |
211 | 159 | 48 | 21 |
212 | 175 | 63 | 19 |
213 | 157 | 64 | 20 |
214 | 168 | 63 | 22 |
215 | 160 | 60 | 18 |
Avarage | 170.70 | 70.23 | 25.57 |
Standard Deviation | 9.01 | 14.73 | 9.32 |
The data obtained from the volunteers were used for binary classification in fall detection. In the study, falls were labeled as 1 and activities as 0. The MTw dataset (30 volunteers x 3 repetitions x (15 activities + falls)) consists of 1350 records. These records are obtained from 20 fall activities and 16 activities of the 14 volunteers, each repeated 5 times.
3.1.3. Hybrid Dataset and Data Formation
Both data sets were collected at 25 Hz, resulting in data of equal resolution. Each activity is represented by a two-dimensional matrix consisting of 101 rows and 9 columns of sensor axis data. These matrices include the sensor values of the accelerometer (Ax, Ay, Az), magnetometer (Mx, My, Mz), and gravity acceleration (Gx, Gy, Gz). Each activity data (DLA and fall) contains a duration of 4 seconds. Each activity (101x9 matrix + 1 class column) includes sensor data (Fig. 5).
All activities are combined into a single dataset, resulting in 20 falls + 16 DLAs = 36 activities. Since each activity has 5 repetitions, it is transformed into a two-dimensional matrix of size 2520x910, consisting of 2520 rows and 910 columns. Thus, each data point collected at 25 Hz within 4 seconds is organized as a feature. As a result, each activity is represented in one row of a two-dimensional matrix with dimensions 2520x910. This allows for easy modeling of time series data using machine learning algorithms.
Each activity consists of 1 row and 910 columns, including 909 sensor data columns and 1 label column (Ax0, Ay0, Az0, Gx0, Gy0, Gz0, Mx0, My0, Mz0, ..., Ax100, Ay100, Az100, Gx100, Gy100, Gz100, Mx100, My100, Mz100, and Class Label). The MTw dataset is obtained from two different IMUs and has dimensions of 2520x910, while the ATD dataset is obtained from a single IMU and has dimensions of 1350x910. Both datasets are organized in a time series approach.
For the machine learning approach, the datasets are divided into train and test sets. The train part of the MTw dataset consists of data collected from 10 volunteers (1800x910), while the data collected from 4 volunteers (720x910) forms the MTw_test dataset (Fig. 4). The ATD dataset is divided into ATD_train, consisting of data from 22 individuals (990x910) for training, and ATD_test, consisting of data from 8 volunteers (360x910) for testing. It is crucial to exclude the data from the same individuals in the testing phase to ensure the reliability of the tests.
After the separation of data for training and testing, they are combined within their respective groups to obtain the hybrid_train dataset (2790x910) and the hybrid_test dataset (1080x910).
3.2 Performance Evaluation
The data sets used for training were subjected to training with various machine learning algorithms before being merged, with aim of determining successful algorithm. Models of successful algorithms were then applied to the respective data sets. After data sets were merged, hybrid model was trained, and predictions were made using this model. Performance metrics were extensively analyzed, including success rates in inter-brand test data sets and the hybrid data set. The performance of the hybrid model using different brand sensor data and the hybrid data set was evaluated comparatively.
3.2.1 Evaluation of MTw Dataset
The dataset obtained from the MTw sensor was transformed into a matrix of size 2520x910 and analyzed using a time-series approach. For training purposes, the data of 10 volunteers were used, and the data of 4 volunteers were reserved for testing. As a result of training studies, the Ada Boost algorithm was identified as the most successful model. The confusion matrix used for performance evaluation is shown in Fig. 6. In this matrix, 0.0 represents Activities of Daily Living (ADL) and 1.0 represents fall activities. While the Ada Boost classification algorithm correctly classified 397 out of 400 fall activities as falls, it incorrectly predicted 3 of them as ADL (Fig. 6). Additionally, it correctly classified 318 out of 320 ADL activities as ADL, and incorrectly predicted 2 of them as falls. These results indicate that Ada Boost algorithm has a high success rate.
These high-performance criteria can be expressed as accuracy (99.31%), area under the curve (AUC) (99.94%), recall (99.38%), precision (99.07%), and F1-score (99.22%). The classification performance of the model trained with the MTw dataset on the unused test dataset is presented in Table 7
Table 7
Performance metrics of models trained with MTw dataset and tested on MTw test dataset
Model | Accuracy | AUC | Recall | Precision | F1 |
Ada Boost Classifier | 0.9931 | 0.9994 | 0.9938 | 0.9907 | 0.9922 |
Extra Trees Classifier | 0.9917 | 0.9997 | 0.9875 | 0.9937 | 0.9906 |
Extreme Gradient Boosting | 0.9917 | 0.9998 | 0.9938 | 0.9876 | 0.9907 |
Logistic Regression | 0.9903 | 0.9996 | 0.9875 | 0.9906 | 0.9890 |
Light Gradient Boosting Machine | 0.9889 | 0.9998 | 0.9906 | 0.9845 | 0.9875 |
Gradient Boosting Classifier | 0.9847 | 0.9995 | 0.9875 | 0.9783 | 0.9829 |
K Neighbors Classifier | 0.9847 | 0.9963 | 0.9781 | 0.9874 | 0.9827 |
Ridge Classifier | 0.9750 | 0.9728 | 0.9531 | 0.9903 | 0.9713 |
Linear Discriminant Analysis | 0.9736 | 0.9891 | 0.9500 | 0.9902 | 0.9697 |
Decision Tree Classifier | 0.9736 | 0.9734 | 0.9719 | 0.9688 | 0.9704 |
These 10 algorithms trained on MTw dataset were used to classify ATD test dataset. Thus, performance of a model trained on different brand sensor data was observed on normalized sensor data. These results are presented in Table 8.
Table 8
Classification performance of models trained with MTw dataset tested on ATD test dataset
Model | Accuracy | AUC | Recall | Precision | F1 |
Gradient Boosting Classifier | 0.5333 | 0.4495 | 0.0000 | 0.0000 | 0.0000 |
Extreme Gradient Boosting | 0.5333 | 0.3129 | 0.0000 | 0.0000 | 0.0000 |
Ada Boost Classifier | 0.5167 | 0.2809 | 0.0000 | 0.0000 | 0.0000 |
Decision Tree Classifier | 0.4861 | 0.4788 | 0.3690 | 0.4397 | 0.4013 |
SVM - Linear Kernel | 0.4778 | 0.4881 | 0.6429 | 0.4576 | 0.5347 |
Linear Discriminant Analysis | 0.4750 | 0.4760 | 0.5655 | 0.4502 | 0.5013 |
Ridge Classifier | 0.4583 | 0.4647 | 0.5595 | 0.4372 | 0.4909 |
Logistic Regression | 0.4333 | 0.4473 | 0.6607 | 0.4302 | 0.5211 |
Extra Trees Classifier | 0.4167 | 0.4429 | 0.5595 | 0.4087 | 0.4724 |
Light’ Gradient Boosting Machine | 0.3694 | 0.2222 | 0.5952 | 0.3861 | 0.4684 |
Although IMU data belonged to the same type of sensor group, it was observed that model trained with MTw system failed to recognize data obtained from ATD system sensor group. According to performance criteria, most successful model was determined to be Gradient Boosting Classifier, with an accuracy rate of 53.33%, accuracy in classification (AUC) of 44.95%, recall of 0%, precision of 0%, and F1-score of 0.
Classification performance of models trained using MTw training dataset on a hybrid test dataset consisting of both MTw and ATD test data is shown in Table 9.
Table 9
Classification performance of models trained with MTw dataset tested on hybrid test dataset.
Model | Accuracy | AUC | Recall | Precision | F1 |
Extreme Gradient Boosting | 0.8389 | 0.9093 | 0.6516 | 0.9876 | 0.7852 |
Ada Boost Classifier | 0.8343 | 0.8855 | 0.6516 | 0.9725 | 0.7804 |
Gradient Boosting Classifier | 0.8343 | 0.9291 | 0.6475 | 0.9783 | 0.7793 |
Decision Tree Classifier | 0.8111 | 0.8070 | 0.7643 | 0.8074 | 0.7853 |
Linear Discriminant Analysis | 0.8074 | 0.7686 | 0.8176 | 0.7703 | 0.7932 |
Logistic Regression | 0.8046 | 0.7036 | 0.8750 | 0.7400 | 0.8019 |
SVM - Linear Kernel | 0.8083 | 0.8113 | 0.8422 | 0.7597 | 0.7988 |
Ridge Classifier | 0.8028 | 0.8041 | 0.8176 | 0.7629 | 0.7893 |
Extra Trees Classifier | 0.8000 | 0.9336 | 0.8402 | 0.7482 | 0.7915 |
Light Gradient Boosting Machine | 0.7824 | 0.9078 | 0.8545 | 0.7177 | 0.7802 |
In hybrid dataset, it was observed that higher number of data samples from MTw dataset resulted in improved prediction accuracy for MTw test data. However, significant errors were observed in recognizing ATD data. Ada Boost Classifier emerged as most successful algorithm, with an accuracy rate of 83.43%, accuracy in classification (AUC) of 88.55%, recall of 65.16%, precision of 97.25%, and F1-score of 78.04. A comparison of performance of top-performing Ada Boost Classifier algorithm on MTw dataset with test datasets is presented in Table 10.
Table 10
Performance of AdaBoost Classifier model trained with MTw dataset tested on test datasets
Model | Test Set | Accuracy | AUC | Recall | Precision | F1 |
Ada Boost Classifier | MTw | 0.9931 | 0.9994 | 0.9938 | 0.9907 | 0.9922 |
Ada Boost Classifier | ATD | 0.5167 | 0.2809 | 0.0000 | 0.0000 | 0.0000 |
Ada Boost Classifier | Hibrit | 0.8343 | 0.8855 | 0.6516 | 0.9725 | 0.7804 |
3.2.2 Evaluation of ATD Dataset
Dataset of ATD system has been transformed into a matrix of size 1350x910 to enable application of a time series approach. This dataset has been obtained from a total of 30 volunteers. Data from 22 volunteers were used for training purposes, while data from 8 volunteers were reserved for testing.
At conclusion of training process, it was determined that Extra Tree algorithm was the most successful. confusion matrix, presented in Fig. 7, illustrates performance achieved on test data. In confusion matrix, a value of 0.0 represents Daily Life Activities (DLA), while a value of 1.0 represents Falls. With a 100% success rate, the confusion matrix accurately predicted all 192 instances of falls and 168 instances of daily life activities (Fig. 7). These results indicate an exceptionally high level of accuracy. However, there is a suspicion that model may have simply memorized data, raising concerns about reliability of system. Performances of other trained algorithms can be examined in Table 11.
Table 11
Performance of algorithms trained with ATD dataset tested on ATD test data.
Model | Accuracy | AUC | Recall | Precision | F1 |
Extra Trees Classifier | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
K Neighbors Classifier | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
Logistic Regression | 0.9972 | 1.0000 | 0.9940 | 1.0000 | 0.9970 |
Light Gradient Boosting Machine | 0.9944 | 0.9999 | 0.9881 | 10.000 | 0.9940 |
Ada Boost Classifier | 0.9944 | 0.9997 | 0.9940 | 0.9940 | 0.9940 |
Gradient Boosting Classifier | 0.9944 | 0.9996 | 0.9881 | 10.000 | 0.9940 |
Extreme Gradient Boosting | 0.9944 | 0.9997 | 0.9881 | 10.000 | 0.9940 |
Naive Bayes | 0.9944 | 0.9998 | 0.9940 | 0.9940 | 0.9940 |
SVM - Linear Kernel | 0.9944 | 0.9940 | 0.9881 | 10.000 | 0.9940 |
Decision Tree Classifier | 0.9889 | 0.9885 | 0.9821 | 0.9940 | 0.9880 |
High-performing models trained with MTw dataset provided unsuccessful results in classifying ATD test dataset values, as shown in Table 12.
Table 12
Performance of algorithms trained with ATD dataset tested on MTw test data.
Model | Accuracy | AUC | Recall | Precision | F1 |
Decision Tree Classifier | 0.6213 | 0.5991 | 0.3689 | 0.6406 | 0.4681 |
Ada Boost Classifier | 0.6194 | 0.5350 | 0.3730 | 0.6341 | 0.4697 |
Extreme Gradient Boosting | 0.6167 | 0.4727 | 0.3709 | 0.6285 | 0.4665 |
Light Gradient Boosting Machine | 0.6139 | 0.5567 | 0.3709 | 0.6220 | 0.4647 |
Gradient Boosting Classifier | 0.6083 | 0.5608 | 0.3709 | 0.6094 | 0.4611 |
Extra Trees Classifier | 0.5769 | 0.5947 | 0.3730 | 0.5465 | 0.4434 |
Linear Discriminant Analysis | 0.4361 | 0.4468 | 0.5020 | 0.4010 | 0.4459 |
Ridge Classifier | 0.4361 | 0.4388 | 0.4672 | 0.3951 | 0.4282 |
Logistic Regression | 0.4139 | 0.1989 | 0.3852 | 0.3608 | 0.3726 |
K Neighbors Classifier | 0.4009 | 0.4012 | 0.3689 | 0.3468 | 0.3575 |
Due to the imbalance in data distribution, the hybrid dataset contains fewer ATD dataset values, resulting in a decreased accuracy in predicting the ATD test data within the hybrid dataset. The Decision Tree algorithm emerged as the most successful algorithm, with values of 62.13% for accuracy, 59.91% for area under the curve (AUC) in classification, 36.89% for recall, 64.04% for precision, and 46.81% for F1-score. The accuracy of the K-Neighbors algorithm on the hybrid dataset was 40.09%, indicating the lowest performance and supporting the suspicion of memorization during modeling.
The performance of the most successful model, the Extra Tree Classifier Algorithm, on other datasets can be further compared in Table 14. However, there is a suspicion that model may have memorized some aspects during the learning process with the ATD dataset.
Table 14
Performance of Extra Tree Classifier model trained with ATD tested on test datasets.
Model | Test Set | Accuracy | AUC | Recall | Precision | F1 |
Extra Trees Classifier | ATD | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
Extra Trees Classifier | MTw | 0.3653 | 0.1159 | 0.0438 | 0.0848 | 0.0577 |
Extra Trees Classifier | Hibrit | 0.5769 | 0.5947 | 0.3730 | 0.5465 | 0.4434 |
3.2.3 Evaluation of the Hybrid Dataset
Data obtained from two different brands of inertial sensors were combined without mixing to create a hybrid dataset. Training phase of hybrid dataset utilized data from 10 MTw volunteers and 22 ATD volunteers. For testing, data from 4 MTw volunteers and 8 ATD volunteers were used. At conclusion of training process, Extra Trees Classifier algorithm emerged as the most successful model. Performance on test data was presented using a confusion matrix, which is shown in Fig. 7. In confusion matrix, value of 0.0 represents fall activities correctly predicted as falls, and value of 1.0 represents fall activities incorrectly predicted as ADL (Activities of Daily Living). According to the confusion matrix (in Fig. 8), out of 592 fall activities, 591 were correctly predicted as falls, while 1 was incorrectly predicted as ADL. Additionally, out of 488 fall activities, 4 were correctly predicted as falls, while 484 were correctly predicted as ADL (Fig. 8). These results demonstrate that the model operates with a high success rate.
When evaluated according to performance metrics, the success rates in the confusion matrix were calculated as follows: accuracy 99.54%, classification accuracy (AUC) 100.00%, recall 99.18%, precision 99.79%, and F1-score 99.49%. These values indicate that the model performs with a high level of success. The ATD test dataset was predicted using the top 10 performing algorithms from MTw dataset. A performance comparison is presented in Table 15.
Table 15
Performance metrics of algorithms trained with hybrid dataset test on hybrid test dataset
Model | Accuracy | AUC | Recall | Precision | F1 |
Extra Trees Classifier | 0.9954 | 1.0000 | 0.9918 | 0.9979 | 0.9949 |
Light Gradient Boosting Machine | 0.9935 | 1.0000 | 0.9877 | 0.9979 | 0.9928 |
Extreme Gradient Boosting | 0.9935 | 0.9998 | 0.9918 | 0.9938 | 0.9928 |
Gradient Boosting Classifier | 0.9880 | 0.9994 | 0.9857 | 0.9877 | 0.9867 |
Ada Boost Classifier | 0.9833 | 0.9986 | 0.9857 | 0.9776 | 0.9816 |
K Neighbors Classifier | 0.9796 | 0.9942 | 0.9672 | 0.9874 | 0.9772 |
Decision Tree Classifier | 0.9685 | 0.9661 | 0.9406 | 0.9892 | 0.9643 |
Logistic Regression | 0.9250 | 0.9671 | 0.9180 | 0.9162 | 0.9171 |
Ridge Classifier | 0.9176 | 0.9162 | 0.9016 | 0.9148 | 0.9082 |
Linear Discriminant Analysis | 0.9148 | 0.9590 | 0.8975 | 0.9125 | 0.9050 |
Performance of these highly successful models on MTw test dataset of the hybrid dataset can be seen in Table 16.
Table 16
Performance metrics of algorithms trained with hybrid dataset test on MTw test dataset.
Model | Accuracy | AUC | Recall | Precision | F1 |
Extra Trees Classifier | 0.9944 | 0.9999 | 0.9906 | 0.9969 | 0.9937 |
Light Gradient Boosting Machine | 0.9931 | 10.000 | 0.9875 | 0.9968 | 0.9922 |
Extreme Gradient Boosting | 0.9931 | 0.9999 | 0.9938 | 0.9907 | 0.9922 |
Ada Boost Classifier | 0.9875 | 0.9992 | 0.9875 | 0.9844 | 0.9860 |
Gradient Boosting Classifier | 0.9847 | 0.9993 | 0.9812 | 0.9843 | 0.9828 |
K Neighbors Classifier | 0.9833 | 0.9934 | 0.9812 | 0.9812 | 0.9812 |
Logistic Regression | 0.9819 | 0.9975 | 0.9719 | 0.9873 | 0.9795 |
Ridge Classifier | 0.9819 | 0.9806 | 0.9688 | 0.9904 | 0.9795 |
Linear Discriminant Analysis | 0.9792 | 0.9920 | 0.9625 | 0.9904 | 0.9762 |
Decision Tree Classifier | 0.9556 | 0.9512 | 0.9125 | 0.9865 | 0.9481 |
The performance of hybrid dataset models on ATD test dataset can be seen in Table 17.
Table 17
Performance metrics of algorithms trained with hybrid dataset test on ATD test data.
Model | Accuracy | AUC | Recall | Precision | F1 |
Extra Trees Classifier | 0.9972 | 1.0000 | 0.9940 | 1.0000 | 0.9970 |
Light Gradient Boosting Machine | 0.9944 | 1.0000 | 0.9881 | 1.0000 | 0.9940 |
Gradient Boosting Classifier | 0.9944 | 0.9996 | 0.9940 | 0.9940 | 0.9940 |
Extreme Gradient Boosting | 0.9944 | 0.9995 | 0.9881 | 10000 | 0.9940 |
Decision Tree Classifier | 0.9944 | 0.9944 | 0.9940 | 0.9940 | 0.9940 |
Ada Boost Classifier | 0.9750 | 0.9972 | 0.9821 | 0.9649 | 0.9735 |
K Neighbors Classifier | 0.9722 | 0.9968 | 0.9405 | 10000 | 0.9693 |
Logistic Regression | 0.8111 | 0.8876 | 0.8155 | 0.7874 | 0.8012 |
SVM - Linear Kernel | 0.8111 | 0.8140 | 0.8571 | 0.7660 | 0.8090 |
Linear Discriminant Analysis | 0.7861 | 0.8599 | 0.7738 | 0.7692 | 0.7715 |
It has been observed that the model trained on the hybrid dataset can achieve significantly higher performance compared to the models trained on the MTw and ATD datasets separately. The performance of the hybrid dataset on the test datasets is presented in Table 18.
Table 18
Performance metrics of algorithms trained with hybrid dataset tested on test datasets.
Model | Test Set | Accuracy | AUC | Recall | Precision | F1 |
Extra Trees Classifier | Hibrit | 0.9954 | 1.0000 | 0.9918 | 0.9979 | 0.9949 |
Extra Trees Classifier | MTw | 0.9944 | 0.9999 | 0.9906 | 0.9969 | 0.9937 |
Extra Trees Classifier | ATD | 0.9972 | 1.0000 | 0.9940 | 1.0000 | 0.9970 |
Indeed, the Extra Tree algorithm demonstrates high performance across all datasets, as seen in Table 18. Extra Tree machine learning classification model trained on the hybrid dataset exhibits a high success rate not only on the MTw dataset but also on the ATD dataset. This model, based on hybrid dataset, proves its ability to reliably classify data and can be confidently integrated into a system.