Wearable Technology and Mobile Applications for HealthcareView this Special Issue
Depression Episodes Detection in Unipolar and Bipolar Patients: A Methodology with Feature Extraction and Feature Selection with Genetic Algorithms Using Activity Motion Signal as Information Source
Depression is a mental disorder which typically includes recurrent sadness and loss of interest in the enjoyment of the positive aspects of life, and in severe cases fatigue, causing inability to perform daily activities, leading to a progressive loss of quality of life. Monitoring depression (unipolar and bipolar patients) stats relays on traditional method reports from patients; however, bias is commonly present, given the patients’ interpretation of the experiences. Nevertheless, to overcome this problem, Ecological Momentary Assessment (EMA) reports have been proposed and widely used. These reports includes data of the behaviour, feelings, and other type of activities recorded almost in real time using different types of portable devices, which nowadays include smartphones and other wearables such as smartwatches. In this study is proposed a methodology to detect depressive patients with the motion data generated by patient activity, recorded with a smartband, obtained from the “Depresjon” database. Using this signal as information source, a feature extraction approach of statistical features, in time and spectral evolution of the signal, is done. Subsequently, a clever feature selection with a genetic algorithm approach is done to reduce the amount of information required to give a fast noninvasive diagnostic. Results show that the feature extraction approach can achieve a value of 0.734 of area under the curve (AUC), and after applying feature selection approach, a model comprised by two features from the motion signal can achieve a 0.647 AUC. These results allow us to conclude that using the activity signal from a smartband, it is possible to distinguish between depressive states, providing a preliminary and automated tool to specialists for the diagnosis of depression almost in real time.
The definition of health issued by the World Health Organization (WHO) says that “health is a state of complete physical, mental and social well-being and not only of disease or infirmity.” More than 350 million people in the world suffer from depression, and this can become a serious health problem, especially when it is of long duration and moderate to severe intensity, and can cause great suffering and disrupt work, school, family, economic, and emotional activities, among others. In the worst case, it can lead to suicide, which is the cause of approximately 1 million deaths annually . In Latin America, there is a high rate of mental health problems in the infant and youth population; about 20% of this population has disorders that require interventions of health services, but this number is underestimated due to the tendency of adolescents to hide and disguise their own problems to adults and their lack of confidence to access therapeutic structures . Depression is a mental disorder characterized fundamentally by depressive mood, loss of interest, and enjoyment of the positive aspects of life and fatigue, which impoverish the quality of life and generate difficulties in the family, work, and social environment of those who suffer it.
Depression can manifest itself regardless of age, gender, socioeconomic status, and academic program and can present with primary symptoms that do not encompass mood changes and even change cognitive function, so it is not difficult for any individual to become depressed . There is no doubt that studying the sociodemographic factors of age, gender, socioeconomic stratum, and family in adolescent students is relevant, due to the relationship that may exist between these and the manifestation of depression. This is why, in the global context, we find a series of studies that report high rates of depression in this population .
When referring to depression, we include those mood disorders with depressive symptoms, which include unipolar major depressive disorder, dysthymia, and mood disorders due to medical illness with depressive symptoms, among others. Despite the variety of alterations, by default, when speaking of depression, reference is made to the major unipolar depressive disorder. This disorder is considered the main cause of years of life lost due to disability (AVPD) according to the Global Burden of Disease Study (GBDS), conducted by the WHO .
One in four people suffer from one or more mood or behavioural disorders throughout their lives, and between 50% and 70% of those with a major or minor depressive episode have a predisposition to develop a new one in the next 5 years, which generates great impact on the global economy by the cost of psychotherapeutic and pharmacological management and by the avoidant personality disorder (AVPD), which for 2010 was 2.5 trillion dollars associated with major depressive disorder, with a projected increase to 6 trillion by 2030. The prognosis would improve with timely and adequate psychological, social, and pharmacological management.
The efficacy of the current treatment of depression needs to be increased since its prevalence is very high worldwide and only half of patients experience complete remission with first-line treatments (pharmacotherapy and psychotherapy) within two years [1, 5].
Within the framework of preventive actions or psychological rehabilitation, the instrumentation to obtain a standardized measure of depression is to act from a scientific framework. Therefore, it is necessary to have evaluation instruments that demonstrate the best evidence of validity and reliability to support inferences about the early detection of some symptoms of depression. One of these measures is the State-Rask Depression Inventory , which has advantages over some known instruments, such as the Self-Administered Depression Scale  and the Beck-II Depression Inventory . Another of the most representative characteristics is that it allows us to differentiate between the persons’ current experience (state) and their habitual way of behaving (trait) with regard to the affective component of depression . This is of great clinical value in differentiating experience in two time frames and specifically targeting one of the constituent areas of depression: affective disorders . Classical methods to achieve correct monitoring of depression states (unipolar and bipolar) in patients are done by reports from patients’ recall. Nevertheless, this type of monitoring is prone to bias commonly, in addition to changes in the behaviour and understanding the real world as is reported by Shiffman et al. . Another type of method to overcome these problems is Ecological Momentary Assessment (EMA). This type of report includes behaviour, feelings, and other types of activities as close as possible to the moment of the experience in real-life situations . One improvement to these types of reports is done by the increase of wearable devices (for instance, smartwatches and smartglasses) and smartphones, which includes different types of sensors (motion sensors, gyroscopes, and accelerometers), allowing EMA measurements to be done almost in real time, helping to monitor mental illness and give a close view to provide treatments, interventions, and increase the coverage of mental health services in the population without the need of new specific proposal devices or modifying adding sensors to the environment where the patients are living.
One device currently used to achieve mental health illness supervision is the smartphones and similar ones like smartwatches. One proposal is presented by Gravenhorst et al. ; they discussed how mobile phones can increase the effectiveness of mental disorders treatment by two main approaches: in one hand, the implementation of human-computer interfaces for therapy and secondly, collection of important data from patients’ daily lives to be recorded by the current state and the development of their mental problems; they also discussed the advantages and drawbacks of the most promising technologies for detecting disorders like depression or bipolar disorder.
Other interesting approach is given by Firth et al.’s  study; they demonstrate that psychological interventions with a smartphone as a clinical tool can reduce anxiety in schizophrenia patients. Torous et al. , in their study, provide data on psychiatric patients and the relationship with the use and interest of utilizing mobile applications to monitor their mental health conditions. In this study, results presented show that 50% of patients from all age groups are interested and will use mobile applications to monitor their mental health condition to control their illness. Bayindir et al.  present a systematic review of different works that focus in the use of mobile phone sensors to detect human behavior characteristics, describing activity detection at different abstraction levels of activity and characterizing health-related activities, like physical exercise and sleeping.
Additionally to the use of applications, these devices include several embedded sensors that have been used to acquire contextual information and several niches [14, 15], including activity recognition  and particularly activity that helps to find mental disorders . For instance, Gruenerbl et al.  demonstrate that inertial sensors and GPS traces can be used as a measurement device in psychiatric diagnosis, through a methodology based on a feature extraction of physical motion levels and travel patterns, and a classification analysis using a naïve Bayes technique. Reece et al.  identify depressed subjects using uploaded photos of Instagram based on a random forest technique. Grünerbl et al.  propose the classification of depressed and maniac states in bipolar patients based on smartphone data. Maxhuni et al.  classify bipolar patients through audio, motor activity, and questionnaires. Berle et al.  propose an approach using motor activity information to reveal schizophrenia and depression patterns. Koo et al.  present a research of the utility of the combination of biomarkers related to different approaches, such as motor activity based on actigraphy measurements, showing that the discrimination of patients based on these biomarkers improves in the identification of depressed subjects. Averill et al.  examine the psychomotor change in depressive episodes based on the activity levels measured by actigraphy in order to know the response of the depression treatment, concluding that the early change in simple activity and psychomotor speed allows one to measure the treatment response in depressed patients. Garcia-Ceja et al.  performed an analysis of data collected through actigraphy applying machine learning to classify depressed patients, finding that the data contain information that allows determination of the depression status of a subject. Huguet et al.  present a review to identify self-help apps that are available for depressed people. The apps that offer cognitive behavioural therapy (CBT) or behavioural activation (BA) are evaluated since the low level of adherence to the core ingredients of the CBT and BA models causes that the utility of these apps is questionable. It was possible to conclude that the application of superior scientific, technological, and legal knowledge is required to improve the credibility of the apps for people with depression.
On the other hand, Mohr et al.  provide a review of sensing research related to mental health, where a layered and hierarchical model is provided for the translation of raw sensor data into markers of behaviors and states related to mental health. Finally, in the work of Guntuku et al.  is reviewed the study of predicting mental illness using social media, including screening surveys, public sharing on Twitter, and the membership in an online forum, concluding that automated detection methods are useful to identify depressed or individuals at risk through the monitoring of passive activity in social media.
The aim of this work is to study the signal generated by a smartbands accelerometer to detect depressive states through the activity of patients and to propose a feature extraction (using the temporal and spectral evolution of the signal), as well as a clever feature selection based on a genetic algorithm approach to minimize the data required to identify these depressive states allowing an almost real-time noninvasive diagnosis. In this type of disease, early symptomatic detection can significantly increase the development of an effective treatment and contribute to the prevention of this type of psychopathology.
One of the main advantages proposed in this work is the simplicity in the data acquisition since the device used is noninvasive, has a small size, and does not hinder daily activities, which is a benefit compared to other devices that can be interposed in day-to-day tasks, in addition to using multiple sensors for the acquisition of different types of data, which is not necessary in this approach because the same purpose is achieved with a single source of acquisition, obtaining the information required for the extraction of features that allowed the classification of depressive patients.
This paper is organized as follows: in Section 2 is detailed described the materials used for the development of this research, as well as the set of the stages of the methodology proposed. Then, Section 3 presents the results obtained. Section 4 is referred to the discussion developed based on the results previously, and finally, Section 5 shows the conclusions of this work.
2. Materials and Methods
The methodology proposed in this research consists in five main stages, shown in Figure 1. The data used for the development of this work is acquired from the “Depresjon” dataset (A). These data are initially subjected to a data preprocessing (B) step, in order to select the samples and subjects for further analysis, to normalize the data, and to eliminate missing values. Then, the feature extraction (C) is performed, obtaining temporal and frequency statistical features, which are submitted to a feature selection (D) step, using the genetic algorithm (GA) “Galgo.” Finally, the set of selected features is evaluated, measuring its fitness in the classification of controls and cases, based on a random forest (RF) technique and a statistical analysis (E).
2.1. Data Description
The Depresjon dataset is a collection of data that contains the motor activity of patients monitored with an actigraph watch held on the right wrist. The actigraph watch is called “Actiwatch” (model AW4), developed by Cambridge Neurotechnology Ltd, England. The Actiwatch measures activity levels, and the sampling frequency is 32 Hz, recording movements over 0.05 g. Movements equal a corresponding voltage, which is stored as an activity count in the memory of the Actiwatch, and the number of counts is proportional to the intensity of the movement. The activity counts were recorded in intervals of one minute.
The database contains the data for the controls (absence of depression, 32 subjects) and for the cases (presence of depression, 23 subjects). The features collected for each subject were divided in two categories, actigraph data recorded over time and Montgomery Åsberg Depression Rating Scale (MADRS) scores. The data collected over time include the features “timestamp” (one minute intervals), “date” (date of measurement), and “activity” (activity measurement from the actigraph watch). In addition, MADRS scores include the features “number” (patient identifier), “day” (number of days of measurements), “gender” (1: female/2: male), “age” (age in age groups), “afftype” (1: bipolar II, 2: unipolar depressive, and 3: bipolar I), “melanch” (1: melancholia; 2: no melancholia), “inpatient” (1: inpatient; 2: outpatient), “edu” (education grouped in years), “marriage” (1: married or cohabiting; 2: single), “work” (1: working or studying; 2: unemployed/sick leave/pension), “madrs1” (MADRS score when measurement started), and “madrs2” (MADRS when measurement stopped) .
For this work only the features over time were used.
2.2. Data Preprocessing
The data preprocessing consists in three main steps, the selection of samples and subjects, the normalization of the data, and the elimination of incomplete cases presented as (not available).
The number of samples collected is not consistent, differing in the number of minutes recorded for each subject, so a selection of subjects and samples was made in order to present a balanced amount of data referring to controls and cases. The selection of the samples is carried out keeping only the first value of the 60 acquired data in the minutes equivalent to one hour, counting now the activity in intervals of one hour, whereas the selection of subjects depended on the amount of data resulting from the selection of samples, selecting the first four controls present in the dataset and the first five cases. This number of subjects allows balance of the number of samples for cases and controls.
Then, in the normalization, the data are adjusted in order to obtain a normal distribution, presenting a mean = 1 and a standard deviation = 0, and it is calculated with Equation (1), where represents the normalized value, represents the sample, µ is the mean of the total data, and σ is the standard deviation of the total data:
Finally, the missing data are eliminated, removing all the rows with presence of NAs, in order to avoid problems in the subsequent analysis.
2.3. Feature Extraction
The feature extraction is performed using two types of data, temporal and frequency data. The temporal data are directly used from the time-dependent Depresjon data, which were collected from the activity of the subjects through the actigraph watch.
On the other hand, the frequency data are obtained through the calculation of the Fourier transform of the time-dependent Depresjon data.
Then, for each type of data, 14 statistical parameters are extracted, presented in Table 1, obtaining a total of 38 features.
The names of the features correspondent to the temporal data are “tKurtosis,” “tSesgo,” “tQ01,” “tQ05,” “tQ25,” “tQ75,” “tQ95,” “tQ99,” “tMedia,” “tSD,” “tVarianza,” “tTrimMedia,” “tCV,” and “tICV,” while the names of the features correspondent to the frequency data are “fKurtosis,” “fSesgo,” “fQ01,” “fQ05,” “fQ25,” “fQ75,” “fQ95,” “fQ99,” “fMedia,” “fSD,” “fVarianza,” “fTrimMedia,” “fCV,” and “fICV.”
2.4. Feature Selection
In this stage, the 38 features extracted are subjected to a feature selection based on a GA approach. GAs are a stochastic strategy that has been widely used in the analysis of data and they consist in a sequence of stages that starts with a random set of models and develops good local solutions reproducing the natural selection process using measures such as (1) higher rate of replication of the more accurate feature subsets, (2) mutation to generate different chromosomes, and (3) crossover to improve the combinations of the chromosomes.
A validation measure is calculated in combination during the selection process, testing the sets of chromosomes, ensuring that the multivariate feature selection is suitable. The aim of the GA is to minimize the score calculated by the fitness function, converging then into a solution, being therefore possible to select the most significant predictive subset of n features .
For this work, the genetic algorithm “Galgo” is used. Galgo is a package implemented under the R language, which is oriented to select models with high fitness and to analyze them, as well as for the reconstruction and characterization of representative summary models.
The procedure of Galgo begins with a random population of feature or gene subsets or chromosomes of a defined size (n), which are assessed through a fitness function for their ability to predict or classify the desirable outcome or the dependent variable, obtaining a certain value of accuracy. The classification methods that can be used in the internal procedure of Galgo are k-nearest-neighbors, discriminant functions, nearest centroid, support vector machine, neural networks, and random forest.
The main idea of the process is to replace the first population with a new one, including variants of chromosomes that achieved a higher classification accuracy, and to repeat this procedure until a desired accuracy is reached. The progressive changes of the chromosomes are performed through a series of operators that simulate the process of natural selection, selection, mutation, and crossover.
The proportion of the solution space increases with the evolution of independent chromosome populations in partially isolated environments, known as niches, and chromosomes can migrate from one niche to another, in order to ensure the recombination of good solutions. A set of niches is called world .
This process is carried out in four main steps:(i)First, the analysis is configured, specifying the input and the outcome features, as well as a series of parameters that will guide the behaviour of the process, such as the classification model, the desired accuracy, and the error estimation scheme, among others. The classification model can be selected from the implemented or can be defined by a function of the user; while the error estimation can be defined in two levels, with a training/test validation strategy using variant random splits, and in the internal training process using a k-fold cross-validation, random splits, or re-substitution error.(ii)Then, the search of relevant multivariate models begins with a random population of chromosomes in each cycle of the procedure. The number of chromosomes developed needs to be large enough to ensure that the greatest amount of solutions was found and to achieve this, two approaches are designed to provide information of the chromosome composition, the level of convergence of the solutions, and the evolution of the fitness values, diagnosing the stability of the populations.(iii)A refinement and analysis of the population of the selected chromosomes is carried out, since not all the genes included in the best chromosome may be contributing in a significant way to the fitness value. Therefore, a backward selection strategy is implemented to obtain a model contained by genes that significantly contribute to the accuracy of the result.(iv)Finally, the development of a significant statistical model is obtained from the population of the selected chromosomes. For this step, a forward selection strategy is included, and its operation is based on a stepwise inclusion adding the most frequent genes of the chromosome population.
The configuration of the analysis for this study is composed by 200 generations, five genes per chromosome, a desired accuracy of 0.99, and “nearest centroid” as classification model, and an error estimation scheme was used a cross-validation approach.
2.5. Classification Analysis
The classification analysis was carried out through a RF method, looking for the classification of subjects in two different states, depressed (labeled as “1”) and not depressed (labeled as “0”).
RF is a machine learning technique that presents two main approaches, classification and regression, and its performance is based on decision trees. In the classification option, RF provides estimators of a Bayes classifier, , minimizing the error classification .
Roughly, an ensemble of trees grows, constructed with random vectors that generate each of the trees, deciding the class to which the data correspond by voting, where the majority of the class votes determine the RF prediction. This process causes that the generalization error merges to a limiting value, thus improving the classification accuracy of the system .
Specifically, the trees are created using a subset of bootstrap samples with replacement, (of a training set L), known as a bagging approach, which means that one same sample can be selected several times for the classification analysis while the others samples may not be selected.
Every decision tree is independently constructed without any pruning, and each node is divided through a splitting rule using a specific number of features, , randomly selected.
The splitting rule is added to the estimators calculated from the trees, represented as . A response value is subsequently obtained from the new point, which consists in the construction of the following equation:
The forest is growing up to a defined number of trees, , and by this step, the algorithm creates tree that present two main characteristics, high variance and low bias. The final classification decision is calculated through the arithmetic mean of the class assignment probabilities of the total number of trees. Then, an evaluation step is performed using a new set of unlabelled data input with the decision trees developed in the ensemble, giving each tree a vote for a class. The class that collects the greatest number of votes is the one selected.
Around two thirds of the total samples are usually used for the training of the trees, and they are referred as samples; then, with the remaining one third samples, referred as samples, an internal cross-validation is realized for the estimation of the model performance .
The estimation of this error is known as out-of-bag (OOB) error. This value measures the misclassification rate for the classification of the OOB samples. This means that a feature, , is important if when breaking the relationship between and Y, the error of the prediction increases, and the error of the prediction in each tree, , is evaluated with the OOB sample using
It is important to note that according to literature, the classification accuracy is less sensitive to than to ; therefore, since RF is a computationally efficient classifier that does not present problems of overfitting, can be a number as large as possible. On the other hand, the parameter is usually defined by the square root of the total number of input features .
For the development of this study, the number of trees selected is = 2000, and the number of features at each split, , is calculated according to the number of features as , with being the number of features.
The validation stage is based on three parameters, the AUC as a single value quantity of the ROC curve, specificity, and sensitivity.
The ROC curve has been a widely used tool for the evaluation of binary classification models since it presents a series of characteristics that allow the correct interpretation of the results, such as the intuitive visual interpretation of the curve, easy comparison among multiple models, and the AUC value .
The calculation of the classifier’s performance through the ROC curve provides a suitable operating point, called as decision threshold, for the parameterization of the classification model.
A classification problem presents two possible outputs, “correct” and “incorrect,” for each class of the model. An orderly way to present this information is through a confusion matrix, a table that shows the differences between the real and the predicted classes. The values contained in a confusion matrix are the true positives (), true negatives (), false positives (), and false negatives (); besides, the value of the row totals with the truly negatives () and truly positives () examples, and the value of the column totals with the predicted negative () and the predicted positive () examples .
The sensitivity is a parameter referred to as the ability to correctly identify those data with a condition, and it is calculated with the following equation:
On the other hand, the specificity is a parameter referred to as the ability to correctly identify those data without a condition, and it is calculated with the following equation:
Finally, the plotted values of the sensitivity and the specificity in conjunction represent the decision threshold of the ROC curve. The AUC value of the curve can be calculated through trapezoidal integration, as shown in the following equation:where and .
All the analysis is carried out in “R” (version 3.4.4), a free software environment designed for statistical computing and graphics . The libraries required for this analysis are “Galgo” (version 1.2-01) , “pROC” (version 1.11.0) , “e1071” (version 1.7-0) , “randomForest” (version 4.6-14) , “caret” (version 6.0-79) , and “rminer” (version 1.4.2) .
The results of this research are presented in this section. Through the first step of this methodology, which was the data acquisition, the number of subjects selected for the subsequent analysis was five for cases and four for controls, in order to balance the number of samples in both datasets.
Then, the feature extraction allowed collection of a series of 38 statistical features, which of the total, 14 belong to the time data and the remaining to the frequency data. It is important to remind that the frequency data were calculated through the Fourier transform of the time data.
For the third stage, a feature selection based on the GA, Galgo, is carried out, obtaining a series of graphs that allow observation of the performance of the data through the development of the different models created in the evolution of the algorithm. Figure 2 presents a graph of the frequency percentage with which each feature appeared within the different models developed, positioning each feature according to its order of appearance, from highest to lowest, where those features in black present the highest frequency and those features in gray present the lowest. According to this graph, the most significant features, according to their appearance frequency, are “tCV,” “tQ99,” “fCV,” and “tVarianza.” (Tables 2 and 3).
Then, in Figure 3 is shown a graph of the fitness performance throughout the evolution of the 200 generations of the GA, where it is possible to observe that the average fitness reaches a stable behaviour, with a value of around 0.63.
Figure 4 presents a heat map of the best chromosome presented by the GA, contained by a model of five chromosomes, “tCV,” “tQ25,” “tQ99,” “tICV,” and “tCV.”
Then, the best chromosome is subjected to a forward selection step, where for each feature added to the model, its average fitness was calculated, as shown in Figure 5. According to this graph, the model reaches its best average fitness, as well as stability, with three features, “tCV,” “tQ99,” and “fCV.”
Finally, in Figure 6 is present a heat map of the final model obtained through backward elimination step, contained by two time features, “tCV” and “tQ99.”
In the classification analysis, a RF approach is used, measuring the OOB error in order to know the accuracy classification reached through the model selected in the previous step. In Table 4 is present the confusion matrix obtained through the classification of subjects using the total set of features and the respective error values for each of the classes. The OOB error obtained was of 26.95%. In Table 4 is present the confusion matrix obtained through the classification of subjects using the best chromosome and the respective error values for each of the classes. The OOB error obtained was of 30.52%. Finally, In Table 4 is present the confusion matrix obtained through the classification of subjects using the final model and the respective error values for each of the classes. The OOB error obtained was of 35.97%.
For the last stage of this work, a validation step was performed, calculating the ROC curves of the models, shown in Figure 7, where in Figure 7(a) is present the ROC curve of the model contained by the total features and its respective AUC value, which obtained a sensitivity of 0.751 and a specificity of 0.717. Then, in Figure 7(b) is present the ROC curve of the model contained by the best chromosome and its respective AUC value, which obtained a sensitivity of 0.699 and a specificity of 0.694. Finally, in Figure 7(c) is present the ROC curve of the model contained by the final model and its respective AUC value, which obtained a sensitivity of 0.684 and a specificity of 0.611.
In this section, the results obtained are discussed. From the total 38 statistical features extracted, a feature selection is performed based in Galgo. Initially, Galgo developed the graph shown in Figure 2, which provides the information of the frequency with which the features are part of the different chromosomes developed, ordered by rank from highest to lowest.
According to Figure 2, the four most significant features or the features that presented the highest frequency were those presented in black, of which three correspond to the temporal features and one to the frequency features, which means that temporal data are presenting more significant information than frequency data for the classification of subjects.
Then, in Figure 3 is shown a graph of the average fitness behaviour through the different generations of the GA, being possible to observe that the greatest change occurs at the beginning of the graph, within the first 50 generations, where the GA is in the process of finding the best combination of genes to obtain a chromosome suitable for classification. Subsequently, a relatively stable value is reached around generation 80, obtaining an average fitness value of around 0.63 in the last generation.
At the end of the 200 generations, the best chromosome obtained is presented in Figure 4, contained by the five features presented in the heat map, of which four correspond to the time features and the remaining corresponds to the frequency features. The first feature of the best chromosome corresponds to the frequency feature, “fCV”, referred to the coefficient of variation (CV), which is related to the standard deviation and the mean value, where the higher the value of the standard deviation compared to the mean, the higher will be the CV and vice versa. This feature may imply that the frequency data could be presenting significant variations among its values between cases and controls, being able to distinguish between both classes.
The second feature is “tICV”, referred to the inverse coefficient of variation (ICV), which may imply a similar meaning than the feature “fCV”, where the time data can be presenting significant information in the distribution of the data that allows one to distinguish between depressed and nondepressed subjects.
Then, there are present the features “tQ25” and “tQ99,” which represent the 25 and the 99 quantiles, respectively. Quantiles are points of regular intervals of the distribution function of a random variable. Therefore, these two features may imply that, in these data intervals, the most significant information or the greatest differences between both classes are presented because taking into account that data on the amount of activity carried out as a function of time are being analyzed and that quantile data are arranged in ascending order, it is possible that by comparing the variations in the amount of activity correspondingly, that is, the greater activity of depressed patients against the higher activity of nondepressed patients, a difference is presented meaningful.
The fifth feature is “tCV,” which represents the same as “fCV” but with time data. This feature may imply that the information of the physical activity of patients is presenting differences in the standard deviation and the mean between the two possible classes that provide support for the correct classification.
On the other hand, in Figure 5 is presented a graph of the behaviour of the average fitness when the features are subjected to a forward selection step. For each feature that is added to the model, the average fitness, the fitness of each class, and the total fitness are measured, in order to know the behaviour that the model has when including the information of the features and in this way to select the adequate number of features, avoiding having nonsignificant information. According to the graph, the model achieves stability from the third feature, reaching an average fitness of 0.636.
The last step of the feature selection consisted of a robust gene back elimination (RGBE) step, in order to remove redundant information, obtaining a final model contained by two features, presented in the heat map of Figure 6. This model is dependent on the mean, standard deviation, and the 99 quantile values, which according to the previous steps, these measures provide data that allow the classification of the classes.
Then, a RF approach is used for the classification analysis, comparing the OOB error obtained through three different models: one model is contained by the total set of initial features, and it obtained an OOB error of 26.95%; the second model is contained by the features of the best chromosome obtained with GA, and it obtained an OOB error of 30.52%; and the third model is contained by the final model obtained through the RGBE step, obtaining an OOB error of 35.97%.
The OOB error values allow one to know the percentage of data that was misclassified during the construction of the decision trees that form the random forest, and as is possible to observe, this value increases when the model contains less features, presenting the lowest OOB error in the model contained by the total set of features; nevertheless, even when the final model presents an error 10% higher than the first model, the percentage that is correctly classified remains statistically significant. Besides, it is important to remark that the number of features contained in the final model is significantly smaller than that contained in the first model; therefore, the information required for the classification is much smaller, thus reducing the computational cost for the analysis of the data.
Also, from the validation that RF performs internally, a series of confusion matrices were obtained in order to explain the OOB errors obtained. In Table 4 is present the true positives (= 962) and true negatives (= 819), as well as the error value for controls (= 0.284) and for cases (= 0.255), for the model contained by the total set of features, where it is shown that even when the error is higher for controls, both classes present similar values in the classification error. Then, Table 4 shows the true positives (= 920), true negatives (= 774), error value for controls (= 0.324), and error value for cases (= 0.288), obtained through the classification using the best chromosome, where it is evident that the error values increase for both classes; nevertheless, the classification continues presenting a statistically significant aptitude even though the quantity of features contained in the model was reduced by around 86.84%. In Table 4 is present the true positives (= 690) and true negatives (= 871), and the error value for controls (= 0.397) and for cases (= 0.326), showing that, evidently, by reducing the number of features of the best chromosome, the error is increased again, especially for controls, which may represent that the activity registered by controls could be confused with the activity of cases in specific moments of time, for example, in the hours of sleep. In addition, could also be confused the time of greatest activity for both classes when in the case of controls, physical activity is not very energetic. However, this problem of confusion in the classification can be solved by increasing the number of samples in both classes, taking into account that it is important that the training of the algorithms have a balanced amount of data.
In the validation stage, the specificity and sensitivity allowed support of the previous results, obtaining higher values in the evaluation of sensitivity than in specificity, although it is important to note that the results of the validation presented significant values for the three models evaluated.
Then, the ROC curve is calculated for each of the models, as shown in Figure 7, where Figure 7(a) represents the curve obtained using the total set of features, Figure 7(b) represents the curve obtained using the best chromosome, and Figure 7(c) represents the curve obtained using the final model, obtaining AUC values of 0.734 0.697 0.647, respectively. The AUC value is reduced by decreasing the number of features of the models; however, the difference between the AUC of the model that contains 100% of the features and the final model, which only contains 5.26% of the features, is not representative taking into account that the AUC remains statistically significant in the final model. Therefore, the ability of the final model to classify cases from controls remains significant despite the limited amount of information used, thus benefiting the computational cost necessary to carry out the classification.
Finally, in Table 5, a comparison between different techniques based on the same approach is shown, collecting data through actigraphy in order to identify depressed patients, where according to the results is possible to conclude that all works present statistical significant results; however, the complexity of the methodologies and the quantity of characteristics related to different information used on each research, as well as the information sources, are greater than those proposed in this work, since it was only necessary the extraction of a reduced set of statistical features from a database collected by a single sensor from a small set of patients, presenting as one of the main contributions the simplicity of the experimentation made for the classification of subjects with presence of depression obtaining statistically significant results, in addition to presenting a lower computational cost than that presented in the mentioned works due to the small amount of data.
In this research is proposed a methodology composed by a series of steps which mainly includes a feature selection, a classification analysis, and a validation, in order to find the relationship between a series of statistical features, based on time and frequency continuous values acquired in a specific time and the possible condition of depression.
It is important to remark that the number of subjects allows one to obtain significant results; nevertheless, this number of samples can be increased in order to mainly improve the result of the true negatives, which presents greater error than true positives. On the other hand, the extracted statistical features show that the information they contain provides a description of the main characteristics of a patient’s full-day activity that allows differentiation between depressed and nondepressed subjects.
The feature selection through the GA provides a best chromosome which is subsequently reduced to a model contained by two features. These two features are statistical descriptors of temporary data that according to the validation step, despite presenting a greater error in the differentiation of cases and controls than if the whole set of features is used, the results remain statistically significant, thus allowing having a contained model with a reduced amount of features that automatically classifies depressed subjects of nondepressed subjects with significant fitness.
In addition, it is worth noting that one of the greatest advantages of the model being significantly reduced is that it is also reduced in its computational cost, making it easier to access it, since it does not require specialized software or hardware for its implementation.
Besides, one of the main benefits demonstrated in this work is the values with high precision obtained through a simple methodology using a single source of data, which in comparison with other works, where it is necessary to use more than one source for the data acquisition and a series of different techniques for the classification analysis, this approach provides simplicity and statistically significant results for less processing steps and computational cost.
Then, it is possible to conclude that the methodology implemented in this study allows one to know that evidently, there is an association between the recorded daily activity of a patient and the condition of his depressive state. Besides, the results obtained are sustained according to what is reported in the literature, where among the symptoms presented by patients with depression is the slowness of movement, poor body gesticulation, and the feeling of fatigue, tending to show lower levels of activity than subjects who do not have this condition.
Therefore, through this work is obtained a preliminary tool for the possible support in the diagnosis of the specialists to know the state of health of a patient according to his state of presence or absence of depression, based on the level of activity he has in a full day.
The Depresjon data used to support the findings of this study have been deposited in the “control” and “condition” repositories. This dataset can be accesed in http://datasets.simula.no/depresjon/ and/or can be directly downloaded from http://doi.org/10.5281/zenodo.1219550.
Conflicts of Interest
The authors declare that there are no conflicts of interest regarding the publication of this paper.
Carlos E. Galván-Tejada and Laura A. Zanella-Calzada contributed equally to this work.
J. Fisher, M. Cabral de Mello, V. Patel et al., “Prevalence and determinants of common perinatal mental disorders in women in low- and lower-middle-income countries: a systematic review,” Bulletin of the World Health Organization, vol. 90, no. 2, pp. 139–149H, 2012.View at: Publisher Site | Google Scholar
R. Leyva-Jiménez, A. M. Hernández-Juárez, G. Nava-Jiménez, and V. López-Gaona, “Depresión en adolescentes y funcionamiento familiar,” Revista Médica del Instituto Mexicano del Seguro Social, vol. 45, 2007.View at: Google Scholar
L. Mendoza, L. d. P. S. Peinado, L. A. D. Martínez, and A. Campo-Arias, “Prevalencia de sintomatología depresiva en niños y niñas escolares de bucaramanga, Colombia,” Revista Colombiana de Psiquiatr, vol. 33, pp. 163–171, 2004.View at: Google Scholar
M. E. Buelna Serrano, L. Gutiérrez Herrera, and S. Ávila Saldoval, “El desarrollo de la economía de consumo en el contexto del mundo bipolar de mediados del siglo xx. una visión retrospectiva,” Análisis Económico, vol. 30, 2015.View at: Google Scholar
D. Agudelo, G. Buela-Casal, and C. D. Spielberger, “Ansiedad y depresión: el problema de la diferenciación a través de los síntomas,” Salud Mental, vol. 30, pp. 33–41, 2007.View at: Google Scholar
A. T. Beck, R. A. Steer, and G. K. Brown, Beck Depression Inventory-II, vol. 78, San Antonio, TX, USA, 1996.
A. Gruenerbl, “Using smart phone mobility traces for the diagnosis of depressive and manic episodes in bipolar patients,” in Proceedings of the 5th Augmented Human International Conference, Kobe, Japan, March 2014.View at: Google Scholar
P. C. Koo, C. Berger, G. Kronenberg et al., “Combined cognitive, psychomotor and electrophysiological biomarkers in major depressive disorder,” European Archives of Psychiatry and Clinical Neuroscience, pp. 1–10, 2018, In press.View at: Google Scholar
E. Garcia-Ceja, “Depresjon: a motor activity database of depression episodes in unipolar and bipolar patients,” in Proceedings of the 9th ACM on Multimedia Systems Conference, MMSys’18, ACM, New York, NY, USA, 2018.View at: Google Scholar
N. Dogru and A. Subasi, “Traffic accident detection using random forest classifier,” in Proceedings of the Learning and Technology Conference, Jinan, China, May 2018.View at: Google Scholar
A. Liaw and M. Wiener, Breiman and Cutler’s Random Forests for Classification And Regression, R package Version 4.6-12, R Foundation for Statistical Computing, Vienna, Austria, 2015.
R Core Team, A Language and Environment for Statistical Computing, R Foundation for Statistical Computing, Vienna, Austria, 2018.
V. Trevino and F. Falciani, “Galgo: genetic algorithms for multivariate statistical models from large-scale functional genomics data, R package version 1.4”, 2018.
D. Meyer, E. Dimitriadou, K. Hornik, A. Weingessel, and F. Leisch, E1071: Misc Functions of the Department of Statistics, Probability Theory Group (Formerly: E1071), TU Wien, 2018, R package version 1.7-0.
A. Liaw and M. Wiener, “Classification and regression by randomforest,” R News, vol. 2, pp. 18–22, 2002.View at: Google Scholar
M. K. C. Jed Wing, Caret: Classification and Regression Training, 2018, R package version 6.0-79.
P. Cortez, Rminer: Data Mining Classification and Regression Methods, 2016, R package version 1.4.2.