Processing math: 100%
Research article Special Issues

Semi-supervised random forest regression model based on co-training and grouping with information entropy for evaluation of depression symptoms severity

  • Semi-supervised learning has always been a hot topic in machine learning. It uses a large number of unlabeled data to improve the performance of the model. This paper combines the co-training strategy and random forest to propose a novel semi-supervised regression algorithm: semi-supervised random forest regression model based on co-training and grouping with information entropy (E-CoGRF), and applies it to the evaluation of depression symptoms severity. The algorithm inherits the ensemble characteristics of random forest, and combines well with co-training. In order to balance the accuracy and diversity of co-training random forests, the algorithm proposes a grouping strategy to decision trees. Moreover, the information entropy is used to measure the confidence, which avoids unnecessary repeated training and improves the efficiency of the model. In the practical application of evaluation of depression symptoms severity, we collect cognitive behavioral data of emotional conflict based on the depressive affective disorder. And on this basis, feature construction and normalization preprocessing are carried out. Finally, the test is conducted on 35 labeled and 80 unlabeled depression patients. The result shows that the proposed algorithm obtains MAE (Mean Absolute Error) = 3.63 and RMSE (Root Mean Squared Error) = 4.50, which is better than other semi-supervised regression algorithms. The proposed method effectively solves the modeling difficulties caused by insufficient labeled samples, and has important reference value for the diagnosis of depression symptoms severity.

    Citation: Shengfu Lu, Xin Shi, Mi Li, Jinan Jiao, Lei Feng, Gang Wang. Semi-supervised random forest regression model based on co-training and grouping with information entropy for evaluation of depression symptoms severity[J]. Mathematical Biosciences and Engineering, 2021, 18(4): 4586-4602. doi: 10.3934/mbe.2021233

    Related Papers:

    [1] Zhi Yang, Yadong Yan, Haitao Gan, Jing Zhao, Zhiwei Ye . A safe semi-supervised graph convolution network. Mathematical Biosciences and Engineering, 2022, 19(12): 12677-12692. doi: 10.3934/mbe.2022592
    [2] Tianhui Sha, Yikai Zhang, Yong Peng, Wanzeng Kong . Semi-supervised regression with adaptive graph learning for EEG-based emotion recognition. Mathematical Biosciences and Engineering, 2023, 20(6): 11379-11402. doi: 10.3934/mbe.2023505
    [3] Ansheng Ye, Xiangbing Zhou, Kai Weng, Yu Gong, Fang Miao, Huimin Zhao . Image classification of hyperspectral remote sensing using semi-supervised learning algorithm. Mathematical Biosciences and Engineering, 2023, 20(6): 11502-11527. doi: 10.3934/mbe.2023510
    [4] Zhanhong Qiu, Weiyan Gan, Zhi Yang, Ran Zhou, Haitao Gan . Dual uncertainty-guided multi-model pseudo-label learning for semi-supervised medical image segmentation. Mathematical Biosciences and Engineering, 2024, 21(2): 2212-2232. doi: 10.3934/mbe.2024097
    [5] Xiaobo Zhang, Donghai Zhai, Yan Yang, Yiling Zhang, Chunlin Wang . A novel semi-supervised multi-view clustering framework for screening Parkinson's disease. Mathematical Biosciences and Engineering, 2020, 17(4): 3395-3411. doi: 10.3934/mbe.2020192
    [6] Wenli Cheng, Jiajia Jiao . An adversarially consensus model of augmented unlabeled data for cardiac image segmentation (CAU+). Mathematical Biosciences and Engineering, 2023, 20(8): 13521-13541. doi: 10.3934/mbe.2023603
    [7] R. Loula, L. H. A. Monteiro . On the criteria for diagnosing depression in bereaved individuals: a self-organizing map approach. Mathematical Biosciences and Engineering, 2022, 19(6): 5380-5392. doi: 10.3934/mbe.2022252
    [8] Keying Du, Liuyang Fang, Jie Chen, Dongdong Chen, Hua Lai . CTFusion: CNN-transformer-based self-supervised learning for infrared and visible image fusion. Mathematical Biosciences and Engineering, 2024, 21(7): 6710-6730. doi: 10.3934/mbe.2024294
    [9] Peng Ying, Zhongnian Li, Renke Sun, Xinzheng Xu . Complementary label learning based on knowledge distillation. Mathematical Biosciences and Engineering, 2023, 20(10): 17905-17918. doi: 10.3934/mbe.2023796
    [10] Haitao Gan, Zhi Yang, Ji Wang, Bing Li . 1-norm based safe semi-supervised learning. Mathematical Biosciences and Engineering, 2021, 18(6): 7727-7742. doi: 10.3934/mbe.2021383
  • Semi-supervised learning has always been a hot topic in machine learning. It uses a large number of unlabeled data to improve the performance of the model. This paper combines the co-training strategy and random forest to propose a novel semi-supervised regression algorithm: semi-supervised random forest regression model based on co-training and grouping with information entropy (E-CoGRF), and applies it to the evaluation of depression symptoms severity. The algorithm inherits the ensemble characteristics of random forest, and combines well with co-training. In order to balance the accuracy and diversity of co-training random forests, the algorithm proposes a grouping strategy to decision trees. Moreover, the information entropy is used to measure the confidence, which avoids unnecessary repeated training and improves the efficiency of the model. In the practical application of evaluation of depression symptoms severity, we collect cognitive behavioral data of emotional conflict based on the depressive affective disorder. And on this basis, feature construction and normalization preprocessing are carried out. Finally, the test is conducted on 35 labeled and 80 unlabeled depression patients. The result shows that the proposed algorithm obtains MAE (Mean Absolute Error) = 3.63 and RMSE (Root Mean Squared Error) = 4.50, which is better than other semi-supervised regression algorithms. The proposed method effectively solves the modeling difficulties caused by insufficient labeled samples, and has important reference value for the diagnosis of depression symptoms severity.



    In practical application, data tagging is limited by various factors, and sometimes even needs to pay a high price. Compared with labeled samples, the data acquisition of unlabeled samples is relatively simple. Semi-supervised learning is generated under the drive of practical application. It mainly studies how to make learners automatically use a large number of unlabeled data to assist the learning of a small number of labeled data under the premise of partial information missing in training data. Semi-supervised learning has been widely used in many fields because of its strong practical demand. Especially for psychiatric disorders research, there is a shortage of medical information talents and a low rate of clinical practice, so it is very expensive to obtain labeled samples. For example, in a recent study on bipolar disorder [1], the authors investigated the use of a semi-supervised learning method to process data obtained from patient-smartphone interactions, and considered the evolving time structure. The results shown that the semi-supervised model can effectively derive patient's status even if only 25% of the labeled data are available.

    Depression is a psychiatric disorder characterized by significant and persistent loss of pleasure. To date, the pathogenesis of depression is still unclear. The diagnosis of depression is mainly carried out by psychiatrists through structured interviews based on diagnostic manuals [2]. Clinically, 17-item Hamilton depression rating scale (HAMD-17) [3] is considered as the gold standard for the evaluation of depression symptoms severity. However, the evaluation of the HAMD scores is time-consuming. On the one hand, the evaluation time depends on patients' symptoms and cooperation, and it will take longer if the patient is severely blocked. On the other hand, HAMD scores are evaluated independently by two trained evaluators through conversation and observation, which also requires a high level of experience and skill. Therefore, it is very difficult to obtain the HAMD scores, which results in less labeled (HAMD score) data and more unlabeled data in patients with depression. It may be difficult to train a learning system with strong generalization if only a small number of labeled data is used for supervised learning, while ignoring a large number of unlabeled data will cause a great waste of data resources. In this paper, the semi-supervised learning method is used to construct the evaluation model, which can solve the above problems well and achieve better performance.

    In addition, machine learning methods based on objective biomarkers have become key technologies and effective means to diagnose depression. A large number of studies have shown that depression is related to the abnormalities in brain structure and function [4,5]. At present, the use of functional Magnetic Resonance Imaging (fMRI) and Electroencephalography (EEG) to record brain structure or brain activity has become an important means of identifying depression [6,7,8]. In addition, depression not only causes brain abnormalities, but also leads to behavioral disorders [9,10]. Therefore, in recent years, more and more attention has been paid to the correlation between depression and behavioral patterns such as facial expression, speech, language and body posture [11,12,13].

    This study considers the basic characteristic of depression—affective disorder, and collects cognitive behavioral data derived from emotional conflict tasks. Cognitive behavioral data on conflict tasks are obtained by active attention and conscious cognitive. Compared with EEG/fMRI, or expression/speech/language data, it is purer, reliable, and more likely to show the depression mood of patients with depression [14,15].

    Emotional conflict is caused by the processing conflict between the information arousing different emotions, which leads to the interference of irrelevant emotional stimulation on current emotional cognition. Researches based on cognitive psychology have shown that patients with depression show delayed reaction effect no matter in dealing with emotional conflict or non-conflict information. In particular, the delay is more significant when dealing with emotional conflict information [15]. The reason is that emotional disorder leads to negative emotional attention bias in patients with depression, which is manifested as excessive attention to sadness and insufficient attention to happiness [16,17]. Meanwhile, patients with depression have rumination. That is, they indulge and repeat negative emotions, which leads to slow thinking and movement, and difficult decision-making [18,19]. In addition, due to the sequence adjustment effect, the reaction time in dealing with an emotional conflict task is also affected by the previous task type [20]. Therefore, this study collects four kinds of cognitive behavior data: conflict/non-conflict/conflict monitoring/conflict resolution reaction time.

    Although deep learning is widely studied in image processing, speech and text recognition [21,22], the emotional conflict data in this study is tabular data, which is not suitable for deep learning model. In the future, we will use deep learning to explore the evaluation method of depression severity based on facial expression, speech and language text data obtained under emotional stimulation.

    The rest of this paper is arranged as follows: Section 2 is related work. Section 3 is data acquisition and processing. Section 4 is the proposed model. Section 5 is result and discussion. Section 6 is conclusion of this study.

    At present, automatic depression evaluation methods based on machine learning have been widely studied. For example, Yoshida et al. (2017) adopted partial least squares regression for fMRI-based depression recognition. The Beck Depression Inventory-II (BDI-II) score was used to evaluate depression severity with a prediction error of RMSE = 9.56 [6]. Kang et al. (2017) presented a deep transformation learning (DTL) method for visual-based depression recognition, and the prediction error for BDI-II score was RMSE = 9.43 and MAE = 7.74 [11]. Haque et al. (2018) evaluated depression symptoms severity by using causal convolutional network (C-CNN) based on spoken language and 3D facial expressions. The final prediction error on the Patient Health Questionnaire-8 (PHQ-8) was MAE = 3.67 [12]. Muzammel et al. (2020) proposed a speech-based depression evaluation method through deep learning, with the evaluation error of PHQ-8 was RMSE = 3.22 [13].

    Current studies on the evaluation of depression symptoms severity all use self-rating depression scales like PHQ-8 or BDI-II. However, these self-rating scales are too subjective to be used for clinical symptoms evaluation in patients with depression. HAMD depression scale is a clinical scale specially used to evaluate symptoms of patients with depression. In the evaluation process, two psychiatrists or two trained professionals are required to score independently. So the evaluation has a high accuracy and can be used as the standard for the selection of treatment methods or evaluation of treatment effects. Therefore, the HAMD scale is called the "gold standard" for clinical depression symptoms evaluation. The acquisition of HAMD-17 score is time-consuming and laborious, leading to the existence of a large number of samples without HAMD score labels in reality. Therefore, there is a lack of research on automatic evaluation methods of HAMD-17 score. The semi-supervised learning method introduced in this paper can solve this problem well.

    At present, there are four main paradigms in semi-supervised learning: generative model, transductive SVM, graph-based algorithm and co-training algorithm [23]. In many branches of semi-supervised learning, co-training algorithm has the characteristics of simple, effective, stable and fast convergence, which has attracted the extensive attention of many scholars and achieved a lot of research results.

    In 2005, Zhou and Li first studied semi-supervised regression and proposed co-training regressors (CoReg) [23]. This method generates two different KNN models based on different distance measures. Each regression model labels and selects unlabeled samples with high confidence to join the training set of another regression model. Lei et al. [24] proposed a semi-supervised regression algorithm based on support vector machine co-training. After that, Li et al. [25] proposed a semi-supervised regression algorithm based on co-training with SVR-KNN by combining SVR and KNN. In addition, some people [26,27] proposed the method of combining co-training strategy with partial least squares model (PLS).

    In the above co-training methods, only one model is used to assist the training of the other model. That is, one model selects some high confidence samples from the unlabeled samples to label and add them to the training set of the other model. However, the prediction result of the single model is not reliable and it is easy to introduce noise data into the model. Therefore, Hady et al. [28] constructed a co-training algorithm based on ensemble method—CoBCReg. CoBCReg uses a committee of regressors (RBFNNs constructed by bagging) to predict unlabeled samples instead of a single regressor. The algorithm combines co-training and ensemble learning to generate a more powerful learning system.

    Random forest is an ensemble learning method composed of decision trees, which is easy to implement in parallel. Moreover, randomness is introduced in both row and column directions, which can weaken over-fitting. Therefore, it is a powerful classification and regression method. Co-training of decision trees in random forest does not require the integration of different basic learners, but also ensures the diversity among members. However, there is a lack of research on co-training random forest in regression. Saitoh et al. [29] and Levatić et al. [30] only use random forest for self-training in regression. Therefore, this paper applies random forest to semi-supervised regression, and realizes the combination of ensemble learning and co-training.

    There are still some problems in the semi-supervised random forest regression model based on co-training: 1) a small number of decision trees will lead to low prediction accuracy; a large number of decision trees will reduce the diversity. Therefore, in this paper, the random grouping strategy is used to group the decision trees to achieve the balance between prediction performance and model diversity. 2) The confidence measure method used in the previous research [23,28] has the problems of large amount of calculation and low generalization. Therefore, this paper proposes a confidence measure method based on information entropy.

    All 115 patients with depression in this study were from the clinic. All participants provided signed informed consent and this study was approved by the Ethics Committee at Beijing Anding Hospital, Capital Medical University, Beijing, China. And the experiments have been conducted in accordance with the Helsinki declaration.

    In the decision-making tasks of emotional conflict, we collected the time-related features of all 115 patients with depression while performing the tasks. However, HAMD-17 scores were obtained from only 35 patients. When building the dataset using the acquired values of HAMD-17 and the time-related features of the decision-making tasks, HAMD-17 values for only 35 people are available, and those for the remaining 80 people are absent. Therefore, we adopted a semi-supervised machine learning approach to solve the problem of incomplete HAMD-17 labels in the samples.

    Table 1.  Distribution and depressive symptoms of HAMD-17 scores.
    HAMD-17 score Symptoms severity Samples size
    0–7 none 0
    8–17 mild 7
    18–24 moderate 19
    > 25 severe 9

     | Show Table
    DownLoad: CSV

    In order to obtain the cognitive behavioral data on emotional conflict, we require participants to complete the word-face Stroop tasks [14,31]. Each Stroop task is composed of an emotional face (happy face or sad face) and an emotional word ("HAPPY" or "SAD") in red ink. Emotional face pictures are from the International Standard Expressions Library (NimStim Set of Facial Expressions [32]), including 40 happy face pictures (22 males and 18 females) and 40 sad face pictures (22 males and 18 females). Photoshop is used to process all pictures in a unified way, so that the size of all face pictures are consistent.

    According to whether the emotions expressed by the face picture and the word are consistent, the experimental tasks are divided into two categories: 1) conflict tasks or incongruent tasks (happy face + "SAD" and sad face + "HAPPY"); 2) non-conflict tasks or congruent tasks (happy face + "HAPPY" and sad face + "SAD"), as shown in Figure 1.

    Figure 1.  Experimental tasks.

    In addition, due to the sequence adjustment effect, the reaction time in dealing with an emotional conflict task is also affected by the previous task type. Therefore, conflict tasks can be further divided into two types: 1) conflict monitoring tasks (Figure 1a): a non-conflict task activates a non-conflict expectation mechanism, resulting in the weakening of conflict resolution in the next conflict task (high conflict monitoring and low conflict resolution). In other words, after completing a non-conflict task, the participants will make a decision relatively slowly on the next conflict task. 2) conflict resolution tasks (Figure 1b): a conflict expectation mechanism activated by a conflict task enhances the conflict resolution in the next conflict task (high conflict resolution and low conflict monitoring). In other words, the participants will have conflict awareness after completing a conflict task, so that they will make a decision relatively quickly on the next conflict task.

    In the end, it constitutes a total of 80 conflict tasks and 80 non-conflict tasks. Because of the different types of the previous task, the 80 conflict tasks are divided into 40 conflict monitoring tasks and 40 conflict resolution tasks.

    The data acquisition process is shown in Figure 2. The 160 emotional decision-making tasks (80 conflict tasks and 80 non-conflict tasks) are divided into two sessions, and each session contains 80 trials. At the beginning of each trial, a "+" is displayed at the center of the screen for 3–5 s, followed by an emotional decision-making task for 1s. Participants are required to ignore the word, and quickly recognize the emotional valence (happy or sad) of the face and press the key (left key for happy and right key for sad). The next trial starts automatically after the key or task ends.

    Figure 2.  Data acquisition process.

    When the participants carry out each task, the acquisition program automatically records the start time and end time, and calculates the reaction time.

    Table 2 shows the mean reaction time of patients with different depression symptoms in conflict or non-conflict decision-making tasks. It can be seen that the more severe depression symptoms, the longer the reaction time required for decision-making. The result shows that the decision-making ability decreases with the increase of depression symptoms severity. It also illustrates that the decision-making reaction time under emotional conflict can characterize depression symptoms severity. Since the HAMD scale is a rating scale for depression symptoms severity, if the HAMD values are used as the labels of the samples, we can use the decision-making information on emotional conflict to predict the HAMD value and thus evaluate depression symptoms severity.Table 2. Depression symptoms severity and conflict decision-making.

    Table 2.  Depression symptoms severity and conflict decision-making.
    Symptoms severity Mean of reaction time (ms)
    conflict non-conflict
    mild 752.25 702.48
    moderate 758.15 709.32
    severe 771.84 725.07

     | Show Table
    DownLoad: CSV

    In the process of data acquisition, each participant acquired 80 non-conflict reaction time, 80 conflict reaction time, 40 conflict monitoring reaction time and 40 conflict resolution reaction time.

    From the perspective of statistical distribution, we extract 9 features from each kind of reaction time data of each participant, which are: 4 central tendency features (mean, median, 1st quartile, 3rd quartile), 3 dispersion tendency features (minimum, maximum, standard deviation) and 2 distribution pattern features (skewness coefficient, kurtosis coefficient).

    Therefore, a total of 36 features are constructed in this study, as shown in Table 3. The value of each feature is the mean of samples (ms).

    Table 3.  Feature construction (Q1 = 1st quartile, Q3 = 3rd quartile).
    Central tendency Dispersion tendency Distribution pattern
    Mean Median Q1 Q3 Min Max Std. Skewness Kurtosis
    conflict 743.40 722.71 643.29 819.20 487.50 1226.53 146.84 0.97 4.37
    non-conflict 703.12 675.71 606.55 764.43 467.13 1185.40 141.53 1.14 4.89
    conflict monitoring 744.22 722.02 641.90 824.34 502.63 1161.20 146.64 0.91 4.16
    conflict resolution 742.57 719.04 642.84 821.61 512.10 1162.20 144.96 0.86 3.94

     | Show Table
    DownLoad: CSV

    As shown in Table 3, the values of skewness feature and kurtosis feature are much smaller than other features, so their role in model construction will be ignored. In order for all features to participate in decision-making equally, it is necessary to perform feature normalization.

    In this paper, the min-max conversion function is used for feature normalization, and the feature data is converted to the interval [0, 1]. For each feature, the normalization method is shown in formula (1).

    xi=ximin(X)max(X)min(X) (1)

    where, X={xi},i=1,2,,N. X is the dataset (N is the number of samples), xi is the normalized value of xi. max(X)min(X) are the maximum and minimum of X, respectively.

    In this study, a semi-supervised regression method based on co-training is used to evaluate HAMD scores of patients with depression. The idea of co-training algorithm is as follows: First, train at least two models on the initial labeled sample set; Then, one of them is selected as the main model in turn, and the others constitute the auxiliary model of the main model; The auxiliary model predicts the unlabeled samples and provides the samples with high confidence for the main model; The main model retrains on the updated labeled sample set. Keep training in this way until convergence.

    This research combines co-training algorithm and random forest (CoRF), and on this basis, introduces a grouping strategy and a confidence measure method using information entropy.

    Co-training is effective only when members are diverse. Obviously, if the regression models are similar or even identical, there is no more information to be transmitted between the members, and the co-training is meaningless.

    CoRF: For a random forest with N decision trees, one of the decision trees is selected as the main model by turns in each iteration, and the other N-1 decision trees constitute its auxiliary model. Only one decision tree is different between the auxiliary models of any two main models, and the diversity is 1/ (N-1). The larger N is, the smaller the diversity between the two auxiliary models. The diversity between the two new labeled sample sets predicted and selected by the two auxiliary models will also be small, leading to the similarity of the two main models after retraining. Similarly, any two models will become similar to each other. After many iterations, the diversity between models is destroyed.

    CoGRF: According to the above description, when there are many decision trees, it is difficult for CoRF to ensure its diversity; when there are few decision trees, it is impossible to guarantee the accuracy. Therefore, a grouping strategy is introduced in this paper. N decision trees are divided into m groups, and there are N/m decision trees in each group. In each iteration, one group is taken as the main model, and the other m-1 groups are used as its auxiliary model. One group is different between the auxiliary models of any two main models, including N/m decision trees, and the diversity is 1/ (m-1). The grouping strategy greatly improves the diversity of the co-training model and ensures the number of decision trees.

    An important factor affecting the performance of co-training algorithm is how to measure the confidence of new labeled samples. Incorrect confidence measure method will lead to the selection and addition of samples with wrong labels, which will have a negative impact on the algorithm.

    In previous studies, RMSE difference was often used to measure the reliability of new labeled samples [23,28]. That is, the most reliable new labeled sample should be the one that reduces the regression model's error on the labeled sample set the most, as shown in formula (2).

    Δxu=RMSERMSE (2)

    where xu is the unlabeled sample. RMSE is the prediction error of the original regression model on the validation set. RMSE is the prediction error of the new regression model on the validation set. The new regression model is obtained by adding xu to the original training set and retraining. Finally, the sample with the largest Δxu is regarded as the sample with the highest confidence.

    This method has a significant disadvantage. That is, in each iteration, it has to repeated training model and calculate RMSE after adding each sample of the new labeled sample set. This method is computationally intensive and time-consuming. Therefore, this study proposes a simple and effective method to measure the confidence of samples: information entropy.

    In the actual system, the probability of each possible case in the system may not be the same. So information entropy is used to measure the uncertainty of the whole system. It is also a measure of the consistency of all possible cases in the system. The calculation formula of information entropy is as follows:

    H=ni=1pilogpi (3)

    where, n indicates that there are n possible cases in the system. pi is the probability of each possible case. The larger H is, the higher the system uncertainty is and the higher the consistency of all possible cases is.

    In the auxiliary model, the predicted values of different decision trees are different. Therefore, we use information entropy to measure the consistency of predictions, that is, the similarity degree of the predicted values of all decision trees. For sample xu, The confidence measure method follows formula (3), where pi is defined as follows:

    pi=Ti(xu)/ni=1Ti(xu) (4)

    where n is the number of decision trees in the auxiliary model. Ti(xu) is the predicted value of the i-th decision tree to sample xu. pi represents the proportion of predicted value of the i-th decision tree to predicted values of all decision trees in the auxiliary model.

    The larger the information entropy of the sample is, the higher the consistency of the predictions of the auxiliary decision trees is, and the higher the confidence of the sample is. The confidence measure method based on information entropy, on the one hand, greatly reduces the training process and the time complexity. On the other hand, considering the difference of predictions of all auxiliary decision trees, the selection of samples with the highest consistency is conducive to improving the generalization of the model.

    The algorithm process of semi-supervised random forest regression model based on co-training and grouping with information entropy (E-CoGRF) is shown in Figure 3.

    Figure 3.  E-CoGRF algorithm process.

    After the random forest is trained and randomly grouped, each group is retrained with the help of its auxiliary model (Figure 4). Verify whether the prediction error of the retrained random forest on the validation set decreases, that is, whether RMSE_V(t)<RMSE_V(t1) is correct. If it is true, repeat the Retraining-Verification process until the max iteration is reached; otherwise, regroup and then perform the Retraining-Verification process.

    Figure 4.  Retraining process of G1.

    Figure 4 describes the retraining process of G1 in Figure 3 in detail. In this process, G1 is the main model, and other groups constitute the auxiliary model A1. All auxiliary decision trees in A1 predict and measure the confidence of all unlabeled samples. The k samples with the highest confidence are selected and added to the training set of all decision trees in G1. Verify whether the addition of k samples reduces the prediction error of the main decision tree on the verification set. That is, for the main decision tree Ti, judge whether RMSEi_V(t)<RMSEi_V(t1) is correct. If not, execute k-1, retrain and verify. In the end, New G1 is obtained with the assistance of A1. See Table 4 for the specific algorithm.

    Table 4.  The E-CoGRF pseudo-code.
    Algorithm: E-CoGRF
    1: Initialize: the labeled training set Tr, the labeled validation set V, the unlabeled sample set U, max iteration T, the number of decision trees N, the number of groups m, the number of decision trees within a group n=N/m.
    2: Train: Construct a Random Forest consisting N Decision Tree, {T1,T2,,TN}
    3: Group: Randomly divide all trees into m groups equally, {G1,G2,,Gm}Construct Auxiliary groups for each group, {A1,A2,,Am}
    4:   Repeat for T round:
    5:    for i{1,2,,m}
    6:      for each xU do
    7:        for each TjAi, where j{1,2,,Nn}, do
    8:          predict: ˆyjTj(x)
    9:      end of for
    10:       Mean: ˆyu=Nnj=1ˆyj/(Nn)
    11:      Confidence: pj=ˆyj/Nnj=1ˆyj
    12:            H(x)=Nnj=1pjlogpj
    13:      end of for
    14:      Sort H in descending order
    15:      Select top k, X={x1,x2,,xk}, Y={^y1,^y2,,^yk}
    16:      Add: TriTri(X,Y)
    17:      Retrain: TilearnTree(Tri)
    18:      Verify: RMSEi_V(t)Ti(V)
    19:          if RMSEi_V(t) > = RMSEi_V(t1)
    20:            k=k-1, repeat step 15-19
    21:    end of for
    22:    Verify: RMSE_V(t)RF(V)
    23:      if RMSE_V(t) > = RMSE_V(t-1)
    24:       Regroup and Reconstruct the auxiliary model
    25:  end of Repeat
    26: Output: Regressor RF(x)=1n(T1(x)+T2(x)++TN(x))

     | Show Table
    DownLoad: CSV

    The total number of samples used in the study is 115, including 35 labeled (HAMD score) samples and 80 unlabeled samples. In order to obtain general results, we perform 5 fold cross-validation on 35 labeled samples. One of the folds (7 samples) is used as the test set to evaluate the model each time. The rest are used as the training set to train the model. In the iterative training process, 80 unlabeled samples are reused to achieve semi-supervised learning. Finally, five evaluation models are obtained, and the evaluation results of the five models are averaged to obtain the final evaluation value.

    In E-CoGRF algorithm, the number of decision trees in random forest is set to 12, and the max number of iterations is set to 100.

    Mean Absolute Error (MAE) and Root Mean Square Error (RMSE) are used to evaluate the results of the model.

    MAE=1mmi=1|y(i)ˆy(i)| (5)
    RMSE=1mmi=1(y(i)ˆy(i))2 (6)

    where, m is the sample size, y(i) is the actual value of the i-th sample, ˆy(i) is the predicted value of the i-th sample.

    In order to obtain the best grouping effect, we compare the prediction accuracy of the proposed E-CoGRF algorithm for HAMD scores when the grouping number is 2, 3, 4, and 6, as shown in Table 5 below. Among them, when the number of groups is 4, the lowest prediction error and the largest improvement range are obtained. Therefore, the optimal number of groups of 12 decision trees is 4.

    Table 5.  Performance comparison of E-CoGRF with different number of groups.
    groups MAE RMSE
    Initial Final Improved Initial Final Improved
    2 3.97 ± 0.64 3.82 ± 0.63 3.78% 4.97 ± 0.91 4.73 ± 0.88 4.83%
    3 3.88 ± 0.72 3.7 ± 0.65 4.64% 4.91 ± 0.79 4.63 ± 0.75 5.70%
    4 4.15 ± 0.89 3.63 ± 0.65 12.53% 5.12 ± 1.04 4.50 ± 0.97 12.11%
    6 4.09 ± 0.51 3.78 ± 0.44 7.58% 5.06 ± 0.77 4.71 ± 0.73 6.92%
    Note: Initial: the prediction error of the model trained on the initial labeled sample set.
    Final: the prediction error of the model after 100 iterations.
    Improved: the improvement rate of 'Final' compared to 'Initial'.

     | Show Table
    DownLoad: CSV

    The following experimental analysis is performed with 12 decision trees divided into 4 groups, and the number of decision trees within each group is 3.

    E-CoGRF is compared with CoRF, a co-training random forest regression algorithm, as shown in Table 6. It can be seen from the table that both the CoRF and E-CoGRF improve the evaluation accuracy of HAMD scores, which shows that they can effectively use unlabeled samples for regression estimation. Further analysis shows that compared with CoRF, E-CoGRF has lower error and shorter runtime. This indicates that introduction of the grouping strategy and the information entropy strategy effectively improves the performance of the model.

    Table 6.  Performance comparison between CoRF and E-CoGRF.
    MAE RMSE Runtime(s)
    Initial Final Improved Initial Final Improved
    CoRF 4.22 ± 0.74 3.76 ± 0.60 10.90% 5.34 ± 0.91 4.82 ± 0.84 9.74% 7.77 × 102
    E-CoGRF 4.15 ± 0.89 3.63 ± 0.65 12.53% 5.12 ± 1.04 4.50 ± 0.97 12.11% 1.12 × 102
    Note: Initial: the prediction error of the model trained on the initial labeled sample set.
    Final: the prediction error of the model after 100 iterations.
    Improved: the improvement rate of 'Final' compared to 'Initial'.

     | Show Table
    DownLoad: CSV

    Figure 5 draws the error curves of the two algorithms in iteration process. It can be intuitively seen that E-CoGRF achieves the best convergence accuracy for both MAE and RMSE.

    Figure 5.  Error convergence curves in iteration process.

    In order to further verify the performance of the proposed E-CoGRF, five semi-supervised regression algorithms are applied to evaluate HAMD scores of depression. The comparative algorithms include co-training methods based on different basic learners (#1–#4) and a self-training method based on random forest (#5). As shown in Table 7.

    Table 7.  Comparison of evaluation results of several semi-supervised regression methods.
    # Method Regressor MAE RMSE
    1 CoBCReg (2009) [28] RBFNN 4.07 4.68
    2 SVR-KNN CoREG (2014) [25] SVR and KNN 4.63 5.51
    3 COSVR (2019) [33] SVR 5.22 5.85
    4 Co-training PLS (2020) [27] PLS 4.13 4.81
    5 Self-training RF (2017) [30] RF 3.97 4.54
    6 Proposed RF 3.63 4.50

     | Show Table
    DownLoad: CSV

    Among the various co-training methods, CoBCReg achieves the best accuracy, which reveals the important role of ensemble learning in semi-supervised co-training regression. Compared with RBFNN (Radial Basis Function Neural Networks), KNN (K-Nearest Neighbor), SVR (Support Vector Regression), PLS (Partial Least Squares Regression) and other basic learners, RF shows better performance (#5). The proposed algorithm combines RF with co-training, and achieves better performance than the other five algorithms.

    In this study, a semi-supervised random forest regression model based on co-training and grouping with information entropy is proposed and applied to evaluate depression symptoms severity. Finally, the evaluation error of MAE = 3.63 and RMSE = 4.50 is obtained on 35 labeled (HAMD-17 scores) and 80 unlabeled samples.

    Firstly, the conflict/non-conflict decision-making tasks of two different emotions, happiness and sadness, are constructed based on the basic characteristics of depression affective disorders. Then four kinds of decision-making behavior data (reaction time) of conflict/non-conflict/conflict monitoring/conflict resolution are collected. On this basis, 36 features are constructed according to the statistical distribution. And then the features are normalized.

    Secondly, random forest is introduced into semi-supervised co-training algorithm. On this basis, a grouping strategy is proposed to balance the accuracy and diversity of the model, and a confidence measure method based on information entropy is proposed to reduce unnecessary repeated training in the model. Compared with CoRF, E-CoGRF achieves the lowest error and the shortest runtime.

    In addition, compared with other semi-supervised regression algorithms in recent years, E-CoGRF achieves better evaluation performance. This paper provides an effective method to solve the problem of less labeled samples and more unlabeled samples, and provides technical support for the evaluation of depression symptoms severity.

    The automatic evaluation method of HAMD proposed in this study not only solves the problem of low efficiency of manual evaluation in clinical, but also improves the evaluation accuracy.

    The work only focuses on the evaluation of depression symptoms severity in patients with depression, but does not involve an automatic evaluation method for depression risk. Moreover, due to the limitations of data type, deep learning method is not applied. Based on this, we will further explore this issue by using facial expression, speech, language and other features acquired under emotional stimulation and deep learning methods in the future.

    This work is supported by the National Natural Science Foundation of China (61602017), the National Basic Research Programme of China (2014CB744600), "RixinScientist" Foundation of Beijing University of Technology (2017-RX(1)-03), the Beijing Natural Science Foundation (4164080), the Beijing Outstanding Talent Training Foundation (2014000020124G039), the National Natural Science Foundation of China (61420106005), the International Science & Technology Cooperation Program of China (2013DFA32180), the Special Fund of Beijing Municipal Science and Technology Commission (Z171100000117004), Beijing Hospitals Authority Youth Programme (QML20181904), Beijing Municipal Administration of Hospitals Clinical Medicine Development of Special Funding Support (ZYLX201607) and Beijing Municipal Administration of Hospitals' Ascent Plan (DFL20151801).

    The authors declare there is no conflict of interest.



    [1] G Casalino, G Castellano, F Galetta, K. Kaczmarek-Majer, Dynamic incremental semi-supervised fuzzy clustering for bipolar disorder episode prediction, in International Conference on Discovery Science, Springer, Cham, (2020), 79-93.
    [2] J. C. Wakefield and S. Demazeux, Introduction: Depression, one and many, Sadness or Depression?, Netherlands, Springer, 2016, 1-15.
    [3] M. E. Gerbasi, A. Eldar-Lissai, S. Acaster, M. Fridman, V. Bonthapally, P. Hodgkins, et al., Associations between commonly used patient-reported outcome tools in postpartum depression clinical practice and the Hamilton Rating Scale for Depression, Arch. Women's Mental Health, 23 (2020), 727-735.
    [4] C. L. Allan, C. E. Sexton, N. Filippini, A. Topiwala, A. Mahmood, E. Zsoldos, et al., Sub-threshold depressive symptoms and brain structure: A magnetic resonance imaging study within the Whitehall Ⅱ cohort, J. Affective Disord., 204 (2016), 219-225.
    [5] X. Li, Z. Jing, B. Hu, J. Zhu, N. Zhong, M. Li, et al., A resting-state brain functional network study in MDD based on minimum spanning tree analysis and the hierarchical clustering, Complexity, 2017 (2017), 9514369.
    [6] K. Yoshida, Y. Shimizu, J. Yoshimoto, M. Takamura, G. Okada, Y. Okamoto, et al., Prediction of clinical depression scores and detection of changes in whole-brain using resting-state functional MRI data with partial least squares regression, Plos One, 12 (2017), e0179638.
    [7] S. Sun, X. Li, J. Zhu, Y. Wang, R. La, X. Zhang, et al., Graph theory analysis of functional connectivity in major depression disorder with high-density resting state EEG data, IEEE Trans. Neural Syst. Rehabil. Eng., 27 (2019), 429-439.
    [8] U. R. Acharya, S. L. Oh, Y Hagiwara, J. Tan, H. Adeli, D. P. Subha, Automated EEG-based screening of depression using deep convolutional neural network, Comput. Methods Prog. Biomed., 161 (2018), 103-113. doi: 10.1016/j.cmpb.2018.04.012
    [9] R. W. Lam, S. H. Kennedy, R. S. McIntyre, A. Khullar, Cognitive dysfunction in major depressive disorder: effects on psychosocial functioning and implications for treatment, Can. J. Psychiatry, 59 (2014), 649-654. doi: 10.1177/070674371405901206
    [10] R. S. McIntyre, D. S. Cha, J. K. Soczynska, H. O. Woldeyohannes, L. A. Gallaugher, P. Kudlow, et al., Cognitive deficits and functional outcomes in major depressive disorder: determinants, substrates, and treatment interventions, Depression Anxiety, 30 (2013), 515-527.
    [11] Y. Kang, X. Jiang, Y. Yin, Y. Shang, X. Zhou, Deep transformation learning for depression diagnosis from facial images, in Chinese Conference on Biometric Recognition, Springer, Cham, (2017), 13-22.
    [12] A. Haque, M. Guo, A. S. Miner, F. Li, Measuring depression symptom severity from spoken language and 3D facial expressions, preprint, arXiv: 1811.08592.
    [13] M. Muzammel, H. Salam, Y. Hoffmann, M. Chetouani, A. Othmani, AudVowelConsNet: A phoneme-level based deep CNN architecture for clinical depression diagnosis, Mach. Learn. Appl., 2 (2020), 100005.
    [14] J. Zhu, J. Li, X. Li, J. Rao, Y. Hao, Z. Ding, et al., Neural basis of the emotional conflict processing in major depression: ERPs and source localization analysis on the N450 and P300 components, Front. Human Neurosci., 12 (2018), 214.
    [15] B. W. Haas, K. Omura, R. T. Constable, T. Canli, Interference produced by emotional conflict associated with anterior cingulate activation, Cognit. Affective Behav. Neurosci., 6 (2006), 152-156. doi: 10.3758/CABN.6.2.152
    [16] T. Armstrong, B. O. Olatunji, Eye tracking of attention in the affective disorders: a meta-analytic review and synthesis, Clin. Psychol. Rev., 32 (2012), 704-723. doi: 10.1016/j.cpr.2012.09.004
    [17] A. Duque, C. Vázquez, Double attention bias for positive and negative emotional faces in clinical depression: Evidence from an eye-tracking study, J Behav. Ther. Exp. Psychiatry, 46 (2015), 107-114. doi: 10.1016/j.jbtep.2014.09.005
    [18] S. P. Karparova, A. Kersting, T. Suslow, Disengagement of attention from facial emotion in unipolar depression, Psychiatry Clin. Neurosci., 59 (2005), 723-729. doi: 10.1111/j.1440-1819.2005.01443.x
    [19] M. P. Caligiuri, J. Ellwanger, Motor and cognitive aspects of motor retardation in depression, J. Affective Disord., 57 (2000), 83-93. doi: 10.1016/S0165-0327(99)00068-3
    [20] A. Etkin, T. Egner, D. M. Peraza, E. R. Kandel, J. Hirsch, Resolving emotional conflict: a role for the rostral anterior cingulate cortex in modulating activity in the amygdala, Neuron, 51 (2006), 871-882. doi: 10.1016/j.neuron.2006.07.029
    [21] K Mohan, A Seal, O Krejcar, A. Yazidi, FER-net: facial expression recognition using deep neural net, Neural Comput. Appl., (2021), 1-12.
    [22] K Mohan, A Seal, O Krejcar, A. Yazidi, Facial expression recognition using local gravitational force descriptor-based deep convolution neural networks, IEEE Trans. Instrum. Meas., 70 (2020), 1-12.
    [23] Z. Zhou, M. Li, Semi-supervised regression with co-training, in IJCAI, (2005), 908-913.
    [24] M. A. Lei, W. Xili, Semi-supervised regression based on support vector machine co-training, Comput. Eng. Appl., 47 (2011), 177-180.
    [25] Y. Q. Li, M. Tian, A semi-supervised regression algorithm based on co-training with SVR-KNN, in Advanced Materials Research, Trans Tech Publications Ltd, (2014), 2914-2918.
    [26] L. Bao, X. Yuan, Z. Ge, Co-training partial least squares model for semi-supervised soft sensor development, Chemom. Intell. Lab. Syst., 147 (2015), 75-85. doi: 10.1016/j.chemolab.2015.08.002
    [27] D. Li, Y. Liu, D. Huang, Development of semi-supervised multiple-output soft-sensors with Co-training and tri-training MPLS and MRVM, Chemom. Intell. Lab. Syst., 199 (2020), 103970.
    [28] M. F. A. Hady, F. Schwenker and G. Palm, Semi-supervised learning for regression with co-training by committee, in International Conference on Artificial Neural Networks, Springer, Berlin, Heidelberg, (2009), 121-130.
    [29] F. Saitoh, Predictive modeling of corporate credit ratings using a semi-supervised random forest regression, 2016 IEEE International Conference on Industrial Engineering and Engineering Management (IEEM), IEEE, (2016), 429-433.
    [30] J. Levatić, M. Ceci, D. Kocev, S. Džeroski, Self-training for multi-target regression with tree ensembles, Knowl. Based Syst., 123 (2017), 41-60. doi: 10.1016/j.knosys.2017.02.014
    [31] S. Xue, S. Wang, X. Kong, J. Qiu, Abnormal neural basis of emotional conflict control in treatment-resistant depression: An event-related potential study, Clin. EEG Neurosci., 48 (2017), 103-110. doi: 10.1177/1550059416631658
    [32] N. Tottenham, J. W. Tanaka, A. C. Leon, T. McCarry, M. Nurse, T. A. Hare, et al., The NimStim set of facial expressions: Judgments from untrained research participants, Psychiatry Res., 168 (2009), 242-249.
    [33] M. Lei, J. Yang, S. Wang, L. Zhao, P. Xia, G. Jiang, et al., Semi-supervised modeling and compensation for the thermal error of precision feed axes, Int. J. Adv. Manuf. Technol., 104 (2019), 4629-4640.
  • This article has been cited by:

    1. Meng Han, Xiaojuan Li, Le Wang, Ni Zhang, Haodong Cheng, Review of ensemble classification over data streams based on supervised and semi-supervised, 2022, 43, 10641246, 3859, 10.3233/JIFS-211101
    2. Anibal Thiago Bezerra, Leonardo Antonio Pinto, Diego Samuel Rodrigues, Gabriela Nogueira Bittencourt, Paulo Fernando de Arruda Mancera, José Ricardo de Arruda Miranda, Classification of gastric emptying and orocaecal transit through artificial neural networks, 2021, 18, 1551-0018, 9511, 10.3934/mbe.2021467
    3. Qun-Xiong Zhu, Hong-Tao Zhang, Ye Tian, Ning Zhang, Yuan Xu, Yan-Lin He, Co-training based virtual sample generation for solving the small sample size problem in process industry, 2022, 00190578, 10.1016/j.isatra.2022.08.021
    4. Hongwei Sheng, Ping Wang, Chenglong Tang, Yu Shi, Yang Zheng, Microstructure and mechanical properties evaluation of automotive plate steel based on micromagnetic NDT technologies, 2022, 199, 02632241, 111459, 10.1016/j.measurement.2022.111459
    5. Li Yan Liu, Peng Huang, Fan Min, 2022, Safe Multi-view Co-training for Semi-supervised Regression, 978-1-6654-7330-9, 1, 10.1109/DSAA54385.2022.10032437
    6. R. Loula, L. H. A. Monteiro, On the criteria for diagnosing depression in bereaved individuals: a self-organizing map approach, 2022, 19, 1551-0018, 5380, 10.3934/mbe.2022252
    7. Kwang-Sig Lee, Byung-Joo Ham, Machine Learning on Early Diagnosis of Depression, 2022, 19, 1738-3684, 597, 10.30773/pi.2022.0075
    8. R. Loula, L. H. A. Monteiro, Monoamine neurotransmitters and mood swings: a dynamical systems approach, 2022, 19, 1551-0018, 4075, 10.3934/mbe.2022187
    9. Liyan Liu, Peng Huang, Hong Yu, Fan Min, Safe co-training for semi-supervised regression, 2023, 27, 1088467X, 959, 10.3233/IDA-226718
    10. Mi Li, Zeying Lu, Qishuang Cao, Junlong Gao, Bin Hu, Automatic Assessment Method and Device for Depression Symptom Severity Based on Emotional Facial Expression and Pupil-Wave, 2024, 73, 0018-9456, 1, 10.1109/TIM.2024.3415778
    11. Yan Zhang, Xin Liu, Panrui Tang, Zuping Zhang, 2024, Chapter 3, 978-981-99-8069-7, 31, 10.1007/978-981-99-8070-3_3
    12. Mingyu Li, Lu Gao, Hongwei Xu, Kai Li, Yisong Huang, Equipment Damage Measurement Method of Wartime Based on FCE-PCA-RF, 2024, 35, 1004-4132, 707, 10.23919/JSEE.2024.000065
    13. Zhengguang Tang, Cong Li, Hailong You, Zhixiong Di, Linying Zhang, Xingming Liu, Yu Wang, Yong Dai, Geng Bai, Semi-Supervised Transfer Learning Framework for Aging-Aware Library Characterization, 2024, 71, 1549-7747, 1156, 10.1109/TCSII.2023.3323384
    14. Liyan Liu, Jin Zhang, Kun Qian, Fan Min, Semi-supervised regression via embedding space mapping and pseudo-label smearing, 2024, 54, 0924-669X, 9622, 10.1007/s10489-024-05686-6
  • Reader Comments
  • © 2021 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(3894) PDF downloads(163) Cited by(14)

Figures and Tables

Figures(5)  /  Tables(7)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog