Research article Special Issues

Prediction of hearing preservation after acoustic neuroma surgery based on SMOTE-XGBoost


  • Prior to the surgical removal of an acoustic neuroma, the majority of patients anticipate that their hearing will be preserved to the greatest possible extent following surgery. This paper proposes a postoperative hearing preservation prediction model for the characteristics of class-imbalanced hospital real data based on the extreme gradient boost tree (XGBoost). In order to eliminate sample imbalance, the synthetic minority oversampling technique (SMOTE) is applied to increase the number of underclass samples in the data. Multiple machine learning models are also used for the accurate prediction of surgical hearing preservation in acoustic neuroma patients. In comparison to research results from existing literature, the experimental results found the model proposed in this paper to be superior. In summary, the method this paper proposes can make a significant contribution to the development of personalized preoperative diagnosis and treatment plans for patients, leading to effective judgment for the hearing retention of patients with acoustic neuroma following surgery, a simplified long medical treatment process and saved medical resources.

    Citation: Cenyi Yang. Prediction of hearing preservation after acoustic neuroma surgery based on SMOTE-XGBoost[J]. Mathematical Biosciences and Engineering, 2023, 20(6): 10757-10772. doi: 10.3934/mbe.2023477

    Related Papers:

    [1] Jianing Wang . Prediction of postoperative recovery in patients with acoustic neuroma using machine learning and SMOTE-ENN techniques. Mathematical Biosciences and Engineering, 2022, 19(10): 10407-10423. doi: 10.3934/mbe.2022487
    [2] Yunxiang Meng, Qihong Duan, Kai Jiao, Jiang Xue . A screened predictive model for esophageal squamous cell carcinoma based on salivary flora data. Mathematical Biosciences and Engineering, 2023, 20(10): 18368-18385. doi: 10.3934/mbe.2023816
    [3] Sarth Kanani, Shivam Patel, Rajeev Kumar Gupta, Arti Jain, Jerry Chun-Wei Lin . An AI-Enabled ensemble method for rainfall forecasting using Long-Short term memory. Mathematical Biosciences and Engineering, 2023, 20(5): 8975-9002. doi: 10.3934/mbe.2023394
    [4] Lili Jiang, Sirong Chen, Yuanhui Wu, Da Zhou, Lihua Duan . Prediction of coronary heart disease in gout patients using machine learning models. Mathematical Biosciences and Engineering, 2023, 20(3): 4574-4591. doi: 10.3934/mbe.2023212
    [5] Hu Dong, Gang Liu, Xin Tong . Influence of temperature-dependent acoustic and thermal parameters and nonlinear harmonics on the prediction of thermal lesion under HIFU ablation. Mathematical Biosciences and Engineering, 2021, 18(2): 1340-1351. doi: 10.3934/mbe.2021070
    [6] Ji Zhou, Zhanlin Su, Shahab Hosseini, Qiong Tian, Yijun Lu, Hao Luo, Xingquan Xu, Chupeng Chen, Jiandong Huang . Decision tree models for the estimation of geo-polymer concrete compressive strength. Mathematical Biosciences and Engineering, 2024, 21(1): 1413-1444. doi: 10.3934/mbe.2024061
    [7] Yuanyao Lu, Kexin Li . Research on lip recognition algorithm based on MobileNet + attention-GRU. Mathematical Biosciences and Engineering, 2022, 19(12): 13526-13540. doi: 10.3934/mbe.2022631
    [8] Sidra Abid Syed, Munaf Rashid, Samreen Hussain . Meta-analysis of voice disorders databases and applied machine learning techniques. Mathematical Biosciences and Engineering, 2020, 17(6): 7958-7979. doi: 10.3934/mbe.2020404
    [9] Hyeonjeong Ahn, Hyojung Lee . Predicting the transmission trends of COVID-19: an interpretable machine learning approach based on daily, death, and imported cases. Mathematical Biosciences and Engineering, 2024, 21(5): 6150-6166. doi: 10.3934/mbe.2024270
    [10] Jinyi Tai, Chang Liu, Xing Wu, Jianwei Yang . Bearing fault diagnosis based on wavelet sparse convolutional network and acoustic emission compression signals. Mathematical Biosciences and Engineering, 2022, 19(8): 8057-8080. doi: 10.3934/mbe.2022377
  • Prior to the surgical removal of an acoustic neuroma, the majority of patients anticipate that their hearing will be preserved to the greatest possible extent following surgery. This paper proposes a postoperative hearing preservation prediction model for the characteristics of class-imbalanced hospital real data based on the extreme gradient boost tree (XGBoost). In order to eliminate sample imbalance, the synthetic minority oversampling technique (SMOTE) is applied to increase the number of underclass samples in the data. Multiple machine learning models are also used for the accurate prediction of surgical hearing preservation in acoustic neuroma patients. In comparison to research results from existing literature, the experimental results found the model proposed in this paper to be superior. In summary, the method this paper proposes can make a significant contribution to the development of personalized preoperative diagnosis and treatment plans for patients, leading to effective judgment for the hearing retention of patients with acoustic neuroma following surgery, a simplified long medical treatment process and saved medical resources.



    Acoustic neuroma, which is also known as vestibular schwannoma, is a benign tumor that arises from the vestibular nerve sheath [1]. It is most commonly found in the internal acoustic canal and the cerebellopontine angle, and can result in hearing loss, tinnitus, dizziness, unsteady walking, face numbness and other symptoms [2]. Surgical excision is the most common treatment option as this effectively removes the tumor, but problems such as hearing loss or disabilities can result, which can significantly impact the daily lives of patients [3,4]. As a result, the removal of the entire tumor while maintaining hearing is the optimum therapeutic goal for acoustic neuromas. Based on their clinical and pathological appearance, how can the postoperative hearing status of individuals with acoustic neuroma be accurately predicted? This question has clinical importance in terms of relevant clinical trials and for guiding the individualized treatment and management of patients.

    For postoperative hearing preservation prediction, if only the experience of the doctor is relied upon, it may be time-consuming and laborious, potentially leading to accidental errors. Machine learning is a fast and effective method for the analysis of patient characteristics, obtaining discriminatory results in a timely manner and assessing the validity of the results. It is widely used in the medical field and is widely regarded as being superior to traditional statistical methods.

    For example, in diabetes clinical epidemiology, supervised and unsupervised learning are both widely used, and a study demonstrated that machine learning methods can help improve risk stratification and outcome prediction in diabetes epidemiology applications [5]. In the field of cardiology, scholars summarized 40 research papers relating to machine learning since 2018, comparing seven machine learning models and arguing that machine learning algorithms can improve the healthcare industry by making improvements to the entire disease prediction and treatment recommendation process in the vast majority of cases [6]. A review summarized 14 studies on machine learning for the prediction colorectal cancer (CRC) risk stratification from 2011 to 2022, arguing that as an adjunct to disease prediction, it can help patients and clinicians identify patients who are at high risk for CRC and diagnose and treat them more quickly [7]. In the field of heart disease, scholars have used a variety of machine learning methods for effectively predicting the preincidence of heart disease [8,9,10,11]. In addition, machine learning has achieved desirable results in other fields, including biomedicine [12,13,14,15,16,17,18].

    However, to the best of our knowledge, the study of the prognostic factors that affect postoperative hearing preservation among patients with auditory neuroma remains in its infancy. Most previous research is based on traditional statistical methods such as ANOVA, and this does not provide an available predictive model of postoperative hearing preservation for patients [19,20,21,22,23]. In the scope of this research, it was recognized that Cha et al. took the integration of machine learning methods with hearing preservation studies a step further in 2020, proposing a regression model to facilitate the use of analysis by physicians to provide a referenceable prediction of postoperative hearing preservation for patients prior to surgery [24]. Four models were used to fit and regress the data and obtain the most accurate model through experimental comparison: support vector machine (SVM), gradient growth machine (GBM), deep neural network (DNN), and diffusion random forest (DRF). This work inspired us to expand on the idea that hospital data often has the problem of class imbalance, and obtaining the best prediction results by direct model fitting may be difficult. Therefore, perhaps a data imbalance treatment method could be added. Furthermore, while their best model was DNN, which runs slowly and produces unstable results, other models could be considered for fitting attempts.

    XGBoost is an ensemble learning classification method that is currently one of the most popular methods used in machine learning areas. Its main idea is improving the stability and accuracy of the model by combining multiple weak classifiers to form a strong classifier [25]. It is widely used and has been found to outperform other models in various medical fields. For example, Wimalarathna et al. proved XGBoost is the best model for the classification of auditory brainstem responses recorded from children with auditory processing disorder [26]. Kivrak et al. found XGBoost to be the most suitable model for the prediction of death among COVID-19 patients [27]. Shorthouse et al. concluded that XGBoost is effective for separating cancer from noncancerous cells [28]. Chen et al. proved that XGBoost provides accurate predictions of disease risk progression stratification and long-term prognosis among kidney patients [29]. XGBoost has also demonstrated excellent prediction capabilities in many sectors outside of healthcare [30,31,32,33].

    As was previously mentioned, a detailed study of data from patients with acoustic neuroma is presented in this paper, and three machine learning methods are used: DNN, classification and regression trees (CART), and XGBoost in the study of hearing preservation among patients with acoustic neuroma. A SMOTE-XGBoost model is proposed that incorporates an oversampling imbalance processing method based on SMOTE. The SMOTE-XGBoost model is compared to other machine learning models that use the same data preprocessing, and to the same model both before and after data imbalance processing in this work. The outcome demonstrates the viability of the data processing approach used based on the properties of the data, in addition to the higher performance of the model, which demonstrates its superiority.

    The article is organized in the following manner: Section 2 analyzes the methods used in this article are analyzed in Section 2; Section 3 describes the experimental data and required materials in detail; Section 4 shows the data processing flow and provides a discussion of the experimental results; and Section 5 outlines the conclusions that have been reached.

    Figure 1.  Predictive modeling flowchart for postoperative hearing preservation among patients with auditory neuroma.

    In this section, the two most important methods in the model this paper proposes are described in detail: SMOTE and XGBoost. The second subsection outlines the theory foreshadowing for the third subsection. It is taken as a comparative model in numerical experiments.

    There is an unbalanced sample proportion, which is commonly referred to as data imbalance. With classification problems, assuming the quantity of samples of various classes to be roughly equal is common. However, in the field of acoustic neuroma that is described here, the number of patients with better hearing preservation following surgery is far greater than the number of patients with poorer hearing preservation, and this paper is more concerned with the small percentage of patients who are likely to have poorer hearing recovery.

    There are several approaches for dealing with data imbalance, including oversampling and undersampling, which are the two most prevalent processing methods [34]. As there is such a small sample of actual medical data, utilizing the oversampling method as a means of retaining as much sample information as possible is preferable.

    The initial oversampling approach simply copied the few samples, but this would result in data overfitting and would not increase the accuracy of judgment. This difficulty can be solved by SMOTE [35].

    This is an improved scheme that is based on the random oversampling algorithm. The basic premise is to analyze the minority samples and artificially synthesize new samples based on them to add to the data set, which is represented by the algorithm flow and Figure 2.

    Figure 2.  Schematic diagram of SMOTE algorithm.

    The SMOTE algorithm flow is summarized as follows.

    (1) For each minority sample xi, the Euclidean distance from xi to all other minority samples is calculated to obtain its K nearest neighbors;

    (2) The oversampling ratio is specified according to the sample category imbalance ratio, N samples are randomly selected from the K nearest neighbor samples, and it is assumed that the selected sample is xj;

    (3) For each randomly selected sample xj, a new sample xnew with the original sample xi is generated according to the following formula

    xnew=xi+gap×(xjxi), (2.1)

    where gap is a random number in the interval (0, 1).

    The CART algorithm is an extensive decision tree learning method that was initially proposed by Breiman [36]. It is used for both classification and regression. The following introduces the classification function that is used in this paper.

    A classification decision tree is generated through the process of selecting the optimal feature based on the criterion of minimizing the Gini index and the optimal split point in the optimal feature and recursively generating a binary tree. In the classification problem, if it is assumed that there are K classes, the probability of the sample point belongings to the k-th class is pk, then the Gini index of the probability distribution is

    Gini(p)=Kk=1pk(1pk)=1Kk=1p2k. (2.2)

    The specific CART algorithm calculation process can be seen in Figure 3. This can be summarized as follows.

    Figure 3.  CART tree splitting process.

    (1) For the current node data set D, recursion cannot be performed when one of the following three conditions are met: the number of samples in D is less than the specified threshold; there is no feature in D; the Gini index is less than the specified threshold;

    (2) The Gini index of each value of each existing feature of the current node is calculated, and the feature with the smallest Gini index is then selected as the division point. Taking A=a as an example, it divides the node data set into two sub-nodes D1 and D2 in Figure 3;

    (3) The above steps are performed recursively on the generated sub-nodes until the termination condition is met.

    XGBoost is the integration of trees. The basic classifier is generally selected as the CART tree, which is also the case in this paper. Each tree is added as a means of improving the classification effect. The calculation formula is

    ^yi=Kt=1ft(xi), (2.3)

    where ^yi is the predicted output result of the i-th sample, and ft is the t-th decision tree. For such prediction output, its objective function is

    Obj=ni=1l(yi,^yi)+kt=1Ω(ft(xi)), (2.4)

    where ni=1l(yi,^yi) is the loss function. The commonly used loss function is the sum of squares of the residuals L=ni=1(yi^yi)2, which represents the deviation of the predicted value from the true value. Ω(ft(xi)) is a penalty term, which is, specifically expressed as Ω(ft)=γT+12λTj=1ω2j, where T represents the number of leaves in the tree, ωj represents the score on j-th leaf in Figure 4, γ and λ are both custom coefficients. The equation clearly shows that the more trees there are, the larger the term is, so this is used for preventing overfitting due to too many tree splits.

    Figure 4.  Tree ensemble model (the final prediction for a given example is the sum of predictions from each tree).

    The objective function is expanded with Taylor's formula binomial and the following is obtained,

    Objt=ni=1l(yi,^yit)+kt=1Ω(ft(xi))ni=1[gift(xi)+12hif2t(xi)]+kt=1Ω(ft(xi))=Tj=1[(iIjgi)ωj+12(iIjhi+λ)ω2j]+γT, (2.5)

    where gi=ˆyt1l(yi,ˆyt1i),hi=2ˆyt1l(yi,ˆyt1). The expression can be simplified thusly

    Gj=iIjgi,Hj=iIjhi.

    The objective function is then

    Objt=Tj=1[Gjωj+12(Hj+λ)ω2j]+γT. (2.6)

    In the above formula, each ωj is independent of each other. For the quadratic equation Gjωj+12(Hj+λ)ω2j, it is relatively easy to find ωj under the minimum value of the objective function when the structure of this additional tree is known

    ωj=GjHj+λ. (2.7)

    Substituting Eq (2.7) for Eq (2.6), the objective function can be further simplified as

    Obj=12Tj=1G2jHj+λ+γT, (2.8)

    which is the corresponding optimal objective function value. It is assumed that IL and IR are the instance sets of left and right nodes following the split. I=ILIR as the tree on the left in Figure 4, where Obj1 represents the objective function value of I, and Obj2 represents the objective function value of IL and IR, and the loss reduction following the split can be obtained by

    Gain=Obj1Obj2=[12(GL+GR)2HL+HR+λ+γ][12(G2LHL+λ+G2RHR+λ)+2γ]=12[G2LHL+λ+G2RHR+λ(GL+GR)2HL+HR+λ]γ. (2.9)

    From the above criteria, in order to sort the values obtained by dividing each feature at different values, the various divisions must be traversed to identify the best one and generate a decision tree.

    A brief description of the data set used in this paper is provided in this section and the criteria for comparing model quality in subsequent experiments are identified.

    The data used in this article was gathered by physicians following surgery from the Department of Neurosurgery, Xiangya School of Medicine, Central South University. There is a high level of authenticity and trustworthiness to the data set that includes 79 individuals who were chosen with the help of doctors for 13 characteristics that are potentially related to hearing following acoustic neuroma surgery, including quantitative and categorical parameters. Hearing preservation following surgery is the dependent variable. Negative class 0 refers to the hearing of a patient not changing or improving following the procedure in comparison to preoperative hearing; positive class 1 refers to the hearing of a patient deteriorating following the operation in comparison to preoperative hearing. The distinguishing characteristics are shown in Table 1.

    Table 1.  Data set features.
    Feature no. Feature description Identification Type
    1 Age Patient age Numerical
    2 Hypertension (y/n) History of hypertension Categorical
    3 Diabetes (y/n) History of diabetes Categorical
    4 Others (y/n) History of other illnesses Categorical
    5 Position Tumor position Categorical
    6 Pre-class Preoperative hearing AAO-HNS classification Categorical
    7 Symptom Number of preoperative symptoms Categorical
    8 Tumor-size Tumor volume Numerical
    9 Tumor-status Tumor cystic status Categorical
    10 Shape Ear hearing channel shape Categorical
    11 Samii Samii grading Categorical
    12 TFIAC TFIAC grading Categorical
    13 Take-medicine (y/n) Whether medication is taken on time Categorical
    14 Class (target variable) Whether hearing is preserved following surgery Categorical

     | Show Table
    DownLoad: CSV

    Due to the imbalance in the initial data set and the small sample number, model performance evaluation in this paper employs evaluation metrics that are linked to the ROC curve rather than Accuracy [37]. The confusion matrix, which is shown in Figure 5, can be used for depicting the classification performance of the two classification problems. The majority class is represented by the negative class in this paper, while the minority class is represented by the positive class. The real class label of the sample appears in the first column of the confusion matrix table, while the anticipated class label is in the first row. The accurately predicted positive label is represented by TP and the number of samples with negative labels is represented by TN, while the misclassified positive labels and negative labels are represented by FN and FP.

    Figure 5.  Confusion matrix.

    Based on the confusion matrix, the indicators for the evaluation of the discriminant model of unbalanced data are

    Accuracy=TP+TNTP+FN+FP+TNTP rate= Sensitivity = Recall =TPTP+FNPrecision=TPTP+FPFP rate=FPTN+FPF1=2×Recall×PrecisionRecall+Precision. (3.1)

    Accuracy represents overall accuracy, and TP rate is the percentage of correct classification of positive samples, which refers to the proportion of patients whose hearing is not successfully preserved following surgery and is correctly predicted. Sensitivity represents the percentage of correct classification of negative samples, which refers to the percentage of patients whose hearing is preserved following surgery and is correctly predicted. FP rate is the proportion of patients whose hearing is not preserved following incorrect prediction. Precision is the percentage of samples predicted to be positive, which is the percentage of patients who are predicted to be hearing unreserved. F1 is the harmonic average of Precision and Sensitivity, which simply and clearly reflects the balance between the two.

    The ROC curve is a common indicator for the evaluation of imbalanced data prediction models and is a two-dimensional graph with FP rate as the abscissa and TP rate as the ordinate, where (0, 1) is the ideal point. This curve demonstrates the relationship between FP rate and TP rate. AUC is the area under the ROC curve, AUC=1 means a perfect classifier, and when AUC is less than 0.5, this means the model is not as effective as random guessing.

    At the same time, the PR curve reflects the trade-off between the accuracy of the classifier in recognizing positive examples and its ability to cover positive examples, taking Precision as the vertical coordinate and Recall as the horizontal coordinate. The higher the accuracy of the model, the higher the recall will be, meaning the performance of the model is better. Average precision (AP) is the average accuracy at different recall points and is represented on the PR graph as the area under the PR curve. The larger the AP value, the higher the average accuracy of the model will be.

    The imbalance treatment and model fitting process according to the experimental sequence is detailed in this section and the superiority of the model is demonstrated through a comparison between each index.

    According to the 79 case samples in the data set used in this paper, 68 patients retained or improved their pre-operative hearing level following acoustic neuroma cutting surgery, while 11 patients had poor hearing after surgery. The distribution of the two types of hearing following surgery is obviously quite uneven, which causes the classification results to be biased toward multi-class samples and classification accuracy to be lower. It is apparent that patients are more concerned about hearing deterioration following surgery and doctors are also interested in those who have lost hearing following surgery.

    As a result, this paper performed imbalance processing on the data before classifying and regressing to change the sample classification from 68:11 to 68:68, thereby making training a better classifier easier and enabling a more accurate classification for a small amount of data to be obtained.

    Following the imbalance treatment, five-fold cross-validation was used to more reasonably compare the effects of the models, and predictive entropy was used for estimating the uncertainty of the individual models. DNN and CART were utilized as comparisons for XGBoost in the studies, one of them being the model used in the hearing preservation study by Cha et al. and the other the basis model for XGBoost. In addition to cross-sectional comparisons, these three models provide comparisons with findings before and after imbalance treatment. Table 2 shows that all models were significantly improved in all metrics following SMOTE treatment and the floating range of each metric and entropy(Accuracy,F1) demonstrate more stable performance with the SMOTE-treated model. At the same time, paired t-test was conducted for the F1 of each model both before and after SMOTE treatment. The results presented in Table 3 were obtained, which shows that p<0.1, thereby indicating a significant difference before and after SMOTE treatment. As a result, class imbalanced data requires imbalanced processing.

    Table 2.  Comparison of evaluation indicators of the model.
    Model Accuracy Sensitivity Precision F1 entropy(Accuracy,F1)
    DNN 0.82±0.05 0.27±0.45 0.37±0.37 0.29±0.26 0.15, 0.36
    SMOTE-DNN 0.85±0.12 0.85±0.08 0.83±0.19 0.83±0.14 0.15, 0.15
    CART 0.72±0.24 0.55±0.26 0.53±0.24 0.52±0.24 0.24, 0.34
    SMOTE-CART 0.85±0.10 0.85±0.10 0.87±0.08 0.85±0.11 0.14, 0.14
    XGBoost 0.86±0.17 0.67±0.38 0.64±0.43 0.65±0.41 0.13, 0.28
    SMOTE-XGBoost 0.91±0.13 0.90±0.13 0.92±0.11 0.90±0.14 0.09, 0.09

     | Show Table
    DownLoad: CSV
    Table 3.  Paired t-test.
    Model SD SEM 95%CI t P
    DNN 0.36406 0.16281 (0.993 - 0.08891) 3.322 0.029
    CART 0.16291 0.07286 (0.53233 - 0.12776) 4.530 0.011
    XGBoost 0.23146 0.10351 (0.53992 - 0.03487) 2.440 0.071

     | Show Table
    DownLoad: CSV

    XGBoost also demonstrated significant advantages over other models both before and after SMOTE processing. It achieved the best results for Accuracy, Sensitivity, Precision and F1, while also achieving the minimum predicted entropy to prove the stability of the model. SMOTE- XGBoost outperformed even 0.9 in all evaluation metrics, but also had the lowest predictive entropy of less than 0.1, meaning that it can be considered the best model for the prediction of postoperative hearing preservation among patients with auditory neuroma.

    Due to the imbalance of the original dataset, a 7:3 training set test set cut of the data set was made, and ROC curves and PR curves were used on the test set for further exploring model performance. The joint curves of the six models were then obtained, as can be seen in Figure 6.

    Figure 6.  Comparison of the curves of each model.

    Multiple dashed lines in Figure 6 show that the models without data imbalance treatment and with AUC and AP values around 0.5 and below are close to the random prediction, meaning the prediction results have no reference value. Conversely, the solid line indicates all models with data imbalance treatment performed well. The SMOTE-XGBoost, which is represented by the blue curve, was found to be the best model, achieving an AUC value of 0.96 and an AP value of 0.97.

    Based on the selected model, the importance of each feature in the model was determined, as can be seen in Figure 7. In the figure, F-score represents the weighted summed mean value of the information gain from feature splitting. A higher F-score means the feature has greater importance. Therefore, it is recognized that tumor size, age, pre-class, symptoms and tumor status are the five most important characteristics that affect postoperative hearing preservation.

    Figure 7.  The feature importance of the model.

    All patients underwent microscopic intracranial occupancy resection via the posterior suboccipital sigmoid sinus approach from the same operator (same treatment group), and subsequent analysis excluded the effect of intraoperative variables. For patients with poor predicted postoperative hearing preservation, a more observational, conservative treatment plan based on individualized care must be adopted, so the surgeon is able to assess the risks and benefits in order to make further choices based on the actual situation. For those with better predicted postoperative hearing preservation, their preoperative fears can be largely eliminated and good surgery ensured.

    Based on the previous discussion, the main results of this work can be summarized as follows.

    (1) XGBoost was introduced into the field of auditory neuroma and found to be superior to DNN. The proposed model in this study performed well in all metrics, also outperforming the neural network in terms of model fitting process time.

    (2) The imbalance treatment was applied to patient data in the field of auditory neuroma based on the inherent distribution characteristics of auditory neuroma. The need for its introduction was demonstrated as this would allow for a greater focus on the small number of patients whose hearing is not preserved postoperatively while avoiding serious treatment misjudgments by physicians.

    (3) The preoperative indicators with the greatest impact on postoperative hearing preservation in patients were analyzed using the selected rational prediction model. The importance of age, preoperative hearing AAO-HNS classification, tumor size, and tumor status has been repeatedly mentioned in previous relevant medical studies [38,39,40], while the importance of symptoms, a characteristic summarized to represent the number of preoperative symptoms in patients, has rarely been mentioned in previous studies, but this can be further confirmed in future medical observations.

    Only 79 samples were used in this study, so the potential of the proposed model was not fully demonstrated, potentially leading to some overfitting of the prediction results. This was due to the fact that there are few patients with auditory neuroma and relatively few hospitals where relevant surgery is conducted, which makes the collection of various types of data quite difficult. In future studies, the information collection system should be improved to enable more patient data to be obtained and the model to be updated.

    A practical predictive model for hearing preservation in patients with acoustic neuroma following tumor cutting surgery was proposed by this paper. The model was found to have obvious credibility in many important indicators, thereby validating the feasibility of the introduction of machine learning into the field of acoustic neuromas.

    The model also incorporated a data imbalance processing method based on data distribution characteristics, compensating for the lack of direct model regression in previous research that did not consider data distribution characteristics. This served to improve the accuracy and specificity of the proposed model in a variety of ways. For example, it was able to reliably forecast a larger sample of patient data sets from different categories without the need to account for the flaws of unbalanced data distribution.

    In practical terms, the results of this paper demonstrate the application of ensemble learning algorithms for the prediction of the preservation of patient hearing following surgery, thereby saving considerable time and money for clinicians diagnosing the postoperative hearing of patients. As a result, it is believed that this model could be applied clinically. As more data on patients is gathered, it is anticipated that the model proposed by this paper could be adjusted appropriately as a means of producing better outcomes.

    The author is grateful to Miss Yue Li from The Xiangya Hospital of Central South University for her kind support with data collection for this project. This work is partly supported by the Fundamental Research Funds for the Central Universities of Central South University (No. 1053320213996).

    The author declares there to be no conflict of interest.



    [1] M. Samii, Surgery of Cerebellopontine Lesions, Springer-Verlag Berlin Heidelberg, 2013. https://doi.org/10.1007/978-3-642-35422-9_5
    [2] E. S. Murphy, J. H. Suh, Radiotherapy for vestibular schwannomas: A critical review, Int. J. Radiat. Oncol. Biol. Phys., 79 (2011), 985–997. https://doi.org/10.1016/j.ijrobp.2010.10.010 doi: 10.1016/j.ijrobp.2010.10.010
    [3] M. Samii, V. M. Gerganov, A. Samii, Functional outcome after complete surgical removal of giant vestibular schwannomas, J. Neurosurg., 112 (2010), 860–867. https://doi.org/10.3171/2009.7.JNS0989 doi: 10.3171/2009.7.JNS0989
    [4] M. Samii, V. M. Gerganov, A. Samii, Quasi-morphisms and the Poisson bracket, J. Neurosurg., 40 (1997), 248–262. https://doi.org/10.1097/0006123-199701000-00001 doi: 10.1097/0006123-199701000-00001
    [5] S. Basu, K. T. Johnson, S. A. Berkowitz, Use of machine learning approaches in clinical epidemiological research of diabetes, Curr. Diab. Rep., 20 (2020), 80–99. https://doi.org/10.1007/s11892-020-01353-5 doi: 10.1007/s11892-020-01353-5
    [6] J. Azmi, M. Arif, M. T. Nafis, M. A. Alam, S. Tanweer, G. Wang, A systematic review on machine learning approaches for cardiovascular disease prediction using medical big data, Med. Eng. Phys., 105 (2022), 103825.
    [7] K. Oliver, M. Stuart, B. Richard, Machine learning as a new horizon for colorectal cancer risk prediction? A systematic review, Health Sci. Rev., 105 (2022), 10041.
    [8] O. W. Samuel, G. M. Asogbon, A. K. Sangaiah, P. Fang, G. Li, An integrated decision support system based on ANN and fuzzy-AHP for heart failure risk prediction, Expert Syst. Appl., 68 (2017), 163–172. https://doi.org/10.1016/j.eswa.2016.10.020 doi: 10.1016/j.eswa.2016.10.020
    [9] Z. Arabasadi, R. Alizadehsani, M. Roshanzamir, H. Moosaei, A. A. Yarifard, Computer aided decision making for heart disease detection using hybrid neural network-genetic algorithm, Comput. Methods Programs Biomed., 141 (2017), 19–26. https://doi.org/10.1016/j.cmpb.2017.01.004 doi: 10.1016/j.cmpb.2017.01.004
    [10] Y. J. King, M. Saqlian, J. Y. Lee, Deep learning-based prediction model of occurrences of major adverse cardiac events during 1-year follow-up after hospital discharge in patients with AMI using knowledge mining, Pers. Ubiquit. Comput., 26 (2022), 259–267. https://doi.org/10.1007/s00779-019-01248-7 doi: 10.1007/s00779-019-01248-7
    [11] S. W. A. Sherazi, J. Bae, J. Y. Lee, A soft voting ensemble classifier for early prediction and diagnosis of occurrences of major adverse cardiovascular event for STEMI and NSTEMI during 1-year follow-up in patients with acute coronary syndrome, PLoS One, 16 (2021), e0249338. https://doi.org/10.1371/journal.pone.0249338 doi: 10.1371/journal.pone.0249338
    [12] H. L. T. Lam, N. H. Le, L. van Tuan, H. T. Ban, T. N. K. Hung, N. T. K. Nguyen, et al., Machine learning model for identifying antioxidant proteins using features calculated from primary sequences, Biology (Basel), 9 (2020), 325. https://doi.org/10.3390/biology9100325 doi: 10.3390/biology9100325
    [13] T. H. Vo, N. T. K. Nguyen, Q. H. Kha, N. Q. K. Le, On the road to explainable AI in drug-drug interactions prediction: A systematic review, Comput. Struct. Biotechnol. J., 9 (2020), 325. https://doi.org/10.1016/j.csbj.2022.04.021 doi: 10.1016/j.csbj.2022.04.021
    [14] T. N. K. Hung, N. Q. K. Le, N. H. Le, L. van Tuan, T. P. Nguyen, C. Thi, et al., An AI-based prediction model for drug-drug interactions in osteoporosis and paget's diseases from SMILES, Mol. Inform., 9 (2020), 325. https://doi.org/10.1002/minf.202100264 doi: 10.1002/minf.202100264
    [15] N. Hafeez, X. Du, N. Boulgouris, P. Begg, R. Irving, C. Coulson, et al., Electrical impedance guides electrode array in cochlear implantation using machine learning and robotic feeder, Hear. Res., 412 (2021), 108371. https://doi.org/10.1016/j.heares.2021.108371 doi: 10.1016/j.heares.2021.108371
    [16] W. Duan, S. S. C. Congress, G. Cai, S. Liu, X. Dong, R. Chen, et al., A hybrid GMDH neural network and logistic regression framework for state parameter–based liquefaction evaluation, Can. Geotech. J., 58 (2021), 1801–1811. https://doi.org/10.1139/cgj-2020-0686 doi: 10.1139/cgj-2020-0686
    [17] J. Skidmore, L. Xu, X. Chao, W. J. Riggs, A. Pellittieri, C. Vaughan, et al., Prediction of the functional status of the cochlear nerve in individual cochlear implant users using machine learning and electrophysiological measures, Ear Hear., 42 (2021), 180-192. https://doi.org/10.1097/AUD.0000000000000916 doi: 10.1097/AUD.0000000000000916
    [18] W. Duan, Z. Zhao, G. Cai, S. Pu, S. Liu, X. Dong, Evaluating model uncertainty of an in situ state parameter-based simplified method for reliability analysis of liquefaction potential, Comput. Geotech., 151 (2022), 104957. https://doi.org/10.1016/j.compgeo.2022.104957 doi: 10.1016/j.compgeo.2022.104957
    [19] Y. Bozhkov, J. Shawarba, J. Feulner, F. Winter, S. Rampp, U. Hoppe, er al., Prediction of hearing preservation in vestibular schwannoma surgery according to tumor size and anatomic extension, Otolaryngol. Head Neck Surg., 166 (2021), 530–536. https://doi.org/10.1177/01945998211012674 doi: 10.1177/01945998211012674
    [20] J. H. Han, D. G. Kim, H. T. Chung, S. H. Paek, C. K. Park, Hearing preservation in patients with unilateral vestibular schwannoma who undergo stereotactic radiosurgery: Reinterpretation of the auditory brainstem response, Cancer, 118 (2012), 5441–5447. https://doi.org/10.1002/cncr.27501 doi: 10.1002/cncr.27501
    [21] A. Elliott, A. L. Hebb, S. Walling, D. P. Morris, M. Bance, Hearing preservation in vestibular schwannoma management, Am. J. Otolaryngol., 36 (2015), 526–534. https://doi.org/10.1016/j.amjoto.2015.02.016 doi: 10.1016/j.amjoto.2015.02.016
    [22] Y. Ren, K. O. Tawfik, B. J. Mastrodimos, R. A. Cueva, Preoperative radiographic predictors of hearing preservation after retrosigmoid resection of vestibular schwannomas, Otolaryngol. Head Neck Surg., 165 (2021), 344–353. https://doi.org/10.1177/0194599820978246 doi: 10.1177/0194599820978246
    [23] D. E. Roos, A. E. Potter, A. C. Zacest, Hearing preservation after low dose linac radiosurgery for acoustic neuroma depends on initial hearing and time, Radiother. Oncol., 101 (2011), 420–424. https://doi.org/10.1016/j.radonc.2011.06.035 doi: 10.1016/j.radonc.2011.06.035
    [24] D. Cha, S. H. Shin, S. H. Kim, J. Y. Choi, I. S. Moon, Machine learning approach for prediction of hearing preservation in vestibular schwannoma surgery, Sci. Rep., 10 (2020), 7136. https://doi.org/10.1038/s41598-020-64175-1 doi: 10.1038/s41598-020-64175-1
    [25] T. Q. Chen, C. Guestrin, XGBoost: A scalable tree boosting system, in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, (2016), 785–794.
    [26] H. Wimalarathna, S. Ankmnal-Veeranna, C. Allan, S. K. Agrawal, P. Allen, J. Samarabandu, et al., Comparison of machine learning models to classify Auditory Brainstem Responses recorded from children with auditory processing disorder, Comput. Methods Programs Biomed., 200 (2021), 105942. https://doi.org/10.1016/j.cmpb.2021.105942 doi: 10.1016/j.cmpb.2021.105942
    [27] M. Kivrak, E. Guldogan, C. Colak, Prediction of death status on the course of treatment in SARS-COV-2 patients with deep learning and machine learning methods, Comput. Methods Programs Biomed., 201 (2021), 105951. https://doi.org/10.1016/j.cmpb.2021.105951 doi: 10.1016/j.cmpb.2021.105951
    [28] D. Shorthouse, A. Riedel, E. Kerr, L. Pedro, D. Bihary, S. Samarajiwa, et al., Exploring the role of stromal osmoregulation in cancer and disease using executable modelling, Nat. Commun., 9 (2018), 3011. https://doi.org/10.1038/s41467-018-05414-y doi: 10.1038/s41467-018-05414-y
    [29] T. Y. Chen, X. Li, Y. X. Li, E. Xia, Y. Qin, S. Liang, et al., Prediction and risk stratification of Kidney outcomes in IgA nephropathy, Am. J. Kidney. Dis., 74 (2019), 300–309. https://doi.org/10.1053/j.ajkd.2019.02.016 doi: 10.1053/j.ajkd.2019.02.016
    [30] J. Fan, W. Yue, L. Wu, F. Zhang, H. Cai, X. Wang, et al., Evaluation of SVM, ELM and four tree-based ensemble models for predicting daily reference evapotranspiration using limited meteorological data in different climates of China, Agr. Forest Meteorol., 263 (2018), 225–241. https://doi.org/10.1016/j.agrformet.2018.08.019 doi: 10.1016/j.agrformet.2018.08.019
    [31] S. Wang, S. Liu, J Zhang, X. Che, Y. Yuan, Z. Wang, A new method of diesel fuel brands identification: SMOTE oversampling combined with XGBoost ensemble learning, Fuel, 282 (2020), 118848. https://doi.org/10.1016/j.fuel.2020.118848 doi: 10.1016/j.fuel.2020.118848
    [32] J. Ma, J. Cheng, Z. Xu, K. Chen, C. Lin, F. Jiang, Identification of the most influential areas for air pollution control using XGBoost and grid importance rank, J. Neurosurg., 40 (1997), 248–262. https://doi.org/10.1016/j.jclepro.2020.122835 doi: 10.1016/j.jclepro.2020.122835
    [33] D. Chakraborty, H. Elzarka, Early detection of faults in HVAC systems using an XGBoost model with a dynamic threshold, Energy Build., 185 (2019), 326–344. https://doi.org/10.1016/j.enbuild.2018.12.032 doi: 10.1016/j.enbuild.2018.12.032
    [34] T. Zhu, Y. Lin, Y. Liu, Synthetic minority oversampling technique for multiclass imbalance problems, Pattern Recognit., 72 (2017), 327–340. https://doi.org/10.1016/j.agrformet.2018.08.019 doi: 10.1016/j.agrformet.2018.08.019
    [35] N. V. Chawla, K. W. Bowyer, L. O. Hall, W. Kegelmeyer, SMOTE: Synthetic minority oversampling technique, J. Artif. Intell. Res., 16 (2002), 321–357.
    [36] R. A. Olshen, L. Breiman, J. Friedman, C. Stone, Classification and regression trees, Wadsworth, 40 (1984), 358.
    [37] A. C. Olivieri, Analytical figures of merit: From univariate to multiway calibration, Chem. Rev., 114 (2014), 5358–5378. https://doi.org/10.1007/978-3-319-97097-4_10 doi: 10.1007/978-3-319-97097-4_10
    [38] G. Mohr, B. Sade, J. J. Dufour, J. M. Rappaport, Preservation of hearing in patients undergoing microsurgery for vestibular schwannoma: Degree of meatal filling, J. Neurosurg., 102 (2005), 11–15. https://doi.org/10.3171/jns.2005.102.1.0001 doi: 10.3171/jns.2005.102.1.0001
    [39] H. A. Arts, S. A. Telian, H. E. Kashlan, B. G. Thompson, Hearing preservation and facial nerve outcomes in vestibular schwannoma surgery: Results using the middle cranial fossa approach, Otol. Neurotol., 27 (2006), 234–241. https://doi.org/10.1097/01.mao.0000185153.54457.16 doi: 10.1097/01.mao.0000185153.54457.16
    [40] J. W. Jr Kutz, T. Scoresby, B. Isaacson, B. E. Mickey, C. J. Madden, S. L. Barnett, et al., Hearing preservation using the middle fossa approach for the treatment of vestibular schwannoma, Neurosurgery, 70 (2012), 334–340. https://doi.org/10.1227/NEU.0b013e31823110f1 doi: 10.1227/NEU.0b013e31823110f1
  • This article has been cited by:

    1. Lukasz Przepiorka, Sławomir Kujawski, Katarzyna Wójtowicz, Edyta Maj, Andrzej Marchel, Przemysław Kunert, Development and application of explainable artificial intelligence using machine learning classification for long-term facial nerve function after vestibular schwannoma surgery, 2024, 0167-594X, 10.1007/s11060-024-04844-7
    2. José Alberto Benítez-Andrades, Camino Prada-García, Nicolás Ordás-Reyes, Marta Esteban Blanco, Alicia Merayo, Antonio Serrano-García, Enhanced prediction of spine surgery outcomes using advanced machine learning techniques and oversampling methods, 2025, 13, 2047-2501, 10.1007/s13755-025-00343-9
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2142) PDF downloads(122) Cited by(2)

Figures and Tables

Figures(7)  /  Tables(3)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog