Loading [MathJax]/jax/output/SVG/jax.js
Research article

Pandemic within a pandemic! Policy Implications of community-based Interventions to mitigate violence against women during COVID-19 in Urban Slums of Lucknow, India

  • Background 

    The COVID-19 pandemic has brought an unprecedented adverse impact on women's health. Evidence from the literature suggests that violence against women has increased multifold. Gender-based violence in urban slums has worsened due to a lack of water and sanitation services, overcrowding, deteriorating conditions and a lack of institutional frameworks to address gender inequities.

    Methods 

    The SAMBHAV (Synchronized Action for Marginalized to Improve Behaviors and Vulnerabilities) initiative was launched between June 2020 to December 2020 by collaborating with the Uttar Pradesh state government, UNICEF and UNDP. The program intended to reach 6000 families in 30 UPS (Urban Poor settlements) of 13 city wards. These 30 UPS were divided into 5 clusters. The survey was conducted in 760 households, 397 taken from randomly selected 15 interventions and 363 households from 15 control UPS. This paper utilized data from a baseline assessment of gender and decision-making from a household survey conducted in the selected UPS during July 03–15, 2020. A sample size of 360 completed interviews was calculated for intervention and control areas to measure changes attributable to the SAMBHAV intervention in the behaviours and service utilization (pre- and post-intervention).

    Results 

    The data analysis showed a significant difference (p-value < 0.001) between respondents regarding women's freedom to move alone in the control and intervention area. It also reflected a significant difference between control and intervention areas as the respondents in the intervention area chose to work for the cause of gender-based violence.

    Conclusion 

    The SAMBHAV initiative brought an intersectional lens to gender issues. The community volunteers were trained to approach issues based on gender-based violence with the local public, and various conferences and meetings were organized to sensitize the community. The initiative's overall impact was that it built momentum around the issue of applying the concept of intersectionality for gender issues and building resilience in the community. There is still a need to bring multi-layered and more aggressive approaches to reduce the prevalence of gender-based violence in the community.

    Citation: Fnu Kajal, Ram Manohar Mishra, Amit Mehrotra, Vijay Kumar Chattu. Pandemic within a pandemic! Policy Implications of community-based Interventions to mitigate violence against women during COVID-19 in Urban Slums of Lucknow, India[J]. AIMS Public Health, 2023, 10(2): 297-309. doi: 10.3934/publichealth.2023022

    Related Papers:

    [1] Yan Xie, Rui Zhu, Xiaolong Tan, Yuan Chai . Inhibition of absence seizures in a reduced corticothalamic circuit via closed-loop control. Electronic Research Archive, 2023, 31(5): 2651-2666. doi: 10.3934/era.2023134
    [2] Ariel Leslie, Jianzhong Su . Modeling and simulation of a network of neurons regarding Glucose Transporter Deficiency induced epileptic seizures. Electronic Research Archive, 2022, 30(5): 1813-1835. doi: 10.3934/era.2022092
    [3] Zhihui Wang, Yanying Yang, Lixia Duan . Dynamic mechanism of epileptic seizures induced by excitatory pyramidal neuronal population. Electronic Research Archive, 2023, 31(8): 4427-4442. doi: 10.3934/era.2023226
    [4] Xiaolong Tan, Hudong Zhang, Yan Xie, Yuan Chai . Electromagnetic radiation and electrical stimulation controls of absence seizures in a coupled reduced corticothalamic model. Electronic Research Archive, 2023, 31(1): 58-74. doi: 10.3934/era.2023004
    [5] Saeedreza Tofighi, Farshad Merrikh-Bayat, Farhad Bayat . Designing and tuning MIMO feedforward controllers using iterated LMI restriction. Electronic Research Archive, 2022, 30(7): 2465-2486. doi: 10.3934/era.2022126
    [6] Yang Song, Beiyan Yang, Jimin Wang . Stability analysis and security control of nonlinear singular semi-Markov jump systems. Electronic Research Archive, 2025, 33(1): 1-25. doi: 10.3934/era.2025001
    [7] Jichen Hu, Ming Zhu, Tian Chen . The nonlinear observer-based fault diagnosis method for the high altitude airship. Electronic Research Archive, 2025, 33(2): 907-930. doi: 10.3934/era.2025041
    [8] Ramalingam Sakthivel, Palanisamy Selvaraj, Oh-Min Kwon, Seong-Gon Choi, Rathinasamy Sakthivel . Robust memory control design for semi-Markovian jump systems with cyber attacks. Electronic Research Archive, 2023, 31(12): 7496-7510. doi: 10.3934/era.2023378
    [9] Yu-Jing Shi, Yan Ma . Finite/fixed-time synchronization for complex networks via quantized adaptive control. Electronic Research Archive, 2021, 29(2): 2047-2061. doi: 10.3934/era.2020104
    [10] Huan Luo . Heterogeneous anti-synchronization of stochastic complex dynamical networks involving uncertain dynamics: an approach of the space-time discretizations. Electronic Research Archive, 2025, 33(2): 613-641. doi: 10.3934/era.2025029
  • Background 

    The COVID-19 pandemic has brought an unprecedented adverse impact on women's health. Evidence from the literature suggests that violence against women has increased multifold. Gender-based violence in urban slums has worsened due to a lack of water and sanitation services, overcrowding, deteriorating conditions and a lack of institutional frameworks to address gender inequities.

    Methods 

    The SAMBHAV (Synchronized Action for Marginalized to Improve Behaviors and Vulnerabilities) initiative was launched between June 2020 to December 2020 by collaborating with the Uttar Pradesh state government, UNICEF and UNDP. The program intended to reach 6000 families in 30 UPS (Urban Poor settlements) of 13 city wards. These 30 UPS were divided into 5 clusters. The survey was conducted in 760 households, 397 taken from randomly selected 15 interventions and 363 households from 15 control UPS. This paper utilized data from a baseline assessment of gender and decision-making from a household survey conducted in the selected UPS during July 03–15, 2020. A sample size of 360 completed interviews was calculated for intervention and control areas to measure changes attributable to the SAMBHAV intervention in the behaviours and service utilization (pre- and post-intervention).

    Results 

    The data analysis showed a significant difference (p-value < 0.001) between respondents regarding women's freedom to move alone in the control and intervention area. It also reflected a significant difference between control and intervention areas as the respondents in the intervention area chose to work for the cause of gender-based violence.

    Conclusion 

    The SAMBHAV initiative brought an intersectional lens to gender issues. The community volunteers were trained to approach issues based on gender-based violence with the local public, and various conferences and meetings were organized to sensitize the community. The initiative's overall impact was that it built momentum around the issue of applying the concept of intersectionality for gender issues and building resilience in the community. There is still a need to bring multi-layered and more aggressive approaches to reduce the prevalence of gender-based violence in the community.



    Cardiovascular disease has always been regarded as the most serious and fatal disease for humans in the world. The increase in morbidity and mortality of cardiovascular disease has brought enormous risks and burdens to healthcare systems around the world [1]. Despite the efforts of medical staff to prevent, diagnose, and treat different types of cardiovascular diseases, the number of deaths from cardiovascular disease in the world continues to increase every year, and by 2019 the number has increased to 18.6 million [2]. World Health Organization estimates that the deaths from cardiovascular disease in 2020 accounts for approximately 32% of the total deaths worldwide [3]. According to NHANES report, the prevalence of cardiovascular disease among adults over 20 years old from 2013 to 2016 was 48%, and the prevalence increased with age [4].

    CAD is one of the common clinical cardiovascular diseases. Clinically, Coronary Angiography (CAG) is mainly used to determine the location and extent of arterial stenosis. The CAG technique is to obtain coronary artery images by X-ray after direct injection of contrast medium into the femoral artery. The inspection is time-consuming, expensive, traumatic, and it has high technical threshold and equipment requirements [5]. Coronary Computed Tomography Angiography (CTA) is an emerging examination technique. The CTA technology obtains accurate and clear images of cardiac coronary artery by intravenous injection of contrast medium and computer reconstruction spiral CT scanning, which is non-invasive. However, the patient's respiration, heart rate, cardiac function, and other factors could affect the imaging quality, which causes the inspection effect of this method to be inferior to CAG [6]. To avoid the harm of CAG technology to patients and the limitations of CTA technology on patients' factors, researchers have widely applied machine learning and data mining techniques to diagnose CAD [7].

    This research will build models for diagnosing LAD, LCX, and RCA based on swarm intelligence optimization algorithm and machine learning techniques. All the research in this paper is done on the extension of Z-Alizadeh Sani dataset, which is derived from the Mendeley Data [8]. We could know the severity of the patient's condition by diagnosing the stenosis of individual artery, which can assist physicians to take corresponding treatments according to different degrees of disease. Feature selection optimization based on meta-heuristic optimizer is a new feature selection algorithm proposed in recent years, which can effectively solve global optimization problems and avoid falling into local optimal solutions [9]. For feature selection, we first apply the filtering method to delete the features with the variance of 0. Then we use the KNN-based WOA to select the feature subsets and use the classification accuracy of KNN and the number of features to guarantee the quality of the selected feature subsets [10]. In this study, a two-layer stacking model is established to blend the results of individual and ensemble classifiers. Four classifiers with best performance are selected from KNN, Support Vector Machine (SVM), Decision Tree (DT), Random Forest (RF), Gradient Boosting Decision Tree (GBDT), Extreme Gradient Boosting (XGBoost), and Adaptive Boosting (AdaBoost) algorithms as primary learners, and Logistic Regression (LR) is applied as the secondary learner to reduce the complexity of the model [11]. We adopt accuracy, recall, precision, F1-score, and AUC value as the evaluation metrics of the model.

    The rest of the paper consists of the following contents: Section 2 provides an overview of relevant studies in previous literature; Section 3 introduces the feature selection methods and machine learning algorithms. Section 4 discusses the results of feature selection and the classification performance of the proposed method. Section 5 summarizes this paper and looks forward to future study.

    In this section, we briefly discuss some relevant studies that use machine learning algorithms, data mining techniques, and improved methods to diagnose and predict diseases.

    The methods and results of feature selection determine the classification performance of the model to some extent. In [12,13,17,18,19,20], Roohallah Alizadehsani and Zeinab Arabasadi et al. adopted the weight by SVM method to complete feature selection. This method uses the normal vector coefficients of linear support vector machine as feature weights. In [14,15], Roohallah Alizadehsani et al. used information gain to measure the importance of features, and selected features with information gain higher than a certain value as feature subset. In [16], Roohallah Alizadehsani et al. measured feature importance by calculating the Gini index of each feature. In [21], Roohallah Alizadehsani et al. proposed the assurance feature selection method. This method measured the importance of a feature by calculating the ratio of the number of patients associated with a feature to the total number of patients. In [22], Roohallah Alizadehsani et al. applied Gini index and principal component analysis (PCA) to calculate the weights of features, and determined the threshold of the weights for feature selection through experiments. In the above studies, researchers implemented feature selection by calculating and evaluating the importance of single features. These methods are computationally fast and easy to implement, focusing on the ability to select features that have a great impact on disease classification.

    Metaheuristic optimization algorithm (MOA) mainly simulates natural and human intelligence to solve the optimal solution [23]. MOA can be divided into four main categories: evolutionary, swarm intelligence, human, and physical and chemical based algorithms. Among them, Genetic Algorithm (GA) [24], Particle Swarm Optimization (PSO) [25], Sine Cosine Algorithm (SCA) [26], Moth-flame Optimization (MFO) [27], WOA [28], Grey Wolf Optimizer (GWO) [29] and other algorithms are widely used for feature selection.

    Moloud Abdar et al. [30] used three types SVMs to establish models for CAD diagnosis, and compared the performance of GA and PSO for feature selection and model parameter optimization in parallel. This method could simultaneously select the optimal feature subset and parameter combination of the model. Bayu Adhi Tama et al. [5] combined the Correlation-based Feature Selection (CFS) and Credal Decision Tree (CDT)-based BPSO to identify important features. The CFS could identify the unimportant and unnecessary features for classification. Shafaq Abbas et al. [31] applied Extremely Randomized Tree (ERT)-based WOA to conduct feature selection and classification on breast cancer dataset. Hoda Zamani et al. [32] proposed a FSWOA algorithm for feature selection, which achieved effective dimensionality reduction in medical datasets. In the above studies, researchers used different machine learning algorithms as wrappers to combine MOA to achieve feature selection for disease diagnosis problems. By preliminarily filtering redundant features in the dataset, the initial population of the MOA can be optimized and the efficiency of feature selection can be improved.

    Since most metaheuristic algorithms are proposed to solve continuous problems, researchers have used transfer functions to convert each dimension of the solution vectors into binary form for feature selection, especially for medical data [33]. E. Emary et al. [34] proposed two binary-conversion methods for the GWO algorithm. The second method is to use the S-shaped transfer function to convert the updated gray wolf position vector into binary form. Shokooh Taghian et al. [35] proposed SBSCA and VBSCA method for feature selection, researchers used S-shaped and V-shaped transfer function to achieve binary-conversion of SCA. Mohammad H. Nadimi-Shahraki et al. [23] improved the MFO algorithm using S-shaped, V-shaped and U-shaped transfer functions for feature selection on medical datasets.

    In previous studies, researchers have applied a variety of basic classification algorithms and ensemble classification algorithms to establish CAD diagnostic models. In [12,13], Roohallah Alizadehsani et al. introduced the cost sensitive algorithm into the model construction, and combined the 10-fold cross-validation (cv) method to evaluate the classification performance of the Sequence Minimum Optimization (SMO) algorithm for CAD. In [15], Roohallah Alizadehsani et al. used ensemble learning method to combine classification results of SMO algorithm and Naïve Bayes (NB) to diagnose CAD. In [16], the researchers used bagging algorithm to obtain high accuracy in the diagnose of LAD. In [18,20,21], Roohallah Alizadehsani et al. used SVM to establish diagnostic models for CAD and main coronary artery stenosis. Zeinab Arabasadi et al. [19] applied Neural Network to establish CAD diagnostic model, and adjusted the weight of Neural Network through GA to obtain ideal classification performance. In [22], the improved SVM algorithm was used to diagnose LAD, LCX and RCA stenosis. The researchers combined the distance between the sample and the separating hyperplane with the accuracy of the classifier to improve the model's performance. Md Mamun Ali et al. [36] applied KNN, DT, and RF to establish disease diagnosis models. Bayu Adhi Tama et al. [5] built a two-layer stacking model to diagnose CAD, RF, GBDT, and XGBoost were used to obtain classification results in the first layer, and the Generalized Linear Model (GLM) in the second layer to generate the final predictions.

    According to the above summary, we can find that there are few studies on the diagnosis of each main coronary artery stenosis. In this study, we will apply machine learning and data mining algorithms to diagnose stenosis of each main coronary artery. We will divide the training set and test set at a ratio of 9: 1, combine randomized search and 10-fold cv to train the model on the training set, and then use the trained model to make predictions on the test set. We apply filtering and KNN-based WOA to select the optimal feature subset for each main coronary artery and then build a two-layer stacking model based on the selected feature subset. At last, we compare the performance achieved by the proposed method with the classification performance obtained in the existing literature.

    The extension of Z-Alizadeh Sani dataset contains the clinical record information of 303 patients, and each patient has 55 features [8]. The patients with LAD, LCX, and RCA stenosis were 160,109, and 101. A total of 216 people were diagnosed with CAD. The features can be divided into four fields: demographic features, symptoms and physical examination, electrocardiography (ECG), laboratory and echocardiography [13]. Table 1 shows the information of features in the extension of Z-Alizadeh Sani dataset.

    Table 1.  Features and types of the extension of Z-Alizadeh Sani dataset.
    Feature type Feature name Count
    Continuous Age, Weight, Height, Body Mass Index (BMI), Blood Pressure (BP), Pulse Rate (PR), Fasting Blood Sugar (FBS), Creatine (Cr), Triglyceride (TG), Low Density Lipoprotein (LDL), High Density Lipoprotein (HDL), Blood Urea Nitrogen (BUN), Erythrocyte Sedimentation Rate (ESR), Hemoglobin (HB), Potassium (K), Sodium (Na), White Blood Cell (WBC), Lymphocyte (Lymph), Neutrophil (Neut), Platelet (PLT), Ejection Fraction (EF-TTE) 21
    Binary Sex, Diabetes Mellitus (DM), Hyper Tension (HTN), Current Smoker, EX-Smoker, Family History (FH), Obesity, Chronic Renal Failure (CRF), Cerebrovascular Accident (CVA), Airway disease, Thyroid Disease, Congestive Heart Failure (CHF), Dyslipidemia (DLP), Edema, Weak Peripheral Pulse, Lung Rales, Systolic Murmur, Diastolic Murmur, Typical Chest Pain (Typical CP), Dyspnea, Atypical, Nonanginal Chest Pain (Nonanginal CP), Exertional Chest Pain (Exertional CP), Low Threshold Angina (LowTH Ang), Q Wave, St Elevation, St Depression, T inversion, Left Ventricular Hypertrophy (LVH), Poor R Progression 30
    Categorical Function Class, Bundle Branch Block (BBB), Region with Regional wall motion abnormality (Region RWMA), Valvular Heart Disease (VHD) 4

     | Show Table
    DownLoad: CSV

    The data mining process starts from the preprocessing stage, followed by feature engineering, and finally uses machine learning algorithms to establish models. The algorithm flow of this research is shown in Figure 1.

    Figure 1.  Algorithm flow chart.

    In one-hot encoding, each category of features is represented by a vector whose length is equal to the number of categories. The i-th vector only takes the value of 1 at the i-th component, and the rest are all 0. One-hot encoding of features can extend the value of discrete features to Euclidean space, making the distance calculation between features more reasonable.

    Standardization is to scale the values of a column of numerical features to a state of mean 0 and variance 1. The standardized formula is shown in Eq (1).

    S(x)=xxσ (1)

    Among them, x is the instance in an n-dimensional space, n is the number of features, and x and σ represent the mean and standard deviation of each feature [37].

    In this research, we use accuracy, precision, recall, F1-score, and AUC value to evaluate the classification performance of proposed model. The calculation formulas are shown in Eqs (2)–(5) [38].

    Accuracy=TP+TNTP+FN+FP+TN (2)
    Precision=TPTP+FP (3)
    Recall=TPTP+FN (4)
    F1score=2×Precision×RecallPrecision+Recall (5)

    AUC (Area under curve) is defined as the area under the ROC curve, which is usually greater than 0.5 and less than 1. The larger the AUC value, the better the classification performance of the classifier.

    WOA is a meta-heuristic global optimization algorithm related to swarm intelligence proposed by Mirjalili and Lewis [28]. The algorithm is inspired by the bubble net foraging behavior of humpback whales and finds the optimal solution by simulating this unique behavior. They hunt for food by continuously shrinking the enclosure, spirally updating their positions, and hunting randomly [39].

    The algorithm mainly includes two stages: first, achieve the encirclement of the prey, and update the spiral position (also known as hunting behavior); second, search for the prey randomly [39]. Next, we will introduce each stage in detail:

    1) Surround the prey. Humpback whales can identify the location of their prey and circle around them. In the initial stage of the algorithm, since we don't know the location of the optimal solution in the search space, the WOA will assume that the best candidate solution currently obtained is the target solution or is close to the optimal solution. After defining the best candidate solution, the whales will attempt to move from other candidate positions to the best position and update their positions. This process is represented by Eq (6):

    P(t+1)=P(t)A|CP(t)P(t)| (6)

    Among them, t is the number of iterations in the current state, P(t) is the position vector of the optimal solution currently, P(t) is the current position vector, and A and C are the constant vectors. The vectors A and C are calculated by Eqs (7) and (8):

    A=2ara (7)
    C=2r (8)

    In the above equations, a is decreased linearly from 2 to 0 in the iterative process, and r is a random vector in the range of [0, 1].

    2) Hunting behavior. Humpback whales hunt by swimming towards their prey in a spiral motion. The mathematical model of hunting behavior is as follows:

    P(t+1)=|P(t)P(t)|eblcos(2πl)+P(t) (9)

    Among them, b is a constant used to define the shape of the logarithmic spiral, and l is a random number in the range of [-1, 1]. During the hunting process, each humpback whale randomly chooses to shrink to surround the prey or spiral upward to chase the prey. The probability of each whale choosing these two behaviors is 50%. The researchers simulated this behavior through the following mathematical model:

    P(t+1)={P(t)A|CP(t)P(t)|,if p<0.5|P(t)P(t)|eblcos(2πl)+P(t),if p>0.5 (10)

    3) Search for prey. The algorithm searches for prey according to the change of the value of A. When |A|>1, the algorithm randomly selects a search individual, and then updates the position of other individuals according to the location of the randomly selected individual, forcing the whale to deviate from the prey, thereby finding a more suitable prey and make the WOA realize global search. When |A|<1, the whales attack the preys. The mathematical model is shown in Eq (11):

    P(t+1)=PrandA|CPrandP(t)| (11)

    Among them, Prand is the position vector of the whale randomly selected from the population.

    In this study, we apply the KNN-based WOA for feature selection. WOA is used to adaptively search for the optimal feature subset to maximize classification accuracy, and KNN is used to ensure the quality of the selected feature subset. In WOA, the whale takes any point in space as a starting point, and continuously adjusts its position to the best candidate solution. Each solution obtained by this algorithm is a continuous vector of the same dimension, bounded in [0, 1] [39]. The function of feature selection can be achieved by setting a threshold to perform binary-conversion on the solution vector. In this study, we set the threshold to 0.5, and the value in the solution is 1 when it is greater than 0.5, and 0 when it is less than 0.5. The length of each solution is M and consists of 0 and 1, where M is the total number of features, 1 means that the feature at the corresponding position is selected, and 0 means that the feature is abandoned. Multiple solution vectors are obtained by changing the initial population size and the number of iterations. The quality of the solution is evaluated by the fitness function. Eq (12) shows the fitness function used in this paper.

    f=αE+(1α)mM (12)

    where f is the fitness of a given solution vector of size M, m is the number of selected features, E is the classification error rate of the classifier, and α is a constant that balances the error rate of the classifier with the number of selected features [10]. The smaller the fitness value, the better the performance of the feature, and the closer to the optimal solution. In this research, E is the classification error rate of KNN, and α is set to 0.99. The pseudocode of WOA feature selection algorithm is shown in Algorithm 1.

    Table Algorithm1.  Pseudocode of WOA feature selection algorithm.
    Input: Number of iterations (T), size of initial population (N).
    Output: The global optimal position vector P*.
    1 Initialize the population Pi (i = 1, 2, 3, …, N).
    2 Binary-conversion.
    3 Compute the fitness of each solution.
    4 Set P* as the best solution.
    5 while (t < T) do
    6  for (each solution) do
    7    Update a, A, C, l and p.
    8    if (p < 0.5) then
    9      if (|A| < 1) then
    10        The whale position is updating by the Eq (6).
    11      else
    12        if (|A| ≥ 1) then
    13          Select the random whale Prand.
    14          The whale position is updating by the Eq (11).
    15        end
    16      end
    17    else
    18       if (p ≥ 0.5) then
    19        Modify the whale position by the Eq (9).
    20      end
    21    end
    22   end
    23  Check if any search agent goes beyond the search space and amend it.
    24  Binary-conversion.
    25  Compute the fitness of each search agent.
    26  Update P* if there is a better solution.
    27  t = t + 1
    28  end
    29 return P*

     | Show Table
    DownLoad: CSV

    In this study, we construct a two-layer stacking model for diagnosing each main coronary artery stenosis. The primary learners in this model are selected from KNN, SVM, DT, RF, GBDT, XGBoost, and AdaBoost, and LR is used as a secondary learner to blend the classification results of multiple primary learners to obtain the final prediction results.

    LR is a generalized linear model used to solve binary classification problems. The output value of the linear model is processed by the sigmoid function and positioned between (0, 1) for the task of binary classification [36]. KNN is a supervised algorithm, and the principle of KNN is that when predicting a new sample, it can judge which category the sample belongs to according to the category of the k points closest to it [40]. SVM is a two-class model with superior performance and flexibility, which can minimize both empirical and structural risks. For a sample set in a finite-dimensional space, SVM performs classification by mapping the sample set from the original feature space to a high-dimensional space [41]. DT is a non-parametric supervised learning method, and the generation of DT is to continuously select the optimal features to divide the training set. ID3, C4.5, and CART are the three main DT algorithms [42]. RF is a special bagging method [43]. For each training set, a decision tree is constructed. When nodes are selected for feature splitting, some features are randomly selected from all the features, and the optimal solution is found from the selected features and applied to node splitting. GBDT uses the addition model and forward stepwise algorithm for greedy learning, and learns a CART tree in each iteration to fit the residuals between the predicted results of the previous (t-1) trees and the real values of the training samples [44]. XGBoost is an optimized distributed gradient boosting library [45]. XGBoost supports multiple types of base classifiers. When using CART as a base classifier, XGBoost improves its generalization ability by adding a regular term to control the complexity of the model. AdaBoost is an iterative algorithm implemented by changing the distribution of dataset. The weight of the sample incorrectly classified by the previous classifier in the training set will increase, and the weight of the sample correctly classified will decrease. Then the new dataset with modified weight is passed to the next classifier for training. At last, the algorithm combines the classifiers obtained each time as the final decision classifier [46].

    Stacking is an ensemble learning algorithm that learns how to best combine the prediction results from multiple well-performing machine learning models. In the stacking model, we call the base learners the primary learners and the learner used for blending is called the secondary learner or meta-learner. Specifically, the original dataset is divided into several subsets, which are input into each primary learner of the first layer and predicted by k-fold cv. Then, the output of each primary learner in the first layer is taken as the input value of the secondary learner in the second layer. The final prediction result is obtained by fitting the trained model to the test set [47]. The algorithm flow of the stacking model is shown in Figure 2.

    Figure 2.  Flow chart of stacking model.

    In this section, we systematically describe the implementation of the proposed method on the extension of Z-Alizadeh Sani dataset and report the results. All experiments in this research are carried out on a Windows machine with 8GB memory and Intel (R) Core (TM) i5-7200U CPU @ 2.50GHZ. Python 3.8 is used in the Jupyter Notebook IDE to implement the entire experiment.

    First, for the processing of categorical features, one-hot encoding is performed on the feature Bundle Branch Block (BBB) to obtain three binary features BBB_LBBB, BBB_RBBB and BBB_N. For feature Valvular Heart Disease (VHD), its values "Normal", "Mild", "Moderate", and "Severe" are denoted as 0, 1, 2 and 3 respectively. Second, for the features Function Class and Region with Regional wall motion abnormality (Region RWMA), they are processed according to the discretization range provided in the Braunwald heart book, when the value of the feature is zero, it is recorded as "Normal", and non-zero is recorded as "High" [48]. Then, all the categorical features with two values are transformed into numerical values. Next, the dataset is divided in a ratio of 9: 1 to obtain training data and test data for LAD, LCX and RCA diagnosis respectively. The training data is used to develop the model, while the test data is used to evaluate the classification performance of the model. At last, the training set is standardized, and the mean and variance of the training set are used for standardizing the test set. Mark "Stenotic" as 1, and "Normal" as 0 in the three labels of LAD, LCX and RCA.

    In this study, we first use the filtering method to delete the feature Exertional CP with the variance of 0 and then use KNN-based WOA to select the feature subsets for diagnosing each main coronary artery. The parameters of feature selection algorithm are shown in Table 2.

    Table 2.  Parameters of WOA feature selection algorithm.
    Parameter Value
    Number of iterations (T) 200,300,400
    Size of initial population (N) 20, 30
    Search domain [0, 1]
    K value 5

     | Show Table
    DownLoad: CSV

    We first compare the results of feature selection by KNN-based WOA, BPSO, GA and BGWO, and the number of iterations and the size of initial population of each algorithm are the same as WOA, as shown in Table 2. Each method is run for 60 times. Table 3 shows the average fitness of the feature subsets obtained by these four algorithms, the average classification accuracy, and the average AUC value on validation set. It can be seen from Table 3 that the KNN-based WOA method has the best performance, and Friedman test results show that there are significant differences in the performance of different feature selection methods. The bold font indicates the best result in the following tables.

    Table 3.  Comparison of KNN-based WOA, BPSO, GA, and BGWO feature selection methods.
    Artery Metrics WOA BPSO GA BGWO
    LAD Fitness 0.0685 0.0999 0.1306 0.0783
    Accuracy 0.9801 0.9024 0.8714 0.9238
    AUC 0.9627 0.8581 0.8251 0.9080
    Friedman rank 1 3 4 2
    LCX Fitness 0.0764 0.0771 0.1341 0.0737
    Accuracy 0.9464 0.9262 0.8679 0.9262
    AUC 0.9382 0.9039 0.8325 0.9077
    Friedman rank 1.33 2.83 4 1.84
    RCA Fitness 0.0761 0.0899 0.1006 0.0519
    Accuracy 0.9559 0.9428 0.9269 0.9452
    AUC 0.9287 0.9089 0.8876 0.9052
    Friedman rank 1.33 2.67 4 2

     | Show Table
    DownLoad: CSV

    Next, we compare WOA feature selection methods based on different wrappers. We choose SVM, DT, and RF to compare with KNN, and run each method 60 times to generate multiple feature subsets. The average performance of these four algorithms on validation set and Friedman test results are given in Table 4. We can conclude from Table 4 that the KNN-based WOA is more suitable for feature selection of the problems in this study. Because the WOA has low complexity, fast convergence, and good optimization performance, and the KNN has low computational complexity and is easy to repeat, the KNN-based WOA method is suitable for the feature selection problem of this study.

    Table 4.  Comparison of WOA feature selection methods based on different wrappers.
    Artery Metrics KNN-WOA SVM-WOA DT-WOA RF-WOA
    LAD Fitness 0.0685 0.1702 0.1171 0.2506
    Accuracy 0.9801 0.8309 0.8857 0.75
    AUC 0.9627 0.7652 0.8823 0.7599
    Friedman rank 1 3 2 4
    LCX Fitness 0.0764 0.1615 0.1401 0.1729
    Accuracy 0.9464 0.8405 0.8619 0.8286
    AUC 0.9382 0.7811 0.8606 0.8570
    Friedman rank 1 3.33 2 3.67
    RCA Fitness 0.0761 0.1217 0.1463 0.1727
    Accuracy 0.9559 0.8809 0.8549 0.8286
    AUC 0.9287 0.8307 0.8269 0.8043
    Friedman rank 1 2 3 4

     | Show Table
    DownLoad: CSV

    In the feature selection process, we obtain multiple feature subsets by setting different iterations and the sizes of the initial population for WOA. We run 20 times for each parameter combination to find the corresponding optimal feature subset. So, for all parameter combinations, we run a total of 120 times. The number of features, classification accuracy and AUC values of the optimal feature subsets on test sets obtained by multiple operations and comparison under different parameter combinations are given in Table 5.

    Table 5.  Classification performance of feature subsets obtained under different parameter combinations.
    Artery Metrics N = 20
    T = 200
    N = 20
    T = 300
    N = 20
    T = 400
    N = 30
    T = 200
    N = 30
    T = 300
    N = 30
    T = 400
    LAD Number 17 8 17 17 19 18
    Accuracy 0.8129 0.7290 0.7742 0.8968 0.6548 0.7742
    Recall 0.9118 0.8235 0.8824 1 0.7235 0.9412
    AUC 0.9005 0.7546 0.7761 0.9067 0.7386 0.8739
    Friedman rank 2.5 4 3.625 1.5 6 13.5
    LCX Number 10 11 17 21 17 12
    Accuracy 0.6451 0.6871 0.8871 0.5903 0.7419 0.7774
    Recall 0.77 0.63 0.82 0.52 0.5 0.6
    AUC 0.7057 0.6381 0.9019 0.5519 0.6333 0.7048
    Friedman rank 2.5 3.25 1.875 5.75 4.625 3
    RCA Number 18 17 22 18 13 10
    Accuracy 0.5613 0.8581 0.6129 0.7452 0.6829 0.7097
    Recall 0.4923 0.7923 0.5 0.5371 0.4219 0.3846
    AUC 0.5184 0.9252 0.6731 0.7603 0.7134 0.7692
    Friedman rank 5.125 1.5 4.75 2.875 3.25 3

     | Show Table
    DownLoad: CSV

    For the diagnosis of three main coronary artery stenosis, the optimal features with better performance are obtained under the parameter combinations of (N = 30, T = 200), (N = 20, T = 400), and (N = 20, T = 300) respectively. Through the KNN-based WOA feature selection method, 17 features are randomly obtained for the diagnosis of LAD, LCX and RCA. The results of feature selection and the Pearson correlation coefficient between features and labels are shown in Table 6.

    Table 6.  Feature selection results and Pearson correlation coefficient of LAD, LCX and RCA.
    LAD (N = 30, T = 200) LCX (N = 20, T = 400) RCA (N = 20, T = 300)
    Feature Coefficient Feature Coefficient Feature Coefficient
    Typical CP 0.47 Typical CP 0.31 Lymph -0.23
    Atypical -0.36 Age 0.29 Age 0.22
    Region RWMA 0.36 HTN 0.18 Atypical -0.22
    Age 0.27 PLT -0.16 Poor R Progression 0.2
    Nonanginal CP -0.22 FBS 0.13 ESR 0.17
    T inversion 0.2 K 0.11 HTN 0.16
    Lymph -0.19 EF-TTE -0.11 FBS 0.15
    Neut 0.18 Nonanginal CP -0.099 WBC 0.14
    PR 0.18 Sex 0.086 Na -0.12
    HTN 0.16 Thyroid Disease -0.086 Q wave 0.11
    BP 0.16 ESR 0.08 K 0.089
    Na -0.15 T inversion 0.043 PR 0.088
    FBS 0.14 Dyspnea -0.038 Dyspnea -0.075
    Poor R Progression 0.11 LDL 0.035 Airway disease 0.052
    BMI -0.087 BMI -0.033 Current Smoker 0.023
    DLP -0.086 Lung rales 0.023 BMI -0.021
    LDL 0.011 LVH 0.019 BBB -0.017

     | Show Table
    DownLoad: CSV

    In this section, we first train multiple basic classifiers, then evaluate and compare the performance of these classifiers. The parameters of each classifier are adjusted on the training set by randomized search combined with 10-fold cv, and the average AUC value is taken as the parameter tuning metric to obtain the optimal hyperparameter combination. We compare the KNN, SVM, DT, RF, GBDT, XGBoost, and AdaBoost algorithms to select four best classifiers as the primary learners to build stacking models. The performance of each algorithm on each main coronary artery test set is shown in Table 7.

    Table 7.  Performance of each algorithm in LAD, LCX, and RCA diagnosis.
    Artery Metrics Algorithm
    KNN SVM DT RF GBDT XGBoost AdaBoost
    LAD Accuracy (%) 80.65 83.87 77.42 83.87 87.1 87.1 87.1
    Recall (%) 94.12 100 88.24 100 94.12 94.12 88.24
    Precision (%) 76.19 77.27 75 77.28 84.21 84.21 88.24
    F1-score (%) 84.21 87.18 81.08 87.18 88.89 88.89 88.24
    AUC 0.7374 0.916 0.8361 0.8782 0.9118 0.9076 0.916
    LCX Accuracy (%) 83.87 83.87 80.65 74.19 87.1 83.87 87.1
    Recall (%) 70 80 100 50 80 90 90
    Precision (%) 77.78 72.73 62.5 62.5 80 69.23 75
    F1-score (%) 73.68 76.19 76.92 55.56 80 78.26 81.82
    AUC 0.9 0.8667 0.8 0.819 0.9 0.8333 0.8952
    RCA Accuracy (%) 83.87 80.65 64.52 67.74 83.87 80.65 83.87
    Recall (%) 84.62 84.62 30.77 30.77 69.23 69.23 76.92
    Precision (%) 78.57 73.33 66.67 80 90 81.82 83.33
    F1-score (%) 81.48 78.57 42.11 44.44 78.26 75 80
    AUC 0.9316 0.8205 0.7607 0.7009 0.8248 0.7991 0.8419

     | Show Table
    DownLoad: CSV

    According to the "strong-strong combination" strategy of the stacking model, we should try to choose classifiers with good performance and maintain the diversity of primary learners. According to the performance indicators of each machine learning algorithm in the above table, we choose SVM, GBDT, XGBoost, and AdaBoost as primary learners for LAD diagnosis, choose KNN, SVM, GBDT, and AdaBoost for LCX diagnosis, choose KNN, SVM, GBDT, and AdaBoost for RCA diagnosis, and use LR as the secondary learner of these three models. For these three stacking models, we compare the performance of the stacking algorithm in the case of 3-fold, 5-fold, and 8-fold cv to select the best cross-validation fold. Moreover, we generate three sets of 10 random numbers between 0 and 1000 to control the random state of the stacking algorithm in each model. Then, we calculate the average value of the model evaluation indicators under 10 different random states to test the stability of the models.

    Table 8 shows the performance of the model in the diagnosis of main coronary arteries when different cross-validation folds are set for the stacking algorithm. From the data in the table, we can conclude that the 5-fold cv of the stacking algorithm can make our model obtain the best classification performance.

    Table 8.  The performance of the stacking model under different cross-validation folds.
    Artery CV Metrics
    Accuracy (%) Recall (%) Precision (%) F1-score (%) AUC
    LAD cv = 3 87.42 ± 2.82 98.24 ± 2.84 82.44 ± 3.59 89.58 ± 2.11 0.9029 ± 0.0169
    cv = 5 89.68 ± 2.04 100 ± 0.00 84.23 ± 2.44 91.42 ± 1.49 0.9067 ± 0.0018
    cv = 8 87.42 ± 3.21 99.41 ± 1.86 81.83 ± 3.93 89.71 ± 2.36 0.9029 ± 0.0169
    LCX cv = 3 87.1 ± 2.15 73 ± 11.6 87.62 ± 11.59 78.23 ± 4.06 0.9038 ± 0.0406
    cv = 5 88.71 ± 1.7 82 ± 4.22 83.03 ± 4.11 82.41 ± 2.59 0.9019 ± 0.0356
    cv = 8 86.45 ± 2.96 76 ± 8.43 81.63 ± 6.95 78.24 ± 5.19 0.9062 ± 0.0127
    RCA cv = 3 82.9 ± 4.57 73.85 ± 7.43 85.33 ± 11.61 78.47 ± 5.03 0.9154 ± 0.0112
    cv = 5 85.81 ± 2.26 79.23 ± 5.2 86.81 ± 7.55 82.44 ± 1.97 0.9252 ± 0.0088
    cv = 8 84.19 ± 3.21 76.92 ± 5.13 84.98 ± 8.25 80.39 ± 3.33 0.9218 ± 0.0962

     | Show Table
    DownLoad: CSV

    In Tables 911, we compare the classification performance of the proposed stacking models with each primary learner on the test set. It can be seen that the stacking models have the highest accuracy and F1-score in all prediction results, and the stacking models can also stabilize the recall, precision, and AUC value in a good range.

    Table 9.  Performance of stacking model and primary learners in LAD diagnosis.
    Metrics Algorithm
    SVM GBDT XGBoost AdaBoost Stacking
    Accuracy (%) 83.87 87.1 87.1 87.1 89.68 ± 2.04
    Recall (%) 100 94.12 94.12 88.24 100 ± 0.00
    Precision (%) 77.27 84.21 84.21 88.24 84.23 ± 2.44
    F1-score (%) 87.18 88.89 88.89 88.24 91.42 ± 1.49
    AUC 0.916 0.9118 0.9076 0.916 0.9067 ± 0.0018
    Friedman rank 3.6 2.5 3.3 2.9 2.1

     | Show Table
    DownLoad: CSV
    Table 10.  Performance of stacking model and primary learners in LCX diagnosis.
    Metrics Algorithm
    KNN SVM GBDT AdaBoost Stacking
    Accuracy (%) 83.87 83.87 87.1 87.1 88.71 ± 1.7
    Recall (%) 70 80 80 90 82 ± 4.22
    Precision (%) 77.78 72.73 80 75 83.03 ± 4.11
    F1-score (%) 73.68 76.19 80 81.82 82.41 ± 2.59
    AUC 0.9 0.8667 0.9 0.8952 0.9019 ± 0.0356
    Friedman rank 4 4.4 2.7 2.7 1.2

     | Show Table
    DownLoad: CSV
    Table 11.  Performance of stacking model and primary learners in RCA diagnosis.
    Metrics Algorithm
    KNN SVM GBDT AdaBoost Stacking
    Accuracy (%) 83.87 80.65 83.87 83.87 85.81 ± 2.26
    Recall (%) 84.62 84.62 69.23 76.92 79.23 ± 5.2
    Precision (%) 78.57 73.33 90 83.33 86.81 ± 7.55
    F1-score (%) 81.48 78.57 78.26 80 82.44 ± 1.97
    AUC 0.9316 0.8205 0.8248 0.8419 0.9252 ± 0.0088
    Friedman rank 2.3 4.1 3.6 3 2

     | Show Table
    DownLoad: CSV

    From Table 9 we can see that compared with other machine learning algorithms, stacking model has the highest accuracy (89.68%), recall (100%), and F1-score (91.42%) in the diagnosis of LAD stenosis, especially the recall of 100%. This shows that the model can identify all patients with artery stenosis in the LAD test set, and minimize the possibility of missed diagnosis in patients with CAD. The high accuracy and F1-score indicate that the model also has good ability to identify patients without CAD. In Table 10, compared with other algorithms, the stacking model achieves the highest accuracy (88.71%), precision (83.03%), F1-score (82.41%), and AUC value (0.9019) in the diagnosis of LCX, indicating that the model can accurately distinguish patients with LCX stenosis from those without stenosis. In Table 11, compared with other machine learning algorithms, stacking model in diagnosis of RCA obtains the highest accuracy (85.81%) and F1-score (82.44%). The AUC value (0.9252) of the model remains at a relatively good level, which also shows that the model can accurately distinguish the patients with RCA stenosis and non-stenosis, and effectively reduce the possibility of missed diagnosis in patients with CAD. From the above analysis, we can know that the classification performance of the stacking model is better than that of individual classifiers and ensemble classifiers based on bagging and boosting. The stacking model can combine the advantages of each primary learner to improve the prediction performance to the highest level.

    We compare the classification performance of the model proposed in this paper with the performance of the model established by using the recursive feature elimination cross-validation (RFE-CV) method based on SVM for feature selection [49]. The results are shown in Table 12. It can be seen that the proposed method has better classification performance.

    Table 12.  Performance of stacking model and primary learners in RCA diagnosis.
    Artery Method Metrics
    Accuracy (%) Recall (%) Precision (%) F1-score (%) AUC
    LAD WOA 89.68 100 84.23 91.42 0.9067
    RFE-CV 77.41 91.18 73.81 81.5 0.7908
    LCX WOA 88.71 82 83.03 82.41 0.9019
    RFE-CV 78.06 51 73.68 59.53 0.7881
    RCA WOA 85.81 79.23 86.81 82.44 0.9252
    RFE-CV 71.61 46.15 76.48 55.53 0.7299

     | Show Table
    DownLoad: CSV

    Figure 3 shows the comparison diagrams of the classification performance between the stacking model and each primary learner in diagnosing LAD, LCX and RCA stenosis.

    Figure 3.  Classification performance of LAD, LCX, and RCA diagnostics using stacking model versus primary learners.

    Figure 4 shows the ROC curves of the stacking model and other machine learning algorithms in diagnosing LAD, LCX and RCA stenosis, respectively.

    Figure 4.  ROC curves of LAD, LCX and RCA diagnostics.

    We compare the number of features and classification accuracy of the proposed model in this study with previous studies and show the differences between the methods and results in Table 13. It can be seen that the feature selection method used in this paper selects fewer features, and the classification accuracy of the proposed model is significantly better than other methods in diagnosing individual LAD, LCX and RCA stenosis.

    Table 13.  Comparison of proposed method with previous studies for detecting main coronary arteries stenosis.
    Study Feature selection method Model Number of features Accuracy (%)
    LAD LCX RCA LAD LCX RCA
    [50] All features C4.5 30 30 30 74.2 63.76 68.33
    [18] Weight by SVM SVM 24 24 24 86.14 83.17 83.5
    [21] Assurance SVM 30 40 40 86.64 83.47 82.85
    [22] Gini index and PCA SVM 23 16 26 86.43 83.67 82.67
    Proposed method KNN-WOA Stacking 17 17 17 89.68 88.71 85.81

     | Show Table
    DownLoad: CSV

    We apply the proposed model to the well-known Cleveland dataset to diagnose heart disease [51]. The dataset contains 303 patient records, each of which has 13 features for diagnosing heart disease [52]. In this study, we select 297 records with no missing values to build the model, of which 137 patients were diagnosed with heart disease. We use the same processing method as the Z-Alizadeh Sani dataset to complete the preprocessing of the Cleveland dataset, and apply the KNN-based WOA method for feature selection. We run 20 times for each parameter combination and obtain the optimal feature subset under parameter combination of (T = 20, N = 300). The feature subset containing 6 features including resting electrocardiographic results (restecg), maximum heart rate achieved (thalach), exercise induced angina (exang), the slope of the peak exercise ST segment (slope), number of major vessels colored by flourosopy (ca), and thallium scan (thal). Then, we build a stacking model based on the selected features and choose KNN, SVM, GBDT, and AdaBoost as primary learners. The proposed method achieves a classification accuracy of 89.67% and an AUC value of 0.9129 on the Cleveland dataset. We compare the classification performance of the stacking model with each primary learner and show the results in Table 14.

    Table 14.  Performance of stacking model and primary learners on the Cleveland dataset.
    Metrics Algorithm
    KNN SVM GBDT AdaBoost Stacking
    Accuracy (%) 86.67 86.67 86.67 86.67 89.67 ± 1.05
    Recall (%) 78.57 71.42 78.57 78.57 77.86 ± 2.26
    Precision (%) 91.67 100 91.67 91.67 100 ± 0.00
    F1-score (%) 84.62 83.33 84.62 84.62 87.53 ± 1.48
    AUC 0.9196 0.9018 0.9063 0.8973 0.9129 ± 0.0212
    Friedman rank 2.7 3.8 3.1 3.5 1.9

     | Show Table
    DownLoad: CSV

    Table 15 shows the comparison of the performance of the proposed model and the models in previous studies on the Cleveland dataset. It can be seen from the comparison results that our proposed method selects fewer features on the Cleveland dataset, and the model can achieve high precision and AUC value in the diagnosis of heart disease.

    Table 15.  Comparison of proposed method with previous studies on the Cleveland dataset.
    Study Feature selection method Model Number of features Accuracy (%) Precision (%) F1-score (%) AUC
    [5] PSO-based Two-tier ensemble 7 85.71 -- 86.49 0.8586
    [53] All features HRFLM 13 88.47 87.5 90 --
    [54] MAFW CNN 8 90.1 88.9 87.8 --
    [55] Chi-Squared SMO 11 86.47 86.5 86.4 0.861
    Proposed method KNN-WOA Stacking 6 89.67 100 87.53 0.9129
    Note:(The "--" in Table 15 indicates that this performance metric was not reported in the study.)

     | Show Table
    DownLoad: CSV

    This paper proposes new models for main coronary arteries diagnosis. We use the KNN-based WOA for feature selection on the extension of Z-Alizadeh Sani dataset and apply stacking models to diagnose LAD, LCX and RCA stenosis.

    In the feature selection process, first, we delete the feature with zero variance in the dataset. Then, by comparing the feature selection results of multiple meta-heuristic optimization algorithms and different wrappers, the KNN-based WOA method is used to select the optimal feature subsets. By using this method, we obtain three optimal feature subsets for diagnosing each main coronary artery, each of which contains 17 features. According to the results of feature selection, it can be seen that the features Age, BMI, FBS, and HTN appear in the feature subsets of each main coronary artery, which indicates that these features are important indicators affecting CAD.

    In the stacking model, we choose different primary learners for the three coronary arteries, and use LR as the secondary learner. In this paper, the average classification performance of the stacking algorithm in multiple random states is calculated to obtain stable results. The diagnostic accuracy of the proposed method in this research for LAD, LCX and RCA stenosis is 89.68, 88.71 and 85.81%, respectively. The classification performance of the stacking model is more stable than that of other machine learning algorithms. Compared with previous studies, we select relatively fewer features in this study, and the diagnostic accuracy of the proposed model is also significantly improved. Our results show that the proposed method can be well applied to CAD datasets and provide a reliable and robust model for clinical diagnosis.

    In the future study, we intend to use improved WOA to select fewer features and make accurate predictions of each main coronary artery stenosis on larger CAD datasets.

    Thanks to our families and colleagues who supported us morally.

    All authors declare no conflicts of interest in this paper.


    Acknowledgments



    We thank UNICEF-India and UNFPA-India for wholeheartedly supporting this intervention. We also appreciate Mr Sanjeev Kumar, Research Specialist, Health Systems Transformation Platform, for reviewing this paper and improving the content.

    Conflict of interest



    There are no financial interests to disclose.

    [1] Ministry of Home AffairsCensus of India Website: Office of the Registrar General & Census Commissioner, India (2011). Available from: https://censusindia.gov.in/census.website/data/HH2011.
    [2] UN Habitat, Slums: Some Definitions, State of the World's Cities. Available from: https://unhabitat.org/state-of-the-worlds-cities-20062007.
    [3] Ministry of Home Affairs G of ICensus of India Website: SRS Statistical Report (2011). Available from: https://censusindia.gov.in/nada/index.php/catalog/42684.
    [4] Bailey RK Intimate Partner Violence: An Evidence-Based Approach (2021). https://doi.org/10.1007/978-3-030-55864-2
    [5] Edwards KM (2015) Intimate Partner Violence and the Rural–Urban–Suburban Divide: Myth or Reality? A Critical Review of the Literature. Trauma Violence Abuse 16: 359-373. https://doi.org/10.1177/1524838014557289
    [6] Yakubovich AR, Heron J, Feder G, et al. (2020) Long-term Exposure to Neighborhood Deprivation and Intimate Partner Violence among Women: A UK Birth Cohort Study. Epidemiology 31: 272-281. https://doi.org/10.1097/EDE.0000000000001144
    [7] Jesmin SS (2017) Social Determinants of Married Women's Attitudinal Acceptance of Intimate Partner Violence. J Interpers Violence 32: 3226-3244. https://doi.org/10.1177/0886260515597436
    [8] Chakraborty D (2021) The “living dead” within “death-worlds”: Gender crisis and covid-19 in India. Gend Work Organ 28: 330-339. https://doi.org/10.1111/gwao.12585
    [9] International Institute for Population SciencesNational Family Health Survey (NFHS-4) 2015–16 India (2017). Available from: http://rchiips.org/NFHS/NFHS-4Reports/India.pdf.
    [10] UN WomenViolence Against Women and Women's Economic Empowerment (2017). Available from: https://asiapacific.unwomen.org/en/news-and-events/stories/2017/10/addressing-violence-against-women-essential-to-economic-empowerment.
    [11] Gaikwad V, Rao DH (2014) A cross-sectional study of domestic violence perpetrated by intimate partner against married women in the reproductive age group in an urban slum area in Mumbai. Indian J Public Heal Res Dev 5: 49-54. https://doi.org/10.5958/j.0976-5506.5.1.013
    [12] Begum S, Donta B, Nair S, et al. (2015) Socio-demographic factors associated with domestic violence in urban slums, Mumbai, Maharashtra, India. Indian J Med Res 141: 783-788. https://doi.org/10.4103/0971-5916.160701
    [13] Maji S, Bansod S, Singh T (2021) Domestic violence during COVID-19 pandemic: The case for Indian women. J Community Appl Soc Psychol 32: 374-381. https://doi.org/10.1002/casp.2501
    [14] Johnson K, Green L, Volpellier M, et al. (2020) The impact of COVID-19 on services for people affected by sexual and gender-based violence. Int J Gynecol Obstet 150: 285-287. https://doi.org/10.1002/ijgo.13285
    [15] Sharma P, Sharma S, Singh N (2020) COVID-19: Endangering women's mental and reproductive health. Indian J Public Health 64: 251-252. https://doi.org/10.4103/ijph.IJPH_498_20
    [16] Usher K, Bradbury Jones C, Bhullar N, et al. (2021) COVID-19 and family violence: Is this a perfect storm?. Int J Ment Health Nurs 30: 1022-1032. https://doi.org/10.1111/inm.12876
    [17] Government of India D portalPradhan Mantri Garib Kalyan Package (PMGKP) | National Portal of India (2021). Available from: https://www.india.gov.in/spotlight/pradhan-mantri-garib-kalyan-package-pmgkp.
    [18] Surma SA, Hakim SS, Rahman Lushan MS (2021) Planning for Pandemic Resilience: COVID-19 experience from urban slums in Khulna, Bangladesh. J Urban Manag 10: 325-344. https://doi.org/10.1016/j.jum.2021.08.003
    [19] George CE, Inbaraj LR (2021) Re: George et al. High seroprevalence of COVID-19 infection in a large slum in South India; what does it tell us about managing a pandemic and beyond?. Epidemiol Infect 149: e152. https://doi.org/10.1017/S0950268821001382
    [20] WHOResponding to intimate partner violence and sexual violence against women (2011). Available from: https://www.who.int/publications/i/item/WHO-RHR-13.10.
    [21] IndiaCensus.net, Lucknow Population 2021. Available from: https://www.indiacensus.net/district/lucknow.
    [22] World Health Organization (WHO)Violence against women (2021). Available from: https://www.who.int/news-room/fact-sheets/detail/violence-against-women.
    [23] Krishnakumar A, Verma S (2021) Understanding domestic violence in India during COVID-19: a routine activity approach. Asian J Criminol 16: 19-35. https://doi.org/10.1007/s11417-020-09340-1
    [24] Onditi F, Obimbo M, Muchina SK, et al. (2020) Modeling a pandemic (COVID-19) management strategy for urban slums using social geometry framework. Eur J Dev Res 32: 1450-1475. https://doi.org/10.1057/s41287-020-00317-5
    [25] Ryan NE, El Ayadi AM (2020) A call for a gender-responsive, intersectional approach to address COVID-19. Glob Public Health 15: 1404-1412. https://doi.org/10.1080/17441692.2020.1791214
  • This article has been cited by:

    1. Ying Xu, Huixin Qin, Dynamical Response in an Electromechanical Arm Driven by Temperature-Dependent Neural Circuit, 2024, 05779073, 10.1016/j.cjph.2024.11.010
    2. D. Vignesh, Shaobo He, Santo Banerjee, A review on the complexities of brain activity: insights from nonlinear dynamics in neuroscience, 2024, 0924-090X, 10.1007/s11071-024-10558-2
    3. Yitong Guo, Chunni Wang, Jun Ma, Model approach of artificial muscle and leg movements, 2025, 529, 03759601, 130069, 10.1016/j.physleta.2024.130069
    4. Denggui Fan, Jin Chen, Songan Hou, Zhengyong Song, Gerold Baier, Qingyun Wang, From experimental phenomena to computational models: Exploring the synchronization mechanisms of phase-locked stimulation in the hippocampal–thalamic–cortical circuit for memory consolidation, 2025, 191, 09600779, 115754, 10.1016/j.chaos.2024.115754
    5. Qun Guo, Guodong Ren, Chunni Wang, Zhigang Zhu, Reliability and energy function of an oscillator and map neuron, 2025, 251, 03032647, 105443, 10.1016/j.biosystems.2025.105443
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2210) PDF downloads(96) Cited by(0)

Figures and Tables

Figures(4)  /  Tables(3)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog