
The beneficial effects of Prosopis africana (PA) on human health have been demonstrated; however, its protective effects against heavy metals (HM) are not yet understood. This study evaluated the potential neuroprotective effects of PA in the cerebral cortex and cerebellum. To accomplish this, we divided 35 albino Sprague Dawley rats into five groups. Group I did not receive either heavy metal mixture (HMM) or PA. Group II received a HMM of PbCl2 (20 mg/kg), CdCl2 (1.61 mg/kg), HgCl2 (0.40 mg/kg), and NaAsO3 (10 mg/kg) orally for a period of two months. Groups III, IV, and V received HMM along with PA at doses of 500, 1000, and 1500 mg/kg, respectively. PA caused decreased levels of HM accumulation in the cerebral cortex and cerebellum and improved performance in the Barnes maze and rotarod tests. PA significantly reduced levels of IL-6 and TNF-α. PA increased concentrations of SOD, CAT, GSH, and Hmox-1 and decreased the activities of AChE and Nrf2. In addition, levels of MDA and NO decreased in groups III, IV, and V, along with an increase in the number of live neurons. In conclusion, PA demonstrates a complex neuroprotective effect with the potential to alleviate various aspects of HM-induced neurotoxicity.
Citation: Orish E. Orisakwe, Evelyn Utomoibor Ikpeama, Chinna N. Orish, Anthonet N. Ezejiofor, Kenneth O. Okolo, Aleksandar Cirovic, Ana Cirovic, Ify L. Nwaogazie, Chinekwu Samson Onoyima. Prosopis africana exerts neuroprotective activity against quaternary metal mixture-induced memory impairment mediated by oxido-inflammatory response via Nrf2 pathway[J]. AIMS Neuroscience, 2024, 11(2): 118-143. doi: 10.3934/Neuroscience.2024008
[1] | Moquddsa Zahra, Dina Abuzaid, Ghulam Farid, Kamsing Nonlaopon . On Hadamard inequalities for refined convex functions via strictly monotone functions. AIMS Mathematics, 2022, 7(11): 20043-20057. doi: 10.3934/math.20221096 |
[2] | Ghulam Farid, Hafsa Yasmeen, Hijaz Ahmad, Chahn Yong Jung . Riemann-Liouville Fractional integral operators with respect to increasing functions and strongly (α,m)-convex functions. AIMS Mathematics, 2021, 6(10): 11403-11424. doi: 10.3934/math.2021661 |
[3] | Hengxiao Qi, Muhammad Yussouf, Sajid Mehmood, Yu-Ming Chu, Ghulam Farid . Fractional integral versions of Hermite-Hadamard type inequality for generalized exponentially convexity. AIMS Mathematics, 2020, 5(6): 6030-6042. doi: 10.3934/math.2020386 |
[4] | Maryam Saddiqa, Ghulam Farid, Saleem Ullah, Chahn Yong Jung, Soo Hak Shim . On Bounds of fractional integral operators containing Mittag-Leffler functions for generalized exponentially convex functions. AIMS Mathematics, 2021, 6(6): 6454-6468. doi: 10.3934/math.2021379 |
[5] | Maryam Saddiqa, Saleem Ullah, Ferdous M. O. Tawfiq, Jong-Suk Ro, Ghulam Farid, Saira Zainab . k-Fractional inequalities associated with a generalized convexity. AIMS Mathematics, 2023, 8(12): 28540-28557. doi: 10.3934/math.20231460 |
[6] | Atiq Ur Rehman, Ghulam Farid, Sidra Bibi, Chahn Yong Jung, Shin Min Kang . k-fractional integral inequalities of Hadamard type for exponentially (s,m)-convex functions. AIMS Mathematics, 2021, 6(1): 882-892. doi: 10.3934/math.2021052 |
[7] | Yue Wang, Ghulam Farid, Babar Khan Bangash, Weiwei Wang . Generalized inequalities for integral operators via several kinds of convex functions. AIMS Mathematics, 2020, 5(5): 4624-4643. doi: 10.3934/math.2020297 |
[8] | Manar A. Alqudah, Artion Kashuri, Pshtiwan Othman Mohammed, Muhammad Raees, Thabet Abdeljawad, Matloob Anwar, Y. S. Hamed . On modified convex interval valued functions and related inclusions via the interval valued generalized fractional integrals in extended interval space. AIMS Mathematics, 2021, 6(5): 4638-4663. doi: 10.3934/math.2021273 |
[9] | Moquddsa Zahra, Muhammad Ashraf, Ghulam Farid, Kamsing Nonlaopon . Inequalities for unified integral operators of generalized refined convex functions. AIMS Mathematics, 2022, 7(4): 6218-6233. doi: 10.3934/math.2022346 |
[10] | Wenfeng He, Ghulam Farid, Kahkashan Mahreen, Moquddsa Zahra, Nana Chen . On an integral and consequent fractional integral operators via generalized convexity. AIMS Mathematics, 2020, 5(6): 7632-7648. doi: 10.3934/math.2020488 |
The beneficial effects of Prosopis africana (PA) on human health have been demonstrated; however, its protective effects against heavy metals (HM) are not yet understood. This study evaluated the potential neuroprotective effects of PA in the cerebral cortex and cerebellum. To accomplish this, we divided 35 albino Sprague Dawley rats into five groups. Group I did not receive either heavy metal mixture (HMM) or PA. Group II received a HMM of PbCl2 (20 mg/kg), CdCl2 (1.61 mg/kg), HgCl2 (0.40 mg/kg), and NaAsO3 (10 mg/kg) orally for a period of two months. Groups III, IV, and V received HMM along with PA at doses of 500, 1000, and 1500 mg/kg, respectively. PA caused decreased levels of HM accumulation in the cerebral cortex and cerebellum and improved performance in the Barnes maze and rotarod tests. PA significantly reduced levels of IL-6 and TNF-α. PA increased concentrations of SOD, CAT, GSH, and Hmox-1 and decreased the activities of AChE and Nrf2. In addition, levels of MDA and NO decreased in groups III, IV, and V, along with an increase in the number of live neurons. In conclusion, PA demonstrates a complex neuroprotective effect with the potential to alleviate various aspects of HM-induced neurotoxicity.
Object recognition aims at detecting the objects and predicting the class label of a given image, which has been widely used in classification [1,2], localization [3,4,5,6,7], segmentation [8,9], retrieval [10,11,12] and natural language processing [13,14], etc. The significant advances have been reported in a large number of deep learning literatures [15,16,17,18,19]. Despite the exciting success, most methods proposed in those papers are based on supervised learning, which is driven by the availability of manually annotated instances with powerful low-level visual features [7]. However, the frequencies of objects in the wild follow a long-tailed distribution that consist of a few common classes and most rare classes [20]. On one hand, it is difficult for rare classes without sufficient representative labeled instances to train a classifier effectively. Moreover, it is extremely challenging to collect large-scale labeled instances, even if the performance of the model is improved by adding more instances. Taking the large-scale dataset ImageNet [21] as an example, it contains a total of 14M images in 21,841 classes. It is unrealistic to exhaustively annotate hundreds of instances for each class. On the other hand, the labeled instances of certain classes are precious and difficult to obtain significant amount of the corresponding annotated instances, e.g., endangered bird breed in fine-grained datasets, which is hard to annotate images without expert knowledge [22], let alone collecting instances. In addition, new objects emerge over time that are not covered by known classes and have no labeled instances beforehand, e.g., the high-quality radiology images of the patients infected by COVID-19 are not available before 2019. As a result, the conventional approaches cannot tackle above problems. There are increasing efforts to address the problem of insufficient or even no labeled instances, such as one-shot and few-shot learning [23] deal with the classes of few labeled instances; open world recognition performs the tasks: detecting the novelty of the test classes via open set recognition that was initially proposed by [24], progressively labeling instances of novel unseen classes by class-incremental learning, and adapting the model to classify the acquired labeled instances [25]. The above-mentioned techniques reduce the dependence on labeled instances and improve the accuracy, but still require at least some labeled instances for model learning. Unfortunately, the aforementioned strategies fail to determine the class labels of the instances belonging to unseen classes that have no labeled data.
In contrary, humans have the ability to recognize unseen classes by intelligently utilizing the previously learned knowledge extracted from the seen classes. For example, a learner can easily recognize the Persian fallow deer, if he/she has ever seen fallow deer and is aware that it resembles the fallow deer with bigger antlers and white spots around the neck. Therefore, they are capable of distinguishing beyond 30,000 objects [26] as well as varieties of the subordinates. Inspired by the mechanism of human's ability to recognize new objects without seeing all classes in advance, Zero-Shot Learning (ZSL) [27,28,29,30,31]has drawn significant attention and is proposed to recognize the entirely novel classes omitted from training instances by extrapolation from the knowledge contained in the observed classes [32]. More specifically, given labeled training instances of seen classes in the source domain, ZSL aims to establish a model to classify the instances of unseen classes in the target domain, which increasingly reduces the resources in labor and time expenses. In addition to computer vision related to images, the applications of ZSL has been emerging in various fields, such as zero-shot translation [33], bilingual dictionary induction [34] and molecular compound analysis [35].
In the absence of the labeled instances of unseen classes, the key idea underpinning ZSL methods is to explore the knowledge that transfer via shared auxiliary information. Seen classes are associated with unseen classes in a common space, i.e., semantic space, and the high-level semantic representations are considered as auxiliary information among these classes. Thereby, they can act as a bridge to guarantee the feasibility of ZSL. Generally, there are multiple types of semantic information: attributes [36], word vectors [37,38,39], textual descriptions [40], hierarchical ontology [41,42], etc. The commonly used semantic space nowadays is attribute space [27]. Each class is endowed with a unique semantic prototype [43] in this space. The prototype is specified by a binary or continuous attribute vector that indicates the class properties manually designed by experts. The relatedness of the classes is represented by the similarity of the semantic prototypes, e.g., the semantic prototype of zebra is closer to that of horse instead of pig, which is agreement to the reality that zebra is semantically related to horse. Therefore, ZSL can learn a model properly with the aid of semantic representations. Most existing ZSL approaches [27,40,41,44,45,46] exploit a visual-semantic projection to reflect the relationship among the classes. Specifically, the projection is learned to map the low-level visual features of the labeled instances consisting of seen classes only to semantic space during training. At test stage, the learned projection function is applied to map the target instances of unseen classes to the same semantic embedding space where seen and unseen classes reside. Then, the similarities of the predicted semantic presentations and prototypes are measured by certain matric. Employing the nearest neighbor (NN) search, the classification of the target instance is realized by aligning the semantic prototype of unseen class that yields the highest score.
Despite the success of those semantic embedding models, the largest challenge in ZSL is the projection domain shift problem [43] among the disjoint seen and unseen classes and is manifested through the following aspects. On the one hand, the visual feature space is mutually independent of semantic space, and they have distinct distributions. Hence, there is great difficulty in learning an effective and compatible projection function between the two spaces. On the other hand, the visual appearance of the same attributes in seen and unseen classes are fairly different. The discrepancy is analyzed empirically in [43]. It can be seen that shared characteristic "has tail" in target unseen class Pig is visually different from the source seen class Zebra. Thus, there are significant differences in the underlying distributions of the classes that leads to poor performance on novel classes. In other words, if the projection functions learned with the training instances in seen classes are directly adopted to the unseen classes without adaptation, the target instance tends to be shifted far away from the corresponding class prototype, resulting in the unsatisfactory recognition by NN search at test stage.
There is a recent surge of interest in building a better generalizable projection function on the novel classes to be less susceptible to domain shift. Firstly, a large volume of the literatures belongs to inductive setting are published to overcome this problem [27,32,47,48,49]. The most representative one is SAE [48] and it learns the linear projection function from visual feature space to semantic space based on auto-encoder paradigm, in which the decoder is the transpose of encoder and imposed by a reconstruction constraint of the original visual features. However, inductive methods only have access to the seen data, the projection is likely to capture the characteristics of the seen classes rather the unseen ones. As a result, it hinders the effective generalization. Secondly, the generative models are proposed to compensate for the visual features of target unseen data. The two prominent members are Generative Adversarial Networks (GANs) [50,51,52,53,54] and Variational Auto-Encoders (VAEs) [55] that synthesize visual features by utilizing the semantic prototypes of unseen classes. While, it is noticed that the choice of the semantic prototype is essential, as low quality may degrade the effectiveness of the generator. Afterwards, various methods resort to transductive learning [43,56,57,58,59,60,61,62] leverage the unlabeled target instances during training. The existing transductive learning methods are classified into three categories. The first one is label propagation. For example, Fu et al. [43] combines multiple semantic representations with visual features of unseen classes to learn a joint embedding space, in which the target data are aligned with the label embeddings and then the recognition is performed via label propagation. The second one is self-training that progressively improves the classification capacity in an iterative refining process [59]. The last one termed domain adaptation is the most relevant method to our model and has been well-investigated to uncover the common knowledge of the source and target domains [63,64,65]. Different from the above first two strategies, [65] simultaneously utilize the visual features and semantic prototypes of unseen classes, our model only utilizes the visual features of the unlabeled instances. Furthermore, our model is not concerned with the distribution alignment of the projected and original domains [63] or that of the features in the immediate space [64], whereas the latent space in our model is semantically meaningful and encourages learning the generalizable semantics containing sufficient information of the visual features through a reconstruction task.
In this paper, we develop a novel model by exploiting the idea of autoencoder framework to solve zero-shot challenges and reveal the relationship between visual features and semantic representations. We assume that the majority semantic properties of the unseen classes are shared with that of seen classes. Following the previous work, we adopt the semantic space as the latent embedding space to preserve the semantic relatedness between the classes. Motivated by [48,63], our model takes advantage of the bi-shifting linear auto-encoder framework. In specific, the common encoder shared by source and target domains tries to learn the projection from visual feature space to semantic space. Considering the distribution divergence of the disjoint domains, the original features are reconstructed by two different decoders based on the learned semantic representations. It is worth mentioning that there are two regularization terms in our model. Inspired by [48], the first term is designed to inherit the properties of semantic space by incorporating the semantic prototypes of the seen classes and then the projections are constrained to force the learned semantics of unlabeled target instances as close as possible to their class prototypes. Consequently, the semantic mismatch between visual features and semantic representations can be refined. The second regularization term is adopted to enforce that the decoder in target domain is derived from, rather than the same as the decoder supervised learned via semantics of instances in source domain, resulting in truthfully reconstruct the visual features. To that end, we design a novel Bi-shifting Semantic Auto-Encoder (hereafter referred as BSAE) architecture that integrates the merits of both domain adaptation and discriminative ability of class semantics, as shown in Figure 1. However, BSAE is based on linear auto-encoder and quite shallow, experimental results reported in section 4 prove its outstanding performance. For example, the average accuracy on five benchmark datasets under different protocols are whopping 6.9% improvement over the current state-of-the-art. In conclusion, the main contributions are three-fold:
● A simple and effective ZSL model termed BSAE is developed in an auto-encoder framework. Our model not only alleviates the domain shift problem, but also recovers the interaction between visual feature and semantic representation.
● We consider ZSL as the problem of learning the projection functions to explore shared discriminative semantic representations of instances, which are supervised by the semantic prototypes of the seen instances. Meanwhile, the generalizable capability of the learned semantics are enhanced by exploiting the visual features of unlabeled instances.
● An iterative algorithm with high computational efficiency is introduced to solve the problem. Extensive experimental results demonstrate that our approach achieves superior performance on five benchmark datasets, even if the class prototypes of the unseen classes are not available.
The remainder of the paper is organized as follows. In Section 2, we briefly review the related work proposed to overcome the challenges in ZSL. In Section 3, we describe the proposed model and deduce an efficiently iterative algorithm. The results are reported and discussed in Section 4. Finally in Section 5, we present the conclusion and propose several research directions to be investigate.
In this section, we firstly introduce a review about the semantic space exploited in current zero-shot learning. Then, we briefly review of projection learning concerned with our work. Finally, we present the related advances to relive the domain shift problem.
Semantic representation shared between classes bridges the gap in ZSL and enables the transmission of common knowledge from seen to unseen classes. There are various semantic spaces formed by different class embeddings. Attribute space [27] is the most popular and effective one [46,66,67], in which the properties of the classes are described as attributes. However, manually collecting and annotating attributes are heavy reliance on the efforts of experts. The word vector [38,39] and text description [40] based semantic space are proposed because of relatively less labor intensive. The semantic representations are automatically extracted by embedding models from text corpus (e.g., Wikipedia). In spite of the inconvenience for humans to incorporate the knowledge of the classes into the semantics, as reported in [40], 10 sentence descriptions are collected for each image to construct the semantic space, which is even more expensive than annotating attributes. Moreover. SJE [41] and ESZSL [47] have shown that attribute space is more effective than word vector space. Besides, several ZSL methods take advantage of them via combining the aforementioned semantic spaces [50,60,68]. In our work, we consider the attribute space as semantic space.
The existing ZSL models can be sub-categorized into three groups, depending on how the projection function is established.
The first group learns a forward projection from visual feature space to semantic space. Lampert et al. [27] proposed two attribute-based classifiers includes direct attribute predictor (DAP) and indirect attribute predictor (IAP) that exploit the attributes to predict the class labels of instances in a two-stage schema. SOC [32] firstly projects the visual features into the semantic space, and then determines the class label through KNN. CONSE [37] exploits a probabilistic model and then predicts the unseen classes via the convex combination of the class-embeddding vectors. DeViSE [38] applies linear corresponding function by combining similarity and hinge ranking loss. ESZSL [47] learns the bilinear compatibility function by optimizing square loss. To optimize the ranking loss, ALE [58] employs a bilinear mapping compatibility function.
The second group learns a reverse projection from the semantic space to the visual space to rectify the hubness problem [69]. The hubness refers to the phenomenon of some semantic prototypes are nearest neighbors of instances from different classes, which is a curse of demensionality. Zhang et al. [70] proposed a deep end-to-end neural network to embed the class prototypes into the visual feature space that suffer much less from the hubness problem, as discussed in [71]. In addition to the embedding models, generative-based methods are proposed recently to generate instances for unseen classes by leveraging the semantic prototypes, then the ZSL problem is converted into a traditionally supervised problem. f-CLSWGAN [50] and LisGAN [68] explore the conditional generator on semantics to synthesize the visual features. However, it is hard to train generative models because of the min-max optimization. Auto-encoder is an effective framework to extract the representative features in a unsupervised manner and alleviate the domain shift problem. Xu et al. [52] construct visual feature space as latent layer and learns two different regressors for semantic reconstructions. The latent layer of [48] is semantic space and the linear projection between the visual feature and semantic space is learned with the semantic constraint of seen classes to reconstruct the original data. [72] improves the model in [48] by adding a regularization constraint of the projection function, thereby ensuring that the structural risk of the model is minimized.
In the last group, both visual features and semantic representations are projected into a common space. SYNC [49] learns classifiers of unseen classes by linearly combining base classifiers. Zhang and Saligrama [67] leverage similar class relationships in the common space, which is defined by the seen classes proportions.
Taking full advantage of the first two groups, our model is close to [63] in which a bi-shifting auto-encoder is employed for reconstructing visual features in different domains. Different with [63] that apply nonlinear projections to learn the representations in latent space, our model reinforces the latent space as semantic space with class semantic prototypes and exploits the linear projection functions to fit the distributions of visual and semantic spaces, respectively.
Domain shift problem was firstly reported in [43] and is an open issue in ZSL. It describes that the projection functions learned from the seen classes are biased when exploit them to map instances of unseen classes from visual feature space to semantic space. It is essentially caused by the disjoint seen and unseen classes with different underlying data distributions. The researchers have investigated how to rectify the domain shift problem and obtain competitive results, for instance, SAE [48] imposes an additional reconstruction constraint to the training seen data, resulting in the learned projection function more generalizable across seen and unseen classes. LisGAN [68] refines the domain shift via generating the soul instances related to the semantic representations. However, as the unseen class data are not involved in the model learning, the generalizable ability of the inductive methods is limited. Transductive ZSL is an emerging topic to mitigate the domain shift problem where not only labeled seen class data are available, but also has access to unlabeled unseen class data, which potentially leads to improvements in classification performance. Fu et al. [43] first propose a transductive multi-view embedding framework, and then generate the class labels for unseen class data via label propagation. Kodirov et al. [65] formulate a regularized sparse coding framework to solve the domain shift problem. A measure of inter-class semantic consistency is proposed by [73] to explore the relation between the semantic manifold and visual-semantic projection on seen classes. VCL [56] proposes a visual structure constraint on class centers. Unlike SAE [48] exploits one decoder to reconstruct the features without domain adaptation, our model employs transductive setting and adopts two decoders to reconstruct the visual features in the source and target domains. Additionally, we restrict the similarity constraint of the two different decoders as a regularizer by considering the amount of adaptation from the labeled seen class data rather than being deviated freely. Although we only use the visual features of unseen data rather than the combination of the visual features and semantic representations of target unseen data like others [41,70], our model boosts ZSL performance.
In this section, we describe the procedures and methods used in this paper. we firstly set up the zero-shot learning problem, then develop our novel model BSAE for this task, and finally derive an efficient algorithm to solve it. Subsequently, the classification of unseen classes can be performed in the original feature space and semantic space.
We start by introducing some notations and problem definition of our interest. Considering m labeled source instances S={(xsi,ysi,ssi)|xsi∈X,ysi∈Cs}mi=1 are given as training data, where xsi∈Rd denotes the d-dimensional visual feature, ysi is the corresponding class label in Cs consisting of τ discrete seen classes, ssi is the semantic representation of ith instance. In addition, given p unlabeled target data T={(xti,yti,sti)|xti∈X,yti∈Ct}pi=1 of unseen classes, where xti∈Rd denotes the d-dimensional visual feature, yti is the corresponding label and belongs to the μ unseen classes set Ct. While the seen and unseen classes are disjoint, i.e., Cs∩Ct=∅, the semantic space A are associated with mitigating this challenge, which is spanned by attribute vector or word vector derived from text for each class. The k-dimensional semantic prototypes of seen and unseen classes are denoted as As=[as1,as2,…,asτ]∈Rk×τ and At=[at1,at2,…,atμ]∈Rk×μ. Therefore, Str=[ss1,ss2,…,ssm]∈Rk×m is given because the source data Xs=[xs1,xs2,…,xsm]∈Rd×m are labeled by either binary or continuous attributes indicating the corresponding class labels Ys={ysi}mi=1. On the contrary, as the target instances Xt=[xt1,xt2,…,xtp]∈Rd×p are unlabeled, Ste=[st1,st2,…,stp]∈Rk×p that stands for the semantic prototypes and Yt={yti}pi=1 that denotes the class labels have to be predicted.
The goal of standard zero-shot learning is to predict the correct class of Xt by learning a classifier f:T→Ct. The key notations used in this paper are listed in Table 1.
Notations | Descriptions |
Xs∈Rd×m | Visual feature of source instances |
Xt∈Rd×p | Visual feature of target instances |
Ys={ysi}mi=1 | Class labels of Xs |
Yt={yti}pi=1 | Class labels of Xt |
Cs={1,2,…,τ} | Set of τ seen classes |
Ct={1,2,…,μ} | Set of μ unseen classes |
At∈Rk×μ | Semantic prototypes of Ct |
Str∈Rk×m | Semantic representations of Xs |
Ste∈Rk×p | Semantic representations of Xt |
λ1,λ2 | Hyper-parameters |
X | d-dimensional visual feature space |
A | k-dimensional semantic space |
m,p | Number of source and target instances |
d,k | Dimensionality of visual feature and semantic space |
We begin our discussion with auto-encoder (AE) for it being the basis of our model. The simplest form of AE is linear and has one hidden layer [48] that is responsible to truthfully reconstruct the input data as similar as possible. We force the semantic space in hidden layer as [48] so that the latent space is semantically meaningful, e.g., each column of Str stands for the attribute vector of the corresponding labeled source instance.
Assume that labeled seen-class training set S and unlabeled target data Xt are available. The proposed BSAE aims to learn a model to estimate the discriminative semantic representations Ste and reconstructed features ˆXt of the target instances and then obtain their class labels Yt in semantic space and visual feature space, respectively. Specifically, considering the seen classes and unseen classes are related in the same class embedding space (e.g., attribute), BSAE consists of three components: (1) It attempts to learn the encoder parameterized by W0∈Rk×d (k<d) to project both domains from visual feature space X to the common semantic space A. In order to guarantee whether the learned semantic representations capture sufficient discriminative information, in terms of the distribution discrepancy between domains, (2) on one hand, the decoder W1∈Rd×k reconstructs the original visual features of source domain exactly. (3) On the other hand, the mapped class embeddings of target domain are projected to the visual features by decoder W2∈Rd×k. We simultaneously minimize the reconstruction errors in different domains by utilizing the unlabeled instances from unseen classes to narrow down the domain gap. Therefore, it is applicable to better generalize the learned regression model to unseen classes. As observed from Figure 2, our model preserves enough discriminative information across unseen classes, even in the low dimensional semantic space that exacerbates the hubness problem.
Our model is learned by optimizing the following objective:
minW0,W1,W2 J=‖Xs−W1W0Xs‖2F+‖Xt−W2W0Xt‖2Fs.t. W0Xs=Str, | (3.1) |
where ‖⋅‖F is the Frobenius norm of a matrix. Eq 3.1 denotes the loss of the autoencoder. It is difficult to solve the objective Eq 3.1 with a hard constraint. To fight off the constraint W0Xs=Str efficiently, we relax the constraint through incorporating a semantic similarity term into Eq 3.1:
minW0,W1,W2 J=‖Xs−W1Str‖2F+‖Xt−W2W0Xt‖2F+λ1‖W0Xs−Str‖2F | (3.2) |
and λ1 is the hyper-parameter. The first two terms are regarded as the losses of different decoders. The last term is the loss of encoder. W2 is unsupervised because of the unknown semantic representation Ste of target data and [65] proves that W2 adapted from W1 is efficient to this issue. To this end, we adds the regularization term ‖W2−W1‖2F to Eq 3.2 to restrict the amount of adaptation of the two projections. It is worth noting that W1 is considered as a basis to ensure W2 cannot deviate freely from W1. The full objective of our proposed model then becomes:
minW0,W1,W2 J=‖Xs−W1Str‖2F+‖Xt−W2W0Xt‖2F+λ1‖W0Xs−Str‖2F+λ2‖W2−W1‖2F, | (3.3) |
where λ2 is a hyper-parameter used to balance the importance of different terms.
Next, we will formulate our solver as a novel gradient-based algorithm to alternately update projection functions W0, W1 and W2. Note that the conventional iterative algorithms (e.g., Gradient Descent) have been widely exploited to directly solve such problems without computationally efficiency. Whilst our solver depends on the dimension of the features, not the number of instances and hence is more effective than the conventional iterative algorithms. To solve the Eq 3.3, we calculate the partial derivative of it and set it to zero:
∂JW0=−WT2(Xt−W2W0Xt)XTt+λ1(W0Xs−Str)XTs, | (3.4) |
∂JW1=−(Xs−W1Str)STtr−λ2(W2−W1), | (3.5) |
∂JW2=−(Xt−W2W0Xt)XTtWT0+λ2(W2−W1), | (3.6) |
and optimize the following sub-problems through alternative optimization methods.
Setting Eq 3.4 to zero, we obtain:
WT2W2W0+λ1W0XsXTs(XtXTt)−1=WT2+λ1StrXTs(XtXTt)−1. | (3.7) |
Let M0=WT2W2,H0=λ1XsXTs(XtXTt)−1,Q0=WT2+λ1StrXTs(XtXTt)−1, we have the Sylvester equation:
M0W0+W0H0=Q0, | (3.8) |
where M0∈Rk×k and H0∈Rd×d are square matrices, Q0∈Rk×d is a rectangle matrix. The above matrix function has a unique solution if it satisfies conditions of the Theorem 3.1 quoted in [74]. Obviously, it is easy to meet in practical applications.
Theorem 3.1. Eq 3.8 has a unique solution if and only if the matrices M0 and H0 have distinct eigenvalues, that is, the eigenvalues γ1,γ2,…,γk of M0 and ζ1,ζ2,…,ζd of H0 satisfy γi+ζj≠0 (i=1,...,k;j=1,...,d).
As a result, Eq 3.8 can be easily solved by Bartels-Stewart algorithm, which is implemented with a single line of code: sylvester in MATLAB:
W0=sylvester(M0,H0,Q0). | (3.9) |
Setting Eq 3.5 to zero, then we have the following Sylvester equation:
λ2W1+W1StrSTtr=XsSTtr+λ2W2. | (3.10) |
Let M1=λ2Ik, Ik is the k×k identity matrix, H1=StrSTtr, Q1=XsSTtr+λ2W2. The Eq 3.10 can be efficiently solved in MATLAB:
W1=sylvester(M1,H1,Q1). | (3.11) |
Similarly, we set Eq 3.6 to zero, and have the following formulation:
λ2W2+W2W0XtXTtWT0=XtXTtWT0+λ2W1. | (3.12) |
If we denote M2=M1,H2=W0XtXTtWT0,Q2=XtXTtWT0+λ2W1, the above Sylvester equation (3.12) can be solved in MATLAB:
W2=sylvester(M2,H2,Q2). | (3.13) |
Algorithm 1 summarizes the implementation of our algorithm. We simply initialize W2 with all elements of 0.1 for coarse-grained datasets (e.g., AWA1) and all elements of 0.01 for fine-grained datasets (e.g., CUB). The hyper-parameters λ1 and λ2 are selected by cross-validations. Details are listed in section 4. The iterations will terminate when the Eq 3.3 converges or reaches a fixed number of iterations.
Algorithm 1 Bi-shifting Semantic Auto-Encoder |
Input: Training data Xs, Str Test data Xt Hyper-parameters λ1, λ2 Output: Projection matrices W0, W2 1: Initialize W2 2: while not converge do 3: Update W0 by Eq 3.9 4: Update W1 by Eq 3.11 5: Update W2 by Eq 3.13 6: Check the converge condition 7: end while 8: Return W0, W2 |
To this end, we propose to briefly explain the analysis of the time complexity and convergence of our algorithm. As mentioned in Algorithm 1, optimizing the objective function Eq 3.3 is actually the process of solving three Sylvester equations. The time complexity of computing each Sylvester equation, e.g., Eq 3.8, given M0 and H0, is O(k3+d3)(d,k≪min(m,p)), which is independent of number of instances. In other words, it can be effectively applied to large-scale datasets. As can be observed from Eq 3.7 to Eq 3.13, due to linear formulations, it is easier to solve three sub-problems with respect to three projection functions W0, W1 and W2 in our proposed model. Concretely, updating each projection function is regarded to solve Sylvester equation. Hence, the objective function Eq 3.3 is non-increasing with a lower bound during the alternative optimization.
According to Algorithm 1, we obtain the optimal projection functions W0 and W2. We measure the similarity score between the estimated value of target instance and its prototype, and then predict the class label.
In the semantic space, considering a target instance xti, we could firstly calculate the estimated semantic representation with (Ste)i=W0xti, then compare with the prototypes At of classes in Ct by calculating the cosine distance between them:
l(xti)=argminj d((Ste)i,(At)j), | (3.14) |
where j∈[μ], (At)j is the prototype attribute vector of j–th unseen class and d(⋅,⋅) is a distance function. l(⋅) returns the class label of a target instance.
In the feature space, it is worth mentioning that the predicted visual features ˆXt of unseen classes are easily synthesized by embedding the semantic prototypes of Ct to the visual feature space with ˆXt=W2At. Hence, the ZSL is converted to a conventional classification problem. Empirically, any supervised classifier can be utilized. We simply exploit k-Nearest Neighbor (KNN) to demonstrate the capability of our decoder W2. Similar to the process in 1), the class label of target instance can be inferred by calculating the cosine distance between the prototype projections and the original visual feature xti:
l(xti)=argminj d(xti,ˆXtj), | (3.15) |
where ˆXtj is the j–th unseen class prototype projected into the visual feature space.
In this section, we firstly introduce our experimental protocols in detail, then we present our results that are compared with the state-of-the-art approaches on five small-scale benchmark datasets (AWA1, AWA2, CUB, SUN, aP & Y) for conventional zero-shot learning (CZSL) task.
Five benchmark datasets are selected from the widely used datasets for ZSL: AwA1 (Animals with Attributes 1) [27], AWA2 (Animals with Attributes 2) [75], aP & Y (Attribute Pascal and Yahoo) [76], CUB (Caltech-UCSD-Birds 200-2011) [22] and SUN (Scene UNderstanding) [77]. We exploit two typical protocols to evaluate the performance of our model: standard splits (SS) [27] and proposed splits (PS) [75]. More concretely, SS is widely used in previous works, but the weakness is that unseen classes are subset of ImageNet during training, resulting in violating the true zero-shot rule. On the contrary, PS ensures that none of the unseen classes used for pre-training the ResNet belong to 1K classes of ImageNet. The fact that PS is much more difficult than SS on account of low correlation between seen and unseen classes. For clarity, the statistics of these datasets are briefly reported in Table 2.
At Training Time | At Testing Time | |||||||||
Datasets | Granularity | Size | Attributes | Cs/Ct | Images | SS (Cs) | PS (Cs) | SS (Ct) | PS (Ct) | |
AWA1 | coarse | medium | 85 | 40/10 | 30,475 | 24,295 | 19,832 | 6180 | 10,643 | |
AWA2 | coarse | medium | 85 | 40/10 | 37,322 | 30,337 | 23,527 | 6985 | 13,795 | |
aP & Y | coarse | small | 64 | 20/12 | 15,339 | 12,695 | 5932 | 2644 | 9407 | |
CUB | fine | medium | 312 | 150/50 | 11,788 | 8855 | 7057 | 2933 | 4731 | |
SUN | fine | medium | 102 | 645/72 | 14,340 | 12,900 | 10,320 | 1440 | 4020 |
We take advantage of semantic space spanned by continuous attributes like the pioneering works [27,47,48]. Each instance is associated with the corresponding continuous class-level attribute. The dimension of the semantic space equals to that of the attributes, e.g., the semantic space of AWA1 is formed by 85-dim attributes. The dimensions of the attributes of all datasets are listed in Table 2.
Following the general procedure in other literatures, we use visual features extracted by deep convolutional neural networks (CNNs) and GoogleNet features [78] which is the 1024-dim activation of the final pooling layer as in [41]. Furthermore, the latest works adopt the pre-trained 2048-dim ResNet features, which are extracted by 2048-dim top layer pooling units of the 101-layered ResNet, to achieve improved performance [75]. It is worth noting that the ResNet features has two protocols, namely, SS and PS. For GoogleNet features, only SS is provided. For fair comparison, we do not perform any image pre-processing or any other data augmentation techniques, and conduct extensive experiments on the above two types of features.
To demonstrate the capability of BSAE, we evaluate on the conventional ZSL (CZSL) setting: Assume that the search space is restricted to the unseen classes, the goal is to predict the class labels of Ct at test stage. We use SS and PS protocols in this setting.
Our BSAE model has two hyper-parameters: λ1 and λ2 (see Eq 3.3). We select λ1 and λ2 from {10−2,10−1,1,10,102,103,104} for cross-validation. Considering the two split protocols, we propose tuning these parameters in different ways. For SS protocol, the parameters are chosen by means of class-wise cross-validation on Cs as in [67], that is, two seen classes are randomly selected form a validation set in each iteration to choose the best hyper-parameter {λ1, λ2} and use them for testing on unseen classes. For PS protocol, we perform hyper-parameter search on a disjoint set of validation set of 13 (AWA1/AWA2), 5 (AP & Y), 50 (CUB) and 65 (SUN) classes respectively [75]. Note that we report the average performance for ensuring the significance of the results.
Most ZSL methods use Top-1 accuracy (e.g., [48]) averaged for all images, where the prediction is correct for the predicted class is coincide with ground-truth. However we are concentrated on high performance of both densely and sparsely populated classes. Therefore, under CZSL setting, we evaluate our method on the benchmark datasets by using per-class top-1 accuracy proposed in [75]. We compute the top-1 accuracy independently for each class, and then average for all unseen classes:
accCt=1μμ∑c=1#correct predictions in c#instances in c. | (4.1) |
We evaluate our proposed framework for zero-shot learning on several benchmark datasets. The competitors are representative, competitive state-of-the-art and recently published that encompass a wide range in zero-shot learning.
In these experiments, the test instances only come from Ct disjoint with the seen classes Cs. We use both SS and PS protocols for more convincing results and the qualitative results are shown in Table 3 and Table 4.
Method | Feature | AWA1 | AWA2 | SUN | CUB | aP & Y | Average |
DAP [27] | R | 57.1 | 58.7 | 38.9 | 37.5 | 35.2 | 45.5 |
IAP [27] | R | 48.1 | 46.9 | 17.4 | 27.1 | 22.4 | 32.4 |
CONSE [37] | R | 63.6 | 67.9 | 44.2 | 36.7 | 25.9 | 47.7 |
DEVISE [38] | R | 72.9 | 68.6 | 57.5 | 53.2 | 35.4 | 57.5 |
CMT [39] | R | 58.9 | 66.3 | 41.9 | 37.3 | 26.9 | 46.3 |
SSE [67] | R | 68.8 | 67.5 | 54.5 | 43.7 | 31.1 | 53.1 |
SJE [41] | R | 76.7 | 69.5 | 57.1 | 55.3 | 32 | 58.1 |
ESZSL [47] | R | 74.7 | 75.6 | 57.3 | 55.1 | 34.4 | 59.4 |
LATEM [42] | R | 74.8 | 68.7 | 56.9 | 49.4 | 34.5 | 56.9 |
ALE [58] | R | 78.6 | 80.3 | 59.1 | 53.2 | 30.9 | 60.4 |
SYNC [49] | R | 72.2 | 71.2 | 59.1 | 54.1 | 39.7 | 59.3 |
G | 72.9 | - | 62.7 | 54.7 | - | - | |
SAE [48] | R | 80.6 | 80.7 | 42.4 | 33.4 | 8.3 | 49.1 |
G | 81.9 | - | 59.7 | 53.6 | 34.5 | - | |
SSZSL [79] | V | 88.6 | - | - | 58.8 | 49.9 | - |
DSRL [57] | V | 87.2 | - | - | 57.1 | 56.3 | - |
STZSL [80] | V | 83.7 | - | - | 58.7 | 54.4 | - |
GFZSL [62] | R | 80.5 | 79.3 | 62.9 | 53 | 51.3 | 65.4 |
TSTD [59] | V | 90.3 | - | - | 58.2 | - | - |
QFSL [61] | R | - | 84.8 | 61.7 | 69.7 | - | - |
VCL [56] | R | 82.0 | 82.5 | 63.8 | 60.1 | - | - |
DEARF [60] | R | 81 | 81.2 | 64.3 | 56.1 | - | - |
BSAE | R | 90.7 | 85.2 | 60.4 | 61.8 | 63.2 | 72.3 |
G | 91 | - | 67.1 | 62.4 | 57.7 | - |
Method | AWA1 | AWA2 | CUB | SUN | aP & Y | Average |
DAP [27] | 44.1 | 46.1 | 40 | 39.9 | 33.8 | 40.8 |
IAP [27] | 35.9 | 35.9 | 24 | 19.4 | 36.6 | 30.4 |
CONSE [37] | 45.6 | 44.5 | 34.3 | 38.8 | 26.9 | 38 |
DEVISE [38] | 54.2 | 59.7 | 52 | 56.5 | 39.8 | 52.4 |
CMT [39] | 39.5 | 37.9 | 34.6 | 33.9 | 28 | 34.8 |
SSE [67] | 60.1 | 61 | 43.9 | 51.5 | 34 | 50.1 |
SJE [41] | 65.6 | 61.9 | 53.9 | 53.7 | 32.9 | 53.6 |
ESZSL [47] | 58.2 | 58.6 | 53.9 | 54.5 | 38.3 | 52.7 |
LATEM [42] | 55.1 | 55.8 | 49.3 | 55.3 | 35.2 | 50.1 |
ALE [58] | 59.9 | 62.5 | 54.9 | 58.1 | 39.7 | 55 |
SYNC [49] | 54 | 46.6 | 55.6 | 56.3 | 23.9 | 47.3 |
SAE [48] | 53 | 54.1 | 33.3 | 40.3 | 8.3 | 37.8 |
DEM [70] | 68.4 | 67.1 | 51.7 | 61.9 | 35 | 56.8 |
GFZSL [62] | 68.3 | 63.8 | 49.3 | 60.6 | 38.4 | 56.1 |
CAVE [55] | 71.4 | 65.8 | 52.1 | 61.7 | - | - |
PSR [81] | - | 63.8 | 56 | 61.4 | 38.4 | - |
TVN [82] | - | 68.8 | 58.1 | 60.7 | - | - |
GAZSL [54] | - | 68.4 | 55.8 | 61.3 | 41.1 | - |
f-CLSWGAN [50] | - | 68.8 | 57.3 | 60.8 | 40.5 | - |
GDAN [51] | - | 67.7 | 51 | 54.8 | 40.4 | - |
LESAE [66] | 66.1 | 68.4 | 53.9 | 60 | 40.8 | 57.8 |
LisGAN [68] | 70.6 | - | 58.8 | 61.7 | 43.1 | - |
SRGAN [53] | 71.9 | - | 55.4 | 62.2 | - | - |
DEARF [60] | 72.1 | 69.3 | 38.5 | 48.6 | - | - |
BSR [52] | - | 68.4 | 57.7 | 61.2 | 41.3 | - |
BSAE | 72.3 | 69.4 | 59.7 | 58.6 | 53 | 62.6 |
For SS protocol, the SAE [48] is close to our model while lacks of the results of per-class top-1 accuracy with GoogleNet features. For fair comparison, we recreate SAE by following the settings in their original paper and exploit the same classifier to predict the class labels. Leveraging the code available online, we re-implement GFZSL [62] and SYNC [49] to obtain the recognition results. Note that the first 10 methods in Table 3 are cited from [75] and the rest are copied from the original paper. To further verify that our method is not only effective to specific visual features, we implement our model under the SS protocol with 1024-dim GoogleNet features (G) and 2048-dim ResNet features (R).
From comprehensive comparison in Table 3, we witness that: (1) our model achieves the best four of the five evaluations, i.e., AWA1, AWA2, SUN and aP & Y. Specifically, the improvements over the strongest competitor achieve 0.7%, 0.4% and 6.9% on AWA1, AWA2 and aP & Y. For fine-grained dataset SUN that contains more classes and relatively fewer instances per class, while our result of 67.1% is 2.8% higher than the strongest competitor [60]. The accuracy boost can be attributed to the combination of semantic representations and domain adaptation constraints significantly improving the ability for classification. (2) Meanwhile, from Figure 3, we can observe that the overlap between unseen classes of CUB, which is regarded as the well-known complicated dataset, is particularly striking. Moreover, it is hard to learn the visual-semantic projection for the reason that the sparsity of training instances (∼ 60 instances per class). However, our model still performs well on this dataset. It is worth pointing out that our model learns a more effective and stable visual-semantic relation from seen data for unseen data analysis. (3) [73] and [75] demonstrate that VggNet and ResNet features lead to improved results in ZSL than GoogleNet features. While our model using GoogleNet features consistently performs favorably against state-of-the-art on the five benchmarks, especially the best result of 91% on AWA1 and 67.1% on SUN. This provide further evidence that our model achieves good performance on coarse-grained and fine-grained datasets even if the features are not the strongest.
For PS protocol, we keep the same setting as in [75] to make sure the unseen classes at test time do not overlap with the 1K training classes of ImageNet. The first 12 reported results are cited from [75] and others copied from their original paper. Generally, the performance is expected degrade under this stricter settings. From the comparative results listed in Table 4, we can make the observation that the average top-1 per-class accuracy of our model performs 4.8% higher than all others and drops least on coarse-to-fine grained datasets among all methods, which illustrates more significant. Due to the similar idea between auto-encoder and GAN-based model, we compare several representative methods, e.g., LisGAN [68] and SRGAN [53]. Our model outperforms these competitors on four of five datasets, while 3.6% less than SRGAN [53] under the challenging split of SUN. We conjecture that our regression model is over-fitted in terms of scarce training instances of each class.
It is noting that there is no single approach claims the best results on all datasets simultaneously [60]. The aforementioned improvements actually create new baselines in the area of ZSL, given that most of the compared models utilize more complicated nonlinear formulations and some of them combine complementary semantic spaces or even generate richer features for unseen classes. In contrast, we apply only one type of semantic space as well as computational fast linear projection functions, but gain a significant performance boost.
We aim to evaluation the effectiveness of the decoder of BSAE via the image retrieval task, which is defined as searching top matched images by taking the provided semantic prototypes of unseen classes as queries. The ratio of the number of accurately retrieved images to that of all retrieved images, namely precision, is regarded as the measurement. Table 5 reports 5 out of 50 classes in CUB and 5 out of 65 classes in SUN and depicts the qualitative results of our designed model with highest anterior and posterior scores for each unseen class. Specifically, each column is a category, with class name and precision are shown at the top. The first three rows in the middle are the top-3 correctly retrieved instances. The following three rows are the top-3 misclassified instances in each unseen class. Observing form the top correct images, BSAE reasonably captures discriminative visual information only using its semantic prototype. It suggests that the adaptation regularization helps make approxiamate inference of unseen instances. Meanwhile, taking the class in the second column as an example, Pomarine Jaeger and Rhinoceros Auklet are visually similar to Pacific Loon, the discriminative ability of the decoder is not enough to distinguish the visual appearances between them. Due to the strong visual similarity and only a few different attributes among the classes, we further notice that it is hardly recognize these classes without expert knowledge, even for humans.
Grasshopper Sparrow 50.5% | Pacific Loon 88.2% | Rhinoceros Auklet 86.8% | Western Grebe 85.7% | Least Auklet 84.8% | market outdoor 82.6% | recycling plant outdoor 37.8% | van interior 44.7% | subway station platform 54.8% | lecture room 54.5% |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
For straightforward illustration of BSAE in ZSL. We explore t-SNE visualization [83] to compare the visual features with genuine class label (left) and semantic representations with the predicted class labels (right) in Figure 4. Each color represents clustering in the same class and all the features are embedding into two dimensions using t-SNE. It suggests that our model captures the underlying global distribution in the semantic space and performs better on the dataset. It is worth that our model alleviates the hubness problem in the lower dimensional semantic space. Moreover, the instances of the same class are grouped into one cluster in Figure 4, which confirms that the discriminative semantic representations learned by our model are able to cluster visually similar instances. Therefore, our proposed model preserves the local information of target unseen classes that the closeness are kept in the projected semantic representations.
The ROC curve and AUC value depict the tradeoff between specificity (False Positive Rate) and sensitivity (True Positive Rate) as a metric of the performance of our proposed BSAE. Figure 5 shows the results of the ROC curve and AUC value on AWA1 under PS protocol. We can observe that the ROC curve of the 10 unseen classes are close to the top-left corner of the plot, even though using the simplest KNN classifier.
We observed that almost all methods performed worst on aP & Y compared to other datasets. In order to show our experimental results in a more fine-grained manner, we take the PS protocol of aP & Y dataset as an example, compared with the best competitor LisGAN [68]. Figure 6 shows the confusion matrix of LisGAN and our model i.e., BSAE on the 12 unseen classes. The value in the diagonal of the confusion matrix indicates the ratio of the correctly predicted of each class. The darker color represents the higher class-wise accuracy. It can be seen that BSAE generally performs better on the most classes. Concretely, we boosts 38%, 10%, 24%, 49%, 41%, 3% and 3% on "horse", "motorbike", "person", "sheep", "goat", "jetski" and "statue" against LisGAN respectively. Although the GAN model directly handle zero-shot problem by converting it to a supervised task, we find that our model perform better than GAN-based model, i.e., LisGAN. In addition, it is common that one model does not have the highest accuracy on each unseen class. There will be great improvement in the future.
We measure the inter-class and intra-class distances to investigate BSAE can alleviate domain shift and hubness problem. We follow the two measurements provided by [84]:
Dcintra=1nc∑iD(φ(Ac),ψ(sci)), | (4.2) |
Dcinter=1C−1∑j≠cD(φ(Ac),φ(Aj)). | (4.3) |
Where nc represents the data size of the cth class. C means the number of the classes. φ(⋅) and ψ(⋅) denote the two dimensional outputs of t-SNE [83]. D(⋅) is the cosine distance, which reflects the degree of similarity between actual and the compared one [81]. Dcintra stands for the mean distance between the cth class prototype Ac and semantic representations of instances in that class. Dcinter stands for the mean distance between the cth class prototype Ac and all other classes. We compare our proposed model with TSTD [59], which is the best competitor under SS protocol on AWA1 dataset. For consideration of fairness, we re-implement the experiments of the two methods under the same settings in AWA1, i.e., use ResNet-101 features and take continuous class-level attributes as semantic representations. Different from TSTD [59] that applies the attributes of unseen classes during training to improve the performance, BSAE obtains smaller intra-class distances and larger inter-class distances with a large margin as illustrated in Table 6. Thus, BSAE is capable of alleviate domain shift problem as well as hubness problem in lower dimensional semantic space.
Class Name | Dcintra | Dcinter | |||
BSAE | TSTD | BSAE | TSTD | ||
chimpanzee | 0.275 | 0.444 | 1.663 | 1.361 | |
giant+panda | 0.441 | 0.442 | 1.583 | 1.264 | |
leopard | 0.36 | 0.39 | 1.69 | 1.246 | |
persian+cat | 0.077 | 0.388 | 1.887 | 1.501 | |
pig | 0.199 | 0.43 | 1.868 | 1.399 | |
hippopotamus | 0.412 | 0.498 | 1.594 | 1.337 | |
humpback+whale | 0.318 | 0.35 | 1.28 | 1.257 | |
raccoon | 0.224 | 0.333 | 1.326 | 1.192 | |
rat | 0.215 | 0.28 | 1.28 | 1.234 | |
seal | 0.257 | 0.228 | 1.31 | 1.243 |
In this section, we analyze the complexity and convergence of BSAE. It is remarkable that the operation of Algorithm 1 mostly comes from matrix multiplication. Obviously, it can accelerate the training process greatly. Additionally, we set 200 as the maximum iterations. The F-norm of the parameter variation with respect to the iteration on fine-grained datasets are reported.
From Figure 7a, it is notable that our model reaches 80% of the accuracy within 4 iterations and is close to the highest accuracy around 10 iterations on coarse-grained dataset, e.g., AWA1 and around 20 iterations on fine-grained datasets. These demonstrate that our algorithm has a good practical application for its low complexity and good performance.
Figure 7b and 7c shows that the algorithm converges within 40 iterations. It is obvious that the decoder W2 of the target domain is well restricted by the decoder W1 of the source domain, which verifies the significance of the adaptation regularization. Moreover, these observations finally support the theoretical analysis of complexity and convergence in Section 3.3.
To provide further insights into the role of the two regularization terms: ‖W0Xs−Str‖2F and ‖W2−W1‖2F in our proposed objective function in helping the model to achieve better performance, we simplify our full model BSAE with various stripped-down versions of the model on the PS protocol of CZSL. Specially, for λ1=0, when the similarity constraint of he predicted and actual semantic representations are not exploited to encoder, i.e, Eq 3.3 without ‖W0Xs−Str‖2F term (denoted BSAE-SR), BSAE degrades to only contain adaptation regularization. The encoder does not ensure that the learned semantic representations of each instance is close to its class prototype. For λ2=0, i.e., BSAE without the adaptation regularization ‖W2−W1‖2F (denoted BSAE-DR), the decoder in the target domain is not restricted to derive from the decoder in the source domain, which is supervised by the semantic prototypes of the source instances. Figure 8 shows clearly that the two terms contribute to the superior performance of proposed model. We achieved up to around 10% improvements on five datasets. It is reasonable to believe that the learning of semantics will help the learning of domain adaptation among seen and unseen classes.
In this paper, we have proposed a novel model called Bi-shifting Auto-Encoder to perform efficient zero-shot recognition in semantic and visual space by taking advantage of autoencoder network. Our model learns the generalizable and computationally fast projection functions in transductive settings, which leverages the labeled source data and the visual features of the unlabeled target data. In particular, to improve the discriminability of the semantic embeddings, the encoder is constrained by aligning the semantic representations of the labeled source instances with their corresponding prototypes of the seen classes. Furthermore, to guarantee the generalizability of the projected semantic representations, two different decoders reconstruct the visual features of the instances in source and target domain simultaneously with the adaptation regularization. Thus, our model recovers the interactions between visual features and semantics, and is able to alleviate the projection shift problem. Extensive experiments are conducted on five benchmark datasets and comparative evaluations demonstrate that our model yields superior performance on zero-shot learning. The major limitation of our model lies in the fact that each class is represented by one attribute prototype in the semantic space, which is insufficient to completely characterize the features of the class, resulting in the semantics of the instances may be misplaced from the class prototype. Therefore, our research work will put effort in exploring different types of semantic representations to investigate the relationships between classes, especially the subtle differences among the classes of fine-grained datasets. An additional limitation of this study is that the full set of unlabeled target instances are utilized, ignoring their distinctive effects on the model learning. A natural processing of this work is to explore the most useful unseen instances that facilitate the zero-shot classification.
This work was supported by the Fundamental Research Funds for the Central Universities.
The authors declare there is no conflicts of interest.
[1] |
Bjørklund G, Chartrand MS, Aaseth J (2017) Manganese exposure and neurotoxic effects in children. Environ Res 155: 380-384. https://doi.org/10.1016/j.envres.2017.03.003 ![]() |
[2] | Mason LH, Harp JP, Han DY (2014) Pb neurotoxicity: neuropsychological effects of lead toxicity. Biomed Res Int 2014: 840547. https://doi.org/10.1155/2014/840547 |
[3] |
Florea AM, Büsselberg D (2006) Occurrence, use and potential toxic effects of metals and metal compounds. Biometals 19: 419-427. https://doi.org/10.1007/s10534-005-4451-x ![]() |
[4] |
Jaishankar M, Tseten T, Anbalagan N, et al. (2014) Toxicity, mechanism and health effects of some heavy metals. Interdiscip Toxicol 7: 60-72. https://doi.org/10.2478/intox-2014-0009 ![]() |
[5] |
Aherrera A, Olmedo P, Grau-Perez M, et al. (2017) The association of e-cigarette use with exposure to nickel and chromium: A preliminary study of non-invasive biomarkers. Environ Res 159: 313-320. https://doi.org/10.1016/j.envres.2017.08.014 ![]() |
[6] |
Olmedo P, Goessler W, Tanda S, et al. (2018) Metal Concentrations in e-Cigarette Liquid and Aerosol Samples: The Contribution of Metallic Coils. Environ Health Perspect 126: 027010. https://doi.org/10.1289/EHP2175 ![]() |
[7] |
Zhao D, Navas-Acien A, Ilievski V, et al. (2019) Metal concentrations in electronic cigarette aerosol: Effect of open-system and closed-system devices and power settings. Environ Res 174: 125-134. https://doi.org/10.1016/j.envres.2019.04.003 ![]() |
[8] |
Gade M, Comfort N, Re DB (2021) Sex-specific neurotoxic effects of heavy metal pollutants: Epidemiological, experimental evidence and candidate mechanisms. Environ Res 201: 111558. https://doi.org/10.1016/j.envres.2021.111558 ![]() |
[9] |
Zheng W, Aschner M, Ghersi-Egea JF (2003) Brain barrier systems: a new frontier in metal neurotoxicological research. Toxicol Appl Pharmacol 192: 1-11. https://doi.org/10.1016/S0041-008X(03)00251-5 ![]() |
[10] |
Singh G, Singh V, Sobolewski M, et al. (2018) Sex-Dependent Effects of Developmental Lead Exposure on the Brain. Front Genet 9: 89. https://doi.org/10.3389/fgene.2018.00089 ![]() |
[11] |
Lidsky TI, Schneider JS (2003) Lead neurotoxicity in children: basic mechanisms and clinical correlates. Brain 126: 5-19. https://doi.org/10.1093/brain/awg014 ![]() |
[12] |
Aliomrani M, Sahraian MA, Shirkhanloo H, et al. (2017) Correlation between heavy metal exposure and GSTM1 polymorphism in Iranian multiple sclerosis patients. Neurol Sci 38: 1271-1278. https://doi.org/10.1007/s10072-017-2934-5 ![]() |
[13] |
Tyler CR, Allan AM (2014) The Effects of Arsenic Exposure on Neurological and Cognitive Dysfunction in Human and Rodent Studies: A Review. Curr Environ Health Rep 1: 132-147. https://doi.org/10.1007/s40572-014-0012-1 ![]() |
[14] |
Amin-Zaki L, Elhassani S, Majeed MA, et al. (1974) Intra-uterine methylmercury poisoning in Iraq. Pediatrics 54: 587-595. https://doi.org/10.1542/peds.54.5.587 ![]() |
[15] | Fløtre CH, Varsi K, Helm T, et al. Predictors of mercury, lead, cadmium and antimony status in Norwegian never-pregnant women of fertile age (2017)12: e0189169. https://doi.org/10.1371/journal.pone.0189169 |
[16] |
Zheng G, Zhong H, Guo Z, et al. (2014) Levels of heavy metals and trace elements in umbilical cord blood and the risk of adverse pregnancy outcomes: a population-based study. Biol Trace Elem Res 160: 437-444. https://doi.org/10.1007/s12011-014-0057-x ![]() |
[17] |
Okoye EA, Bocca B, Ruggieri F, et al. (2021) Metal pollution of soil, plants, feed and food in the Niger Delta, Nigeria: Health risk assessment through meat and fish consumption. Environ Res 198: 111273. https://doi.org/10.1016/j.envres.2021.111273 ![]() |
[18] |
Okoye EA, Bocca B, Ruggieri F, et al. (2022) Arsenic and toxic metals in meat and fish consumed in Niger delta, Nigeria: Employing the margin of exposure approach in human health risk assessment. Food Chem Toxicol 159: 112767. https://doi.org/10.1016/j.fct.2021.112767 ![]() |
[19] |
Abbaoui A, Chatoui H, El Hiba O, et al. (2017) Neuroprotective effect of curcumin-I in copper-induced dopaminergic neurotoxicity in rats: A possible link with Parkinson's disease. Neurosci Lett 660: 103-108. https://doi.org/10.1016/j.neulet.2017.09.032 ![]() |
[20] |
Albarracin SL, Stab B, Casas Z, et al. (2012) Effects of natural antioxidants in neurodegenerative disease. Nutr Neurosci 15: 1-9. https://doi.org/10.1179/1476830511Y.0000000028 ![]() |
[21] |
Moosmann B, Behl C (2000) Dietary phenols: antioxidants for the brain?. Nutr Neurosci 3: 1-10. https://doi.org/10.1080/1028415X.2000.11747298 ![]() |
[22] | Li H-w, Lan T-j, Yun C-x, et al. (2020) Mangiferin exerts neuroprotective activity against lead-induced toxicity and oxidative stress via Nrf2 pathway. Chin Herb Med 12: 36-46. https://doi.org/10.1016/j.chmed.2019.12.002 |
[23] |
Falade KO, Akeem SA (2020) Physicochemical properties, protein digestibility and thermal stability of processed African mesquite bean (Prosopis africana) flours and protein isolates. J Food Meas Charact 14: 1481-1496. https://doi.org/10.1007/s11694-020-00398-0 ![]() |
[24] | Aremu M, Awala E, Opaluwa O, et al. (2015) Effect of processing on nutritional composition of African locust bean (Parkia biglobosa) and mesquite bean (Prosopis africana) seeds. Commun Appl Sci 3. |
[25] |
Aremu M, Olonisakin A, Atolaye B, et al. (2007) Some nutritional composition and functional properties of Prosopis africana. Bangladesh J Sci Ind Res 42: 269-280. https://doi.org/10.3329/bjsir.v42i3.665 ![]() |
[26] | Keay R (1989) Trees of Nigeria. Clarendon. Oxford: Oxford University) Press. |
[27] |
Ozoani H, Ezejiofor AN, Okolo KO, et al. (2024) Ameliorative Effects of Zn and Se Supplementation on Heavy Metal Mixture Burden via Increased Renal Metal Excretion and Restoration of Redoxo-Inflammatory Alterations. Biol Trace Elem Res 202: 643-658. https://doi.org/10.1007/s12011-023-03709-w ![]() |
[28] |
Murakami A (2022) Novel mechanisms underlying bioactivities of polyphenols via hormesis. Curr Opinion Toxicol 30: 100337. https://doi.org/10.1016/j.cotox.2022.02.010 ![]() |
[29] |
Murakami A (2024) Impact of hormesis to deepen our understanding of the mechanisms underlying the bioactivities of polyphenols. Curr Opinion Biotechnol 86: 103074. https://doi.org/10.1016/j.copbio.2024.103074 ![]() |
[30] |
Hannan MA, Dash R, Sohag AAM, et al. (2020) Neuroprotection Against Oxidative Stress: Phytochemicals Targeting TrkB Signaling and the Nrf2-ARE Antioxidant System. Front Mol Neurosci 13: 116. https://doi.org/10.3389/fnmol.2020.00116 ![]() |
[31] |
Ezike AC, Akah PA, Okoli CO, et al. (2010) Medicinal Plants Used in Wound Care: A Study of Prosopis africana (Fabaceae) Stem Bark. Indian J Pharm Sci 72: 334-339. https://doi.org/10.4103/0250-474X.70479 ![]() |
[32] |
Anyanwu BO, Orish CN, Ezejiofor AN, et al. (2020) Neuroprotective effect of Costus afer on low dose heavy metal mixture (lead, cadmium and mercury) induced neurotoxicity via antioxidant, anti-inflammatory activities. Toxicol Rep 7: 1032-1038. https://doi.org/10.1016/j.toxrep.2020.08.008 ![]() |
[33] |
Messarah M, Klibet F, Boumendjel A, et al. (2012) Hepatoprotective role and antioxidant capacity of selenium on arsenic-induced liver injury in rats. Exp Toxicol Pathol 64: 167-174. https://doi.org/10.1016/j.etp.2010.08.002 ![]() |
[34] |
Tarantino LM, Gould TJ, Druhan JP, et al. (2000) Behavior and mutagenesis screens: the importance of baseline analysis of inbred strains. Mamm Genome 11: 555-564. https://doi.org/10.1007/s003350010107 ![]() |
[35] |
Popović N, Madrid JA, Rol MÁ, et al. (2010) Barnes maze performance of Octodon degus is gender dependent. Behav Brain Res 212: 159-167. https://doi.org/10.1016/j.bbr.2010.04.005 ![]() |
[36] |
Popović N, Baño-Otalora B, Rol MÁ, et al. (2023) Effects of long-term individual housing of middle-aged female Octodon degus on spatial learning and memory in the Barnes maze task. Front Behav Neurosci 17: 1221090. https://doi.org/10.3389/fnbeh.2023.1221090 ![]() |
[37] |
van den Berg R, Laman JD, van Meurs M, et al. (2016) Rotarod motor performance and advanced spinal cord lesion image analysis refine assessment of neurodegeneration in experimental autoimmune encephalomyelitis. J Neurosci Meth 262: 66-76. https://doi.org/10.1016/j.jneumeth.2016.01.013 ![]() |
[38] | Pritchett K, Mulder GB (2003) The rotarod. J Am Assoc Lab Anim 42: 49-49. |
[39] |
Eddie-Amadi BF, Ezejiofor AN, Orish CN, et al. (2023) Zn and Se abrogate heavy metal mixture induced ovarian and thyroid oxido-inflammatory effects mediated by activation of NRF2-HMOX-1 in female albino rats. Curr Res Toxicol 4: 100098. https://doi.org/10.1016/j.crtox.2022.100098 ![]() |
[40] |
Eddie-Amadi BF, Ezejiofor AN, Orish CN, et al. (2022) Zinc and selenium mitigated heavy metals mixture (Pb, Al, Hg and Mn) mediated hepatic-nephropathy via modulation of oxido-inflammatory status and NF‑kB signaling in female albino rats. Toxicology 481: 153350. https://doi.org/10.1016/j.tox.2022.153350 ![]() |
[41] | Paglia DE, Valentine WN (1967) Studies on the quantitative and qualitative characterization of erythrocyte glutathione peroxidase. J Lab Clin Med 70: 158-169. |
[42] |
Jollow D, Mitchell J, Zampaglione Na, et al. (1974) Bromobenzene-induced liver necrosis. Protective role of glutathione and evidence for 3, 4-bromobenzene oxide as the hepatotoxic metabolite. Pharmacology 11: 151-169. https://doi.org/10.1159/000136485 ![]() |
[43] |
Wh H (1974) Glutathione S-transferase. J Biol Chem 249: 7130-7139. https://doi.org/10.1016/S0021-9258(19)42083-8 ![]() |
[44] |
Marklund S, Marklund G (1974) Involvement of the superoxide anion radical in the autoxidation of pyrogallol and a convenient assay for superoxide dismutase. Eur J Biochem 47: 469-474. https://doi.org/10.1111/j.1432-1033.1974.tb03714.x ![]() |
[45] |
Bergmeyer HU, Bernt E (1974) UV-assay with pyruvate and NADH. Methods of enzymatic analysis.Elsevier pp. 574-579. https://doi.org/10.1016/B978-0-12-091302-2.50010-4 ![]() |
[46] |
Esterbauer H, Cheeseman KH (1990) [42] Determination of aldehydic lipid peroxidation products: malonaldehyde and 4-hydroxynonenal. Methods in enzymology.Elsevier pp. 407-421. https://doi.org/10.1016/0076-6879(90)86134-H ![]() |
[47] |
Sosroseno W, Sugiatno E, Samsudin AR, et al. (2008) The role of nitric oxide on the proliferation of a human osteoblast cell line stimulated with hydroxyapatite. Jb Oral Implantol 34: 196-202. https://doi.org/10.1563/0.910.1 ![]() |
[48] |
Ikpeama EU, Orish CN, Ezejiofor AN, et al. (2023) Essential Trace Elements Prevent the Impairment in the Retention Memory, Cerebral Cortex, and Cerebellum Damage in Male Rats Exposed to Quaternary Metal Mixture by Up-regulation, of Heme Oxygynase-1 and Down-regulation of Nuclear Factor Erythroid 2-related Factor 2-NOs Signaling Pathways. Neuroscience 512: 70-84. https://doi.org/10.1016/j.neuroscience.2023.01.002 ![]() |
[49] |
Okoye EA, Ezejiofor AN, Nwaogazie IL, et al. (2022) Heavy metals and arsenic in soil and vegetation of Niger Delta, Nigeria: Ecological risk assessment. Case Studies in Chemical and Environmental Engineering 6: 100222. https://doi.org/10.1016/j.cscee.2022.100222 ![]() |
[50] |
Doungue HT, Kengne APN, Kuate D (2018) Neuroprotective effect and antioxidant activity of Passiflora edulis fruit flavonoid fraction, aqueous extract, and juice in aluminum chloride-induced Alzheimer's disease rats. Nutrire 43: 1-12. https://doi.org/10.1186/s41110-018-0082-1 ![]() |
[51] |
Trott O, Olson AJ (2010) AutoDock Vina: improving the speed and accuracy of docking with a new scoring function, efficient optimization, and multithreading. J Comput Chem 31: 455-461. https://doi.org/10.1002/jcc.21334 ![]() |
[52] | Ahmed MQ, Alenazi FS, Fazaludeen MF, et al. (2018) Pathology and Management of Alzheimer's disease: A review. Int J Pharm Res Alli 7. |
[53] | Colizzi C (2019) The protective effects of polyphenols on Alzheimer's disease: a systematic review. Alzh Dement-TRCI 5: 184-196. https://doi.org/10.1016/j.trci.2018.09.002 |
[54] |
Qu Z, Zhang J, Yang H, et al. (2016) Protective effect of tetrahydropalmatine against d-galactose induced memory impairment in rat. Physiol Behav 154: 114-125. https://doi.org/10.1016/j.physbeh.2015.11.016 ![]() |
[55] |
Mattson MP (2004) Pathways towards and away from Alzheimer's disease. Nature 430: 631-639. https://doi.org/10.1038/nature02621 ![]() |
[56] |
Afzal S, Abdul Manap AS, Attiq A, et al. (2023) From imbalance to impairment: the central role of reactive oxygen species in oxidative stress-induced disorders and therapeutic exploration. Front Pharmacol 14: 1269581. https://doi.org/10.3389/fphar.2023.1269581 ![]() |
[57] | Hu B, Ouyang Y, Zhao T, et al. (2024) Antioxidant Hydrogels: Antioxidant Mechanisms, Design Strategies, and Applications in the Treatment of Oxidative Stress-Related Diseases. Adv Healthc Mater 2303817. https://doi.org/10.1002/adhm.202303817 |
[58] | Oluwafemi R, Agubosi O, Alagbe J (2021) Proximate, minerals, vitamins and amino acid composition of Prosopis africana (African mesquite) seed oil. Asian J Adv Res 4: 1011-1017. |
[59] |
Alagbe JO, Agubosi OC, Oluwafemi RA (2023) Histopathology of broiler chickens fed diets supplemented with Prosopis africana (African mesquite) essential oil. Brazilian J Sci 2: 49-59. https://doi.org/10.14295/bjs.v2i9.385 ![]() |
[60] | Hu H-C, Lei Y-H, Zhang W-H, et al. (2022) Antioxidant and Anti-inflammatory Properties of Resveratrol in Diabetic Nephropathy: A Systematic Review and Meta-analysis of Animal Studies. Front Pharmacol 13. https://doi.org/10.3389/fphar.2022.841818 |
[61] |
Musial C, Kuban-Jankowska A, Gorska-Ponikowska M (2020) Beneficial Properties of Green Tea Catechins. Int J Mol Sci 21. https://doi.org/10.3390/ijms21051744 ![]() |
[62] |
Tvrda E, Straka P, Galbavy D, et al. (2019) Epicatechin Provides Antioxidant Protection to Bovine Spermatozoa Subjected to Induced Oxidative Stress. Molecules 24. https://doi.org/10.3390/molecules24183226 ![]() |
[63] |
Qu Z, Liu A, Li P, et al. (2021) Advances in physiological functions and mechanisms of (−)-epicatechin. Crit Rev Food Sci 61: 211-233. https://doi.org/10.1080/10408398.2020.1723057 ![]() |
[64] |
Auti ST, Kulkarni YA (2019) Neuroprotective Effect of Cardamom Oil Against Aluminum Induced Neurotoxicity in Rats. Front Neurol 10: 399. https://doi.org/10.3389/fneur.2019.00399 ![]() |
[65] |
El-Hawary SS, Sobeh M, Badr WK, et al. (2020) HPLC-PDA-MS/MS profiling of secondary metabolites from Opuntia ficus-indica cladode, peel and fruit pulp extracts and their antioxidant, neuroprotective effect in rats with aluminum chloride induced neurotoxicity. Saudi J Biol Sci 27: 2829-2838. https://doi.org/10.1016/j.sjbs.2020.07.003 ![]() |
[66] | Elsawi SA, Aly HF, Elbatanony MM, et al. (2018) Phytochemical evaluation of Lagerstroemia indica (L.) Pers leaves as anti-Alzheimer's. J Mater Environ Sci 9: 2575-2586. |
[67] |
Olennikov DN, Kashchenko NI, Chirikova NK, et al. (2017) Isorhamnetin and quercetin derivatives as anti-acetylcholinesterase principles of marigold (Calendula officinalis) flowers and preparations. Int J Mol Sci 18: 1685. https://doi.org/10.3390/ijms18081685 ![]() |
[68] |
Anwar HM, Georgy GS, Hamad SR, et al. (2021) A leaf extract of harrisonia abyssinica ameliorates neurobehavioral, histological and biochemical changes in the hippocampus of rats with aluminum chloride-induced alzheimer's disease. Antioxidants 10: 947. https://doi.org/10.3390/antiox10060947 ![]() |
[69] |
Taïr K, Kharoubi O, Taïr OA, et al. (2016) Aluminium-induced acute neurotoxicity in rats: Treatment with aqueous extract of Arthrophytum (Hammada scoparia). J Acute Dis 5: 470-482. https://doi.org/10.1016/j.joad.2016.08.028 ![]() |
[70] |
El-Hawary S, Abd El-Kader E, Rabeh M, et al. (2020) Eliciting callus culture for production of hepatoprotective flavonoids and phenolics from Sequoia sempervirens (D. Don Endl). Nat Prod Res 34: 3125-3129. https://doi.org/10.1080/14786419.2019.1607334 ![]() |
[71] |
Amri Z, Ghorbel A, Turki M, et al. (2017) Effect of pomegranate extracts on brain antioxidant markers and cholinesterase activity in high fat-high fructose diet induced obesity in rat model. BMC Complement Altern Med 17: 339. https://doi.org/10.1186/s12906-017-1842-9 ![]() |
[72] |
Kujawska M, Jourdes M (2020) Neuroprotective Effects of Pomegranate Juice against Parkinson's Disease and Presence of Ellagitannins-Derived Metabolite-Urolithin A-In the Brain. Int J Mol Sci 21: 202. https://doi.org/10.3390/ijms21010202 ![]() |
[73] |
Zhang H, Wei M, Lu X, et al. (2020) Aluminum trichloride caused hippocampal neural cells death and subsequent depression-like behavior in rats via the activation of IL-1β/JNK signaling pathway. Sci Total Environ 715: 136942. https://doi.org/10.1016/j.scitotenv.2020.136942 ![]() |
[74] |
Yaseen AA, Al-Okbi SY, Hussein AM, et al. (2019) Potential protection from Alzheimer's disease by wheat germ and rice bran nano-form in rat model. J Appl Pharm Sci 9: 067-076. https://doi.org/10.7324/JAPS.2019.S108 ![]() |
[75] |
Lu T-H, Tseng T-J, Su C-C, et al. (2014) Arsenic induces reactive oxygen species-caused neuronal cell apoptosis through JNK/ERK-mediated mitochondria-dependent and GRP 78/CHOP-regulated pathways. Toxicol Lett 224: 130-140. https://doi.org/10.1016/j.toxlet.2013.10.013 ![]() |
[76] |
Chakraborti D, Singh SK, Rahman MM, et al. (2018) Groundwater arsenic contamination in the Ganga River Basin: a future health danger. Int J Env Res Pub He 15: 180. https://doi.org/10.3390/ijerph15020180 ![]() |
[77] |
Firdaus F, Zafeer MF, Anis E, et al. (2018) Ellagic acid attenuates arsenic induced neuro-inflammation and mitochondrial dysfunction associated apoptosis. Toxicol Rep 5: 411-417. https://doi.org/10.1016/j.toxrep.2018.02.017 ![]() |
[78] | Jahan-Abad AJ, Morteza-Zadeh P, Negah SS, et al. (2017) Curcumin attenuates harmful effects of arsenic on neural stem/progenitor cells. Avicenna J Phytomedi 7: 376. |
[79] |
Essa AF, Teleb M, El-Kersh DM, et al. (2023) Natural acylated flavonoids: Their chemistry and biological merits in context to molecular docking studies. Phytochem Rev 22: 1469-1508. https://doi.org/10.1007/s11101-022-09840-1 ![]() |
[80] |
Naoi M, Inaba-Hasegawa K, Shamoto-Nagai M, et al. (2017) Neurotrophic function of phytochemicals for neuroprotection in aging and neurodegenerative disorders: modulation of intracellular signaling and gene expression. J Neural Transm 124: 1515-1527. https://doi.org/10.1007/s00702-017-1797-5 ![]() |
[81] |
Hannan MA, Sohag AAM, Dash R, et al. (2020) Phytosterols of marine algae: Insights into the potential health benefits and molecular pharmacology. Phytomedicine 69: 153201. https://doi.org/10.1016/j.phymed.2020.153201 ![]() |
[82] |
Gao Y, Xu X, Chang S, et al. (2015) Totarol prevents neuronal injury in vitro and ameliorates brain ischemic stroke: Potential roles of Akt activation and HO-1 induction. Toxicol Appl Pharmacol 289: 142-154. https://doi.org/10.1016/j.taap.2015.10.001 ![]() |
[83] |
Fang J, Wang H, Zhou J, et al. (2018) Baicalin provides neuroprotection in traumatic brain injury mice model through Akt/Nrf2 pathway. Drug Des Devel Ther 12: 2497-2508. https://doi.org/10.2147/DDDT.S163951 ![]() |
[84] |
Cui HY, Zhang XJ, Yang Y, et al. (2018) Rosmarinic acid elicits neuroprotection in ischemic stroke via Nrf2 and heme oxygenase 1 signaling. Neural Regen Res 13: 2119-2128. https://doi.org/10.4103/1673-5374.241463 ![]() |
[85] |
Hui Y, Chengyong T, Cheng L, et al. (2018) Resveratrol Attenuates the Cytotoxicity Induced by Amyloid-β(1-42) in PC12 Cells by Upregulating Heme Oxygenase-1 via the PI3K/Akt/Nrf2 Pathway. Neurochem Res 43: 297-305. https://doi.org/10.1007/s11064-017-2421-7 ![]() |
[86] |
Dinkova-Kostova AT, Kostov RV, Kazantsev AG (2018) The role of Nrf2 signaling in counteracting neurodegenerative diseases. Febs J 285: 3576-3590. https://doi.org/10.1111/febs.14379 ![]() |
[87] |
Joshi G, Johnson JA (2012) The Nrf2-ARE pathway: a valuable therapeutic target for the treatment of neurodegenerative diseases. Recent Pat CNS Drug Discov 7: 218-229. https://doi.org/10.2174/157488912803252023 ![]() |
[88] |
Cuadrado A, Rojo AI (2008) Heme oxygenase-1 as a therapeutic target in neurodegenerative diseases and brain infections. Curr Pharm Des 14: 429-442. https://doi.org/10.2174/138161208783597407 ![]() |
[89] |
Schipper HM (2004) Heme oxygenase expression in human central nervous system disorders. Free Radical Bio Med 37: 1995-2011. https://doi.org/10.1016/j.freeradbiomed.2004.09.015 ![]() |
1. | Muhammad Bilal Khan, Muhammad Aslam Noor, Nehad Ali Shah, Khadijah M. Abualnaja, Thongchai Botmart, Some New Versions of Hermite–Hadamard Integral Inequalities in Fuzzy Fractional Calculus for Generalized Pre-Invex Functions via Fuzzy-Interval-Valued Settings, 2022, 6, 2504-3110, 83, 10.3390/fractalfract6020083 | |
2. | Soubhagya Kumar Sahoo, Hijaz Ahmad, Muhammad Tariq, Bibhakar Kodamasingh, Hassen Aydi, Manuel De la Sen, Hermite–Hadamard Type Inequalities Involving k-Fractional Operator for (h¯,m)-Convex Functions, 2021, 13, 2073-8994, 1686, 10.3390/sym13091686 | |
3. | Hongling Zhou, Muhammad Shoaib Saleem, Waqas Nazeer, Ahsan Fareed Shah, Hermite-Hadamard type inequalities for interval-valued exponential type pre-invex functions via Riemann-Liouville fractional integrals, 2022, 7, 2473-6988, 2602, 10.3934/math.2022146 | |
4. | Tareq Saeed, Muhammad Bilal Khan, Savin Treanțǎ, Hamed H. Alsulami, Mohammed Sh. Alhodaly, Interval Fejér-Type Inequalities for Left and Right-λ-Preinvex Functions in Interval-Valued Settings, 2022, 11, 2075-1680, 368, 10.3390/axioms11080368 |
Notations | Descriptions |
Xs∈Rd×m | Visual feature of source instances |
Xt∈Rd×p | Visual feature of target instances |
Ys={ysi}mi=1 | Class labels of Xs |
Yt={yti}pi=1 | Class labels of Xt |
Cs={1,2,…,τ} | Set of τ seen classes |
Ct={1,2,…,μ} | Set of μ unseen classes |
At∈Rk×μ | Semantic prototypes of Ct |
Str∈Rk×m | Semantic representations of Xs |
Ste∈Rk×p | Semantic representations of Xt |
λ1,λ2 | Hyper-parameters |
X | d-dimensional visual feature space |
A | k-dimensional semantic space |
m,p | Number of source and target instances |
d,k | Dimensionality of visual feature and semantic space |
Algorithm 1 Bi-shifting Semantic Auto-Encoder |
Input: Training data Xs, Str Test data Xt Hyper-parameters λ1, λ2 Output: Projection matrices W0, W2 1: Initialize W2 2: while not converge do 3: Update W0 by Eq 3.9 4: Update W1 by Eq 3.11 5: Update W2 by Eq 3.13 6: Check the converge condition 7: end while 8: Return W0, W2 |
At Training Time | At Testing Time | |||||||||
Datasets | Granularity | Size | Attributes | Cs/Ct | Images | SS (Cs) | PS (Cs) | SS (Ct) | PS (Ct) | |
AWA1 | coarse | medium | 85 | 40/10 | 30,475 | 24,295 | 19,832 | 6180 | 10,643 | |
AWA2 | coarse | medium | 85 | 40/10 | 37,322 | 30,337 | 23,527 | 6985 | 13,795 | |
aP & Y | coarse | small | 64 | 20/12 | 15,339 | 12,695 | 5932 | 2644 | 9407 | |
CUB | fine | medium | 312 | 150/50 | 11,788 | 8855 | 7057 | 2933 | 4731 | |
SUN | fine | medium | 102 | 645/72 | 14,340 | 12,900 | 10,320 | 1440 | 4020 |
Method | Feature | AWA1 | AWA2 | SUN | CUB | aP & Y | Average |
DAP [27] | R | 57.1 | 58.7 | 38.9 | 37.5 | 35.2 | 45.5 |
IAP [27] | R | 48.1 | 46.9 | 17.4 | 27.1 | 22.4 | 32.4 |
CONSE [37] | R | 63.6 | 67.9 | 44.2 | 36.7 | 25.9 | 47.7 |
DEVISE [38] | R | 72.9 | 68.6 | 57.5 | 53.2 | 35.4 | 57.5 |
CMT [39] | R | 58.9 | 66.3 | 41.9 | 37.3 | 26.9 | 46.3 |
SSE [67] | R | 68.8 | 67.5 | 54.5 | 43.7 | 31.1 | 53.1 |
SJE [41] | R | 76.7 | 69.5 | 57.1 | 55.3 | 32 | 58.1 |
ESZSL [47] | R | 74.7 | 75.6 | 57.3 | 55.1 | 34.4 | 59.4 |
LATEM [42] | R | 74.8 | 68.7 | 56.9 | 49.4 | 34.5 | 56.9 |
ALE [58] | R | 78.6 | 80.3 | 59.1 | 53.2 | 30.9 | 60.4 |
SYNC [49] | R | 72.2 | 71.2 | 59.1 | 54.1 | 39.7 | 59.3 |
G | 72.9 | - | 62.7 | 54.7 | - | - | |
SAE [48] | R | 80.6 | 80.7 | 42.4 | 33.4 | 8.3 | 49.1 |
G | 81.9 | - | 59.7 | 53.6 | 34.5 | - | |
SSZSL [79] | V | 88.6 | - | - | 58.8 | 49.9 | - |
DSRL [57] | V | 87.2 | - | - | 57.1 | 56.3 | - |
STZSL [80] | V | 83.7 | - | - | 58.7 | 54.4 | - |
GFZSL [62] | R | 80.5 | 79.3 | 62.9 | 53 | 51.3 | 65.4 |
TSTD [59] | V | 90.3 | - | - | 58.2 | - | - |
QFSL [61] | R | - | 84.8 | 61.7 | 69.7 | - | - |
VCL [56] | R | 82.0 | 82.5 | 63.8 | 60.1 | - | - |
DEARF [60] | R | 81 | 81.2 | 64.3 | 56.1 | - | - |
BSAE | R | 90.7 | 85.2 | 60.4 | 61.8 | 63.2 | 72.3 |
G | 91 | - | 67.1 | 62.4 | 57.7 | - |
Method | AWA1 | AWA2 | CUB | SUN | aP & Y | Average |
DAP [27] | 44.1 | 46.1 | 40 | 39.9 | 33.8 | 40.8 |
IAP [27] | 35.9 | 35.9 | 24 | 19.4 | 36.6 | 30.4 |
CONSE [37] | 45.6 | 44.5 | 34.3 | 38.8 | 26.9 | 38 |
DEVISE [38] | 54.2 | 59.7 | 52 | 56.5 | 39.8 | 52.4 |
CMT [39] | 39.5 | 37.9 | 34.6 | 33.9 | 28 | 34.8 |
SSE [67] | 60.1 | 61 | 43.9 | 51.5 | 34 | 50.1 |
SJE [41] | 65.6 | 61.9 | 53.9 | 53.7 | 32.9 | 53.6 |
ESZSL [47] | 58.2 | 58.6 | 53.9 | 54.5 | 38.3 | 52.7 |
LATEM [42] | 55.1 | 55.8 | 49.3 | 55.3 | 35.2 | 50.1 |
ALE [58] | 59.9 | 62.5 | 54.9 | 58.1 | 39.7 | 55 |
SYNC [49] | 54 | 46.6 | 55.6 | 56.3 | 23.9 | 47.3 |
SAE [48] | 53 | 54.1 | 33.3 | 40.3 | 8.3 | 37.8 |
DEM [70] | 68.4 | 67.1 | 51.7 | 61.9 | 35 | 56.8 |
GFZSL [62] | 68.3 | 63.8 | 49.3 | 60.6 | 38.4 | 56.1 |
CAVE [55] | 71.4 | 65.8 | 52.1 | 61.7 | - | - |
PSR [81] | - | 63.8 | 56 | 61.4 | 38.4 | - |
TVN [82] | - | 68.8 | 58.1 | 60.7 | - | - |
GAZSL [54] | - | 68.4 | 55.8 | 61.3 | 41.1 | - |
f-CLSWGAN [50] | - | 68.8 | 57.3 | 60.8 | 40.5 | - |
GDAN [51] | - | 67.7 | 51 | 54.8 | 40.4 | - |
LESAE [66] | 66.1 | 68.4 | 53.9 | 60 | 40.8 | 57.8 |
LisGAN [68] | 70.6 | - | 58.8 | 61.7 | 43.1 | - |
SRGAN [53] | 71.9 | - | 55.4 | 62.2 | - | - |
DEARF [60] | 72.1 | 69.3 | 38.5 | 48.6 | - | - |
BSR [52] | - | 68.4 | 57.7 | 61.2 | 41.3 | - |
BSAE | 72.3 | 69.4 | 59.7 | 58.6 | 53 | 62.6 |
Grasshopper Sparrow 50.5% | Pacific Loon 88.2% | Rhinoceros Auklet 86.8% | Western Grebe 85.7% | Least Auklet 84.8% | market outdoor 82.6% | recycling plant outdoor 37.8% | van interior 44.7% | subway station platform 54.8% | lecture room 54.5% |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
Class Name | Dcintra | Dcinter | |||
BSAE | TSTD | BSAE | TSTD | ||
chimpanzee | 0.275 | 0.444 | 1.663 | 1.361 | |
giant+panda | 0.441 | 0.442 | 1.583 | 1.264 | |
leopard | 0.36 | 0.39 | 1.69 | 1.246 | |
persian+cat | 0.077 | 0.388 | 1.887 | 1.501 | |
pig | 0.199 | 0.43 | 1.868 | 1.399 | |
hippopotamus | 0.412 | 0.498 | 1.594 | 1.337 | |
humpback+whale | 0.318 | 0.35 | 1.28 | 1.257 | |
raccoon | 0.224 | 0.333 | 1.326 | 1.192 | |
rat | 0.215 | 0.28 | 1.28 | 1.234 | |
seal | 0.257 | 0.228 | 1.31 | 1.243 |
Notations | Descriptions |
Xs∈Rd×m | Visual feature of source instances |
Xt∈Rd×p | Visual feature of target instances |
Ys={ysi}mi=1 | Class labels of Xs |
Yt={yti}pi=1 | Class labels of Xt |
Cs={1,2,…,τ} | Set of τ seen classes |
Ct={1,2,…,μ} | Set of μ unseen classes |
At∈Rk×μ | Semantic prototypes of Ct |
Str∈Rk×m | Semantic representations of Xs |
Ste∈Rk×p | Semantic representations of Xt |
λ1,λ2 | Hyper-parameters |
X | d-dimensional visual feature space |
A | k-dimensional semantic space |
m,p | Number of source and target instances |
d,k | Dimensionality of visual feature and semantic space |
Algorithm 1 Bi-shifting Semantic Auto-Encoder |
Input: Training data Xs, Str Test data Xt Hyper-parameters λ1, λ2 Output: Projection matrices W0, W2 1: Initialize W2 2: while not converge do 3: Update W0 by Eq 3.9 4: Update W1 by Eq 3.11 5: Update W2 by Eq 3.13 6: Check the converge condition 7: end while 8: Return W0, W2 |
At Training Time | At Testing Time | |||||||||
Datasets | Granularity | Size | Attributes | Cs/Ct | Images | SS (Cs) | PS (Cs) | SS (Ct) | PS (Ct) | |
AWA1 | coarse | medium | 85 | 40/10 | 30,475 | 24,295 | 19,832 | 6180 | 10,643 | |
AWA2 | coarse | medium | 85 | 40/10 | 37,322 | 30,337 | 23,527 | 6985 | 13,795 | |
aP & Y | coarse | small | 64 | 20/12 | 15,339 | 12,695 | 5932 | 2644 | 9407 | |
CUB | fine | medium | 312 | 150/50 | 11,788 | 8855 | 7057 | 2933 | 4731 | |
SUN | fine | medium | 102 | 645/72 | 14,340 | 12,900 | 10,320 | 1440 | 4020 |
Method | Feature | AWA1 | AWA2 | SUN | CUB | aP & Y | Average |
DAP [27] | R | 57.1 | 58.7 | 38.9 | 37.5 | 35.2 | 45.5 |
IAP [27] | R | 48.1 | 46.9 | 17.4 | 27.1 | 22.4 | 32.4 |
CONSE [37] | R | 63.6 | 67.9 | 44.2 | 36.7 | 25.9 | 47.7 |
DEVISE [38] | R | 72.9 | 68.6 | 57.5 | 53.2 | 35.4 | 57.5 |
CMT [39] | R | 58.9 | 66.3 | 41.9 | 37.3 | 26.9 | 46.3 |
SSE [67] | R | 68.8 | 67.5 | 54.5 | 43.7 | 31.1 | 53.1 |
SJE [41] | R | 76.7 | 69.5 | 57.1 | 55.3 | 32 | 58.1 |
ESZSL [47] | R | 74.7 | 75.6 | 57.3 | 55.1 | 34.4 | 59.4 |
LATEM [42] | R | 74.8 | 68.7 | 56.9 | 49.4 | 34.5 | 56.9 |
ALE [58] | R | 78.6 | 80.3 | 59.1 | 53.2 | 30.9 | 60.4 |
SYNC [49] | R | 72.2 | 71.2 | 59.1 | 54.1 | 39.7 | 59.3 |
G | 72.9 | - | 62.7 | 54.7 | - | - | |
SAE [48] | R | 80.6 | 80.7 | 42.4 | 33.4 | 8.3 | 49.1 |
G | 81.9 | - | 59.7 | 53.6 | 34.5 | - | |
SSZSL [79] | V | 88.6 | - | - | 58.8 | 49.9 | - |
DSRL [57] | V | 87.2 | - | - | 57.1 | 56.3 | - |
STZSL [80] | V | 83.7 | - | - | 58.7 | 54.4 | - |
GFZSL [62] | R | 80.5 | 79.3 | 62.9 | 53 | 51.3 | 65.4 |
TSTD [59] | V | 90.3 | - | - | 58.2 | - | - |
QFSL [61] | R | - | 84.8 | 61.7 | 69.7 | - | - |
VCL [56] | R | 82.0 | 82.5 | 63.8 | 60.1 | - | - |
DEARF [60] | R | 81 | 81.2 | 64.3 | 56.1 | - | - |
BSAE | R | 90.7 | 85.2 | 60.4 | 61.8 | 63.2 | 72.3 |
G | 91 | - | 67.1 | 62.4 | 57.7 | - |
Method | AWA1 | AWA2 | CUB | SUN | aP & Y | Average |
DAP [27] | 44.1 | 46.1 | 40 | 39.9 | 33.8 | 40.8 |
IAP [27] | 35.9 | 35.9 | 24 | 19.4 | 36.6 | 30.4 |
CONSE [37] | 45.6 | 44.5 | 34.3 | 38.8 | 26.9 | 38 |
DEVISE [38] | 54.2 | 59.7 | 52 | 56.5 | 39.8 | 52.4 |
CMT [39] | 39.5 | 37.9 | 34.6 | 33.9 | 28 | 34.8 |
SSE [67] | 60.1 | 61 | 43.9 | 51.5 | 34 | 50.1 |
SJE [41] | 65.6 | 61.9 | 53.9 | 53.7 | 32.9 | 53.6 |
ESZSL [47] | 58.2 | 58.6 | 53.9 | 54.5 | 38.3 | 52.7 |
LATEM [42] | 55.1 | 55.8 | 49.3 | 55.3 | 35.2 | 50.1 |
ALE [58] | 59.9 | 62.5 | 54.9 | 58.1 | 39.7 | 55 |
SYNC [49] | 54 | 46.6 | 55.6 | 56.3 | 23.9 | 47.3 |
SAE [48] | 53 | 54.1 | 33.3 | 40.3 | 8.3 | 37.8 |
DEM [70] | 68.4 | 67.1 | 51.7 | 61.9 | 35 | 56.8 |
GFZSL [62] | 68.3 | 63.8 | 49.3 | 60.6 | 38.4 | 56.1 |
CAVE [55] | 71.4 | 65.8 | 52.1 | 61.7 | - | - |
PSR [81] | - | 63.8 | 56 | 61.4 | 38.4 | - |
TVN [82] | - | 68.8 | 58.1 | 60.7 | - | - |
GAZSL [54] | - | 68.4 | 55.8 | 61.3 | 41.1 | - |
f-CLSWGAN [50] | - | 68.8 | 57.3 | 60.8 | 40.5 | - |
GDAN [51] | - | 67.7 | 51 | 54.8 | 40.4 | - |
LESAE [66] | 66.1 | 68.4 | 53.9 | 60 | 40.8 | 57.8 |
LisGAN [68] | 70.6 | - | 58.8 | 61.7 | 43.1 | - |
SRGAN [53] | 71.9 | - | 55.4 | 62.2 | - | - |
DEARF [60] | 72.1 | 69.3 | 38.5 | 48.6 | - | - |
BSR [52] | - | 68.4 | 57.7 | 61.2 | 41.3 | - |
BSAE | 72.3 | 69.4 | 59.7 | 58.6 | 53 | 62.6 |
Grasshopper Sparrow 50.5% | Pacific Loon 88.2% | Rhinoceros Auklet 86.8% | Western Grebe 85.7% | Least Auklet 84.8% | market outdoor 82.6% | recycling plant outdoor 37.8% | van interior 44.7% | subway station platform 54.8% | lecture room 54.5% |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
Class Name | Dcintra | Dcinter | |||
BSAE | TSTD | BSAE | TSTD | ||
chimpanzee | 0.275 | 0.444 | 1.663 | 1.361 | |
giant+panda | 0.441 | 0.442 | 1.583 | 1.264 | |
leopard | 0.36 | 0.39 | 1.69 | 1.246 | |
persian+cat | 0.077 | 0.388 | 1.887 | 1.501 | |
pig | 0.199 | 0.43 | 1.868 | 1.399 | |
hippopotamus | 0.412 | 0.498 | 1.594 | 1.337 | |
humpback+whale | 0.318 | 0.35 | 1.28 | 1.257 | |
raccoon | 0.224 | 0.333 | 1.326 | 1.192 | |
rat | 0.215 | 0.28 | 1.28 | 1.234 | |
seal | 0.257 | 0.228 | 1.31 | 1.243 |