Processing math: 100%
Research article Special Issues

A comprehensive transfer news headline generation method based on semantic prototype transduction


  • Most current deep learning-based news headline generation models only target domain-specific news data. When a new news domain appears, it is usually costly to obtain a large amount of data with reference truth on the new domain for model training, so text generation models trained by traditional supervised approaches often do not generalize well on the new domain—inspired by the idea of transfer learning, this paper designs a cross-domain transfer text generation method based on domain data distribution alignment, intermediate domain redistribution, and zero-shot learning semantic prototype transduction, focusing on the data problem with no reference truth in the target domain. Eventually, the model can be guided by the most relevant source domain data to generate headlines from the target domain news text through the semantic correlation between source and target domain data during the training process of generating headlines for the target domain news, even without any reference truth of the news headlines in the target domain, which improves the usability of the text generation model in real scenarios. The experimental results show that the proposed transfer text generation method has a good domain transfer effect and outperforms other existing transfer text generation methods in various text generation evaluation indexes, proving the proposed method's effectiveness in this paper.

    Citation: Ting-Huai Ma, Xin Yu, Huan Rong. A comprehensive transfer news headline generation method based on semantic prototype transduction[J]. Mathematical Biosciences and Engineering, 2023, 20(1): 1195-1228. doi: 10.3934/mbe.2023055

    Related Papers:

    [1] Tianshu Wei, Jinjie Huang, Cong Jin . Zero-shot learning via visual-semantic aligned autoencoder. Mathematical Biosciences and Engineering, 2023, 20(8): 14081-14095. doi: 10.3934/mbe.2023629
    [2] Ziyue Wang, Junjun Guo . Self-adaptive attention fusion for multimodal aspect-based sentiment analysis. Mathematical Biosciences and Engineering, 2024, 21(1): 1305-1320. doi: 10.3934/mbe.2024056
    [3] Xuwen Wang, Yu Zhang, Zhen Guo, Jiao Li . Identifying concepts from medical images via transfer learning and image retrieval. Mathematical Biosciences and Engineering, 2019, 16(4): 1978-1991. doi: 10.3934/mbe.2019097
    [4] Wanru Du, Xiaochuan Jing, Quan Zhu, Xiaoyin Wang, Xuan Liu . A cross-modal conditional mechanism based on attention for text-video retrieval. Mathematical Biosciences and Engineering, 2023, 20(11): 20073-20092. doi: 10.3934/mbe.2023889
    [5] Long Wen, Liang Gao, Yan Dong, Zheng Zhu . A negative correlation ensemble transfer learning method for fault diagnosis based on convolutional neural network. Mathematical Biosciences and Engineering, 2019, 16(5): 3311-3330. doi: 10.3934/mbe.2019165
    [6] Kaixiong Zhang, Yongbing Zhang, Zhengtao Yu, Yuxin Huang, Kaiwen Tan . A two-stage fine-tuning method for low-resource cross-lingual summarization. Mathematical Biosciences and Engineering, 2024, 21(1): 1125-1143. doi: 10.3934/mbe.2024047
    [7] Luqi Li, Yunkai Zhai, Jinghong Gao, Linlin Wang, Li Hou, Jie Zhao . Stacking-BERT model for Chinese medical procedure entity normalization. Mathematical Biosciences and Engineering, 2023, 20(1): 1018-1036. doi: 10.3934/mbe.2023047
    [8] Quan Zhu, Xiaoyin Wang, Xuan Liu, Wanru Du, Xingxing Ding . Multi-task learning for aspect level semantic classification combining complex aspect target semantic enhancement and adaptive local focus. Mathematical Biosciences and Engineering, 2023, 20(10): 18566-18591. doi: 10.3934/mbe.2023824
    [9] R Nandhini Abiram, P M Durai Raj Vincent . Identity preserving multi-pose facial expression recognition using fine tuned VGG on the latent space vector of generative adversarial network. Mathematical Biosciences and Engineering, 2021, 18(4): 3699-3717. doi: 10.3934/mbe.2021186
    [10] Jun Yan, Tengsheng Jiang, Junkai Liu, Yaoyao Lu, Shixuan Guan, Haiou Li, Hongjie Wu, Yijie Ding . DNA-binding protein prediction based on deep transfer learning. Mathematical Biosciences and Engineering, 2022, 19(8): 7719-7736. doi: 10.3934/mbe.2022362
  • Most current deep learning-based news headline generation models only target domain-specific news data. When a new news domain appears, it is usually costly to obtain a large amount of data with reference truth on the new domain for model training, so text generation models trained by traditional supervised approaches often do not generalize well on the new domain—inspired by the idea of transfer learning, this paper designs a cross-domain transfer text generation method based on domain data distribution alignment, intermediate domain redistribution, and zero-shot learning semantic prototype transduction, focusing on the data problem with no reference truth in the target domain. Eventually, the model can be guided by the most relevant source domain data to generate headlines from the target domain news text through the semantic correlation between source and target domain data during the training process of generating headlines for the target domain news, even without any reference truth of the news headlines in the target domain, which improves the usability of the text generation model in real scenarios. The experimental results show that the proposed transfer text generation method has a good domain transfer effect and outperforms other existing transfer text generation methods in various text generation evaluation indexes, proving the proposed method's effectiveness in this paper.



    With the rapid development of the Internet in recent years, the total amount of news on the Internet has proliferated. When users receive a large amount of news information, they usually want to filter out the news content of interest quickly. Therefore, a simple and attractive headline for news can significantly increase the number of news views. In traditional news media, headlines are usually generated by human editors. However, in the face of the massive news base on the Internet, it is essential to reduce the cost of human editing and improve the news dissemination rate if a more suitable headline can be automatically generated for given news content.

    Automated news headline generation can be addressed using automatic text summary generation methods in natural language processing [1]. However, training deep neural network-based news headline generation models often relies on many labeled data [2]. In practical application scenarios, it is challenging to train news headline generation models by relying entirely on the labeled data in the target topic domain because of too many news topic domains. Therefore, it is worth further research on effectively training a deep text generation model to achieve better domain generalization in the absence of labeled data in the target domain.

    In this regard, existing work often adopts the "Pre-train & Fine-tune" approach in transfer learning to alleviate the limitation of the lack of labeled data in the target domain [2]. However, the transfer learning paradigm shown in Figure 1 still has some shortcomings. First, the data differences between the source and target domains are relatively noticeable. In addition to fine-tuning, it is necessary to eliminate further the negative impact of data differences on the domain transfer effect from the data distribution perspective. Second, when it is insufficient, or no labeled data in the target domain for fine-tuning, the given deep text generation model cannot be effectively adapted to the target domain by fine-tuning, which leads to poor performance of the transfer text generation and directly weakens the adaptability of the text generation model to the target domain. In this regard, zero-shot learning provides a good inspiration [3] to construct a "domain prototype" for each domain to describe the semantics of the data under the domain through the feature attributes. The semantic correlation between different domain prototypes by the labeled data of the most relevant source domains assists in processing the unlabeled data of the target domain (i.e., semantic prototype transduction). Then for the task of news headline generation, even without any manually labeled data in the target domain, it is still possible to use a deep text generation model to generate domain-adapted news headlines, following the zero-shot learning semantic prototype transduction principle for many news bodies without reference truth summary in the target domain [4].

    Figure 1.  Transfer learning paradigm.

    In summary, this paper proposes a cross-domain news headline generation method based on semantic prototype transduction in the intermediate domain, with the following main contributions:

    1) Aligning the data representation distribution of source and target domain data in the kernel space to alleviate the negative impact of the distribution difference of data representation between different domains on inter-domain transfer and enhance inter-domain transfer at the data representation level.

    2) The source and target domain data are divided into K intermediate transition domain according to the text similarity index, forming an intermediate domain of "new source domain + target domain." In the intermediate domain, the target domain data can refer to the source domain data with more semantic similarity in the generation process through more appropriate domain data selection.

    3) To construct semantic prototypes for source and target domain data and improve the internal structure of the deep text generation model to enhance the model encoding and decoding process, so that the improved deep text generation model can receive all elements of the data-level semantic prototypes and improve inter-domain transferability from the model structure.

    4) Based on the improved text generation model, the corresponding text generation loss functions are constructed for different elements in the semantic prototypes of the data according to the idea of semantic prototype transduction. In this way, the model is guided to capture the approximate reference relations of cross-domain data on the semantic prototypes so that the model can further learn the semantic association between cross-domain data and thus use the relevant new source domain reference news headlines as the reference truth data of news bodies with no reference truth in the target domain in the intermediate domain.

    This paper is organized as follows: Section 1 summarizes the existing work on news headline generation and the current state of research on traditional transfer learning and zero-shot learning approaches in text generation tasks. Section 2 describes the internal mechanism of the proposed approach and the transfer strategy adopted in this paper. Section 3 describes the experimental data, setup, and comparison model and analyzes the experimental results. Finally, a conclusion and an outlook on future work are presented.

    The task of news headline generation is a branch of automatic text summarization [1]. The current mainstream automatic text summarization models mainly rely on many reference truth summary data for supervised training of the generative models to obtain a model with good generative performance. However, the problem of missing ground-truth text often occurs in practical application scenarios, and thus transfer learning-related methods are introduced to solve the above problem. The following is summarized the news headline generation task, the traditional transfer learning methods in the text generation task, and the zero-shot learning approach in the text generation task.

    The news headline generation task aims to output the corresponding news headlines based on the input news body through an automatic text summary generation model. The output headlines need to be semantically fluent and coherent while maintaining the semantic novelty of the headlines [1]. Current mainstream automatic text summary generation methods can be divided into Extractive and Abstractive. The extractive methods extract salient sentences or phrases from the original article [5]. In contrast, the generative methods generate new words or phrases that may be rewritten or use words that are not in the original article [6]. This paper focuses on the generative text summary model to generate corresponding headlines based on a given news body.

    Many researchers have recently used Sequence to Sequence structures to build generative text summarization models. For example, Krishna et al. [7] use a Recurrent Neural Network (RNN) with an attention mechanism in the "encoder-decoder" form to generate summary content for different topics of interest for a single text. Ma et al. [8] proposed a critical sentence selection method combining topic words and TF-IDF to obtain the score of the topic corresponding to each text in the original text data and then select the sentence with the highest score as the summary of the topic. Jadhav et al. [9] used a pointer generator network [10] to identify salient sentences and keywords in the input document and combine them to form the final summary. In addition, text summary generation models can also be constructed by neural network components based on the Self-Attention mechanism, such as Transformer [11]. The Transformer-based text generation model is also built as an "encoder-decoder, " which solves the problem that the traditional RNN architecture cannot compute in parallel and improves the efficiency of text generation. Du et al. [12] used the Transformer as a feature extractor to obtain text representations. They realized context-based bidirectional text coding, which embedded the location information of words into text representations according to their different positions in the text by contextual information and distinguished the semantic differences of similar expressions in the context at a fine granularity.

    As a result, the mainstream text generation models are still structured as "encoder-decoder." However, the current generative text summarization models built using recurrent neural networks (RNN) or Transformer with an "encoder-decoder" structure are usually trained in a traditional supervised manner [13]. They are not suitable for applications where the target domain is missing ground-truth data [14], which means that transfer text generation methods for such scenarios need to be investigated to overcome the limitation of sparse ground-truth data in the target domain.

    For the application of transfer learning methods in text generation tasks, current research work has shown that models trained using specific corpus data cannot be generalized across domains [15]. Currently, traditional transfer learning approaches focus on the source domain data to assist the target domain in accomplishing a specific task through some transfer strategy [13]. An example is the use of large-scale pre-trained language models for transfer. Transferring from the source domain to the target domain is achieved by obtaining a pre-trained model using a large-scale corpus and fine-tuning the pre-trained model using a relatively small amount of training data in the target domain [16]. Various pre-trained language models have been proposed according to the "pre-train & fine-tune" model. Specifically, Raffel et al. [17] proposed a pre-trained text generation model T5 using a large-scale Common Crawl database containing data from multiple domains to pre-train the model for different span masking padding tasks. Lewis et al. [18] pre-trained the sequence-to-sequence model BART using a denoising auto-encoder. A noise function was used to mask a random text span during the model's pre-training, guiding the model to learn how to reconstruct the original text. Zhang et al. [19] proposed the pre-trained text generation model PEGASUS to learn how to repopulate multiple masked sentences in a corpus for pre-training. Zhang et al. [20] proposed a novel encoder-decoder architecture based on a pre-trained language model that integrates contextual representations into Chinese character embeddings to aid the model in semantic understanding for migration. In addition, some researchers have obtained migratory representations of text or features to migrate domain information between domains in different feature spaces. Specifically, Chen et al. [21] designed a generalized covariate transfer assumption method to model the unsupervised domain adaptation problem by applying a distribution adaptation function in the subspace and using a convex optimization loss function to adapt the source domain data distribution to the target domain data distribution, thus solving the problem that the traditional feature transformation methods cannot approximate the transformed source domain distribution and the target domain distribution when the domain differences are significant. Li et al. [22] proposed a semi-supervised heterogeneous domain adaptation method based on matrix decomposition within the Reproducing Kernel Hilbert Space (RKHS) to learn heterogeneous features in the source and target domains using the nonlinear relationship between features and data instances to compensate for the feature differences between the source and target domains in the kernel space. Zellinger et al. [23] proposed a metric-based regularization method, representing similar latent features in different domains by maximizing the similarity between specific activation distributions to achieve unsupervised domain adaptation. In addition, there are also unsupervised domain adaptation methods [24,25,26,27,28,29] that extend the feature space by connecting common features and specific features of two domains in a cross-filling way to eliminate the heterogeneity of features between domains, making the feature space homogeneous and thus achieving the goal of reducing the variability of features between domains.

    Existing research shows that [16], on the one hand, fine-tuning the pre-trained language model with a "small amount" of target domain data allows for effective domain adaptation of the language model. On the other hand, when applying the pre-trained language model to the target domain, a certain amount of data is still needed to fine-tune the model to achieve better domain adaptation [30]. If the target domain lacks ground-truth data, it will directly affect the generalization effect of the model in the target domain, and the limitation of missing labeled data in the new domain still exists. Therefore, more and more researchers are focusing on more effective methods to transfer text generation models from the source domain to the target domain in the absence of labeled data in the target domain to achieve better text generation results in the target domain.

    In terms of relationship-based transfer strategies, many researchers have applied zero-shot learning [31] related methods to transfer text generation tasks in recent years. The zero-shot learning method is more targeted at solving the problem of missing labeled data in the target domain than the traditional transfer learning method. In the absence of ground-truth data for the target domain, the zero-shot learning approach usually constructs a corresponding "prototype description" for each domain. Thus, even if the input data is unlabeled, the category label of a given input data can be inferred if a set of attributes of the input data is "close" to the "prototype description" of a domain [4]. Thus, the problem of the lack of ground-truth data in the target domain can be solved using domain semantic prototype transduction. Specifically, Zhao et al. [32] used a cross-domain encoder to encode the domain prototypes shared between the source and target domains by selecting several representative conversation texts from each domain data, using the corresponding ground-truth texts as seeds as well as the critical entity words in the representative conversation texts as annotations, and then generating the conversation texts through a decoder, thereby achieving transfer from the source domain to the target domain based on the similarity of the domain prototypes between different domains. Liu et al. [33] first collected semantically similar terms in the source and target languages (including words collected from the target language ground-truth text) as domain semantic prototypes in a multilingual scenario; based on this, the hidden variable model was used to deal with the differences in domain distribution of similar sentences across languages. The transfer text generation model proposed by Shen et al. [34] and Duan et al. [35] takes the original document of the source domain as input but generates text for the target domain directly and uses the ground-truth text of the target domain to train the generation model. By building a streamlined text generation model with the same structure, the semantic prototype mapping from the source domain to the target domain is established by imitating the above "input→output" process. Finally, the original document of the target domain is used as input to generate the text generation result corresponding to the target domain.

    The following aspects still need further research by summarizing the existing transfer text generation methods. First, the language model pre-trained by the large-scale corpus still requires a certain amount of labeled data in the target domain for fine-tuning when applied to the target domain so that the model can be adapted to the target domain. This means the limitation of missing ground-truth data in the target domain still exists. Secondly, the variability in the distribution of data representations across different domains can hurt the model across domains [15]. This implies the need to reduce the variability between data representations across domains through practical methods. Finally, in the process of cross-domain model generation, the target domain data should be assisted by the source domain data as much as possible to improve the text generation effect. This means that we need to mine the existing source domain data to find the information that is helpful to the target domain data and find the most valuable source domain data for the target domain data by obtaining the correlation between the data and improving the model's ability to get the correlation information, to assist the generation of the target domain data.

    The main challenge of the transferable text generation task using the zero-shot learning approach is how to fully leverage the existing data with reference truth in the source domain to help the text generation of the target domain data without reference truth. This section describes the proposed transferable news headline generation method in four aspects: domain data distribution alignment, intermediate domain redistribution, transferable news headline generation model construction, and semantic prototype transduction, respectively. Table 1 lists the symbols used in this method and the corresponding descriptions.

    Table 1.  The explain of key symbols used in the paper.
    Symbols Explanations
    Xsrc The news body word embedding representation for the source domain
    Xtar The news body word embedding representation for the target domain
    Sc/Tc Common feature for source and target domains
    Ss Source domain specific features
    Tt Target domain specific features
    Ψ The feature mappings as the translations to link up the common features with the domain specific features
    Φ The feature mappings to RKHS
    X'src Aligned source domain news body word embedding representation
    X'tar Aligned target domain news body word embedding representation
    D'i Intermediate domains
    xsrc/xtar News text for source/target domains
    ysrc/ytar News headlines for source/target domains
    asrc/atar Semantic annotation of source/target domains
    ysrc/ytar Generated news headlines for source/target domains

     | Show Table
    DownLoad: CSV

    The problem to be solved in this paper can be defined as when the news body text of the source domain, the news headline text of the source domain, and the news body text of the target domain are given. The proposed method for generating news headlines text based on semantic prototype transduction generates news headlines text for the target domain when there is no reference truth headline text in the target domain.

    In this regard, the overall flow of the proposed method is shown in Figure 2:

    Figure 2.  The overall flow chart of the method proposed in this paper.

    First, the textual representations of the source and target domain data are projected into the Reproducing Hilbert Kernel Space (RKHS) to align the data distribution of the source domain with the data distribution of the target domain, thus reducing the negative impact caused by the difference in data distribution between different domains and improving inter-domain transferability from the data representation level.

    Second, the intermediate domains are established, and the data in the source and target domains are redistributed into several intermediate domains based on the composite index of text similarity, forming a "new source domain + target domain" intermediate domain. This enables more appropriate domain data selection in the intermediate domains and assigns source domain data with more semantic similarity to the target domain data.

    Third, the text generation model with the "encoder-decoder" structure is improved for the cross-domain transfer scenario to realize the transfer from the source domain to the target domain by applying the semantic prototype transfer method in zero-shot learning.

    Fourth, in each intermediate domain, "semantic prototypes" are constructed for each piece of news data in different domains. Through semantic prototype transduction, the news body without reference truth headline of the target domain in the intermediate domain is semantically associated with the most relevant headlines in the new source domain, and news headlines are generated for the target domain news body in a transferable manner based on the similarity or proximity of the semantic prototypes.

    Eventually, in the transferred text generation process, as shown in Figure 2, the news headlines in the relevant source domains will act as ground-truth data for the target domain news headline generation, thus no longer relying on manual annotation of the target domain data.

    In practical application scenarios, the feature spaces of the two domains have similarities and differences. For example, in a news headline generation task, news headlines on different topics will have some common descriptors, such as "good, " "bad, " etc. News on different topics also have specific descriptors, such as "expensive" for cars and "exciting" for sports [3]. We, therefore, define these common descriptors as domain common features and these specific descriptors as domain specific features. Specifically, there are some common characteristics among the different domains, but each domain has specific characteristics. In domain adaptation, the variability between data distributions in different domains can be effectively reduced by using the common features of different domains to link them together. As shown in Figure 3, there will be some common features between the two domains Sc/Tc, where Sc denotes the common features of the source and target domains contained inside the source domain and Tc denotes the common features of the source and target domains contained inside the target domain. While each domain has domain-specific features Ss/Tt, where Ss denotes the source domain-specific features and Tt represents the target domain-specific features. Therefore, to achieve better performance metrics on transferable text generation, it is necessary first to align the data distribution representation between the source and target domains to reduce the impact of the distribution differences in data representation between domains on transferable text generation.

    Figure 3.  Data distribution cross-fill process.

    In Figure 3, the source domain news body is denoted as Xsrc, and the embedding of the input features is represented as Xsrc=[Sc;Ss], where Sc represents the feature matrix of the source domain containing c common features and s source domain-specific features. The target domain news body data is represented as Xtar, and the embedding of the input features is represented as Xtar=[Tc;Tt], where Tc represents the feature matrix of the target domain containing c common features and the feature matrix of t target domain-specific features Tt. The data distribution between Xsrc and Xtar is first aligned with feature filling using class cross-filling in Figure 3 to reduce the influence of domain-specific features. On this basis, Maximum Mean Discrepancy (MMD) is used to align the filled source and target domain data from the data distribution level by minimizing the maximum mean discrepancy to reduce the difference in the distribution of the filled domain data within the Reproducing Hilbert Kernel Space (RKHS).

    {minΨsrcΨsrc(Sc)Ss2minΨtar Ψtar (Tc)Tt2 (1)
    {Sa=Ψtar(Sc)Ta=Ψsrc(Tc) (2)
    {Xsf=[Sc;Ss;Sa]Xtf=[Tc;Tt;Ta] (3)

    Specifically, first, as shown in Eq (1), the feature mapping functions Ψsrc and Ψsrc map the common features in the source and target domains to associate with features specific to their respective domains. Second, as shown in Eq (2), the resulting feature mappings Ψsrc and Ψtar are cross-acted on Tc and Sc for feature filling, i.e., the feature mapping Ψtar obtained from the target domain is applied to the common feature Sc in the source domain as shown in Figure 3 to obtain the domain-adapted feature matrix Sa. The same crossover operation is done for the target domain to obtain the domain adaptation-specific feature matrix Ta. Third, as shown in Eq (3), the original feature matrix, the specific feature matrix and the adapted feature matrix of the source and target domains are filled to obtain the filled feature matrices Xsf and Xtf, respectively. And the 2 feature mappings Ψsrc and Ψtar in Eq (1) can be expressed respectively as Ψsrc(Sc)=WsSc and Ψtar(Tc)=WtTc, then Sa=WtTc,Ta=WsSc. Thus Eq (1) can be further derived as Eq (4):

    (4)

    Further, to better adapt the source domain to the target domain, it is also necessary to ensure that the feature matrices Xsf and Xtf (Xsfi and Xtfi is the single instance in the source and target domains, respectively; n1 and n2 denote the number of samples input in the source and target domains respectively) of the source and target domains output by Eq (3) are as close as possible in distribution.

    Dist(Xsf,Xtf)=1n1n1i=1Φs(Xsfi)1n2n2i=1Φt(Xtfi)H (5)

    As shown in Eq (5), the representation after filling alignment is mapped into the Reproducing Kernel Hilbert Space (RKHS) H using feature mapping Φ:RH respectively. In this kernel space, the distribution distance Dist of different domain data after mapping into the kernel space is measured by the maximum mean difference (MMD). By reducing the distribution distance Dist between Φs(Xsf) and Φt(Xtf) mapping results thus reducing the distribution difference between the source and target domain data.

    As shown in Figure 4, the source domain text embedding representation and the target domain text embedding representation are output through BERT. Since the maximum number of Tokens input to BERT is 512 [36], we consider truncating the excess directly if the length of the input news body exceeds 512. Since we show information about the dataset we used in Section 4.1, most of the news bodies in the dataset we used were less than 512 in length, so taking a direct truncation approach would not have a significant impact on the results.

    Figure 4.  The process of domain data distribution alignment.

    The source domain text embedding representation is feature transformed through the fully connected layer with the Sigmoid activation function, and then the result is projected into the kernel space H, while the target domain text embedding representation is projected directly into the kernel space H. Because the textual representation output from the source domain feature mapping is treated as consistent with the target domain distribution after alignment of the domain data distribution, the textual representation of the target domain will be output directly through the updated source domain feature mapping, so only the source domain feature mapping needs to be updated here, and no changes to the target domain feature mapping are required.

    By minimizing the objective function Dist(Xsf,Xtf) in Eq (5), the data distribution of the source domain and the target domain are made close to each other. As a result, the parameters of the fully connected layer in Figure 4 will be updated as Φs to minimize the objective function of Eq (5).

    {Xsrc=Φs(Xsf)Xtar=1N1SourceΦs(Xtf) (6)

    After training according to Eq (5), the source domain text representation Xsrc output from the source domain fully connected layer mapping Φs(X) is used as the result of the representation aligned with the target domain distribution as shown in Eq (6). And the text representation Xtar of the target domain itself is calculated by inputting the original embedding representation of the target domain into the source domain mapping Φs(X). When there are multiple source domains, as shown in Eq (6), the text representation of the target domain will be the average representation on multiple source domains. Here, N in Eq (6) denotes the total number of all domains. In summary, the overall process of aligning the domain data distribution for the source domain news body Xsrc and the target domain news body Xtar is shown in Algorithm 1.

    Algorithm 1. Domain data distribution alignment process.
    Input: Source domain news body Xsrc. Target domain news body Xtar;
    Source domain feature representation Xsrc=[Sc;Ss], Target domain feature representation Xtar=[Tc;Tt].
    Output: Source domain distribution alignment representation Xsrc, Target domain distribution alignment representation Xtar.
    Step 1: The feature mapping functions Ψsrc and Ψtar are obtained by minimizing the objective function of Eq (1).
    Step 2: The feature mappings Ψsrc and Ψtar are cross-acted on Tc and Sc to obtain the domain-adapted feature matrices Sa and Ta in Eq (2).
    Step 3: The feature filling operation in Eq (3) is performed to obtain the feature matrices Xsf and Xsf after the source and target domains are filled and aligned.
    Step 4: The distribution difference is reduced by minimizing the maximum mean difference Dist in Eq (5) to obtain the source domain fully connected layer mapping Φs(X).
    Step 5: The Xsf obtained in Step 3 is input into the source domain fully connected layer mapping Φs(X) in Eq (6) to obtain the aligned source domain distribution alignment representation Xsrc.
    Step 6: Input the Xtf obtained in Step 3 into the source domain fully connected layer mapping Φs(X) in Eq (6) to obtain the aligned target domain distribution alignment representation Xtar after alignment. If there are multiple source domains, then take the average representation.

    In order to enhance the transferability between source and target domains, improve the relevance of domain data in the transfer process, and find more suitable source domain news headlines for the target domain news body as a reference, all the data in the source and target domains are further reassigned to K intermediate transition domains according to the comprehensive index of text similarity, forming a new intermediate domain of "new source domain + target domain". Thus, a more appropriate source domain data is assigned to the target domain data in the intermediate domain, i.e., a more relevant domain data selection.

    Specifically, as shown in Figure 5, each redistributed intermediate domain contains the source and target domain data with the most similarity. Due to the semantic differences between different domain data, an inappropriate intermediate domain division can lead to negative transfer problems between the source and target domain data it contains [3]. Therefore, the data within each intermediate domain should have as many similar features as possible. First, as shown in the left side of Figure 5, the distribution alignment representations Xsrc and Xtar of each source domain and target domain are obtained from Eq (6), and then the distribution alignment representations of all data in each source domain are averaged to obtain the average distribution alignment representation vector in each source domain. Next, the data point in each source domain that is closest to the average distribution alignment representation vector is used as the starting point of each intermediate domain, and thus the number of source domains N-1 intermediate domain starting points is obtained. Next, this paper investigates and selects four similarity calculation metrics for intermediate domain redistribution in terms of text content similarity:

    Figure 5.  The overall process of domain text intermediate domain reclassification based on the comprehensive text similarity index. (Note: * is the average distribution alignment representation of each source domain; arrows point to the starting data points of the intermediate domain).

    Word-specific overlap Soverlap: Calculates the similarity of a given text pair, i.e., the higher the overlap of specific words used in the text, the more similar the main message conveyed by the text. The cosine similarity is used to quantify this metric, as shown in Eq (7). where xi and yi denote the values of the word frequency vectors x and y phases at the same position i, i.e., the number of occurrences of each participle, after the source and target domain texts have been encoded by OneHot.

    cos(θ)=ni=1(xi×yi)ni=1(xi)2×ni=1(yi)2 (7)

    Word coverage Scoverage: the number of overlapping words in each text pair is divided by the number of words in the target domain text, i.e., more identical words in the text indicates that the source domain text is more like the target domain text. The co-occurrence of source and target domain text on a single word was measured using ROUGE-1 [37] based on Recall as shown in Eq (8). Where, 1-gram means the co-occurrence word granularity is 1; 1-gramS Countmatch(1-gram) means the number of 1-gram occurring in the source text and the target text at the same time; 1-gramS Count(1-gram) means the number of 1-gram occurring in the target text.

     ROUGE1=1grameCountmatch  (1 - gram)1grameCount (1 - gram) (8)

    Information density Sdensity: divides the number of overlapping words in each text pair by the number of words in the source domain text, i.e., a high information density indicates that there is a large amount of information in the source domain text that can be transferred to the target domain. Using BLEU1 [37] the similarity between the source and target domain text on a single word is measured according to the Precision as shown in Eq (9). Where 1-gram indicates a co-occurrence word granularity of 1; 1-gramC Countclip(1-gram) indicates the number of simultaneous 1-gram in the source text and the target text; 1-gram'C Count(1-gram) denotes the number of 1-gram appearing in the source domain text.

    BLEU1=1gramCCountclip (1gram)1gramCCount(1gram) (9)

    Text length Slength: The text length can reflect the amount of information contained, i.e., the amount of information contained in text pairs having similar length or approximately the same. This metric is quantified using the negative value of the absolute difference between the length of the source and target domain text tokens and the ratio of the text token lengths, as shown in Eq (10). Where, Star_len denotes the number of words in the word sequence obtained after the target domain text has been split; Ssrc_slen denotes the number of words in the word sequence obtained after the source domain text has been split.

    Slength =|Star_lenSsrc_len|Star_len+Ssrc_len (10)

    Finally, as shown in Eq (11), the word-specific overlap, word coverage, information density and text length are summed to obtain the comprehensive index S used to calculate the content similarity between the source domain text and the target domain text:

    S=Soverlap +Scoverage +Sdensity +Slength  (11)

    Then, after obtaining the number of source domains N-1 intermediate domain starts, the number of starts is evaluated using the Silhouette Coefficient [38], which is commonly used in clustering methods, to determine the best K intermediate domain starts from the N-1 intermediate domain starts.

    Assume that the data in the source and target domains have been divided into several intermediate domains according to the composite index S of text content similarity in the source domain, and for each sample point i in each intermediate domain, its silhouette coefficient is calculated separately. Specifically, the following two metrics need to be calculated for each sample point i: a(i) denotes the average of the distances from sample point i to other sample points belonging to the same intermediate domain as it; b(i) denotes the mean bij of the average distance from sample point i to all samples in the other intermediate domains Cj, where b(i)=min(bi1,bi2,,bik). Then the Silhouette Coefficient of sample point i is

    s(i)=b(i)a(i)max(a(i),b(i)) (12)

    The average of the silhouette coefficients of all sample points i in the intermediate domain is the total silhouette coefficient S[1,1] of the intermediate domain, and the closer S is to 1, the better the intermediate domain partitioning effect is. Then the silhouette coefficients of each intermediate domain are summed and ranked to obtain the combination of intermediate domains with the highest silhouette coefficient sum score, at which the number of combined intermediate domains is the optimal K value for intermediate domain partitioning. Finally, the remaining news bodies in the source and target domains are divided into the intermediate domains with the first scores, and the remaining news bodies in the source and target domains are divided into the intermediate domains with the most similar scores, as shown in Figure 5, forming the K intermediate domains Di of the "new source domain + target domain".

    On the other hand, each intermediate domain in Figure 5 contains both the most similar source and target domain data, so that the most relevant new source domain news headlines can be used as the model training reference ground-truth data according to the similarity of semantic prototypes for the target domain news body in the intermediate domain. Finally, the overall process of intermediate domain redistribution based on the text similarity index in Figure 5 is shown in Algorithm 2.

    Algorithm 2. Intermediate transition domain redistribution process.
    Input: source domain news body, number of source domains = N-1,
    target domain news body, number of target domains = 1.
    Output: Redistribution into K intermediate domains Di of the new source domain news body and the target domain news body.
    Step 1: The source domain distribution alignment embedding representation obtained by Eq (6) is averaged to obtain the average distribution alignment representation in the source domain.
    Step 2: Get the news body text in the source domain that is closest to the mean distribution alignment representation as the starting text and get N-1 news text data of the starting point of the intermediate domain.
    Step 3: According to the Silhouette Coefficient Eq (12), the silhouette coefficient S of the new intermediate domain centered at each starting point is obtained.
    Step 4: According to N-1 silhouette coefficient, the highest ranked silhouette coefficient S score combination is derived, and the number of intermediate domains at this point is the optimal K take value.
    Step 5: The data in the remaining source and target domains were compared with the K intermediate domain starting news texts respectively by Eq (11) to calculate the text similarity composite index S and ranked according to the scores.
    Step 6: Based on the index scores, the text is divided into the highest scoring intermediate domains.
    Step 7: Repeat Steps 5 and 6 operations for the remaining news bodies in the source and target domains until all data are divided into the new K intermediate domains Di.

    The transfer text generation model can effectively cope with the problem of missing ground-truth data in the target domain during the generation process. A semantic prototype is a cognitive reference point, i.e., the proto image of all representatives of the meaning of a word or a category. Accordingly, the members of a category can be graded according to their typicality. Accordingly, the correlation between different domains can be obtained indirectly by obtaining similar high-frequency usage keywords between different domains. Based on the distribution-aligned source domain data representation Xsrc and target domain data representation Xtar in Figure 4, and the data in K intermediate domains Di redivided in Figure 5, this paper designs a zero-shot learning semantic prototype transduction transfer text generation model based on intermediate domains. Through the semantic prototype transduction strategy, the transfer text generation model can learn the semantic association of texts between different domains.

    First, a semantic prototype is constructed for each (news x, headline y) data in the source and target domains from the news body x, the corresponding news headline y and the semantic annotation a obtained based on the news body x, denote as z=[xd,yd,ad],d{src,tar}. Where d indicates that the data comes from the source domain src or the target domain tar. The news body of the source and target domains in the semantic prototype z is denoted as xsrc and xtar; the news headline of the source domain is denoted as ysrc. When it comes to the news headline data ytar of the target domain, the top n clauses with the highest scores are selected from the news body xtar as the "pseudo-truth data" of the current target domain news body xtar (ytar, i.e., the target domain pseudo-news headline) based on the ROUGE-L index score between each clause in the corresponding news body xtar and the whole news body xtar. Here, the number of extracted clauses n = len(headline y)/len(news x) × number of clauses of news x, based on the (news x, headline y) data information of the intermediate domain source domain to which the current target domain news body xtar belongs. The semantic annotations asr c and atar of the source and target domains are obtained by converting the news body xsrc and xtar of the source and target domains into a keyword sequence in which the lexical properties of each word belong to {noun, verb, adjective, adverb}, and each word is assigned the corresponding sentiment polarity value (i.e., between [-1, 1]). Thus, by the above process, a semantic prototype is constructed for each "news x - (pseudo-)headline y" pair in the source and target domains, denoted as z=[xd,yd,ad],d{src,tar}.

    It is worth noting that, on the one hand, the news body xd, news headline yd and semantic annotations ad are obtained from their embedding representations by the BERT model. On the other hand, in constructing the semantic prototype z=[xd,yd,ad],d{src,tar}, all domain data have been divided into K intermediate domains following the domain redistribution principle shown in Figure 5. And the news body representation xd has been aligned with the domain data distribution according to Eq (6). Finally, the constructed semantic prototype z=[xd,yd,ad],d{src,tar} will be input to the transfer text generation model for semantic prototype transduction as shown in Figure 6.

    Figure 6.  Structure of the transfer text generation model for semantic prototype transduction.

    As shown in Figure 6, the proposed transfer text generation model is built based on the "encoder-decoder" format. Among them, the encoder side consists of two encoder modules ET and ER with the same structure. The encoder modules ET and ER, and the decoder module DT at the decoder side are built by combining the Transformer model [11] with a bi-directional long and short-term memory network (Bi-LSTM), which allows the transfer text generation model to integrate a self-attentive mechanism with a recurrent neural network. In addition, a pointer generator network [10] was added to the decoding side of the model to address the Out-Of-Vocabulary (OOV) problem in the text generation. It has been shown [39] that the combination of recurrent networks and transformers can further improve the model performance of the transformer, and that the bi-directional LSTM in the decoder helps downstream sequential prediction tasks to achieve better accuracy in NLP tasks.

    Specifically, first, the encoder module, as well as the decoder module of the transfer text generation model in Figure 6, are designed with reference to the original Transformer model [11], and each module includes N stacked sublayers, each of which consists of a Multi-Head Attention mechanism and a fully connected Feed Forward Network, while both employ the residual connection renormalization (Add & Norm) processing. Second, the Bi-LSTM layer is added to each sub-layer of the encoder modules ET, ER and decoder module DT to construct the enhanced encoder and decoder. In each sub-layer so designed, the input of the Bi-LSTM layer is the same as the original input of the sub-layer, while the output is added to the original output of the sub-layer before the final normalization of the sub-layer (Add & Norm). In addition, if the Bi-LSTM uses the same number of hidden units H as the Transformer model, a Bi-LSTM output of dimension 2H is obtained, so the design adds a Linear layer that projects the output dimension 2H of the Bi-LSTM to dimension H to match the output dimension of the Transformer. As a result, the semantic correlation (provided by the self-attention mechanism in the Transformer) and the temporal dependency (provided by the Bi-LSTM) in the input data can be preserved simultaneously. Third, during the model training process the encoder module ET at the encoder side is used to receive news body xd as input. Another encoder module ER is used to receive news headline yd or semantic annotation ad as input. In contrast, the module DT at the decoder side will receive news headline yd to participate in the model training. When the news headline yd is from the source domain, the ground-truth news headline ysrc from the source domain is used; when the news headline yd is from the target domain, the pseudo news headline ytar from the target domain is used.

    In the above way, the news body xd and news headline yd of the source and target domains are fed to the encoder and decoder simultaneously, thus establishing the semantic association between the source and target domain data in the zero-shot learning semantic prototype transduction phase. As a result, during the training process of the transferred text generation model, the decoder module performs Multi-head Attention computation [11] with the outputs of the two encoder modules at the encoder side, respectively, thus capturing the global dependencies between the news body xd, the semantic annotation ad and the news headline yd at the encoder and decoder sides, as shown in Figure 6. In addition, due to the inclusion of the pointer generator network, the decoder uses the "copy mechanism" [10] provided by the pointer generator network during the text generation process to decide at each time step of the news headline generation whether to copy words from the input text at the encoder side or to generate words from the word list, thus completing the final News headline generation.

    As shown in Figure 6, the text generation model constructed in this paper applicable to semantic prototype transduction receives semantic prototype z=[xd,yd,ad],d{src,tar} as input and outputs the generated news headline text yd'. Specifically, the model encoder receives the semantic prototype z=[xd,yd,ad],d{src,tar} as input, and in the encoding stage, the encoder receives the input v=[Wv1,,Wvn] to obtain the encoder hidden state h=[h1,h2,,hn]. In the decoding stage, given the input xt, the decoded hidden state st at time step t can be derived and the attention distribution at of the encoder hidden state h is calculated to combine the linear transformation of the encoder hidden state h and the decoder state st. Next, at time step t, a context vector representation ct is computed from the weighted sum of the encoder hidden state to the attention distribution. The word distribution Pvocab(wt) can then be obtained, where Wp and bv are learnable parameters, and Pvocab(wt) represents the probability distribution of all words in the word list when words are predicted at time step t.

    In addition, the pointer generator network is used to employ the pointer ptgen as a soft switch at time step t of decoding to choose whether to generate a word from the vocabulary by selecting it with probability Pvocab(wt) or to copy a word from the input text according to the attention weight at. Thus, the probability distribution P(wt) of the final extended word list is obtained. where ptgen is calculated based on the context vector ct, decoder state st and decoder input xt. The specific process of generating news headlines yd' by the model shown in Figure 6 is shown in Eq (13):

    {at=softmax(tanh(Whh+Wsst+batt))Pvocab(wt)=tanh(Wp[st,ct]+bv)Ptgen=softmax(tanh(Wcct+Wsst+Wyyt+bgen))P(wt)=ptgenPvocab(wt)+(1ptgen)wtat,j (13)

    As a result, in the training process of the model shown in Figure 6, the model receives inputs xd/yd/ad and obtains the final word order distribution probability Pvocab by weighting and summing the word generation probability distribution Pvocab and attention probability distribution at with the pointer switch ptgen according to Eq (13) to generate the corresponding news headline yd.

    Further, based on the generation process shown in Eq (13), the loss function Lxy is designed for (xd,yd) in the semantic prototype z according to Eq (14), so that the news headline yd generated by the input news body xd is "close" to the reference news headline yd corresponding to xd, and thus the semantic prototype transduction relationship among the news body xd, the truth news headline yd and the generated news headline yd is deduced.

    {Lxy(xd,yd)=logP(ydET(xd))+D[ET(xd)||ER(yd)],d{Source, Target} s.tydyd (14)

    Specifically, as shown in Eq (14), ET(xd) indicates that the news body xd is input into the encoder module ET at the encoder end; ER(yd) indicates that the news headline yd is input into another encoder module ER at the encoder end. When minimizing the loss function Lxy defined in Eq (13), D[ET(xd)||ER(yd)] means that the hidden state of ET(xd) output should be "close" to the hidden state of ER(yd) output, where D means that the mean-square error (MSE) is used as the distance measure. In the process of xd generating yd', i.e. xdyd', xd and yd' are naturally semantically similar to the reference truth yd', so it can be approximated as xdyd and ydyd', then an implicit semantic prototype transduction relation (ydxd)ydyd can be established in the intermediate domain Di by minimizing the loss function Lxy.

    Similarly, based on the generation process shown in Eq (13), the loss function Lay is designed for (ad,yd) in the semantic prototype z according to Eq (15) for the zero-shot learning semantic prototype transduction, so that the headline yd' generated by the input semantic annotation ad is "close" to the reference news headline yd corresponding to ad, and thus the semantic prototype transduction relationship among the semantic annotation ad, the news headline yd and the generated news headline yd' is deduced.

    {Lay(ad,yd)=logP(ydER(ad))+D[ER(ad)||ER(yd)],d{Source, Target} s.tydyd (15)

    Specifically, as shown in Eq (15), after inputting the semantic annotation ad corresponding to the news body xd to the encoder module ER, the model is still made to generate the news headline yd. Meanwhile, by minimizing D[ER(ad)||ER(yd)], the hidden state ER(ad) of the output of the encoder module ER is guided to be "close" to the hidden state of the output of ER(yd). In the process of ad generating yd', i.e. adyd', ad and yd' are naturally semantically similar to the reference truth yd', so it can be approximated as adyd and ydyd', then an implicit semantic prototype transduction relation (ydad)ydyd can be established in the intermediate domain Di by minimizing the loss function Lay.

    Lmain =Lxy+Lay (16)

    Finally, as shown in Eq (16), a composite generative loss function Lmain is constructed by combining the loss function Lxy and Lay, thus indirectly responding to the principle of transfer text generation based on semantic prototype transduction, i.e., when the input semantic prototype z=[xd,yd,ad],d{src,tar}, the parameters of the transfer text generation model in Figure 6 will be trained by the compound loss function Lmain in Eq (16).

    As shown in Figure 7, when it comes to semantic prototype transduction between the new source and target domains, based on the similar semantic prototypes x/a/y between different domains, if any semantic prototypes ztar=[xtar,ytar,atar] in target domain has a semantic association with a semantic prototypes zsrc=[xsrc,ysrc,asrc] in source domain, then a (xdydad)ydyd relationship can be established through the association between ydxdydyd and ydadydyd, which in turn forms an latent relationship of xtarysrc. Therefore, when given the target domain news body xtar, the relevant ground-truth news headlines ysrc in the new source domain can be referred to assist in generating the news headlines ytar in the target domain. Thus, even if there is no ground-truth news headline data in the target domain, the news headline can be generated with the help of the news body in the target domain by means of zero-shot learning semantic prototype transduction with the new source domain data. The overall process is shown in Algorithm 3.

    Figure 7.  The diagram of the principle of zero-shot learning semantic prototype transduction.

      Algorithm 3. Transfer news headline generation process based on zero-shot learning semantic prototype transduction.
    Input: source domain semantic prototype z=[xsrc,ysrc,asrc],
    target domain semantic prototype z=[xtar,ytar,atar];
    Output: Generated news headlines yd,d{source,target}.
    Step 1: Within the intermediate domain Di, the transfer text generation model is trained by Lxy in Eq (14) to construct semantic associations in the source domain:
    xsrcysrcasrcysrc.
    Step 2: Within the intermediate domain Di, the transfer text generation model is trained by Lay in Eq (15) to construct semantic associations in the target domain:
    xtarytaratar.
    Step 3: Within the intermediate domain Di, the transfer text generation model is trained by Lmain in Eq (16) to construct cross-domain semantic associations:
    xtarysrcasrcysrc, i.e., xtarysrc.
    Step 4: The model generates news headlines yd, d{src,tar} by Eq (13). The parameters of the transfer text generation model are updated during the generation process.

    In the experiment, for the task of news headline generation, this paper selects the publicly available dataset PENS (PErsonalized News headlineS) [1]. 113,762 news items are included in PENS, which are divided into 15 topics, and each news item contains a headline and body. In this paper, 8 news topics are randomly selected from the PENS dataset, including Sport, Finance, Music, Weather, Auto, Movie, Health, and Kid. 8000 news data are randomly chosen as the training dataset in each domain.

    The information about the datasets used in the experiments is described in Table 2. The "Average Length" and "Maximum Length" represent the maximum and average length of the word sequences in each domain after the pre-trained BERT model has been used to split the words of all news bodies and news headlines. The "Compression Ratio" represents the ratio of the average length of the text of news headlines to the average length of the news body in a domain.

    Table 2.  Statistical information of news data extracted from the PENS dataset.
    NO. Topic News Body News Headline Compression Ratio (%)
    Average Length Maximum Length Average Length Maximum Length
    1 Sport 480.5 537 12.9 17 2.22%
    2 Finance 482.7 588 8.9 19 1.84%
    3 Music 528.0 557 10.5 18 1.99%
    4 Weather 484.9 566 12.3 19 2.53%
    5 Auto 511.7 580 9.1 15 1.77%
    6 Movie 483.1 636 10.1 19 2.09%
    7 Health 483.8 544 9.0 15 1.53%
    8 Kid 509.4 560 13.3 16 2.61%

     | Show Table
    DownLoad: CSV

    In the experiments, the number of sub-layers of both the encoder and decoder modules of the transfer text generation model in Figure 6 is 4, the input and output dimensions of the sub-layers are 512, and the number of attention heads of the multi-head attention is 8; the pre-trained BERT model used to obtain the word embedding representation uses BERT-Medium with dimension size 512; the number of hidden units of Bi-LSTM is 512; the model training was performed using Adam optimizer with custom learning rate [11]; the number of iterations (epochs) trained on each domain was 1000; all experiments in this paper were implemented using Python 3.8 and TensorFlow-GPU 2.5.0, and the experimental platform was configured with Windows 10 operating system, NVIDIA 2080Ti GPU graphics card, 32GB RAM, and Intel Core i7-11700K CPU.

    To evaluate the effectiveness of the transfer text generation model proposed in this paper when applied to the task of news headline generation, the transfer text generation model presented in this paper is compared with existing pre-trained language models and text generation models related to zero/few-shot learning that perform well.

    For the pre-trained language models, T5 [17], BART [18], PEGASUS [19], and BertSum [40] were selected. All these four pre-trained language models used the pre-training parameters as the initial parameters of the models. The training was continued based on the pre-training initial parameters for these four models using the data in Table 2 without changing the other hyperparameters.

    For zero/few shot text generation models, TransferRL [41], ZSDG [32], DAML [42], and MTL-ABS [43] are chosen. Among them, TransferRL contains a decoder shared among different domains and maximizes the "reward" of generalizing the decoder to different domains by reinforcement learning self-critic strategy to improve the domain adaptation of the model so that it only needs to be fine-tuned on small batches of data for fast adaptation to the target domain. ZSDG uses zero-shot learning to generate transferred text with zero data in the target domain by projecting "seed-level" data descriptions into a subspace and then transferring the semantic descriptions at the domain level. DAML and MTL-ABS build generative models using a sequence-to-sequence format based on Meta-Learning principles. Still, DAML uses Gate Recurrent Neural Network (GRU) as encoder and decoder, while MTL-ABS uses Transformer as encoder and decoder. The above two models search for the most promising parameter fetches for the model from the gradient optimization level using meta-learning, which makes the model more responsive to small data in the target domain and improves the domain generalization of the model. In contrast to the pre-trained language models, both zero/few-shot learning models directly use the data in Table 2 to train the models according to their respective transfer strategies.

    The generation effectiveness of the above comparison models was evaluated using the evaluation metrics ROUGE-1/2/L [35], BLEU [35], and METEOR [35], which are commonly used in text generation tasks. The news body in the target domain is input to the trained model. The scores of evaluation metrics between the news headlines generated by the model and the corresponding ground-truth news headlines are calculated. Among them, the ground-truth news headlines in the target domain are only used for evaluation and are not involved in the model training process. Based on the above metrics scores, we investigate whether the proposed transferred text generation model can effectively obtain relevant reference knowledge from the source domain data to effectively assist the target domain in completing the text generation task without giving the reference ground-truth data of the target domain text.

    To demonstrate the actual effect of the internal mechanism of each stage of the proposed transfer text generation model more directly, the alignment effect of the data distribution in the domain of Figure 4 is further demonstrated with the "kid" news topic as the target domain, as shown in Figure 8. In Figure 8, the source and target domain data are mapped according to Eq (6); in Figure 8, the original embedding representations Xsrc and Xtar of the source and target domains (Figure 8, left panel), and the aligned representations Xsrc and Xtar obtained by Eq (6) (Figure 8, right panel) are represented by Principal Component Analysis (PCA) method for dimensionality reduction.

    Figure 8.  Comparison of the effect of domain distribution alignment with "kid" news topic as the target domain. (Note: the left figure indicates the original distribution before alignment, and the right figure indicates the alignment distribution after alignment).

    Specifically, in Figure 8, the data representations for different domains are shown in different colors. The top dark blue area indicates the data distribution when the "kid" news topic is the target domain. In Figure 8, the left panel shows the original representation distribution of the text data of the eight domains, which is output by the pre-trained BERT model without any cross-feature filling and data distribution alignment. It can be found that there are significant differences in the original representation distributions for the given eight domains. After that, the other seven domains, except "kid, " are used as source domains, as shown on the right side of Figure 8. The data in the source domain and the news data in the target domain "kid" are first filled with cross features between the source and target domains according to Eqs (1)–(3); on this basis, the alignment of the domain data distribution is done by Eq (6) according to the process shown in Figure 4, and the result is shown in the right side of Figure 8. After the domain data distribution alignment, although there is still a slight difference between the source and target domain data, the distribution difference between different domains has been significantly reduced. Comparing the left plot and right plot in Figure 8, it can be found that the proposed model involves the alignment of domain data distribution between different domains by first using cross-filling to fill the features for the source and target domain data and then minimizing the maximum mean difference distance metric between the source and target domains, which effectively reduces the data distribution difference between the source and target domains.

    For zero-shot learning semantic prototype transduction, one of the eight domains listed in Table 2 is selected as the target domain, and the remaining seven are used as source domains. The seven source domains and one target domain are formed into K intermediate domains, as shown in Figure 5 for experiments according to the intermediate domain redistribution. In the process of target domain rotation, the silhouette coefficient commonly used in the K-means method [38] is used to evaluate the effect of intermediate domain redistribution under different values of K to determine the value of K. Currently, the value of K does not exceed the number of source domains. The range of the silhouette coefficient is [-1, 1]; if the value of the silhouette coefficient tends to 1, it means that the cohesion and separation are relatively better and the clustering effect is better, thus determining the number of intermediate domains.

    Figure 9 shows that the size of the silhouette coefficients is determined by Algorithm 2 for different values of K for each domain as the target domain. The optimal number of intermediate domains is the point with the most considerable K value of the silhouette coefficient. After obtaining the optimal number of intermediate domains K taken for each domain as the target domain, the scores of ROUGE-1/2/L, BLEU, and METEOR metrics in Table 3 are calculated from the news headlines generated by the model and the corresponding headline ground-truth data under the corresponding intermediate domain redistribution scheme after each target domain is determined in the round-robin experiment. Specifically, the effect of text generation in each target domain is first evaluated. In this case, only the ground-truth news headline data in the source domain are involved in the model training, no headline ground-truth data are involved in the target domain, and only the pseudo news headline text extracted from the news body is used in the target domain. As a result, based on the distribution-aligned representation of domain data obtained by Eq (6) and the zero-shot learning semantic prototype transduction by Eqs (14) and (15), the news body in each target domain can generate news headlines directly without relying on any manually ground-truth values.

    Figure 9.  Silhouette coefficient for different values of the number of intermediate domains for each domain as a target domain.
    Table 3.  Comparison of news headline generation performance of transfer text generation models in different domains.
    No. Target Domain K ROUGE-1 ROUGE-2 ROUGE-L BLEU METEOR
    1 Sport 6 74.52 59.33 73.83 43.51 68.84
    2 Health 7 79.03 64.38 78.25 49.95 73.34
    3 Finance 7 76.36 60.99 75.82 48.36 70.02
    4 Music 6 72.54 61.53 72.24 47.02 67.76
    5 Weather 5 77.60 62.00 76.92 45.45 70.80
    6 Movie 2 63.65 41.70 62.19 24.51 55.04
    7 Auto 5 78.20 64.05 77.68 49.59 73.04
    8 Kid 4 75.15 59.44 74.28 45.28 69.48
    Note: bold indicates the top 3 domain data in terms of text generation evaluation metric score; underline indicates the data with the best text generation evaluation metric score.

     | Show Table
    DownLoad: CSV

    Table 3 lists the performance of the transfer text generation model proposed in this paper applicable to semantic prototype transduction for news headline generation in different target domains. As we can see, the performance of indicators in all fields is relatively stable except for "Movie"; "Health", "Car" and "Weather" rank in the top 3 in terms of performance. As a result, although the model does not refer to the headline truth data in the target domain in the process of generating training, the semantic correlation of data between different domains is obtained through the alignment of domain data distribution used according to Eq (6) in Figure 4 and the semantic prototype transduction transfer based on (news x, headline y) in Figure 7, which results in better scores on each evaluation index in the process of rotating through different target domains. This phenomenon can be attributed to the fact that, firstly, based on Figure 4, the differences in the distribution of data across domains are reduced after the alignment of the domain data distribution. Therefore, the negative impact of the differences in the distribution of data across domains can be reduced during the transfer of the model from the source domain to the target domain. Then, through zero-shot learning semantic prototype transduction, the transfer text generation model proposed in this paper simultaneously acquires the semantic correlations between different domain data through the attention mechanism and temporal dependency in the enhanced encoder and decoder in Figure 6 to adjust the model parameters to improve the model domain transfer effect.

    Further, Figure 9 shows the training performance of the text generation model during the zero-shot learning semantic prototype transduction phase when part of the domain is used as the target domain. In this phase, the model is trained through 1000 epochs by the loss function Lmain defined in Eq (16). The accuracy is calculated as the ratio of identical words between the generated text and the ground-truth text at each time step of the generated text. As can be seen from Figure 9, even for the three domains with the lowest evaluation index of text generation, the loss function Lmian in training is gradually decreasing, which proves that the model can fully parse the semantic prototypes of each domain data by the loss function Lxy designed for (xd,yd) in the semantic prototype z and the loss function Lay designed for (ad,yd) in the target domain without the ground-truth data, so that the model captures the correlation between the semantic prototypes of different domain data during the generation process, and thus performs effective transfer from the source domain to the target domain. The smooth increase of accuracy proves the accuracy of the text generated by the transfer text generation model proposed in this paper after transfer from the source domain to the target domain, in which the pointer generator network is responsible for handling the out-of-vocabulary (OOV) problem, which further improves the text quality.

    Figure 10.  Effect of transferable text generation model trained by semantic prototype transduction strategy: loss function curve vs. accuracy curve.

    From Table 3, we can see that the best performance of transfer text generation is achieved when the three domains of "Health, " "Auto, " and "Weather" are used as target domains. Therefore, further ablation experiments are conducted on the proposed transfer text generation method using these three domains, and the results are shown in Table 4.

    Table 4.  Results of the ablation experiments of the transferred text generation model.
    Target Domain ID Ablation combination K ROUGE-1 ROUGE-2 ROUGE-L BLEU METEOR
    Health 1 Semantic prototype transduction 0 72.36 59.85 71.34 42.85 61.47
    2 Intermediate Domain Redistribution + Semantic prototype transduction 7 77.38 63.12 76.43 46.62 68.67
    3 Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction 7 79.03 64.38 78.25 49.95 73.34
    Auto 1 Semantic prototype transduction 0 57.19 55.72 56.53 36.90 46.45
    2 Intermediate Domain Redistribution + Semantic prototype transduction 5 68.98 60.78 68.36 43.31 60.63
    3 Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction 5 78.20 64.05 77.68 49.59 73.04
    Weather 1 Semantic prototype transduction 0 57.04 48.43 55.71 39.72 39.52
    2 Intermediate Domain Redistribution + Semantic prototype transduction 5 68.58 55.87 67.53 42.62 55.82
    3 Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction 5 77.60 62.00 76.92 45.45 70.80
    Note: bold indicates the data with higher scores under each evaluation index in the same ablation combination.

     | Show Table
    DownLoad: CSV

    In Table 4, "Semantic prototype transduction" indicates that the original representation of the pre-trained BERT model output is directly used without intermediate domain redistribution. The model is directly trained using the semantic prototype transduction based on Eqs (14)–(16) in Figure 7; "Intermediate Domain Redistribution + Semantic prototype transduction" indicates that the original representation of the output of the pre-trained BERT model is directly adopted, and the model is trained using the semantic prototype transduction based on Eqs (14)–(16) in Figure 7, after the intermediate domain redistribution according to the optimal number of intermediate domains, "Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction" indicates that the model is trained with the semantic prototype transduction based on Eqs (14)–(16) in Figure 7 after dividing the intermediate domain by the optimal number of intermediate domains based on the distribution-aligned data representation of Eq (6) in Figure 4.

    From the experimental results in Table 4, the text generation effect is better than that of using the original representation directly after adopting the distribution representation alignment method in each target domain, which means that the distribution alignment of domain data can effectively eliminate the data distribution differences between domains and improve the transferability from the source domain to the target domain. In addition, comparing the experimental results in Table 4 with those in the comparison experiments in Table 5 below, the model proposed in this paper can also obtain higher evaluation metric scores compared with most other transfer text generation models using only the semantic prototype transduction method for training. This phenomenon shows that in the transfer scheme proposed in this paper, zero-shot learning semantic prototype transduction explores data semantic correlation between different domains. The enhanced encoder and decoder in the "encoder-decoder" structure correlate the no-ground-truth news text in the target domain with the most relevant news headlines in the source domain and obtain semantic prototype similarity or proximity according to the attention mechanism and temporal dependence and derive the reference of the target domain to the source domain data in text generation, thus improving the text generation effect of transfer.

    Table 5.  Performance comparison between transfer text generation models and existing other transfer text generation models in the "Movie" domain.
    Category Model/Method ROUGE-1 ROUGE-2 ROUGE-L BLEU METEOR
    Pre-training Model (Comparison Models) T5 45.34 30.25 43.27 23.63 41.06
    BART 35.73 20.06 33.78 18.81 33.91
    PEGASUS 35.23 29.62 33.29 18.48 33.49
    BertSum 39.57 20.70 27.29 14.51 38.83
    Zero/Few-shot Learning Models (Comparison Models) MTL-ABS 39.56 34.74 38.68 19.16 32.55
    TransferRL 32.79 31.32 28.10 14.39 26.40
    ZSDG 30.89 33.03 23.91 11.67 30.98
    DAML 34.90 32.09 22.14 13.86 27.33
    Transfer Text Generation Model (Our Model) Semantic prototype transduction 56.98 37.17 55.28 17.41 43.17
    Intermediate Domain Redistribution + Semantic prototype transduction 62.00 40.44 60.37 11.18 50.37
    Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction 63.65↑ 41.70↑ 62.19↑ 24.51↑ 55.04↑
    Performance Improvement Rate (%) 18.31 11.45 18.92 0.88 13.99
    Note: bold indicates the highest score for each evaluation metric in each group of model categories.

     | Show Table
    DownLoad: CSV

    In addition, it can be seen from the data in the table that the combination of "Intermediate Domain Redistribution + Semantic prototype transduction" achieves higher evaluation index scores than the "Semantic prototype transduction" method only, which indicates that the "intermediate domain redistribution" by the content similarity composite index, the target domain text is generated based on the relevant source domain data with more semantic similarity, achieving the better performance of transferred text generation. Meanwhile, the complete "Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction" approach in Table 4 can achieve the optimal text generation results of the model, which means that the model can obtain the optimal transferred text generation performance on the target domain by getting the composite transfer strategy of distribution alignment representation of domain data and zero-shot learning semantic prototype transduction via Eq (16) while obtaining helpful information from the relevant source domain in the intermediate domain without reference to the ground-truth data in the target domain. The pointer generator network will also improve the accuracy of the generated text.

    As the experimental results in Table 3 show, the text generation performance of the model is the worst when the "Movie" domain is the target domain. Therefore, we further compare the performance of the proposed text generation model approach for zero-shot learning semantic prototype transduction with other transfer text generation model approaches in terms of pre-training models (i.e., T5, BART, PEGASUS, BertSum) and "zero/few-shot learning models" (i.e., TransferRL, ZSDG, DAML, MTL-ABS) for the "Movie" domain. The results are shown in Table 5.

    All the models in Table 5 are trained with the intermediate domain data shown in Figure 5 after the domain data distribution alignment. All the models do not use the ground-truth data in the target domain during the training process. The performance improvement rate refers to the difference between the proposed "Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction" method in the performance evaluation metric score compared with the highest score in the comparison model.

    Specifically, with the "Movie" domain as the target domain, which is the least effective of our approach, the transfer text generation model proposed in this paper achieves the best performance in the comparison based on the evaluation metric scores, followed by the pre-trained language model approach, and finally the zero/few-shot learning model approach. This phenomenon can be attributed to the fact that the transfer scheme proposed in this paper first alleviates the data distribution differences between domains by aligning the domain data distribution at the text representation level according to Eq (6) based on Figure 4 and then improves the text generation model structure based on Figure 6 by making it more applicable to the zero-shot learning semantic prototype transduction according to Eq (15) in Figure 7 so that the model can more effectively acquire a priori knowledge from the relevant source domains that can help transfer and improve the model's text generation performance in the target domain.

    Second, furthermore, the pre-trained language models T5, BART, PEGASUS, and BertSum in Table 5 have been pre-trained in a large-scale corpus, so more prior knowledge has been incorporated into the parameters of such pre-trained language models in advance. However, the experimental results in Table 5 show that the scores of T5, BART, PEGASUS and BertSum for each evaluation metric are lower than those of the transfer approach proposed in this paper. This phenomenon can be attributed to the fact that although the pre-trained language model has acquired a large amount of domain prior knowledge through the large-scale corpus pre-training, this knowledge is not specific to the target domain and its underlying tasks. In contrast, the transfer text generation model proposed in this paper first reduces the distribution difference in data representation with other related source domain data from the target domain perspective by domain data distribution alignment, and establishes cross-domain semantic association xtarysrcasrcysrc, i.e., xtarysrc, based on semantic prototypes zsrc=[xsrc,ysrc,asrc] and ztar=[xtar,ytar,atar] by zero-shot learning semantic prototype transduction, which maximizes the semantic correlation between different domain data and exploits the semantic correlation. It ensures that the target domain can help the target domain generate text with the help of source domain data using semantic prototype transduction even if there is no reference ground-truth data, thus having better domain transfer adaptability for a specific target domain and the tasks under it.

    Third, for the zero/few-shot learning models TransferRL, ZSDG, DAML, and MTL-ABS in Table 5, these models use reinforcement learning, zero-shot learning, or meta-learning methods for transfer respectively. However, the results in Table 5 show that the scores of all evaluation metrics of these methods are lower than those of the transferred text generation models proposed in this paper. This phenomenon can be attributed to the structural improvements adopted in this paper for the transferred text generation model in Figure 6. Specifically, as shown in Figure 6, the improved text generation model can explore the semantic meaning of the text to a greater extent by adding a Bi-LSTM layer to resolve the text serialization dependencies. At the same time, the Transformer multi-head attention mechanism increases the internal contextual observation of the text. It processes the out-of-vocabulary (OOV) with the help of the pointer generator network. On this basis, by constructing data-level semantic prototypes, the no ground-truth news body in the target domain is associated with the most relevant news headlines in the source domain, and the semantic relatedness of the cross-domain text is captured according to the approximation on the semantic prototypes; thus, when the news body xtar in the target domain is given, the most relevant ground-truth news headlines ysrc in the source domain will be referenced to assist in generating news headlines ytar in the target domain, and thus the higher score performance on the evaluation metrics of ROUGE-1/2/L, BLEU, and METEOR.

    In this paper, we propose a transfer text generation model based on distribution alignment of domain data and zero-shot learning semantic prototype transduction for the task of news headline generation, whose primary principle is to assist the target domain in text generation with the help of data in the relevant source domain to overcome the problem of missing reference truth in the target domain. When the model is applied to a new domain, even if the data of the target domain (new domain) lacks the reference truth, the method proposed in this paper can fully utilize the existing source domain data information and the data information available in the target domain to solve the problem of model generalization caused by data when the common text generation model is applied in the new domain and realize the excellent application of the model in the new domain. The transfer text generation method proposed in this paper can be effectively applied to the task of news headline generation to solve the reference truth of data missing problems in the target domain by transferring the domain data information.

    We believe that the work in this paper still has some limitations in practical applications. First, the proposed method is mainly applied to tasks with several different but similar domains, and the performance of the proposed method may be reduced if the differences between domains are too significant. In addition, the proposed method is since the target domain provides a certain amount of data with no reference truth that can be used in the domain transfer method. The target domain will also affect the proposed method if it does not provide any data. Therefore, through our research, we believe several aspects of future work deserve further exploration. First, we would like to investigate further the effect of the proposed method on transfer text generation between different domains that differ significantly. Second, to further investigate how to supplement new data when there is a lack of available data in a new domain. Third, selecting relevant source domains is crucial to the final transfer generation performance when the target domain is given. More suitable methods for domain data selection need to be further investigated. Fourth, the source domain data often provide noisy information that is not relevant to the target domain during the transfer process, which affects the transfer effect and brings about the negative transfer problem, so avoiding the "negative transfer" problem is also a direction worthy of further research.

    This work was supported in part by National Science Foundation of China (No. 62102187), Jiangsu Natural Science Foundation (Basic Research Program) (BK20210639) and National Key Research and Development Program (2021YFE0104400). This research project was also supported by a grant from the Research Center of the Female Scientific and Medical Colleges, Deanship of Scientific Research, King Saud University, Saudi Arabia.

    The authors declare no conflict of interest.

    Table A1.  Results of the ablation experiments of the transferred text generation model for all domains. (The full version of Table 4).
    Target domain ID Ablation combination K ROUGE-1 ROUGE-2 ROUGE-L BLEU METEOR
    Health 1 Semantic prototype transduction 0 72.36 59.85 71.34 42.85 61.47
    2 Intermediate Domain Redistribution + Semantic prototype transduction 7 77.38 63.12 76.43 46.62 68.67
    3 Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction 7 79.03 64.38 78.25 49.95 73.34
    Auto 1 Semantic prototype transduction 0 57.19 55.72 56.53 36.90 46.45
    2 Intermediate Domain Redistribution + Semantic prototype transduction 5 68.98 60.78 68.36 43.31 60.63
    3 Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction 5 78.20 64.05 77.68 49.59 73.04
    Weather 1 Semantic prototype transduction 0 57.04 48.43 55.71 39.72 39.52
    2 Intermediate Domain Redistribution + Semantic prototype transduction 5 68.58 55.87 67.53 42.62 55.82
    3 Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction 5 77.60 62.00 76.92 45.45 70.80
    Finance 1 Semantic prototype transduction 0 66.78 55.13 64.21 39.56 50.43
    2 Intermediate Domain Redistribution + Semantic prototype transduction 7 71.98 54.42 65.12 41.50 61.32
    3 Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction 7 73.92 58.61 74.21 41.58 69.23
    Kid 1 Semantic prototype transduction 0 55.84 53.78 59.31 36.85 47.94
    2 Intermediate Domain Redistribution + Semantic prototype transduction 5 64.91 52.15 63.62 41.49 52.30
    3 Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction 5 66.01 57.77 66.04 40.26 63.43
    Sport 1 Semantic prototype transduction 0 61.16 52.29 56.83 36.43 46.61
    2 Intermediate Domain Redistribution + Semantic prototype transduction 6 58.77 56.46 60.40 40.40 50.32
    3 Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction 6 69.74 55.43 61.36 38.39 65.30
    Music 1 Semantic prototype transduction 0 54.56 50.43 52.83 36.20 43.55
    2 Intermediate Domain Redistribution + Semantic prototype transduction 6 63.68 53.78 56.24 39.88 47.61
    3 Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction 6 65.42 55.40 58.88 36.61 60.26
    Movie 1 Semantic prototype transduction 0 47.19 45.72 46.53 32.90 36.45
    2 Intermediate Domain Redistribution + Semantic prototype transduction 2 50.99 50.79 48.37 33.32 40.64
    3 Distribution Alignment + Intermediate Domain Redistribution + Semantic prototype transduction 2 58.21 54.06 57.69 34.59 53.05

     | Show Table
    DownLoad: CSV
    Table A2.  Examples of news headlines generated by the proposed model in different domains.
    Target Domain News Headline
    Sport Maury Wills, who helped the Los Angeles Dodgers win three World Series titles with his base-stealing prowess, has died. LOS ANGELES (AP) — Maury Wills, who intimidated pitchers with his base-stealing prowess as a shortstop for the Los Angeles Dodgers on three World Series championship teams, has died. He was 89. Wills died Monday night at home in Sedona, Arizona, the team said Tuesday after being informed by family members. No cause of death was given.
    Generation: Los Angeles Dodgers shortstop Maury Wills dies at home in Arizona.
    Reference: Maury Wills, Base-Stealing Shortstop for Dodgers, Dies At 89
    Finance Following the market's worst week since the 2008 financial crisis, the Federal Reserve announced early this month that it was slashing the federal rate by 0.50% to a new target range of 1% to 1.25%. The hope is that as the world economy slows due to problems caused by the coronavirus, lower interest rates will encourage Americans to keep spending and borrowing, helping the U.S. economy stay ahead of the damage. And as the coronavirus continues to disrupt manufacturing, supply chains, travel, and other important industries, it's rumored that the Fed may cut rates again ― all the way down to 0%.
    Generation: Now You Can Refinance Your Debt, here's what you should know
    Reference: Thanks To the Coronavirus, Now Is A Great Time To Refinance Your Debt
    Movie James Corden didn't turn tail on the bad reviews for his new movie "Cats." The film version of the Andrew Lloyd Webber musical has been deemed a "cat-tastrophe" by some outlets. HuffPost called it a "growling nightmare." Corden plays Bustopher Jones in the movie. "I've heard it's terrible, " Corden said on Zoe Ball's BBC 2 radio show Monday. Corden, host of "The Late Late Show, " said he had yet to see the film. "I'll catch it one day, I imagine, " he said. To which Ball replied, "We'll let you know what it's like, James." Corden was a guest on the broadcast to tout his "Gavin and Stacey Christmas Special" ― but the movie had to come up at some point. "Has anyone seen 'Cats'?" Ball asked to laughter, raising the subject. Generation: James Corden's "Cats" movie has been deemed a "cat-tastrophe"
    Reference: 'Cats' Star James Corden On The Film: 'I've Heard It's Terrible
    Weather Americans from Texas to Maine sweated out a steamy Saturday as a heat wave canceled events from festivals to horse races and pushed New York City to order power-saving steps to avoid overtaxing the electrical grid.
    Generation: Weather service says the worst heat Hits Much Of East, Central U.S.
    Reference: 'Dangerous Heat Wave' Hits Much Of East, Central U.S.
    Auto The prices of cars are going up. New vehicle prices in May 2019 climbed nearly 4 percent on average — or about $1,320 — over the past year, according to new data from Kelley Blue Book. And impending U.S. tariffs against Mexico may soon ramp prices up even further.
    Generation: The Trump administration is planning a 5 percent tariff on all goods coming into the U.S. from Mexico
    Reference: Prepare To Pay More For New Cars Thanks To Trump's Tariffs Against Mexico
    Music At the Stagecoach music festival over the weekend, hundreds of country music fans paid heartfelt tribute to the victims and survivors of the Route 91 and Borderline Bar mass shootings. And they did it in the best way they know how: with some song and a whole lot of dance.
    Generation: Country music that speaks of unity and strength in the face of heartache
    Reference: Watch How Country Music Fans Pay Heartfelt Tribute To Route 91, Borderline Victims
    Health Sri Lanka on Thursday lowered the death toll from the Easter suicide bombings by nearly one-third, to 253, as authorities hunted urgently for a least five more suspects and braced for the possibility of more attacks in the coming days.
    Generation: Sri Lanka reduces death toll from Easter suicide bombings by nearly one-third, to 253.
    Reference: Sri Lanka Lowers Death Toll From Easter Sunday Bombings To 253
    Kid "Saturday Night Live" was certainly making the most of Jason Momoa this week, with the "Aquaman" star featured in several of the show's sketches.
    Generation: "Saturday Night Live" with the "Aquaman" star featured in several of the show's sketches This Christmas
    Reference: 'SNL's' Elf On The Shelf Wants A New Kid This Christmas

     | Show Table
    DownLoad: CSV


    [1] X. Ao, X. Wang, L. Luo, PENS: A dataset and generic framework for personalized news headline generation, in Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, 1 (2021), 82–92. https://doi.org/10.18653/v1/2021.acl-long.7
    [2] F. Z. Zhuang, P. Luo, Q. He, Z. Z. Shi, Survey on transfer learning, J. Software, 26 (2015), 26–39. https://doi.org/10.13328/j.cnki.jos.004631 doi: 10.13328/j.cnki.jos.004631
    [3] H. Choi, J. Kim, S. Joe, Analyzing Zero-shot cross-lingual transfer in supervised NLP tasks, in 2020 25th International Conference on Pattern Recognition (ICPR), (2021), 9608–9613. https://doi.org/10.1109/icpr48806.2021.9412570
    [4] W. Wang, V. W. Zheng, H. Yu, A survey of Zero-shot learning: Settings, methods, and applications, ACM Trans. Intell. Syst. Technol., 10 (2019), 1–37. https://doi.org/10.1145/3293318 doi: 10.1145/3293318
    [5] N. Y. Wang, Y. X. Ye, L. Liu, L. Z. Feng, T. Bao, T. Peng, Advances in deep learning-based language modeling research, J. Software, 32 (2021), 1082–1115. https://doi.org/10.13328/j.cnki.jos.006169 doi: 10.13328/j.cnki.jos.006169
    [6] S. Bae, T. Kim, J. Kim, Summary level training of sentence rewriting for abstractive summarization, in Proceedings of the 2nd Workshop on New Frontiers in Summarization, (2019), 10–20. https://doi.org/10.18653/v1/d19-5402
    [7] K. Krishna, B. V. Srinivasan, Generating topic-oriented summaries using neural attention, in Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 1 (2018), 1697–1705. https://doi.org/10.18653/v1/n18-1153
    [8] T. Ma, H. Wang, Y. Zhao, Topic-based automatic summarization algorithm for Chinese short text, Math. Biosci. Eng., 17 (2020), 3582–3600. https://doi.org/10.3934/mbe.2020202 doi: 10.3934/mbe.2020202
    [9] S. Narayan, J. Maynez, J. Adamek, Stepwise extractive summarization and planning with structured transformers, preprint, arXiv: 1810.04805.
    [10] A. See, P. J. Liu, C. D. Manning, Get to the point: Summarization with pointer-generator networks, in Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, 1 (2017), 1073–1083. https://doi.org/10.18653/v1/p17-1099
    [11] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, et al., Attention is all you need, in Advances in Neural Information Processing Systems, (2017), 1–30.
    [12] P. F. Du, X. Y. Li, Y. L. Gao, Survey on multimodal visual language representation learning, J. Software, 32 (2021), 327–348. https://doi.org/10.13328/j.cnki.jos.006125 doi: 10.13328/j.cnki.jos.006125
    [13] S. Golovanov, R. Kurbanov, S. Nikolenko, Large-scale transfer learning for natural language generation, in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, (2019), 6053–6058. https://doi.org/10.18653/v1/p19-1608
    [14] J. J. Huang, P. W. Li, M. Peng, Q. Q. Xie, C. Xu, Research on deep learning-based topic models, Chin. J. Comput., 43 (2020), 827–855.
    [15] N. Dethlefs, Domain transfer for deep natural language generation from abstract meaning representations, IEEE Comput. Intell. Mag., 12 (2017), 18–28. https://doi.org/10.1109/mci.2017.2708558 doi: 10.1109/mci.2017.2708558
    [16] X. Qiu, T. Sun, Y. Xu, Pre-trained models for natural language processing: A survey, Sci. Chin. Technol. Sci., 63 (2020), 1872–1897. https://doi.org/10.1109/iceib53692.2021.9686420 doi: 10.1109/iceib53692.2021.9686420
    [17] C. Raffel, N. Shazeer, A. Roberts, Exploring the limits of transfer learning with a unified text-to-text transformer, J. Mach. Learn. Res., 21 (2020), 1–67.
    [18] M. Lewis, Y. Liu, N. Goyal, BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension, in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, (2020), 7871–7880. https://doi.org/10.18653/v1/2020.acl-main.703
    [19] J. Zhang, Y. Zhao, M. Saleh, PEGASUS: Pre-training with extracted gap-sentences for abstractive summarization, in International Conference on Machine Learning, (2020), 11328–11339.
    [20] Z. C. Zhang, M. Y. Zhang, T. Zhou, Pre-trained language model augmented adversarial training network for Chinese clinical event detection, Math. Biosci. Eng, 17 (2020), 2825–2841. https://doi.org/10.3934/mbe.2020157 doi: 10.3934/mbe.2020157
    [21] S. Chen, L. Han, X. Liu, Subspace distribution adaptation frameworks for domain adaptation, IEEE Trans. Neural Networks Learn. Syst., 31 (2020), 5204–5218. https://doi.org/10.1109/tnnls.2020.2964790 doi: 10.1109/tnnls.2020.2964790
    [22] H. Li, S. J. Pan, S. Wang, Heterogeneous domain adaptation via nonlinear matrix factorization, IEEE Trans. Neural Networks Learn. Syst., 31 (2020), 984–996. https://doi.org/10.1109/tnnls.2019.2913723 doi: 10.1109/tnnls.2019.2913723
    [23] W. Zellinger, B. A. Moser, T. Grubinger, Robust unsupervised domain adaptation for neural networks via moment alignment, Inf. Sci., 483 (2019), 174–191. https://doi.org/10.1016/j.ins.2019.01.025 doi: 10.1016/j.ins.2019.01.025
    [24] X. Glorot, A. Bordes, Y. Bengio, Domain adaptation for large-scale sentiment classification: A deep learning approach, in International Conference on Machine Learning, (2011), 513–520.
    [25] J. Blitzer, M. Dredze, F. Pereira, Biographies, bollywood, boom-boxes, blenders: Domain adaptation for sentiment classification, in Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, 7 (2007), 440–447.
    [26] F. Wu, Y. Huang, Sentiment domain adaptation with multiple sources, in Proceedings of the 54th Annual Meeting of the Association of Computational Linguistics, (2016), 301–310, https://doi.org/10.18653/v1/p16-1029
    [27] J. Blitzer, R. McDonald, F. Pereira, Domain adaptation with structural correspondence learning, in Proceedings of the 2006 Conference on Empirical Methods in Natural Language Processing, (2006), 120–128. https://doi.org/10.3115/1610075.1610094
    [28] J. Pan, X. Hu, P. Li, H. Li, W. He, Y. Zhang, Y. Lin, Domain adaptation via multi-layer transfer learning, Neurocomputing, 190 (2016), 10–24. https://doi.org/10.1016/j.neucom.2015.12.097 doi: 10.1016/j.neucom.2015.12.097
    [29] P. Wei, R. Sagarna, Y. Ke, Y. S. Ong, C. K. Goh, Source-target similarity modelings for multi-source transfer gaussian process regression, in Proceedings of the 34th International Conference on Machine Learning, (2017), 3722–3731.
    [30] N. Houlsby, A. Giurgiu, S. Jastrzebski, Parameter-efficient transfer learning for NLP, in PMLR, (2019), 2790–2799.
    [31] H. Zhang, L. Liu, Y. Long, Deep transductive network for generalized zero shot learning, Pattern Recogn., 105 (2020), 107370. https://doi.org/10.1016/j.patcog.2020.107370 doi: 10.1016/j.patcog.2020.107370
    [32] T. Zhao, M. Eskenazi, Zero-shot dialog generation with cross-domain latent actions, in Proceedings of the 19th Annual SIGdial Meeting on Discourse and Dialogue, (2018), 1–10. https://doi.org/10.18653/v1/w18-5001
    [33] Z. Liu, J. Shin, Y. Xu, Zero-shot cross-lingual dialogue systems with transferable latent variables, preprint, arXiv: 1911.04081.
    [34] Ayana, S. Shen, Y. Chen, Zero-shot cross-lingual neural headline generation, IEEE/ACM Trans. Audio Speech Lang. Process., 26 (2018), 2319–2327. https://doi.org/10.1109/taslp.2018.2842432 doi: 10.1109/taslp.2018.2842432
    [35] X. Duan, M. Yin, M. Zhang, Zero-shot cross-lingual abstractive sentence summarization through teaching generation and attention, in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, (2019), 3162–3172. https://doi.org/10.18653/v1/p19-1305
    [36] J. Devlin, M. W. Chang, K. Lee, BERT: Pre-training of deep bidirectional transformers for language understanding, preprint, arXiv: 1810.04805.
    [37] X. T. Song, H. L. Sun, A review of neural network-based automatic source code abstraction techniques, J. Software, 33 (2022), 55–77. https://doi.org/10.13328/j.cnki.jos.006337 doi: 10.13328/j.cnki.jos.006337
    [38] P. J. Rousseeuw, Silhouettes: A graphical aid to the interpretation and validation of cluster analysis, J. Comput. Appl. Math., 20 (1987), 53–65. https://doi.org/10.1016/0377-0427(87)90125-7 doi: 10.1016/0377-0427(87)90125-7
    [39] Z. Huang, P. Xu, D. Liang, TRANS-BLSTM: Transformer with bidirectional LSTM for language understanding, preprint, arXiv: 2003.07000.
    [40] Y. Liu, M. Lapata, Text summarization with pretrained encoders, preprint, arXiv: 1908.08345.
    [41] K. Yaser, R. Naren, K. R. Chandan, Deep transfer reinforcement learning for text summarization, in Proceedings of the 2019 SIAM International Conference on Data Mining, (2019), 675–683. https://doi.org/10.1137/1.9781611975673.76
    [42] K. Qian, Z. Yu, Domain adaptive dialog generation via meta learning, in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, (2019), 2639–2649. https://doi.org/10.18653/v1/p19-1253
    [43] Y. S. Chen, H. H. Shuai, Meta-transfer learning for low-resource abstractive summarization, preprint, arXiv: 2102.09397.
  • This article has been cited by:

    1. Han Ren, Xiaona Chang, Xia Li, Neural headline generation: A comprehensive survey, 2025, 635, 09252312, 129633, 10.1016/j.neucom.2025.129633
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(12542) PDF downloads(110) Cited by(1)

Figures and Tables

Figures(10)  /  Tables(7)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog