Loading [MathJax]/jax/output/SVG/jax.js
Research article

Transferring monolingual model to low-resource language: the case of Tigrinya


  • In recent years, transformer models have achieved great success in natural language processing (NLP) tasks. Most of the current results are achieved by using monolingual transformer models, where the model is pre-trained using a single-language unlabelled text corpus. Then, the model is fine-tuned to the specific downstream task. However, the cost of pre-training a new transformer model is high for most languages. In this work, we propose a cost-effective transfer learning method to adopt a strong source language model, trained from a large monolingual corpus to a low-resource language. Thus, using the XLNet language model, we demonstrate competitive performance with mBERT and a pre-trained target language model on the cross-lingual sentiment (CLS) dataset and on a new sentiment analysis dataset for the low-resource language Tigrinya. With only 10k examples of the given Tigrinya sentiment analysis dataset, English XLNet achieved 78.88% F1-Score, outperforming BERT and mBERT by 10% and 7%, respectively. More interestingly, fine-tuning (English) XLNet model on the CLS dataset showed promising results compared to mBERT and even outperformed mBERT for one dataset of the Japanese language.

    Citation: Abrhalei Tela, Abraham Woubie, Ville Hautamäki. Transferring monolingual model to low-resource language: the case of Tigrinya[J]. Applied Computing and Intelligence, 2024, 4(2): 184-194. doi: 10.3934/aci.2024011

    Related Papers:

    [1] Juan L. G. Guirao, Pshtiwan Othman Mohammed, Hari Mohan Srivastava, Dumitru Baleanu, Marwan S. Abualrub . Relationships between the discrete Riemann-Liouville and Liouville-Caputo fractional differences and their associated convexity results. AIMS Mathematics, 2022, 7(10): 18127-18141. doi: 10.3934/math.2022997
    [2] Shuqin Zhang, Jie Wang, Lei Hu . On definition of solution of initial value problem for fractional differential equation of variable order. AIMS Mathematics, 2021, 6(7): 6845-6867. doi: 10.3934/math.2021401
    [3] Khalid K. Ali, K. R. Raslan, Amira Abd-Elall Ibrahim, Mohamed S. Mohamed . On study the fractional Caputo-Fabrizio integro differential equation including the fractional q-integral of the Riemann-Liouville type. AIMS Mathematics, 2023, 8(8): 18206-18222. doi: 10.3934/math.2023925
    [4] Abdul Samad, Imran Siddique, Zareen A. Khan . Meshfree numerical approach for some time-space dependent order partial differential equations in porous media. AIMS Mathematics, 2023, 8(6): 13162-13180. doi: 10.3934/math.2023665
    [5] Ravi P. Agarwal, Snezhana Hristova . Stability of delay Hopfield neural networks with generalized proportional Riemann-Liouville fractional derivative. AIMS Mathematics, 2023, 8(11): 26801-26820. doi: 10.3934/math.20231372
    [6] Adisorn Kittisopaporn, Pattrawut Chansangiam . Approximate solutions of the 2D space-time fractional diffusion equation via a gradient-descent iterative algorithm with Grünwald-Letnikov approximation. AIMS Mathematics, 2022, 7(5): 8471-8490. doi: 10.3934/math.2022472
    [7] Deepak B. Pachpatte . On some ψ Caputo fractional Čebyšev like inequalities for functions of two and three variables. AIMS Mathematics, 2020, 5(3): 2244-2260. doi: 10.3934/math.2020148
    [8] Ravi Agarwal, Snezhana Hristova, Donal O'Regan . Integral presentations of the solution of a boundary value problem for impulsive fractional integro-differential equations with Riemann-Liouville derivatives. AIMS Mathematics, 2022, 7(2): 2973-2988. doi: 10.3934/math.2022164
    [9] Snezhana Hristova, Antonia Dobreva . Existence, continuous dependence and finite time stability for Riemann-Liouville fractional differential equations with a constant delay. AIMS Mathematics, 2020, 5(4): 3809-3824. doi: 10.3934/math.2020247
    [10] Erdal Bas, Ramazan Ozarslan . Theory of discrete fractional Sturm–Liouville equations and visual results. AIMS Mathematics, 2019, 4(3): 593-612. doi: 10.3934/math.2019.3.593
  • In recent years, transformer models have achieved great success in natural language processing (NLP) tasks. Most of the current results are achieved by using monolingual transformer models, where the model is pre-trained using a single-language unlabelled text corpus. Then, the model is fine-tuned to the specific downstream task. However, the cost of pre-training a new transformer model is high for most languages. In this work, we propose a cost-effective transfer learning method to adopt a strong source language model, trained from a large monolingual corpus to a low-resource language. Thus, using the XLNet language model, we demonstrate competitive performance with mBERT and a pre-trained target language model on the cross-lingual sentiment (CLS) dataset and on a new sentiment analysis dataset for the low-resource language Tigrinya. With only 10k examples of the given Tigrinya sentiment analysis dataset, English XLNet achieved 78.88% F1-Score, outperforming BERT and mBERT by 10% and 7%, respectively. More interestingly, fine-tuning (English) XLNet model on the CLS dataset showed promising results compared to mBERT and even outperformed mBERT for one dataset of the Japanese language.



    In much of the literature, time fractional models are defined using the Caputo definition [32,33,34,35,36], in which time fractional models are models described by fractional differential equations or pseudo state space descriptions. The Caputo definition is widely acclaimed because it makes it possible to define initial conditions that relate to the integer derivatives of the derived functions in the models considered. However, this paper shows that this definition does not take initial conditions properly into account if used to define a time fractional model.

    The problem was analysed for the first time by Lorenzo and Hartley [1,2]. To take the past of the model into account in a convenient way in a finite interval, they introduced an initialization function. The idea of replacing the commonly used initial values by an initial function was further developed in [3]. In [4], the need to consider the "prehistories" before the initial instant of the derivate functions was shown, making it possible to address the initialisation of fractional visco-elastic equations to reach a unique solution. In [5,6], a counter example was used to demonstrate that initial conditions cannot be correctly taken into account in a dynamical model whether by Caputo or Riemann-Liouville definitions. This led to the conclusion in [7] that fractional derivative and time fractional model initializations are two distinct problems. Still using an initial time shifting method, counter examples were proposed in [8] to show similar initialisation problems with the Caputo definition for partial differential equations. A time shifting technique was also recently used in [9] to analyse a groundwater flow model with time Caputo or Riemann-Liouville fractional partial derivatives. The non-objectivity of these models was demonstrated in this paper. The authors in [9] did not address the problem of initialization, but this objectivity can be restored by also introducing an initialization function (instead of initial conditions).

    As previously mentioned, several studies and several solutions have already been published on initialisation of fractional models, but many papers in which the initial conditions are taken into account incorrectly are also still published. Thus the novelties and the contributions of the paper are new demonstrations and new simulations that highlight how initialisations must be done with a time fractional model. Thus, in this paper, two examples are used to show that the Caputo definition does not enable initial conditions to be correctly handled when this definition is used to define a time fractional model. In the first example, the response of a simple model, assumed to be at rest, is calculated analytically on a given time interval. Then inside this interval, a second response is computed by considering initial conditions resulting from the first simulation, and ignoring the model past before the considered initial time. This is the initialisation currently found in the literature and this example shows that it is unable to ensure the correct model trajectories. In the second example, two different histories are generated that produce the same initial conditions for the model. This example shows that in spite of equal initial conditions, the model response is different, thus showing that all the model past must be taken into account to define its future. A similar analysis is also carried out with the Riemann-Liouville and the Grünwald-Letnikov's definitions, suggesting that other definitions should also be problematic. Note that all the analyzes carried out and conclusions obtained in this paper relate to models involving only time fractional derivatives and not space fractional derivatives as in [29,30,31].

    The fractional integral of order ν, 0<ν<1, of a function y(t) is defined by [10]:

    Iνt0y(t)=1Γ(ν)tt0y(τ)(tτ)1νdτ. (1)

    Γ(.) being Euler's gamma function. From this definition, the Caputo derivative definition of order ν, 0<ν<1, of a function y(t) is defined by [11]:

    CDνt0y(t)=I1νt0(ddty(t))=1Γ(1ν)tt01(tτ)νdy(τ)dτdτ. (2)

    Laplace transform applied to relation (2) reveals how initial conditions are associated to this definition:

    L{CDνt0y(t)}=1s1ν(sY(s)y(t0)) = sνY(s)y(t0). (3)

    To demonstrate that Caputo definition is not able to take initial conditions correctly into account when used to define a time fractional model (a fractional differential equation or a pseudo state space description), the following model is considered

    Dνy(t)=ay(t)+u(t)  0<ν<1  a>0. (4)

    In relation (4), Dν denotes the Caputo definition in this section but denotes the Riemann-Liouville or Grünwald-Letnikov definitions in the next section. Then, the following algorithm is used to study model (4).

    Algorithm 1

    1-Simulation on the time interval [0,t1] of the time fractional model (for instance model (4)) with null initial conditions (for t],0]). Let S1 denote this simulation.

    2-Record the model output y(t) and the integer derivatives of y(t) (denoted y'(t), y''(t), ….) at time t0 such that 0<t0<t1.

    3-Simulate the model again on [t0,t1], using y(t0), y'(t0), y''(t0) … as initial conditions. Let S2 denote this simulation.

    4-Compare S1 and S2 on [t0,t1] and notice if they are different.

    Algorithm 1 is now applied to model (4) with a=0. The model is assumed to be at rest before t=0, and the input u(t) is assumed to be a Heaviside function H(t). In such conditions, relation (4) is equivalent to [11]

    y(t)=y(t0)+Iνt0{H(t)}  0<ν<1. (5)

    As a consequence, the simulation defined in Algorithm 1 provides the following solutions:

    S1:y(t)=tνΓ(ν+1)  0<t<t1 (6)
    S2:y(t)=(tt0)νΓ(ν+1)+y(t0)  t0<t<t1. (7)

    Figure 1 proposes a comparison of S1 and S2 and reveals a difference, thus demonstrating that the Caputo definition does not correctly take initial conditions into account.

    Figure 1.  Comparison of the exact response of model (4) with the responses obtained with Caputo definitions with initial conditions (t0=5s, a=0, ν=0.6).

    Another way to illustrate this result is to consider two different input signals u1(t) and u2(t) that create two different histories with:

    ui(t)=AiH(t+ti)AiH(t)  with  ti>0,  i={1,2}. (8)

    The model is assumed to be at rest on t],ti]. A constraint is also imposed on these signals so that at t=0, the two resulting model outputs coincide:

    y1(0)=y2(0). (9)

    The output yi(t) is thus defined by:

    yi(t)=Aia(1E1ν,1(a(t+ti)ν))H(t+ti)Aia(1E1ν,1(atν))H(t). (10)

    where Eγα,β(z) is the Mittag-Leffler function defined by [12]:

    Eγα,β(z)=k=0Γ(γ+k)Γ(αk+β)Γ(γ)zkk!. (11)

    Condition (9) thus leads to

    A1a[(1E1ν,1(a(t1)ν))(1E1ν,1(0))]=A2a[(1E1ν,1(a(t2)ν))(1E1ν,1(0))] (12)

    thus leading to the condition:

    A1=A21E1ν,1(a(t2)ν)1E1ν,1(a(t1)ν). (13)

    With ν=0.4, a=1, t1=8s, t2=2s, A2=5 and thus A14.17, Figure 2 shows the signal inputs u1(t) and u2(t) used for the analysis and proposes a comparison of the resulting outputs. This figure shows that the two responses have the same values at t=0, but that the evolutions for t>0 are not the same. The information at t=0 is thus not enough to predict the future of the model. All the past must be taken into account to predict the future of the model, which confirms that initialization as defined by the Caputo definition is not acceptable if used to define a time fractional model such as (4).

    Figure 2.  Comparison of the responses y1(t) and y2(t) of model (4) to two inputs that provide the same initial conditions.

    The previous section showed that the Caputo definition should no longer be used to define time fractional models such as (4). What about other definitions?

    The Riemann-Liouville derivative of order ν, 0<ν<1, of a function y(t) is defined by [11]:

    RLDνt0y(t)=1Γ(1ν)ddttt0y(τ)(tτ)νdτ. (14)

    Laplace transform applied to relation (14) reveals how initial conditions are associated to this definition:

    L{RLDνt0y(t)}=L{ddt(1Γ(1ν)tt0y(τ)(tτ)νdτ)}=s1s1νY(s)[I1νt0y(t)]t=t0. (15)

    As a consequence, in [11,13], the initialisation of relation (4) is defined by

    dνdtνy(t)=ay(t)+u(t)  I1νt0{y(t)}|t=t0=y0 (16)

    and thus the initialisation problem of relation (4) is equivalent to the integral equation

    y(t)=y0Γ(ν)(tt0)ν1+Iνt0{ay(t)+u(t)}. (17)

    Algorithm 1 is applied again to model (4) with a=0. The model is assumed to be at rest before t=0, and the input u(t) is assumed to be a Heaviside function H(t). Algorithm 1 provides the following solutions:

    S1:y(t)=tνΓ(ν+1)  0<t<t1 (18)
    S2:y(t)=(tt0)νΓ(ν+1) + y0Γ(ν)(tt0)ν1  t0<t<t1. (19)

    Relation (19) seems to say that any value of y0 can be chosen, but whatever the value selected, for S2 y(t) tends toward infinity as t tends toward t0 if y00 whereas y(t)=t0ν/Γ(ν+1) for S1. The two simulations thus give different results. This is illustrated by Figure 3 for various values of y0.

    Figure 3.  Comparison of the exact response of model (4) with the responses obtained with the Riemann-Liouville definition (t0 = 5s, a=0,ν=0.7).

    The Grünwald-Letnikov derivative of order ν, 0<ν<1, of a function y(t) is defined by:

    GLDνt0y(t)=h0_lim1hν0m<(1)m(νm)f(tmh)  t>t0 (20)

    with (νm)=Γ(ν+1)m!Γ(νm+1)=ν(ν1)(ν2)(νm+1)m(m1)(m2)(mm+1).

    This definition is often used in the literature as it provides a simple numerical scheme for fractional derivative implementation. In some research [14,15,16], these numerical schemes are used to solve the initialisation problem:

    Dvt0y(t)=ay(t)+u(t)0<v<1a>0 for t0<v<T, (21)
    y(t0)=y0.

    In this case, it is not the Grünwald-Letnikov derivative definition which is questionable, but the idea that a time fractional model can be initialized solely with information on the initial moment. From relation (20), it is possible to observe that variable m goes from 0 to infinity, and thus this definition is able to take into account the past of the derivative function, prior to t0. In (21), the problem is the way the initial conditions are defined.

    To illustrate this problem, Algorithm 1 is applied to model (4) with a=1. The model is assumed to be at rest before t=0, and the input u(t) is assumed to be a Heaviside function H(t). In such conditions, the simulation S1 defined in Algorithm 1 provides the following solution:

    S1:y(t)=tνΓ(ν+1)  0<t<t1. (22)

    Simulation S2 is done using the Grünwald-Letnikov formula (20) and provides

    S2:y(t)=1hν1m<(1)m(νm)y(tmh)+H(t)1hν+1  t>t0. (23)

    This simulation is done under two conditions:

    - S21: by taking into account all the past of the model (all the values of y(t) on t[0,t0], provided by S1)

    - S22: by considering only an initial condition at t0 (value of y(t) at t0 provided by S1).

    The comparison of the three simulations is done in Figure 4 and reveals that the Grünwald-Letnikov definition produces an exact solution provided that all the past of the model is taken into account.

    Figure 4.  Comparison of the exact response of model (4) with the responses obtained with the Grünwald-Letnikov definition (t0 = 3s, a=1,ν=0.6).

    Relation (23) is particularly interesting because it shows that a time fractional model (here a fractional integrator) is represented by an infinite difference equation, and therefore an initialization of all its terms is necessary for a prediction of the output y(t).

    This remark could also apply to the Caputo and Riemann-Liouville definitions which would lead to their reformulations with integrals on the interval ],t] as suggested in [27],

    The need to take into account the all past of a time fractional model and not just the knowledge of its pseudo state at a single point in the past can be demonstrated quite simply on relation (4) (a particular case of fractional differential equation or of pseudo state space description). Contrary to what relation (4) might suggest, Figure 5 highlights that the implementation of fractional differential equations does not explicitly involve the fractional differentiation operator but the fractional order integration operator Iν. Thus in practice, it is not necessary to specify which particular definition is used for Dν in equation (4). Moreover, even if the system is assumed to have zero initial conditions at t=0, namely if the system is supposed at rest (u(t)=y(t)=0, t<0), it is important to note that y(t) cannot be considered as a state for the time fractional model and that all the past of y(t) is required to compute the model evolution.

    Figure 5.  Block diagram Eq (4).

    To better illustrate such a concept, a simple time fractional model is used: a fractional integrator supposed at rest at t=0. The corresponding block diagram is shown in Figure 6.

    Figure 6.  Block diagram of an order ν fractional integrator.

    For an integer integrator, ν=1, relation (4) is really a state space description. At t1>0, state y(t) can be computed if the input between 0 and t1 is known:

    y(t1)=t10x_(τ)dτ=y1=cst. (24)

    Values of y(t) at later times than t1 are given by:

    y(t)=t0x_(τ)dτ=t10x_(τ)dτy1=cst+tt1x_(τ)dτ,t>t1. (25)

    Thus, y(t) can be computed if x_(t) is known within t1 and t. Integrator output at time t1 thus summarizes the whole model past. y(t) is really the state of the dynamic model, in agreement with the definition given in [26].

    Let us apply the same reasoning to the fractional integrator case of order ν. From the definition of fractional integration, value of y(t) at t1>0 can be computed if the input between t=0 and t1 is known:

    y(t1)=1Γ(ν)t10(t1τ)ν1x_(τ)dτ=y1=cst. (26)

    Variable y(t), t>t1, is thus given by:

    y(t1)=1Γ(v)t0(tτ)v1x_(τ)dτ=1Γ(v)t10(tτ)v1x_(τ)dτα(t)y1+1Γ(v)tt1(tτ)v1x_(τ)dτ. (27)

    Two notable differences can be highlighted with respect to the integer case. First, term α(t) in equation (27) is not a constant but depends on the considered time t. Moreover, even if y1=y(t1) is known, it is not enough to compute α(t). Output y(t) of the fractional integrator is thus not a state. The same analysis can be held for the general case of a pseudo state description or a fractional differential equation.

    Beyond discussions on the concept of state, computation of α(t) in relation (27) whatever time t, requires to know y(t) t[0..t1], thus all the model past. This clearly shows that knowledge of y(t) at a unique point of the past is not enough.

    Fractional operators and the resulting time fractional models are known for their memory property. However, for the following two reasons, many studies proposed in the literature seem to ignore this property when the model initialization problem is considered:

    -they use the Caputo definition that involves only integer derivatives of the derivate function at the initial time,

    -they use other definitions but initialization is done by taking only an initial value for the initial time into consideration.

    This kind of initialization means that the operator or model memory exists everywhere on the time axis, except at the initial time. This is not consistent. Memory is an intrinsic property that exists all the time and that is proved in this paper with very simple examples. If from a mathematical point of view, most of the fractional derivative definitions encountered in the literature [17] are not problematic, this paper shows that the Caputo and Riemann-Liouville definitions are not able to ensure a proper initialization when used in a model definition. The paper also shows that this problem is not encountered with the Grünwald-Letnikov definition, provided that all the past of the model (from t) is taken into account. And this is precisely one of the drawbacks of time fractional models that induces a physical inconsistency and many analysis problems [17].

    What are the possible solutions? One solution can be to add an initialization function to the definition of the model. This is what was proposed by Lorenzo and Hartley [1,2]. Yet again however, it requires the knowledge of all the model past (from t). Another solution consists in introducing new kernels for the definition of fractional integration as in [19]. But the goal would not be to solve only a singularity problem as in [19], but to reach a finite memory length as was done for instance in [20]. Note that while it was claimed in [21] that this class of kernels was too restrictive, it is linked to the problem analyzed in this paper: the inability of the Caputo definition to take into account initial conditions properly if used to define a time fractional model [22]. The other solution is to introduce new solutions for fractional behavior modeling, without the drawbacks associated to time fractional models [18]:

    -distributed time delay models [23];

    -non-linear models [24];

    -partial differential equation with spatially varying coefficients [25].

    All the conclusions presented in this paper can they be extended to models involving space fractional derivatives as in [29,30,31]? As shown in [28], whatever the variable on which the derivative relates, a fractional model remains a doubly infinite dimensional model and as such requires an infinite amount of information for its initialization. The question remains open, however, the authors will seek to answer it in their future work.



    [1] S. Bird, Ewan Klein, Edward Loper, Natural language processing with Python: analyzing text with the natural language toolkit, 1 Ed., Sebastopol: O'Reilly Media, Inc., 2009.
    [2] Y. Wu, M. Schuster, Z. Chen, Q. V. Le, M. Norouzi, W. Macherey, et al., Google's neural machine translation system: bridging the gap between human and machine translation, arXiv: 1609.08144. http://dx.doi.org/10.48550/arXiv.1609.08144
    [3] B. Liu, Sentiment analysis and opinion mining, Cham: Springer, 2012. http://dx.doi.org/10.1007/978-3-031-02145-9
    [4] P. Rajpurkar, J. Zhang, K. Lopyrev, P. Liang, SQuAD: 100,000+ questions for machine comprehension of text, Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, 2016, 2383–2392. http://dx.doi.org/10.18653/v1/D16-1264 doi: 10.18653/v1/D16-1264
    [5] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, et al., Attention is all you need, Proceedings of the 31st International Conference on Neural Information Processing Systems, 2017, 6000–6010.
    [6] J. Devlin, M. Chang, K. Lee, K. Toutanova, BERT: pre-training of deep bidirectional transformers for language understanding, Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 2019, 4171–4186. http://dx.doi.org/10.18653/v1/N19-1423 doi: 10.18653/v1/N19-1423
    [7] Z. Yang, Z. Dai, Y. Yang, J. G. Carbonell, R. Salakhutdinov, Q. V. Le, XLNet: generalized autoregressive pretraining for language understanding, Proceedings of the 33rd International Conference on Neural Information Processing Systems, 2019, 5753–5763.
    [8] S. Ruder, A. Søgaard, I. Vulić, Unsupervised cross-lingual representation learning, Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Tutorial Abstracts, 2019, 31–38. http://dx.doi.org/10.18653/v1/P19-4007 doi: 10.18653/v1/P19-4007
    [9] C. Wang, M. Li, A. J. Smola, Language models with transformers, arXiv: 1904.09408. http://dx.doi.org/10.48550/arXiv.1904.09408
    [10] G. Lample, A. Conneau, Cross-lingual language model pretraining, Proceedings of the 33rd International Conference on Neural Information Processing Systems, 2019, 7059–7069.
    [11] A. Conneau, K. Khandelwal, N. Goyal, V. Chaudhary, G. Wenzek, F. Guzmán, et al., Unsupervised cross-lingual representation learning at scale, Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, 8440–8451. http://dx.doi.org/10.18653/v1/2020.acl-main.747 doi: 10.18653/v1/2020.acl-main.747
    [12] W. Vries, A. Cranenburgh, A. Bisazza. T. Caselli, G. Noord, M. Nissim, BERTje: a dutch BERT model, arXiv: 1912.09582. http://dx.doi.org/10.48550/arXiv.1912.09582
    [13] A. Virtanen, J. Kanerva, R. Ilo, J. Luoma. J. Luotolahti, T. Salakoski, et al., Multilingual is not enough: BERT for Finnish, arXiv: 1912.07076. http://dx.doi.org/10.48550/arXiv.1912.07076
    [14] K. K, Z. Wang, S. Mayhew, D. Roth, Cross-lingual ability of multilingual BERT: an empirical study, arXiv: 1912.07840. http://dx.doi.org/10.48550/arXiv.1912.07840
    [15] M. Artetxe, S. Ruder, D. Yogatama, On the cross-lingual transferability of monolingual representations, Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, 4623–4637. http://dx.doi.org/10.18653/v1/2020.acl-main.421 doi: 10.18653/v1/2020.acl-main.421
    [16] Y. Tedla, K. Yamamoto, Morphological segmentation with LSTM neural networks for Tigrinya, IJNLC, 7 (2018), 29–44. http://dx.doi.org/10.5121/ijnlc.2018.7203 doi: 10.5121/ijnlc.2018.7203
    [17] R. Hetzron, The Semitic languages, New York: Routledge, 1997.
    [18] O. Osman, Y. Mikami, Stemming Tigrinya words for information retrieval, Proceedings of COLING 2012: Demonstration Papers, 2012, 345–352.
    [19] M. Tadesse, Trilingual sentiment analysis on social media, Master Thesis, Univeristy of Addis Ababa, 2018.
    [20] Y. K. Tedla, K. Yamamoto, A. Marasinghe, Tigrinya part-of-speech tagging with morphological patterns and the new Nagaoka Tigrinya corpus, International Journal of Computer Applications, 146 (2016), 33–41. http://dx.doi.org/10.5120/IJCA2016910943 doi: 10.5120/IJCA2016910943
    [21] A. Sahle, Sewasiw Tigrinya B'sefihu/a comprehensive Tigrinya grammar, Lawrenceville: Red Sea Press, Inc., 1998.
    [22] T. Kudo, J. Richardson, SentencePiece: a simple and language independent subword tokenizer and detokenizer for neural text processing, Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, 2018, 66–71. http://dx.doi.org/10.18653/v1/D18-2012 doi: 10.18653/v1/D18-2012
    [23] Z. Chi, L. Dong, F. Wei, X. Mao, H. Huang, Can monolingual pretrained models help cross-lingual classification? Proceedings of the 1st Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, 2020, 12–17.
    [24] T. Mikolov, K. Chen, G. Corrado, J. Dean, Efficient estimation of word representations in vector space, arXiv: 1301.3781. http://dx.doi.org/10.48550/arXiv.1301.3781
    [25] P. Prettenhofer, B. Stein, Cross-language text classification using structural correspondence learning, Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, 2010, 1118–1127.
    [26] A. Sugiyama, N. Yoshinaga, Data augmentation using back-translation for context-aware neural machine translation, Proceedings of the Fourth Workshop on Discourse in Machine Translation, 2019, 35–44. http://dx.doi.org/10.18653/v1/D19-6504 doi: 10.18653/v1/D19-6504
    [27] J. Wei, K. Zou, EDA: easy data augmentation techniques for boosting performance on text classification tasks, Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 2019, 6382–6388. http://dx.doi.org/10.18653/v1/D19-1670 doi: 10.18653/v1/D19-1670
  • This article has been cited by:

    1. Vincent Tartaglione, Christophe Farges, Jocelyn Sabatier, Fractional Behaviours Modelling with Volterra Equations: Application to a Lithium-Ion Cell and Comparison with a Fractional Model, 2022, 6, 2504-3110, 137, 10.3390/fractalfract6030137
    2. Juan A. López-Villanueva, Pablo Rodríguez-Iturriaga, Luis Parrilla, Salvador Rodríguez-Bolívar, A compact model of the ZARC for circuit simulators in the frequency and time domains, 2022, 153, 14348411, 154293, 10.1016/j.aeue.2022.154293
    3. Nezha Maamri, Jean-Claude Trigeassou, A Plea for the Integration of Fractional Differential Systems: The Initial Value Problem, 2022, 6, 2504-3110, 550, 10.3390/fractalfract6100550
    4. Jocelyn Sabatier, 2022, Chapter 1, 978-3-031-04382-6, 1, 10.1007/978-3-031-04383-3_1
    5. Zdeněk Biolek, Dalibor Biolek, Viera Biolková, Zdeněk Kolka, Extended Higher-Order Elements with Frequency-Doubled Parameters: The Hysteresis Loops Are Always of Type II, 2023, 23, 1424-8220, 7179, 10.3390/s23167179
    6. Abdelkader Moumen, Abdelaziz Mennouni, Mohamed Bouye, Contributions to the Numerical Solutions of a Caputo Fractional Differential and Integro-Differential System, 2024, 8, 2504-3110, 201, 10.3390/fractalfract8040201
    7. Reny George, Fahad Al-shammari, Mehran Ghaderi, Shahram Rezapour, On the boundedness of the solution set for the ψ-Caputo fractional pantograph equation with a measure of non-compactness via simulation analysis, 2023, 8, 2473-6988, 20125, 10.3934/math.20231025
    8. Shabnam Tashakori, Andres San-Millan, Vahid Vaziri, Sumeet S. Aphale, Fast Parameter Identification of the Fractional-Order Creep Model, 2024, 13, 2076-0825, 534, 10.3390/act13120534
    9. Hacen Serrai, Brahim Tellab, A generalized contraction mapping applied for existence results for ordinary hybrid version of a generalized Sturm–Liouville–Langevin equations under Ψ-Hilfer fractional-order derivative, 2025, 74, 0009-725X, 10.1007/s12215-024-01175-4
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(686) PDF downloads(29) Cited by(0)

Figures and Tables

Figures(4)  /  Tables(4)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog