Loading [MathJax]/jax/element/mml/optable/SupplementalArrowsB.js
Research article

Challenges of the digital age for privacy and personal data protection

  • Digital age can be described as a collection of different technological solutions as virtual environments, digital services, intelligent applications, machine learning, knowledge-based systems, etc., determining the specific characteristics of contemporary world globalization, e-communications, information sharing, virtualization, etc. However, there is an opportunity the technologies of the digital age to violate some basic principles of the information security and privacy by unregulated access to information and personal data, stored in different nodes of the global network. The goal of the article is to determine some special features of information and personal data protection and to summarise the main challenges of the digital age for the user's security and privacy. A brief presentation of the fundamental legislation in the fields of privacy and personal data protection is made in the introduction, followed by a review of related work on the topic. Components of information security for counteracting threats and attacks and basic principles in the organization of personal data protection are discussed. A summary of the basic challenges of the digital age is made by systematizing the negatives for user's privacy of the contemporary technologies as social computing, cloud services, Internet of Things, Big Data and Big Data Analytics and separate requirements to secure privacy of the participants based on General Data Protection Regulation principles are formulated.

    Citation: Radi P. Romansky, Irina S. Noninska. Challenges of the digital age for privacy and personal data protection[J]. Mathematical Biosciences and Engineering, 2020, 17(5): 5288-5303. doi: 10.3934/mbe.2020286

    Related Papers:

    [1] Minlong Lin, Ke Tang . Selective further learning of hybrid ensemble for class imbalanced increment learning. Big Data and Information Analytics, 2017, 2(1): 1-21. doi: 10.3934/bdia.2017005
    [2] Subrata Dasgupta . Disentangling data, information and knowledge. Big Data and Information Analytics, 2016, 1(4): 377-390. doi: 10.3934/bdia.2016016
    [3] Qinglei Zhang, Wenying Feng . Detecting Coalition Attacks in Online Advertising: A hybrid data mining approach. Big Data and Information Analytics, 2016, 1(2): 227-245. doi: 10.3934/bdia.2016006
    [4] Tieliang Gong, Qian Zhao, Deyu Meng, Zongben Xu . Why Curriculum Learning & Self-paced Learning Work in Big/Noisy Data: A Theoretical Perspective. Big Data and Information Analytics, 2016, 1(1): 111-127. doi: 10.3934/bdia.2016.1.111
    [5] Xin Yun, Myung Hwan Chun . The impact of personalized recommendation on purchase intention under the background of big data. Big Data and Information Analytics, 2024, 8(0): 80-108. doi: 10.3934/bdia.2024005
    [6] Pankaj Sharma, David Baglee, Jaime Campos, Erkki Jantunen . Big data collection and analysis for manufacturing organisations. Big Data and Information Analytics, 2017, 2(2): 127-139. doi: 10.3934/bdia.2017002
    [7] Zhen Mei . Manifold Data Mining Helps Businesses Grow More Effectively. Big Data and Information Analytics, 2016, 1(2): 275-276. doi: 10.3934/bdia.2016009
    [8] Ricky Fok, Agnieszka Lasek, Jiye Li, Aijun An . Modeling daily guest count prediction. Big Data and Information Analytics, 2016, 1(4): 299-308. doi: 10.3934/bdia.2016012
    [9] M Supriya, AJ Deepa . Machine learning approach on healthcare big data: a review. Big Data and Information Analytics, 2020, 5(1): 58-75. doi: 10.3934/bdia.2020005
    [10] Sunmoo Yoon, Maria Patrao, Debbie Schauer, Jose Gutierrez . Prediction Models for Burden of Caregivers Applying Data Mining Techniques. Big Data and Information Analytics, 2017, 2(3): 209-217. doi: 10.3934/bdia.2017014
  • Digital age can be described as a collection of different technological solutions as virtual environments, digital services, intelligent applications, machine learning, knowledge-based systems, etc., determining the specific characteristics of contemporary world globalization, e-communications, information sharing, virtualization, etc. However, there is an opportunity the technologies of the digital age to violate some basic principles of the information security and privacy by unregulated access to information and personal data, stored in different nodes of the global network. The goal of the article is to determine some special features of information and personal data protection and to summarise the main challenges of the digital age for the user's security and privacy. A brief presentation of the fundamental legislation in the fields of privacy and personal data protection is made in the introduction, followed by a review of related work on the topic. Components of information security for counteracting threats and attacks and basic principles in the organization of personal data protection are discussed. A summary of the basic challenges of the digital age is made by systematizing the negatives for user's privacy of the contemporary technologies as social computing, cloud services, Internet of Things, Big Data and Big Data Analytics and separate requirements to secure privacy of the participants based on General Data Protection Regulation principles are formulated.


    For a continuous risk outcome 0<y<1, a model with a random effect has potentially a wide application in portfolio risk management, especially, for stress testing [1,2,7,16,19], capital allocation, conditional expected shortfall estimation [3,11,17].

    Given fixed effects x=(x1,x2,,xk), two widely used regression models to estimate the expected value E(y|x) are: the fraction response model [10] and Beta regression model [4,6,8]. There are cases, however, where tail behaviours or severity levels of the risk outcome are relevant. In those cases, a regression model may no longer fit in for the requirements. In addition, a fraction response model of the form E(y|x)=Φ(a0+a1x1++akxk) may not be adequate when data exhibits significant heteroscedasticity, where Φ is a map from R1 to the open interval (0,1).

    In this paper, we assume that the risk outcome y is driven by a model:

    y=Φ(a0+a1x1++akxk+bs), (1.1)

    where s is a random continuous variable following a known distribution, independent of fixed effects (x1,x2,,xk). Parameters a0,a1,,ak are constant, while parameter b can be chosen to be dependent on (x1,x2,,xk) when required, for example, for addressing data heteroscedasticity.

    Given random effect model (1.1), the expected value E(y|x) can be deduced accordingly. It is given by the integral ΩΦ(a0+a1x1++akxk+bs)f(s)ds over the domain Ω of s, where f is the probability density of s. Given the routine QUAD implemented in SAS and Python, this integral can be evaluated as quickly as other function calls. Relative error tolerance for QUAD is 1.49e-8 in Python and is 1e-7 in SAS. But one can rescale the default tolerance to a desired level when necessary. This leads to an alternative regression tool to the fraction response model and Beta regression model.

    We introduce a family of interval distributions based on variable transformations. Probability densities for these distributions are provided (Proposition 2.1). Parameters of model (1.1) can then be estimated by maximum likelihood approaches assuming an interval distribution. In some cases, these parameters get an analytical solution without the needs for a model fitting (Proposition 4.1). We call a model with a random effect, where parameters are estimated by maximum likelihood assuming an interval distribution, an interval distribution model.

    In its simplest form, the interval distribution model y=Φ(a+bs), where a and b, are constant, can be used to model the loss rate as a random distribution for a homogeneous portfolio. Let yα and sα denote the α -quantiles for y and s at level α, 0<α<1. Then yα=Φ(a+bsα). The conditional expected shortfall for loss rate y, at level α, can then be estimated as the integral 11α[sα,+)Φ(a+bs)f(s)ds, where f is the density of s. Meanwhile, a stress testing loss estimate, derived from a model on a specific scenario, can be compared in loss rate to severity yα(=Φ(a+bsα)), to position its level of severity. A loss estimate may not have reached the desired, for example, 99% level yet, if it is far below y0.99, and far below the maximum historical loss rate. In which case, further recalibrations for the model may be required.

    The paper is organized as follows: in section 2, we introduce a family of interval distributions. A measure for tail fatness is defined. In section 3, we show examples of interval distributions and investigate their tail behaviours. We propose in section 4 an algorithm for estimating the parameters in model (1.1).

    Interval distributions introduced in this section are defined for a risk outcome over a finite open interval (c0,c1), where c0< c1 are finite numbers. These interval distributions can potentially be used for modeling a risk outcome over an arbitrary finite interval, including interval (0, 1), by maximum likelihood approaches.

    Let D=(d0,d1), d0<d1, be an open interval, where d0 can be finite or and d1 can be finite or +.

    Let

    Φ:D(c0,c1) (2.1)

    be a transformation with continuous and positive derivatives Φ(x)=ϕ(x). A special example is (c0,c1)=(0,1), and Φ:D(0,1) is the cumulative distribution function (CDF) of a random variable with a continuous and positive density.

    Given a continuous random variable s, let f and F be respectively its density and CDF. For constants a and b>0, let

    y=Φ(a+bs), (2.2)

    where we assume that the range of variable (a+bs) is in the domain D of Φ. Let g(y,a,b) and G(y,a,b) denote respectively the density and CDF of y in (2.2).

    Proposition 2.1. Given Φ1(y), functions g(y,a,b) and G(y,a,b) are given as:

    g(y,a,b)=U1/(bU2) (2.3)
    G(y,a,b)=F[Φ1(y)ab]. (2.4)

    where

    U1=f{[Φ1(y)a]/b},U2=ϕ[Φ1(y)] (2.5)

    Proof. A proof for the case when (c0,c1)=(0,1) can be found in [18]. The proof here is similar. Since G(y,a,b) is the CDF of y, it follows:

    G(y,a,b)=P[Φ(a+bs)y]
    =P{s[Φ1(y)a]/b}
    =F{[Φ1(y)a]/b}.

    By chain rule and the relationship Φ[Φ1(y)]=y, the derivative of Φ1(y) with respect to y is

    Φ1(y)y=1ϕ[Φ1(y)]. (2.6)

    Taking the derivative of G(y,a,b) with respect to y, we have

    G(y,a,b)y=f{[Φ1(y)a]/b}bϕ[Φ1(y)]=U1bU2.

    One can explore into these interval distributions for their shapes, including skewness and modality. For stress testing purposes, we are more interested in tail risk behaviours for these distributions.

    Recall that, for a variable X over (− ,+), we say that the distribution of X has a fat right tail if there is a positive exponent α>0, called tailed index, such that P(X>x)xα. The relation refers to the asymptotic equivalence of functions, meaning that their ratio tends to a positive constant. Note that, when the density is a continuous function, it tends to 0 when x+. Hence, by L’Hospital’s rule, the existence of tailed index is equivalent to saying that the density decays like a power law, whenever the density is a continuous function.

    For a risk outcome over a finite interval (c0,c1), c0,<c1, however, its density can be + when approaching boundaries c0 and c1. Let y0 be the largest lower bound for all values of y under (2.2), and y1 the smallest upper bound. We assume y0=c0 and y1=c1.

    We say that an interval distribution has a fat right tail if the limit \mathit{li}{m}_{y⤍{y}_{1}^{-}} \ \ g \left(y, a, b\right) = +\infty , and a fat left tail if \mathit{li}{m}_{y⤍{y}_{0}^{+}} \ \ g \left(y, a, b\right) = +\infty , where y⤍{y}_{0}^{+} and y⤍{y}_{1}^{-} denote respectively y approaching {y}_{0} from the right-hand-side, and {y}_{1} from the left-hand-side. For simplicity, we write y⤍{y}_{0} for y⤍{y}_{0}^{+}, and y⤍{y}_{1} for y⤍{y}_{1}^{-}.

    Given \alpha >0, we say that an interval distribution has a fat right tail with tailed index \alpha if \mathit{li}{m}_{y⤍{y}_{1}} \ \ g \left(y, a, b\right){\left({y}_{1}-y\right)}^{\beta } = +\infty whenever 0<\beta <\alpha , and li{m}_{y⤍{y}_{1}} \ \ g \left(y, a, b\right){\left({y}_{1}-y\right)}^{\beta } = 0 for \beta >\alpha . Similarly, an interval distribution has a fat left tail with tailed index \alpha if li{m}_{y⤍{y}_{0}} \ \ g \left(y, a, b\right){\left({y-y}_{0}\right)}^{\beta } = +\infty whenever 0<\beta <\alpha , and li{m}_{y⤍{y}_{0}} \ \ g \left(y, a, b\right)({y-{y}_{0})}^{\beta } = 0 for \beta >\alpha . Here the status at \beta = \alpha is left open. There are examples (Remark 3.4), where an interval distribution has a fat right tail with tailed index \alpha , but the limit li{m}_{y⤍{y}_{1}} \ \ g \left(y, a, b\right){\left({y}_{1}-y\right)}^{\alpha } can either be +\infty or 0. Under this definition, a tailed index of an interval distribution with a continuous density is always larger than 0 and less or equal to 1, if it exists.

    Recall that, for a Beta distribution with parameters \alpha >0 and \beta >0, its density is given by f\left(x\right) = \frac{{x}^{\alpha -1}{\left(1-x\right)}^{\beta -1}}{B(\alpha , \beta )}, where B(\alpha , \beta ) is the Beta function . Under the above definition, Beta distribution has a fat right tail with tailed index (1-\beta ) when 0<\beta <1 , and a fat left tail with tailed index (1-\alpha ) when 0<\alpha <1.

    Next, because the derivative of \mathrm{\Phi } is assumed to be continuous and positive, it is strictly monotonic. Hence {\mathrm{\Phi }}^{-1}\left(y\right) is defined. Let

    {z = \mathrm{\Phi }}^{-1}\left(y\right) (2.7)

    Then li{m}_{y⤍{y}_{0}}z exists (can be -\infty ) , and the same for li{m}_{y⤍{y}_{1}}z (can be +\infty ) . Let li{m}_{y⤍{y}_{0}} \ \ z = {z}_{0} , and li{m}_{y⤍{y}_{1}} \ \ z = {z}_{1} . Rewrite g\left(y, a, b\right) as g\left(\mathrm{\Phi }\left(z\right), a, b\right) by (2.7). Let {\partial \left[g\left(\mathrm{\Phi }\left(\mathrm{z}\right), a, b\right)\right]}^{-\frac{1}{\beta }}/\partial z denote the derivative of {\left[g\left(\mathrm{\Phi }\left(\mathrm{z}\right), a, b\right)\right]}^{-1/\beta } with respect to z .

    Lemma 2.2. Given \beta >0 , the following statements hold:

    (ⅰ) \mathit{li}{m}_{y⤍{y}_{0}} \ \ g \left(y, a, b\right){\left({y-y}_{0}\right)}^{\beta } = \mathit{li}{m}_{z⤍{z}_{0}} \ \ g \left(\mathrm{\Phi }\left(\mathrm{z}\right), a, b\right){\left(\mathrm{\Phi }\left(\mathrm{z}\right)-{\mathrm{y}}_{0}\right)}^{\beta } and li{m}_{y⤍{y}_{1}} \ \ g \left(y, a, b\right){\left({y}_{1}-y\right)}^{\beta } = li{m}_{z⤍{z}_{1}} \ \ g \left(\mathrm{\Phi }\left(z\right), a, b\right){\left({y}_{1}-\mathrm{\Phi }\left(z\right)\right)}^{\beta } .

    (ⅱ) If li{m}_{y⤍{y}_{0}} \ \ g \left(y, a, b\right) = +\infty and li{m}_{z⤍{z}_{0}}{\{\partial \left[g\left(\mathrm{\Phi }\left(\mathrm{z}\right), a, b\right)\right]}^{-\frac{1}{\beta }}/\partial z\}/{\rm{ \mathsf{ ϕ} }}(\mathrm{z}) is 0 (resp. +\infty ) , then li{m}_{y⤍{y}_{0}} \ \ g \left(y, a, b\right){(y-{y}_{0})}^{\beta } = +\infty (resp. 0).

    (ⅲ) If li{m}_{y⤍{y}_{1}} \ \ g \left(y, a, b\right) = +\infty and li{m}_{z⤍{z}_{1}}-{\{\partial \left[g\left(\mathrm{\Phi }\left(\mathrm{z}\right), a, b\right)\right]}^{-\frac{1}{\beta }}/\partial z\}/{\rm{ \mathsf{ ϕ} }}(\mathrm{z}\left)\right) is 0 (resp. +\infty ) , then li{m}_{y⤍{y}_{1}} \ \ g \left(y, a, b\right){\left({y}_{1}-y\right)}^{\beta } = +\infty (resp. 0).

    Proof. The first statement follows from the relationship y = \mathrm{\Phi }(\mathrm{z} ). For statements (ⅱ) and (ⅲ), we show only (ⅲ). The proof for (ⅱ) is similar. Notice that

    {\left[g\left(y, a, b\right){\left({y}_{1}-y\right)}^{\beta }\right]}^{-1/\beta } = \frac{{\left[g\left(y, a, b\right)\right]}^{-1/\beta }}{{y}_{1}-y} = \frac{{\left[g\left(\mathrm{\Phi }\left(\mathrm{z}\right), a, b\right)\right]}^{-1/\beta }}{{y}_{1}-\mathrm{\Phi }\left(\mathrm{z}\right)}. (2.8)

    By L’Hospital’s rule and taking the derivatives of the numerator and the denominator of (2.8) with respect to z , we have {li{m}_{y⤍{y}_{1}}\left[g\left(y, a, b\right){\left({y}_{1}-y\right)}^{\beta }\right]}^{-1/\beta } = 0 (resp. +\infty ) if li{m}_{z⤍{z}_{0}}{-\{\partial \left[g\left(\mathrm{\Phi }\left(z\right), a, b\right)\right]}^{-1/\beta }/\partial z\}/{\rm{ \mathsf{ ϕ} }}(z) is 0 (resp. +\infty ) . Hence li{m}_{y⤍{y}_{1}} \ \ g \left(y, a, b\right){\left({y}_{1}-y\right)}^{\beta } = +\infty (resp. 0).

    For tail convexity, we say that the right tail of an interval distribution is convex if g\left(y, a, b\right) is convex for {y}_{1}-є<y< {y}_{1} for sufficiently small є>0. Similarly, the left tail is convex if g\left(y, a, b\right) is convex for {y}_{0}<y< {y}_{0}+є for sufficiently small є>0. One sufficient condition for convexity for the right (resp. left) tail is {g}_{yy}^{''}(y, a, b)\ge 0 when y is sufficiently close to {y}_{1} (resp. {y}_{0}) .

    Again, write g\left(y, a, b\right) = g\left(\mathrm{\Phi }\left(z\right), a, b\right) . Let

    h\left(z, a, b\right) = \mathrm{log}\left[g\left(\mathrm{\Phi }\left(z\right), a, b\right)\right], (2.9)

    where \mathrm{l}\mathrm{o}\mathrm{g}\left(x\right) denotes the natural logarithmic function. Then

    g\left(y, a, b\right) = \mathrm{exp}\left[h\left(z, a, b\right)\right]. (2.10)

    By (2.9), (2.10), using (2.6) and the relationship {z = \mathrm{\Phi }}^{-1}\left(y\right) , we have

    {g}_{y}^{'} = {[h}_{z}^{'}\left(z\right)/{\rm{ \mathsf{ ϕ} }}\left(\mathrm{z}\right)]\mathrm{e}\mathrm{x}\mathrm{p}[h({\mathrm{\Phi }}^{-1}\left(y\right), a, b)], \\ {g}_{yy}^{''} = \left[\frac{{h}_{zz}^{''}\left(z\right)}{{{\rm{ \mathsf{ ϕ} }}}^{2}\left(\mathrm{z}\right)}-\frac{{h}_{z}^{'}\left(z\right){{\rm{ \mathsf{ ϕ} }}}_{\mathrm{z}}^{'}\left(z\right)}{{{\rm{ \mathsf{ ϕ} }}}^{3}\left(\mathrm{z}\right)}+\frac{{h}_{\mathrm{z}}^{\mathrm{'}}\left(\mathrm{z}\right){h}_{\mathrm{z}}^{\mathrm{'}}\left(\mathrm{z}\right)}{{{\rm{ \mathsf{ ϕ} }}}^{2}\left(\mathrm{z}\right)}\right]\mathrm{e}\mathrm{x}\mathrm{p}\left[h\right({\mathrm{\Phi }}^{-1}\left(y\right), a, b) ]. (2.11)

    The following lemma is useful for checking tail convexity, it follows from (2.11).

    Lemma 2.3. Suppose {\rm{ \mathsf{ ϕ} }}\left(\mathrm{z}\right)>0 , and derivatives {h}_{\mathrm{z}}^{\mathrm{'}}\left(\mathrm{z}\right), {h}_{\mathrm{z}}^{\mathrm{'}\mathrm{'}}\left(\mathrm{z}\right), and {{\rm{ \mathsf{ ϕ} }}}_{\mathrm{z}}^{\mathrm{'}}\left(\mathrm{z}\right) , with respect to z, all exist. If {h}_{zz}^{''}\left(z\right)\ge 0 and {h}_{z}^{'}\left(z\right){{\rm{ \mathsf{ ϕ} }}}_{\mathrm{z}}^{'}\left(z\right)\le 0, then {g}_{yy}^{''}(y, a, b)\ge 0.

    In this section, we focus on the case where \left({c}_{0}, {c}_{1}\right) = \left(0, 1\right), and \mathrm{\Phi }: D\to (0, 1) in (2.2) is the CDF of a continuous distribution . This includes, for example, the CDFs for standard normal and standard logistic distributions.

    One can explore into a wide list of densities with different choices for \mathrm{\Phi } and s under (2.2). We consider here only the following four interval distributions:

    A. s \sim N\left(\mathrm{0, 1}\right) and \mathrm{\Phi } is the CDF for the standard normal distribution.

    B. s follows the standard logistic distribution and \mathrm{\Phi } is the CDF for the standard normal distribution.

    C. s follows the standard logistic distribution and \mathrm{\Phi } is its CDF.

    D.D. s \sim N\left(\mathrm{0, 1}\right) and \mathrm{\Phi } is the CDF for standard logistic distribution.

    Densities for cases A, B, C, and D are given respectively in (3.3) (section 3.1), (A.1), (A.3), and (A5) (Appendix A). Tail behaviour study is summarized in Propositions 3.3, 3.5, and Remark 3.6. Sketches of density plots are provided in Appendix B for distributions A, B, and C.

    Using the notations of section 2, we have {\rm{ \mathsf{ ϕ} }} = f and \mathrm{\Phi } = F . We claim that y = \mathrm{\Phi }\left(a+bs\right) under (2.2) follows the Vasicek distribution [13,14].

    By (2.5), we have

    \mathrm{log}\left(\frac{{U}_{1}}{{U}_{2}}\right) = \frac{{-z}^{2}+2az-{a}^{2}+{b}^{2}{z}^{2}}{2{b}^{2}} (3.1)
    = \frac{{-\left(1-{b}^{2}\right)\left(z-\frac{a}{1-{b}^{2}}\right)}^{2}+\frac{{b}^{2}}{1-{b}^{2}}{a}^{2}}{2{b}^{2}}\text{.} (3.2)

    Therefore, we have

    g\left(\mathrm{y}, a, b\right) = \frac{1}{b}\mathrm{e}\mathrm{x}\mathrm{p}\left\{\frac{{-\left(1-{b}^{2}\right)\left(z-\frac{a}{1-{b}^{2}}\right)}^{2}+\frac{{b}^{2}}{1-{b}^{2}}{a}^{2}}{2{b}^{2}}\right\}\text{.} (3.3)

    Again, using the notations of section 2, we have {y}_{0} = 0 and {y}_{1} = 1 . With z = {\mathrm{\Phi }}^{-1}\left(y\right), we have li{m}_{y⤍0} \ \ z = -\infty and li{m}_{y⤍1} \ \ z = +\infty . Recall that a variable 0<y<1 follows a Vasicek distribution [13,14] if its density has the form:

    g\left(y, p, \rho \right) = \sqrt{\frac{1-\rho }{\rho }}\mathrm{e}\mathrm{x}\mathrm{p}\{-\frac{1}{2\rho }{\left[{\sqrt{1-\rho }{\mathrm{\Phi }}^{-1}\left(y\right)-\mathrm{\Phi }}^{-1}\left(p\right)\right]}^{2}+\frac{1}{2}{\left[{\mathrm{\Phi }}^{-1}\left(y\right)\right]}^{2}\}\text{, } (3.4)

    where p is the mean of y , and \rho is a parameter called asset correlation.

    Proposition 3.1. Density (3.3) is equivalent to (3.4) under the relationships:

    a = \frac{{\Phi }^{-1}\left(p\right)}{\sqrt{1-\rho }} \ \ \text{and}\ \ b = \sqrt{\frac{\rho }{1-\rho }}. (3.5)

    Proof. A similar proof can be found in [19]. By (3.4), we have

    g\left(y, p, \rho \right) = \sqrt{\frac{1-\rho }{\rho }}\mathrm{e}\mathrm{x}\mathrm{p}\{-\frac{1-\rho }{2\rho }{\left[{{\mathrm{\Phi }}^{-1}\left(y\right)-\mathrm{\Phi }}^{-1}\left(p\right)/\sqrt{1-\rho }\right]}^{2}+\frac{1}{2}{\left[{\mathrm{\Phi }}^{-1}\left(y\right)\right]}^{2}\}
    = \frac{1}{b}\mathrm{exp}\left\{-\frac{1}{2}{\left[\frac{{\Phi }^{-1}\left(y\right)-a}{b}\right]}^{2}\right\}\mathrm{e}\mathrm{x}\mathrm{p}\left\{\frac{1}{2}{\left[{\mathrm{\Phi }}^{-1}\left(y\right)\right]}^{2}\right\}
    = {U}_{1}/{(bU}_{2}) = g(y, a, b)\text{.}

    The following relationships are implied by (3.5):

    \rho = \frac{{b}^{2}}{1{+b}^{2}}, (3.6)
    a = {\Phi }^{-1}\left(p\right)\sqrt{1+{b}^{2}}\text{.} (3.7)

    Remark 3.2. The mode of g\left(y, p, \rho \right) in (3.4) is given in [14] as \mathrm{\Phi }\left(\frac{\sqrt{1-\rho }}{1-2\rho }{\mathrm{\Phi }}^{-1}\left(p\right)\right) . We claim this is the same as \mathrm{\Phi }\left(\frac{a}{1-{b}^{2}}\right) . By (3.6), 1-2\rho = \frac{1-{b}^{2}}{1+{b}^{2}} and \sqrt{1-\rho } = \frac{1}{\sqrt{1+{b}^{2}}}. Therefore, we have

    \frac{\sqrt{1-\rho }}{1-2\rho }{\mathrm{\Phi }}^{-1}\left(p\right) = \frac{\sqrt{1+{b}^{2}}}{1-{b}^{2}}{\mathrm{\Phi }}^{-1}\left(p\right) = \frac{a}{1-{b}^{2}}.

    This means \mathrm{\Phi }\left(\frac{\sqrt{1-\rho }}{1-2\rho }{\mathrm{\Phi }}^{-1}\left(p\right)\right) = \mathrm{\Phi }\left(\frac{a}{1-{b}^{2}}\right).

    Proposition 3.3. The following statements hold for g(y, a, b) given in (3.3):

    (ⅰ) g(y, a, b) is unimodal if 0<b<1 with mode given by \mathrm{\Phi }\left(\frac{a}{1-{b}^{2}}\right) , and is in U-shape if b>1 .

    (ⅱ) \mathrm{I}\mathrm{f}\ b>1, \mathrm{t}\mathrm{h}\mathrm{e}\mathrm{n} \ \ g(y, a, b) has a fat left tail and a fat right tail with tailed index (1-1/{b}^{2}) .

    (ⅲ) If b>1, both tails of g(y, a, b) are convex , and is globally convex if in addition a = 0.

    Proof. For statement (ⅰ), we have -\left(1-{b}^{2}\right)<0 when 0<b<1 . Therefore by (3.2) function \mathrm{log}\left(\frac{{U}_{1}}{{U}_{2}}\right) reaches its unique maximum at z = \frac{a}{1-{b}^{2}} , resulting in a value for the mode at \mathrm{\Phi }\left(\frac{a}{1-{b}^{2}}\right). If b>1 , then -\left(1-{b}^{2}\right)>0, thus by (3.2), g(y, a, b) is first decreasing and then increasing when y varying from 0 to 1. This means (y, a, b ) is in U-shape.

    Consider statement (ⅱ). First by (3.3), if b>1, then li{m}_{y⤍1}\ \ g\left(y, a, b\right) = +\infty and li{m}_{y⤍0} \ \ g \left(y, a, b\right) = +\infty . Thus g\left(\mathrm{y}, a, b\right) has a fat right and a fat left tail. Next for tailed index, we use Lemma 2.2 (ⅱ) and (ⅲ). By (3.1),

    {\left[g\left(\mathrm{\Phi }\left(\mathrm{z}\right), a, b\right)\right]}^{-1/\beta } = {b}^{1/\beta }\mathrm{e}\mathrm{x}\mathrm{p}(-\frac{{\left({b}^{2}-1\right)z}^{2}+2az-{a}^{2}}{2\beta {b}^{2}}) (3.8)

    By taking the derivative of (3.8) with respect to z and noting that {\rm{ \mathsf{ ϕ} }}\left(\mathrm{z}\right) = \frac{1}{\sqrt{2\pi }}\mathrm{exp}\left(-\frac{{z}^{2}}{2}\right), we have

    -\left\{\partial {\left[g\left(\mathrm{\Phi }\left(\mathrm{z}\right), a, b\right)\right]}^{-\frac{1}{\beta }}/\partial z\right\}/{\rm{ \mathsf{ ϕ} }}\left(\mathrm{z}\right) = \sqrt{2\pi }{b}^{\frac{1}{\beta }}\frac{\left({b}^{2}-1\right)z+a}{\beta {b}^{2}}\mathrm{e}\mathrm{x}\mathrm{p}(-\frac{{\left({b}^{2}-1\right)z}^{2}+2az-{a}^{2}}{2\beta {b}^{2}}+\frac{{z}^{2}}{2})\text{.} (3.9)

    Thus li{m}_{z⤍+\infty }-\left\{\partial {\left[g\left(\mathrm{\Phi }\left(\mathrm{z}\right), a, b\right)\right]}^{-\frac{1}{\beta }}/\partial z\right\}/{\rm{ \mathsf{ ϕ} }}\left(\mathrm{z}\right) is 0 if \frac{{b}^{2}-1}{\beta {b}^{2}}>1 , and is +\infty if \frac{{b}^{2}-1}{\beta {b}^{2}}<1. Hence by Lemma 2.2 (ⅲ), g\left(y, a, b\right) has a fat right tail with tailed index (1-1/{b}^{2}) . Similarly, for the left tail, we have by (3.9)

    \left\{\partial {\left[g\left(\mathrm{\Phi }\left(\mathrm{z}\right), a, b\right)\right]}^{-\frac{1}{\beta }}/\partial z\right\}/{\rm{ \mathsf{ ϕ} }}\left(\mathrm{z}\right) = -\sqrt{2\pi }{b}^{\frac{1}{\beta }}\frac{\left({b}^{2}-1\right)z+a}{\beta {b}^{2}}\mathrm{e}\mathrm{x}\mathrm{p}(-\frac{{\left({b}^{2}-1\right)z}^{2}+2az-{a}^{2}}{2\beta {b}^{2}}+\frac{{z}^{2}}{2})\text{.} (3.10)

    Thus li{m}_{z⤍-\infty }\left\{\partial {\left[g\left(\mathrm{\Phi }\left(\mathrm{z}\right), a, b\right)\right]}^{-\frac{1}{\beta }}/\partial z\right\}/{\rm{ \mathsf{ ϕ} }}\left(\mathrm{z}\right) is 0 if \frac{{b}^{2}-1}{\beta {b}^{2}}>1 , and is +\infty if \frac{{b}^{2}-1}{\beta {b}^{2}}<1. Hence g\left(y, a, b\right) has a fat left tail with tailed index (1-1/{b}^{2}) by Lemma 2.2 (ⅱ).

    For statement (ⅲ), we use Lemma 2.3. By (2.9) and using (3.2), we have

    h\left(z, a, b\right) = \mathrm{log}\left(\frac{{U}_{1}}{{bU}_{2}}\right) = \frac{{-\left(1-{b}^{2}\right)\left(z-\frac{a}{1-{b}^{2}}\right)}^{2}+\frac{{b}^{2}}{1-{b}^{2}}{a}^{2}}{2{b}^{2}}-\mathrm{l}\mathrm{o}\mathrm{g}\left(b\right)\text{.}

    When b>1, it is not difficult to check out that {h}_{zz}^{''}\left(z\right)\ge 0 and {h}_{z}^{'}\left(z\right){{\rm{ \mathsf{ ϕ} }}}_{\mathrm{z}}^{'}\left(z\right)\le 0 when z⤍\pm \infty or when a = 0 .

    Remark 3.4. Assume \beta = (1-1/{b}^{2}) and b>1. By (3.9), we see

    li{m}_{z⤍+\infty }-\left\{{\partial \left[g\left(\mathrm{\Phi }\left(\mathrm{z}\right), a, b\right)\right]}^{-\frac{1}{\beta }}/\partial z\right\}/{\rm{ \mathsf{ ϕ} }}\left(\mathrm{z}\right)

    is +\infty for a = 0 , and is 0 for a>0. Hence for this \beta , the limit li{m}_{y⤍1} \ \ g \left(y, a, b\right){\left(1-y\right)}^{\beta } can be either 0 or +\infty , depending on the value of a .

    For these distributions, we again focus on their tail behaviours. A proof for the next proposition can be found in Appendix A.

    Proposition 3.5. The following statements hold:

    (a) Density g\left(y, a, b\right) has a fat left tail and a fat right tail for case B for all b>0 , and for case C if b>1. For case D, it does not have a fat right tail nor a fat left tail for any b>0.

    (b) The tailed index of g\left(y, a, b\right) for both right and left tails is 1 for case B for all b>0 , and is (1-\frac{1}{b}) for case C for B for b>1 .

    Remark 3.6. Among distributions A, B, C, and Beta distribution, distribution B gets the highest tailed index of 1, independent of the choices of b>0 .

    In this section, we assume that \mathrm{\Phi } in (2.2) is a function from {R}^{1} to (0, 1) with positive continuous derivatives. We focus on parameter estimation algorithms for model (1.1).

    First, we consider a simple case, where risk outcome y is driven by a model:

    y = \mathrm{\Phi }\left(v+bs\right), (4.1)

    where b>0 is a constant, v = {a}_{0}+{a}_{1}{x}_{1}+\dots +{a}_{k}{x}_{k} , and s \sim N\left(0, 1\right), independent of fixed effects {x = (x}_{1}, {x}_{2}, \dots , {x}_{k}) . The function \mathrm{\Phi } does not have to be the standard normal CDF. But when \mathrm{\Phi } is the standard normal CDF, the expected value E\left(y\right|x) can be evaluated by the formula {E}_{S}\left[\mathrm{\Phi }\left(a+bs\right)\right] = \mathrm{\Phi }\left(\frac{a}{\sqrt{1+{b}^{2}}}\right) [12].

    Given a sample {\left\{({x}_{1i}, {x}_{2i}, \dots , {x}_{ki}, {y}_{i})\right\}}_{i = 1}^{n}, where ({x}_{1i}, {x}_{2i}, \dots , {x}_{ki}, {y}_{i}) denotes the {i}^{th} data point of the sample, let {z}_{i} = {\mathrm{\Phi }}^{-1}\left({y}_{i}\right). and {v}_{i} = {a}_{0}+{a}_{1}{x}_{1i}+\dots +{a}_{k}{x}_{ki}. By (2.3), the log-likelihood function for model (4.1) is:

    LL = \sum _{i = 1}^{n}\left\{\mathrm{log}f\left(\frac{{z}_{i}-{v}_{i}}{b}\right)-\mathrm{l}\mathrm{o}\mathrm{g}{\rm{ \mathsf{ ϕ} }}\left({z}_{i}\right)-logb\right\}\text{, } (4.2)

    where f is the density of s. The part of \sum _{i = 1}^{n}\mathrm{l}\mathrm{o}\mathrm{g}{\rm{ \mathsf{ ϕ} }}\left({z}_{i}\right) is constant, which can be dropped off from the maximization.

    Recall that the least squares estimators of {{a}_{0}, a}_{1}, \dots , {a}_{k} , as a row vector, that minimize the sum squares

    SS = \sum _{i = 1}^{n}{({z}_{i}-{v}_{i})}^{2} (4.3)

    has a closed form solution given by the transpose of {\left({\mathrm{X}}^{T}\mathrm{X}\right)}^{-1}{\mathrm{X}}^{\mathrm{T}}\mathrm{Z} [5,9] whenever the design matrix \mathrm{X} has a rank of k, where

    {\rm{X}} = \left\lceil {\begin{array}{*{20}{c}} {\begin{array}{*{20}{c}} {1\;\;{x_{11}} \ldots {x_{k1}}}\\ {1\;\;{x_{12}} \ldots {x_{k2}}} \end{array}}\\ \ldots \\ {1\;\;{x_{1n}} \ldots {x_{kn}}} \end{array}} \right\rceil , {\rm{Z}} = \left\lceil {\begin{array}{*{20}{c}} {\begin{array}{*{20}{c}} {{z_1}}\\ {{z_2}} \end{array}}\\ \ldots \\ {{z_n}} \end{array}} \right\rceil .

    The next proposition shows there exists an analytical solution for the parameters of model (4.1).

    Proposition 4.1. Given a sample {\left\{({x}_{1i}, {x}_{2i}, \dots , {x}_{ki}, {y}_{i})\right\}}_{i = 1}^{n} , assume that the design matrix has a rank of k. If s \sim N(0, 1), then the maximum likelihood estimates of parameters {(a}_{0}, {a}_{1}, \dots , a) , as a row vector, and parameter b are respectively given by the transpose of {\left({\mathrm{X}}^{T}\mathrm{X}\right)}^{-1}{\mathrm{X}}^{\mathrm{T}}\mathrm{Z}, and {b}^{2} = \frac{1}{n}\sum _{i = 1}^{n}{({z}_{i}-{v}_{i})}^{2}. In absence of fixed effects {\{x}_{1}, {x}_{2}, \dots , {x}_{k}\} , parameters {a}_{0} and {b}^{2} degenerate respectively to the sample mean and variance of {z}_{1} , {z}_{2}, \dots , {z}_{n}.

    Proof. Dropping off the constant term from (4.2) and noting f\left(z\right) = \frac{1}{\sqrt{2\pi }}\mathrm{exp}\left(-\frac{{z}^{2}}{2}\right) , we have

    LL = -\frac{1}{2{b}^{2}}\sum _{i = 1}^{n}{({z}_{i}-{v}_{i})}^{2}-nlogb, (4.4)

    Hence the maximum likelihood estimates ({a}_{0}, {a}_{1}, \dots , {a}_{k}) are the same as least squares estimators of (4.3), which are given by the transpose of {\left({\mathrm{X}}^{T}\mathrm{X}\right)}^{-1}{\mathrm{X}}^{\mathrm{T}}\mathrm{Z}. By taking the derivative of (4.4) with respect to b and setting it to zero, we have {b}^{2} = \frac{1}{n}\sum _{i = 1}^{n}{({z}_{i}-{v}_{i})}^{2}.

    Next, we consider the general case of model (1.1), where the risk outcome y is driven by a model:

    y = \mathrm{\Phi }[v+ws], (4.5)

    where parameter w is formulated as w = \mathrm{exp}\left(u\right), and u = {b}_{0}+{b}_{1}{x}_{1}+ \dots +{b}_{k}{x}_{k}. We focus on the following two cases:

    (a) s \sim N\left(0, 1\right),

    (b) s is standard logistic.

    Given a sample {\left\{({x}_{1i}, {x}_{2i}, \dots , {x}_{ki}, {y}_{i})\right\}}_{i = 1}^{n}, let {w}_{i} = \mathrm{e}\mathrm{x}\mathrm{p}({b}_{0}+{b}_{1}{x}_{1i}+ \dots +{b}_{k}{x}_{ki}) and {{u}_{i} = b}_{0}+{b}_{1}{x}_{1i}+ \dots +{b}_{k}{x}_{ki}. The log-likelihood functions for model (4.5), dropping off the constant part \mathrm{log}\left({U}_{2}\right), for cases (a) and (b) are given respectively by (4.6) and (4.7):

    LL = \sum _{i = 1}^{n}-{\frac{1}{2}[\left({z}_{i}-{v}_{i}\right)}^{2}/{w}_{i}^{2}-{u}_{i}], (4.6)
    LL = \sum _{i = 1}^{n}\{-\left({z}_{i}-{v}_{i}\right)/{w}_{\mathrm{i}}-2\mathrm{log}[1+\mathrm{e}\mathrm{x}\mathrm{p}[-({z}_{i}-{v}_{i})/{w}_{i}]-{u}_{i}\}, (4.7)

    Recall that a function is log-concave if its logarithm is concave. If a function is concave, a local maximum is a global maximum, and the function is unimodal. This property is useful for searching maximum likelihood estimates.

    Proposition 4.2. The functions (4.6) and (4.7) are concave as a function of {(a}_{0}, {a}_{1}, \dots , {a}_{k}) . As a function of {(b}_{0}, {b}_{1}, \dots , {b}_{k}), (4.6) is concave.

    Proof. It is well-known that, if f(x) is log-concave, then so is f(Az + b), where Az + b : {R^m} \to {R^1} is any affine transformation from the m-dimensional Euclidean space to the 1-dimensional Euclidean space. For (4.6), the function f\left(x\right) = -({z-v)}^{2}\mathrm{exp}(-2u) is concave as a function of v, thus function (4.6) is concave as a function of {(a}_{0}, {a}_{1}, \dots , {a}_{k}) . Similarly, this function f\left(x\right) is concave as a function of u, so (4.6) is concave as a function of {(b}_{0}, {b}_{1}, \dots , {b}_{k}).

    For (4.7), the linear part -\left({z}_{i}-{v}_{i}\right)\mathrm{e}\mathrm{x}\mathrm{p}(-{u}_{i}) , as a function of {(a}_{0}, {a}_{1}, \dots , {a}_{k}) , in (4.7) is ignored. For the second part in (4.7), we know -\mathrm{log}\{1+\mathrm{e}\mathrm{x}\mathrm{p}[-(z-v)/\mathrm{e}\mathrm{x}\mathrm{p}\left(u\right)\left]\right\} , as a function of v , is the logarithm of the CDF of a logistic distribution. It is well-known that the CDF for a logistic distribution is log-concave. Thus (4.7) is concave with respect to {(a}_{0}, {a}_{1}, \dots , {a}_{k}) .

    In general, parameters {(a}_{0}, {a}_{1}, \dots , {a}_{k}) and {(b}_{0}, {b}_{1}, \dots , {b}_{k}) in model (4.5) can be estimated by the algorithm below.

    Algorithm 4.3. Follow the steps below to estimate parameters of model (4.5):

    (a) Given {(b}_{0}, {b}_{1}, \dots , {b}_{k}) , estimate {(a}_{0}, {a}_{1}, \dots , {a}_{k}) by maximizing the log-likelihood function;

    (b) Given {(a}_{0}, {a}_{1}, \dots , {a}_{k}) , estimate {(b}_{0}, {b}_{1}, \dots , {b}_{k}) by maximizing the log-likelihood function;

    (c) Iterate (a) and (b) until a convergence is reached.

    With the interval distributions introduced in this paper, models with a random effect can be fitted for a continuous risk outcome by maximum likelihood approaches assuming an interval distribution. These models provide an alternative regression tool to the Beta regression model and fraction response model, and a tool for tail risk assessment as well.

    Authors are very grateful to the third reviewer for many constructive comments. The first author is grateful to Biao Wu for many valuable conversations. Thanks also go to Clovis Sukam for his critical reading for the manuscript.

    We would like to thank you for following the instructions above very closely in advance. It will definitely save us lot of time and expedite the process of your paper's publication.

    The views expressed in this article are not necessarily those of Royal Bank of Canada and Scotiabank or any of their affiliates. Please direct any comments to Bill Huajian Yang at h_y02@yahoo.ca.



    [1] E. J. Bloustein, N. J. Pallone, Individual and Group Privacy, Routledge, New York, 2017.
    [2] M. Oostveen, U. Irion, The golden age of personal data: How to regulate an enabling fundamental right?, in Personal Data in Competition, Consumer Protection and Intellectual Property Law (eds. M. Bakhoum, B. Conde Gallego, M. O. Mackenrodt, G. Surblytė-Namavičienė), Springer, (2018), 7-26. Available from: https://link.springer.com/chapter/10.1007/978-3-662-57646-5_2.
    [3] R. Romansky, A survey of digital world opportunities and challenges for user's privacy, Int. J. Inform. Technol. Secur., 9 (2017), 97-112.
    [4] Regulation (EU) 2916/679 of the European Parliament and the council of 27 April 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and repealing Directive 95/46/EC (General Data Protect Regulation), European Commission, 2016. Available from: https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=celex%3A32016R0679.
    [5] J. J. Hanus, H. G. Relyea, A policy assessment of the privacy act of 1974, Am. Univ. Law Rev., 25 (1976), 555.
    [6] M. Shabani, P. Borry, Rules for processing genetic data for research purposes in view of the new EU general data protection regulation, Eur. J. Human Genet., 26 (2018), 149-156.
    [7] A. V. Tsaregorodtsev, O. Ja. Kravets, O. N. Choporov, A. N. Zelenina, Information security risk estimation for cloud infrastructure, Int. J. Inform. Technol. Secur., 10 (2018), 67-76.
    [8] O. Yu. Zaslavskaya, l. A. Zaslavskiy, V. E. Bolnokin, O. Ja. Kravets, Features of ensuring information security when using cloud technologies in educational institutions, Int. J. Inform. Technol. Secur., 10 (2018), 93-102.
    [9] P. Wandra, H. Jie, DeepProfile: Finding fake profile in online social network using dynamic CNN, J. Inform. Secur. Appl., 52 (2020), article 102465. Available from: https://www.sciencedirect.com/science/article/abs/pii/S2214212619303801.
    [10] V. Kharchenko, Big Data and Internet of Things for safety critical applications: Challenges, methodology and industry cases, Int. J. Inform. Technol. Secur., 10 (2018), 3-16.
    [11] I. Alsmadi, R. Burdwell, A. Aleroud, A. Wahbeh, M. Al-Qudah, A. Al-Omari, Introduction to information security, in Practical Information Security (eds. I. Alsmadi, R. Burdwell, A. Aleroud, A. Wahbeh, M. Al-Qudah, A. Al-Omari), Springer, (2018), 1-16. Available from: https://www.springer.com/gp/book/9783319721187.
    [12] H. Paanen, M. Lapke, M. Siponen, State of the art in information security policy development. Comp. Secur., 88 (2020), article 101608. Available from: https://www.sciencedirect.com/science/article/pii/S0167404818313002.
    [13] M. A. Ferrag, H. Janicke, Deep learning for cyber security intrusion detection: Approaches, datasets, and comparative study, J. Inform. Secur. Appl., 50 (2020), article 102418. Available from: https://www.sciencedirect.com/science/article/pii/S2214212619305046.
    [14] A. R. Mahlous, SSR: A framework for a secure software reuse, Int. J. Inform. Technol. Secur., 10 (2018), 87-98.
    [15] Y. A. Ivanova, Assessment of the probability of cyberattacks on Transport Management Systems, Int. J. Inform. Technol. Secur., 10 (2018), 99-106.
    [16] M. A. P. Chamikara, P. Bertok, D. Liu, S. Camtepe, I. Khalil, An efficient and scalable privacy preserving algorithm for big data and data streams. Comp. & Security, Special issue "Security and Privacy in Smart Cyber-physical Systems" (2019), article 101570. Available from: https://www.sciencedirect.com/journal/computers-and-security/special-issue/109XHWZ5JSX.
    [17] Tz. Tzolov, Data model in the context of the general data protection regulation, Int. J. Inform. Technol. Secur., 9 (2017), 113-122.
    [18] R. Romansky, I. Noninska, Principles of secure access and privacy in combined e-learnng environment: Architecture, formalization and modelling, in Multidisciplinary Perspectives on Human Capital and Information Technology Professionals (eds. V. Ahuja, S. Rathore), IGI Global Publ., USA (2018), 152-178.
    [19] M. Aminzade, Confidentiality, integrity and availability—finding a balanced IT framework, Netw. Secur., 50 (2018), 9-11. Available from: https://www.sciencedirect.com/science/article/pii/S1353485818300436.
    [20] Thales, 2020 Data Threat Report - Global Edition. Survey and Analysis from IDC, 2020. Available from: https://cpl.thalesgroup.com/data-threat-report.
    [21] Guidelines on the Use of Cloud Computing Services by the European Institutions and Bodies, European Data Protection Supervisor, 2018. Available from: https://edps.europa.eu/data-protection/our-ork/publications/guidelines/guidelines-use-cloud-computing-services-european_en.
    [22] Maximizing the value of your data privacy investments - data privacy benchmark study, CISCO Cybersecurity Series, 2019. Available from: https://www.cisco.com/c/dam/en_us/about/doing_business/trust-center/docs/dpbs-2019.pdf.
    [23] Casey Crane, 20 surprising IoT statistics you don't already know, Security Boulevard, 5 Sep 2019. Available from: https://securityboulevard.com/2019/09/20-surprising-iot-statistics-you-dont-already-know/.
    [24] A. Azmoodeh, A. Dehghantanha. Big data and privacy: Challenges and opportunities, in Handbook of Big Data Privacy (ed. K-K. R. Choo, A. Dehghantanha), Springer-Cham, Switzerland, (2020), 1-6.
  • Reader Comments
  • © 2020 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(32202) PDF downloads(3268) Cited by(39)

Figures and Tables

Figures(2)  /  Tables(2)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog