Probability Distributions I

B.R. Martin , in Statistics for Physical Science, 2012

3.3.1 Joint Probability Distributions

The multivariate joint density function f ( x 1 , x 2 , , x n ) of the n continuous random variables x 1 , x 2 , , x n is a single-valued non-negative real number for all real values of x 1 , x 2 , , x n , normalized so that

(3.17) + + f ( x 1 , x 2 , , x n ) i = 1 n d x i = 1 ,

and the probability that x 1 falls between any two numbers a 1 and b 1 , x 2 falls between any two numbers a 2 and b 2 , , and x n falls between any two numbers a n and b n , simultaneously, is defined by

(3.18) P [ a 1 x 1 b 1 ; ; a n x n b n ] a n b n a 1 b 1 f ( x 1 , x 2 , , x n ) i = 1 n d x i .

Similarly, the multivariate joint distribution function F ( x 1 , x 2 , , x n ) of the n random variables x 1 , x 2 , , x m is

(3.19) F ( x 1 , x 2 , , x n ) x n x 1 f ( t 1 , t 2 , , t n ) i = 1 n d t i .

For simplicity, consider the case of just two random variables x and y. These could correspond to the energy and angle of emission of a particle emitted in a nuclear scattering reaction. If an event A corresponds to the variable x being observed in the range ( x , x + d x ) and the event B corresponds to the variable y being observed in the range ( y , y + d y ) , then

P [ A B ] = probability of x being in ( x , x + d x ) and y being in ( y , y + d y ) = f ( x , y ) d x d y .

As noted in Chapter 1, the joint density function corresponds to the density of points on a scatter plot of x and y in the limit of an infinite number of points. This is illustrated in Fig. 3.3, using the data shown on the scatter plot of Fig. 1.3(b).

FIGURE 3.3. A scatter plot of 1000 events that are functions of two random variables x and y showing two infinitesimal bands dx and dy. The area of intersection of the bands is d x d y and f ( x , y ) x y is the probability of finding x in the interval ( x , x + d x ) and y in the interval ( y , y + d y ) .

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780123877604000032

Random Variables

Sheldon M. Ross , in Introduction to Probability Models (Twelfth Edition), 2019

2.5.3 Covariance and Variance of Sums of Random Variables

The covariance of any two random variables X and Y, denoted by Cov ( X , Y ) , is defined by

Cov ( X , Y ) = E [ ( X E [ X ] ) ( Y E [ Y ] ) ] = E [ X Y Y E [ X ] X E [ Y ] + E [ X ] E [ Y ] ] = E [ X Y ] E [ Y ] E [ X ] E [ X ] E [ Y ] + E [ X ] E [ Y ] = E [ X Y ] E [ X ] E [ Y ]

Note that if X and Y are independent, then by Proposition 2.3 it follows that Cov ( X , Y ) = 0 .

Let us consider now the special case where X and Y are indicator variables for whether or not the events A and B occur. That is, for events A and B, define

X = { 1 , if A occurs 0 , otherwise , Y = { 1 , if B occurs 0 , otherwise

Then,

Cov ( X , Y ) = E [ X Y ] E [ X ] E [ Y ]

and, because XY will equal 1 or 0 depending on whether or not both X and Y equal 1, we see that

Cov ( X , Y ) = P { X = 1 , Y = 1 } P { X = 1 } P { Y = 1 }

From this we see that

Cov ( X , Y ) > 0 P { X = 1 , Y = 1 } > P { X = 1 } P { Y = 1 } P { X = 1 , Y = 1 } P { X = 1 } > P { Y = 1 } P { Y = 1 | X = 1 } > P { Y = 1 }

That is, the covariance of X and Y is positive if the outcome X = 1 makes it more likely that Y = 1 (which, as is easily seen by symmetry, also implies the reverse).

In general it can be shown that a positive value of Cov ( X , Y ) is an indication that Y tends to increase as X does, whereas a negative value indicates that Y tends to decrease as X increases.

Example 2.33

The joint density function of X , Y is

f ( x , y ) = 1 y e ( y + x / y ) , 0 < x , y <

(a)

Verify that the preceding is a joint density function.

(b)

Find Cov ( X , Y ) .

Solution:  To show that f ( x , y ) is a joint density function we need to show it is nonnegative, which is immediate, and that f ( x , y ) d y d x = 1 . We prove the latter as follows:

f ( x , y ) d y d x = 0 0 1 y e ( y + x / y ) d y d x = 0 e y 0 1 y e x / y d x d y = 0 e y d y = 1

To obtain Cov ( X , Y ) , note that the density function of Y is

f Y ( y ) = e y 0 1 y e x / y d x = e y

Thus, Y is an exponential random variable with parameter 1, showing (see Example 2.21) that

E [ Y ] = 1

We compute E [ X ] and E [ X Y ] as follows:

E [ X ] = x f ( x , y ) d y d x = 0 e y 0 x y e x / y d x d y

Now, 0 x y e x / y d x is the expected value of an exponential random variable with parameter 1 / y , and thus is equal to y. Consequently,

E [ X ] = 0 y e y d y = 1

Also

E [ X Y ] = x y f ( x , y ) d y d x = 0 y e y 0 x y e x / y d x d y = 0 y 2 e y d y

Integration by parts ( d v = e y d y , u = y 2 ) gives

E [ X Y ] = 0 y 2 e y d y = y 2 e y | 0 + 0 2 y e y d y = 2 E [ Y ] = 2

Consequently,

Cov ( X , Y ) = E [ X Y ] E [ X ] E [ Y ] = 1

The following are important properties of covariance.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B978012814346900007X

Elements of Probability

Sheldon Ross , in Simulation (Fifth Edition), 2013

2.10 Conditional Expectation and Conditional Variance

If X and Y are jointly discrete random variables, we define E X | Y = y , the conditional expectation of X given that Y = y , by

E X | Y = y = x xP { X = x | Y = y } = x xP { X = x , Y = y } P { Y = y }

In other words, the conditional expectation of X , given that Y = y , is defined like E X as a weighted average of all the possible values of X , but now with the weight given to the value x being equal to the conditional probability that X equals x given that Y equals y .

Similarly, if X and Y are jointly continuous with joint density function f ( x , y ) , we define the conditional expectation of X , given that Y = y , by

E X | Y = y = xf ( x , y ) dx f ( x , y ) dx

Let E X | Y denote that function of the random variable Y whose value at Y = y is E X | Y = y ; and note that E X | Y is itself a random variable. The following proposition is quite useful.

Proposition 7

(2.11) E E X | Y = E X

If Y is a discrete random variable, then Equation (2.11) states that

E X = y E X | Y = y P { Y = y }

whereas if Y is continuous with density g , then (2.11) states

E X = E X | Y = y g ( y ) dy

We now give a proof of the preceding proposition when X and Y are discrete:

y E X | Y = y P { Y = y } = y x xP { X = x | Y = y } P { Y = y } = y x xP { X = x , Y = y } = x x y P { X = x , Y = y } = x xP { X = x } = E X

We can also define the conditional variance of X , given the value of Y , as follows:

Var ( X | Y ) = E ( X - E X | Y ) 2 | Y

That is, Var ( X | Y ) is a function of Y , which at Y = y is equal to the variance of X given that Y = y . By the same reasoning that yields the identity Var ( X ) = E X 2 - ( E X ) 2 we have that

Var ( X | Y ) = E X 2 | Y - ( E X | Y ) 2

Taking expectations of both sides of the above equation gives

(2.12) E Var ( X | Y ) = E E X 2 | Y - E ( E X | Y ) 2 = E X 2 - E ( E X | Y ) 2

Also, because E E X | Y = E X , we have that

(2.13) Var ( E X | Y ) = E ( E X | Y ) 2 - ( E X ) 2

Upon adding Equations (2.12) and (2.13) we obtain the following identity, known as the conditional variance formula.

The Conditional Variance Formula

Var ( X ) = E Var ( X | Y ) + Var ( E X | Y )

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780124158252000024

The local Gaussian partial correlation

Dag Tjøstheim , ... Bård Støve , in Statistical Modeling Using Local Gaussian Approximation, 2022

11.2 The local Gaussian partial correlation

Let X = ( X 1 , , X p ) T be a random vector. Denote by ( X 1 , X 2 , X 3 ) a partition of X into vectors of dimensions p 1 , p 2 , and p 3 , respectively, such that X ( 1 ) = ( X 1 , X 2 ) = ( X 1 , , X p 1 + p 2 ) T consists of the first p 1 + p 2 components in X , and X ( 2 ) = X 3 = ( X p 1 + p 2 + 1 , , X p 1 + p 2 + p 3 ) T contains the remaining p 3 variables, where p = p 1 + p 2 + p 3 . We assume that the mean vector μ and covariance matrix Σ of X exist and partition them correspondingly writing

(11.1) μ = ( μ 1 μ 2 μ 3 ) and Σ = ( Σ 11 Σ 12 Σ 13 Σ 21 Σ 22 Σ 23 Σ 31 Σ 32 Σ 33 ) ,

where Σ i j is the covariance matrices of ( X i , X j ) , i , j = 1 , 2 , 3 . There are two main concepts of correlation when X ( 2 ) = X 3 is given, the partial and conditional correlations, which coincide in several joint distributions, among them, the Gaussian. See, for example, Baba et al. (2004) for details. We will use the partial correlation as a starting point when defining the LGPC. The partial variance–covariance matrix of X ( 1 ) = ( X 1 , X 2 ) given X ( 2 ) = X 3 is

(11.2) Σ 12 | 3 = Σ 11 Σ 12 ( Σ 22 ) 1 Σ 21 ,

where

Σ 11 = ( Σ 11 Σ 12 Σ 21 Σ 22 ) , Σ 12 = ( Σ 13 Σ 23 ) , Σ 21 = ( Σ 31 Σ 32 ) , and Σ 22 = Σ 33 ,

and Σ 12 | 3 is the covariance matrix in the conditional (Gaussian) distribution of X ( 1 ) given X 3 if X is jointly normal. The partial correlation matrix between X 1 and X 2 given X 3 is naturally defined as

(11.3) R 12 | 3 = D 1 / 2 Σ 12 | 3 D 1 / 2 ,

where D = diag ( Σ 12 | 3 ) . We identify in the same way the partial correlation matrix (11.3) with the correlation matrix in the conditional (Gaussian) distribution of X ( 1 ) given X ( 2 ) if X is jointly normal. Eqs. (11.2) and (11.3) will serve as the starting point for our definition of the local partial correlation.

11.2.1 Definition

We further assume that the components of X are continuous with joint density function f X , and we again set up a local likelihood framework for obtaining local estimates of the parameters in the multivariate normal distribution. Given a point x , we approximate f X in a neighborhood of x by a multivariate Gaussian density

(11.4) ψ ( x , v ) = 1 ( 2 π ) p / 2 | Σ ( x ) | 1 / 2 exp { 1 2 ( v μ ( x ) ) T Σ 1 ( x ) ( v μ ( x ) ) } ,

where x = ( x 1 , x p ) T , μ ( x ) = { μ j ( x ) } , and Σ ( x ) = { σ j k ( x ) } for j , k = 1 , , p . Moving to another point y , there is another (generally, different) Gaussian approximation ψ ( y , v ) . In this way, we approximate f X by a family of multivariate Gaussian densities defined by a set of smooth parameter functions { μ ( x ) , Σ ( x ) } , and if f X is itself a Gaussian density, then the parameter functions collapse to constants corresponding to the true parameter values, and ψ ( x ) f X ( x ) . Hjort and Jones (1996) provide a general framework for estimating such parameter functions non-parametrically from a given data set using a local likelihood procedure, and the basic idea in the following treatment is replacing the components in the partial covariance matrix (11.2) by their locally estimated counterparts to obtain a local measure of conditional dependence.

In this chapter, we use the same transformation technique as that introduced already in Chapter 4.7 and subsequently used in the following chapters. This improves and simplifies the estimation of the LGPC. It is shown that the estimation of the local parameter functions { μ ( x ) , Σ ( x ) } becomes easier by transforming each X j to a standard normal variable Z j = Φ 1 ( U j ) , where U j is a uniform variable, U j = F j ( X j ) with F j being the cumulative distribution function of X j . Define the random vector Z by this transformation of X = ( X ( 1 ) , X ( 2 ) ) = ( X 1 , X 2 , X 3 ) = ( X 1 , X 2 , , X p ) T to marginal standard normality:

(11.5) Z = ( Φ 1 ( F 1 ( X 1 ) ) , Φ 1 ( F 2 ( X 2 ) ) , , Φ 1 ( F p ( X p ) ) ) T .

The transformation enables us to simplify the local Gaussian approximation (11.4) by writing the density f Z of Z at the point v = z as

(11.6) f Z ( z ) = ψ ( z , R ( z ) ) = 1 | 2 π R ( z ) | 1 / 2 exp { 1 2 z T R 1 ( z ) z } ,

where, as in Otneim and Tjøstheim (2017, 2018) and in Eqs. (9.6) and (9.7) in Chapter 9, in a further simplified approximation, we have fixed local means and standard deviations μ j ( z ) 0 and σ j 2 ( z ) 1 , j = 1 , , p , and where R ( z ) = { ρ j k ( z ) } is the local correlation matrix.

In practice, we do not know F j , but we can instead use the empirical distribution function

F ˆ j ( x ) = 1 n i = 1 n I ( X j i x ) ,

where I is the indicator function, n is the number of observations, and X j i is the ith observation of X j , and where 1 / n can be replaced by 1 / ( n + 1 ) for small or moderate sample sizes. This results in pseudo-standard normal variables Z ˆ j = Φ 1 ( F ˆ j ( X j ) ) .

In the following, we will not always distinguish between Z j and Z ˆ j . In fact, by using the technique of proof of Theorem 4.5 under the regularity conditions of that theorem, the error made by estimating R ( z ) using the empirically transformed variables Z ˆ j instead of Z j is smaller in the limit than the estimation error made when estimating the local correlations themselves.

In this chapter, we refer to X and its probability density function f X as being on the x -scale and to Z and its probability density function f Z as being on the z -scale. For further discussion of the simplified z -approximation, we refer to Chapters 9.2.2 and 9.7.

Denote by ( Z ( 1 ) , Z ( 2 ) ) = ( Z 1 , Z 2 , Z 3 ) the partitioning of Z corresponding to the partitioning ( X ( 1 ) , X ( 2 ) ) = ( X 1 , X 2 , X 3 ) of X . A natural definition of the local partial covariance matrix of Z ( 1 ) | Z ( 2 ) is the local version of Eq. (11.2):

(11.7) Σ 12 | 3 ( z ) = R 11 ( z ( 1 ) ) R 12 ( z ) ( R 22 ( z ( 2 ) ) ) 1 R 21 ( z ) .

If p 1 = p 2 = 1 , then Σ 12 | 3 ( z ) is a 2 × 2 matrix, and we define the local Gaussian partial correlation α ( z ) between the two variables in Z ( 1 ) = ( Z 1 , Z 2 ) given Z ( 2 ) = Z 3 in accordance with the ordinary (global) partial correlation provided by Eq. (11.3):

(11.8) α ( z ) = R 12 | 3 ( z ) = { Σ 12 | 3 ( z ) } 12 { Σ 12 | 3 ( z ) } 11 1 / 2 { Σ 12 | 3 ( z ) } 22 1 / 2 ,

which, when Z ( 2 ) = Z 3 is scalar, reduces to

(11.9) α ( z ) = ρ 12 | 3 ( z 1 , z 2 | z 3 ) = ρ 12 ( z 1 , z 2 ) ρ 13 ( z 1 , z 3 ) ρ 23 ( z 2 , z 3 ) 1 ρ 13 2 ( z 1 , z 3 ) 1 ρ 23 2 ( z 2 , z 3 ) .

This is easily recognizable as a local version of the standard global partial correlation coefficient. It is of course possible to introduce an LGPC α ( x ) directly on the x -scale, but that representation is in many ways harder to handle both computationally and asymptotically. For a multivariate Gaussian distribution, we have that α X ( x ) = α Z ( z ) = α . In the remainder of this chapter, we write mainly in terms of the z -representation using the LGPC α ( z ) = α Z ( z ) , but when we write the local partial correlation between X 1 and X 2 given X 3 = x 3 at the point ( x 1 , x 2 , x 3 ) , this is simply α ( z ) with inserted z j = Φ 1 ( F j ( x j ) ) , j = 1 , , p .

In the more general case where p 1 > 1 and/or p 2 > 1 , Σ 12 | 3 ( z ) is a ( p 1 + p 2 ) × ( p 1 + p 2 ) matrix that describes the non-linear conditional dependence among the variables in Z ( 1 ) = ( Z 1 , Z 2 ) given Z ( 2 ) = Z 3 (or, alternatively, X ( 1 ) = ( X 1 , X 2 ) given X ( 2 ) = X 3 ), which in particular can be used to analyze the conditional dependence between two sets of random variables of dimensions p 1 and p 2 , respectively, given a third set of variables of dimension p 3 . This case poses no conceptual challenges to our approach, but it requires a rather large investment in new notation and does not lead to simple formulas like Eq. (11.9). We will therefore, for the most part, focus on the local Gaussian partial correlation between two stochastic scalar variables in the remainder of this chapter. A complete description of the general multivariate case can be found in the online supplement of Otneim and Tjøstheim (2021) and, somewhat briefly, in Section 11.8.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780128158616000183

RANDOM VARIABLE THEORY

Daniel T. Gillespie , in Markov Processes, 1992

Random Variable Transformation (RVT) Theorem

If the n random variables X 1, …, Xn have the joint density function P(x 1,…,xn ), and if the m random variables Y 1, …, Ym are defined by Yi =fi (X 1,…,Xn ) [i=1 to m], then the joint density function Q(y 1,…,ym ) of Y 1, …, Ym is given by

(1.6-4) Q ( y 1 , , y m ) = d x 1 d x n P ( x 1 , , x n ) i = 1 m δ ( y i f i ( x 1 , , x n ) ) .

A useful special case of the RVT theorem is the following result:

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780080918372500065

SPECIAL RANDOM VARIABLES

Sheldon M. Ross , in Introduction to Probability and Statistics for Engineers and Scientists (Fourth Edition), 2009

EXAMPLE 5.4e

The random vector X, Y is said to have a uniform distribution over the two-dimensional region R if its joint density function is constant for points in R, and is 0 for points outside of R. That is, if

f ( x , y ) = { c if ( x , y ) R 0 if otherwise

Because

1 = R f ( x , y ) d x d y = R c d x d y = c × Area of R

it follows that

c = 1 Area of R

For any region AR,

P { ( X , Y ) A } = ( x , y ) A f ( x , y ) d x d y = ( x , y ) A c d x d y = Area of A Area of R

Suppose now that X, Y is uniformly distributed over the following rectangular region R:

Its joint density function is

f ( x , y ) = { c if 0 x a , 0 y b 0 otherwise

where c = 1 Area of rectangle = 1 a b . In this case, X and Y are independent uniform random variables. To show this, note that for 0 ≤ xa, 0yb

(5.4.5) P { X x , Y y } = c 0 x 0 y d y d x = x y a b

First letting y = b, and then letting x = a, in the preceding shows that

(5.4.6) P { X x } = x a , P { Y y } = y b

Thus, from Equations 5.4.5 and 5.4.6 we can conclude that X and Y are independent, with X being uniform on (0, a) and Y being uniform on (0, b).

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780123704832000102

Multivariate stochastic orders

Félix Belzunce , ... Julio Mulero , in An Introduction to Stochastic Orders, 2016

3.5 The multivariate likelihood ratio order

In a similar way to the univariate case, it is possible to check the multivariate stochastic and hazard rate orders in terms of a property of the joint density functions. This property leads to the definition of the multivariate likelihood ratio order, which is a generalization of the univariate likelihood ratio order. In this section, the results are given in the continuous case but they can be restated for the general case. The main intention of the multivariate likelihood ratio order is to provide a sufficient condition for the multivariate hazard rate order.

Definition 3.5.1

Given two continuous random vectors X = (X 1,…,X n ) and Y = (Y 1,…,Y n ) with joint density functions f and g, respectively, we say that X is smaller than Y in the multivariate likelihood ratio order, denoted by Xlr Y, if

f ( x ) g ( y ) f ( x y ) g ( x y ) , for all x , y R n .

Clearly, this is a generalization of the likelihood ratio order in the univariate case. However, the multivariate likelihood ratio order is not necessarily reflexive. When a random vector X satisfies Xlr X, we have the MTP2 property introduced in Section 1.3.

The following result provides a set of sufficient conditions for the multivariate likelihood ratio order.

Theorem 3.5.2

Let X = (X 1,…,X n ) and Y = (Y 1,…,Y n ) be two continuous random vectors with joint density functions f and g, respectively. If X or Y or both are MTP2, and

(3.13) f ( y ) g ( x ) f ( x ) g ( y ) , for all x , y R n , such that x y ,

then

X lr Y .

Proof

Let us assume that X is MTP2 (in the other case the proof is similar), then we have the following inequalities:

f ( x ) g ( y ) f ( x y ) f ( x y ) f ( y ) g ( y ) f ( x y ) g ( x y ) ,

for all x , y R n , such that xy, where the first inequality follows from the MTP2 property and the second one from (3.13). Therefore, Xlr Y.

The multivariate likelihood ratio order is preserved under conditioning on sublattices, as we see next. Recall that a subset A R n is called a sublattice if x,yA implies xyA and xyA. This result will be used to show the relationship among the multivariate likelihood ratio order and the multivariate dynamic hazard rate order. The proof is obvious from the definition.

Theorem 3.5.3

Let X = (X 1,…,X n ) and Y = (Y 1,…,Y n ) be two continuous random vectors. If Xlr Y, then

[ X | X A ] lr [ Y | Y A ] , for all sublattice A R n .

In particular, from the previous theorem, the multivariate likelihood ratio order is preserved under marginalization. This result is useful because, in some cases, it is easier to provide a result in the multivariate case rather that in the univariate case.

Theorem 3.5.4

Let X = (X 1,…,X n ) and Y = (Y 1,…,Y n ) be two continuous random vectors. If Xlr Y, then

X I lr Y I , for all I { 1 , , n } .

Next, it is showed that the multivariate likelihood ratio order is stronger than the multivariate hazard rate order.

Theorem 3.5.5

Let X = (X 1,…,X n ) and Y = (Y 1,…,Y n ) be two continuous random vectors. If Xlr Y, then

X hr Y .

Proof

Let us check the conditions for the definition of the multivariate dynamic hazard rate order. In particular, denoting by η and λ the multivariate dynamic hazard rates of X and Y, respectively, let us see if

η k ( t | h t ) λ k ( t | h t ) , for all t 0 ,

where

h t = { X I J = x I J , X I J ¯ > t e } ,

and

h t = { Y I = y I , Y I ¯ > t e } ,

whenever I J = , 0x I y I t e, 0x J t e, and for all k I J ¯ .

The result will follow by proving that, as we shall see later,

(3.14) X I J ¯ h t lr Y I J ¯ h t .

Condition (3.14) will follow if

(3.15) X I J ¯ X I J = x I J lr Y I J ¯ Y I = y I , Y J > t e ,

holds.

Let us see that (3.15) follows if Xlr Y holds. Denoting by f and g the joint density functions of ( X I , X J , X I J ) and ( Y I , Y J , Y I J ) , respectively, and by f I,J and g I,J the joint densities of X I,J and Y I,J , respectively, we see that the joint density function of X I J ¯ X I J = x I J is given by

f I J ¯ ( x I J ¯ ) = f ( x I , x J , x I J ¯ ) f I , J ( x I , x J ) ,

and the joint density function of Y I J ¯ Y I = y I , Y J > t e is given by

g I J ¯ ( x I J ¯ ) = y J > t e g ( y I , y J , x I J ¯ ) d y J y J > t e g I , J ( y I , y J ) d y J .

Given y J > t e, we see that x J < t e <y J and, analogously, x I y I . Since Xlr Y, given y I J ¯ , we see that

f ( x I , x J , x I J ¯ ) g ( y I , y J , y I J ¯ ) f ( x I , x J , x I J ¯ y I J ¯ ) g ( y I , y J , x I J ¯ y I J ¯ ) ,

which, upon integration, yields

f I J ¯ ( x I J ¯ ) g I J ¯ ( y I J ¯ ) f I J ¯ ( x I J ¯ y I J ¯ ) g I J ¯ ( x I J ¯ y I J ¯ ) .

Therefore, from previous inequality, we see that (3.15) holds. Now, from Theorem 3.5.3, we see that (3.14) also holds. In particular, X k h t lr Y k h t holds, for all k I J ¯ . Now, since (2.25), we see that the hazard rates of X k h t and Y k h t are ordered and, consequently, we see that

η k ( t | h t ) λ k ( t | h t ) .

Finally, a result for the multivariate likelihood ratio order among random vectors with conditionally independent components is provided. In particular, we consider the same background as that in Theorem 3.3.8.

Theorem 3.5.6

Assume that

(i)

X i (θ) =st Y i (θ), for all θ and for all i = 1,…,n,

(ii)

X i (θ) ≤lr Y j (θ′), for all θθ′ and for all 1 ≤ ijn, and

(iii)

θ 1lr θ 2.

Then,

( X 1 , , X n ) lr ( Y 1 , , Y n ) .

Proof

Let f i (x|θ) be the density function of X i (θ). From condition (ii), we see that

i = 1 n f i ( x i | θ i ) is TP 2 in ( x 1 , , x n , θ 1 , , θ n ) .

Furthermore, condition (iii) is equivalent to the fact that h i (θ 1,…,θ n ) is TP2 in (θ 1,…,θ n ,i) ∈ S ×{1,2}. Due to these facts and from Theorem 1.2.3, we see that f ( x 1 , , x n | θ 1 , , θ n ) d H i ( θ ) is TP2 in (x 1,…,x n ,i), and we get the result.

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B978012803768300003X

Convex Functions, Partial Orderings, and Statistical Applications

In Mathematics in Science and Engineering, 1992

13.6 Some Properties of Log-Concave Density Functions

Log-concave density functions which satisfy (13.19) play an important role in statistics and probability. In the following we observe some known facts concerning this class of densities.

13.24 Fact

Let X 1 …, X n be i.i.d. univariate random variables with a common density function h(x). If h(x) is a log-concave function of x for x ∈ ℝ, then the joint density function of (X 1, …, X n ) is a log-concave function of x for x ∈ ℝ n .

13.25 Fact

If f(x) = g(T(x)) where g: ℝ → [0, ∞) is decreasing and T(x) is a convex function of x for x ∈ ℝ n , then f is a log-concave function of x for x ∈ ℝ n .

The following theorem, due to Prékopa (1971) and Brascamp and Lieb (1975), shows that the integral of a log-concave function is log-concave:

13.26 Theorem

Let f(x, y): ℝ n+m → [0, ∞) be a log-concave function of (x, y) for x ∈ ℝ n and y ∈ ℝ m . Then the function g: ℝ n → [0, ∞) given by

(13.27) g ( x ) = m f ( x , y ) d y

is log-concave.

Proof

We adopt the proof in Brascamp and Lieb (1975). First note that it suffices to prove the theorem for m = n = 1 because the general case follows by Fubini's theorem and induction. Let x 1, x 2 be two points in ℝ such that g(x 1)g(x 2) ≠ 0. For convenience we may assume that

sup y f ( x , y ) = sup y f ( x , y ) ;

because otherwise we can replace f(x, y) by e bx f(x, y) for suitably chosen b and the problem remains unchanged. For each fixed λ > 0, denote

C 1 ( λ ) = { ( x , y ) : f ( x , y ) λ } 2 , C 2 ( x , λ ) = { y : f ( x , y ) λ } .

Then, by log-concavity of f, C 1(λ) is convex and C 2(x, λ) is an interval. (For the convexity of C 1(λ), see Fact 13.28). Letting v(x, y) = ∫ C 2(x,λ) dy be the Lebesgue measure of the set C 2(x, λ), we have, by Theorem 13.18,

v ( α x 1 + ( 1 α ) x 2 , λ ) α v ( x 1 , λ ) + ( 1 α ) v ( x 2 , λ ) for all α [ 0 , 1 ] .

Since g(x) can be expressed as g(x) = ∫0 v(x, λ) dλ, we have

g ( α x 1 ) + ( 1 α ) x 2 ) α g ( x 1 ) + ( 1 α ) g ( x 2 ) ( g ( x 1 ) ) α ( g ( x 2 ) ) 1 α

for all α ∈ [0, 1], where the second inequality follows from the arithmetic mean-geometric mean inequality.

A simple application of Theorem 13.26 is (Brascamp and Lieb, 1975; see also Barlow and Proschan, 1981, p. 104):

13.27 Corollary

The convolution of two log-concave density functions in n is log-concave.

Proof

Let f 1, f 2 be log-concave density functions. Then f 1(xy)f 2(y) is jointly log-concave in (x, y) ∈ ℝ2n . Thus by Theorem 13.26

g ( x ) = n f 1 ( x y ) f 2 ( y ) d y

is log-concave.

A density function f is said to be unimodal if the set

(13.28) D λ = { x : x n , f ( x ) λ }

is a convex set in ℝ n for all λ > 0. The following facts show how log-concavity and unimodality are related.

13.28 Fact

If f: ℝ n → [0, ∞) is a probability density function, then log-concavity off implies unimodality of f.

Proof

Let x 1, x 2 ∈ ℝ n be in D λ. Then for every α ∈ [0, 1], we have, by (13.20),

(13.29) log f ( α x 1 + ( 1 α ) x 2 ) α log f ( x 1 ) + ( 1 α ) log f ( x 2 ) α log λ + ( 1 α ) log λ log λ .

Thus αx 1 + (1 − α)x 2 is also in D λ.

A function f is said to be Schur-concave if yx implies f(x) ≥ f(y) for all x, y ∈ ℝ n (see Definition 12.23). It is known that all Schur-concave functions are permutation invariant (see Theorem 12.24). Furthermore, it is known that

13.29 Fact

If f: ℝ n → [0, ∞) is a permutation-invariant and log-concave function of x ∈ ℝ n , then it is a Schur-concave function of x ∈ ℝ n .

Proof

Assume yx and, without loss of generality, it may be assumed that x, y are of the form

x = ( x 1 , x 2 , x 3 , , x n ) , y = ( y 1 , y 2 , x 3 , , x n ) ,

where y 2 < x 2x 1 < y 1 and x 1 + x 2 = y 1 + y 2. Let y * = (y 2, y 1, x 3, …, x n ). Then there exists an α ∈ (0, 1) such that x = αy + (1 − α)y *. Thus by the permutation-invariance and log-concavity properties of f, we have

log f ( x ) = log f ( α y + ( 1 α ) y * ) α log f ( y ) + ( 1 α ) log f ( y ) = log f ( y ) .

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/S0076539208628258

Preliminaries

Félix Belzunce , ... Julio Mulero , in An Introduction to Stochastic Orders, 2016

1.3.5 Parametric families of multivariate distributions

In this section, some known multivariate distributions are recalled. These models will be compared in Chapter 3 in some multivariate stochastic orders.

It is worth mentioning that the mean of a random vector X = (X 1,…,X n ) is given by

E [ X ] = E [ X 1 ] E [ X n ] ,

and the covariance matrix of X is

Cov ( X ) = E [ ( X E [ X ] ) ( X E [ X ] ) T ] .

Next, we consider some multivariate distributions. First, the definition of the multivariate normal distribution is recalled.

Definition 1.3.2

Given a random vector X = (X 1,…,X n ), it is said that X follows a multivariate normal distribution with mean vector μ R n and covariance matrix Σ R n × R n , denoted by XN n (μ,Σ ), if its joint density function is given by

f ( x ) = 1 ( 2 π ) n 2 | Σ | 1 2 exp 1 2 ( x μ ) T Σ 1 ( x μ ) , for all x R n .

The marginal distribution functions follow univariate normal models. Furthermore, the copula of a N n (μ,Σ) is the same to that of N n (0,P) where P is the correlation matrix obtained through the covariance matrix Σ. In this sense, all multivariate normal distributions with the same dimension and correlation matrix have the same (Gaussian) copula.

Another well-known multivariate family is the elliptically contoured model. Let us consider the formal definition.

Definition 1.3.3

Given a random vector X = (X 1,…,X n ), it is said that X follows an elliptically contoured distribution, denoted by E n (μ,Σ,g), if its joint density function is given by

(1.17) f ( x ) = 1 | Σ | g ( ( x μ ) t Σ 1 ( x μ ) ) , for all x R n ,

where μ is the median vector (which is also the mean vector if the latter exists), Σ is a symmetric positive definite matrix which is proportional to the covariance matrix, if the latter exists, and g : R + R + such that g ( x ) d x < + .

A particular case of elliptically contoured distributions is the case of a multivariate normal taking

g ( x ) = 1 ( 2 π ) n exp 1 2 x .

Some other particular cases are described in Ref. [52].

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780128037683000016

GENERAL FEATURES OF A MARKOV PROCESS

Daniel T. Gillespie , in Markov Processes, 1992

2.1 The Markov State Density Function

We consider a time-evolving or 'dynamical' system whose possible states can be represented by points on the real axis, and we let

(2.1-1) X ( t ) the state point, or state, of the system at time t.

We shall assume that the value of X at some initial time t 0 is fixed,

(2.1-2) X ( t 0 ) = x 0 ,

but that X(t) for any t> t 0 can be predicted only probabilistically; more specifically, we assume that X(t) for any given t> t 0 is a random variable, as defined in Section 1.2 Since it makes sense to inquire about the state of the system at successive instants t 1, t 2, …, tn , where t 0 < t 1 < t 2 < … < tn , then we can ascribe to the corresponding n random variables X(t 1), X(t 2), …, X(tn ) a joint density function P n (1), which is defined as follows:

(2.1-3) P n ( 1 ) ( x n , t n ; x n 1 , t n 1 ; ; x 1 , t 1 | x 0 , t 0 ) d x n d x n 1 d x 1 P r o b { X ( t i ) [ x i , x i + d x i ) for i = 1 , 2 , , n given that X ( t 0 ) = x 0 , with t 0 t 1 t n }.

If all these assumptions are satisfied, then we say that X(t) is a stochastic process.

It is evident that a stochastic process X(t) has infinitely many joint density functions Pn (1), corresponding to n = 1, 2, … . And associated with each of these joint density functions is a plethora of subordinate density functions; for example,

P n j ( j + 1 ) ( x n , t n ; ; x j + 1 , t j + 1 | x j , t J ; ; x 1 , t 1 ; x 0 , t 0 ) ,

is defined to be the joint density function of the nj random variables X ( t j + 1 ) , , X ( t n ) given the j + 1 conditions X ( t 0 ) = x 0 , X ( t 1 ) = x 1 , , X ( t j ) = x j . Notice that the subscript on the density function Pk (j) refers to the number of (x,t) pairs to the left of the 'given' bar, while the superscript refers to the number of (x,t) pairs to the right of the 'given' bar; thus, Pk (j) is a k-variate joint density function with j conditionings.

It is always possible to calculate the function P n-1 (1) from the function Pn (1) by simply integrating the latter over any one of the variables x 1, …, xn . However, it is not in general possible to deduce the function P n+1 (1) from the function Pn (1). This 'open-ended' nature of the density functions for a general stochastic process usually makes any substantive analysis extremely difficult. But we shall be concerned here with only a very restricted subclass of stochastic processes, namely those that have the 'past-forgetting' property that, for all j ≥ 2 and t i-1ti ,

(2.1-4) P 1 ( j ) ( x j , t j | x j 1 , t j 1 ; ; x 1 , t 1 ; x 0 , t 0 ) = P 1 ( 1 ) ( x j , t j | x j 1 , t j 1 ) P ( x j , t j | x j 1 , t J 1 ) .

This is called the Markov property, and it says that only the most recent conditioning matters: Given that X(t') = x', then our ability to predict X(t) for any t > t' will not be enhanced by a knowledge of any values of the process earlier than t'. Any stochastic process X(t) that has this past-forgetting property is called a Markovian stochastic process, or more simply, a Markov process. In what follows it may always be assumed, unless explicitly stated otherwise, that the stochastic process X(t) under consideration is a Markov process.

The Markov property (2.1-4) breaks the open-endedness of the hierarchy of joint state density functions in a dramatic way. For the joint density function P 2 (1) we have

P 2 ( 1 ) ( x 2 , t 2 ; x 1 , t 1 | x 0 , t 0 ) = P 1 ( 1 ) ( x 1 , t 1 | x 0 , t 0 ) P 1 ( 2 ) ( x 2 , t 2 | x 1 , t 1 ; x 0 , t 0 ) [ by ( 1 . 5 9 d ) ]

= P 1 ( 1 ) ( x 1 , t 1 | x 0 , t 0 ) P 1 ( 1 ) ( x 2 , t 2 | x 1 , t 1 ) . [ by ( 2 . 1 4 ) ]

Hence, writing P 1 (1)P in accordance with the notation suggested in Eq. (2.1-4), we have

(2.1-5) P 2 ( 1 ) ( x 2 , t 2 ; x 1 , t 1 | x 0 , t 0 ) = P ( x 2 , t 2 | x 1 , t 1 ) P ( x 1 , t 1 | x 0 , t 0 ) .

The same kind of reasoning shows that

P 3 ( 1 ) ( x 3 , t 3 ; x 2 , t 2 ; x 1 , t 1 | x 0 , t 0 ) = P ( x 3 , t 3 | x 2 , t 2 ) P ( x 2 , t 2 | x 1 , t 1 ) P ( x 1 , t 1 | x 0 , t 0 ) ,

and more generally, for any set of times t n t n 1 t 0 ,

(2.1-6) P n ( 1 ) ( x n , t n ; ; x 1 , t 1 | x 0 , t 0 ) = i =1 n P ( x i , t i | x i 1 , t i 1 )

So for a Markov process, every conditioned state density function P n ( 1 ) can be written solely in terms of the particular conditioned state density function P 1 ( 1 ) P The function P 1 ( 1 ) P thus becomes the principle focus of our study, and we shall henceforth refer to it as the Markov state density function. For future reference, the formal definition of the Markov state density function is [cf.Eq. (2.1-3)]

(2.1-7) P ( x 2 , t 2 | x 1 , t 1 ) d x 2 Prob { X ( t 2 ) [ x 2 , x 2 + d x 2 ) , given X ( t 1 ) = x 1 , with t 2 t 1 } .

Read full chapter

URL:

https://www.sciencedirect.com/science/article/pii/B9780080918372500077