SlideShare une entreprise Scribd logo
1  sur  6
Télécharger pour lire hors ligne
Advanced Communication’s Theory
Session 2
H.Amindavar
February 2021
Multivariate Normal Distribution
Definitions
The multivariate normal distribution is often used to describe, at least approximately, any set of (possibly)
correlated real-valued random variables each of which clusters around a mean value. The multivariate normal
distribution of a k−dimensional random vector
X = (X1, . . . , Xk)T
can be written in the following notation:
X ∼ N (µ, Σ),
or to make it explicitly known that X is k−dimensional,
Xk ∼ N (µk, σk),
with k−dimensional mean vector
µ = E[X] = (E[X1], E[X2], . . . , E[Xk])T
,
and k × k covariance matrix
Σi,j := E[(Xi − µi)(Xj − µj)] = Cov[Xi, Xj]
such that 1 ≤ i, j ≤ k.
The inverse of the covariance matrix is called the precision matrix, denoted by Q = Σ−1
.
A real random vector X = (X1, . . . , Xk)T
is called a standard normal random vector if all of its
components Xn are independent and each is a zero-mean unit-variance normally distributed random vari-
able, i.e. if Xn ∼ N (0, 1) for all n.
A real random vector X = (X1, . . . , Xk)T
is called a centered normal random vector if there exists
a deterministic k × ` matrix A such that AZ has the same distribution as X where Z is a standard normal
random vector with ` components.
A real random vector X = (X1, . . . , Xk)T
is called a normal random vector if there exists a random
` -vector Z, which is a standard normal random vector, a k−vector µ, and a k × ` matrix A, such that
X = AZ + µ.
X ∼ N (µ, Σ) ⇐⇒ there exist µ ∈ Rk
, A ∈ Rk×`
such that X = AZ + µ for Zn ∼ N (0, 1), i.i.d.
1
Figure 1: joint probability density function of bivariate Gaussian with zero mean and [0.25 0.3; 0.3 1]
covariance matrix
Here the covariance matrix is Σ = AAT
.
In the degenerate case where the covariance matrix is singular, the corresponding distribution has no
density. This case arises frequently in statistics; for example, in the distribution of the vector of residuals in
the ordinary least squares regression. The Xi are in general not independent; they can be seen as the result
of applying the matrix A to a collection of independent Gaussian variables Z.
A random vector X = (X1, . . . , Xk)T
has a multivariate normal distribution if it satisfies one of
the following equivalent conditions.
• Every linear combination Y = a1X1 + · · · + akXk of its components is normally distributed. That
is, for any constant vector a ∈ Rk
, the random variable Y = aT
X has a univariate normal distribution,
where a univariate normal distribution with zero variance is a point mass on its mean.
• There is a k-vector µ and a symmetric, positive semidefinite k × k matrix ΣΣ, such that the charac-
teristic function of X is
ϕX(u) = exp

iuT
µ − 1
2
uT
Σu

.
The spherical normal distribution can be characterised as the unique distribution where components are
independent in any orthogonal coordinate system.
2
Density function
Non-degenerate case
The multivariate normal distribution is said to be ”non-degenerate” when the symmetric covariance matrix
Σ is positive definite. In this case the distribution has density:
fX(x1, . . . , xk) =
exp −1
2
(x − µ)T
Σ−1
(x − µ)

p
(2π)k | Σ|
where x is a real k−dimensional column vector and |Σ| ≡ det Σ is the determinant of Σ.
The equation above reduces to that of the univariate normal distribution if Σ is a 1 × 1 matrix (i.e. a single
real number).
the locus of points in k−dimensional space each of which gives the same particular value of the density
— is an ellipse or its higher-dimensional generalization; hence the multivariate normal is a special case of
the elliptical distributions.
The quantity
q
(x − µ)TΣ−1
(x − µ) is known as the Mahalanobis distance, which represents the
distance of the test point x from the mean µ. In the case when k = 1, the distribution reduces to a univariate
normal distribution and the Mahalanobis distance reduces to the absolute value of the standard score.
Bivariate case
In the 2-dimensional nonsingular case k = rank (Σ) = 2, the probability density function of a vector [X Y]’
is:
f(x, y) =
1
2πσXσY
p
1 − ρ2
e
− 1
2(1−ρ2)
h
(
x−µX
σX
)2
−2ρ(
x−µX
σX
)(
y−µY
σY
)+(
y−µY
σY
)2
i
where ρ is the correlation between X and Y and where σX  0 and σY  0. In this case,
µ =

µX
µY

, Σ =

σ2
X ρσXσY
ρσXσY σ2
Y

.
In the bivariate case, the first equivalent condition for multivariate reconstruction of normality can be
made less restrictive as it is sufficient to verify that countably many distinct linear combinations of X and
Y are normal in order to conclude that the vector of [X Y]’ is bivariate normal.
The bivariate iso-density loci plotted in the x, y-plane are ellipses, whose principal axes are defined by
the eigenvectors of the covariance matrix Σ (the major and minor semidiameters of the ellipse equal the
square-root of the ordered eigenvalues).
-3 -2 -1 0 1 2 3
-3
-2
-1
0
1
2
3
x1
x2
Figure 2: samples of bivariate normal distribution and the contour plot of it’s pdf
3
As the absolute value of the correlation parameter ρ increases, these loci are squeezed toward the following
line :
y(x) = sgn(ρ)
σY
σX
(x − µX) + µY .
is the best linear unbiased prediction of Y given a value of X.
The equidensity contours of a non-singular multivariate normal distribution are ellipsoids (i.e. linear
transformations of hyperspheres) centered at the mean. Hence the multivariate normal distribution is an
example of the class of elliptical distributions.
The directions of the principal axes of the ellipsoids are given by the eigen-vectors of the covariance matrix
Σ. The squared relative lengths of the principal axes are given by the corresponding eigenvalues.
If Σ = UΛUT
is an eigendecomposition where the columns of U are unit eigenvectors and Λ is a diagonal
matrix of the eigenvalues, then we have
X ∼ N (µ, Σ) ⇐⇒ X ∼ µ + UΛ1/2
N (0, I) ⇐⇒ X ∼ µ + UN (0, Λ)
Moreover, U can be chosen to be a rotation matrix, as inverting an axis does not have any effect on
N (0, Λ), but inverting a column changes the sign of Us determinant. The distribution N (µ, Σ) is in effect
N (0, I) scaled by Λ, rotated by U and translated by µ.
-3 -2 -1 0 1 2 3
-3
-2
-1
0
1
2
3
x2
x1
Figure 3: the iso-density loci of pdf and the eigen-vectors of covariance matrix of the bivariate normal
distribution
4
Rayleigh Distribution
In probability theory and statistics, the Rayleigh distribution is a continuous probability distribution
for nonnegative-valued random variables. It is essentially a chi distribution with two degrees of freedom.
The probability density function of the Rayleigh distribution is
f(x; σ) =
x
σ2
e−x2
/(2σ2
)
, x ≥ 0,
where σ is the scale parameter of the distribution. The cumulative distribution function is
F (x; σ) = 1 − e−x2
/(2σ2
)
for x ∈ [0, ∞).
0
0.2
0.4
0.6
0.8
1
1.2
0 2 4 6 8 10
0
0.2
0.4
0.6
0.8
1
1.2
0 2 4 6 8 10
σ=0.5
σ=0.5
σ=0.5
σ=1
σ=1
σ=1
σ=2
σ=2
σ=2
σ=3
σ=3
σ=3
σ=4
σ=4
σ=4
0
0.2
0.4
0.6
0.8
1
1.2
0 2 4 6 8 10
σ=0.5
σ=1
σ=2
σ=3
σ=4
0
0.2
0.4
0.6
0.8
1
1.2
0 2 4 6 8 10
σ=0.5
σ=1
σ=2
σ=3
σ=4
Figure 4: probability density function of Rayleigh distribution
Consider the two-dimensional vector Y = (U, V ) which has components that are normally distributed,
centered at zero, and independent. Then U and V have density functions
fU (x; σ) = fV (x; σ) =
e−x2
/(2σ2
)
√
2πσ2
.
Let X be the length of Y That is, X =
p
U2 + V 2 Then X has cumulative distribution function
FX(x; σ) =
ZZ
Dx
fU (u; σ)fV (v; σ) dA
where Dx is the disk
Dx =
n
(u, v) :
p
u2 + v2  x
o
Writing the double integral in polar coordinates, it becomes
FX(x; σ) =
1
2πσ2
Z 2π
0
Z x
0
re−r2
/(2σ2
)
dr dθ =
1
σ2
Z x
0
re−r2
/(2σ2
)
dr.
Finally, the probability density function for X is the derivative of its cumulative distribution function,
which is
fX(x; σ) =
d
dx
FX(x; σ) =
x
σ2
e−x2
/(2σ2
)
5
which is the Rayleigh distribution. It is straightforward to generalize to vectors of dimension other than
2. There are also generalizations when the components have unequal variance or correlations, or when the
vector Y follows a bivariate Student t-distribution.
Rice Distribution
In probability theory, the Rice distribution or Rician distribution (or, less commonly, Ricean
distribution) is the probability distribution of the magnitude of a circularly-symmetric bivariate normal
random variable, possibly with non-zero mean (noncentral). It was named after Stephen O. Rice.
f(x | ν, σ) =
x
σ2
exp

−(x2
+ ν2
)
2σ2

I0

xν
σ2

,
Figure 5: probability density function of Rayleigh distribution
where I0(z) is the modified Bessel function of the first kind with order zero,
Iα(x) = i−α
Jα(ix) =
∞
X
m=0
1
m! Γ(m + α + 1)

x
2
2m+α
,
Kα(x) =
π
2
I−α(x) − Iα(x)
sin απ
,
α is not an integer; when α is an integer, then the limit is used. Iα(x) and Kα(x) are the two linearly
independent solutions to the modified Bessel’s equation:
x2 d2
y
dx2
+ x
dy
dx
− x2
+ α2

y = 0.
R ∼ Rice (|ν|, σ) if R =
p
X2 + Y 2 where N ν cos θ, σ2

and N ν sin θ, σ2

are statistically in-
dependent normal random variables and θ is any real number.
6

Contenu connexe

Tendances

Double Side band Suppressed carrier (DSB-SC) Modulation and Demodulation.
 Double Side band Suppressed carrier (DSB-SC) Modulation and Demodulation. Double Side band Suppressed carrier (DSB-SC) Modulation and Demodulation.
Double Side band Suppressed carrier (DSB-SC) Modulation and Demodulation.
SAiFul IslAm
 

Tendances (20)

Multiple access techniques for wireless communications
Multiple access techniques for wireless communicationsMultiple access techniques for wireless communications
Multiple access techniques for wireless communications
 
3.2 modulation formats bpsk, qpsk, oqpsk,
3.2 modulation formats   bpsk, qpsk, oqpsk,3.2 modulation formats   bpsk, qpsk, oqpsk,
3.2 modulation formats bpsk, qpsk, oqpsk,
 
ASk,FSK,PSK
ASk,FSK,PSKASk,FSK,PSK
ASk,FSK,PSK
 
Coherent and Non-coherent detection of ASK, FSK AND QASK
Coherent and Non-coherent detection of ASK, FSK AND QASKCoherent and Non-coherent detection of ASK, FSK AND QASK
Coherent and Non-coherent detection of ASK, FSK AND QASK
 
Quadrature phase shift keying
Quadrature phase shift keyingQuadrature phase shift keying
Quadrature phase shift keying
 
Matched filter
Matched filterMatched filter
Matched filter
 
Modulation of LED
Modulation of LEDModulation of LED
Modulation of LED
 
6 slides
6 slides6 slides
6 slides
 
Antennas and Wave Propagation
Antennas and Wave Propagation Antennas and Wave Propagation
Antennas and Wave Propagation
 
OKUMURA, HATA and COST231 Propagation Models
OKUMURA, HATA and COST231 Propagation ModelsOKUMURA, HATA and COST231 Propagation Models
OKUMURA, HATA and COST231 Propagation Models
 
Antennas and Wave Propagation
Antennas and Wave PropagationAntennas and Wave Propagation
Antennas and Wave Propagation
 
5. 2 ray propagation model part 1
5. 2 ray propagation model   part 15. 2 ray propagation model   part 1
5. 2 ray propagation model part 1
 
quantization
quantizationquantization
quantization
 
Digital modulation techniques...
Digital modulation techniques...Digital modulation techniques...
Digital modulation techniques...
 
Gaussian noise
Gaussian noiseGaussian noise
Gaussian noise
 
Double Side band Suppressed carrier (DSB-SC) Modulation and Demodulation.
 Double Side band Suppressed carrier (DSB-SC) Modulation and Demodulation. Double Side band Suppressed carrier (DSB-SC) Modulation and Demodulation.
Double Side band Suppressed carrier (DSB-SC) Modulation and Demodulation.
 
Small Scale Multi path measurements
Small Scale Multi path measurements Small Scale Multi path measurements
Small Scale Multi path measurements
 
Gunn Diode
Gunn Diode Gunn Diode
Gunn Diode
 
7. log distance and log normal shadowing
7. log distance and log normal shadowing7. log distance and log normal shadowing
7. log distance and log normal shadowing
 
Channel capacity
Channel capacityChannel capacity
Channel capacity
 

Similaire à Multivariate Gaussin, Rayleigh & Rician distributions

Machine learning (10)
Machine learning (10)Machine learning (10)
Machine learning (10)
NYversity
 
Machine learning (2)
Machine learning (2)Machine learning (2)
Machine learning (2)
NYversity
 

Similaire à Multivariate Gaussin, Rayleigh & Rician distributions (20)

multivariate normal distribution.pdf
multivariate normal distribution.pdfmultivariate normal distribution.pdf
multivariate normal distribution.pdf
 
The Wishart and inverse-wishart distribution
 The Wishart and inverse-wishart distribution The Wishart and inverse-wishart distribution
The Wishart and inverse-wishart distribution
 
Section3 stochastic
Section3 stochasticSection3 stochastic
Section3 stochastic
 
The Multivariate Gaussian Probability Distribution
The Multivariate Gaussian Probability DistributionThe Multivariate Gaussian Probability Distribution
The Multivariate Gaussian Probability Distribution
 
Distributions
DistributionsDistributions
Distributions
 
Machine learning (10)
Machine learning (10)Machine learning (10)
Machine learning (10)
 
Tensor 1
Tensor  1Tensor  1
Tensor 1
 
Multivariate Distributions, an overview
Multivariate Distributions, an overviewMultivariate Distributions, an overview
Multivariate Distributions, an overview
 
Multivriada ppt ms
Multivriada   ppt msMultivriada   ppt ms
Multivriada ppt ms
 
Cs229 notes9
Cs229 notes9Cs229 notes9
Cs229 notes9
 
Probability distribution
Probability distributionProbability distribution
Probability distribution
 
Metodo gauss_newton.pdf
Metodo gauss_newton.pdfMetodo gauss_newton.pdf
Metodo gauss_newton.pdf
 
Doe02 statistics
Doe02 statisticsDoe02 statistics
Doe02 statistics
 
A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”
 
Probability And Random Variable Lecture6
Probability And Random Variable Lecture6Probability And Random Variable Lecture6
Probability And Random Variable Lecture6
 
2 vectors notes
2 vectors notes2 vectors notes
2 vectors notes
 
Lecture_note2.pdf
Lecture_note2.pdfLecture_note2.pdf
Lecture_note2.pdf
 
Machine learning (2)
Machine learning (2)Machine learning (2)
Machine learning (2)
 
Ch07 5
Ch07 5Ch07 5
Ch07 5
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 

Plus de HAmindavarLectures

Plus de HAmindavarLectures (13)

"Digital communications" undergarduate course lecture notes
"Digital communications" undergarduate course lecture notes"Digital communications" undergarduate course lecture notes
"Digital communications" undergarduate course lecture notes
 
Wavelet Signal Processing
Wavelet Signal ProcessingWavelet Signal Processing
Wavelet Signal Processing
 
Stochastic Processes - part 6
Stochastic Processes - part 6Stochastic Processes - part 6
Stochastic Processes - part 6
 
Stochastic Processes - part 5
Stochastic Processes - part 5Stochastic Processes - part 5
Stochastic Processes - part 5
 
Stochastic Processes - part 4
Stochastic Processes - part 4Stochastic Processes - part 4
Stochastic Processes - part 4
 
Stochastic Processes - part 3
Stochastic Processes - part 3Stochastic Processes - part 3
Stochastic Processes - part 3
 
Stochastic Processes - part 2
Stochastic Processes - part 2Stochastic Processes - part 2
Stochastic Processes - part 2
 
Stochastic Processes - part 1
Stochastic Processes - part 1Stochastic Processes - part 1
Stochastic Processes - part 1
 
Random Variables
Random VariablesRandom Variables
Random Variables
 
Detection & Estimation Theory
Detection & Estimation TheoryDetection & Estimation Theory
Detection & Estimation Theory
 
Cyclo-stationary processes
Cyclo-stationary processesCyclo-stationary processes
Cyclo-stationary processes
 
Introduction to communication systems
Introduction to communication systemsIntroduction to communication systems
Introduction to communication systems
 
Advanced Communications Theory
Advanced Communications Theory Advanced Communications Theory
Advanced Communications Theory
 

Dernier

Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global Impact
PECB
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
QucHHunhnh
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
kauryashika82
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
ciinovamais
 

Dernier (20)

Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global Impact
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
 
Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docx
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
Role Of Transgenic Animal In Target Validation-1.pptx
Role Of Transgenic Animal In Target Validation-1.pptxRole Of Transgenic Animal In Target Validation-1.pptx
Role Of Transgenic Animal In Target Validation-1.pptx
 
Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SD
 
Web & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdfWeb & Social Media Analytics Previous Year Question Paper.pdf
Web & Social Media Analytics Previous Year Question Paper.pdf
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...
 
ComPTIA Overview | Comptia Security+ Book SY0-701
ComPTIA Overview | Comptia Security+ Book SY0-701ComPTIA Overview | Comptia Security+ Book SY0-701
ComPTIA Overview | Comptia Security+ Book SY0-701
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 

Multivariate Gaussin, Rayleigh & Rician distributions

  • 1. Advanced Communication’s Theory Session 2 H.Amindavar February 2021 Multivariate Normal Distribution Definitions The multivariate normal distribution is often used to describe, at least approximately, any set of (possibly) correlated real-valued random variables each of which clusters around a mean value. The multivariate normal distribution of a k−dimensional random vector X = (X1, . . . , Xk)T can be written in the following notation: X ∼ N (µ, Σ), or to make it explicitly known that X is k−dimensional, Xk ∼ N (µk, σk), with k−dimensional mean vector µ = E[X] = (E[X1], E[X2], . . . , E[Xk])T , and k × k covariance matrix Σi,j := E[(Xi − µi)(Xj − µj)] = Cov[Xi, Xj] such that 1 ≤ i, j ≤ k. The inverse of the covariance matrix is called the precision matrix, denoted by Q = Σ−1 . A real random vector X = (X1, . . . , Xk)T is called a standard normal random vector if all of its components Xn are independent and each is a zero-mean unit-variance normally distributed random vari- able, i.e. if Xn ∼ N (0, 1) for all n. A real random vector X = (X1, . . . , Xk)T is called a centered normal random vector if there exists a deterministic k × ` matrix A such that AZ has the same distribution as X where Z is a standard normal random vector with ` components. A real random vector X = (X1, . . . , Xk)T is called a normal random vector if there exists a random ` -vector Z, which is a standard normal random vector, a k−vector µ, and a k × ` matrix A, such that X = AZ + µ. X ∼ N (µ, Σ) ⇐⇒ there exist µ ∈ Rk , A ∈ Rk×` such that X = AZ + µ for Zn ∼ N (0, 1), i.i.d. 1
  • 2. Figure 1: joint probability density function of bivariate Gaussian with zero mean and [0.25 0.3; 0.3 1] covariance matrix Here the covariance matrix is Σ = AAT . In the degenerate case where the covariance matrix is singular, the corresponding distribution has no density. This case arises frequently in statistics; for example, in the distribution of the vector of residuals in the ordinary least squares regression. The Xi are in general not independent; they can be seen as the result of applying the matrix A to a collection of independent Gaussian variables Z. A random vector X = (X1, . . . , Xk)T has a multivariate normal distribution if it satisfies one of the following equivalent conditions. • Every linear combination Y = a1X1 + · · · + akXk of its components is normally distributed. That is, for any constant vector a ∈ Rk , the random variable Y = aT X has a univariate normal distribution, where a univariate normal distribution with zero variance is a point mass on its mean. • There is a k-vector µ and a symmetric, positive semidefinite k × k matrix ΣΣ, such that the charac- teristic function of X is ϕX(u) = exp iuT µ − 1 2 uT Σu . The spherical normal distribution can be characterised as the unique distribution where components are independent in any orthogonal coordinate system. 2
  • 3. Density function Non-degenerate case The multivariate normal distribution is said to be ”non-degenerate” when the symmetric covariance matrix Σ is positive definite. In this case the distribution has density: fX(x1, . . . , xk) = exp −1 2 (x − µ)T Σ−1 (x − µ) p (2π)k | Σ| where x is a real k−dimensional column vector and |Σ| ≡ det Σ is the determinant of Σ. The equation above reduces to that of the univariate normal distribution if Σ is a 1 × 1 matrix (i.e. a single real number). the locus of points in k−dimensional space each of which gives the same particular value of the density — is an ellipse or its higher-dimensional generalization; hence the multivariate normal is a special case of the elliptical distributions. The quantity q (x − µ)TΣ−1 (x − µ) is known as the Mahalanobis distance, which represents the distance of the test point x from the mean µ. In the case when k = 1, the distribution reduces to a univariate normal distribution and the Mahalanobis distance reduces to the absolute value of the standard score. Bivariate case In the 2-dimensional nonsingular case k = rank (Σ) = 2, the probability density function of a vector [X Y]’ is: f(x, y) = 1 2πσXσY p 1 − ρ2 e − 1 2(1−ρ2) h ( x−µX σX )2 −2ρ( x−µX σX )( y−µY σY )+( y−µY σY )2 i where ρ is the correlation between X and Y and where σX 0 and σY 0. In this case, µ = µX µY , Σ = σ2 X ρσXσY ρσXσY σ2 Y . In the bivariate case, the first equivalent condition for multivariate reconstruction of normality can be made less restrictive as it is sufficient to verify that countably many distinct linear combinations of X and Y are normal in order to conclude that the vector of [X Y]’ is bivariate normal. The bivariate iso-density loci plotted in the x, y-plane are ellipses, whose principal axes are defined by the eigenvectors of the covariance matrix Σ (the major and minor semidiameters of the ellipse equal the square-root of the ordered eigenvalues). -3 -2 -1 0 1 2 3 -3 -2 -1 0 1 2 3 x1 x2 Figure 2: samples of bivariate normal distribution and the contour plot of it’s pdf 3
  • 4. As the absolute value of the correlation parameter ρ increases, these loci are squeezed toward the following line : y(x) = sgn(ρ) σY σX (x − µX) + µY . is the best linear unbiased prediction of Y given a value of X. The equidensity contours of a non-singular multivariate normal distribution are ellipsoids (i.e. linear transformations of hyperspheres) centered at the mean. Hence the multivariate normal distribution is an example of the class of elliptical distributions. The directions of the principal axes of the ellipsoids are given by the eigen-vectors of the covariance matrix Σ. The squared relative lengths of the principal axes are given by the corresponding eigenvalues. If Σ = UΛUT is an eigendecomposition where the columns of U are unit eigenvectors and Λ is a diagonal matrix of the eigenvalues, then we have X ∼ N (µ, Σ) ⇐⇒ X ∼ µ + UΛ1/2 N (0, I) ⇐⇒ X ∼ µ + UN (0, Λ) Moreover, U can be chosen to be a rotation matrix, as inverting an axis does not have any effect on N (0, Λ), but inverting a column changes the sign of Us determinant. The distribution N (µ, Σ) is in effect N (0, I) scaled by Λ, rotated by U and translated by µ. -3 -2 -1 0 1 2 3 -3 -2 -1 0 1 2 3 x2 x1 Figure 3: the iso-density loci of pdf and the eigen-vectors of covariance matrix of the bivariate normal distribution 4
  • 5. Rayleigh Distribution In probability theory and statistics, the Rayleigh distribution is a continuous probability distribution for nonnegative-valued random variables. It is essentially a chi distribution with two degrees of freedom. The probability density function of the Rayleigh distribution is f(x; σ) = x σ2 e−x2 /(2σ2 ) , x ≥ 0, where σ is the scale parameter of the distribution. The cumulative distribution function is F (x; σ) = 1 − e−x2 /(2σ2 ) for x ∈ [0, ∞). 0 0.2 0.4 0.6 0.8 1 1.2 0 2 4 6 8 10 0 0.2 0.4 0.6 0.8 1 1.2 0 2 4 6 8 10 σ=0.5 σ=0.5 σ=0.5 σ=1 σ=1 σ=1 σ=2 σ=2 σ=2 σ=3 σ=3 σ=3 σ=4 σ=4 σ=4 0 0.2 0.4 0.6 0.8 1 1.2 0 2 4 6 8 10 σ=0.5 σ=1 σ=2 σ=3 σ=4 0 0.2 0.4 0.6 0.8 1 1.2 0 2 4 6 8 10 σ=0.5 σ=1 σ=2 σ=3 σ=4 Figure 4: probability density function of Rayleigh distribution Consider the two-dimensional vector Y = (U, V ) which has components that are normally distributed, centered at zero, and independent. Then U and V have density functions fU (x; σ) = fV (x; σ) = e−x2 /(2σ2 ) √ 2πσ2 . Let X be the length of Y That is, X = p U2 + V 2 Then X has cumulative distribution function FX(x; σ) = ZZ Dx fU (u; σ)fV (v; σ) dA where Dx is the disk Dx = n (u, v) : p u2 + v2 x o Writing the double integral in polar coordinates, it becomes FX(x; σ) = 1 2πσ2 Z 2π 0 Z x 0 re−r2 /(2σ2 ) dr dθ = 1 σ2 Z x 0 re−r2 /(2σ2 ) dr. Finally, the probability density function for X is the derivative of its cumulative distribution function, which is fX(x; σ) = d dx FX(x; σ) = x σ2 e−x2 /(2σ2 ) 5
  • 6. which is the Rayleigh distribution. It is straightforward to generalize to vectors of dimension other than 2. There are also generalizations when the components have unequal variance or correlations, or when the vector Y follows a bivariate Student t-distribution. Rice Distribution In probability theory, the Rice distribution or Rician distribution (or, less commonly, Ricean distribution) is the probability distribution of the magnitude of a circularly-symmetric bivariate normal random variable, possibly with non-zero mean (noncentral). It was named after Stephen O. Rice. f(x | ν, σ) = x σ2 exp −(x2 + ν2 ) 2σ2 I0 xν σ2 , Figure 5: probability density function of Rayleigh distribution where I0(z) is the modified Bessel function of the first kind with order zero, Iα(x) = i−α Jα(ix) = ∞ X m=0 1 m! Γ(m + α + 1) x 2 2m+α , Kα(x) = π 2 I−α(x) − Iα(x) sin απ , α is not an integer; when α is an integer, then the limit is used. Iα(x) and Kα(x) are the two linearly independent solutions to the modified Bessel’s equation: x2 d2 y dx2 + x dy dx − x2 + α2 y = 0. R ∼ Rice (|ν|, σ) if R = p X2 + Y 2 where N ν cos θ, σ2 and N ν sin θ, σ2 are statistically in- dependent normal random variables and θ is any real number. 6