SlideShare a Scribd company logo
1 of 39
Download to read offline
Chalmers & TELECOM Bretagne
Coding for phase noise channels

IVAN LELLOUCH
Department of Signals & Systems
Chalmers University of Technology
Gothenburg, Sweden 2011
Master’s Thesis 2011:1
Abstract
Acknowledgements
Contents



1 Introduction                                                                                                                        1

2 Bounds and capacity                                                                                                                 3
  2.1 Capacity and information density . . .          .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   3
  2.2 Binary hypothesis testing . . . . . . .         .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   4
  2.3 Bounds . . . . . . . . . . . . . . . . .        .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   4
      2.3.1 Dependence Testing bound [5]              .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   5
      2.3.2 Meta converse bound [5] . . . .           .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   7

3 AWGN channel                                                                                                                       9
  3.1 The AWGN channel . . . . . . .     .    .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   . 9
      3.1.1 Information density . . .    .    .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   . 10
      3.1.2 Depending testing bound      .    .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   . 10
      3.1.3 Meta converse bound . . .    .    .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   . 10

4 Phase noise channels                                                                                                                13
  4.1 Uniform phase noise channel . . . .         .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   13
      4.1.1 Information density . . . . .         .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   14
      4.1.2 Depending testing bound . .           .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   14
  4.2 Uniform phase noise AWGN channel            .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   17
      4.2.1 Information density . . . . .         .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   17
      4.2.2 Dpending testing bound . . .          .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   19
  4.3 Tikhonov phase noise channel . . . .        .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   20
      4.3.1 Information density . . . . .         .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   21
      4.3.2 Depending testing bound . .           .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   24
      4.3.3 Meta converse bound . . . . .         .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   .   29

5 Conclusion                                                                                                                          30

    Bibliography                                                                                                                      31



                                          i
List of Figures



3.1   DT and converse for AWGN channel - SN R = 0 - Pe = 10−3 . . . . . . . 12

4.1   DT and constraint capacities for three uniform AM constellations. . . .        .   20
4.2   Tikhonov probability density function . . . . . . . . . . . . . . . . . . .    .   22
4.3   Two 64-QAM constellations in AWGN phase noise channel . . . . . . .            .   25
4.4   Robust Circular QAM constellation with phase noise . . . . . . . . . . .       .   26
4.5   DT curves for two 64-QAM constellations in AWGN phase noise channel
      with SNR = 0dB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .   . 26
4.6   DT curves for two 64-QAM constellations in AWGN phase noise channel
      with SNR = 15dB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .    . 27
4.7   Comparaison of DT curves for different phase noise power . . . . . . . .        . 27
4.8   Comparaison of DT curves for different probabilities of error . . . . . .       . 28




                                         ii
1
                             Introduction

Since Shannon’s landmark paper [1], there has been a lot of studies regarding the chan-
nel capacity, which is the amount of information we can reliably sent through a channel.
This result is a theoretical limit, that required to use an infinite block length. In practice,
we want to know, for a given communication system, how far our system is from this
upper bound.
When we design a system, there are two main parameters we need to determine. The
error probability that the system can tolerate and the delay constraint, which is related
to the size of the message we want to send, ie., the block length. Therefore, we want to
find, given these parameters what is the new upper bound for our system. Thus we will
work with finite block length and a given probability of error.
Two bounds are defined in order to determine this new limit, the achievability bound
and the converse bound.
Achievability bound is a lower bound on the size of the codebook, given a block length
and error probability.
Converse bound is an upper bound on the size of the codebook, given a block length
and error probability.
By using both of this bounds, we can determine an approximation of the theoretical limit
of information we can send through a channel, for a given block length and a probability
of error.
Achievability bounds already exist in information theory studies. Three main bounds
were defined by Feinstein [2], Shannon [3] and Gallager [4]. An optimization of auxiliary
constants was needed in order to compute those bounds. Thanks to this work we have
some new insights regarding how far systems can work from the capacity of the channel
with a finite blocklength.
In a recent work [5] Polyanskiy et al defined a new achievability bound that do not
required any auxiliary constant optimization and is tighter than the three bounds in
[2, 3, 4] and one converse bound.


                                              1
CHAPTER 1. INTRODUCTION



This thesis is in the framework of the MAGIC project involving Chalmers University of
Technology, Ericsson AB and Qamcom Technology AB and the context is the microwave
backhauling for IMT advanced and beyond. A part of this project is to investigate the
modulations and coding techniques for channels impaired by phase noise. In digital com-
munications systems, the use of low-cost oscillators at the receiver causes phase noise
which can become a severe problem for high symbol rates and constellation sizes. For
channels impaired by phase noise, the codes we are usually using do not perform as good
as they do in classical channels.

In this thesis, we will deal with two bounds from [5], an achievable bound and a converse
bound. We will apply those bounds to phase noise channels and see how far we are from
the capacity and which rates we can reach given a block length and an error probability.

The outline of the Thesis is as follows.
In Chapter 2, we introduce the capacity and the bounds that will be used in the follow-
ing chapters. The main result is the expression of the depending testing (DT) bound
that we used to find a lower bound on the maximal coding rate. We will explain how
Polyanskiy et al derived this bound and show how we can use it for our channel.
In Chapter 3, we first apply our results to the additive white Gaussian noise (AWGN)
channel. It is useful to see how the equations works for continuous noise channel, but
also to determine the impact the phase noise will cause on the maximal coding rate.
In Chapter 4, we find the main results of the thesis. We apply the DT bound on sev-
eral partially coherent additive white Gaussian noise (PC-AWGN) channels, ie., AWGN
channel impaired by phase noise, and compare them with the bound for the AWGN
channel and the constrained capacity. Therefore, the loss induce by the phase noise will
be estimated, for a given channel, and this will leads to an approximation of the maximal
coding rate for the PC-AWGN we investigated. We also present a constellation designed
for phase noise channel and show the performance improvements.




                                           2
2
                 Bounds and capacity

2.1     Capacity and information density
We denote by A and B the input and output sets. If X and Y are random variables from
A and B respectively then x and y are their particular realizations. X and Y denote
the probability spaces and PX and PY are the probability density functions of X and
Y respectively. We also define PY |X , the conditional probability from A to B given a
codebook (c1 ,...,cM ) and we denote by M its size.
Since we are interested in finite block length analysis, a realisation x of a random variable
X represent a n-dimensional vector, ie., x = (x1 ,...,xn ).
    The capacity C of a channel is the maximum of information we can reliably send
through it with a vanishing error probability and for an infinite block length. For input
and output X and Y distributed according to PX and PY respectively, the capacity C
is given by

                                    C = max {I(X; Y )}                                 (2.1)
                                             X

   where I(X,Y ) is the mutual information between X and Y . The expression is max-
imized with respect to the choice of the input distribution PX .

                                                            pXY (x,y)
                    I(X; Y ) =            pXY (x,y) log2                dxdy
                                   X ,Y                    pX (x)pY (y)
   where the logarithmic term is called the information density :

                                                      pX,Y (x,y)
                                 i(x; y) = log2                                        (2.2)
                                                     pX (x)pY (y)
   It is proven in [6] that the capacity for the AWGN channel is achieved by a Gaussian
input distribution and is given by


                                                 3
CHAPTER 2. BOUNDS AND CAPACITY



                                              1
                                    C=          log2 (1 + SNR)                           (2.3)
                                              2
                                                     P
where SNR is the signal-to-noise ratio, ie., SNR = N , where P and N are the input and
noise power respectively.
    The capacity can be computed when we know which distribution maximized (2.1).
In this case we say that the distribution is capacity achieving. For some channels, such
as phase noise channels, we have few informations regarding the capacity. For those
channels, we will determine an input and use it for our calculations. Thus, we will work
with a constrained capacity, ie., constrained to a specific input distribution, which is an
upper bound of the information that can be sent through the channel for a given input
distribution.
                                                         log2 M
The capacity can also be define by using the rate R = n

                                                    1
                                C = lim lim           log2 M ∗ (n, )                     (2.4)
                                        ←0 n←inf    n
where n is the block length,     the probability of error and M ∗ defined as follow

                         M ∗ (n, ) = max {M : ∃(n,M, ) − code}                           (2.5)


2.2    Binary hypothesis testing
Later in this thesis we will need to use a binary hypothesis testing in order to define an
upper bound for the rate. We consider a random variable R defined as follow

                                        R : {P,Q} → {1,0}                                (2.6)

where 1 indicates that P is chosen. We also consider the random transformation

                                        PZ|R : R → {1,0}                                 (2.7)

We define βα (P,Q), the maximal probability of error under Q if the probability of error
under P is lower than α. We can denote this test by

                 βα (P,Q) =                   inf                       PZ|R (1|r)Q(r)   (2.8)
                               PZ|R :   r∈R   PZ|R (1|r)P (r)≥α
                                                                  r∈R


2.3    Bounds
Yury Polyanskiy defined in [5] the DT bound and the meta converse bound over random
codes . In this chapter we will start by describing those bounds and apply them over
continuous channels.




                                                    4
CHAPTER 2. BOUNDS AND CAPACITY


2.3.1    Dependence Testing bound [5]
We will present the technique proposed in [5] regarding the bounding of the error prob-
ability for any input distribution given a channel.

Theorem 1. Depending testing bound
Given an input distribution PX on A, there exists a code with codebook size M , and the
average probability of error is bounded by
                                                                         +
                                                 M −1
                         ≤ E exp − i(x,y) − log2                                            (2.9)
                                                   2

where
                                     |u|+ = max(0,u)                                       (2.10)

Proof. Let Zx (y) be the following function

                                 Zx (y) = 1(i(x,y)>log M −1 )                              (2.11)
                                                                 2


where 1A (.) is an indicator function :

                                                  1, if x ∈ A,
                                 1A (x) =                                                  (2.12)
                                                  0, otherwise.

    For a given codebook (c1 ,...,cM ), the decoder computes (2.11) for the codeword cj
starting with c1 , until it finds Zcj (y) = 1, or the decoder returns an error. Therefore,
there is no error with probability
                                                            

                           Pr {Zcj (y) = 1}             {Zci (y) = 0}                    (2.13)
                                                   i<j


Then, we can write the error for the j th codeword as
                                                                            

                        (cj ) = Pr {Zcj (y) = 0}               {Zci (y) = 1}             (2.14)
                                                          i<j

Using the union bound on this expression and (2.11)

             (cj ) ≤ Pr {Zcj (y) = 0} +           Pr [{Zci (y) = 1}]                       (2.15)
                                            i<j
                                       M −1                                         M −1
                 = Pr i(cj ,y) ≤ log        +                   Pr i(ci ,y) > log          (2.16)
                                         2                                            2
                                                         i<j




                                                  5
CHAPTER 2. BOUNDS AND CAPACITY


The codebook is generated randomly according to the distribution PX , and we denote by
y a realization of the random variable Y , but independent of the transmitted codeword.
¯
Thus, the probability of error, if we send the codeword cj , is bounded by

                                    M −1                          M −1
          (cj ) ≤ Pr i(x,y) ≤ log        + (j − 1)Pr i(x,¯) > log
                                                         y                                             (2.17)
                                      2                             2
                                       1
Then, if we suppose that Pr(cj ) =     M,   we have

                                                M
                                            1
                                          =           (cj )                                            (2.18)
                                            M
                                                j=1

and
               M
          1                           M −1                          M −1
        ≤           Pr i(x,y) ≤ log        + (j − 1)Pr i(x,¯) > log
                                                           y                                           (2.19)
          M                             2                             2
              j=1

which give us finally the following expression for the average error probability

                                      M −1   M −1                 M −1
                ≤ Pr i(x,y) ≤ log          +      Pr i(x,¯) > log
                                                         y                                             (2.20)
                                        2      2                    2

We know that
                                                              p(x)p(y)
           exp −|i(x,y) − log u|+ = 1(i(x,y)≤log u) + u                1                               (2.21)
                                                               p(x,y) (i(x,y)>log u)

By averaging over p(x,y), and using y , we have
                                    ¯

  exp −|i(x,y) − log u|+ = Pr(i(x,y) ≤ log u) + u                             p(x)p(¯)1(i(x,¯)>log u) (2.22)
                                                                                    y       y
                                                              x       y
                                                                      ¯

and knowing that y is independent of x it leads us to
                 ¯

      exp −|i(x,y) − log u|+ = Pr(i(x,y) ≤ log u) + u                          p(x,¯)1(i(x,¯)>log u)
                                                                                   y       y           (2.23)
                                                                  x       y
                                                                          ¯

and finally

          exp −|i(x,y) − log u|+ = Pr(i(x,y) ≤ log u) + uPr(i(x,¯) > log u)
                                                                y                                      (2.24)
                       M −1
Thus, replacing u =     2     and using (2.20) we obtain (2.9) which completes the proof.

    This expression needs no auxiliary constant optimization and can be computed for
a given channel by knowing the information density. Applications over AWGN channels
and phase noise channels will be shown in following chapters.


                                                6
CHAPTER 2. BOUNDS AND CAPACITY


2.3.2   Meta converse bound [5]
The meta converse bound is an upper bound on the size of the codebook for a given error
probability and block length. To define this bound, we will use the binary hypothesis
testing defined in (2.8).
Theorem 2. Let denote by A and B the input and output alphabets respectively. We
consider two random transformations PY |X and QY |X from A to B, and a code (f,g)
with average probability of error under PY |X and under QY |X .
The probability distribution induced by the encoder is PX = QX . Then we have
                               β1− (PY |X ,QY |X ) ≤ 1 −                          (2.25)
where β is the binary hypothesis testing defined in (2.8).
Proof. We denote by s the input message chosen in (s1 ,...,sM ) and by x = f (s) the
encoded message. Also y is the message before decoding and z = g(z) the decoded
message. We define the following random variable to represent an error-free transmission
                                           Z = 1s=z                               (2.26)
First we notice that the conditional distribution of Z given (X,Y ) is the same for both
channels PY |X and QY |X .
                                            M
                      P [Z = 1|X,Y ] =            P [s = si ,z = si |X,Y ]        (2.27)
                                            i=1
Then, given (X,Y ), since the input and output messages are independent we have
                                      M
                  P [Z = 1|X,Y ] =         P [s = si |X,Y ] P [z = si |X,Y ]      (2.28)
                                     i=1
We can simplify the expression as follows
                                           M
                     P [Z = 1|X,Y ] =           P [s = si |X] P [z = si |Y ]      (2.29)
                                          i=1
We recognize in the second term of the product the decoding function, while the first
term is independent of the choice of the channel given the definition of the probability
distributions induced by the encoder PX = QX that we defined earlier.
Then, using
                                   PZ|XY = QZ|XY                                (2.30)
we define the following binary hypothesis testing
                                  PZ|XY (1|xy)PXY (x,y) = 1 −                     (2.31)
                        x∈A y∈B

                                  PZ|XY (1|xy)QXY (x,y) = 1 −                     (2.32)
                        x∈A y∈B

and using the definition in (2.8) we get (2.25).

                                                 7
CHAPTER 2. BOUNDS AND CAPACITY


Theorem 3. Every code with M codeword in A and an average probability of error
satisfies
                                           1
                      M ≤ sup inf                                         (2.33)
                            PX QY β1− (PXY ,PX × QY )

where PX describes all input distributions on A and QY all output distributions on B.




                                          8
3
                       AWGN channel

In this chapter we will apply the bounds discussed in the previous chapter to the AWGN
channel. We know several results for this channel, such as the capacity. We will see
how far the DT and the meta-converse bounds are from the capacity and if they are
tight enough to have an idea of the maximum achievable rate given a block length and
error probability. The results for the AWGN channel will be useful in the next chapter
to evaluate the effect of the phase noise on the achievable rates for AWGN channels
impaired by phase noise.


3.1    The AWGN channel
Let us consider x ∈ X , y ∈ Y and the transition probability between X and Y , PXY .
We have the following expression for the AWGN channel

                                       y =x+t                                     (3.1)

where the noise samples t ∼ N (0,σIn ) are independent and identically-distributed.
   Thus, we know the conditional output probability function
                                               n/2        (y−x)T (y−x)
                                        1
                          PY |X=x =                  e−       2σ 2                (3.2)
                                      2πσ 2
where .T is the transpose operation. We know from [6] that the Gaussian input distri-
bution achieves the capacity for this channel. So we will consider x ∼ N (0,P In ) and
denote by PX the corresponding probability distribution.
   Given the conditional output probability function and the input, the output distri-
bution is defined as follow (summation of Gaussian variables)

                                 y ∼ N (0,(σ 2 + P )In )                          (3.3)


                                           9
CHAPTER 3. AWGN CHANNEL


3.1.1    Information density
We can now define the information density of the channel using distributions y|x ∼
N (x,σ 2 In ) and y ∼ N (0,(σ 2 + P )In ).

                        log2 (e)  yT y    (y − x)T (y − x)  n     P + σ2
             i(x,y) =                   −                  + log2                      (3.4)
                           2     P + σ2         σ2          2       σ2
which can be rewritten as
                                                         n
                              n      P + σ 2 log2 (e)            2
                                                                yi   n2i
                   i(x,y) =     log2        +                       − 2                (3.5)
                              2        σ2       2             P + σ2 σ
                                                        i=1


3.1.2    Depending testing bound
To compute the DT bound for the AWGN channel, we are using (3.5) in (2.9).
                                             n                                  +
            n     P + σ 2 log2 (e)                   2
                                                    yi    n2i       M −1
   ≤ E exp − log2        +                               − 2 − log2                    (3.6)
            2       σ2       2                    P +σ 2  σ           2
                                            i=1

Then to compute the expectation we can use a Monte Carlo simulation. The samples
are generated according to the model description in section 3.1.

   For this simulation, we use the input distribution x ∼ N (0,P ). This leads to the
DT bound for the maximal coding rate on this channel, ie., this bound will be an upper
bound for all other DT’s for this channel. In practice, discrete constellations are used for
our systems. Therefore we also look at the results for a know discrete input constellation
which will be useful in the next chapter when we will compare results for the AWGN
and the partially coherent AWGN channels.

3.1.3    Meta converse bound
We know that the input distribution and the noise are Gaussian with parameters P and
σ 2 respectively. We also know that the summation of two Gaussian random variable is
a Gaussian random variable, thus we chose y ∼ N (0,σY In ) as the output distribution
for the computation of the converse bound.
    We can now define the information density
                                                  n      2
                            n      2   log2 e           yi             2
                    i(x,y) = log2 σY +
                            2            2               2 − (yi − xi )
                                                        σY
                                                                                       (3.7)
                                                  i=1

   We choose the input such that ||x||2 = nP . To simplify calculations, we are using
          √       √
x = x0 = ( P ,..., P ). This is possible because of the symmetry of the problem.
   Thus, using Zi ∼ N (0,1), Hn and Gn have the following distributions



                                            10
CHAPTER 3. AWGN CHANNEL



                                                             n                    √
                                 P         1
          Hn = n log2 σY − n       log2 e + log2 e                (1 − σY )Zi2 + 2 P σY Zi
                                                                        2
                                                                                              (3.8)
                                 2         2
                                                           i=1
and
                                                   n                  √
                             P           1
      Gn = n log2 σY + n      2 log2 e + 2 log2 e     (1 − σY )Zi2 + 2 P Zi
                                                            2
                                                                                              (3.9)
                            2σY         2σY       i=1
where Hn and Gn are the information density under PY |X and PY respectively.
                   2
Then, by choosing σY = 1 + P , we have
                                                                  n
                 n                    P                                           2
             Hn = log2 (1 + P ) +           log2 e                     1 − Zi2 + √ Zi        (3.10)
                 2                2(1 + P )                                        P
                                                                 i=1
and
                                                       n
                       n                P                                          1
                Gn =     log2 (1 + P ) − log2 e                  1 + Zi2 − 2 1 +     Zi      (3.11)
                       2                2                                          P
                                                       i=1
Notice that Hn and Gn are non-central        χ2   distributions, thus we have

                                 n                            P log2 e
                        Hn =       (log2 (1 + P ) + log2 e) −           yn                   (3.12)
                                 2                            2(1 + P )
               n
with yn ∼ χ2 ( P ), and
           n
                                 n                            P log2 e
                          Gn =     (log2 (1 + P ) + log2 e) −          yn                    (3.13)
                                 2                               2
                  n
with yn ∼ χ2 (n + P ).
             n
Finally, to compute the bound, we find γn such as
                                      Pr [Hn ≥ γn ] = 1 −                                    (3.14)
which lead to
                                                       1
                                       M≤                                            (3.15)
                                            Pr [Gn ≥ γn ]
    Those expressions can be computed directly using closed-form expressions. For some
channels, when we do not have them, therefore we have to compute the bound with a
Monte Carlo simulation and we will discuss the issue of calculate the second probability
Pr [Gn ≥ γn ], which decreases exponentially to 0, by this method.
    In Fig. 3.1, we plot the results for a real-valued AWGN channel of the rate, in bit per
channel use against the block length n. For this example, we use the capacity achieving
input distribution x ∼ N (0,σ) to compute the DT bound. For the following chapters,
we will use discrete input distribution for the AWGN channel in order to compare with
the results we will find for partially coherent AWGN channels.
    We see that the gap between both curves, the DT and the converse, get smaller when
the block length get larger. This result give a good approximation of the maximal coding
rate for this channel given the error probability. We also know from the definition of the
capacity, that both curves will tend toward it when n grows to infinity.

                                                  11
CHAPTER 3. AWGN CHANNEL




                   0.6



                   0.5



                   0.4
Rate, bit/ch.use




                   0.3



                   0.2



                   0.1                                                       Meta−converse
                                                                             DT
                                                                             Capacity
                    0
                         0   200   400   600   800   1000     1200   1400   1600   1800   2000
                                                 Blocklength, n


                    Figure 3.1: DT and converse for AWGN channel - SN R = 0 - Pe = 10−3




                                                     12
4
                 Phase noise channels

In this chapter we will focus on channels impaired by phase noise. First, we will see
some results with a uniform phase noise and the equations that lead to the DT bound.
Then we will focus on two more realistic channels, the AWGN channel impaired with
uniform phase noise, and with Tikhonov phase noise.


4.1    Uniform phase noise channel
We consider a uniform phase noise channel where θ is the additive noise on the phase.
This noise is distributed uniformly between −a and a, ie., θ ∼ U (−a,a). If x ∈ X and
y ∈ Y we have the following expressions

                                        y = xeiθ                                    (4.1)
and
                                      yk = xk eiθk                                  (4.2)
   Using an interleaver, we can assume that the noise is memoryless, then we have the
conditional output distribution
                                            n
                                 p(y|x) =         p(yk |xk )                        (4.3)
                                            k=1

    Notice that regardless the phase noise, the noise cannot change the magnitude of the
sample, ie., |x| = |y|. Then, for this channel we will only consider a constellation with
all points belonging to the same ring.




                                            13
CHAPTER 4. PHASE NOISE CHANNELS


4.1.1   Information density
The information density is defined by

                                                     PY |X=x (y)
                               i(x,y) = log2                                       (4.4)
                                                       PY (y)

   We know that p(yk ) = p(θk ), thus we have

                                       1 , if |θ | ≤ a
                                      
                                                 k
                             p(θk ) = 2a                                           (4.5)
                                        0, otherwise
                                      

   Using (4.3) and (4.5) we obtain the expression of the conditional output distribution

                                                  1 , if y ∈ Yx
                                                 
                                  n
                   PY |X=x (y) =     p(yi |xi ) = (2a)n                            (4.6)
                                                   0, otherwise
                                                 
                                 i=1

where
                                                              yi
                        Yx =    y ∈ Y, ∀i ∈ [1,n] ,| arg         |≤a               (4.7)
                                                              xi
   Then, using the law of total probability, we obtain the output distribution

                                      1                       |Xy |
                          PY (y) =              p(y|x) =                           (4.8)
                                     |X |   x
                                                            |X |(2a)n
where
                                                               yi
                        Xy =   x ∈ X , ∀i ∈ [1,N ] ,| arg         |≤a              (4.9)
                                                               xi
    Finally, the information density for the uniform phase noise channel is given by the
following expression
                                     
                                      log |X | , if y ∈ Yx
                                     
                                           2
                            i(x,y) =         |Xy |                                (4.10)
                                     
                                      0, otherwise


4.1.2   Depending testing bound
Since the capacity achieving distribution for this channel is not known, we will work
with a given input constellation.
   Let the input alphabet be distributed according to an m-PSK modulation. Let n
be the block length, M the size of the codebook M and E(M ) the set of all M -size
codebooks.
The codebook is randomly chosen.



                                                14
CHAPTER 4. PHASE NOISE CHANNELS


   Given a probability of error , we want to find the highest M such as the following
expression stands
                                                   M −1 +
                               ≤ E e−(i(x,y)−log2 ( 2 ))                       (4.11)

   Since we are using discrete input, we can rewrite the expression as follows by ex-
panding the expectation over PXY
                                                                       M −1         +
                          ≤                       p(x,y)e(i(x,y)−log2 ( 2 )) dy                           (4.12)
                               x∈X       y∈Y


   Let z(x,y) = |Xy |, we obtain
                                                                             +
                                                                     2M
                                                          − log2
                                         ≤         P (z)e          (M −1)z
                                                                                                          (4.13)
                                             z∈N

   Then, the probability P (z) can be expanded as follows

                          P (z) =                        P (z|x,y)p(y|x)p(x)dy                            (4.14)
                                         x∈X       y∈Y

which in (4.13) gives
                                                                                                      +
                                                                                             2M
                                                                                 − log2
                ≤                            P (z|x,y)p(y|x)p(x)dy           e             (M −1)z
                                                                                                          (4.15)
                    z∈N   x∈X       y∈Y


   Since the input is an m-PSK modulation and the phase noise is uniform we know the
expressions of p(y|x) and p(x)
                                                                                                      +
                                                            1    1     −            log2     2M
                ≤                            P (z|x,y)        n mn
                                                                   dy e                    (M −1)z
                                                                                                          (4.16)
                                    y∈Yx                  (2a)
                    z∈N   x∈X

   Then, we can simplify the equation using the symmetry of the problem, by choosing
x0 a realisation of X
                                                                                                  +
                                                              1       −          log2     2M
                    ≤                        P (z|x0 ,y)          dy e                  (M −1)z
                                                                                                          (4.17)
                                   y∈Yx0                    (2a)n
                        z∈N

and by expanding the integration over y we obtain
                              a              a                                                        +
              1                                                                   − log2      2M
          ≤                        ···           P (z|x0 ,y)dy1 · · · dyn e                 (M −1)z
                                                                                                          (4.18)
            (2a)n             −a         −a
                    z∈N

   Let V (x,y) be the number of neighbours in X for y ∈ Yx . V = V (x,y) − 1.
Then the probability P (z|x,y) can be written as follows




                                                         15
CHAPTER 4. PHASE NOISE CHANNELS




                           V −1                       z−1                 V −z−1
                      V                1
     P (z|x,y) =                                            (M − 1 − j)            (mn − M − j)              (4.19)
                      z           (mn − 1 − j)
                            j=0                       j=0                  j=0

P (z|x,y) = 0 if z > V (x,y).

   Given the phase noise parameter a, and the number of points m in one ring of the
constellation, we can determine the function v(yk ).
v(yk ) define the number of points the output yk can come from, in one ring of the
constellation. This function is a simple function with two values v1 and v2 as soon as
the points in each ring are equally spaced.
Then we can define two constant d1 and d2 by the following expressions
                                               yk +a
                                       d1 =            1(v(yk ) = v1 )dyk                                    (4.20)
                                              yk −a
                                               yk +a
                                       d2 =            1(v(yk ) = v2 )dyk                                    (4.21)
                                              yk −a

   Finally we have the following expression to compute the bound

               max(v1 n ,v2 n )    n                                                                          +
         1                               n                                            − log2       2M
     ≤                                     d1 u d2 n−u p z|V = v1 u v2 n−u           e         (M −1)(z+1)
       (2a)n        z=0           u=0
                                         u
                                                                                   (4.22)
    The complexity of this calculation depends on the complexity of p (z|V = v1 u v2 n−u ).
This is a product of V terms, so the complexity is O(2n ). Of course, we can’t compute
this expression because of its complexity. In the next chapters, we are using partially
coherent AWGN channels, and the expressions do not need to calculate the probability
p(z|x,y) which make the computation much faster.




                                                        16
CHAPTER 4. PHASE NOISE CHANNELS


4.2      Uniform phase noise AWGN channel
We consider an AWGN channel impaired with a uniform phase noise θ ∼ U (−a,a). If
                            2
x ∈ X , y ∈ Y and t ∼ N (0,σN ) we have

                                            y = xeiθ + t                                  (4.23)
and
                                          yk = xk eiθk + tk                               (4.24)
   For this channel we can define the information density as follows.

4.2.1     Information density
We need both expressions of PY and PY |X to determine the expression of i(x,y). First,
we know that the noise is memoryless, which allows us to write
                                                        n
                                  PY |X=x (y) =             p(yk |xk )                    (4.25)
                                                      k=1

   where xk , yk and tk can be written as

                                            xk = ak eibk
                                            tk = ck eidk
                                            yk = αk eiβk

which give us the following expression for the conditional output distribution (using
polar coordinates)


                           p(yk |xk ) =        αk p(θk |xk )p(tk |θk ,xk )dθk             (4.26)
                                          θk
                                           a
                                               αk 1          |tk |2
                                    =                   exp − 2 dθk                       (4.27)
                                          −a   2a 2πσ 2      2σ

where
                                     |tk |2 = |yk − xk eiθk |2                            (4.28)
   We develop (4.28) as follows

 |tk |2 = |αk cos(βk ) + iαk sin(βk ) − |xk | cos(θk + arg(xk )) + i|xk | sin(θk + arg(xk ))|2
                                                                                           (4.29)
        = (αk cos(βk ) − |xk | cos(θk ) + arg(xk ))2 + (αk sin(βk ) − |xk | sin(θk + arg(xk )))2
                                                                                           (4.30)
        = α2 + a2 − 2αk ak cos(θk + bk − βk )
           k    k                                                                         (4.31)


                                                   17
CHAPTER 4. PHASE NOISE CHANNELS


which used in (4.26) gives
                                         (a2 +α2 )           a
                            exp(− k 2 k )
                                    2σ                                  αk ak cos(θk + bk − βk )
            p(yk |xk ) = αk                                      exp                             dθk              (4.32)
                              (2a)(2πσ 2 )                 −a                      σ2
   Then, using the law of total probabilities we obtain the expression of the output
density probability
                                                   m−1
                                   p(yk ) =                p(xu,k )p(yk |xu,k )                                   (4.33)
                                                   u=0
    Now, we need to choose the input distribution to determine the expression of the
information density. We will consider a set of M codewords (c1 ,...,cM ) with the same
probability. Then we have

                  m−1             (a2 +α2 )            a
                        αk exp(− u,k 2 k )
                                  2σ                                 αk au,k cos(θk + bu,k − βk )
      p(yk ) =                                             exp                                    dθk             (4.34)
                        m (2a)(2πσ 2 )               −a                           σ2
                  u=0

which finally gives the following expression for the information density

                                            (a2 +α2 )
                                                                                                             
                                     exp(− k 2 k )            a      αk ak cos(θk +bk −βk )
                  N
                                                              −a exp
                                             2σ
                                       (2a)(2πσ2 )                             σ2           dθk
 i(x,αeiβ ) =
                                                                                                             
                       log2 
                                              (a2 +α2 )
                                                                                                               (4.35)
                                                                                                              
                k=1              m−1 1   exp(− u,k 2 k )              a      αk au,k cos(θk +bu,k −βk )
                                                                      −a exp
                                                  2σ
                                 u=0 m      (2a)(2πσ2 )                                  σ2
                                                                                                        dθk

and with some simplifications we obtain

                                                                                                        
                                            a2     a
                   N
                                     exp(− 2σ2 ) −a exp αk ak cos(θk +bk −βk ) dθk
                                             k
                                                                  σ2
   i(x,αeiβ ) =         log2                  a2
                                                                                                                 (4.36)
                                  m−1 1         u,k  a      αk au,k cos(θk +bu,k −βk )
                  k=1             u=0 m exp(− 2σ2 ) −a exp              σ2
                                                                                       dθk
   We recognize in the information density expression the following integral
                                               s
                                                   ek cos(x) dx, s ≤ π.                                           (4.37)
                                           0

   We can find a closed-form expression only if we choose s = π, by using the Bessel
function of the first kind. For this case (4.32) becomes,

                                     (a2 +α2 )                   π
                            exp(− k 2 k ) 1
                                   2σ                                      αk ak cos(θk + bk − βk )
    p(αk ,βk |ak ,bk ) = αk                                          exp                            dθk           (4.38)
                               (2πσ 2 )  2π                  −π                       σ2
and using the properties of trigonometric functions we can rewrite the expression as
follows
                                        (a2 +α2 )     π
                                   exp(− k 2 k ) 1
                                          2σ                i2 αk ak sin(θk )
           p(αk ,βk |ak ,bk ) = αk                      exp                   dθk (4.39)
                                      (2πσ 2 )    2π −π            σ2

                                                            18
CHAPTER 4. PHASE NOISE CHANNELS


   We notice that the expression is independent of βk , the angle of y
                                                      π
                            p(αk |ak ,bk ) =              p(αk ,βk |ak ,bk )dβk                                 (4.40)
                                                   −π
                                              = (2π)p(αk ,βk |ak ,bk )                                          (4.41)

which leads to
                                        αk      a2 + α2                            αk ak
                         p(αk |ak ) =      exp − k 2 k                     I0                                   (4.42)
                                        σ2        2σ                                σ2
using I0 (.) the Bessel function of the first kind.

M-PSK
                                              √
Considering a M-PSK, we have ak =                 P and PY can be defined as follows
                                        m
                          PY (αk ) =          p(au,k )p(αk |ak ) = p(αk |ak )                                   (4.43)
                                        u=1

which leads to i(x,y) = 0. Given the fact that we have a non-coherent AWGN chan-
nel with an m-PSK modulation, we easily understand that no information can be sent
through the channel.

Amplitude modulation
Now, we consider an amplitude modulation. If we have R points in our constellation,
           √
and if ar = Pr then we have
                                                                          a2
                                                                                                  
                                                                                         αk ak
                                    N                       exp − 2σ2
                                                                   k
                                                                                   I0     σ2
              i(x,y) = i(a,α) =          log2                                               √                 (4.44)
                                                          R                 P              αk Pr
                                   k=1                    r=1 exp         − 2σr2    I0       σ2

   Once again, given the channel, there is no information on the phase. So we can work
by using only the magnitude of each point.

4.2.2    Dpending testing bound
Now we want to determine the upper bound for this channel. First, we pick an input
constellation and then we use it in (4.44) to determine the equation of the DT (2.9),
which leads us to the expression

                                                                                                       +
                                                                                                         
                                                 a2
                                                                                  
                                                                  αk ak
                    N               exp       − 2σ2
                                                  k
                                                           I0
   ≤ E exp −
           
                         log2 
                                                                   σ2
                                                                      √             − log2 M − 1           (4.45)
                                                                                                           
                                  R                 P
                                                  − 2σr2            αk Pr                     2
                   k=1            r=1 exp                    I0       σ2


   We use a Monte Carlo simulation to calculate this expression.

                                                      19
CHAPTER 4. PHASE NOISE CHANNELS


Amplitude modulation input
For this constellation, we consider m points equally spaced with average power P = 1.
In Fig. 4.1 we plot the rate, in bit per channel use, against the block length n. We choose
to present 3 constellations, with 8, 16 and 32 points. The Gaussian noise is defined by
SNR = 15dB, and the error probability is Pe = 10−3 .
For each constellation, the depending testing bound and the constrained capacity are
plotted.


                         4.5

                          4

                         3.5

                          3
      Rate, bit/Ch.use




                         2.5

                          2

                         1.5

                          1
                                                     DT and constraint capacity : 8−AM constellation
                         0.5                         DT and constraint capacity : 16−AM constellation
                                                     DT and constraint capacity : 32−AM constellation
                          0
                               0   200   400   600   800    1000 1200       1400    1600    1800    2000
                                                        blocklength, n


                   Figure 4.1: DT and constraint capacities for three uniform AM constellations.


   We see in Fig. 4.1 that, for a given constellation, both curves, the DT bound and
the constrained capacity, are tight when the block length is large. We can also notice
that the gap between both curves decreases faster when we have fewer points in our
constellation.


4.3                      Tikhonov phase noise channel
A more realistic channel to describe a system impaired by a phase noise is the Tikhonov
AWGN channel. We have a closed-form expression for the noise and, using Lapidoth’s
result in [7] we can also have one for the conditional output.


                                                           20
CHAPTER 4. PHASE NOISE CHANNELS


We choose to study this model because it is a good approximation of the phase noise
error induced by a first-order phase-locked loop [? ].
We consider t ∼ N (0,σn ), the Gaussian noise, and θ the phase noise distributed according
to the Tikhonov distribution. This distribution is presented in section 4.3.

                                              y = xeiθ + t                                     (4.46)
and
                                            yk = xk eiθk + tk                                  (4.47)

Tikhonov distribution
The Tikhonov distribution, also known as Von Mises distribution [? ], is an approxima-
tion of the wrapped Gaussian which is defined as follows

                                                       1                        −(θ−2kπ)2
                       pW (θ) =       pΘ (θ + 2kπ) = √                      e      2σ 2        (4.48)
                                  k∈Z
                                                      2πσ 2           k∈Z

   Its support is [−π; π] and is function of a parameter ρ. The probability density
function is given by

                                                       eρ cos(x)
                                            p(x|ρ) =                                           (4.49)
                                                       2πI0 (ρ)
    In Fig. 4.2 we see the Tikhonov distribution for 3 values of the parameter ρ. The
larger the parameter is, the smaller the noise is.

4.3.1    Information density
First, we need to determine the expression for the conditional output distribution. We
know that the noise is memoryless, so we can focus on p(yk |xk ).
                                               π
                              p(yk |xk ) =         pn (yk |xk ,θk )pθ (θk )dθk                 (4.50)
                                              −π

Using both expressions of the Gaussian pdf and the Tikhonov pdf, we have

                                                             2
                        π
                              1         yk − xk ejθk               eρ cos(θk )
        p(yk |xk ) =              exp −                                        dθk             (4.51)
                       −π   2πσ 2           2σ 2                   2πI0 (ρ)
                                                                        2
                         1              π      yk − xk ejθk
                 =     2 σ 2 I (ρ)
                                         exp −                              + ρ cos(θk ) dθk   (4.52)
                   (2π)       0       −π           2σ 2




                                                     21
CHAPTER 4. PHASE NOISE CHANNELS



          9
                                                                                         rho=500
          8                                                                              rho=100
                                                                                         rho=10

          7

          6

          5

          4

          3

          2

          1

          0
              −3             −2           −1             0           1           2             3
                                                     angle (rad)


                           Figure 4.2: Tikhonov probability density function



we can now expand the expression in the exponential
                     2
    yk − xk ejθk         = |yk |2 + |xk |2 − yk xk ejθk − yk x∗ e−jθk
                                              ∗
                                                              k                                      (4.53)
                         = |yk |2 + |xk |2 − yk xk (cos(θk ) + sin(θk )) − yk x∗ (cos(θk ) − sin(θk ))
                                              ∗
                                                                               k
                                                                                                   (4.54)
                         = |yk |2 + |xk |2 − 2 (yk xk ) cos(θk ) + 2 (yk xk ) sin(θk )
                                                 ∗                     ∗
                                                                                                     (4.55)
which gives us the following expression for the conditional output distribution

                     −(|yk |2 +|xk |2 )
               exp         2σ2
                                           π              ∗                        ∗
                                                      ( (yk xk ) + ρ) cos(θk ) − (yk xk ) sin(θk )
p(yk |xk ) =           2 σ 2 I (ρ)
                                               exp                                                    dθk
                   (2π)       0           −π                              σ2
                                                                              (4.56)
   Because of the symmetry of the problem, we choose to work with polar coordinates,
thus we define xk and yk by
                                                xk = ak eibk
                                                yk = αk eiβk


                                                       22
CHAPTER 4. PHASE NOISE CHANNELS


    Then we have

                                              yk xk = aαei(b−β)
                                               ∗
                                                                                                        (4.57)
and


                                          ∗
                                        (yk xk ) = aα cos (i(b − β))                                    (4.58)
                                          ∗
                                        (yk xk )   = aα sin (i(b − β))                                  (4.59)

    Using both equations we define
                                                           aα
                                                    u=                                                  (4.60)
                                                           σ2
and
                                   A = ( (yk xk ) + ρ)2 + ( (yk xk ))2
                                           ∗                  ∗
                                                                                                        (4.61)
which leads to

                      −(a2 +α2 )
               α exp       2σ2
                                        π          √                    u cos(b − β)            u sin(b − β)
p(yk |xk ) =         2 σ 2 I (ρ)
                                            exp        A cos(θk )           √        − sin(θk )      √          dθk
                 (2π)       0         −π                                      A                        A
                                                                                                         (4.62)
    We defined A such that

                                A = (u cos(b − β))2 + (u sin(b − β))2                                   (4.63)

so we can find z such that
                                                     u cos(b − β)
                                            cos(z) =      √                                             (4.64)
                                                            A
                                                     u sin(b − β)
                                            sin(z) =      √                                             (4.65)
                                                            A
    Then we have
                                        −(a2 +α2 )
                                α exp        2σ2
                                                              π         √
                 p(yk |xk ) =          2 σ 2 I (ρ)
                                                                  exp        A (cos(θk + z)) dθk        (4.66)
                                   (2π)       0           −π

which is equal to
                                           −(a2 +α2 )
                                    α exp       2σ2
                                                                  π          √
                    p(yk |xk ) =                                       exp       A cos(θk ) dθk         (4.67)
                                      (2π)2 σ 2 I0 (ρ)            −π

given that z and θk are independent.
    Finally, we recognize in (4.67) the Bessel function of the first kind which gives the
following expression for the conditional output distribution

                                                         23
CHAPTER 4. PHASE NOISE CHANNELS


                                                                          √
                                         α          α2 + a2           I0 ( A)
                      pY |X (yk ,xk ) =       exp −                                    (4.68)
                                        2πσ 2         2σ 2             I0 (ρ)
    To find an expression for the information density, we also need the output distribution
PY . In this thesis, we consider a discrete input constellation with M codewords (c1 ,...,cM )
               1
and P (ci ) = M .
    Given this input, the output distribution can be computed as follows
                                             M
                                                   1
                                PY (yk ) =           p    (yk |ci )                    (4.69)
                                                   M Y |X
                                             i=1

and the information density is

                                                               a2
                                                                      √
                                                        exp − 2σ2 I0 ( A)
                i(x,y) = i(aeib ,αeiβ ) = log2                     a2     √            (4.70)
                                                      M 1
                                                      i=1 M exp − 2σ2 I0 ( Ai )
                                                                    i



where

                                  a2 α2     aα
                             A=      4
                                        + 2ρ 2 cos(b − β) + ρ2                         (4.71)
                                   σ        σ

4.3.2    Depending testing bound
We compare two constellations for the AWGN channel with Tikhonov phase noise. We
have the classic 64-QAM constellation and a robust circular QAM constellation designed
specifically for this channel [8].
The second constellation is designed in order to minimize the average minimum distance
between two points of the constellation. The algorithm presented in [8] gives an example
of the constellation for a given phase noise.
In Fig. 4.3 we plot both constellations and what happen to them through an AWGN
channel with SNR = 30dB impaired by the given phase noise ρ = 625 (σph = 0.04).

    In Fig. 4.4 we plot the robust circular 64-QAM constellation impaired by a Tikhonov
phase noise with parameter ρ = 625.
    In Fig. 4.5 we plot the DT curve and the constrained capacity for both constellations.
We choose SNR = 0dB, ρ = 625 and Pe = 10−3 for this simulation.
    In Fig. 4.6 we plot the DT bound and the constrained capacity for both constella-
tions. We choose SNR = 15dB, ρ = 625 and Pe = 10−3 for this simulation.
    In Fig. 4.7 we plot the DT bound for the robust circular 64-QAM constellation for
two power of phase noise. We also plot the DT bound and the constrained capacity
without phase noise, ie., Gaussian noise only, and both constrained and unconstrained
capacity for this channel. We choose SNR = 15dB and Pe = 10−3 for this simulation.




                                                 24
CHAPTER 4. PHASE NOISE CHANNELS


                   64−QAM constellation                         Robust Circular QAM constellation
         2                                                 2

         1                                                 1

         0                                                 0

        −1                                                −1

        −2                                                −2
         −2        −1       0        1          2          −2         −1       0        1           2

              64−QAM constellation with noise            Robust Circular QAM constellation with noise
         2                                                 2

         1                                                 1

         0                                                 0

        −1                                                −1

        −2                                                −2
         −2        −1       0        1          2          −2         −1       0        1           2




          Figure 4.3: Two 64-QAM constellations in AWGN phase noise channel



   In Fig. 4.8 we plot the DT bound for the robust circular 64-QAM constellation for
two probabilities of error. We also plot the constrained capacity for this channel. We
choose SNR = 0dB and ρ = 100 for this simulation.




                                                    25
CHAPTER 4. PHASE NOISE CHANNELS




                      1.5



                       1



                      0.5



                       0



                 −0.5



                      −1



                 −1.5
                   −1.5                  −1          −0.5           0             0.5        1          1.5




                                Figure 4.4: Robust Circular QAM constellation with phase noise




                       1

                      0.9

                      0.8

                      0.7
   Rate, bit/Ch.use




                      0.6

                      0.5

                      0.4

                      0.3

                      0.2
                                                                          DT for Robust Circular 64−QAM
                                                                          DT for regular 64−QAM
                      0.1
                                                                          Constrained−capacity for 64−QAM

                       0
                            0      200        400   600     800   1000     1200    1400   1600   1800   2000
                                                              Blocklength, n


Figure 4.5: DT curves for two 64-QAM constellations in AWGN phase noise channel with
SNR = 0dB




                                                                  26
CHAPTER 4. PHASE NOISE CHANNELS




                       5

                      4.5

                       4

                      3.5
   Rate, bit/Ch.use




                       3

                      2.5

                       2

                      1.5

                                                      DT for Robust Circular 64−QAM
                       1
                                                      DT for regular 64−QAM
                      0.5                             Constrained−capacity for Robust Circular 64−QAM
                                                      Constrained−capacity for Regular 64−QAM
                       0
                            0    200    400    600    800   1000     1200    1400   1600    1800    2000
                                                        Blocklength, n


Figure 4.6: DT curves for two 64-QAM constellations in AWGN phase noise channel with
SNR = 15dB




                      5.5

                       5

                      4.5

                       4

                      3.5
   Rate bit/Ch.use




                       3

                      2.5

                       2

                      1.5
                                                      DT for rho=100
                       1                              DT for rho=625
                                                      DT and constrained−capacity without phase noise
                      0.5                             Capacity
                       0
                            0    200    400    600    800   1000     1200    1400   1600    1800    2000
                                                        Blocklength, n


                            Figure 4.7: Comparaison of DT curves for different phase noise power




                                                            27
CHAPTER 4. PHASE NOISE CHANNELS



                         1

                        0.9

                        0.8

                        0.7
     Rate, bit/Ch.use



                        0.6

                        0.5

                        0.4

                        0.3

                        0.2                                                  DT with Pe=10E−3
                                                                             DT with Pe=10E−9
                        0.1
                                                                             Constrained−capacity
                         0
                              0   100   200   300    400    500     600    700    800     900       1000
                                                       Blocklength, n


                         Figure 4.8: Comparaison of DT curves for different probabilities of error



     In Fig. 4.3 we see both constellations we used in our simulations. The robust circular
64-QAM has been designed for phase noise channels with a noise power ρ = 625. The
optimization criteria used is the maximization of the minimal distance between two
adjacent rings.
In Fig. 4.5 we notice that both DT curves are the same for both constellations. Despite
the differences between these constellations, the power of the Gaussian noise is too high
to make a difference between these constellations. We can also notice that, even for a
large block length (n = 2000), their is still a gap between the DT and the constrained
capacity.
In Fig. 4.6 we notice a difference between both constellations. For large block length
(n ≤ 100) the robust circular 64-QAM performs better than the regular 64-QAM. We
also notice that the DT bound and the constrained capacity are tight, which gives us a
better approximation of the maximal coding rate for this channel.
From these curves, we can also see that for high SNR we reach the capacity much faster
than for small SNR. Then the gap between the DT bound and the constrained capacity
is tighter for large SNR.
In Fig. 4.7 we see the impact of the phase noise power on the DT bound. We present
the loss of coding rate between two channels with different phase noise power. We also
see that with the parameter ρ = 625, the maximal coding rate is very close to the coding
rate without any phase noise. We also notice the loss induce by our constellation in
regard to the capacity achieving distribution.
In Fig. 4.8 we see the impact of the probability of error on the coding rate. We notice


                                                           28
CHAPTER 4. PHASE NOISE CHANNELS


that the difference between these curves exist for small block length, and we will need a
larger block length to have a smaller probability of error.

4.3.3   Meta converse bound
As we define earlier, we know the conditional output for our channel

                                               R          R2 + r 2              I0 (ν)
                       pY |X (r,φ,R,ψ) =            exp −                                  (4.72)
                                              2πσ 2        2σ 2                 I0 (ρ)
where
                                 R2 r 2     Rr
                             ν=         + 2ρ 2 cos(φ − ψ) + ρ2
                                  σ4        σ
   The meta converse bound requires to pick an output distribution. For our case, we
use the following distribution, which is capacity achieving for high SNR [7].

                                                  R 2 ∼ χ1
                                                         2
                                                                                           (4.73)
                                            ψ ∼ U (−π,π)                                   (4.74)
                                                     1                     R2
                          PY (R,ψ) =             √
                                                 exp −                                     (4.75)
                                  2π 2R2 Γ 1 2
                                                         2
   Thus, we can define the information density given those two distributions
                                             N
                                                          pY |X=xi (yi )
                               i(x,y) =           log2                                     (4.76)
                                                             PY (yi )
                                            i=1
                                                                                      
                                                   Ri           R2 +r 2     I0 (νi )
                                  N
                                                  2πσ2   exp   − 2σ2 i
                                                                 i
                                                                            I0 (ρ)
                       i(x,y) =         log2                                              (4.77)
                                                                                      
                                                                           R2
                                                                                       
                                  i=1                √ 1 1         exp    − 2i
                                                   2π 2R2 Γ( 2 )
                                                        i

                         √       1          N
                           2Γ 2                                             2
                                                                           Ri + ri2  R2
     i(x,y) = N log2                    +         log2 I0 (νi ) exp −               − i    (4.78)
                         σ 2 I0 (ρ)                                          2σ 2     2
                                            i=1
    Then, we denote by Gn and Hn the information density under PY and PY |X respec-
tively.
    To compute the converse bound, we have to find the parameter γn given the following
condition
                                P [Hn ≥ γn ] = 1 −                               (4.79)
and then, we use this parameter to determine the following probability
                                             P [Gn ≥ γn ]                                  (4.80)
   The main issue for this bound is the calculation of the probability P [Gn ≥ γn ].
We know that this value decreases exponentially to 0 and we do not have any closed-
form expression to compute it. In the real-value Gaussian case, we found a closed-form
expression using the chi-square distribution.

                                                     29
5
                              Conclusion

In this work we applied an achievability bound on phase noise channels in order to
determine the maximum coding rate for such channel.
First, we focused on a simple model with a uniform phase noise. We managed to find
a closed-form expression for the DT bound, but the computation complexity was an
issue. Then we moved on to two partially coherent AWGN channels. For the AWGN
channel impaired by uniform phase noise, a closed form expression has be found for
the non coherent case which gave some results. And finally, we found some results for
the AWGN channel impaired by a Tikhonov phase noise. We investigated the impact
of all parameters (Noises power and probability of error) on the curves. Through both
applications on phase noise channels we can see that the DT bound and the constraint
capacity associated with a constellation are really close for high SNR. This give us a
good idea of the achievable rate for a given block length and an error probability.
We also investigated the impact of different power of phase noise and the loss induced on
this rate. Moreover, we can see on the curves that for large block length (n > 500) and
high SNR (SNR> 15), more than 95% of the constrained capacity is already achieved.
Given those informations, we can evaluate the performances of codes and discuss the
interest of using larger block.
For small SNR, the gap between the achievability bound and the constrained capacity is
still large, therefore, we do not have a tight approximation of the maximal coding rate.
As a future work for our thesis, we can study the meta-converse and try to find an
approximation for the binary hypothesis testing. In that way we could compute the
upper bound and have a tighter approximation.
Another following of this thesis could be to investigate all the codes we already have and
see their performances over PC-AWGN channels.




                                           30
Bibliography


[1] C. E. Shannon, A mathematical theory of communication, Bell System Technical
    Journal (1948) 379–423.

[2] A. Feinstein, A new basic theorem of information theory, IRE trans. Inform. Theory
    (1954) pp. 2–22.

[3] C. E. Shannon, Certain results in coding theory for noisy channels, Inf. Contr., vol.
    1 (1957) pp. 6–25.

[4] R. G. Gallager, A simple derivation of the coding theorem and some applications,
    IEEE Trans. Inf. Theory, vol. 40 (1965) pp.3–18.

[5] Y. Polyanskiy, H. V. Poor, S. Verdu, Channel coding rate in the finite blocklength
    regime, IEEE Trans. Inf. Theory.

[6] T. Cover, J. Thomas, Elements of Information Theory, Wiley, 2006.

[7] A. Lapidoth, On phase noise channels at high snr, IEEE Trans. Inf. Theory.

[8] A. Papadopoulos, K. N. Pappi, G. K. Karagiannidis, H. Mehrpouyan, Robust circular
    qam constellations in the presence of phase noise.




                                           31

More Related Content

What's hot

Blind, Non-stationary Source Separation Using Variational Mode Decomposition ...
Blind, Non-stationary Source Separation Using Variational Mode Decomposition ...Blind, Non-stationary Source Separation Using Variational Mode Decomposition ...
Blind, Non-stationary Source Separation Using Variational Mode Decomposition ...CSCJournals
 
SIR Analysis of Overloaded CDMA System Using Orthogonal Gold Codes
SIR Analysis of Overloaded CDMA System Using Orthogonal Gold CodesSIR Analysis of Overloaded CDMA System Using Orthogonal Gold Codes
SIR Analysis of Overloaded CDMA System Using Orthogonal Gold CodesIDES Editor
 
Multihop Routing In Camera Sensor Networks
Multihop Routing In Camera Sensor NetworksMultihop Routing In Camera Sensor Networks
Multihop Routing In Camera Sensor NetworksChuka Okoye
 
Performance of MMSE Denoise Signal Using LS-MMSE Technique
Performance of MMSE Denoise Signal Using LS-MMSE  TechniquePerformance of MMSE Denoise Signal Using LS-MMSE  Technique
Performance of MMSE Denoise Signal Using LS-MMSE TechniqueIJMER
 
Thesis multiuser + zf sic pic thesis mincheol park
Thesis multiuser + zf sic pic thesis mincheol parkThesis multiuser + zf sic pic thesis mincheol park
Thesis multiuser + zf sic pic thesis mincheol parknhusang26
 
phd Thesis Mengdi Zheng (Summer) Brown Applied Maths
phd Thesis Mengdi Zheng (Summer) Brown Applied Mathsphd Thesis Mengdi Zheng (Summer) Brown Applied Maths
phd Thesis Mengdi Zheng (Summer) Brown Applied MathsZheng Mengdi
 
EE443 - Communications 1 - Lab 3 - Loren Schwappach.pdf
EE443 - Communications 1 - Lab 3 - Loren Schwappach.pdfEE443 - Communications 1 - Lab 3 - Loren Schwappach.pdf
EE443 - Communications 1 - Lab 3 - Loren Schwappach.pdfLoren Schwappach
 
Tree Based Collaboration For Target Tracking
Tree Based Collaboration For Target TrackingTree Based Collaboration For Target Tracking
Tree Based Collaboration For Target TrackingChuka Okoye
 
VECTOR VS PIECEWISE-LINEAR FITTING FOR SIGNAL AND POWER INTEGRITY SIMULATION
VECTOR VS PIECEWISE-LINEAR FITTING FOR SIGNAL AND POWER INTEGRITY SIMULATIONVECTOR VS PIECEWISE-LINEAR FITTING FOR SIGNAL AND POWER INTEGRITY SIMULATION
VECTOR VS PIECEWISE-LINEAR FITTING FOR SIGNAL AND POWER INTEGRITY SIMULATIONPiero Belforte
 
Multiuser detection new
Multiuser detection newMultiuser detection new
Multiuser detection newNebiye Slmn
 
104623 time division multiplexing (transmitter, receiver,commutator)
104623 time  division multiplexing (transmitter, receiver,commutator)104623 time  division multiplexing (transmitter, receiver,commutator)
104623 time division multiplexing (transmitter, receiver,commutator)Devyani Gera
 

What's hot (19)

Thesis
ThesisThesis
Thesis
 
Dissertation A. Sklavos
Dissertation A. SklavosDissertation A. Sklavos
Dissertation A. Sklavos
 
diplom_master
diplom_masterdiplom_master
diplom_master
 
Blind, Non-stationary Source Separation Using Variational Mode Decomposition ...
Blind, Non-stationary Source Separation Using Variational Mode Decomposition ...Blind, Non-stationary Source Separation Using Variational Mode Decomposition ...
Blind, Non-stationary Source Separation Using Variational Mode Decomposition ...
 
SIR Analysis of Overloaded CDMA System Using Orthogonal Gold Codes
SIR Analysis of Overloaded CDMA System Using Orthogonal Gold CodesSIR Analysis of Overloaded CDMA System Using Orthogonal Gold Codes
SIR Analysis of Overloaded CDMA System Using Orthogonal Gold Codes
 
Dr35672675
Dr35672675Dr35672675
Dr35672675
 
wavelet packets
wavelet packetswavelet packets
wavelet packets
 
Ch12
Ch12Ch12
Ch12
 
Multihop Routing In Camera Sensor Networks
Multihop Routing In Camera Sensor NetworksMultihop Routing In Camera Sensor Networks
Multihop Routing In Camera Sensor Networks
 
Performance of MMSE Denoise Signal Using LS-MMSE Technique
Performance of MMSE Denoise Signal Using LS-MMSE  TechniquePerformance of MMSE Denoise Signal Using LS-MMSE  Technique
Performance of MMSE Denoise Signal Using LS-MMSE Technique
 
1570265619
15702656191570265619
1570265619
 
Thesis multiuser + zf sic pic thesis mincheol park
Thesis multiuser + zf sic pic thesis mincheol parkThesis multiuser + zf sic pic thesis mincheol park
Thesis multiuser + zf sic pic thesis mincheol park
 
phd Thesis Mengdi Zheng (Summer) Brown Applied Maths
phd Thesis Mengdi Zheng (Summer) Brown Applied Mathsphd Thesis Mengdi Zheng (Summer) Brown Applied Maths
phd Thesis Mengdi Zheng (Summer) Brown Applied Maths
 
Bojja_thesis_2015
Bojja_thesis_2015Bojja_thesis_2015
Bojja_thesis_2015
 
EE443 - Communications 1 - Lab 3 - Loren Schwappach.pdf
EE443 - Communications 1 - Lab 3 - Loren Schwappach.pdfEE443 - Communications 1 - Lab 3 - Loren Schwappach.pdf
EE443 - Communications 1 - Lab 3 - Loren Schwappach.pdf
 
Tree Based Collaboration For Target Tracking
Tree Based Collaboration For Target TrackingTree Based Collaboration For Target Tracking
Tree Based Collaboration For Target Tracking
 
VECTOR VS PIECEWISE-LINEAR FITTING FOR SIGNAL AND POWER INTEGRITY SIMULATION
VECTOR VS PIECEWISE-LINEAR FITTING FOR SIGNAL AND POWER INTEGRITY SIMULATIONVECTOR VS PIECEWISE-LINEAR FITTING FOR SIGNAL AND POWER INTEGRITY SIMULATION
VECTOR VS PIECEWISE-LINEAR FITTING FOR SIGNAL AND POWER INTEGRITY SIMULATION
 
Multiuser detection new
Multiuser detection newMultiuser detection new
Multiuser detection new
 
104623 time division multiplexing (transmitter, receiver,commutator)
104623 time  division multiplexing (transmitter, receiver,commutator)104623 time  division multiplexing (transmitter, receiver,commutator)
104623 time division multiplexing (transmitter, receiver,commutator)
 

Similar to Coding for phase noise channels

PhD thesis - Decision feedback equalization and channel estimation for SC-FDMA
PhD thesis - Decision feedback equalization and channel estimation for SC-FDMAPhD thesis - Decision feedback equalization and channel estimation for SC-FDMA
PhD thesis - Decision feedback equalization and channel estimation for SC-FDMAGillian Huang
 
Audio Equalization Using LMS Adaptive Filtering
Audio Equalization Using LMS Adaptive FilteringAudio Equalization Using LMS Adaptive Filtering
Audio Equalization Using LMS Adaptive FilteringBob Minnich
 
Homodyne OLT-ONU Design for Access Optical Networks
Homodyne OLT-ONU Design for Access Optical NetworksHomodyne OLT-ONU Design for Access Optical Networks
Homodyne OLT-ONU Design for Access Optical NetworksJosep Fabrega
 
02 whole
02 whole02 whole
02 wholevino205
 
Extended LTE Coverage for Indoor Machine Type Communication.pdf
Extended LTE Coverage for Indoor Machine Type Communication.pdfExtended LTE Coverage for Indoor Machine Type Communication.pdf
Extended LTE Coverage for Indoor Machine Type Communication.pdfYAAKOVSOLOMON1
 
MIMO-OFDM communication systems_ channel estimation and wireless.pdf
MIMO-OFDM communication systems_  channel estimation and wireless.pdfMIMO-OFDM communication systems_  channel estimation and wireless.pdf
MIMO-OFDM communication systems_ channel estimation and wireless.pdfSamerSamerM
 
Distance Measurements using Ultra Wide-Band (UWB)
Distance Measurements using Ultra Wide-Band (UWB)Distance Measurements using Ultra Wide-Band (UWB)
Distance Measurements using Ultra Wide-Band (UWB)Iqbal Hossain
 
disertation_Pavel_Prochazka_A1
disertation_Pavel_Prochazka_A1disertation_Pavel_Prochazka_A1
disertation_Pavel_Prochazka_A1Pavel Prochazka
 
OFDM Based Cognitive radio
OFDM Based Cognitive radioOFDM Based Cognitive radio
OFDM Based Cognitive radioDeepesh Sinnya
 
checking from app
checking from appchecking from app
checking from appDeepa Bman
 
Microwave Engineering
Microwave EngineeringMicrowave Engineering
Microwave EngineeringSam Joey
 
Multi tone test
Multi tone testMulti tone test
Multi tone testslmnsvn
 

Similar to Coding for phase noise channels (20)

論文
論文論文
論文
 
PhD thesis - Decision feedback equalization and channel estimation for SC-FDMA
PhD thesis - Decision feedback equalization and channel estimation for SC-FDMAPhD thesis - Decision feedback equalization and channel estimation for SC-FDMA
PhD thesis - Decision feedback equalization and channel estimation for SC-FDMA
 
Audio Equalization Using LMS Adaptive Filtering
Audio Equalization Using LMS Adaptive FilteringAudio Equalization Using LMS Adaptive Filtering
Audio Equalization Using LMS Adaptive Filtering
 
Homodyne OLT-ONU Design for Access Optical Networks
Homodyne OLT-ONU Design for Access Optical NetworksHomodyne OLT-ONU Design for Access Optical Networks
Homodyne OLT-ONU Design for Access Optical Networks
 
02 whole
02 whole02 whole
02 whole
 
Extended LTE Coverage for Indoor Machine Type Communication.pdf
Extended LTE Coverage for Indoor Machine Type Communication.pdfExtended LTE Coverage for Indoor Machine Type Communication.pdf
Extended LTE Coverage for Indoor Machine Type Communication.pdf
 
Final Thesis Report
Final Thesis ReportFinal Thesis Report
Final Thesis Report
 
MastersThesis
MastersThesisMastersThesis
MastersThesis
 
mythesis
mythesismythesis
mythesis
 
MIMO-OFDM communication systems_ channel estimation and wireless.pdf
MIMO-OFDM communication systems_  channel estimation and wireless.pdfMIMO-OFDM communication systems_  channel estimation and wireless.pdf
MIMO-OFDM communication systems_ channel estimation and wireless.pdf
 
Distance Measurements using Ultra Wide-Band (UWB)
Distance Measurements using Ultra Wide-Band (UWB)Distance Measurements using Ultra Wide-Band (UWB)
Distance Measurements using Ultra Wide-Band (UWB)
 
disertation_Pavel_Prochazka_A1
disertation_Pavel_Prochazka_A1disertation_Pavel_Prochazka_A1
disertation_Pavel_Prochazka_A1
 
OFDM Based Cognitive radio
OFDM Based Cognitive radioOFDM Based Cognitive radio
OFDM Based Cognitive radio
 
Cdma pdf
Cdma pdfCdma pdf
Cdma pdf
 
checking from app
checking from appchecking from app
checking from app
 
shifas_thesis
shifas_thesisshifas_thesis
shifas_thesis
 
repport christian el hajj
repport christian el hajjrepport christian el hajj
repport christian el hajj
 
Microwave Engineering
Microwave EngineeringMicrowave Engineering
Microwave Engineering
 
Multi tone test
Multi tone testMulti tone test
Multi tone test
 
master thesis
master thesismaster thesis
master thesis
 

Coding for phase noise channels

  • 1. Chalmers & TELECOM Bretagne Coding for phase noise channels IVAN LELLOUCH Department of Signals & Systems Chalmers University of Technology Gothenburg, Sweden 2011 Master’s Thesis 2011:1
  • 2.
  • 4.
  • 6.
  • 7. Contents 1 Introduction 1 2 Bounds and capacity 3 2.1 Capacity and information density . . . . . . . . . . . . . . . . . . . . . . . 3 2.2 Binary hypothesis testing . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 2.3 Bounds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 2.3.1 Dependence Testing bound [5] . . . . . . . . . . . . . . . . . . . . 5 2.3.2 Meta converse bound [5] . . . . . . . . . . . . . . . . . . . . . . . . 7 3 AWGN channel 9 3.1 The AWGN channel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 3.1.1 Information density . . . . . . . . . . . . . . . . . . . . . . . . . . 10 3.1.2 Depending testing bound . . . . . . . . . . . . . . . . . . . . . . . 10 3.1.3 Meta converse bound . . . . . . . . . . . . . . . . . . . . . . . . . . 10 4 Phase noise channels 13 4.1 Uniform phase noise channel . . . . . . . . . . . . . . . . . . . . . . . . . 13 4.1.1 Information density . . . . . . . . . . . . . . . . . . . . . . . . . . 14 4.1.2 Depending testing bound . . . . . . . . . . . . . . . . . . . . . . . 14 4.2 Uniform phase noise AWGN channel . . . . . . . . . . . . . . . . . . . . . 17 4.2.1 Information density . . . . . . . . . . . . . . . . . . . . . . . . . . 17 4.2.2 Dpending testing bound . . . . . . . . . . . . . . . . . . . . . . . . 19 4.3 Tikhonov phase noise channel . . . . . . . . . . . . . . . . . . . . . . . . . 20 4.3.1 Information density . . . . . . . . . . . . . . . . . . . . . . . . . . 21 4.3.2 Depending testing bound . . . . . . . . . . . . . . . . . . . . . . . 24 4.3.3 Meta converse bound . . . . . . . . . . . . . . . . . . . . . . . . . . 29 5 Conclusion 30 Bibliography 31 i
  • 8. List of Figures 3.1 DT and converse for AWGN channel - SN R = 0 - Pe = 10−3 . . . . . . . 12 4.1 DT and constraint capacities for three uniform AM constellations. . . . . 20 4.2 Tikhonov probability density function . . . . . . . . . . . . . . . . . . . . 22 4.3 Two 64-QAM constellations in AWGN phase noise channel . . . . . . . . 25 4.4 Robust Circular QAM constellation with phase noise . . . . . . . . . . . . 26 4.5 DT curves for two 64-QAM constellations in AWGN phase noise channel with SNR = 0dB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 4.6 DT curves for two 64-QAM constellations in AWGN phase noise channel with SNR = 15dB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 4.7 Comparaison of DT curves for different phase noise power . . . . . . . . . 27 4.8 Comparaison of DT curves for different probabilities of error . . . . . . . 28 ii
  • 9. 1 Introduction Since Shannon’s landmark paper [1], there has been a lot of studies regarding the chan- nel capacity, which is the amount of information we can reliably sent through a channel. This result is a theoretical limit, that required to use an infinite block length. In practice, we want to know, for a given communication system, how far our system is from this upper bound. When we design a system, there are two main parameters we need to determine. The error probability that the system can tolerate and the delay constraint, which is related to the size of the message we want to send, ie., the block length. Therefore, we want to find, given these parameters what is the new upper bound for our system. Thus we will work with finite block length and a given probability of error. Two bounds are defined in order to determine this new limit, the achievability bound and the converse bound. Achievability bound is a lower bound on the size of the codebook, given a block length and error probability. Converse bound is an upper bound on the size of the codebook, given a block length and error probability. By using both of this bounds, we can determine an approximation of the theoretical limit of information we can send through a channel, for a given block length and a probability of error. Achievability bounds already exist in information theory studies. Three main bounds were defined by Feinstein [2], Shannon [3] and Gallager [4]. An optimization of auxiliary constants was needed in order to compute those bounds. Thanks to this work we have some new insights regarding how far systems can work from the capacity of the channel with a finite blocklength. In a recent work [5] Polyanskiy et al defined a new achievability bound that do not required any auxiliary constant optimization and is tighter than the three bounds in [2, 3, 4] and one converse bound. 1
  • 10. CHAPTER 1. INTRODUCTION This thesis is in the framework of the MAGIC project involving Chalmers University of Technology, Ericsson AB and Qamcom Technology AB and the context is the microwave backhauling for IMT advanced and beyond. A part of this project is to investigate the modulations and coding techniques for channels impaired by phase noise. In digital com- munications systems, the use of low-cost oscillators at the receiver causes phase noise which can become a severe problem for high symbol rates and constellation sizes. For channels impaired by phase noise, the codes we are usually using do not perform as good as they do in classical channels. In this thesis, we will deal with two bounds from [5], an achievable bound and a converse bound. We will apply those bounds to phase noise channels and see how far we are from the capacity and which rates we can reach given a block length and an error probability. The outline of the Thesis is as follows. In Chapter 2, we introduce the capacity and the bounds that will be used in the follow- ing chapters. The main result is the expression of the depending testing (DT) bound that we used to find a lower bound on the maximal coding rate. We will explain how Polyanskiy et al derived this bound and show how we can use it for our channel. In Chapter 3, we first apply our results to the additive white Gaussian noise (AWGN) channel. It is useful to see how the equations works for continuous noise channel, but also to determine the impact the phase noise will cause on the maximal coding rate. In Chapter 4, we find the main results of the thesis. We apply the DT bound on sev- eral partially coherent additive white Gaussian noise (PC-AWGN) channels, ie., AWGN channel impaired by phase noise, and compare them with the bound for the AWGN channel and the constrained capacity. Therefore, the loss induce by the phase noise will be estimated, for a given channel, and this will leads to an approximation of the maximal coding rate for the PC-AWGN we investigated. We also present a constellation designed for phase noise channel and show the performance improvements. 2
  • 11. 2 Bounds and capacity 2.1 Capacity and information density We denote by A and B the input and output sets. If X and Y are random variables from A and B respectively then x and y are their particular realizations. X and Y denote the probability spaces and PX and PY are the probability density functions of X and Y respectively. We also define PY |X , the conditional probability from A to B given a codebook (c1 ,...,cM ) and we denote by M its size. Since we are interested in finite block length analysis, a realisation x of a random variable X represent a n-dimensional vector, ie., x = (x1 ,...,xn ). The capacity C of a channel is the maximum of information we can reliably send through it with a vanishing error probability and for an infinite block length. For input and output X and Y distributed according to PX and PY respectively, the capacity C is given by C = max {I(X; Y )} (2.1) X where I(X,Y ) is the mutual information between X and Y . The expression is max- imized with respect to the choice of the input distribution PX . pXY (x,y) I(X; Y ) = pXY (x,y) log2 dxdy X ,Y pX (x)pY (y) where the logarithmic term is called the information density : pX,Y (x,y) i(x; y) = log2 (2.2) pX (x)pY (y) It is proven in [6] that the capacity for the AWGN channel is achieved by a Gaussian input distribution and is given by 3
  • 12. CHAPTER 2. BOUNDS AND CAPACITY 1 C= log2 (1 + SNR) (2.3) 2 P where SNR is the signal-to-noise ratio, ie., SNR = N , where P and N are the input and noise power respectively. The capacity can be computed when we know which distribution maximized (2.1). In this case we say that the distribution is capacity achieving. For some channels, such as phase noise channels, we have few informations regarding the capacity. For those channels, we will determine an input and use it for our calculations. Thus, we will work with a constrained capacity, ie., constrained to a specific input distribution, which is an upper bound of the information that can be sent through the channel for a given input distribution. log2 M The capacity can also be define by using the rate R = n 1 C = lim lim log2 M ∗ (n, ) (2.4) ←0 n←inf n where n is the block length, the probability of error and M ∗ defined as follow M ∗ (n, ) = max {M : ∃(n,M, ) − code} (2.5) 2.2 Binary hypothesis testing Later in this thesis we will need to use a binary hypothesis testing in order to define an upper bound for the rate. We consider a random variable R defined as follow R : {P,Q} → {1,0} (2.6) where 1 indicates that P is chosen. We also consider the random transformation PZ|R : R → {1,0} (2.7) We define βα (P,Q), the maximal probability of error under Q if the probability of error under P is lower than α. We can denote this test by βα (P,Q) = inf PZ|R (1|r)Q(r) (2.8) PZ|R : r∈R PZ|R (1|r)P (r)≥α r∈R 2.3 Bounds Yury Polyanskiy defined in [5] the DT bound and the meta converse bound over random codes . In this chapter we will start by describing those bounds and apply them over continuous channels. 4
  • 13. CHAPTER 2. BOUNDS AND CAPACITY 2.3.1 Dependence Testing bound [5] We will present the technique proposed in [5] regarding the bounding of the error prob- ability for any input distribution given a channel. Theorem 1. Depending testing bound Given an input distribution PX on A, there exists a code with codebook size M , and the average probability of error is bounded by + M −1 ≤ E exp − i(x,y) − log2 (2.9) 2 where |u|+ = max(0,u) (2.10) Proof. Let Zx (y) be the following function Zx (y) = 1(i(x,y)>log M −1 ) (2.11) 2 where 1A (.) is an indicator function : 1, if x ∈ A, 1A (x) = (2.12) 0, otherwise. For a given codebook (c1 ,...,cM ), the decoder computes (2.11) for the codeword cj starting with c1 , until it finds Zcj (y) = 1, or the decoder returns an error. Therefore, there is no error with probability   Pr {Zcj (y) = 1} {Zci (y) = 0} (2.13) i<j Then, we can write the error for the j th codeword as   (cj ) = Pr {Zcj (y) = 0} {Zci (y) = 1} (2.14) i<j Using the union bound on this expression and (2.11) (cj ) ≤ Pr {Zcj (y) = 0} + Pr [{Zci (y) = 1}] (2.15) i<j M −1 M −1 = Pr i(cj ,y) ≤ log + Pr i(ci ,y) > log (2.16) 2 2 i<j 5
  • 14. CHAPTER 2. BOUNDS AND CAPACITY The codebook is generated randomly according to the distribution PX , and we denote by y a realization of the random variable Y , but independent of the transmitted codeword. ¯ Thus, the probability of error, if we send the codeword cj , is bounded by M −1 M −1 (cj ) ≤ Pr i(x,y) ≤ log + (j − 1)Pr i(x,¯) > log y (2.17) 2 2 1 Then, if we suppose that Pr(cj ) = M, we have M 1 = (cj ) (2.18) M j=1 and M 1 M −1 M −1 ≤ Pr i(x,y) ≤ log + (j − 1)Pr i(x,¯) > log y (2.19) M 2 2 j=1 which give us finally the following expression for the average error probability M −1 M −1 M −1 ≤ Pr i(x,y) ≤ log + Pr i(x,¯) > log y (2.20) 2 2 2 We know that p(x)p(y) exp −|i(x,y) − log u|+ = 1(i(x,y)≤log u) + u 1 (2.21) p(x,y) (i(x,y)>log u) By averaging over p(x,y), and using y , we have ¯ exp −|i(x,y) − log u|+ = Pr(i(x,y) ≤ log u) + u p(x)p(¯)1(i(x,¯)>log u) (2.22) y y x y ¯ and knowing that y is independent of x it leads us to ¯ exp −|i(x,y) − log u|+ = Pr(i(x,y) ≤ log u) + u p(x,¯)1(i(x,¯)>log u) y y (2.23) x y ¯ and finally exp −|i(x,y) − log u|+ = Pr(i(x,y) ≤ log u) + uPr(i(x,¯) > log u) y (2.24) M −1 Thus, replacing u = 2 and using (2.20) we obtain (2.9) which completes the proof. This expression needs no auxiliary constant optimization and can be computed for a given channel by knowing the information density. Applications over AWGN channels and phase noise channels will be shown in following chapters. 6
  • 15. CHAPTER 2. BOUNDS AND CAPACITY 2.3.2 Meta converse bound [5] The meta converse bound is an upper bound on the size of the codebook for a given error probability and block length. To define this bound, we will use the binary hypothesis testing defined in (2.8). Theorem 2. Let denote by A and B the input and output alphabets respectively. We consider two random transformations PY |X and QY |X from A to B, and a code (f,g) with average probability of error under PY |X and under QY |X . The probability distribution induced by the encoder is PX = QX . Then we have β1− (PY |X ,QY |X ) ≤ 1 − (2.25) where β is the binary hypothesis testing defined in (2.8). Proof. We denote by s the input message chosen in (s1 ,...,sM ) and by x = f (s) the encoded message. Also y is the message before decoding and z = g(z) the decoded message. We define the following random variable to represent an error-free transmission Z = 1s=z (2.26) First we notice that the conditional distribution of Z given (X,Y ) is the same for both channels PY |X and QY |X . M P [Z = 1|X,Y ] = P [s = si ,z = si |X,Y ] (2.27) i=1 Then, given (X,Y ), since the input and output messages are independent we have M P [Z = 1|X,Y ] = P [s = si |X,Y ] P [z = si |X,Y ] (2.28) i=1 We can simplify the expression as follows M P [Z = 1|X,Y ] = P [s = si |X] P [z = si |Y ] (2.29) i=1 We recognize in the second term of the product the decoding function, while the first term is independent of the choice of the channel given the definition of the probability distributions induced by the encoder PX = QX that we defined earlier. Then, using PZ|XY = QZ|XY (2.30) we define the following binary hypothesis testing PZ|XY (1|xy)PXY (x,y) = 1 − (2.31) x∈A y∈B PZ|XY (1|xy)QXY (x,y) = 1 − (2.32) x∈A y∈B and using the definition in (2.8) we get (2.25). 7
  • 16. CHAPTER 2. BOUNDS AND CAPACITY Theorem 3. Every code with M codeword in A and an average probability of error satisfies 1 M ≤ sup inf (2.33) PX QY β1− (PXY ,PX × QY ) where PX describes all input distributions on A and QY all output distributions on B. 8
  • 17. 3 AWGN channel In this chapter we will apply the bounds discussed in the previous chapter to the AWGN channel. We know several results for this channel, such as the capacity. We will see how far the DT and the meta-converse bounds are from the capacity and if they are tight enough to have an idea of the maximum achievable rate given a block length and error probability. The results for the AWGN channel will be useful in the next chapter to evaluate the effect of the phase noise on the achievable rates for AWGN channels impaired by phase noise. 3.1 The AWGN channel Let us consider x ∈ X , y ∈ Y and the transition probability between X and Y , PXY . We have the following expression for the AWGN channel y =x+t (3.1) where the noise samples t ∼ N (0,σIn ) are independent and identically-distributed. Thus, we know the conditional output probability function n/2 (y−x)T (y−x) 1 PY |X=x = e− 2σ 2 (3.2) 2πσ 2 where .T is the transpose operation. We know from [6] that the Gaussian input distri- bution achieves the capacity for this channel. So we will consider x ∼ N (0,P In ) and denote by PX the corresponding probability distribution. Given the conditional output probability function and the input, the output distri- bution is defined as follow (summation of Gaussian variables) y ∼ N (0,(σ 2 + P )In ) (3.3) 9
  • 18. CHAPTER 3. AWGN CHANNEL 3.1.1 Information density We can now define the information density of the channel using distributions y|x ∼ N (x,σ 2 In ) and y ∼ N (0,(σ 2 + P )In ). log2 (e) yT y (y − x)T (y − x) n P + σ2 i(x,y) = − + log2 (3.4) 2 P + σ2 σ2 2 σ2 which can be rewritten as n n P + σ 2 log2 (e) 2 yi n2i i(x,y) = log2 + − 2 (3.5) 2 σ2 2 P + σ2 σ i=1 3.1.2 Depending testing bound To compute the DT bound for the AWGN channel, we are using (3.5) in (2.9). n + n P + σ 2 log2 (e) 2 yi n2i M −1 ≤ E exp − log2 + − 2 − log2 (3.6) 2 σ2 2 P +σ 2 σ 2 i=1 Then to compute the expectation we can use a Monte Carlo simulation. The samples are generated according to the model description in section 3.1. For this simulation, we use the input distribution x ∼ N (0,P ). This leads to the DT bound for the maximal coding rate on this channel, ie., this bound will be an upper bound for all other DT’s for this channel. In practice, discrete constellations are used for our systems. Therefore we also look at the results for a know discrete input constellation which will be useful in the next chapter when we will compare results for the AWGN and the partially coherent AWGN channels. 3.1.3 Meta converse bound We know that the input distribution and the noise are Gaussian with parameters P and σ 2 respectively. We also know that the summation of two Gaussian random variable is a Gaussian random variable, thus we chose y ∼ N (0,σY In ) as the output distribution for the computation of the converse bound. We can now define the information density n 2 n 2 log2 e yi 2 i(x,y) = log2 σY + 2 2 2 − (yi − xi ) σY (3.7) i=1 We choose the input such that ||x||2 = nP . To simplify calculations, we are using √ √ x = x0 = ( P ,..., P ). This is possible because of the symmetry of the problem. Thus, using Zi ∼ N (0,1), Hn and Gn have the following distributions 10
  • 19. CHAPTER 3. AWGN CHANNEL n √ P 1 Hn = n log2 σY − n log2 e + log2 e (1 − σY )Zi2 + 2 P σY Zi 2 (3.8) 2 2 i=1 and n √ P 1 Gn = n log2 σY + n 2 log2 e + 2 log2 e (1 − σY )Zi2 + 2 P Zi 2 (3.9) 2σY 2σY i=1 where Hn and Gn are the information density under PY |X and PY respectively. 2 Then, by choosing σY = 1 + P , we have n n P 2 Hn = log2 (1 + P ) + log2 e 1 − Zi2 + √ Zi (3.10) 2 2(1 + P ) P i=1 and n n P 1 Gn = log2 (1 + P ) − log2 e 1 + Zi2 − 2 1 + Zi (3.11) 2 2 P i=1 Notice that Hn and Gn are non-central χ2 distributions, thus we have n P log2 e Hn = (log2 (1 + P ) + log2 e) − yn (3.12) 2 2(1 + P ) n with yn ∼ χ2 ( P ), and n n P log2 e Gn = (log2 (1 + P ) + log2 e) − yn (3.13) 2 2 n with yn ∼ χ2 (n + P ). n Finally, to compute the bound, we find γn such as Pr [Hn ≥ γn ] = 1 − (3.14) which lead to 1 M≤ (3.15) Pr [Gn ≥ γn ] Those expressions can be computed directly using closed-form expressions. For some channels, when we do not have them, therefore we have to compute the bound with a Monte Carlo simulation and we will discuss the issue of calculate the second probability Pr [Gn ≥ γn ], which decreases exponentially to 0, by this method. In Fig. 3.1, we plot the results for a real-valued AWGN channel of the rate, in bit per channel use against the block length n. For this example, we use the capacity achieving input distribution x ∼ N (0,σ) to compute the DT bound. For the following chapters, we will use discrete input distribution for the AWGN channel in order to compare with the results we will find for partially coherent AWGN channels. We see that the gap between both curves, the DT and the converse, get smaller when the block length get larger. This result give a good approximation of the maximal coding rate for this channel given the error probability. We also know from the definition of the capacity, that both curves will tend toward it when n grows to infinity. 11
  • 20. CHAPTER 3. AWGN CHANNEL 0.6 0.5 0.4 Rate, bit/ch.use 0.3 0.2 0.1 Meta−converse DT Capacity 0 0 200 400 600 800 1000 1200 1400 1600 1800 2000 Blocklength, n Figure 3.1: DT and converse for AWGN channel - SN R = 0 - Pe = 10−3 12
  • 21. 4 Phase noise channels In this chapter we will focus on channels impaired by phase noise. First, we will see some results with a uniform phase noise and the equations that lead to the DT bound. Then we will focus on two more realistic channels, the AWGN channel impaired with uniform phase noise, and with Tikhonov phase noise. 4.1 Uniform phase noise channel We consider a uniform phase noise channel where θ is the additive noise on the phase. This noise is distributed uniformly between −a and a, ie., θ ∼ U (−a,a). If x ∈ X and y ∈ Y we have the following expressions y = xeiθ (4.1) and yk = xk eiθk (4.2) Using an interleaver, we can assume that the noise is memoryless, then we have the conditional output distribution n p(y|x) = p(yk |xk ) (4.3) k=1 Notice that regardless the phase noise, the noise cannot change the magnitude of the sample, ie., |x| = |y|. Then, for this channel we will only consider a constellation with all points belonging to the same ring. 13
  • 22. CHAPTER 4. PHASE NOISE CHANNELS 4.1.1 Information density The information density is defined by PY |X=x (y) i(x,y) = log2 (4.4) PY (y) We know that p(yk ) = p(θk ), thus we have  1 , if |θ | ≤ a  k p(θk ) = 2a (4.5) 0, otherwise  Using (4.3) and (4.5) we obtain the expression of the conditional output distribution  1 , if y ∈ Yx  n PY |X=x (y) = p(yi |xi ) = (2a)n (4.6) 0, otherwise  i=1 where yi Yx = y ∈ Y, ∀i ∈ [1,n] ,| arg |≤a (4.7) xi Then, using the law of total probability, we obtain the output distribution 1 |Xy | PY (y) = p(y|x) = (4.8) |X | x |X |(2a)n where yi Xy = x ∈ X , ∀i ∈ [1,N ] ,| arg |≤a (4.9) xi Finally, the information density for the uniform phase noise channel is given by the following expression   log |X | , if y ∈ Yx  2 i(x,y) = |Xy | (4.10)   0, otherwise 4.1.2 Depending testing bound Since the capacity achieving distribution for this channel is not known, we will work with a given input constellation. Let the input alphabet be distributed according to an m-PSK modulation. Let n be the block length, M the size of the codebook M and E(M ) the set of all M -size codebooks. The codebook is randomly chosen. 14
  • 23. CHAPTER 4. PHASE NOISE CHANNELS Given a probability of error , we want to find the highest M such as the following expression stands M −1 + ≤ E e−(i(x,y)−log2 ( 2 )) (4.11) Since we are using discrete input, we can rewrite the expression as follows by ex- panding the expectation over PXY M −1 + ≤ p(x,y)e(i(x,y)−log2 ( 2 )) dy (4.12) x∈X y∈Y Let z(x,y) = |Xy |, we obtain + 2M − log2 ≤ P (z)e (M −1)z (4.13) z∈N Then, the probability P (z) can be expanded as follows P (z) = P (z|x,y)p(y|x)p(x)dy (4.14) x∈X y∈Y which in (4.13) gives + 2M − log2 ≤ P (z|x,y)p(y|x)p(x)dy e (M −1)z (4.15) z∈N x∈X y∈Y Since the input is an m-PSK modulation and the phase noise is uniform we know the expressions of p(y|x) and p(x) + 1 1 − log2 2M ≤ P (z|x,y) n mn dy e (M −1)z (4.16) y∈Yx (2a) z∈N x∈X Then, we can simplify the equation using the symmetry of the problem, by choosing x0 a realisation of X + 1 − log2 2M ≤ P (z|x0 ,y) dy e (M −1)z (4.17) y∈Yx0 (2a)n z∈N and by expanding the integration over y we obtain a a + 1 − log2 2M ≤ ··· P (z|x0 ,y)dy1 · · · dyn e (M −1)z (4.18) (2a)n −a −a z∈N Let V (x,y) be the number of neighbours in X for y ∈ Yx . V = V (x,y) − 1. Then the probability P (z|x,y) can be written as follows 15
  • 24. CHAPTER 4. PHASE NOISE CHANNELS V −1 z−1 V −z−1 V 1 P (z|x,y) = (M − 1 − j) (mn − M − j) (4.19) z (mn − 1 − j) j=0 j=0 j=0 P (z|x,y) = 0 if z > V (x,y). Given the phase noise parameter a, and the number of points m in one ring of the constellation, we can determine the function v(yk ). v(yk ) define the number of points the output yk can come from, in one ring of the constellation. This function is a simple function with two values v1 and v2 as soon as the points in each ring are equally spaced. Then we can define two constant d1 and d2 by the following expressions yk +a d1 = 1(v(yk ) = v1 )dyk (4.20) yk −a yk +a d2 = 1(v(yk ) = v2 )dyk (4.21) yk −a Finally we have the following expression to compute the bound max(v1 n ,v2 n ) n + 1 n − log2 2M ≤ d1 u d2 n−u p z|V = v1 u v2 n−u e (M −1)(z+1) (2a)n z=0 u=0 u (4.22) The complexity of this calculation depends on the complexity of p (z|V = v1 u v2 n−u ). This is a product of V terms, so the complexity is O(2n ). Of course, we can’t compute this expression because of its complexity. In the next chapters, we are using partially coherent AWGN channels, and the expressions do not need to calculate the probability p(z|x,y) which make the computation much faster. 16
  • 25. CHAPTER 4. PHASE NOISE CHANNELS 4.2 Uniform phase noise AWGN channel We consider an AWGN channel impaired with a uniform phase noise θ ∼ U (−a,a). If 2 x ∈ X , y ∈ Y and t ∼ N (0,σN ) we have y = xeiθ + t (4.23) and yk = xk eiθk + tk (4.24) For this channel we can define the information density as follows. 4.2.1 Information density We need both expressions of PY and PY |X to determine the expression of i(x,y). First, we know that the noise is memoryless, which allows us to write n PY |X=x (y) = p(yk |xk ) (4.25) k=1 where xk , yk and tk can be written as xk = ak eibk tk = ck eidk yk = αk eiβk which give us the following expression for the conditional output distribution (using polar coordinates) p(yk |xk ) = αk p(θk |xk )p(tk |θk ,xk )dθk (4.26) θk a αk 1 |tk |2 = exp − 2 dθk (4.27) −a 2a 2πσ 2 2σ where |tk |2 = |yk − xk eiθk |2 (4.28) We develop (4.28) as follows |tk |2 = |αk cos(βk ) + iαk sin(βk ) − |xk | cos(θk + arg(xk )) + i|xk | sin(θk + arg(xk ))|2 (4.29) = (αk cos(βk ) − |xk | cos(θk ) + arg(xk ))2 + (αk sin(βk ) − |xk | sin(θk + arg(xk )))2 (4.30) = α2 + a2 − 2αk ak cos(θk + bk − βk ) k k (4.31) 17
  • 26. CHAPTER 4. PHASE NOISE CHANNELS which used in (4.26) gives (a2 +α2 ) a exp(− k 2 k ) 2σ αk ak cos(θk + bk − βk ) p(yk |xk ) = αk exp dθk (4.32) (2a)(2πσ 2 ) −a σ2 Then, using the law of total probabilities we obtain the expression of the output density probability m−1 p(yk ) = p(xu,k )p(yk |xu,k ) (4.33) u=0 Now, we need to choose the input distribution to determine the expression of the information density. We will consider a set of M codewords (c1 ,...,cM ) with the same probability. Then we have m−1 (a2 +α2 ) a αk exp(− u,k 2 k ) 2σ αk au,k cos(θk + bu,k − βk ) p(yk ) = exp dθk (4.34) m (2a)(2πσ 2 ) −a σ2 u=0 which finally gives the following expression for the information density (a2 +α2 )   exp(− k 2 k ) a αk ak cos(θk +bk −βk ) N −a exp 2σ (2a)(2πσ2 ) σ2 dθk i(x,αeiβ ) =   log2   (a2 +α2 )  (4.35)  k=1 m−1 1 exp(− u,k 2 k ) a αk au,k cos(θk +bu,k −βk ) −a exp 2σ u=0 m (2a)(2πσ2 ) σ2 dθk and with some simplifications we obtain   a2 a N exp(− 2σ2 ) −a exp αk ak cos(θk +bk −βk ) dθk k σ2 i(x,αeiβ ) = log2  a2  (4.36) m−1 1 u,k a αk au,k cos(θk +bu,k −βk ) k=1 u=0 m exp(− 2σ2 ) −a exp σ2 dθk We recognize in the information density expression the following integral s ek cos(x) dx, s ≤ π. (4.37) 0 We can find a closed-form expression only if we choose s = π, by using the Bessel function of the first kind. For this case (4.32) becomes, (a2 +α2 ) π exp(− k 2 k ) 1 2σ αk ak cos(θk + bk − βk ) p(αk ,βk |ak ,bk ) = αk exp dθk (4.38) (2πσ 2 ) 2π −π σ2 and using the properties of trigonometric functions we can rewrite the expression as follows (a2 +α2 ) π exp(− k 2 k ) 1 2σ i2 αk ak sin(θk ) p(αk ,βk |ak ,bk ) = αk exp dθk (4.39) (2πσ 2 ) 2π −π σ2 18
  • 27. CHAPTER 4. PHASE NOISE CHANNELS We notice that the expression is independent of βk , the angle of y π p(αk |ak ,bk ) = p(αk ,βk |ak ,bk )dβk (4.40) −π = (2π)p(αk ,βk |ak ,bk ) (4.41) which leads to αk a2 + α2 αk ak p(αk |ak ) = exp − k 2 k I0 (4.42) σ2 2σ σ2 using I0 (.) the Bessel function of the first kind. M-PSK √ Considering a M-PSK, we have ak = P and PY can be defined as follows m PY (αk ) = p(au,k )p(αk |ak ) = p(αk |ak ) (4.43) u=1 which leads to i(x,y) = 0. Given the fact that we have a non-coherent AWGN chan- nel with an m-PSK modulation, we easily understand that no information can be sent through the channel. Amplitude modulation Now, we consider an amplitude modulation. If we have R points in our constellation, √ and if ar = Pr then we have a2   αk ak N exp − 2σ2 k I0 σ2 i(x,y) = i(a,α) = log2  √  (4.44) R P αk Pr k=1 r=1 exp − 2σr2 I0 σ2 Once again, given the channel, there is no information on the phase. So we can work by using only the magnitude of each point. 4.2.2 Dpending testing bound Now we want to determine the upper bound for this channel. First, we pick an input constellation and then we use it in (4.44) to determine the equation of the DT (2.9), which leads us to the expression +    a2   αk ak N exp − 2σ2 k I0 ≤ E exp −   log2  σ2 √  − log2 M − 1  (4.45)  R P − 2σr2 αk Pr 2 k=1 r=1 exp I0 σ2 We use a Monte Carlo simulation to calculate this expression. 19
  • 28. CHAPTER 4. PHASE NOISE CHANNELS Amplitude modulation input For this constellation, we consider m points equally spaced with average power P = 1. In Fig. 4.1 we plot the rate, in bit per channel use, against the block length n. We choose to present 3 constellations, with 8, 16 and 32 points. The Gaussian noise is defined by SNR = 15dB, and the error probability is Pe = 10−3 . For each constellation, the depending testing bound and the constrained capacity are plotted. 4.5 4 3.5 3 Rate, bit/Ch.use 2.5 2 1.5 1 DT and constraint capacity : 8−AM constellation 0.5 DT and constraint capacity : 16−AM constellation DT and constraint capacity : 32−AM constellation 0 0 200 400 600 800 1000 1200 1400 1600 1800 2000 blocklength, n Figure 4.1: DT and constraint capacities for three uniform AM constellations. We see in Fig. 4.1 that, for a given constellation, both curves, the DT bound and the constrained capacity, are tight when the block length is large. We can also notice that the gap between both curves decreases faster when we have fewer points in our constellation. 4.3 Tikhonov phase noise channel A more realistic channel to describe a system impaired by a phase noise is the Tikhonov AWGN channel. We have a closed-form expression for the noise and, using Lapidoth’s result in [7] we can also have one for the conditional output. 20
  • 29. CHAPTER 4. PHASE NOISE CHANNELS We choose to study this model because it is a good approximation of the phase noise error induced by a first-order phase-locked loop [? ]. We consider t ∼ N (0,σn ), the Gaussian noise, and θ the phase noise distributed according to the Tikhonov distribution. This distribution is presented in section 4.3. y = xeiθ + t (4.46) and yk = xk eiθk + tk (4.47) Tikhonov distribution The Tikhonov distribution, also known as Von Mises distribution [? ], is an approxima- tion of the wrapped Gaussian which is defined as follows 1 −(θ−2kπ)2 pW (θ) = pΘ (θ + 2kπ) = √ e 2σ 2 (4.48) k∈Z 2πσ 2 k∈Z Its support is [−π; π] and is function of a parameter ρ. The probability density function is given by eρ cos(x) p(x|ρ) = (4.49) 2πI0 (ρ) In Fig. 4.2 we see the Tikhonov distribution for 3 values of the parameter ρ. The larger the parameter is, the smaller the noise is. 4.3.1 Information density First, we need to determine the expression for the conditional output distribution. We know that the noise is memoryless, so we can focus on p(yk |xk ). π p(yk |xk ) = pn (yk |xk ,θk )pθ (θk )dθk (4.50) −π Using both expressions of the Gaussian pdf and the Tikhonov pdf, we have 2 π 1 yk − xk ejθk eρ cos(θk ) p(yk |xk ) = exp − dθk (4.51) −π 2πσ 2 2σ 2 2πI0 (ρ) 2 1 π yk − xk ejθk = 2 σ 2 I (ρ) exp − + ρ cos(θk ) dθk (4.52) (2π) 0 −π 2σ 2 21
  • 30. CHAPTER 4. PHASE NOISE CHANNELS 9 rho=500 8 rho=100 rho=10 7 6 5 4 3 2 1 0 −3 −2 −1 0 1 2 3 angle (rad) Figure 4.2: Tikhonov probability density function we can now expand the expression in the exponential 2 yk − xk ejθk = |yk |2 + |xk |2 − yk xk ejθk − yk x∗ e−jθk ∗ k (4.53) = |yk |2 + |xk |2 − yk xk (cos(θk ) + sin(θk )) − yk x∗ (cos(θk ) − sin(θk )) ∗ k (4.54) = |yk |2 + |xk |2 − 2 (yk xk ) cos(θk ) + 2 (yk xk ) sin(θk ) ∗ ∗ (4.55) which gives us the following expression for the conditional output distribution −(|yk |2 +|xk |2 ) exp 2σ2 π ∗ ∗ ( (yk xk ) + ρ) cos(θk ) − (yk xk ) sin(θk ) p(yk |xk ) = 2 σ 2 I (ρ) exp dθk (2π) 0 −π σ2 (4.56) Because of the symmetry of the problem, we choose to work with polar coordinates, thus we define xk and yk by xk = ak eibk yk = αk eiβk 22
  • 31. CHAPTER 4. PHASE NOISE CHANNELS Then we have yk xk = aαei(b−β) ∗ (4.57) and ∗ (yk xk ) = aα cos (i(b − β)) (4.58) ∗ (yk xk ) = aα sin (i(b − β)) (4.59) Using both equations we define aα u= (4.60) σ2 and A = ( (yk xk ) + ρ)2 + ( (yk xk ))2 ∗ ∗ (4.61) which leads to −(a2 +α2 ) α exp 2σ2 π √ u cos(b − β) u sin(b − β) p(yk |xk ) = 2 σ 2 I (ρ) exp A cos(θk ) √ − sin(θk ) √ dθk (2π) 0 −π A A (4.62) We defined A such that A = (u cos(b − β))2 + (u sin(b − β))2 (4.63) so we can find z such that u cos(b − β) cos(z) = √ (4.64) A u sin(b − β) sin(z) = √ (4.65) A Then we have −(a2 +α2 ) α exp 2σ2 π √ p(yk |xk ) = 2 σ 2 I (ρ) exp A (cos(θk + z)) dθk (4.66) (2π) 0 −π which is equal to −(a2 +α2 ) α exp 2σ2 π √ p(yk |xk ) = exp A cos(θk ) dθk (4.67) (2π)2 σ 2 I0 (ρ) −π given that z and θk are independent. Finally, we recognize in (4.67) the Bessel function of the first kind which gives the following expression for the conditional output distribution 23
  • 32. CHAPTER 4. PHASE NOISE CHANNELS √ α α2 + a2 I0 ( A) pY |X (yk ,xk ) = exp − (4.68) 2πσ 2 2σ 2 I0 (ρ) To find an expression for the information density, we also need the output distribution PY . In this thesis, we consider a discrete input constellation with M codewords (c1 ,...,cM ) 1 and P (ci ) = M . Given this input, the output distribution can be computed as follows M 1 PY (yk ) = p (yk |ci ) (4.69) M Y |X i=1 and the information density is a2 √ exp − 2σ2 I0 ( A) i(x,y) = i(aeib ,αeiβ ) = log2 a2 √ (4.70) M 1 i=1 M exp − 2σ2 I0 ( Ai ) i where a2 α2 aα A= 4 + 2ρ 2 cos(b − β) + ρ2 (4.71) σ σ 4.3.2 Depending testing bound We compare two constellations for the AWGN channel with Tikhonov phase noise. We have the classic 64-QAM constellation and a robust circular QAM constellation designed specifically for this channel [8]. The second constellation is designed in order to minimize the average minimum distance between two points of the constellation. The algorithm presented in [8] gives an example of the constellation for a given phase noise. In Fig. 4.3 we plot both constellations and what happen to them through an AWGN channel with SNR = 30dB impaired by the given phase noise ρ = 625 (σph = 0.04). In Fig. 4.4 we plot the robust circular 64-QAM constellation impaired by a Tikhonov phase noise with parameter ρ = 625. In Fig. 4.5 we plot the DT curve and the constrained capacity for both constellations. We choose SNR = 0dB, ρ = 625 and Pe = 10−3 for this simulation. In Fig. 4.6 we plot the DT bound and the constrained capacity for both constella- tions. We choose SNR = 15dB, ρ = 625 and Pe = 10−3 for this simulation. In Fig. 4.7 we plot the DT bound for the robust circular 64-QAM constellation for two power of phase noise. We also plot the DT bound and the constrained capacity without phase noise, ie., Gaussian noise only, and both constrained and unconstrained capacity for this channel. We choose SNR = 15dB and Pe = 10−3 for this simulation. 24
  • 33. CHAPTER 4. PHASE NOISE CHANNELS 64−QAM constellation Robust Circular QAM constellation 2 2 1 1 0 0 −1 −1 −2 −2 −2 −1 0 1 2 −2 −1 0 1 2 64−QAM constellation with noise Robust Circular QAM constellation with noise 2 2 1 1 0 0 −1 −1 −2 −2 −2 −1 0 1 2 −2 −1 0 1 2 Figure 4.3: Two 64-QAM constellations in AWGN phase noise channel In Fig. 4.8 we plot the DT bound for the robust circular 64-QAM constellation for two probabilities of error. We also plot the constrained capacity for this channel. We choose SNR = 0dB and ρ = 100 for this simulation. 25
  • 34. CHAPTER 4. PHASE NOISE CHANNELS 1.5 1 0.5 0 −0.5 −1 −1.5 −1.5 −1 −0.5 0 0.5 1 1.5 Figure 4.4: Robust Circular QAM constellation with phase noise 1 0.9 0.8 0.7 Rate, bit/Ch.use 0.6 0.5 0.4 0.3 0.2 DT for Robust Circular 64−QAM DT for regular 64−QAM 0.1 Constrained−capacity for 64−QAM 0 0 200 400 600 800 1000 1200 1400 1600 1800 2000 Blocklength, n Figure 4.5: DT curves for two 64-QAM constellations in AWGN phase noise channel with SNR = 0dB 26
  • 35. CHAPTER 4. PHASE NOISE CHANNELS 5 4.5 4 3.5 Rate, bit/Ch.use 3 2.5 2 1.5 DT for Robust Circular 64−QAM 1 DT for regular 64−QAM 0.5 Constrained−capacity for Robust Circular 64−QAM Constrained−capacity for Regular 64−QAM 0 0 200 400 600 800 1000 1200 1400 1600 1800 2000 Blocklength, n Figure 4.6: DT curves for two 64-QAM constellations in AWGN phase noise channel with SNR = 15dB 5.5 5 4.5 4 3.5 Rate bit/Ch.use 3 2.5 2 1.5 DT for rho=100 1 DT for rho=625 DT and constrained−capacity without phase noise 0.5 Capacity 0 0 200 400 600 800 1000 1200 1400 1600 1800 2000 Blocklength, n Figure 4.7: Comparaison of DT curves for different phase noise power 27
  • 36. CHAPTER 4. PHASE NOISE CHANNELS 1 0.9 0.8 0.7 Rate, bit/Ch.use 0.6 0.5 0.4 0.3 0.2 DT with Pe=10E−3 DT with Pe=10E−9 0.1 Constrained−capacity 0 0 100 200 300 400 500 600 700 800 900 1000 Blocklength, n Figure 4.8: Comparaison of DT curves for different probabilities of error In Fig. 4.3 we see both constellations we used in our simulations. The robust circular 64-QAM has been designed for phase noise channels with a noise power ρ = 625. The optimization criteria used is the maximization of the minimal distance between two adjacent rings. In Fig. 4.5 we notice that both DT curves are the same for both constellations. Despite the differences between these constellations, the power of the Gaussian noise is too high to make a difference between these constellations. We can also notice that, even for a large block length (n = 2000), their is still a gap between the DT and the constrained capacity. In Fig. 4.6 we notice a difference between both constellations. For large block length (n ≤ 100) the robust circular 64-QAM performs better than the regular 64-QAM. We also notice that the DT bound and the constrained capacity are tight, which gives us a better approximation of the maximal coding rate for this channel. From these curves, we can also see that for high SNR we reach the capacity much faster than for small SNR. Then the gap between the DT bound and the constrained capacity is tighter for large SNR. In Fig. 4.7 we see the impact of the phase noise power on the DT bound. We present the loss of coding rate between two channels with different phase noise power. We also see that with the parameter ρ = 625, the maximal coding rate is very close to the coding rate without any phase noise. We also notice the loss induce by our constellation in regard to the capacity achieving distribution. In Fig. 4.8 we see the impact of the probability of error on the coding rate. We notice 28
  • 37. CHAPTER 4. PHASE NOISE CHANNELS that the difference between these curves exist for small block length, and we will need a larger block length to have a smaller probability of error. 4.3.3 Meta converse bound As we define earlier, we know the conditional output for our channel R R2 + r 2 I0 (ν) pY |X (r,φ,R,ψ) = exp − (4.72) 2πσ 2 2σ 2 I0 (ρ) where R2 r 2 Rr ν= + 2ρ 2 cos(φ − ψ) + ρ2 σ4 σ The meta converse bound requires to pick an output distribution. For our case, we use the following distribution, which is capacity achieving for high SNR [7]. R 2 ∼ χ1 2 (4.73) ψ ∼ U (−π,π) (4.74) 1 R2 PY (R,ψ) = √ exp − (4.75) 2π 2R2 Γ 1 2 2 Thus, we can define the information density given those two distributions N pY |X=xi (yi ) i(x,y) = log2 (4.76) PY (yi ) i=1   Ri R2 +r 2 I0 (νi ) N 2πσ2 exp − 2σ2 i i I0 (ρ) i(x,y) = log2  (4.77)   R2  i=1 √ 1 1 exp − 2i 2π 2R2 Γ( 2 ) i √ 1 N 2Γ 2 2 Ri + ri2 R2 i(x,y) = N log2 + log2 I0 (νi ) exp − − i (4.78) σ 2 I0 (ρ) 2σ 2 2 i=1 Then, we denote by Gn and Hn the information density under PY and PY |X respec- tively. To compute the converse bound, we have to find the parameter γn given the following condition P [Hn ≥ γn ] = 1 − (4.79) and then, we use this parameter to determine the following probability P [Gn ≥ γn ] (4.80) The main issue for this bound is the calculation of the probability P [Gn ≥ γn ]. We know that this value decreases exponentially to 0 and we do not have any closed- form expression to compute it. In the real-value Gaussian case, we found a closed-form expression using the chi-square distribution. 29
  • 38. 5 Conclusion In this work we applied an achievability bound on phase noise channels in order to determine the maximum coding rate for such channel. First, we focused on a simple model with a uniform phase noise. We managed to find a closed-form expression for the DT bound, but the computation complexity was an issue. Then we moved on to two partially coherent AWGN channels. For the AWGN channel impaired by uniform phase noise, a closed form expression has be found for the non coherent case which gave some results. And finally, we found some results for the AWGN channel impaired by a Tikhonov phase noise. We investigated the impact of all parameters (Noises power and probability of error) on the curves. Through both applications on phase noise channels we can see that the DT bound and the constraint capacity associated with a constellation are really close for high SNR. This give us a good idea of the achievable rate for a given block length and an error probability. We also investigated the impact of different power of phase noise and the loss induced on this rate. Moreover, we can see on the curves that for large block length (n > 500) and high SNR (SNR> 15), more than 95% of the constrained capacity is already achieved. Given those informations, we can evaluate the performances of codes and discuss the interest of using larger block. For small SNR, the gap between the achievability bound and the constrained capacity is still large, therefore, we do not have a tight approximation of the maximal coding rate. As a future work for our thesis, we can study the meta-converse and try to find an approximation for the binary hypothesis testing. In that way we could compute the upper bound and have a tighter approximation. Another following of this thesis could be to investigate all the codes we already have and see their performances over PC-AWGN channels. 30
  • 39. Bibliography [1] C. E. Shannon, A mathematical theory of communication, Bell System Technical Journal (1948) 379–423. [2] A. Feinstein, A new basic theorem of information theory, IRE trans. Inform. Theory (1954) pp. 2–22. [3] C. E. Shannon, Certain results in coding theory for noisy channels, Inf. Contr., vol. 1 (1957) pp. 6–25. [4] R. G. Gallager, A simple derivation of the coding theorem and some applications, IEEE Trans. Inf. Theory, vol. 40 (1965) pp.3–18. [5] Y. Polyanskiy, H. V. Poor, S. Verdu, Channel coding rate in the finite blocklength regime, IEEE Trans. Inf. Theory. [6] T. Cover, J. Thomas, Elements of Information Theory, Wiley, 2006. [7] A. Lapidoth, On phase noise channels at high snr, IEEE Trans. Inf. Theory. [8] A. Papadopoulos, K. N. Pappi, G. K. Karagiannidis, H. Mehrpouyan, Robust circular qam constellations in the presence of phase noise. 31