2. CONTENTS
What Is Noise And Noise Cancellation?
Adaptive Filter
Basic Adaptive Filters
Applications Of Adaptive Filters
Problem Statement
Various Adaptive Algorithms For Noise Cancellation
LMS Algorithm
NLMS Algorithm
RLS Algorithm
Affine Projection Algorithm
SNRI Table
Outputs
Comparison
Conclusion
References
3. WHAT IS NOISE AND NOISE
CANCELLATION?
•Noise consists of unwanted waveforms that can
interfere with communication.
• Noise can be internal or external to the system.
•Sound noise: interferes with your normal hearing
•Colored Noise
•Impulsive Noise
•White noise (AWGN)
•NOISE CANCELLATION: Noise cancellation is
a method to reduce or cancel out undesirable
components of the signal
4. ADAPTIVE FILTERS
A filter which adapts itself to the input signal given to it.
It is Non Linear And Time Variant.
Best suited when signal conditions are slowly changing.
Relies on recursive algorithm.
It has Adaptation algorithm for adjusting parameters for
improved performance
It is meant to monitor the environment and varies the
filter transfer function accordingly.
5. CONTINUED…
The basic operation of adaptive
filter involves two processes :
Filtering process
•produces an output signal in response to a given input signal.
Adaptation process
•aims to adjust the filter parameters to the environment.
7. Applications of Adaptive
Filters:
NOISE
CANCELLATION
:Subtracts Noise
from Received
Signal adaptively
to improve SNR.
SIGNAL
PREDICTION :
Used to provide a
prediction of the
present value of a
random signal
SYSTEM
IDENTIFICATION:
Designs an
adaptive filter
that provides an
approximation for
an unknown
system.
ECHO
CANCELLATION:
Used to cancel
unknown
interference from
a primary signal
8. VARIOUS ADAPTIVE ALGORITHMS FOR
NOISE CANCELLATION
Properties of an ideal algorithm:
Practical to implement
Adapt to coefficient quickly to minimize error
Provide the Desired Performance
Different algorithms used are:
Least Mean Squares (LMS) algorithm
The Normalized Least Mean Squares(NLMS) algorithm
The Recursive Least Squares (RLS) algorithm
Affine Projection Algorithm(APA)
9. PROBLEM STATEMENT
We have taken an input random signal of N samples as reference
signal.
We have taken random noise or adaptive Gaussian noise.
Then we are adding the noise signal with the input signal.
So the problem is to extract the input signal from output signal by
eliminating the noise.
10. LMS ALGORITHM:
Adjusts the weight w(n) of the filter
Adaptively adjusts the filter tap weights according to equation:
w(n+1)=w(n)+µe(n)x(n)
Acts as negative feedback to minimize error signal
It is robust in nature.
Slow in convergence and sensitive to variations in step size
parameter.
Requires number of iterations equals to dimensionality of the
input.
13. ADVANTAGES AND DISADVANTAGES:
• It is simple in implementation.
• Stable and robust performance
against different signal condition.
ADVANTAGES
•Slow Convergence.DISADVANTAGES
16. NLMS ALGORITHM:-
• In structural terms both NLMS filter is exactly same as a
standard LMS filter.
• From one iteration to the next, the weight of an adaptive filter
should be changed in a minimal manner.
17. NLMS CONTINUED..
One of the drawback of LMS is selection of step size
parameter µ.
In order to solve this difficulty, we can use the NLMS
(Normalized Least Mean Square) algorithm. Here the step size
parameter is normalised.
So the NLMS algorithm is a time varying step-size algorithm,
calculating the convergence factor μ as :
18. NLMS PARAMETERS:
Where || x(n) ||² is the squared Euclidean norm of the input
x(n).
Here α is the adaption constant, which optimizes the
convergence rate of the algorithm
Range of alpha is: 0<α<2
c is the constant term for normalization and always c<1.
The updated filter weight is:
19. ADVANTAGES AND DISADVANTAGES:
• As here µ is normalized this
algorithm converges faster than
LMS.
• Here estimated error value between
the desired signal and filter output
is less than LMS.
ADVANTAGES
•But LMS is less complex than
NLMS and more stable.DISADVANTAGES
21. RLS ALGORITHM
Recursively finds the filter coefficients that minimize a
weighted linear least squares cost function relating to the input
signals.
In this algorithm the filter tap weight vector is updated
using:
22. CONTINUED…
Whitens the input data by using inverse correlation
matrix of data.
The Cost function C(n) should be minimized.
C(n)=
e(i)=d(i)-wH(n)u(i)
where,
β(n,i) is weighting vector
0<β(n,i)<=1 i=1,2,3……,n
β(n,i)=λn-i , where λ=forgetting factor
24. CONTINUED…
Let Φ(n) is the correlation matrix of input u(i)
Φ(n)= λn-i u(i) uH(i)+ δλnI
Then the average cross correlation vector z(n) is given by-
z(n)=Φ(n)ŵ(n) , n=1,2………
Using the matrix inversion lemma, we can find the inverse of correlation
matrix,
Φ-1(n) = P(n) (let)
Cost function is always expressed in terms of gain where ,K(n) is the
gain vector.
k(n) = P(n)u(n) = Φ-1(n)u(n)
25. CONTINUED…
The tap weight vector ŵ(n)
ŵ(n)= Φ-1(n)z(n)
From the above equations, we summarize the RLS Algorithm as-
k(n) =
π(n) = P(n-1)u(n)
ξ(n) = d(n) – ŵH(n-1)u(n)
ŵ(n) = ŵ(n-1) + k(n)ξ*(n)
P(n) = λ-1 P(n-1) – λ-1 k(n) uH(n) P(n-1)
26. ADVANTAGES AND DISADVANTAGES OF
RLS
•RLS converges faster than
LMS, NLMS and APA.
•Its noise cancellation
capacity is the most.
ADVANTAGES
•This is the most complex
algorithm of all the four
algorithms.
DISADVANTAGES
28. AFFINE PROJECTION ALGORITHM
Generalization of the well known normalized least mean
square (NLMS) adaptive filtering algorithm.
Fast convergence compared to NLMS.
Computational complexity increases.
Convergence gets better with increase in filter order N.
Faster tracking capabilities than NLMS.
Better performance in steady state mean square error (MSE)
than other algorithms.
29. APA MATHEMATICAL
IMPLEMENTATION…
A(n) = input data matrix [N*N]
AH(n) =input data matrix in hermitian transpose[N*N]
d(n)= desired response [N*1]
Error can be computed as-
e(n)=d(n)-A(n)ŵ(n)
The updated tap weight vector can be calculated as-
ŵ(n+1)=ŵ(n)+μ AH(n)(A(n) AH(n))-1e(n)
30. CONVERGENCE & STABILITY OF
APA
The learning curve of an APA consists of the sum of exponential
terms.
It converges at a rate faster than that of a NLMS filter.
As more delayed versions of tap input vector is used, the rate of
convergence improves, so does the computational complexity.
APA is less stable than LMS and NLMS algorithms, whereas it
is more stable than RLS algorithm.
35. COMPARISON OF
LMS,NLMS,APA AND RLS-----
• RLS converges faster than APA,
APA converges faster than NLMS and
NLMS converges faster than LMS.
CONVERGENCE:
• RLS is the most complex algorithm
among the four algorithms. Hence ,
complexity is inversely proportional to
convergence.
COMPLEXITY:
• Difference between final and initial
SNR is highest in case of RLS then
APA then NLMS then LMS.
SNR
IMPROVEMENT:
36. CONCLUSION
We studied the behavior of LMS, NLMS, APA and RLS algorithms
by implementing them in the adaptive filter for noise cancellation.
LMS was the simplest and easiest to implement but it converges at
the slowest rate.
NLMS has a normalized step size making it converge faster than
LMS but complexity also increases along with convergence rate.
APA is the improved version of NLMS with increasing
convergence rate.
37. CONTINUED….
RLS is the fastest converging algorithm with maximum
computational complexity. But it cancels maximum noise by
minimizing error with the rapidest rate.
So we are making a tradeoff between computational complexity and
convergence rate here to get the most noise free signal.
RLS is the best algorithm as it is faster than the other three.
38. REFERENCES
Adaptive Filter Theory by Simon Haykin: 3rd edition, Pearson Education
Asia.LPE.
Adaptive Signal Processing by John G Proakis, 3rd edition, Perntice Hall
of India.
B. Widow, "Adaptive noise canceling: principles and applications",
Proceedings of the IEEE, vol. 63, pp. 1692-1716, 1975.
A Family of Adaptive Filter Algorithms in Noise cancellation for Speech
Enhancement By Sayed. A. Hadei, Student Member IEEE and M. lotfizad.
Steven L. Gay and Sanjeev Tavathia, “The Fast Affine Projection
Algorithm”, Acoustics Research Department, AT&T Bell Laboratories.
Sundar G. Sankaran, Student Member, IEEE, and A. A. (Louis) Beex,
Senior Member, IEEE” Convergence Behavior of Affine Projection
Algorithms”.