Slideshare uses cookies to improve functionality and performance, and to provide you with relevant advertising. If you continue browsing the site, you agree to the use of cookies on this website. See our User Agreement and Privacy Policy.

Slideshare uses cookies to improve functionality and performance, and to provide you with relevant advertising. If you continue browsing the site, you agree to the use of cookies on this website. See our Privacy Policy and User Agreement for details.

Successfully reported this slideshow.

17,134 views

Published on

No Downloads

Total views

17,134

On SlideShare

0

From Embeds

0

Number of Embeds

27

Shares

0

Downloads

1,405

Comments

7

Likes

31

No notes for slide

- 1. Sweta Mohanty -1011016060 Anwesha Samal -1011016057 Brati Sundar Nanda -1011016238 Abhilash Mishra -1011016237 Guided By:- P.SHIVANI SAHOO
- 2. CONTENTS What Is Noise And Noise Cancellation? Adaptive Filter Basic Adaptive Filters Applications Of Adaptive Filters Problem Statement Various Adaptive Algorithms For Noise Cancellation LMS Algorithm NLMS Algorithm RLS Algorithm Affine Projection Algorithm SNRI Table Outputs Comparison Conclusion References
- 3. WHAT IS NOISE AND NOISE CANCELLATION? •Noise consists of unwanted waveforms that can interfere with communication. • Noise can be internal or external to the system. •Sound noise: interferes with your normal hearing •Colored Noise •Impulsive Noise •White noise (AWGN) •NOISE CANCELLATION: Noise cancellation is a method to reduce or cancel out undesirable components of the signal
- 4. ADAPTIVE FILTERS A filter which adapts itself to the input signal given to it. It is Non Linear And Time Variant. Best suited when signal conditions are slowly changing. Relies on recursive algorithm. It has Adaptation algorithm for adjusting parameters for improved performance It is meant to monitor the environment and varies the filter transfer function accordingly.
- 5. CONTINUED… The basic operation of adaptive filter involves two processes : Filtering process •produces an output signal in response to a given input signal. Adaptation process •aims to adjust the filter parameters to the environment.
- 6. BASIC ADAPTIVE FILTER It contains 4 signals: Reference Signal [x(n)] Input Signal[d(n)] Filter output Signal[y(n)] Error Signal[e(n)]
- 7. Applications of Adaptive Filters: NOISE CANCELLATION :Subtracts Noise from Received Signal adaptively to improve SNR. SIGNAL PREDICTION : Used to provide a prediction of the present value of a random signal SYSTEM IDENTIFICATION: Designs an adaptive filter that provides an approximation for an unknown system. ECHO CANCELLATION: Used to cancel unknown interference from a primary signal
- 8. VARIOUS ADAPTIVE ALGORITHMS FOR NOISE CANCELLATION Properties of an ideal algorithm: Practical to implement Adapt to coefficient quickly to minimize error Provide the Desired Performance Different algorithms used are: Least Mean Squares (LMS) algorithm The Normalized Least Mean Squares(NLMS) algorithm The Recursive Least Squares (RLS) algorithm Affine Projection Algorithm(APA)
- 9. PROBLEM STATEMENT We have taken an input random signal of N samples as reference signal. We have taken random noise or adaptive Gaussian noise. Then we are adding the noise signal with the input signal. So the problem is to extract the input signal from output signal by eliminating the noise.
- 10. LMS ALGORITHM: Adjusts the weight w(n) of the filter Adaptively adjusts the filter tap weights according to equation: w(n+1)=w(n)+µe(n)x(n) Acts as negative feedback to minimize error signal It is robust in nature. Slow in convergence and sensitive to variations in step size parameter. Requires number of iterations equals to dimensionality of the input.
- 11. LMS ADAPTIVE FILTER BLOCK DIAGRAM
- 12. LMS ALGORITHM STEPS: nenxnwnw 1 nyndne 1 0 ][ N n nwnxny Filter output : Estimation error: Tap-weight adaptation: •Each iteration of LMS involves three steps: STABILITY: •Condition for stability is: •Larger values for step size:- •Increases adaptation rate (faster adaptation) •Increases residual mean-squared error powersignalinput 2 0
- 13. ADVANTAGES AND DISADVANTAGES: • It is simple in implementation. • Stable and robust performance against different signal condition. ADVANTAGES •Slow Convergence.DISADVANTAGES
- 14. OUTPUT: MEAN SQUARE ERROR FOR LMS-----
- 15. VARIATION OF MSE WITH RESPECT TO Μ:
- 16. NLMS ALGORITHM:- • In structural terms both NLMS filter is exactly same as a standard LMS filter. • From one iteration to the next, the weight of an adaptive filter should be changed in a minimal manner.
- 17. NLMS CONTINUED.. One of the drawback of LMS is selection of step size parameter µ. In order to solve this difficulty, we can use the NLMS (Normalized Least Mean Square) algorithm. Here the step size parameter is normalised. So the NLMS algorithm is a time varying step-size algorithm, calculating the convergence factor μ as :
- 18. NLMS PARAMETERS: Where || x(n) ||² is the squared Euclidean norm of the input x(n). Here α is the adaption constant, which optimizes the convergence rate of the algorithm Range of alpha is: 0<α<2 c is the constant term for normalization and always c<1. The updated filter weight is:
- 19. ADVANTAGES AND DISADVANTAGES: • As here µ is normalized this algorithm converges faster than LMS. • Here estimated error value between the desired signal and filter output is less than LMS. ADVANTAGES •But LMS is less complex than NLMS and more stable.DISADVANTAGES
- 20. OUTPUT: MEAN SQUARE ERROR FOR NLMS-----
- 21. RLS ALGORITHM Recursively finds the filter coefficients that minimize a weighted linear least squares cost function relating to the input signals. In this algorithm the filter tap weight vector is updated using:
- 22. CONTINUED… Whitens the input data by using inverse correlation matrix of data. The Cost function C(n) should be minimized. C(n)= e(i)=d(i)-wH(n)u(i) where, β(n,i) is weighting vector 0<β(n,i)<=1 i=1,2,3……,n β(n,i)=λn-i , where λ=forgetting factor
- 23. CONTINUED… REGULARISATION: C(n)= The sum of weighted error squares: A regularizing term:
- 24. CONTINUED… Let Φ(n) is the correlation matrix of input u(i) Φ(n)= λn-i u(i) uH(i)+ δλnI Then the average cross correlation vector z(n) is given by- z(n)=Φ(n)ŵ(n) , n=1,2……… Using the matrix inversion lemma, we can find the inverse of correlation matrix, Φ-1(n) = P(n) (let) Cost function is always expressed in terms of gain where ,K(n) is the gain vector. k(n) = P(n)u(n) = Φ-1(n)u(n)
- 25. CONTINUED… The tap weight vector ŵ(n) ŵ(n)= Φ-1(n)z(n) From the above equations, we summarize the RLS Algorithm as- k(n) = π(n) = P(n-1)u(n) ξ(n) = d(n) – ŵH(n-1)u(n) ŵ(n) = ŵ(n-1) + k(n)ξ*(n) P(n) = λ-1 P(n-1) – λ-1 k(n) uH(n) P(n-1)
- 26. ADVANTAGES AND DISADVANTAGES OF RLS •RLS converges faster than LMS, NLMS and APA. •Its noise cancellation capacity is the most. ADVANTAGES •This is the most complex algorithm of all the four algorithms. DISADVANTAGES
- 27. MEAN SQUARE ERROR FOR RLS-----
- 28. AFFINE PROJECTION ALGORITHM Generalization of the well known normalized least mean square (NLMS) adaptive filtering algorithm. Fast convergence compared to NLMS. Computational complexity increases. Convergence gets better with increase in filter order N. Faster tracking capabilities than NLMS. Better performance in steady state mean square error (MSE) than other algorithms.
- 29. APA MATHEMATICAL IMPLEMENTATION… A(n) = input data matrix [N*N] AH(n) =input data matrix in hermitian transpose[N*N] d(n)= desired response [N*1] Error can be computed as- e(n)=d(n)-A(n)ŵ(n) The updated tap weight vector can be calculated as- ŵ(n+1)=ŵ(n)+μ AH(n)(A(n) AH(n))-1e(n)
- 30. CONVERGENCE & STABILITY OF APA The learning curve of an APA consists of the sum of exponential terms. It converges at a rate faster than that of a NLMS filter. As more delayed versions of tap input vector is used, the rate of convergence improves, so does the computational complexity. APA is less stable than LMS and NLMS algorithms, whereas it is more stable than RLS algorithm.
- 31. OUTPUT: MEAN SQUARE ERROR FOR APA-
- 32. SNRI TABLE:- •Signal to Noise ratio improvement= Final SNR-Original SNR ALGORITHM SNRI LMS 13.69 NLMS 18.009 APA 20.39 RLS 29.09
- 33. COMPARISION FOR CONVERGENCE FOR DIFFERENT ALGORITHMS:
- 34. COMPARISON OF MSE FOR DIFFERENT ALGORITHMS:
- 35. COMPARISON OF LMS,NLMS,APA AND RLS----- • RLS converges faster than APA, APA converges faster than NLMS and NLMS converges faster than LMS. CONVERGENCE: • RLS is the most complex algorithm among the four algorithms. Hence , complexity is inversely proportional to convergence. COMPLEXITY: • Difference between final and initial SNR is highest in case of RLS then APA then NLMS then LMS. SNR IMPROVEMENT:
- 36. CONCLUSION We studied the behavior of LMS, NLMS, APA and RLS algorithms by implementing them in the adaptive filter for noise cancellation. LMS was the simplest and easiest to implement but it converges at the slowest rate. NLMS has a normalized step size making it converge faster than LMS but complexity also increases along with convergence rate. APA is the improved version of NLMS with increasing convergence rate.
- 37. CONTINUED…. RLS is the fastest converging algorithm with maximum computational complexity. But it cancels maximum noise by minimizing error with the rapidest rate. So we are making a tradeoff between computational complexity and convergence rate here to get the most noise free signal. RLS is the best algorithm as it is faster than the other three.
- 38. REFERENCES Adaptive Filter Theory by Simon Haykin: 3rd edition, Pearson Education Asia.LPE. Adaptive Signal Processing by John G Proakis, 3rd edition, Perntice Hall of India. B. Widow, "Adaptive noise canceling: principles and applications", Proceedings of the IEEE, vol. 63, pp. 1692-1716, 1975. A Family of Adaptive Filter Algorithms in Noise cancellation for Speech Enhancement By Sayed. A. Hadei, Student Member IEEE and M. lotfizad. Steven L. Gay and Sanjeev Tavathia, “The Fast Affine Projection Algorithm”, Acoustics Research Department, AT&T Bell Laboratories. Sundar G. Sankaran, Student Member, IEEE, and A. A. (Louis) Beex, Senior Member, IEEE” Convergence Behavior of Affine Projection Algorithms”.

No public clipboards found for this slide

Login to see the comments