Abstract

We consider the convergence analysis of the sign algorithm for adaptive filtering when the input processes are uncorrelated and Gaussian and a fixed step size /spl mu/>0 is used. Exact recursive equations for the covariance matrix of the deviation error are established for any step size /spl mu/>0. Asymptotic time-averaged convergence for the mean-absolute deviation error, mean-square deviation error, and for the signal mean-square estimation error are established. These results are shown to hold for arbitrary step size /spl mu/>0. >

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call