Abstract
A numerical algorithm is presented for the purpose of reducing noise from a discretely sampled input signal where the underlying signal of interest has a broadband spectrum. It is designed to be useful even if the clean signal is contaminated with 100% or more noise (signal to noise ratio less than or equal to zero). The method is based on time delay embedding using coordinates generated by local low-pass filtering, which we call a low-pass embedding. The singular value decomposition is then used locally in embedding space to distinguish between the dynamics and the noise. The algorithm is evaluated for chaotic signals generated by the Lorenz and Rössler systems, to which Gaussian white noise has been added.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.