Abstract
The problem of differentiating a function with bounded second derivative in the presence of bounded measurement noise is considered in both continuous-time and sampled-data settings. Fundamental performance limitations of causal differentiators, in terms of the smallest achievable worst-case differentiation error, are shown. A robust exact differentiator is then constructed via the adaptation of a single parameter of a linear differentiator. It is demonstrated that the resulting differentiator exhibits a combination of properties that outperforms existing continuous-time differentiators: it is robust with respect to noise, it instantaneously converges to the exact derivative in the absence of noise, and it attains the smallest possible—hence optimal—upper bound on its differentiation error under noisy measurements. For sample-based differentiators, the concept of quasi-exactness is introduced to classify differentiators that achieve the lowest possible worst-case error based on sampled measurements in the absence of noise. A straightforward sample-based implementation of the proposed linear adaptive continuous-time differentiator is shown to achieve quasi-exactness after a single sampling step as well as a theoretically optimal differentiation error bound that, in addition, converges to the continuous-time optimal one as the sampling period becomes arbitrarily small. A numerical simulation illustrates the presented formal results.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.