Abstract
The field of data-driven, neural-network-based machine learning (ML) has seen significant growth, with applications in various information and control systems. Despite promising real-world uses, the reliability of models remains questionable. Conventionally, reliability is assessed based on predictive fidelity, accuracy, and training effectiveness; however, quality developmental procedures and excellent training performance metrics do not guarantee operational reliability. Instead, an ML model's predictive performance depends on the training set's representativeness to the intended operational space. It is known that ML algorithms excel at interpolation but struggle with extrapolation tasks. Anomalies and feature drift can also reduce operational performance. Determining whether a new sample is an interpolation or extrapolation task involves out-of-distribution (OOD) detection for assessing its proximity to the existing training data. Thus, we present a real-time, model-agnostic individual prediction reliability evaluation method called Data Auditing for Reliability Evaluation (DARE) for applying OOD detection to the training dataset. We demonstrate on a feedforward neural network ML-integrated digital twin for predicting fuel centerline temperatures during loss-of-flow transients. DARE acts as a “data supervisor” in determining the model's applicability under different operating conditions. In this manner, we demonstrate how training data can serve as inductive evidence to support the reliability of ML predictions.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have