Abstract

We present a technique for translating a black-box machine-learned classifier operating on a high-dimensional input space into a small set of human-interpretable observables that can be combined to make the same classification decisions. We iteratively select these observables from a large space of high-level discriminants by finding those with the highest decision similarity relative to the black box, quantified via a metric we introduce that evaluates the relative ordering of pairs of inputs. Successive iterations focus only on the subset of input pairs that are misordered by the current set of observables. This method enables simplification of the machine-learning strategy, interpretation of the results in terms of well-understood physical concepts, validation of the physical model, and the potential for new insights into the nature of the problem itself. As a demonstration, we apply our approach to the benchmark task of jet classification in collider physics, where a convolutional neural network acting on calorimeter jet images outperforms a set of six well-known jet substructure observables. Our method maps the convolutional neural network into a set of observables called energy flow polynomials, and it closes the performance gap by identifying a class of observables with an interesting physical interpretation that has been previously overlooked in the jet substructure literature.

Highlights

  • It is widely appreciated that neural networks (NNs) and related machine-learning (ML) tools can provide powerful solutions to important and difficult problems in high-energy physics [1,2]

  • There are many such metrics one could use, but we introduce the average decision ordering (ADO), in part because it shares the conceptual simplicity of the area under the curve (AUC) metric often used to benchmark classifiers against ground truth

  • We have proposed a new technique for mapping an ML solution into a space of human-interpretable observables

Read more

Summary

INTRODUCTION

It is widely appreciated that neural networks (NNs) and related machine-learning (ML) tools can provide powerful solutions to important and difficult problems in high-energy physics [1,2]. We present a technique for translating a black-box ML strategy based on low-level inputs into a human-readable space of HL observables. Our mapping strategy suggests a new approach to the application of deep learning to high-energy physics data In this approach, training a powerful deep neural network (DNN) on low-level inputs is just the first step, which helps gauge the effective upper limit on possible ML performance and determine asymptotically optimal decision boundaries. The new second step is translating as much of the ML strategy as possible to a well-understood set of HL observables This allows for physical interpretation of the information being used, validation of the modeling, definition of reasonable systematic uncertainties, as well as computational benefits due to dimensionality reduction.

TRANSLATING FROM MACHINE TO HUMAN
Average decision ordering
Black-box guided search strategy
A CASE STUDY IN JET SUBSTRUCTURE
Boosted boson classification
Energy flow polynomials
Black-box guiding
Physics interpretation
ITERATIVELY MAPPING FROM MINIMAL FEATURES
Comparison to brute force search
Comparison to truth-label guiding
DISCUSSION
Baseline convolutional neural network
Baseline dense neural network
Findings
K-fold validation
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call