Abstract

There is no uniform approach in the literature for modelling sequential correlations in sequence classification problems. It is easy to find examples of unstructured models (e.g. logistic regression) where correlations are not taken into account at all, but there are also many examples where the correlations are explicitly incorporated into a--potentially computationally expensive--structured classification model (e.g. conditional random fields). In this paper we lay theoretical and empirical foundations for clarifying the types of problem which necessitate direct modelling of correlations in sequences, and the types of problem where unstructured models that capture sequential aspects solely through features are sufficient. The theoretical work in this paper shows that the rate of decay of auto-correlations within a sequence is related to the excess classification risk that is incurred by ignoring the structural aspect of the data. This is an intuitively appealing result, demonstrating the intimate link between the auto-correlations and excess classification risk. Drawing directly on this theory, we develop well-founded visual analytics tools that can be applied a priori on data sequences and we demonstrate how these tools can guide practitioners in specifying feature representations based on auto-correlation profiles. Empirical analysis is performed on three sequential datasets. With baseline feature templates, structured and unstructured models achieve similar performance, indicating no initial preference for either model. We then apply the visual analytics tools to the datasets, and show that classification performance in all cases is improved over baseline results when our tools are involved in defining feature representations.

Highlights

  • Structure modelling permits target variables to collaborate so that ‘informed’ decisions about a set of random variables are based on a collection of beliefs linked together in a graphical structure (Lafferty et al 2001; Sutton and McCallum 2011)

  • To help us understand the use of Logistic Regression (LR) for sequence prediction, we show in Theorem 1 that given certain conditions on transition potentials of Conditional Random Field (CRF), unconditional independence can be proved between adjacent nodes

  • Our first experiments assess the difference in classification performance between LR and CRF models over the Word Hyphenation (WH), Activity Recognition (AR) and Occasionally Dishonest Casino (ODC) datasets

Read more

Summary

Introduction

Structure modelling permits target variables to collaborate so that ‘informed’ decisions about a set of random variables are based on a collection of beliefs linked together in a graphical structure (Lafferty et al 2001; Sutton and McCallum 2011). In such frameworks, instances can be a list of vectors each relating to a single target variable in the graph. Marginal distributions in a structured model, are explicitly influenced by all possible target permutations over the graph This can be expensive to compute, but, in some applications, superior classification performance admonishes time complexity. The abandonment of structure might be considered sub-optimal for many of these applications, yet some are considered ‘solved’ with the unstructured model choice

Objectives
Methods
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.