Abstract

A Finite State Machine (FSM) is a mathematical model of computation which can effectively model a sequence of words or tokens. A grammar representing a collection of tokens in a finite alphabet might contain regularities that are not fully captured by a deterministic formal grammar. Therefore, the simple FSM model is extended to include some probabilistic structure in the grammar which is now termed as Probabilistic Finite State Machine (PFSM). We extend earlier work on inferring PFSMs using the Bayesian informationtheoretic Minimum Message Length (MML) principle to the case of inferring hierarchical PFSMs (HPFSMs). HPFSMs consist of an outer PFSM whose states can internally contain a PFSM (or, recursively, an HPFSM). The alphabet of each such internally contained PFSM can be smaller than the complete HPFSM. HPFSMs can often represent the behaviour of a PFSM more concisely, and MML's ability to deal with both discrete structures and continuous probabilities renders MML well suited to this more general inference. We empirically compare on pseudo-random data-sets.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.