Abstract

A new algorithm, the Partial Correlation Statistic (PCS) algorithm, is presented for structure learning under linear Structural Equation Models. The PCS algorithm can deal with continuous data following linear arbitrary distribution rather than only a Gaussian distribution. This paper makes two specific contributions. First, for linear arbitrarily distributed datasets, which are generated by the linear structural equation models, if the sample size is sufficiently large, partial correlation coefficient statistic is proved to follow a Student's t-distribution. Second, the PCS algorithm combines hypothesis testing of partial correlation statistic and local learning to select potential neighbors of the target node. This significantly reduces the search space and achieves good time performance. The PCS algorithm does not need to choose optimal threshold of partial correlation by large amount of experiments. Especially, the PCS algorithm redefines the relevance from statistic theory and measure the relevance of the variables based on $p$ -value. The effectiveness of the algorithm is compared with current state of the art methods on seven networks. A simulation shows that the PCS algorithm outperforms existing algorithms in terms of both accuracy and time performance on average.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.