Abstract

Graphs provide powerful representations for statistical modeling of interrelated variables (observed or latent) in a broad range of machine learning applications. Examples include learning and inference based on the dependency structures among words, documents, topics, users, items, web sites, and more. How to best leverage such dependency structures from multiple graphs with massive and heterogeneous types of nodes and relations has posed grand challenges to machine learning theory and algorithms. This talk presents our recent work in this direction focusing on three significant tasks, including 1) a novel framework for fusing multiple heterogeneous graphs into a unified product graph to enable semi-supervised multi-relational learning, 2) the first algorithmic solution for imposing analogical structures in graph-based entity/relation embedding, and 3) a new formulation of neural architecture search as a graph topology optimization problem, with simple yet powerful algorithms that automatically discover high-performing convolutional neural architectures on image recognition benchmarks, and reduce the computational cost over state-of-the-art non-differentiable techniques by several orders of magnitude.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call