Abstract

We consider biclustering that clusters both samples and features and propose efficient convex biclustering procedures. The convex biclustering algorithm (COBRA) procedure solves twice the standard convex clustering problem that contains a non-differentiable function optimization. We instead convert the original optimization problem to a differentiable one and improve another approach based on the augmented Lagrangian method (ALM). Our proposed method combines the basic procedures in the ALM with the accelerated gradient descent method (Nesterov’s accelerated gradient method), which can attain O(1/k2) convergence rate. It only uses first-order gradient information, and the efficiency is not influenced by the tuning parameter λ so much. This advantage allows users to quickly iterate among the various tuning parameters λ and explore the resulting changes in the biclustering solutions. The numerical experiments demonstrate that our proposed method has high accuracy and is much faster than the currently known algorithms, even for large-scale problems.

Highlights

  • By clustering, such as k-means clustering [1] and hierarchical clustering [2,3], we usually mean dividing N samples, each consisting of p covariate values, into several categories, where N, p ≥ 1.In this paper, we consider biclustering [4] that is an extended notion of clustering

  • This paper proposes an efficient algorithm with simple subproblems and a fast convergence rate to solve the convex biclustering problem

  • We proposed a method to find a solution to the convex biclustering problem

Read more

Summary

Introduction

By clustering, such as k-means clustering [1] and hierarchical clustering [2,3], we usually mean dividing N samples, each consisting of p covariate values, into several categories, where N, p ≥ 1. The main challenge of solving the optimization problem is the two fused penalty terms: indecomposable and non-differentiable. It is necessary to propose an efficient algorithm to solve the convex biclustering problem. This paper proposes an efficient algorithm with simple subproblems and a fast convergence rate to solve the convex biclustering problem. Rather than update each variable alternately, like ADMM, we use the augmented Lagrangian method (ALM) to update the primal variables simultaneously In this way, we can transform the optimization problem to be differentiable, solve the problem via an efficient gradient descent method and further simplify the subproblems. Our proposed method does not require as much computation, even when the tuning parameter λ is large, as the existing approaches do, which means that it is easier to obtain biclustering results simultaneously for several λ values. Rp×q, we use ||X||F ∑ip=1 ∑qj=1 to denote |xij| if not the Frobenius specified

Preliminaries
ADMM and ALM
Nesterov’s Accelerated Gradient Method
Convex Biclustering
The Proposed Method and Theoretical Analysis
The ALM Formulation
The Proposed Method
4: Update iterate
Lipschitz Constant and Convergence Rate
Experiments
Artificial Data Analysis
Objective function error:
Comparisons
Assessment
Real Data Analysis
Findings
Discussion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call