Skip to yearly menu bar Skip to main content


Poster

Learning-Augmented Hierarchical Clustering

Vladimir Braverman · Jon C. Ergun · Chen Wang · Samson Zhou

East Exhibition Hall A-B #E-3600
[ ] [ ]
Wed 16 Jul 11 a.m. PDT — 1:30 p.m. PDT

Abstract:

Hierarchical clustering (HC) is an important data analysis technique in which the goal is to recursively partition a dataset into a tree-like structure while grouping together similar data points at each level of granularity. Unfortunately, for many of the proposed HC objectives, there exist strong barriers to approximation algorithms with the hardness of approximation. We consider the problem of hierarchical clustering given auxiliary information from natural oracles in the learning-augmented framework. Our main results are algorithms that, given learning-augmented oracles, compute efficient approximate HC trees for the celebrated Dasgupta's and Moseley-Wang objectives that overcome known hardness barriers.

Lay Summary:

We studied the hierarchical clustering (HC) problem, which aims to summarize data into increasingly precise sub-clusters based on similarity and dissimilarity. The optimization of hierarchical clustering is a major problem in machine learning with applications in artificial intelligence and medical science. Since the problem is hard in general, we considered a learning-augmented version that takes advantage of a (potentially erroneous) oracle obtained from machine learning models.This oracle can provide some 'hints' to the HC algorithm in the hope of getting improved results.Our paper showed that it is indeed possible to obtain such improved HC algorithms with the learning-augmented oracle. Our results shed light on advancing the optimization of HC using machine learning models, and these algorithms have the potential to be applied to a broad range of applications, e.g., in medical modeling.

Chat is not available.