Fast Computation of Content-Sensitive Superpixels and Supervoxels Using Q-Distances

Zipeng Ye, Ran Yi, Minjing Yu, Yong-Jin Liu, Ying He; The IEEE International Conference on Computer Vision (ICCV), 2019, pp. 3770-3779


State-of-the-art researches model the data of images and videos as low-dimensional manifolds and generate superpixels/supervoxels in a content-sensitive way, which is achieved by computing geodesic centroidal Voronoi tessellation (GCVT) on manifolds. However, computing exact GCVTs is slow due to computationally expensive geodesic distances. In this paper, we propose a much faster queue-based graph distance (called q-distance). Our key idea is that for manifold regions in which q-distances are different from geodesic distances, GCVT is prone to placing more generators in them, and therefore after few iterations, the q-distance-induced tessellation is an exact GCVT. This idea works well in practice and we also prove it theoretically under moderate assumption. Our method is simple and easy to implement. It runs 6-8 times faster than state-of-the-art GCVT computation, and has an optimal approximation ratio O(1) and a linear time complexity O(N) for N-pixel images or N-voxel videos. A thorough evaluation of 31 superpixel methods on five image datasets and 8 supervoxel methods on four video datasets shows that our method consistently achieves the best over-segmentation accuracy. We also demonstrate the advantage of our method on one image and two video applications.

Related Material

[pdf] [supp]
author = {Ye, Zipeng and Yi, Ran and Yu, Minjing and Liu, Yong-Jin and He, Ying},
title = {Fast Computation of Content-Sensitive Superpixels and Supervoxels Using Q-Distances},
booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
month = {October},
year = {2019}