Discovering Style Trends Through Deep Visually Aware Latent Item Embeddings

Murium Iqbal, Adair Kovac, Kamelia Aryafar; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2018, pp. 1899-1901

Abstract


In this paper, we explore Latent Dirichlet Allocation (LDA) and Polylingual Latent Dirichlet Allocation (PolyLDA), as a means to discover trending styles in Overstock from deep visual semantic features transferred from a pretrained convolutional neural network and text-based item attributes. To utilize deep visual semantic features in conjunction with LDA, we develop a method for creating a bag of words representation of unrolled image vectors. By viewing the channels within the convolutional layers of a Resnet-50 as being representative of a word, we can index these activations to create visual documents. We then train LDA over these documents to discover the latent style in the images. We also incorporate text-based data with PolyLDA, where each representation is viewed as an independent language attempting to describe the same style. The resulting topics are shown to be excellent indicators of visual style across our platform.

Related Material


[pdf] [arXiv]
[bibtex]
@InProceedings{Iqbal_2018_CVPR_Workshops,
author = {Iqbal, Murium and Kovac, Adair and Aryafar, Kamelia},
title = {Discovering Style Trends Through Deep Visually Aware Latent Item Embeddings},
booktitle = {Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2018}
}