Generating Images with Perceptual Similarity Metrics based on Deep Networks
Alexey Dosovitskiy
and
Thomas Brox
arXiv e-Print archive - 2016 via Local arXiv
Keywords:
cs.LG, cs.CV, cs.NE
First published: 2016/02/08 (8 years ago) Abstract: Image-generating machine learning models are typically trained with loss
functions based on distance in the image space. This often leads to
over-smoothed results. We propose a class of loss functions, which we call deep
perceptual similarity metrics (DeePSiM), that mitigate this problem. Instead of
computing distances in the image space, we compute distances between image
features extracted by deep neural networks. This metric better reflects
perceptually similarity of images and thus leads to better results. We show
three applications: autoencoder training, a modification of a variational
autoencoder, and inversion of deep convolutional networks. In all cases, the
generated images look sharp and resemble natural images.