I want to use SSIM metric as my loss function for the model I'm working on in tensorflow. SSIM should measure the similarity between my reconstructed output image of my denoising autoencoder and the input uncorrupted image (RGB).
As of what I understood, for using the SSIM metric in tensorflow, the images should be normalized to [0,1] or [0,255] and not [-1,1]. After converting my tensors to [0,1] and implementing SSIM as my loss function, the reconstructed image is black and white instead of a colorful RGB image.
tf.reduce_mean(tf.image.ssim(reconstructed, truth, 1.0))
My model is working fine with MSE (mean squared error), the reconstructed images are colorful (RGB).
using tf.losses.mean_squared_error(truth, reconstructed)
the reconstructed image would be RGB image, while using SSIM would give me a one dimensional image.
Why using SSIM as loss function gives me different result than MSE (in terms of reconstructed image channels) in tensorflow?
I was capable of solving the issue by changing the dynamic range of the images to 2.0, since I have images scaled between [-1, 1] by:
loss_rec = tf.reduce_mean(tf.image.ssim(truth, reconstructed, 2.0))
And since a better image quality is shown by a higher SSIM value, I had to minimize the negative of my loss function (SSIM) to optimize my model:
optimizer = tf.train.AdamOptimizer(learning_rate).minimize(-1 * loss_rec)