Learning joint latent representations based on information maximization

Research output: Contribution to journalArticlepeer-review


Learning disentangled and interpretable representations is an important aspect of information understanding. In this paper, we propose a novel deep learning model representing both discrete and continuous latent variable spaces which can be used in either supervised or unsupervised learning. The proposed model is trained using an optimization function employing the mutual information maximization criterion. For the unsupervised learning setting we define a lower bound to the mutual information between the joint distribution of the latent variables corresponding to the real data and those generated by the model. The maximization of this lower bound during the training induces the learning of disentangled and interpretable data representations. Such representations can be used for attribute manipulation and image editing tasks.
Original languageEnglish
Pages (from-to)216-236
Number of pages21
JournalInformation Sciences
Early online date5 Apr 2021
Publication statusPublished - Aug 2021

Bibliographical note

© 2021 Elsevier Inc. All rights reserved. This is an author-produced version of the published paper. Uploaded in accordance with the publisher’s self-archiving policy.

Cite this