Channel-Recurrent Autoencoding for Image Modeling

2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2018)(2018)

引用 15|浏览95
暂无评分
摘要
Despite recent successes in synthesizing faces and bedrooms, existing generative models struggle to capture more complex image types (Figure 1), potentially due to the oversimplification of their latent space constructions. To tackle this issue, building on Variational Autoencoders (VAEs), we integrate recurrent connections across channels to both inference and generation steps, allowing the high-level features to be captured in global-to-local, coarse-to-fine manners. Combined with adversarial loss, our channel-recurrent VAE-GAN (crVAE-GAN) outperforms VAE-GAN in generating a diverse spectrum of high resolution images while maintaining the same level of computational efficacy. Our model produces interpretable and expressive latent representations to benefit downstream tasks such as image completion. Moreover, we propose two novel regularizations, namely the KL objective weighting scheme over time steps and mutual information maximization between transformed latent variables and the outputs, to enhance the training.
更多
查看译文
关键词
oversimplification,latent space constructions,Variational Autoencoders,VAEs,recurrent connections,inference,generation steps,high-level features,coarse-to-fine manners,adversarial loss,channel-recurrent VAE-GAN,crVAE-GAN,diverse spectrum,high resolution images,latent representations,image completion,time steps,transformed latent variables,complex image types,generative models struggle,bedrooms,faces,image modeling,channel-recurrent autoencoding
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要