下载链接:https://pan.baidu.com/s/1U2NgEzbNxCAinZNp2cvOPQ
提取码:4oi0
分类目录归档:论文
GAN论文分类整理-ICML2018
下载链接:https://pan.baidu.com/s/1xtT0K-2N88nebX6ocnhLnQ
提取码:vnmp
GAN论文分类整理-CVPR2018
下载链接:https://pan.baidu.com/s/1bCamc3JeAe5Au4WH9hVGFg
提取码:ut33
2018人工智能国际学术会议论文集
AAAI (AAAI Conference on Artificial Intelligence):https://aaai.org/Library/AAAI/aaai18contents.php
CVPR (IEEE Conference on Computer Vision and Pattern Recognition): http://openaccess.thecvf.com/CVPR2018.py
ICCV (International Conference on Computer Vision): https://waset.org/Publications
ICML (International Conference on Machine Learning):https://icml.cc/Conferences/2018/Schedule?type=Poster
IJCAI (International Joint Conference on Artificial Intelligence):https://www.ijcai.org/proceedings/2018/
NeurIPS (Annual Conference on Neural Information Processing Systems):https://papers.nips.cc/book/advances-in-neural-information-processing-systems-31-2018
ACL (Annual Meeting of the Association for Computational Linguistics):https://acl2018.org/programme/papers/
ECCV (European Conference on Computer Vision):http://openaccess.thecvf.com/ECCV2018.py
COLING (International Conference on Computational Linguistics):http://coling2018.org/coling-2018-accepted-papers/
UAI (International Conference on Uncertainty in Artificial Intelligence):http://www.auai.org/uai2018/accepted.php#
Latent Alignment and Variational Attention
Abstract: Neural attention has become central to many state-of-the-art models in naturallanguage processing and related domains. Attention networks are an easy-to-trainand effective method for softly simulating alignment; however, the approach doesnot marginalize over latent alignments in a probabilistic sense. This property makesit difficult to compare attention to other alignment approaches, to compose it withprobabilistic models, and to perform posterior inference conditioned on observeddata. A related latent approach, hard attention, fixes these issues, but is generallyharder to train and less accurate. This work considers variational attention networks,alternatives to soft and hard attention for learning latent variable alignmentmodels, with tighter approximation bounds based on amortized variational inference.We further propose methods for reducing the variance of gradients to makethese approaches computationally feasible. Experiments show that for machinetranslation and visual question answering, inefficient exact latent variable modelsoutperform standard neural attention, but these gains go away when using hardattention based training. On the other hand, variational attention retains most ofthe performance gain but with training speed comparable to neural attention.
Implicit Autoencoders
Abstract: In this paper, we describe the “implicit autoencoder” (IAE), a generative autoencoderin which both the generative path and the recognition path are parametrizedby implicit distributions. We use two generative adversarial networks to define thereconstruction and the regularization cost functions of the implicit autoencoder,and derive the learning rules based on maximum-likelihood learning. Using implicitdistributions allows us to learn more expressive posterior and conditionallikelihood distributions for the autoencoder. Learning an expressive conditional likelihood distribution enables the latent code to only capture the abstract andhigh-level information of the data, while the remaining information is capturedby the implicit conditional likelihood distribution. For example, we show thatimplicit autoencoders can disentangle the global and local information, and performdeterministic or stochastic reconstructions of the images. We further showthat implicit autoencoders can disentangle discrete underlying factors of variationfrom the continuous factors in an unsupervised fashion, and perform clustering andsemi-supervised learning.
Wasserstein Auto-Encoders
Abstract: We propose the Wasserstein Auto-Encoder (WAE)—a new algorithm for building a generative model of the data distribution. WAE minimizes a penalized form of the Wasserstein distance between the model distribution and the target distribution, which leads to a different regularizer than the one used by the Variational Auto-Encoder (VAE) [1]. This regularizer encourages the encoded training distribution to match the prior. We compare our algorithm with several other techniques and show that it is a generalization of adversarial auto-encoders(AAE) [2]. Our experiments show that WAE shares many of the properties of VAEs (stable training, encoder-decoder architecture, nice latent manifold structure) while generating samples of better quality, as measured by the FID score.
Variational Inference: A Unified Framework of Generative Models and Some Revelations
We reinterpreting the variational inference in a new perspective.Via this way, we can easily prove that EM algorithm,VAE, GAN, AAE, ALI(BiGAN) are all special cases of variationalinference. The proof also reveals the loss of standard GAN is incomplete and it explains why we need to train GAN cautiously. From that, we find out a regularization term to improvestability of GAN training.
Variational Inference: A Unified Framework of Generative Models and Some Revelations