分类目录归档:论文

2018人工智能国际学术会议论文集

AAAI (AAAI Conference on Artificial Intelligence):https://aaai.org/Library/AAAI/aaai18contents.php

CVPR (IEEE Conference on Computer Vision and Pattern Recognition): http://openaccess.thecvf.com/CVPR2018.py

ICCV (International Conference on Computer Vision): https://waset.org/Publications

ICML (International Conference on Machine Learning):https://icml.cc/Conferences/2018/Schedule?type=Poster

IJCAI (International Joint Conference on Artificial Intelligence):https://www.ijcai.org/proceedings/2018/

NeurIPS (Annual Conference on Neural Information Processing Systems):https://papers.nips.cc/book/advances-in-neural-information-processing-systems-31-2018

ACL (Annual Meeting of the Association for Computational Linguistics):https://acl2018.org/programme/papers/

ECCV (European Conference on Computer Vision):http://openaccess.thecvf.com/ECCV2018.py

COLING (International Conference on Computational Linguistics):http://coling2018.org/coling-2018-accepted-papers/

UAI (International Conference on Uncertainty in Artificial Intelligence):http://www.auai.org/uai2018/accepted.php#

Latent Alignment and Variational Attention

Abstract: Neural attention has become central to many state-of-the-art models in naturallanguage processing and related domains. Attention networks are an easy-to-trainand effective method for softly simulating alignment; however, the approach doesnot marginalize over latent alignments in a probabilistic sense. This property makesit difficult to compare attention to other alignment approaches, to compose it withprobabilistic models, and to perform posterior inference conditioned on observeddata. A related latent approach, hard attention, fixes these issues, but is generallyharder to train and less accurate. This work considers variational attention networks,alternatives to soft and hard attention for learning latent variable alignmentmodels, with tighter approximation bounds based on amortized variational inference.We further propose methods for reducing the variance of gradients to makethese approaches computationally feasible. Experiments show that for machinetranslation and visual question answering, inefficient exact latent variable modelsoutperform standard neural attention, but these gains go away when using hardattention based training. On the other hand, variational attention retains most ofthe performance gain but with training speed comparable to neural attention.

Latent Alignment and Variational Attention

Implicit Autoencoders

Abstract: In this paper, we describe the “implicit autoencoder” (IAE), a generative autoencoderin which both the generative path and the recognition path are parametrizedby implicit distributions. We use two generative adversarial networks to define thereconstruction and the regularization cost functions of the implicit autoencoder,and derive the learning rules based on maximum-likelihood learning. Using implicitdistributions allows us to learn more expressive posterior and conditionallikelihood distributions for the autoencoder. Learning an expressive conditional likelihood distribution enables the latent code to only capture the abstract andhigh-level information of the data, while the remaining information is capturedby the implicit conditional likelihood distribution. For example, we show thatimplicit autoencoders can disentangle the global and local information, and performdeterministic or stochastic reconstructions of the images. We further showthat implicit autoencoders can disentangle discrete underlying factors of variationfrom the continuous factors in an unsupervised fashion, and perform clustering andsemi-supervised learning.

Implicit Autoencoders

Wasserstein Auto-Encoders

Abstract: We propose the Wasserstein Auto-Encoder (WAE)—a new algorithm for building a generative model of the data distribution. WAE minimizes a penalized form of the Wasserstein distance between the model distribution and the target distribution, which leads to a different regularizer than the one used by the Variational Auto-Encoder (VAE) [1]. This regularizer encourages the encoded training distribution to match the prior. We compare our algorithm with several other techniques and show that it is a generalization of adversarial auto-encoders(AAE) [2]. Our experiments show that WAE shares many of the properties of VAEs (stable training, encoder-decoder architecture, nice latent manifold structure) while generating samples of better quality, as measured by the FID score.

Wasserstein Auto-Encoders

Variational Inference: A Unified Framework of Generative Models and Some Revelations

We reinterpreting the variational inference in a new perspective.Via this way, we can easily prove that EM algorithm,VAE, GAN, AAE, ALI(BiGAN) are all special cases of variationalinference. The proof also reveals the loss of standard GAN is incomplete and it explains why we need to train GAN cautiously. From that, we find out a regularization term to improvestability of GAN training.

Variational Inference: A Unified Framework of Generative Models and Some Revelations