官术网_书友最值得收藏!

Variational autoencoders

Auto-encoder models are used to model , the joint probability of , the observed data, and , the latent variable. The joint probability is normally factorized as . During inference, we are interested in finding good values of to produce observed data – that is, we are interested in learning the posterior probability of  given . Using Bayes' rule, we can rewrite the posterior as follows:

Close inspection of this equation reveals that computing the evidence, the marginal distribution of the data , is hardly possible and normally intractable. We first try to circumvent this barrier by computing an approximation of the evidence. We do it by using variational inference, and, instead, estimating the parameters of a known distribution  that is the least divergent from the posterior Variational inference approximates the posterior  with a family of distributions , where the variational  parameter indexes the family of distributions. This can be done by minimizing the KL divergence between  and , as described in the following equation:

Unfortunately, using does not circumvent the problem, and we are still faced with computing the evidence . At this point, we give up on computing the exact evidence and focus on estimating an Evidence Lower Bound (ELBO). The ELBO sets the perfect scenario for Variational Autoencoders, and is computed by removing of the previous equation and inverting the signs, giving:

VAEs consist of an encoder   parametrized by , and a decoder   parametrized by . The encoder is trained to maximize the posterior probability of a  latent vector, given the , . data. The decoder is trained to maximize the probability of the data  given the latent representation  latent vector, . Informally speaking, the encoder learns to compress the data into a latent representation, and the decoder learns to decompress the data from the latent representation. The VAE loss is defined as follows:

The first term represents the reconstruction loss, or the expectation of the negative probability. The second term is a regularize term that was derived in our problem setup. 

Unlike autoregressive models, VAEs are normally easy to run in parallel during training and inference. Conversely, they are normally harder to optimize than autoregressive models. 

Deep feature-consistent VAE is one of the best models for image generation using VAEs. The following figure shows the faces generated by the model. From a qualitative perspective, image samples produced with VAEs tend to be blurry. 

Source: Deep feature-consistent VAE (https://arxiv.org/abs/1610.00291)
主站蜘蛛池模板: 泊头市| 依安县| 平潭县| 醴陵市| 沾益县| 同德县| 五莲县| 苍梧县| 清水河县| 长白| 定襄县| 通渭县| 宣威市| 长春市| 绥德县| 江安县| 志丹县| 休宁县| 姜堰市| 溧水县| 迁安市| 新乡市| 古田县| 黄冈市| 扶绥县| 寿宁县| 武山县| 海门市| 颍上县| 沾化县| 会东县| 连南| 南和县| 汤原县| 高州市| 洪江市| 彩票| 五常市| 通河县| 武平县| 沈阳市|