Academics | Yann LeCun Latest research: can help Gan to use discrete data arae

Source: Internet
Author: User

AI Technology Review News, Fair, head of the Facebook AI Research Center, Yann LeCun, one of the deep-learning troika, commented on a dynamic on Facebook yesterday. Yann LeCun forwarded the news from a New York University who worked with him, who was reading a doctoral student, Jake Zhao, introducing a paper that had been uploaded to ArXiv. This paper introduces a kind of "antagonistic regularization Automatic encoder" (arae,adversarially regularized autoencoders) which can help Gan to use discrete data. Yann LeCun, as one of the co-authors of this paper, forwarded the message and was interpreted and evaluated.

According to the original paper, AI Science and Technology Review compiled such a diagram, can be arae to solve the problem to do a general understanding:

As shown in the original dynamic and Yann LeCun forwarding comments, please see below

Yann LeCun's forwarding comments

Adversarially regularized autoencoders (adversarial regularization automatic encoder, Arae) is a new method that restricts the encoded content information to an automatic encoder and prevents the automatic encoder from learning the identity function. In order to achieve this goal, there have been some sparse ae (sparse automatic encoder), variational AE (variational automatic encoder, add noise in the code), Denoising AE (Noise reduction automatic encoder) Such an attempt, There is also a comparison between these comparisons and other methods based on MCMC (Markov Chain Monte Carlo, random simulations). This time, we use a set of adversarial generators and discriminator to generate a sample of the specified entropy, and we use the discriminator to make the coded distribution in the Automatic encoder match the distribution of the generator discriminator.

After training, the generator and encoder parts of this automatic encoder can be used as a generative model.

Arae can have a powerful usage, which is the ability to generate discrete structures (such as text) in a confrontational setting. Since the process of confrontation occurs on (continuous) encoding rather than output, this can alleviate the problems that the adversarial discriminator may encounter when it is used directly on discrete outputs.

The study was attended by New York University, Harvard University, Fair, several participants, a master of data science graduate from New York University, PhD in Reading Jake Zhao, Harvard University alumni, PhD in Reading Yoon Kim, New York University undergraduate Kelly Zhang, Harvard University student Sasha Rush, as well as myself. Jake Zhao's message content

Just published this article "Adversarially regularized autoencoders for generating discrete structures" (Auto encoder for the adversarial regularization of discrete structures), and I felt very excited, Several co-authors Yoon Kim, Kelly Zhang, Sasha Rush and Yann LeCun are also very powerful.

Paper Address: https://arxiv.org/abs/1706.04223

Code address (with torch and pytorch two versions): Https://github.com/jakezhaojb/ARAE

For adversarial training on discrete structures, we present a simple solution: train a sequence-to-sequence transformation model in a rebuilt manner, and then let Gan work in a contiguous coded space. At the same time, the gradient from Gan to the Sequence transformation model is helpful to train a meaningful automatic sequence conversion model. By effort, we got an image Gan network should have a set of ideal properties: Generative, Z-space interpolation, vector computing. We also make a quantitative evaluation of this method based on the quality of the sample generated by the language model.

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.