All Research Areas
Research Areas
Year Published

444 Results

April 24, 2017

An Analytical Formula of Population Gradient for Two-Layered ReLU network and its Applications in Convergence and Critical Point Analysis

International Conference on Learning Representations (ICLR) 2017

In this paper, we explore theoretical properties of training a two-layered ReLU network g(x; w) = PK j=1 σ(w | j x) with centered d-dimensional spherical Gaussian input x (σ=ReLU). We train our network with gradient descent on w to mimic the output of a teacher network with the same architecture and fixed parameters w∗.

By: Yuandong Tian
April 24, 2017

Connective recovery in social networks after the death of a friend

Nature Human Behavior

Most individuals have few close friends, leading to potential isolation after a friend’s death. Do social networks heal to fill the space left by the loss? We conduct such a study of self-healing and resilience in social networks.

By: William Hobbs, Moira Burke
April 24, 2017

LR-GAN: Layered Recursive Generative Adversarial Networks for Image Generation

International Conference on Learning Representations (ICLR)

We present LR-GAN: an adversarial image generation model which takes scene structure and context into account.

By: Jianwei Yang, Anitha Kannan, Dhruv Batra, Devi Parikh
April 24, 2017

Towards Principled Methods for Training Generative Adversarial Networks

International Conference on Learning Representations (ICLR) 2017

The goal of this paper is not to introduce a single algorithm or method, but to make theoretical steps towards fully understanding the training dynamics of generative adversarial networks.

By: Martin Arjovsky, Leon Bottou
April 24, 2017

Improving Neural Language Models with a Continuous Cache

International Conference on Learning Representations (ICLR) 2017

We propose an extension to neural network language models to adapt their prediction to the recent history. Our model is a simplified version of memory augmented networks, which stores past hidden activations as memory and accesses them through a dot product with the current hidden activation.

By: Edouard Grave, Armand Joulin, Nicolas Usunier
April 24, 2017

Revisiting Classifier Two-Sample Tests for GAN Evaluation and Causal Discovery

International Conference on Learning Representations (ICLR) 2017

In this paper, we aim to revive interest in the use of binary classifiers for two-sample testing. To this end, we review their fundamentals, previous literature on their use, compare their performance against alternative state-of-the-art two-sample tests, and propose them to evaluate generative adversarial network models applied to image synthesis.

By: David Lopez-Paz, Maxime Oquab
April 24, 2017

CommAI: Evaluating the First Steps Towards a Useful General AI

ICLR 2017 Workshop

We propose a set of concrete desiderata for general AI, together with a platform to test machines on how well they satisfy such desiderata, while keeping all further complexities to a minimum.

By: Marco Baroni, Armand Joulin, Allan Jabri, Germán Kruszewski, Angeliki Lazaridou, Klemen Simonic, Tomas Mikolov
April 24, 2017

Automatic Rule Extraction from Long Short Term Memory Networks

International Conference on Learning Representations (ICLR) 2017

In this paper we consider Long Short Term Memory networks (LSTMs) and demonstrate a new approach for tracking the importance of a given input to the LSTM for a given output.

By: W. James Murdoch, Arthur Szlam
April 24, 2017

Learning End-to-End Goal-Oriented Dialog

International Conference on Learning Representations (ICLR)

This paper proposes a testbed to break down the strengths and shortcomings of end-to-end dialog systems in goal-oriented applications.

By: Antoine Bordes, Y-Lan Boureau, Jason Weston
April 24, 2017

Variable Computation in Recurrent Neural Networks

International Conference on Learning Representations (ICLR) 2017

In this paper, we explore a modification to existing recurrent units which allows them to learn to vary the amount of computation they perform at each step, without prior knowledge of the sequence’s time structure.

By: Yacine Jernite, Edouard Grave, Armand Joulin, Tomas Mikolov