【ICLR2019】Poster 论文汇总

ICLR2019 Poster 文章汇总, 共478 papers

在这里插入图片描述

Convolutional Neural Networks on Non-uniform Geometrical Signals Using Euclidean Spectral Transformation
Keywords:Non-uniform Fourier Transform, 3D Learning, CNN, surface reconstruction
TL;DR:We use non-Euclidean Fourier Transformation of shapes defined by a simplicial complex for deep learning, achieving significantly better results than point-based sampling techiques used in current 3D learning literature.
Augmented Cyclic Adversarial Learning for Low Resource Domain Adaptation
Keywords:Domain adaptation, generative adversarial network, cyclic adversarial learning, speech
TL;DR:A new cyclic adversarial learning augmented with auxiliary task model which improves domain adaptation performance in low resource supervised and unsupervised situations
Variance Networks: When Expectation Does Not Meet Your Expectations
Keywords:deep learning, variational inference, variational dropout
TL;DR:It is possible to learn a zero-centered Gaussian distribution over the weights of a neural network by learning only variances, and it works surprisingly well.
Initialized Equilibrium Propagation for Backprop-Free Training
Keywords:credit assignment, energy-based models, biologically plausible learning
TL;DR:We train a feedforward network without backprop by using an energy-based model to provide local targets
Explaining Image Classifiers by Counterfactual Generation
Keywords:Explainability, Interpretability, Generative Models, Saliency Map, Machine Learning, Deep Learning
TL;DR:We compute saliency by using a strong generative model to efficiently marginalize over plausible alternative inputs, revealing concentrated pixel areas that preserve label information.
SNIP: SINGLE-SHOT NETWORK PRUNING BASED ON CONNECTION SENSITIVITY
Keywords:neural network pruning, connection sensitivity
TL;DR:We present a new approach, SNIP, that is simple, versatile and interpretable; it prunes irrelevant connections for a given task at single-shot prior to training and is applicable to a variety of neural network models without modifications.
Diagnosing and Enhancing VAE Models
Keywords:variational autoencoder, generative models
TL;DR:We closely analyze the VAE objective function and draw novel conclusions that lead to simple enhancements.
Disjoint Mapping Network for Cross-modal Matching of Voices and Faces
Keywords:None
TL;DR:None
Automatically Composing Representation Transformations as a Means for Generalization
Keywords:compositionality, deep learning, metareasoning
TL;DR:We explore the problem of compositional generalization and propose a means for endowing neural network architectures with the ability to compose themselves to solve these problems.
Visual Reasoning by Progressive Module Networks
Keywords:None
TL;DR:None
Bayesian Deep Convolutional Networks with Many Channels are Gaussian Processes
Keywords:Deep Convolutional Neural Networks, Gaussian Processes, Bayesian
TL;DR:Finite-width SGD trained CNNs vs. infinitely wide fully Bayesian CNNs. Who wins?
Learning to Learn without Forgetting by Maximizing Transfer and Minimizing Interference
Keywords:None
TL;DR:None
Sparse Dictionary Learning by Dynamical Neural Networks
Keywords:None
TL;DR:None
Eidetic 3D LSTM: A Model for Video Prediction and Beyond
Keywords:None
TL;DR:None
ALISTA: Analytic Weights Are As Good As Learned Weights in LISTA
Keywords:None
TL;DR:None
Three Mechanisms of Weight Decay Regularization
Keywords:Generalization, Regularization, Optimization
TL;DR:We investigate weight decay regularization for different optimizers and identify three distinct mechanisms by which weight decay improves generalization.
Learning Multimodal Graph-to-Graph Translation for Molecule Optimization
Keywords:graph-to-graph translation, graph generation, molecular optimization
TL;DR:We introduce a graph-to-graph encoder-decoder framework for learning diverse graph translations.
A Data-Driven and Distributed Approach to Sparse Signal Representation and Recovery
Keywords:Sparsity, Compressive Sensing, Convolutional Network
TL;DR:We use deep learning techniques to solve the sparse signal representation and recovery problem.
On the Minimal Supervision for Training Any Binary Classifier from Only Unlabeled Data
Keywords:learning from only unlabeled data, empirical risk minimization, unbiased risk estimator
TL;DR:Three class priors are all you need to train deep models from only U data, while any two should not be enough.
Neural Logic Machines
Keywords:Neural-Symbolic Computation, Rule Induction, First-Order Logic
TL;DR:We propose the Neural Logic Machine (NLM), a neural-symbolic architecture for both inductive learning and logic reasoning.
Neural Speed Reading with Structural-Jump-LSTM
Keywords:natural language processing, speed reading, recurrent neural network, classification
TL;DR:We propose a new model for neural speed reading that utilizes the inherent punctuation structure of a text to define effective jumping and skipping behavior.
Rigorous Agent Evaluation: An Adversarial Approach to Uncover Catastrophic Failures
Keywords:agent evaluation, adversarial examples, robustness, safety, reinforcement learning
TL;DR:We show that rare but catastrophic failures may be missed entirely by random testing, which poses issues for safe deployment. Our proposed approach for adversarial testing fixes this.
Woulda, Coulda, Shoulda: Counterfactually-Guided Policy Search
Keywords:None
TL;DR:None
signSGD via Zeroth-Order Oracle
Keywords:nonconvex optimization, zeroth-order algorithm, black-box adversarial attack
TL;DR:We design and analyze a new zeroth-order stochastic optimization algorithm, ZO-signSGD, and demonstrate its connection and application to black-box adversarial attacks in robust deep learning
Preventing Posterior Collapse with delta-VAEs
Keywords:Posterior Collapse, VAE, Autoregressive Models
TL;DR: Avoid posterior collapse by lower bounding the rate.
Algorithmic Framework for Model-based Deep Reinforcement Learning with Theoretical Guarantees
Keywords:model-based reinforcement learning, sample efficiency, deep reinforcement learning
TL;DR:We design model-based reinforcement learning algorithms with theoretical guarantees and achieve state-of-the-art results on Mujuco benchmark tasks when one million or fewer samples are permitted.
Knowledge Flow: Improve Upon Your Teachers
Keywords:Transfer Learning, Reinforcement Learning
TL;DR:‘Knowledge Flow’ trains a deep net (student) by injecting information from multiple nets (teachers). The student is independent upon training and performs very well on learned tasks irrespective of the setting (reinforcement or supervised learning).
Directed-Info GAIL: Learning Hierarchical Policies from Unsegmented Demonstrations using Directed Information
Keywords:Imitation Learning, Reinforcement Learning, Deep Learning
TL;DR:Learning Hierarchical Policies from Unsegmented Demonstrations using Directed Information
A Max-Affine Spline Perspective of Recurrent Neural Networks
Keywords:RNN, max-affine spline operators
TL;DR:We provide new insights and interpretations of RNNs from a max-affine spline operators perspective.
Learning to Navigate the Web
Keywords:navigating web pages, reinforcement learning, q learning, curriculum learning, meta training
TL;DR:We train reinforcement learning policies using reward augmentation, curriculum learning, and meta-learning to successfully navigate web pages.
Training for Faster Adversarial Robustness Verification via Inducing ReLU Stability
Keywords:verification, adversarial robustness, adversarial examples, stability, deep learning, regularization
TL;DR:We develop methods to train deep neural models that are both robust to adversarial perturbations and whose robustness is significantly easier to verify.
Learning to Learn with Conditional Class Dependencies
Keywords:meta-learning, learning to learn, few-shot learning
TL;DR:CAML is an instance of MAML with conditional class dependencies.
Hierarchical Visuomotor Control of Humanoids
Keywords:hierarchical reinforcement learning, motor control, motion capture
TL;DR:Solve tasks involving vision-guided humanoid locomotion, reusing locomotion behavior from motion capture data.
Unsupervised Adversarial Image Reconstruction
Keywords:None
TL;DR:None
Max-MIG: an Information Theoretic Approach for Joint Learning from Crowds
Keywords:None
TL;DR:None
AutoLoss: Learning Discrete Schedule for Alternate Optimization
Keywords:Meta Learning, AutoML, Optimization Schedule
TL;DR:We propose a unified formulation for iterative alternate optimization and develop AutoLoss, a framework to automatically learn and generate optimization schedules.
Learning what and where to attend
Keywords:Attention models, human feature importance, object recognition, cognitive science
TL;DR:A large-scale dataset for training attention models for object recognition leads to more accurate, interpretable, and human-like object recognition.
ROBUST ESTIMATION VIA GENERATIVE ADVERSARIAL NETWORKS
Keywords:robust statistics, neural networks, minimax rate, data depth, contamination model, Tukey median, GAN
TL;DR:GANs are shown to provide us a new effective robust mean estimate against agnostic contaminations with both statistical optimality and practical tractability.
INVASE: Instance-wise Variable Selection using Neural Networks
Keywords:None
TL;DR:None
Meta-Learning with Latent Embedding Optimization
Keywords:meta-learning, few-shot, miniImageNet, tieredImageNet, hypernetworks, generative, latent embedding, optimization
TL;DR:Latent Embedding Optimization (LEO) is a novel gradient-based meta-learner with state-of-the-art performance on the challenging 5-way 1-shot and 5-shot miniImageNet and tieredImageNet classification tasks.
Non-vacuous Generalization Bounds at the ImageNet Scale: a PAC-Bayesian Compression Approach
Keywords:generalization, deep-learning, pac-bayes
TL;DR:We obtain non-vacuous generalization bounds on ImageNet-scale deep neural networks by combining an original PAC-Bayes bound and an off-the-shelf neural network compression method.
Learning to Represent Edits
Keywords:None
TL;DR:None
Neural Probabilistic Motor Primitives for Humanoid Control
Keywords:Motor Primitives, Distillation, Reinforcement Learning, Continuous Control, Humanoid Control, Motion Capture, One-Shot Imitation
TL;DR:Neural Probabilistic Motor Primitives compress motion capture tracking policies into one flexible model capable of one-shot imitation and reuse as a low-level controller.
Differentiable Perturb-and-Parse: Semi-Supervised Parsing with a Structured Variational Autoencoder
Keywords:differentiable dynamic programming, variational auto-encoder, dependency parsing, semi-supervised learning
TL;DR:Differentiable dynamic programming over perturbed input weights with application to semi-supervised VAE
Janossy Pooling: Learning Deep Permutation-Invariant Functions for Variable-Size Inputs
Keywords:representation learning, permutation invariance, set functions, feature pooling
TL;DR:We propose Janossy pooling, a method for learning deep permutation invariant functions designed to exploit relationships within the input sequence and tractable inference strategies such as a stochastic optimization procedure we call piSGD
An Empirical Study of Example Forgetting during Deep Neural Network Learning
Keywords:catastrophic forgetting, sample weighting, deep generalization
TL;DR:We show that catastrophic forgetting occurs within what is considered to be a single task and find that examples that are not prone to forgetting can be removed from the training set without loss of generalization.
RNNs implicitly implement tensor-product representations
Keywords:tensor-product representations, compositionality, neural network interpretability, recurrent neural networks
TL;DR:RNNs implicitly implement tensor-product representations, a principled and interpretable method for representing symbolic structures in continuous space.
Learning To Solve Circuit-SAT: An Unsupervised Differentiable Approach
Keywords:Neuro-Symbolic Methods, Circuit Satisfiability, Neural SAT Solver, Graph Neural Networks
TL;DR:We propose a neural framework that can learn to solve the Circuit Satisfiability problem from (unlabeled) circuit instances.
Dynamic Channel Pruning: Feature Boosting and Suppression
Keywords:dynamic network, faster CNNs, channel pruning
TL;DR:We make convolutional layers run faster by dynamically boosting and suppressing channels in feature computation.
signSGD with Majority Vote is Communication Efficient and Fault Tolerant
Keywords:large-scale learning, distributed systems, communication efficiency, convergence rate analysis, robust optimisation
TL;DR:Workers send gradient signs to the server, and the update is decided by majority vote. We show that this algorithm is convergent, communication efficient and fault tolerant, both in theory and in practice.
Bounce and Learn: Modeling Scene Dynamics with Real-World Bounces
Keywords:None
TL;DR:None
K for the Price of 1: Parameter-efficient Multi-task and Transfer Learning
Keywords:deep learning, mobile, transfer learning, multi-task learning, computer vision, small models, imagenet, inception, batch normalization
TL;DR:A novel and practically effective method to adapt pretrained neural networks to new tasks by retraining a minimal (e.g., less than 2%) number of parameters
Towards Metamerism via Foveated Style Transfer
Keywords:Metamerism, foveation, perception, style transfer, psychophysics
TL;DR:We introduce a novel feed-forward framework to generate visual metamers
Post Selection Inference with Incomplete Maximum Mean Discrepancy Estimator
Keywords:None
TL;DR:None
Emergent Coordination Through Competition
Keywords:Multi-agent learning, Reinforcement Learning
TL;DR:We introduce a new MuJoCo soccer environment for continuous multi-agent reinforcement learning research, and show that population-based training of independent reinforcement learners can learn cooperative behaviors
Prior Convictions: Black-box Adversarial Attacks with Bandits and Priors
Keywords:adversarial examples, gradient estimation, black-box attacks, model-based optimization, bandit optimization
TL;DR:We present a unifying view on black-box adversarial attacks as a gradient estimation problem, and then present a framework (based on bandits optimization) to integrate priors into gradient estimation, leading to significantly increased performance.
Sample Efficient Imitation Learning for Continuous Control
Keywords:Imitation Learning, Continuous Control, Reinforcement Learning, Inverse Reinforcement Learning, Conditional Generative Adversarial Network
TL;DR:In this paper, we proposed a model-free, off-policy IL algorithm for continuous control. Experimental results showed that our algorithm achieves competitive results with GAIL while significantly reducing the environment interactions.
Generative Code Modeling with Graphs
Keywords:Generative Model, Source Code, Graph Learning
TL;DR:Representing programs as graphs including semantics helps when generating programs
Critical Learning Periods in Deep Networks
Keywords:Critical Period, Deep Learning, Information Theory, Artificial Neuroscience, Information Plasticity
TL;DR:Sensory deficits in early training phases can lead to irreversible performance loss in both artificial and neuronal networks, suggesting information phenomena as the common cause, and point to the importance of the initial transient and forgetting.
CEM-RL: Combining evolutionary and gradient-based methods for policy search
Keywords:evolution strategy, deep reinforcement learning
TL;DR:We propose a new combination of evolution strategy and deep reinforcement learning which takes the best of both worlds
LanczosNet: Multi-Scale Deep Graph Convolutional Networks
Keywords:None
TL;DR:None
Excessive Invariance Causes Adversarial Vulnerability
Keywords:Generalization, Adversarial Examples, Invariance, Information Theory, Invertible Networks
TL;DR:We show deep networks are not only too sensitive to task-irrelevant changes of their input, but also too invariant to a wide range of task-relevant changes, thus making vast regions in input space vulnerable to adversarial attacks.
Hindsight policy gradients
Keywords:reinforcement learning, policy gradients, multi-goal reinforcement learning
TL;DR:We introduce the capacity to exploit information about the degree to which an arbitrary goal has been achieved while another goal was intended to policy gradient methods.
Adaptive Gradient Methods with Dynamic Bound of Learning Rate
Keywords:Optimization, SGD, Adam, Generalization
TL;DR:Novel variants of optimization methods that combine the benefits of both adaptive and non-adaptive methods.
Decoupled Weight Decay Regularization
Keywords:None
TL;DR:None
Optimistic mirror descent in saddle-point problems: Going the extra (gradient) mile
Keywords:Mirror descent, extra-gradient, generative adversarial networks, saddle-point problems
TL;DR:We show how the inclusion of an extra-gradient step in first-order GAN training methods can improve stability and lead to improved convergence results.
DialogWAE: Multimodal Response Generation with Conditional Wasserstein Auto-Encoder
Keywords:None
TL;DR:None
No Training Required: Exploring Random Encoders for Sentence Classification
Keywords:None
TL;DR:None
Neural Graph Evolution: Automatic Robot Design
Keywords:Reinforcement learning, graph neural networks, robotics, deep learning, transfer learning
TL;DR:Automatic robotic design search with graph neural networks
Function Space Particle Optimization for Bayesian Neural Networks
Keywords:None
TL;DR:None
Structured Adversarial Attack: Towards General Implementation and Better Interpretability
Keywords:None
TL;DR:None
Spherical CNNs on Unstructured Grids
Keywords:Spherical CNN, unstructured grid, panoramic, semantic segmentation, parameter efficiency
TL;DR:We present a new CNN kernel for unstructured grids for spherical signals, and show significant accuracy and parameter efficiency gain on tasks such as 3D classfication and omnidirectional image segmentation.
Optimal Transport Maps For Distribution Preserving Operations on Latent Spaces of Generative Models
Keywords:generative models, optimal transport, distribution preserving operations
TL;DR:We propose a framework for modifying the latent space operations such that the distribution mismatch between the resulting outputs and the prior distribution the generative model was trained on is fully eliminated.
Deep Lagrangian Networks: Using Physics as Model Prior for Deep Learning
Keywords:Deep Model Learning, Robot Control
TL;DR:This paper introduces a physics prior for Deep Learning and applies the resulting network topology for model-based control.
Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Keywords:reduced precision floating-point, partial sum accumulation bit-width, deep learning, training
TL;DR:We present an analytical framework to determine accumulation bit-width requirements in all three deep learning training GEMMs and verify the validity and tightness of our method via benchmarking experiments.
Deep Convolutional Networks as shallow Gaussian Processes
Keywords:Gaussian process, CNN, ResNet, Bayesian
TL;DR:We show that CNNs and ResNets with appropriate priors on the parameters are Gaussian processes in the limit of infinitely many convolutional filters.
Unsupervised Domain Adaptation for Distance Metric Learning
Keywords:domain adaptation, distance metric learning, face recognition
TL;DR:A new theory of unsupervised domain adaptation for distance metric learning and its application to face recognition across diverse ethnicity variations.
A comprehensive, application-oriented study of catastrophic forgetting in DNNs
Keywords:incremental learning, deep neural networks, catatrophic forgetting, sequential learning
TL;DR:We check DNN models for catastrophic forgetting using a new evaluation scheme that reflects typical application conditions, with surprising results.
Posterior Attention Models for Sequence to Sequence Learning
Keywords:posterior inference, attention, seq2seq learning, translation
TL;DR:Computing attention based on posterior distribution leads to more meaningful attention and better performance
Generative Question Answering: Learning to Answer the Whole Question
Keywords:Question answering, question generation, reasoning, squad, clevr
TL;DR:Question answering models that model the joint distribution of questions and answers can learn more than discriminative models
Diversity and Depth in Per-Example Routing Models
Keywords:conditional computation, routing models, depth
TL;DR:Per-example routing models benefit from architectural diversity, but still struggle to scale to a large number of routing decisions.
Selfless Sequential Learning
Keywords:Lifelong learning, Continual Learning, Sequential learning, Regularization
TL;DR:A regularization strategy for improving the performance of sequential learning
M^3RL: Mind-aware Multi-agent Management Reinforcement Learning
Keywords:Multi-agent Reinforcement Learning, Deep Reinforcement Learning
TL;DR:We propose Mind-aware Multi-agent Management Reinforcement Learning (M^3RL) for training a manager to motivate self-interested workers to achieve optimal collaboration by assigning suitable contracts to them.
The Deep Weight Prior
Keywords:deep learning, variational inference, prior distributions
TL;DR:The deep weight prior learns a generative model for kernels of convolutional neural networks, that acts as a prior distribution while training on new datasets.
Efficient Multi-Objective Neural Architecture Search via Lamarckian Evolution
Keywords:Neural Architecture Search, AutoML, AutoDL, Deep Learning, Evolutionary Algorithms, Multi-Objective Optimization
TL;DR:We propose a method for efficient Multi-Objective Neural Architecture Search based on Lamarckian inheritance and evolutionary algorithms.
Quaternion Recurrent Neural Networks
Keywords:None
TL;DR:None
Adversarial Audio Synthesis
Keywords:audio, waveform, spectrogram, GAN, adversarial, WaveGAN, SpecGAN
TL;DR:Learning to synthesize raw waveform audio with GANs
Preconditioner on Matrix Lie Group for SGD
Keywords:preconditioner, stochastic gradient descent, Newton method, Fisher information, natural gradient, Lie group
TL;DR:We propose a new framework for preconditioner learning, derive new forms of preconditioners and learning methods, and reveal the relationship to methods like RMSProp, Adam, Adagrad, ESGD, KFAC, batch normalization, etc.
Learning to Screen for Fast Softmax Inference on Large Vocabulary Neural Networks
Keywords:None
TL;DR:None
Adaptive Posterior Learning: few-shot learning with a surprise-based memory module
Keywords:metalearning, memory, few-shot, relational, self-attention, classification, sequential, reasoning, working memory, episodic memory
TL;DR:We introduce a model which generalizes quickly from few observations by storing surprising information and attending over the most relevant data at each time point.
Probabilistic Planning with Sequential Monte Carlo methods
Keywords:control as inference, probabilistic planning, sequential monte carlo, model based reinforcement learning
TL;DR:Leveraging control as inference and Sequential Monte Carlo methods, we proposed a probabilistic planning algorithm.
Plan Online, Learn Offline: Efficient Learning and Exploration via Model-Based Control
Keywords:deep reinforcement learning, exploration, model-based
TL;DR:We propose a framework that incorporates planning for efficient exploration and learning in complex environments.
DHER: Hindsight Experience Replay for Dynamic Goals
Keywords:None
TL;DR:None
FlowQA: Grasping Flow in History for Conversational Machine Comprehension
Keywords:Machine Comprehension, Conversational Agent, Natural Language Processing, Deep Learning
TL;DR:We propose the Flow mechanism and an end-to-end architecture, FlowQA, that achieves SotA on two conversational QA datasets and a sequential instruction understanding task.
Learning to Design RNA
Keywords:matter engineering, bioinformatics, rna design, reinforcement learning, meta learning, neural architecture search, hyperparameter optimization
TL;DR:We learn to solve the RNA Design problem with reinforcement learning using meta learning and autoML approaches.
Robust Conditional Generative Adversarial Networks
Keywords:conditional GAN, unsupervised pathway, autoencoder, robustness
TL;DR:We introduce a new type of conditional GAN, which aims to leverage structure in the target space of the generator. We augment the generator with a new, unsupervised pathway to learn the target structure.
Top-Down Neural Model For Formulae
Keywords:logic, formula, recursive neural networks, recurrent neural networks
TL;DR:A top-down approach how to recursively represent propositional formulae by neural networks is presented.
Cost-Sensitive Robustness against Adversarial Examples
Keywords:Certified robustness, Adversarial examples, Cost-sensitive learning
TL;DR:A general method for training certified cost-sensitive robust classifier against adversarial perturbations
The role of over-parametrization in generalization of neural networks
Keywords:Generalization, Over-Parametrization, Neural Networks, Deep Learning
TL;DR:We suggest a generalization bound that could partly explain the improvement in generalization with over-parametrization.
Diffusion Scattering Transforms on Graphs
Keywords:graph neural networks, deep learning, stability, scattering transforms, convolutional neural networks
TL;DR:Stability of scattering transform representations of graph data to deformations of the underlying graph support.
Capsule Graph Neural Network
Keywords:CapsNet, Graph embedding, GNN
TL;DR:Inspired by CapsNet, we propose a novel architecture for graph embeddings on the basis of node features extracted from GNN.
Energy-Constrained Compression for Deep Neural Networks via Weighted Sparse Projection and Layer Input Masking
Keywords:None
TL;DR:None
Emerging Disentanglement in Auto-Encoder Based Unsupervised Image Content Transfer
Keywords:Image-to-image Translation, Disentanglement, Autoencoders, Faces
TL;DR:An image to image translation method which adds to one image the content of another thereby creating a new image.
SGD Converges to Global Minimum in Deep Learning via Star-convex Path
Keywords:None
TL;DR:None
Toward Understanding the Impact of Staleness in Distributed Machine Learning
Keywords:None
TL;DR:None
Transfer Learning for Sequences via Learning to Collocate
Keywords:transfer learning, recurrent neural network, attention, natural language processing
TL;DR:Transfer learning for sequence via learning to align cell-level information across domains.
Learning Procedural Abstractions and Evaluating Discrete Latent Temporal Structure
Keywords:None
TL;DR:None
Unsupervised Speech Recognition via Segmental Empirical Output Distribution Matching
Keywords:None
TL;DR:None
Adversarial Attacks on Graph Neural Networks via Meta Learning
Keywords:graph mining, adversarial attacks, meta learning, graph neural networks, node classification
TL;DR:We use meta-gradients to attack the training procedure of deep neural networks for graphs.
Maximal Divergence Sequential Autoencoder for Binary Software Vulnerability Detection
Keywords:Vulnerabilities Detection, Sequential Auto-Encoder, Separable Representation
TL;DR:We propose a novel method named Maximal Divergence Sequential Auto-Encoder that leverages Variational AutoEncoder representation for binary code vulnerability detection.
Neural Program Repair by Jointly Learning to Localize and Repair
Keywords:neural program repair, neural program embeddings, pointer networks
TL;DR:Multi-headed Pointer Networks for jointly learning to localize and repair Variable Misuse bugs
Information-Directed Exploration for Deep Reinforcement Learning
Keywords:reinforcement learning, exploration, information directed sampling
TL;DR:We develop a practical extension of Information-Directed Sampling for Reinforcement Learning, which accounts for parametric uncertainty and heteroscedasticity in the return distribution for exploration.
Attention, Learn to Solve Routing Problems!
Keywords:learning, routing problems, heuristics, attention, reinforce, travelling salesman problem, vehicle routing problem, orienteering problem, prize collecting travelling salesman problem
TL;DR:Attention based model trained with REINFORCE with greedy rollout baseline to learn heuristics with competitive results on TSP and other routing problems
L2-Nonexpansive Neural Networks
Keywords:None
TL;DR:None
Improving Generalization and Stability of Generative Adversarial Networks
Keywords:GAN, generalization, gradient penalty, zero centered, convergence
TL;DR:We propose a zero-centered gradient penalty for improving generalization and stability of GANs
Adaptive Input Representations for Neural Language Modeling
Keywords:Neural language modeling
TL;DR:Variable capacity input word embeddings and SOTA on WikiText-103, Billion Word benchmarks.
Neural Persistence: A Complexity Measure for Deep Neural Networks Using Algebraic Topology
Keywords:Algebraic topology, persistent homology, network complexity, neural network
TL;DR:We develop a new topological complexity measure for deep neural networks and demonstrate that it captures their salient properties.
Efficient Augmentation via Data Subsampling
Keywords:data augmentation, invariance, subsampling, influence
TL;DR:Selectively augmenting difficult to classify points results in efficient training.
Neural TTS Stylization with Adversarial and Collaborative Games
Keywords:Text-To-Speech synthesis, GANs
TL;DR:a generative adversarial network for style modeling in a text-to-speech system
Optimal Control Via Neural Networks: A Convex Approach
Keywords:None
TL;DR:None
CBOW Is Not All You Need: Combining CBOW with the Compositional Matrix Space Model
Keywords:Text representation learning, Sentence embedding, Efficient training scheme, word2vec
TL;DR:We present a novel training scheme for efficiently obtaining order-aware sentence representations.
Stochastic Optimization of Sorting Networks via Continuous Relaxations
Keywords:continuous relaxations, sorting, permutation, stochastic computation graphs, Plackett-Luce
TL;DR:We provide a continuous relaxation to the sorting operator, enabling end-to-end, gradient-based stochastic optimization.
Adaptivity of deep ReLU network for learning in Besov and mixed smooth Besov spaces: optimal rate and curse of dimensionality
Keywords:None
TL;DR:None
Generating Multiple Objects at Spatially Distinct Locations
Keywords:controllable image generation, text-to-image synthesis, generative model, generative adversarial network, gan
TL;DR:Extend GAN architecture to obtain control over locations and identities of multiple objects within generated images.
Near-Optimal Representation Learning for Hierarchical Reinforcement Learning
Keywords:representation hierarchy reinforcement learning
TL;DR:We translate a bound on sub-optimality of representations to a practical training objective in the context of hierarchical reinforcement learning.
Understanding Composition of Word Embeddings via Tensor Decomposition
Keywords:word embeddings, semantic composition, tensor decomposition
TL;DR:We present a generative model for compositional word embeddings that captures syntactic relations, and provide empirical verification and evaluation.
Structured Neural Summarization
Keywords:Summarization, Graphs, Source Code
TL;DR:One simple trick to improve sequence models: Compose them with a graph model
Graph Wavelet Neural Network
Keywords:graph convolution, graph wavelet transform, graph Fourier transform, semi-supervised learning
TL;DR:We present graph wavelet neural network (GWNN), a novel graph convolutional neural network (CNN), leveraging graph wavelet transform to address the shortcoming of previous spectral graph CNN methods that depend on graph Fourier transform.
A rotation-equivariant convolutional neural network model of primary visual cortex
Keywords:rotation equivariance, equivariance, primary visual cortex, V1, neuroscience, system identification
TL;DR:A rotation-equivariant CNN model of V1 that outperforms previous models and suggest functional groupings of V1 neurons.
Supervised Community Detection with Line Graph Neural Networks
Keywords:community detection, graph neural networks, belief propagation, energy landscape, non-backtracking matrix
TL;DR:We propose a novel graph neural network architecture based on the non-backtracking matrix defined over the edge adjacencies and demonstrate its effectiveness in community detection tasks on graphs.
Multiple-Attribute Text Rewriting
Keywords:controllable text generation, generative models, conditional generative models, style transfer
TL;DR:A system for rewriting text conditioned on multiple controllable attributes
Wasserstein Barycenter Model Ensembling
Keywords:Wasserstein barycenter model ensembling
TL;DR:we propose to use Wasserstein barycenters for semantic model ensembling
Policy Transfer with Strategy Optimization
Keywords:transfer learning, reinforcement learning, modeling error, strategy optimization
TL;DR:We propose a policy transfer algorithm that can overcome large and challenging discrepancies in the system dynamics such as latency, actuator modeling error, etc.
code2seq: Generating Sequences from Structured Representations of Code
Keywords:source code, programs, code2seq
TL;DR:We leverage the syntactic structure of source code to generate natural language sequences.
Predict then Propagate: Graph Neural Networks meet Personalized PageRank
Keywords:Graph, GCN, GNN, Neural network, Graph neural network, Message passing neural network, Semi-supervised classification, Semi-supervised learning, PageRank, Personalized PageRank
TL;DR:Personalized propagation of neural predictions (PPNP) improves graph neural networks by separating them into prediction and propagation via personalized PageRank.
Slimmable Neural Networks
Keywords:Slimmable neural networks, mobile deep learning, accuracy-efficiency trade-offs
TL;DR:We present a simple and general method to train a single neural network executable at different widths (number of channels in a layer), permitting instant and adaptive accuracy-efficiency trade-offs at runtime.
Analysing Mathematical Reasoning Abilities of Neural Models
Keywords:mathematics, dataset, algebraic, reasoning
TL;DR:A dataset for testing mathematical reasoning (and algebraic generalization), and results on current sequence-to-sequence models.
RotDCF: Decomposition of Convolutional Filters for Rotation-Equivariant Deep Networks
Keywords:None
TL;DR:None
Execution-Guided Neural Program Synthesis
Keywords:None
TL;DR:None
Dynamic Sparse Graph for Efficient Deep Learning
Keywords:Sparsity, compression, training, acceleration
TL;DR:We construct dynamic sparse graph via dimension-reduction search to reduce compute and memory cost in both DNN training and inference.
Fixup Initialization: Residual Learning Without Normalization
Keywords:deep learning, residual networks, initialization, batch normalization, layer normalization
TL;DR:All you need to train deep residual networks is a good initialization; normalization layers are not necessary.
ProbGAN: Towards Probabilistic GAN with Theoretical Guarantees
Keywords:Generative Adversarial Networks, Bayesian Deep Learning, Mode Collapse, Inception Score, Generator, Discriminator, CIFAR-10, STL-10, ImageNet
TL;DR:A novel probabilistic treatment for GAN with theoretical guarantee.
Exploration by random network distillation
Keywords:reinforcement learning, exploration, curiosity
TL;DR:A simple exploration bonus is introduced and achieves state of the art performance in 3 hard exploration Atari games.
Unsupervised Learning of the Set of Local Maxima
Keywords:None
TL;DR:None
On the Convergence of A Class of Adam-Type Algorithms for Non-Convex Optimization
Keywords:nonconvex optimization, Adam, convergence analysis
TL;DR:We analyze convergence of Adam-type algorithms and provide mild sufficient conditions to guarantee their convergence, we also show violating the conditions can makes an algorithm diverge.
Minimum Divergence vs. Maximum Margin: an Empirical Comparison on Seq2Seq Models
Keywords:None
TL;DR:None
GANSynth: Adversarial Neural Audio Synthesis
Keywords:GAN, Audio, WaveNet, NSynth, Music
TL;DR:High-quality audio synthesis with GANs
Sliced Wasserstein Auto-Encoders
Keywords:optimal transport, Wasserstein distances, auto-encoders, unsupervised learning
TL;DR:In this paper we use the sliced-Wasserstein distance to shape the latent distribution of an auto-encoder into any samplable prior distribution.
Learning Two-layer Neural Networks with Symmetric Inputs
Keywords:Neural Network, Optimization, Symmetric Inputs, Moment-of-moments
TL;DR:We give an algorithm for learning a two-layer neural network with symmetric input distribution.
Learning to Understand Goal Specifications by Modelling Reward
Keywords:instruction following, reward modelling, language understanding
TL;DR:We propose AGILE, a framework for training agents to perform instructions from examples of respective goal-states.
Do Deep Generative Models Know What They Don't Know?
Keywords:None
TL;DR:None
Identifying and Controlling Important Neurons in Neural Machine Translation
Keywords:neural machine translation, individual neurons, unsupervised, analysis, correlation, translation control, distributivity, localization
TL;DR:Unsupervised methods for finding, analyzing, and controlling important neurons in NMT
Representing Formal Languages: A Comparison Between Finite Automata and Recurrent Neural Networks
Keywords:Language recognition, Recurrent Neural Networks, Representation Learning, deterministic finite automaton, automaton
TL;DR:Finite Automata Can be Linearly decoded from Language-Recognizing RNNs using low coarseness abstraction functions and high accuracy decoders.
Visual Explanation by Interpretation: Improving Visual Feedback Capabilities of Deep Neural Networks
Keywords:model explanation, model interpretation, explainable ai, evaluation
TL;DR:Interpretation by Identifying model-learned features that serve as indicators for the task of interest. Explain model decisions by highlighting the response of these features in test data. Evaluate explanations objectively with a controlled dataset.
Don't let your Discriminator be fooled
Keywords:GAN, generative models, computer vision
TL;DR:A discriminator that is not easily fooled by adversarial example makes GAN training more robust and leads to a smoother objective.
Latent Convolutional Models
Keywords:latent models, convolutional networks, unsupervised learning, deep learning, modeling natural images, image restoration
TL;DR:We present a new deep latent model of natural images that can be trained from unlabeled datasets and can be utilized to solve various image restoration tasks.
A Universal Music Translation Network
Keywords:None
TL;DR:None
How to train your MAML
Keywords:meta-learning, deep-learning, few-shot learning, supervised learning, neural-networks, stochastic optimization
TL;DR:MAML is great, but it has many problems, we solve many of those problems and as a result we learn most hyper parameters end to end, speed-up training and inference and set a new SOTA in few-shot learning
Learning a SAT Solver from Single-Bit Supervision
Keywords:sat, search, graph neural network, theorem proving, proof
TL;DR:We train a graph network to predict boolean satisfiability and show that it learns to search for solutions, and that the solutions it finds can be decoded from its activations.
Learning Representations of Sets through Optimized Permutations
Keywords:sets, representation learning, permutation invariance
TL;DR:Learn how to permute a set, then encode permuted set with RNN to obtain a set representation.
Big-Little Net: An Efficient Multi-Scale Feature Representation for Visual and Speech Recognition
Keywords:None
TL;DR:None
Unsupervised Hyper-alignment for Multilingual Word Embeddings
Keywords:None
TL;DR:None
Visual Semantic Navigation using Scene Priors
Keywords:None
TL;DR:None
NOODL: Provable Online Dictionary Learning and Sparse Coding
Keywords:provable dictionary learning, sparse coding, support recovery, iterative hard thresholding, matrix factorization, neural architectures, noodl
TL;DR:We present a provable algorithm for exactly recovering both factors of the dictionary learning model.
Stochastic Gradient/Mirror Descent: Minimax Optimality and Implicit Regularization
Keywords:None
TL;DR:None
Active Learning with Partial Feedback
Keywords:None
TL;DR:None
Gradient descent aligns the layers of deep linear networks
Keywords:None
TL;DR:None
Data-Dependent Coresets for Compressing Neural Networks with Applications to Generalization Bounds
Keywords:None
TL;DR:None
On the loss landscape of a class of deep neural networks with no bad local valleys
Keywords:None
TL;DR:None
DOM-Q-NET: Grounded RL on Structured Language
Keywords:Reinforcement Learning, Web Navigation, Graph Neural Networks
TL;DR:Graph-based Deep Q Network for Web Navigation
Boosting Robustness Certification of Neural Networks
Keywords:Robustness certification, Adversarial Attacks, Abstract Interpretation, MILP Solvers, Verification of Neural Networks
TL;DR:We refine the over-approximation results from incomplete verifiers using MILP solvers to prove more robustness properties than state-of-the-art.
Learning To Simulate
Keywords:Simulation in machine learning, reinforcement learning, policy gradients, image rendering
TL;DR:We propose an algorithm that automatically adjusts parameters of a simulation engine to generate training data for a neural network such that validation accuracy is maximized.
Towards Understanding Regularization in Batch Normalization
Keywords:None
TL;DR:None
The Laplacian in RL: Learning Representations with Efficient Approximations
Keywords:Laplacian, reinforcement learning, representation
TL;DR:We propose a scalable method to approximate the eigenvectors of the Laplacian in the reinforcement learning context and we show that the learned representations can improve the performance of an RL agent.
Predicting the Generalization Gap in Deep Networks with Margin Distributions
Keywords:Deep learning, large margin, generalization bounds, generalization gap.
TL;DR:We develop a new scheme to predict the generalization gap in deep networks with high accuracy.
Adversarial Imitation via Variational Inverse Reinforcement Learning
Keywords:Our method introduces the empowerment-regularized maximum-entropy inverse reinforcement learning to learn near-optimal rewards and policies from expert demonstrations.
TL;DR:Inverse Reinforcement Learning, Imitation learning, Variational lnference, Learning from demonstrations
Reasoning About Physical Interactions with Object-Oriented Prediction and Planning
Keywords:structured scene representation, predictive models, intuitive physics, self-supervised learning
TL;DR:We present a framework for learning object-centric representations suitable for planning in tasks that require an understanding of physics.
LayoutGAN: Generating Graphic Layouts with Wireframe Discriminators
Keywords:None
TL;DR:None
Learning Mixed-Curvature Representations in Product Spaces
Keywords:embeddings, non-Euclidean geometry, manifolds, geometry of data
TL;DR:Product manifold embedding spaces with heterogenous curvature yield improved representations compared to traditional embedding spaces for a variety of structures.
StrokeNet: A Neural Painting Environment
Keywords:image generation, differentiable model, reinforcement learning, deep learning, model based
TL;DR:StrokeNet is a novel architecture where the agent is trained to draw by strokes on a differentiable simulation of the environment, which could effectively exploit the power of back-propagation.
Harmonizing Maximum Likelihood with GANs for Multimodal Conditional Generation
Keywords:conditional GANs, conditional image generation, multimodal generation, reconstruction loss, maximum likelihood estimation, moment matching
TL;DR:We prove that the mode collapse in conditional GANs is largely attributed to a mismatch between reconstruction loss and GAN loss and introduce a set of novel loss functions as alternatives for reconstruction loss.
Measuring Compositionality in Representation Learning
Keywords:compositionality, representation learning, evaluation
TL;DR:This paper proposes a simple procedure for evaluating compositional structure in learned representations, and uses the procedure to explore the role of compositionality in four learning problems.
Benchmarking Neural Network Robustness to Common Corruptions and Perturbations
Keywords:robustness, benchmark, convnets, perturbations
TL;DR:We propose ImageNet-C to measure classifier corruption robustness and ImageNet-P to measure perturbation robustness
ADef: an Iterative Algorithm to Construct Adversarial Deformations
Keywords:Adversarial examples, deformations, deep neural networks, computer vision
TL;DR:We propose a new, efficient algorithm to construct adversarial examples by means of deformations, rather than additive perturbations.
Discriminator-Actor-Critic: Addressing Sample Inefficiency and Reward Bias in Adversarial Imitation Learning
Keywords:deep learning, reinforcement learning, imitation learning, adversarial learning
TL;DR:We address sample inefficiency and reward bias in adversarial imitation learning algorithms such as GAIL and AIRL.
Doubly Reparameterized Gradient Estimators for Monte Carlo Objectives
Keywords:variational autoencoder, reparameterization trick, IWAE, VAE, RWS, JVI
TL;DR:Doubly reparameterized gradient estimators provide unbiased variance reduction which leads to improved performance.
Learning Recurrent Binary/Ternary Weights
Keywords:Quantized Recurrent Neural Network, Hardware Implementation, Deep Learning
TL;DR:We propose high-performance LSTMs with binary/ternary weights, that can greatly reduce implementation complexity
Learning concise representations for regression by evolving networks of trees
Keywords:regression, stochastic optimization, evolutionary compution, feature engineering
TL;DR:Representing the network architecture as a set of syntax trees and optimizing their structure leads to accurate and concise regression models.
Efficient Training on Very Large Corpora via Gramian Estimation
Keywords:similarity learning, pairwise learning, matrix factorization, Gramian estimation, variance reduction, neural embedding models, recommender systems
TL;DR:We develop efficient methods to train neural embedding models with a dot-product structure, by reformulating the objective function in terms of generalized Gram matrices, and maintaining estimates of those matrices.
MAE: Mutual Posterior-Divergence Regularization for Variational AutoEncoders
Keywords:None
TL;DR:None
Residual Non-local Attention Networks for Image Restoration
Keywords:Non-local network, attention network, image restoration, residual learning
TL;DR:New state-of-the-art framework for image restoration
Meta-Learning For Stochastic Gradient MCMC
Keywords:Meta Learning, MCMC
TL;DR:This paper proposes a method to automate the design of stochastic gradient MCMC proposal using meta learning approach.
Systematic Generalization: What Is Required and Can It Be Learned?
Keywords:systematic generalization, language understanding, visual questions answering, neural module networks
TL;DR:We show that modular structured models are the best in terms of systematic generalization and that their end-to-end versions don't generalize as well.
Efficient Lifelong Learning with A-GEM
Keywords:Lifelong Learning, Continual Learning, Catastrophic Forgetting, Few-shot Transfer
TL;DR:An efficient lifelong learning algorithm that provides a better trade-off between accuracy and time/ memory complexity compared to other algorithms.
Multi-step Retriever-Reader Interaction for Scalable Open-domain Question Answering
Keywords:Open domain Question Answering, Reinforcement Learning, Query reformulation
TL;DR:Paragraph retriever and machine reader interacts with each other via reinforcement learning to yield large improvements on open domain datasets
Double Viterbi: Weight Encoding for High Compression Ratio and Fast On-Chip Reconstruction for Deep Neural Network
Keywords:quantization, pruning, memory footprint, model compression, sparse matrix
TL;DR:We present a new weight encoding scheme which enables high compression ratio and fast sparse-to-dense matrix conversion.
Overcoming the Disentanglement vs Reconstruction Trade-off via Jacobian Supervision
Keywords:disentangling, autoencoders, jacobian, face manipulation
TL;DR:A method for learning image representations that are good for both disentangling factors of variation and obtaining faithful reconstructions.
RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space
Keywords:knowledge graph embedding, knowledge graph completion, adversarial sampling
TL;DR:A new state-of-the-art approach for knowledge graph embedding.
Guiding Policies with Language via Meta-Learning
Keywords:meta-learning, language grounding, interactive
TL;DR:We propose a meta-learning method for interactively correcting policies with natural language.
AdaShift: Decorrelation and Convergence of Adaptive Learning Rate Methods
Keywords:optimizer, Adam, convergence, decorrelation
TL;DR:We analysis and solve the non-convergence issue of Adam.
AD-VAT: An Asymmetric Dueling mechanism for learning Visual Active Tracking
Keywords:Active tracking, reinforcement learning, adversarial learning, multi agent
TL;DR:We propose AD-VAT, where the tracker and the target object, viewed as two learnable agents, are opponents and can mutually enhance during training.
Marginal Policy Gradients: A Unified Family of Estimators for Bounded Action Spaces with Applications
Keywords:None
TL;DR:None
On Self Modulation for Generative Adversarial Networks
Keywords:unsupervised learning, generative adversarial networks, deep generative modelling
TL;DR:A simple GAN modification that improves performance across many losses, architectures, regularization schemes, and datasets.
Off-Policy Evaluation and Learning from Logged Bandit Feedback: Error Reduction via Surrogate Policy
Keywords:None
TL;DR:None
Subgradient Descent Learns Orthogonal Dictionaries
Keywords:Dictionary learning, Sparse coding, Non-convex optimization, Theory
TL;DR:Efficient dictionary learning by L1 minimization via a novel analysis of the non-convex non-smooth geometry.
ClariNet: Parallel Wave Generation in End-to-End Text-to-Speech
Keywords:None
TL;DR:None
MARGINALIZED AVERAGE ATTENTIONAL NETWORK FOR WEAKLY-SUPERVISED LEARNING
Keywords:feature aggregation, weakly supervised learning, temporal action localization
TL;DR:A novel marginalized average attentional network for weakly-supervised temporal action localization
Towards GAN Benchmarks Which Require Generalization
Keywords:evaluation, generative adversarial networks, adversarial divergences
TL;DR:We argue that GAN benchmarks must require a large sample from the model to penalize memorization and investigate whether neural network divergences have this property.
A Closer Look at Few-shot Classification
Keywords:few shot classification, meta-learning
TL;DR: A detailed empirical study in few-shot classification that revealing challenges in standard evaluation setting and showing a new direction.
Meta-Learning Probabilistic Inference for Prediction
Keywords:probabilistic models, approximate inference, few-shot learning, meta-learning
TL;DR:Novel framework for meta-learning that unifies and extends a broad class of existing few-shot learning methods. Achieves strong performance on few-shot learning benchmarks without requiring iterative test-time inference.
Deep reinforcement learning with relational inductive biases
Keywords:relational reasoning, reinforcement learning, graph neural networks, starcraft, generalization, inductive bias
TL;DR:Relational inductive biases improve out-of-distribution generalization capacities in model-free reinforcement learning agents
Relaxed Quantization for Discretized Neural Networks
Keywords:Quantization, Compression, Neural Networks, Efficiency
TL;DR:We introduce a technique that allows for gradient based training of quantized neural networks.
Tree-Structured Recurrent Switching Linear Dynamical Systems for Multi-Scale Modeling
Keywords:None
TL;DR:None
STCN: Stochastic Temporal Convolutional Networks
Keywords:latent variables, variational inference, temporal convolutional networks, sequence modeling, auto-regressive modeling
TL;DR:We combine the computational advantages of temporal convolutional architectures with the expressiveness of stochastic latent variables.
Soft Q-Learning with Mutual-Information Regularization
Keywords:None
TL;DR:None
On the Turing Completeness of Modern Neural Network Architectures
Keywords:Transformer, NeuralGPU, Turing completeness
TL;DR:We show that the Transformer architecture and the Neural GPU are Turing complete.
Improving Differentiable Neural Computers Through Memory Masking, De-allocation, and Link Distribution Sharpness Control
Keywords:None
TL;DR:None
Evaluating Robustness of Neural Networks with Mixed Integer Programming
Keywords:verification, adversarial robustness, adversarial examples, deep learning
TL;DR:We efficiently verify the robustness of deep neural models with over 100,000 ReLUs, certifying more samples than the state-of-the-art and finding more adversarial examples than a strong first-order attack.
Random mesh projectors for inverse problems
Keywords:imaging, inverse problems, subspace projections, random Delaunay triangulations, CNN, geophysics, regularization
TL;DR:We solve ill-posed inverse problems with scarce ground truth examples by estimating an ensemble of random projections of the model instead of the model itself.
Multi-Agent Dual Learning
Keywords:None
TL;DR:None
Complement Objective Training
Keywords:optimization, entropy, image recognition, natural language understanding, adversarial attacks, deep learning
TL;DR:We propose Complement Objective Training (COT), a new training paradigm that optimizes both the primary and complement objectives for effectively learning the parameters of neural networks.
Mode Normalization
Keywords:Deep Learning, Expert Models, Normalization, Computer Vision
TL;DR:We present a novel normalization method for deep neural networks that is robust to multi-modalities in intermediate feature distributions.
Detecting Egregious Responses in Neural Sequence-to-sequence Models
Keywords:Deep Learning, Natural Language Processing, Adversarial Attacks, Dialogue Response Generation
TL;DR:This paper aims to provide an empirical answer to the question of whether well-trained dialogue response model can output malicious responses.
Learning Actionable Representations with Goal Conditioned Policies
Keywords:Representation Learning, Reinforcement Learning
TL;DR:Learning state representations which capture factors necessary for control
Verification of Non-Linear Specifications for Neural Networks
Keywords:None
TL;DR:None
Generating Liquid Simulations with Deformation-aware Neural Networks
Keywords:Learning weighting and deformations of space-time data sets for highly efficient approximations of liquid behavior.
TL;DR:deformation learning, spatial transformer networks, fluid simulation
DyRep: Learning Representations over Dynamic Graphs
Keywords:Dynamic Graphs, Representation Learning, Dynamic Processes, Temporal Point Process, Attention, Latent Representation
TL;DR:Models Representation Learning over dynamic graphs as latent hidden process bridging two observed processes of Topological Evolution of and Interactions on dynamic graphs.
Trellis Networks for Sequence Modeling
Keywords:sequence modeling, language modeling, recurrent networks, convolutional networks, trellis networks
TL;DR:Trellis networks are a new sequence modeling architecture that bridges recurrent and convolutional models and sets a new state of the art on word- and character-level language modeling.
Scalable Unbalanced Optimal Transport using Generative Adversarial Networks
Keywords:unbalanced optimal transport, generative adversarial networks, population modeling
TL;DR:We propose new methodology for unbalanced optimal transport using generative adversarial networks.
Solving the Rubik's Cube with Approximate Policy Iteration
Keywords:reinforcement learning, Rubik's Cube, approximate policy iteration, deep learning, deep reinforcement learning
TL;DR:We solve the Rubik's Cube with pure reinforcement learning
Variance Reduction for Reinforcement Learning in Input-Driven Environments
Keywords:reinforcement learning, policy gradient, input-driven environments, variance reduction, baseline
TL;DR:For environments dictated partially by external input processes, we derive an input-dependent baseline that provably reduces the variance for policy gradient methods and improves the policy performance in a wide range of RL tasks.
Model-Predictive Policy Learning with Uncertainty Regularization for Driving in Dense Traffic
Keywords:model-based reinforcement learning, stochastic video prediction, autonomous driving
TL;DR:A model-based RL approach which uses a differentiable uncertainty penalty to learn driving policies from purely observational data.
GAN Dissection: Visualizing and Understanding Generative Adversarial Networks
Keywords:GAN representations are examined in detail, and sets of representation units are found that control the generation of semantic concepts in the output.
TL;DR:GANs, representation, interpretability, causality
Improving MMD-GAN Training with Repulsive Loss Function
Keywords:generative adversarial nets, loss function, maximum mean discrepancy, image generation, unsupervised learning
TL;DR:Rearranging the terms in maximum mean discrepancy yields a much better loss function for the discriminator of generative adversarial nets
Deterministic PAC-Bayesian generalization bounds for deep networks via generalizing noise-resilience
Keywords:generalization, PAC-Bayes, SGD, learning theory, implicit regularization
TL;DR:We provide a PAC-Bayes based generalization guarantee for uncompressed, deterministic deep networks by generalizing noise-resilience of the network on the training data to the test data.
Recall Traces: Backtracking Models for Efficient Reinforcement Learning
Keywords:Model free RL, Variational Inference
TL;DR:A backward model of previous (state, action) given the next state, i.e. P(s_t, a_t | s_{t+1}), can be used to simulate additional trajectories terminating at states of interest! Improves RL learning efficiency.
Stable Recurrent Models
Keywords:stability, gradient descent, non-convex optimization, recurrent neural networks
TL;DR:Stable recurrent models can be approximated by feed-forward networks and empirically perform as well as unstable models on benchmark tasks.
The Limitations of Adversarial Training and the Blind-Spot Attack
Keywords:Adversarial Examples, Adversarial Training, Blind-Spot Attack
TL;DR:We show that even the strongest adversarial training methods cannot defend against adversarial examples crafted on slightly scaled and shifted test images.
Efficiently testing local optimality and escaping saddles for ReLU networks
Keywords:local optimality, second-order stationary point, escaping saddle points, nondifferentiability, ReLU, empirical risk
TL;DR:A theoretical algorithm for testing local optimality and extracting descent directions at nondifferentiable points of empirical risks of one-hidden-layer ReLU networks.
ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware
Keywords:Neural Architecture Search, Efficient Neural Networks
TL;DR:Proxy-less neural architecture search for directly learning architectures on large-scale target task (ImageNet) while reducing the cost to the same level of normal training.
Hierarchical Reinforcement Learning via Advantage-Weighted Information Maximization
Keywords:Hierarchical reinforcement learning, Representation learning, Continuous control
TL;DR:This paper presents a hierarchical reinforcement learning framework based on deterministic option policies and mutual information maximization.
Generalizable Adversarial Training via Spectral Normalization
Keywords:None
TL;DR:None
Adversarial Domain Adaptation for Stable Brain-Machine Interfaces
Keywords:Brain-Machine Interfaces, Domain Adaptation, Adversarial Networks
TL;DR:We implement an adversarial domain adaptation network to stabilize a fixed Brain-Machine Interface against gradual changes in the recorded neural signals.
Deep Online Learning Via Meta-Learning: Continual Adaptation for Model-Based RL
Keywords:None
TL;DR:None
Deep Anomaly Detection with Outlier Exposure
Keywords:confidence, uncertainty, anomaly, robustness
TL;DR:OE teaches anomaly detectors to learn heuristics for detecting unseen anomalies; experiments are in classification, density estimation, and calibration in NLP and vision settings; we do not tune on test distribution samples, unlike previous work
Contingency-Aware Exploration in Reinforcement Learning
Keywords:Reinforcement Learning, Exploration, Contingency-Awareness
TL;DR:We investigate contingency-awareness and controllable aspects in exploration and achieve state-of-the-art performance on Montezuma's Revenge without expert demonstrations.
Context-adaptive Entropy Model for End-to-end Optimized Image Compression
Keywords:image compression, deep learning, entropy model
TL;DR:Context-adaptive entropy model for use in end-to-end optimized image compression, which significantly improves compression performance
Variational Discriminator Bottleneck: Improving Imitation Learning, Inverse RL, and GANs by Constraining Information Flow
Keywords:reinforcement learning, generative adversarial networks, imitation learning, inverse reinforcement learning, information bottleneck
TL;DR:Regularizing adversarial learning with an information bottleneck, applied to imitation learning, inverse reinforcement learning, and generative adversarial networks.
Meta-learning with differentiable closed-form solvers
Keywords:few-shot learning, one-shot learning, meta-learning, deep learning, ridge regression, classification
TL;DR:We propose a meta-learning approach for few-shot classification that achieves strong performance at high-speed by back-propagating through the solution of fast solvers, such as ridge regression or logistic regression.
Learning Self-Imitating Diverse Policies
Keywords:Reinforcement-learning, Imitation-learning, Ensemble-training
TL;DR:Policy optimization by using past good rollouts from the agent; learning shaped rewards via divergence minimization; SVPG with JS-kernel for population-based exploration.
ProxQuant: Quantized Neural Networks via Proximal Operators
Keywords:Model quantization, Optimization, Regularization
TL;DR:A principled framework for model quantization using the proximal gradient method, with empirical evaluation and theoretical convergence analyses.
Universal Transformers
Keywords:sequence-to-sequence, rnn, transformer, machine translation, language understanding, learning to execute
TL;DR:We introduce the Universal Transformer, a self-attentive parallel-in-time recurrent sequence model that outperforms Transformers and LSTMs on a wide range of sequence-to-sequence tasks, including machine translation.
Learning to Adapt in Dynamic, Real-World Environments through Meta-Reinforcement Learning
Keywords:meta-learning, reinforcement learning, meta reinforcement learning, online adaptation
TL;DR:A model-based meta-RL algorithm that enables a real robot to adapt online in dynamic environments
L-Shapley and C-Shapley: Efficient Model Interpretation for Structured Data
Keywords:Model Interpretation, Feature Selection
TL;DR:We develop two linear-complexity algorithms for model-agnostic model interpretation based on the Shapley value, in the settings where the contribution of features to the target is well-approximated by a graph-structured factorization.
Discovery of Natural Language Concepts in Individual Units of CNNs
Keywords:interpretability of deep neural networks, natural language representation
TL;DR:We show that individual units in CNN representations learned in NLP tasks are selectively responsive to natural language concepts.
Towards the first adversarially robust neural network model on MNIST
Keywords:None
TL;DR:None
Discriminator Rejection Sampling
Keywords:GANs, rejection sampling
TL;DR:We use a GAN discriminator to perform an approximate rejection sampling scheme on the output of the GAN generator.
Harmonic Unpaired Image-to-image Translation
Keywords:unpaired image-to-image translation, cyclegan, smoothness constraint
TL;DR:Smooth regularization over sample graph for unpaired image-to-image translation results in significantly improved consistency
Universal Successor Features Approximators
Keywords:None
TL;DR:None
Gradient Descent Provably Optimizes Over-parameterized Neural Networks
Keywords:theory, non-convex optimization, overparameterization, gradient descent
TL;DR:We prove gradient descent achieves zero training loss with a linear rate on over-parameterized neural networks.
Opportunistic Learning: Budgeted Cost-Sensitive Learning from Data Streams
Keywords:Cost-Aware Learning, Feature Acquisition, Reinforcement Learning, Stream Learning, Deep Q-Learning
TL;DR:An online algorithm for cost-aware feature acquisition and prediction
DARTS: Differentiable Architecture Search
Keywords:deep learning, autoML, neural architecture search, image classification, language modeling
TL;DR:We propose a differentiable architecture search algorithm for both convolutional and recurrent networks, achieving competitive performance with the state of the art using orders of magnitude less computation resources.
Feature-Wise Bias Amplification
Keywords:None
TL;DR:None
The relativistic discriminator: a key element missing from standard GAN
Keywords:Improving the quality and stability of GANs using a relativistic discriminator; IPM GANs (such as WGAN-GP) are a special case.
TL;DR:AI, deep learning, generative models, GAN
Understanding and Improving Interpolation in Autoencoders via an Adversarial Regularizer
Keywords:We propose a regularizer that improves interpolation and autoencoders and show that it also improves the learned representation for downstream tasks.
TL;DR:autoencoders, interpolation, unsupervised learning, representation learning, adversarial learning
Quasi-hyperbolic momentum and Adam for deep learning
Keywords:sgd, momentum, nesterov, adam, qhm, qhadam, optimization
TL;DR:Mix plain SGD and momentum (or do something similar with Adam) for great profit.
Local SGD Converges Fast and Communicates Little
Keywords:optimization, communication, theory, stochastic gradient descent, SGD, mini-batch, local SGD, parallel restart SGD, distributed training
TL;DR:We prove that parallel local SGD achieves linear speedup with much lesser communication than parallel mini-batch SGD.
Learning Finite State Representations of Recurrent Policy Networks
Keywords:recurrent neural networks, finite state machine, quantization, interpretability, autoencoder, moore machine, reinforcement learning, imitation learning, representation, Atari, Tomita
TL;DR:Extracting a finite state machine from a recurrent neural network via quantization for the purpose of interpretability with experiments on Atari.
Multilingual Neural Machine Translation with Knowledge Distillation
Keywords:NMT, Multilingual NMT, Knowledge Distillation
TL;DR:We proposed a knowledge distillation based method to boost the accuracy of multilingual neural machine translation.
MisGAN: Learning from Incomplete Data with Generative Adversarial Networks
Keywords:generative models, missing data
TL;DR:This paper presents a GAN-based framework for learning the distribution from high-dimensional incomplete data.
A Direct Approach to Robust Deep Learning Using Adversarial Networks
Keywords:deep learning, adversarial learning, generative adversarial networks
TL;DR:Jointly train an adversarial noise generating network with a classification network to provide better robustness to adversarial attacks.
Combinatorial Attacks on Binarized Neural Networks
Keywords:binarized neural networks, combinatorial optimization, integer programming
TL;DR:Gradient-based attacks on binarized neural networks are not effective due to the non-differentiability of such networks; Our IPROP algorithm solves this problem using integer optimization
Exemplar Guided Unsupervised Image-to-Image Translation with Semantic Consistency
Keywords:image-to-image translation, image generation, domain adaptation
TL;DR:We propose the Exemplar Guided & Semantically Consistent Image-to-image Translation (EGSC-IT) network which conditions the translation process on an exemplar image in the target domain.
ARM: Augment-REINFORCE-Merge Gradient for Stochastic Binary Networks
Keywords:Antithetic sampling, variable augmentation, deep discrete latent variable models, variance reduction, variational auto-encoder
TL;DR:An unbiased and low-variance gradient estimator for discrete latent variable models
Building Dynamic Knowledge Graphs from Text using Machine Reading Comprehension
Keywords:None
TL;DR:None
Information asymmetry in KL-regularized RL
Keywords:Deep Reinforcement Learning, Continuous Control, RL as Inference
TL;DR:Limiting state information for the default policy can improvement performance, in a KL-regularized RL framework where both agent and default policy are optimized together
TimbreTron: A WaveNet(CycleGAN(CQT(Audio))) Pipeline for Musical Timbre Transfer
Keywords:Generative models, Timbre Transfer, Wavenet, CycleGAN
TL;DR:We present the TimbreTron, a pipeline for perfoming high-quality timbre transfer on musical waveforms using CQT-domain style transfer.
Whitening and Coloring Batch Transform for GANs
Keywords:None
TL;DR:None
Learnable Embedding Space for Efficient Neural Architecture Compression
Keywords:Network Compression, Neural Architecture Search, Bayesian Optimization, Architecture Embedding
TL;DR:We propose a method to incrementally learn an embedding space over the domain of network architectures, to enable the careful selection of architectures for evaluation during compressed architecture search.
On the Sensitivity of Adversarial Robustness to Input Data Distributions
Keywords:adversarial robustness, adversarial training, PGD training, adversarial perturbation, input data distribution
TL;DR:Robustness performance of PGD trained models are sensitive to semantics-preserving transformation of image datasets, which implies the trickiness of evaluation of robust learning algorithms in practice.
Minimal Images in Deep Neural Networks: Fragile Object Recognition in Natural Images
Keywords:None
TL;DR:None
A Statistical Approach to Assessing Neural Network Robustness
Keywords:neural network verification, multi-level splitting, formal verification
TL;DR:We introduce a statistical approach to assessing neural network robustness that provides an informative notion of how robust a network is, rather than just the conventional binary assertion of whether or not of property is violated.
Improving Sequence-to-Sequence Learning via Optimal Transport
Keywords:None
TL;DR:None
PATE-GAN: Generating Synthetic Data with Differential Privacy Guarantees
Keywords:None
TL;DR:None
Integer Networks for Data Compression with Latent-Variable Models
Keywords:data compression, variational models, network quantization
TL;DR:We train variational models with quantized networks for computational determinism. This enables using them for cross-platform data compression.
Value Propagation Networks
Keywords:Reinforcement Learning, Value Iteration, Navigation, Convolutional Neural Networks, Learning to plan
TL;DR:We present planners based on convnets that are sample-efficient and that generalize to larger instances of navigation and pathfinding problems.
Bayesian Policy Optimization for Model Uncertainty
Keywords:Bayes-Adaptive Markov Decision Process, Model Uncertainty, Bayes Policy Optimization
TL;DR:We formulate model uncertainty in Reinforcement Learning as a continuous Bayes-Adaptive Markov Decision Process and present a method for practical and scalable Bayesian policy optimization.
Variational Bayesian Phylogenetic Inference
Keywords:Bayesian phylogenetic inference, Variational inference, Subsplit Bayesian networks
TL;DR:The first variational Bayes formulation of phylogenetic inference, a challenging inference problem over structures with intertwined discrete and continuous components
LEARNING FACTORIZED REPRESENTATIONS FOR OPEN-SET DOMAIN ADAPTATION
Keywords:None
TL;DR:None
On the Universal Approximability and Complexity Bounds of Quantized ReLU Neural Networks
Keywords:Quantized Neural Networks, Universial Approximability, Complexity Bounds, Optimal Bit-width
TL;DR:This paper proves the universal approximability of quantized ReLU neural networks and puts forward the complexity bound given arbitrary error.
Learning Localized Generative Models for 3D Point Clouds via Graph Convolution
Keywords:A GAN using graph convolution operations with dynamically computed graphs from hidden features
TL;DR:GAN, graph convolution, point clouds
ACCELERATING NONCONVEX LEARNING VIA REPLICA EXCHANGE LANGEVIN DIFFUSION
Keywords:None
TL;DR:None
Dynamically Unfolding Recurrent Restorer: A Moving Endpoint Control Method for Image Restoration
Keywords:image restoration, differential equation
TL;DR:We propose a novel method to handle image degradations of different levels by learning a diffusion terminal time. Our model can generalize to unseen degradation level and different noise statistic.
Bias-Reduced Uncertainty Estimation for Deep Neural Classifiers
Keywords:Uncertainty estimation, Deep learning
TL;DR:We use snapshots from the training process to improve any uncertainty estimation method of a DNN classifier.
CAMOU: Learning Physical Vehicle Camouflages to Adversarially Attack Detectors in the Wild
Keywords:Adversarial Attack, Object Detection, Synthetic Simulation
TL;DR:We propose a method to learn physical vehicle camouflage to adversarially attack object detectors in the wild. We find our camouflage effective and transferable.
Learning Latent Superstructures in Variational Autoencoders for Deep Multidimensional Clustering
Keywords:latent tree model, variational autoencoder, deep learning, latent variable model, bayesian network, structure learning, stepwise em, message passing, graphical model, multidimensional clustering, unsupervised learning
TL;DR:We investigate a variant of variational autoencoders where there is a superstructure of discrete latent variables on top of the latent features.
Learning Programmatically Structured Representations with Perceptor Gradients
Keywords:None
TL;DR:None
Variational Autoencoders with Jointly Optimized Latent Dependency Structure
Keywords:deep generative models, structure learning
TL;DR:We propose a method for learning latent dependency structure in variational autoencoders.
The Unusual Effectiveness of Averaging in GAN Training
Keywords:None
TL;DR:None
Beyond Pixel Norm-Balls: Parametric Adversaries using an Analytically Differentiable Renderer
Keywords:adversarial examples, norm-balls, differentiable renderer
TL;DR:Enabled by a novel differentiable renderer, we propose a new metric that has real-world implications for evaluating adversarial machine learning algorithms, resolving the lack of realism of the existing metric based on pixel norms.
Diversity is All You Need: Learning Skills without a Reward Function
Keywords:reinforcement learning, unsupervised learning, skill discovery
TL;DR:We propose an algorithm for learning useful skills without a reward function, and show how these skills can be used to solve downstream tasks.
Supervised Policy Update for Deep Reinforcement Learning
Keywords:Deep Reinforcement Learning
TL;DR:first posing and solving the sample efficiency optimization problem in the non-parameterized policy space, and then solving a supervised regression problem to find a parameterized policy that is near the optimal non-parameterized policy.
Learning sparse relational transition models
Keywords:Deictic reference, relational model, rule-based transition model
TL;DR:A new approach that learns a representation for describing transition models in complex uncertaindomains using relational rules.
Learning to Schedule Communication in Multi-agent Reinforcement Learning
Keywords:None
TL;DR:None
Hierarchical RL Using an Ensemble of Proprioceptive Periodic Policies
Keywords:None
TL;DR:None
Multi-class classification without multi-class labels
Keywords:None
TL;DR:None
What do you learn from context? Probing for sentence structure in contextualized word representations
Keywords:natural language processing, word embeddings, transfer learning, interpretability
TL;DR:We probe for sentence structure in ELMo and related contextual embedding models. We find existing models efficiently encode syntax and show evidence of long-range dependencies, but only offer small improvements on semantic tasks.
Spectral Inference Networks: Unifying Deep and Spectral Learning
Keywords:spectral learning, unsupervised learning, manifold learning, dimensionality reduction
TL;DR:We show how to learn spectral decompositions of linear operators with deep learning, and use it for unsupervised learning without a generative model.
PeerNets: Exploiting Peer Wisdom Against Adversarial Attacks
Keywords:None
TL;DR:None
Attentive Neural Processes
Keywords:Neural Processes, Conditional Neural Processes, Stochastic Processes, Regression, Attention
TL;DR:A model for regression that learns conditional distributions of a stochastic process, by incorporating attention into Neural Processes.
Representation Degeneration Problem in Training Natural Language Generation Models
Keywords:None
TL;DR:None
Hierarchical interpretations for neural network predictions
Keywords:interpretability, natural language processing, computer vision
TL;DR:We introduce and validate hierarchical local interpretations, the first technique to automatically search for and display important interactions for individual predictions made by LSTMs and CNNs.
Spreading vectors for similarity search
Keywords:dimensionality reduction, similarity search, indexing, differential entropy
TL;DR:We learn a neural network that uniformizes the input distribution, which leads to competitive indexing performance in high-dimensional space
A Convergence Analysis of Gradient Descent for Deep Linear Neural Networks
Keywords:Deep Learning, Learning Theory, Non-Convex Optimization
TL;DR:We analyze gradient descent for deep linear neural networks, providing a guarantee of convergence to global optimum at a linear rate.
Feed-forward Propagation in Probabilistic Neural Networks with Categorical and Max Layers
Keywords:probabilistic neural network, uncertainty, dropout, bayesian, softmax, argmax, logsumexp
TL;DR:Approximating mean and variance of the NN output over noisy input / dropout / uncertain parameters. Analytic approximations for argmax, softmax and max layers.
Measuring and regularizing networks in function space
Keywords:function space, Hilbert space, empirical characterization, multitask learning, catastrophic forgetting, optimization, natural gradient
TL;DR:We find movement in function space is not proportional to movement in parameter space during optimization. We propose a new natural-gradient style optimizer to address this.
Fluctuation-dissipation relations for stochastic gradient descent
Keywords:stochastic gradient descent, adaptive method, loss surface, Hessian
TL;DR:We prove fluctuation-dissipation relations for SGD, which can be used to (i) adaptively set learning rates and (ii) probe loss surfaces.
Poincare Glove: Hyperbolic Word Embeddings
Keywords:word embeddings, hyperbolic spaces, poincare ball, hypernymy, analogy, similarity, gaussian embeddings
TL;DR:We embed words in the hyperbolic space and make the connection with the Gaussian word embeddings.
Episodic Curiosity through Reachability
Keywords:deep learning, reinforcement learning, curiosity, exploration, episodic memory
TL;DR:We propose a novel model of curiosity based on episodic memory and the ideas of reachability which allows us to overcome the known "couch-potato" issues of prior work.
Phase-Aware Speech Enhancement with Deep Complex U-Net
Keywords:speech enhancement, deep learning, complex neural networks, phase estimation
TL;DR:This paper proposes a novel complex masking method for speech enhancement along with a loss function for efficient phase estimation.
Generative predecessor models for sample-efficient imitation learning
Keywords:None
TL;DR:None
Adaptive Estimators Show Information Compression in Deep Neural Networks
Keywords:deep neural networks, mutual information, information bottleneck, noise, L2 regularization
TL;DR:We developed robust mutual information estimates for DNNs and used them to observe compression in networks with non-saturating activation functions
Multilingual Neural Machine Translation With Soft Decoupled Encoding
Keywords:None
TL;DR:None
Approximating CNNs with Bag-of-local-Features models works surprisingly well on ImageNet
Keywords:interpretability, representation learning, bag of features, deep learning, object recognition
TL;DR:Aggregating class evidence from many small image patches suffices to solve ImageNet, yields more interpretable models and can explain aspects of the decision-making of popular DNNs.
Reward Constrained Policy Optimization
Keywords:reinforcement learning, markov decision process, constrained markov decision process, deep learning
TL;DR:For complex constraints in which it is not easy to estimate the gradient, we use the discounted penalty as a guiding signal. We prove that under certain assumptions it converges to a feasible solution.
On the Relation Between the Sharpest Directions of DNN Loss and the SGD Step Length
Keywords:optimization, generalization, theory of deep learning, SGD, hessian
TL;DR:SGD is steered early on in training towards a region in which its step is too large compared to curvature, which impacts the rest of training.
Modeling the Long Term Future in Model-Based Reinforcement Learning
Keywords:model-based reinforcement learning, variation inference
TL;DR:incorporating, in the model, latent variables that encode future content improves the long-term prediction accuracy, which is critical for better planning in model-based RL.
Understanding Straight-Through Estimator in Training Activation Quantized Neural Nets
Keywords:straight-through estimator, quantized activation, binary neuron
TL;DR:We make theoretical justification for the concept of straight-through estimator.
DISTRIBUTIONAL CONCAVITY REGULARIZATION FOR GANS
Keywords:None
TL;DR:None
LeMoNADe: Learned Motif and Neuronal Assembly Detection in calcium imaging videos
Keywords:VAE, unsupervised learning, neuronal assemblies, calcium imaging analysis
TL;DR:We present LeMoNADe, an end-to-end learned motif detection method directly operating on calcium imaging videos.
Competitive experience replay
Keywords:reinforcement learning, sparse reward, goal-based learning
TL;DR:a novel method to learn with sparse reward using adversarial reward re-labeling
Multi-Domain Adversarial Learning
Keywords:multi-domain learning, domain adaptation, adversarial learning, H-divergence, deep representation learning, high-content microscopy
TL;DR:Adversarial Domain adaptation and Multi-domain learning: a new loss to handle multi- and single-domain classes in the semi-supervised setting.
ProMP: Proximal Meta-Policy Search
Keywords:Meta-Reinforcement Learning, Meta-Learning, Reinforcement-Learning
TL;DR:A novel and theoretically grounded meta-reinforcement learning algorithm
Don't Settle for Average, Go for the Max: Fuzzy Sets and Max-Pooled Word Vectors
Keywords:word vectors, sentence representations, distributed representations, fuzzy sets, bag-of-words, unsupervised learning, word vector compositionality, max-pooling, Jaccard index
TL;DR:Max-pooled word vectors with fuzzy Jaccard set similarity are an extremely competitive baseline for semantic similarity; we propose a simple dynamic variant that performs even better.
Stable Opponent Shaping in Differentiable Games
Keywords:multi-agent learning, multiple interacting losses, opponent shaping, exploitation, convergence
TL;DR:Opponent shaping is a powerful approach to multi-agent learning but can prevent convergence; our SOS algorithm fixes this with strong guarantees in all differentiable games.
A Mean Field Theory of Batch Normalization
Keywords:theory, batch normalization, mean field theory, trainability
TL;DR:Batch normalization causes exploding gradients in vanilla feedforward networks.
Learning Exploration Policies for Navigation
Keywords:None
TL;DR:None
Distribution-Interpolation Trade off in Generative Models
Keywords:generative models, latent distribution, Cauchy distribution, interpolations
TL;DR:We theoretically prove that linear interpolations are unsuitable for analysis of trained implicit generative models.
Learning to Describe Scenes with Programs
Keywords:Structured scene representations, program synthesis
TL;DR:We present scene programs, a structured scene representation that captures both low-level object appearance and high-level regularity in the scene.
Visceral Machines: Risk-Aversion in Reinforcement Learning with Intrinsic Physiological Rewards
Keywords:Reinforcement Learning, Simulation, Affective Computing
TL;DR:We present a novel approach to reinforcement learning that leverages a task-independent intrinsic reward function trained on peripheral pulse measurements that are correlated with human autonomic nervous system responses.
Deep Frank-Wolfe For Neural Network Optimization
Keywords:optimization, conditional gradient, Frank-Wolfe, SVM
TL;DR:We train neural networks by locally linearizing them and using a linear SVM solver (Frank-Wolfe) at each iteration.
LEARNING TO PROPAGATE LABELS: TRANSDUCTIVE PROPAGATION NETWORK FOR FEW-SHOT LEARNING
Keywords:few-shot learning, meta-learning, label propagation, manifold learning
TL;DR:We propose a novel meta-learning framework for transductive inference that classifies the entire test set at once to alleviate the low-data problem.
Improving the Generalization of Adversarial Training with Domain Adaptation
Keywords:adversarial training, domain adaptation, adversarial example, deep learning
TL;DR:We propose a novel adversarial training with domain adaptation method that significantly improves the generalization ability on adversarial examples from different attacks.
Dimensionality Reduction for Representing the Knowledge of Probabilistic Models
Keywords:metric learning, distance learning, dimensionality reduction, bound guarantees
TL;DR:dimensionality reduction for cases where examples can be represented as soft probability distributions
Learning protein sequence embeddings using information from structure
Keywords:sequence embedding, sequence alignment, RNN, LSTM, protein structure, amino acid sequence, contextual embeddings, transmembrane prediction
TL;DR:We present a method for learning protein sequence embedding models using structural information in the form of global structural similarity between proteins and within protein residue-residue contacts.
Variational Smoothing in Recurrent Neural Network Language Models
Keywords:None
TL;DR:None
Biologically-Plausible Learning Algorithms Can Scale to Large Datasets
Keywords:biologically plausible learning algorithm, ImageNet, sign-symmetry, feedback alignment
TL;DR:Biologically plausible learning algorithms, particularly sign-symmetry, work well on ImageNet
Coarse-grain Fine-grain Coattention Network for Multi-evidence Question Answering
Keywords:question answering, reading comprehension, nlp, natural language processing, attention, representation learning
TL;DR:A new state-of-the-art model for multi-evidence question answering using coarse-grain fine-grain hierarchical attention.
Learning a Meta-Solver for Syntax-Guided Program Synthesis
Keywords:Syntax-guided Synthesis, Context Free Grammar, Logical Specification, Representation Learning, Meta Learning, Reinforcement Learning
TL;DR:We propose a meta-learning framework that learns a transferable policy from only weak supervision to solve synthesis tasks with different logical specifications and grammars.
Towards Robust, Locally Linear Deep Networks
Keywords:robust derivatives, transparency, interpretability
TL;DR:A scalable algorithm to establish robust derivatives of deep networks w.r.t. the inputs.
How Important is a Neuron
Keywords:None
TL;DR:None
Learning to Make Analogies by Contrasting Abstract Relational Structure
Keywords:cognitive science, analogy, psychology, cognitive theory, cognition, abstraction, generalization
TL;DR:The most robust capacity for analogical reasoning is induced when networks learn analogies by contrasting abstract relational structures in their input domains.
Learning what you can do before doing anything
Keywords:unsupervised learning, vision, motion, action space, video prediction, variational models
TL;DR:We learn a representation of an agent's action space from pure visual observations. We use a recurrent latent variable approach with a novel composability loss.
Learning Grid Cells as Vector Representation of Self-Position Coupled with Matrix Representation of Self-Motion
Keywords:None
TL;DR:None
Universal Stagewise Learning for Non-Convex Problems with Convergence on Averaged Solutions
Keywords:None
TL;DR:None
Invariant and Equivariant Graph Networks
Keywords:graph learning, equivariance, deep learning
TL;DR:The paper provides a full characterization of permutation invariant and equivariant linear layers for graph data.
Robustness May Be at Odds with Accuracy
Keywords:adversarial examples, robust machine learning, robust optimization, deep feature representations
TL;DR:We show that adversarial robustness might come at the cost of standard classification performance, but also yields unexpected benefits.
Feature Intertwiner for Object Detection
Keywords:feature learning, computer vision, deep learning
TL;DR:(Camera-ready version) A feature intertwiner module to leverage features from one accurate set to help the learning of another less reliable set.
Adversarial Reprogramming of Neural Networks
Keywords:Adversarial, Neural Networks, Machine Learning Security
TL;DR:We introduce the first instance of adversarial attacks that reprogram the target model to perform a task chosen by the attacker---without the attacker needing to specify or compute the desired output for each test-time input.
G-SGD: Optimizing ReLU Neural Networks in its Positively Scale-Invariant Space
Keywords:None
TL;DR:None
From Hard to Soft: Understanding Deep Network Nonlinearities via Vector Quantization and Statistical Inference
Keywords:Spline, Vector Quantization, Inference, Nonlinearities, Deep Network
TL;DR:Reformulate deep networks nonlinearities from a vector quantization scope and bridge most known nonlinearities together.
Aggregated Momentum: Stability Through Passive Damping
Keywords:momentum, optimization, deep learning, neural networks
TL;DR:We introduce a simple variant of momentum optimization which is able to outperform classical momentum, Nesterov, and Adam on deep learning tasks with minimal hyperparameter tuning.
Variational Autoencoder with Arbitrary Conditioning
Keywords:unsupervised learning, generative models, conditional variational autoencoder, variational autoencoder, missing features multiple imputation, inpainting
TL;DR:We propose an extension of conditional variational autoencoder that allows conditioning on an arbitrary subset of the features and sampling the remaining ones.
Time-Agnostic Prediction: Predicting Predictable Video Frames
Keywords:visual prediction, subgoal generation, bottleneck states, time-agnostic
TL;DR:In visual prediction tasks, letting your predictive model choose which times to predict does two things: (i) improves prediction quality, and (ii) leads to semantically coherent "bottleneck state" predictions, which are useful for planning.
A Closer Look at Deep Learning Heuristics: Learning rate restarts, Warmup and Distillation
Keywords:deep learning heuristics, learning rate restarts, learning rate warmup, knowledge distillation, mode connectivity, SVCCA
TL;DR:We use empirical tools of mode connectivity and SVCCA to investigate neural network training heuristics of learning rate restarts, warmup and knowledge distillation.
Self-Monitoring Navigation Agent via Auxiliary Progress Estimation
Keywords:visual grounding, textual grounding, instruction-following, navigation agent
TL;DR:We propose a self-monitoring agent for the Vision-and-Language Navigation task.
Kernel Change-point Detection with Auxiliary Deep Generative Models
Keywords:deep kernel learning, generative models, kernel two-sample test, time series change-point detection
TL;DR:In this paper, we propose KL-CPD, a novel kernel learning framework for time series CPD that optimizes a lower bound of test power via an auxiliary generative model as a surrogate to the abnormal distribution.
Unsupervised Learning via Meta-Learning
Keywords:unsupervised learning, meta-learning
TL;DR:An unsupervised learning method that uses meta-learning to enable efficient learning of downstream image classification tasks, outperforming state-of-the-art methods.
Auxiliary Variational MCMC
Keywords:None
TL;DR:None
Neural network gradient-based learning of black-box function interfaces
Keywords:neural networks, black box functions, gradient descent
TL;DR:Training DNNs to interface w black box functions wo intermediate labels by using an estimator sub-network that can be replaced with the black box after training
Self-Tuning Networks: Bilevel Optimization of Hyperparameters using Structured Best-Response Functions
Keywords:hyperparameter optimization, game theory, optimization
TL;DR:We use a hypernetwork to predict optimal weights given hyperparameters, and jointly train everything together.
Unsupervised Control Through Non-Parametric Discriminative Rewards
Keywords:deep reinforcement learning, goals, UVFA, mutual information
TL;DR:Unsupervised reinforcement learning method for learning a policy to robustly achieve perceptually specified goals.
Interpolation-Prediction Networks for Irregularly Sampled Time Series
Keywords:irregular sampling, multivariate time series, supervised learning, interpolation, missing data
TL;DR:This paper presents a new deep learning architecture for addressing the problem of supervised learning with sparse and irregularly sampled multivariate time series.
Riemannian Adaptive Optimization Methods
Keywords:Riemannian optimization, adaptive, hyperbolic, curvature, manifold, adam, amsgrad, adagrad, rsgd, convergence
TL;DR:Adapting Adam, Amsgrad, Adagrad to Riemannian manifolds.
Minimal Random Code Learning: Getting Bits Back from Compressed Model Parameters
Keywords:compression, neural networks, bits-back argument, Bayesian, Shannon, information theory
TL;DR:This paper proposes an effective method to compress neural networks based on recent results in information theory.
Characterizing Audio Adversarial Examples Using Temporal Dependency
Keywords:audio adversarial example, mitigation, detection, machine learning
TL;DR:Adversarial audio discrimination using temporal dependency
Equi-normalization of Neural Networks
Keywords:convolutional neural networks, Normalization, Sinkhorn, Regularization
TL;DR:Fast iterative algorithm to balance the energy of a network while staying in the same functional equivalence class
Generalized Tensor Models for Recurrent Neural Networks
Keywords:expressive power, recurrent neural networks, Tensor-Train decomposition
TL;DR:Analysis of expressivity and generality of recurrent neural networks with ReLu nonlinearities using Tensor-Train decomposition.
Wizard of Wikipedia: Knowledge-Powered Conversational Agents
Keywords:dialogue, knowledge, language, conversation
TL;DR:We build knowledgeable conversational agents by conditioning on Wikipedia + a new supervised task.
Are adversarial examples inevitable?
Keywords:adversarial examples, neural networks, security
TL;DR:This paper identifies classes of problems for which adversarial examples are inescapable, and derives fundamental bounds on the susceptibility of any classifier to adversarial examples.
A Variational Inequality Perspective on Generative Adversarial Networks
Keywords:optimization, variational inequality, games, saddle point, extrapolation, averaging, extragradient, generative modeling, generative adversarial network
TL;DR:We cast GANs in the variational inequality framework and import techniques from this literature to optimize GANs better; we give algorithmic extensions and empirically test their performance for training GANs.
Learning-Based Frequency Estimation Algorithms
Keywords:streaming algorithms, heavy-hitters, Count-Min, Count-Sketch
TL;DR:Data stream algorithms can be improved using deep learning, while retaining performance guarantees.
From Language to Goals: Inverse Reinforcement Learning for Vision-Based Instruction Following
Keywords:inverse reinforcement learning, language grounding, instruction following, language-based learning
TL;DR:We ground language commands in a high-dimensional visual environment by learning language-conditioned rewards using inverse reinforcement learning.
Backpropamine: training self-modifying neural networks with differentiable neuromodulated plasticity
Keywords:meta-learning, reinforcement learning, plasticity, neuromodulation, Hebbian learning, recurrent neural networks
TL;DR:Neural networks can be trained to modify their own connectivity, improving their online learning performance on challenging tasks.
Recurrent Experience Replay in Distributed Reinforcement Learning
Keywords:RNN, LSTM, experience replay, distributed training, reinforcement learning
TL;DR:Investigation on combining recurrent neural networks and experience replay leading to state-of-the-art agent on both Atari-57 and DMLab-30 using single set of hyper-parameters.
A Generative Model For Electron Paths
Keywords:Molecules, Reaction Prediction, Graph Neural Networks, Deep Generative Models
TL;DR:A generative model for reaction prediction that learns the mechanistic electron steps of a reaction directly from raw reaction data.
Modeling Uncertainty with Hedged Instance Embeddings
Keywords:uncertainty, instance embedding, metric learning, probabilistic embedding
TL;DR:The paper proposes using probability distributions instead of points for instance embeddings tasks such as recognition and verification.
Beyond Greedy Ranking: Slate Optimization via List-CVAE
Keywords:CVAE, VAE, recommendation system, slate optimization, whole page optimization
TL;DR:We used a CVAE type model structure to learn to directly generate slates/whole pages for recommendation systems.
Stochastic Prediction of Multi-Agent Interactions from Partial Observations
Keywords:Dynamics modeling, partial observations, multi-agent interactions, predictive models
TL;DR:We present a method which learns to integrate temporal information and ambiguous visual information in the context of interacting agents.
GamePad: A Learning Environment for Theorem Proving
Keywords:Theorem proving, ITP, systems, neural embeddings
TL;DR:We introduce a system called GamePad to explore the application of machine learning methods to theorem proving in the Coq proof assistant.
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Keywords:natural language understanding, multi-task learning, evaluation
TL;DR:We present a multi-task benchmark and analysis platform for evaluating generalization in natural language understanding systems.
On Computation and Generalization of Generative Adversarial Networks under Spectrum Control
Keywords:None
TL;DR:None
Large-Scale Study of Curiosity-Driven Learning
Keywords:exploration, curiosity, intrinsic reward, no extrinsic reward, unsupervised, no-reward, skills
TL;DR:An agent trained only with curiosity, and no extrinsic reward, does surprisingly well on 54 popular environments, including the suite of Atari games, Mario etc.
Unsupervised Discovery of Parts, Structure, and Dynamics
Keywords:Self-Supervised Learning, Visual Prediction, Hierarchical Models
TL;DR:Learning object parts, hierarchical structure, and dynamics by watching how they move
Music Transformer: Generating Music with Long-Term Structure
Keywords:music generation
TL;DR:We show the first successful use of Transformer in generating music that exhibits long-term structure.
BabyAI: A Platform to Study the Sample Efficiency of Grounded Language Learning
Keywords:language, learning, efficiency, imitation learning, reinforcement learning
TL;DR:We present the BabyAI platform for studying data efficiency of language learning with a human in the loop
Analyzing Inverse Problems with Invertible Neural Networks
Keywords:Inverse problems, Neural Networks, Uncertainty, Invertible Neural Networks
TL;DR:To analyze inverse problems with Invertible Neural Networks
RelGAN: Relational Generative Adversarial Networks for Text Generation
Keywords:None
TL;DR:None
The Singular Values of Convolutional Layers
Keywords:singular values, operator norm, convolutional layers, regularization
TL;DR:We characterize the singular values of the linear transformation associated with a standard 2D multi-channel convolutional layer, enabling their efficient computation.
An Empirical study of Binary Neural Networks' Optimisation
Keywords:None
TL;DR:None
Approximability of Discriminators Implies Diversity in GANs
Keywords:Theory, Generative adversarial networks, Mode collapse, Generalization
TL;DR:GANs can in principle learn distributions sample-efficiently, if the discriminator class is compact and has strong distinguishing power against the particular generator class.
Learning Embeddings into Entropic Wasserstein Spaces
Keywords:Embedding, Wasserstein, Sinkhorn, Optimal Transport
TL;DR:We show that Wasserstein spaces are good targets for embedding data with complex semantic structure.
DeepOBS: A Deep Learning Optimizer Benchmark Suite
Keywords:deep learning, optimization
TL;DR:We provide a software package that drastically simplifies, automates, and improves the evaluation of deep learning optimizers.
InfoBot: Transfer and Exploration via the Information Bottleneck
Keywords:Information bottleneck, policy transfer, policy generalization, exploration
TL;DR:Training agents with goal-policy information bottlenecks promotes transfer and yields a powerful exploration bonus
The Comparative Power of ReLU Networks and Polynomial Kernels in the Presence of Sparse Latent Structure
Keywords:theory, representational power, universal approximators, polynomial kernels, latent sparsity, beyond worst case, separation result
TL;DR:Beyond-worst-case analysis of the representational power of ReLU nets & polynomial kernels -- in particular in the presence of sparse latent structure.
Learning Implicitly Recurrent CNNs Through Parameter Sharing
Keywords:deep learning, architecture search, computer vision
TL;DR:We propose a method that enables CNN folding to create recurrent connections
Learning Particle Dynamics for Manipulating Rigid Bodies, Deformable Objects, and Fluids
Keywords:Dynamics modeling, Control, Particle-Based Representation
TL;DR:Learning particle dynamics with dynamic interaction graphs for simulating and control rigid bodies, deformable objects, and fluids.
Regularized Learning for Domain Adaptation under Label Shifts
Keywords:Deep Learning, Domain Adaptation, Label Shift, Importance Weights, Generalization
TL;DR:A practical and provably guaranteed approach for training efficiently classifiers in the presence of label shifts between Source and Target data sets
Von Mises-Fisher Loss for Training Sequence to Sequence Models with Continuous Outputs
Keywords:Language Generation, Regression, Word Embeddings, Machine Translation
TL;DR:Language generation using seq2seq models which produce word embeddings instead of a softmax based distribution over the vocabulary at each step enabling much faster training while maintaining generation quality
Relational Forward Models for Multi-Agent Learning
Keywords:multi-agent reinforcement learning, relational reasoning, forward models
TL;DR:Relational Forward Models for multi-agent learning make accurate predictions of agents' future behavior, they produce intepretable representations and can be used inside agents.
Imposing Category Trees Onto Word-Embeddings Using A Geometric Construction
Keywords:category tree, word-embeddings, geometry
TL;DR:we show a geometric method to perfectly encode categroy tree information into pre-trained word-embeddings.
Two-Timescale Networks for Nonlinear Value Function Approximation
Keywords:Reinforcement learning, policy evaluation, nonlinear function approximation
TL;DR:We propose an architecture for learning value functions which allows the use of any linear policy evaluation algorithm in tandem with nonlinear feature learning.
Diversity-Sensitive Conditional Generative Adversarial Networks
Keywords:Conditional Generative Adversarial Network, mode-collapse, multi-modal generation, image-to-image translation, image in-painting, video prediction
TL;DR:We propose a simple and general approach that avoids a mode collapse problem in various conditional GANs.
Query-Efficient Hard-label Black-box Attack: An Optimization-based Approach
Keywords:None
TL;DR:None
Rethinking the Value of Network Pruning
Keywords:In structured network pruning, fine-tuning a pruned model only gives comparable performance with training it from scratch.
TL;DR:network pruning, network compression, architecture search, train from scratch
Hyperbolic Attention Networks
Keywords:Hyperbolic Geometry, Attention Methods, Reasoning on Graphs, Relation Learning, Scale Free Graphs, Transformers, Power Law
TL;DR:We propose to incorporate inductive biases and operations coming from hyperbolic geometry to improve the attention mechanism of the neural networks.
Learning from Positive and Unlabeled Data with a Selection Bias
Keywords:None
TL;DR:None
Adv-BNN: Improved Adversarial Defense through Robust Bayesian Neural Network
Keywords:None
TL;DR:None
Optimal Completion Distillation for Sequence Learning
Keywords:Sequence Learning, Edit Distance, Speech Recognition, Deep Reinforcement Learning
TL;DR:Optimal Completion Distillation (OCD) is a training procedure for optimizing sequence to sequence models based on edit distance which achieves state-of-the-art on end-to-end Speech Recognition tasks.
Caveats for information bottleneck in deterministic scenarios
Keywords:Information bottleneck behaves in surprising ways whenever the output is a deterministic function of the input.
TL;DR:information bottleneck, supervised learning, deep learning, information theory
Deep Learning 3D Shapes Using Alt-az Anisotropic 2-Sphere Convolution
Keywords:Spherical Convolution, Geometric deep learning, 3D shape analysis
TL;DR:A method for applying deep learning to 3D surfaces using their spherical descriptors and alt-az anisotropic convolution on 2-sphere.
Small nonlinearities in activation functions create bad local minima in neural networks
Keywords:spurious local minima, loss surface, optimization landscape, neural network
TL;DR:We constructively prove that even the slightest nonlinear activation functions introduce spurious local minima, for general datasets and activation functions.
Information Theoretic lower bounds on negative log likelihood
Keywords:latent variable modeling, rate-distortion theory, log likelihood bounds
TL;DR:Use rate-distortion theory to bound how much a latent variable model can be improved
Preferences Implicit in the State of the World
Keywords:Preference learning, Inverse reinforcement learning, Inverse optimal stochastic control, Maximum entropy reinforcement learning, Apprenticeship learning
TL;DR:When a robot is deployed in an environment that humans have been acting in, the state of the environment is already optimized for what humans want, and we can use this to infer human preferences.
A Kernel Random Matrix-Based Approach for Sparse PCA
Keywords:None
TL;DR:None
Bayesian Prediction of Future Street Scenes using Synthetic Likelihoods
Keywords:bayesian inference, segmentation, anticipation, multi-modality
TL;DR:Dropout based Bayesian inference is extended to deal with multi-modality and is evaluated on scene anticipation tasks.
There Are Many Consistent Explanations of Unlabeled Data: Why You Should Average
Keywords:semi-supervised learning, computer vision, classification, consistency regularization, flatness, weight averaging, stochastic weight averaging
TL;DR:Consistency-based models for semi-supervised learning do not converge to a single point but continue to explore a diverse set of plausible solutions on the perimeter of a flat region. Weight averaging helps improve generalization performance.
Large-Scale Answerer in Questioner's Mind for Visual Dialog Question Generation
Keywords:None
TL;DR:None
Graph HyperNetworks for Neural Architecture Search
Keywords:None
TL;DR:None
DELTA: DEEP LEARNING TRANSFER USING FEATURE MAP WITH ATTENTION FOR CONVOLUTIONAL NETWORKS
Keywords:transfer learning, deep learning, regularization, attention, cnn
TL;DR:improving deep transfer learning with regularization using attention based feature maps
textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE WITH DISTRIBUTED COMPOSITIONAL PRIOR
Keywords:neural topic model, natural language processing, text representation, language modeling, information retrieval, deep learning
TL;DR:Unified neural model of topic and language modeling to introduce language structure in topic models for contextualized topic vectors
Amortized Bayesian Meta-Learning
Keywords:variational inference, meta-learning, few-shot learning, uncertainty quantification
TL;DR:We propose a meta-learning method which efficiently amortizes hierarchical variational inference across training episodes.
Probabilistic Recursive Reasoning for Multi-Agent Reinforcement Learning
Keywords:Multi-agent Reinforcement Learning, Recursive Reasoning
TL;DR:We proposed a novel probabilisitic recursive reasoning (PR2) framework for multi-agent deep reinforcement learning tasks.
Learning Neural PDE Solvers with Convergence Guarantees
Keywords:Partial differential equation, deep learning
TL;DR:We learn a fast neural solver for PDEs that has convergence guarantees.
A new dog learns old tricks: RL finds classic optimization algorithms
Keywords:reinforcement learning, algorithms, adwords, knapsack, secretary
TL;DR:By combining ideas from traditional algorithms design and reinforcement learning, we introduce a novel framework for learning algorithms that solve online combinatorial optimization problems.
Deep Graph Infomax
Keywords:Unsupervised Learning, Graph Neural Networks, Graph Convolutions, Mutual Information, Infomax, Deep Learning
TL;DR:A new method for unsupervised representation learning on graphs, relying on maximizing mutual information between local and global representations in a graph. State-of-the-art results, competitive with supervised learning.
Theoretical Analysis of Auto Rate-Tuning by Batch Normalization
Keywords:batch normalization, scale invariance, learning rate, stationary point
TL;DR:We give a theoretical analysis of the ability of batch normalization to automatically tune learning rates, in the context of finding stationary points for a deep learning objective.
Per-Tensor Fixed-Point Quantization of the Back-Propagation Algorithm
Keywords:deep learning, reduced precision, fixed-point, quantization, back-propagation algorithm
TL;DR:We analyze and determine the precision requirements for training neural networks when all tensors, including back-propagated signals and weight accumulators, are quantized to fixed-point format.
FUNCTIONAL VARIATIONAL BAYESIAN NEURAL NETWORKS
Keywords:functional variational inference, Bayesian neural networks, stochastic processes
TL;DR:We perform functional variational inference on the stochastic processes defined by Bayesian neural networks.
NADPEx: An on-policy temporally consistent exploration method for deep reinforcement learning
Keywords:None
TL;DR:None
SPIGAN: Privileged Adversarial Learning from Simulation
Keywords:domain adaptation, GAN, semantic segmentation, simulation, privileged information
TL;DR:An unsupervised sim-to-real domain adaptation method for semantic segmentation using privileged information from a simulator with GAN-based image translation.
Generating Multi-Agent Trajectories using Programmatic Weak Supervision
Keywords:deep learning, generative models, imitation learning, hierarchical methods, data programming, weak supervision, spatiotemporal
TL;DR:We blend deep generative models with programmatic weak supervision to generate coordinated multi-agent trajectories of significantly higher quality than previous baselines.
Label super-resolution networks
Keywords:weakly supervised segmentation, land cover mapping, medical imaging
TL;DR:Super-resolving coarse labels into pixel-level labels, applied to aerial imagery and medical scans.
ANYTIME MINIBATCH: EXPLOITING STRAGGLERS IN ONLINE DISTRIBUTED OPTIMIZATION
Keywords:distributed optimization, gradient descent, minibatch, stragglers
TL;DR:Accelerate distributed optimization by exploiting stragglers.
Sample Efficient Adaptive Text-to-Speech
Keywords:few shot, meta learning, text to speech, wavenet
TL;DR:Sample efficient algorithms to adapt a text-to-speech model to a new voice style with the state-of-the-art performance.
Practical lossless compression with latent variables using bits back coding
Keywords:compression, variational auto-encoders, deep latent gaussian models, lossless compression, latent variables, approximate inference, variational inference
TL;DR:We do lossless compression of large image datasets using a VAE, beat existing compression algorithms.
Kernel RNN Learning (KeRNL)
Keywords:RNNs, Biologically plausible learning rules, Algorithm, Neural Networks, Supervised Learning
TL;DR:A biologically plausible learning rule for training recurrent neural networks
Deep, Skinny Neural Networks are not Universal Approximators
Keywords:This paper proves that skinny neural networks cannot approximate certain functions, no matter how deep they are.
TL;DR:neural network, universality, expressability
Large Scale Graph Learning From Smooth Signals
Keywords:None
TL;DR:None
Overcoming Catastrophic Forgetting for Continual Learning via Model Adaptation
Keywords:None
TL;DR:None
Analysis of Quantized Models
Keywords:weight quantization, gradient quantization, distributed learning
TL;DR:In this paper, we studied efficient training of loss-aware weight-quantized networks with quantized gradient in a distributed environment, both theoretically and empirically.
Deep learning generalizes because the parameter-function map is biased towards simple functions
Keywords:generalization, deep learning theory, PAC-Bayes, Gaussian processes, parameter-function map, simplicity bias
TL;DR:The parameter-function map of deep networks is hugely biased; this can explain why they generalize. We use PAC-Bayes and Gaussian processes to obtain nonvacuous bounds.
Learning when to Communicate at Scale in Multiagent Cooperative and Competitive Tasks
Keywords:multiagent, communication, competitive, cooperative, continuous, emergent, reinforcement learning
TL;DR:We introduce IC3Net, a single network which can be used to train agents in cooperative, competitive and mixed scenarios. We also show that agents can learn when to communicate using our model.
Synthetic Datasets for Neural Program Synthesis
Keywords:None
TL;DR:None
DPSNet: End-to-end Deep Plane Sweep Stereo
Keywords:Deep Learning, Stereo, Depth, Geometry
TL;DR:A convolution neural network for multi-view stereo matching whose design is inspired by best practices of traditional geometry-based approaches
Conditional Network Embeddings
Keywords:Network embedding, graph embedding, learning node representations, link prediction, multi-label classification of nodes
TL;DR:We introduce a network embedding method that accounts for prior information about the network, yielding superior empirical performance.
Defensive Quantization: When Efficiency Meets Robustness
Keywords:defensive quantization, model quantization, adversarial attack, efficiency, robustness
TL;DR:We designed a novel quantization methodology to jointly optimize the efficiency and robustness of deep learning models.
GO Gradient for Expectation-Based Objectives
Keywords:generalized reparameterization gradient, variance reduction, non-reparameterizable, discrete random variable, GO gradient, general and one-sample gradient, expectation-based objective, variable nabla, statistical back-propagation, hierarchical, graphical model
TL;DR:a Rep-like gradient for non-reparameterizable continuous/discrete distributions; further generalized to deep probabilistic models, yielding statistical back-propagation
h-detach: Modifying the LSTM Gradient Towards Better Optimization
Keywords:LSTM, Optimization, Long term dependencies, Back-propagation through time
TL;DR:A simple algorithm to improve optimization and handling of long term dependencies in LSTM
An analytic theory of generalization dynamics and transfer learning in deep linear networks
Keywords:Generalization, Theory, Transfer, Multi-task, Linear
TL;DR:We provide many insights into neural network generalization from the theoretically tractable linear case.
Differentiable Learning-to-Normalize via Switchable Normalization
Keywords:None
TL;DR:None
SOM-VAE: Interpretable Discrete Representation Learning on Time Series
Keywords:deep learning, self-organizing map, variational autoencoder, representation learning, time series, machine learning, interpretability
TL;DR:We present a method to learn interpretable representations on time series using ideas from variational autoencoders, self-organizing maps and probabilistic models.
Hierarchical Generative Modeling for Controllable Speech Synthesis
Keywords:speech synthesis, representation learning, deep generative model, sequence-to-sequence model
TL;DR:Building a TTS model with Gaussian Mixture VAEs enables fine-grained control of speaking style, noise condition, and more.
Learning Factorized Multimodal Representations
Keywords:multimodal learning, representation learning
TL;DR:We propose a model to learn factorized multimodal representations that are discriminative, generative, and interpretable.
Composing Complex Skills by Learning Transition Policies
Keywords:reinforcement learning, hierarchical reinforcement learning, continuous control, modular framework
TL;DR:Transition policies enable agents to compose complex skills by smoothly connecting previously acquired primitive skills.
Human-level Protein Localization with Convolutional Neural Networks
Keywords:None
TL;DR:None
Environment Probing Interaction Policies
Keywords:None
TL;DR:None
Lagging Inference Networks and Posterior Collapse in Variational Autoencoders
Keywords:variational autoencoders, posterior collapse, generative models
TL;DR:To address posterior collapse in VAEs, we propose a novel yet simple training procedure that aggressively optimizes inference network with more updates. This new training procedure mitigates posterior collapse and leads to a better VAE model.
A2BCD: Asynchronous Acceleration with Optimal Complexity
Keywords:asynchronous, optimization, parallel, accelerated, complexity
TL;DR:We prove the first-ever convergence proof of an asynchronous accelerated algorithm that attains a speedup.
Learning to Infer and Execute 3D Shape Programs
Keywords:Program Synthesis, 3D Shape Modeling, Self-supervised Learning
TL;DR:We propose 3D shape programs, a structured, compositional shape representation. Our model learns to infer and execute shape programs to explain 3D shapes.
Deep Decoder: Concise Image Representations from Untrained Non-convolutional Networks
Keywords:natural image model, image prior, under-determined neural networks, untrained network, non-convolutional network, denoising, inverse problem
TL;DR:We introduce an underparameterized, nonconvolutional, and simple deep neural network that can, without training, effectively represent natural images and solve image processing tasks like compression and denoising competitively.
SNAS: stochastic neural architecture search
Keywords:None
TL;DR:None
Revealing interpretable object representations from human behavior
Keywords:category representation, sparse coding, representation learning, interpretable representations
TL;DR:Human behavioral judgments are used to obtain sparse and interpretable representations of objects that generalize to other tasks
AntisymmetricRNN: A Dynamical System View on Recurrent Neural Networks
Keywords:None
TL;DR:None
Global-to-local Memory Pointer Networks for Task-Oriented Dialogue
Keywords:pointer networks, memory networks, task-oriented dialogue systems, natural language processing
TL;DR:GLMP: Global memory encoder (context RNN, global pointer) and local memory decoder (sketch RNN, local pointer) that share external knowledge (MemNN) are proposed to strengthen response generation in task-oriented dialogue.
InstaGAN: Instance-aware Image-to-Image Translation
Keywords:Image-to-Image Translation, Generative Adversarial Networks
TL;DR:We propose a novel method to incorporate the set of instance attributes for image-to-image translation.
Deep Layers as Stochastic Solvers
Keywords:deep networks, optimization
TL;DR:A framework that links deep network layers to stochastic optimization algorithms; can be used to improve model accuracy and inform network design.
Learning Multi-Level Hierarchies with Hindsight
Keywords:Hierarchical Reinforcement Learning, Reinforcement Learning, Deep Reinforcement Learning
TL;DR:We introduce the first Hierarchical RL approach to successfully learn 3-level hierarchies in parallel in tasks with continuous state and action spaces.
原文地址:https://www.cnblogs.com/Tom-Ren/p/11054650.html