April 1, 2020

2747 words 13 mins read

Paper Group NANR 72

Paper Group NANR 72

Efficacy of Pixel-Level OOD Detection for Semantic Segmentation. Customizing Sequence Generation with Multi-Task Dynamical Systems. Learning Representations in Reinforcement Learning: an Information Bottleneck Approach. RAT-SQL: Relation-Aware Schema Encoding and Linking for Text-to-SQL Parsers. LEX-GAN: Layered Explainable Rumor Detector Based on …

Efficacy of Pixel-Level OOD Detection for Semantic Segmentation

Title Efficacy of Pixel-Level OOD Detection for Semantic Segmentation
Authors Anonymous
Abstract The detection of out of distribution samples for image classification has been widely researched. Safety critical applications, such as autonomous driving, would benefit from the ability to localise the unusual objects causing the image to be out of distribution. This paper adapts state-of-the-art methods for detecting out of distribution images for image classification to the new task of detecting out of distribution pixels, which can localise the unusual objects. It further experimentally compares the adapted methods on two new datasets derived from existing semantic segmentation datasets using PSPNet and DeeplabV3+ architectures, as well as proposing a new metric for the task. The evaluation shows that the performance ranking of the compared methods does not transfer to the new task and every method performs significantly worse than their image-level counterparts.
Tasks Autonomous Driving, Image Classification, Semantic Segmentation
Published 2020-01-01
URL https://openreview.net/forum?id=SJlxglSFPB
PDF https://openreview.net/pdf?id=SJlxglSFPB
PWC https://paperswithcode.com/paper/efficacy-of-pixel-level-ood-detection-for
Repo
Framework

Customizing Sequence Generation with Multi-Task Dynamical Systems

Title Customizing Sequence Generation with Multi-Task Dynamical Systems
Authors Anonymous
Abstract Dynamical system models (including RNNs) often lack the ability to adapt the sequence generation or prediction to a given context, limiting their real-world application. In this paper we show that hierarchical multi-task dynamical systems (MTDSs) provide direct user control over sequence generation, via use of a latent code z that specifies the customization to the individual data sequence. This enables style transfer, interpolation and morphing within generated sequences. We show the MTDS can improve predictions via latent code interpolation, and avoid the long-term performance degradation of standard RNN approaches.
Tasks Style Transfer
Published 2020-01-01
URL https://openreview.net/forum?id=Bkln2a4tPB
PDF https://openreview.net/pdf?id=Bkln2a4tPB
PWC https://paperswithcode.com/paper/customizing-sequence-generation-with-multi-1
Repo
Framework

Learning Representations in Reinforcement Learning: an Information Bottleneck Approach

Title Learning Representations in Reinforcement Learning: an Information Bottleneck Approach
Authors Yingjun Pei, Xinwen Hou
Abstract The information bottleneck principle is an elegant and useful approach to representation learning. In this paper, we investigate the problem of representation learning in the context of reinforcement learning using the information bottleneck framework, aiming at improving the sample efficiency of the learning algorithms.We analytically derive the optimal conditional distribution of the representation, and provide a variational lower bound. Then, we maximize this lower bound with the Stein variational (SV) gradient method. We incorporate this framework in the advantageous actor critic algorithm (A2C) and the proximal policy optimization algorithm (PPO). Our experimental results show that our framework can improve the sample efficiency of vanilla A2C and PPO significantly. Finally, we study the information-bottleneck (IB) perspective in deep RL with the algorithm called mutual information neural estimation(MINE). We experimentally verify that the information extraction-compression process also exists in deep RL and our framework is capable of accelerating this process. We also analyze the relationship between MINE and our method, through this relationship, we theoretically derive an algorithm to optimize our IB framework without constructing the lower bound.
Tasks Representation Learning
Published 2020-01-01
URL https://openreview.net/forum?id=Syl-xpNtwS
PDF https://openreview.net/pdf?id=Syl-xpNtwS
PWC https://paperswithcode.com/paper/learning-representations-in-reinforcement-1
Repo
Framework

RAT-SQL: Relation-Aware Schema Encoding and Linking for Text-to-SQL Parsers

Title RAT-SQL: Relation-Aware Schema Encoding and Linking for Text-to-SQL Parsers
Authors Anonymous
Abstract When translating natural language questions into SQL queries to answer questions from a database, contemporary semantic parsing models struggle to generalize to unseen database schemas. The generalization challenge lies in (a) encoding the database relations in an accessible way for the semantic parser, and (b) modeling alignment between database columns and their mentions in a given query. We present a unified framework, based on the relation-aware self-attention mechanism,to address schema encoding, schema linking, and feature representation within a text-to-SQL encoder. On the challenging Spider dataset this framework boosts the exact match accuracy to 53.7%, compared to 47.4% for the previous state-of-the-art model unaugmented with BERT embeddings. In addition, we observe qualitative improvements in the model’s understanding of schema linking and alignment.
Tasks Semantic Parsing, Text-To-Sql
Published 2020-01-01
URL https://openreview.net/forum?id=H1egcgHtvB
PDF https://openreview.net/pdf?id=H1egcgHtvB
PWC https://paperswithcode.com/paper/rat-sql-relation-aware-schema-encoding-and
Repo
Framework

LEX-GAN: Layered Explainable Rumor Detector Based on Generative Adversarial Networks

Title LEX-GAN: Layered Explainable Rumor Detector Based on Generative Adversarial Networks
Authors Anonymous
Abstract Social media have emerged to be increasingly popular and have been used as tools for gathering and propagating information. However, the vigorous growth of social media contributes to the fast-spreading and far-reaching rumors. Rumor detection has become a necessary defense. Traditional rumor detection methods based on hand-crafted feature selection are replaced by automatic approaches that are based on Artificial Intelligence (AI). AI decision making systems need to have the necessary means, such as explainability to assure users their trustworthiness. Inspired by the thriving development of Generative Adversarial Networks (GANs) on text applications, we propose LEX-GAN, a GAN-based layered explainable rumor detector to improve the detection quality and provide explainability. Unlike fake news detection that needs a previously collected verified news database, LEX-GAN realizes explainable rumor detection based on only tweet-level text. LEX-GAN is trained with generated non-rumor-looking rumors. The generators produce rumors by intelligently inserting controversial information in non-rumors, and force the discriminators to detect detailed glitches and deduce exactly which parts in the sentence are problematic. The layered structures in both generative and discriminative model contributes to the high performance. We show LEX-GAN’s mutation detection ability in textural sequences by performing a gene classification and mutation detection task.
Tasks Decision Making, Fake News Detection, Feature Selection
Published 2020-01-01
URL https://openreview.net/forum?id=S1lukyrKPr
PDF https://openreview.net/pdf?id=S1lukyrKPr
PWC https://paperswithcode.com/paper/lex-gan-layered-explainable-rumor-detector
Repo
Framework

The Usual Suspects? Reassessing Blame for VAE Posterior Collapse

Title The Usual Suspects? Reassessing Blame for VAE Posterior Collapse
Authors Anonymous
Abstract In narrow asymptotic settings Gaussian VAE models of continuous data have been shown to possess global optima aligned with ground-truth distributions. Even so, it is well known that poor solutions whereby the latent posterior collapses to an uninformative prior are sometimes obtained in practice. However, contrary to conventional wisdom that largely assigns blame for this phenomena on the undue influence of KL-divergence regularization, we will argue that posterior collapse is, at least in part, a direct consequence of bad local minima inherent to the loss surface of deep autoencoder networks. In particular, we prove that even small nonlinear perturbations of affine VAE decoder models can produce such minima, and in deeper models, analogous minima can force the VAE to behave like an aggressive truncation operator, provably discarding information along all latent dimensions in certain circumstances. Regardless, the underlying message here is not meant to undercut valuable existing explanations of posterior collapse, but rather, to refine the discussion and elucidate alternative risk factors that may have been previously underappreciated.
Tasks
Published 2020-01-01
URL https://openreview.net/forum?id=r1lIKlSYvH
PDF https://openreview.net/pdf?id=r1lIKlSYvH
PWC https://paperswithcode.com/paper/the-usual-suspects-reassessing-blame-for-vae
Repo
Framework

Generative Restricted Kernel Machines

Title Generative Restricted Kernel Machines
Authors Anonymous
Abstract We introduce a novel framework for generative models based on Restricted Kernel Machines (RKMs) with multi-view generation and uncorrelated feature learning capabilities, called Gen-RKM. To incorporate multi-view generation, this mechanism uses a shared representation of data from various views. The mechanism is flexible to incorporate both kernel-based, (deep) neural network and convolutional based models within the same setting. To update the parameters of the network, we propose a novel training procedure which jointly learns the features and shared representation. Experiments demonstrate the potential of the framework through qualitative evaluation of generated samples.
Tasks
Published 2020-01-01
URL https://openreview.net/forum?id=ryghPCVYvH
PDF https://openreview.net/pdf?id=ryghPCVYvH
PWC https://paperswithcode.com/paper/generative-restricted-kernel-machines-1
Repo
Framework

Which Tasks Should Be Learned Together in Multi-task Learning?

Title Which Tasks Should Be Learned Together in Multi-task Learning?
Authors Anonymous
Abstract Many computer vision applications require solving multiple tasks in real-time. A neural network can be trained to solve multiple tasks simultaneously using ‘multi-task learning’. This saves computation at inference time as only a single network needs to be evaluated. Unfortunately, this often leads to inferior overall performance as task objectives compete, which consequently poses the question: which tasks should and should not be learned together in one network when employing multi-task learning? We systematically study task cooperation and competition and propose a framework for assigning tasks to a few neural networks such that cooperating tasks are computed by the same neural network, while competing tasks are computed by different networks. Our framework offers a time-accuracy trade-off and can produce better accuracy using less inference time than not only a single large multi-task neural network but also many single-task networks.
Tasks Multi-Task Learning
Published 2020-01-01
URL https://openreview.net/forum?id=HJlTpCEKvS
PDF https://openreview.net/pdf?id=HJlTpCEKvS
PWC https://paperswithcode.com/paper/which-tasks-should-be-learned-together-in-1
Repo
Framework

Infinite-horizon Off-Policy Policy Evaluation with Multiple Behavior Policies

Title Infinite-horizon Off-Policy Policy Evaluation with Multiple Behavior Policies
Authors Anonymous
Abstract We consider off-policy policy evaluation when the trajectory data are generated by multiple behavior policies. Recent work has shown the key role played by the state or state-action stationary distribution corrections in the infinite horizon context for off-policy policy evaluation. We propose estimated mixture policy (EMP), a novel class of partially policy-agnostic methods to accurately estimate those quantities. With careful analysis, we show that EMP gives rise to estimates with reduced variance for estimating the state stationary distribution correction while it also offers a useful induction bias for estimating the state-action stationary distribution correction. In extensive experiments with both continuous and discrete environments, we demonstrate that our algorithm offers significantly improved accuracy compared to the state-of-the-art methods.
Tasks
Published 2020-01-01
URL https://openreview.net/forum?id=rkgU1gHtvr
PDF https://openreview.net/pdf?id=rkgU1gHtvr
PWC https://paperswithcode.com/paper/infinite-horizon-off-policy-policy-evaluation
Repo
Framework

Continuous Meta-Learning without Tasks

Title Continuous Meta-Learning without Tasks
Authors Anonymous
Abstract Meta-learning is a promising strategy for learning to efficiently learn within new tasks, using data gathered from a distribution of tasks. However, the meta-learning literature thus far has focused on the task segmented setting, where at train-time, offline data is assumed to be split according to the underlying task, and at test-time, the algorithms are optimized to learn in a single task. In this work, we enable the application of generic meta-learning algorithms to settings where this task segmentation is unavailable, such as continual online learning with a time-varying task. We present meta-learning via online changepoint analysis (MOCA), an approach which augments a meta-learning algorithm with a differentiable Bayesian changepoint detection scheme. The framework allows both training and testing directly on time series data without segmenting it into discrete tasks. We demonstrate the utility of this approach on a nonlinear meta-regression benchmark as well as two meta-image-classification benchmarks.
Tasks Image Classification, Meta-Learning, Time Series
Published 2020-01-01
URL https://openreview.net/forum?id=r1l1myStwr
PDF https://openreview.net/pdf?id=r1l1myStwr
PWC https://paperswithcode.com/paper/continuous-meta-learning-without-tasks
Repo
Framework

FAST LEARNING VIA EPISODIC MEMORY: A PERSPECTIVE FROM ANIMAL DECISION-MAKING

Title FAST LEARNING VIA EPISODIC MEMORY: A PERSPECTIVE FROM ANIMAL DECISION-MAKING
Authors Anonymous
Abstract A typical experiment to study cognitive function is to train animals to perform tasks, while the researcher records the electrical activity of the animals neurons. The main obstacle faced, when using this type of electrophysiological experiment to uncover the circuit mechanisms underlying complex behaviors, is our incomplete access to relevant circuits in the brain. One promising approach is to model neural circuits using an artificial neural network (ANN), which can provide complete access to the “neural circuits” responsible for a behavior. More recently, reinforcement learning models have been adopted to understand the functions of cortico-basal ganglia circuits as reward-based learning has been found in mammalian brain. In this paper, we propose a Biologically-plausible Actor-Critic with Episodic Memory (B-ACEM) framework to model a prefrontal cortex-basal ganglia-hippocampus (PFC-BG) circuit, which is verified to capture the behavioral findings from a well-known perceptual decision-making task, i.e., random dots motion discrimination. This B-ACEM framework links neural computation to behaviors, on which we can explore how episodic memory should be considered to govern future decision. Experiments are conducted using different settings of the episodic memory and results show that all patterns of episodic memories can speed up learning. In particular, salient events are prioritized to propagate reward information and guide decisions. Our B-ACEM framework and the built-on experiments give inspirations to both designs for more standard decision-making models in biological system and a more biologically-plausible ANN.
Tasks Decision Making
Published 2020-01-01
URL https://openreview.net/forum?id=BJl4pA4Kwr
PDF https://openreview.net/pdf?id=BJl4pA4Kwr
PWC https://paperswithcode.com/paper/fast-learning-via-episodic-memory-a
Repo
Framework

Implicit Generative Modeling for Efficient Exploration

Title Implicit Generative Modeling for Efficient Exploration
Authors Anonymous
Abstract Efficient exploration remains a challenging problem in reinforcement learning, especially for those tasks where rewards from environments are sparse. A commonly used approach for exploring such environments is to introduce some “intrinsic” reward. In this work, we focus on model uncertainty estimation as an intrinsic reward for efficient exploration. In particular, we introduce an implicit generative modeling approach to estimate a Bayesian uncertainty of the agent’s belief of the environment dynamics. Each random draw from our generative model is a neural network that instantiates the dynamic function, hence multiple draws would approximate the posterior, and the variance in the future prediction based on this posterior is used as an intrinsic reward for exploration. We design a training algorithm for our generative model based on the amortized Stein Variational Gradient Descent. In experiments, we compare our implementation with state-of-the-art intrinsic reward-based exploration approaches, including two recent approaches based on an ensemble of dynamic models. In challenging exploration tasks, our implicit generative model consistently outperforms competing approaches regarding data efficiency in exploration.
Tasks Efficient Exploration, Future prediction
Published 2020-01-01
URL https://openreview.net/forum?id=BkgeQ1BYwS
PDF https://openreview.net/pdf?id=BkgeQ1BYwS
PWC https://paperswithcode.com/paper/implicit-generative-modeling-for-efficient
Repo
Framework

Selfish Emergent Communication

Title Selfish Emergent Communication
Authors Anonymous
Abstract Current literature in machine learning holds that unaligned, self-interested agents do not learn to use an emergent communication channel. We introduce a new sender-receiver game to study emergent communication for this spectrum of partially-competitive scenarios and put special care into evaluation. We find that communication can indeed emerge in partially-competitive scenarios, and we discover three things that are tied to improving it. First, that selfish communication is proportional to cooperation, and it naturally occurs for situations that are more cooperative than competitive. Second, that stability and performance are improved by using LOLA (Foerster et al, 2018), especially in more competitive scenarios. And third, that discrete protocols lend themselves better to learning cooperative communication than continuous ones.
Tasks
Published 2020-01-01
URL https://openreview.net/forum?id=B1liIlBKvS
PDF https://openreview.net/pdf?id=B1liIlBKvS
PWC https://paperswithcode.com/paper/selfish-emergent-communication
Repo
Framework

Robust Graph Representation Learning via Neural Sparsification

Title Robust Graph Representation Learning via Neural Sparsification
Authors Anonymous
Abstract Graph representation learning serves as the core of many important prediction tasks, ranging from product recommendation in online marketing to fraud detection in financial domain. Real-life graphs are usually large with complex local neighborhood, where each node is described by a rich set of features and easily connects to dozens or even hundreds of neighbors. Most existing graph learning techniques rely on neighborhood aggregation, however, the complexity on real-life graphs is usually high, posing non-trivial overfitting risk during model training. In this paper, we present Neural Sparsification (NeuralSparse), a supervised graph sparsification technique that mitigates the overfitting risk by reducing the complexity of input graphs. Our method takes both structural and non-structural information as input, utilizes deep neural networks to parameterize the sparsification process, and optimizes the parameters by feedback signals from downstream tasks. Under the NeuralSparse framework, supervised graph sparsification could seamlessly connect with existing graph neural networks for more robust performance on testing data. Experimental results on both benchmark and private datasets show that NeuralSparse can effectively improve testing accuracy and bring up to 7.4% improvement when working with existing graph neural networks on node classification tasks.
Tasks Fraud Detection, Graph Representation Learning, Node Classification, Product Recommendation, Representation Learning
Published 2020-01-01
URL https://openreview.net/forum?id=S1emOTNKvS
PDF https://openreview.net/pdf?id=S1emOTNKvS
PWC https://paperswithcode.com/paper/robust-graph-representation-learning-via
Repo
Framework

NEURAL EXECUTION ENGINES

Title NEURAL EXECUTION ENGINES
Authors Yujun Yan, Kevin Swersky, Danai Koutra, Parthasarathy Ranganathan, Milad Hashemi
Abstract Turing complete computation and reasoning are often regarded as necessary pre- cursors to general intelligence. There has been a significant body of work studying neural networks that mimic general computation, but these networks fail to generalize to data distributions that are outside of their training set. We study this problem through the lens of fundamental computer science problems: sorting and graph processing. We modify the masking mechanism of a transformer in order to allow them to implement rudimentary functions with strong generalization. We call this model the Neural Execution Engine, and show that it learns, through supervision, to numerically compute the basic subroutines comprising these algorithms with near perfect accuracy. Moreover, it retains this level of accuracy while generalizing to unseen data and long sequences outside of the training distribution.
Tasks
Published 2020-01-01
URL https://openreview.net/forum?id=rJg7BA4YDr
PDF https://openreview.net/pdf?id=rJg7BA4YDr
PWC https://paperswithcode.com/paper/neural-execution-engines
Repo
Framework
comments powered by Disqus