May 7, 2019

2698 words 13 mins read

Paper Group AWR 78

Paper Group AWR 78

Referential Uncertainty and Word Learning in High-dimensional, Continuous Meaning Spaces. Clockwork Convnets for Video Semantic Segmentation. RL$^2$: Fast Reinforcement Learning via Slow Reinforcement Learning. Instance Normalization: The Missing Ingredient for Fast Stylization. A Baseline for Detecting Misclassified and Out-of-Distribution Example …

Referential Uncertainty and Word Learning in High-dimensional, Continuous Meaning Spaces

Title Referential Uncertainty and Word Learning in High-dimensional, Continuous Meaning Spaces
Authors Michael Spranger, Katrien Beuls
Abstract This paper discusses lexicon word learning in high-dimensional meaning spaces from the viewpoint of referential uncertainty. We investigate various state-of-the-art Machine Learning algorithms and discuss the impact of scaling, representation and meaning space structure. We demonstrate that current Machine Learning techniques successfully deal with high-dimensional meaning spaces. In particular, we show that exponentially increasing dimensions linearly impact learner performance and that referential uncertainty from word sensitivity has no impact.
Tasks
Published 2016-09-30
URL http://arxiv.org/abs/1609.09580v1
PDF http://arxiv.org/pdf/1609.09580v1.pdf
PWC https://paperswithcode.com/paper/referential-uncertainty-and-word-learning-in
Repo https://github.com/mspranger/icdl2016language
Framework none

Clockwork Convnets for Video Semantic Segmentation

Title Clockwork Convnets for Video Semantic Segmentation
Authors Evan Shelhamer, Kate Rakelly, Judy Hoffman, Trevor Darrell
Abstract Recent years have seen tremendous progress in still-image segmentation; however the na"ive application of these state-of-the-art algorithms to every video frame requires considerable computation and ignores the temporal continuity inherent in video. We propose a video recognition framework that relies on two key observations: 1) while pixels may change rapidly from frame to frame, the semantic content of a scene evolves more slowly, and 2) execution can be viewed as an aspect of architecture, yielding purpose-fit computation schedules for networks. We define a novel family of “clockwork” convnets driven by fixed or adaptive clock signals that schedule the processing of different layers at different update rates according to their semantic stability. We design a pipeline schedule to reduce latency for real-time recognition and a fixed-rate schedule to reduce overall computation. Finally, we extend clockwork scheduling to adaptive video processing by incorporating data-driven clocks that can be tuned on unlabeled video. The accuracy and efficiency of clockwork convnets are evaluated on the Youtube-Objects, NYUD, and Cityscapes video datasets.
Tasks Semantic Segmentation, Video Recognition, Video Semantic Segmentation
Published 2016-08-11
URL http://arxiv.org/abs/1608.03609v1
PDF http://arxiv.org/pdf/1608.03609v1.pdf
PWC https://paperswithcode.com/paper/clockwork-convnets-for-video-semantic
Repo https://github.com/shelhamer/clockwork-fcn
Framework caffe2

RL$^2$: Fast Reinforcement Learning via Slow Reinforcement Learning

Title RL$^2$: Fast Reinforcement Learning via Slow Reinforcement Learning
Authors Yan Duan, John Schulman, Xi Chen, Peter L. Bartlett, Ilya Sutskever, Pieter Abbeel
Abstract Deep reinforcement learning (deep RL) has been successful in learning sophisticated behaviors automatically; however, the learning process requires a huge number of trials. In contrast, animals can learn new tasks in just a few trials, benefiting from their prior knowledge about the world. This paper seeks to bridge this gap. Rather than designing a “fast” reinforcement learning algorithm, we propose to represent it as a recurrent neural network (RNN) and learn it from data. In our proposed method, RL$^2$, the algorithm is encoded in the weights of the RNN, which are learned slowly through a general-purpose (“slow”) RL algorithm. The RNN receives all information a typical RL algorithm would receive, including observations, actions, rewards, and termination flags; and it retains its state across episodes in a given Markov Decision Process (MDP). The activations of the RNN store the state of the “fast” RL algorithm on the current (previously unseen) MDP. We evaluate RL$^2$ experimentally on both small-scale and large-scale problems. On the small-scale side, we train it to solve randomly generated multi-arm bandit problems and finite MDPs. After RL$^2$ is trained, its performance on new MDPs is close to human-designed algorithms with optimality guarantees. On the large-scale side, we test RL$^2$ on a vision-based navigation task and show that it scales up to high-dimensional problems.
Tasks
Published 2016-11-09
URL http://arxiv.org/abs/1611.02779v2
PDF http://arxiv.org/pdf/1611.02779v2.pdf
PWC https://paperswithcode.com/paper/rl2-fast-reinforcement-learning-via-slow
Repo https://github.com/dragen1860/MAML-Pytorch-RL
Framework pytorch

Instance Normalization: The Missing Ingredient for Fast Stylization

Title Instance Normalization: The Missing Ingredient for Fast Stylization
Authors Dmitry Ulyanov, Andrea Vedaldi, Victor Lempitsky
Abstract It this paper we revisit the fast stylization method introduced in Ulyanov et. al. (2016). We show how a small change in the stylization architecture results in a significant qualitative improvement in the generated images. The change is limited to swapping batch normalization with instance normalization, and to apply the latter both at training and testing times. The resulting method can be used to train high-performance architectures for real-time image generation. The code will is made available on github at https://github.com/DmitryUlyanov/texture_nets. Full paper can be found at arXiv:1701.02096.
Tasks Image Generation, Image Stylization, Style Transfer
Published 2016-07-27
URL http://arxiv.org/abs/1607.08022v3
PDF http://arxiv.org/pdf/1607.08022v3.pdf
PWC https://paperswithcode.com/paper/instance-normalization-the-missing-ingredient
Repo https://github.com/tbullmann/imagetranslation-tensorflow
Framework tf

A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks

Title A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks
Authors Dan Hendrycks, Kevin Gimpel
Abstract We consider the two related problems of detecting if an example is misclassified or out-of-distribution. We present a simple baseline that utilizes probabilities from softmax distributions. Correctly classified examples tend to have greater maximum softmax probabilities than erroneously classified and out-of-distribution examples, allowing for their detection. We assess performance by defining several tasks in computer vision, natural language processing, and automatic speech recognition, showing the effectiveness of this baseline across all. We then show the baseline can sometimes be surpassed, demonstrating the room for future research on these underexplored detection tasks.
Tasks Speech Recognition
Published 2016-10-07
URL http://arxiv.org/abs/1610.02136v3
PDF http://arxiv.org/pdf/1610.02136v3.pdf
PWC https://paperswithcode.com/paper/a-baseline-for-detecting-misclassified-and
Repo https://github.com/dabsdamoon/MNIST-Auxiliary-Decoder
Framework none

IRLS and Slime Mold: Equivalence and Convergence

Title IRLS and Slime Mold: Equivalence and Convergence
Authors Damian Straszak, Nisheeth K. Vishnoi
Abstract In this paper we present a connection between two dynamical systems arising in entirely different contexts: one in signal processing and the other in biology. The first is the famous Iteratively Reweighted Least Squares (IRLS) algorithm used in compressed sensing and sparse recovery while the second is the dynamics of a slime mold (Physarum polycephalum). Both of these dynamics are geared towards finding a minimum l1-norm solution in an affine subspace. Despite its simplicity the convergence of the IRLS method has been shown only for a certain regularization of it and remains an important open problem. Our first result shows that the two dynamics are projections of the same dynamical system in higher dimensions. As a consequence, and building on the recent work on Physarum dynamics, we are able to prove convergence and obtain complexity bounds for a damped version of the IRLS algorithm.
Tasks
Published 2016-01-12
URL http://arxiv.org/abs/1601.02712v1
PDF http://arxiv.org/pdf/1601.02712v1.pdf
PWC https://paperswithcode.com/paper/irls-and-slime-mold-equivalence-and
Repo https://github.com/DamianStraszak/IRLS-and-Physarum-Dynamics
Framework none

Real-time Action Recognition with Enhanced Motion Vector CNNs

Title Real-time Action Recognition with Enhanced Motion Vector CNNs
Authors Bowen Zhang, Limin Wang, Zhe Wang, Yu Qiao, Hanli Wang
Abstract The deep two-stream architecture exhibited excellent performance on video based action recognition. The most computationally expensive step in this approach comes from the calculation of optical flow which prevents it to be real-time. This paper accelerates this architecture by replacing optical flow with motion vector which can be obtained directly from compressed videos without extra calculation. However, motion vector lacks fine structures, and contains noisy and inaccurate motion patterns, leading to the evident degradation of recognition performance. Our key insight for relieving this problem is that optical flow and motion vector are inherent correlated. Transferring the knowledge learned with optical flow CNN to motion vector CNN can significantly boost the performance of the latter. Specifically, we introduce three strategies for this, initialization transfer, supervision transfer and their combination. Experimental results show that our method achieves comparable recognition performance to the state-of-the-art, while our method can process 390.7 frames per second, which is 27 times faster than the original two-stream method.
Tasks Optical Flow Estimation, Temporal Action Localization
Published 2016-04-26
URL http://arxiv.org/abs/1604.07669v1
PDF http://arxiv.org/pdf/1604.07669v1.pdf
PWC https://paperswithcode.com/paper/real-time-action-recognition-with-enhanced
Repo https://github.com/yjxiong/caffe
Framework none

High-Resolution Image Inpainting using Multi-Scale Neural Patch Synthesis

Title High-Resolution Image Inpainting using Multi-Scale Neural Patch Synthesis
Authors Chao Yang, Xin Lu, Zhe Lin, Eli Shechtman, Oliver Wang, Hao Li
Abstract Recent advances in deep learning have shown exciting promise in filling large holes in natural images with semantically plausible and context aware details, impacting fundamental image manipulation tasks such as object removal. While these learning-based methods are significantly more effective in capturing high-level features than prior techniques, they can only handle very low-resolution inputs due to memory limitations and difficulty in training. Even for slightly larger images, the inpainted regions would appear blurry and unpleasant boundaries become visible. We propose a multi-scale neural patch synthesis approach based on joint optimization of image content and texture constraints, which not only preserves contextual structures but also produces high-frequency details by matching and adapting patches with the most similar mid-layer feature correlations of a deep classification network. We evaluate our method on the ImageNet and Paris Streetview datasets and achieved state-of-the-art inpainting accuracy. We show our approach produces sharper and more coherent results than prior methods, especially for high-resolution images.
Tasks Image Inpainting
Published 2016-11-30
URL http://arxiv.org/abs/1611.09969v2
PDF http://arxiv.org/pdf/1611.09969v2.pdf
PWC https://paperswithcode.com/paper/high-resolution-image-inpainting-using-multi
Repo https://github.com/leehomyc/Faster-High-Res-Neural-Inpainting
Framework torch

Finding Better Active Learners for Faster Literature Reviews

Title Finding Better Active Learners for Faster Literature Reviews
Authors Zhe Yu, Nicholas A. Kraft, Tim Menzies
Abstract Literature reviews can be time-consuming and tedious to complete. By cataloging and refactoring three state-of-the-art active learning techniques from evidence-based medicine and legal electronic discovery, this paper finds and implements FASTREAD, a faster technique for studying a large corpus of documents. This paper assesses FASTREAD using datasets generated from existing SE literature reviews (Hall, Wahono, Radjenovi'c, Kitchenham et al.). Compared to manual methods, FASTREAD lets researchers find 95% relevant studies after reviewing an order of magnitude fewer papers. Compared to other state-of-the-art automatic methods, FASTREAD reviews 20-50% fewer studies while finding same number of relevant primary studies in a systematic literature review.
Tasks Active Learning
Published 2016-12-10
URL http://arxiv.org/abs/1612.03224v5
PDF http://arxiv.org/pdf/1612.03224v5.pdf
PWC https://paperswithcode.com/paper/finding-better-active-learners-for-faster
Repo https://github.com/fastread/src
Framework none

MPI-FAUN: An MPI-Based Framework for Alternating-Updating Nonnegative Matrix Factorization

Title MPI-FAUN: An MPI-Based Framework for Alternating-Updating Nonnegative Matrix Factorization
Authors Ramakrishnan Kannan, Grey Ballard, Haesun Park
Abstract Non-negative matrix factorization (NMF) is the problem of determining two non-negative low rank factors $W$ and $H$, for the given input matrix $A$, such that $A \approx W H$. NMF is a useful tool for many applications in different domains such as topic modeling in text mining, background separation in video analysis, and community detection in social networks. Despite its popularity in the data mining community, there is a lack of efficient parallel algorithms to solve the problem for big data sets. The main contribution of this work is a new, high-performance parallel computational framework for a broad class of NMF algorithms that iteratively solves alternating non-negative least squares (NLS) subproblems for $W$ and $H$. It maintains the data and factor matrices in memory (distributed across processors), uses MPI for interprocessor communication, and, in the dense case, provably minimizes communication costs (under mild assumptions). The framework is flexible and able to leverage a variety of NMF and NLS algorithms, including Multiplicative Update, Hierarchical Alternating Least Squares, and Block Principal Pivoting. Our implementation allows us to benchmark and compare different algorithms on massive dense and sparse data matrices of size that spans for few hundreds of millions to billions. We demonstrate the scalability of our algorithm and compare it with baseline implementations, showing significant performance improvements. The code and the datasets used for conducting the experiments are available online.
Tasks Community Detection
Published 2016-09-28
URL http://arxiv.org/abs/1609.09154v1
PDF http://arxiv.org/pdf/1609.09154v1.pdf
PWC https://paperswithcode.com/paper/mpi-faun-an-mpi-based-framework-for
Repo https://github.com/ramkikannan/nmflibrary
Framework none

An Actor-Critic Algorithm for Sequence Prediction

Title An Actor-Critic Algorithm for Sequence Prediction
Authors Dzmitry Bahdanau, Philemon Brakel, Kelvin Xu, Anirudh Goyal, Ryan Lowe, Joelle Pineau, Aaron Courville, Yoshua Bengio
Abstract We present an approach to training neural networks to generate sequences using actor-critic methods from reinforcement learning (RL). Current log-likelihood training methods are limited by the discrepancy between their training and testing modes, as models must generate tokens conditioned on their previous guesses rather than the ground-truth tokens. We address this problem by introducing a \textit{critic} network that is trained to predict the value of an output token, given the policy of an \textit{actor} network. This results in a training procedure that is much closer to the test phase, and allows us to directly optimize for a task-specific score such as BLEU. Crucially, since we leverage these techniques in the supervised learning setting rather than the traditional RL setting, we condition the critic network on the ground-truth output. We show that our method leads to improved performance on both a synthetic task, and for German-English machine translation. Our analysis paves the way for such methods to be applied in natural language generation tasks, such as machine translation, caption generation, and dialogue modelling.
Tasks Machine Translation, Spelling Correction, Text Generation
Published 2016-07-24
URL http://arxiv.org/abs/1607.07086v3
PDF http://arxiv.org/pdf/1607.07086v3.pdf
PWC https://paperswithcode.com/paper/an-actor-critic-algorithm-for-sequence
Repo https://github.com/juliakreutzer/joeynmt
Framework pytorch

Quasi-Recurrent Neural Networks

Title Quasi-Recurrent Neural Networks
Authors James Bradbury, Stephen Merity, Caiming Xiong, Richard Socher
Abstract Recurrent neural networks are a powerful tool for modeling sequential data, but the dependence of each timestep’s computation on the previous timestep’s output limits parallelism and makes RNNs unwieldy for very long sequences. We introduce quasi-recurrent neural networks (QRNNs), an approach to neural sequence modeling that alternates convolutional layers, which apply in parallel across timesteps, and a minimalist recurrent pooling function that applies in parallel across channels. Despite lacking trainable recurrent layers, stacked QRNNs have better predictive accuracy than stacked LSTMs of the same hidden size. Due to their increased parallelism, they are up to 16 times faster at train and test time. Experiments on language modeling, sentiment classification, and character-level neural machine translation demonstrate these advantages and underline the viability of QRNNs as a basic building block for a variety of sequence tasks.
Tasks Language Modelling, Machine Translation, Sentiment Analysis
Published 2016-11-05
URL http://arxiv.org/abs/1611.01576v2
PDF http://arxiv.org/pdf/1611.01576v2.pdf
PWC https://paperswithcode.com/paper/quasi-recurrent-neural-networks
Repo https://github.com/salesforce/pytorch-qrnn
Framework pytorch

Authorship clustering using multi-headed recurrent neural networks

Title Authorship clustering using multi-headed recurrent neural networks
Authors Douglas Bagnall
Abstract A recurrent neural network that has been trained to separately model the language of several documents by unknown authors is used to measure similarity between the documents. It is able to find clues of common authorship even when the documents are very short and about disparate topics. While it is easy to make statistically significant predictions regarding authorship, it is difficult to group documents into definite clusters with high accuracy.
Tasks
Published 2016-08-16
URL http://arxiv.org/abs/1608.04485v1
PDF http://arxiv.org/pdf/1608.04485v1.pdf
PWC https://paperswithcode.com/paper/authorship-clustering-using-multi-headed
Repo https://github.com/douglasbagnall/bog
Framework none

Memory-Efficient Backpropagation Through Time

Title Memory-Efficient Backpropagation Through Time
Authors Audrūnas Gruslys, Remi Munos, Ivo Danihelka, Marc Lanctot, Alex Graves
Abstract We propose a novel approach to reduce memory consumption of the backpropagation through time (BPTT) algorithm when training recurrent neural networks (RNNs). Our approach uses dynamic programming to balance a trade-off between caching of intermediate results and recomputation. The algorithm is capable of tightly fitting within almost any user-set memory budget while finding an optimal execution policy minimizing the computational cost. Computational devices have limited memory capacity and maximizing a computational performance given a fixed memory budget is a practical use-case. We provide asymptotic computational upper bounds for various regimes. The algorithm is particularly effective for long sequences. For sequences of length 1000, our algorithm saves 95% of memory usage while using only one third more time per iteration than the standard BPTT.
Tasks
Published 2016-06-10
URL http://arxiv.org/abs/1606.03401v1
PDF http://arxiv.org/pdf/1606.03401v1.pdf
PWC https://paperswithcode.com/paper/memory-efficient-backpropagation-through-time
Repo https://github.com/cybertronai/gradient-checkpointing
Framework tf

Logic Tensor Networks: Deep Learning and Logical Reasoning from Data and Knowledge

Title Logic Tensor Networks: Deep Learning and Logical Reasoning from Data and Knowledge
Authors Luciano Serafini, Artur d’Avila Garcez
Abstract We propose Logic Tensor Networks: a uniform framework for integrating automatic learning and reasoning. A logic formalism called Real Logic is defined on a first-order language whereby formulas have truth-value in the interval [0,1] and semantics defined concretely on the domain of real numbers. Logical constants are interpreted as feature vectors of real numbers. Real Logic promotes a well-founded integration of deductive reasoning on a knowledge-base and efficient data-driven relational machine learning. We show how Real Logic can be implemented in deep Tensor Neural Networks with the use of Google’s tensorflow primitives. The paper concludes with experiments applying Logic Tensor Networks on a simple but representative example of knowledge completion.
Tasks Tensor Networks
Published 2016-06-14
URL http://arxiv.org/abs/1606.04422v2
PDF http://arxiv.org/pdf/1606.04422v2.pdf
PWC https://paperswithcode.com/paper/logic-tensor-networks-deep-learning-and
Repo https://github.com/thadumi/logictensornetworks-keras
Framework tf
comments powered by Disqus