Paper Group AWR 81
Learning More Universal Representations for Transfer-Learning. Conjugate-Computation Variational Inference : Converting Variational Inference in Non-Conjugate Models to Inferences in Conjugate Models. Deep Competitive Pathway Networks. One-Shot Reinforcement Learning for Robot Navigation with Interactive Replay. The Code2Text Challenge: Text Genera …
Learning More Universal Representations for Transfer-Learning
Title | Learning More Universal Representations for Transfer-Learning |
Authors | Youssef Tamaazousti, Hervé Le Borgne, Céline Hudelot, Mohamed El Amine Seddik, Mohamed Tamaazousti |
Abstract | A representation is supposed universal if it encodes any element of the visual world (e.g., objects, scenes) in any configuration (e.g., scale, context). While not expecting pure universal representations, the goal in the literature is to improve the universality level, starting from a representation with a certain level. To do so, the state-of-the-art consists in learning CNN-based representations on a diversified training problem (e.g., ImageNet modified by adding annotated data). While it effectively increases universality, such approach still requires a large amount of efforts to satisfy the needs in annotated data. In this work, we propose two methods to improve universality, but pay special attention to limit the need of annotated data. We also propose a unified framework of the methods based on the diversifying of the training problem. Finally, to better match Atkinson’s cognitive study about universal human representations, we proposed to rely on the transfer-learning scheme as well as a new metric to evaluate universality. This latter, aims us to demonstrates the interest of our methods on 10 target-problems, relating to the classification task and a variety of visual domains. |
Tasks | Transfer Learning |
Published | 2017-12-27 |
URL | http://arxiv.org/abs/1712.09708v5 |
http://arxiv.org/pdf/1712.09708v5.pdf | |
PWC | https://paperswithcode.com/paper/learning-more-universal-representations-for |
Repo | https://github.com/youssefTamaazousti/MulDiPNet |
Framework | tf |
Conjugate-Computation Variational Inference : Converting Variational Inference in Non-Conjugate Models to Inferences in Conjugate Models
Title | Conjugate-Computation Variational Inference : Converting Variational Inference in Non-Conjugate Models to Inferences in Conjugate Models |
Authors | Mohammad Emtiyaz Khan, Wu Lin |
Abstract | Variational inference is computationally challenging in models that contain both conjugate and non-conjugate terms. Methods specifically designed for conjugate models, even though computationally efficient, find it difficult to deal with non-conjugate terms. On the other hand, stochastic-gradient methods can handle the non-conjugate terms but they usually ignore the conjugate structure of the model which might result in slow convergence. In this paper, we propose a new algorithm called Conjugate-computation Variational Inference (CVI) which brings the best of the two worlds together – it uses conjugate computations for the conjugate terms and employs stochastic gradients for the rest. We derive this algorithm by using a stochastic mirror-descent method in the mean-parameter space, and then expressing each gradient step as a variational inference in a conjugate model. We demonstrate our algorithm’s applicability to a large class of models and establish its convergence. Our experimental results show that our method converges much faster than the methods that ignore the conjugate structure of the model. |
Tasks | |
Published | 2017-03-13 |
URL | http://arxiv.org/abs/1703.04265v2 |
http://arxiv.org/pdf/1703.04265v2.pdf | |
PWC | https://paperswithcode.com/paper/conjugate-computation-variational-inference |
Repo | https://github.com/vmasrani/CVI_PLDS |
Framework | none |
Deep Competitive Pathway Networks
Title | Deep Competitive Pathway Networks |
Authors | Jia-Ren Chang, Yong-Sheng Chen |
Abstract | In the design of deep neural architectures, recent studies have demonstrated the benefits of grouping subnetworks into a larger network. For examples, the Inception architecture integrates multi-scale subnetworks and the residual network can be regarded that a residual unit combines a residual subnetwork with an identity shortcut. In this work, we embrace this observation and propose the Competitive Pathway Network (CoPaNet). The CoPaNet comprises a stack of competitive pathway units and each unit contains multiple parallel residual-type subnetworks followed by a max operation for feature competition. This mechanism enhances the model capability by learning a variety of features in subnetworks. The proposed strategy explicitly shows that the features propagate through pathways in various routing patterns, which is referred to as pathway encoding of category information. Moreover, the cross-block shortcut can be added to the CoPaNet to encourage feature reuse. We evaluated the proposed CoPaNet on four object recognition benchmarks: CIFAR-10, CIFAR-100, SVHN, and ImageNet. CoPaNet obtained the state-of-the-art or comparable results using similar amounts of parameters. The code of CoPaNet is available at: https://github.com/JiaRenChang/CoPaNet. |
Tasks | Image Classification, Object Recognition |
Published | 2017-09-29 |
URL | http://arxiv.org/abs/1709.10282v1 |
http://arxiv.org/pdf/1709.10282v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-competitive-pathway-networks |
Repo | https://github.com/JiaRenChang/CoPaNet |
Framework | torch |
One-Shot Reinforcement Learning for Robot Navigation with Interactive Replay
Title | One-Shot Reinforcement Learning for Robot Navigation with Interactive Replay |
Authors | Jake Bruce, Niko Suenderhauf, Piotr Mirowski, Raia Hadsell, Michael Milford |
Abstract | Recently, model-free reinforcement learning algorithms have been shown to solve challenging problems by learning from extensive interaction with the environment. A significant issue with transferring this success to the robotics domain is that interaction with the real world is costly, but training on limited experience is prone to overfitting. We present a method for learning to navigate, to a fixed goal and in a known environment, on a mobile robot. The robot leverages an interactive world model built from a single traversal of the environment, a pre-trained visual feature encoder, and stochastic environmental augmentation, to demonstrate successful zero-shot transfer under real-world environmental variations without fine-tuning. |
Tasks | Robot Navigation |
Published | 2017-11-28 |
URL | http://arxiv.org/abs/1711.10137v2 |
http://arxiv.org/pdf/1711.10137v2.pdf | |
PWC | https://paperswithcode.com/paper/one-shot-reinforcement-learning-for-robot |
Repo | https://github.com/ayusefi/Localization-Papers |
Framework | none |
The Code2Text Challenge: Text Generation in Source Code Libraries
Title | The Code2Text Challenge: Text Generation in Source Code Libraries |
Authors | Kyle Richardson, Sina Zarrieß, Jonas Kuhn |
Abstract | We propose a new shared task for tactical data-to-text generation in the domain of source code libraries. Specifically, we focus on text generation of function descriptions from example software projects. Data is drawn from existing resources used for studying the related problem of semantic parser induction (Richardson and Kuhn, 2017b; Richardson and Kuhn, 2017a), and spans a wide variety of both natural languages and programming languages. In this paper, we describe these existing resources, which will serve as training and development data for the task, and discuss plans for building new independent test sets. |
Tasks | Data-to-Text Generation, Text Generation |
Published | 2017-07-31 |
URL | http://arxiv.org/abs/1708.00098v1 |
http://arxiv.org/pdf/1708.00098v1.pdf | |
PWC | https://paperswithcode.com/paper/the-code2text-challenge-text-generation-in |
Repo | https://github.com/yakazimir/Code-Datasets |
Framework | none |
Multi-Scale Dense Networks for Resource Efficient Image Classification
Title | Multi-Scale Dense Networks for Resource Efficient Image Classification |
Authors | Gao Huang, Danlu Chen, Tianhong Li, Felix Wu, Laurens van der Maaten, Kilian Q. Weinberger |
Abstract | In this paper we investigate image classification with computational resource limits at test time. Two such settings are: 1. anytime classification, where the network’s prediction for a test example is progressively updated, facilitating the output of a prediction at any time; and 2. budgeted batch classification, where a fixed amount of computation is available to classify a set of examples that can be spent unevenly across “easier” and “harder” inputs. In contrast to most prior work, such as the popular Viola and Jones algorithm, our approach is based on convolutional neural networks. We train multiple classifiers with varying resource demands, which we adaptively apply during test time. To maximally re-use computation between the classifiers, we incorporate them as early-exits into a single deep convolutional neural network and inter-connect them with dense connectivity. To facilitate high quality classification early on, we use a two-dimensional multi-scale network architecture that maintains coarse and fine level features all-throughout the network. Experiments on three image-classification tasks demonstrate that our framework substantially improves the existing state-of-the-art in both settings. |
Tasks | Image Classification |
Published | 2017-03-29 |
URL | http://arxiv.org/abs/1703.09844v5 |
http://arxiv.org/pdf/1703.09844v5.pdf | |
PWC | https://paperswithcode.com/paper/multi-scale-dense-networks-for-resource |
Repo | https://github.com/gaohuang/MSDNet |
Framework | pytorch |
LEPOR: An Augmented Machine Translation Evaluation Metric
Title | LEPOR: An Augmented Machine Translation Evaluation Metric |
Authors | Aaron Li-Feng Han |
Abstract | Machine translation (MT) was developed as one of the hottest research topics in the natural language processing (NLP) literature. One important issue in MT is that how to evaluate the MT system reasonably and tell us whether the translation system makes an improvement or not. The traditional manual judgment methods are expensive, time-consuming, unrepeatable, and sometimes with low agreement. On the other hand, the popular automatic MT evaluation methods have some weaknesses. Firstly, they tend to perform well on the language pairs with English as the target language, but weak when English is used as source. Secondly, some methods rely on many additional linguistic features to achieve good performance, which makes the metric unable to replicate and apply to other language pairs easily. Thirdly, some popular metrics utilize incomprehensive factors, which result in low performance on some practical tasks. In this thesis, to address the existing problems, we design novel MT evaluation methods and investigate their performances on different languages. Firstly, we design augmented factors to yield highly accurate evaluation.Secondly, we design a tunable evaluation model where weighting of factors can be optimised according to the characteristics of languages. Thirdly, in the enhanced version of our methods, we design concise linguistic feature using POS to show that our methods can yield even higher performance when using some external linguistic resources. Finally, we introduce the practical performance of our metrics in the ACL-WMT workshop shared tasks, which show that the proposed methods are robust across different languages. |
Tasks | Machine Translation |
Published | 2017-03-26 |
URL | http://arxiv.org/abs/1703.08748v1 |
http://arxiv.org/pdf/1703.08748v1.pdf | |
PWC | https://paperswithcode.com/paper/lepor-an-augmented-machine-translation |
Repo | https://github.com/poethan/LEPOR |
Framework | none |
Safety-Aware Apprenticeship Learning
Title | Safety-Aware Apprenticeship Learning |
Authors | Weichao Zhou, Wenchao Li |
Abstract | Apprenticeship learning (AL) is a kind of Learning from Demonstration techniques where the reward function of a Markov Decision Process (MDP) is unknown to the learning agent and the agent has to derive a good policy by observing an expert’s demonstrations. In this paper, we study the problem of how to make AL algorithms inherently safe while still meeting its learning objective. We consider a setting where the unknown reward function is assumed to be a linear combination of a set of state features, and the safety property is specified in Probabilistic Computation Tree Logic (PCTL). By embedding probabilistic model checking inside AL, we propose a novel counterexample-guided approach that can ensure safety while retaining performance of the learnt policy. We demonstrate the effectiveness of our approach on several challenging AL scenarios where safety is essential. |
Tasks | |
Published | 2017-10-22 |
URL | http://arxiv.org/abs/1710.07983v4 |
http://arxiv.org/pdf/1710.07983v4.pdf | |
PWC | https://paperswithcode.com/paper/safety-aware-apprenticeship-learning |
Repo | https://github.com/zwc662/CAV2018 |
Framework | none |
LIUM Machine Translation Systems for WMT17 News Translation Task
Title | LIUM Machine Translation Systems for WMT17 News Translation Task |
Authors | Mercedes García-Martínez, Ozan Caglayan, Walid Aransa, Adrien Bardet, Fethi Bougares, Loïc Barrault |
Abstract | This paper describes LIUM submissions to WMT17 News Translation Task for English-German, English-Turkish, English-Czech and English-Latvian language pairs. We train BPE-based attentive Neural Machine Translation systems with and without factored outputs using the open source nmtpy framework. Competitive scores were obtained by ensembling various systems and exploiting the availability of target monolingual corpora for back-translation. The impact of back-translation quantity and quality is also analyzed for English-Turkish where our post-deadline submission surpassed the best entry by +1.6 BLEU. |
Tasks | Machine Translation |
Published | 2017-07-14 |
URL | http://arxiv.org/abs/1707.04499v1 |
http://arxiv.org/pdf/1707.04499v1.pdf | |
PWC | https://paperswithcode.com/paper/lium-machine-translation-systems-for-wmt17 |
Repo | https://github.com/lium-lst/wmt17-newstask |
Framework | none |
Sharing deep generative representation for perceived image reconstruction from human brain activity
Title | Sharing deep generative representation for perceived image reconstruction from human brain activity |
Authors | Changde Du, Changying Du, Huiguang He |
Abstract | Decoding human brain activities via functional magnetic resonance imaging (fMRI) has gained increasing attention in recent years. While encouraging results have been reported in brain states classification tasks, reconstructing the details of human visual experience still remains difficult. Two main challenges that hinder the development of effective models are the perplexing fMRI measurement noise and the high dimensionality of limited data instances. Existing methods generally suffer from one or both of these issues and yield dissatisfactory results. In this paper, we tackle this problem by casting the reconstruction of visual stimulus as the Bayesian inference of missing view in a multiview latent variable model. Sharing a common latent representation, our joint generative model of external stimulus and brain response is not only “deep” in extracting nonlinear features from visual images, but also powerful in capturing correlations among voxel activities of fMRI recordings. The nonlinearity and deep structure endow our model with strong representation ability, while the correlations of voxel activities are critical for suppressing noise and improving prediction. We devise an efficient variational Bayesian method to infer the latent variables and the model parameters. To further improve the reconstruction accuracy, the latent representations of testing instances are enforced to be close to that of their neighbours from the training set via posterior regularization. Experiments on three fMRI recording datasets demonstrate that our approach can more accurately reconstruct visual stimuli. |
Tasks | Bayesian Inference, Image Reconstruction |
Published | 2017-04-25 |
URL | http://arxiv.org/abs/1704.07575v3 |
http://arxiv.org/pdf/1704.07575v3.pdf | |
PWC | https://paperswithcode.com/paper/sharing-deep-generative-representation-for |
Repo | https://github.com/ChangdeDu/DGMM |
Framework | none |
Unsupervised Learning by Predicting Noise
Title | Unsupervised Learning by Predicting Noise |
Authors | Piotr Bojanowski, Armand Joulin |
Abstract | Convolutional neural networks provide visual features that perform remarkably well in many computer vision applications. However, training these networks requires significant amounts of supervision. This paper introduces a generic framework to train deep networks, end-to-end, with no supervision. We propose to fix a set of target representations, called Noise As Targets (NAT), and to constrain the deep features to align to them. This domain agnostic approach avoids the standard unsupervised learning issues of trivial solutions and collapsing of features. Thanks to a stochastic batch reassignment strategy and a separable square loss function, it scales to millions of images. The proposed approach produces representations that perform on par with state-of-the-art unsupervised methods on ImageNet and Pascal VOC. |
Tasks | |
Published | 2017-04-18 |
URL | http://arxiv.org/abs/1704.05310v1 |
http://arxiv.org/pdf/1704.05310v1.pdf | |
PWC | https://paperswithcode.com/paper/unsupervised-learning-by-predicting-noise |
Repo | https://github.com/facebookresearch/noise-as-targets |
Framework | torch |
A Generalization of Convolutional Neural Networks to Graph-Structured Data
Title | A Generalization of Convolutional Neural Networks to Graph-Structured Data |
Authors | Yotam Hechtlinger, Purvasha Chakravarti, Jining Qin |
Abstract | This paper introduces a generalization of Convolutional Neural Networks (CNNs) from low-dimensional grid data, such as images, to graph-structured data. We propose a novel spatial convolution utilizing a random walk to uncover the relations within the input, analogous to the way the standard convolution uses the spatial neighborhood of a pixel on the grid. The convolution has an intuitive interpretation, is efficient and scalable and can also be used on data with varying graph structure. Furthermore, this generalization can be applied to many standard regression or classification problems, by learning the the underlying graph. We empirically demonstrate the performance of the proposed CNN on MNIST, and challenge the state-of-the-art on Merck molecular activity data set. |
Tasks | |
Published | 2017-04-26 |
URL | http://arxiv.org/abs/1704.08165v1 |
http://arxiv.org/pdf/1704.08165v1.pdf | |
PWC | https://paperswithcode.com/paper/a-generalization-of-convolutional-neural |
Repo | https://github.com/hechtlinger/graph_cnn |
Framework | none |
Image reconstruction by domain transform manifold learning
Title | Image reconstruction by domain transform manifold learning |
Authors | Bo Zhu, Jeremiah Z. Liu, Bruce R. Rosen, Matthew S. Rosen |
Abstract | Image reconstruction plays a critical role in the implementation of all contemporary imaging modalities across the physical and life sciences including optical, MRI, CT, PET, and radio astronomy. During an image acquisition, the sensor encodes an intermediate representation of an object in the sensor domain, which is subsequently reconstructed into an image by an inversion of the encoding function. Image reconstruction is challenging because analytic knowledge of the inverse transform may not exist a priori, especially in the presence of sensor non-idealities and noise. Thus, the standard reconstruction approach involves approximating the inverse function with multiple ad hoc stages in a signal processing chain whose composition depends on the details of each acquisition strategy, and often requires expert parameter tuning to optimize reconstruction performance. We present here a unified framework for image reconstruction, AUtomated TransfOrm by Manifold APproximation (AUTOMAP), which recasts image reconstruction as a data-driven, supervised learning task that allows a mapping between sensor and image domain to emerge from an appropriate corpus of training data. We implement AUTOMAP with a deep neural network and exhibit its flexibility in learning reconstruction transforms for a variety of MRI acquisition strategies, using the same network architecture and hyperparameters. We further demonstrate its efficiency in sparsely representing transforms along low-dimensional manifolds, resulting in superior immunity to noise and reconstruction artifacts compared with conventional handcrafted reconstruction methods. In addition to improving the reconstruction performance of existing acquisition methodologies, we anticipate accelerating the discovery of new acquisition strategies across modalities as the burden of reconstruction becomes lifted by AUTOMAP and learned-reconstruction approaches. |
Tasks | Image Reconstruction |
Published | 2017-04-28 |
URL | http://arxiv.org/abs/1704.08841v1 |
http://arxiv.org/pdf/1704.08841v1.pdf | |
PWC | https://paperswithcode.com/paper/image-reconstruction-by-domain-transform |
Repo | https://github.com/chongduan/MRI-AUTOMAP |
Framework | tf |
Selective Encoding for Abstractive Sentence Summarization
Title | Selective Encoding for Abstractive Sentence Summarization |
Authors | Qingyu Zhou, Nan Yang, Furu Wei, Ming Zhou |
Abstract | We propose a selective encoding model to extend the sequence-to-sequence framework for abstractive sentence summarization. It consists of a sentence encoder, a selective gate network, and an attention equipped decoder. The sentence encoder and decoder are built with recurrent neural networks. The selective gate network constructs a second level sentence representation by controlling the information flow from encoder to decoder. The second level representation is tailored for sentence summarization task, which leads to better performance. We evaluate our model on the English Gigaword, DUC 2004 and MSR abstractive sentence summarization datasets. The experimental results show that the proposed selective encoding model outperforms the state-of-the-art baseline models. |
Tasks | Abstractive Sentence Summarization |
Published | 2017-04-24 |
URL | http://arxiv.org/abs/1704.07073v1 |
http://arxiv.org/pdf/1704.07073v1.pdf | |
PWC | https://paperswithcode.com/paper/selective-encoding-for-abstractive-sentence |
Repo | https://github.com/toru34/zhou_acl_2017 |
Framework | none |
A Tutorial on Deep Learning for Music Information Retrieval
Title | A Tutorial on Deep Learning for Music Information Retrieval |
Authors | Keunwoo Choi, György Fazekas, Kyunghyun Cho, Mark Sandler |
Abstract | Following their success in Computer Vision and other areas, deep learning techniques have recently become widely adopted in Music Information Retrieval (MIR) research. However, the majority of works aim to adopt and assess methods that have been shown to be effective in other domains, while there is still a great need for more original research focusing on music primarily and utilising musical knowledge and insight. The goal of this paper is to boost the interest of beginners by providing a comprehensive tutorial and reducing the barriers to entry into deep learning for MIR. We lay out the basic principles and review prominent works in this hard to navigate the field. We then outline the network structures that have been successful in MIR problems and facilitate the selection of building blocks for the problems at hand. Finally, guidelines for new tasks and some advanced topics in deep learning are discussed to stimulate new research in this fascinating field. |
Tasks | Information Retrieval, Music Information Retrieval |
Published | 2017-09-13 |
URL | http://arxiv.org/abs/1709.04396v2 |
http://arxiv.org/pdf/1709.04396v2.pdf | |
PWC | https://paperswithcode.com/paper/a-tutorial-on-deep-learning-for-music |
Repo | https://github.com/keunwoochoi/dl4mir |
Framework | none |