Paper Group AWR 36
A Mathematical Formalization of Hierarchical Temporal Memory’s Spatial Pooler. Specific Differential Entropy Rate Estimation for Continuous-Valued Time Series. Benchmarking Deep Reinforcement Learning for Continuous Control. RNN Approaches to Text Normalization: A Challenge. Towards Viewpoint Invariant 3D Human Pose Estimation. A budget-constrained …
A Mathematical Formalization of Hierarchical Temporal Memory’s Spatial Pooler
Title | A Mathematical Formalization of Hierarchical Temporal Memory’s Spatial Pooler |
Authors | James Mnatzaganian, Ernest Fokoué, Dhireesha Kudithipudi |
Abstract | Hierarchical temporal memory (HTM) is an emerging machine learning algorithm, with the potential to provide a means to perform predictions on spatiotemporal data. The algorithm, inspired by the neocortex, currently does not have a comprehensive mathematical framework. This work brings together all aspects of the spatial pooler (SP), a critical learning component in HTM, under a single unifying framework. The primary learning mechanism is explored, where a maximum likelihood estimator for determining the degree of permanence update is proposed. The boosting mechanisms are studied and found to be only relevant during the initial few iterations of the network. Observations are made relating HTM to well-known algorithms such as competitive learning and attribute bagging. Methods are provided for using the SP for classification as well as dimensionality reduction. Empirical evidence verifies that given the proper parameterizations, the SP may be used for feature learning. |
Tasks | Dimensionality Reduction |
Published | 2016-01-22 |
URL | http://arxiv.org/abs/1601.06116v3 |
http://arxiv.org/pdf/1601.06116v3.pdf | |
PWC | https://paperswithcode.com/paper/a-mathematical-formalization-of-hierarchical |
Repo | https://github.com/mrkrynmdsco/htm-python |
Framework | none |
Specific Differential Entropy Rate Estimation for Continuous-Valued Time Series
Title | Specific Differential Entropy Rate Estimation for Continuous-Valued Time Series |
Authors | David Darmon |
Abstract | We introduce a method for quantifying the inherent unpredictability of a continuous-valued time series via an extension of the differential Shannon entropy rate. Our extension, the specific entropy rate, quantifies the amount of predictive uncertainty associated with a specific state, rather than averaged over all states. We relate the specific entropy rate to popular `complexity’ measures such as Approximate and Sample Entropies. We provide a data-driven approach for estimating the specific entropy rate of an observed time series. Finally, we consider three case studies of estimating specific entropy rate from synthetic and physiological data relevant to the analysis of heart rate variability. | |
Tasks | Heart Rate Variability, Time Series |
Published | 2016-06-08 |
URL | http://arxiv.org/abs/1606.02615v1 |
http://arxiv.org/pdf/1606.02615v1.pdf | |
PWC | https://paperswithcode.com/paper/specific-differential-entropy-rate-estimation |
Repo | https://github.com/ddarmon/spenra |
Framework | none |
Benchmarking Deep Reinforcement Learning for Continuous Control
Title | Benchmarking Deep Reinforcement Learning for Continuous Control |
Authors | Yan Duan, Xi Chen, Rein Houthooft, John Schulman, Pieter Abbeel |
Abstract | Recently, researchers have made significant progress combining the advances in deep learning for learning feature representations with reinforcement learning. Some notable examples include training agents to play Atari games based on raw pixel data and to acquire advanced manipulation skills using raw sensory inputs. However, it has been difficult to quantify progress in the domain of continuous control due to the lack of a commonly adopted benchmark. In this work, we present a benchmark suite of continuous control tasks, including classic tasks like cart-pole swing-up, tasks with very high state and action dimensionality such as 3D humanoid locomotion, tasks with partial observations, and tasks with hierarchical structure. We report novel findings based on the systematic evaluation of a range of implemented reinforcement learning algorithms. Both the benchmark and reference implementations are released at https://github.com/rllab/rllab in order to facilitate experimental reproducibility and to encourage adoption by other researchers. |
Tasks | Atari Games, Continuous Control |
Published | 2016-04-22 |
URL | http://arxiv.org/abs/1604.06778v3 |
http://arxiv.org/pdf/1604.06778v3.pdf | |
PWC | https://paperswithcode.com/paper/benchmarking-deep-reinforcement-learning-for |
Repo | https://github.com/richardrl/cartpole-request-for-research |
Framework | pytorch |
RNN Approaches to Text Normalization: A Challenge
Title | RNN Approaches to Text Normalization: A Challenge |
Authors | Richard Sproat, Navdeep Jaitly |
Abstract | This paper presents a challenge to the community: given a large corpus of written text aligned to its normalized spoken form, train an RNN to learn the correct normalization function. We present a data set of general text where the normalizations were generated using an existing text normalization component of a text-to-speech system. This data set will be released open-source in the near future. We also present our own experiments with this data set with a variety of different RNN architectures. While some of the architectures do in fact produce very good results when measured in terms of overall accuracy, the errors that are produced are problematic, since they would convey completely the wrong message if such a system were deployed in a speech application. On the other hand, we show that a simple FST-based filter can mitigate those errors, and achieve a level of accuracy not achievable by the RNN alone. Though our conclusions are largely negative on this point, we are actually not arguing that the text normalization problem is intractable using an pure RNN approach, merely that it is not going to be something that can be solved merely by having huge amounts of annotated text data and feeding that to a general RNN model. And when we open-source our data, we will be providing a novel data set for sequence-to-sequence modeling in the hopes that the the community can find better solutions. The data used in this work have been released and are available at: https://github.com/rwsproat/text-normalization-data |
Tasks | |
Published | 2016-10-31 |
URL | http://arxiv.org/abs/1611.00068v2 |
http://arxiv.org/pdf/1611.00068v2.pdf | |
PWC | https://paperswithcode.com/paper/rnn-approaches-to-text-normalization-a |
Repo | https://github.com/rwsproat/text-normalization-data |
Framework | none |
Towards Viewpoint Invariant 3D Human Pose Estimation
Title | Towards Viewpoint Invariant 3D Human Pose Estimation |
Authors | Albert Haque, Boya Peng, Zelun Luo, Alexandre Alahi, Serena Yeung, Li Fei-Fei |
Abstract | We propose a viewpoint invariant model for 3D human pose estimation from a single depth image. To achieve this, our discriminative model embeds local regions into a learned viewpoint invariant feature space. Formulated as a multi-task learning problem, our model is able to selectively predict partial poses in the presence of noise and occlusion. Our approach leverages a convolutional and recurrent network architecture with a top-down error feedback mechanism to self-correct previous pose estimates in an end-to-end manner. We evaluate our model on a previously published depth dataset and a newly collected human pose dataset containing 100K annotated depth images from extreme viewpoints. Experiments show that our model achieves competitive performance on frontal views while achieving state-of-the-art performance on alternate viewpoints. |
Tasks | 3D Human Pose Estimation, Multi-Task Learning, Pose Estimation |
Published | 2016-03-23 |
URL | http://arxiv.org/abs/1603.07076v3 |
http://arxiv.org/pdf/1603.07076v3.pdf | |
PWC | https://paperswithcode.com/paper/towards-viewpoint-invariant-3d-human-pose |
Repo | https://github.com/mks0601/V2V-PoseNet_RELEASE |
Framework | pytorch |
A budget-constrained inverse classification framework for smooth classifiers
Title | A budget-constrained inverse classification framework for smooth classifiers |
Authors | Michael T. Lash, Qihang Lin, W. Nick Street, Jennifer G. Robinson |
Abstract | Inverse classification is the process of manipulating an instance such that it is more likely to conform to a specific class. Past methods that address such a problem have shortcomings. Greedy methods make changes that are overly radical, often relying on data that is strictly discrete. Other methods rely on certain data points, the presence of which cannot be guaranteed. In this paper we propose a general framework and method that overcomes these and other limitations. The formulation of our method can use any differentiable classification function. We demonstrate the method by using logistic regression and Gaussian kernel SVMs. We constrain the inverse classification to occur on features that can actually be changed, each of which incurs an individual cost. We further subject such changes to fall within a certain level of cumulative change (budget). Our framework can also accommodate the estimation of (indirectly changeable) features whose values change as a consequence of actions taken. Furthermore, we propose two methods for specifying feature-value ranges that result in different algorithmic behavior. We apply our method, and a proposed sensitivity analysis-based benchmark method, to two freely available datasets: Student Performance from the UCI Machine Learning Repository and a real world cardiovascular disease dataset. The results obtained demonstrate the validity and benefits of our framework and method. |
Tasks | |
Published | 2016-05-29 |
URL | http://arxiv.org/abs/1605.09068v3 |
http://arxiv.org/pdf/1605.09068v3.pdf | |
PWC | https://paperswithcode.com/paper/a-budget-constrained-inverse-classification |
Repo | https://github.com/michael-lash/BCIC |
Framework | none |
A Systematic Approach to Blocking Convolutional Neural Networks
Title | A Systematic Approach to Blocking Convolutional Neural Networks |
Authors | Xuan Yang, Jing Pu, Blaine Burton Rister, Nikhil Bhagdikar, Stephen Richardson, Shahar Kvatinsky, Jonathan Ragan-Kelley, Ardavan Pedram, Mark Horowitz |
Abstract | Convolutional Neural Networks (CNNs) are the state of the art solution for many computer vision problems, and many researchers have explored optimized implementations. Most implementations heuristically block the computation to deal with the large data sizes and high data reuse of CNNs. This paper explores how to block CNN computations for memory locality by creating an analytical model for CNN-like loop nests. Using this model we automatically derive optimized blockings for common networks that improve the energy efficiency of custom hardware implementations by up to an order of magnitude. Compared to traditional CNN CPU implementations based on highly-tuned, hand-optimized BLAS libraries,our x86 programs implementing the optimal blocking reduce the number of memory accesses by up to 90%. |
Tasks | |
Published | 2016-06-14 |
URL | http://arxiv.org/abs/1606.04209v1 |
http://arxiv.org/pdf/1606.04209v1.pdf | |
PWC | https://paperswithcode.com/paper/a-systematic-approach-to-blocking |
Repo | https://github.com/stanford-mast/nn_dataflow |
Framework | none |
TorchCraft: a Library for Machine Learning Research on Real-Time Strategy Games
Title | TorchCraft: a Library for Machine Learning Research on Real-Time Strategy Games |
Authors | Gabriel Synnaeve, Nantas Nardelli, Alex Auvolat, Soumith Chintala, Timothée Lacroix, Zeming Lin, Florian Richoux, Nicolas Usunier |
Abstract | We present TorchCraft, a library that enables deep learning research on Real-Time Strategy (RTS) games such as StarCraft: Brood War, by making it easier to control these games from a machine learning framework, here Torch. This white paper argues for using RTS games as a benchmark for AI research, and describes the design and components of TorchCraft. |
Tasks | Real-Time Strategy Games, Starcraft |
Published | 2016-11-01 |
URL | http://arxiv.org/abs/1611.00625v2 |
http://arxiv.org/pdf/1611.00625v2.pdf | |
PWC | https://paperswithcode.com/paper/torchcraft-a-library-for-machine-learning |
Repo | https://github.com/TorchCraft/TorchCraft |
Framework | torch |
Towards a Job Title Classification System
Title | Towards a Job Title Classification System |
Authors | Faizan Javed, Matt McNair, Ferosh Jacob, Meng Zhao |
Abstract | Document classification for text, images and other applicable entities has long been a focus of research in academia and also finds application in many industrial settings. Amidst a plethora of approaches to solve such problems, machine-learning techniques have found success in a variety of scenarios. In this paper we discuss the design of a machine learning-based semi-supervised job title classification system for the online job recruitment domain currently in production at CareerBuilder.com and propose enhancements to it. The system leverages a varied collection of classification as well clustering algorithms. These algorithms are encompassed in an architecture that facilitates leveraging existing off-the-shelf machine learning tools and techniques while keeping into consideration the challenges of constructing a scalable classification system for a large taxonomy of categories. As a continuously evolving system that is still under development we first discuss the existing semi-supervised classification system which is composed of both clustering and classification components in a proximity-based classifier setup and results of which are already used across numerous products at CareerBuilder. We then elucidate our long-term goals for job title classification and propose enhancements to the existing system in the form of a two-stage coarse and fine level classifier augmentation to construct a cascade of hierarchical vertical classifiers. Preliminary results are presented using experimental evaluation on real world industrial data. |
Tasks | Document Classification |
Published | 2016-06-02 |
URL | http://arxiv.org/abs/1606.00917v1 |
http://arxiv.org/pdf/1606.00917v1.pdf | |
PWC | https://paperswithcode.com/paper/towards-a-job-title-classification-system |
Repo | https://github.com/kmamykin/askamanager_salary_survey |
Framework | none |
Texture Synthesis with Spatial Generative Adversarial Networks
Title | Texture Synthesis with Spatial Generative Adversarial Networks |
Authors | Nikolay Jetchev, Urs Bergmann, Roland Vollgraf |
Abstract | Generative adversarial networks (GANs) are a recent approach to train generative models of data, which have been shown to work particularly well on image data. In the current paper we introduce a new model for texture synthesis based on GAN learning. By extending the input noise distribution space from a single vector to a whole spatial tensor, we create an architecture with properties well suited to the task of texture synthesis, which we call spatial GAN (SGAN). To our knowledge, this is the first successful completely data-driven texture synthesis method based on GANs. Our method has the following features which make it a state of the art algorithm for texture synthesis: high image quality of the generated textures, very high scalability w.r.t. the output texture size, fast real-time forward generation, the ability to fuse multiple diverse source images in complex textures. To illustrate these capabilities we present multiple experiments with different classes of texture images and use cases. We also discuss some limitations of our method with respect to the types of texture images it can synthesize, and compare it to other neural techniques for texture generation. |
Tasks | Texture Synthesis |
Published | 2016-11-24 |
URL | http://arxiv.org/abs/1611.08207v4 |
http://arxiv.org/pdf/1611.08207v4.pdf | |
PWC | https://paperswithcode.com/paper/texture-synthesis-with-spatial-generative |
Repo | https://github.com/zalandoresearch/spatial_gan |
Framework | none |
Deep Multi-fidelity Gaussian Processes
Title | Deep Multi-fidelity Gaussian Processes |
Authors | Maziar Raissi, George Karniadakis |
Abstract | We develop a novel multi-fidelity framework that goes far beyond the classical AR(1) Co-kriging scheme of Kennedy and O’Hagan (2000). Our method can handle general discontinuous cross-correlations among systems with different levels of fidelity. A combination of multi-fidelity Gaussian Processes (AR(1) Co-kriging) and deep neural networks enables us to construct a method that is immune to discontinuities. We demonstrate the effectiveness of the new technology using standard benchmark problems designed to resemble the outputs of complicated high- and low-fidelity codes. |
Tasks | Gaussian Processes |
Published | 2016-04-26 |
URL | http://arxiv.org/abs/1604.07484v1 |
http://arxiv.org/pdf/1604.07484v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-multi-fidelity-gaussian-processes |
Repo | https://github.com/maziarraissi/TutorialGP |
Framework | none |
Incorporating Relation Paths in Neural Relation Extraction
Title | Incorporating Relation Paths in Neural Relation Extraction |
Authors | Wenyuan Zeng, Yankai Lin, Zhiyuan Liu, Maosong Sun |
Abstract | Distantly supervised relation extraction has been widely used to find novel relational facts from plain text. To predict the relation between a pair of two target entities, existing methods solely rely on those direct sentences containing both entities. In fact, there are also many sentences containing only one of the target entities, which provide rich and useful information for relation extraction. To address this issue, we build inference chains between two target entities via intermediate entities, and propose a path-based neural relation extraction model to encode the relational semantics from both direct sentences and inference chains. Experimental results on real-world datasets show that, our model can make full use of those sentences containing only one target entity, and achieves significant and consistent improvements on relation extraction as compared with baselines. The source code of this paper can be obtained from https: //github.com/thunlp/PathNRE. |
Tasks | Relation Extraction |
Published | 2016-09-23 |
URL | http://arxiv.org/abs/1609.07479v3 |
http://arxiv.org/pdf/1609.07479v3.pdf | |
PWC | https://paperswithcode.com/paper/incorporating-relation-paths-in-neural |
Repo | https://github.com/thunlp/PathNRE |
Framework | none |
DeepMath - Deep Sequence Models for Premise Selection
Title | DeepMath - Deep Sequence Models for Premise Selection |
Authors | Alex A. Alemi, Francois Chollet, Niklas Een, Geoffrey Irving, Christian Szegedy, Josef Urban |
Abstract | We study the effectiveness of neural sequence models for premise selection in automated theorem proving, one of the main bottlenecks in the formalization of mathematics. We propose a two stage approach for this task that yields good results for the premise selection task on the Mizar corpus while avoiding the hand-engineered features of existing state-of-the-art models. To our knowledge, this is the first time deep learning has been applied to theorem proving on a large scale. |
Tasks | Automated Theorem Proving |
Published | 2016-06-14 |
URL | http://arxiv.org/abs/1606.04442v2 |
http://arxiv.org/pdf/1606.04442v2.pdf | |
PWC | https://paperswithcode.com/paper/deepmath-deep-sequence-models-for-premise |
Repo | https://github.com/BartoszPiotrowski/ATPboost |
Framework | none |
Deep Reinforcement Learning with a Combinatorial Action Space for Predicting Popular Reddit Threads
Title | Deep Reinforcement Learning with a Combinatorial Action Space for Predicting Popular Reddit Threads |
Authors | Ji He, Mari Ostendorf, Xiaodong He, Jianshu Chen, Jianfeng Gao, Lihong Li, Li Deng |
Abstract | We introduce an online popularity prediction and tracking task as a benchmark task for reinforcement learning with a combinatorial, natural language action space. A specified number of discussion threads predicted to be popular are recommended, chosen from a fixed window of recent comments to track. Novel deep reinforcement learning architectures are studied for effective modeling of the value function associated with actions comprised of interdependent sub-actions. The proposed model, which represents dependence between sub-actions through a bi-directional LSTM, gives the best performance across different experimental configurations and domains, and it also generalizes well with varying numbers of recommendation requests. |
Tasks | |
Published | 2016-06-12 |
URL | http://arxiv.org/abs/1606.03667v4 |
http://arxiv.org/pdf/1606.03667v4.pdf | |
PWC | https://paperswithcode.com/paper/deep-reinforcement-learning-with-a |
Repo | https://github.com/jvking/reddit-RL-simulator |
Framework | none |
AC-BLSTM: Asymmetric Convolutional Bidirectional LSTM Networks for Text Classification
Title | AC-BLSTM: Asymmetric Convolutional Bidirectional LSTM Networks for Text Classification |
Authors | Depeng Liang, Yongdong Zhang |
Abstract | Recently deeplearning models have been shown to be capable of making remarkable performance in sentences and documents classification tasks. In this work, we propose a novel framework called AC-BLSTM for modeling sentences and documents, which combines the asymmetric convolution neural network (ACNN) with the Bidirectional Long Short-Term Memory network (BLSTM). Experiment results demonstrate that our model achieves state-of-the-art results on five tasks, including sentiment analysis, question type classification, and subjectivity classification. In order to further improve the performance of AC-BLSTM, we propose a semi-supervised learning framework called G-AC-BLSTM for text classification by combining the generative model with AC-BLSTM. |
Tasks | Sentence Embeddings, Sentiment Analysis, Text Classification |
Published | 2016-11-07 |
URL | http://arxiv.org/abs/1611.01884v3 |
http://arxiv.org/pdf/1611.01884v3.pdf | |
PWC | https://paperswithcode.com/paper/ac-blstm-asymmetric-convolutional |
Repo | https://github.com/Ldpe2G/AC-BLSTM |
Framework | mxnet |