July 28, 2019

2706 words 13 mins read

Paper Group ANR 290

Paper Group ANR 290

A Parameter-Free Learning Automaton Scheme. Rapid Probabilistic Interest Learning from Domain-Specific Pairwise Image Comparisons. Neural Graph Machines: Learning Neural Networks Using Graphs. Lower Bound On the Computational Complexity of Discounted Markov Decision Problems. Path-Based Attention Neural Model for Fine-Grained Entity Typing. Exploit …

A Parameter-Free Learning Automaton Scheme

Title A Parameter-Free Learning Automaton Scheme
Authors Hao Ge
Abstract For a learning automaton, a proper configuration of its learning parameters, which are crucial for the automaton’s performance, is relatively difficult due to the necessity of a manual parameter tuning before real applications. To ensure a stable and reliable performance in stochastic environments, parameter tuning can be a time-consuming and interaction-costing procedure in the field of LA. Especially, it is a fatal limitation for LA-based applications where the interactions with environments are expensive. In this paper, we propose a parameter-free learning automaton scheme to avoid parameter tuning by a Bayesian inference method. In contrast to existing schemes where the parameters should be carefully tuned according to the environment, the performance of this scheme is not sensitive to external environments because a set of parameters can be consistently applied to various environments, which dramatically reduce the difficulty of applying a learning automaton to an unknown stochastic environment. A rigorous proof of $\epsilon$-optimality for the proposed scheme is provided and numeric experiments are carried out on benchmark environments to verify its effectiveness. The results show that, without any parameter tuning cost, the proposed parameter-free learning automaton (PFLA) can achieve a competitive performance compared with other well-tuned schemes and outperform untuned schemes on consistency of performance.
Tasks Bayesian Inference
Published 2017-11-28
URL http://arxiv.org/abs/1711.10111v1
PDF http://arxiv.org/pdf/1711.10111v1.pdf
PWC https://paperswithcode.com/paper/a-parameter-free-learning-automaton-scheme
Repo
Framework

Rapid Probabilistic Interest Learning from Domain-Specific Pairwise Image Comparisons

Title Rapid Probabilistic Interest Learning from Domain-Specific Pairwise Image Comparisons
Authors Michael Burke
Abstract A great deal of work aims to discover general purpose models of image interest or memorability for visual search and information retrieval. This paper argues that image interest is often domain and user specific, and that mechanisms for learning about this domain-specific image interest as quickly as possible, while limiting the amount of data-labelling required, are often more useful to end-users. Specifically, this paper is concerned with the small to medium-sized data regime regularly faced by practising data scientists, who are often required to build turnkey models for end-users with domain-specific challenges. This work uses pairwise image comparisons to reduce the labelling burden on these users, and shows that Gaussian process smoothing in image feature space can be used to build probabilistic models of image interest extremely quickly for a wide range of problems, and performs similarly to recent deep learning approaches trained using pairwise ranking losses. The Gaussian process model used in this work interpolates image interest inferred using a Bayesian ranking approach over image features extracted using a pre-trained convolutional neural network. This probabilistic approach produces image interests paired with uncertainties that can be used to identify images for which additional labelling is required and measure inference convergence. Results obtained on five distinct datasets reinforce recent findings that pre-trained convolutional neural networks can be used to extract useful representations applicable across multiple domains, and highlight the fact that domain-specific image interest does not always correlate with concepts like image memorability.
Tasks Information Retrieval
Published 2017-06-19
URL http://arxiv.org/abs/1706.05850v2
PDF http://arxiv.org/pdf/1706.05850v2.pdf
PWC https://paperswithcode.com/paper/user-driven-mobile-robot-storyboarding
Repo
Framework

Neural Graph Machines: Learning Neural Networks Using Graphs

Title Neural Graph Machines: Learning Neural Networks Using Graphs
Authors Thang D. Bui, Sujith Ravi, Vivek Ramavajjala
Abstract Label propagation is a powerful and flexible semi-supervised learning technique on graphs. Neural networks, on the other hand, have proven track records in many supervised learning tasks. In this work, we propose a training framework with a graph-regularised objective, namely “Neural Graph Machines”, that can combine the power of neural networks and label propagation. This work generalises previous literature on graph-augmented training of neural networks, enabling it to be applied to multiple neural architectures (Feed-forward NNs, CNNs and LSTM RNNs) and a wide range of graphs. The new objective allows the neural networks to harness both labeled and unlabeled data by: (a) allowing the network to train using labeled data as in the supervised setting, (b) biasing the network to learn similar hidden representations for neighboring nodes on a graph, in the same vein as label propagation. Such architectures with the proposed objective can be trained efficiently using stochastic gradient descent and scaled to large graphs, with a runtime that is linear in the number of edges. The proposed joint training approach convincingly outperforms many existing methods on a wide range of tasks (multi-label classification on social graphs, news categorization, document classification and semantic intent classification), with multiple forms of graph inputs (including graphs with and without node-level features) and using different types of neural networks.
Tasks Document Classification, Intent Classification, Multi-Label Classification
Published 2017-03-14
URL http://arxiv.org/abs/1703.04818v1
PDF http://arxiv.org/pdf/1703.04818v1.pdf
PWC https://paperswithcode.com/paper/neural-graph-machines-learning-neural
Repo
Framework

Lower Bound On the Computational Complexity of Discounted Markov Decision Problems

Title Lower Bound On the Computational Complexity of Discounted Markov Decision Problems
Authors Yichen Chen, Mengdi Wang
Abstract We study the computational complexity of the infinite-horizon discounted-reward Markov Decision Problem (MDP) with a finite state space $\mathcal{S}$ and a finite action space $\mathcal{A}$. We show that any randomized algorithm needs a running time at least $\Omega(\mathcal{S}^2\mathcal{A})$ to compute an $\epsilon$-optimal policy with high probability. We consider two variants of the MDP where the input is given in specific data structures, including arrays of cumulative probabilities and binary trees of transition probabilities. For these cases, we show that the complexity lower bound reduces to $\Omega\left( \frac{\mathcal{S} \mathcal{A}}{\epsilon} \right)$. These results reveal a surprising observation that the computational complexity of the MDP depends on the data structure of input.
Tasks
Published 2017-05-20
URL http://arxiv.org/abs/1705.07312v1
PDF http://arxiv.org/pdf/1705.07312v1.pdf
PWC https://paperswithcode.com/paper/lower-bound-on-the-computational-complexity
Repo
Framework

Path-Based Attention Neural Model for Fine-Grained Entity Typing

Title Path-Based Attention Neural Model for Fine-Grained Entity Typing
Authors Denghui Zhang, Pengshan Cai, Yantao Jia, Manling Li, Yuanzhuo Wang, Xueqi Cheng
Abstract Fine-grained entity typing aims to assign entity mentions in the free text with types arranged in a hierarchical structure. Traditional distant supervision based methods employ a structured data source as a weak supervision and do not need hand-labeled data, but they neglect the label noise in the automatically labeled training corpus. Although recent studies use many features to prune wrong data ahead of training, they suffer from error propagation and bring much complexity. In this paper, we propose an end-to-end typing model, called the path-based attention neural model (PAN), to learn a noise- robust performance by leveraging the hierarchical structure of types. Experiments demonstrate its effectiveness.
Tasks Entity Typing
Published 2017-10-29
URL http://arxiv.org/abs/1710.10585v2
PDF http://arxiv.org/pdf/1710.10585v2.pdf
PWC https://paperswithcode.com/paper/path-based-attention-neural-model-for-fine
Repo
Framework

Exploiting Restricted Boltzmann Machines and Deep Belief Networks in Compressed Sensing

Title Exploiting Restricted Boltzmann Machines and Deep Belief Networks in Compressed Sensing
Authors Luisa F. Polania, Kenneth E. Barner
Abstract This paper proposes a CS scheme that exploits the representational power of restricted Boltzmann machines and deep learning architectures to model the prior distribution of the sparsity pattern of signals belonging to the same class. The determined probability distribution is then used in a maximum a posteriori (MAP) approach for the reconstruction. The parameters of the prior distribution are learned from training data. The motivation behind this approach is to model the higher-order statistical dependencies between the coefficients of the sparse representation, with the final goal of improving the reconstruction. The performance of the proposed method is validated on the Berkeley Segmentation Dataset and the MNIST Database of handwritten digits.
Tasks
Published 2017-05-30
URL http://arxiv.org/abs/1705.10500v1
PDF http://arxiv.org/pdf/1705.10500v1.pdf
PWC https://paperswithcode.com/paper/exploiting-restricted-boltzmann-machines-and
Repo
Framework

Systèmes du LIA à DEFT’13

Title Systèmes du LIA à DEFT’13
Authors Xavier Bost, Ilaria Brunetti, Luis Adrián Cabrera-Diego, Jean-Valère Cossu, Andréa Linhares, Mohamed Morchid, Juan-Manuel Torres-Moreno, Marc El-Bèze, Richard Dufour
Abstract The 2013 D'efi de Fouille de Textes (DEFT) campaign is interested in two types of language analysis tasks, the document classification and the information extraction in the specialized domain of cuisine recipes. We present the systems that the LIA has used in DEFT 2013. Our systems show interesting results, even though the complexity of the proposed tasks.
Tasks Document Classification
Published 2017-02-21
URL http://arxiv.org/abs/1702.06478v1
PDF http://arxiv.org/pdf/1702.06478v1.pdf
PWC https://paperswithcode.com/paper/systemes-du-lia-a-deft13
Repo
Framework

Semantic classifier approach to document classification

Title Semantic classifier approach to document classification
Authors Piotr Borkowski, Krzysztof Ciesielski, Mieczysław A. Kłopotek
Abstract In this paper we propose a new document classification method, bridging discrepancies (so-called semantic gap) between the training set and the application sets of textual data. We demonstrate its superiority over classical text classification approaches, including traditional classifier ensembles. The method consists in combining a document categorization technique with a single classifier or a classifier ensemble (SEMCOM algorithm - Committee with Semantic Categorizer).
Tasks Document Classification, Text Classification
Published 2017-01-16
URL http://arxiv.org/abs/1701.04292v1
PDF http://arxiv.org/pdf/1701.04292v1.pdf
PWC https://paperswithcode.com/paper/semantic-classifier-approach-to-document
Repo
Framework

Semantic Visual Localization

Title Semantic Visual Localization
Authors Johannes L. Schönberger, Marc Pollefeys, Andreas Geiger, Torsten Sattler
Abstract Robust visual localization under a wide range of viewing conditions is a fundamental problem in computer vision. Handling the difficult cases of this problem is not only very challenging but also of high practical relevance, e.g., in the context of life-long localization for augmented reality or autonomous robots. In this paper, we propose a novel approach based on a joint 3D geometric and semantic understanding of the world, enabling it to succeed under conditions where previous approaches failed. Our method leverages a novel generative model for descriptor learning, trained on semantic scene completion as an auxiliary task. The resulting 3D descriptors are robust to missing observations by encoding high-level 3D geometric and semantic information. Experiments on several challenging large-scale localization datasets demonstrate reliable localization under extreme viewpoint, illumination, and geometry changes.
Tasks Visual Localization
Published 2017-12-15
URL http://arxiv.org/abs/1712.05773v2
PDF http://arxiv.org/pdf/1712.05773v2.pdf
PWC https://paperswithcode.com/paper/semantic-visual-localization
Repo
Framework

Deep learning with spatiotemporal consistency for nerve segmentation in ultrasound images

Title Deep learning with spatiotemporal consistency for nerve segmentation in ultrasound images
Authors Adel Hafiane, Pierre Vieyres, Alain Delbos
Abstract Ultrasound-Guided Regional Anesthesia (UGRA) has been gaining importance in the last few years, offering numerous advantages over alternative methods of nerve localization (neurostimulation or paraesthesia). However, nerve detection is one of the most tasks that anaesthetists can encounter in the UGRA procedure. Computer aided system that can detect automatically region of nerve, would help practitioner to concentrate more in anaesthetic delivery. In this paper we propose a new method based on deep learning combined with spatiotemporal information to robustly segment the nerve region. The proposed method is based on two phases, localisation and segmentation. The first phase, consists in using convolutional neural network combined with spatial and temporal consistency to detect the nerve zone. The second phase utilises active contour model to delineate the region of interest. Obtained results show the validity of the proposed approach and its robustness.
Tasks
Published 2017-06-19
URL http://arxiv.org/abs/1706.05870v1
PDF http://arxiv.org/pdf/1706.05870v1.pdf
PWC https://paperswithcode.com/paper/deep-learning-with-spatiotemporal-consistency
Repo
Framework

The Intentional Unintentional Agent: Learning to Solve Many Continuous Control Tasks Simultaneously

Title The Intentional Unintentional Agent: Learning to Solve Many Continuous Control Tasks Simultaneously
Authors Serkan Cabi, Sergio Gómez Colmenarejo, Matthew W. Hoffman, Misha Denil, Ziyu Wang, Nando de Freitas
Abstract This paper introduces the Intentional Unintentional (IU) agent. This agent endows the deep deterministic policy gradients (DDPG) agent for continuous control with the ability to solve several tasks simultaneously. Learning to solve many tasks simultaneously has been a long-standing, core goal of artificial intelligence, inspired by infant development and motivated by the desire to build flexible robot manipulators capable of many diverse behaviours. We show that the IU agent not only learns to solve many tasks simultaneously but it also learns faster than agents that target a single task at-a-time. In some cases, where the single task DDPG method completely fails, the IU agent successfully solves the task. To demonstrate this, we build a playroom environment using the MuJoCo physics engine, and introduce a grounded formal language to automatically generate tasks.
Tasks Continuous Control
Published 2017-07-11
URL http://arxiv.org/abs/1707.03300v1
PDF http://arxiv.org/pdf/1707.03300v1.pdf
PWC https://paperswithcode.com/paper/the-intentional-unintentional-agent-learning
Repo
Framework

Implications of Decentralized Q-learning Resource Allocation in Wireless Networks

Title Implications of Decentralized Q-learning Resource Allocation in Wireless Networks
Authors Francesc Wilhelmi, Boris Bellalta, Cristina Cano, Anders Jonsson
Abstract Reinforcement Learning is gaining attention by the wireless networking community due to its potential to learn good-performing configurations only from the observed results. In this work we propose a stateless variation of Q-learning, which we apply to exploit spatial reuse in a wireless network. In particular, we allow networks to modify both their transmission power and the channel used solely based on the experienced throughput. We concentrate in a completely decentralized scenario in which no information about neighbouring nodes is available to the learners. Our results show that although the algorithm is able to find the best-performing actions to enhance aggregate throughput, there is high variability in the throughput experienced by the individual networks. We identify the cause of this variability as the adversarial setting of our setup, in which the most played actions provide intermittent good/poor performance depending on the neighbouring decisions. We also evaluate the effect of the intrinsic learning parameters of the algorithm on this variability.
Tasks Q-Learning
Published 2017-05-30
URL http://arxiv.org/abs/1705.10508v2
PDF http://arxiv.org/pdf/1705.10508v2.pdf
PWC https://paperswithcode.com/paper/implications-of-decentralized-q-learning
Repo
Framework

Unsupervised Latent Behavior Manifold Learning from Acoustic Features: audio2behavior

Title Unsupervised Latent Behavior Manifold Learning from Acoustic Features: audio2behavior
Authors Haoqi Li, Brian Baucom, Panayiotis Georgiou
Abstract Behavioral annotation using signal processing and machine learning is highly dependent on training data and manual annotations of behavioral labels. Previous studies have shown that speech information encodes significant behavioral information and be used in a variety of automated behavior recognition tasks. However, extracting behavior information from speech is still a difficult task due to the sparseness of training data coupled with the complex, high-dimensionality of speech, and the complex and multiple information streams it encodes. In this work we exploit the slow varying properties of human behavior. We hypothesize that nearby segments of speech share the same behavioral context and hence share a similar underlying representation in a latent space. Specifically, we propose a Deep Neural Network (DNN) model to connect behavioral context and derive the behavioral manifold in an unsupervised manner. We evaluate the proposed manifold in the couples therapy domain and also provide examples from publicly available data (e.g. stand-up comedy). We further investigate training within the couples’ therapy domain and from movie data. The results are extremely encouraging and promise improved behavioral quantification in an unsupervised manner and warrants further investigation in a range of applications.
Tasks
Published 2017-01-12
URL http://arxiv.org/abs/1701.03198v1
PDF http://arxiv.org/pdf/1701.03198v1.pdf
PWC https://paperswithcode.com/paper/unsupervised-latent-behavior-manifold
Repo
Framework

Attention-based Mixture Density Recurrent Networks for History-based Recommendation

Title Attention-based Mixture Density Recurrent Networks for History-based Recommendation
Authors Tian Wang, Kyunghyun Cho
Abstract The goal of personalized history-based recommendation is to automatically output a distribution over all the items given a sequence of previous purchases of a user. In this work, we present a novel approach that uses a recurrent network for summarizing the history of purchases, continuous vectors representing items for scalability, and a novel attention-based recurrent mixture density network, which outputs each component in a mixture sequentially, for modelling a multi-modal conditional distribution. We evaluate the proposed approach on two publicly available datasets, MovieLens-20M and RecSys15. The experiments show that the proposed approach, which explicitly models the multi-modal nature of the predictive distribution, is able to improve the performance over various baselines in terms of precision, recall and nDCG.
Tasks
Published 2017-09-22
URL http://arxiv.org/abs/1709.07545v1
PDF http://arxiv.org/pdf/1709.07545v1.pdf
PWC https://paperswithcode.com/paper/attention-based-mixture-density-recurrent
Repo
Framework

Generating OWA weights using truncated distributions

Title Generating OWA weights using truncated distributions
Authors Maxime Lenormand
Abstract Ordered weighted averaging (OWA) operators have been widely used in decision making these past few years. An important issue facing the OWA operators’ users is the determination of the OWA weights. This paper introduces an OWA determination method based on truncated distributions that enables intuitive generation of OWA weights according to a certain level of risk and trade-off. These two dimensions are represented by the two first moments of the truncated distribution. We illustrate our approach with the well-know normal distribution and the definition of a continuous parabolic decision-strategy space. We finally study the impact of the number of criteria on the results.
Tasks Decision Making
Published 2017-09-13
URL http://arxiv.org/abs/1709.04328v2
PDF http://arxiv.org/pdf/1709.04328v2.pdf
PWC https://paperswithcode.com/paper/generating-owa-weights-using-truncated
Repo
Framework
comments powered by Disqus