January 26, 2020

3072 words 15 mins read

Paper Group ANR 1363

Paper Group ANR 1363

Towards hardware acceleration for parton densities estimation. Fast Adaptation with Meta-Reinforcement Learning for Trust Modelling in Human-Robot Interaction. Robust Deep Sensing Through Transfer Learning in Cognitive Radio. Error Analysis of Elitist Random Search Heuristics. ZK-GanDef: A GAN based Zero Knowledge Adversarial Training Defense for N …

Towards hardware acceleration for parton densities estimation

Title Towards hardware acceleration for parton densities estimation
Authors Stefano Carrazza, Juan Cruz-Martinez, Jesús Urtasun-Elizari, Emilio Villa
Abstract In this proceedings we describe the computational challenges associated to the determination of parton distribution functions (PDFs). We compare the performance of the convolution of the parton distributions with matrix elements using different hardware instructions. We quantify and identify the most promising data-model configurations to increase PDF fitting performance in adapting the current code frameworks to hardware accelerators such as graphics processing units.
Tasks
Published 2019-09-23
URL https://arxiv.org/abs/1909.10547v1
PDF https://arxiv.org/pdf/1909.10547v1.pdf
PWC https://paperswithcode.com/paper/towards-hardware-acceleration-for-parton
Repo
Framework

Fast Adaptation with Meta-Reinforcement Learning for Trust Modelling in Human-Robot Interaction

Title Fast Adaptation with Meta-Reinforcement Learning for Trust Modelling in Human-Robot Interaction
Authors Yuan Gao, Elena Sibirtseva, Ginevra Castellano, Danica Kragic
Abstract In socially assistive robotics, an important research area is the development of adaptation techniques and their effect on human-robot interaction. We present a meta-learning based policy gradient method for addressing the problem of adaptation in human-robot interaction and also investigate its role as a mechanism for trust modelling. By building an escape room scenario in mixed reality with a robot, we test our hypothesis that bi-directional trust can be influenced by different adaptation algorithms. We found that our proposed model increased the perceived trustworthiness of the robot and influenced the dynamics of gaining human’s trust. Additionally, participants evaluated that the robot perceived them as more trustworthy during the interactions with the meta-learning based adaptation compared to the previously studied statistical adaptation model.
Tasks Meta-Learning
Published 2019-08-12
URL https://arxiv.org/abs/1908.04087v1
PDF https://arxiv.org/pdf/1908.04087v1.pdf
PWC https://paperswithcode.com/paper/fast-adaptation-with-meta-reinforcement
Repo
Framework

Robust Deep Sensing Through Transfer Learning in Cognitive Radio

Title Robust Deep Sensing Through Transfer Learning in Cognitive Radio
Authors Qihang Peng, Andrew Gilman, Nuno Vasconcelos, Pamela C. Cosman, Laurence B. Milstein
Abstract We propose a robust spectrum sensing framework based on deep learning. The received signals at the secondary user’s receiver are filtered, sampled and then directly fed into a convolutional neural network. Although this deep sensing is effective when operating in the same scenario as the collected training data, the sensing performance is degraded when it is applied in a different scenario with different wireless signals and propagation. We incorporate transfer learning into the framework to improve the robustness. Results validate the effectiveness as well as the robustness of the proposed deep spectrum sensing framework.
Tasks Transfer Learning
Published 2019-08-01
URL https://arxiv.org/abs/1908.00658v1
PDF https://arxiv.org/pdf/1908.00658v1.pdf
PWC https://paperswithcode.com/paper/robust-deep-sensing-through-transfer-learning
Repo
Framework

Error Analysis of Elitist Random Search Heuristics

Title Error Analysis of Elitist Random Search Heuristics
Authors Yu Chen, Cong Wang, Jun He, Chengwang Xie
Abstract When random search heuristics (RSH) cannot locate precise globally optimal solutions with satisfactory performances, hitting time/running time analysis method is not flexible enough to accommodate the requirement of theoretical performance study. Thus, an alternative routine is needed to bridge the gap between theoretical analysis and algorithm implementation. Inspired by this idea, this paper is dedicated to perform an error analysis for elitist RSHs. By diagonalizing the transition matrix of a Markov chain model, the $t$-step transition matrix can be computed for any iteration budget $t$. Then, A general framework for estimation of expected approximation errors can be proposed. Case studies indicate that the error analysis works well for both uni- and multi-modal benchmark problems. It leads to precise estimations of approximation error instead of asymptotic results on fitness values, which demonstrates its competitiveness to fixed budget analysis.
Tasks
Published 2019-09-03
URL https://arxiv.org/abs/1909.00894v3
PDF https://arxiv.org/pdf/1909.00894v3.pdf
PWC https://paperswithcode.com/paper/estimating-approximation-errors-of-elitist
Repo
Framework

ZK-GanDef: A GAN based Zero Knowledge Adversarial Training Defense for Neural Networks

Title ZK-GanDef: A GAN based Zero Knowledge Adversarial Training Defense for Neural Networks
Authors Guanxiong Liu, Issa Khalil, Abdallah Khreishah
Abstract Neural Network classifiers have been used successfully in a wide range of applications. However, their underlying assumption of attack free environment has been defied by adversarial examples. Researchers tried to develop defenses; however, existing approaches are still far from providing effective solutions to this evolving problem. In this paper, we design a generative adversarial net (GAN) based zero knowledge adversarial training defense, dubbed ZK-GanDef, which does not consume adversarial examples during training. Therefore, ZK-GanDef is not only efficient in training but also adaptive to new adversarial examples. This advantage comes at the cost of small degradation in test accuracy compared to full knowledge approaches. Our experiments show that ZK-GanDef enhances test accuracy on adversarial examples by up-to 49.17% compared to zero knowledge approaches. More importantly, its test accuracy is close to that of the state-of-the-art full knowledge approaches (maximum degradation of 8.46%), while taking much less training time.
Tasks
Published 2019-04-17
URL http://arxiv.org/abs/1904.08516v1
PDF http://arxiv.org/pdf/1904.08516v1.pdf
PWC https://paperswithcode.com/paper/zk-gandef-a-gan-based-zero-knowledge
Repo
Framework

Planning in Hierarchical Reinforcement Learning: Guarantees for Using Local Policies

Title Planning in Hierarchical Reinforcement Learning: Guarantees for Using Local Policies
Authors Tom Zahavy, Avinatan Hasidim, Haim Kaplan, Yishay Mansour
Abstract We consider a settings of hierarchical reinforcement learning, in which the reward is a sum of components. For each component we are given a policy that maximizes it and our goal is to assemble a policy from the individual policies that maximizes the sum of the components. We provide theoretical guarantees for assembling such policies in deterministic MDPs with collectible rewards. Our approach builds on formulating this problem as a traveling salesman problem with discounted reward. We focus on local solutions, i.e., policies that only use information from the current state; thus, they are easy to implement and do not require substantial computational resources. We propose three local stochastic policies and prove that they guarantee better performance than any deterministic local policy in the worst case; experimental results suggest that they also perform better on average.
Tasks Hierarchical Reinforcement Learning
Published 2019-02-26
URL https://arxiv.org/abs/1902.10140v2
PDF https://arxiv.org/pdf/1902.10140v2.pdf
PWC https://paperswithcode.com/paper/planning-in-hierarchical-reinforcement
Repo
Framework

BIGPATENT: A Large-Scale Dataset for Abstractive and Coherent Summarization

Title BIGPATENT: A Large-Scale Dataset for Abstractive and Coherent Summarization
Authors Eva Sharma, Chen Li, Lu Wang
Abstract Most existing text summarization datasets are compiled from the news domain, where summaries have a flattened discourse structure. In such datasets, summary-worthy content often appears in the beginning of input articles. Moreover, large segments from input articles are present verbatim in their respective summaries. These issues impede the learning and evaluation of systems that can understand an article’s global content structure as well as produce abstractive summaries with high compression ratio. In this work, we present a novel dataset, BIGPATENT, consisting of 1.3 million records of U.S. patent documents along with human written abstractive summaries. Compared to existing summarization datasets, BIGPATENT has the following properties: i) summaries contain a richer discourse structure with more recurring entities, ii) salient content is evenly distributed in the input, and iii) lesser and shorter extractive fragments are present in the summaries. Finally, we train and evaluate baselines and popular learning models on BIGPATENT to shed light on new challenges and motivate future directions for summarization research.
Tasks Text Summarization
Published 2019-06-10
URL https://arxiv.org/abs/1906.03741v1
PDF https://arxiv.org/pdf/1906.03741v1.pdf
PWC https://paperswithcode.com/paper/bigpatent-a-large-scale-dataset-for
Repo
Framework

Scalable Gaussian Process Classification with Additive Noise for Various Likelihoods

Title Scalable Gaussian Process Classification with Additive Noise for Various Likelihoods
Authors Haitao Liu, Yew-Soon Ong, Ziwei Yu, Jianfei Cai, Xiaobo Shen
Abstract Gaussian process classification (GPC) provides a flexible and powerful statistical framework describing joint distributions over function space. Conventional GPCs however suffer from (i) poor scalability for big data due to the full kernel matrix, and (ii) intractable inference due to the non-Gaussian likelihoods. Hence, various scalable GPCs have been proposed through (i) the sparse approximation built upon a small inducing set to reduce the time complexity; and (ii) the approximate inference to derive analytical evidence lower bound (ELBO). However, these scalable GPCs equipped with analytical ELBO are limited to specific likelihoods or additional assumptions. In this work, we present a unifying framework which accommodates scalable GPCs using various likelihoods. Analogous to GP regression (GPR), we introduce additive noises to augment the probability space for (i) the GPCs with step, (multinomial) probit and logit likelihoods via the internal variables; and particularly, (ii) the GPC using softmax likelihood via the noise variables themselves. This leads to unified scalable GPCs with analytical ELBO by using variational inference. Empirically, our GPCs showcase better results than state-of-the-art scalable GPCs for extensive binary/multi-class classification tasks with up to two million data points.
Tasks
Published 2019-09-14
URL https://arxiv.org/abs/1909.06541v1
PDF https://arxiv.org/pdf/1909.06541v1.pdf
PWC https://paperswithcode.com/paper/scalable-gaussian-process-classification-with
Repo
Framework

Multi-modal Active Learning From Human Data: A Deep Reinforcement Learning Approach

Title Multi-modal Active Learning From Human Data: A Deep Reinforcement Learning Approach
Authors Ognjen Rudovic, Meiru Zhang, Bjorn Schuller, Rosalind W. Picard
Abstract Human behavior expression and experience are inherently multi-modal, and characterized by vast individual and contextual heterogeneity. To achieve meaningful human-computer and human-robot interactions, multi-modal models of the users states (e.g., engagement) are therefore needed. Most of the existing works that try to build classifiers for the users states assume that the data to train the models are fully labeled. Nevertheless, data labeling is costly and tedious, and also prone to subjective interpretations by the human coders. This is even more pronounced when the data are multi-modal (e.g., some users are more expressive with their facial expressions, some with their voice). Thus, building models that can accurately estimate the users states during an interaction is challenging. To tackle this, we propose a novel multi-modal active learning (AL) approach that uses the notion of deep reinforcement learning (RL) to find an optimal policy for active selection of the users data, needed to train the target (modality-specific) models. We investigate different strategies for multi-modal data fusion, and show that the proposed model-level fusion coupled with RL outperforms the feature-level and modality-specific models, and the naive AL strategies such as random sampling, and the standard heuristics such as uncertainty sampling. We show the benefits of this approach on the task of engagement estimation from real-world child-robot interactions during an autism therapy. Importantly, we show that the proposed multi-modal AL approach can be used to efficiently personalize the engagement classifiers to the target user using a small amount of actively selected users data.
Tasks Active Learning
Published 2019-06-07
URL https://arxiv.org/abs/1906.03098v1
PDF https://arxiv.org/pdf/1906.03098v1.pdf
PWC https://paperswithcode.com/paper/multi-modal-active-learning-from-human-data-a
Repo
Framework

Unsupervised Segmentation of Fire and Smoke from Infra-Red Videos

Title Unsupervised Segmentation of Fire and Smoke from Infra-Red Videos
Authors Meenu Ajith, Manel Martínez-Ramón
Abstract This paper proposes a vision-based fire and smoke segmentation system which use spatial, temporal and motion information to extract the desired regions from the video frames. The fusion of information is done using multiple features such as optical flow, divergence and intensity values. These features extracted from the images are used to segment the pixels into different classes in an unsupervised way. A comparative analysis is done by using multiple clustering algorithms for segmentation. Here the Markov Random Field performs more accurately than other segmentation algorithms since it characterizes the spatial interactions of pixels using a finite number of parameters. It builds a probabilistic image model that selects the most likely labeling using the maximum a posteriori (MAP) estimation. This unsupervised approach is tested on various images and achieves a frame-wise fire detection rate of 95.39%. Hence this method can be used for early detection of fire in real-time and it can be incorporated into an indoor or outdoor surveillance system.
Tasks Optical Flow Estimation
Published 2019-09-18
URL https://arxiv.org/abs/1909.12937v1
PDF https://arxiv.org/pdf/1909.12937v1.pdf
PWC https://paperswithcode.com/paper/unsupervised-segmentation-of-fire-and-smoke
Repo
Framework

Fuzzy Rule Interpolation Toolbox for the GNU Open-Source OCTAVE

Title Fuzzy Rule Interpolation Toolbox for the GNU Open-Source OCTAVE
Authors Maen Alzubi, Mohammad Almseidin, Mohd Aaqib Lone, Szilveszter Kovacs
Abstract In most fuzzy control applications (applying classical fuzzy reasoning), the reasoning method requires a complete fuzzy rule-base, i.e all the possible observations must be covered by the antecedents of the fuzzy rules, which is not always available. Fuzzy control systems based on the Fuzzy Rule Interpolation (FRI) concept play a major role in different platforms, in case if only a sparse fuzzy rule-base is available. This cases the fuzzy model contains only the most relevant rules, without covering all the antecedent universes. The first FRI toolbox being able to handle different FRI methods was developed by Johanyak et. al. in 2006 for the MATLAB environment. The goal of this paper is to introduce some details of the adaptation of the FRI toolbox to support the GNU/OCTAVE programming language. The OCTAVE Fuzzy Rule Interpolation (OCTFRI) Toolbox is an open-source toolbox for OCTAVE programming language, providing a large functionally compatible subset of the MATLAB FRI toolbox as well as many extensions. The OCTFRI Toolbox includes functions that enable the user to evaluate Fuzzy Inference Systems (FISs) from the command line and from OCTAVE scripts, read/write FISs and OBS to/from files, and produce a graphical visualisation of both the membership functions and the FIS outputs. Future work will focus on implementing advanced fuzzy inference techniques and GUI tools.
Tasks
Published 2019-12-10
URL https://arxiv.org/abs/1912.04999v1
PDF https://arxiv.org/pdf/1912.04999v1.pdf
PWC https://paperswithcode.com/paper/fuzzy-rule-interpolation-toolbox-for-the-gnu
Repo
Framework

An Internal Learning Approach to Video Inpainting

Title An Internal Learning Approach to Video Inpainting
Authors Haotian Zhang, Long Mai, Ning Xu, Zhaowen Wang, John Collomosse, Hailin Jin
Abstract We propose a novel video inpainting algorithm that simultaneously hallucinates missing appearance and motion (optical flow) information, building upon the recent ‘Deep Image Prior’ (DIP) that exploits convolutional network architectures to enforce plausible texture in static images. In extending DIP to video we make two important contributions. First, we show that coherent video inpainting is possible without a priori training. We take a generative approach to inpainting based on internal (within-video) learning without reliance upon an external corpus of visual data to train a one-size-fits-all model for the large space of general videos. Second, we show that such a framework can jointly generate both appearance and flow, whilst exploiting these complementary modalities to ensure mutual consistency. We show that leveraging appearance statistics specific to each video achieves visually plausible results whilst handling the challenging problem of long-term consistency.
Tasks Optical Flow Estimation, Video Inpainting
Published 2019-09-17
URL https://arxiv.org/abs/1909.07957v1
PDF https://arxiv.org/pdf/1909.07957v1.pdf
PWC https://paperswithcode.com/paper/an-internal-learning-approach-to-video
Repo
Framework

Visuomotor Understanding for Representation Learning of Driving Scenes

Title Visuomotor Understanding for Representation Learning of Driving Scenes
Authors Seokju Lee, Junsik Kim, Tae-Hyun Oh, Yongseop Jeong, Donggeun Yoo, Stephen Lin, In So Kweon
Abstract Dashboard cameras capture a tremendous amount of driving scene video each day. These videos are purposefully coupled with vehicle sensing data, such as from the speedometer and inertial sensors, providing an additional sensing modality for free. In this work, we leverage the large-scale unlabeled yet naturally paired data for visual representation learning in the driving scenario. A representation is learned in an end-to-end self-supervised framework for predicting dense optical flow from a single frame with paired sensing data. We postulate that success on this task requires the network to learn semantic and geometric knowledge in the ego-centric view. For example, forecasting a future view to be seen from a moving vehicle requires an understanding of scene depth, scale, and movement of objects. We demonstrate that our learned representation can benefit other tasks that require detailed scene understanding and outperforms competing unsupervised representations on semantic segmentation.
Tasks Optical Flow Estimation, Representation Learning, Scene Understanding, Semantic Segmentation
Published 2019-09-16
URL https://arxiv.org/abs/1909.06979v1
PDF https://arxiv.org/pdf/1909.06979v1.pdf
PWC https://paperswithcode.com/paper/visuomotor-understanding-for-representation
Repo
Framework

Feature-Fused Context-Encoding Network for Neuroanatomy Segmentation

Title Feature-Fused Context-Encoding Network for Neuroanatomy Segmentation
Authors Yuemeng Li, Hangfan Liu, Hongming Li, Yong Fan
Abstract Automatic segmentation of fine-grained brain structures remains a challenging task. Current segmentation methods mainly utilize 2D and 3D deep neural networks. The 2D networks take image slices as input to produce coarse segmentation in less processing time, whereas the 3D networks take the whole image volumes to generated fine-detailed segmentation with more computational burden. In order to obtain accurate fine-grained segmentation efficiently, in this paper, we propose an end-to-end Feature-Fused Context-Encoding Network for brain structure segmentation from MR (magnetic resonance) images. Our model is implemented based on a 2D convolutional backbone, which integrates a 2D encoding module to acquire planar image features and a spatial encoding module to extract spatial context information. A global context encoding module is further introduced to capture global context semantics from the fused 2D encoding and spatial features. The proposed network aims to fully leverage the global anatomical prior knowledge learned from context semantics, which is represented by a structure-aware attention factor to recalibrate the outputs of the network. In this way, the network is guaranteed to be aware of the class-dependent feature maps to facilitate the segmentation. We evaluate our model on 2012 Brain Multi-Atlas Labelling Challenge dataset for 134 fine-grained structure segmentation. Besides, we validate our network on 27 coarse structure segmentation tasks. Experimental results have demonstrated that our model can achieve improved performance compared with the state-of-the-art approaches.
Tasks
Published 2019-05-07
URL https://arxiv.org/abs/1905.02686v1
PDF https://arxiv.org/pdf/1905.02686v1.pdf
PWC https://paperswithcode.com/paper/feature-fused-context-encoding-network-for
Repo
Framework

Embedding Human Heuristics in Machine-Learning-Enabled Probe Microscopy

Title Embedding Human Heuristics in Machine-Learning-Enabled Probe Microscopy
Authors O. Gordon, F. Junqueira, P. Moriarty
Abstract Scanning probe microscopists generally do not rely on complete images to assess the quality of data acquired during a scan. Instead, assessments of the state of the tip apex, which not only determines the resolution in any scanning probe technique but can also generate a wide array of frustrating artefacts, are carried out in real time on the basis of a few lines of an image (and, typically, their associated line profiles.) The very small number of machine learning approaches to probe microscopy published to date, however, involve classifications based on full images. Given that data acquisition is the most time-consuming task during routine tip conditioning, automated methods are thus currently extremely slow in comparison to the tried-and-trusted strategies and heuristics used routinely by probe microscopists. Here, we explore various strategies by which different STM image classes (arising from changes in the tip state) can be correctly identified from partial scans. By employing a secondary temporal network and a rolling window of a small group of individual scanlines, we find that tip assessment is possible with a small fraction of a complete image. We achieve this with little-to-no performance penalty – or, indeed, markedly improved performance in some cases – and introduce a protocol to detect the state of the tip apex in real time.
Tasks
Published 2019-07-31
URL https://arxiv.org/abs/1907.13401v1
PDF https://arxiv.org/pdf/1907.13401v1.pdf
PWC https://paperswithcode.com/paper/embedding-human-heuristics-in-machine
Repo
Framework
comments powered by Disqus