January 30, 2020

3186 words 15 mins read

Paper Group ANR 381

Paper Group ANR 381

An Improved Convergence Analysis of Stochastic Variance-Reduced Policy Gradient. FusionMapping: Learning Depth Prediction with Monocular Images and 2D Laser Scans. Identify the cells’ nuclei based on the deep learning neural network. The Query Translation Landscape: a Survey. CopyCAT: Taking Control of Neural Policies with Constant Attacks. Differe …

An Improved Convergence Analysis of Stochastic Variance-Reduced Policy Gradient

Title An Improved Convergence Analysis of Stochastic Variance-Reduced Policy Gradient
Authors Pan Xu, Felicia Gao, Quanquan Gu
Abstract We revisit the stochastic variance-reduced policy gradient (SVRPG) method proposed by Papini et al. (2018) for reinforcement learning. We provide an improved convergence analysis of SVRPG and show that it can find an $\epsilon$-approximate stationary point of the performance function within $O(1/\epsilon^{5/3})$ trajectories. This sample complexity improves upon the best known result $O(1/\epsilon^2)$ by a factor of $O(1/\epsilon^{1/3})$. At the core of our analysis is (i) a tighter upper bound for the variance of importance sampling weights, where we prove that the variance can be controlled by the parameter distance between different policies; and (ii) a fine-grained analysis of the epoch length and batch size parameters such that we can significantly reduce the number of trajectories required in each iteration of SVRPG. We also empirically demonstrate the effectiveness of our theoretical claims of batch sizes on reinforcement learning benchmark tasks.
Tasks
Published 2019-05-29
URL https://arxiv.org/abs/1905.12615v1
PDF https://arxiv.org/pdf/1905.12615v1.pdf
PWC https://paperswithcode.com/paper/an-improved-convergence-analysis-of
Repo
Framework

FusionMapping: Learning Depth Prediction with Monocular Images and 2D Laser Scans

Title FusionMapping: Learning Depth Prediction with Monocular Images and 2D Laser Scans
Authors Peng Yin, Jianing Qian, Yibo Cao, David Held, Howie Choset
Abstract Acquiring accurate three-dimensional depth information conventionally requires expensive multibeam LiDAR devices. Recently, researchers have developed a less expensive option by predicting depth information from two-dimensional color imagery. However, there still exists a substantial gap in accuracy between depth information estimated from two-dimensional images and real LiDAR point-cloud. In this paper, we introduce a fusion-based depth prediction method, called FusionMapping. This is the first method that fuses colored imagery and two-dimensional laser scan to estimate depth in-formation. More specifically, we propose an autoencoder-based depth prediction network and a novel point-cloud refinement network for depth estimation. We analyze the performance of our FusionMapping approach on the KITTI LiDAR odometry dataset and an indoor mobile robot system. The results show that our introduced approach estimates depth with better accuracy when compared to existing methods.
Tasks Depth Estimation
Published 2019-11-29
URL https://arxiv.org/abs/1912.00096v1
PDF https://arxiv.org/pdf/1912.00096v1.pdf
PWC https://paperswithcode.com/paper/fusionmapping-learning-depth-prediction-with
Repo
Framework

Identify the cells’ nuclei based on the deep learning neural network

Title Identify the cells’ nuclei based on the deep learning neural network
Authors Tianyang Zhang, Rui Ma
Abstract Identify the cells’ nuclei is the important point for most medical analyses. To assist doctors finding the accurate cell’ nuclei location automatically is highly demanded in the clinical practice. Recently, fully convolutional neural network (FCNs) serve as the back-bone in many image segmentation, like liver and tumer segmentation in medical field, human body block in technical filed. The cells’ nuclei identification task is also kind of image segmentation. To achieve this, we prefer to use deep learning algorithms. we construct three general frameworks, one is Mask Region-based Convolutional Neural Network (Mask RCNN), which has the high performance in many image segmentations, one is U-net, which has the high generalization performance on small dataset and the other is DenseUNet, which is mixture network architecture with Dense Net and U-net. we compare the performance of these three frameworks. And we evaluated our method on the dataset of data science bowl 2018 challenge. For single model without any ensemble, they all have good performance.
Tasks Semantic Segmentation
Published 2019-11-22
URL https://arxiv.org/abs/1911.09830v1
PDF https://arxiv.org/pdf/1911.09830v1.pdf
PWC https://paperswithcode.com/paper/identify-the-cells-nuclei-based-on-the-deep
Repo
Framework

The Query Translation Landscape: a Survey

Title The Query Translation Landscape: a Survey
Authors Mohamed Nadjib Mami, Damien Graux, Harsh Thakkar, Simon Scerri, Sören Auer, Jens Lehmann
Abstract Whereas the availability of data has seen a manyfold increase in past years, its value can be only shown if the data variety is effectively tackled —one of the prominent Big Data challenges. The lack of data interoperability limits the potential of its collective use for novel applications. Achieving interoperability through the full transformation and integration of diverse data structures remains an ideal that is hard, if not impossible, to achieve. Instead, methods that can simultaneously interpret different types of data available in different data structures and formats have been explored. On the other hand, many query languages have been designed to enable users to interact with the data, from relational, to object-oriented, to hierarchical, to the multitude emerging NoSQL languages. Therefore, the interoperability issue could be solved not by enforcing physical data transformation, but by looking at techniques that are able to query heterogeneous sources using one uniform language. Both industry and research communities have been keen to develop such techniques, which require the translation of a chosen ‘universal’ query language to the various data model specific query languages that make the underlying data accessible. In this article, we survey more than forty query translation methods and tools for popular query languages, and classify them according to eight criteria. In particular, we study which query language is a most suitable candidate for that ‘universal’ query language. Further, the results enable us to discover the weakly addressed and unexplored translation paths, to discover gaps and to learn lessons that can benefit future research in the area.
Tasks
Published 2019-10-07
URL https://arxiv.org/abs/1910.03118v1
PDF https://arxiv.org/pdf/1910.03118v1.pdf
PWC https://paperswithcode.com/paper/the-query-translation-landscape-a-survey
Repo
Framework

CopyCAT: Taking Control of Neural Policies with Constant Attacks

Title CopyCAT: Taking Control of Neural Policies with Constant Attacks
Authors Léonard Hussenot, Matthieu Geist, Olivier Pietquin
Abstract We propose a new perspective on adversarial attacks against deep reinforcement learning agents. Our main contribution is CopyCAT, a targeted attack able to consistently lure an agent into following an outsider’s policy. It is pre-computed, therefore fast inferred, and could thus be usable in a real-time scenario. We show its effectiveness on Atari 2600 games in the novel read-only setting. In this setting, the adversary cannot directly modify the agent’s state – its representation of the environment – but can only attack the agent’s observation – its perception of the environment. Directly modifying the agent’s state would require a write-access to the agent’s inner workings and we argue that this assumption is too strong in realistic settings.
Tasks Atari Games
Published 2019-05-29
URL https://arxiv.org/abs/1905.12282v2
PDF https://arxiv.org/pdf/1905.12282v2.pdf
PWC https://paperswithcode.com/paper/targeted-attacks-on-deep-reinforcement
Repo
Framework

Differentially Private Summation with Multi-Message Shuffling

Title Differentially Private Summation with Multi-Message Shuffling
Authors Borja Balle, James Bell, Adria Gascon, Kobbi Nissim
Abstract In recent work, Cheu et al. (Eurocrypt 2019) proposed a protocol for $n$-party real summation in the shuffle model of differential privacy with $O_{\epsilon, \delta}(1)$ error and $\Theta(\epsilon\sqrt{n})$ one-bit messages per party. In contrast, every local model protocol for real summation must incur error $\Omega(1/\sqrt{n})$, and there exist protocols matching this lower bound which require just one bit of communication per party. Whether this gap in number of messages is necessary was left open by Cheu et al. In this note we show a protocol with $O(1/\epsilon)$ error and $O(\log(n/\delta))$ messages of size $O(\log(n))$ per party. This protocol is based on the work of Ishai et al.\ (FOCS 2006) showing how to implement distributed summation from secure shuffling, and the observation that this allows simulating the Laplace mechanism in the shuffle model.
Tasks
Published 2019-06-20
URL https://arxiv.org/abs/1906.09116v3
PDF https://arxiv.org/pdf/1906.09116v3.pdf
PWC https://paperswithcode.com/paper/differentially-private-summation-with-multi
Repo
Framework

Putting Humans in a Scene: Learning Affordance in 3D Indoor Environments

Title Putting Humans in a Scene: Learning Affordance in 3D Indoor Environments
Authors Xueting Li, Sifei Liu, Kihwan Kim, Xiaolong Wang, Ming-Hsuan Yang, Jan Kautz
Abstract Affordance modeling plays an important role in visual understanding. In this paper, we aim to predict affordances of 3D indoor scenes, specifically what human poses are afforded by a given indoor environment, such as sitting on a chair or standing on the floor. In order to predict valid affordances and learn possible 3D human poses in indoor scenes, we need to understand the semantic and geometric structure of a scene as well as its potential interactions with a human. To learn such a model, a large-scale dataset of 3D indoor affordances is required. In this work, we build a fully automatic 3D pose synthesizer that fuses semantic knowledge from a large number of 2D poses extracted from TV shows as well as 3D geometric knowledge from voxel representations of indoor scenes. With the data created by the synthesizer, we introduce a 3D pose generative model to predict semantically plausible and physically feasible human poses within a given scene (provided as a single RGB, RGB-D, or depth image). We demonstrate that our human affordance prediction method consistently outperforms existing state-of-the-art methods.
Tasks
Published 2019-03-13
URL http://arxiv.org/abs/1903.05690v2
PDF http://arxiv.org/pdf/1903.05690v2.pdf
PWC https://paperswithcode.com/paper/putting-humans-in-a-scene-learning-affordance
Repo
Framework

Refined Complexity of PCA with Outliers

Title Refined Complexity of PCA with Outliers
Authors Fedor V. Fomin, Petr A. Golovach, Fahad Panolan, Kirill Simonov
Abstract Principal component analysis (PCA) is one of the most fundamental procedures in exploratory data analysis and is the basic step in applications ranging from quantitative finance and bioinformatics to image analysis and neuroscience. However, it is well-documented that the applicability of PCA in many real scenarios could be constrained by an “immune deficiency” to outliers such as corrupted observations. We consider the following algorithmic question about the PCA with outliers. For a set of $n$ points in $\mathbb{R}^{d}$, how to learn a subset of points, say 1% of the total number of points, such that the remaining part of the points is best fit into some unknown $r$-dimensional subspace? We provide a rigorous algorithmic analysis of the problem. We show that the problem is solvable in time $n^{O(d^2)}$. In particular, for constant dimension the problem is solvable in polynomial time. We complement the algorithmic result by the lower bound, showing that unless Exponential Time Hypothesis fails, in time $f(d)n^{o(d)}$, for any function $f$ of $d$, it is impossible not only to solve the problem exactly but even to approximate it within a constant factor.
Tasks
Published 2019-05-10
URL https://arxiv.org/abs/1905.04124v1
PDF https://arxiv.org/pdf/1905.04124v1.pdf
PWC https://paperswithcode.com/paper/refined-complexity-of-pca-with-outliers
Repo
Framework

Explaining a prediction in some nonlinear models

Title Explaining a prediction in some nonlinear models
Authors Cosimo Izzo
Abstract In this article we will analyse how to compute the contribution of each input value to its aggregate output in some nonlinear models. Regression and classification applications, together with related algorithms for deep neural networks are presented. The proposed approach merges two methods currently present in the literature: integrated gradient and deep Taylor decomposition. Compared to DeepLIFT and Deep SHAP, it provides a natural choice of the reference point peculiar to the model at use.
Tasks
Published 2019-04-21
URL https://arxiv.org/abs/1904.09615v3
PDF https://arxiv.org/pdf/1904.09615v3.pdf
PWC https://paperswithcode.com/paper/explaining-a-prediction-in-some-nonlinear
Repo
Framework

Graph Spectral Embedding for Parsimonious Transmission of Multivariate Time Series

Title Graph Spectral Embedding for Parsimonious Transmission of Multivariate Time Series
Authors Lihan Yao, Paul Bendich
Abstract We propose a graph spectral representation of time series data that 1) is parsimoniously encoded to user-demanded resolution; 2) is unsupervised and performant in data-constrained scenarios; 3) captures event and event-transition structure within the time series; and 4) has near-linear computational complexity in both signal length and ambient dimension. This representation, which we call Laplacian Events Signal Segmentation (LESS), can be computed on time series of arbitrary dimension and originating from sensors of arbitrary type. Hence, time series originating from sensors of heterogeneous type can be compressed to levels demanded by constrained-communication environments, before being fused at a common center. Temporal dynamics of the data is summarized without explicit partitioning or probabilistic modeling. As a proof-of-principle, we apply this technique on high dimensional wavelet coefficients computed from the Free Spoken Digit Dataset to generate a memory efficient representation that is interpretable. Due to its unsupervised and non-parametric nature, LESS representations remain performant in the digit classification task despite the absence of labels and limited data.
Tasks Time Series
Published 2019-10-10
URL https://arxiv.org/abs/1910.04689v1
PDF https://arxiv.org/pdf/1910.04689v1.pdf
PWC https://paperswithcode.com/paper/graph-spectral-embedding-for-parsimonious
Repo
Framework

Improving Malaria Parasite Detection from Red Blood Cell using Deep Convolutional Neural Networks

Title Improving Malaria Parasite Detection from Red Blood Cell using Deep Convolutional Neural Networks
Authors Aimon Rahman, Hasib Zunair, M Sohel Rahman, Jesia Quader Yuki, Sabyasachi Biswas, Md Ashraful Alam, Nabila Binte Alam, M. R. C. Mahdy
Abstract Malaria is a female anopheles mosquito-bite inflicted life-threatening disease which is considered endemic in many parts of the world. This article focuses on improving malaria detection from patches segmented from microscopic images of red blood cell smears by introducing a deep convolutional neural network. Compared to the traditional methods that use tedious hand engineering feature extraction, the proposed method uses deep learning in an end-to-end arrangement that performs both feature extraction and classification directly from the raw segmented patches of the red blood smears. The dataset used in this study was taken from National Institute of Health named NIH Malaria Dataset. The evaluation metric accuracy and loss along with 5-fold cross validation was used to compare and select the best performing architecture. To maximize the performance, existing standard pre-processing techniques from the literature has also been experimented. In addition, several other complex architectures have been implemented and tested to pick the best performing model. A holdout test has also been conducted to verify how well the proposed model generalizes on unseen data. Our best model achieves an accuracy of almost 97.77%.
Tasks
Published 2019-07-23
URL https://arxiv.org/abs/1907.10418v1
PDF https://arxiv.org/pdf/1907.10418v1.pdf
PWC https://paperswithcode.com/paper/improving-malaria-parasite-detection-from-red
Repo
Framework

A different take on the best-first game tree pruning algorithms

Title A different take on the best-first game tree pruning algorithms
Authors Ishan Srivastava
Abstract The alpha-beta pruning algorithms have been popular in game tree searching ever since they were discovered. Numerous enhancements are proposed in literature and it is often overwhelming as to which would be the best for implementation. A certain enhancement can take far too long to fine tune its hyper parameters or to decide whether it is going to not make much of a difference due to the memory limitations. On the other hand are the best first pruning techniques, mostly the counterparts of the infamous SSS* algorithm, the algorithm which proved out to be disruptive at the time of its discovery but gradually became outcast as being too memory intensive and having a higher time complexity. Later research doesn’t see the best first approaches to be completely different from the depth first based enhancements but both seem to be transitionary in the sense that a best first approach could be looked as a depth first approach with a certain set of enhancements and with the growing power of the computers, SSS* didn’t seem to be as taxing on the memory either. Even so, there seems to be quite difficulty in understanding the nature of the SSS* algorithm, why it does what it does and it being termed as being too complex to fathom, visualize and understand on an intellectual level. This article tries to bridge this gap and provide some experimental results comparing the two with the most promising advances.
Tasks
Published 2019-11-08
URL https://arxiv.org/abs/1911.03388v1
PDF https://arxiv.org/pdf/1911.03388v1.pdf
PWC https://paperswithcode.com/paper/a-different-take-on-the-best-first-game-tree
Repo
Framework

An Efficient B-spline-Based Kinodynamic Replanning Framework for Quadrotors

Title An Efficient B-spline-Based Kinodynamic Replanning Framework for Quadrotors
Authors Wenchao Ding, Wenliang Gao, Kaixuan Wang, Shaojie Shen
Abstract Trajectory replanning for quadrotors is essential to enable fully autonomous flight in unknown environments. Hierarchical motion planning frameworks, which combine path planning with path parameterization, are popular due to their time efficiency. However, the path planning cannot properly deal with non-static initial states of the quadrotor, which may result in non-smooth or even dynamically infeasible trajectories. In this paper, we present an efficient kinodynamic replanning framework by exploiting the advantageous properties of the B-spline, which facilitates dealing with the non-static state and guarantees safety and dynamical feasibility. Our framework starts with an efficient B-spline-based kinodynamic (EBK) search algorithm which finds a feasible trajectory with minimum control effort and time. To compensate for the discretization induced by the EBK search, an elastic optimization (EO) approach is proposed to refine the control point placement to the optimal location. Systematic comparisons against the state-of-the-art are conducted to validate the performance. Comprehensive onboard experiments using two different vision-based quadrotors are carried out showing the general applicability of the framework.
Tasks Motion Planning
Published 2019-06-24
URL https://arxiv.org/abs/1906.09785v1
PDF https://arxiv.org/pdf/1906.09785v1.pdf
PWC https://paperswithcode.com/paper/an-efficient-b-spline-based-kinodynamic
Repo
Framework

Designing Trustworthy AI: A Human-Machine Teaming Framework to Guide Development

Title Designing Trustworthy AI: A Human-Machine Teaming Framework to Guide Development
Authors Carol J. Smith
Abstract Artificial intelligence (AI) holds great promise to empower us with knowledge and augment our effectiveness. We can – and must – ensure that we keep humans safe and in control, particularly with regard to government and public sector applications that affect broad populations. How can AI development teams harness the power of AI systems and design them to be valuable to humans? Diverse teams are needed to build trustworthy artificial intelligent systems, and those teams need to coalesce around a shared set of ethics. There are many discussions in the AI field about ethics and trust, but there are few frameworks available for people to use as guidance when creating these systems. The Human-Machine Teaming (HMT) Framework for Designing Ethical AI Experiences described in this paper, when used with a set of technical ethics, will guide AI development teams to create AI systems that are accountable, de-risked, respectful, secure, honest, and usable. To support the team’s efforts, activities to understand people’s needs and concerns will be introduced along with the themes to support the team’s efforts. For example, usability testing can help determine if the audience understands how the AI system works and complies with the HMT Framework. The HMT Framework is based on reviews of existing ethical codes and best practices in human-computer interaction and software development. Human-machine teams are strongest when human users can trust AI systems to behave as expected, safely, securely, and understandably. Using the HMT Framework to design trustworthy AI systems will provide support to teams in identifying potential issues ahead of time and making great experiences for humans.
Tasks
Published 2019-10-08
URL https://arxiv.org/abs/1910.03515v1
PDF https://arxiv.org/pdf/1910.03515v1.pdf
PWC https://paperswithcode.com/paper/designing-trustworthy-ai-a-human-machine
Repo
Framework

A Kaczmarz Algorithm for Solving Tree Based Distributed Systems of Equations

Title A Kaczmarz Algorithm for Solving Tree Based Distributed Systems of Equations
Authors Chinmay Hegde, Fritz Keinert, Eric S. Weber
Abstract The Kaczmarz algorithm is an iterative method for solving systems of linear equations. We introduce a modified Kaczmarz algorithm for solving systems of linear equations in a distributed environment, i.e. the equations within the system are distributed over multiple nodes within a network. The modification we introduce is designed for a network with a tree structure that allows for passage of solution estimates between the nodes in the network. We prove that the modified algorithm converges under no additional assumptions on the equations. We demonstrate that the algorithm converges to the solution, or the solution of minimal norm, when the system is consistent. We also demonstrate that in the case of an inconsistent system of equations, the modified relaxed Kaczmarz algorithm converges to a weighted least squares solution as the relaxation parameter approaches $0$.
Tasks
Published 2019-04-11
URL http://arxiv.org/abs/1904.05732v1
PDF http://arxiv.org/pdf/1904.05732v1.pdf
PWC https://paperswithcode.com/paper/a-kaczmarz-algorithm-for-solving-tree-based
Repo
Framework
comments powered by Disqus