January 26, 2020

3114 words 15 mins read

Paper Group ANR 1603

Paper Group ANR 1603

The statistical Minkowski distances: Closed-form formula for Gaussian Mixture Models. Single Image based Head Pose Estimation with Spherical Parameterization and 3D Morphing. Towards Real-Time Head Pose Estimation: Exploring Parameter-Reduced Residual Networks on In-the-wild Datasets. Deep Features for Tissue-Fold Detection in Histopathology Images …

The statistical Minkowski distances: Closed-form formula for Gaussian Mixture Models

Title The statistical Minkowski distances: Closed-form formula for Gaussian Mixture Models
Authors Frank Nielsen
Abstract The traditional Minkowski distances are induced by the corresponding Minkowski norms in real-valued vector spaces. In this work, we propose novel statistical symmetric distances based on the Minkowski’s inequality for probability densities belonging to Lebesgue spaces. These statistical Minkowski distances admit closed-form formula for Gaussian mixture models when parameterized by integer exponents. This result extends to arbitrary mixtures of exponential families with natural parameter spaces being cones: This includes the binomial, the multinomial, the zero-centered Laplacian, the Gaussian and the Wishart mixtures, among others. We also derive a Minkowski’s diversity index of a normalized weighted set of probability distributions from Minkowski’s inequality.
Tasks
Published 2019-01-09
URL http://arxiv.org/abs/1901.03732v2
PDF http://arxiv.org/pdf/1901.03732v2.pdf
PWC https://paperswithcode.com/paper/the-statistical-minkowski-distances-closed
Repo
Framework

Single Image based Head Pose Estimation with Spherical Parameterization and 3D Morphing

Title Single Image based Head Pose Estimation with Spherical Parameterization and 3D Morphing
Authors Hui Yuan, Mengyu Li, Junhui Hou, Jimin Xiao
Abstract Head pose estimation plays a vital role in various applications, e.g., driverassistance systems, human-computer interaction, virtual reality technology, and so on. We propose a novel geometry based algorithm for accurately estimating the head pose from a single 2D face image at a very low computational cost. Specifically, the rectangular coordinates of only four non-coplanar feature points from a predefined 3D facial model as well as the corresponding ones automatically/ manually extracted from a 2D face image are first normalized to exclude the effect of external factors (i.e., scale factor and translation parameters). Then, the four normalized 3D feature points are represented in spherical coordinates with reference to the uniquely determined sphere by themselves. Due to the spherical parameterization, the coordinates of feature points can then be morphed along all the three directions in the rectangular coordinates effectively. Finally, the rotation matrix indicating the head pose is obtained by minimizing the Euclidean distance between the normalized 2D feature points and the 2D re-projections of morphed 3D feature points. Comprehensive experimental results over two popular databases, i.e., Pointing’04 and Biwi Kinect, demonstrate that the proposed algorithm can estimate head poses with higher accuracy and lower run time than state-of-the-art geometry based methods. Even compared with start-of-the-art learning based methods or geometry based methods with additional depth information, our algorithm still produces comparable performance.
Tasks Head Pose Estimation, Pose Estimation
Published 2019-07-22
URL https://arxiv.org/abs/1907.09217v3
PDF https://arxiv.org/pdf/1907.09217v3.pdf
PWC https://paperswithcode.com/paper/a-single-image-based-head-pose-estimation
Repo
Framework

Towards Real-Time Head Pose Estimation: Exploring Parameter-Reduced Residual Networks on In-the-wild Datasets

Title Towards Real-Time Head Pose Estimation: Exploring Parameter-Reduced Residual Networks on In-the-wild Datasets
Authors Ines Rieger, Thomas Hauenstein, Sebastian Hettenkofer, Jens-Uwe Garbas
Abstract Head poses are a key component of human bodily communication and thus a decisive element of human-computer interaction. Real-time head pose estimation is crucial in the context of human-robot interaction or driver assistance systems. The most promising approaches for head pose estimation are based on Convolutional Neural Networks (CNNs). However, CNN models are often too complex to achieve real-time performance. To face this challenge, we explore a popular subgroup of CNNs, the Residual Networks (ResNets) and modify them in order to reduce their number of parameters. The ResNets are modifed for different image sizes including low-resolution images and combined with a varying number of layers. They are trained on in-the-wild datasets to ensure real-world applicability. As a result, we demonstrate that the performance of the ResNets can be maintained while reducing the number of parameters. The modified ResNets achieve state-of-the-art accuracy and provide fast inference for real-time applicability.
Tasks Head Pose Estimation, Pose Estimation
Published 2019-06-12
URL https://arxiv.org/abs/1906.05203v2
PDF https://arxiv.org/pdf/1906.05203v2.pdf
PWC https://paperswithcode.com/paper/towards-real-time-head-pose-estimation
Repo
Framework

Deep Features for Tissue-Fold Detection in Histopathology Images

Title Deep Features for Tissue-Fold Detection in Histopathology Images
Authors Morteza Babaie, H. R. Tizhoosh
Abstract Whole slide imaging (WSI) refers to the digitization of a tissue specimen which enables pathologists to explore high-resolution images on a monitor rather than through a microscope. The formation of tissue folds occur during tissue processing. Their presence may not only cause out-of-focus digitization but can also negatively affect the diagnosis in some cases. In this paper, we have compared five pre-trained convolutional neural networks (CNNs) of different depths as feature extractors to characterize tissue folds. We have also explored common classifiers to discriminate folded tissue against the normal tissue in hematoxylin and eosin (H&E) stained biopsy samples. In our experiments, we manually select the folded area in roughly 2.5mm $\times$ 2.5mm patches at $20$x magnification level as the training data. The ``DenseNet’’ with 201 layers alongside an SVM classifier outperformed all other configurations. Based on the leave-one-out validation strategy, we achieved $96.3%$ accuracy, whereas with augmentation the accuracy increased to $97.2%$. We have tested the generalization of our method with five unseen WSIs from the NIH (National Cancer Institute) dataset. The accuracy for patch-wise detection was $81%$. One folded patch within an image suffices to flag the entire specimen for visual inspection. |
Tasks
Published 2019-03-17
URL http://arxiv.org/abs/1903.07011v1
PDF http://arxiv.org/pdf/1903.07011v1.pdf
PWC https://paperswithcode.com/paper/deep-features-for-tissue-fold-detection-in
Repo
Framework

How to “DODGE” Complex Software Analytics?

Title How to “DODGE” Complex Software Analytics?
Authors Amritanshu Agrawal, Wei Fu, Di Chen, Xipeng Shen, Tim Menzies
Abstract Machine learning techniques applied to software engineering tasks can be improved by hyperparameter optimization, i.e., automatic tools that find good settings for a learner’s control parameters. We show that such hyperparameter optimization can be unnecessarily slow, particularly when the optimizers waste time exploring “redundant tunings”', i.e., pairs of tunings which lead to indistinguishable results. By ignoring redundant tunings, DODGE, a tuning tool, runs orders of magnitude faster, while also generating learners with more accurate predictions than seen in prior state-of-the-art approaches.
Tasks Hyperparameter Optimization
Published 2019-02-05
URL https://arxiv.org/abs/1902.01838v2
PDF https://arxiv.org/pdf/1902.01838v2.pdf
PWC https://paperswithcode.com/paper/how-to-dodge-complex-software-analytics
Repo
Framework

BioSNet: A Fast-Learning and High-Robustness Unsupervised Biomimetic Spiking Neural Network

Title BioSNet: A Fast-Learning and High-Robustness Unsupervised Biomimetic Spiking Neural Network
Authors Mingyuan Meng, Xingyu Yang, Shanlin Xiao, Zhiyi Yu
Abstract Spiking Neural Network (SNN), as a brain-inspired machine learning algorithm, is closer to the computing mechanism of human brain and more suitable to reveal the essence of intelligence compared with Artificial Neural Networks (ANN), attracting more and more attention in recent years. In addition, the information processed by SNN is in the form of discrete spikes, which makes SNN have low power consumption characteristics. In this paper, we propose an efficient and strong unsupervised SNN named BioSNet with high biological plausibility to handle image classification tasks. In BioSNet, we propose a new biomimetic spiking neuron model named MRON inspired by ‘recognition memory’ in the human brain, design an efficient and robust network architecture corresponding to biological characteristics of the human brain as well, and extend the traditional voting mechanism to the Vote-for-All (VFA) decoding layer so as to reduce information loss during decoding. Simulation results show that BioSNet not only achieves state-of-the-art unsupervised classification accuracy on MNIST/EMNIST data sets, but also exhibits superior learning efficiency and high robustness. Specifically, the BioSNet trained with only dozens of samples per class can achieve a favorable classification accuracy over 80% and randomly deleting even 95% of synapses or neurons in the BioSNet only leads to slight performance degradation.
Tasks Image Classification
Published 2019-12-02
URL https://arxiv.org/abs/2001.01680v2
PDF https://arxiv.org/pdf/2001.01680v2.pdf
PWC https://paperswithcode.com/paper/biosnet-a-fast-learning-and-high-robustness
Repo
Framework

Fast Neural Network Approach for Direct Covariant Forces Prediction in Complex Multi-Element Extended Systems

Title Fast Neural Network Approach for Direct Covariant Forces Prediction in Complex Multi-Element Extended Systems
Authors Jonathan P. Mailoa, Mordechai Kornbluth, Simon L. Batzner, Georgy Samsonidze, Stephen T. Lam, Chris Ablitt, Nicola Molinari, Boris Kozinsky
Abstract Neural network force field (NNFF) is a method for performing regression on atomic structure-force relationships, bypassing expensive quantum mechanics calculation which prevents the execution of long ab-initio quality molecular dynamics simulations. However, most NNFF methods for complex multi-element atomic systems indirectly predict atomic force vectors by exploiting just atomic structure rotation-invariant features and the network-feature spatial derivatives which are computationally expensive. We develop a staggered NNFF architecture exploiting both rotation-invariant and covariant features separately to directly predict atomic force vectors without using spatial derivatives, thereby reducing expensive structural feature calculation by ~180-480x. This acceleration enables us to develop NNFF which directly predicts atomic forces in complex ternary and quaternary-element extended systems comprised of long polymer chains, amorphous oxide, and surface chemical reactions. The staggered rotation-invariant-covariant architecture described here can also directly predict complex covariant vector outputs from local physical structures in domains beyond computational material science.
Tasks
Published 2019-05-07
URL https://arxiv.org/abs/1905.02791v1
PDF https://arxiv.org/pdf/1905.02791v1.pdf
PWC https://paperswithcode.com/paper/fast-neural-network-approach-for-direct
Repo
Framework

Frequentist Regret Bounds for Randomized Least-Squares Value Iteration

Title Frequentist Regret Bounds for Randomized Least-Squares Value Iteration
Authors Andrea Zanette, David Brandfonbrener, Emma Brunskill, Matteo Pirotta, Alessandro Lazaric
Abstract We consider the exploration-exploitation dilemma in finite-horizon reinforcement learning (RL). When the state space is large or continuous, traditional tabular approaches are unfeasible and some form of function approximation is mandatory. In this paper, we introduce an optimistically-initialized variant of the popular randomized least-squares value iteration (RLSVI), a model-free algorithm where exploration is induced by perturbing the least-squares approximation of the action-value function. Under the assumption that the Markov decision process has low-rank transition dynamics, we prove that the frequentist regret of RLSVI is upper-bounded by $\widetilde O(d^2 H^2 \sqrt{T})$ where $ d $ are the feature dimension, $ H $ is the horizon, and $ T $ is the total number of steps. To the best of our knowledge, this is the first frequentist regret analysis for randomized exploration with function approximation.
Tasks
Published 2019-11-01
URL https://arxiv.org/abs/1911.00567v4
PDF https://arxiv.org/pdf/1911.00567v4.pdf
PWC https://paperswithcode.com/paper/frequentist-regret-bounds-for-randomized
Repo
Framework

On Uniform Equivalence of Epistemic Logic Programs

Title On Uniform Equivalence of Epistemic Logic Programs
Authors Wolfgang Faber, Michael Morak, Stefan Woltran
Abstract Epistemic Logic Programs (ELPs) extend Answer Set Programming (ASP) with epistemic negation and have received renewed interest in recent years. This led to the development of new research and efficient solving systems for ELPs. In practice, ELPs are often written in a modular way, where each module interacts with other modules by accepting sets of facts as input, and passing on sets of facts as output. An interesting question then presents itself: under which conditions can such a module be replaced by another one without changing the outcome, for any set of input facts? This problem is known as uniform equivalence, and has been studied extensively for ASP. For ELPs, however, such an investigation is, as of yet, missing. In this paper, we therefore propose a characterization of uniform equivalence that can be directly applied to the language of state-of-the-art ELP solvers. We also investigate the computational complexity of deciding uniform equivalence for two ELPs, and show that it is on the third level of the polynomial hierarchy.
Tasks
Published 2019-07-25
URL https://arxiv.org/abs/1907.10925v1
PDF https://arxiv.org/pdf/1907.10925v1.pdf
PWC https://paperswithcode.com/paper/on-uniform-equivalence-of-epistemic-logic
Repo
Framework

Consistent Risk Estimation in High-Dimensional Linear Regression

Title Consistent Risk Estimation in High-Dimensional Linear Regression
Authors Ji Xu, Arian Maleki, Kamiar Rahnama Rad
Abstract Risk estimation is at the core of many learning systems. The importance of this problem has motivated researchers to propose different schemes, such as cross validation, generalized cross validation, and Bootstrap. The theoretical properties of such estimates have been extensively studied in the low-dimensional settings, where the number of predictors $p$ is much smaller than the number of observations $n$. However, a unifying methodology accompanied with a rigorous theory is lacking in high-dimensional settings. This paper studies the problem of risk estimation under the high-dimensional asymptotic setting $n,p \rightarrow \infty$ and $n/p \rightarrow \delta$ ($\delta$ is a fixed number), and proves the consistency of three risk estimates that have been successful in numerical studies, i.e., leave-one-out cross validation (LOOCV), approximate leave-one-out (ALO), and approximate message passing (AMP)-based techniques. A corner stone of our analysis is a bound that we obtain on the discrepancy of the `residuals’ obtained from AMP and LOOCV. This connection not only enables us to obtain a more refined information on the estimates of AMP, ALO, and LOOCV, but also offers an upper bound on the convergence rate of each estimate. |
Tasks
Published 2019-02-05
URL http://arxiv.org/abs/1902.01753v2
PDF http://arxiv.org/pdf/1902.01753v2.pdf
PWC https://paperswithcode.com/paper/consistent-risk-estimation-in-high
Repo
Framework

Trading-Off Static and Dynamic Regret in Online Least-Squares and Beyond

Title Trading-Off Static and Dynamic Regret in Online Least-Squares and Beyond
Authors Jianjun Yuan, Andrew Lamperski
Abstract Recursive least-squares algorithms often use forgetting factors as a heuristic to adapt to non-stationary data streams. The first contribution of this paper rigorously characterizes the effect of forgetting factors for a class of online Newton algorithms. For exp-concave and strongly convex objectives, the algorithms achieve the dynamic regret of $\max{O(\log T),O(\sqrt{TV})}$, where $V$ is a bound on the path length of the comparison sequence. In particular, we show how classic recursive least-squares with a forgetting factor achieves this dynamic regret bound. By varying $V$, we obtain a trade-off between static and dynamic regret. In order to obtain more computationally efficient algorithms, our second contribution is a novel gradient descent step size rule for strongly convex functions. Our gradient descent rule recovers the order optimal dynamic regret bounds described above. For smooth problems, we can also obtain static regret of $O(T^{1-\beta})$ and dynamic regret of $O(T^\beta V^*)$, where $\beta \in (0,1)$ and $V^*$ is the path length of the sequence of minimizers. By varying $\beta$, we obtain a trade-off between static and dynamic regret.
Tasks
Published 2019-09-06
URL https://arxiv.org/abs/1909.03118v2
PDF https://arxiv.org/pdf/1909.03118v2.pdf
PWC https://paperswithcode.com/paper/trading-off-static-and-dynamic-regret-in
Repo
Framework

Recommendations on Designing Practical Interval Type-2 Fuzzy Systems

Title Recommendations on Designing Practical Interval Type-2 Fuzzy Systems
Authors Dongrui Wu, Jerry Mendel
Abstract Interval type-2 (IT2) fuzzy systems have become increasingly popular in the last 20 years. They have demonstrated superior performance in many applications. However, the operation of an IT2 fuzzy system is more complex than that of its type-1 counterpart. There are many questions to be answered in designing an IT2 fuzzy system: Should singleton or non-singleton fuzzifier be used? How many membership functions (MFs) should be used for each input? Should Gaussian or piecewise linear MFs be used? Should Mamdani or Takagi-Sugeno-Kang (TSK) inference be used? Should minimum or product $t$-norm be used? Should type-reduction be used or not? How to optimize the IT2 fuzzy system? These questions may look overwhelming and confusing to IT2 beginners. In this paper we recommend some representative starting choices for an IT2 fuzzy system design, which hopefully will make IT2 fuzzy systems more accessible to IT2 fuzzy system designers.
Tasks
Published 2019-07-03
URL https://arxiv.org/abs/1907.01697v1
PDF https://arxiv.org/pdf/1907.01697v1.pdf
PWC https://paperswithcode.com/paper/recommendations-on-designing-practical
Repo
Framework

Incremental Improvement of a Question Answering System by Re-ranking Answer Candidates using Machine Learning

Title Incremental Improvement of a Question Answering System by Re-ranking Answer Candidates using Machine Learning
Authors Michael Barz, Daniel Sonntag
Abstract We implement a method for re-ranking top-10 results of a state-of-the-art question answering (QA) system. The goal of our re-ranking approach is to improve the answer selection given the user question and the top-10 candidates. We focus on improving deployed QA systems that do not allow re-training or re-training comes at a high cost. Our re-ranking approach learns a similarity function using n-gram based features using the query, the answer and the initial system confidence as input. Our contributions are: (1) we generate a QA training corpus starting from 877 answers from the customer care domain of T-Mobile Austria, (2) we implement a state-of-the-art QA pipeline using neural sentence embeddings that encode queries in the same space than the answer index, and (3) we evaluate the QA pipeline and our re-ranking approach using a separately provided test set. The test set can be considered to be available after deployment of the system, e.g., based on feedback of users. Our results show that the system performance, in terms of top-n accuracy and the mean reciprocal rank, benefits from re-ranking using gradient boosted regression trees. On average, the mean reciprocal rank improves by 9.15%.
Tasks Answer Selection, Question Answering, Sentence Embeddings
Published 2019-08-27
URL https://arxiv.org/abs/1908.10149v1
PDF https://arxiv.org/pdf/1908.10149v1.pdf
PWC https://paperswithcode.com/paper/incremental-improvement-of-a-question
Repo
Framework

Active Learning for Black-Box Adversarial Attacks in EEG-Based Brain-Computer Interfaces

Title Active Learning for Black-Box Adversarial Attacks in EEG-Based Brain-Computer Interfaces
Authors Xue Jiang, Xiao Zhang, Dongrui Wu
Abstract Deep learning has made significant breakthroughs in many fields, including electroencephalogram (EEG) based brain-computer interfaces (BCIs). However, deep learning models are vulnerable to adversarial attacks, in which deliberately designed small perturbations are added to the benign input samples to fool the deep learning model and degrade its performance. This paper considers transferability-based black-box attacks, where the attacker trains a substitute model to approximate the target model, and then generates adversarial examples from the substitute model to attack the target model. Learning a good substitute model is critical to the success of these attacks, but it requires a large number of queries to the target model. We propose a novel framework which uses query synthesis based active learning to improve the query efficiency in training the substitute model. Experiments on three convolutional neural network (CNN) classifiers and three EEG datasets demonstrated that our method can improve the attack success rate with the same number of queries, or, in other words, our method requires fewer queries to achieve a desired attack performance. To our knowledge, this is the first work that integrates active learning and adversarial attacks for EEG-based BCIs.
Tasks Active Learning, EEG
Published 2019-11-07
URL https://arxiv.org/abs/1911.04338v1
PDF https://arxiv.org/pdf/1911.04338v1.pdf
PWC https://paperswithcode.com/paper/active-learning-for-black-box-adversarial
Repo
Framework

Sequential Learning of Active Subspaces

Title Sequential Learning of Active Subspaces
Authors Nathan Wycoff, Mickael Binois, Stefan M. Wild
Abstract In recent years, active subspace methods (ASMs) have become a popular means of performing subspace sensitivity analysis on black-box functions. Naively applied, however, ASMs require gradient evaluations of the target function. In the event of noisy, expensive, or stochastic simulators, evaluating gradients via finite differencing may be infeasible. In such cases, often a surrogate model is employed, on which finite differencing is performed. When the surrogate model is a Gaussian process, we show that the ASM estimator is available in closed form, rendering the finite-difference approximation unnecessary. We use our closed-form solution to develop acquisition functions focused on sequential learning tailored to sensitivity analysis on top of ASMs. We also show that the traditional ASM estimator may be viewed as a method of moments estimator for a certain class of Gaussian processes. We demonstrate how uncertainty on Gaussian process hyperparameters may be propagated to uncertainty on the sensitivity analysis, allowing model-based confidence intervals on the active subspace. Our methodological developments are illustrated on several examples.
Tasks Gaussian Processes
Published 2019-07-26
URL https://arxiv.org/abs/1907.11572v1
PDF https://arxiv.org/pdf/1907.11572v1.pdf
PWC https://paperswithcode.com/paper/sequential-learning-of-active-subspaces
Repo
Framework
comments powered by Disqus