January 28, 2020

2854 words 14 mins read

Paper Group ANR 850

Paper Group ANR 850

Geometry of the Hough transforms with applications to synthetic data. On Dropout and Nuclear Norm Regularization. Accelerating Training using Tensor Decomposition. Weighted Sampling for Combined Model Selection and Hyperparameter Tuning. Sequential Dynamic Resource Allocation for Epidemic Control. Latent-Space Laplacian Pyramids for Adversarial Rep …

Geometry of the Hough transforms with applications to synthetic data

Title Geometry of the Hough transforms with applications to synthetic data
Authors Mauro C. Beltrametti, Cristina Campi, Anna Maria Massone, Maria-Laura Torrente
Abstract In the framework of the Hough transform technique to detect curves in images, we provide a bound for the number of Hough transforms to be considered for a successful optimization of the accumulator function in the recognition algorithm. Such a bound is consequence of geometrical arguments. We also show the robustness of the results when applied to synthetic datasets strongly perturbed by noise. An algebraic approach, discussed in the appendix, leads to a better bound of theoretical interest in the exact case.
Tasks
Published 2019-04-04
URL http://arxiv.org/abs/1904.02587v1
PDF http://arxiv.org/pdf/1904.02587v1.pdf
PWC https://paperswithcode.com/paper/geometry-of-the-hough-transforms-with
Repo
Framework

On Dropout and Nuclear Norm Regularization

Title On Dropout and Nuclear Norm Regularization
Authors Poorya Mianjy, Raman Arora
Abstract We give a formal and complete characterization of the explicit regularizer induced by dropout in deep linear networks with squared loss. We show that (a) the explicit regularizer is composed of an $\ell_2$-path regularizer and other terms that are also re-scaling invariant, (b) the convex envelope of the induced regularizer is the squared nuclear norm of the network map, and (c) for a sufficiently large dropout rate, we characterize the global optima of the dropout objective. We validate our theoretical findings with empirical results.
Tasks
Published 2019-05-28
URL https://arxiv.org/abs/1905.11887v1
PDF https://arxiv.org/pdf/1905.11887v1.pdf
PWC https://paperswithcode.com/paper/on-dropout-and-nuclear-norm-regularization
Repo
Framework

Accelerating Training using Tensor Decomposition

Title Accelerating Training using Tensor Decomposition
Authors Mostafa Elhoushi, Ye Henry Tian, Zihao Chen, Farhan Shafiq, Joey Yiwei Li
Abstract Tensor decomposition is one of the well-known approaches to reduce the latency time and number of parameters of a pre-trained model. However, in this paper, we propose an approach to use tensor decomposition to reduce training time of training a model from scratch. In our approach, we train the model from scratch (i.e., randomly initialized weights) with its original architecture for a small number of epochs, then the model is decomposed, and then continue training the decomposed model till the end. There is an optional step in our approach to convert the decomposed architecture back to the original architecture. We present results of using this approach on both CIFAR10 and Imagenet datasets, and show that there can be upto 2x speed up in training time with accuracy drop of upto 1.5% only, and in other cases no accuracy drop. This training acceleration approach is independent of hardware and is expected to have similar speed ups on both CPU and GPU platforms.
Tasks
Published 2019-09-10
URL https://arxiv.org/abs/1909.05675v1
PDF https://arxiv.org/pdf/1909.05675v1.pdf
PWC https://paperswithcode.com/paper/accelerating-training-using-tensor
Repo
Framework

Weighted Sampling for Combined Model Selection and Hyperparameter Tuning

Title Weighted Sampling for Combined Model Selection and Hyperparameter Tuning
Authors Dimitrios Sarigiannis, Thomas Parnell, Haris Pozidis
Abstract The combined algorithm selection and hyperparameter tuning (CASH) problem is characterized by large hierarchical hyperparameter spaces. Model-free hyperparameter tuning methods can explore such large spaces efficiently since they are highly parallelizable across multiple machines. When no prior knowledge or meta-data exists to boost their performance, these methods commonly sample random configurations following a uniform distribution. In this work, we propose a novel sampling distribution as an alternative to uniform sampling and prove theoretically that it has a better chance of finding the best configuration in a worst-case setting. In order to compare competing methods rigorously in an experimental setting, one must perform statistical hypothesis testing. We show that there is little-to-no agreement in the automated machine learning literature regarding which methods should be used. We contrast this disparity with the methods recommended by the broader statistics literature, and identify a suitable approach. We then select three popular model-free solutions to CASH and evaluate their performance, with uniform sampling as well as the proposed sampling scheme, across 67 datasets from the OpenML platform. We investigate the trade-off between exploration and exploitation across the three algorithms, and verify empirically that the proposed sampling distribution improves performance in all cases.
Tasks Model Selection
Published 2019-09-16
URL https://arxiv.org/abs/1909.07140v3
PDF https://arxiv.org/pdf/1909.07140v3.pdf
PWC https://paperswithcode.com/paper/weighted-sampling-for-combined-model
Repo
Framework

Sequential Dynamic Resource Allocation for Epidemic Control

Title Sequential Dynamic Resource Allocation for Epidemic Control
Authors Mathilde Fekom, Nicolas Vayatis, Argyris Kalogeratos
Abstract Under the Dynamic Resource Allocation (DRA) model, an administrator has the mission to allocate dynamically a limited budget of resources to the nodes of a network in order to reduce a diffusion process (DP) (e.g. an epidemic). The standard DRA assumes that the administrator has constantly full information and instantaneous access to the entire network. Towards bringing such strategies closer to real-life constraints, we first present the Restricted DRA model extension where, at each intervention round, the access is restricted to only a fraction of the network nodes, called sample. Then, inspired by sequential selection problems such as the well-known Secretary Problem, we propose the Sequential DRA (SDRA) model. Our model introduces a sequential aspect in the decision process over the sample of each round, offering a completely new perspective to the dynamic DP control. Finally, we incorporate several sequential selection algorithms to SDRA control strategies and compare their performance in SIS epidemic simulations.
Tasks
Published 2019-09-20
URL https://arxiv.org/abs/1909.09678v1
PDF https://arxiv.org/pdf/1909.09678v1.pdf
PWC https://paperswithcode.com/paper/190909678
Repo
Framework

Latent-Space Laplacian Pyramids for Adversarial Representation Learning with 3D Point Clouds

Title Latent-Space Laplacian Pyramids for Adversarial Representation Learning with 3D Point Clouds
Authors Vage Egiazarian, Savva Ignatyev, Alexey Artemov, Oleg Voynov, Andrey Kravchenko, Youyi Zheng, Luiz Velho, Evgeny Burnaev
Abstract Constructing high-quality generative models for 3D shapes is a fundamental task in computer vision with diverse applications in geometry processing, engineering, and design. Despite the recent progress in deep generative modelling, synthesis of finely detailed 3D surfaces, such as high-resolution point clouds, from scratch has not been achieved with existing approaches. In this work, we propose to employ the latent-space Laplacian pyramid representation within a hierarchical generative model for 3D point clouds. We combine the recently proposed latent-space GAN and Laplacian GAN architectures to form a multi-scale model capable of generating 3D point clouds at increasing levels of detail. Our evaluation demonstrates that our model outperforms the existing generative models for 3D point clouds.
Tasks Generating 3D Point Clouds, Representation Learning
Published 2019-12-13
URL https://arxiv.org/abs/1912.06466v1
PDF https://arxiv.org/pdf/1912.06466v1.pdf
PWC https://paperswithcode.com/paper/latent-space-laplacian-pyramids-for
Repo
Framework

Time Masking: Leveraging Temporal Information in Spoken Dialogue Systems

Title Time Masking: Leveraging Temporal Information in Spoken Dialogue Systems
Authors Rylan Conway, Lambert Mathias
Abstract In a spoken dialogue system, dialogue state tracker (DST) components track the state of the conversation by updating a distribution of values associated with each of the slots being tracked for the current user turn, using the interactions until then. Much of the previous work has relied on modeling the natural order of the conversation, using distance based offsets as an approximation of time. In this work, we hypothesize that leveraging the wall-clock temporal difference between turns is crucial for finer-grained control of dialogue scenarios. We develop a novel approach that applies a {\it time mask}, based on the wall-clock time difference, to the associated slot embeddings and empirically demonstrate that our proposed approach outperforms existing approaches that leverage distance offsets, on both an internal benchmark dataset as well as DSTC2.
Tasks Spoken Dialogue Systems
Published 2019-07-25
URL https://arxiv.org/abs/1907.11315v1
PDF https://arxiv.org/pdf/1907.11315v1.pdf
PWC https://paperswithcode.com/paper/time-masking-leveraging-temporal-information
Repo
Framework

Closed-loop Model Selection for Kernel-based Models using Bayesian Optimization

Title Closed-loop Model Selection for Kernel-based Models using Bayesian Optimization
Authors Thomas Beckers, Somil Bansal, Claire J. Tomlin, Sandra Hirche
Abstract Kernel-based nonparametric models have become very attractive for model-based control approaches for nonlinear systems. However, the selection of the kernel and its hyperparameters strongly influences the quality of the learned model. Classically, these hyperparameters are optimized to minimize the prediction error of the model but this process totally neglects its later usage in the control loop. In this work, we present a framework to optimize the kernel and hyperparameters of a kernel-based model directly with respect to the closed-loop performance of the model. Our framework uses Bayesian optimization to iteratively refine the kernel-based model using the observed performance on the actual system until a desired performance is achieved. We demonstrate the proposed approach in a simulation and on a 3-DoF robotic arm.
Tasks Model Selection
Published 2019-09-12
URL https://arxiv.org/abs/1909.05699v1
PDF https://arxiv.org/pdf/1909.05699v1.pdf
PWC https://paperswithcode.com/paper/closed-loop-model-selection-for-kernel-based
Repo
Framework

Incrementalizing RASA’s Open-Source Natural Language Understanding Pipeline

Title Incrementalizing RASA’s Open-Source Natural Language Understanding Pipeline
Authors Andrew Rafla, Casey Kennington
Abstract As spoken dialogue systems and chatbots are gaining more widespread adoption, commercial and open-sourced services for natural language understanding are emerging. In this paper, we explain how we altered the open-source RASA natural language understanding pipeline to process incrementally (i.e., word-by-word), following the incremental unit framework proposed by Schlangen and Skantze. To do so, we altered existing RASA components to process incrementally, and added an update-incremental intent recognition model as a component to RASA. Our evaluations on the Snips dataset show that our changes allow RASA to function as an effective incremental natural language understanding service.
Tasks Spoken Dialogue Systems
Published 2019-07-11
URL https://arxiv.org/abs/1907.05403v1
PDF https://arxiv.org/pdf/1907.05403v1.pdf
PWC https://paperswithcode.com/paper/incrementalizing-rasas-open-source-natural
Repo
Framework

Technical Considerations for Semantic Segmentation in MRI using Convolutional Neural Networks

Title Technical Considerations for Semantic Segmentation in MRI using Convolutional Neural Networks
Authors Arjun D. Desai, Garry E. Gold, Brian A. Hargreaves, Akshay S. Chaudhari
Abstract High-fidelity semantic segmentation of magnetic resonance volumes is critical for estimating tissue morphometry and relaxation parameters in both clinical and research applications. While manual segmentation is accepted as the gold-standard, recent advances in deep learning and convolutional neural networks (CNNs) have shown promise for efficient automatic segmentation of soft tissues. However, due to the stochastic nature of deep learning and the multitude of hyperparameters in training networks, predicting network behavior is challenging. In this paper, we quantify the impact of three factors associated with CNN segmentation performance: network architecture, training loss functions, and training data characteristics. We evaluate the impact of these variations on the segmentation of femoral cartilage and propose potential modifications to CNN architectures and training protocols to train these models with confidence.
Tasks Semantic Segmentation
Published 2019-02-05
URL http://arxiv.org/abs/1902.01977v1
PDF http://arxiv.org/pdf/1902.01977v1.pdf
PWC https://paperswithcode.com/paper/technical-considerations-for-semantic
Repo
Framework

Meta-Learning via Learned Loss

Title Meta-Learning via Learned Loss
Authors Sarah Bechtle, Artem Molchanov, Yevgen Chebotar, Edward Grefenstette, Ludovic Righetti, Gaurav Sukhatme, Franziska Meier
Abstract Typically, loss functions, regularization mechanisms and other important aspects of training parametric models are chosen heuristically from a limited set of options. In this paper, we take the first step towards automating this process, with the view of producing models which train faster and more robustly. Concretely, we present a meta-learning method for learning parametric loss functions that can generalize across different tasks and model architectures. We develop a pipeline for meta-training such loss functions, targeted at maximizing the performance of the model trained under them. The loss landscape produced by our learned losses significantly improves upon the original task-specific losses in both supervised and reinforcement learning tasks. Furthermore, we show that our meta-learning framework is flexible enough to incorporate additional information at meta-train time. This information shapes the learned loss function such that the environment does not need to provide this information during meta-test time.
Tasks Meta-Learning
Published 2019-06-12
URL https://arxiv.org/abs/1906.05374v3
PDF https://arxiv.org/pdf/1906.05374v3.pdf
PWC https://paperswithcode.com/paper/meta-learning-via-learned-loss
Repo
Framework

Hybrid Machine Learning Forecasts for the FIFA Women’s World Cup 2019

Title Hybrid Machine Learning Forecasts for the FIFA Women’s World Cup 2019
Authors Andreas Groll, Christophe Ley, Gunther Schauberger, Hans Van Eetvelde, Achim Zeileis
Abstract In this work, we combine two different ranking methods together with several other predictors in a joint random forest approach for the scores of soccer matches. The first ranking method is based on the bookmaker consensus, the second ranking method estimates adequate ability parameters that reflect the current strength of the teams best. The proposed combined approach is then applied to the data from the two previous FIFA Women’s World Cups 2011 and 2015. Finally, based on the resulting estimates, the FIFA Women’s World Cup 2019 is simulated repeatedly and winning probabilities are obtained for all teams. The model clearly favors the defending champion USA before the host France.
Tasks
Published 2019-06-03
URL https://arxiv.org/abs/1906.01131v1
PDF https://arxiv.org/pdf/1906.01131v1.pdf
PWC https://paperswithcode.com/paper/hybrid-machine-learning-forecasts-for-the
Repo
Framework

Subspace Clustering of Very Sparse High-Dimensional Data

Title Subspace Clustering of Very Sparse High-Dimensional Data
Authors Hankui Peng, Nicos Pavlidis, Idris Eckley, Ioannis Tsalamanis
Abstract In this paper we consider the problem of clustering collections of very short texts using subspace clustering. This problem arises in many applications such as product categorisation, fraud detection, and sentiment analysis. The main challenge lies in the fact that the vectorial representation of short texts is both high-dimensional, due to the large number of unique terms in the corpus, and extremely sparse, as each text contains a very small number of words with no repetition. We propose a new, simple subspace clustering algorithm that relies on linear algebra to cluster such datasets. Experimental results on identifying product categories from product names obtained from the US Amazon website indicate that the algorithm can be competitive against state-of-the-art clustering algorithms.
Tasks Fraud Detection, Sentiment Analysis
Published 2019-01-25
URL http://arxiv.org/abs/1901.09108v1
PDF http://arxiv.org/pdf/1901.09108v1.pdf
PWC https://paperswithcode.com/paper/subspace-clustering-of-very-sparse-high
Repo
Framework

Adversarially Trained Convolutional Neural Networks for Semantic Segmentation of Ischaemic Stroke Lesion using Multisequence Magnetic Resonance Imaging

Title Adversarially Trained Convolutional Neural Networks for Semantic Segmentation of Ischaemic Stroke Lesion using Multisequence Magnetic Resonance Imaging
Authors Rachana Sathish, Ronnie Rajan, Anusha Vupputuri, Nirmalya Ghosh, Debdoot Sheet
Abstract Ischaemic stroke is a medical condition caused by occlusion of blood supply to the brain tissue thus forming a lesion. A lesion is zoned into a core associated with irreversible necrosis typically located at the center of the lesion, while reversible hypoxic changes in the outer regions of the lesion are termed as the penumbra. Early estimation of core and penumbra in ischaemic stroke is crucial for timely intervention with thrombolytic therapy to reverse the damage and restore normalcy. Multisequence magnetic resonance imaging (MRI) is commonly employed for clinical diagnosis. However, a sequence singly has not been found to be sufficiently able to differentiate between core and penumbra, while a combination of sequences is required to determine the extent of the damage. The challenge, however, is that with an increase in the number of sequences, it cognitively taxes the clinician to discover symptomatic biomarkers in these images. In this paper, we present a data-driven fully automated method for estimation of core and penumbra in ischaemic lesions using diffusion-weighted imaging (DWI) and perfusion-weighted imaging (PWI) sequence maps of MRI. The method employs recent developments in convolutional neural networks (CNN) for semantic segmentation in medical images. In the absence of availability of a large amount of labeled data, the CNN is trained using an adversarial approach employing cross-entropy as a segmentation loss along with losses aggregated from three discriminators of which two employ relativistic visual Turing test. This method is experimentally validated on the ISLES-2015 dataset through three-fold cross-validation to obtain with an average Dice score of 0.82 and 0.73 for segmentation of penumbra and core respectively.
Tasks Semantic Segmentation
Published 2019-08-03
URL https://arxiv.org/abs/1908.01176v1
PDF https://arxiv.org/pdf/1908.01176v1.pdf
PWC https://paperswithcode.com/paper/adversarially-trained-convolutional-neural
Repo
Framework

Accurate and Fast reconstruction of Porous Media from Extremely Limited Information Using Conditional Generative Adversarial Network

Title Accurate and Fast reconstruction of Porous Media from Extremely Limited Information Using Conditional Generative Adversarial Network
Authors Junxi Feng, Xiaohai He, Qizhi Teng, Chao Ren, Honggang Chen, Yang Li
Abstract Porous media are ubiquitous in both nature and engineering applications, thus their modelling and understanding is of vital importance. In contrast to direct acquisition of three-dimensional (3D) images of such medium, obtaining its sub-region (s) like two-dimensional (2D) images or several small areas could be much feasible. Therefore, reconstructing whole images from the limited information is a primary technique in such cases. Specially, in practice the given data cannot generally be determined by users and may be incomplete or partially informed, thus making existing reconstruction methods inaccurate or even ineffective. To overcome this shortcoming, in this study we proposed a deep learning-based framework for reconstructing full image from its much smaller sub-area(s). Particularly, conditional generative adversarial network (CGAN) is utilized to learn the mapping between input (partial image) and output (full image). To preserve the reconstruction accuracy, two simple but effective objective functions are proposed and then coupled with the other two functions to jointly constrain the training procedure. Due to the inherent essence of this ill-posed problem, a Gaussian noise is introduced for producing reconstruction diversity, thus allowing for providing multiple candidate outputs. Extensively tested on a variety of porous materials and demonstrated by both visual inspection and quantitative comparison, the method is shown to be accurate, stable yet fast ($\sim0.08s$ for a $128 \times 128$ image reconstruction). We highlight that the proposed approach can be readily extended, such as incorporating any user-define conditional data and an arbitrary number of object functions into reconstruction, and being coupled with other reconstruction methods.
Tasks Image Reconstruction
Published 2019-04-04
URL http://arxiv.org/abs/1905.02135v1
PDF http://arxiv.org/pdf/1905.02135v1.pdf
PWC https://paperswithcode.com/paper/190502135
Repo
Framework
comments powered by Disqus