February 1, 2020

2774 words 14 mins read

Paper Group AWR 360

Paper Group AWR 360

A Multi-level procedure for enhancing accuracy of machine learning algorithms. Template-free Data-to-Text Generation of Finnish Sports News. On the (In)fidelity and Sensitivity for Explanations. Regression Concept Vectors for Bidirectional Explanations in Histopathology. ASSERT: Anti-Spoofing with Squeeze-Excitation and Residual neTworks. BMN: Boun …

A Multi-level procedure for enhancing accuracy of machine learning algorithms

Title A Multi-level procedure for enhancing accuracy of machine learning algorithms
Authors Kjetil O. Lye, Siddhartha Mishra, Roberto Molinaro
Abstract We propose a multi-level method to increase the accuracy of machine learning algorithms for approximating observables in scientific computing, particularly those that arise in systems modeled by differential equations. The algorithm relies on judiciously combining a large number of computationally cheap training data on coarse resolutions with a few expensive training samples on fine grid resolutions. Theoretical arguments for lowering the generalization error, based on reducing the variance of the underlying maps, are provided and numerical evidence, indicating significant gains over underlying single-level machine learning algorithms, are presented. Moreover, we also apply the multi-level algorithm in the context of forward uncertainty quantification and observe a considerable speed-up over competing algorithms.
Tasks
Published 2019-09-20
URL https://arxiv.org/abs/1909.09448v1
PDF https://arxiv.org/pdf/1909.09448v1.pdf
PWC https://paperswithcode.com/paper/a-multi-level-procedure-for-enhancing
Repo https://github.com/mroberto166/MultilevelMachineLearning
Framework tf

Template-free Data-to-Text Generation of Finnish Sports News

Title Template-free Data-to-Text Generation of Finnish Sports News
Authors Jenna Kanerva, Samuel Rönnqvist, Riina Kekki, Tapio Salakoski, Filip Ginter
Abstract News articles such as sports game reports are often thought to closely follow the underlying game statistics, but in practice they contain a notable amount of background knowledge, interpretation, insight into the game, and quotes that are not present in the official statistics. This poses a challenge for automated data-to-text news generation with real-world news corpora as training data. We report on the development of a corpus of Finnish ice hockey news, edited to be suitable for training of end-to-end news generation methods, as well as demonstrate generation of text, which was judged by journalists to be relatively close to a viable product. The new dataset and system source code are available for research purposes at https://github.com/scoopmatic/finnish-hockey-news-generation-paper.
Tasks Data-to-Text Generation, News Generation, Text Generation
Published 2019-10-04
URL https://arxiv.org/abs/1910.01863v1
PDF https://arxiv.org/pdf/1910.01863v1.pdf
PWC https://paperswithcode.com/paper/template-free-data-to-text-generation-of
Repo https://github.com/scoopmatic/finnish-hockey-news-generation-paper
Framework none

On the (In)fidelity and Sensitivity for Explanations

Title On the (In)fidelity and Sensitivity for Explanations
Authors Chih-Kuan Yeh, Cheng-Yu Hsieh, Arun Sai Suggala, David I. Inouye, Pradeep Ravikumar
Abstract We consider objective evaluation measures of saliency explanations for complex black-box machine learning models. We propose simple robust variants of two notions that have been considered in recent literature: (in)fidelity, and sensitivity. We analyze optimal explanations with respect to both these measures, and while the optimal explanation for sensitivity is a vacuous constant explanation, the optimal explanation for infidelity is a novel combination of two popular explanation methods. By varying the perturbation distribution that defines infidelity, we obtain novel explanations by optimizing infidelity, which we show to out-perform existing explanations in both quantitative and qualitative measurements. Another salient question given these measures is how to modify any given explanation to have better values with respect to these measures. We propose a simple modification based on lowering sensitivity, and moreover show that when done appropriately, we could simultaneously improve both sensitivity as well as fidelity.
Tasks
Published 2019-01-27
URL https://arxiv.org/abs/1901.09392v4
PDF https://arxiv.org/pdf/1901.09392v4.pdf
PWC https://paperswithcode.com/paper/how-sensitive-are-sensitivity-based
Repo https://github.com/chihkuanyeh/saliency_evaluation
Framework pytorch

Regression Concept Vectors for Bidirectional Explanations in Histopathology

Title Regression Concept Vectors for Bidirectional Explanations in Histopathology
Authors Mara Graziani, Vincent Andrearczyk, Henning Müller
Abstract Explanations for deep neural network predictions in terms of domain-related concepts can be valuable in medical applications, where justifications are important for confidence in the decision-making. In this work, we propose a methodology to exploit continuous concept measures as Regression Concept Vectors (RCVs) in the activation space of a layer. The directional derivative of the decision function along the RCVs represents the network sensitivity to increasing values of a given concept measure. When applied to breast cancer grading, nuclei texture emerges as a relevant concept in the detection of tumor tissue in breast lymph node samples. We evaluate score robustness and consistency by statistical analysis.
Tasks Breast Cancer Detection, Breast Cancer Histology Image Classification, Decision Making, Histopathological Image Classification
Published 2019-04-09
URL http://arxiv.org/abs/1904.04520v1
PDF http://arxiv.org/pdf/1904.04520v1.pdf
PWC https://paperswithcode.com/paper/regression-concept-vectors-for-bidirectional
Repo https://github.com/medgift/iMIMIC-RCVs
Framework tf

ASSERT: Anti-Spoofing with Squeeze-Excitation and Residual neTworks

Title ASSERT: Anti-Spoofing with Squeeze-Excitation and Residual neTworks
Authors Cheng-I Lai, Nanxin Chen, Jesús Villalba, Najim Dehak
Abstract We present JHU’s system submission to the ASVspoof 2019 Challenge: Anti-Spoofing with Squeeze-Excitation and Residual neTworks (ASSERT). Anti-spoofing has gathered more and more attention since the inauguration of the ASVspoof Challenges, and ASVspoof 2019 dedicates to address attacks from all three major types: text-to-speech, voice conversion, and replay. Built upon previous research work on Deep Neural Network (DNN), ASSERT is a pipeline for DNN-based approach to anti-spoofing. ASSERT has four components: feature engineering, DNN models, network optimization and system combination, where the DNN models are variants of squeeze-excitation and residual networks. We conducted an ablation study of the effectiveness of each component on the ASVspoof 2019 corpus, and experimental results showed that ASSERT obtained more than 93% and 17% relative improvements over the baseline systems in the two sub-challenges in ASVspooof 2019, ranking ASSERT one of the top performing systems. Code and pretrained models will be made publicly available.
Tasks Feature Engineering, Voice Conversion
Published 2019-04-01
URL http://arxiv.org/abs/1904.01120v1
PDF http://arxiv.org/pdf/1904.01120v1.pdf
PWC https://paperswithcode.com/paper/assert-anti-spoofing-with-squeeze-excitation
Repo https://github.com/jefflai108/ASSERT
Framework pytorch

BMN: Boundary-Matching Network for Temporal Action Proposal Generation

Title BMN: Boundary-Matching Network for Temporal Action Proposal Generation
Authors Tianwei Lin, Xiao Liu, Xin Li, Errui Ding, Shilei Wen
Abstract Temporal action proposal generation is an challenging and promising task which aims to locate temporal regions in real-world videos where action or event may occur. Current bottom-up proposal generation methods can generate proposals with precise boundary, but cannot efficiently generate adequately reliable confidence scores for retrieving proposals. To address these difficulties, we introduce the Boundary-Matching (BM) mechanism to evaluate confidence scores of densely distributed proposals, which denote a proposal as a matching pair of starting and ending boundaries and combine all densely distributed BM pairs into the BM confidence map. Based on BM mechanism, we propose an effective, efficient and end-to-end proposal generation method, named Boundary-Matching Network (BMN), which generates proposals with precise temporal boundaries as well as reliable confidence scores simultaneously. The two-branches of BMN are jointly trained in an unified framework. We conduct experiments on two challenging datasets: THUMOS-14 and ActivityNet-1.3, where BMN shows significant performance improvement with remarkable efficiency and generalizability. Further, combining with existing action classifier, BMN can achieve state-of-the-art temporal action detection performance.
Tasks Action Detection, Temporal Action Localization, Temporal Action Proposal Generation
Published 2019-07-23
URL https://arxiv.org/abs/1907.09702v1
PDF https://arxiv.org/pdf/1907.09702v1.pdf
PWC https://paperswithcode.com/paper/bmn-boundary-matching-network-for-temporal
Repo https://github.com/PaddlePaddle/models/tree/develop/dygraph/bmn
Framework none

Fast Low-rank Metric Learning for Large-scale and High-dimensional Data

Title Fast Low-rank Metric Learning for Large-scale and High-dimensional Data
Authors Han Liu, Zhizhong Han, Yu-Shen Liu, Ming Gu
Abstract Low-rank metric learning aims to learn better discrimination of data subject to low-rank constraints. It keeps the intrinsic low-rank structure of datasets and reduces the time cost and memory usage in metric learning. However, it is still a challenge for current methods to handle datasets with both high dimensions and large numbers of samples. To address this issue, we present a novel fast low-rank metric learning (FLRML) method.FLRML casts the low-rank metric learning problem into an unconstrained optimization on the Stiefel manifold, which can be efficiently solved by searching along the descent curves of the manifold.FLRML significantly reduces the complexity and memory usage in optimization, which makes the method scalable to both high dimensions and large numbers of samples.Furthermore, we introduce a mini-batch version of FLRML to make the method scalable to larger datasets which are hard to be loaded and decomposed in limited memory. The outperforming experimental results show that our method is with high accuracy and much faster than the state-of-the-art methods under several benchmarks with large numbers of high-dimensional data. Code has been made available at https://github.com/highan911/FLRML
Tasks Metric Learning
Published 2019-09-13
URL https://arxiv.org/abs/1909.06297v1
PDF https://arxiv.org/pdf/1909.06297v1.pdf
PWC https://paperswithcode.com/paper/fast-low-rank-metric-learning-for-large-scale
Repo https://github.com/highan911/FLRML
Framework none

Representation Learning for Attributed Multiplex Heterogeneous Network

Title Representation Learning for Attributed Multiplex Heterogeneous Network
Authors Yukuo Cen, Xu Zou, Jianwei Zhang, Hongxia Yang, Jingren Zhou, Jie Tang
Abstract Network embedding (or graph embedding) has been widely used in many real-world applications. However, existing methods mainly focus on networks with single-typed nodes/edges and cannot scale well to handle large networks. Many real-world networks consist of billions of nodes and edges of multiple types, and each node is associated with different attributes. In this paper, we formalize the problem of embedding learning for the Attributed Multiplex Heterogeneous Network and propose a unified framework to address this problem. The framework supports both transductive and inductive learning. We also give the theoretical analysis of the proposed framework, showing its connection with previous works and proving its better expressiveness. We conduct systematical evaluations for the proposed framework on four different genres of challenging datasets: Amazon, YouTube, Twitter, and Alibaba. Experimental results demonstrate that with the learned embeddings from the proposed framework, we can achieve statistically significant improvements (e.g., 5.99-28.23% lift by F1 scores; p«0.01, t-test) over previous state-of-the-art methods for link prediction. The framework has also been successfully deployed on the recommendation system of a worldwide leading e-commerce company, Alibaba Group. Results of the offline A/B tests on product recommendation further confirm the effectiveness and efficiency of the framework in practice.
Tasks Graph Embedding, Link Prediction, Network Embedding, Product Recommendation, Representation Learning
Published 2019-05-05
URL https://arxiv.org/abs/1905.01669v2
PDF https://arxiv.org/pdf/1905.01669v2.pdf
PWC https://paperswithcode.com/paper/190501669
Repo https://github.com/cenyk1230/GATNE
Framework tf

On the design of convolutional neural networks for automatic detection of Alzheimer’s disease

Title On the design of convolutional neural networks for automatic detection of Alzheimer’s disease
Authors Sheng Liu, Chhavi Yadav, Carlos Fernandez-Granda, Narges Razavian
Abstract Early detection is a crucial goal in the study of Alzheimer’s Disease (AD). In this work, we describe several techniques to boost the performance of 3D convolutional neural networks trained to detect AD using structural brain MRI scans. Specifically, we provide evidence that (1) instance normalization outperforms batch normalization, (2) early spatial downsampling negatively affects performance, (3) widening the model brings consistent gains while increasing the depth does not, and (4) incorporating age information yields moderate improvement. Together, these insights yield an increment of approximately 14% in test accuracy over existing models when distinguishing between patients with AD, mild cognitive impairment, and controls in the ADNI dataset. Similar performance is achieved on an independent dataset.
Tasks
Published 2019-11-09
URL https://arxiv.org/abs/1911.03740v2
PDF https://arxiv.org/pdf/1911.03740v2.pdf
PWC https://paperswithcode.com/paper/on-the-design-of-convolutional-neural
Repo https://github.com/NYUMedML/CNN_design_for_AD
Framework pytorch

DALS: Deep Active Lesion Segmentation

Title DALS: Deep Active Lesion Segmentation
Authors Ali Hatamizadeh, Assaf Hoogi, Debleena Sengupta, Wuyue Lu, Brian Wilcox, Daniel Rubin, Demetri Terzopoulos
Abstract Lesion segmentation is an important problem in computer-assisted diagnosis that remains challenging due to the prevalence of low contrast, irregular boundaries that are unamenable to shape priors. We introduce Deep Active Lesion Segmentation (DALS), a fully automated segmentation framework for that leverages the powerful nonlinear feature extraction abilities of fully Convolutional Neural Networks (CNNs) and the precise boundary delineation abilities of Active Contour Models (ACMs). Our DALS framework benefits from an improved level-set ACM formulation with a per-pixel-parameterized energy functional and a novel multiscale encoder-decoder CNN that learns an initialization probability map along with parameter maps for the ACM. We evaluate our lesion segmentation model on a new Multiorgan Lesion Segmentation (MLS) dataset that contains images of various organs, including brain, liver, and lung, across different imaging modalities—MR and CT. Our results demonstrate favorable performance compared to competing methods, especially for small training datasets.
Tasks Lesion Segmentation
Published 2019-08-19
URL https://arxiv.org/abs/1908.06933v3
PDF https://arxiv.org/pdf/1908.06933v3.pdf
PWC https://paperswithcode.com/paper/deep-active-lesion-segmentation
Repo https://github.com/ahatamiz/DALS
Framework none

Low-variance Black-box Gradient Estimates for the Plackett-Luce Distribution

Title Low-variance Black-box Gradient Estimates for the Plackett-Luce Distribution
Authors Artyom Gadetsky, Kirill Struminsky, Christopher Robinson, Novi Quadrianto, Dmitry Vetrov
Abstract Learning models with discrete latent variables using stochastic gradient descent remains a challenge due to the high variance of gradient estimates. Modern variance reduction techniques mostly consider categorical distributions and have limited applicability when the number of possible outcomes becomes large. In this work, we consider models with latent permutations and propose control variates for the Plackett-Luce distribution. In particular, the control variates allow us to optimize black-box functions over permutations using stochastic gradient descent. To illustrate the approach, we consider a variety of causal structure learning tasks for continuous and discrete data. We show that our method outperforms competitive relaxation-based optimization methods and is also applicable to non-differentiable score functions.
Tasks
Published 2019-11-22
URL https://arxiv.org/abs/1911.10036v1
PDF https://arxiv.org/pdf/1911.10036v1.pdf
PWC https://paperswithcode.com/paper/low-variance-black-box-gradient-estimates-for
Repo https://github.com/agadetsky/pytorch-pl-variance-reduction
Framework pytorch

Transfer learning in hybrid classical-quantum neural networks

Title Transfer learning in hybrid classical-quantum neural networks
Authors Andrea Mari, Thomas R. Bromley, Josh Izaac, Maria Schuld, Nathan Killoran
Abstract We extend the concept of transfer learning, widely applied in modern machine learning algorithms, to the emerging context of hybrid neural networks composed of classical and quantum elements. We propose different implementations of hybrid transfer learning, but we focus mainly on the paradigm in which a pre-trained classical network is modified and augmented by a final variational quantum circuit. This approach is particularly attractive in the current era of intermediate-scale quantum technology since it allows to optimally pre-process high dimensional data (e.g., images) with any state-of-the-art classical network and to embed a select set of highly informative features into a quantum processor. We present several proof-of-concept examples of the convenient application of quantum transfer learning for image recognition and quantum state classification. We use the cross-platform software library PennyLane to experimentally test a high-resolution image classifier with two different quantum computers, respectively provided by IBM and Rigetti.
Tasks Transfer Learning
Published 2019-12-17
URL https://arxiv.org/abs/1912.08278v1
PDF https://arxiv.org/pdf/1912.08278v1.pdf
PWC https://paperswithcode.com/paper/transfer-learning-in-hybrid-classical-quantum
Repo https://github.com/XanaduAI/quantum-transfer-learning
Framework pytorch

Mitigating Information Leakage in Image Representations: A Maximum Entropy Approach

Title Mitigating Information Leakage in Image Representations: A Maximum Entropy Approach
Authors Proteek Chandan Roy, Vishnu Naresh Boddeti
Abstract Image recognition systems have demonstrated tremendous progress over the past few decades thanks, in part, to our ability of learning compact and robust representations of images. As we witness the wide spread adoption of these systems, it is imperative to consider the problem of unintended leakage of information from an image representation, which might compromise the privacy of the data owner. This paper investigates the problem of learning an image representation that minimizes such leakage of user information. We formulate the problem as an adversarial non-zero sum game of finding a good embedding function with two competing goals: to retain as much task dependent discriminative image information as possible, while simultaneously minimizing the amount of information, as measured by entropy, about other sensitive attributes of the user. We analyze the stability and convergence dynamics of the proposed formulation using tools from non-linear systems theory and compare to that of the corresponding adversarial zero-sum game formulation that optimizes likelihood as a measure of information content. Numerical experiments on UCI, Extended Yale B, CIFAR-10 and CIFAR-100 datasets indicate that our proposed approach is able to learn image representations that exhibit high task performance while mitigating leakage of predefined sensitive information.
Tasks
Published 2019-04-11
URL http://arxiv.org/abs/1904.05514v1
PDF http://arxiv.org/pdf/1904.05514v1.pdf
PWC https://paperswithcode.com/paper/mitigating-information-leakage-in-image
Repo https://github.com/human-analysis/MaxEnt-ARL
Framework pytorch

Computational Mirrors: Blind Inverse Light Transport by Deep Matrix Factorization

Title Computational Mirrors: Blind Inverse Light Transport by Deep Matrix Factorization
Authors Miika Aittala, Prafull Sharma, Lukas Murmann, Adam B. Yedidia, Gregory W. Wornell, William T. Freeman, Fredo Durand
Abstract We recover a video of the motion taking place in a hidden scene by observing changes in indirect illumination in a nearby uncalibrated visible region. We solve this problem by factoring the observed video into a matrix product between the unknown hidden scene video and an unknown light transport matrix. This task is extremely ill-posed, as any non-negative factorization will satisfy the data. Inspired by recent work on the Deep Image Prior, we parameterize the factor matrices using randomly initialized convolutional neural networks trained in a one-off manner, and show that this results in decompositions that reflect the true motion in the hidden scene.
Tasks
Published 2019-12-05
URL https://arxiv.org/abs/1912.02314v1
PDF https://arxiv.org/pdf/1912.02314v1.pdf
PWC https://paperswithcode.com/paper/computational-mirrors-blind-inverse-light-1
Repo https://github.com/prafull7/compmirrors
Framework pytorch

dAUTOMAP: decomposing AUTOMAP to achieve scalability and enhance performance

Title dAUTOMAP: decomposing AUTOMAP to achieve scalability and enhance performance
Authors Jo Schlemper, Ilkay Oksuz, James R. Clough, Jinming Duan, Andrew P. King, Julia A. Schnabel, Joseph V. Hajnal, Daniel Rueckert
Abstract AUTOMAP is a promising generalized reconstruction approach, however, it is not scalable and hence the practicality is limited. We present dAUTOMAP, a novel way for decomposing the domain transformation of AUTOMAP, making the model scale linearly. We show dAUTOMAP outperforms AUTOMAP with significantly fewer parameters.
Tasks
Published 2019-09-24
URL https://arxiv.org/abs/1909.10995v2
PDF https://arxiv.org/pdf/1909.10995v2.pdf
PWC https://paperswithcode.com/paper/dautomap-decomposing-automap-to-achieve
Repo https://github.com/js3611/dAUTOMAP
Framework pytorch
comments powered by Disqus