February 1, 2020

3022 words 15 mins read

Paper Group AWR 175

Paper Group AWR 175

Improving Document Classification with Multi-Sense Embeddings. Cost-Sensitive Diagnosis and Learning Leveraging Public Health Data. Retinal Vessel Segmentation based on Fully Convolutional Networks. Quantization-Based Regularization for Autoencoders. Feature Selective Anchor-Free Module for Single-Shot Object Detection. Additive Noise Annealing and …

Improving Document Classification with Multi-Sense Embeddings

Title Improving Document Classification with Multi-Sense Embeddings
Authors Vivek Gupta, Ankit Saw, Pegah Nokhiz, Harshit Gupta, Partha Talukdar
Abstract Efficient representation of text documents is an important building block in many NLP tasks. Research on long text categorization has shown that simple weighted averaging of word vectors for sentence representation often outperforms more sophisticated neural models. Recently proposed Sparse Composite Document Vector (SCDV) (Mekala et. al, 2017) extends this approach from sentences to documents using soft clustering over word vectors. However, SCDV disregards the multi-sense nature of words, and it also suffers from the curse of higher dimensionality. In this work, we address these shortcomings and propose SCDV-MS. SCDV-MS utilizes multi-sense word embeddings and learns a lower dimensional manifold. Through extensive experiments on multiple real-world datasets, we show that SCDV-MS embeddings outperform previous state-of-the-art embeddings on multi-class and multi-label text categorization tasks. Furthermore, SCDV-MS embeddings are more efficient than SCDV in terms of time and space complexity on textual classification tasks.
Tasks Document Classification, Text Categorization, Word Embeddings
Published 2019-11-18
URL https://arxiv.org/abs/1911.07918v1
PDF https://arxiv.org/pdf/1911.07918v1.pdf
PWC https://paperswithcode.com/paper/improving-document-classification-with-multi
Repo https://github.com/vgupta123/SCDV-MS
Framework none

Cost-Sensitive Diagnosis and Learning Leveraging Public Health Data

Title Cost-Sensitive Diagnosis and Learning Leveraging Public Health Data
Authors Mohammad Kachuee, Kimmo Karkkainen, Orpaz Goldstein, Davina Zamanzadeh, Majid Sarrafzadeh
Abstract Traditionally, machine learning algorithms rely on the assumption that all features of a given dataset are available for free. However, there are many concerns such as monetary data collection costs, patient discomfort in medical procedures, and privacy impacts of data collection that require careful consideration in any real-world health analytics system. An efficient solution would only acquire a subset of features based on the value it provides while considering acquisition costs. Moreover, datasets that provide feature costs are very limited, especially in healthcare. In this paper, we provide a health dataset as well as a method for assigning feature costs based on the total level of inconvenience asking for each feature entails. Furthermore, based on the suggested dataset, we provide a comparison of recent and state-of-the-art approaches to cost-sensitive feature acquisition and learning. Specifically, we analyze the performance of major sensitivity-based and reinforcement learning based methods in the literature on three different problems in the health domain, including diabetes, heart disease, and hypertension classification.
Tasks
Published 2019-02-19
URL https://arxiv.org/abs/1902.07102v2
PDF https://arxiv.org/pdf/1902.07102v2.pdf
PWC https://paperswithcode.com/paper/nutrition-and-health-data-for-cost-sensitive
Repo https://github.com/mkachuee/Opportunistic
Framework pytorch

Retinal Vessel Segmentation based on Fully Convolutional Networks

Title Retinal Vessel Segmentation based on Fully Convolutional Networks
Authors Zhengyuan Liu
Abstract The morphological attributes of retinal vessels, such as length, width, tortuosity and branching pattern and angles, play an important role in diagnosis, screening, treatment, and evaluation of various cardiovascular and ophthalmologic diseases such as diabetes, hypertension and arteriosclerosis. The crucial step before extracting these morphological characteristics of retinal vessels from retinal fundus images is vessel segmentation. In this work, we propose a method for retinal vessel segmentation based on fully convolutional networks. Thousands of patches are extracted from each retinal image and then fed into the network, and data argumentation is applied by rotating extracted patches. Two architectures of fully convolutional networks, U-Net and LadderNet, are used for vessel segmentation. The performance of our method is evaluated on three public datasets: DRIVE, STARE, and CHASE_DB1. Experimental results of our method show superior performance compared to recent state-of-the-art methods.
Tasks Retinal Vessel Segmentation
Published 2019-11-22
URL https://arxiv.org/abs/1911.09915v1
PDF https://arxiv.org/pdf/1911.09915v1.pdf
PWC https://paperswithcode.com/paper/retinal-vessel-segmentation-based-on-fully-1
Repo https://github.com/americofmoliveira/VesselSegmentation_ESWA
Framework none

Quantization-Based Regularization for Autoencoders

Title Quantization-Based Regularization for Autoencoders
Authors Hanwei Wu, Markus Flierl
Abstract Autoencoders and their variations provide unsupervised models for learning low-dimensional representations for downstream tasks. Without proper regularization, autoencoder models are susceptible to the overfitting problem and the so-called posterior collapse phenomenon. In this paper, we introduce a quantization-based regularizer in the bottleneck stage of autoencoder models to learn meaningful latent representations. We combine both perspectives of Vector Quantized-Variational AutoEncoders (VQ-VAE) and classical denoising regularization methods of neural networks. We interpret quantizers as regularizers that constrain latent representations while fostering a similarity-preserving mapping at the encoder. Before quantization, we impose noise on the latent codes and use a Bayesian estimator to optimize the quantizer-based representation. The introduced bottleneck Bayesian estimator outputs the posterior mean of the centroids to the decoder, and thus, is performing soft quantization of the noisy latent codes. We show that our proposed regularization method results in improved latent representations for both supervised learning and clustering downstream tasks when compared to autoencoders using other bottleneck structures.
Tasks Denoising, Quantization
Published 2019-05-27
URL https://arxiv.org/abs/1905.11062v2
PDF https://arxiv.org/pdf/1905.11062v2.pdf
PWC https://paperswithcode.com/paper/quantization-based-regularization-for
Repo https://github.com/AlbertOh90/Soft-VQ-VAE
Framework none

Feature Selective Anchor-Free Module for Single-Shot Object Detection

Title Feature Selective Anchor-Free Module for Single-Shot Object Detection
Authors Chenchen Zhu, Yihui He, Marios Savvides
Abstract We motivate and present feature selective anchor-free (FSAF) module, a simple and effective building block for single-shot object detectors. It can be plugged into single-shot detectors with feature pyramid structure. The FSAF module addresses two limitations brought up by the conventional anchor-based detection: 1) heuristic-guided feature selection; 2) overlap-based anchor sampling. The general concept of the FSAF module is online feature selection applied to the training of multi-level anchor-free branches. Specifically, an anchor-free branch is attached to each level of the feature pyramid, allowing box encoding and decoding in the anchor-free manner at an arbitrary level. During training, we dynamically assign each instance to the most suitable feature level. At the time of inference, the FSAF module can work jointly with anchor-based branches by outputting predictions in parallel. We instantiate this concept with simple implementations of anchor-free branches and online feature selection strategy. Experimental results on the COCO detection track show that our FSAF module performs better than anchor-based counterparts while being faster. When working jointly with anchor-based branches, the FSAF module robustly improves the baseline RetinaNet by a large margin under various settings, while introducing nearly free inference overhead. And the resulting best model can achieve a state-of-the-art 44.6% mAP, outperforming all existing single-shot detectors on COCO.
Tasks Feature Selection, Object Detection
Published 2019-03-02
URL http://arxiv.org/abs/1903.00621v1
PDF http://arxiv.org/pdf/1903.00621v1.pdf
PWC https://paperswithcode.com/paper/feature-selective-anchor-free-module-for
Repo https://github.com/xuannianz/FSAF
Framework tf

Additive Noise Annealing and Approximation Properties of Quantized Neural Networks

Title Additive Noise Annealing and Approximation Properties of Quantized Neural Networks
Authors Matteo Spallanzani, Lukas Cavigelli, Gian Paolo Leonardi, Marko Bertogna, Luca Benini
Abstract We present a theoretical and experimental investigation of the quantization problem for artificial neural networks. We provide a mathematical definition of quantized neural networks and analyze their approximation capabilities, showing in particular that any Lipschitz-continuous map defined on a hypercube can be uniformly approximated by a quantized neural network. We then focus on the regularization effect of additive noise on the arguments of multi-step functions inherent to the quantization of continuous variables. In particular, when the expectation operator is applied to a non-differentiable multi-step random function, and if the underlying probability density is differentiable (in either classical or weak sense), then a differentiable function is retrieved, with explicit bounds on its Lipschitz constant. Based on these results, we propose a novel gradient-based training algorithm for quantized neural networks that generalizes the straight-through estimator, acting on noise applied to the network’s parameters. We evaluate our algorithm on the CIFAR-10 and ImageNet image classification benchmarks, showing state-of-the-art performance on AlexNet and MobileNetV2 for ternary networks.
Tasks Image Classification, Quantization
Published 2019-05-24
URL https://arxiv.org/abs/1905.10452v1
PDF https://arxiv.org/pdf/1905.10452v1.pdf
PWC https://paperswithcode.com/paper/additive-noise-annealing-and-approximation
Repo https://github.com/spallanzanimatteo/QuantLab
Framework pytorch

Cross-Lingual Dependency Parsing Using Code-Mixed TreeBank

Title Cross-Lingual Dependency Parsing Using Code-Mixed TreeBank
Authors Zhang Meishan, Zhang Yue, Fu Guohong
Abstract Treebank translation is a promising method for cross-lingual transfer of syntactic dependency knowledge. The basic idea is to map dependency arcs from a source treebank to its target translation according to word alignments. This method, however, can suffer from imperfect alignment between source and target words. To address this problem, we investigate syntactic transfer by code mixing, translating only confident words in a source treebank. Cross-lingual word embeddings are leveraged for transferring syntactic knowledge to the target from the resulting code-mixed treebank. Experiments on University Dependency Treebanks show that code-mixed treebanks are more effective than translated treebanks, giving highly competitive performances among cross-lingual parsing methods.
Tasks Cross-Lingual Transfer, Dependency Parsing, Word Embeddings
Published 2019-09-05
URL https://arxiv.org/abs/1909.02235v1
PDF https://arxiv.org/pdf/1909.02235v1.pdf
PWC https://paperswithcode.com/paper/cross-lingual-dependency-parsing-using-code
Repo https://github.com/zhangmeishan/CodeMixedTreebank
Framework none

Adversarially Robust Few-Shot Learning: A Meta-Learning Approach

Title Adversarially Robust Few-Shot Learning: A Meta-Learning Approach
Authors Micah Goldblum, Liam Fowl, Tom Goldstein
Abstract Previous work on adversarially robust neural networks for image classification requires large training sets and computationally expensive training procedures. On the other hand, few-shot learning methods are highly vulnerable to adversarial examples. The goal of our work is to produce networks which both perform well at few-shot classification tasks and are simultaneously robust to adversarial examples. We develop an algorithm for producing adversarially robust meta-learners, and we thoroughly investigate factors which contribute to adversarial vulnerability. Moreover, our method achieves far superior robust performance on few-shot image classification tasks, such as Mini-ImageNet and CIFAR-FS, than robust transfer learning.
Tasks Few-Shot Image Classification, Few-Shot Learning, Image Classification, Meta-Learning, Transfer Learning
Published 2019-10-02
URL https://arxiv.org/abs/1910.00982v2
PDF https://arxiv.org/pdf/1910.00982v2.pdf
PWC https://paperswithcode.com/paper/robust-few-shot-learning-with-adversarially-1
Repo https://github.com/goldblum/AdversarialQuerying
Framework pytorch

The Art of Food: Meal Image Synthesis from Ingredients

Title The Art of Food: Meal Image Synthesis from Ingredients
Authors Fangda Han, Ricardo Guerrero, Vladimir Pavlovic
Abstract In this work we propose a new computational framework, based on generative deep models, for synthesis of photo-realistic food meal images from textual descriptions of its ingredients. Previous works on synthesis of images from text typically rely on pre-trained text models to extract text features, followed by a generative neural networks (GANs) aimed to generate realistic images conditioned on the text features. These works mainly focus on generating spatially compact and well-defined categories of objects, such as birds or flowers. In contrast, meal images are significantly more complex, consisting of multiple ingredients whose appearance and spatial qualities are further modified by cooking methods. We propose a method that first builds an attention-based ingredients-image association model, which is then used to condition a generative neural network tasked with synthesizing meal images. Furthermore, a cycle-consistent constraint is added to further improve image quality and control appearance. Extensive experiments show our model is able to generate meal image corresponding to the ingredients, which could be used to augment existing dataset for solving other computational food analysis problems.
Tasks Image Generation
Published 2019-05-09
URL https://arxiv.org/abs/1905.13149v1
PDF https://arxiv.org/pdf/1905.13149v1.pdf
PWC https://paperswithcode.com/paper/190513149
Repo https://github.com/CorneliusHsiao/FoodMethodGAN
Framework pytorch

Synthetic-Neuroscore: Using A Neuro-AI Interface for Evaluating Generative Adversarial Networks

Title Synthetic-Neuroscore: Using A Neuro-AI Interface for Evaluating Generative Adversarial Networks
Authors Zhengwei Wang, Qi She, Alan F. Smeaton, Tomas E. Ward, Graham Healy
Abstract Generative adversarial networks (GANs) are increasingly attracting attention in the computer vision, natural language processing, speech synthesis and similar domains. Arguably the most striking results have been in the area of image synthesis. However, evaluating the performance of GANs is still an open and challenging problem. Existing evaluation metrics primarily measure the dissimilarity between real and generated images using automated statistical methods. They often require large sample sizes for evaluation and do not directly reflect human perception of image quality. In this work, we describe an evaluation metric we call Neuroscore, for evaluating the performance of GANs, that more directly reflects psychoperceptual image quality through the utilization of brain signals. Our results show that Neuroscore has superior performance to the current evaluation metrics in that: (1) It is more consistent with human judgment; (2) The evaluation process needs much smaller numbers of samples; and (3) It is able to rank the quality of images on a per GAN basis. A convolutional neural network (CNN) based neuro-AI interface is proposed to predict Neuroscore from GAN-generated images directly without the need for neural responses. Importantly, we show that including neural responses during the training phase of the network can significantly improve the prediction capability of the proposed model. Materials related to this work are provided at https://github.com/villawang/Neuro-AI-Interface.
Tasks Image Generation, Speech Synthesis
Published 2019-05-10
URL https://arxiv.org/abs/1905.04243v2
PDF https://arxiv.org/pdf/1905.04243v2.pdf
PWC https://paperswithcode.com/paper/neuroscore-a-brain-inspired-evaluation-metric
Repo https://github.com/villawang/Neuro-AI-Interface
Framework tf

On the Accuracy of Influence Functions for Measuring Group Effects

Title On the Accuracy of Influence Functions for Measuring Group Effects
Authors Pang Wei Koh, Kai-Siang Ang, Hubert H. K. Teo, Percy Liang
Abstract Influence functions estimate the effect of removing a training point on a model without the need to retrain. They are based on a first-order Taylor approximation that is guaranteed to be accurate for sufficiently small changes to the model, and so are commonly used to study the effect of individual points in large datasets. However, we often want to study the effects of large groups of training points, e.g., to diagnose batch effects or apportion credit between different data sources. Removing such large groups can result in significant changes to the model. Are influence functions still accurate in this setting? In this paper, we find that across many different types of groups and for a range of real-world datasets, the predicted effect (using influence functions) of a group correlates surprisingly well with its actual effect, even if the absolute and relative errors are large. Our theoretical analysis shows that such strong correlation arises only under certain settings and need not hold in general, indicating that real-world datasets have particular properties that allow the influence approximation to be accurate.
Tasks
Published 2019-05-30
URL https://arxiv.org/abs/1905.13289v2
PDF https://arxiv.org/pdf/1905.13289v2.pdf
PWC https://paperswithcode.com/paper/on-the-accuracy-of-influence-functions-for
Repo https://github.com/kohpangwei/group-influence-release
Framework tf

Extrapolating paths with graph neural networks

Title Extrapolating paths with graph neural networks
Authors Jean-Baptiste Cordonnier, Andreas Loukas
Abstract We consider the problem of path inference: given a path prefix, i.e., a partially observed sequence of nodes in a graph, we want to predict which nodes are in the missing suffix. In particular, we focus on natural paths occurring as a by-product of the interaction of an agent with a network—a driver on the transportation network, an information seeker in Wikipedia, or a client in an online shop. Our interest is sparked by the realization that, in contrast to shortest-path problems, natural paths are usually not optimal in any graph-theoretic sense, but might still follow predictable patterns. Our main contribution is a graph neural network called Gretel. Conditioned on a path prefix, this network can efficiently extrapolate path suffixes, evaluate path likelihood, and sample from the future path distribution. Our experiments with GPS traces on a road network and user-navigation paths in Wikipedia confirm that Gretel is able to adapt to graphs with very different properties, while also comparing favorably to previous solutions.
Tasks
Published 2019-03-18
URL http://arxiv.org/abs/1903.07518v1
PDF http://arxiv.org/pdf/1903.07518v1.pdf
PWC https://paperswithcode.com/paper/extrapolating-paths-with-graph-neural
Repo https://github.com/jbcdnr/gretel-path-extrapolation
Framework pytorch

Blind Universal Bayesian Image Denoising with Gaussian Noise Level Learning

Title Blind Universal Bayesian Image Denoising with Gaussian Noise Level Learning
Authors Majed El Helou, Sabine Süsstrunk
Abstract Blind and universal image denoising consists of using a unique model that denoises images with any level of noise. It is especially practical as noise levels do not need to be known when the model is developed or at test time. We propose a theoretically-grounded blind and universal deep learning image denoiser for additive Gaussian noise removal. Our network is based on an optimal denoising solution, which we call fusion denoising. It is derived theoretically with a Gaussian image prior assumption. Synthetic experiments show our network’s generalization strength to unseen additive noise levels. We also adapt the fusion denoising network architecture for image denoising on real images. Our approach improves real-world grayscale additive image denoising PSNR results for training noise levels and further on noise levels not seen during training. It also improves state-of-the-art color image denoising performance on every single noise level, by an average of 0.1dB, whether trained on or not.
Tasks Denoising, Image Denoising
Published 2019-07-05
URL https://arxiv.org/abs/1907.03029v2
PDF https://arxiv.org/pdf/1907.03029v2.pdf
PWC https://paperswithcode.com/paper/blind-universal-bayesian-image-denoising-with
Repo https://github.com/IVRL/BUIFD
Framework pytorch

Check-It: A Plugin for Detecting and Reducing the Spread of Fake News and Misinformation on the Web

Title Check-It: A Plugin for Detecting and Reducing the Spread of Fake News and Misinformation on the Web
Authors Demetris Paschalides, Alexandros Kornilakis, Chrysovalantis Christodoulou, Rafael Andreou, George Pallis, Marios D. Dikaiakos, Evangelos Markatos
Abstract Over the past few years, we have been witnessing the rise of misinformation on the Web. People fall victims of fake news during their daily lives and assist their further propagation knowingly and inadvertently. There have been many initiatives that are trying to mitigate the damage caused by fake news, focusing on signals from either domain flag-lists, online social networks or artificial intelligence. In this work, we present Check-It, a system that combines, in an intelligent way, a variety of signals into a pipeline for fake news identification. Check-It is developed as a web browser plugin with the objective of efficient and timely fake news detection, respecting the user’s privacy. Experimental results show that Check-It is able to outperform the state-of-the-art methods. On a dataset, consisting of 9 millions of articles labeled as fake and real, Check-It obtains classification accuracies that exceed 99%.
Tasks Fake News Detection
Published 2019-05-10
URL https://arxiv.org/abs/1905.04260v1
PDF https://arxiv.org/pdf/1905.04260v1.pdf
PWC https://paperswithcode.com/paper/check-it-a-plugin-for-detecting-and-reducing
Repo https://github.com/anguyen120/fake-news-in-time
Framework none

OmniFold: A Method to Simultaneously Unfold All Observables

Title OmniFold: A Method to Simultaneously Unfold All Observables
Authors Anders Andreassen, Patrick T. Komiske, Eric M. Metodiev, Benjamin Nachman, Jesse Thaler
Abstract Collider data must be corrected for detector effects (“unfolded”) to be compared with theoretical calculations and measurements from other experiments. Unfolding is traditionally done for individual, binned observables without including all information relevant for characterizing the detector response. We introduce OmniFold, an unfolding method that iteratively reweights a simulated dataset, using machine learning to capitalize on all available information. Our approach is unbinned, works for arbitrarily high-dimensional data, and naturally incorporates information from the full phase space. We illustrate this technique on a realistic jet substructure example from the Large Hadron Collider and compare it to standard binned unfolding methods. This new paradigm enables the simultaneous measurement of all observables, including those not yet invented at the time of the analysis.
Tasks
Published 2019-11-20
URL https://arxiv.org/abs/1911.09107v1
PDF https://arxiv.org/pdf/1911.09107v1.pdf
PWC https://paperswithcode.com/paper/omnifold-a-method-to-simultaneously-unfold
Repo https://github.com/ericmetodiev/OmniFold
Framework none
comments powered by Disqus