January 28, 2020

3003 words 15 mins read

Paper Group ANR 779

Paper Group ANR 779

Blind Motion Deblurring with Cycle Generative Adversarial Networks. Variational approach to unsupervised learning. A Collaborative Approach using Ridge-Valley Minutiae for More Accurate Contactless Fingerprint Identification. Making a Case for Social Media Corpus for Detecting Depression. Rapid Learning or Feature Reuse? Towards Understanding the E …

Blind Motion Deblurring with Cycle Generative Adversarial Networks

Title Blind Motion Deblurring with Cycle Generative Adversarial Networks
Authors Quan Yuan, Junxia Li, Lingwei Zhang, Zhefu Wu, Guangyu Liu
Abstract Blind motion deblurring is one of the most basic and challenging problems in image processing and computer vision. It aims to recover a sharp image from its blurred version knowing nothing about the blur process. Many existing methods use Maximum A Posteriori (MAP) or Expectation Maximization (EM) frameworks to deal with this kind of problems, but they cannot handle well the figh frequency features of natural images. Most recently, deep neural networks have been emerging as a powerful tool for image deblurring. In this paper, we prove that encoder-decoder architecture gives better results for image deblurring tasks. In addition, we propose a novel end-to-end learning model which refines generative adversarial network by many novel training strategies so as to tackle the problem of deblurring. Experimental results show that our model can capture high frequency features well, and the results on benchmark dataset show that proposed model achieves the competitive performance.
Tasks Deblurring
Published 2019-01-07
URL http://arxiv.org/abs/1901.01641v2
PDF http://arxiv.org/pdf/1901.01641v2.pdf
PWC https://paperswithcode.com/paper/blind-motion-deblurring-with-cycle-generative
Repo
Framework

Variational approach to unsupervised learning

Title Variational approach to unsupervised learning
Authors Swapnil Nitin Shah
Abstract Deep belief networks are used extensively for unsupervised stochastic learning on large datasets. Compared to other deep learning approaches their layer-by-layer learning makes them highly scalable. Unfortunately, the principles by which they achieve efficient learning are not well understood. Numerous attempts have been made to explain their efficiency and applicability to a wide class of learning problems in terms of principles drawn from cognitive psychology, statistics, information theory, and more recently physics, but quite often these imported principles lack strong scientific foundation. Here we demonstrate how one can arrive at convolutional deep belief networks as potential solution to unsupervised learning problems without making assumptions about the underlying framework. To do this, we exploit the notion of symmetry that is fundamental in machine learning, physics and other fields, utilizing the particular form of the functional renormalization group in physics.
Tasks
Published 2019-04-24
URL http://arxiv.org/abs/1904.10869v1
PDF http://arxiv.org/pdf/1904.10869v1.pdf
PWC https://paperswithcode.com/paper/variational-approach-to-unsupervised-learning
Repo
Framework

A Collaborative Approach using Ridge-Valley Minutiae for More Accurate Contactless Fingerprint Identification

Title A Collaborative Approach using Ridge-Valley Minutiae for More Accurate Contactless Fingerprint Identification
Authors Ritesh Vyas, Ajay Kumar
Abstract Contactless fingerprint identification has emerged as an reliable and user friendly alternative for the personal identification in a range of e-business and law-enforcement applications. It is however quite known from the literature that the contactless fingerprint images deliver remarkably low matching accuracies as compared with those obtained from the contact-based fingerprint sensors. This paper develops a new approach to significantly improve contactless fingerprint matching capabilities available today. We systematically analyze the extent of complimentary ridge-valley information and introduce new approaches to achieve significantly higher matching accuracy over state-of-art fingerprint matchers commonly employed today. We also investigate least explored options for the fingerprint color-space conversions, which can play a key-role for more accurate contactless fingerprint matching. This paper presents experimental results from different publicly available contactless fingerprint databases using NBIS, MCC and COTS matchers. Our consistently outperforming results validate the effectiveness of the proposed approach for more accurate contactless fingerprint identification.
Tasks
Published 2019-09-13
URL https://arxiv.org/abs/1909.06045v2
PDF https://arxiv.org/pdf/1909.06045v2.pdf
PWC https://paperswithcode.com/paper/a-collaborative-approach-using-ridge-valley
Repo
Framework

Making a Case for Social Media Corpus for Detecting Depression

Title Making a Case for Social Media Corpus for Detecting Depression
Authors Adil Rajput, Samara Ahmed
Abstract The social media platform provides an opportunity to gain valuable insights into user behaviour. Users mimic their internal feelings and emotions in a disinhibited fashion using natural language. Techniques in Natural Language Processing have helped researchers decipher standard documents and cull together inferences from massive amount of data. A representative corpus is a prerequisite for NLP and one of the challenges we face today is the non-standard and noisy language that exists on the internet. Our work focuses on building a corpus from social media that is focused on detecting mental illness. We use depression as a case study and demonstrate the effectiveness of using such a corpus for helping practitioners detect such cases. Our results show a high correlation between our Social Media Corpus and the standard corpus for depression.
Tasks
Published 2019-02-02
URL http://arxiv.org/abs/1902.00702v1
PDF http://arxiv.org/pdf/1902.00702v1.pdf
PWC https://paperswithcode.com/paper/making-a-case-for-social-media-corpus-for
Repo
Framework

Rapid Learning or Feature Reuse? Towards Understanding the Effectiveness of MAML

Title Rapid Learning or Feature Reuse? Towards Understanding the Effectiveness of MAML
Authors Aniruddh Raghu, Maithra Raghu, Samy Bengio, Oriol Vinyals
Abstract An important research direction in machine learning has centered around developing meta-learning algorithms to tackle few-shot learning. An especially successful algorithm has been Model Agnostic Meta-Learning (MAML), a method that consists of two optimization loops, with the outer loop finding a meta-initialization, from which the inner loop can efficiently learn new tasks. Despite MAML’s popularity, a fundamental open question remains – is the effectiveness of MAML due to the meta-initialization being primed for rapid learning (large, efficient changes in the representations) or due to feature reuse, with the meta initialization already containing high quality features? We investigate this question, via ablation studies and analysis of the latent representations, finding that feature reuse is the dominant factor. This leads to the ANIL (Almost No Inner Loop) algorithm, a simplification of MAML where we remove the inner loop for all but the (task-specific) head of a MAML-trained network. ANIL matches MAML’s performance on benchmark few-shot image classification and RL and offers computational improvements over MAML. We further study the precise contributions of the head and body of the network, showing that performance on the test tasks is entirely determined by the quality of the learned features, and we can remove even the head of the network (the NIL algorithm). We conclude with a discussion of the rapid learning vs feature reuse question for meta-learning algorithms more broadly.
Tasks Few-Shot Image Classification, Few-Shot Learning, Image Classification, Meta-Learning
Published 2019-09-19
URL https://arxiv.org/abs/1909.09157v2
PDF https://arxiv.org/pdf/1909.09157v2.pdf
PWC https://paperswithcode.com/paper/rapid-learning-or-feature-reuse-towards
Repo
Framework

Investigating Constraint Programming and Hybrid Methods for Real World Industrial Test Laboratory Scheduling

Title Investigating Constraint Programming and Hybrid Methods for Real World Industrial Test Laboratory Scheduling
Authors Tobias Geibinger, Florian Mischek, Nysret Musliu
Abstract In this paper we deal with a complex real world scheduling problem closely related to the well-known Resource-Constrained Project Scheduling Problem (RCPSP). The problem concerns industrial test laboratories in which a large number of tests has to be performed by qualified personnel using specialised equipment, while respecting deadlines and other constraints. We present different constraint programming models and search strategies for this problem. Furthermore, we propose a Very Large Neighborhood Search approach based on our CP methods. Our models are evaluated using CP solvers and a MIP solver both on real-world test laboratory data and on a set of generated instances of different sizes based on the real-world data. Further, we compare the exact approaches with VLNS and a Simulated Annealing heuristic. We could find feasible solutions for all instances and several optimal solutions and we show that using VLNS we can improve upon the results of the other approaches.
Tasks
Published 2019-11-12
URL https://arxiv.org/abs/1911.04766v1
PDF https://arxiv.org/pdf/1911.04766v1.pdf
PWC https://paperswithcode.com/paper/investigating-constraint-programming-and
Repo
Framework

Discriminative Sentence Modeling for Story Ending Prediction

Title Discriminative Sentence Modeling for Story Ending Prediction
Authors Yiming Cui, Wanxiang Che, Wei-Nan Zhang, Ting Liu, Shijin Wang, Guoping Hu
Abstract Story Ending Prediction is a task that needs to select an appropriate ending for the given story, which requires the machine to understand the story and sometimes needs commonsense knowledge. To tackle this task, we propose a new neural network called Diff-Net for better modeling the differences of each ending in this task. The proposed model could discriminate two endings in three semantic levels: contextual representation, story-aware representation, and discriminative representation. Experimental results on the Story Cloze Test dataset show that the proposed model siginificantly outperforms various systems by a large margin, and detailed ablation studies are given for better understanding our model. We also carefully examine the traditional and BERT-based models on both SCT v1.0 and v1.5 with interesting findings that may potentially help future studies.
Tasks
Published 2019-12-19
URL https://arxiv.org/abs/1912.09008v1
PDF https://arxiv.org/pdf/1912.09008v1.pdf
PWC https://paperswithcode.com/paper/discriminative-sentence-modeling-for-story
Repo
Framework

Selective Brain Damage: Measuring the Disparate Impact of Model Pruning

Title Selective Brain Damage: Measuring the Disparate Impact of Model Pruning
Authors Sara Hooker, Aaron Courville, Yann Dauphin, Andrea Frome
Abstract Neural network pruning techniques have demonstrated it is possible to remove the majority of weights in a network with surprisingly little degradation to test set accuracy. However, this measure of performance conceals significant differences in how different classes and images are impacted by pruning. We find that certain examples, which we term pruning identified exemplars (PIEs), and classes are systematically more impacted by the introduction of sparsity. Removing PIE images from the test-set greatly improves top-1 accuracy for both pruned and non-pruned models. These hard-to-generalize-to images tend to be mislabelled, of lower image quality, depict multiple objects or require fine-grained classification. These findings shed light on previously unknown trade-offs, and suggest that a high degree of caution should be exercised before pruning is used in sensitive domains.
Tasks Network Pruning
Published 2019-11-13
URL https://arxiv.org/abs/1911.05248v1
PDF https://arxiv.org/pdf/1911.05248v1.pdf
PWC https://paperswithcode.com/paper/selective-brain-damage-measuring-the
Repo
Framework
Title Detecting Target-Area Link-Flooding DDoS Attacks using Traffic Analysis and Supervised Learning
Authors Mostafa Rezazad, Matthias R. Brust, Mohammad Akbari, Pascal Bouvry, Ngai-Man Cheung
Abstract A novel class of extreme link-flooding DDoS (Distributed Denial of Service) attacks is designed to cut off entire geographical areas such as cities and even countries from the Internet by simultaneously targeting a selected set of network links. The Crossfire attack is a target-area link-flooding attack, which is orchestrated in three complex phases. The attack uses a massively distributed large-scale botnet to generate low-rate benign traffic aiming to congest selected network links, so-called target links. The adoption of benign traffic, while simultaneously targeting multiple network links, makes detecting the Crossfire attack a serious challenge. In this paper, we present analytical and emulated results showing hitherto unidentified vulnerabilities in the execution of the attack, such as a correlation between coordination of the botnet traffic and the quality of the attack, and a correlation between the attack distribution and detectability of the attack. Additionally, we identified a warm-up period due to the bot synchronization. For attack detection, we report results of using two supervised machine learning approaches: Support Vector Machine (SVM) and Random Forest (RF) for classification of network traffic to normal and abnormal traffic, i.e, attack traffic. These machine learning models have been trained in various scenarios using the link volume as the main feature set.
Tasks
Published 2019-03-01
URL http://arxiv.org/abs/1903.01550v1
PDF http://arxiv.org/pdf/1903.01550v1.pdf
PWC https://paperswithcode.com/paper/detecting-target-area-link-flooding-ddos
Repo
Framework

One-Pass Incomplete Multi-view Clustering

Title One-Pass Incomplete Multi-view Clustering
Authors Menglei Hu, Songcan Chen
Abstract Real data are often with multiple modalities or from multiple heterogeneous sources, thus forming so-called multi-view data, which receives more and more attentions in machine learning. Multi-view clustering (MVC) becomes its important paradigm. In real-world applications, some views often suffer from instances missing. Clustering on such multi-view datasets is called incomplete multi-view clustering (IMC) and quite challenging. To date, though many approaches have been developed, most of them are offline and have high computational and memory costs especially for large scale datasets. To address this problem, in this paper, we propose an One-Pass Incomplete Multi-view Clustering framework (OPIMC). With the help of regularized matrix factorization and weighted matrix factorization, OPIMC can relatively easily deal with such problem. Different from the existing and sole online IMC method, OPIMC can directly get clustering results and effectively determine the termination of iteration process by introducing two global statistics. Finally, extensive experiments conducted on four real datasets demonstrate the efficiency and effectiveness of the proposed OPIMC method.
Tasks
Published 2019-03-02
URL http://arxiv.org/abs/1903.00637v1
PDF http://arxiv.org/pdf/1903.00637v1.pdf
PWC https://paperswithcode.com/paper/one-pass-incomplete-multi-view-clustering
Repo
Framework

Self-Adaptive Network Pruning

Title Self-Adaptive Network Pruning
Authors Jinting Chen, Zhaocheng Zhu, Cheng Li, Yuming Zhao
Abstract Deep convolutional neural networks have been proved successful on a wide range of tasks, yet they are still hindered by their large computation cost in many industrial scenarios. In this paper, we propose to reduce such cost for CNNs through a self-adaptive network pruning method (SANP). Our method introduces a general Saliency-and-Pruning Module (SPM) for each convolutional layer, which learns to predict saliency scores and applies pruning for each channel. Given a total computation budget, SANP adaptively determines the pruning strategy with respect to each layer and each sample, such that the average computation cost meets the budget. This design allows SANP to be more efficient in computation, as well as more robust to datasets and backbones. Extensive experiments on 2 datasets and 3 backbones show that SANP surpasses state-of-the-art methods in both classification accuracy and pruning rate.
Tasks Network Pruning
Published 2019-10-20
URL https://arxiv.org/abs/1910.08906v1
PDF https://arxiv.org/pdf/1910.08906v1.pdf
PWC https://paperswithcode.com/paper/self-adaptive-network-pruning
Repo
Framework

Learning Matchable Image Transformations for Long-term Metric Visual Localization

Title Learning Matchable Image Transformations for Long-term Metric Visual Localization
Authors Lee Clement, Mona Gridseth, Justin Tomasi, Jonathan Kelly
Abstract Long-term metric self-localization is an essential capability of autonomous mobile robots, but remains challenging for vision-based systems due to appearance changes caused by lighting, weather, or seasonal variations. While experience-based mapping has proven to be an effective technique for bridging the `appearance gap,’ the number of experiences required for reliable metric localization over days or months can be very large, and methods for reducing the necessary number of experiences are needed for this approach to scale. Taking inspiration from color constancy theory, we learn a nonlinear RGB-to-grayscale mapping that explicitly maximizes the number of inlier feature matches for images captured under different lighting and weather conditions, and use it as a pre-processing step in a conventional single-experience localization pipeline to improve its robustness to appearance change. We train this mapping by approximating the target non-differentiable localization pipeline with a deep neural network, and find that incorporating a learned low-dimensional context feature can further improve cross-appearance feature matching. Using synthetic and real-world datasets, we demonstrate substantial improvements in localization performance across day-night cycles, enabling continuous metric localization over a 30-hour period using a single mapping experience, and allowing experience-based localization to scale to long deployments with dramatically reduced data requirements. |
Tasks Color Constancy, Visual Localization
Published 2019-04-01
URL https://arxiv.org/abs/1904.01080v4
PDF https://arxiv.org/pdf/1904.01080v4.pdf
PWC https://paperswithcode.com/paper/learning-matchable-colorspace-transformations
Repo
Framework

RASNet: Segmentation for Tracking Surgical Instruments in Surgical Videos Using Refined Attention Segmentation Network

Title RASNet: Segmentation for Tracking Surgical Instruments in Surgical Videos Using Refined Attention Segmentation Network
Authors Zhen-Liang Ni, Gui-Bin Bian, Xiao-Liang Xie, Zeng-Guang Hou, Xiao-Hu Zhou, Yan-Jie Zhou
Abstract Segmentation for tracking surgical instruments plays an important role in robot-assisted surgery. Segmentation of surgical instruments contributes to capturing accurate spatial information for tracking. In this paper, a novel network, Refined Attention Segmentation Network, is proposed to simultaneously segment surgical instruments and identify their categories. The U-shape network which is popular in segmentation is used. Different from previous work, an attention module is adopted to help the network focus on key regions, which can improve the segmentation accuracy. To solve the class imbalance problem, the weighted sum of the cross entropy loss and the logarithm of the Jaccard index is used as loss function. Furthermore, transfer learning is adopted in our network. The encoder is pre-trained on ImageNet. The dataset from the MICCAI EndoVis Challenge 2017 is used to evaluate our network. Based on this dataset, our network achieves state-of-the-art performance 94.65% mean Dice and 90.33% mean IOU.
Tasks Transfer Learning
Published 2019-05-21
URL https://arxiv.org/abs/1905.08663v2
PDF https://arxiv.org/pdf/1905.08663v2.pdf
PWC https://paperswithcode.com/paper/rasnet-segmentation-for-tracking-surgical
Repo
Framework

Probabilistic Multilayer Regularization Network for Unsupervised 3D Brain Image Registration

Title Probabilistic Multilayer Regularization Network for Unsupervised 3D Brain Image Registration
Authors Lihao Liu, Xiaowei Hu, Lei Zhu, Pheng-Ann Heng
Abstract Brain image registration transforms a pair of images into one system with the matched imaging contents, which is of essential importance for brain image analysis. This paper presents a novel framework for unsupervised 3D brain image registration by capturing the feature-level transformation relationships between the unaligned image and reference image. To achieve this, we develop a feature-level probabilistic model to provide the direct regularization to the hidden layers of two deep convolutional neural networks, which are constructed from two input images. This model design is developed into multiple layers of these two networks to capture the transformation relationships at different levels. We employ two common benchmark datasets for 3D brain image registration and perform various experiments to evaluate our method. Experimental results show that our method clearly outperforms state-of-the-art methods on both benchmark datasets by a large margin.
Tasks Image Registration
Published 2019-07-03
URL https://arxiv.org/abs/1907.01922v1
PDF https://arxiv.org/pdf/1907.01922v1.pdf
PWC https://paperswithcode.com/paper/probabilistic-multilayer-regularization
Repo
Framework

Identity-Free Facial Expression Recognition using conditional Generative Adversarial Network

Title Identity-Free Facial Expression Recognition using conditional Generative Adversarial Network
Authors Jie Cai, Zibo Meng, Ahmed Shehab Khan, Zhiyuan Li, James O’Reilly, Yan Tong
Abstract In this paper, we proposed a novel Identity-free conditional Generative Adversarial Network (IF-GAN) to explicitly reduce inter-subject variations for facial expression recognition. Specifically, for any given input face image, a conditional generative model was developed to transform an average neutral face, which is calculated from various subjects showing neutral expressions, to an average expressive face with the same expression as the input image. Since the transformed images have the same synthetic “average” identity, they differ from each other by only their expressions and thus, can be used for identity-free expression classification. In this work, an end-to-end system was developed to perform expression transformation and expression recognition in the IF-GAN framework. Experimental results on three facial expression datasets have demonstrated that the proposed IF-GAN outperforms the baseline CNN model and achieves comparable or better performance compared with the state-of-the-art methods for facial expression recognition.
Tasks Facial Expression Recognition
Published 2019-03-19
URL http://arxiv.org/abs/1903.08051v1
PDF http://arxiv.org/pdf/1903.08051v1.pdf
PWC https://paperswithcode.com/paper/identity-free-facial-expression-recognition
Repo
Framework
comments powered by Disqus