January 27, 2020

3096 words 15 mins read

Paper Group ANR 1231

Paper Group ANR 1231

Abnormal Client Behavior Detection in Federated Learning. Interrupted and cascaded permutation invariant training for speech separation. Mixup-breakdown: a consistency training method for improving generalization of speech separation models. Robust Deep Neural Networks Inspired by Fuzzy Logic. A Reinforcement Learning Approach for the Multichannel …

Abnormal Client Behavior Detection in Federated Learning

Title Abnormal Client Behavior Detection in Federated Learning
Authors Suyi Li, Yong Cheng, Yang Liu, Wei Wang, Tianjian Chen
Abstract In federated learning systems, clients are autonomous in that their behaviors are not fully governed by the server. Consequently, a client may intentionally or unintentionally deviate from the prescribed course of federated model training, resulting in abnormal behaviors, such as turning into a malicious attacker or a malfunctioning client. Timely detecting those anomalous clients is therefore critical to minimize their adverse impacts. In this work, we propose to detect anomalous clients at the server side. In particular, we generate low-dimensional surrogates of model weight vectors and use them to perform anomaly detection. We evaluate our solution through experiments on image classification model training over the FEMNIST dataset. Experimental results show that the proposed detection-based approach significantly outperforms the conventional defense-based methods.
Tasks Anomaly Detection, Image Classification
Published 2019-10-22
URL https://arxiv.org/abs/1910.09933v2
PDF https://arxiv.org/pdf/1910.09933v2.pdf
PWC https://paperswithcode.com/paper/abnormal-client-behavior-detection-in
Repo
Framework

Interrupted and cascaded permutation invariant training for speech separation

Title Interrupted and cascaded permutation invariant training for speech separation
Authors Gene-Ping Yang, Szu-Lin Wu, Yao-Wen Mao, Hung-yi Lee, Lin-shan Lee
Abstract Permutation Invariant Training (PIT) has long been a stepping stone method for training speech separation model in handling the label ambiguity problem. With PIT selecting the minimum cost label assignments dynamically, very few studies considered the separation problem to be optimizing both the model parameters and the label assignments, but focused on searching for good model architecture and parameters. In this paper, we investigate instead for a given model architecture the various flexible label assignment strategies for training the model, rather than directly using PIT. Surprisingly, we discover a significant performance boost compared to PIT is possible if the model is trained with fixed label assignments and a good set of labels is chosen. With fixed label training cascaded between two sections of PIT, we achieved the state-of-the-art performance on WSJ0-2mix without changing the model architecture at all.
Tasks Speech Separation
Published 2019-10-28
URL https://arxiv.org/abs/1910.12706v1
PDF https://arxiv.org/pdf/1910.12706v1.pdf
PWC https://paperswithcode.com/paper/interrupted-and-cascaded-permutation
Repo
Framework

Mixup-breakdown: a consistency training method for improving generalization of speech separation models

Title Mixup-breakdown: a consistency training method for improving generalization of speech separation models
Authors Max W. Y. Lam, Jun Wang, Dan Su, Dong Yu
Abstract Deep-learning based speech separation models confront poor generalization problem that even the state-of-the-art models could abruptly fail when evaluating them in mismatch conditions. To address this problem, we propose an easy-to-implement yet effective consistency based semi-supervised learning (SSL) approach, namely Mixup-Breakdown training (MBT). It learns a teacher model to “breakdown” unlabeled inputs, and the estimated separations are interpolated to produce more useful pseudo “mixup” input-output pairs, on which the consistency regularization could apply for learning a student model. In our experiment, we evaluate MBT under various conditions with ascending degrees of mismatch, including unseen interfering speech, noise, and music, and compare MBT’s generalization capability against state-of-the-art supervised learning and SSL approaches. The result indicates that MBT significantly outperforms several strong baselines with up to 13.77% relative SI-SNRi improvement. Moreover, MBT only adds negligible computational overhead to standard training schemes.
Tasks Speech Separation
Published 2019-10-28
URL https://arxiv.org/abs/1910.13253v3
PDF https://arxiv.org/pdf/1910.13253v3.pdf
PWC https://paperswithcode.com/paper/mixup-breakdown-a-consistency-training-method
Repo
Framework

Robust Deep Neural Networks Inspired by Fuzzy Logic

Title Robust Deep Neural Networks Inspired by Fuzzy Logic
Authors Minh Le
Abstract Deep neural networks have achieved impressive performance and become the de-facto standard in many tasks. However, troubling phenomena such as adversarial and fooling examples suggest that the generalization they make is flawed. I argue that among the roots of the phenomena are two geometric properties of common deep learning architectures: their distributed nature and the connectedness of their decision regions. As a remedy, I propose new architectures inspired by fuzzy logic that combine several alternative design elements. Through experiments on MNIST and CIFAR-10, the new models are shown to be more local, better at rejecting noise samples, and more robust against adversarial examples. Ablation analyses reveal behaviors on adversarial examples that cannot be explained by the linearity hypothesis but are consistent with the hypothesis that logic-inspired traits create more robust models.
Tasks
Published 2019-11-20
URL https://arxiv.org/abs/1911.08635v2
PDF https://arxiv.org/pdf/1911.08635v2.pdf
PWC https://paperswithcode.com/paper/logic-inspired-deep-neural-networks
Repo
Framework

A Reinforcement Learning Approach for the Multichannel Rendezvous Problem

Title A Reinforcement Learning Approach for the Multichannel Rendezvous Problem
Authors Jen-Hung Wang, Ping-En Lu, Cheng-Shang Chang, Duan-Shin Lee
Abstract In this paper, we consider the multichannel rendezvous problem in cognitive radio networks (CRNs) where the probability that two users hopping on the same channel have a successful rendezvous is a function of channel states. The channel states are modelled by two-state Markov chains that have a good state and a bad state. These channel states are not observable by the users. For such a multichannel rendezvous problem, we are interested in finding the optimal policy to minimize the expected time-to-rendezvous (ETTR) among the class of {\em dynamic blind rendezvous policies}, i.e., at the $t^{th}$ time slot each user selects channel $i$ independently with probability $p_i(t)$, $i=1,2, \ldots, N$. By formulating such a multichannel rendezvous problem as an adversarial bandit problem, we propose using a reinforcement learning approach to learn the channel selection probabilities $p_i(t)$, $i=1,2, \ldots, N$. Our experimental results show that the reinforcement learning approach is very effective and yields comparable ETTRs when comparing to various approximation policies in the literature.
Tasks
Published 2019-07-02
URL https://arxiv.org/abs/1907.01919v2
PDF https://arxiv.org/pdf/1907.01919v2.pdf
PWC https://paperswithcode.com/paper/a-reinforcement-learning-approach-for-the
Repo
Framework

Conditionally Learn to Pay Attention for Sequential Visual Task

Title Conditionally Learn to Pay Attention for Sequential Visual Task
Authors Jun He, Quan-Jie Cao, Lei Zhang
Abstract Sequential visual task usually requires to pay attention to its current interested object conditional on its previous observations. Different from popular soft attention mechanism, we propose a new attention framework by introducing a novel conditional global feature which represents the weak feature descriptor of the current focused object. Specifically, for a standard CNN (Convolutional Neural Network) pipeline, the convolutional layers with different receptive fields are used to produce the attention maps by measuring how the convolutional features align to the conditional global feature. The conditional global feature can be generated by different recurrent structure according to different visual tasks, such as a simple recurrent neural network for multiple objects recognition, or a moderate complex language model for image caption. Experiments show that our proposed conditional attention model achieves the best performance on the SVHN (Street View House Numbers) dataset with / without extra bounding box; and for image caption, our attention model generates better scores than the popular soft attention model.
Tasks Language Modelling
Published 2019-11-11
URL https://arxiv.org/abs/1911.04365v1
PDF https://arxiv.org/pdf/1911.04365v1.pdf
PWC https://paperswithcode.com/paper/conditionally-learn-to-pay-attention-for
Repo
Framework

COCO_TS Dataset: Pixel-level Annotations Based on Weak Supervision for Scene Text Segmentation

Title COCO_TS Dataset: Pixel-level Annotations Based on Weak Supervision for Scene Text Segmentation
Authors Simone Bonechi, Paolo Andreini, Monica Bianchini, Franco Scarselli
Abstract The absence of large scale datasets with pixel-level supervisions is a significant obstacle for the training of deep convolutional networks for scene text segmentation. For this reason, synthetic data generation is normally employed to enlarge the training dataset. Nonetheless, synthetic data cannot reproduce the complexity and variability of natural images. In this paper, a weakly supervised learning approach is used to reduce the shift between training on real and synthetic data. Pixel-level supervisions for a text detection dataset (i.e. where only bounding-box annotations are available) are generated. In particular, the COCO-Text-Segmentation (COCO_TS) dataset, which provides pixel-level supervisions for the COCO-Text dataset, is created and released. The generated annotations are used to train a deep convolutional neural network for semantic segmentation. Experiments show that the proposed dataset can be used instead of synthetic data, allowing us to use only a fraction of the training samples and significantly improving the performances.
Tasks Semantic Segmentation, Synthetic Data Generation
Published 2019-04-01
URL https://arxiv.org/abs/1904.00818v6
PDF https://arxiv.org/pdf/1904.00818v6.pdf
PWC https://paperswithcode.com/paper/coco_ts-dataset-pixel-level-annotations-based
Repo
Framework

Learning deep forest with multi-scale Local Binary Pattern features for face anti-spoofing

Title Learning deep forest with multi-scale Local Binary Pattern features for face anti-spoofing
Authors Rizhao Cai, Changsheng Chen
Abstract Face Anti-Spoofing (FAS) is significant for the security of face recognition systems. Convolutional Neural Networks (CNNs) have been introduced to the field of the FAS and have achieved competitive performance. However, CNN-based methods are vulnerable to the adversarial attack. Attackers could generate adversarial-spoofing examples to circumvent a CNN-based face liveness detector. Studies about the transferability of the adversarial attack reveal that utilizing handcrafted feature-based methods could improve security in a system-level. Therefore, handcrafted feature-based methods are worth our exploration. In this paper, we introduce the deep forest, which is proposed as an alternative towards CNNs by Zhou et al., in the problem of the FAS. To the best of our knowledge, this is the first attempt at exploiting the deep forest in the problem of FAS. Moreover, we propose to re-devise the representation constructing by using LBP descriptors rather than the Grained-Scanning Mechanism in the original scheme. Our method achieves competitive results. On the benchmark database IDIAP REPLAY-ATTACK, 0% Equal Error Rate (EER) is achieved. This work provides a competitive option in a fusing scheme for improving system-level security and offers important ideas to those who want to explore methods besides CNNs.
Tasks Adversarial Attack, Face Anti-Spoofing, Face Recognition
Published 2019-10-09
URL https://arxiv.org/abs/1910.03850v1
PDF https://arxiv.org/pdf/1910.03850v1.pdf
PWC https://paperswithcode.com/paper/learning-deep-forest-with-multi-scale-local
Repo
Framework

Yet another but more efficient black-box adversarial attack: tiling and evolution strategies

Title Yet another but more efficient black-box adversarial attack: tiling and evolution strategies
Authors Laurent Meunier, Jamal Atif, Olivier Teytaud
Abstract We introduce a new black-box attack achieving state of the art performances. Our approach is based on a new objective function, borrowing ideas from $\ell_\infty$-white box attacks, and particularly designed to fit derivative-free optimization requirements. It only requires to have access to the logits of the classifier without any other information which is a more realistic scenario. Not only we introduce a new objective function, we extend previous works on black box adversarial attacks to a larger spectrum of evolution strategies and other derivative-free optimization methods. We also highlight a new intriguing property that deep neural networks are not robust to single shot tiled attacks. Our models achieve, with a budget limited to $10,000$ queries, results up to $99.2%$ of success rate against InceptionV3 classifier with $630$ queries to the network on average in the untargeted attacks setting, which is an improvement by $90$ queries of the current state of the art. In the targeted setting, we are able to reach, with a limited budget of $100,000$, $100%$ of success rate with a budget of $6,662$ queries on average, i.e. we need $800$ queries less than the current state of the art.
Tasks Adversarial Attack
Published 2019-10-05
URL https://arxiv.org/abs/1910.02244v2
PDF https://arxiv.org/pdf/1910.02244v2.pdf
PWC https://paperswithcode.com/paper/yet-another-but-more-efficient-black-box-1
Repo
Framework

Sim-to-Real Domain Adaptation For High Energy Physics

Title Sim-to-Real Domain Adaptation For High Energy Physics
Authors Marouen Baalouch, Maxime Defurne, Jean-Philippe Poli, Noëlie Cherrier
Abstract Particle physics or High Energy Physics (HEP) studies the elementary constituents of matter and their interactions with each other. Machine Learning (ML) has played an important role in HEP analysis and has proven extremely successful in this area. Usually, the ML algorithms are trained on numerical simulations of the experimental setup and then applied to the real experimental data. However, any discrepancy between the simulation and real data may lead to dramatic consequences concerning the performances of the algorithm on real data. In this paper, we present an application of domain adaptation using a Domain Adversarial Neural Network trained on public HEP data. We demonstrate the success of this approach to achieve sim-to-real transfer and ensure the consistency of the ML algorithms performances on real and simulated HEP datasets.
Tasks Domain Adaptation
Published 2019-12-17
URL https://arxiv.org/abs/1912.08001v1
PDF https://arxiv.org/pdf/1912.08001v1.pdf
PWC https://paperswithcode.com/paper/sim-to-real-domain-adaptation-for-high-energy
Repo
Framework

AutoML using Metadata Language Embeddings

Title AutoML using Metadata Language Embeddings
Authors Iddo Drori, Lu Liu, Yi Nian, Sharath C. Koorathota, Jie S. Li, Antonio Khalil Moretti, Juliana Freire, Madeleine Udell
Abstract As a human choosing a supervised learning algorithm, it is natural to begin by reading a text description of the dataset and documentation for the algorithms you might use. We demonstrate that the same idea improves the performance of automated machine learning methods. We use language embeddings from modern NLP to improve state-of-the-art AutoML systems by augmenting their recommendations with vector embeddings of datasets and of algorithms. We use these embeddings in a neural architecture to learn the distance between best-performing pipelines. The resulting (meta-)AutoML framework improves on the performance of existing AutoML frameworks. Our zero-shot AutoML system using dataset metadata embeddings provides good solutions instantaneously, running in under one second of computation. Performance is competitive with AutoML systems OBOE, AutoSklearn, AlphaD3M, and TPOT when each framework is allocated a minute of computation. We make our data, models, and code publicly available.
Tasks AutoML
Published 2019-10-08
URL https://arxiv.org/abs/1910.03698v1
PDF https://arxiv.org/pdf/1910.03698v1.pdf
PWC https://paperswithcode.com/paper/automl-using-metadata-language-embeddings
Repo
Framework

Perturbations are not Enough: Generating Adversarial Examples with Spatial Distortions

Title Perturbations are not Enough: Generating Adversarial Examples with Spatial Distortions
Authors He Zhao, Trung Le, Paul Montague, Olivier De Vel, Tamas Abraham, Dinh Phung
Abstract Deep neural network image classifiers are reported to be susceptible to adversarial evasion attacks, which use carefully crafted images created to mislead a classifier. Recently, various kinds of adversarial attack methods have been proposed, most of which focus on adding small perturbations to input images. Despite the success of existing approaches, the way to generate realistic adversarial images with small perturbations remains a challenging problem. In this paper, we aim to address this problem by proposing a novel adversarial method, which generates adversarial examples by imposing not only perturbations but also spatial distortions on input images, including scaling, rotation, shear, and translation. As humans are less susceptible to small spatial distortions, the proposed approach can produce visually more realistic attacks with smaller perturbations, able to deceive classifiers without affecting human predictions. We learn our method by amortized techniques with neural networks and generate adversarial examples efficiently by a forward pass of the networks. Extensive experiments on attacking different types of non-robustified classifiers and robust classifiers with defence show that our method has state-of-the-art performance in comparison with advanced attack parallels.
Tasks Adversarial Attack
Published 2019-10-03
URL https://arxiv.org/abs/1910.01329v1
PDF https://arxiv.org/pdf/1910.01329v1.pdf
PWC https://paperswithcode.com/paper/perturbations-are-not-enough-generating-1
Repo
Framework

Private Federated Learning with Domain Adaptation

Title Private Federated Learning with Domain Adaptation
Authors Daniel Peterson, Pallika Kanani, Virendra J. Marathe
Abstract Federated Learning (FL) is a distributed machine learning (ML) paradigm that enables multiple parties to jointly re-train a shared model without sharing their data with any other parties, offering advantages in both scale and privacy. We propose a framework to augment this collaborative model-building with per-user domain adaptation. We show that this technique improves model accuracy for all users, using both real and synthetic data, and that this improvement is much more pronounced when differential privacy bounds are imposed on the FL model.
Tasks Domain Adaptation
Published 2019-12-13
URL https://arxiv.org/abs/1912.06733v1
PDF https://arxiv.org/pdf/1912.06733v1.pdf
PWC https://paperswithcode.com/paper/private-federated-learning-with-domain
Repo
Framework

Open-plan Glare Evaluator (OGE): A New Glare Prediction Model for Open-Plan Offices Using Machine Learning Algorithms

Title Open-plan Glare Evaluator (OGE): A New Glare Prediction Model for Open-Plan Offices Using Machine Learning Algorithms
Authors Ayman Wagdy, Veronica Garcia-Hansen, Mohammed Elhenawy, Gillian Isoardi, Robin Drogemuller, Fatma Fathy
Abstract Predicting discomfort glare in open-plan offices is a challenging problem since most of available glare metrics are developed for cellular offices which are typically daylight dominated. The problem with open-plan offices is that they are mainly dependent on electric lighting rather than daylight even when they have a fully glazed facade. In addition, the contrast between bright windows and the buildings interior can be problematic and may cause discomfort glare to the building occupants. These problems can affect occupant productivity and wellbeing. Thus, it is important to develop a predictive model to avoid discomfort glare when designing open plan offices. To the best of our knowledge, we are the first to adopt Machine Learning (ML) models to predict discomfort glare. In order to develop new glare predictive models for these types of offices, Post-Occupancy Evaluation (POE) and High Dynamic Range (HDR) images were collected from 80 occupants (n=80) in four different open-plan offices. Consequently, various multi-region luminance values, luminance and glare indices were calculated and used as input features to train ML models. The accuracy of the ML model was compared to the accuracy of 24 indices which were also evaluated using a Receiver Operating Characteristic (ROC) analysis to identify the best cutoff values (thresholds) for each index for open-plan configurations. Results showed that the ML glare model could predict glare in open-plan offices with an accuracy of 83.8% (0.80 true positive rate and 0.86 true negative rate) which outperformed the accuracy of the previously developed glare metrics.
Tasks
Published 2019-10-12
URL https://arxiv.org/abs/1910.05594v1
PDF https://arxiv.org/pdf/1910.05594v1.pdf
PWC https://paperswithcode.com/paper/open-plan-glare-evaluator-oge-a-new-glare
Repo
Framework

Restoring Images with Unknown Degradation Factors by Recurrent Use of a Multi-branch Network

Title Restoring Images with Unknown Degradation Factors by Recurrent Use of a Multi-branch Network
Authors Xing Liu, Masanori Suganuma, Xiyang Luo, Takayuki Okatani
Abstract The employment of convolutional neural networks has achieved unprecedented performance in the task of image restoration for a variety of degradation factors. However, high-performance networks have been specifically designed for a single degradation factor. In this paper, we tackle a harder problem, restoring a clean image from its degraded version with an unknown degradation factor, subject to the condition that it is one of the known factors. Toward this end, we design a network having multiple pairs of input and output branches and use it in a recurrent fashion such that a different branch pair is used at each of the recurrent paths. We reinforce the shared part of the network with improved components so that it can handle different degradation factors. We also propose a two-step training method for the network, which consists of multi-task learning and finetuning. The experimental results show that the proposed network yields at least comparable or sometimes even better performance on four degradation factors as compared with the best dedicated network for each of the four. We also test it on a further harder task where the input image contains multiple degradation factors that are mixed with unknown mixture ratios, showing that it achieves better performance than the previous state-of-the-art method designed for the task.
Tasks Deblurring, Image Restoration, JPEG Artifact Removal, Multi-Task Learning
Published 2019-07-10
URL https://arxiv.org/abs/1907.04508v2
PDF https://arxiv.org/pdf/1907.04508v2.pdf
PWC https://paperswithcode.com/paper/joint-learning-of-multiple-image-restoration
Repo
Framework
comments powered by Disqus