January 31, 2020

3362 words 16 mins read

Paper Group ANR 96

Paper Group ANR 96

Exponential Separations in Local Differential Privacy. Semantic filtering through deep source separation on microscopy images. Adaptive Robust Optimization with Nearly Submodular Structure. Analysis and Improvement of Adversarial Training in DQN Agents With Adversarially-Guided Exploration (AGE). Localized Debiased Machine Learning: Efficient Estim …

Exponential Separations in Local Differential Privacy

Title Exponential Separations in Local Differential Privacy
Authors Matthew Joseph, Jieming Mao, Aaron Roth
Abstract We prove a general connection between the communication complexity of two-player games and the sample complexity of their multi-player locally private analogues. We use this connection to prove sample complexity lower bounds for locally differentially private protocols as straightforward corollaries of results from communication complexity. In particular, we 1) use a communication lower bound for the hidden layers problem to prove an exponential sample complexity separation between sequentially and fully interactive locally private protocols, and 2) use a communication lower bound for the pointer chasing problem to prove an exponential sample complexity separation between $k$ round and $k+1$ round sequentially interactive locally private protocols, for every $k$.
Tasks
Published 2019-07-01
URL https://arxiv.org/abs/1907.00813v3
PDF https://arxiv.org/pdf/1907.00813v3.pdf
PWC https://paperswithcode.com/paper/exponential-separations-in-local-differential
Repo
Framework

Semantic filtering through deep source separation on microscopy images

Title Semantic filtering through deep source separation on microscopy images
Authors Avelino Javer, Jens Rittscher
Abstract By their very nature microscopy images of cells and tissues consist of a limited number of object types or components. In contrast to most natural scenes, the composition is known a priori. Decomposing biological images into semantically meaningful objects and layers is the aim of this paper. Building on recent approaches to image de-noising we present a framework that achieves state-of-the-art segmentation results requiring little or no manual annotations. Here, synthetic images generated by adding cell crops are sufficient to train the model. Extensive experiments on cellular images, a histology data set, and small animal videos demonstrate that our approach generalizes to a broad range of experimental settings. As the proposed methodology does not require densely labelled training images and is capable of resolving the partially overlapping objects it holds the promise of being of use in a number of different applications.
Tasks
Published 2019-09-02
URL https://arxiv.org/abs/1909.00691v1
PDF https://arxiv.org/pdf/1909.00691v1.pdf
PWC https://paperswithcode.com/paper/semantic-filtering-through-deep-source
Repo
Framework

Adaptive Robust Optimization with Nearly Submodular Structure

Title Adaptive Robust Optimization with Nearly Submodular Structure
Authors Shaojie Tang, Jing Yuan
Abstract Constrained submodular maximization has been extensively studied in the recent years. In this paper, we study adaptive robust optimization with nearly submodular structure (ARONSS). Our objective is to randomly select a subset of items that maximizes the worst-case value of several reward functions simultaneously. Our work differs from existing studies in two ways: (1) we study the robust optimization problem under the adaptive setting, i.e., one needs to adaptively select items based on the feedback collected from picked items, and (2) our results apply to a broad range of reward functions characterized by $\epsilon$-nearly submodular function. We first analyze the adaptvity gap of ARONSS and show that the gap between the best adaptive solution and the best non-adaptive solution is bounded. Then we propose a approximate solution to this problem when all reward functions are submodular. Our algorithm achieves approximation ratio $(1-1/e)$ when considering matroid constraint. At last, we present two heuristics for the general case. All proposed solutions are non-adaptive which are easy to implement.
Tasks
Published 2019-05-14
URL https://arxiv.org/abs/1905.05339v3
PDF https://arxiv.org/pdf/1905.05339v3.pdf
PWC https://paperswithcode.com/paper/adaptive-robust-optimization-with-nearly
Repo
Framework

Analysis and Improvement of Adversarial Training in DQN Agents With Adversarially-Guided Exploration (AGE)

Title Analysis and Improvement of Adversarial Training in DQN Agents With Adversarially-Guided Exploration (AGE)
Authors Vahid Behzadan, William Hsu
Abstract This paper investigates the effectiveness of adversarial training in enhancing the robustness of Deep Q-Network (DQN) policies to state-space perturbations. We first present a formal analysis of adversarial training in DQN agents and its performance with respect to the proportion of adversarial perturbations to nominal observations used for training. Next, we consider the sample-inefficiency of current adversarial training techniques, and propose a novel Adversarially-Guided Exploration (AGE) mechanism based on a modified hybrid of the $\epsilon$-greedy algorithm and Boltzmann exploration. We verify the feasibility of this exploration mechanism through experimental evaluation of its performance in comparison with the traditional decaying $\epsilon$-greedy and parameter-space noise exploration algorithms.
Tasks
Published 2019-06-03
URL https://arxiv.org/abs/1906.01119v1
PDF https://arxiv.org/pdf/1906.01119v1.pdf
PWC https://paperswithcode.com/paper/analysis-and-improvement-of-adversarial
Repo
Framework

Localized Debiased Machine Learning: Efficient Estimation of Quantile Treatment Effects, Conditional Value at Risk, and Beyond

Title Localized Debiased Machine Learning: Efficient Estimation of Quantile Treatment Effects, Conditional Value at Risk, and Beyond
Authors Nathan Kallus, Xiaojie Mao, Masatoshi Uehara
Abstract We consider the efficient estimation of a low-dimensional parameter in the presence of very high-dimensional nuisances that may depend on the parameter of interest. An important example is the quantile treatment effect (QTE) in causal inference, where the efficient estimation equation involves as a nuisance the conditional cumulative distribution evaluated at the quantile to be estimated. Debiased machine learning (DML) is a data-splitting approach to address the need to estimate nuisances using flexible machine learning methods that may not satisfy strong metric entropy conditions, but applying it to problems with estimand-dependent nuisances would require estimating too many nuisances to be practical. For the QTE estimation, DML requires we learn the whole conditional cumulative distribution function, which may be challenging in practice and stands in contrast to only needing to estimate just two regression functions as in the efficient estimation of average treatment effects. Instead, we propose localized debiased machine learning (LDML), a new three-way data-splitting approach that avoids this burdensome step and needs only estimate the nuisances at a single initial bad guess for the parameters. In particular, under a Frechet-derivative orthogonality condition, we show the oracle estimation equation is asymptotically equivalent to one where the nuisance is evaluated at the true parameter value and we provide a strategy to target this alternative formulation. In the case of QTE estimation, this involves only learning two binary regression models, for which many standard, time-tested machine learning methods exist. We prove that under certain lax rate conditions, our estimator has the same favorable asymptotic behavior as the infeasible oracle estimator that solves the estimating equation with the true nuisance functions.
Tasks Causal Inference
Published 2019-12-30
URL https://arxiv.org/abs/1912.12945v1
PDF https://arxiv.org/pdf/1912.12945v1.pdf
PWC https://paperswithcode.com/paper/localized-debiased-machine-learning-efficient
Repo
Framework

Deep Reinforcement Learning for Imbalanced Classification

Title Deep Reinforcement Learning for Imbalanced Classification
Authors Enlu Lin, Qiong Chen, Xiaoming Qi
Abstract Data in real-world application often exhibit skewed class distribution which poses an intense challenge for machine learning. Conventional classification algorithms are not effective in the case of imbalanced data distribution, and may fail when the data distribution is highly imbalanced. To address this issue, we propose a general imbalanced classification model based on deep reinforcement learning. We formulate the classification problem as a sequential decision-making process and solve it by deep Q-learning network. The agent performs a classification action on one sample at each time step, and the environment evaluates the classification action and returns a reward to the agent. The reward from minority class sample is larger so the agent is more sensitive to the minority class. The agent finally finds an optimal classification policy in imbalanced data under the guidance of specific reward function and beneficial learning environment. Experiments show that our proposed model outperforms the other imbalanced classification algorithms, and it can identify more minority samples and has great classification performance.
Tasks Decision Making, Q-Learning
Published 2019-01-05
URL http://arxiv.org/abs/1901.01379v1
PDF http://arxiv.org/pdf/1901.01379v1.pdf
PWC https://paperswithcode.com/paper/deep-reinforcement-learning-for-imbalanced
Repo
Framework

Domain Adaptation for Vehicle Detection from Bird’s Eye View LiDAR Point Cloud Data

Title Domain Adaptation for Vehicle Detection from Bird’s Eye View LiDAR Point Cloud Data
Authors Khaled Saleh, Ahmed Abobakr, Mohammed Attia, Julie Iskander, Darius Nahavandi, Mohammed Hossny
Abstract Point cloud data from 3D LiDAR sensors are one of the most crucial sensor modalities for versatile safety-critical applications such as self-driving vehicles. Since the annotations of point cloud data is an expensive and time-consuming process, therefore recently the utilisation of simulated environments and 3D LiDAR sensors for this task started to get some popularity. With simulated sensors and environments, the process for obtaining an annotated synthetic point cloud data became much easier. However, the generated synthetic point cloud data are still missing the artefacts usually exist in point cloud data from real 3D LiDAR sensors. As a result, the performance of the trained models on this data for perception tasks when tested on real point cloud data is degraded due to the domain shift between simulated and real environments. Thus, in this work, we are proposing a domain adaptation framework for bridging this gap between synthetic and real point cloud data. Our proposed framework is based on the deep cycle-consistent generative adversarial networks (CycleGAN) architecture. We have evaluated the performance of our proposed framework on the task of vehicle detection from a bird’s eye view (BEV) point cloud images coming from real 3D LiDAR sensors. The framework has shown competitive results with an improvement of more than 7% in average precision score over other baseline approaches when tested on real BEV point cloud images.
Tasks Domain Adaptation, Unsupervised Domain Adaptation
Published 2019-05-22
URL https://arxiv.org/abs/1905.08955v1
PDF https://arxiv.org/pdf/1905.08955v1.pdf
PWC https://paperswithcode.com/paper/domain-adaptation-for-vehicle-detection-from
Repo
Framework

Learning Credible Deep Neural Networks with Rationale Regularization

Title Learning Credible Deep Neural Networks with Rationale Regularization
Authors Mengnan Du, Ninghao Liu, Fan Yang, Xia Hu
Abstract Recent explainability related studies have shown that state-of-the-art DNNs do not always adopt correct evidences to make decisions. It not only hampers their generalization but also makes them less likely to be trusted by end-users. In pursuit of developing more credible DNNs, in this paper we propose CREX, which encourages DNN models to focus more on evidences that actually matter for the task at hand, and to avoid overfitting to data-dependent bias and artifacts. Specifically, CREX regularizes the training process of DNNs with rationales, i.e., a subset of features highlighted by domain experts as justifications for predictions, to enforce DNNs to generate local explanations that conform with expert rationales. Even when rationales are not available, CREX still could be useful by requiring the generated explanations to be sparse. Experimental results on two text classification datasets demonstrate the increased credibility of DNNs trained with CREX. Comprehensive analysis further shows that while CREX does not always improve prediction accuracy on the held-out test set, it significantly increases DNN accuracy on new and previously unseen data beyond test set, highlighting the advantage of the increased credibility.
Tasks Text Classification
Published 2019-08-13
URL https://arxiv.org/abs/1908.05601v1
PDF https://arxiv.org/pdf/1908.05601v1.pdf
PWC https://paperswithcode.com/paper/learning-credible-deep-neural-networks-with
Repo
Framework

Diagnosis of Celiac Disease and Environmental Enteropathy on Biopsy Images Using Color Balancing on Convolutional Neural Networks

Title Diagnosis of Celiac Disease and Environmental Enteropathy on Biopsy Images Using Color Balancing on Convolutional Neural Networks
Authors Kamran Kowsari, Rasoul Sali, Marium N. Khan, William Adorno, S. Asad Ali, Sean R. Moore, Beatrice C. Amadi, Paul Kelly, Sana Syed, Donald E. Brown
Abstract Celiac Disease (CD) and Environmental Enteropathy (EE) are common causes of malnutrition and adversely impact normal childhood development. CD is an autoimmune disorder that is prevalent worldwide and is caused by an increased sensitivity to gluten. Gluten exposure destructs the small intestinal epithelial barrier, resulting in nutrient mal-absorption and childhood under-nutrition. EE also results in barrier dysfunction but is thought to be caused by an increased vulnerability to infections. EE has been implicated as the predominant cause of under-nutrition, oral vaccine failure, and impaired cognitive development in low-and-middle-income countries. Both conditions require a tissue biopsy for diagnosis, and a major challenge of interpreting clinical biopsy images to differentiate between these gastrointestinal diseases is striking histopathologic overlap between them. In the current study, we propose a convolutional neural network (CNN) to classify duodenal biopsy images from subjects with CD, EE, and healthy controls. We evaluated the performance of our proposed model using a large cohort containing 1000 biopsy images. Our evaluations show that the proposed model achieves an area under ROC of 0.99, 1.00, and 0.97 for CD, EE, and healthy controls, respectively. These results demonstrate the discriminative power of the proposed model in duodenal biopsies classification.
Tasks
Published 2019-04-10
URL https://arxiv.org/abs/1904.05773v5
PDF https://arxiv.org/pdf/1904.05773v5.pdf
PWC https://paperswithcode.com/paper/diagnosis-of-celiac-disease-and-environmental
Repo
Framework

Optimal Use of Experience in First Person Shooter Environments

Title Optimal Use of Experience in First Person Shooter Environments
Authors Matthew Aitchison
Abstract Although reinforcement learning has made great strides recently, a continuing limitation is that it requires an extremely high number of interactions with the environment. In this paper, we explore the effectiveness of reusing experience from the experience replay buffer in the Deep Q-Learning algorithm. We test the effectiveness of applying learning update steps multiple times per environmental step in the VizDoom environment and show first, this requires a change in the learning rate, and second that it does not improve the performance of the agent. Furthermore, we show that updating less frequently is effective up to a ratio of 4:1, after which performance degrades significantly. These results quantitatively confirm the widespread practice of performing learning updates every 4th environmental step.
Tasks Q-Learning
Published 2019-06-24
URL https://arxiv.org/abs/1906.09734v1
PDF https://arxiv.org/pdf/1906.09734v1.pdf
PWC https://paperswithcode.com/paper/optimal-use-of-experience-in-first-person
Repo
Framework

On Intra-Class Variance for Deep Learning of Classifiers

Title On Intra-Class Variance for Deep Learning of Classifiers
Authors Rafal Pilarczyk, Wladyslaw Skarbek
Abstract A novel technique for deep learning of image classifiers is presented. The learned CNN models offer better separation of deep features (also known as embedded vectors) measured by Euclidean proximity and also no deterioration of the classification results by class membership probability. The latter feature can be used for enhancing image classifiers having the classes at the model’s exploiting stage different from from classes during the training stage. While the Shannon information of SoftMax probability for target class is extended for mini-batch by the intra-class variance, the trained network itself is extended by the Hadamard layer with the parameters representing the class centers. Contrary to the existing solutions, this extra neural layer enables interfacing of the training algorithm to the standard stochastic gradient optimizers, e.g. AdaM algorithm. Moreover, this approach makes the computed centroids immediately adapting to the updating embedded vectors and finally getting the comparable accuracy in less epochs.
Tasks
Published 2019-01-31
URL http://arxiv.org/abs/1901.11186v2
PDF http://arxiv.org/pdf/1901.11186v2.pdf
PWC https://paperswithcode.com/paper/on-intra-class-variance-for-deep-learning-of
Repo
Framework

Few Labeled Atlases are Necessary for Deep-Learning-Based Segmentation

Title Few Labeled Atlases are Necessary for Deep-Learning-Based Segmentation
Authors Hyeon Woo Lee, Mert R. Sabuncu, Adrian V. Dalca
Abstract We tackle biomedical image segmentation in the scenario of only a few labeled brain MR images. This is an important and challenging task in medical applications, where manual annotations are time-consuming. Current multi-atlas based segmentation methods use image registration to warp segments from labeled images onto a new scan. In a different paradigm, supervised learning-based segmentation strategies have gained popularity. These method consistently use relatively large sets of labeled training data, and their behavior in the regime of a few labeled biomedical images has not been thoroughly evaluated. In this work, we provide two important results for segmentation in the scenario where few labeled images are available. First, we propose a straightforward implementation of efficient semi-supervised learning-based registration method, which we showcase in a multi-atlas segmentation framework. Second, through an extensive empirical study, we evaluate the performance of a supervised segmentation approach, where the training images are augmented via random deformations. Surprisingly, we find that in both paradigms, accurate segmentation is generally possible even in the context of few labeled images.
Tasks Deformable Medical Image Registration, Image Registration, Medical Image Registration, Semantic Segmentation, Semi-Supervised Semantic Segmentation
Published 2019-08-13
URL https://arxiv.org/abs/1908.04466v4
PDF https://arxiv.org/pdf/1908.04466v4.pdf
PWC https://paperswithcode.com/paper/few-labeled-atlases-are-necessary-for-deep
Repo
Framework

RNN-Test: Adversarial Testing Framework for Recurrent Neural Network Systems

Title RNN-Test: Adversarial Testing Framework for Recurrent Neural Network Systems
Authors Jianmin Guo, Yue Zhao, Xueying Han, Yu Jiang, Jiaguang Sun
Abstract While huge efforts have been investigated in the adversarial testing of convolutional neural networks (CNN), the testing for recurrent neural networks (RNN) is still limited to the classification context and leave threats for vast sequential application domains. In this work, we propose a generic adversarial testing framework RNN-Test. First, based on the distinctive structure of RNNs, we define three novel coverage metrics to measure the testing completeness and guide the generation of adversarial inputs. Second, we propose the state inconsistency orientation to generate the perturbations by maximizing the inconsistency of the hidden states of RNN cells. Finally, we combine orientations with coverage guidance to produce minute perturbations. Given the RNN model and the sequential inputs, RNN-Test will modify one character or one word out of the whole inputs based on the perturbations obtained, so as to lead the RNN to produce wrong outputs. For evaluation, we apply RNN-Test on two models of common RNN structure - the PTB language model and the spell checker model. RNN-Test efficiently reduces the performance of the PTB language model by increasing its test perplexity by 58.11%, and finds numbers of incorrect behaviors of the spell checker model with the success rate of 73.44% on average. With our customization, RNN-Test using the redefined neuron coverage as guidance could achieve 35.71% higher perplexity than original strategy of DeepXplore.
Tasks Language Modelling
Published 2019-11-11
URL https://arxiv.org/abs/1911.06155v1
PDF https://arxiv.org/pdf/1911.06155v1.pdf
PWC https://paperswithcode.com/paper/rnn-test-adversarial-testing-framework-for
Repo
Framework

Reinforcement Learning-based Automatic Diagnosis of Acute Appendicitis in Abdominal CT

Title Reinforcement Learning-based Automatic Diagnosis of Acute Appendicitis in Abdominal CT
Authors Walid Abdullah Al, Il Dong Yun, Kyong Joon Lee
Abstract Acute appendicitis characterized by a painful inflammation of the vermiform appendix is one of the most common surgical emergencies. Localizing the appendix is challenging due to its unclear anatomy amidst the complex colon-structure as observed in the conventional CT views, resulting in a time-consuming diagnosis. End-to-end learning of a convolutional neural network (CNN) is also not likely to be useful because of the negligible size of the appendix compared with the abdominal CT volume. With no prior computational approaches to the best of our knowledge, we propose the first computerized automation for acute appendicitis diagnosis. In our approach, we utilize a reinforcement learning agent deployed in the lower abdominal region to obtain the appendix location first to reduce the search space for diagnosis. Then, we obtain the classification scores (i.e., the likelihood of acute appendicitis) for the local neighborhood around the localized position, using a CNN trained only on a small appendix patch per volume. From the spatial representation of the resultant scores, we finally define a region of low-entropy (RLE) to choose the optimal diagnosis score, which helps improve the classification accuracy showing robustness even under high appendix localization error cases. In our experiment with 319 abdominal CT volumes, the proposed RLE-based decision with prior localization showed significant improvement over the standard CNN-based diagnosis approaches.
Tasks
Published 2019-09-02
URL https://arxiv.org/abs/1909.00617v1
PDF https://arxiv.org/pdf/1909.00617v1.pdf
PWC https://paperswithcode.com/paper/reinforcement-learning-based-automatic
Repo
Framework

Intrinsically Sparse Long Short-Term Memory Networks

Title Intrinsically Sparse Long Short-Term Memory Networks
Authors Shiwei Liu, Decebal Constantin Mocanu, Mykola Pechenizkiy
Abstract Long Short-Term Memory (LSTM) has achieved state-of-the-art performances on a wide range of tasks. Its outstanding performance is guaranteed by the long-term memory ability which matches the sequential data perfectly and the gating structure controlling the information flow. However, LSTMs are prone to be memory-bandwidth limited in realistic applications and need an unbearable period of training and inference time as the model size is ever-increasing. To tackle this problem, various efficient model compression methods have been proposed. Most of them need a big and expensive pre-trained model which is a nightmare for resource-limited devices where the memory budget is strictly limited. To remedy this situation, in this paper, we incorporate the Sparse Evolutionary Training (SET) procedure into LSTM, proposing a novel model dubbed SET-LSTM. Rather than starting with a fully-connected architecture, SET-LSTM has a sparse topology and dramatically fewer parameters in both phases, training and inference. Considering the specific architecture of LSTMs, we replace the LSTM cells and embedding layers with sparse structures and further on, use an evolutionary strategy to adapt the sparse connectivity to the data. Additionally, we find that SET-LSTM can provide many different good combinations of sparse connectivity to substitute the overparameterized optimization problem of dense neural networks. Evaluated on four sentiment analysis classification datasets, the results demonstrate that our proposed model is able to achieve usually better performance than its fully connected counterpart while having less than 4% of its parameters.
Tasks Model Compression, Sentiment Analysis
Published 2019-01-26
URL http://arxiv.org/abs/1901.09208v1
PDF http://arxiv.org/pdf/1901.09208v1.pdf
PWC https://paperswithcode.com/paper/intrinsically-sparse-long-short-term-memory
Repo
Framework
comments powered by Disqus