January 28, 2020

2871 words 14 mins read

Paper Group ANR 827

Paper Group ANR 827

Towards Generalizing Sensorimotor Control Across Weather Conditions. Proximal Reliability Optimization for Reinforcement Learning. Balancing Accuracy and Diversity in Recommendations using Matrix Completion Framework. Influenza Modeling Based on Massive Feature Engineering and International Flow Deconvolution. Kernel density estimation based sampli …

Towards Generalizing Sensorimotor Control Across Weather Conditions

Title Towards Generalizing Sensorimotor Control Across Weather Conditions
Authors Qadeer Khan, Patrick Wenzel, Daniel Cremers, Laura Leal-Taixé
Abstract The ability of deep learning models to generalize well across different scenarios depends primarily on the quality and quantity of annotated data. Labeling large amounts of data for all possible scenarios that a model may encounter would not be feasible; if even possible. We propose a framework to deal with limited labeled training data and demonstrate it on the application of vision-based vehicle control. We show how limited steering angle data available for only one condition can be transferred to multiple different weather scenarios. This is done by leveraging unlabeled images in a teacher-student learning paradigm complemented with an image-to-image translation network. The translation network transfers the images to a new domain, whereas the teacher provides soft supervised targets to train the student on this domain. Furthermore, we demonstrate how utilization of auxiliary networks can reduce the size of a model at inference time, without affecting the accuracy. The experiments show that our approach generalizes well across multiple different weather conditions using only ground truth labels from one domain.
Tasks Image-to-Image Translation
Published 2019-07-25
URL https://arxiv.org/abs/1907.11025v1
PDF https://arxiv.org/pdf/1907.11025v1.pdf
PWC https://paperswithcode.com/paper/towards-generalizing-sensorimotor-control
Repo
Framework

Proximal Reliability Optimization for Reinforcement Learning

Title Proximal Reliability Optimization for Reinforcement Learning
Authors Narendra Patwardhan, Zequn Wang
Abstract Despite the numerous advances, reinforcement learning remains away from widespread acceptance for autonomous controller design as compared to classical methods due to lack of ability to effectively tackle the reality gap. The reliance on absolute or deterministic reward as a metric for optimization process renders reinforcement learning highly susceptible to changes in problem dynamics. We introduce a novel framework that effectively quantizes the uncertainty of the design space and induces robustness in controllers by switching to a reliability-based optimization routine. The data efficiency of the method is maintained to match reward based optimization methods by employing a model-based approach. We prove the stability of learned neuro-controllers in both static and dynamic environments on classical reinforcement learning tasks such as Cart Pole balancing and Inverted Pendulum.
Tasks
Published 2019-06-03
URL https://arxiv.org/abs/1906.01127v1
PDF https://arxiv.org/pdf/1906.01127v1.pdf
PWC https://paperswithcode.com/paper/proximal-reliability-optimization-for
Repo
Framework

Balancing Accuracy and Diversity in Recommendations using Matrix Completion Framework

Title Balancing Accuracy and Diversity in Recommendations using Matrix Completion Framework
Authors Anupriya Gogna, Angshul Majumdar
Abstract Design of recommender systems aimed at achieving high prediction accuracy is a widely researched area. However, several studies have suggested the need for diversified recommendations, with acceptable level of accuracy, to avoid monotony and improve customers experience. However, increasing diversity comes with an associated reduction in recommendation accuracy; thereby necessitating an optimum tradeoff between the two. In this work, we attempt to achieve accuracy vs diversity balance, by exploiting available ratings and item metadata, through a single, joint optimization model built over the matrix completion framework. Most existing works, unlike our formulation, propose a 2 stage model, a heuristic item ranking scheme on top of an existing collaborative filtering technique. Experimental evaluation on a movie recommender system indicates that our model achieves higher diversity for a given drop in accuracy as compared to existing state of the art techniques.
Tasks Matrix Completion, Recommendation Systems
Published 2019-12-11
URL https://arxiv.org/abs/2001.04349v1
PDF https://arxiv.org/pdf/2001.04349v1.pdf
PWC https://paperswithcode.com/paper/balancing-accuracy-and-diversity-in
Repo
Framework

Influenza Modeling Based on Massive Feature Engineering and International Flow Deconvolution

Title Influenza Modeling Based on Massive Feature Engineering and International Flow Deconvolution
Authors Ziming Liu, Yixuan Wang, Zizhao Han, Dian Wu
Abstract In this article, we focus on the analysis of the potential factors driving the spread of influenza, and possible policies to mitigate the adverse effects of the disease. To be precise, we first invoke discrete Fourier transform (DFT) to conclude a yearly periodic regional structure in the influenza activity, thus safely restricting ourselves to the analysis of the yearly influenza behavior. Then we collect a massive number of possible region-wise indicators contributing to the influenza mortality, such as consumption, immunization, sanitation, water quality, and other indicators from external data, with $1170$ dimensions in total. We extract significant features from the high dimensional indicators using a combination of data analysis techniques, including matrix completion, support vector machines (SVM), autoencoders, and principal component analysis (PCA). Furthermore, we model the international flow of migration and trade as a convolution on regional influenza activity, and solve the deconvolution problem as higher-order perturbations to the linear regression, thus separating regional and international factors related to the influenza mortality. Finally, both the original model and the perturbed model are tested on regional examples, as validations of our models. Pertaining to the policy, we make a proposal based on the connectivity data along with the previously extracted significant features to alleviate the impact of influenza, as well as efficiently propagate and carry out the policies. We conclude that environmental features and economic features are of significance to the influenza mortality. The model can be easily adapted to model other types of infectious diseases.
Tasks Feature Engineering, Matrix Completion
Published 2019-12-06
URL https://arxiv.org/abs/1912.02989v1
PDF https://arxiv.org/pdf/1912.02989v1.pdf
PWC https://paperswithcode.com/paper/influenza-modeling-based-on-massive-feature
Repo
Framework

Kernel density estimation based sampling for imbalanced class distribution

Title Kernel density estimation based sampling for imbalanced class distribution
Authors Firuz Kamalov
Abstract Imbalanced response variable distribution is a common occurrence in data science. In fields such as fraud detection, medical diagnostics, system intrusion detection and many others where abnormal behavior is rarely observed the data under study often features disproportionate target class distribution. One common way to combat class imbalance is through resampling the minority class to achieve a more balanced distribution. In this paper, we investigate the performance of the sampling method based on kernel density estimation (KDE). We believe that KDE offers a more natural way of generating new instances of minority class that is less prone to overfitting than other standard sampling techniques. It is based on a well established theory of nonparametric statistical estimation. Numerical experiments show that KDE can outperform other sampling techniques on a range of real life datasets as measured by F1-score and G-mean. The results remain consistent across a number of classification algorithms used in the experiments. Furthermore, the proposed method outperforms the benchmark methods irregardless of the class distribution ratio. We conclude, based on the solid theoretical foundation and strong experimental results, that the proposed method would be a valuable tool in problems involving imbalanced class distribution.
Tasks Density Estimation, Fraud Detection, Intrusion Detection
Published 2019-10-17
URL https://arxiv.org/abs/1910.07842v2
PDF https://arxiv.org/pdf/1910.07842v2.pdf
PWC https://paperswithcode.com/paper/kde-sampling-for-imbalanced-class
Repo
Framework

Inertial nonconvex alternating minimizations for the image deblurring

Title Inertial nonconvex alternating minimizations for the image deblurring
Authors Tao Sun, Roberto Barrio, Marcos Rodriguez, Hao Jiang
Abstract In image processing, Total Variation (TV) regularization models are commonly used to recover blurred images. One of the most efficient and popular methods to solve the convex TV problem is the Alternating Direction Method of Multipliers (ADMM) algorithm, recently extended using the inertial proximal point method. Although all the classical studies focus on only a convex formulation, recent articles are paying increasing attention to the nonconvex methodology due to its good numerical performance and properties. In this paper, we propose to extend the classical formulation with a novel nonconvex Alternating Direction Method of Multipliers with the Inertial technique (IADMM). Under certain assumptions on the parameters, we prove the convergence of the algorithm with the help of the Kurdyka-{\L}ojasiewicz property. We also present numerical simulations on classical TV image reconstruction problems to illustrate the efficiency of the new algorithm and its behavior compared with the well established ADMM method.
Tasks Deblurring, Image Reconstruction
Published 2019-07-27
URL https://arxiv.org/abs/1907.12945v1
PDF https://arxiv.org/pdf/1907.12945v1.pdf
PWC https://paperswithcode.com/paper/inertial-nonconvex-alternating-minimizations
Repo
Framework

Inverse-Transform AutoEncoder for Anomaly Detection

Title Inverse-Transform AutoEncoder for Anomaly Detection
Authors Chaoqing Huang, Jinkun Cao, Fei Ye, Maosen Li, Ya Zhang, Cewu Lu
Abstract Reconstruction-based methods have recently shown great promise for anomaly detection. We here propose a new transform-based framework for anomaly detection. A selected set of transformations based on human priors is used to erase certain targeted information from input data. An inverse-transform autoencoder is trained with the normal data only to embed corresponding erased information during the restoration of the original data. The normal and anomalous data are thus expected to be differentiable based on restoration errors. Extensive experiments have demonstrated that the proposed method significantly outperforms several state-of-the-arts on multiple benchmark datasets, especially on ImageNet, increasing the AUROC of the top-performing baseline by 10.1%. We also evaluate our method on a real-world anomaly detection dataset MVTec AD and a video anomaly detection dataset ShanghaiTech to validate the effectiveness of the method in real-world environments.
Tasks Anomaly Detection
Published 2019-11-25
URL https://arxiv.org/abs/1911.10676v1
PDF https://arxiv.org/pdf/1911.10676v1.pdf
PWC https://paperswithcode.com/paper/inverse-transform-autoencoder-for-anomaly
Repo
Framework

Early Classification for Agricultural Monitoring from Satellite Time Series

Title Early Classification for Agricultural Monitoring from Satellite Time Series
Authors Marc Rußwurm, Romain Tavenard, Sébastien Lefèvre, Marco Körner
Abstract In this work, we introduce a recently developed early classification mechanism to satellite-based agricultural monitoring. It augments existing classification models by an additional stopping probability based on the previously seen information. This mechanism is end-to-end trainable and derives its stopping decision solely from the observed satellite data. We show results on field parcels in central Europe where sufficient ground truth data is available for an empiric evaluation of the results with local phenological information obtained from authorities. We observe that the recurrent neural network outfitted with this early classification mechanism was able to distinguish the many of the crop types before the end of the vegetative period. Further, we associated these stopping times with evaluated ground truth information and saw that the times of classification were related to characteristic events of the observed plants’ phenology.
Tasks Time Series
Published 2019-08-27
URL https://arxiv.org/abs/1908.10283v1
PDF https://arxiv.org/pdf/1908.10283v1.pdf
PWC https://paperswithcode.com/paper/early-classification-for-agricultural
Repo
Framework

Blind Deblurring using Deep Learning: A Survey

Title Blind Deblurring using Deep Learning: A Survey
Authors Siddhant Sahu, Manoj Kumar Lenka, Pankaj Kumar Sa
Abstract We inspect all the deep learning based solutions and provide holistic understanding of various architectures that have evolved over the past few years to solve blind deblurring. The introductory work used deep learning to estimate some features of the blur kernel and then moved onto predicting the blur kernel entirely, which converts the problem into non-blind deblurring. The recent state of the art techniques are end to end, i.e., they don’t estimate the blur kernel rather try to estimate the latent sharp image directly from the blurred image. The benchmarking PSNR and SSIM values on standard datasets of GOPRO and Kohler using various architectures are also provided.
Tasks Deblurring
Published 2019-07-23
URL https://arxiv.org/abs/1907.10128v1
PDF https://arxiv.org/pdf/1907.10128v1.pdf
PWC https://paperswithcode.com/paper/blind-deblurring-using-deep-learning-a-survey
Repo
Framework

A Case for Backward Compatibility for Human-AI Teams

Title A Case for Backward Compatibility for Human-AI Teams
Authors Gagan Bansal, Besmira Nushi, Ece Kamar, Dan Weld, Walter Lasecki, Eric Horvitz
Abstract AI systems are being deployed to support human decision making in high-stakes domains. In many cases, the human and AI form a team, in which the human makes decisions after reviewing the AI’s inferences. A successful partnership requires that the human develops insights into the performance of the AI system, including its failures. We study the influence of updates to an AI system in this setting. While updates can increase the AI’s predictive performance, they may also lead to changes that are at odds with the user’s prior experiences and confidence in the AI’s inferences, hurting therefore the overall team performance. We introduce the notion of the compatibility of an AI update with prior user experience and present methods for studying the role of compatibility in human-AI teams. Empirical results on three high-stakes domains show that current machine learning algorithms do not produce compatible updates. We propose a re-training objective to improve the compatibility of an update by penalizing new errors. The objective offers full leverage of the performance/compatibility tradeoff, enabling more compatible yet accurate updates.
Tasks Decision Making
Published 2019-06-04
URL https://arxiv.org/abs/1906.01148v1
PDF https://arxiv.org/pdf/1906.01148v1.pdf
PWC https://paperswithcode.com/paper/a-case-for-backward-compatibility-for-human
Repo
Framework

Dynamically Composing Domain-Data Selection with Clean-Data Selection by “Co-Curricular Learning” for Neural Machine Translation

Title Dynamically Composing Domain-Data Selection with Clean-Data Selection by “Co-Curricular Learning” for Neural Machine Translation
Authors Wei Wang, Isaac Caswell, Ciprian Chelba
Abstract Noise and domain are important aspects of data quality for neural machine translation. Existing research focus separately on domain-data selection, clean-data selection, or their static combination, leaving the dynamic interaction across them not explicitly examined. This paper introduces a “co-curricular learning” method to compose dynamic domain-data selection with dynamic clean-data selection, for transfer learning across both capabilities. We apply an EM-style optimization procedure to further refine the “co-curriculum”. Experiment results and analysis with two domains demonstrate the effectiveness of the method and the properties of data scheduled by the co-curriculum.
Tasks Machine Translation, Transfer Learning
Published 2019-06-03
URL https://arxiv.org/abs/1906.01130v1
PDF https://arxiv.org/pdf/1906.01130v1.pdf
PWC https://paperswithcode.com/paper/dynamically-composing-domain-data-selection
Repo
Framework

Improving Low-Resource Cross-lingual Document Retrieval by Reranking with Deep Bilingual Representations

Title Improving Low-Resource Cross-lingual Document Retrieval by Reranking with Deep Bilingual Representations
Authors Rui Zhang, Caitlin Westerfield, Sungrok Shim, Garrett Bingham, Alexander Fabbri, Neha Verma, William Hu, Dragomir Radev
Abstract In this paper, we propose to boost low-resource cross-lingual document retrieval performance with deep bilingual query-document representations. We match queries and documents in both source and target languages with four components, each of which is implemented as a term interaction-based deep neural network with cross-lingual word embeddings as input. By including query likelihood scores as extra features, our model effectively learns to rerank the retrieved documents by using a small number of relevance labels for low-resource language pairs. Due to the shared cross-lingual word embedding space, the model can also be directly applied to another language pair without any training label. Experimental results on the MATERIAL dataset show that our model outperforms the competitive translation-based baselines on English-Swahili, English-Tagalog, and English-Somali cross-lingual information retrieval tasks.
Tasks Information Retrieval, Word Embeddings
Published 2019-06-08
URL https://arxiv.org/abs/1906.03492v1
PDF https://arxiv.org/pdf/1906.03492v1.pdf
PWC https://paperswithcode.com/paper/improving-low-resource-cross-lingual-document
Repo
Framework

Novel Uncertainty Framework for Deep Learning Ensembles

Title Novel Uncertainty Framework for Deep Learning Ensembles
Authors Tal Kachman, Michal Moshkovitz, Michal Rosen-Zvi
Abstract Deep neural networks have become the default choice for many of the machine learning tasks such as classification and regression. Dropout, a method commonly used to improve the convergence of deep neural networks, generates an ensemble of thinned networks with extensive weight sharing. Recent studies that dropout can be viewed as an approximate variational inference in Gaussian processes, and used as a practical tool to obtain uncertainty estimates of the network. We propose a novel statistical mechanics based framework to dropout and use this framework to propose a new generic algorithm that focuses on estimates of the variance of the loss as measured by the ensemble of thinned networks. Our approach can be applied to a wide range of deep neural network architectures and machine learning tasks. In classification, this algorithm allows the generation of a don’t-know answer to be generated, which can increase the reliability of the classifier. Empirically we demonstrate state-of-the-art AUC results on publicly available benchmarks.
Tasks Gaussian Processes
Published 2019-04-09
URL http://arxiv.org/abs/1904.04917v1
PDF http://arxiv.org/pdf/1904.04917v1.pdf
PWC https://paperswithcode.com/paper/novel-uncertainty-framework-for-deep-learning
Repo
Framework

Robust Deep Gaussian Processes

Title Robust Deep Gaussian Processes
Authors Jeremias Knoblauch
Abstract This report provides an in-depth overview over the implications and novelty Generalized Variational Inference (GVI) (Knoblauch et al., 2019) brings to Deep Gaussian Processes (DGPs) (Damianou & Lawrence, 2013). Specifically, robustness to model misspecification as well as principled alternatives for uncertainty quantification are motivated with an information-geometric view. These modifications have clear interpretations and can be implemented in less than 100 lines of Python code. Most importantly, the corresponding empirical results show that DGPs can greatly benefit from the presented enhancements.
Tasks Gaussian Processes
Published 2019-04-04
URL https://arxiv.org/abs/1904.02303v2
PDF https://arxiv.org/pdf/1904.02303v2.pdf
PWC https://paperswithcode.com/paper/robust-deep-gaussian-processes
Repo
Framework

Quantifying the Effects of the 2008 Recession using the Zillow Dataset

Title Quantifying the Effects of the 2008 Recession using the Zillow Dataset
Authors Arunav Gupta, Lucas Nguyen, Camille Dunning, Ka Ming Chan
Abstract This report explores the use of Zillow’s housing metrics dataset to investigate the effects of the 2008 US subprime mortgage crisis on various US locales. We begin by exploring the causes of the recession and the metrics available to us in the dataset. We settle on using the Zillow Home Value Index (ZHVI) because it is seasonally adjusted and able to account for a variety of inventory factors. Then, we explore three methodologies for quantifying recession impact: (a) Principal Components Analysis, (b) Area Under Baseline, and (c) ARIMA modeling and Confidence Intervals. While PCA does not yield useable results, we ended up with six cities from both AUB and ARIMA analysis, the top 3 “losers” and “gainers” of the 2008 recession, as determined by each analysis. This gave us 12 cities in total. Finally, we tested the robustness of our analysis against three “common knowledge” metrics for the recession: geographic clustering, population trends, and unemployment rate. While we did find some overlap between the results of our analysis and geographic clustering, there was no positive regression outcome from comparing our methodologies to population trends and the unemployment rate.
Tasks
Published 2019-12-23
URL https://arxiv.org/abs/1912.11341v1
PDF https://arxiv.org/pdf/1912.11341v1.pdf
PWC https://paperswithcode.com/paper/quantifying-the-effects-of-the-2008-recession
Repo
Framework
comments powered by Disqus