Paper Group AWR 24
DeepTest: Automated Testing of Deep-Neural-Network-driven Autonomous Cars. A Distributed Frank-Wolfe Framework for Learning Low-Rank Matrices with the Trace Norm. Smooth Neighbors on Teacher Graphs for Semi-supervised Learning. Bayesian Nonlinear Support Vector Machines for Big Data. The biglasso Package: A Memory- and Computation-Efficient Solver …
DeepTest: Automated Testing of Deep-Neural-Network-driven Autonomous Cars
Title | DeepTest: Automated Testing of Deep-Neural-Network-driven Autonomous Cars |
Authors | Yuchi Tian, Kexin Pei, Suman Jana, Baishakhi Ray |
Abstract | Recent advances in Deep Neural Networks (DNNs) have led to the development of DNN-driven autonomous cars that, using sensors like camera, LiDAR, etc., can drive without any human intervention. Most major manufacturers including Tesla, GM, Ford, BMW, and Waymo/Google are working on building and testing different types of autonomous vehicles. The lawmakers of several US states including California, Texas, and New York have passed new legislation to fast-track the process of testing and deployment of autonomous vehicles on their roads. However, despite their spectacular progress, DNNs, just like traditional software, often demonstrate incorrect or unexpected corner case behaviors that can lead to potentially fatal collisions. Several such real-world accidents involving autonomous cars have already happened including one which resulted in a fatality. Most existing testing techniques for DNN-driven vehicles are heavily dependent on the manual collection of test data under different driving conditions which become prohibitively expensive as the number of test conditions increases. In this paper, we design, implement and evaluate DeepTest, a systematic testing tool for automatically detecting erroneous behaviors of DNN-driven vehicles that can potentially lead to fatal crashes. First, our tool is designed to automatically generated test cases leveraging real-world changes in driving conditions like rain, fog, lighting conditions, etc. DeepTest systematically explores different parts of the DNN logic by generating test inputs that maximize the numbers of activated neurons. DeepTest found thousands of erroneous behaviors under different realistic driving conditions (e.g., blurring, rain, fog, etc.) many of which lead to potentially fatal crashes in three top performing DNNs in the Udacity self-driving car challenge. |
Tasks | Autonomous Vehicles |
Published | 2017-08-28 |
URL | http://arxiv.org/abs/1708.08559v2 |
http://arxiv.org/pdf/1708.08559v2.pdf | |
PWC | https://paperswithcode.com/paper/deeptest-automated-testing-of-deep-neural |
Repo | https://github.com/ARiSE-Lab/deepTest |
Framework | none |
A Distributed Frank-Wolfe Framework for Learning Low-Rank Matrices with the Trace Norm
Title | A Distributed Frank-Wolfe Framework for Learning Low-Rank Matrices with the Trace Norm |
Authors | Wenjie Zheng, Aurélien Bellet, Patrick Gallinari |
Abstract | We consider the problem of learning a high-dimensional but low-rank matrix from a large-scale dataset distributed over several machines, where low-rankness is enforced by a convex trace norm constraint. We propose DFW-Trace, a distributed Frank-Wolfe algorithm which leverages the low-rank structure of its updates to achieve efficiency in time, memory and communication usage. The step at the heart of DFW-Trace is solved approximately using a distributed version of the power method. We provide a theoretical analysis of the convergence of DFW-Trace, showing that we can ensure sublinear convergence in expectation to an optimal solution with few power iterations per epoch. We implement DFW-Trace in the Apache Spark distributed programming framework and validate the usefulness of our approach on synthetic and real data, including the ImageNet dataset with high-dimensional features extracted from a deep neural network. |
Tasks | |
Published | 2017-12-20 |
URL | http://arxiv.org/abs/1712.07495v2 |
http://arxiv.org/pdf/1712.07495v2.pdf | |
PWC | https://paperswithcode.com/paper/a-distributed-frank-wolfe-framework-for |
Repo | https://github.com/WenjieZ/distributed-frank-wolfe |
Framework | none |
Smooth Neighbors on Teacher Graphs for Semi-supervised Learning
Title | Smooth Neighbors on Teacher Graphs for Semi-supervised Learning |
Authors | Yucen Luo, Jun Zhu, Mengxi Li, Yong Ren, Bo Zhang |
Abstract | The recently proposed self-ensembling methods have achieved promising results in deep semi-supervised learning, which penalize inconsistent predictions of unlabeled data under different perturbations. However, they only consider adding perturbations to each single data point, while ignoring the connections between data samples. In this paper, we propose a novel method, called Smooth Neighbors on Teacher Graphs (SNTG). In SNTG, a graph is constructed based on the predictions of the teacher model, i.e., the implicit self-ensemble of models. Then the graph serves as a similarity measure with respect to which the representations of “similar” neighboring points are learned to be smooth on the low-dimensional manifold. We achieve state-of-the-art results on semi-supervised learning benchmarks. The error rates are 9.89%, 3.99% for CIFAR-10 with 4000 labels, SVHN with 500 labels, respectively. In particular, the improvements are significant when the labels are fewer. For the non-augmented MNIST with only 20 labels, the error rate is reduced from previous 4.81% to 1.36%. Our method also shows robustness to noisy labels. |
Tasks | |
Published | 2017-11-01 |
URL | http://arxiv.org/abs/1711.00258v2 |
http://arxiv.org/pdf/1711.00258v2.pdf | |
PWC | https://paperswithcode.com/paper/smooth-neighbors-on-teacher-graphs-for-semi |
Repo | https://github.com/xinmei9322/SNTG |
Framework | none |
Bayesian Nonlinear Support Vector Machines for Big Data
Title | Bayesian Nonlinear Support Vector Machines for Big Data |
Authors | Florian Wenzel, Theo Galy-Fajou, Matthaeus Deutsch, Marius Kloft |
Abstract | We propose a fast inference method for Bayesian nonlinear support vector machines that leverages stochastic variational inference and inducing points. Our experiments show that the proposed method is faster than competing Bayesian approaches and scales easily to millions of data points. It provides additional features over frequentist competitors such as accurate predictive uncertainty estimates and automatic hyperparameter search. |
Tasks | |
Published | 2017-07-18 |
URL | http://arxiv.org/abs/1707.05532v1 |
http://arxiv.org/pdf/1707.05532v1.pdf | |
PWC | https://paperswithcode.com/paper/bayesian-nonlinear-support-vector-machines |
Repo | https://github.com/theogf/BayesianSVM |
Framework | tf |
The biglasso Package: A Memory- and Computation-Efficient Solver for Lasso Model Fitting with Big Data in R
Title | The biglasso Package: A Memory- and Computation-Efficient Solver for Lasso Model Fitting with Big Data in R |
Authors | Yaohui Zeng, Patrick Breheny |
Abstract | Penalized regression models such as the lasso have been extensively applied to analyzing high-dimensional data sets. However, due to memory limitations, existing R packages like glmnet and ncvreg are not capable of fitting lasso-type models for ultrahigh-dimensional, multi-gigabyte data sets that are increasingly seen in many areas such as genetics, genomics, biomedical imaging, and high-frequency finance. In this research, we implement an R package called biglasso that tackles this challenge. biglasso utilizes memory-mapped files to store the massive data on the disk, only reading data into memory when necessary during model fitting, and is thus able to handle out-of-core computation seamlessly. Moreover, it’s equipped with newly proposed, more efficient feature screening rules, which substantially accelerate the computation. Benchmarking experiments show that our biglasso package, as compared to existing popular ones like glmnet, is much more memory- and computation-efficient. We further analyze a 31 GB real data set on a laptop with only 16 GB RAM to demonstrate the out-of-core computation capability of biglasso in analyzing massive data sets that cannot be accommodated by existing R packages. |
Tasks | |
Published | 2017-01-20 |
URL | http://arxiv.org/abs/1701.05936v2 |
http://arxiv.org/pdf/1701.05936v2.pdf | |
PWC | https://paperswithcode.com/paper/the-biglasso-package-a-memory-and-computation |
Repo | https://github.com/YaohuiZeng/biglasso |
Framework | none |
DeepXplore: Automated Whitebox Testing of Deep Learning Systems
Title | DeepXplore: Automated Whitebox Testing of Deep Learning Systems |
Authors | Kexin Pei, Yinzhi Cao, Junfeng Yang, Suman Jana |
Abstract | Deep learning (DL) systems are increasingly deployed in safety- and security-critical domains including self-driving cars and malware detection, where the correctness and predictability of a system’s behavior for corner case inputs are of great importance. Existing DL testing depends heavily on manually labeled data and therefore often fails to expose erroneous behaviors for rare inputs. We design, implement, and evaluate DeepXplore, the first whitebox framework for systematically testing real-world DL systems. First, we introduce neuron coverage for systematically measuring the parts of a DL system exercised by test inputs. Next, we leverage multiple DL systems with similar functionality as cross-referencing oracles to avoid manual checking. Finally, we demonstrate how finding inputs for DL systems that both trigger many differential behaviors and achieve high neuron coverage can be represented as a joint optimization problem and solved efficiently using gradient-based search techniques. DeepXplore efficiently finds thousands of incorrect corner case behaviors (e.g., self-driving cars crashing into guard rails and malware masquerading as benign software) in state-of-the-art DL models with thousands of neurons trained on five popular datasets including ImageNet and Udacity self-driving challenge data. For all tested DL models, on average, DeepXplore generated one test input demonstrating incorrect behavior within one second while running only on a commodity laptop. We further show that the test inputs generated by DeepXplore can also be used to retrain the corresponding DL model to improve the model’s accuracy by up to 3%. |
Tasks | Malware Detection, Self-Driving Cars |
Published | 2017-05-18 |
URL | http://arxiv.org/abs/1705.06640v4 |
http://arxiv.org/pdf/1705.06640v4.pdf | |
PWC | https://paperswithcode.com/paper/deepxplore-automated-whitebox-testing-of-deep |
Repo | https://github.com/peikexin9/deepxplore |
Framework | tf |
Semantic Specialisation of Distributional Word Vector Spaces using Monolingual and Cross-Lingual Constraints
Title | Semantic Specialisation of Distributional Word Vector Spaces using Monolingual and Cross-Lingual Constraints |
Authors | Nikola Mrkšić, Ivan Vulić, Diarmuid Ó Séaghdha, Ira Leviant, Roi Reichart, Milica Gašić, Anna Korhonen, Steve Young |
Abstract | We present Attract-Repel, an algorithm for improving the semantic quality of word vectors by injecting constraints extracted from lexical resources. Attract-Repel facilitates the use of constraints from mono- and cross-lingual resources, yielding semantically specialised cross-lingual vector spaces. Our evaluation shows that the method can make use of existing cross-lingual lexicons to construct high-quality vector spaces for a plethora of different languages, facilitating semantic transfer from high- to lower-resource ones. The effectiveness of our approach is demonstrated with state-of-the-art results on semantic similarity datasets in six languages. We next show that Attract-Repel-specialised vectors boost performance in the downstream task of dialogue state tracking (DST) across multiple languages. Finally, we show that cross-lingual vector spaces produced by our algorithm facilitate the training of multilingual DST models, which brings further performance improvements. |
Tasks | Dialogue State Tracking, Semantic Similarity, Semantic Textual Similarity |
Published | 2017-06-01 |
URL | http://arxiv.org/abs/1706.00374v1 |
http://arxiv.org/pdf/1706.00374v1.pdf | |
PWC | https://paperswithcode.com/paper/semantic-specialisation-of-distributional |
Repo | https://github.com/zliucr/mixed-language-training |
Framework | pytorch |
Stable and Controllable Neural Texture Synthesis and Style Transfer Using Histogram Losses
Title | Stable and Controllable Neural Texture Synthesis and Style Transfer Using Histogram Losses |
Authors | Eric Risser, Pierre Wilmot, Connelly Barnes |
Abstract | Recently, methods have been proposed that perform texture synthesis and style transfer by using convolutional neural networks (e.g. Gatys et al. [2015,2016]). These methods are exciting because they can in some cases create results with state-of-the-art quality. However, in this paper, we show these methods also have limitations in texture quality, stability, requisite parameter tuning, and lack of user controls. This paper presents a multiscale synthesis pipeline based on convolutional neural networks that ameliorates these issues. We first give a mathematical explanation of the source of instabilities in many previous approaches. We then improve these instabilities by using histogram losses to synthesize textures that better statistically match the exemplar. We also show how to integrate localized style losses in our multiscale framework. These losses can improve the quality of large features, improve the separation of content and style, and offer artistic controls such as paint by numbers. We demonstrate that our approach offers improved quality, convergence in fewer iterations, and more stability over the optimization. |
Tasks | Style Transfer, Texture Synthesis |
Published | 2017-01-31 |
URL | http://arxiv.org/abs/1701.08893v2 |
http://arxiv.org/pdf/1701.08893v2.pdf | |
PWC | https://paperswithcode.com/paper/stable-and-controllable-neural-texture |
Repo | https://github.com/Asteur/paint_harmonization |
Framework | torch |
Comparing deep neural networks against humans: object recognition when the signal gets weaker
Title | Comparing deep neural networks against humans: object recognition when the signal gets weaker |
Authors | Robert Geirhos, David H. J. Janssen, Heiko H. Schütt, Jonas Rauber, Matthias Bethge, Felix A. Wichmann |
Abstract | Human visual object recognition is typically rapid and seemingly effortless, as well as largely independent of viewpoint and object orientation. Until very recently, animate visual systems were the only ones capable of this remarkable computational feat. This has changed with the rise of a class of computer vision algorithms called deep neural networks (DNNs) that achieve human-level classification performance on object recognition tasks. Furthermore, a growing number of studies report similarities in the way DNNs and the human visual system process objects, suggesting that current DNNs may be good models of human visual object recognition. Yet there clearly exist important architectural and processing differences between state-of-the-art DNNs and the primate visual system. The potential behavioural consequences of these differences are not well understood. We aim to address this issue by comparing human and DNN generalisation abilities towards image degradations. We find the human visual system to be more robust to image manipulations like contrast reduction, additive noise or novel eidolon-distortions. In addition, we find progressively diverging classification error-patterns between humans and DNNs when the signal gets weaker, indicating that there may still be marked differences in the way humans and current DNNs perform visual object recognition. We envision that our findings as well as our carefully measured and freely available behavioural datasets provide a new useful benchmark for the computer vision community to improve the robustness of DNNs and a motivation for neuroscientists to search for mechanisms in the brain that could facilitate this robustness. |
Tasks | Object Recognition |
Published | 2017-06-21 |
URL | http://arxiv.org/abs/1706.06969v2 |
http://arxiv.org/pdf/1706.06969v2.pdf | |
PWC | https://paperswithcode.com/paper/comparing-deep-neural-networks-against-humans |
Repo | https://github.com/rgeirhos/object-recognition |
Framework | none |
Insights into Analogy Completion from the Biomedical Domain
Title | Insights into Analogy Completion from the Biomedical Domain |
Authors | Denis Newman-Griffis, Albert M Lai, Eric Fosler-Lussier |
Abstract | Analogy completion has been a popular task in recent years for evaluating the semantic properties of word embeddings, but the standard methodology makes a number of assumptions about analogies that do not always hold, either in recent benchmark datasets or when expanding into other domains. Through an analysis of analogies in the biomedical domain, we identify three assumptions: that of a Single Answer for any given analogy, that the pairs involved describe the Same Relationship, and that each pair is Informative with respect to the other. We propose modifying the standard methodology to relax these assumptions by allowing for multiple correct answers, reporting MAP and MRR in addition to accuracy, and using multiple example pairs. We further present BMASS, a novel dataset for evaluating linguistic regularities in biomedical embeddings, and demonstrate that the relationships described in the dataset pose significant semantic challenges to current word embedding methods. |
Tasks | Word Embeddings |
Published | 2017-06-07 |
URL | http://arxiv.org/abs/1706.02241v1 |
http://arxiv.org/pdf/1706.02241v1.pdf | |
PWC | https://paperswithcode.com/paper/insights-into-analogy-completion-from-the |
Repo | https://github.com/OSU-slatelab/BMASS |
Framework | tf |
Long-term Blood Pressure Prediction with Deep Recurrent Neural Networks
Title | Long-term Blood Pressure Prediction with Deep Recurrent Neural Networks |
Authors | Peng Su, Xiao-Rong Ding, Yuan-Ting Zhang, Jing Liu, Fen Miao, Ni Zhao |
Abstract | Existing methods for arterial blood pressure (BP) estimation directly map the input physiological signals to output BP values without explicitly modeling the underlying temporal dependencies in BP dynamics. As a result, these models suffer from accuracy decay over a long time and thus require frequent calibration. In this work, we address this issue by formulating BP estimation as a sequence prediction problem in which both the input and target are temporal sequences. We propose a novel deep recurrent neural network (RNN) consisting of multilayered Long Short-Term Memory (LSTM) networks, which are incorporated with (1) a bidirectional structure to access larger-scale context information of input sequence, and (2) residual connections to allow gradients in deep RNN to propagate more effectively. The proposed deep RNN model was tested on a static BP dataset, and it achieved root mean square error (RMSE) of 3.90 and 2.66 mmHg for systolic BP (SBP) and diastolic BP (DBP) prediction respectively, surpassing the accuracy of traditional BP prediction models. On a multi-day BP dataset, the deep RNN achieved RMSE of 3.84, 5.25, 5.80 and 5.81 mmHg for the 1st day, 2nd day, 4th day and 6th month after the 1st day SBP prediction, and 1.80, 4.78, 5.0, 5.21 mmHg for corresponding DBP prediction, respectively, which outperforms all previous models with notable improvement. The experimental results suggest that modeling the temporal dependencies in BP dynamics significantly improves the long-term BP prediction accuracy. |
Tasks | Blood pressure estimation, Electrocardiography (ECG), Photoplethysmography (PPG) |
Published | 2017-05-12 |
URL | http://arxiv.org/abs/1705.04524v3 |
http://arxiv.org/pdf/1705.04524v3.pdf | |
PWC | https://paperswithcode.com/paper/long-term-blood-pressure-prediction-with-deep |
Repo | https://github.com/psu1/DeepRNN |
Framework | tf |
ChainerCV: a Library for Deep Learning in Computer Vision
Title | ChainerCV: a Library for Deep Learning in Computer Vision |
Authors | Yusuke Niitani, Toru Ogawa, Shunta Saito, Masaki Saito |
Abstract | Despite significant progress of deep learning in the field of computer vision, there has not been a software library that covers these methods in a unifying manner. We introduce ChainerCV, a software library that is intended to fill this gap. ChainerCV supports numerous neural network models as well as software components needed to conduct research in computer vision. These implementations emphasize simplicity, flexibility and good software engineering practices. The library is designed to perform on par with the results reported in published papers and its tools can be used as a baseline for future research in computer vision. Our implementation includes sophisticated models like Faster R-CNN and SSD, and covers tasks such as object detection and semantic segmentation. |
Tasks | Object Detection, Semantic Segmentation |
Published | 2017-08-28 |
URL | http://arxiv.org/abs/1708.08169v1 |
http://arxiv.org/pdf/1708.08169v1.pdf | |
PWC | https://paperswithcode.com/paper/chainercv-a-library-for-deep-learning-in |
Repo | https://github.com/chainer/chainercv |
Framework | none |
Probabilistic RGB-D Odometry based on Points, Lines and Planes Under Depth Uncertainty
Title | Probabilistic RGB-D Odometry based on Points, Lines and Planes Under Depth Uncertainty |
Authors | Pedro F. Proenca, Yang Gao |
Abstract | This work proposes a robust visual odometry method for structured environments that combines point features with line and plane segments, extracted through an RGB-D camera. Noisy depth maps are processed by a probabilistic depth fusion framework based on Mixtures of Gaussians to denoise and derive the depth uncertainty, which is then propagated throughout the visual odometry pipeline. Probabilistic 3D plane and line fitting solutions are used to model the uncertainties of the feature parameters and pose is estimated by combining the three types of primitives based on their uncertainties. Performance evaluation on RGB-D sequences collected in this work and two public RGB-D datasets: TUM and ICL-NUIM show the benefit of using the proposed depth fusion framework and combining the three feature-types, particularly in scenes with low-textured surfaces, dynamic objects and missing depth measurements. |
Tasks | Visual Odometry |
Published | 2017-06-13 |
URL | http://arxiv.org/abs/1706.04034v3 |
http://arxiv.org/pdf/1706.04034v3.pdf | |
PWC | https://paperswithcode.com/paper/probabilistic-rgb-d-odometry-based-on-points |
Repo | https://github.com/pedropro/OMG_Depth_Fusion |
Framework | none |
Multi-View Data Generation Without View Supervision
Title | Multi-View Data Generation Without View Supervision |
Authors | Mickaël Chen, Ludovic Denoyer, Thierry Artières |
Abstract | The development of high-dimensional generative models has recently gained a great surge of interest with the introduction of variational auto-encoders and generative adversarial neural networks. Different variants have been proposed where the underlying latent space is structured, for example, based on attributes describing the data to generate. We focus on a particular problem where one aims at generating samples corresponding to a number of objects under various views. We assume that the distribution of the data is driven by two independent latent factors: the content, which represents the intrinsic features of an object, and the view, which stands for the settings of a particular observation of that object. Therefore, we propose a generative model and a conditional variant built on such a disentangled latent space. This approach allows us to generate realistic samples corresponding to various objects in a high variety of views. Unlike many multi-view approaches, our model doesn’t need any supervision on the views but only on the content. Compared to other conditional generation approaches that are mostly based on binary or categorical attributes, we make no such assumption about the factors of variations. Our model can be used on problems with a huge, potentially infinite, number of categories. We experiment it on four image datasets on which we demonstrate the effectiveness of the model and its ability to generalize. |
Tasks | |
Published | 2017-11-01 |
URL | http://arxiv.org/abs/1711.00305v2 |
http://arxiv.org/pdf/1711.00305v2.pdf | |
PWC | https://paperswithcode.com/paper/multi-view-data-generation-without-view |
Repo | https://github.com/mickaelChen/GMV |
Framework | pytorch |
A Large Dimensional Study of Regularized Discriminant Analysis Classifiers
Title | A Large Dimensional Study of Regularized Discriminant Analysis Classifiers |
Authors | Khalil Elkhalil, Abla Kammoun, Romain Couillet, Tareq Y. Al-Naffouri, Mohamed-Slim Alouini |
Abstract | This article carries out a large dimensional analysis of standard regularized discriminant analysis classifiers designed on the assumption that data arise from a Gaussian mixture model with different means and covariances. The analysis relies on fundamental results from random matrix theory (RMT) when both the number of features and the cardinality of the training data within each class grow large at the same pace. Under mild assumptions, we show that the asymptotic classification error approaches a deterministic quantity that depends only on the means and covariances associated with each class as well as the problem dimensions. Such a result permits a better understanding of the performance of regularized discriminant analsysis, in practical large but finite dimensions, and can be used to determine and pre-estimate the optimal regularization parameter that minimizes the misclassification error probability. Despite being theoretically valid only for Gaussian data, our findings are shown to yield a high accuracy in predicting the performances achieved with real data sets drawn from the popular USPS data base, thereby making an interesting connection between theory and practice. |
Tasks | |
Published | 2017-11-01 |
URL | https://arxiv.org/abs/1711.00382v4 |
https://arxiv.org/pdf/1711.00382v4.pdf | |
PWC | https://paperswithcode.com/paper/a-large-dimensional-study-of-regularized |
Repo | https://github.com/KhalilElkhalil/Large-Dimensional-Discriminant-Analysis-Classifiers-with-Random-Matrix-Theory |
Framework | none |