Paper Group ANR 1133
Separable Convolutional LSTMs for Faster Video Segmentation. Risk-Averse Action Selection Using Extreme Value Theory Estimates of the CVaR. Quantifying and Leveraging Classification Uncertainty for Chest Radiograph Assessment. An analysis of the cost of hyper-parameter selection via split-sample validation, with applications to penalized regression …
Separable Convolutional LSTMs for Faster Video Segmentation
Title | Separable Convolutional LSTMs for Faster Video Segmentation |
Authors | Andreas Pfeuffer, Klaus Dietmayer |
Abstract | Semantic Segmentation is an important module for autonomous robots such as self-driving cars. The advantage of video segmentation approaches compared to single image segmentation is that temporal image information is considered, and their performance increases due to this. Hence, single image segmentation approaches are extended by recurrent units such as convolutional LSTM (convLSTM) cells, which are placed at suitable positions in the basic network architecture. However, a major critique of video segmentation approaches based on recurrent neural networks is their large parameter count and their computational complexity, and so, their inference time of one video frame takes up to 66 percent longer than their basic version. Inspired by the success of the spatial and depthwise separable convolutional neural networks, we generalize these techniques for convLSTMs in this work, so that the number of parameters and the required FLOPs are reduced significantly. Experiments on different datasets show that the segmentation approaches using the proposed, modified convLSTM cells achieve similar or slightly worse accuracy, but are up to 15 percent faster on a GPU than the ones using the standard convLSTM cells. Furthermore, a new evaluation metric is introduced, which measures the amount of flickering pixels in the segmented video sequence. |
Tasks | Self-Driving Cars, Semantic Segmentation, Video Semantic Segmentation |
Published | 2019-07-16 |
URL | https://arxiv.org/abs/1907.06876v1 |
https://arxiv.org/pdf/1907.06876v1.pdf | |
PWC | https://paperswithcode.com/paper/separable-convolutional-lstms-for-faster |
Repo | |
Framework | |
Risk-Averse Action Selection Using Extreme Value Theory Estimates of the CVaR
Title | Risk-Averse Action Selection Using Extreme Value Theory Estimates of the CVaR |
Authors | Dylan Troop, Frédéric Godin, Jia Yuan Yu |
Abstract | The Conditional Value-at-Risk (CVaR) is a useful risk measure in machine learning, finance, insurance, energy, etc. When the CVaR confidence parameter is very high, estimation by sample averaging exhibits high variance due to the limited number of samples above the corresponding threshold. To mitigate this problem, we present an estimation procedure for the CVaR that combines extreme value theory and a recently introduced method of automated threshold selection by Bader et al. (2018). Under appropriate conditions, we estimate the tail risk using a generalized Pareto distribution. We compare empirically this estimation procedure with the naive method of sample averaging, and show an improvement in accuracy for some specific cases. We also show how the estimation procedure can be used in reinforcement learning by applying our method to the multi-armed bandit problem where the goal is to avoid catastrophic risk. |
Tasks | |
Published | 2019-12-03 |
URL | https://arxiv.org/abs/1912.01718v1 |
https://arxiv.org/pdf/1912.01718v1.pdf | |
PWC | https://paperswithcode.com/paper/risk-averse-action-selection-using-extreme |
Repo | |
Framework | |
Quantifying and Leveraging Classification Uncertainty for Chest Radiograph Assessment
Title | Quantifying and Leveraging Classification Uncertainty for Chest Radiograph Assessment |
Authors | Florin C. Ghesu, Bogdan Georgescu, Eli Gibson, Sebastian Guendel, Mannudeep K. Kalra, Ramandeep Singh, Subba R. Digumarthy, Sasa Grbic, Dorin Comaniciu |
Abstract | The interpretation of chest radiographs is an essential task for the detection of thoracic diseases and abnormalities. However, it is a challenging problem with high inter-rater variability and inherent ambiguity due to inconclusive evidence in the data, limited data quality or subjective definitions of disease appearance. Current deep learning solutions for chest radiograph abnormality classification are typically limited to providing probabilistic predictions, relying on the capacity of learning models to adapt to the high degree of label noise and become robust to the enumerated causal factors. In practice, however, this leads to overconfident systems with poor generalization on unseen data. To account for this, we propose an automatic system that learns not only the probabilistic estimate on the presence of an abnormality, but also an explicit uncertainty measure which captures the confidence of the system in the predicted output. We argue that explicitly learning the classification uncertainty as an orthogonal measure to the predicted output, is essential to account for the inherent variability characteristic of this data. Experiments were conducted on two datasets of chest radiographs of over 85,000 patients. Sample rejection based on the predicted uncertainty can significantly improve the ROC-AUC, e.g., by 8% to 0.91 with an expected rejection rate of under 25%. Eliminating training samples using uncertainty-driven bootstrapping, enables a significant increase in robustness and accuracy. In addition, we present a multi-reader study showing that the predictive uncertainty is indicative of reader errors. |
Tasks | |
Published | 2019-06-18 |
URL | https://arxiv.org/abs/1906.07775v1 |
https://arxiv.org/pdf/1906.07775v1.pdf | |
PWC | https://paperswithcode.com/paper/quantifying-and-leveraging-classification |
Repo | |
Framework | |
An analysis of the cost of hyper-parameter selection via split-sample validation, with applications to penalized regression
Title | An analysis of the cost of hyper-parameter selection via split-sample validation, with applications to penalized regression |
Authors | Jean Feng, Noah Simon |
Abstract | In the regression setting, given a set of hyper-parameters, a model-estimation procedure constructs a model from training data. The optimal hyper-parameters that minimize generalization error of the model are usually unknown. In practice they are often estimated using split-sample validation. Up to now, there is an open question regarding how the generalization error of the selected model grows with the number of hyper-parameters to be estimated. To answer this question, we establish finite-sample oracle inequalities for selection based on a single training/test split and based on cross-validation. We show that if the model-estimation procedures are smoothly parameterized by the hyper-parameters, the error incurred from tuning hyper-parameters shrinks at nearly a parametric rate. Hence for semi- and non-parametric model-estimation procedures with a fixed number of hyper-parameters, this additional error is negligible. For parametric model-estimation procedures, adding a hyper-parameter is roughly equivalent to adding a parameter to the model itself. In addition, we specialize these ideas for penalized regression problems with multiple penalty parameters. We establish that the fitted models are Lipschitz in the penalty parameters and thus our oracle inequalities apply. This result encourages development of regularization methods with many penalty parameters. |
Tasks | |
Published | 2019-03-28 |
URL | http://arxiv.org/abs/1903.12297v1 |
http://arxiv.org/pdf/1903.12297v1.pdf | |
PWC | https://paperswithcode.com/paper/an-analysis-of-the-cost-of-hyper-parameter |
Repo | |
Framework | |
Experimentation on the motion of an obstacle avoiding robot
Title | Experimentation on the motion of an obstacle avoiding robot |
Authors | Rakhmanov Ochilbek, Nzurumike Obianuju, Amina Sani, Rukayya Umar |
Abstract | An intelligent robot can be used for applications where a human is at significant risk (like nuclear, space, military), the economics or menial nature of the application result in inefficient use of human workers (service industry, agriculture), for humanitarian uses where there is great risk (demining an area of land mines, urban search and rescue). This paper implements an experiment on one of important fields of AI Searching Algorithms, to find shortest possible solution by searching the produced tree. We will concentrate on Hill climbing algorithm, which is one of simplest searching algorithms in AI. This algorithm is one of most suitable searching methods to help expert system to make decision at every state, at every node. The experimental robot will traverse the maze by using sensors plugged on it. The robot used is E.V.3 Lego Mind storms, with native software for programming LabView. The reason we chose this robot is that it interacts quickly with sensors and can be reconstructed in many ways. This programmed robot will calculate the best possibilities to find way out of maze. The maze is made of wood, and it is adjustable, as robot should be able to leave the maze in any design. |
Tasks | |
Published | 2019-07-25 |
URL | https://arxiv.org/abs/1907.11021v1 |
https://arxiv.org/pdf/1907.11021v1.pdf | |
PWC | https://paperswithcode.com/paper/experimentation-on-the-motion-of-an-obstacle |
Repo | |
Framework | |
Soft-Bayes: Prod for Mixtures of Experts with Log-Loss
Title | Soft-Bayes: Prod for Mixtures of Experts with Log-Loss |
Authors | Laurent Orseau, Tor Lattimore, Shane Legg |
Abstract | We consider prediction with expert advice under the log-loss with the goal of deriving efficient and robust algorithms. We argue that existing algorithms such as exponentiated gradient, online gradient descent and online Newton step do not adequately satisfy both requirements. Our main contribution is an analysis of the Prod algorithm that is robust to any data sequence and runs in linear time relative to the number of experts in each round. Despite the unbounded nature of the log-loss, we derive a bound that is independent of the largest loss and of the largest gradient, and depends only on the number of experts and the time horizon. Furthermore we give a Bayesian interpretation of Prod and adapt the algorithm to derive a tracking regret. |
Tasks | |
Published | 2019-01-08 |
URL | http://arxiv.org/abs/1901.02230v1 |
http://arxiv.org/pdf/1901.02230v1.pdf | |
PWC | https://paperswithcode.com/paper/soft-bayes-prod-for-mixtures-of-experts-with |
Repo | |
Framework | |
Image Super Resolution via Bilinear Pooling: Application to Confocal Endomicroscopy
Title | Image Super Resolution via Bilinear Pooling: Application to Confocal Endomicroscopy |
Authors | Saeed Izadi, Darren Sutton, Ghassan Hamarneh |
Abstract | Recent developments in image acquisition literature have miniaturized the confocal laser endomicroscopes to improve usability and flexibility of the apparatus in actual clinical settings. However, miniaturized devices collect less light and have fewer optical components, resulting in pixelation artifacts and low resolution images. Owing to the strength of deep networks, many supervised methods known as super resolution have achieved considerable success in restoring low resolution images by generating the missing high frequency details. In this work, we propose a novel attention mechanism that, for the first time, combines 1st- and 2nd-order statistics for pooling operation, in the spatial and channel-wise dimensions. We compare the efficacy of our method to 11 other existing single image super resolution techniques that compensate for the reduction in image quality caused by the necessity of endomicroscope miniaturization. All evaluations are carried out on three publicly available datasets. Experimental results show that our method can produce competitive results against state-of-the-art in terms of PSNR, SSIM, and IFC metrics. Additionally, our proposed method contains small number of parameters, which makes it lightweight and fast for real-time applications. |
Tasks | Image Super-Resolution, Super-Resolution |
Published | 2019-06-18 |
URL | https://arxiv.org/abs/1906.07802v2 |
https://arxiv.org/pdf/1906.07802v2.pdf | |
PWC | https://paperswithcode.com/paper/image-super-resolution-via-bilinear-pooling |
Repo | |
Framework | |
Kidney Recognition in CT Using YOLOv3
Title | Kidney Recognition in CT Using YOLOv3 |
Authors | Andréanne Lemay |
Abstract | Organ localization can be challenging considering the heterogeneity of medical images and the biological diversity from one individual to another. The contribution of this paper is to overview the performance of the object detection model, YOLOv3, on kidney localization in 2D and in 3D from CT scans. The model obtained a 0.851 Dice score in 2D and 0.742 in 3D. The SSD, a similar state-of-the-art object detection model, showed similar scores on the test set. YOLOv3 and SSD demonstrated the ability to detect kidneys on a wide variety of CT scans including patients suffering from different renal conditions. |
Tasks | Object Detection |
Published | 2019-10-03 |
URL | https://arxiv.org/abs/1910.01268v1 |
https://arxiv.org/pdf/1910.01268v1.pdf | |
PWC | https://paperswithcode.com/paper/kidney-recognition-in-ct-using-yolov3 |
Repo | |
Framework | |
Using ConceptNet to Teach Common Sense to an Automated Theorem Prover
Title | Using ConceptNet to Teach Common Sense to an Automated Theorem Prover |
Authors | Claudia Schon, Sophie Siebert, Frieder Stolzenburg |
Abstract | The CoRg system is a system to solve commonsense reasoning problems. The core of the CoRg system is the automated theorem prover Hyper that is fed with large amounts of background knowledge. This background knowledge plays a crucial role in solving commonsense reasoning problems. In this paper we present different ways to use knowledge graphs as background knowledge and discuss challenges that arise. |
Tasks | Common Sense Reasoning, Knowledge Graphs |
Published | 2019-12-30 |
URL | https://arxiv.org/abs/1912.12957v1 |
https://arxiv.org/pdf/1912.12957v1.pdf | |
PWC | https://paperswithcode.com/paper/using-conceptnet-to-teach-common-sense-to-an |
Repo | |
Framework | |
ODE-Inspired Analysis for the Biological Version of Oja’s Rule in Solving Streaming PCA
Title | ODE-Inspired Analysis for the Biological Version of Oja’s Rule in Solving Streaming PCA |
Authors | Chi-Ning Chou, Mien Brabeeba Wang |
Abstract | Oja’s rule [Oja, Journal of mathematical biology 1982] is a well-known biologically-plausible algorithm using a Hebbian-type synaptic update rule to solve streaming principal component analysis (PCA). Computational neuroscientists have known that this biological version of Oja’s rule converges to the top eigenvector of the covariance matrix of the input in the limit. However, prior to this work, it was open to prove any convergence rate guarantee. In this work, we give the first convergence rate analysis for the biological version of Oja’s rule in solving streaming PCA. Moreover, our convergence rate matches the information theoretical lower bound up to logarithmic factors and outperforms the state-of-the-art upper bound for streaming PCA. Furthermore, we develop a novel framework inspired by ordinary differential equations (ODE) to analyze general stochastic dynamics. The framework abandons the traditional step-by-step analysis and instead analyzes a stochastic dynamic in one-shot by giving a closed-form solution to the entire dynamic. The one-shot framework allows us to apply stopping time and martingale techniques to have a flexible and precise control on the dynamic. We believe that this general framework is powerful and should lead to effective yet simple analysis for a large class of problems with stochastic dynamics. |
Tasks | |
Published | 2019-11-04 |
URL | https://arxiv.org/abs/1911.02363v1 |
https://arxiv.org/pdf/1911.02363v1.pdf | |
PWC | https://paperswithcode.com/paper/ode-inspired-analysis-for-the-biological |
Repo | |
Framework | |
A Unified Analytical Framework for Trustable Machine Learning and Automation Running with Blockchain
Title | A Unified Analytical Framework for Trustable Machine Learning and Automation Running with Blockchain |
Authors | Tao Wang |
Abstract | Traditional machine learning algorithms use data from databases that are mutable, and therefore the data cannot be fully trusted. Also, the machine learning process is difficult to automate. This paper proposes building a trustable machine learning system by using blockchain technology, which can store data in a permanent and immutable way. In addition, smart contracts are used to automate the machine learning process. This paper makes three contributions. First, it establishes a link between machine learning technology and blockchain technology. Previously, machine learning and blockchain have been considered two independent technologies without an obvious link. Second, it proposes a unified analytical framework for trustable machine learning by using blockchain technology. This unified framework solves both the trustability and automation issues in machine learning. Third, it enables a computer to translate core machine learning implementation from a single thread on a single machine to multiple threads on multiple machines running with blockchain by using a unified approach. The paper uses association rule mining as an example to demonstrate how trustable machine learning can be implemented with blockchain, and it shows how this approach can be used to analyze opioid prescriptions to help combat the opioid crisis. |
Tasks | |
Published | 2019-03-21 |
URL | http://arxiv.org/abs/1903.08801v1 |
http://arxiv.org/pdf/1903.08801v1.pdf | |
PWC | https://paperswithcode.com/paper/a-unified-analytical-framework-for-trustable |
Repo | |
Framework | |
ST-UNet: A Spatio-Temporal U-Network for Graph-structured Time Series Modeling
Title | ST-UNet: A Spatio-Temporal U-Network for Graph-structured Time Series Modeling |
Authors | Bing Yu, Haoteng Yin, Zhanxing Zhu |
Abstract | The spatio-temporal graph learning is becoming an increasingly important object of graph study. Many application domains involve highly dynamic graphs where temporal information is crucial, e.g. traffic networks and financial transaction graphs. Despite the constant progress made on learning structured data, there is still a lack of effective means to extract dynamic complex features from spatio-temporal structures. Particularly, conventional models such as convolutional networks or recurrent neural networks are incapable of revealing the temporal patterns in short or long terms and exploring the spatial properties in local or global scope from spatio-temporal graphs simultaneously. To tackle this problem, we design a novel multi-scale architecture, Spatio-Temporal U-Net (ST-UNet), for graph-structured time series modeling. In this U-shaped network, a paired sampling operation is proposed in spacetime domain accordingly: the pooling (ST-Pool) coarsens the input graph in spatial from its deterministic partition while abstracts multi-resolution temporal dependencies through dilated recurrent skip connections; based on previous settings in the downsampling, the unpooling (ST-Unpool) restores the original structure of spatio-temporal graphs and resumes regular intervals within graph sequences. Experiments on spatio-temporal prediction tasks demonstrate that our model effectively captures comprehensive features in multiple scales and achieves substantial improvements over mainstream methods on several real-world datasets. |
Tasks | Time Series, Traffic Prediction |
Published | 2019-03-13 |
URL | http://arxiv.org/abs/1903.05631v1 |
http://arxiv.org/pdf/1903.05631v1.pdf | |
PWC | https://paperswithcode.com/paper/st-unet-a-spatio-temporal-u-network-for-graph |
Repo | |
Framework | |
Cross-Resolution Face Recognition via Prior-Aided Face Hallucination and Residual Knowledge Distillation
Title | Cross-Resolution Face Recognition via Prior-Aided Face Hallucination and Residual Knowledge Distillation |
Authors | Hanyang Kong, Jian Zhao, Xiaoguang Tu, Junliang Xing, Shengmei Shen, Jiashi Feng |
Abstract | Recent deep learning based face recognition methods have achieved great performance, but it still remains challenging to recognize very low-resolution query face like 28x28 pixels when CCTV camera is far from the captured subject. Such face with very low-resolution is totally out of detail information of the face identity compared to normal resolution in a gallery and hard to find corresponding faces therein. To this end, we propose a Resolution Invariant Model (RIM) for addressing such cross-resolution face recognition problems, with three distinct novelties. First, RIM is a novel and unified deep architecture, containing a Face Hallucination sub-Net (FHN) and a Heterogeneous Recognition sub-Net (HRN), which are jointly learned end to end. Second, FHN is a well-designed tri-path Generative Adversarial Network (GAN) which simultaneously perceives facial structure and geometry prior information, i.e. landmark heatmaps and parsing maps, incorporated with an unsupervised cross-domain adversarial training strategy to super-resolve very low-resolution query image to its 8x larger ones without requiring them to be well aligned. Third, HRN is a generic Convolutional Neural Network (CNN) for heterogeneous face recognition with our proposed residual knowledge distillation strategy for learning discriminative yet generalized feature representation. Quantitative and qualitative experiments on several benchmarks demonstrate the superiority of the proposed model over the state-of-the-arts. Codes and models will be released upon acceptance. |
Tasks | Face Hallucination, Face Recognition, Heterogeneous Face Recognition |
Published | 2019-05-26 |
URL | https://arxiv.org/abs/1905.10777v1 |
https://arxiv.org/pdf/1905.10777v1.pdf | |
PWC | https://paperswithcode.com/paper/cross-resolution-face-recognition-via-prior |
Repo | |
Framework | |
Carving out the low surface brightness universe with NoiseChisel
Title | Carving out the low surface brightness universe with NoiseChisel |
Authors | Mohammad Akhlaghi |
Abstract | NoiseChisel is a program to detect very low signal-to-noise ratio (S/N) features with minimal assumptions on their morphology. It was introduced in 2015 and released within a collection of data analysis programs and libraries known as GNU Astronomy Utilities (Gnuastro). Over the last ten stable releases of Gnuastro, NoiseChisel has significantly improved: detecting even fainter signal, enabling better user control over its inner workings, and many bug fixes. The most important change may be that NoiseChisel’s segmentation features have been moved into a new program called Segment. Another major change is the final growth strategy of its true detections, for example NoiseChisel is able to detect the outer wings of M51 down to S/N of 0.25, or 28.27 mag/arcsec2 on a single-exposure SDSS image (r-band). Segment is also able to detect the localized HII regions as “clumps” much more successfully. Finally, to orchestrate a controlled analysis, the concept of a “reproducible paper” is discussed: this paper itself is exactly reproducible (snapshot v4-0-g8505cfd). |
Tasks | |
Published | 2019-09-24 |
URL | https://arxiv.org/abs/1909.11230v1 |
https://arxiv.org/pdf/1909.11230v1.pdf | |
PWC | https://paperswithcode.com/paper/carving-out-the-low-surface-brightness |
Repo | |
Framework | |
Low-Power Neuromorphic Hardware for Signal Processing Applications
Title | Low-Power Neuromorphic Hardware for Signal Processing Applications |
Authors | Bipin Rajendran, Abu Sebastian, Michael Schmuker, Narayan Srinivasa, Evangelos Eleftheriou |
Abstract | Machine learning has emerged as the dominant tool for implementing complex cognitive tasks that require supervised, unsupervised, and reinforcement learning. While the resulting machines have demonstrated in some cases even super-human performance, their energy consumption has often proved to be prohibitive in the absence of costly super-computers. Most state-of-the-art machine learning solutions are based on memory-less models of neurons. This is unlike the neurons in the human brain, which encode and process information using temporal information in spike events. The different computing principles underlying biological neurons and how they combine together to efficiently process information is believed to be a key factor behind their superior efficiency compared to current machine learning systems. Inspired by the time-encoding mechanism used by the brain, third generation spiking neural networks (SNNs) are being studied for building a new class of information processing engines. Modern computing systems based on the von Neumann architecture, however, are ill-suited for efficiently implementing SNNs, since their performance is limited by the need to constantly shuttle data between physically separated logic and memory units. Hence, novel computational architectures that address the von Neumann bottleneck are necessary in order to build systems that can implement SNNs with low energy budgets. In this paper, we review some of the architectural and system level design aspects involved in developing a new class of brain-inspired information processing engines that mimic the time-based information encoding and processing aspects of the brain. |
Tasks | |
Published | 2019-01-11 |
URL | https://arxiv.org/abs/1901.03690v3 |
https://arxiv.org/pdf/1901.03690v3.pdf | |
PWC | https://paperswithcode.com/paper/low-power-neuromorphic-hardware-for-signal |
Repo | |
Framework | |