May 6, 2019

3211 words 16 mins read

Paper Group ANR 259

Paper Group ANR 259

Cognitive Deep Machine Can Train Itself. TODIM and TOPSIS with Z-numbers. A General Framework for Density Based Time Series Clustering Exploiting a Novel Admissible Pruning Strategy. Deep Learning with Sets and Point Clouds. Optimized clothes segmentation to boost gender classification in unconstrained scenarios. Kernel Bandwidth Selection for SVDD …

Cognitive Deep Machine Can Train Itself

Title Cognitive Deep Machine Can Train Itself
Authors András Lőrincz, Máté Csákvári, Áron Fóthi, Zoltán Ádám Milacski, András Sárkány, Zoltán Tősér
Abstract Machine learning is making substantial progress in diverse applications. The success is mostly due to advances in deep learning. However, deep learning can make mistakes and its generalization abilities to new tasks are questionable. We ask when and how one can combine network outputs, when (i) details of the observations are evaluated by learned deep components and (ii) facts and confirmation rules are available in knowledge based systems. We show that in limited contexts the required number of training samples can be low and self-improvement of pre-trained networks in more general context is possible. We argue that the combination of sparse outlier detection with deep components that can support each other diminish the fragility of deep methods, an important requirement for engineering applications. We argue that supervised learning of labels may be fully eliminated under certain conditions: a component based architecture together with a knowledge based system can train itself and provide high quality answers. We demonstrate these concepts on the State Farm Distracted Driver Detection benchmark. We argue that the view of the Study Panel (2016) may overestimate the requirements on years of focused research' and careful, unique construction’ for `AI systems’. |
Tasks Outlier Detection
Published 2016-12-02
URL http://arxiv.org/abs/1612.00745v1
PDF http://arxiv.org/pdf/1612.00745v1.pdf
PWC https://paperswithcode.com/paper/cognitive-deep-machine-can-train-itself
Repo
Framework

TODIM and TOPSIS with Z-numbers

Title TODIM and TOPSIS with Z-numbers
Authors R. A. Krohling, Artem dos Santos, A. G. C. Pacheco
Abstract In this paper, we present an approach that is able to handle with Z-numbers in the context of Multi-Criteria Decision Making (MCDM) problems. Z-numbers are composed of two parts, the first one is a restriction on the values that can be assumed, and the second part is the reliability of the information. As human beings we communicate with other people by means of natural language using sentences like: the journey time from home to university takes about half hour, very likely. Firstly, Z-numbers are converted to fuzzy numbers using a standard procedure. Next, the Z-TODIM and Z-TOPSIS are presented as a direct extension of the fuzzy TODIM and fuzzy TOPSIS, respectively. The proposed methods are applied to two case studies and compared with the standard approach using crisp values. Results obtained show the feasibility of the approach. In addition, a graphical interface was built to handle with both methods Z- TODIM and Z-TOPSIS allowing ease of use for user in other areas of knowledge.
Tasks Decision Making
Published 2016-09-19
URL http://arxiv.org/abs/1609.05705v1
PDF http://arxiv.org/pdf/1609.05705v1.pdf
PWC https://paperswithcode.com/paper/todim-and-topsis-with-z-numbers
Repo
Framework

A General Framework for Density Based Time Series Clustering Exploiting a Novel Admissible Pruning Strategy

Title A General Framework for Density Based Time Series Clustering Exploiting a Novel Admissible Pruning Strategy
Authors Nurjahan Begum, Liudmila Ulanova, Hoang Anh Dau, Jun Wang, Eamonn Keogh
Abstract Time Series Clustering is an important subroutine in many higher-level data mining analyses, including data editing for classifiers, summarization, and outlier detection. It is well known that for similarity search the superiority of Dynamic Time Warping (DTW) over Euclidean distance gradually diminishes as we consider ever larger datasets. However, as we shall show, the same is not true for clustering. Clustering time series under DTW remains a computationally expensive operation. In this work, we address this issue in two ways. We propose a novel pruning strategy that exploits both the upper and lower bounds to prune off a very large fraction of the expensive distance calculations. This pruning strategy is admissible and gives us provably identical results to the brute force algorithm, but is at least an order of magnitude faster. For datasets where even this level of speedup is inadequate, we show that we can use a simple heuristic to order the unavoidable calculations in a most-useful-first ordering, thus casting the clustering into an anytime framework. We demonstrate the utility of our ideas with both single and multidimensional case studies in the domains of astronomy, speech physiology, medicine and entomology. In addition, we show the generality of our clustering framework to other domains by efficiently obtaining semantically significant clusters in protein sequences using the Edit Distance, the discrete data analogue of DTW.
Tasks Outlier Detection, Time Series, Time Series Clustering
Published 2016-12-02
URL http://arxiv.org/abs/1612.00637v1
PDF http://arxiv.org/pdf/1612.00637v1.pdf
PWC https://paperswithcode.com/paper/a-general-framework-for-density-based-time
Repo
Framework

Deep Learning with Sets and Point Clouds

Title Deep Learning with Sets and Point Clouds
Authors Siamak Ravanbakhsh, Jeff Schneider, Barnabas Poczos
Abstract We introduce a simple permutation equivariant layer for deep learning with set structure.This type of layer, obtained by parameter-sharing, has a simple implementation and linear-time complexity in the size of each set. We use deep permutation-invariant networks to perform point-could classification and MNIST-digit summation, where in both cases the output is invariant to permutations of the input. In a semi-supervised setting, where the goal is make predictions for each instance within a set, we demonstrate the usefulness of this type of layer in set-outlier detection as well as semi-supervised learning with clustering side-information.
Tasks Outlier Detection
Published 2016-11-14
URL http://arxiv.org/abs/1611.04500v3
PDF http://arxiv.org/pdf/1611.04500v3.pdf
PWC https://paperswithcode.com/paper/deep-learning-with-sets-and-point-clouds
Repo
Framework

Optimized clothes segmentation to boost gender classification in unconstrained scenarios

Title Optimized clothes segmentation to boost gender classification in unconstrained scenarios
Authors D. Freire-Obregón, M. Castrillón-Santana, J. Lorenzo-Navarro
Abstract Several applications require demographic information of ordinary people in unconstrained scenarios. This is not a trivial task due to significant human appearance variations. In this work, we introduce trixels for clustering image regions, enumerating their advantages compared to superpixels. The classical GrabCut algorithm is later modified to segment trixels instead of pixels in an unsupervised context. Combining with face detection lead us to a clothes segmentation approach close to real time. The study uses the challenging Pascal VOC dataset for segmentation evaluation experiments. A final experiment analyzes the fusion of clothes features with state-of-the-art gender classifiers in ClothesDB, revealing a significant performance improvement in gender classification.
Tasks Face Detection
Published 2016-11-12
URL http://arxiv.org/abs/1611.03999v1
PDF http://arxiv.org/pdf/1611.03999v1.pdf
PWC https://paperswithcode.com/paper/optimized-clothes-segmentation-to-boost
Repo
Framework

Kernel Bandwidth Selection for SVDD: Peak Criterion Approach for Large Data

Title Kernel Bandwidth Selection for SVDD: Peak Criterion Approach for Large Data
Authors Sergiy Peredriy, Deovrat Kakde, Arin Chaudhuri
Abstract Support Vector Data Description (SVDD) provides a useful approach to construct a description of multivariate data for single-class classification and outlier detection with various practical applications. Gaussian kernel used in SVDD formulation allows flexible data description defined by observations designated as support vectors. The data boundary of such description is non-spherical and conforms to the geometric features of the data. By varying the Gaussian kernel bandwidth parameter, the SVDD-generated boundary can be made either smoother (more spherical) or tighter/jagged. The former case may lead to under-fitting, whereas the latter may result in overfitting. Peak criterion has been proposed to select an optimal value of the kernel bandwidth to strike the balance between the data boundary smoothness and its ability to capture the general geometric shape of the data. Peak criterion involves training SVDD at various values of the kernel bandwidth parameter. When training datasets are large, the time required to obtain the optimal value of the Gaussian kernel bandwidth parameter according to Peak method can become prohibitively large. This paper proposes an extension of Peak method for the case of large data. The proposed method gives good results when applied to several datasets. Two existing alternative methods of computing the Gaussian kernel bandwidth parameter (Coefficient of Variation and Distance to the Farthest Neighbor) were modified to allow comparison with the proposed method on convergence. Empirical comparison demonstrates the advantage of the proposed method.
Tasks Outlier Detection
Published 2016-10-31
URL http://arxiv.org/abs/1611.00058v3
PDF http://arxiv.org/pdf/1611.00058v3.pdf
PWC https://paperswithcode.com/paper/kernel-bandwidth-selection-for-svdd-peak
Repo
Framework

Binary Codes for Tagging X-Ray Images via Deep De-Noising Autoencoders

Title Binary Codes for Tagging X-Ray Images via Deep De-Noising Autoencoders
Authors Antonio Sze-To, Hamid R. Tizhoosh, Andrew K. C. Wong
Abstract A Content-Based Image Retrieval (CBIR) system which identifies similar medical images based on a query image can assist clinicians for more accurate diagnosis. The recent CBIR research trend favors the construction and use of binary codes to represent images. Deep architectures could learn the non-linear relationship among image pixels adaptively, allowing the automatic learning of high-level features from raw pixels. However, most of them require class labels, which are expensive to obtain, particularly for medical images. The methods which do not need class labels utilize a deep autoencoder for binary hashing, but the code construction involves a specific training algorithm and an ad-hoc regularization technique. In this study, we explored using a deep de-noising autoencoder (DDA), with a new unsupervised training scheme using only backpropagation and dropout, to hash images into binary codes. We conducted experiments on more than 14,000 x-ray images. By using class labels only for evaluating the retrieval results, we constructed a 16-bit DDA and a 512-bit DDA independently. Comparing to other unsupervised methods, we succeeded to obtain the lowest total error by using the 512-bit codes for retrieval via exhaustive search, and speed up 9.27 times with the use of the 16-bit codes while keeping a comparable total error. We found that our new training scheme could reduce the total retrieval error significantly by 21.9%. To further boost the image retrieval performance, we developed Radon Autoencoder Barcode (RABC) which are learned from the Radon projections of images using a de-noising autoencoder. Experimental results demonstrated its superior performance in retrieval when it was combined with DDA binary codes.
Tasks Content-Based Image Retrieval, Image Retrieval
Published 2016-04-24
URL http://arxiv.org/abs/1604.07060v1
PDF http://arxiv.org/pdf/1604.07060v1.pdf
PWC https://paperswithcode.com/paper/binary-codes-for-tagging-x-ray-images-via
Repo
Framework

Detecting facial landmarks in the video based on a hybrid framework

Title Detecting facial landmarks in the video based on a hybrid framework
Authors Nian Cai, Zhineng Lin, Fu Zhang, Guandong Cen, Han Wang
Abstract To dynamically detect the facial landmarks in the video, we propose a novel hybrid framework termed as detection-tracking-detection (DTD). First, the face bounding box is achieved from the first frame of the video sequence based on a traditional face detection method. Then, a landmark detector detects the facial landmarks, which is based on a cascaded deep convolution neural network (DCNN). Next, the face bounding box in the current frame is estimated and validated after the facial landmarks in the previous frame are tracked based on the median flow. Finally, the facial landmarks in the current frame are exactly detected from the validated face bounding box via the landmark detector. Experimental results indicate that the proposed framework can detect the facial landmarks in the video sequence more effectively and with lower consuming time compared to the frame-by-frame method via the DCNN.
Tasks Face Detection
Published 2016-09-21
URL http://arxiv.org/abs/1609.06441v1
PDF http://arxiv.org/pdf/1609.06441v1.pdf
PWC https://paperswithcode.com/paper/detecting-facial-landmarks-in-the-video-based
Repo
Framework

Depth Estimation from Single Image using Sparse Representations

Title Depth Estimation from Single Image using Sparse Representations
Authors Yigit Oktar
Abstract Monocular depth estimation is an interesting and challenging problem as there is no analytic mapping known between an intensity image and its depth map. Recently there has been a lot of data accumulated through depth-sensing cameras, in parallel to that researchers started to tackle this task using various learning algorithms. In this paper, a deep sparse coding method is proposed for monocular depth estimation along with an approach for deterministic dictionary initialization.
Tasks Depth Estimation, Monocular Depth Estimation
Published 2016-06-27
URL http://arxiv.org/abs/1606.08315v1
PDF http://arxiv.org/pdf/1606.08315v1.pdf
PWC https://paperswithcode.com/paper/depth-estimation-from-single-image-using
Repo
Framework

On the Use of Deep Learning for Blind Image Quality Assessment

Title On the Use of Deep Learning for Blind Image Quality Assessment
Authors Simone Bianco, Luigi Celona, Paolo Napoletano, Raimondo Schettini
Abstract In this work we investigate the use of deep learning for distortion-generic blind image quality assessment. We report on different design choices, ranging from the use of features extracted from pre-trained Convolutional Neural Networks (CNNs) as a generic image description, to the use of features extracted from a CNN fine-tuned for the image quality task. Our best proposal, named DeepBIQ, estimates the image quality by average pooling the scores predicted on multiple sub-regions of the original image. The score of each sub-region is computed using a Support Vector Regression (SVR) machine taking as input features extracted using a CNN fine-tuned for category-based image quality assessment. Experimental results on the LIVE In the Wild Image Quality Challenge Database and on the LIVE Image Quality Assessment Database show that DeepBIQ outperforms the state-of-the-art methods compared, having a Linear Correlation Coefficient (LCC) with human subjective scores of almost 0.91 and 0.98 respectively. Furthermore, in most of the cases, the quality score predictions of DeepBIQ are closer to the average observer than those of a generic human observer.
Tasks Blind Image Quality Assessment, Image Quality Assessment
Published 2016-02-17
URL http://arxiv.org/abs/1602.05531v5
PDF http://arxiv.org/pdf/1602.05531v5.pdf
PWC https://paperswithcode.com/paper/on-the-use-of-deep-learning-for-blind-image
Repo
Framework

Automatic Sleep Stage Scoring with Single-Channel EEG Using Convolutional Neural Networks

Title Automatic Sleep Stage Scoring with Single-Channel EEG Using Convolutional Neural Networks
Authors Orestis Tsinalis, Paul M. Matthews, Yike Guo, Stefanos Zafeiriou
Abstract We used convolutional neural networks (CNNs) for automatic sleep stage scoring based on single-channel electroencephalography (EEG) to learn task-specific filters for classification without using prior domain knowledge. We used an openly available dataset from 20 healthy young adults for evaluation and applied 20-fold cross-validation. We used class-balanced random sampling within the stochastic gradient descent (SGD) optimization of the CNN to avoid skewed performance in favor of the most represented sleep stages. We achieved high mean F1-score (81%, range 79-83%), mean accuracy across individual sleep stages (82%, range 80-84%) and overall accuracy (74%, range 71-76%) over all subjects. By analyzing and visualizing the filters that our CNN learns, we found that rules learned by the filters correspond to sleep scoring criteria in the American Academy of Sleep Medicine (AASM) manual that human experts follow. Our method’s performance is balanced across classes and our results are comparable to state-of-the-art methods with hand-engineered features. We show that, without using prior domain knowledge, a CNN can automatically learn to distinguish among different normal sleep stages.
Tasks EEG
Published 2016-10-05
URL http://arxiv.org/abs/1610.01683v1
PDF http://arxiv.org/pdf/1610.01683v1.pdf
PWC https://paperswithcode.com/paper/automatic-sleep-stage-scoring-with-single
Repo
Framework

Automatic Lymphocyte Detection in H&E Images with Deep Neural Networks

Title Automatic Lymphocyte Detection in H&E Images with Deep Neural Networks
Authors Jianxu Chen, Chukka Srinivas
Abstract Automatic detection of lymphocyte in H&E images is a necessary first step in lots of tissue image analysis algorithms. An accurate and robust automated lymphocyte detection approach is of great importance in both computer science and clinical studies. Most of the existing approaches for lymphocyte detection are based on traditional image processing algorithms and/or classic machine learning methods. In the recent years, deep learning techniques have fundamentally transformed the way that a computer interprets images and have become a matchless solution in various pattern recognition problems. In this work, we design a new deep neural network model which extends the fully convolutional network by combining the ideas in several recent techniques, such as shortcut links. Also, we design a new training scheme taking the prior knowledge about lymphocytes into consideration. The training scheme not only efficiently exploits the limited amount of free-form annotations from pathologists, but also naturally supports efficient fine-tuning. As a consequence, our model has the potential of self-improvement by leveraging the errors collected during real applications. Our experiments show that our deep neural network model achieves good performance in the images of different staining conditions or different types of tissues.
Tasks
Published 2016-12-09
URL http://arxiv.org/abs/1612.03217v1
PDF http://arxiv.org/pdf/1612.03217v1.pdf
PWC https://paperswithcode.com/paper/automatic-lymphocyte-detection-in-he-images
Repo
Framework

Camera Pose Estimation from Lines using Plücker Coordinates

Title Camera Pose Estimation from Lines using Plücker Coordinates
Authors Bronislav Přibyl, Pavel Zemčík, Martin Čadík
Abstract Correspondences between 3D lines and their 2D images captured by a camera are often used to determine position and orientation of the camera in space. In this work, we propose a novel algebraic algorithm to estimate the camera pose. We parameterize 3D lines using Pl"ucker coordinates that allow linear projection of the lines into the image. A line projection matrix is estimated using Linear Least Squares and the camera pose is then extracted from the matrix. An algebraic approach to handle mismatched line correspondences is also included. The proposed algorithm is an order of magnitude faster yet comparably accurate and robust to the state-of-the-art, it does not require initialization, and it yields only one solution. The described method requires at least 9 lines and is particularly suitable for scenarios with 25 and more lines, as also shown in the results.
Tasks Pose Estimation
Published 2016-08-09
URL http://arxiv.org/abs/1608.02824v1
PDF http://arxiv.org/pdf/1608.02824v1.pdf
PWC https://paperswithcode.com/paper/camera-pose-estimation-from-lines-using
Repo
Framework

Faster Sublinear Algorithms using Conditional Sampling

Title Faster Sublinear Algorithms using Conditional Sampling
Authors Themistoklis Gouleakis, Christos Tzamos, Manolis Zampetakis
Abstract A conditional sampling oracle for a probability distribution D returns samples from the conditional distribution of D restricted to a specified subset of the domain. A recent line of work (Chakraborty et al. 2013 and Cannone et al. 2014) has shown that having access to such a conditional sampling oracle requires only polylogarithmic or even constant number of samples to solve distribution testing problems like identity and uniformity. This significantly improves over the standard sampling model where polynomially many samples are necessary. Inspired by these results, we introduce a computational model based on conditional sampling to develop sublinear algorithms with exponentially faster runtimes compared to standard sublinear algorithms. We focus on geometric optimization problems over points in high dimensional Euclidean space. Access to these points is provided via a conditional sampling oracle that takes as input a succinct representation of a subset of the domain and outputs a uniformly random point in that subset. We study two well studied problems: k-means clustering and estimating the weight of the minimum spanning tree. In contrast to prior algorithms for the classic model, our algorithms have time, space and sample complexity that is polynomial in the dimension and polylogarithmic in the number of points. Finally, we comment on the applicability of the model and compare with existing ones like streaming, parallel and distributed computational models.
Tasks
Published 2016-08-16
URL http://arxiv.org/abs/1608.04759v1
PDF http://arxiv.org/pdf/1608.04759v1.pdf
PWC https://paperswithcode.com/paper/faster-sublinear-algorithms-using-conditional
Repo
Framework

Deep Learning for the Classification of Lung Nodules

Title Deep Learning for the Classification of Lung Nodules
Authors He Yang, Hengyong Yu, Ge Wang
Abstract Deep learning, as a promising new area of machine learning, has attracted a rapidly increasing attention in the field of medical imaging. Compared to the conventional machine learning methods, deep learning requires no hand-tuned feature extractor, and has shown a superior performance in many visual object recognition applications. In this study, we develop a deep convolutional neural network (CNN) and apply it to thoracic CT images for the classification of lung nodules. We present the CNN architecture and classification accuracy for the original images of lung nodules. In order to understand the features of lung nodules, we further construct new datasets, based on the combination of artificial geometric nodules and some transformations of the original images, as well as a stochastic nodule shape model. It is found that simplistic geometric nodules cannot capture the important features of lung nodules.
Tasks Object Recognition
Published 2016-11-21
URL http://arxiv.org/abs/1611.06651v2
PDF http://arxiv.org/pdf/1611.06651v2.pdf
PWC https://paperswithcode.com/paper/deep-learning-for-the-classification-of-lung
Repo
Framework
comments powered by Disqus