Paper Group ANR 277
Screen Content Image Segmentation Using Sparse Decomposition and Total Variation Minimization. Secure Approximation Guarantee for Cryptographically Private Empirical Risk Minimization. A Benchmark and Comparison of Active Learning for Logistic Regression. Convex Formulation for Kernel PCA and its Use in Semi-Supervised Learning. Harmonic Networks: …
Screen Content Image Segmentation Using Sparse Decomposition and Total Variation Minimization
Title | Screen Content Image Segmentation Using Sparse Decomposition and Total Variation Minimization |
Authors | Shervin Minaee, Yao Wang |
Abstract | Sparse decomposition has been widely used for different applications, such as source separation, image classification, image denoising and more. This paper presents a new algorithm for segmentation of an image into background and foreground text and graphics using sparse decomposition and total variation minimization. The proposed method is designed based on the assumption that the background part of the image is smoothly varying and can be represented by a linear combination of a few smoothly varying basis functions, while the foreground text and graphics can be modeled with a sparse component overlaid on the smooth background. The background and foreground are separated using a sparse decomposition framework regularized with a few suitable regularization terms which promotes the sparsity and connectivity of foreground pixels. This algorithm has been tested on a dataset of images extracted from HEVC standard test sequences for screen content coding, and is shown to have superior performance over some prior methods, including least absolute deviation fitting, k-means clustering based segmentation in DjVu and shape primitive extraction and coding (SPEC) algorithm. |
Tasks | Denoising, Image Classification, Image Denoising, Semantic Segmentation |
Published | 2016-02-07 |
URL | http://arxiv.org/abs/1602.02434v2 |
http://arxiv.org/pdf/1602.02434v2.pdf | |
PWC | https://paperswithcode.com/paper/screen-content-image-segmentation-using-1 |
Repo | |
Framework | |
Secure Approximation Guarantee for Cryptographically Private Empirical Risk Minimization
Title | Secure Approximation Guarantee for Cryptographically Private Empirical Risk Minimization |
Authors | Toshiyuki Takada, Hiroyuki Hanada, Yoshiji Yamada, Jun Sakuma, Ichiro Takeuchi |
Abstract | Privacy concern has been increasingly important in many machine learning (ML) problems. We study empirical risk minimization (ERM) problems under secure multi-party computation (MPC) frameworks. Main technical tools for MPC have been developed based on cryptography. One of limitations in current cryptographically private ML is that it is computationally intractable to evaluate non-linear functions such as logarithmic functions or exponential functions. Therefore, for a class of ERM problems such as logistic regression in which non-linear function evaluations are required, one can only obtain approximate solutions. In this paper, we introduce a novel cryptographically private tool called secure approximation guarantee (SAG) method. The key property of SAG method is that, given an arbitrary approximate solution, it can provide a non-probabilistic assumption-free bound on the approximation quality under cryptographically secure computation framework. We demonstrate the benefit of the SAG method by applying it to several problems including a practical privacy-preserving data analysis task on genomic and clinical information. |
Tasks | |
Published | 2016-02-15 |
URL | http://arxiv.org/abs/1602.04579v1 |
http://arxiv.org/pdf/1602.04579v1.pdf | |
PWC | https://paperswithcode.com/paper/secure-approximation-guarantee-for |
Repo | |
Framework | |
A Benchmark and Comparison of Active Learning for Logistic Regression
Title | A Benchmark and Comparison of Active Learning for Logistic Regression |
Authors | Yazhou Yang, Marco Loog |
Abstract | Logistic regression is by far the most widely used classifier in real-world applications. In this paper, we benchmark the state-of-the-art active learning methods for logistic regression and discuss and illustrate their underlying characteristics. Experiments are carried out on three synthetic datasets and 44 real-world datasets, providing insight into the behaviors of these active learning methods with respect to the area of the learning curve (which plots classification accuracy as a function of the number of queried examples) and their computational costs. Surprisingly, one of the earliest and simplest suggested active learning methods, i.e., uncertainty sampling, performs exceptionally well overall. Another remarkable finding is that random sampling, which is the rudimentary baseline to improve upon, is not overwhelmed by individual active learning techniques in many cases. |
Tasks | Active Learning |
Published | 2016-11-25 |
URL | http://arxiv.org/abs/1611.08618v2 |
http://arxiv.org/pdf/1611.08618v2.pdf | |
PWC | https://paperswithcode.com/paper/a-benchmark-and-comparison-of-active-learning |
Repo | |
Framework | |
Convex Formulation for Kernel PCA and its Use in Semi-Supervised Learning
Title | Convex Formulation for Kernel PCA and its Use in Semi-Supervised Learning |
Authors | Carlos M. Alaíz, Michaël Fanuel, Johan A. K. Suykens |
Abstract | In this paper, Kernel PCA is reinterpreted as the solution to a convex optimization problem. Actually, there is a constrained convex problem for each principal component, so that the constraints guarantee that the principal component is indeed a solution, and not a mere saddle point. Although these insights do not imply any algorithmic improvement, they can be used to further understand the method, formulate possible extensions and properly address them. As an example, a new convex optimization problem for semi-supervised classification is proposed, which seems particularly well-suited whenever the number of known labels is small. Our formulation resembles a Least Squares SVM problem with a regularization parameter multiplied by a negative sign, combined with a variational principle for Kernel PCA. Our primal optimization principle for semi-supervised learning is solved in terms of the Lagrange multipliers. Numerical experiments in several classification tasks illustrate the performance of the proposed model in problems with only a few labeled data. |
Tasks | |
Published | 2016-10-21 |
URL | http://arxiv.org/abs/1610.06811v1 |
http://arxiv.org/pdf/1610.06811v1.pdf | |
PWC | https://paperswithcode.com/paper/convex-formulation-for-kernel-pca-and-its-use |
Repo | |
Framework | |
Harmonic Networks: Deep Translation and Rotation Equivariance
Title | Harmonic Networks: Deep Translation and Rotation Equivariance |
Authors | Daniel E. Worrall, Stephan J. Garbin, Daniyar Turmukhambetov, Gabriel J. Brostow |
Abstract | Translating or rotating an input image should not affect the results of many computer vision tasks. Convolutional neural networks (CNNs) are already translation equivariant: input image translations produce proportionate feature map translations. This is not the case for rotations. Global rotation equivariance is typically sought through data augmentation, but patch-wise equivariance is more difficult. We present Harmonic Networks or H-Nets, a CNN exhibiting equivariance to patch-wise translation and 360-rotation. We achieve this by replacing regular CNN filters with circular harmonics, returning a maximal response and orientation for every receptive field patch. H-Nets use a rich, parameter-efficient and low computational complexity representation, and we show that deep feature maps within the network encode complicated rotational invariants. We demonstrate that our layers are general enough to be used in conjunction with the latest architectures and techniques, such as deep supervision and batch normalization. We also achieve state-of-the-art classification on rotated-MNIST, and competitive results on other benchmark challenges. |
Tasks | Data Augmentation |
Published | 2016-12-14 |
URL | http://arxiv.org/abs/1612.04642v2 |
http://arxiv.org/pdf/1612.04642v2.pdf | |
PWC | https://paperswithcode.com/paper/harmonic-networks-deep-translation-and |
Repo | |
Framework | |
Inverting The Generator Of A Generative Adversarial Network
Title | Inverting The Generator Of A Generative Adversarial Network |
Authors | Antonia Creswell, Anil Anthony Bharath |
Abstract | Generative adversarial networks (GANs) learn to synthesise new samples from a high-dimensional distribution by passing samples drawn from a latent space through a generative network. When the high-dimensional distribution describes images of a particular data set, the network should learn to generate visually similar image samples for latent variables that are close to each other in the latent space. For tasks such as image retrieval and image classification, it may be useful to exploit the arrangement of the latent space by projecting images into it, and using this as a representation for discriminative tasks. GANs often consist of multiple layers of non-linear computations, making them very difficult to invert. This paper introduces techniques for projecting image samples into the latent space using any pre-trained GAN, provided that the computational graph is available. We evaluate these techniques on both MNIST digits and Omniglot handwritten characters. In the case of MNIST digits, we show that projections into the latent space maintain information about the style and the identity of the digit. In the case of Omniglot characters, we show that even characters from alphabets that have not been seen during training may be projected well into the latent space; this suggests that this approach may have applications in one-shot learning. |
Tasks | Image Classification, Image Retrieval, Omniglot, One-Shot Learning |
Published | 2016-11-17 |
URL | http://arxiv.org/abs/1611.05644v1 |
http://arxiv.org/pdf/1611.05644v1.pdf | |
PWC | https://paperswithcode.com/paper/inverting-the-generator-of-a-generative |
Repo | |
Framework | |
Masking Strategies for Image Manifolds
Title | Masking Strategies for Image Manifolds |
Authors | Hamid Dadkhahi, Marco F. Duarte |
Abstract | We consider the problem of selecting an optimal mask for an image manifold, i.e., choosing a subset of the pixels of the image that preserves the manifold’s geometric structure present in the original data. Such masking implements a form of compressive sensing through emerging imaging sensor platforms for which the power expense grows with the number of pixels acquired. Our goal is for the manifold learned from masked images to resemble its full image counterpart as closely as possible. More precisely, we show that one can indeed accurately learn an image manifold without having to consider a large majority of the image pixels. In doing so, we consider two masking methods that preserve the local and global geometric structure of the manifold, respectively. In each case, the process of finding the optimal masking pattern can be cast as a binary integer program, which is computationally expensive but can be approximated by a fast greedy algorithm. Numerical experiments show that the relevant manifold structure is preserved through the data-dependent masking process, even for modest mask sizes. |
Tasks | Compressive Sensing |
Published | 2016-06-15 |
URL | http://arxiv.org/abs/1606.04618v1 |
http://arxiv.org/pdf/1606.04618v1.pdf | |
PWC | https://paperswithcode.com/paper/masking-strategies-for-image-manifolds |
Repo | |
Framework | |
Tensor Completion by Alternating Minimization under the Tensor Train (TT) Model
Title | Tensor Completion by Alternating Minimization under the Tensor Train (TT) Model |
Authors | Wenqi Wang, Vaneet Aggarwal, Shuchin Aeron |
Abstract | Using the matrix product state (MPS) representation of tensor train decompositions, in this paper we propose a tensor completion algorithm which alternates over the matrices (tensors) in the MPS representation. This development is motivated in part by the success of matrix completion algorithms which alternate over the (low-rank) factors. We comment on the computational complexity of the proposed algorithm and numerically compare it with existing methods employing low rank tensor train approximation for data completion as well as several other recently proposed methods. We show that our method is superior to existing ones for a variety of real settings. |
Tasks | Matrix Completion |
Published | 2016-09-19 |
URL | http://arxiv.org/abs/1609.05587v1 |
http://arxiv.org/pdf/1609.05587v1.pdf | |
PWC | https://paperswithcode.com/paper/tensor-completion-by-alternating-minimization |
Repo | |
Framework | |
Evaluating Causal Models by Comparing Interventional Distributions
Title | Evaluating Causal Models by Comparing Interventional Distributions |
Authors | Dan Garant, David Jensen |
Abstract | The predominant method for evaluating the quality of causal models is to measure the graphical accuracy of the learned model structure. We present an alternative method for evaluating causal models that directly measures the accuracy of estimated interventional distributions. We contrast such distributional measures with structural measures, such as structural Hamming distance and structural intervention distance, showing that structural measures often correspond poorly to the accuracy of estimated interventional distributions. We use a number of real and synthetic datasets to illustrate various scenarios in which structural measures provide misleading results with respect to algorithm selection and parameter tuning, and we recommend that distributional measures become the new standard for evaluating causal models. |
Tasks | |
Published | 2016-08-16 |
URL | http://arxiv.org/abs/1608.04698v1 |
http://arxiv.org/pdf/1608.04698v1.pdf | |
PWC | https://paperswithcode.com/paper/evaluating-causal-models-by-comparing |
Repo | |
Framework | |
Learning Attributes Equals Multi-Source Domain Generalization
Title | Learning Attributes Equals Multi-Source Domain Generalization |
Authors | Chuang Gan, Tianbao Yang, Boqing Gong |
Abstract | Attributes possess appealing properties and benefit many computer vision problems, such as object recognition, learning with humans in the loop, and image retrieval. Whereas the existing work mainly pursues utilizing attributes for various computer vision problems, we contend that the most basic problem—how to accurately and robustly detect attributes from images—has been left under explored. Especially, the existing work rarely explicitly tackles the need that attribute detectors should generalize well across different categories, including those previously unseen. Noting that this is analogous to the objective of multi-source domain generalization, if we treat each category as a domain, we provide a novel perspective to attribute detection and propose to gear the techniques in multi-source domain generalization for the purpose of learning cross-category generalizable attribute detectors. We validate our understanding and approach with extensive experiments on four challenging datasets and three different problems. |
Tasks | Domain Generalization, Image Retrieval, Object Recognition |
Published | 2016-05-03 |
URL | http://arxiv.org/abs/1605.00743v1 |
http://arxiv.org/pdf/1605.00743v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-attributes-equals-multi-source |
Repo | |
Framework | |
Theoretically-Grounded Policy Advice from Multiple Teachers in Reinforcement Learning Settings with Applications to Negative Transfer
Title | Theoretically-Grounded Policy Advice from Multiple Teachers in Reinforcement Learning Settings with Applications to Negative Transfer |
Authors | Yusen Zhan, Haitham Bou Ammar, Matthew E. taylor |
Abstract | Policy advice is a transfer learning method where a student agent is able to learn faster via advice from a teacher. However, both this and other reinforcement learning transfer methods have little theoretical analysis. This paper formally defines a setting where multiple teacher agents can provide advice to a student and introduces an algorithm to leverage both autonomous exploration and teacher’s advice. Our regret bounds justify the intuition that good teachers help while bad teachers hurt. Using our formalization, we are also able to quantify, for the first time, when negative transfer can occur within such a reinforcement learning setting. |
Tasks | Transfer Learning |
Published | 2016-04-13 |
URL | http://arxiv.org/abs/1604.03986v1 |
http://arxiv.org/pdf/1604.03986v1.pdf | |
PWC | https://paperswithcode.com/paper/theoretically-grounded-policy-advice-from |
Repo | |
Framework | |
Leveraging Mid-Level Deep Representations For Predicting Face Attributes in the Wild
Title | Leveraging Mid-Level Deep Representations For Predicting Face Attributes in the Wild |
Authors | Yang Zhong, Josephine Sullivan, Haibo Li |
Abstract | Predicting facial attributes from faces in the wild is very challenging due to pose and lighting variations in the real world. The key to this problem is to build proper feature representations to cope with these unfavourable conditions. Given the success of Convolutional Neural Network (CNN) in image classification, the high-level CNN feature, as an intuitive and reasonable choice, has been widely utilized for this problem. In this paper, however, we consider the mid-level CNN features as an alternative to the high-level ones for attribute prediction. This is based on the observation that face attributes are different: some of them are locally oriented while others are globally defined. Our investigations reveal that the mid-level deep representations outperform the prediction accuracy achieved by the (fine-tuned) high-level abstractions. We empirically demonstrate that the midlevel representations achieve state-of-the-art prediction performance on CelebA and LFWA datasets. Our investigations also show that by utilizing the mid-level representations one can employ a single deep network to achieve both face recognition and attribute prediction. |
Tasks | Face Recognition, Image Classification |
Published | 2016-02-04 |
URL | http://arxiv.org/abs/1602.01827v3 |
http://arxiv.org/pdf/1602.01827v3.pdf | |
PWC | https://paperswithcode.com/paper/leveraging-mid-level-deep-representations-for |
Repo | |
Framework | |
Shape-aware Surface Reconstruction from Sparse 3D Point-Clouds
Title | Shape-aware Surface Reconstruction from Sparse 3D Point-Clouds |
Authors | Florian Bernard, Luis Salamanca, Johan Thunberg, Alexander Tack, Dennis Jentsch, Hans Lamecker, Stefan Zachow, Frank Hertel, Jorge Goncalves, Peter Gemmar |
Abstract | The reconstruction of an object’s shape or surface from a set of 3D points plays an important role in medical image analysis, e.g. in anatomy reconstruction from tomographic measurements or in the process of aligning intra-operative navigation and preoperative planning data. In such scenarios, one usually has to deal with sparse data, which significantly aggravates the problem of reconstruction. However, medical applications often provide contextual information about the 3D point data that allow to incorporate prior knowledge about the shape that is to be reconstructed. To this end, we propose the use of a statistical shape model (SSM) as a prior for surface reconstruction. The SSM is represented by a point distribution model (PDM), which is associated with a surface mesh. Using the shape distribution that is modelled by the PDM, we formulate the problem of surface reconstruction from a probabilistic perspective based on a Gaussian Mixture Model (GMM). In order to do so, the given points are interpreted as samples of the GMM. By using mixture components with anisotropic covariances that are “oriented” according to the surface normals at the PDM points, a surface-based fitting is accomplished. Estimating the parameters of the GMM in a maximum a posteriori manner yields the reconstruction of the surface from the given data points. We compare our method to the extensively used Iterative Closest Points method on several different anatomical datasets/SSMs (brain, femur, tibia, hip, liver) and demonstrate superior accuracy and robustness on sparse data. |
Tasks | |
Published | 2016-02-26 |
URL | http://arxiv.org/abs/1602.08425v2 |
http://arxiv.org/pdf/1602.08425v2.pdf | |
PWC | https://paperswithcode.com/paper/shape-aware-surface-reconstruction-from |
Repo | |
Framework | |
A Review on Algorithms for Constraint-based Causal Discovery
Title | A Review on Algorithms for Constraint-based Causal Discovery |
Authors | Kui Yu, Jiuyong Li, Lin Liu |
Abstract | Causal discovery studies the problem of mining causal relationships between variables from data, which is of primary interest in science. During the past decades, significant amount of progresses have been made toward this fundamental data mining paradigm. Recent years, as the availability of abundant large-sized and complex observational data, the constrain-based approaches have gradually attracted a lot of interest and have been widely applied to many diverse real-world problems due to the fast running speed and easy generalizing to the problem of causal insufficiency. In this paper, we aim to review the constraint-based causal discovery algorithms. Firstly, we discuss the learning paradigm of the constraint-based approaches. Secondly and primarily, the state-of-the-art constraint-based casual inference algorithms are surveyed with the detailed analysis. Thirdly, several related open-source software packages and benchmark data repositories are briefly summarized. As a conclusion, some open problems in constraint-based causal discovery are outlined for future research. |
Tasks | Causal Discovery |
Published | 2016-11-12 |
URL | http://arxiv.org/abs/1611.03977v2 |
http://arxiv.org/pdf/1611.03977v2.pdf | |
PWC | https://paperswithcode.com/paper/a-review-on-algorithms-for-constraint-based |
Repo | |
Framework | |
Improved Optimistic Mirror Descent for Sparsity and Curvature
Title | Improved Optimistic Mirror Descent for Sparsity and Curvature |
Authors | Parameswaran Kamalaruban |
Abstract | Online Convex Optimization plays a key role in large scale machine learning. Early approaches to this problem were conservative, in which the main focus was protection against the worst case scenario. But recently several algorithms have been developed for tightening the regret bounds in easy data instances such as sparsity, predictable sequences, and curved losses. In this work we unify some of these existing techniques to obtain new update rules for the cases when these easy instances occur together. First we analyse an adaptive and optimistic update rule which achieves tighter regret bound when the loss sequence is sparse and predictable. Then we explain an update rule that dynamically adapts to the curvature of the loss function and utilizes the predictable nature of the loss sequence as well. Finally we extend these results to composite losses. |
Tasks | |
Published | 2016-09-08 |
URL | http://arxiv.org/abs/1609.02383v1 |
http://arxiv.org/pdf/1609.02383v1.pdf | |
PWC | https://paperswithcode.com/paper/improved-optimistic-mirror-descent-for |
Repo | |
Framework | |