Paper Group ANR 183
![Paper Group ANR 183](/2017/images/pwc/paper-arxiv_hu144ec288a26b3e360d673e256787de3e_28623_900x500_fit_q75_box.jpg)
Manifold-valued Image Generation with Wasserstein Generative Adversarial Nets. Estimating Mutual Information for Discrete-Continuous Mixtures. Exact Inference for Relational Graphical Models with Interpreted Functions: Lifted Probabilistic Inference Modulo Theories. View Independent Vehicle Make, Model and Color Recognition Using Convolutional Neur …
Manifold-valued Image Generation with Wasserstein Generative Adversarial Nets
Title | Manifold-valued Image Generation with Wasserstein Generative Adversarial Nets |
Authors | Zhiwu Huang, Jiqing Wu, Luc Van Gool |
Abstract | Generative modeling over natural images is one of the most fundamental machine learning problems. However, few modern generative models, including Wasserstein Generative Adversarial Nets (WGANs), are studied on manifold-valued images that are frequently encountered in real-world applications. To fill the gap, this paper first formulates the problem of generating manifold-valued images and exploits three typical instances: hue-saturation-value (HSV) color image generation, chromaticity-brightness (CB) color image generation, and diffusion-tensor (DT) image generation. For the proposed generative modeling problem, we then introduce a theorem of optimal transport to derive a new Wasserstein distance of data distributions on complete manifolds, enabling us to achieve a tractable objective under the WGAN framework. In addition, we recommend three benchmark datasets that are CIFAR-10 HSV/CB color images, ImageNet HSV/CB color images, UCL DT image datasets. On the three datasets, we experimentally demonstrate the proposed manifold-aware WGAN model can generate more plausible manifold-valued images than its competitors. |
Tasks | Image Generation |
Published | 2017-12-05 |
URL | http://arxiv.org/abs/1712.01551v2 |
http://arxiv.org/pdf/1712.01551v2.pdf | |
PWC | https://paperswithcode.com/paper/manifold-valued-image-generation-with |
Repo | |
Framework | |
Estimating Mutual Information for Discrete-Continuous Mixtures
Title | Estimating Mutual Information for Discrete-Continuous Mixtures |
Authors | Weihao Gao, Sreeram Kannan, Sewoong Oh, Pramod Viswanath |
Abstract | Estimating mutual information from observed samples is a basic primitive, useful in several machine learning tasks including correlation mining, information bottleneck clustering, learning a Chow-Liu tree, and conditional independence testing in (causal) graphical models. While mutual information is a well-defined quantity in general probability spaces, existing estimators can only handle two special cases of purely discrete or purely continuous pairs of random variables. The main challenge is that these methods first estimate the (differential) entropies of X, Y and the pair (X;Y) and add them up with appropriate signs to get an estimate of the mutual information. These 3H-estimators cannot be applied in general mixture spaces, where entropy is not well-defined. In this paper, we design a novel estimator for mutual information of discrete-continuous mixtures. We prove that the proposed estimator is consistent. We provide numerical experiments suggesting superiority of the proposed estimator compared to other heuristics of adding small continuous noise to all the samples and applying standard estimators tailored for purely continuous variables, and quantizing the samples and applying standard estimators tailored for purely discrete variables. This significantly widens the applicability of mutual information estimation in real-world applications, where some variables are discrete, some continuous, and others are a mixture between continuous and discrete components. |
Tasks | |
Published | 2017-09-19 |
URL | http://arxiv.org/abs/1709.06212v3 |
http://arxiv.org/pdf/1709.06212v3.pdf | |
PWC | https://paperswithcode.com/paper/estimating-mutual-information-for-discrete |
Repo | |
Framework | |
Exact Inference for Relational Graphical Models with Interpreted Functions: Lifted Probabilistic Inference Modulo Theories
Title | Exact Inference for Relational Graphical Models with Interpreted Functions: Lifted Probabilistic Inference Modulo Theories |
Authors | Rodrigo de Salvo Braz, Ciaran O’Reilly |
Abstract | Probabilistic Inference Modulo Theories (PIMT) is a recent framework that expands exact inference on graphical models to use richer languages that include arithmetic, equalities, and inequalities on both integers and real numbers. In this paper, we expand PIMT to a lifted version that also processes random functions and relations. This enhancement is achieved by adapting Inversion, a method from Lifted First-Order Probabilistic Inference literature, to also be modulo theories. This results in the first algorithm for exact probabilistic inference that efficiently and simultaneously exploits random relations and functions, arithmetic, equalities and inequalities. |
Tasks | |
Published | 2017-09-04 |
URL | http://arxiv.org/abs/1709.01122v1 |
http://arxiv.org/pdf/1709.01122v1.pdf | |
PWC | https://paperswithcode.com/paper/exact-inference-for-relational-graphical |
Repo | |
Framework | |
View Independent Vehicle Make, Model and Color Recognition Using Convolutional Neural Network
Title | View Independent Vehicle Make, Model and Color Recognition Using Convolutional Neural Network |
Authors | Afshin Dehghan, Syed Zain Masood, Guang Shu, Enrique. G. Ortiz |
Abstract | This paper describes the details of Sighthound’s fully automated vehicle make, model and color recognition system. The backbone of our system is a deep convolutional neural network that is not only computationally inexpensive, but also provides state-of-the-art results on several competitive benchmarks. Additionally, our deep network is trained on a large dataset of several million images which are labeled through a semi-automated process. Finally we test our system on several public datasets as well as our own internal test dataset. Our results show that we outperform other methods on all benchmarks by significant margins. Our model is available to developers through the Sighthound Cloud API at https://www.sighthound.com/products/cloud |
Tasks | |
Published | 2017-02-06 |
URL | http://arxiv.org/abs/1702.01721v1 |
http://arxiv.org/pdf/1702.01721v1.pdf | |
PWC | https://paperswithcode.com/paper/view-independent-vehicle-make-model-and-color |
Repo | |
Framework | |
Deep Learning-Guided Image Reconstruction from Incomplete Data
Title | Deep Learning-Guided Image Reconstruction from Incomplete Data |
Authors | Brendan Kelly, Thomas P. Matthews, Mark A. Anastasio |
Abstract | An approach to incorporate deep learning within an iterative image reconstruction framework to reconstruct images from severely incomplete measurement data is presented. Specifically, we utilize a convolutional neural network (CNN) as a quasi-projection operator within a least squares minimization procedure. The CNN is trained to encode high level information about the class of images being imaged; this information is utilized to mitigate artifacts in intermediate images produced by use of an iterative method. The structure of the method was inspired by the proximal gradient descent method, where the proximal operator is replaced by a deep CNN and the gradient descent step is generalized by use of a linear reconstruction operator. It is demonstrated that this approach improves image quality for several cases of limited-view image reconstruction and that using a CNN in an iterative method increases performance compared to conventional image reconstruction approaches. We test our method on several limited-view image reconstruction problems. Qualitative and quantitative results demonstrate state-of-the-art performance. |
Tasks | Image Reconstruction |
Published | 2017-09-02 |
URL | http://arxiv.org/abs/1709.00584v1 |
http://arxiv.org/pdf/1709.00584v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-learning-guided-image-reconstruction |
Repo | |
Framework | |
A Meta-Learning Approach to One-Step Active Learning
Title | A Meta-Learning Approach to One-Step Active Learning |
Authors | Gabriella Contardo, Ludovic Denoyer, Thierry Artieres |
Abstract | We consider the problem of learning when obtaining the training labels is costly, which is usually tackled in the literature using active-learning techniques. These approaches provide strategies to choose the examples to label before or during training. These strategies are usually based on heuristics or even theoretical measures, but are not learned as they are directly used during training. We design a model which aims at \textit{learning active-learning strategies} using a meta-learning setting. More specifically, we consider a pool-based setting, where the system observes all the examples of the dataset of a problem and has to choose the subset of examples to label in a single shot. Experiments show encouraging results. |
Tasks | Active Learning, Meta-Learning |
Published | 2017-06-26 |
URL | http://arxiv.org/abs/1706.08334v2 |
http://arxiv.org/pdf/1706.08334v2.pdf | |
PWC | https://paperswithcode.com/paper/a-meta-learning-approach-to-one-step-active |
Repo | |
Framework | |
Tackling Over-pruning in Variational Autoencoders
Title | Tackling Over-pruning in Variational Autoencoders |
Authors | Serena Yeung, Anitha Kannan, Yann Dauphin, Li Fei-Fei |
Abstract | Variational autoencoders (VAE) are directed generative models that learn factorial latent variables. As noted by Burda et al. (2015), these models exhibit the problem of factor over-pruning where a significant number of stochastic factors fail to learn anything and become inactive. This can limit their modeling power and their ability to learn diverse and meaningful latent representations. In this paper, we evaluate several methods to address this problem and propose a more effective model-based approach called the epitomic variational autoencoder (eVAE). The so-called epitomes of this model are groups of mutually exclusive latent factors that compete to explain the data. This approach helps prevent inactive units since each group is pressured to explain the data. We compare the approaches with qualitative and quantitative results on MNIST and TFD datasets. Our results show that eVAE makes efficient use of model capacity and generalizes better than VAE. |
Tasks | |
Published | 2017-06-09 |
URL | http://arxiv.org/abs/1706.03643v2 |
http://arxiv.org/pdf/1706.03643v2.pdf | |
PWC | https://paperswithcode.com/paper/tackling-over-pruning-in-variational |
Repo | |
Framework | |
Pixelwise Instance Segmentation with a Dynamically Instantiated Network
Title | Pixelwise Instance Segmentation with a Dynamically Instantiated Network |
Authors | Anurag Arnab, Philip H. S Torr |
Abstract | Semantic segmentation and object detection research have recently achieved rapid progress. However, the former task has no notion of different instances of the same object, and the latter operates at a coarse, bounding-box level. We propose an Instance Segmentation system that produces a segmentation map where each pixel is assigned an object class and instance identity label. Most approaches adapt object detectors to produce segments instead of boxes. In contrast, our method is based on an initial semantic segmentation module, which feeds into an instance subnetwork. This subnetwork uses the initial category-level segmentation, along with cues from the output of an object detector, within an end-to-end CRF to predict instances. This part of our model is dynamically instantiated to produce a variable number of instances per image. Our end-to-end approach requires no post-processing and considers the image holistically, instead of processing independent proposals. Therefore, unlike some related work, a pixel cannot belong to multiple instances. Furthermore, far more precise segmentations are achieved, as shown by our state-of-the-art results (particularly at high IoU thresholds) on the Pascal VOC and Cityscapes datasets. |
Tasks | Instance Segmentation, Object Detection, Panoptic Segmentation, Semantic Segmentation |
Published | 2017-04-07 |
URL | http://arxiv.org/abs/1704.02386v1 |
http://arxiv.org/pdf/1704.02386v1.pdf | |
PWC | https://paperswithcode.com/paper/pixelwise-instance-segmentation-with-a |
Repo | |
Framework | |
Adversarial Generation of Training Examples: Applications to Moving Vehicle License Plate Recognition
Title | Adversarial Generation of Training Examples: Applications to Moving Vehicle License Plate Recognition |
Authors | Xinlong Wang, Zhipeng Man, Mingyu You, Chunhua Shen |
Abstract | Generative Adversarial Networks (GAN) have attracted much research attention recently, leading to impressive results for natural image generation. However, to date little success was observed in using GAN generated images for improving classification tasks. Here we attempt to explore, in the context of car license plate recognition, whether it is possible to generate synthetic training data using GAN to improve recognition accuracy. With a carefully-designed pipeline, we show that the answer is affirmative. First, a large-scale image set is generated using the generator of GAN, without manual annotation. Then, these images are fed to a deep convolutional neural network (DCNN) followed by a bidirectional recurrent neural network (BRNN) with long short-term memory (LSTM), which performs the feature learning and sequence labelling. Finally, the pre-trained model is fine-tuned on real images. Our experimental results on a few data sets demonstrate the effectiveness of using GAN images: an improvement of 7.5% over a strong baseline with moderate-sized real data being available. We show that the proposed framework achieves competitive recognition accuracy on challenging test datasets. We also leverage the depthwise separate convolution to construct a lightweight convolutional RNN, which is about half size and 2x faster on CPU. Combining this framework and the proposed pipeline, we make progress in performing accurate recognition on mobile and embedded devices. |
Tasks | Image Generation, License Plate Recognition |
Published | 2017-07-11 |
URL | http://arxiv.org/abs/1707.03124v3 |
http://arxiv.org/pdf/1707.03124v3.pdf | |
PWC | https://paperswithcode.com/paper/adversarial-generation-of-training-examples |
Repo | |
Framework | |
Image Generation from Sketch Constraint Using Contextual GAN
Title | Image Generation from Sketch Constraint Using Contextual GAN |
Authors | Yongyi Lu, Shangzhe Wu, Yu-Wing Tai, Chi-Keung Tang |
Abstract | In this paper we investigate image generation guided by hand sketch. When the input sketch is badly drawn, the output of common image-to-image translation follows the input edges due to the hard condition imposed by the translation process. Instead, we propose to use sketch as weak constraint, where the output edges do not necessarily follow the input edges. We address this problem using a novel joint image completion approach, where the sketch provides the image context for completing, or generating the output image. We train a generated adversarial network, i.e, contextual GAN to learn the joint distribution of sketch and the corresponding image by using joint images. Our contextual GAN has several advantages. First, the simple joint image representation allows for simple and effective learning of joint distribution in the same image-sketch space, which avoids complicated issues in cross-domain learning. Second, while the output is related to its input overall, the generated features exhibit more freedom in appearance and do not strictly align with the input features as previous conditional GANs do. Third, from the joint image’s point of view, image and sketch are of no difference, thus exactly the same deep joint image completion network can be used for image-to-sketch generation. Experiments evaluated on three different datasets show that our contextual GAN can generate more realistic images than state-of-the-art conditional GANs on challenging inputs and generalize well on common categories. |
Tasks | Image Generation, Image-to-Image Translation |
Published | 2017-11-24 |
URL | http://arxiv.org/abs/1711.08972v2 |
http://arxiv.org/pdf/1711.08972v2.pdf | |
PWC | https://paperswithcode.com/paper/image-generation-from-sketch-constraint-using |
Repo | |
Framework | |
Submodular Function Maximization for Group Elevator Scheduling
Title | Submodular Function Maximization for Group Elevator Scheduling |
Authors | Srikumar Ramalingam, Arvind U. Raghunathan, Daniel Nikovski |
Abstract | We propose a novel approach for group elevator scheduling by formulating it as the maximization of submodular function under a matroid constraint. In particular, we propose to model the total waiting time of passengers using a quadratic Boolean function. The unary and pairwise terms in the function denote the waiting time for single and pairwise allocation of passengers to elevators, respectively. We show that this objective function is submodular. The matroid constraints ensure that every passenger is allocated to exactly one elevator. We use a greedy algorithm to maximize the submodular objective function, and derive provable guarantees on the optimality of the solution. We tested our algorithm using Elevate 8, a commercial-grade elevator simulator that allows simulation with a wide range of elevator settings. We achieve significant improvement over the existing algorithms. |
Tasks | |
Published | 2017-06-28 |
URL | http://arxiv.org/abs/1707.00617v1 |
http://arxiv.org/pdf/1707.00617v1.pdf | |
PWC | https://paperswithcode.com/paper/submodular-function-maximization-for-group |
Repo | |
Framework | |
Using Argument-based Features to Predict and Analyse Review Helpfulness
Title | Using Argument-based Features to Predict and Analyse Review Helpfulness |
Authors | Haijing Liu, Yang Gao, Pin Lv, Mengxue Li, Shiqiang Geng, Minglan Li, Hao Wang |
Abstract | We study the helpful product reviews identification problem in this paper. We observe that the evidence-conclusion discourse relations, also known as arguments, often appear in product reviews, and we hypothesise that some argument-based features, e.g. the percentage of argumentative sentences, the evidences-conclusions ratios, are good indicators of helpful reviews. To validate this hypothesis, we manually annotate arguments in 110 hotel reviews, and investigate the effectiveness of several combinations of argument-based features. Experiments suggest that, when being used together with the argument-based features, the state-of-the-art baseline features can enjoy a performance boost (in terms of F1) of 11.01% in average. |
Tasks | |
Published | 2017-07-23 |
URL | http://arxiv.org/abs/1707.07279v1 |
http://arxiv.org/pdf/1707.07279v1.pdf | |
PWC | https://paperswithcode.com/paper/using-argument-based-features-to-predict-and |
Repo | |
Framework | |
Greedy Algorithms for Cone Constrained Optimization with Convergence Guarantees
Title | Greedy Algorithms for Cone Constrained Optimization with Convergence Guarantees |
Authors | Francesco Locatello, Michael Tschannen, Gunnar Rätsch, Martin Jaggi |
Abstract | Greedy optimization methods such as Matching Pursuit (MP) and Frank-Wolfe (FW) algorithms regained popularity in recent years due to their simplicity, effectiveness and theoretical guarantees. MP and FW address optimization over the linear span and the convex hull of a set of atoms, respectively. In this paper, we consider the intermediate case of optimization over the convex cone, parametrized as the conic hull of a generic atom set, leading to the first principled definitions of non-negative MP algorithms for which we give explicit convergence rates and demonstrate excellent empirical performance. In particular, we derive sublinear ($\mathcal{O}(1/t)$) convergence on general smooth and convex objectives, and linear convergence ($\mathcal{O}(e^{-t})$) on strongly convex objectives, in both cases for general sets of atoms. Furthermore, we establish a clear correspondence of our algorithms to known algorithms from the MP and FW literature. Our novel algorithms and analyses target general atom sets and general objective functions, and hence are directly applicable to a large variety of learning settings. |
Tasks | |
Published | 2017-05-31 |
URL | http://arxiv.org/abs/1705.11041v3 |
http://arxiv.org/pdf/1705.11041v3.pdf | |
PWC | https://paperswithcode.com/paper/greedy-algorithms-for-cone-constrained |
Repo | |
Framework | |
A General Theory of Sample Complexity for Multi-Item Profit Maximization
Title | A General Theory of Sample Complexity for Multi-Item Profit Maximization |
Authors | Maria-Florina Balcan, Tuomas Sandholm, Ellen Vitercik |
Abstract | The design of profit-maximizing multi-item mechanisms is a notoriously challenging problem with tremendous real-world impact. The mechanism designer’s goal is to field a mechanism with high expected profit on the distribution over buyers’ values. Unfortunately, if the set of mechanisms he optimizes over is complex, a mechanism may have high empirical profit over a small set of samples but low expected profit. This raises the question, how many samples are sufficient to ensure that the empirically optimal mechanism is nearly optimal in expectation? We uncover structure shared by a myriad of pricing, auction, and lottery mechanisms that allows us to prove strong sample complexity bounds: for any set of buyers’ values, profit is a piecewise linear function of the mechanism’s parameters. We prove new bounds for mechanism classes not yet studied in the sample-based mechanism design literature and match or improve over the best known guarantees for many classes. The profit functions we study are significantly different from well-understood functions in machine learning, so our analysis requires a sharp understanding of the interplay between mechanism parameters and buyer values. We strengthen our main results with data-dependent bounds when the distribution over buyers’ values is “well-behaved.” Finally, we investigate a fundamental tradeoff in sample-based mechanism design: complex mechanisms often have higher profit than simple mechanisms, but more samples are required to ensure that empirical and expected profit are close. We provide techniques for optimizing this tradeoff. |
Tasks | |
Published | 2017-04-29 |
URL | http://arxiv.org/abs/1705.00243v4 |
http://arxiv.org/pdf/1705.00243v4.pdf | |
PWC | https://paperswithcode.com/paper/a-general-theory-of-sample-complexity-for |
Repo | |
Framework | |
N2N Learning: Network to Network Compression via Policy Gradient Reinforcement Learning
Title | N2N Learning: Network to Network Compression via Policy Gradient Reinforcement Learning |
Authors | Anubhav Ashok, Nicholas Rhinehart, Fares Beainy, Kris M. Kitani |
Abstract | While bigger and deeper neural network architectures continue to advance the state-of-the-art for many computer vision tasks, real-world adoption of these networks is impeded by hardware and speed constraints. Conventional model compression methods attempt to address this problem by modifying the architecture manually or using pre-defined heuristics. Since the space of all reduced architectures is very large, modifying the architecture of a deep neural network in this way is a difficult task. In this paper, we tackle this issue by introducing a principled method for learning reduced network architectures in a data-driven way using reinforcement learning. Our approach takes a larger teacher' network as input and outputs a compressed student’ network derived from the teacher' network. In the first stage of our method, a recurrent policy network aggressively removes layers from the large teacher’ model. In the second stage, another recurrent policy network carefully reduces the size of each remaining layer. The resulting network is then evaluated to obtain a reward – a score based on the accuracy and compression of the network. Our approach uses this reward signal with policy gradients to train the policies to find a locally optimal student network. Our experiments show that we can achieve compression rates of more than 10x for models such as ResNet-34 while maintaining similar performance to the input teacher' network. We also present a valuable transfer learning result which shows that policies which are pre-trained on smaller teacher’ networks can be used to rapidly speed up training on larger `teacher’ networks. | |
Tasks | Model Compression, Transfer Learning |
Published | 2017-09-18 |
URL | http://arxiv.org/abs/1709.06030v2 |
http://arxiv.org/pdf/1709.06030v2.pdf | |
PWC | https://paperswithcode.com/paper/n2n-learning-network-to-network-compression |
Repo | |
Framework | |