October 17, 2019

2881 words 14 mins read

Paper Group ANR 714

Paper Group ANR 714

CRRN: Multi-Scale Guided Concurrent Reflection Removal Network. MBA: Mini-Batch AUC Optimization. Packaging and Sharing Machine Learning Models via the Acumos AI Open Platform. On the Limitations of Unsupervised Bilingual Dictionary Induction. Efficient Discovery of Heterogeneous Treatment Effects in Randomized Experiments via Anomalous Pattern Det …

CRRN: Multi-Scale Guided Concurrent Reflection Removal Network

Title CRRN: Multi-Scale Guided Concurrent Reflection Removal Network
Authors Renjie Wan, Boxin Shi, Ling-Yu Duan, Ah-Hwee Tan, Alex C. Kot
Abstract Removing the undesired reflections from images taken through the glass is of broad application to various computer vision tasks. Non-learning based methods utilize different handcrafted priors such as the separable sparse gradients caused by different levels of blurs, which often fail due to their limited description capability to the properties of real-world reflections. In this paper, we propose the Concurrent Reflection Removal Network (CRRN) to tackle this problem in a unified framework. Our proposed network integrates image appearance information and multi-scale gradient information with human perception inspired loss function, and is trained on a new dataset with 3250 reflection images taken under diverse real-world scenes. Extensive experiments on a public benchmark dataset show that the proposed method performs favorably against state-of-the-art methods.
Tasks
Published 2018-05-30
URL http://arxiv.org/abs/1805.11802v1
PDF http://arxiv.org/pdf/1805.11802v1.pdf
PWC https://paperswithcode.com/paper/crrn-multi-scale-guided-concurrent-reflection
Repo
Framework

MBA: Mini-Batch AUC Optimization

Title MBA: Mini-Batch AUC Optimization
Authors San Gultekin, Avishek Saha, Adwait Ratnaparkhi, John Paisley
Abstract Area under the receiver operating characteristics curve (AUC) is an important metric for a wide range of signal processing and machine learning problems, and scalable methods for optimizing AUC have recently been proposed. However, handling very large datasets remains an open challenge for this problem. This paper proposes a novel approach to AUC maximization, based on sampling mini-batches of positive/negative instance pairs and computing U-statistics to approximate a global risk minimization problem. The resulting algorithm is simple, fast, and learning-rate free. We show that the number of samples required for good performance is independent of the number of pairs available, which is a quadratic function of the positive and negative instances. Extensive experiments show the practical utility of the proposed method.
Tasks
Published 2018-05-29
URL http://arxiv.org/abs/1805.11221v2
PDF http://arxiv.org/pdf/1805.11221v2.pdf
PWC https://paperswithcode.com/paper/mba-mini-batch-auc-optimization
Repo
Framework

Packaging and Sharing Machine Learning Models via the Acumos AI Open Platform

Title Packaging and Sharing Machine Learning Models via the Acumos AI Open Platform
Authors Shuai Zhao, Manoop Talasila, Guy Jacobson, Cristian Borcea, Syed Anwar Aftab, John F Murray
Abstract Applying Machine Learning (ML) to business applications for automation usually faces difficulties when integrating diverse ML dependencies and services, mainly because of the lack of a common ML framework. In most cases, the ML models are developed for applications which are targeted for specific business domain use cases, leading to duplicated effort, and making reuse impossible. This paper presents Acumos, an open platform capable of packaging ML models into portable containerized microservices which can be easily shared via the platform’s catalog, and can be integrated into various business applications. We present a case study of packaging sentiment analysis and classification ML models via the Acumos platform, permitting easy sharing with others. We demonstrate that the Acumos platform reduces the technical burden on application developers when applying machine learning models to their business applications. Furthermore, the platform allows the reuse of readily available ML microservices in various business domains.
Tasks Sentiment Analysis
Published 2018-10-16
URL http://arxiv.org/abs/1810.07159v1
PDF http://arxiv.org/pdf/1810.07159v1.pdf
PWC https://paperswithcode.com/paper/packaging-and-sharing-machine-learning-models
Repo
Framework

On the Limitations of Unsupervised Bilingual Dictionary Induction

Title On the Limitations of Unsupervised Bilingual Dictionary Induction
Authors Anders Søgaard, Sebastian Ruder, Ivan Vulić
Abstract Unsupervised machine translation—i.e., not assuming any cross-lingual supervision signal, whether a dictionary, translations, or comparable corpora—seems impossible, but nevertheless, Lample et al. (2018) recently proposed a fully unsupervised machine translation (MT) model. The model relies heavily on an adversarial, unsupervised alignment of word embedding spaces for bilingual dictionary induction (Conneau et al., 2018), which we examine here. Our results identify the limitations of current unsupervised MT: unsupervised bilingual dictionary induction performs much worse on morphologically rich languages that are not dependent marking, when monolingual corpora from different domains or different embedding algorithms are used. We show that a simple trick, exploiting a weak supervision signal from identical words, enables more robust induction, and establish a near-perfect correlation between unsupervised bilingual dictionary induction performance and a previously unexplored graph similarity metric.
Tasks Graph Similarity, Machine Translation, Unsupervised Machine Translation
Published 2018-05-09
URL http://arxiv.org/abs/1805.03620v1
PDF http://arxiv.org/pdf/1805.03620v1.pdf
PWC https://paperswithcode.com/paper/on-the-limitations-of-unsupervised-bilingual
Repo
Framework

Efficient Discovery of Heterogeneous Treatment Effects in Randomized Experiments via Anomalous Pattern Detection

Title Efficient Discovery of Heterogeneous Treatment Effects in Randomized Experiments via Anomalous Pattern Detection
Authors Edward McFowland III, Sriram Somanchi, Daniel B. Neill
Abstract In the recent literature on estimating heterogeneous treatment effects, each proposed method makes its own set of restrictive assumptions about the intervention’s effects and which subpopulations to explicitly estimate. Moreover, the majority of the literature provides no mechanism to identify which subpopulations are the most affected–beyond manual inspection–and provides little guarantee on the correctness of the identified subpopulations. Therefore, we propose Treatment Effect Subset Scan (TESS), a new method for discovering which subpopulation in a randomized experiment is most significantly affected by a treatment. We frame this challenge as a pattern detection problem where we efficiently maximize a nonparametric scan statistic over subpopulations. Furthermore, we identify the subpopulation which experiences the largest distributional change as a result of the intervention, while making minimal assumptions about the intervention’s effects or the underlying data generating process. In addition to the algorithm, we demonstrate that the asymptotic Type I and II error can be controlled, and provide sufficient conditions for detection consistency–i.e., exact identification of the affected subpopulation. Finally, we validate the efficacy of the method by discovering heterogeneous treatment effects in simulations and in real-world data from a well-known program evaluation study.
Tasks
Published 2018-03-24
URL http://arxiv.org/abs/1803.09159v2
PDF http://arxiv.org/pdf/1803.09159v2.pdf
PWC https://paperswithcode.com/paper/efficient-discovery-of-heterogeneous
Repo
Framework

Re-Identification with Consistent Attentive Siamese Networks

Title Re-Identification with Consistent Attentive Siamese Networks
Authors Meng Zheng, Srikrishna Karanam, Ziyan Wu, Richard J. Radke
Abstract We propose a new deep architecture for person re-identification (re-id). While re-id has seen much recent progress, spatial localization and view-invariant representation learning for robust cross-view matching remain key, unsolved problems. We address these questions by means of a new attention-driven Siamese learning architecture, called the Consistent Attentive Siamese Network. Our key innovations compared to existing, competing methods include (a) a flexible framework design that produces attention with only identity labels as supervision, (b) explicit mechanisms to enforce attention consistency among images of the same person, and (c) a new Siamese framework that integrates attention and attention consistency, producing principled supervisory signals as well as the first mechanism that can explain the reasoning behind the Siamese framework’s predictions. We conduct extensive evaluations on the CUHK03-NP, DukeMTMC-ReID, and Market-1501 datasets and report competitive performance.
Tasks Person Re-Identification, Representation Learning
Published 2018-11-19
URL http://arxiv.org/abs/1811.07487v4
PDF http://arxiv.org/pdf/1811.07487v4.pdf
PWC https://paperswithcode.com/paper/re-identification-with-consistent-attentive
Repo
Framework

Evaluating approaches for supervised semantic labeling

Title Evaluating approaches for supervised semantic labeling
Authors Natalia Ruemmele, Yuriy Tyshetskiy, Alex Collins
Abstract Relational data sources are still one of the most popular ways to store enterprise or Web data, however, the issue with relational schema is the lack of a well-defined semantic description. A common ontology provides a way to represent the meaning of a relational schema and can facilitate the integration of heterogeneous data sources within a domain. Semantic labeling is achieved by mapping attributes from the data sources to the classes and properties in the ontology. We formulate this problem as a multi-class classification problem where previously labeled data sources are used to learn rules for labeling new data sources. The majority of existing approaches for semantic labeling have focused on data integration challenges such as naming conflicts and semantic heterogeneity. In addition, machine learning approaches typically have issues around class imbalance, lack of labeled instances and relative importance of attributes. To address these issues, we develop a new machine learning model with engineered features as well as two deep learning models which do not require extensive feature engineering. We evaluate our new approaches with the state-of-the-art.
Tasks Feature Engineering
Published 2018-01-29
URL http://arxiv.org/abs/1801.09788v1
PDF http://arxiv.org/pdf/1801.09788v1.pdf
PWC https://paperswithcode.com/paper/evaluating-approaches-for-supervised-semantic
Repo
Framework

Random Polyhedral Scenes: An Image Generator for Active Vision System Experiments

Title Random Polyhedral Scenes: An Image Generator for Active Vision System Experiments
Authors Markus D. Solbach, Stephen Voland, Jeff Edmonds, John K. Tsotsos
Abstract We present a Polyhedral Scene Generator system which creates a random scene based on a few user parameters, renders the scene from random view points and creates a dataset containing the renderings and corresponding annotation files. We hope that this generator will enable research on how a program could parse a scene if it had multiple viewpoints to consider. For ambiguous scenes, typically people move their head or change their position to see the scene from different angles as well as seeing how it changes while they move; this research field is called active perception. The random scene generator presented is designed to support research in this field by generating images of scenes with known complexity characteristics and with verifiable properties with respect to the distribution of features across a population. Thus, it is well-suited for research in active perception without the requirement of a live 3D environment and mobile sensing agent, including comparative performance evaluations. The system is publicly available at https://polyhedral.eecs.yorku.ca.
Tasks
Published 2018-03-27
URL http://arxiv.org/abs/1803.10100v1
PDF http://arxiv.org/pdf/1803.10100v1.pdf
PWC https://paperswithcode.com/paper/random-polyhedral-scenes-an-image-generator
Repo
Framework

Conditional Video Generation Using Action-Appearance Captions

Title Conditional Video Generation Using Action-Appearance Captions
Authors Shohei Yamamoto, Antonio Tejero-de-Pablos, Yoshitaka Ushiku, Tatsuya Harada
Abstract The field of automatic video generation has received a boost thanks to the recent Generative Adversarial Networks (GANs). However, most existing methods cannot control the contents of the generated video using a text caption, losing their usefulness to a large extent. This particularly affects human videos due to their great variety of actions and appearances. This paper presents Conditional Flow and Texture GAN (CFT-GAN), a GAN-based video generation method from action-appearance captions. We propose a novel way of generating video by encoding a caption (e.g., “a man in blue jeans is playing golf”) in a two-stage generation pipeline. Our CFT-GAN uses such caption to generate an optical flow (action) and a texture (appearance) for each frame. As a result, the output video reflects the content specified in the caption in a plausible way. Moreover, to train our method, we constructed a new dataset for human video generation with captions. We evaluated the proposed method qualitatively and quantitatively via an ablation study and a user study. The results demonstrate that CFT-GAN is able to successfully generate videos containing the action and appearances indicated in the captions.
Tasks Optical Flow Estimation, Video Generation
Published 2018-12-04
URL http://arxiv.org/abs/1812.01261v2
PDF http://arxiv.org/pdf/1812.01261v2.pdf
PWC https://paperswithcode.com/paper/conditional-video-generation-using-action
Repo
Framework

Geometric Scattering for Graph Data Analysis

Title Geometric Scattering for Graph Data Analysis
Authors Feng Gao, Guy Wolf, Matthew Hirn
Abstract We explore the generalization of scattering transforms from traditional (e.g., image or audio) signals to graph data, analogous to the generalization of ConvNets in geometric deep learning, and the utility of extracted graph features in graph data analysis. In particular, we focus on the capacity of these features to retain informative variability and relations in the data (e.g., between individual graphs, or in aggregate), while relating our construction to previous theoretical results that establish the stability of similar transforms to families of graph deformations. We demonstrate the application the our geometric scattering features in graph classification of social network data, and in data exploration of biochemistry data.
Tasks Graph Classification, Image Classification
Published 2018-10-07
URL http://arxiv.org/abs/1810.03068v2
PDF http://arxiv.org/pdf/1810.03068v2.pdf
PWC https://paperswithcode.com/paper/geometric-scattering-for-graph-data-analysis
Repo
Framework

Weakly supervised learning of indoor geometry by dual warping

Title Weakly supervised learning of indoor geometry by dual warping
Authors Pulak Purkait, Ujwal Bonde, Christopher Zach
Abstract A major element of depth perception and 3D understanding is the ability to predict the 3D layout of a scene and its contained objects for a novel pose. Indoor environments are particularly suitable for novel view prediction, since the set of objects in such environments is relatively restricted. In this work we address the task of 3D prediction especially for indoor scenes by leveraging only weak supervision. In the literature 3D scene prediction is usually solved via a 3D voxel grid. However, such methods are limited to estimating rather coarse 3D voxel grids, since predicting entire voxel spaces has large computational costs. Hence, our method operates in image-space rather than in voxel space, and the task of 3D estimation essentially becomes a depth image completion problem. We propose a novel approach to easily generate training data containing depth maps with realistic occlusions, and subsequently train a network for completing those occluded regions. Using multiple publicly available dataset~\cite{song2017semantic,Silberman:ECCV12} we benchmark our method against existing approaches and are able to obtain superior performance. We further demonstrate the flexibility of our method by presenting results for new view synthesis of RGB-D images.
Tasks
Published 2018-08-10
URL http://arxiv.org/abs/1808.03609v1
PDF http://arxiv.org/pdf/1808.03609v1.pdf
PWC https://paperswithcode.com/paper/weakly-supervised-learning-of-indoor-geometry
Repo
Framework

Complementary Segmentation of Primary Video Objects with Reversible Flows

Title Complementary Segmentation of Primary Video Objects with Reversible Flows
Authors Jia Li, Junjie Wu, Anlin Zheng, Yafei Song, Yu Zhang, Xiaowu Chen
Abstract Segmenting primary objects in a video is an important yet challenging problem in computer vision, as it exhibits various levels of foreground/background ambiguities. To reduce such ambiguities, we propose a novel formulation via exploiting foreground and background context as well as their complementary constraint. Under this formulation, a unified objective function is further defined to encode each cue. For implementation, we design a Complementary Segmentation Network (CSNet) with two separate branches, which can simultaneously encode the foreground and background information along with joint spatial constraints. The CSNet is trained on massive images with manually annotated salient objects in an end-to-end manner. By applying CSNet on each video frame, the spatial foreground and background maps can be initialized. To enforce temporal consistency effectively and efficiently, we divide each frame into superpixels and construct neighborhood reversible flow that reflects the most reliable temporal correspondences between superpixels in far-away frames. With such flow, the initialized foregroundness and backgroundness can be propagated along the temporal dimension so that primary video objects gradually pop-out and distractors are well suppressed. Extensive experimental results on three video datasets show that the proposed approach achieves impressive performance in comparisons with 18 state-of-the-art models.
Tasks Video Semantic Segmentation
Published 2018-11-23
URL http://arxiv.org/abs/1811.09521v1
PDF http://arxiv.org/pdf/1811.09521v1.pdf
PWC https://paperswithcode.com/paper/complementary-segmentation-of-primary-video
Repo
Framework

Learning in Memristive Neural Network Architectures using Analog Backpropagation Circuits

Title Learning in Memristive Neural Network Architectures using Analog Backpropagation Circuits
Authors Olga Krestinskaya, Khaled Nabil Salama, Alex Pappachen James
Abstract The on-chip implementation of learning algorithms would speed-up the training of neural networks in crossbar arrays. The circuit level design and implementation of backpropagation algorithm using gradient descent operation for neural network architectures is an open problem. In this paper, we proposed the analog backpropagation learning circuits for various memristive learning architectures, such as Deep Neural Network (DNN), Binary Neural Network (BNN), Multiple Neural Network (MNN), Hierarchical Temporal Memory (HTM) and Long-Short Term Memory (LSTM). The circuit design and verification is done using TSMC 180nm CMOS process models, and TiO2 based memristor models. The application level validations of the system are done using XOR problem, MNIST character and Yale face image databases
Tasks
Published 2018-08-31
URL http://arxiv.org/abs/1808.10631v1
PDF http://arxiv.org/pdf/1808.10631v1.pdf
PWC https://paperswithcode.com/paper/learning-in-memristive-neural-network
Repo
Framework

Creatures great and SMAL: Recovering the shape and motion of animals from video

Title Creatures great and SMAL: Recovering the shape and motion of animals from video
Authors Benjamin Biggs, Thomas Roddick, Andrew Fitzgibbon, Roberto Cipolla
Abstract We present a system to recover the 3D shape and motion of a wide variety of quadrupeds from video. The system comprises a machine learning front-end which predicts candidate 2D joint positions, a discrete optimization which finds kinematically plausible joint correspondences, and an energy minimization stage which fits a detailed 3D model to the image. In order to overcome the limited availability of motion capture training data from animals, and the difficulty of generating realistic synthetic training images, the system is designed to work on silhouette data. The joint candidate predictor is trained on synthetically generated silhouette images, and at test time, deep learning methods or standard video segmentation tools are used to extract silhouettes from real data. The system is tested on animal videos from several species, and shows accurate reconstructions of 3D shape and pose.
Tasks Motion Capture, Video Semantic Segmentation
Published 2018-11-14
URL http://arxiv.org/abs/1811.05804v1
PDF http://arxiv.org/pdf/1811.05804v1.pdf
PWC https://paperswithcode.com/paper/creatures-great-and-smal-recovering-the-shape
Repo
Framework

Sequential Learning of Principal Curves: Summarizing Data Streams on the Fly

Title Sequential Learning of Principal Curves: Summarizing Data Streams on the Fly
Authors Benjamin Guedj, Le Li
Abstract When confronted with massive data streams, summarizing data with dimension reduction methods such as PCA raises theoretical and algorithmic pitfalls. Principal curves act as a nonlinear generalization of PCA and the present paper proposes a novel algorithm to automatically and sequentially learn principal curves from data streams. We show that our procedure is supported by regret bounds with optimal sublinear remainder terms. A greedy local search implementation (called \texttt{slpc}, for Sequential Learning Principal Curves) that incorporates both sleeping experts and multi-armed bandit ingredients is presented, along with its regret computation and performance on synthetic and real-life data.
Tasks Dimensionality Reduction
Published 2018-05-18
URL https://arxiv.org/abs/1805.07418v2
PDF https://arxiv.org/pdf/1805.07418v2.pdf
PWC https://paperswithcode.com/paper/sequential-learning-of-principal-curves
Repo
Framework
comments powered by Disqus