October 18, 2019

2851 words 14 mins read

Paper Group ANR 536

Paper Group ANR 536

Code-division multiplexed resistive pulse sensor networks for spatio-temporal detection of particles in microfluidic devices. Decoupled Novel Object Captioner. Massively Distributed SGD: ImageNet/ResNet-50 Training in a Flash. Multi-robot Path Planning in Well-formed Infrastructures: Prioritized Planning vs. Prioritized Wait Adjustment (Preliminary …

Code-division multiplexed resistive pulse sensor networks for spatio-temporal detection of particles in microfluidic devices

Title Code-division multiplexed resistive pulse sensor networks for spatio-temporal detection of particles in microfluidic devices
Authors Ningquan Wang, Ruxiu Liu, Roozbeh Khodambashi, Norh Asmare, A. Fatih Sarioglu
Abstract Spatial separation of suspended particles based on contrast in their physical or chemical properties forms the basis of various biological assays performed on lab-on-achip devices. To electronically acquire this information, we have recently introduced a microfluidic sensing platform, called Microfluidic CODES, which combines the resistive pulse sensing with the code division multiple access in multiplexing a network of integrated electrical sensors. In this paper, we enhance the multiplexing capacity of the Microfluidic CODES by employing sensors that generate non-orthogonal code waveforms and a new decoding algorithm that combines machine learning techniques with minimum mean-squared error estimation. As a proof of principle, we fabricated a microfluidic device with a network of 10 code-multiplexed sensors and characterized it using cells suspended in phosphate buffer saline solution.
Tasks
Published 2018-08-10
URL http://arxiv.org/abs/1808.03388v1
PDF http://arxiv.org/pdf/1808.03388v1.pdf
PWC https://paperswithcode.com/paper/code-division-multiplexed-resistive-pulse
Repo
Framework

Decoupled Novel Object Captioner

Title Decoupled Novel Object Captioner
Authors Yu Wu, Linchao Zhu, Lu Jiang, Yi Yang
Abstract Image captioning is a challenging task where the machine automatically describes an image by sentences or phrases. It often requires a large number of paired image-sentence annotations for training. However, a pre-trained captioning model can hardly be applied to a new domain in which some novel object categories exist, i.e., the objects and their description words are unseen during model training. To correctly caption the novel object, it requires professional human workers to annotate the images by sentences with the novel words. It is labor expensive and thus limits its usage in real-world applications. In this paper, we introduce the zero-shot novel object captioning task where the machine generates descriptions without extra sentences about the novel object. To tackle the challenging problem, we propose a Decoupled Novel Object Captioner (DNOC) framework that can fully decouple the language sequence model from the object descriptions. DNOC has two components. 1) A Sequence Model with the Placeholder (SM-P) generates a sentence containing placeholders. The placeholder represents an unseen novel object. Thus, the sequence model can be decoupled from the novel object descriptions. 2) A key-value object memory built upon the freely available detection model, contains the visual information and the corresponding word for each object. The SM-P will generate a query to retrieve the words from the object memory. The placeholder will then be filled with the correct word, resulting in a caption with novel object descriptions. The experimental results on the held-out MSCOCO dataset demonstrate the ability of DNOC in describing novel concepts in the zero-shot novel object captioning task.
Tasks Image Captioning
Published 2018-04-11
URL http://arxiv.org/abs/1804.03803v2
PDF http://arxiv.org/pdf/1804.03803v2.pdf
PWC https://paperswithcode.com/paper/decoupled-novel-object-captioner
Repo
Framework

Massively Distributed SGD: ImageNet/ResNet-50 Training in a Flash

Title Massively Distributed SGD: ImageNet/ResNet-50 Training in a Flash
Authors Hiroaki Mikami, Hisahiro Suganuma, Pongsakorn U-chupala, Yoshiki Tanaka, Yuichi Kageyama
Abstract Scaling the distributed deep learning to a massive GPU cluster level is challenging due to the instability of the large mini-batch training and the overhead of the gradient synchronization. We address the instability of the large mini-batch training with batch-size control and label smoothing. We address the overhead of the gradient synchronization with 2D-Torus all-reduce. Specifically, 2D-Torus all-reduce arranges GPUs in a logical 2D grid and performs a series of collective operation in different orientations. These two techniques are implemented with Neural Network Libraries (NNL). We have successfully trained ImageNet/ResNet-50 in 122 seconds without significant accuracy loss on ABCI cluster.
Tasks
Published 2018-11-13
URL http://arxiv.org/abs/1811.05233v2
PDF http://arxiv.org/pdf/1811.05233v2.pdf
PWC https://paperswithcode.com/paper/massively-distributed-sgd-imagenetresnet-50
Repo
Framework

Multi-robot Path Planning in Well-formed Infrastructures: Prioritized Planning vs. Prioritized Wait Adjustment (Preliminary Results)

Title Multi-robot Path Planning in Well-formed Infrastructures: Prioritized Planning vs. Prioritized Wait Adjustment (Preliminary Results)
Authors Anton Andreychuk, Konstantin Yakovlev
Abstract We study the problem of planning collision-free paths for a group of homogeneous robots. We propose a novel approach for turning the paths that were planned egocentrically by the robots, e.g. without taking other robots’ moves into account, into collision-free trajectories and evaluate it empirically. Suggested algorithm is much faster (up to one order of magnitude) than state-of-the-art but this comes at the price of notable drop-down of the solution cost.
Tasks
Published 2018-07-05
URL http://arxiv.org/abs/1807.01909v1
PDF http://arxiv.org/pdf/1807.01909v1.pdf
PWC https://paperswithcode.com/paper/multi-robot-path-planning-in-well-formed
Repo
Framework

Sample Efficient Deep Reinforcement Learning for Dialogue Systems with Large Action Spaces

Title Sample Efficient Deep Reinforcement Learning for Dialogue Systems with Large Action Spaces
Authors Gellért Weisz, Paweł Budzianowski, Pei-Hao Su, Milica Gašić
Abstract In spoken dialogue systems, we aim to deploy artificial intelligence to build automated dialogue agents that can converse with humans. A part of this effort is the policy optimisation task, which attempts to find a policy describing how to respond to humans, in the form of a function taking the current state of the dialogue and returning the response of the system. In this paper, we investigate deep reinforcement learning approaches to solve this problem. Particular attention is given to actor-critic methods, off-policy reinforcement learning with experience replay, and various methods aimed at reducing the bias and variance of estimators. When combined, these methods result in the previously proposed ACER algorithm that gave competitive results in gaming environments. These environments however are fully observable and have a relatively small action set so in this paper we examine the application of ACER to dialogue policy optimisation. We show that this method beats the current state-of-the-art in deep learning approaches for spoken dialogue systems. This not only leads to a more sample efficient algorithm that can train faster, but also allows us to apply the algorithm in more difficult environments than before. We thus experiment with learning in a very large action space, which has two orders of magnitude more actions than previously considered. We find that ACER trains significantly faster than the current state-of-the-art.
Tasks Spoken Dialogue Systems
Published 2018-02-11
URL http://arxiv.org/abs/1802.03753v1
PDF http://arxiv.org/pdf/1802.03753v1.pdf
PWC https://paperswithcode.com/paper/sample-efficient-deep-reinforcement-learning
Repo
Framework

Novel Single View Constraints for Manhattan 3D Line Reconstruction

Title Novel Single View Constraints for Manhattan 3D Line Reconstruction
Authors Siddhant Ranade, Srikumar Ramalingam
Abstract This paper proposes a novel and exact method to reconstruct line-based 3D structure from a single image using Manhattan world assumption. This problem is a distinctly unsolved problem because there can be multiple 3D reconstructions from a single image. Thus, we are often forced to look for priors like Manhattan world assumption and common scene structures. In addition to the standard orthogonality, perspective projection, and parallelism constraints, we investigate a few novel constraints based on the physical realizability of the 3D scene structure. We treat the line segments in the image to be part of a graph similar to straws and connectors game, where the goal is to back-project the line segments in 3D space and while ensuring that some of these 3D line segments connect with each other (i.e., truly intersect in 3D space) to form the 3D structure. We consider three sets of novel constraints while solving the reconstruction: (1) constraints on a series of Manhattan line intersections that form cycles, but are not all physically realizable, (2) constraints on true and false intersections in the case of nearby lines lying on the same Manhattan plane, and (3) constraints from the intersections on boundary and non-boundary line segments. The reconstruction is achieved using mixed integer linear programming (MILP), and we show compelling results on real images. Along with this paper, we will release a challenging Single View Line Reconstruction dataset with ground truth 3D line models for research purposes.
Tasks
Published 2018-10-08
URL http://arxiv.org/abs/1810.03737v1
PDF http://arxiv.org/pdf/1810.03737v1.pdf
PWC https://paperswithcode.com/paper/novel-single-view-constraints-for-manhattan
Repo
Framework

Integrating domain knowledge: using hierarchies to improve deep classifiers

Title Integrating domain knowledge: using hierarchies to improve deep classifiers
Authors Clemens-Alexander Brust, Joachim Denzler
Abstract One of the most prominent problems in machine learning in the age of deep learning is the availability of sufficiently large annotated datasets. For specific domains, e.g. animal species, a long-tail distribution means that some classes are observed and annotated insufficiently. Additional labels can be prohibitively expensive, e.g. because domain experts need to be involved. However, there is more information available that is to the best of our knowledge not exploited accordingly. In this paper, we propose to make use of preexisting class hierarchies like WordNet to integrate additional domain knowledge into classification. We encode the properties of such a class hierarchy into a probabilistic model. From there, we derive a novel label encoding and a corresponding loss function. On the ImageNet and NABirds datasets our method offers a relative improvement of 10.4% and 9.6% in accuracy over the baseline respectively. After less than a third of training time, it is already able to match the baseline’s fine-grained recognition performance. Both results show that our suggested method is efficient and effective.
Tasks Data Augmentation
Published 2018-11-17
URL https://arxiv.org/abs/1811.07125v2
PDF https://arxiv.org/pdf/1811.07125v2.pdf
PWC https://paperswithcode.com/paper/integrating-domain-knowledge-using
Repo
Framework

A tutorial on MDL hypothesis testing for graph analysis

Title A tutorial on MDL hypothesis testing for graph analysis
Authors Peter Bloem, Steven de Rooij
Abstract This document provides a tutorial description of the use of the MDL principle in complex graph analysis. We give a brief summary of the preliminary subjects, and describe the basic principle, using the example of analysing the size of the largest clique in a graph. We also provide a discussion of how to interpret the results of such an analysis, making note of several common pitfalls.
Tasks
Published 2018-10-31
URL http://arxiv.org/abs/1810.13163v1
PDF http://arxiv.org/pdf/1810.13163v1.pdf
PWC https://paperswithcode.com/paper/a-tutorial-on-mdl-hypothesis-testing-for
Repo
Framework

A Tutorial on Distance Metric Learning: Mathematical Foundations, Algorithms and Experiments

Title A Tutorial on Distance Metric Learning: Mathematical Foundations, Algorithms and Experiments
Authors Juan Luis Suárez, Salvador García, Francisco Herrera
Abstract Distance metric learning is a branch of machine learning that aims to learn distances from the data. Distance metric learning can be useful to improve similarity learning algorithms, and also has applications in dimensionality reduction. This paper describes the distance metric learning problem and analyzes its main mathematical foundations. In addition, it also discusses some of the most popular distance metric learning techniques used in classification, showing their goals and the required information to understand and use them. Furthermore, some experiments to evaluate the performance of the different algorithms are also provided. Finally, this paper discusses several possibilities of future work in this topic.
Tasks Dimensionality Reduction, Metric Learning
Published 2018-12-14
URL https://arxiv.org/abs/1812.05944v2
PDF https://arxiv.org/pdf/1812.05944v2.pdf
PWC https://paperswithcode.com/paper/a-tutorial-on-distance-metric-learning
Repo
Framework

End-to-End Diagnosis and Segmentation Learning from Cardiac Magnetic Resonance Imaging

Title End-to-End Diagnosis and Segmentation Learning from Cardiac Magnetic Resonance Imaging
Authors Gerard Snaauw, Dong Gong, Gabriel Maicas, Anton van den Hengel, Wiro J. Niessen, Johan Verjans, Gustavo Carneiro
Abstract Cardiac magnetic resonance (CMR) is used extensively in the diagnosis and management of cardiovascular disease. Deep learning methods have proven to deliver segmentation results comparable to human experts in CMR imaging, but there have been no convincing results for the problem of end-to-end segmentation and diagnosis from CMR. This is in part due to a lack of sufficiently large datasets required to train robust diagnosis models. In this paper, we propose a learning method to train diagnosis models, where our approach is designed to work with relatively small datasets. In particular, the optimisation loss is based on multi-task learning that jointly trains for the tasks of segmentation and diagnosis classification. We hypothesize that segmentation has a regularizing effect on the learning of features relevant for diagnosis. Using the 100 training and 50 testing samples available from the Automated Cardiac Diagnosis Challenge (ACDC) dataset, which has a balanced distribution of 5 cardiac diagnoses, we observe a reduction of the classification error from 32% to 22%, and a faster convergence compared to a baseline without segmentation. To the best of our knowledge, this is the best diagnosis results from CMR using an end-to-end diagnosis and segmentation learning method.
Tasks Multi-Task Learning
Published 2018-10-23
URL http://arxiv.org/abs/1810.10117v1
PDF http://arxiv.org/pdf/1810.10117v1.pdf
PWC https://paperswithcode.com/paper/end-to-end-diagnosis-and-segmentation
Repo
Framework

On the Reduction of Biases in Big Data Sets for the Detection of Irregular Power Usage

Title On the Reduction of Biases in Big Data Sets for the Detection of Irregular Power Usage
Authors Patrick Glauner, Radu State, Petko Valtchev, Diogo Duarte
Abstract In machine learning, a bias occurs whenever training sets are not representative for the test data, which results in unreliable models. The most common biases in data are arguably class imbalance and covariate shift. In this work, we aim to shed light on this topic in order to increase the overall attention to this issue in the field of machine learning. We propose a scalable novel framework for reducing multiple biases in high-dimensional data sets in order to train more reliable predictors. We apply our methodology to the detection of irregular power usage from real, noisy industrial data. In emerging markets, irregular power usage, and electricity theft in particular, may range up to 40% of the total electricity distributed. Biased data sets are of particular issue in this domain. We show that reducing these biases increases the accuracy of the trained predictors. Our models have the potential to generate significant economic value in a real world application, as they are being deployed in a commercial software for the detection of irregular power usage.
Tasks
Published 2018-01-17
URL http://arxiv.org/abs/1801.05627v2
PDF http://arxiv.org/pdf/1801.05627v2.pdf
PWC https://paperswithcode.com/paper/on-the-reduction-of-biases-in-big-data-sets
Repo
Framework

Modeling Online Discourse with Coupled Distributed Topics

Title Modeling Online Discourse with Coupled Distributed Topics
Authors Akshay Srivatsan, Zachary Wojtowicz, Taylor Berg-Kirkpatrick
Abstract In this paper, we propose a deep, globally normalized topic model that incorporates structural relationships connecting documents in socially generated corpora, such as online forums. Our model (1) captures discursive interactions along observed reply links in addition to traditional topic information, and (2) incorporates latent distributed representations arranged in a deep architecture, which enables a GPU-based mean-field inference procedure that scales efficiently to large data. We apply our model to a new social media dataset consisting of 13M comments mined from the popular internet forum Reddit, a domain that poses significant challenges to models that do not account for relationships connecting user comments. We evaluate against existing methods across multiple metrics including perplexity and metadata prediction, and qualitatively analyze the learned interaction patterns.
Tasks
Published 2018-09-19
URL http://arxiv.org/abs/1809.07282v2
PDF http://arxiv.org/pdf/1809.07282v2.pdf
PWC https://paperswithcode.com/paper/modeling-online-discourse-with-coupled
Repo
Framework

Temporal Event Knowledge Acquisition via Identifying Narratives

Title Temporal Event Knowledge Acquisition via Identifying Narratives
Authors Wenlin Yao, Ruihong Huang
Abstract Inspired by the double temporality characteristic of narrative texts, we propose a novel approach for acquiring rich temporal “before/after” event knowledge across sentences in narrative stories. The double temporality states that a narrative story often describes a sequence of events following the chronological order and therefore, the temporal order of events matches with their textual order. We explored narratology principles and built a weakly supervised approach that identifies 287k narrative paragraphs from three large text corpora. We then extracted rich temporal event knowledge from these narrative paragraphs. Such event knowledge is shown useful to improve temporal relation classification and outperform several recent neural network models on the narrative cloze task.
Tasks Relation Classification
Published 2018-05-28
URL http://arxiv.org/abs/1805.10956v1
PDF http://arxiv.org/pdf/1805.10956v1.pdf
PWC https://paperswithcode.com/paper/temporal-event-knowledge-acquisition-via
Repo
Framework

Image Retargetability

Title Image Retargetability
Authors Fan Tang, Weiming Dong, Yiping Meng, Chongyang Ma, Fuzhang Wu, Xinrui Li, Tong-Yee Lee
Abstract Real-world applications could benefit from the ability to automatically retarget an image to different aspect ratios and resolutions, while preserving its visually and semantically important content. However, not all images can be equally well processed that way. In this work, we introduce the notion of image retargetability to describe how well a particular image can be handled by content-aware image retargeting. We propose to learn a deep convolutional neural network to rank photo retargetability in which the relative ranking of photo retargetability is directly modeled in the loss function. Our model incorporates joint learning of meaningful photographic attributes and image content information which can help regularize the complicated retargetability rating problem. To train and analyze this model, we have collected a database which contains retargetability scores and meaningful image attributes assigned by six expert raters. Experiments demonstrate that our unified model can generate retargetability rankings that are highly consistent with human labels. To further validate our model, we show applications of image retargetability in retargeting method selection, retargeting method assessment and photo collage generation.
Tasks
Published 2018-02-12
URL https://arxiv.org/abs/1802.04392v2
PDF https://arxiv.org/pdf/1802.04392v2.pdf
PWC https://paperswithcode.com/paper/image-retargetability
Repo
Framework

Gender Bias in Coreference Resolution

Title Gender Bias in Coreference Resolution
Authors Rachel Rudinger, Jason Naradowsky, Brian Leonard, Benjamin Van Durme
Abstract We present an empirical study of gender bias in coreference resolution systems. We first introduce a novel, Winograd schema-style set of minimal pair sentences that differ only by pronoun gender. With these “Winogender schemas,” we evaluate and confirm systematic gender bias in three publicly-available coreference resolution systems, and correlate this bias with real-world and textual gender statistics.
Tasks Coreference Resolution
Published 2018-04-25
URL http://arxiv.org/abs/1804.09301v1
PDF http://arxiv.org/pdf/1804.09301v1.pdf
PWC https://paperswithcode.com/paper/gender-bias-in-coreference-resolution
Repo
Framework
comments powered by Disqus