January 25, 2020

3033 words 15 mins read

Paper Group ANR 1758

Paper Group ANR 1758

Estimation of Personalized Heterogeneous Treatment Effects Using Concatenation and Augmentation of Feature Vectors. Towards a Wearable Interface for Food Quality Grading through ERP Analysis. Task-Agnostic Dynamics Priors for Deep Reinforcement Learning. Dependency Parsing for Spoken Dialog Systems. Progressive Learning Algorithm for Efficient Pers …

Estimation of Personalized Heterogeneous Treatment Effects Using Concatenation and Augmentation of Feature Vectors

Title Estimation of Personalized Heterogeneous Treatment Effects Using Concatenation and Augmentation of Feature Vectors
Authors Lev V. Utkin, Mikhail V. Kots, Viacheslav S. Chukanov
Abstract A new meta-algorithm for estimating the conditional average treatment effects is proposed in the paper. The main idea underlying the algorithm is to consider a new dataset consisting of feature vectors produced by means of concatenation of examples from control and treatment groups, which are close to each other. Outcomes of new data are defined as the difference between outcomes of the corresponding examples comprising new feature vectors. The second idea is based on the assumption that the number of controls is rather large and the control outcome function is precisely determined. This assumption allows us to augment treatments by generating feature vectors which are closed to available treatments. The outcome regression function constructed on the augmented set of concatenated feature vectors can be viewed as an estimator of the conditional average treatment effects. A simple modification of the Co-learner based on the random subspace method or the feature bagging is also proposed. Various numerical simulation experiments illustrate the proposed algorithm and show its outperformance in comparison with the well-known T-learner and X-learner for several types of the control and treatment outcome functions.
Tasks
Published 2019-09-09
URL https://arxiv.org/abs/1909.03894v1
PDF https://arxiv.org/pdf/1909.03894v1.pdf
PWC https://paperswithcode.com/paper/estimation-of-personalized-heterogeneous
Repo
Framework

Towards a Wearable Interface for Food Quality Grading through ERP Analysis

Title Towards a Wearable Interface for Food Quality Grading through ERP Analysis
Authors M. Guermandi, S. Benatti, D. Brunelli, V. Kartsch, L. Benini
Abstract Sensory evaluation is used to assess the consumer acceptance of foods or other consumer products, so as to improve industrial processes and marketing strategies. The procedures currently involved are time-consuming because they require a statistical approach from measurements and feedback reports from a wide set of evaluators under a well-established measurement setup. In this paper, we propose to collect directly the signal of the perceived quality of the food from Event-related potentials (ERPs) that are the outcome of the processing of visual stimuli. This permits to narrow the number of evaluators since errors related to psychological factors are by-passed. We present the design of a wearable system for ERP measurement and we present preliminary results on the use of ERP to give a quantitative measure to the appearance of a food product. The system is developed to be wearable and our experiments demonstrate that is possible to use it to identify and classify the grade of acceptance of the food.
Tasks
Published 2019-05-28
URL https://arxiv.org/abs/1905.11633v1
PDF https://arxiv.org/pdf/1905.11633v1.pdf
PWC https://paperswithcode.com/paper/towards-a-wearable-interface-for-food-quality
Repo
Framework

Task-Agnostic Dynamics Priors for Deep Reinforcement Learning

Title Task-Agnostic Dynamics Priors for Deep Reinforcement Learning
Authors Yilun Du, Karthik Narasimhan
Abstract While model-based deep reinforcement learning (RL) holds great promise for sample efficiency and generalization, learning an accurate dynamics model is often challenging and requires substantial interaction with the environment. A wide variety of domains have dynamics that share common foundations like the laws of classical mechanics, which are rarely exploited by existing algorithms. In fact, humans continuously acquire and use such dynamics priors to easily adapt to operating in new environments. In this work, we propose an approach to learn task-agnostic dynamics priors from videos and incorporate them into an RL agent. Our method involves pre-training a frame predictor on task-agnostic physics videos to initialize dynamics models (and fine-tune them) for unseen target environments. Our frame prediction architecture, SpatialNet, is designed specifically to capture localized physical phenomena and interactions. Our approach allows for both faster policy learning and convergence to better policies, outperforming competitive approaches on several different environments. We also demonstrate that incorporating this prior allows for more effective transfer between environments.
Tasks
Published 2019-05-13
URL https://arxiv.org/abs/1905.04819v4
PDF https://arxiv.org/pdf/1905.04819v4.pdf
PWC https://paperswithcode.com/paper/task-agnostic-dynamics-priors-for-deep
Repo
Framework

Dependency Parsing for Spoken Dialog Systems

Title Dependency Parsing for Spoken Dialog Systems
Authors Sam Davidson, Dian Yu, Zhou Yu
Abstract Dependency parsing of conversational input can play an important role in language understanding for dialog systems by identifying the relationships between entities extracted from user utterances. Additionally, effective dependency parsing can elucidate differences in language structure and usage for discourse analysis of human-human versus human-machine dialogs. However, models trained on datasets based on news articles and web data do not perform well on spoken human-machine dialog, and currently available annotation schemes do not adapt well to dialog data. Therefore, we propose the Spoken Conversation Universal Dependencies (SCUD) annotation scheme that extends the Universal Dependencies (UD) (Nivre et al., 2016) guidelines to spoken human-machine dialogs. We also provide ConvBank, a conversation dataset between humans and an open-domain conversational dialog system with SCUD annotation. Finally, to demonstrate the utility of the dataset, we train a dependency parser on the ConvBank dataset. We demonstrate that by pre-training a dependency parser on a set of larger public datasets and fine-tuning on ConvBank data, we achieved the best result, 85.05% unlabeled and 77.82% labeled attachment accuracy.
Tasks Dependency Parsing
Published 2019-09-07
URL https://arxiv.org/abs/1909.03317v1
PDF https://arxiv.org/pdf/1909.03317v1.pdf
PWC https://paperswithcode.com/paper/dependency-parsing-for-spoken-dialog-systems
Repo
Framework

Progressive Learning Algorithm for Efficient Person Re-Identification

Title Progressive Learning Algorithm for Efficient Person Re-Identification
Authors Zhen Li, Hanyang Shao, Nian Xue, Liang Niu, LiangLiang Cao
Abstract This paper studies the problem of Person Re-Identification (ReID)for large-scale applications. Recent research efforts have been devoted to building complicated part models, which introduce considerably high computational cost and memory consumption, inhibiting its practicability in large-scale applications. This paper aims to develop a novel learning strategy to find efficient feature embeddings while maintaining the balance of accuracy and model complexity. More specifically, we find by enhancing the classical triplet loss together with cross-entropy loss, our method can explore the hard examples and build a discriminant feature embedding yet compact enough for large-scale applications. Our method is carried out progressively using Bayesian optimization, and we call it the Progressive Learning Algorithm (PLA). Extensive experiments on three large-scale datasets show that our PLA is comparable or better than the-state-of-the-arts. Especially, on the challenging Market-1501 dataset, we achieve Rank-1=94.7%/mAP=89.4% while saving at least 30% parameters than strong part models.
Tasks Person Re-Identification
Published 2019-12-16
URL https://arxiv.org/abs/1912.07447v1
PDF https://arxiv.org/pdf/1912.07447v1.pdf
PWC https://paperswithcode.com/paper/progressive-learning-algorithm-for-efficient
Repo
Framework

Network Embedding: An Overview

Title Network Embedding: An Overview
Authors Nino Arsov, Georgina Mirceva
Abstract Networks are one of the most powerful structures for modeling problems in the real world. Downstream machine learning tasks defined on networks have the potential to solve a variety of problems. With link prediction, for instance, one can predict whether two persons will become friends on a social network. Many machine learning algorithms, however, require that each input example is a real vector. Network embedding encompasses various methods for unsupervised, and sometimes supervised, learning of feature representations of nodes and links in a network. Typically, embedding methods are based on the assumption that the similarity between nodes in the network should be reflected in the learned feature representations. In this paper, we review significant contributions to network embedding in the last decade. In particular, we look at four methods: Spectral Clustering, DeepWalk, Large-scale Information Network Embedding (LINE), and node2vec. We describe each method and list its advantages and shortcomings. In addition, we give examples of real-world machine learning problems on networks in which the embedding is critical in order to maximize the predictive performance of the machine learning task. Finally, we take a look at research trends and state-of-the art methods in the research on network embedding.
Tasks Link Prediction, Network Embedding
Published 2019-11-26
URL https://arxiv.org/abs/1911.11726v1
PDF https://arxiv.org/pdf/1911.11726v1.pdf
PWC https://paperswithcode.com/paper/network-embedding-an-overview
Repo
Framework

Iris Recognition for Personal Identification using LAMSTAR neural network

Title Iris Recognition for Personal Identification using LAMSTAR neural network
Authors Shideh Homayon, Mahdi Salarian
Abstract Iris recognition is one of the most important biometric recognition method. This is because the iris texture provides many features such as freckles, coronas, stripes, furrows, crypts, etc. Those features are unique for different people and distinguishable. Such unique features in the anatomical structure of the iris make it possible the differentiation among individuals. So during last years huge number of people have been trying to improve its performance. In this article first different common steps for the Iris recognition system is explained. Then a special type of neural network is used for recognition part. Experimental results show high accuracy can be obtained especially when the primary steps are done well.
Tasks Iris Recognition
Published 2019-07-28
URL https://arxiv.org/abs/1907.12145v1
PDF https://arxiv.org/pdf/1907.12145v1.pdf
PWC https://paperswithcode.com/paper/iris-recognition-for-personal-identification
Repo
Framework

PT-ResNet: Perspective Transformation-Based Residual Network for Semantic Road Image Segmentation

Title PT-ResNet: Perspective Transformation-Based Residual Network for Semantic Road Image Segmentation
Authors Rui Fan, Yuan Wang, Lei Qiao, Ruiwen Yao, Peng Han, Weidong Zhang, Ioannis Pitas, Ming Liu
Abstract Semantic road region segmentation is a high-level task, which paves the way towards road scene understanding. This paper presents a residual network trained for semantic road segmentation. Firstly, we represent the projections of road disparities in the v-disparity map as a linear model, which can be estimated by optimizing the v-disparity map using dynamic programming. This linear model is then utilized to reduce the redundant information in the left and right road images. The right image is also transformed into the left perspective view, which greatly enhances the road surface similarity between the two images. Finally, the processed stereo images and their disparity maps are concatenated to create a set of 3D images, which are then utilized to train our neural network. The experimental results illustrate that our network achieves a maximum F1-measure of approximately 91.19% when analyzing the images from the KITTI road dataset.
Tasks Scene Understanding, Semantic Segmentation
Published 2019-10-29
URL https://arxiv.org/abs/1910.13055v1
PDF https://arxiv.org/pdf/1910.13055v1.pdf
PWC https://paperswithcode.com/paper/pt-resnet-perspective-transformation-based
Repo
Framework

Deep-SLAM++: Object-level RGBD SLAM based on class-specific deep shape priors

Title Deep-SLAM++: Object-level RGBD SLAM based on class-specific deep shape priors
Authors Lan Hu, Wanting Xu, Kun Huang, Laurent Kneip
Abstract In an effort to increase the capabilities of SLAM systems and produce object-level representations, the community increasingly investigates the imposition of higher-level priors into the estimation process. One such example is given by employing object detectors to load and register full CAD models. Our work extends this idea to environments with unknown objects and imposes object priors by employing modern class-specific neural networks to generate complete model geometry proposals. The difficulty of using such predictions in a real SLAM scenario is that the prediction performance depends on the view-point and measurement quality, with even small changes of the input data sometimes leading to a large variability in the network output. We propose a discrete selection strategy that finds the best among multiple proposals from different registered views by re-enforcing the agreement with the online depth measurements. The result is an effective object-level RGBD SLAM system that produces compact, high-fidelity, and dense 3D maps with semantic annotations. It outperforms traditional fusion strategies in terms of map completeness and resilience against degrading measurement quality.
Tasks
Published 2019-07-23
URL https://arxiv.org/abs/1907.09691v2
PDF https://arxiv.org/pdf/1907.09691v2.pdf
PWC https://paperswithcode.com/paper/deep-slam-object-level-rgbd-slam-based-on
Repo
Framework

Detection of perfusion ROI as a quality control in perfusion analysis

Title Detection of perfusion ROI as a quality control in perfusion analysis
Authors Svitlana Alkhimova
Abstract In perfusion analysis automated approaches for image processing is preferable due to reduce time-consuming tasks for radiologists. Assessment of perfusion results quality is important step in development of algorithms for automated processing. One of them is an assessment of perfusion maps quality based on detection of perfusion ROI.
Tasks
Published 2019-02-04
URL http://arxiv.org/abs/1902.01855v1
PDF http://arxiv.org/pdf/1902.01855v1.pdf
PWC https://paperswithcode.com/paper/detection-of-perfusion-roi-as-a-quality
Repo
Framework

Crowd Density Estimation using Novel Feature Descriptor

Title Crowd Density Estimation using Novel Feature Descriptor
Authors Adwan Alownie Alanazi, Muhammad Bilal
Abstract Crowd density estimation is an important task for crowd monitoring. Many efforts have been done to automate the process of estimating crowd density from images and videos. Despite series of efforts, it remains a challenging task. In this paper, we proposes a new texture feature-based approach for the estimation of crowd density based on Completed Local Binary Pattern (CLBP). We first divide the image into blocks and then re-divide the blocks into cells. For each cell, we compute CLBP and then concatenate them to describe the texture of the corresponding block. We then train a multi-class Support Vector Machine (SVM) classifier, which classifies each block of image into one of four categories, i.e. Very Low, Low, Medium, and High. We evaluate our technique on the PETS 2009 dataset, and from the experiments, we show to achieve 95% accuracy for the proposed descriptor. We also compare other state-of-the-art texture descriptors and from the experimental results, we show that our proposed method outperforms other state-of-the-art methods.
Tasks Density Estimation
Published 2019-05-15
URL https://arxiv.org/abs/1905.05891v1
PDF https://arxiv.org/pdf/1905.05891v1.pdf
PWC https://paperswithcode.com/paper/crowd-density-estimation-using-novel-feature
Repo
Framework

RLScheduler: Learn to Schedule HPC Batch Jobs Using Deep Reinforcement Learning

Title RLScheduler: Learn to Schedule HPC Batch Jobs Using Deep Reinforcement Learning
Authors Di Zhang, Dong Dai, Youbiao He, Forrest Sheng Bao
Abstract We present RLScheduler, a deep reinforcement learning based job scheduler for scheduling independent batch jobs in high-performance computing (HPC) environment. From knowing nothing about scheduling at beginning, RLScheduler is able to autonomously learn how to effectively schedule HPC batch jobs, targeting a given optimization goal. This is achieved by deep reinforcement learning with the help of specially designed neural network structures and various optimizations to stabilize and accelerate the learning. Our results show that RLScheduler can outperform existing heuristic scheduling algorithms, including a manually fine-tuned machine learning-based scheduler on the same workload. More importantly, we show that RLScheduler does not blindly over-fit the given workload to achieve such optimization, instead, it learns general rules for scheduling batch jobs which can be further applied to different workloads and systems to achieve similarly optimized performance. We also demonstrate that RLScheduler is capable of adjusting itself along with changing goals and workloads, making it an attractive solution for the future autonomous HPC management.
Tasks
Published 2019-10-20
URL https://arxiv.org/abs/1910.08925v1
PDF https://arxiv.org/pdf/1910.08925v1.pdf
PWC https://paperswithcode.com/paper/rlscheduler-learn-to-schedule-hpc-batch-jobs
Repo
Framework

Filling Factors of Sunspots in SODISM Images

Title Filling Factors of Sunspots in SODISM Images
Authors Amro F. Alasta, Abdrazag Algamudi, Fatma Almesrati, Mustapha Meftah, Rami Qahwaji
Abstract Received: 1st December 2018; Accepted: 18th February 2019; Published: 1st April 2019 Abstract: The calculated filling factors (FFs) for a feature reflect the fraction of the solar disc covered by that feature, and the assignment of reference synthetic spectra. In this paper, the FFs, specified as a function of radial position on the solar disc, are computed for each image in a tabular form. The filling factor (FF) is an important parameter and is defined as the fraction of area in a pixel covered with the magnetic field, whereas the rest of the area in the pixel is field-free. However, this does not provide extensive information about the experiments conducted on tens or hundreds of such images. This is the first time that filling factors for SODISM images have been catalogued in tabular formation. This paper presents a new method that provides the means to detect sunspots on full-disk solar images recorded by the Solar Diameter Imager and Surface Mapper (SODISM) on the PICARD satellite. The method is a totally automated detection process that achieves a sunspot recognition rate of 97.6%. The number of sunspots detected by this method strongly agrees with the NOAA catalogue. The sunspot areas calculated by this method have a 99% correlation with SOHO over the same period, and thus help to calculate the filling factor for wavelength (W.L.) 607nm.
Tasks
Published 2019-04-01
URL http://arxiv.org/abs/1904.01133v1
PDF http://arxiv.org/pdf/1904.01133v1.pdf
PWC https://paperswithcode.com/paper/filling-factors-of-sunspots-in-sodism-images
Repo
Framework

Liver Lesion Segmentation with slice-wise 2D Tiramisu and Tversky loss function

Title Liver Lesion Segmentation with slice-wise 2D Tiramisu and Tversky loss function
Authors Karsten Roth, Tomasz Konopczyński, Jürgen Hesser
Abstract At present, lesion segmentation is still performed manually (or semi-automatically) by medical experts. To facilitate this process, we contribute a fully-automatic lesion segmentation pipeline. This work proposes a method as a part of the LiTS (Liver Tumor Segmentation Challenge) competition for ISBI 17 and MICCAI 17 comparing methods for automatics egmentation of liver lesions in CT scans. By utilizing cascaded, densely connected 2D U-Nets and a Tversky-coefficient based loss function, our framework achieves very good shape extractions with high detection sensitivity, with competitive scores at time of publication. In addition, adjusting hyperparameters in our Tversky-loss allows to tune the network towards higher sensitivity or robustness.
Tasks Lesion Segmentation
Published 2019-05-09
URL https://arxiv.org/abs/1905.03639v1
PDF https://arxiv.org/pdf/1905.03639v1.pdf
PWC https://paperswithcode.com/paper/190503639
Repo
Framework

A Programmatic and Semantic Approach to Explaining and DebuggingNeural Network Based Object Detectors

Title A Programmatic and Semantic Approach to Explaining and DebuggingNeural Network Based Object Detectors
Authors Edward Kim, Divya Gopinath, Corina Pasareanu, Sanjit Seshia
Abstract Even as deep neural networks have become very effective for tasks in vision and perception, it remains difficult to explain and debug their behavior. In this paper, we present a programmatic and semantic approach to explaining, understanding, and debugging the correct and incorrect behaviors of a neural network based perception system. Our approach is semantic in that it employs a high-level representation of the distribution of environment scenarios that the detector is intended to work on. It is programmatic in that the representation is a program in a domain-specific probabilistic programming language using which synthetic data can be generated to train and test the neural network. We present a framework that assesses the performance of the neural network to identify correct and incorrect detections, extracts rules from those results that semantically characterizes the correct and incorrect scenarios, and then specializes the probabilistic program with those rules in order to more precisely characterize the scenarios in which the neural network operates correctly or not, without human intervention to identify important features. We demonstrate our results using the SCENIC probabilistic programming language and a neural network-based object detector. Our experiments show that it is possible to automatically generate compact rules that significantly increase the correct detection rate (or conversely the incorrect detection rate) of the network and can thus help with debugging and understanding its behavior.
Tasks Probabilistic Programming
Published 2019-12-01
URL https://arxiv.org/abs/1912.00289v1
PDF https://arxiv.org/pdf/1912.00289v1.pdf
PWC https://paperswithcode.com/paper/a-programmatic-and-semantic-approach-to
Repo
Framework
comments powered by Disqus