July 26, 2019

2856 words 14 mins read

Paper Group ANR 792

Paper Group ANR 792

Hyperspectral Unmixing: Ground Truth Labeling, Datasets, Benchmark Performances and Survey. Deep Residual Text Detection Network for Scene Text. Multitask Learning with Deep Neural Networks for Community Question Answering. R2N2: Residual Recurrent Neural Networks for Multivariate Time Series Forecasting. Automatic Classification of Cancerous Tissu …

Hyperspectral Unmixing: Ground Truth Labeling, Datasets, Benchmark Performances and Survey

Title Hyperspectral Unmixing: Ground Truth Labeling, Datasets, Benchmark Performances and Survey
Authors Feiyun Zhu
Abstract Hyperspectral unmixing (HU) is a very useful and increasingly popular preprocessing step for a wide range of hyperspectral applications. However, the HU research has been constrained a lot by three factors: (a) the number of hyperspectral images (especially the ones with ground truths) are very limited; (b) the ground truths of most hyperspectral images are not shared on the web, which may cause lots of unnecessary troubles for researchers to evaluate their algorithms; (c) the codes of most state-of-the-art methods are not shared, which may also delay the testing of new methods. Accordingly, this paper deals with the above issues from the following three perspectives: (1) as a profound contribution, we provide a general labeling method for the HU. With it, we labeled up to 15 hyperspectral images, providing 18 versions of ground truths. To the best of our knowledge, this is the first paper to summarize and share up to 15 hyperspectral images and their 18 versions of ground truths for the HU. Observing that the hyperspectral classification (HyC) has much more standard datasets (whose ground truths are generally publicly shared) than the HU, we propose an interesting method to transform the HyC datasets for the HU research. (2) To further facilitate the evaluation of HU methods under different conditions, we reviewed and implemented the algorithm to generate a complex synthetic hyperspectral image. By tuning the hyper-parameters in the code, we may verify the HU methods from four perspectives. The code would also be shared on the web. (3) To provide a standard comparison, we reviewed up to 10 state-of-the-art HU algorithms, then selected the 5 most benchmark HU algorithms, and compared them on the 15 real hyperspectral datasets. The experiment results are surely reproducible; the implemented codes would be shared on the web.
Tasks Hyperspectral Unmixing
Published 2017-08-17
URL http://arxiv.org/abs/1708.05125v2
PDF http://arxiv.org/pdf/1708.05125v2.pdf
PWC https://paperswithcode.com/paper/hyperspectral-unmixing-ground-truth-labeling
Repo
Framework

Deep Residual Text Detection Network for Scene Text

Title Deep Residual Text Detection Network for Scene Text
Authors Xiangyu Zhu, Yingying Jiang, Shuli Yang, Xiaobing Wang, Wei Li, Pei Fu, Hua Wang, Zhenbo Luo
Abstract Scene text detection is a challenging problem in computer vision. In this paper, we propose a novel text detection network based on prevalent object detection frameworks. In order to obtain stronger semantic feature, we adopt ResNet as feature extraction layers and exploit multi-level feature by combining hierarchical convolutional networks. A vertical proposal mechanism is utilized to avoid proposal classification, while regression layer remains working to improve localization accuracy. Our approach evaluated on ICDAR2013 dataset achieves F-measure of 0.91, which outperforms previous state-of-the-art results in scene text detection.
Tasks Object Detection, Scene Text Detection
Published 2017-11-11
URL http://arxiv.org/abs/1711.04147v1
PDF http://arxiv.org/pdf/1711.04147v1.pdf
PWC https://paperswithcode.com/paper/deep-residual-text-detection-network-for
Repo
Framework

Multitask Learning with Deep Neural Networks for Community Question Answering

Title Multitask Learning with Deep Neural Networks for Community Question Answering
Authors Daniele Bonadiman, Antonio Uva, Alessandro Moschitti
Abstract In this paper, we developed a deep neural network (DNN) that learns to solve simultaneously the three tasks of the cQA challenge proposed by the SemEval-2016 Task 3, i.e., question-comment similarity, question-question similarity and new question-comment similarity. The latter is the main task, which can exploit the previous two for achieving better results. Our DNN is trained jointly on all the three cQA tasks and learns to encode questions and comments into a single vector representation shared across the multiple tasks. The results on the official challenge test set show that our approach produces higher accuracy and faster convergence rates than the individual neural networks. Additionally, our method, which does not use any manual feature engineering, approaches the state of the art established with methods that make heavy use of it.
Tasks Community Question Answering, Feature Engineering, Question Answering, Question Similarity
Published 2017-02-13
URL http://arxiv.org/abs/1702.03706v1
PDF http://arxiv.org/pdf/1702.03706v1.pdf
PWC https://paperswithcode.com/paper/multitask-learning-with-deep-neural-networks
Repo
Framework

R2N2: Residual Recurrent Neural Networks for Multivariate Time Series Forecasting

Title R2N2: Residual Recurrent Neural Networks for Multivariate Time Series Forecasting
Authors Hardik Goel, Igor Melnyk, Arindam Banerjee
Abstract Multivariate time-series modeling and forecasting is an important problem with numerous applications. Traditional approaches such as VAR (vector auto-regressive) models and more recent approaches such as RNNs (recurrent neural networks) are indispensable tools in modeling time-series data. In many multivariate time series modeling problems, there is usually a significant linear dependency component, for which VARs are suitable, and a nonlinear component, for which RNNs are suitable. Modeling such times series with only VAR or only RNNs can lead to poor predictive performance or complex models with large training times. In this work, we propose a hybrid model called R2N2 (Residual RNN), which first models the time series with a simple linear model (like VAR) and then models its residual errors using RNNs. R2N2s can be trained using existing algorithms for VARs and RNNs. Through an extensive empirical evaluation on two real world datasets (aviation and climate domains), we show that R2N2 is competitive, usually better than VAR or RNN, used alone. We also show that R2N2 is faster to train as compared to an RNN, while requiring less number of hidden units.
Tasks Multivariate Time Series Forecasting, Time Series, Time Series Forecasting
Published 2017-09-10
URL http://arxiv.org/abs/1709.03159v1
PDF http://arxiv.org/pdf/1709.03159v1.pdf
PWC https://paperswithcode.com/paper/r2n2-residual-recurrent-neural-networks-for
Repo
Framework

Automatic Classification of Cancerous Tissue in Laserendomicroscopy Images of the Oral Cavity using Deep Learning

Title Automatic Classification of Cancerous Tissue in Laserendomicroscopy Images of the Oral Cavity using Deep Learning
Authors Marc Aubreville, Christian Knipfer, Nicolai Oetter, Christian Jaremenko, Erik Rodner, Joachim Denzler, Christopher Bohr, Helmut Neumann, Florian Stelzle, Andreas Maier
Abstract Oral Squamous Cell Carcinoma (OSCC) is a common type of cancer of the oral epithelium. Despite their high impact on mortality, sufficient screening methods for early diagnosis of OSCC often lack accuracy and thus OSCCs are mostly diagnosed at a late stage. Early detection and accurate outline estimation of OSCCs would lead to a better curative outcome and an reduction in recurrence rates after surgical treatment. Confocal Laser Endomicroscopy (CLE) records sub-surface micro-anatomical images for in vivo cell structure analysis. Recent CLE studies showed great prospects for a reliable, real-time ultrastructural imaging of OSCC in situ. We present and evaluate a novel automatic approach for a highly accurate OSCC diagnosis using deep learning technologies on CLE images. The method is compared against textural feature-based machine learning approaches that represent the current state of the art. For this work, CLE image sequences (7894 images) from patients diagnosed with OSCC were obtained from 4 specific locations in the oral cavity, including the OSCC lesion. The present approach is found to outperform the state of the art in CLE image recognition with an area under the curve (AUC) of 0.96 and a mean accuracy of 88.3% (sensitivity 86.6%, specificity 90%).
Tasks
Published 2017-03-05
URL http://arxiv.org/abs/1703.01622v2
PDF http://arxiv.org/pdf/1703.01622v2.pdf
PWC https://paperswithcode.com/paper/automatic-classification-of-cancerous-tissue
Repo
Framework

Deep learning for undersampled MRI reconstruction

Title Deep learning for undersampled MRI reconstruction
Authors Chang Min Hyun, Hwa Pyung Kim, Sung Min Lee, Sungchul Lee, Jin Keun Seo
Abstract This paper presents a deep learning method for faster magnetic resonance imaging (MRI) by reducing k-space data with sub-Nyquist sampling strategies and provides a rationale for why the proposed approach works well. Uniform subsampling is used in the time-consuming phase-encoding direction to capture high-resolution image information, while permitting the image-folding problem dictated by the Poisson summation formula. To deal with the localization uncertainty due to image folding, very few low-frequency k-space data are added. Training the deep learning net involves input and output images that are pairs of Fourier transforms of the subsampled and fully sampled k-space data. Numerous experiments show the remarkable performance of the proposed method; only 29% of k-space data can generate images of high quality as effectively as standard MRI reconstruction with fully sampled data.
Tasks
Published 2017-09-08
URL https://arxiv.org/abs/1709.02576v3
PDF https://arxiv.org/pdf/1709.02576v3.pdf
PWC https://paperswithcode.com/paper/deep-learning-for-undersampled-mri
Repo
Framework

Discourse-Based Objectives for Fast Unsupervised Sentence Representation Learning

Title Discourse-Based Objectives for Fast Unsupervised Sentence Representation Learning
Authors Yacine Jernite, Samuel R. Bowman, David Sontag
Abstract This work presents a novel objective function for the unsupervised training of neural network sentence encoders. It exploits signals from paragraph-level discourse coherence to train these models to understand text. Our objective is purely discriminative, allowing us to train models many times faster than was possible under prior methods, and it yields models which perform well in extrinsic evaluations.
Tasks Representation Learning
Published 2017-04-23
URL http://arxiv.org/abs/1705.00557v1
PDF http://arxiv.org/pdf/1705.00557v1.pdf
PWC https://paperswithcode.com/paper/discourse-based-objectives-for-fast
Repo
Framework

Arc-Standard Spinal Parsing with Stack-LSTMs

Title Arc-Standard Spinal Parsing with Stack-LSTMs
Authors Miguel Ballesteros, Xavier Carreras
Abstract We present a neural transition-based parser for spinal trees, a dependency representation of constituent trees. The parser uses Stack-LSTMs that compose constituent nodes with dependency-based derivations. In experiments, we show that this model adapts to different styles of dependency relations, but this choice has little effect for predicting constituent structure, suggesting that LSTMs induce useful states by themselves.
Tasks
Published 2017-09-01
URL http://arxiv.org/abs/1709.00489v1
PDF http://arxiv.org/pdf/1709.00489v1.pdf
PWC https://paperswithcode.com/paper/arc-standard-spinal-parsing-with-stack-lstms
Repo
Framework

A Human and Group Behaviour Simulation Evaluation Framework utilising Composition and Video Analysis

Title A Human and Group Behaviour Simulation Evaluation Framework utilising Composition and Video Analysis
Authors Rob Dupre, Vasileios Argyriou
Abstract In this work we present the modular Crowd Simulation Evaluation through Composition framework (CSEC) which provides a quantitative comparison between different pedestrian and crowd simulation approaches. Evaluation is made based on the comparison of source footage against synthetic video created through novel composition techniques. The proposed framework seeks to reduce the complexity of simulation evaluation and provide a platform from which the comparison of differing simulation algorithms as well as parametric tuning can be conducted to improve simulation accuracy or providing measures of similarity between crowd simulation algorithms and source data. Through the use of features designed to mimic the Human Visual System (HVS), specific simulation properties can be evaluated relative to sample footage. Validation was performed on a number of popular crowd datasets and through comparisons of multiple pedestrian and crowd simulation algorithms.
Tasks
Published 2017-07-09
URL http://arxiv.org/abs/1707.02655v3
PDF http://arxiv.org/pdf/1707.02655v3.pdf
PWC https://paperswithcode.com/paper/a-human-and-group-behaviour-simulation
Repo
Framework

Variational Inference using Implicit Distributions

Title Variational Inference using Implicit Distributions
Authors Ferenc Huszár
Abstract Generative adversarial networks (GANs) have given us a great tool to fit implicit generative models to data. Implicit distributions are ones we can sample from easily, and take derivatives of samples with respect to model parameters. These models are highly expressive and we argue they can prove just as useful for variational inference (VI) as they are for generative modelling. Several papers have proposed GAN-like algorithms for inference, however, connections to the theory of VI are not always well understood. This paper provides a unifying review of existing algorithms establishing connections between variational autoencoders, adversarially learned inference, operator VI, GAN-based image reconstruction, and more. Secondly, the paper provides a framework for building new algorithms: depending on the way the variational bound is expressed we introduce prior-contrastive and joint-contrastive methods, and show practical inference algorithms based on either density ratio estimation or denoising.
Tasks Denoising, Image Reconstruction
Published 2017-02-27
URL http://arxiv.org/abs/1702.08235v1
PDF http://arxiv.org/pdf/1702.08235v1.pdf
PWC https://paperswithcode.com/paper/variational-inference-using-implicit
Repo
Framework

C3A: A Cognitive Collaborative Control Architecture For an Intelligent Wheelchair

Title C3A: A Cognitive Collaborative Control Architecture For an Intelligent Wheelchair
Authors Rupam Bhattacharyya, Adity Saikia, Shyamanta M. Hazarika
Abstract Retention of residual skills for persons who partially lose their cognitive or physical ability is of utmost importance. Research is focused on developing systems that provide need-based assistance for retention of such residual skills. This paper describes a novel cognitive collaborative control architecture C3A, designed to address the challenges of developing need- based assistance for wheelchair navigation. Organization of C3A is detailed and results from simulation of the proposed architecture is presented. For simulation of our proposed architecture, we have used ROS (Robot Operating System) as a control framework and a 3D robotic simulator called USARSim (Unified System for Automation and Robot Simulation).
Tasks
Published 2017-01-30
URL http://arxiv.org/abs/1701.08761v1
PDF http://arxiv.org/pdf/1701.08761v1.pdf
PWC https://paperswithcode.com/paper/c3a-a-cognitive-collaborative-control
Repo
Framework

MaMaDroid: Detecting Android Malware by Building Markov Chains of Behavioral Models (Extended Version)

Title MaMaDroid: Detecting Android Malware by Building Markov Chains of Behavioral Models (Extended Version)
Authors Lucky Onwuzurike, Enrico Mariconti, Panagiotis Andriotis, Emiliano De Cristofaro, Gordon Ross, Gianluca Stringhini
Abstract As Android has become increasingly popular, so has malware targeting it, thus pushing the research community to propose different detection techniques. However, the constant evolution of the Android ecosystem, and of malware itself, makes it hard to design robust tools that can operate for long periods of time without the need for modifications or costly re-training. Aiming to address this issue, we set to detect malware from a behavioral point of view, modeled as the sequence of abstracted API calls. We introduce MaMaDroid, a static-analysis based system that abstracts the API calls performed by an app to their class, package, or family, and builds a model from their sequences obtained from the call graph of an app as Markov chains. This ensures that the model is more resilient to API changes and the features set is of manageable size. We evaluate MaMaDroid using a dataset of 8.5K benign and 35.5K malicious apps collected over a period of six years, showing that it effectively detects malware (with up to 0.99 F-measure) and keeps its detection capabilities for long periods of time (up to 0.87 F-measure two years after training). We also show that MaMaDroid remarkably outperforms DroidAPIMiner, a state-of-the-art detection system that relies on the frequency of (raw) API calls. Aiming to assess whether MaMaDroid’s effectiveness mainly stems from the API abstraction or from the sequencing modeling, we also evaluate a variant of it that uses frequency (instead of sequences), of abstracted API calls. We find that it is not as accurate, failing to capture maliciousness when trained on malware samples that include API calls that are equally or more frequently used by benign apps.
Tasks
Published 2017-11-20
URL http://arxiv.org/abs/1711.07477v2
PDF http://arxiv.org/pdf/1711.07477v2.pdf
PWC https://paperswithcode.com/paper/mamadroid-detecting-android-malware-by-1
Repo
Framework

Context aware saliency map generation using semantic segmentation

Title Context aware saliency map generation using semantic segmentation
Authors Mahdi Ahmadi, Mohsen Hajabdollahi, Nader Karimi, Shadrokh Samavi
Abstract Saliency map detection, as a method for detecting important regions of an image, is used in many applications such as image classification and recognition. We propose that context detection could have an essential role in image saliency detection. This requires extraction of high level features. In this paper a saliency map is proposed, based on image context detection using semantic segmentation as a high level feature. Saliency map from semantic information is fused with color and contrast based saliency maps. The final saliency map is then generated. Simulation results for Pascal-voc11 image dataset show 99% accuracy in context detection. Also final saliency map produced by our proposed method shows acceptable results in detecting salient points.
Tasks Image Classification, Saliency Detection, Semantic Segmentation
Published 2017-12-31
URL http://arxiv.org/abs/1801.00256v2
PDF http://arxiv.org/pdf/1801.00256v2.pdf
PWC https://paperswithcode.com/paper/context-aware-saliency-map-generation-using
Repo
Framework

Synergistic Union of Word2Vec and Lexicon for Domain Specific Semantic Similarity

Title Synergistic Union of Word2Vec and Lexicon for Domain Specific Semantic Similarity
Authors Keet Sugathadasa, Buddhi Ayesha, Nisansa de Silva, Amal Shehan Perera, Vindula Jayawardana, Dimuthu Lakmal, Madhavi Perera
Abstract Semantic similarity measures are an important part in Natural Language Processing tasks. However Semantic similarity measures built for general use do not perform well within specific domains. Therefore in this study we introduce a domain specific semantic similarity measure that was created by the synergistic union of word2vec, a word embedding method that is used for semantic similarity calculation and lexicon based (lexical) semantic similarity methods. We prove that this proposed methodology out performs word embedding methods trained on generic corpus and methods trained on domain specific corpus but do not use lexical semantic similarity methods to augment the results. Further, we prove that text lemmatization can improve the performance of word embedding methods.
Tasks Lemmatization, Semantic Similarity, Semantic Textual Similarity
Published 2017-06-06
URL http://arxiv.org/abs/1706.01967v2
PDF http://arxiv.org/pdf/1706.01967v2.pdf
PWC https://paperswithcode.com/paper/synergistic-union-of-word2vec-and-lexicon-for
Repo
Framework

Bayesian Decision Making in Groups is Hard

Title Bayesian Decision Making in Groups is Hard
Authors Jan Hązła, Ali Jadbabaie, Elchanan Mossel, M. Amin Rahimian
Abstract We study the computations that Bayesian agents undertake when exchanging opinions over a network. The agents act repeatedly on their private information and take myopic actions that maximize their expected utility according to a fully rational posterior belief. We show that such computations are NP-hard for two natural utility functions: one with binary actions, and another where agents reveal their posterior beliefs. In fact, we show that distinguishing between posteriors that are concentrated on different states of the world is NP-hard. Therefore, even approximating the Bayesian posterior beliefs is hard. We also describe a natural search algorithm to compute agents’ actions, which we call elimination of impossible signals, and show that if the network is transitive, the algorithm can be modified to run in polynomial time.
Tasks Decision Making
Published 2017-05-12
URL https://arxiv.org/abs/1705.04770v4
PDF https://arxiv.org/pdf/1705.04770v4.pdf
PWC https://paperswithcode.com/paper/bayesian-decision-making-in-groups-is-hard
Repo
Framework
comments powered by Disqus