July 26, 2019

2562 words 13 mins read

Paper Group NANR 6

Paper Group NANR 6

IoT Data Analytics Using Deep Learning. The Effect of Error Rate in Artificially Generated Data for Automatic Preposition and Determiner Correction. Nearly Optimal Robust Matrix Completion. Deep Learning in Semantic Kernel Spaces. Extraction of Gene-Environment Interaction from the Biomedical Literature. Non-Deterministic Segmentation for Chinese L …

IoT Data Analytics Using Deep Learning

Title IoT Data Analytics Using Deep Learning
Authors Xiaofeng Xie, Di Wu, Siping Liu, Renfa Li
Abstract Deep learning is a popular machine learning approach which has achieved a lot of progress in all traditional machine learning areas. Internet of thing (IoT) and Smart City deployments are generating large amounts of time-series sensor data in need of analysis. Applying deep learning to these domains has been an important topic of research. The Long-Short Term Memory (LSTM) network has been proven to be well suited for dealing with and predicting important events with long intervals and delays in the time series. LTSM networks have the ability to maintain long-term memory. In an LTSM network, a stacked LSTM hidden layer also makes it possible to learn a high level temporal feature without the need of any fine tuning and preprocessing which would be required by other techniques. In this paper, we construct a long-short term memory (LSTM) recurrent neural network structure, use the normal time series training set to build the prediction model. And then we use the predicted error from the prediction model to construct a Gaussian naive Bayes model to detect whether the original sample is abnormal. This method is called LSTM-Gauss-NBayes for short. We use three real-world data sets, each of which involve long-term time-dependence or short-term time-dependence, even very weak time dependence. The experimental results show that LSTM-Gauss-NBayes is an effective and robust model.
Tasks Time Series
Published 2017-08-13
URL https://arxiv.org/abs/1708.03854
PDF https://arxiv.org/ftp/arxiv/papers/1708/1708.03854.pdf
PWC https://paperswithcode.com/paper/iot-data-analytics-using-deep-learning
Repo
Framework

The Effect of Error Rate in Artificially Generated Data for Automatic Preposition and Determiner Correction

Title The Effect of Error Rate in Artificially Generated Data for Automatic Preposition and Determiner Correction
Authors Fraser Bowen, Jon Dehdari, Josef van Genabith
Abstract In this research we investigate the impact of mismatches in the density and type of error between training and test data on a neural system correcting preposition and determiner errors. We use synthetically produced training data to control error density and type, and {``}real{''} error data for testing. Our results show it is possible to combine error types, although prepositions and determiners behave differently in terms of how much error should be artificially introduced into the training data in order to get the best results. |
Tasks Grammatical Error Correction, Machine Translation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4410/
PDF https://www.aclweb.org/anthology/W17-4410
PWC https://paperswithcode.com/paper/the-effect-of-error-rate-in-artificially
Repo
Framework

Nearly Optimal Robust Matrix Completion

Title Nearly Optimal Robust Matrix Completion
Authors Yeshwanth Cherapanamjeri, Kartik Gupta, Prateek Jain
Abstract In this paper, we consider the problem of Robust Matrix Completion (RMC) where the goal is to recover a low-rank matrix by observing a small number of its entries out of which a few can be arbitrarily corrupted. We propose a simple projected gradient descent-based method to estimate the low-rank matrix that alternately performs a projected gradient descent step and cleans up a few of the corrupted entries using hard-thresholding. Our algorithm solves RMC using nearly optimal number of observations while tolerating a nearly optimal number of corruptions. Our result also implies significant improvement over the existing time complexity bounds for the low-rank matrix completion problem. Finally, an application of our result to the robust PCA problem (low-rank+sparse matrix separation) leads to nearly linear time (in matrix dimensions) algorithm for the same; existing state-of-the-art methods require quadratic time. Our empirical results corroborate our theoretical results and show that even for moderate sized problems, our method for robust PCA is an order of magnitude faster than the existing methods.
Tasks Low-Rank Matrix Completion, Matrix Completion
Published 2017-08-01
URL https://icml.cc/Conferences/2017/Schedule?showEvent=772
PDF http://proceedings.mlr.press/v70/cherapanamjeri17a/cherapanamjeri17a.pdf
PWC https://paperswithcode.com/paper/nearly-optimal-robust-matrix-completion-1
Repo
Framework

Deep Learning in Semantic Kernel Spaces

Title Deep Learning in Semantic Kernel Spaces
Authors Danilo Croce, Simone Filice, Giuseppe Castellucci, Roberto Basili
Abstract Kernel methods enable the direct usage of structured representations of textual data during language learning and inference tasks. Expressive kernels, such as Tree Kernels, achieve excellent performance in NLP. On the other side, deep neural networks have been demonstrated effective in automatically learning feature representations during training. However, their input is tensor data, i.e., they can not manage rich structured information. In this paper, we show that expressive kernels and deep neural networks can be combined in a common framework in order to (i) explicitly model structured information and (ii) learn non-linear decision functions. We show that the input layer of a deep architecture can be pre-trained through the application of the Nystrom low-rank approximation of kernel spaces. The resulting {``}kernelized{''} neural network achieves state-of-the-art accuracy in three different tasks. |
Tasks Feature Engineering, Information Retrieval, Question Answering
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-1032/
PDF https://www.aclweb.org/anthology/P17-1032
PWC https://paperswithcode.com/paper/deep-learning-in-semantic-kernel-spaces
Repo
Framework

Extraction of Gene-Environment Interaction from the Biomedical Literature

Title Extraction of Gene-Environment Interaction from the Biomedical Literature
Authors Jinseon You, Jin-Woo Chung, Wonsuk Yang, Jong C. Park
Abstract Genetic information in the literature has been extensively looked into for the purpose of discovering the etiology of a disease. As the gene-disease relation is sensitive to external factors, their identification is important to study a disease. Environmental influences, which are usually called Gene-Environment interaction (GxE), have been considered as important factors and have extensively been researched in biology. Nevertheless, there is still a lack of systems for automatic GxE extraction from the biomedical literature due to new challenges: (1) there are no preprocessing tools and corpora for GxE, (2) expressions of GxE are often quite implicit, and (3) document-level comprehension is usually required. We propose to overcome these challenges with neural network models and show that a modified sequence-to-sequence model with a static RNN decoder produces a good performance in GxE recognition.
Tasks Named Entity Recognition
Published 2017-11-01
URL https://www.aclweb.org/anthology/I17-1087/
PDF https://www.aclweb.org/anthology/I17-1087
PWC https://paperswithcode.com/paper/extraction-of-gene-environment-interaction
Repo
Framework

Non-Deterministic Segmentation for Chinese Lattice Parsing

Title Non-Deterministic Segmentation for Chinese Lattice Parsing
Authors Hai Hu, Daniel Dakota, S K{"u}bler, ra
Abstract Parsing Chinese critically depends on correct word segmentation for the parser since incorrect segmentation inevitably causes incorrect parses. We investigate a pipeline approach to segmentation and parsing using word lattices as parser input. We compare CRF-based and lexicon-based approaches to word segmentation. Our results show that the lattice parser is capable of selecting the correction segmentation from thousands of options, thus drastically reducing the number of unparsed sentence. Lexicon-based parsing models have a better coverage than the CRF-based approach, but the many options are more difficult to handle. We reach our best result by using a lexicon from the n-best CRF analyses, combined with highly probable words.
Tasks Morphological Analysis
Published 2017-09-01
URL https://www.aclweb.org/anthology/R17-1043/
PDF https://doi.org/10.26615/978-954-452-049-6_043
PWC https://paperswithcode.com/paper/non-deterministic-segmentation-for-chinese
Repo
Framework

Supervised Attention for Sequence-to-Sequence Constituency Parsing

Title Supervised Attention for Sequence-to-Sequence Constituency Parsing
Authors Hidetaka Kamigaito, Katsuhiko Hayashi, Tsutomu Hirao, Hiroya Takamura, Manabu Okumura, Masaaki Nagata
Abstract The sequence-to-sequence (Seq2Seq) model has been successfully applied to machine translation (MT). Recently, MT performances were improved by incorporating supervised attention into the model. In this paper, we introduce supervised attention to constituency parsing that can be regarded as another translation task. Evaluation results on the PTB corpus showed that the bracketing F-measure was improved by supervised attention.
Tasks Constituency Parsing, Machine Translation, Text Generation, Text Summarization
Published 2017-11-01
URL https://www.aclweb.org/anthology/I17-2002/
PDF https://www.aclweb.org/anthology/I17-2002
PWC https://paperswithcode.com/paper/supervised-attention-for-sequence-to-sequence
Repo
Framework

Coherence Pursuit: Fast, Simple, and Robust Subspace Recovery

Title Coherence Pursuit: Fast, Simple, and Robust Subspace Recovery
Authors Mostafa Rahmani, George Atia
Abstract This paper presents a remarkably simple, yet powerful, algorithm for robust Principal Component Analysis (PCA). In the proposed approach, an outlier is set apart from an inlier by comparing their coherence with the rest of the data points. As inliers lie on a low dimensional subspace, they are likely to have strong mutual coherence provided there are enough inliers. By contrast, outliers do not typically admit low dimensional structures, wherefore an outlier is unlikely to bear strong resemblance with a large number of data points. The mutual coherences are computed by forming the Gram matrix of normalized data points. Subsequently, the subspace is recovered from the span of a small subset of the data points that exhibit strong coherence with the rest of the data. As coherence pursuit only involves one simple matrix multiplication, it is significantly faster than the state of-the-art robust PCA algorithms. We provide a mathematical analysis of the proposed algorithm under a random model for the distribution of the inliers and outliers. It is shown that the proposed method can recover the correct subspace even if the data is predominantly outliers. To the best of our knowledge, this is the first provable robust PCA algorithm that is simultaneously non-iterative, can tolerate a large number of outliers and is robust to linearly dependent outliers.
Tasks
Published 2017-08-01
URL https://icml.cc/Conferences/2017/Schedule?showEvent=527
PDF http://proceedings.mlr.press/v70/rahmani17a/rahmani17a.pdf
PWC https://paperswithcode.com/paper/coherence-pursuit-fast-simple-and-robust
Repo
Framework

Toward a Comparable Corpus of Latvian, Russian and English Tweets

Title Toward a Comparable Corpus of Latvian, Russian and English Tweets
Authors Dmitrijs Milajevs
Abstract Twitter has become a rich source for linguistic data. Here, a possibility of building a trilingual Latvian-Russian-English corpus of tweets from Riga, Latvia is investigated. Such a corpus, once constructed, might be of great use for multiple purposes including training machine translation models, examining cross-lingual phenomena and studying the population of Riga. This pilot study shows that it is feasible to build such a resource by collecting and analysing a pilot corpus, which is made publicly available and can be used to construct a large comparable corpus.
Tasks Information Retrieval, Machine Translation
Published 2017-08-01
URL https://www.aclweb.org/anthology/W17-2505/
PDF https://www.aclweb.org/anthology/W17-2505
PWC https://paperswithcode.com/paper/toward-a-comparable-corpus-of-latvian-russian
Repo
Framework

Deception detection in Russian texts

Title Deception detection in Russian texts
Authors Olga Litvinova, Pavel Seredin, Tatiana Litvinova, John Lyell
Abstract Humans are known to detect deception in speech randomly and it is therefore important to develop tools to enable them to detect deception. The problem of deception detection has been studied for a significant amount of time, however the last 10-15 years have seen methods of computational linguistics being employed. Texts are processed using different NLP tools and then classified as deceptive/truthful using machine learning methods. While most research has been performed for English, Slavic languages have never been a focus of detection deception studies. The paper deals with deception detection in Russian narratives. It employs a specially designed corpus of truthful and deceptive texts on the same topic from each respondent, N = 113. The texts were processed using Linguistic Inquiry and Word Count software that is used in most studies of text-based deception detection. The list of parameters computed using the software was expanded due to the designed users{'} dictionaries. A variety of text classification methods was employed. The accuracy of the model was found to depend on the author{'}s gender and text type (deceptive/truthful).
Tasks Deception Detection, Text Classification
Published 2017-04-01
URL https://www.aclweb.org/anthology/E17-4005/
PDF https://www.aclweb.org/anthology/E17-4005
PWC https://paperswithcode.com/paper/deception-detection-in-russian-texts
Repo
Framework

MT/IE: Cross-lingual Open Information Extraction with Neural Sequence-to-Sequence Models

Title MT/IE: Cross-lingual Open Information Extraction with Neural Sequence-to-Sequence Models
Authors Sheng Zhang, Kevin Duh, Benjamin Van Durme
Abstract Cross-lingual information extraction is the task of distilling facts from foreign language (e.g. Chinese text) into representations in another language that is preferred by the user (e.g. English tuples). Conventional pipeline solutions decompose the task as machine translation followed by information extraction (or vice versa). We propose a joint solution with a neural sequence model, and show that it outperforms the pipeline in a cross-lingual open information extraction setting by 1-4 BLEU and 0.5-0.8 F1.
Tasks Machine Translation, Open Information Extraction, Question Answering, Semantic Parsing
Published 2017-04-01
URL https://www.aclweb.org/anthology/E17-2011/
PDF https://www.aclweb.org/anthology/E17-2011
PWC https://paperswithcode.com/paper/mtie-cross-lingual-open-information
Repo
Framework

Exploring Properties of Intralingual and Interlingual Association Measures Visually

Title Exploring Properties of Intralingual and Interlingual Association Measures Visually
Authors Johannes Gra{"e}n, Christof Bless
Abstract
Tasks Language Modelling, Lemmatization, Word Alignment
Published 2017-05-01
URL https://www.aclweb.org/anthology/W17-0245/
PDF https://www.aclweb.org/anthology/W17-0245
PWC https://paperswithcode.com/paper/exploring-properties-of-intralingual-and
Repo
Framework

Proceedings of the Linguistic Resources for Automatic Natural Language Generation - LiRA@NLG

Title Proceedings of the Linguistic Resources for Automatic Natural Language Generation - LiRA@NLG
Authors
Abstract
Tasks Text Generation
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-3800/
PDF https://www.aclweb.org/anthology/W17-3800
PWC https://paperswithcode.com/paper/proceedings-of-the-linguistic-resources-for
Repo
Framework

A Framework for Software Defect Prediction and Metric Selection

Title A Framework for Software Defect Prediction and Metric Selection
Authors SHAMSUL HUDA 1, SULTAN ALYAHYA2, MD MOHSIN ALI3, SHAFIQ AHMAD 4, JEMAL ABAWAJY1, HMOOD AL-DOSSARI2, AND JOHN YEARWOOD1
Abstract ABSTRACT Automated software defect prediction is an important and fundamental activity in the domain of software development. However, modern software systems are inherently large and complex with numerous correlated metrics that capture different aspects of the software components. This large number of correlated metrics makes building a software defect prediction model very complex. Thus, identifying and selecting a subset of metrics that enhance the software defect prediction method’s performance are an important but challenging problem that has received little attention in the literature. The main objective of this paper is to identify significant software metrics, to build and evaluate an automated software defect prediction model. We propose two novel hybrid software defect prediction models to identify the significant attributes (metrics) using a combination of wrapper and filter techniques. The novelty of our approach is that it embeds the metric selection and training processes of software defect prediction as a single process while reducing the measurement overhead significantly. Different wrapper approaches were combined, including SVM and ANN, with a maximum relevance filter approach to find the significant metrics. A filter score was injected into the wrapper selection process in the proposed approaches to direct the search process efficiently to identify significant metrics. Experimental results with real defect-prone software data sets show that the proposed hybrid approaches achieve significantly compact metrics (i.e., selecting the most significant metrics) with high prediction accuracy compared with conventional wrapper or filter approaches. The performance of the proposed framework has also been verified using a statistical multivariate quality control process using multivariate exponentially weighted moving average. The proposed framework demonstrates that the hybrid heuristic can guide the metric selection process in a computationally efficient way by integrating the intrinsic characteristics from the filters into the wrapper and using the advantages of both the filter and wrapper approaches.
Tasks
Published 2017-12-27
URL https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=8240899
PDF https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=8240899
PWC https://paperswithcode.com/paper/a-framework-for-software-defect-prediction
Repo
Framework

Bregman Divergence for Stochastic Variance Reduction: Saddle-Point and Adversarial Prediction

Title Bregman Divergence for Stochastic Variance Reduction: Saddle-Point and Adversarial Prediction
Authors Zhan Shi, Xinhua Zhang, Yaoliang Yu
Abstract Adversarial machines, where a learner competes against an adversary, have regained much recent interest in machine learning. They are naturally in the form of saddle-point optimization, often with separable structure but sometimes also with unmanageably large dimension. In this work we show that adversarial prediction under multivariate losses can be solved much faster than they used to be. We first reduce the problem size exponentially by using appropriate sufficient statistics, and then we adapt the new stochastic variance-reduced algorithm of Balamurugan & Bach (2016) to allow any Bregman divergence. We prove that the same linear rate of convergence is retained and we show that for adversarial prediction using KL-divergence we can further achieve a speedup of #example times compared with the Euclidean alternative. We verify the theoretical findings through extensive experiments on two example applications: adversarial prediction and LPboosting.
Tasks
Published 2017-12-01
URL http://papers.nips.cc/paper/7184-bregman-divergence-for-stochastic-variance-reduction-saddle-point-and-adversarial-prediction
PDF http://papers.nips.cc/paper/7184-bregman-divergence-for-stochastic-variance-reduction-saddle-point-and-adversarial-prediction.pdf
PWC https://paperswithcode.com/paper/bregman-divergence-for-stochastic-variance
Repo
Framework
comments powered by Disqus