July 26, 2019

1878 words 9 mins read

Paper Group NANR 69

Paper Group NANR 69

Hybrid Neural Networks for Learning the Trend in Time Series. Exploring Vector Spaces for Semantic Relations. An Object-oriented Model of Role Framing and Attitude Prediction. Maximum Margin Reward Networks for Learning from Explicit and Implicit Supervision. Neural Sequence Learning Models for Word Sense Disambiguation. If mice were reptiles, then …

Hybrid Neural Networks for Learning the Trend in Time Series

Title Hybrid Neural Networks for Learning the Trend in Time Series
Authors Tao Lin∗,Tian Guo∗,Karl Aberer
Abstract Trend of time series characterizes the intermediate upward and downward behaviour of time series. Learning and forecasting the trend in time series data play an important role in many real applica- tions, ranging from resource allocation in data cen- ters, load schedule in smart grid, and so on. In- spired by the recent successes of neural networks, in this paper we propose TreNet, a novel end-to- end hybrid neural network to learn local and global contextual features for predicting the trend of time series. TreNet leverages convolutional neural net- works (CNNs) to extract salient features from local raw data of time series. Meanwhile, considering the long-range dependency existing in the sequence of historical trends, TreNet uses a long-short term memory recurrent neural network (LSTM) to cap- ture such dependency. Then, a feature fusion layer is to learn joint representation for predicting the trend. TreNet demonstrates its effectiveness by out- performing CNN, LSTM, the cascade of CNN and LSTM, Hidden Markov Model based method and various kernel based baselines on real datasets.
Tasks Time Series
Published 2017-07-01
URL https://www.ijcai.org/Proceedings/2017/316
PDF https://www.ijcai.org/Proceedings/2017/316
PWC https://paperswithcode.com/paper/hybrid-neural-networks-for-learning-the-trend
Repo
Framework

Exploring Vector Spaces for Semantic Relations

Title Exploring Vector Spaces for Semantic Relations
Authors Kata G{'a}bor, Ha{"\i}fa Zargayouna, Isabelle Tellier, Davide Buscaldi, Thierry Charnois
Abstract Word embeddings are used with success for a variety of tasks involving lexical semantic similarities between individual words. Using unsupervised methods and just cosine similarity, encouraging results were obtained for analogical similarities. In this paper, we explore the potential of pre-trained word embeddings to identify generic types of semantic relations in an unsupervised experiment. We propose a new relational similarity measure based on the combination of word2vec{'}s CBOW input and output vectors which outperforms concurrent vector representations, when used for unsupervised clustering on SemEval 2010 Relation Classification data.
Tasks Relation Classification, Relation Extraction, Semantic Textual Similarity, Word Embeddings
Published 2017-09-01
URL https://www.aclweb.org/anthology/D17-1193/
PDF https://www.aclweb.org/anthology/D17-1193
PWC https://paperswithcode.com/paper/exploring-vector-spaces-for-semantic
Repo
Framework

An Object-oriented Model of Role Framing and Attitude Prediction

Title An Object-oriented Model of Role Framing and Attitude Prediction
Authors Manfred Klenner
Abstract
Tasks
Published 2017-01-01
URL https://www.aclweb.org/anthology/W17-6917/
PDF https://www.aclweb.org/anthology/W17-6917
PWC https://paperswithcode.com/paper/an-object-oriented-model-of-role-framing-and
Repo
Framework

Maximum Margin Reward Networks for Learning from Explicit and Implicit Supervision

Title Maximum Margin Reward Networks for Learning from Explicit and Implicit Supervision
Authors Haoruo Peng, Ming-Wei Chang, Wen-tau Yih
Abstract Neural networks have achieved state-of-the-art performance on several structured-output prediction tasks, trained in a fully supervised fashion. However, annotated examples in structured domains are often costly to obtain, which thus limits the applications of neural networks. In this work, we propose Maximum Margin Reward Networks, a neural network-based framework that aims to learn from both explicit (full structures) and implicit supervision signals (delayed feedback on the correctness of the predicted structure). On named entity recognition and semantic parsing, our model outperforms previous systems on the benchmark datasets, CoNLL-2003 and WebQuestionsSP.
Tasks Dependency Parsing, Named Entity Recognition, Part-Of-Speech Tagging, Semantic Parsing
Published 2017-09-01
URL https://www.aclweb.org/anthology/D17-1252/
PDF https://www.aclweb.org/anthology/D17-1252
PWC https://paperswithcode.com/paper/maximum-margin-reward-networks-for-learning
Repo
Framework

Neural Sequence Learning Models for Word Sense Disambiguation

Title Neural Sequence Learning Models for Word Sense Disambiguation
Authors Aless Raganato, ro, Claudio Delli Bovi, Roberto Navigli
Abstract Word Sense Disambiguation models exist in many flavors. Even though supervised ones tend to perform best in terms of accuracy, they often lose ground to more flexible knowledge-based solutions, which do not require training by a word expert for every disambiguation target. To bridge this gap we adopt a different perspective and rely on sequence learning to frame the disambiguation problem: we propose and study in depth a series of end-to-end neural architectures directly tailored to the task, from bidirectional Long Short-Term Memory to encoder-decoder models. Our extensive evaluation over standard benchmarks and in multiple languages shows that sequence learning enables more versatile all-words models that consistently lead to state-of-the-art results, even against word experts with engineered features.
Tasks Information Retrieval, Machine Translation, Word Sense Disambiguation
Published 2017-09-01
URL https://www.aclweb.org/anthology/D17-1120/
PDF https://www.aclweb.org/anthology/D17-1120
PWC https://paperswithcode.com/paper/neural-sequence-learning-models-for-word
Repo
Framework

If mice were reptiles, then reptiles could be mammals or How to detect errors in the JeuxDeMots lexical network?

Title If mice were reptiles, then reptiles could be mammals or How to detect errors in the JeuxDeMots lexical network?
Authors Mathieu Lafourcade, Alain Joubert, Nathalie Le Brun
Abstract Correcting errors in a data set is a critical issue. This task can be either hand-made by experts, or by crowdsourcing methods, or automatically done using algorithms. Although the rate of errors present in the JeuxDeMots network is rather low, it is important to reduce it. We present here automatic methods for detecting potential secondary errors that would result from automatic inference mechanisms when they rely on an initial error manually detected. Encouraging results also invite us to consider strategies that would automatically detect {``}erroneous{''} initial relations, which could lead to the automatic detection of the majority of errors in the network. |
Tasks
Published 2017-09-01
URL https://www.aclweb.org/anthology/R17-1056/
PDF https://doi.org/10.26615/978-954-452-049-6_056
PWC https://paperswithcode.com/paper/if-mice-were-reptiles-then-reptiles-could-be
Repo
Framework

Training Data Augmentation for Low-Resource Morphological Inflection

Title Training Data Augmentation for Low-Resource Morphological Inflection
Authors Toms Bergmanis, Katharina Kann, Hinrich Sch{"u}tze, Sharon Goldwater
Abstract
Tasks Data Augmentation, Morphological Inflection, Transfer Learning
Published 2017-08-01
URL https://www.aclweb.org/anthology/K17-2002/
PDF https://www.aclweb.org/anthology/K17-2002
PWC https://paperswithcode.com/paper/training-data-augmentation-for-low-resource
Repo
Framework

Cross-Lingual Pronoun Prediction with Deep Recurrent Neural Networks v2.0

Title Cross-Lingual Pronoun Prediction with Deep Recurrent Neural Networks v2.0
Authors Juhani Luotolahti, Jenna Kanerva, Filip Ginter
Abstract In this paper we present our system in the DiscoMT 2017 Shared Task on Crosslingual Pronoun Prediction. Our entry builds on our last year{'}s success, our system based on deep recurrent neural networks outperformed all the other systems with a clear margin. This year we investigate whether different pre-trained word embeddings can be used to improve the neural systems, and whether the recently published Gated Convolutions outperform the Gated Recurrent Units used last year.
Tasks Machine Translation, Word Embeddings
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4808/
PDF https://www.aclweb.org/anthology/W17-4808
PWC https://paperswithcode.com/paper/cross-lingual-pronoun-prediction-with-deep-1
Repo
Framework

Aligning phonemes using finte-state methods

Title Aligning phonemes using finte-state methods
Authors Kimmo Koskenniemi
Abstract
Tasks Speech Synthesis, Spelling Correction
Published 2017-05-01
URL https://www.aclweb.org/anthology/W17-0207/
PDF https://www.aclweb.org/anthology/W17-0207
PWC https://paperswithcode.com/paper/aligning-phonemes-using-finte-state-methods
Repo
Framework

Replacing OOV Words For Dependency Parsing With Distributional Semantics

Title Replacing OOV Words For Dependency Parsing With Distributional Semantics
Authors Prasanth Kolachina, Martin Riedl, Chris Biemann
Abstract
Tasks Dependency Parsing
Published 2017-05-01
URL https://www.aclweb.org/anthology/W17-0202/
PDF https://www.aclweb.org/anthology/W17-0202
PWC https://paperswithcode.com/paper/replacing-oov-words-for-dependency-parsing
Repo
Framework

Coarse-to-Fine Question Answering for Long Documents

Title Coarse-to-Fine Question Answering for Long Documents
Authors Eunsol Choi, Daniel Hewlett, Jakob Uszkoreit, Illia Polosukhin, Alex Lacoste, re, Jonathan Berant
Abstract We present a framework for question answering that can efficiently scale to longer documents while maintaining or even improving performance of state-of-the-art models. While most successful approaches for reading comprehension rely on recurrent neural networks (RNNs), running them over long documents is prohibitively slow because it is difficult to parallelize over sequences. Inspired by how people first skim the document, identify relevant parts, and carefully read these parts to produce an answer, we combine a coarse, fast model for selecting relevant sentences and a more expensive RNN for producing the answer from those sentences. We treat sentence selection as a latent variable trained jointly from the answer only using reinforcement learning. Experiments demonstrate state-of-the-art performance on a challenging subset of the WikiReading dataset and on a new dataset, while speeding up the model by 3.5x-6.7x.
Tasks Question Answering, Reading Comprehension
Published 2017-07-01
URL https://www.aclweb.org/anthology/P17-1020/
PDF https://www.aclweb.org/anthology/P17-1020
PWC https://paperswithcode.com/paper/coarse-to-fine-question-answering-for-long
Repo
Framework

Imitation learning for structured prediction in natural language processing

Title Imitation learning for structured prediction in natural language processing
Authors Andreas Vlachos, Gerasimos Lampouras, Sebastian Riedel
Abstract Imitation learning is a learning paradigm originally developed to learn robotic controllers from demonstrations by humans, e.g. autonomous flight from pilot demonstrations. Recently, algorithms for structured prediction were proposed under this paradigm and have been applied successfully to a number of tasks including syntactic dependency parsing, information extraction, coreference resolution, dynamic feature selection, semantic parsing and natural language generation. Key advantages are the ability to handle large output search spaces and to learn with non-decomposable loss functions. Our aim in this tutorial is to have a unified presentation of the various imitation algorithms for structure prediction, and show how they can be applied to a variety of NLP tasks.All material associated with the tutorial will be made available through https://sheffieldnlp.github.io/ImitationLearningTutorialEACL2017/.
Tasks Coreference Resolution, Dependency Parsing, Feature Selection, Imitation Learning, Semantic Parsing, Structured Prediction, Text Generation
Published 2017-04-01
URL https://www.aclweb.org/anthology/E17-5003/
PDF https://www.aclweb.org/anthology/E17-5003
PWC https://paperswithcode.com/paper/imitation-learning-for-structured-prediction
Repo
Framework

Enriching ASR Lattices with POS Tags for Dependency Parsing

Title Enriching ASR Lattices with POS Tags for Dependency Parsing
Authors Moritz Stiefel, Ngoc Thang Vu
Abstract Parsing speech requires a richer representation than 1-best or n-best hypotheses, e.g. lattices. Moreover, previous work shows that part-of-speech (POS) tags are a valuable resource for parsing. In this paper, we therefore explore a joint modeling approach of automatic speech recognition (ASR) and POS tagging to enrich ASR word lattices. To that end, we manipulate the ASR process from the pronouncing dictionary onward to use word-POS pairs instead of words. We evaluate ASR, POS tagging and dependency parsing (DP) performance demonstrating a successful lattice-based integration of ASR and POS tagging.
Tasks Dependency Parsing, Speech Recognition, Spoken Language Understanding
Published 2017-09-01
URL https://www.aclweb.org/anthology/W17-4605/
PDF https://www.aclweb.org/anthology/W17-4605
PWC https://paperswithcode.com/paper/enriching-asr-lattices-with-pos-tags-for
Repo
Framework

Say the Right Thing Right: Ethics Issues in Natural Language Generation Systems

Title Say the Right Thing Right: Ethics Issues in Natural Language Generation Systems
Authors Charese Smiley, Frank Schilder, Vassilis Plachouras, Jochen L. Leidner
Abstract We discuss the ethical implications of Natural Language Generation systems. We use one particular system as a case study to identify and classify issues, and we provide an ethics checklist, in the hope that future system designers may benefit from conducting their own ethics reviews based on our checklist.
Tasks Text Generation
Published 2017-04-01
URL https://www.aclweb.org/anthology/W17-1613/
PDF https://www.aclweb.org/anthology/W17-1613
PWC https://paperswithcode.com/paper/say-the-right-thing-right-ethics-issues-in
Repo
Framework

Group Linguistic Bias Aware Neural Response Generation

Title Group Linguistic Bias Aware Neural Response Generation
Authors Jianan Wang, Xin Wang, Fang Li, Zhen Xu, Zhuoran Wang, Baoxun Wang
Abstract For practical chatbots, one of the essential factor for improving user experience is the capability of customizing the talking style of the agents, that is, to make chatbots provide responses meeting users{'} preference on language styles, topics, etc. To address this issue, this paper proposes to incorporate linguistic biases, which implicitly involved in the conversation corpora generated by human groups in the Social Network Services (SNS), into the encoder-decoder based response generator. By attaching a specially designed neural component to dynamically control the impact of linguistic biases in response generation, a Group Linguistic Bias Aware Neural Response Generation (GLBA-NRG) model is eventually presented. The experimental results on the dataset from the Chinese SNS show that the proposed architecture outperforms the current response generating models by producing both meaningful and vivid responses with customized styles.
Tasks
Published 2017-12-01
URL https://www.aclweb.org/anthology/W17-6001/
PDF https://www.aclweb.org/anthology/W17-6001
PWC https://paperswithcode.com/paper/group-linguistic-bias-aware-neural-response
Repo
Framework
comments powered by Disqus