October 15, 2019

2488 words 12 mins read

Paper Group NANR 274

Paper Group NANR 274

Neural Network Language Modeling with Letter-based Features and Importance Sampling. YNU-HPCC at SemEval-2018 Task 12: The Argument Reasoning Comprehension Task Using a Bi-directional LSTM with Attention Model. Lyb3b at SemEval-2018 Task 12: Ensemble-based Deep Learning Models for Argument Reasoning Comprehension Task. Learning distributed event re …

Neural Network Language Modeling with Letter-based Features and Importance Sampling

Title Neural Network Language Modeling with Letter-based Features and Importance Sampling
Authors Hainan Xu, Ke Li, Yiming Wang, Jian Wang, Shiyin Kang, Xie Chen, Daniel Povey, Sanjeev Khudanpur
Abstract In this paper we describe an extension of the Kaldi software toolkit to support neural-based language modeling, intended for use in automatic speech recognition (ASR) and related tasks. We combine the use of subword features (letter n-grams) and one-hot encoding of frequent words so that the models can handle large vocabularies containing infrequent words. We propose a new objective function that allows for training of unnormalized probabilities. An importance sampling based method is supported to speed up training when the vocabulary is large. Experimental results on five corpora show that Kaldi-RNNLM rivals other recurrent neural network language model toolkits both on performance and training speed.
Tasks Language Modelling, Speech Recognition
Published 2018-04-15
URL https://www.cs.jhu.edu/~hxu/neural-network-language.pdf
PDF https://www.cs.jhu.edu/~hxu/neural-network-language.pdf
PWC https://paperswithcode.com/paper/neural-network-language-modeling-with-letter
Repo
Framework

YNU-HPCC at SemEval-2018 Task 12: The Argument Reasoning Comprehension Task Using a Bi-directional LSTM with Attention Model

Title YNU-HPCC at SemEval-2018 Task 12: The Argument Reasoning Comprehension Task Using a Bi-directional LSTM with Attention Model
Authors Quanlei Liao, Xutao Yang, Jin Wang, Xuejie Zhang
Abstract An argument is divided into two parts, the claim and the reason. To obtain a clearer conclusion, some additional explanation is required. In this task, the explanations are called warrants. This paper introduces a bi-directional long short term memory (Bi-LSTM) with an attention model to select a correct warrant from two to explain an argument. We address this question as a question-answering system. For each warrant, the model produces a probability that it is correct. Finally, the system chooses the highest correct probability as the answer. Ensemble learning is used to enhance the performance of the model. Among all of the participants, we ranked 15th on the test results.
Tasks Question Answering
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-1187/
PDF https://www.aclweb.org/anthology/S18-1187
PWC https://paperswithcode.com/paper/ynu-hpcc-at-semeval-2018-task-12-the-argument
Repo
Framework

Lyb3b at SemEval-2018 Task 12: Ensemble-based Deep Learning Models for Argument Reasoning Comprehension Task

Title Lyb3b at SemEval-2018 Task 12: Ensemble-based Deep Learning Models for Argument Reasoning Comprehension Task
Authors Yongbin Li, Xiaobing Zhou
Abstract Reasoning is a crucial part of natural language argumentation. In order to comprehend an argument, we have to reconstruct and analyze its reasoning. In this task, given a natural language argument with a reason and a claim, the goal is to choose the correct implicit reasoning from two options, in order to form a reasonable structure of (Reason, Warrant, Claim). Our approach is to build distributed word embedding of reason, warrant and claim respectively, meanwhile, we use a series of frameworks such as CNN model, LSTM model, GRU with attention model and biLSTM with attention model for processing word vector. Finally, ensemble mechanism is used to integrate the results of each framework to improve the final accuracy. Experiments demonstrate superior performance of ensemble mechanism compared to each separate framework. We are the 11th in official results, the final model can reach a 0.568 accuracy rate on the test dataset.
Tasks
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-1193/
PDF https://www.aclweb.org/anthology/S18-1193
PWC https://paperswithcode.com/paper/lyb3b-at-semeval-2018-task-12-ensemble-based
Repo
Framework

Learning distributed event representations with a multi-task approach

Title Learning distributed event representations with a multi-task approach
Authors Xudong Hong, Asad Sayeed, Vera Demberg
Abstract Human world knowledge contains information about prototypical events and their participants and locations. In this paper, we train the first models using multi-task learning that can both predict missing event participants and also perform semantic role classification based on semantic plausibility. Our best-performing model is an improvement over the previous state-of-the-art on thematic fit modelling tasks. The event embeddings learned by the model can additionally be used effectively in an event similarity task, also outperforming the state-of-the-art.
Tasks Multi-Task Learning
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-2002/
PDF https://www.aclweb.org/anthology/S18-2002
PWC https://paperswithcode.com/paper/learning-distributed-event-representations
Repo
Framework

Graph Algebraic Combinatory Categorial Grammar

Title Graph Algebraic Combinatory Categorial Grammar
Authors Sebastian Beschke, Wolfgang Menzel
Abstract This paper describes CCG/AMR, a novel grammar for semantic parsing of Abstract Meaning Representations. CCG/AMR equips Combinatory Categorial Grammar derivations with graph semantics by assigning each CCG combinator an interpretation in terms of a graph algebra. We provide an algorithm that induces a CCG/AMR from a corpus and show that it creates a compact lexicon with low ambiguity and achieves a robust coverage of 78{%} of the examined sentences under ideal conditions. We also identify several phenomena that affect any approach relying either on CCG or graph algebraic approaches for AMR parsing. This includes differences of representation between CCG and AMR, as well as non-compositional constructions that are not expressible through a monotonous construction process. To our knowledge, this paper provides the first analysis of these corpus issues.
Tasks Amr Parsing, Machine Translation, Semantic Parsing
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-2006/
PDF https://www.aclweb.org/anthology/S18-2006
PWC https://paperswithcode.com/paper/graph-algebraic-combinatory-categorial
Repo
Framework

GBD-NER at PARSEME Shared Task 2018: Multi-Word Expression Detection Using Bidirectional Long-Short-Term Memory Networks and Graph-Based Decoding

Title GBD-NER at PARSEME Shared Task 2018: Multi-Word Expression Detection Using Bidirectional Long-Short-Term Memory Networks and Graph-Based Decoding
Authors Tiberiu Boros, Rux Burtica, ra
Abstract This paper addresses the issue of multi-word expression (MWE) detection by employing a new decoding strategy inspired after graph-based parsing. We show that this architecture achieves state-of-the-art results with minimum feature-engineering, just by relying on lexicalized and morphological attributes. We validate our approach in a multilingual setting, using standard MWE corpora supplied in the PARSEME Shared Task.
Tasks Feature Engineering
Published 2018-08-01
URL https://www.aclweb.org/anthology/W18-4928/
PDF https://www.aclweb.org/anthology/W18-4928
PWC https://paperswithcode.com/paper/gbd-ner-at-parseme-shared-task-2018-multi
Repo
Framework

Quantitative Semantic Variation in the Contexts of Concrete and Abstract Words

Title Quantitative Semantic Variation in the Contexts of Concrete and Abstract Words
Authors Daniela Naumann, Diego Frassinelli, Sabine Schulte im Walde
Abstract Across disciplines, researchers are eager to gain insight into empirical features of abstract vs. concrete concepts. In this work, we provide a detailed characterisation of the distributional nature of abstract and concrete words across 16,620 English nouns, verbs and adjectives. Specifically, we investigate the following questions: (1) What is the distribution of concreteness in the contexts of concrete and abstract target words? (2) What are the differences between concrete and abstract words in terms of contextual semantic diversity? (3) How does the entropy of concrete and abstract word contexts differ? Overall, our studies show consistent differences in the distributional representation of concrete and abstract words, thus challenging existing theories of cognition and providing a more fine-grained description of their nature.
Tasks
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-2008/
PDF https://www.aclweb.org/anthology/S18-2008
PWC https://paperswithcode.com/paper/quantitative-semantic-variation-in-the
Repo
Framework

EmoWordNet: Automatic Expansion of Emotion Lexicon Using English WordNet

Title EmoWordNet: Automatic Expansion of Emotion Lexicon Using English WordNet
Authors Gilbert Badaro, Hussein Jundi, Hazem Hajj, Wassim El-Hajj
Abstract Nowadays, social media have become a platform where people can easily express their opinions and emotions about any topic such as politics, movies, music, electronic products and many others. On the other hand, politicians, companies, and businesses are interested in analyzing automatically people{'}s opinions and emotions. In the last decade, a lot of efforts has been put into extracting sentiment polarity from texts. Recently, the focus has expanded to also cover emotion recognition from texts. In this work, we expand an existing emotion lexicon, DepecheMood, by leveraging semantic knowledge from English WordNet (EWN). We create an expanded lexicon, EmoWordNet, consisting of 67K terms aligned with EWN, almost 1.8 times the size of DepecheMood. We also evaluate EmoWordNet in an emotion recognition task using SemEval 2007 news headlines dataset and we achieve an improvement compared to the use of DepecheMood. EmoWordNet is publicly available to speed up research in the field on \url{http://oma-project.com}.
Tasks Emotion Classification, Emotion Recognition, Recommendation Systems, Sentiment Analysis
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-2009/
PDF https://www.aclweb.org/anthology/S18-2009
PWC https://paperswithcode.com/paper/emowordnet-automatic-expansion-of-emotion
Repo
Framework

Named Graphs for Semantic Representation

Title Named Graphs for Semantic Representation
Authors Richard Crouch, Aikaterini-Lida Kalouli
Abstract A position paper arguing that purely graphical representations for natural language semantics lack a fundamental degree of expressiveness, and cannot deal with even basic Boolean operations like negation or disjunction. Moving from graphs to named graphs leads to representations that stand some chance of having sufficient expressive power. Named $\mathcal{FL}_0$ graphs are of particular interest.
Tasks Knowledge Graphs
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-2013/
PDF https://www.aclweb.org/anthology/S18-2013
PWC https://paperswithcode.com/paper/named-graphs-for-semantic-representation
Repo
Framework

Coarse Lexical Frame Acquisition at the Syntax–Semantics Interface Using a Latent-Variable PCFG Model

Title Coarse Lexical Frame Acquisition at the Syntax–Semantics Interface Using a Latent-Variable PCFG Model
Authors Laura Kallmeyer, Behrang QasemiZadeh, Jackie Chi Kit Cheung
Abstract We present a method for unsupervised lexical frame acquisition at the syntax{–}semantics interface. Given a set of input strings derived from dependency parses, our method generates a set of clusters that resemble lexical frame structures. Our work is motivated not only by its practical applications (e.g., to build, or expand the coverage of lexical frame databases), but also to gain linguistic insight into frame structures with respect to lexical distributions in relation to grammatical structures. We model our task using a hierarchical Bayesian network and employ tools and methods from latent variable probabilistic context free grammars (L-PCFGs) for statistical inference and parameter fitting, for which we propose a new split and merge procedure. We show that our model outperforms several baselines on a portion of the Wall Street Journal sentences that we have newly annotated for evaluation purposes.
Tasks Question Answering, Text Summarization
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-2016/
PDF https://www.aclweb.org/anthology/S18-2016
PWC https://paperswithcode.com/paper/coarse-lexical-frame-acquisition-at-the
Repo
Framework

In-domain Context-aware Token Embeddings Improve Biomedical Named Entity Recognition

Title In-domain Context-aware Token Embeddings Improve Biomedical Named Entity Recognition
Authors Golnar Sheikhshabbafghi, Inanc Birol, Anoop Sarkar
Abstract Rapidly expanding volume of publications in the biomedical domain makes it increasingly difficult for a timely evaluation of the latest literature. That, along with a push for automated evaluation of clinical reports, present opportunities for effective natural language processing methods. In this study we target the problem of named entity recognition, where texts are processed to annotate terms that are relevant for biomedical studies. Terms of interest in the domain include gene and protein names, and cell lines and types. Here we report on a pipeline built on Embeddings from Language Models (ELMo) and a deep learning package for natural language processing (AllenNLP). We trained context-aware token embeddings on a dataset of biomedical papers using ELMo, and incorporated these embeddings in the LSTM-CRF model used by AllenNLP for named entity recognition. We show these representations improve named entity recognition for different types of biomedical named entities. We also achieve a new state of the art in gene mention detection on the BioCreative II gene mention shared task.
Tasks Language Modelling, Named Entity Recognition, Question Answering, Sentiment Analysis, Word Embeddings, Word Sense Disambiguation
Published 2018-10-01
URL https://www.aclweb.org/anthology/W18-5618/
PDF https://www.aclweb.org/anthology/W18-5618
PWC https://paperswithcode.com/paper/in-domain-context-aware-token-embeddings
Repo
Framework

Indigenous language technologies in Canada: Assessment, challenges, and successes

Title Indigenous language technologies in Canada: Assessment, challenges, and successes
Authors Patrick Littell, Anna Kazantseva, Rol Kuhn, , Aidan Pine, Antti Arppe, Christopher Cox, Marie-Odile Junker
Abstract In this article, we discuss which text, speech, and image technologies have been developed, and would be feasible to develop, for the approximately 60 Indigenous languages spoken in Canada. In particular, we concentrate on technologies that may be feasible to develop for most or all of these languages, not just those that may be feasible for the few most-resourced of these. We assess past achievements and consider future horizons for Indigenous language transliteration, text prediction, spell-checking, approximate search, machine translation, speech recognition, speaker diarization, speech synthesis, optical character recognition, and computer-aided language learning.
Tasks Machine Translation, Optical Character Recognition, Speaker Diarization, Speech Recognition, Speech Synthesis, Transliteration
Published 2018-08-01
URL https://www.aclweb.org/anthology/C18-1222/
PDF https://www.aclweb.org/anthology/C18-1222
PWC https://paperswithcode.com/paper/indigenous-language-technologies-in-canada
Repo
Framework

Putting Semantics into Semantic Roles

Title Putting Semantics into Semantic Roles
Authors James Allen, Choh Man Teng
Abstract While there have been many proposals for theories of semantic roles over the years, these models are mostly justified by intuition and the only evaluation methods have been inter-annotator agreement. We explore three different ideas for providing more rigorous theories of semantic roles. These ideas give rise to more objective criteria for designing role sets, and lend themselves to some experimental evaluation. We illustrate the discussion by examining the semantic roles in TRIPS.
Tasks
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-2028/
PDF https://www.aclweb.org/anthology/S18-2028
PWC https://paperswithcode.com/paper/putting-semantics-into-semantic-roles
Repo
Framework

Pose Transferrable Person Re-Identification

Title Pose Transferrable Person Re-Identification
Authors Jinxian Liu, Bingbing Ni, Yichao Yan, Peng Zhou, Shuo Cheng, Jianguo Hu
Abstract Person re-identification (ReID) is an important task in the field of intelligent security. A key challenge is how to capture human pose variations, while existing benchmarks (i.e., Market1501, DukeMTMC-reID, CUHK03, etc.) do NOT provide sufficient pose coverage to train a robust ReID system. To address this issue, we propose a pose-transferrable person ReID framework which utilizes pose-transferred sample augmentations (i.e., with ID supervision) to enhance ReID model training. On one hand, novel training samples with rich pose variations are generated via transferring pose instances from MARS dataset, and they are added into the target dataset to facilitate robust training. On the other hand, in addition to the conventional discriminator of GAN (i.e., to distinguish between REAL/FAKE samples), we propose a novel guider sub-network which encourages the generated sample (i.e., with novel pose) towards better satisfying the ReID loss (i.e., cross-entropy ReID loss, triplet ReID loss). In the meantime, an alternative optimization procedure is proposed to train the proposed Generator-Guider-Discriminator network. Experimental results on Market-1501, DukeMTMC-reID and CUHK03 show that our method achieves great performance improvement, and outperforms most state-of-the-art methods without elaborate designing the ReID model.
Tasks Person Re-Identification
Published 2018-06-01
URL http://openaccess.thecvf.com/content_cvpr_2018/html/Liu_Pose_Transferrable_Person_CVPR_2018_paper.html
PDF http://openaccess.thecvf.com/content_cvpr_2018/papers/Liu_Pose_Transferrable_Person_CVPR_2018_paper.pdf
PWC https://paperswithcode.com/paper/pose-transferrable-person-re-identification
Repo
Framework

Solving Feature Sparseness in Text Classification using Core-Periphery Decomposition

Title Solving Feature Sparseness in Text Classification using Core-Periphery Decomposition
Authors Xia Cui, Sadamori Kojaku, Naoki Masuda, Danushka Bollegala
Abstract Feature sparseness is a problem common to cross-domain and short-text classification tasks. To overcome this feature sparseness problem, we propose a novel method based on graph decomposition to find candidate features for expanding feature vectors. Specifically, we first create a feature-relatedness graph, which is subsequently decomposed into core-periphery (CP) pairs and use the peripheries as the expansion candidates of the cores. We expand both training and test instances using the computed related features and use them to train a text classifier. We observe that prioritising features that are common to both training and test instances as cores during the CP decomposition to further improve the accuracy of text classification. We evaluate the proposed CP-decomposition-based feature expansion method on benchmark datasets for cross-domain sentiment classification and short-text classification. Our experimental results show that the proposed method consistently outperforms all baselines on short-text classification tasks, and perform competitively with pivot-based cross-domain sentiment classification methods.
Tasks Domain Adaptation, Sentiment Analysis, Text Classification
Published 2018-06-01
URL https://www.aclweb.org/anthology/S18-2030/
PDF https://www.aclweb.org/anthology/S18-2030
PWC https://paperswithcode.com/paper/solving-feature-sparseness-in-text
Repo
Framework
comments powered by Disqus