January 24, 2020

2383 words 12 mins read

Paper Group NANR 141

Paper Group NANR 141

Leveraging syntactic parsing to improve event annotation matching. Automatic Detection of Translation Direction. HITS-SBD at the FinSBD Task: Machine Learning vs. Rule-based Sentence Boundary Detection. Differences between SMT and NMT Output - a Translators’ Point of View. JRC TMA-CC: Slavic Named Entity Recognition and Linking. Participation in th …

Leveraging syntactic parsing to improve event annotation matching

Title Leveraging syntactic parsing to improve event annotation matching
Authors Camiel Colruyt, Orph{'e}e De Clercq, V{'e}ronique Hoste
Abstract Detecting event mentions is the first step in event extraction from text and annotating them is a notoriously difficult task. Evaluating annotator consistency is crucial when building datasets for mention detection. When event mentions are allowed to cover many tokens, annotators may disagree on their span, which means that overlapping annotations may then refer to the same event or to different events. This paper explores different fuzzy-matching functions which aim to resolve this ambiguity. The functions extract the sets of syntactic heads present in the annotations, use the Dice coefficient to measure the similarity between sets and return a judgment based on a given threshold. The functions are tested against the judgment of a human evaluator and a comparison is made between sets of tokens and sets of syntactic heads. The best-performing function is a head-based function that is found to agree with the human evaluator in 89{%} of cases.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5903/
PDF https://www.aclweb.org/anthology/D19-5903
PWC https://paperswithcode.com/paper/leveraging-syntactic-parsing-to-improve-event
Repo
Framework

Automatic Detection of Translation Direction

Title Automatic Detection of Translation Direction
Authors Ilia Sominsky, Shuly Wintner
Abstract Parallel corpora are crucial resources for NLP applications, most notably for machine translation. The direction of the (human) translation of parallel corpora has been shown to have significant implications for the quality of statistical machine translation systems that are trained with such corpora. We describe a method for determining the direction of the (manual) translation of parallel corpora at the sentence-pair level. Using several linguistically-motivated features, coupled with a neural network model, we obtain high accuracy on several language pairs. Furthermore, we demonstrate that the accuracy is correlated with the (typological) distance between the two languages.
Tasks Machine Translation
Published 2019-09-01
URL https://www.aclweb.org/anthology/R19-1130/
PDF https://www.aclweb.org/anthology/R19-1130
PWC https://paperswithcode.com/paper/automatic-detection-of-translation-direction
Repo
Framework

HITS-SBD at the FinSBD Task: Machine Learning vs. Rule-based Sentence Boundary Detection

Title HITS-SBD at the FinSBD Task: Machine Learning vs. Rule-based Sentence Boundary Detection
Authors Mehwish Fatima, Mark-Christoph Mueller
Abstract
Tasks Boundary Detection
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5520/
PDF https://www.aclweb.org/anthology/W19-5520
PWC https://paperswithcode.com/paper/hits-sbd-at-the-finsbd-task-machine-learning
Repo
Framework

Differences between SMT and NMT Output - a Translators’ Point of View

Title Differences between SMT and NMT Output - a Translators’ Point of View
Authors Jonathan Mutal, Lise Volkart, Pierrette Bouillon, Sabrina Girletti, Paula Estrella
Abstract In this study, we compare the output quality of two MT systems, a statistical (SMT) and a neural (NMT) engine, customised for Swiss Post{'}s Language Service using the same training data. We focus on the point of view of professional translators and investigate how they perceive the differences between the MT output and a human reference (namely deletions, substitutions, insertions and word order). Our findings show that translators more frequently consider these differences to be errors in SMT than NMT, and that deletions are the most serious errors in both architectures. We also observe lower agreement on differences to be corrected in NMT than in SMT, suggesting that errors are easier to identify in SMT. These findings confirm the ability of NMT to produce correct paraphrases, which could also explain why BLEU is often considered as an inadequate metric to evaluate the performance of NMT systems.
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-8709/
PDF https://www.aclweb.org/anthology/W19-8709
PWC https://paperswithcode.com/paper/differences-between-smt-and-nmt-output-a
Repo
Framework

JRC TMA-CC: Slavic Named Entity Recognition and Linking. Participation in the BSNLP-2019 shared task

Title JRC TMA-CC: Slavic Named Entity Recognition and Linking. Participation in the BSNLP-2019 shared task
Authors Guillaume Jacquet, Jakub Piskorski, Hristo Tanev, Ralf Steinberger
Abstract We report on the participation of the JRC Text Mining and Analysis Competence Centre (TMA-CC) in the BSNLP-2019 Shared Task, which focuses on named-entity recognition, lemmatisation and cross-lingual linking. We propose a hybrid system combining a rule-based approach and light ML techniques. We use multilingual lexical resources such as JRC-NAMES and BABELNET together with a named entity guesser to recognise names. In a second step, we combine known names with wild cards to increase recognition recall by also capturing inflection variants. In a third step, we increase precision by filtering these name candidates with automatically learnt inflection patterns derived from name occurrences in large news article collections. Our major requirement is to achieve high precision. We achieved an average of 65{%} F-measure with 93{%} precision on the four languages.
Tasks Named Entity Recognition
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-3714/
PDF https://www.aclweb.org/anthology/W19-3714
PWC https://paperswithcode.com/paper/jrc-tma-cc-slavic-named-entity-recognition
Repo
Framework

Discourse Representation Parsing for Sentences and Documents

Title Discourse Representation Parsing for Sentences and Documents
Authors Jiangming Liu, Shay B. Cohen, Mirella Lapata
Abstract We introduce a novel semantic parsing task based on Discourse Representation Theory (DRT; Kamp and Reyle 1993). Our model operates over Discourse Representation Tree Structures which we formally define for sentences and documents. We present a general framework for parsing discourse structures of arbitrary length and granularity. We achieve this with a neural model equipped with a supervised hierarchical attention mechanism and a linguistically-motivated copy strategy. Experimental results on sentence- and document-level benchmarks show that our model outperforms competitive baselines by a wide margin.
Tasks Semantic Parsing
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1629/
PDF https://www.aclweb.org/anthology/P19-1629
PWC https://paperswithcode.com/paper/discourse-representation-parsing-for
Repo
Framework

SPI-GCN: A Simple Permutation-Invariant Graph Convolutional Network

Title SPI-GCN: A Simple Permutation-Invariant Graph Convolutional Network
Authors Asma Atamna, Nataliya Sokolovska, Jean-Claude Crivello
Abstract A wide range of machine learning problems involve handling graph-structured data. Existing machine learning approaches for graphs, however, often imply computing expensive graph similarity measures, preprocessing input graphs, or explicitly ordering graph nodes. In this work, we present a novel and simple convolutional neural network architecture for supervised learning on graphs that is provably invariant to node permutation. The proposed architecture operates directly on arbitrary graphs and performs no node sorting. It also uses a simple multi-layer perceptron for prediction as opposed to conventional convolution layers commonly used in other deep learning approaches for graphs. Despite its simplicity, our architecture is competitive with state-of-the-art graph kernels and existing graph neural networks on benchmark graph classification data sets. Our approach clearly outperforms other deep learning algorithms for graphs on multiple multiclass classification tasks. We also evaluate our approach on a real-world original application in materials science, on which we achieve extremely reasonable results.
Tasks Graph Classification, Graph Similarity
Published 2019-04-08
URL https://hal.archives-ouvertes.fr/hal-02093451/
PDF https://hal.archives-ouvertes.fr/hal-02093451/document
PWC https://paperswithcode.com/paper/spi-gcn-a-simple-permutation-invariant-graph
Repo
Framework

Neural Self-Training through Spaced Repetition

Title Neural Self-Training through Spaced Repetition
Authors Hadi Amiri
Abstract Self-training is a semi-supervised learning approach for utilizing unlabeled data to create better learners. The efficacy of self-training algorithms depends on their data sampling techniques. The majority of current sampling techniques are based on predetermined policies which may not effectively explore the data space or improve model generalizability. In this work, we tackle the above challenges by introducing a new data sampling technique based on spaced repetition that dynamically samples informative and diverse unlabeled instances with respect to individual learner and instance characteristics. The proposed model is specifically effective in the context of neural models which can suffer from overfitting and high-variance gradients when trained with small amount of labeled data. Our model outperforms current semi-supervised learning approaches developed for neural networks on publicly-available datasets.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/N19-1003/
PDF https://www.aclweb.org/anthology/N19-1003
PWC https://paperswithcode.com/paper/neural-self-training-through-spaced
Repo
Framework

FiST – towards a free Semantic Tagger of modern standard Finnish

Title FiST – towards a free Semantic Tagger of modern standard Finnish
Authors Kimmo Kettunen
Abstract
Tasks Morphological Analysis
Published 2019-01-01
URL https://www.aclweb.org/anthology/W19-0306/
PDF https://www.aclweb.org/anthology/W19-0306
PWC https://paperswithcode.com/paper/fist-extendash-towards-a-free-semantic-tagger
Repo
Framework

DEBUG: A Dense Bottom-Up Grounding Approach for Natural Language Video Localization

Title DEBUG: A Dense Bottom-Up Grounding Approach for Natural Language Video Localization
Authors Chujie Lu, Long Chen, Chilie Tan, Xiaolin Li, Jun Xiao
Abstract In this paper, we focus on natural language video localization: localizing (ie, grounding) a natural language description in a long and untrimmed video sequence. All currently published models for addressing this problem can be categorized into two types: (i) top-down approach: it does classification and regression for a set of pre-cut video segment candidates; (ii) bottom-up approach: it directly predicts probabilities for each video frame as the temporal boundaries (ie, start and end time point). However, both two approaches suffer several limitations: the former is computation-intensive for densely placed candidates, while the latter has trailed the performance of the top-down counterpart thus far. To this end, we propose a novel dense bottom-up framework: DEnse Bottom-Up Grounding (DEBUG). DEBUG regards all frames falling in the ground truth segment as foreground, and each foreground frame regresses the unique distances from its location to bi-directional ground truth boundaries. Extensive experiments on three challenging benchmarks (TACoS, Charades-STA, and ActivityNet Captions) show that DEBUG is able to match the speed of bottom-up models while surpassing the performance of the state-of-the-art top-down models.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1518/
PDF https://www.aclweb.org/anthology/D19-1518
PWC https://paperswithcode.com/paper/debug-a-dense-bottom-up-grounding-approach
Repo
Framework

Investigating CNNs’ Learning Representation under label noise

Title Investigating CNNs’ Learning Representation under label noise
Authors Ryuichiro Hataya, Hideki Nakayama
Abstract Deep convolutional neural networks (CNNs) are known to be robust against label noise on extensive datasets. However, at the same time, CNNs are capable of memorizing all labels even if they are random, which means they can memorize corrupted labels. Are CNNs robust or fragile to label noise? Much of researches focusing on such memorization uses class-independent label noise to simulate label corruption, but this setting is simple and unrealistic. In this paper, we investigate the behavior of CNNs under class-dependently simulated label noise, which is generated based on the conceptual distance between classes of a large dataset (i.e., ImageNet-1k). Contrary to previous knowledge, we reveal CNNs are more robust to such class-dependent label noise than class-independent label noise. We also demonstrate the networks under class-dependent noise situations learn similar representation to the no noise situation, compared to class-independent noise situations.
Tasks
Published 2019-05-01
URL https://openreview.net/forum?id=H1xmqiAqFm
PDF https://openreview.net/pdf?id=H1xmqiAqFm
PWC https://paperswithcode.com/paper/investigating-cnns-learning-representation
Repo
Framework

Towards discourse annotation and sentiment analysis of the Basque Opinion Corpus

Title Towards discourse annotation and sentiment analysis of the Basque Opinion Corpus
Authors Jon Alkorta, Koldo Gojenola, Mikel Iruskieta
Abstract Discourse information is crucial for a better understanding of the text structure and it is also necessary to describe which part of an opinionated text is more relevant or to decide how a text span can change the polarity (strengthen or weaken) of other span by means of coherence relations. This work presents the first results on the annotation of the Basque Opinion Corpus using Rhetorical Structure Theory (RST). Our evaluation results and analysis show us the main avenues to improve on a future annotation process. We have also extracted the subjectivity of several rhetorical relations and the results show the effect of sentiment words in relations and the influence of each relation in the semantic orientation value.
Tasks Sentiment Analysis
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-2718/
PDF https://www.aclweb.org/anthology/W19-2718
PWC https://paperswithcode.com/paper/towards-discourse-annotation-and-sentiment
Repo
Framework

What do phone embeddings learn about Phonology?

Title What do phone embeddings learn about Phonology?
Authors Sudheer Kolachina, Lilla Magyar
Abstract Recent work has looked at evaluation of phone embeddings using sound analogies and correlations between distinctive feature space and embedding space. It has not been clear what aspects of natural language phonology are learnt by neural network inspired distributed representational models such as word2vec. To study the kinds of phonological relationships learnt by phone embeddings, we present artificial phonology experiments that show that phone embeddings learn paradigmatic relationships such as phonemic and allophonic distribution quite well. They are also able to capture co-occurrence restrictions among vowels such as those observed in languages with vowel harmony. However, they are unable to learn co-occurrence restrictions among the class of consonants.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4219/
PDF https://www.aclweb.org/anthology/W19-4219
PWC https://paperswithcode.com/paper/what-do-phone-embeddings-learn-about
Repo
Framework

SenZi: A Sentiment Analysis Lexicon for the Latinised Arabic (Arabizi)

Title SenZi: A Sentiment Analysis Lexicon for the Latinised Arabic (Arabizi)
Authors Taha Tobaili, Fern, Miriam ez, Harith Alani, Sanaa Sharafeddine, Hazem Hajj, Goran Glava{\v{s}}
Abstract Arabizi is an informal written form of dialectal Arabic transcribed in Latin alphanumeric characters. It has a proven popularity on chat platforms and social media, yet it suffers from a severe lack of natural language processing (NLP) resources. As such, texts written in Arabizi are often disregarded in sentiment analysis tasks for Arabic. In this paper we describe the creation of a sentiment lexicon for Arabizi that was enriched with word embeddings. The result is a new Arabizi lexicon consisting of 11.3K positive and 13.3K negative words. We evaluated this lexicon by classifying the sentiment of Arabizi tweets achieving an F1-score of 0.72. We provide a detailed error analysis to present the challenges that impact the sentiment analysis of Arabizi.
Tasks Sentiment Analysis, Word Embeddings
Published 2019-09-01
URL https://www.aclweb.org/anthology/R19-1138/
PDF https://www.aclweb.org/anthology/R19-1138
PWC https://paperswithcode.com/paper/senzi-a-sentiment-analysis-lexicon-for-the
Repo
Framework

Extracting Common Inference Patterns from Semi-Structured Explanations

Title Extracting Common Inference Patterns from Semi-Structured Explanations
Authors Sebastian Thiem, Peter Jansen
Abstract Complex questions often require combining multiple facts to correctly answer, particularly when generating detailed explanations for why those answers are correct. Combining multiple facts to answer questions is often modeled as a {}multi-hop{''} graph traversal problem, where a given solver must find a series of interconnected facts in a knowledge graph that, taken together, answer the question and explain the reasoning behind that answer. Multi-hop inference currently suffers from semantic drift, or the tendency for chains of reasoning to {}drift{''}{'} to unrelated topics, and this semantic drift greatly limits the number of facts that can be combined in both free text or knowledge base inference. In this work we present our effort to mitigate semantic drift by extracting large high-confidence multi-hop inference patterns, generated by abstracting large-scale explanatory structure from a corpus of detailed explanations. We represent these inference patterns as sets of generalized constraints over sentences represented as rows in a knowledge base of semi-structured tables. We present a prototype tool for identifying common inference patterns from corpora of semi-structured explanations, and use it to successfully extract 67 inference patterns from a {``}matter{''} subset of standardized elementary science exam questions that span scientific and world knowledge. |
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-6006/
PDF https://www.aclweb.org/anthology/D19-6006
PWC https://paperswithcode.com/paper/extracting-common-inference-patterns-from
Repo
Framework
comments powered by Disqus