January 24, 2020

2272 words 11 mins read

Paper Group NANR 256

Paper Group NANR 256

A Pointer Network Architecture for Context-Dependent Semantic Parsing. Investigating Multilingual Abusive Language Detection: A Cautionary Tale. Neural Network to Identify Personal Health Experience Mention in Tweets Using BioBERT Embeddings. UHH-LT at SemEval-2019 Task 6: Supervised vs. Unsupervised Transfer Learning for Offensive Language Detecti …

A Pointer Network Architecture for Context-Dependent Semantic Parsing

Title A Pointer Network Architecture for Context-Dependent Semantic Parsing
Authors Xuanli He, Quan Tran, Gholamreza Haffari
Abstract
Tasks Semantic Parsing
Published 2019-04-01
URL https://www.aclweb.org/anthology/U19-1013/
PDF https://www.aclweb.org/anthology/U19-1013
PWC https://paperswithcode.com/paper/a-pointer-network-architecture-for-context
Repo
Framework

Investigating Multilingual Abusive Language Detection: A Cautionary Tale

Title Investigating Multilingual Abusive Language Detection: A Cautionary Tale
Authors Kenneth Steimel, Daniel Dakota, Yue Chen, S K{"u}bler, ra
Abstract Abusive language detection has received much attention in the last years, and recent approaches perform the task in a number of different languages. We investigate which factors have an effect on multilingual settings, focusing on the compatibility of data and annotations. In the current paper, we focus on English and German. Our findings show large differences in performance between the two languages. We find that the best performance is achieved by different classification algorithms. Sampling to address class imbalance issues is detrimental for German and beneficial for English. The only similarity that we find is that neither data set shows clear topics when we compare the results of topic modeling to the gold standard. Based on our findings, we can conclude that a multilingual optimization of classifiers is not possible even in settings where comparable data sets are used.
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/R19-1132/
PDF https://www.aclweb.org/anthology/R19-1132
PWC https://paperswithcode.com/paper/investigating-multilingual-abusive-language
Repo
Framework

Neural Network to Identify Personal Health Experience Mention in Tweets Using BioBERT Embeddings

Title Neural Network to Identify Personal Health Experience Mention in Tweets Using BioBERT Embeddings
Authors Shubham Gondane
Abstract This paper describes the system developed by team ASU-NLP for the Social Media Mining for Health Applications(SMM4H) shared task 4. We extract feature embeddings from the BioBERT (Lee et al., 2019) model which has been fine-tuned on the training dataset and use that as inputs to a dense fully connected neural network. We achieve above average scores among the participant systems with the overall F1-score, accuracy, precision, recall as 0.8036, 0.8456, 0.9783, 0.6818 respectively.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-3218/
PDF https://www.aclweb.org/anthology/W19-3218
PWC https://paperswithcode.com/paper/neural-network-to-identify-personal-health
Repo
Framework

UHH-LT at SemEval-2019 Task 6: Supervised vs. Unsupervised Transfer Learning for Offensive Language Detection

Title UHH-LT at SemEval-2019 Task 6: Supervised vs. Unsupervised Transfer Learning for Offensive Language Detection
Authors Gregor Wiedemann, Eugen Ruppert, Chris Biemann
Abstract We present a neural network based approach of transfer learning for offensive language detection. For our system, we compare two types of knowledge transfer: supervised and unsupervised pre-training. Supervised pre-training of our bidirectional GRU-3-CNN architecture is performed as multi-task learning of parallel training of five different tasks. The selected tasks are supervised classification problems from public NLP resources with some overlap to offensive language such as sentiment detection, emoji classification, and aggressive language classification. Unsupervised transfer learning is performed with a thematic clustering of 40M unlabeled tweets via LDA. Based on this dataset, pre-training is performed by predicting the main topic of a tweet. Results indicate that unsupervised transfer from large datasets performs slightly better than supervised training on small {`}near target category{'} datasets. In the SemEval Task, our system ranks 14 out of 103 participants. |
Tasks Multi-Task Learning, Transfer Learning
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2137/
PDF https://www.aclweb.org/anthology/S19-2137
PWC https://paperswithcode.com/paper/uhh-lt-at-semeval-2019-task-6-supervised-vs
Repo
Framework

Using NLG for speech synthesis of mathematical sentences

Title Using NLG for speech synthesis of mathematical sentences
Authors Aless Mazzei, ro, Michele Monticone, Cristian Bernareggi
Abstract People with sight impairments can access to a mathematical expression by using its LaTeX source. However, this mechanisms have several drawbacks: (1) it assumes the knowledge of the LaTeX, (2) it is slow, since LaTeX is verbose and (3) it is error-prone since LATEX is a typographical language. In this paper we study the design of a natural language generation system for producing a mathematical sentence, i.e. a natural language sentence expressing the semantics of a mathematical expression. Moreover, we describe the main results of a first human based evaluation experiment of the system for Italian language.
Tasks Speech Synthesis, Text Generation
Published 2019-10-01
URL https://www.aclweb.org/anthology/W19-8658/
PDF https://www.aclweb.org/anthology/W19-8658
PWC https://paperswithcode.com/paper/using-nlg-for-speech-synthesis-of
Repo
Framework

SC-LSTM: Learning Task-Specific Representations in Multi-Task Learning for Sequence Labeling

Title SC-LSTM: Learning Task-Specific Representations in Multi-Task Learning for Sequence Labeling
Authors Peng Lu, Ting Bai, Philippe Langlais
Abstract Multi-task learning (MTL) has been studied recently for sequence labeling. Typically, auxiliary tasks are selected specifically in order to improve the performance of a target task. Jointly learning multiple tasks in a way that benefit all of them simultaneously can increase the utility of MTL. In order to do so, we propose a new LSTM cell which contains both shared parameters that can learn from all tasks, and task-specific parameters that can learn task-specific information. We name it a Shared-Cell Long-Short Term Memory SC-LSTM. Experimental results on three sequence labeling benchmarks (named-entity recognition, text chunking, and part-of-speech tagging) demonstrate the effectiveness of our SC-LSTM cell.
Tasks Chunking, Multi-Task Learning, Named Entity Recognition, Part-Of-Speech Tagging
Published 2019-06-01
URL https://www.aclweb.org/anthology/N19-1249/
PDF https://www.aclweb.org/anthology/N19-1249
PWC https://paperswithcode.com/paper/sc-lstm-learning-task-specific
Repo
Framework

PharmaCoNER: Pharmacological Substances, Compounds and proteins Named Entity Recognition track

Title PharmaCoNER: Pharmacological Substances, Compounds and proteins Named Entity Recognition track
Authors Aitor Gonzalez-Agirre, Montserrat Marimon, Ander Intxaurrondo, Obdulia Rabal, Marta Villegas, Martin Krallinger
Abstract One of the biomedical entity types of relevance for medicine or biosciences are chemical compounds and drugs. The correct detection these entities is critical for other text mining applications building on them, such as adverse drug-reaction detection, medication-related fake news or drug-target extraction. Although a significant effort was made to detect mentions of drugs/chemicals in English texts, so far only very limited attempts were made to recognize them in medical documents in other languages. Taking into account the growing amount of medical publications and clinical records written in Spanish, we have organized the first shared task on detecting drug and chemical entities in Spanish medical documents. Additionally, we included a clinical concept-indexing sub-track asking teams to return SNOMED-CT identifiers related to drugs/chemicals for a collection of documents. For this task, named PharmaCoNER, we generated annotation guidelines together with a corpus of 1,000 manually annotated clinical case studies. A total of 22 teams participated in the sub-track 1, (77 system runs), and 7 teams in the sub-track 2 (19 system runs). Top scoring teams used sophisticated deep learning approaches yielding very competitive results with F-measures above 0.91. These results indicate that there is a real interest in promoting biomedical text mining efforts beyond English. We foresee that the PharmaCoNER annotation guidelines, corpus and participant systems will foster the development of new resources for clinical and biomedical text mining systems of Spanish medical data.
Tasks Named Entity Recognition
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5701/
PDF https://www.aclweb.org/anthology/D19-5701
PWC https://paperswithcode.com/paper/pharmaconer-pharmacological-substances
Repo
Framework

Sample Size in Arabic Authorship Verification

Title Sample Size in Arabic Authorship Verification
Authors Hossam Ahmed
Abstract
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-7412/
PDF https://www.aclweb.org/anthology/W19-7412
PWC https://paperswithcode.com/paper/sample-size-in-arabic-authorship-verification
Repo
Framework

LDA Topic Modeling for pram=aṇa Texts: A Case Study in Sanskrit NLP Corpus Building

Title LDA Topic Modeling for pram=aṇa Texts: A Case Study in Sanskrit NLP Corpus Building
Authors Tyler Neill
Abstract
Tasks
Published 2019-10-01
URL https://www.aclweb.org/anthology/W19-7505/
PDF https://www.aclweb.org/anthology/W19-7505
PWC https://paperswithcode.com/paper/lda-topic-modeling-for-pramaa1a-texts-a-case
Repo
Framework

Cross-Lingual Lemmatization and Morphology Tagging with Two-Stage Multilingual BERT Fine-Tuning

Title Cross-Lingual Lemmatization and Morphology Tagging with Two-Stage Multilingual BERT Fine-Tuning
Authors Dan Kondratyuk
Abstract We present our CHARLES-SAARLAND system for the SIGMORPHON 2019 Shared Task on Crosslinguality and Context in Morphology, in task 2, Morphological Analysis and Lemmatization in Context. We leverage the multilingual BERT model and apply several fine-tuning strategies introduced by UDify demonstrating exceptional evaluation performance on morpho-syntactic tasks. Our results show that fine-tuning multilingual BERT on the concatenation of all available treebanks allows the model to learn cross-lingual information that is able to boost lemmatization and morphology tagging accuracy over fine-tuning it purely monolingually. Unlike UDify, however, we show that when paired with additional character-level and word-level LSTM layers, a second stage of fine-tuning on each treebank individually can improve evaluation even further. Out of all submissions for this shared task, our system achieves the highest average accuracy and f1 score in morphology tagging and places second in average lemmatization accuracy.
Tasks Lemmatization, Morphological Analysis
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4203/
PDF https://www.aclweb.org/anthology/W19-4203
PWC https://paperswithcode.com/paper/cross-lingual-lemmatization-and-morphology
Repo
Framework

Investigating the Stability of Concrete Nouns in Word Embeddings

Title Investigating the Stability of Concrete Nouns in Word Embeddings
Authors B{'e}n{'e}dicte Pierrejean, Ludovic Tanguy
Abstract We know that word embeddings trained using neural-based methods (such as word2vec SGNS) are sensitive to stability problems and that across two models trained using the exact same set of parameters, the nearest neighbors of a word are likely to change. All words are not equally impacted by this internal instability and recent studies have investigated features influencing the stability of word embeddings. This stability can be seen as a clue for the reliability of the semantic representation of a word. In this work, we investigate the influence of the degree of concreteness of nouns on the stability of their semantic representation. We show that for English generic corpora, abstract words are more affected by stability problems than concrete words. We also found that to a certain extent, the difference between the degree of concreteness of a noun and its nearest neighbors can partly explain the stability or instability of its neighbors.
Tasks Word Embeddings
Published 2019-05-01
URL https://www.aclweb.org/anthology/W19-0510/
PDF https://www.aclweb.org/anthology/W19-0510
PWC https://paperswithcode.com/paper/investigating-the-stability-of-concrete-nouns
Repo
Framework

Convolutional Neural Networks for Financial Text Regression

Title Convolutional Neural Networks for Financial Text Regression
Authors Ne{\c{s}}at Dereli, Murat Saraclar
Abstract Forecasting financial volatility of a publicly-traded company from its annual reports has been previously defined as a text regression problem. Recent studies use a manually labeled lexicon to filter the annual reports by keeping sentiment words only. In order to remove the lexicon dependency without decreasing the performance, we replace bag-of-words model word features by word embedding vectors. Using word vectors increases the number of parameters. Considering the increase in number of parameters and excessive lengths of annual reports, a convolutional neural network model is proposed and transfer learning is applied. Experimental results show that the convolutional neural network model provides more accurate volatility predictions than lexicon based models.
Tasks Transfer Learning
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-2046/
PDF https://www.aclweb.org/anthology/P19-2046
PWC https://paperswithcode.com/paper/convolutional-neural-networks-for-financial
Repo
Framework

eventAI at SemEval-2019 Task 7: Rumor Detection on Social Media by Exploiting Content, User Credibility and Propagation Information

Title eventAI at SemEval-2019 Task 7: Rumor Detection on Social Media by Exploiting Content, User Credibility and Propagation Information
Authors Quanzhi Li, Qiong Zhang, Luo Si
Abstract This paper describes our system for SemEval 2019 RumorEval: Determining rumor veracity and support for rumors (SemEval 2019 Task 7). This track has two tasks: Task A is to determine a user{'}s stance towards the source rumor, and Task B is to detect the veracity of the rumor: true, false or unverified. For stance classification, a neural network model with language features is utilized. For rumor verification, our approach exploits information from different dimensions: rumor content, source credibility, user credibility, user stance, event propagation path, etc. We use an ensemble approach in both tasks, which includes neural network models as well as the traditional classification algorithms. Our system is ranked 1st place in the rumor verification task by both the macro F1 measure and the RMSE measure.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/S19-2148/
PDF https://www.aclweb.org/anthology/S19-2148
PWC https://paperswithcode.com/paper/eventai-at-semeval-2019-task-7-rumor
Repo
Framework

Variance Reduction in Bipartite Experiments through Correlation Clustering

Title Variance Reduction in Bipartite Experiments through Correlation Clustering
Authors Jean Pouget-Abadie, Kevin Aydin, Warren Schudy, Kay Brodersen, Vahab Mirrokni
Abstract Causal inference in randomized experiments typically assumes that the units of randomization and the units of analysis are one and the same. In some applications, however, these two roles are played by distinct entities linked by a bipartite graph. The key challenge in such bipartite settings is how to avoid interference bias, which would typically arise if we simply randomized the treatment at the level of analysis units. One effective way of minimizing interference bias in standard experiments is through cluster randomization, but this design has not been studied in the bipartite setting where conventional clustering schemes can lead to poorly powered experiments. This paper introduces a novel clustering objective and a corresponding algorithm that partitions a bipartite graph so as to maximize the statistical power of a bipartite experiment on that graph. Whereas previous work relied on balanced partitioning, our formulation suggests the use of a correlation clustering objective. We use a publicly-available graph of Amazon user-item reviews to validate our solution and illustrate how it substantially increases the statistical power in bipartite experiments.
Tasks Causal Inference
Published 2019-12-01
URL http://papers.nips.cc/paper/9486-variance-reduction-in-bipartite-experiments-through-correlation-clustering
PDF http://papers.nips.cc/paper/9486-variance-reduction-in-bipartite-experiments-through-correlation-clustering.pdf
PWC https://paperswithcode.com/paper/variance-reduction-in-bipartite-experiments
Repo
Framework

What Influences the Features of Post-editese? A Preliminary Study

Title What Influences the Features of Post-editese? A Preliminary Study
Authors Sheila Castilho, Nat{'a}lia Resende, Ruslan Mitkov
Abstract While a number of studies have shown evidence of translationese phenomena, that is, statistical differences between original texts and translated texts (Gellerstam, 1986), results of studies searching for translationese features in postedited texts (what has been called {''}posteditese{''} (Daems et al., 2017)) have presented mixed results. This paper reports a preliminary study aimed at identifying the presence of post-editese features in machine-translated post-edited texts and at understanding how they differ from translationese features. We test the influence of factors such as post-editing (PE) levels (full vs. light), translation proficiency (professionals vs. students) and text domain (news vs. literary). Results show evidence of post-editese features, especially in light PE texts and in certain domains.
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-8703/
PDF https://www.aclweb.org/anthology/W19-8703
PWC https://paperswithcode.com/paper/what-influences-the-features-of-post-editese
Repo
Framework
comments powered by Disqus