January 24, 2020

2599 words 13 mins read

Paper Group NANR 129

Paper Group NANR 129

Idiap NMT System for WAT 2019 Multimodal Translation Task. Similarity Based Auxiliary Classifier for Named Entity Recognition. NLP Automation to Read Radiological Reports to Detect the Stage of Cancer Among Lung Cancer Patients. TDDiscourse: A Dataset for Discourse-Level Temporal Ordering of Events. Sentence-Level Propaganda Detection in News Artic …

Idiap NMT System for WAT 2019 Multimodal Translation Task

Title Idiap NMT System for WAT 2019 Multimodal Translation Task
Authors Shantipriya Parida, Ond{\v{r}}ej Bojar, Petr Motlicek
Abstract This paper describes the Idiap submission to WAT 2019 for the English-Hindi Multi-Modal Translation Task. We have used the state-of-the-art Transformer model and utilized the IITB English-Hindi parallel corpus as an additional data source. Among the different tracks of the multi-modal task, we have participated in the {}Text-Only{''} track for the evaluation and challenge test sets. Our submission tops in its track among the competitors in terms of both automatic and manual evaluation. Based on automatic scores, our text-only submission also outperforms systems that consider visual information in the {}multi-modal translation{''} task.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5223/
PDF https://www.aclweb.org/anthology/D19-5223
PWC https://paperswithcode.com/paper/idiap-nmt-system-for-wat-2019-multimodal
Repo
Framework

Similarity Based Auxiliary Classifier for Named Entity Recognition

Title Similarity Based Auxiliary Classifier for Named Entity Recognition
Authors Shiyuan Xiao, Yuanxin Ouyang, Wenge Rong, Jianxin Yang, Zhang Xiong
Abstract The segmentation problem is one of the fundamental challenges associated with name entity recognition (NER) tasks that aim to reduce the boundary error when detecting a sequence of entity words. A considerable number of advanced approaches have been proposed and most of them exhibit performance deterioration when entities become longer. Inspired by previous work in which a multi-task strategy is used to solve segmentation problems, we design a similarity based auxiliary classifier (SAC), which can distinguish entity words from non-entity words. Unlike conventional classifiers, SAC uses vectors to indicate tags. Therefore, SAC can calculate the similarities between words and tags, and then compute a weighted sum of the tag vectors, which can be considered a useful feature for NER tasks. Empirical results are used to verify the rationality of the SAC structure and demonstrate the SAC model{'}s potential in performance improvement against our baseline approaches.
Tasks Named Entity Recognition
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1105/
PDF https://www.aclweb.org/anthology/D19-1105
PWC https://paperswithcode.com/paper/similarity-based-auxiliary-classifier-for
Repo
Framework

NLP Automation to Read Radiological Reports to Detect the Stage of Cancer Among Lung Cancer Patients

Title NLP Automation to Read Radiological Reports to Detect the Stage of Cancer Among Lung Cancer Patients
Authors Khushbu Gupta, Ratchainant Thammasudjarit, Ammarin Thakkinstian
Abstract A common challenge in the healthcare industry today is physicians have access to massive amounts of healthcare data but have little time and no appropriate tools. For instance, the risk prediction model generated by logistic regression could predict the probability of diseases occurrence and thus prioritizing patients{'} waiting list for further investigations. However, many medical reports available in current clinical practice system are not yet ready for analysis using either statistics or machine learning as they are in unstructured text format. The complexity of medical information makes the annotation or validation of data very challenging and thus acts as a bottleneck to apply machine learning techniques in medical data. This study is therefore conducted to create such annotations automatically where the computer can read radiological reports for oncologists and mark the staging of lung cancer. This staging information is obtained using the rule-based method implemented using the standards of Tumor Node Metastasis (TNM) staging along with deep learning technology called Long Short Term Memory (LSTM) to extract clinical information from the Computed Tomography (CT) text report. The empirical experiment shows promising results being the accuracy of up to 85{%}.
Tasks Computed Tomography (CT)
Published 2019-08-01
URL https://www.aclweb.org/anthology/papers/W/W19/W19-3643/
PDF https://www.aclweb.org/anthology/W19-3643
PWC https://paperswithcode.com/paper/nlp-automation-to-read-radiological-reports
Repo
Framework

TDDiscourse: A Dataset for Discourse-Level Temporal Ordering of Events

Title TDDiscourse: A Dataset for Discourse-Level Temporal Ordering of Events
Authors Aakanksha Naik, Luke Breitfeller, Carolyn Rose
Abstract Prior work on temporal relation classification has focused extensively on event pairs in the same or adjacent sentences (local), paying scant attention to discourse-level (global) pairs. This restricts the ability of systems to learn temporal links between global pairs, since reliance on local syntactic features suffices to achieve reasonable performance on existing datasets. However, systems should be capable of incorporating cues from document-level structure to assign temporal relations. In this work, we take a first step towards discourse-level temporal ordering by creating TDDiscourse, the first dataset focusing specifically on temporal links between event pairs which are more than one sentence apart. We create TDDiscourse by augmenting TimeBank-Dense, a corpus of English news articles, manually annotating global pairs that cannot be inferred automatically from existing annotations. Our annotations double the number of temporal links in TimeBank-Dense, while possessing several desirable properties such as focusing on long-distance pairs and not being automatically inferable. We adapt and benchmark the performance of three state-of-the-art models on TDDiscourse and observe that existing systems indeed find discourse-level temporal ordering harder.
Tasks Relation Classification
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-5929/
PDF https://www.aclweb.org/anthology/W19-5929
PWC https://paperswithcode.com/paper/tddiscourse-a-dataset-for-discourse-level
Repo
Framework

Sentence-Level Propaganda Detection in News Articles with Transfer Learning and BERT-BiLSTM-Capsule Model

Title Sentence-Level Propaganda Detection in News Articles with Transfer Learning and BERT-BiLSTM-Capsule Model
Authors George-Alex Vlad, ru, Mircea-Adrian Tanase, Cristian Onose, Dumitru-Clementin Cercel
Abstract In recent years, the need for communication increased in online social media. Propaganda is a mechanism which was used throughout history to influence public opinion and it is gaining a new dimension with the rising interest of online social media. This paper presents our submission to NLP4IF-2019 Shared Task SLC: Sentence-level Propaganda Detection in news articles. The challenge of this task is to build a robust binary classifier able to provide corresponding propaganda labels, propaganda or non-propaganda. Our model relies on a unified neural network, which consists of several deep leaning modules, namely BERT, BiLSTM and Capsule, to solve the sentencelevel propaganda classification problem. In addition, we take a pre-training approach on a somewhat similar task (i.e., emotion classification) improving results against the cold-start model. Among the 26 participant teams in the NLP4IF-2019 Task SLC, our solution ranked 12th with an F1-score 0.5868 on the official test data. Our proposed solution indicates promising results since our system significantly exceeds the baseline approach of the organizers by 0.1521 and is slightly lower than the winning system by 0.0454.
Tasks Emotion Classification, Transfer Learning
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5022/
PDF https://www.aclweb.org/anthology/D19-5022
PWC https://paperswithcode.com/paper/sentence-level-propaganda-detection-in-news
Repo
Framework

Sentiment Aware Neural Machine Translation

Title Sentiment Aware Neural Machine Translation
Authors Chenglei Si, Kui Wu, Ai Ti Aw, Min-Yen Kan
Abstract Sentiment ambiguous lexicons refer to words where their polarity depends strongly on con- text. As such, when the context is absent, their translations or their embedded sentence ends up (incorrectly) being dependent on the training data. While neural machine translation (NMT) has achieved great progress in recent years, most systems aim to produce one single correct translation for a given source sentence. We investigate the translation variation in two sentiment scenarios. We perform experiments to study the preservation of sentiment during translation with three different methods that we propose. We conducted tests with both sentiment and non-sentiment bearing contexts to examine the effectiveness of our methods. We show that NMT can generate both positive- and negative-valent translations of a source sentence, based on a given input sentiment label. Empirical evaluations show that our valence-sensitive embedding (VSE) method significantly outperforms a sequence-to-sequence (seq2seq) baseline, both in terms of BLEU score and ambiguous word translation accuracy in test, given non-sentiment bearing contexts.
Tasks Machine Translation
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5227/
PDF https://www.aclweb.org/anthology/D19-5227
PWC https://paperswithcode.com/paper/sentiment-aware-neural-machine-translation
Repo
Framework

Comparative Analysis of Errors in MT Output and Computer-assisted Translation: Effect of the Human Factor

Title Comparative Analysis of Errors in MT Output and Computer-assisted Translation: Effect of the Human Factor
Authors Irina Ovchinnikova, Daria Morozova
Abstract
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-6713/
PDF https://www.aclweb.org/anthology/W19-6713
PWC https://paperswithcode.com/paper/comparative-analysis-of-errors-in-mt-output
Repo
Framework

Deep neural model with enhanced embeddings for pharmaceutical and chemical entities recognition in Spanish clinical text

Title Deep neural model with enhanced embeddings for pharmaceutical and chemical entities recognition in Spanish clinical text
Authors Renzo Rivera, Paloma Mart{'\i}nez
Abstract In this work, we introduce a Deep Learning architecture for pharmaceutical and chemical Named Entity Recognition in Spanish clinical cases texts. We propose a hybrid model approach based on two Bidirectional Long Short-Term Memory (Bi-LSTM) network and Conditional Random Field (CRF) network using character, word, concept and sense embeddings to deal with the extraction of semantic, syntactic and morphological features. The approach was evaluated on the PharmaCoNER Corpus obtaining an F-measure of 85.24{%} for subtask 1 and 49.36{%} for subtask2. These results prove that deep learning methods with specific domain embedding representations can outperform the state-of-the-art approaches.
Tasks Named Entity Recognition
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5707/
PDF https://www.aclweb.org/anthology/D19-5707
PWC https://paperswithcode.com/paper/deep-neural-model-with-enhanced-embeddings
Repo
Framework

A Neural Pipeline Approach for the PharmaCoNER Shared Task using Contextual Exhaustive Models

Title A Neural Pipeline Approach for the PharmaCoNER Shared Task using Contextual Exhaustive Models
Authors Mohammad Golam Sohrab, Minh Thang Pham, Makoto Miwa, Hiroya Takamura
Abstract We present a neural pipeline approach that performs named entity recognition (NER) and concept indexing (CI), which links them to concept unique identifiers (CUIs) in a knowledge base, for the PharmaCoNER shared task on pharmaceutical drugs and chemical entities. We proposed a neural NER model that captures the surrounding semantic information of a given sequence by capturing the forward- and backward-context of bidirectional LSTM (Bi-LSTM) output of a target span using contextual span representation-based exhaustive approach. The NER model enumerates all possible spans as potential entity mentions and classify them into entity types or no entity with deep neural networks. For representing span, we compare several different neural network architectures and their ensembling for the NER model. We then perform dictionary matching for CI and, if there is no matching, we further compute similarity scores between a mention and CUIs using entity embeddings to assign the CUI with the highest score to the mention. We evaluate our approach on the two sub-tasks in the shared task. Among the five submitted runs, the best run for each sub-task achieved the F-score of 86.76{%} on Sub-task 1 (NER) and the F-score of 79.97{%} (strict) on Sub-task 2 (CI).
Tasks Entity Embeddings, Named Entity Recognition
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5708/
PDF https://www.aclweb.org/anthology/D19-5708
PWC https://paperswithcode.com/paper/a-neural-pipeline-approach-for-the
Repo
Framework

Context-Aware Feature and Label Fusion for Facial Action Unit Intensity Estimation With Partially Labeled Data

Title Context-Aware Feature and Label Fusion for Facial Action Unit Intensity Estimation With Partially Labeled Data
Authors Yong Zhang, Haiyong Jiang, Baoyuan Wu, Yanbo Fan, Qiang Ji
Abstract Facial action unit (AU) intensity estimation is a fundamental task for facial behaviour analysis. Most previous methods use a whole face image as input for intensity prediction. Considering that AUs are defined according to their corresponding local appearance, a few patch-based methods utilize image features of local patches. However, fusion of local features is always performed via straightforward feature concatenation or summation. Besides, these methods require fully annotated databases for model learning, which is expensive to acquire. In this paper, we propose a novel weakly supervised patch-based deep model on basis of two types of attention mechanisms for joint intensity estimation of multiple AUs. The model consists of a feature fusion module and a label fusion module. And we augment attention mechanisms of these two modules with a learnable task-related context, as one patch may play different roles in analyzing different AUs and each AU has its own temporal evolution rule. The context-aware feature fusion module is used to capture spatial relationships among local patches while the context-aware label fusion module is used to capture the temporal dynamics of AUs. The latter enables the model to be trained on a partially annotated database. Experimental evaluations on two benchmark expression databases demonstrate the superior performance of the proposed method.
Tasks
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Zhang_Context-Aware_Feature_and_Label_Fusion_for_Facial_Action_Unit_Intensity_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Zhang_Context-Aware_Feature_and_Label_Fusion_for_Facial_Action_Unit_Intensity_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/context-aware-feature-and-label-fusion-for
Repo
Framework

Transfer Learning in Biomedical Named Entity Recognition: An Evaluation of BERT in the PharmaCoNER task

Title Transfer Learning in Biomedical Named Entity Recognition: An Evaluation of BERT in the PharmaCoNER task
Authors Cong Sun, Zhihao Yang
Abstract To date, a large amount of biomedical content has been published in non-English texts, especially for clinical documents. Therefore, it is of considerable significance to conduct Natural Language Processing (NLP) research in non-English literature. PharmaCoNER is the first Named Entity Recognition (NER) task to recognize chemical and protein entities from Spanish biomedical texts. Since there have been abundant resources in the NLP field, how to exploit these existing resources to a new task to obtain competitive performance is a meaningful study. Inspired by the success of transfer learning with language models, we introduce the BERT benchmark to facilitate the research of PharmaCoNER task. In this paper, we evaluate two baselines based on Multilingual BERT and BioBERT on the PharmaCoNER corpus. Experimental results show that transferring the knowledge learned from source large-scale datasets to the target domain offers an effective solution for the PharmaCoNER task.
Tasks Named Entity Recognition, Transfer Learning
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5715/
PDF https://www.aclweb.org/anthology/D19-5715
PWC https://paperswithcode.com/paper/transfer-learning-in-biomedical-named-entity
Repo
Framework

Easy First Relation Extraction with Information Redundancy

Title Easy First Relation Extraction with Information Redundancy
Authors Shuai Ma, Gang Wang, Yansong Feng, Jinpeng Huai
Abstract Many existing relation extraction (RE) models make decisions globally using integer linear programming (ILP). However, it is nontrivial to make use of integer linear programming as a blackbox solver for RE. Its cost of time and memory may become unacceptable with the increase of data scale, and redundant information needs to be encoded cautiously for ILP. In this paper, we propose an easy first approach for relation extraction with information redundancies, embedded in the results produced by local sentence level extractors, during which conflict decisions are resolved with domain and uniqueness constraints. Information redundancies are leveraged to support both easy first collective inference for easy decisions in the first stage and ILP for hard decisions in a subsequent stage. Experimental study shows that our approach improves the efficiency and accuracy of RE, and outperforms both ILP and neural network-based methods.
Tasks Relation Extraction
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1398/
PDF https://www.aclweb.org/anthology/D19-1398
PWC https://paperswithcode.com/paper/easy-first-relation-extraction-with
Repo
Framework

Learning Word Embeddings without Context Vectors

Title Learning Word Embeddings without Context Vectors
Authors Alexey Zobnin, Evgenia Elistratova
Abstract Most word embedding algorithms such as word2vec or fastText construct two sort of vectors: for words and for contexts. Naive use of vectors of only one sort leads to poor results. We suggest using indefinite inner product in skip-gram negative sampling algorithm. This allows us to use only one sort of vectors without loss of quality. Our {``}context-free{''} cf algorithm performs on par with SGNS on word similarity datasets |
Tasks Learning Word Embeddings, Word Embeddings
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4329/
PDF https://www.aclweb.org/anthology/W19-4329
PWC https://paperswithcode.com/paper/learning-word-embeddings-without-context
Repo
Framework

ParaDis and D'emonette: From Theory to Resources for Derivational Paradigms

Title ParaDis and D'emonette: From Theory to Resources for Derivational Paradigms
Authors Fiammetta Namer, Nabil Hathout
Abstract
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/W19-8502/
PDF https://www.aclweb.org/anthology/W19-8502
PWC https://paperswithcode.com/paper/paradis-and-demonette-from-theory-to
Repo
Framework

Personality Traits Recognition in Literary Texts

Title Personality Traits Recognition in Literary Texts
Authors Daniele Pizzolli, Carlo Strapparava
Abstract Interesting stories often are built around interesting characters. Finding and detailing what makes an interesting character is a real challenge, but certainly a significant cue is the character personality traits. Our exploratory work tests the adaptability of the current personality traits theories to literal characters, focusing on the analysis of utterances in theatre scripts. And, at the opposite, we try to find significant traits for interesting characters. The preliminary results demonstrate that our approach is reasonable. Using machine learning for gaining insight into the personality traits of fictional characters can make sense.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-3411/
PDF https://www.aclweb.org/anthology/W19-3411
PWC https://paperswithcode.com/paper/personality-traits-recognition-in-literary
Repo
Framework
comments powered by Disqus