January 24, 2020

2519 words 12 mins read

Paper Group NANR 138

Paper Group NANR 138

A Deep Learning-Based System for PharmaCoNER. Term Based Semantic Clusters for Very Short Text Classification. A Robust Self-Learning Framework for Cross-Lingual Text Classification. Similar Minds Post Alike: Assessment of Suicide Risk Using a Hybrid Model. Controlling Japanese Honorifics in English-to-Japanese Neural Machine Translation. Surface R …

A Deep Learning-Based System for PharmaCoNER

Title A Deep Learning-Based System for PharmaCoNER
Authors Ying Xiong, Yedan Shen, Yuanhang Huang, Shuai Chen, Buzhou Tang, Xiaolong Wang, Qingcai Chen, Jun Yan, Yi Zhou
Abstract The Biological Text Mining Unit at BSC and CNIO organized the first shared task on chemical {&} drug mention recognition from Spanish medical texts called PharmaCoNER (Pharmacological Substances, Compounds and proteins and Named Entity Recognition track) in 2019, which includes two tracks: one for NER offset and entity classification (track 1) and the other one for concept indexing (track 2). We developed a pipeline system based on deep learning methods for this shared task, specifically, a subsystem based on BERT (Bidirectional Encoder Representations from Transformers) for NER offset and entity classification and a subsystem based on Bpool (Bi-LSTM with max/mean pooling) for concept indexing. Evaluation conducted on the shared task data showed that our system achieves a micro-average F1-score of 0.9105 on track 1 and a micro-average F1-score of 0.8391 on track 2.
Tasks Named Entity Recognition
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5706/
PDF https://www.aclweb.org/anthology/D19-5706
PWC https://paperswithcode.com/paper/a-deep-learning-based-system-for-pharmaconer
Repo
Framework

Term Based Semantic Clusters for Very Short Text Classification

Title Term Based Semantic Clusters for Very Short Text Classification
Authors Jasper Paalman, Shantanu Mullick, Kalliopi Zervanou, Yingqian Zhang
Abstract Very short texts, such as tweets and invoices, present challenges in classification. Although term occurrences are strong indicators of content, in very short texts, the sparsity of these texts makes it difficult to capture important semantic relationships. A solution calls for a method that not only considers term occurrence, but also handles sparseness well. In this work, we introduce such an approach, the Term Based Semantic Clusters (TBSeC) that employs terms to create distinctive semantic concept clusters. These clusters are ranked using a semantic similarity function which in turn defines a semantic feature space that can be used for text classification. Our method is evaluated in an invoice classification task. Compared to well-known content representation methods the proposed method performs competitively.
Tasks Semantic Similarity, Semantic Textual Similarity, Text Classification
Published 2019-09-01
URL https://www.aclweb.org/anthology/R19-1102/
PDF https://www.aclweb.org/anthology/R19-1102
PWC https://paperswithcode.com/paper/term-based-semantic-clusters-for-very-short
Repo
Framework

A Robust Self-Learning Framework for Cross-Lingual Text Classification

Title A Robust Self-Learning Framework for Cross-Lingual Text Classification
Authors Xin Dong, Gerard de Melo
Abstract Based on massive amounts of data, recent pretrained contextual representation models have made significant strides in advancing a number of different English NLP tasks. However, for other languages, relevant training data may be lacking, while state-of-the-art deep learning methods are known to be data-hungry. In this paper, we present an elegantly simple robust self-learning framework to include unlabeled non-English samples in the fine-tuning process of pretrained multilingual representation models. We leverage a multilingual model{'}s own predictions on unlabeled non-English data in order to obtain additional information that can be used during further fine-tuning. Compared with original multilingual models and other cross-lingual classification models, we observe significant gains in effectiveness on document and sentiment classification for a range of diverse languages.
Tasks Sentiment Analysis, Text Classification
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1658/
PDF https://www.aclweb.org/anthology/D19-1658
PWC https://paperswithcode.com/paper/a-robust-self-learning-framework-for-cross
Repo
Framework

Similar Minds Post Alike: Assessment of Suicide Risk Using a Hybrid Model

Title Similar Minds Post Alike: Assessment of Suicide Risk Using a Hybrid Model
Authors Lushi Chen, Abeer Aldayel, Nikolay Bogoychev, Tao Gong
Abstract This paper describes our system submission for the CLPsych 2019 shared task B on suicide risk assessment. We approached the problem with three separate models: a behaviour model; a language model and a hybrid model. For the behavioral model approach, we model each user{'}s behaviour and thoughts with four groups of features: posting behaviour, sentiment, motivation, and content of the user{'}s posting. We use these features as an input in a support vector machine (SVM). For the language model approach, we trained a language model for each risk level using all the posts from the users as the training corpora. Then, we computed the perplexity of each user{'}s posts to determine how likely his/her posts were to belong to each risk level. Finally, we built a hybrid model that combines both the language model and the behavioral model, which demonstrates the best performance in detecting the suicide risk level.
Tasks Language Modelling
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-3018/
PDF https://www.aclweb.org/anthology/W19-3018
PWC https://paperswithcode.com/paper/similar-minds-post-alike-assessment-of
Repo
Framework

Controlling Japanese Honorifics in English-to-Japanese Neural Machine Translation

Title Controlling Japanese Honorifics in English-to-Japanese Neural Machine Translation
Authors Weston Feely, Eva Hasler, Adri{`a} de Gispert
Abstract In the Japanese language different levels of honorific speech are used to convey respect, deference, humility, formality and social distance. In this paper, we present a method for controlling the level of formality of Japanese output in English-to-Japanese neural machine translation (NMT). By using heuristics to identify honorific verb forms, we classify Japanese sentences as being one of three levels of informal, polite, or formal speech in parallel text. The English source side is marked with a feature that identifies the level of honorific speech present in the Japanese target side. We use this parallel text to train an English-Japanese NMT model capable of producing Japanese translations in different honorific speech styles for the same English input sentence.
Tasks Machine Translation
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5203/
PDF https://www.aclweb.org/anthology/D19-5203
PWC https://paperswithcode.com/paper/controlling-japanese-honorifics-in-english-to
Repo
Framework

Surface Reconstruction From Normals: A Robust DGP-Based Discontinuity Preservation Approach

Title Surface Reconstruction From Normals: A Robust DGP-Based Discontinuity Preservation Approach
Authors Wuyuan Xie, Miaohui Wang, Mingqiang Wei, Jianmin Jiang, Jing Qin
Abstract In 3D surface reconstruction from normals, discontinuity preservation is an important but challenging task. However, existing studies fail to address the discontinuous normal maps by enforcing the surface integrability in the continuous domain. This paper introduces a robust approach to preserve the surface discontinuity in the discrete geometry way. Firstly, we design two representative normal incompatibility features and propose an efficient discontinuity detection scheme to determine the splitting pattern for a discrete mesh. Secondly, we model the discontinuity preservation problem as a light-weight energy optimization framework by jointly considering the discontinuity detection and the overall reconstruction error. Lastly, we further shrink the feasible solution space to reduce the complexity based on the prior knowledge. Experiments show that the proposed method achieves the best performance on an extensive 3D dataset compared with the state-of-the-arts in terms of mean angular error and computational complexity.
Tasks
Published 2019-06-01
URL http://openaccess.thecvf.com/content_CVPR_2019/html/Xie_Surface_Reconstruction_From_Normals_A_Robust_DGP-Based_Discontinuity_Preservation_Approach_CVPR_2019_paper.html
PDF http://openaccess.thecvf.com/content_CVPR_2019/papers/Xie_Surface_Reconstruction_From_Normals_A_Robust_DGP-Based_Discontinuity_Preservation_Approach_CVPR_2019_paper.pdf
PWC https://paperswithcode.com/paper/surface-reconstruction-from-normals-a-robust
Repo
Framework

A New Perspective on Pool-Based Active Classification and False-Discovery Control

Title A New Perspective on Pool-Based Active Classification and False-Discovery Control
Authors Lalit Jain, Kevin G. Jamieson
Abstract In many scientific settings there is a need for adaptive experimental design to guide the process of identifying regions of the search space that contain as many true positives as possible subject to a low rate of false discoveries (i.e. false alarms). Such regions of the search space could differ drastically from a predicted set that minimizes 0/1 error and accurate identification could require very different sampling strategies. Like active learning for binary classification, this experimental design cannot be optimally chosen a priori, but rather the data must be taken sequentially and adaptively in a closed loop. However, unlike classification with 0/1 error, collecting data adaptively to find a set with high true positive rate and low false discovery rate (FDR) is not as well understood. In this paper, we provide the first provably sample efficient adaptive algorithm for this problem. Along the way, we highlight connections between classification, combinatorial bandits, and FDR control making contributions to each.
Tasks Active Learning
Published 2019-12-01
URL http://papers.nips.cc/paper/9549-a-new-perspective-on-pool-based-active-classification-and-false-discovery-control
PDF http://papers.nips.cc/paper/9549-a-new-perspective-on-pool-based-active-classification-and-false-discovery-control.pdf
PWC https://paperswithcode.com/paper/a-new-perspective-on-pool-based-active
Repo
Framework

Johns Hopkins University Submission for WMT News Translation Task

Title Johns Hopkins University Submission for WMT News Translation Task
Authors Kelly Marchisio, Yash Kumar Lal, Philipp Koehn
Abstract We describe the work of Johns Hopkins University for the shared task of news translation organized by the Fourth Conference on Machine Translation (2019). We submitted systems for both directions of the English-German language pair. The systems combine multiple techniques {–} sampling, filtering, iterative backtranslation, and continued training {–} previously used to improve performance of neural machine translation models. At submission time, we achieve a BLEU score of 38.1 for De-En and 42.5 for En-De translation directions on newstest2019. Post-submission, the score is 38.4 for De-En and 42.8 for En-De. Various experiments conducted in the process are also described.
Tasks Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5329/
PDF https://www.aclweb.org/anthology/W19-5329
PWC https://paperswithcode.com/paper/johns-hopkins-university-submission-for-wmt
Repo
Framework

Normalization of Kazakh Texts

Title Normalization of Kazakh Texts
Authors Assina Abdussaitova, Alina Amangeldiyeva
Abstract Kazakh language, like other agglutinative languages, has specific difficulties on both recognition of wrong words and generation the corrections for misspelt words. The main goal of this work is to develop a better algorithm for the normalization of Kazakh texts based on traditional and Machine Learning methods, as well as the new approach which is also considered in this paper. The procedure of election among methods of normalization has been conducted in a manner of comparative analysis. The results of the comparative analysis turned up successful and are shown in details.
Tasks
Published 2019-09-01
URL https://www.aclweb.org/anthology/R19-2001/
PDF https://www.aclweb.org/anthology/R19-2001
PWC https://paperswithcode.com/paper/normalization-of-kazakh-texts
Repo
Framework

PROMT Systems for WMT 2019 Shared Translation Task

Title PROMT Systems for WMT 2019 Shared Translation Task
Authors Alex Molchanov, er
Abstract This paper describes the PROMT submissions for the WMT 2019 Shared News Translation Task. This year we participated in two language pairs and in three directions: English-Russian, English-German and German-English. All our submissions are Marian-based neural systems. We use significantly more data compared to the last year. We also present our improved data filtering pipeline.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5331/
PDF https://www.aclweb.org/anthology/W19-5331
PWC https://paperswithcode.com/paper/promt-systems-for-wmt-2019-shared-translation
Repo
Framework

JU-Saarland Submission to the WMT2019 English–Gujarati Translation Shared Task

Title JU-Saarland Submission to the WMT2019 English–Gujarati Translation Shared Task
Authors Riktim Mondal, Shankha Raj Nayek, Aditya Chowdhury, Santanu Pal, Sudip Kumar Naskar, Josef van Genabith
Abstract In this paper we describe our joint submission (JU-Saarland) from Jadavpur University and Saarland University in the WMT 2019 news translation shared task for English{–}Gujarati language pair within the translation task sub-track. Our baseline and primary submissions are built using Recurrent neural network (RNN) based neural machine translation (NMT) system which follows attention mechanism. Given the fact that the two languages belong to different language families and there is not enough parallel data for this language pair, building a high quality NMT system for this language pair is a difficult task. We produced synthetic data through back-translation from available monolingual data. We report the translation quality of our English{–}Gujarati and Gujarati{–}English NMT systems trained at word, byte-pair and character encoding levels where RNN at word level is considered as the baseline and used for comparison purpose. Our English{–}Gujarati system ranked in the second position in the shared task.
Tasks Machine Translation
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5332/
PDF https://www.aclweb.org/anthology/W19-5332
PWC https://paperswithcode.com/paper/ju-saarland-submission-to-the-wmt2019-english
Repo
Framework

Neural Relation Extraction for Knowledge Base Enrichment

Title Neural Relation Extraction for Knowledge Base Enrichment
Authors Bayu Distiawan Trisedya, Gerhard Weikum, Jianzhong Qi, Rui Zhang
Abstract We study relation extraction for knowledge base (KB) enrichment. Specifically, we aim to extract entities and their relationships from sentences in the form of triples and map the elements of the extracted triples to an existing KB in an end-to-end manner. Previous studies focus on the extraction itself and rely on Named Entity Disambiguation (NED) to map triples into the KB space. This way, NED errors may cause extraction errors that affect the overall precision and recall.To address this problem, we propose an end-to-end relation extraction model for KB enrichment based on a neural encoder-decoder model. We collect high-quality training data by distant supervision with co-reference resolution and paraphrase detection. We propose an n-gram based attention model that captures multi-word entity names in a sentence. Our model employs jointly learned word and entity embeddings to support named entity disambiguation. Finally, our model uses a modified beam search and a triple classifier to help generate high-quality triples. Our model outperforms state-of-the-art baselines by 15.51{%} and 8.38{%} in terms of F1 score on two real-world datasets.
Tasks Entity Disambiguation, Entity Embeddings, Relation Extraction
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1023/
PDF https://www.aclweb.org/anthology/P19-1023
PWC https://paperswithcode.com/paper/neural-relation-extraction-for-knowledge-base
Repo
Framework

Language Modeling with Graph Temporal Convolutional Networks

Title Language Modeling with Graph Temporal Convolutional Networks
Authors Hongyin Luo, Yichen Li, Jie Fu, James Glass
Abstract Recently, there have been some attempts to use non-recurrent neural models for language modeling. However, a noticeable performance gap still remains. We propose a non-recurrent neural language model, dubbed graph temporal convolutional network (GTCN), that relies on graph neural network blocks and convolution operations. While the standard recurrent neural network language models encode sentences sequentially without modeling higher-level structural information, our model regards sentences as graphs and processes input words within a message propagation framework, aiming to learn better syntactic information by inferring skip-word connections. Specifically, the graph network blocks operate in parallel and learn the underlying graph structures in sentences without any additional annotation pertaining to structure knowledge. Experiments demonstrate that the model without recurrence can achieve comparable perplexity results in language modeling tasks and successfully learn syntactic information.
Tasks Language Modelling
Published 2019-05-01
URL https://openreview.net/forum?id=HJlYzhR9tm
PDF https://openreview.net/pdf?id=HJlYzhR9tm
PWC https://paperswithcode.com/paper/language-modeling-with-graph-temporal
Repo
Framework

BOUN-ISIK Participation: An Unsupervised Approach for the Named Entity Normalization and Relation Extraction of Bacteria Biotopes

Title BOUN-ISIK Participation: An Unsupervised Approach for the Named Entity Normalization and Relation Extraction of Bacteria Biotopes
Authors {.I}lknur Karadeniz, {"O}mer Faruk Tuna, Arzucan {"O}zg{"u}r
Abstract This paper presents our participation to the Bacteria Biotope Task of the BioNLP Shared Task 2019. Our participation includes two systems for the two subtasks of the Bacteria Biotope Task: the normalization of entities (BB-norm) and the identification of the relations between the entities given a biomedical text (BB-rel). For the normalization of entities, we utilized word embeddings and syntactic re-ranking. For the relation extraction task, pre-defined rules are used. Although both approaches are unsupervised, in the sense that they do not need any labeled data, they achieved promising results. Especially, for the BB-norm task, the results have shown that the proposed method performs as good as deep learning based methods, which require labeled data.
Tasks Relation Extraction, Word Embeddings
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5722/
PDF https://www.aclweb.org/anthology/D19-5722
PWC https://paperswithcode.com/paper/boun-isik-participation-an-unsupervised
Repo
Framework

Toward a deep dialectological representation of Indo-Aryan

Title Toward a deep dialectological representation of Indo-Aryan
Authors Chundra Cathcart
Abstract This paper presents a new approach to disentangling inter-dialectal and intra-dialectal relationships within one such group, the Indo-Aryan subgroup of Indo-European. We draw upon admixture models and deep generative models to tease apart historic language contact and language-specific behavior in the overall patterns of sound change displayed by Indo-Aryan languages. We show that a {}deep{''} model of Indo-Aryan dialectology sheds some light on questions regarding inter-relationships among the Indo-Aryan languages, and performs better than a {}shallow{''} model in terms of certain qualities of the posterior distribution (e.g., entropy of posterior distributions), and outline future pathways for model development.
Tasks
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-1411/
PDF https://www.aclweb.org/anthology/W19-1411
PWC https://paperswithcode.com/paper/toward-a-deep-dialectological-representation
Repo
Framework
comments powered by Disqus