January 24, 2020

2575 words 13 mins read

Paper Group NANR 232

Paper Group NANR 232

Discriminating between Mandarin Chinese and Swiss-German varieties using adaptive language models. Detecting Target of Sarcasm using Ensemble Methods. Overview of the 6th Workshop on Asian Translation. A Stylometry Toolkit for Latin Literature. Learning the Dyck Language with Attention-based Seq2Seq Models. Learning With Unsure Data for Medical Ima …

Discriminating between Mandarin Chinese and Swiss-German varieties using adaptive language models

Title Discriminating between Mandarin Chinese and Swiss-German varieties using adaptive language models
Authors Tommi Jauhiainen, Krister Lind{'e}n, Heidi Jauhiainen
Abstract This paper describes the language identification systems used by the SUKI team in the Discriminating between the Mainland and Taiwan variation of Mandarin Chinese (DMT) and the German Dialect Identification (GDI) shared tasks which were held as part of the third VarDial Evaluation Campaign. The DMT shared task included two separate tracks, one for the simplified Chinese script and one for the traditional Chinese script. We submitted three runs on both tracks of the DMT task as well as on the GDI task. We won the traditional Chinese track using Naive Bayes with language model adaptation, came second on GDI with an adaptive version of the HeLI 2.0 method, and third on the simplified Chinese track using again the adaptive Naive Bayes.
Tasks Language Identification, Language Modelling
Published 2019-06-01
URL https://www.aclweb.org/anthology/W19-1419/
PDF https://www.aclweb.org/anthology/W19-1419
PWC https://paperswithcode.com/paper/discriminating-between-mandarin-chinese-and
Repo
Framework

Detecting Target of Sarcasm using Ensemble Methods

Title Detecting Target of Sarcasm using Ensemble Methods
Authors Pradeesh Parameswaran, Andrew Trotman, Veronica Liesaputra, David Eyers
Abstract We describe our methods in trying to detect the target of sarcasm as part of ALTA 2019 shared task. We use combination of ensemble of clas- sifiers and a rule-based system. Our team ob- tained a Dice-Sorensen Coefficient score of 0.37150, which placed 2nd in the public leader- board. Despite no team beating the baseline score for the private dataset, we present our findings and also some of the challenges and future improvements which can be used in or- der to tackle the problem.
Tasks
Published 2019-04-01
URL https://www.aclweb.org/anthology/U19-1027/
PDF https://www.aclweb.org/anthology/U19-1027
PWC https://paperswithcode.com/paper/detecting-target-of-sarcasm-using-ensemble
Repo
Framework

Overview of the 6th Workshop on Asian Translation

Title Overview of the 6th Workshop on Asian Translation
Authors Toshiaki Nakazawa, Nobushige Doi, Shohei Higashiyama, Chenchen Ding, Raj Dabre, Hideya Mino, Isao Goto, Win Pa Pa, Anoop Kunchukuttan, Shantipriya Parida, Ond{\v{r}}ej Bojar, Sadao Kurohashi
Abstract This paper presents the results of the shared tasks from the 6th workshop on Asian translation (WAT2019) including Ja↔En, Ja↔Zh scientific paper translation subtasks, Ja↔En, Ja↔Ko, Ja↔En patent translation subtasks, Hi↔En, My↔En, Km↔En, Ta↔En mixed domain subtasks and Ru↔Ja news commentary translation task. For the WAT2019, 25 teams participated in the shared tasks. We also received 10 research paper submissions out of which 61 were accepted. About 400 translation results were submitted to the automatic evaluation server, and selected submis- sions were manually evaluated.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-5201/
PDF https://www.aclweb.org/anthology/D19-5201
PWC https://paperswithcode.com/paper/overview-of-the-6th-workshop-on-asian
Repo
Framework

A Stylometry Toolkit for Latin Literature

Title A Stylometry Toolkit for Latin Literature
Authors Thomas J. Bolt, Jeffrey H. Flynt, Pramit Chaudhuri, Joseph P. Dexter
Abstract Computational stylometry has become an increasingly important aspect of literary criticism, but many humanists lack the technical expertise or language-specific NLP resources required to exploit computational methods. We demonstrate a stylometry toolkit for analysis of Latin literary texts, which is freely available at www.qcrit.org/stylometry. Our toolkit generates data for a diverse range of literary features and has an intuitive point-and-click interface. The features included have proven effective for multiple literary studies and are calculated using custom heuristics without the need for syntactic parsing. As such, the toolkit models one approach to the user-friendly generation of stylometric data, which could be extended to other premodern and non-English languages underserved by standard NLP resources.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-3035/
PDF https://www.aclweb.org/anthology/D19-3035
PWC https://paperswithcode.com/paper/a-stylometry-toolkit-for-latin-literature
Repo
Framework

Learning the Dyck Language with Attention-based Seq2Seq Models

Title Learning the Dyck Language with Attention-based Seq2Seq Models
Authors Xiang Yu, Ngoc Thang Vu, Jonas Kuhn
Abstract The generalized Dyck language has been used to analyze the ability of Recurrent Neural Networks (RNNs) to learn context-free grammars (CFGs). Recent studies draw conflicting conclusions on their performance, especially regarding the generalizability of the models with respect to the depth of recursion. In this paper, we revisit several common models and experimental settings, discuss the potential problems of the tasks and analyses. Furthermore, we explore the use of attention mechanisms within the seq2seq framework to learn the Dyck language, which could compensate for the limited encoding ability of RNNs. Our findings reveal that attention mechanisms still cannot truly generalize over the recursion depth, although they perform much better than other models on the closing bracket tagging task. Moreover, this also suggests that this commonly used task is not sufficient to test a model{'}s understanding of CFGs.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-4815/
PDF https://www.aclweb.org/anthology/W19-4815
PWC https://paperswithcode.com/paper/learning-the-dyck-language-with-attention
Repo
Framework

Learning With Unsure Data for Medical Image Diagnosis

Title Learning With Unsure Data for Medical Image Diagnosis
Authors Botong Wu, Xinwei Sun, Lingjing Hu, Yizhou Wang
Abstract In image-based disease prediction, it can be hard to give certain cases a deterministic “disease/normal” label due to lack of enough information, e.g., at its early stage. We call such cases “unsure” data. Labeling such data as unsure suggests follow-up examinations so as to avoid irreversible medical accident/loss in contrast to incautious prediction. This is a common practice in clinical diagnosis, however, mostly neglected by existing methods. Learning with unsure data also interweaves with two other practical issues: (i) data imbalance issue that may incur model-bias towards the majority class, and (ii) conservative/aggressive strategy consideration, i.e., the negative (normal) samples and positive (disease) samples should NOT be treated equally -- the former should be detected with a high precision (conservativeness) and the latter should be detected with a high recall (aggression) to avoid missing opportunity for treatment. Mixed with these issues, learning with unsure data becomes particularly challenging. In this paper, we raise “learning with unsure data” problem and formulate it as an ordinal regression and propose a unified end-to-end learning framework, which also considers the aforementioned two issues: (i) incorporate cost-sensitive parameters to alleviate the data imbalance problem, and (ii) execute the conservative and aggressive strategies by introducing two parameters in the training procedure. The benefits of learning with unsure data and validity of our models are demonstrated on the prediction of Alzheimer’s Disease and lung nodules.
Tasks Disease Prediction
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Wu_Learning_With_Unsure_Data_for_Medical_Image_Diagnosis_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Wu_Learning_With_Unsure_Data_for_Medical_Image_Diagnosis_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/learning-with-unsure-data-for-medical-image
Repo
Framework

TellMeWhy: Learning to Explain Corrective Feedback for Second Language Learners

Title TellMeWhy: Learning to Explain Corrective Feedback for Second Language Learners
Authors Yi-Huei Lai, Jason Chang
Abstract We present a writing prototype feedback system, TellMeWhy, to provide explanations of errors in submitted essays. In our approach, the sentence with corrections is analyzed to identify error types and problem words, aimed at customizing explanations based on the context of the error. The method involves learning the relation of errors and problem words, generating common feedback patterns, and extracting grammar patterns, collocations and example sentences. At run-time, a sentence with corrections is classified, and the problem word and template are identified to provide detailed explanations. Preliminary evaluation shows that the method has potential to improve existing commercial writing services.
Tasks
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-3040/
PDF https://www.aclweb.org/anthology/D19-3040
PWC https://paperswithcode.com/paper/tellmewhy-learning-to-explain-corrective
Repo
Framework

A Deep Learning Approach for Dynamic Survival Analysis with Competing Risks

Title A Deep Learning Approach for Dynamic Survival Analysis with Competing Risks
Authors Changhee Lee, Mihaela van der Schaar
Abstract Currently available survival analysis methods are limited in their ability to deal with complex, heterogeneous, and longitudinal data such as that available in primary care records, or in their ability to deal with multiple competing risks. This paper develops a novel deep learning architecture that flexibly incorporates the available longitudinal data comprising various repeated measurements (rather than only the last available measurements) in order to issue dynamically updated survival predictions for one or multiple competing risk(s). Unlike existing works in the survival analysis on the basis of longitudinal data, the proposed method learns the time-to-event distributions without specifying underlying stochastic assumptions of the longitudinal or the time-to-event processes. Thus, our method is able to learn associations between the longitudinal data and the various associated risks in a fully data-driven fashion. We demonstrate the power of our method by applying it to real-world longitudinal datasets and show a drastic improvement over state-of-the-art methods in discriminative performance. Furthermore, our analysis of the variable importance and dynamic survival predictions will yield a better understanding of the predicted risks which will result in more effective health care.
Tasks Survival Analysis
Published 2019-05-01
URL https://openreview.net/forum?id=rJG8asRqKX
PDF https://openreview.net/pdf?id=rJG8asRqKX
PWC https://paperswithcode.com/paper/a-deep-learning-approach-for-dynamic-survival
Repo
Framework

Extraction of Lactation Frames from Drug Labels and LactMed

Title Extraction of Lactation Frames from Drug Labels and LactMed
Authors Heath Goodrum, Meghana Gudala, Ankita Misra, Kirk Roberts
Abstract This paper describes a natural language processing (NLP) approach to extracting lactation-specific drug information from two sources: FDA-mandated drug labels and the NLM Drugs and Lactation Database (LactMed). A frame semantic approach is utilized, and the paper describes the selected frames, their annotation on a set of 900 sections from drug labels and LactMed articles, and the NLP system to extract such frame instances automatically. The ultimate goal of the project is to use such a system to identify discrepancies in lactation-related drug information between these resources.
Tasks
Published 2019-08-01
URL https://www.aclweb.org/anthology/W19-5020/
PDF https://www.aclweb.org/anthology/W19-5020
PWC https://paperswithcode.com/paper/extraction-of-lactation-frames-from-drug
Repo
Framework

LSTMEmbed: Learning Word and Sense Representations from a Large Semantically Annotated Corpus with Long Short-Term Memories

Title LSTMEmbed: Learning Word and Sense Representations from a Large Semantically Annotated Corpus with Long Short-Term Memories
Authors Ignacio Iacobacci, Roberto Navigli
Abstract While word embeddings are now a de facto standard representation of words in most NLP tasks, recently the attention has been shifting towards vector representations which capture the different meanings, i.e., senses, of words. In this paper we explore the capabilities of a bidirectional LSTM model to learn representations of word senses from semantically annotated corpora. We show that the utilization of an architecture that is aware of word order, like an LSTM, enables us to create better representations. We assess our proposed model on various standard benchmarks for evaluating semantic representations, reaching state-of-the-art performance on the SemEval-2014 word-to-sense similarity task. We release the code and the resulting word and sense embeddings at http://lcl.uniroma1.it/LSTMEmbed.
Tasks Word Embeddings
Published 2019-07-01
URL https://www.aclweb.org/anthology/P19-1165/
PDF https://www.aclweb.org/anthology/P19-1165
PWC https://paperswithcode.com/paper/lstmembed-learning-word-and-sense
Repo
Framework

A System for Diacritizing Four Varieties of Arabic

Title A System for Diacritizing Four Varieties of Arabic
Authors Hamdy Mubarak, Ahmed Abdelali, Kareem Darwish, Mohamed Eldesouki, Younes Samih, Hassan Sajjad
Abstract Short vowels, aka diacritics, are more often omitted when writing different varieties of Arabic including Modern Standard Arabic (MSA), Classical Arabic (CA), and Dialectal Arabic (DA). However, diacritics are required to properly pronounce words, which makes diacritic restoration (a.k.a. diacritization) essential for language learning and text-to-speech applications. In this paper, we present a system for diacritizing MSA, CA, and two varieties of DA, namely Moroccan and Tunisian. The system uses a character level sequence-to-sequence deep learning model that requires no feature engineering and beats all previous SOTA systems for all the Arabic varieties that we test on.
Tasks Feature Engineering
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-3037/
PDF https://www.aclweb.org/anthology/D19-3037
PWC https://paperswithcode.com/paper/a-system-for-diacritizing-four-varieties-of
Repo
Framework

Maximum Likelihood Estimation of Factored Regular Deterministic Stochastic Languages

Title Maximum Likelihood Estimation of Factored Regular Deterministic Stochastic Languages
Authors Chihiro Shibata, Jeffrey Heinz
Abstract
Tasks
Published 2019-07-01
URL https://www.aclweb.org/anthology/W19-5709/
PDF https://www.aclweb.org/anthology/W19-5709
PWC https://paperswithcode.com/paper/maximum-likelihood-estimation-of-factored
Repo
Framework

Hint-based Training for Non-Autoregressive Translation

Title Hint-based Training for Non-Autoregressive Translation
Authors Zhuohan Li, Di He, Fei Tian, Tao Qin, Liwei Wang, Tie-Yan Liu
Abstract Machine translation is an important real-world application, and neural network-based AutoRegressive Translation (ART) models have achieved very promising accuracy. Due to the unparallelizable nature of the autoregressive factorization, ART models have to generate tokens one by one during decoding and thus suffer from high inference latency. Recently, Non-AutoRegressive Translation (NART) models were proposed to reduce the inference time. However, they could only achieve inferior accuracy compared with ART models. To improve the accuracy of NART models, in this paper, we propose to leverage the hints from a well-trained ART model to train the NART model. We define two hints for the machine translation task: hints from hidden states and hints from word alignments, and use such hints to regularize the optimization of NART models. Experimental results show that the NART model trained with hints could achieve significantly better translation performance than previous NART models on several tasks. In particular, for the WMT14 En-De and De-En task, we obtain BLEU scores of 25.20 and 29.52 respectively, which largely outperforms the previous non-autoregressive baselines. It is even comparable to a strong LSTM-based ART model (24.60 on WMT14 En-De), but one order of magnitude faster in inference.
Tasks Machine Translation
Published 2019-05-01
URL https://openreview.net/forum?id=r1gGpjActQ
PDF https://openreview.net/pdf?id=r1gGpjActQ
PWC https://paperswithcode.com/paper/hint-based-training-for-non-autoregressive
Repo
Framework

Deep Reinforcement Active Learning for Human-in-the-Loop Person Re-Identification

Title Deep Reinforcement Active Learning for Human-in-the-Loop Person Re-Identification
Authors Zimo Liu, Jingya Wang, Shaogang Gong, Huchuan Lu, Dacheng Tao
Abstract Most existing person re-identification(Re-ID) approaches achieve superior results based on the assumption that a large amount of pre-labelled data is usually available and can be put into training phrase all at once. However, this assumption is not applicable to most real-world deployment of the Re-ID task. In this work, we propose an alternative reinforcement learning based human-in-the-loop model which releases the restriction of pre-labelling and keeps model upgrading with progressively collected data. The goal is to minimize human annotation efforts while maximizing Re-ID performance. It works in an iteratively updating framework by refining the RL policy and CNN parameters alternately. In particular, we formulate a Deep Reinforcement Active Learning (DRAL) method to guide an agent (a model in a reinforcement learning process) in selecting training samples on-the-fly by a human user/annotator. The reinforcement learning reward is the uncertainty value of each human selected sample. A binary feedback (positive or negative) labelled by the human annotator is used to select the samples of which are used to fine-tune a pre-trained CNN Re-ID model. Extensive experiments demonstrate the superiority of our DRAL method for deep reinforcement learning based human-in-the-loop person Re-ID when compared to existing unsupervised and transfer learning models as well as active learning models.
Tasks Active Learning, Person Re-Identification, Transfer Learning
Published 2019-10-01
URL http://openaccess.thecvf.com/content_ICCV_2019/html/Liu_Deep_Reinforcement_Active_Learning_for_Human-in-the-Loop_Person_Re-Identification_ICCV_2019_paper.html
PDF http://openaccess.thecvf.com/content_ICCV_2019/papers/Liu_Deep_Reinforcement_Active_Learning_for_Human-in-the-Loop_Person_Re-Identification_ICCV_2019_paper.pdf
PWC https://paperswithcode.com/paper/deep-reinforcement-active-learning-for-human
Repo
Framework

Social IQa: Commonsense Reasoning about Social Interactions

Title Social IQa: Commonsense Reasoning about Social Interactions
Authors Maarten Sap, Hannah Rashkin, Derek Chen, Ronan Le Bras, Yejin Choi
Abstract We introduce Social IQa, the first large-scale benchmark for commonsense reasoning about social situations. Social IQa contains 38,000 multiple choice questions for probing emotional and social intelligence in a variety of everyday situations (e.g., Q: {}Jordan wanted to tell Tracy a secret, so Jordan leaned towards Tracy. Why did Jordan do this?{''} A: {}Make sure no one else could hear{''}). Through crowdsourcing, we collect commonsense questions along with correct and incorrect answers about social interactions, using a new framework that mitigates stylistic artifacts in incorrect answers by asking workers to provide the right answer to a different but related question. Empirical results show that our benchmark is challenging for existing question-answering models based on pretrained language models, compared to human performance ({\textgreater}20{%} gap). Notably, we further establish Social IQa as a resource for transfer learning of commonsense knowledge, achieving state-of-the-art performance on multiple commonsense reasoning tasks (Winograd Schemas, COPA).
Tasks Question Answering, Transfer Learning
Published 2019-11-01
URL https://www.aclweb.org/anthology/D19-1454/
PDF https://www.aclweb.org/anthology/D19-1454
PWC https://paperswithcode.com/paper/social-iqa-commonsense-reasoning-about-social
Repo
Framework
comments powered by Disqus